|
{ |
|
"paper_id": "2021", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T14:31:20.679159Z" |
|
}, |
|
"title": "Message from the Local Organisers", |
|
"authors": [ |
|
{ |
|
"first": "Program", |
|
"middle": [], |
|
"last": "Committee", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Mark", |
|
"middle": [], |
|
"last": "Fishel", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Simon", |
|
"middle": [], |
|
"last": "Dobnik", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Johannes", |
|
"middle": [], |
|
"last": "Bjerva", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Hans", |
|
"middle": [], |
|
"last": "Moen", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Lucia", |
|
"middle": [], |
|
"last": "Specia", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "NoDaLiDa 2021 has received financial support from our generous sponsors, which we would also like to thank here: Lingsoft, Tilde, Mideind and Grammatek. Above all, their support made it possible for us to offer this NoDaLiDa free of charge. I hope that this will open the conference up to an even larger audience of NLP researchers in Northern Europe.", |
|
"pdf_parse": { |
|
"paper_id": "2021", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "NoDaLiDa 2021 has received financial support from our generous sponsors, which we would also like to thank here: Lingsoft, Tilde, Mideind and Grammatek. Above all, their support made it possible for us to offer this NoDaLiDa free of charge. I hope that this will open the conference up to an even larger audience of NLP researchers in Northern Europe.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "Welcome to the 23rd Nordic Conference on Computational Linguistics (NoDaLiDa 2021)! I am a great fan of the NoDaLiDa conference, as a friendly, medium sized conference that offers the opportunity for scientific and social interaction with colleagues from neighbouring countries. When I agreed to serve as the general chair for this years NoDaLiDa I was still relatively optimistic that we could all meet in beautiful Reykjavik, Iceland to enjoy two days of scientific talks, posters and socialising in early June. Unfortunately that turned out to not be possible due to the COVID-19 pandemic. Instead, we are for the first time offering NoDaLiDa as a fully virtual event, free of charge. Even so, I am confident that this years conference will offer the same high-quality program as in previous years and hopefully it can also constitute a meeting place, albeit a digital one, for Northern European NLP researchers in these unusual times.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Message from the General Chair", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "As in previous editions, the conference features three different types of papers (long, short and demo papers). We received 91 legal submissions, which represents an increase compared to the previous edition of the conference. In total, we accepted 54 papers, which will be presented as 30 oral presentations, 22 posters and 2 demos at the conference. Each paper was reviewed by three experts. We are extremely grateful to the Programme Committee members for their detailed and helpful reviews. Overall, there are 8 oral sessions with talks and two poster sessions organised into themes over the two days, as well as two exciting keynote talks. I would further like to thank our two great keynote speakers for sharing their work with us: Lucia Specia from Imperial College London will talk about \"Disagreement in human evaluation: blame the task not the annotators\". Adina Williams from Facebook AI Research (FAIR) will talk about \"For Matters Word Order Little MLM\". Two exciting talks that complement each other well! As in previous years, the conference will be preceded by three workshops: Translatology in the Digital Age, NLP for Computer-Assisted Language Learning and Sustainable language representations. I want to thank the workshop organisers for complementing the main program and offering opportunities for in-depth scientific interaction on these diverse and exciting topics.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Message from the General Chair", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "I would like to thank the entire group of people that made NoDaLiDa 2021 possible. First of all, I would like to thank Beata Megyesi for inviting me to take up this exciting (and at times daunting) role and all her valuable input regarding NEALT and previous editions of NoDaLiDa. I am further indebted to Barbara Plank for her encouragement, for the sharing all the great resources from the last NoDaLiDa and willingly answering questions on all aspects of the conference organisation. I want to thank the program chair committee Jurgita Kapo\u010di\u016bt\u0117-Dzikien\u0117, Mark Fishel, J\u00f3n Gudnason, Barbara Plank, Yves Scherrer and Sara Stymne, for working hard on putting the program together. I am particularly grateful to Jurgita Kapo\u010di\u016bt\u0117-Dzikien\u0117, J\u00f3n Gudnason, Yves Scherrer and Sara Stymne for their great effort in leading the reviewing process and shepherding papers from submission to a final decision. I could not have done this without you! Special thanks go to the workshop chairs, Hans Moen and Ildik\u00f3 Pil\u00e1n, who have done an invaluable job with leading the workshop selection and organisation. A big thanks also to Johannes Bjerva for his work as social media chair and Simon Dobnik for leading the publication efforts that led to this volume, as well as the coordination of the workshop proceedings. Thank you! Finally, my ultimate thanks goes to the local organisation committee and team. ", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Message from the General Chair", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Lucia Specia: Disagreement in human evaluation: blame the task not the annotators. It is well known that human evaluators are prone to disagreement and that this is a problem for reliability and reproducibility of evaluation experiments. The reasons for disagreement can fall into two broad categories: (1) human evaluator, including under-trained, under-incentivised, lacking expertise, or illintended individuals, e.g., cheaters; and (2) task, including ill-definition, poor guidelines, sub-optimal setup, or inherent complexity or subjectivity. While in an ideal evaluation experiment many of these elements will be controlled for, in this talk I will argue that task complexity and subjectivity are much harder issues and that in some cases agreement cannot and should not be expected. I will cover several evaluation experiments on tasks with variable degrees of complexity and subjectivity, discuss their levels of disagreement along with other issues. I hope this will lead to an open discussion on possible strategies and directions to address this problem.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Invited Talks", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Adina Williams: For Matters Word Order Little MLM. One possible explanation for the impressive performance of masked language models (MLMs) is that they can learn to represent the syntactic structures prevalent in classical NLP pipelines. Were this correct, we would expect that fine-tuning such models on tasks requiring syntactic structure would lead them to be sensitive to word order at inference time. To address this question, we permute example word order at several steps in the pipeline-during fine-tuning, evaluation, and/or pre-training-and measure the results. We find that permuting word order during fine-tuning has remarkably little effect on downstream performance for several purportedly syntax sensitive NLU tasks (including NLI). Next, we pre-train MLMs on examples with randomly shuffled word order, and find that these models still achieve high accuracy (even after unpermuted fine-tuning) on many downstream tasks-including tasks specifically designed to be challenging for models that ignore word order. Our results show that the success of MLM pre-training is largely due to distributional information not any knowledge of word order per se, and underscores the importance of curating challenging evaluation datasets that require deeper syntactic knowledge.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Invited Talks", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [], |
|
"bib_entries": {}, |
|
"ref_entries": { |
|
"TABREF0": { |
|
"html": null, |
|
"type_str": "table", |
|
"text": "Thank you, Hr\u00e1fn L\u00f3ftsson, Anton Karl Ingason and Stein\u00fe\u00f3r Steingr\u00edmsson. They are the ones who did all the heavy lifting in the switch to a virtual event and did a truly amazing job!", |
|
"content": "<table><tr><td>Elena Volodina, University of Gothenburg</td></tr><tr><td>J\u00fcrgen Wedekind, University of Copenhagen</td></tr><tr><td>Anssi Yli-Jyr\u00e4, University of Helsinki</td></tr><tr><td>Marcos Zampieri, University of Wolverhampton</td></tr><tr><td>Niklas Zechner, University of Gothenburg</td></tr><tr><td>Heike Zinsmeister, University of Hamburg</td></tr><tr><td>x</td></tr></table>", |
|
"num": null |
|
} |
|
} |
|
} |
|
} |