{ "paper_id": "2021", "header": { "generated_with": "S2ORC 1.0.0", "date_generated": "2023-01-19T13:14:13.575011Z" }, "title": "NUS-IDS at CASE 2021 Task 1: Improving Multilingual Event Sentence Coreference Identification With Linguistic Information", "authors": [ { "first": "Fiona", "middle": [ "Anting" ], "last": "Tan", "suffix": "", "affiliation": { "laboratory": "", "institution": "National University of Singapore", "location": { "country": "Singapore" } }, "email": "tan.f@u.nus.edu" }, { "first": "Das", "middle": [], "last": "Gollapalli", "suffix": "", "affiliation": { "laboratory": "", "institution": "National University of Singapore", "location": { "country": "Singapore" } }, "email": "" }, { "first": "See-Kiong", "middle": [], "last": "Ng", "suffix": "", "affiliation": { "laboratory": "", "institution": "National University of Singapore", "location": { "country": "Singapore" } }, "email": "seekiong@nus.edu.sg" } ], "year": "", "venue": null, "identifiers": {}, "abstract": "Event Sentence Coreference Identification (ESCI) aims to cluster event sentences that refer to the same event together for information extraction. We describe our ESCI solution developed for the ACL-CASE 2021 shared tasks on the detection and classification of sociopolitical and crisis event information in a multilingual setting. For a given article, our proposed pipeline comprises of an accurate sentence pair classifier that identifies coreferent sentence pairs and subsequently uses these predicted probabilities to cluster sentences into groups. Sentence pair representations are constructed from fine-tuned BERT embeddings plus POS embeddings fed through a BiLSTM model, and combined with linguistic-based lexical and semantic similarities between sentences. Our best models ranked 2 nd , 1 st and 2 nd and obtained CoNLL F 1 scores of 81.20%, 93.03%, 83.15% for the English, Portuguese and Spanish test sets respectively in the ACL-CASE 2021 competition.", "pdf_parse": { "paper_id": "2021", "_pdf_hash": "", "abstract": [ { "text": "Event Sentence Coreference Identification (ESCI) aims to cluster event sentences that refer to the same event together for information extraction. We describe our ESCI solution developed for the ACL-CASE 2021 shared tasks on the detection and classification of sociopolitical and crisis event information in a multilingual setting. For a given article, our proposed pipeline comprises of an accurate sentence pair classifier that identifies coreferent sentence pairs and subsequently uses these predicted probabilities to cluster sentences into groups. Sentence pair representations are constructed from fine-tuned BERT embeddings plus POS embeddings fed through a BiLSTM model, and combined with linguistic-based lexical and semantic similarities between sentences. Our best models ranked 2 nd , 1 st and 2 nd and obtained CoNLL F 1 scores of 81.20%, 93.03%, 83.15% for the English, Portuguese and Spanish test sets respectively in the ACL-CASE 2021 competition.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Abstract", "sec_num": null } ], "body_text": [ { "text": "The ability to automatically extract sentences that refer to the same event from any given document is useful for downstream information extraction tasks like event extraction and summarization, timeline extraction or cause and effect extraction (\u00d6rs et al., 2020) . Event Sentence Coreference Identification (ESCI) aims to cluster sentences with event mentions such that each cluster comprises of sentences that refer to the same specific event.", "cite_spans": [ { "start": 246, "end": 264, "text": "(\u00d6rs et al., 2020)", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "We address ESCI for news articles referring to socio-political and crisis event information in a multilingual setting, introduced as one of the ACL-CASE 2021's shared tasks (H\u00fcrriyetoglu et al., 2021) . Given that news articles comprise of multiple events spread across a few sentences, and the syntax referring to the same event differs in different contexts, ESCI for news articles is a challenging NLP problem (H\u00fcrriyetoglu et al., 2020) . Furthermore, considering the availability of news in various languages, ESCI techniques that are applicable beyond English and robust across different languages are desirable.", "cite_spans": [ { "start": 173, "end": 200, "text": "(H\u00fcrriyetoglu et al., 2021)", "ref_id": "BIBREF5" }, { "start": 413, "end": 440, "text": "(H\u00fcrriyetoglu et al., 2020)", "ref_id": "BIBREF7" } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "The ESCI task is illustrated using an example article shown in Figure 1 . As shown in this figure, ESCI involves the identification of the event clusters (e.g. {[2], [1,7]} in the figure) based on the content of the individual sentences.", "cite_spans": [], "ref_spans": [ { "start": 63, "end": 71, "text": "Figure 1", "ref_id": "FIGREF0" } ], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "We propose a two-step solution for the ESCI task. In Step-1, we obtain sentence pair embeddings by fine-tuning Bidirectional Encoder Representations from Transformers (BERT) (Devlin et al., 2019) embeddings combined with parts-of-speech (POS) embeddings that are fed through a bi-directional long short-term memory (BiLSTM) model. Next, these sentence pair embeddings are combined with novel features based on lexical and semantic similarities to train a classifier that predicts if the sentence pair is coreferent. Step-2 involves the clustering of sentences using sentence pair probabilities predicted from Step-1. We apply the clustering algorithm from\u00d6rs et al. (2020) to obtain a variable number of clusters for each article.", "cite_spans": [ { "start": 174, "end": 195, "text": "(Devlin et al., 2019)", "ref_id": "BIBREF4" } ], "ref_spans": [], "eq_spans": [], "section": "Contributions:", "sec_num": null }, { "text": "We illustrate the effectiveness of our proposed solution via detailed validation experiments on the training datasets from ACL-CASE 2021. We show that our features are effective on documents from all three languages studied in the competition, viz, English, Portuguese, and Spanish. Indeed, on the ACL-CASE 2021 Shared Task 1 Subtask 3, our best-performing models ranked 2 nd , 1 st and 2 nd and obtained CoNLL F 1 scores of 81.20%, 93.03%, 83.15% for the English, Portuguese and Spanish test sets respectively.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Contributions:", "sec_num": null }, { "text": "Organization: In the next section, we present closely related work on ESCI. Subsequently, Section 3 introduces our features and classification model while Section 4 discusses our dataset, experimental setup, results, and findings. In Section 5, we conclude the paper with some future directions.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Contributions:", "sec_num": null }, { "text": "Most end-to-end event coreference systems approach the task in a two-stage manner: (1) To detect the mention or event of interest, and (2) To resolve the given mentions or events and cluster if coreferent (Zhang et al., 2018) . In our work, we focus only on latter task of coreference resolution and have direct access to identified event sentences.", "cite_spans": [ { "start": 205, "end": 225, "text": "(Zhang et al., 2018)", "ref_id": "BIBREF15" } ], "ref_spans": [], "eq_spans": [], "section": "Related Work", "sec_num": "2" }, { "text": "Early works of ESCI adopted linguistic (Bejan and Harabagiu, 2010) or template-based features (Choubey and Huang, 2017). Subsequently, neural network methods to encode textual events and contexts became increasingly popular (Krause et al., 2016) . The combination of the two methods have also proved to be effective in recent works (Zeng et al., 2020; Barhom et al., 2019) .", "cite_spans": [ { "start": 39, "end": 66, "text": "(Bejan and Harabagiu, 2010)", "ref_id": "BIBREF1" }, { "start": 224, "end": 245, "text": "(Krause et al., 2016)", "ref_id": "BIBREF8" }, { "start": 332, "end": 351, "text": "(Zeng et al., 2020;", "ref_id": "BIBREF14" }, { "start": 352, "end": 372, "text": "Barhom et al., 2019)", "ref_id": "BIBREF0" } ], "ref_spans": [], "eq_spans": [], "section": "Related Work", "sec_num": "2" }, { "text": "In the previous run of ESCI by the same organ-isers (H\u00fcrriyetoglu et al., 2019 (H\u00fcrriyetoglu et al., , 2020 , the bestperforming team (\u00d6rs et al., 2020) deconstructed the task into two steps: (1) To predict if a sentence pair is coreferent or not, and (2) Use the predictions as scores for clustering. This approach is common amongst other event coreference resolution methods too (Barhom et al., 2019) . We employ this general approach and focus on enriching the feature space using linguistic-based similarity measures along with richer text embeddings based on BERT with POS embeddings.", "cite_spans": [ { "start": 52, "end": 78, "text": "(H\u00fcrriyetoglu et al., 2019", "ref_id": "BIBREF6" }, { "start": 79, "end": 107, "text": "(H\u00fcrriyetoglu et al., , 2020", "ref_id": "BIBREF7" }, { "start": 381, "end": 402, "text": "(Barhom et al., 2019)", "ref_id": "BIBREF0" } ], "ref_spans": [], "eq_spans": [], "section": "Related Work", "sec_num": "2" }, { "text": "Figure 2 summarizes our proposed pipeline. In this section, we describe our approach in detail. 1 Let A be an article containing m sentences with event mentions {s 1 , s 2 , ..., s m }. To produce a list of c clusters that group these m sentences, we adopt the approach of\u00d6rs et al. (2020) and first extract sentence pairs from A. Next, binary classification is performed to identify if a given pair is coreferent. Let (h, t) represent a sentence pair, with h and t referring to the lower and higher sentence numbers in A, respectively. The features employed for training a binary classifier that identifies coreferent sentences are described next.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Our Approach", "sec_num": "3" }, { "text": "BERT Embeddings: We utilize BERT, the bidirectional encoder transformer architecture (Devlin et al., 2019) , to obtain sentence pair representations for our task. The models were pretrained with masked language modeling (MLM) and next sentence prediction (NSP) objectives. Our sentence pair input is encoded in this order: the special starting token \"[CLS]\", the head sentence, the separator \"[SEP]\" token, the tail sentence, another separator token, and padding up to a fixed maximum length.", "cite_spans": [ { "start": 85, "end": 106, "text": "(Devlin et al., 2019)", "ref_id": "BIBREF4" } ], "ref_spans": [], "eq_spans": [], "section": "Features for Sentence Pair Classification", "sec_num": "3.1" }, { "text": "The encoded inputs, alongside attention mask and token type indexes, are fed into the BERT model. BERT acts as an encoder by producing sentence pair representations, which is later passed on to the BiLSTM model along with other features to train our classifier. As BERT is exposed to label information in the downstream layers, we are able to obtain fined-tuned representations for our task.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Features for Sentence Pair Classification", "sec_num": "3.1" }, { "text": "POS Embeddings: For each sentence, we obtain parts-of-speech (POS) tags for each word token to represent grammatical structure. To align with tokens of BERT embeddings, we similarly concatenate a starting token, POS tags of the head sentence plus a separator token, POS tags of the tail sentence plus a separator token, followed by padding. These POS tags are subsequently encoded as one-hot vectors and combined with the BERT embeddings per word before feeding them through a BiLSTM.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Features for Sentence Pair Classification", "sec_num": "3.1" }, { "text": "Lexical and Semantic Similarities: Event mentions in a sentence often correspond to specific POS and Named-Entity (NE) tags. Thus, similarity values capturing the overlap of these token types between the two sentences are indicative of whether they are coreferent. We incorporated lexical similarity based on surface-form overlap of POS and NE tags of sentences and semantic similarity based on sentence embeddings and overlap of the dependence trees of the two sentences. We represent the counts of verb, nouns, and entities occurring in both head and tail sentences using two similarity functions: raw counts and Jaccard overlap. These six features are referred to as \"Basic Similarities\" in our experiments.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Features for Sentence Pair Classification", "sec_num": "3.1" }, { "text": "For an \"extended\" set of similarities, we also computed the cosine similarity based on words of the sentences after stopword removal, and normalized dot product of vectors corresponding to words with POS tags pertaining to nouns, adjectives, verbs, and adverbs, and NER tags corresponding to tangible types such as person, organizations, products, geopolitical location. That is, named-entity tags corresponding to concepts such as money, quantities, and dates as well as POS tags corresponding to punctuation, and pronouns were ignored since they are unlikely to refer to event mentions.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Features for Sentence Pair Classification", "sec_num": "3.1" }, { "text": "For semantic similarity we use cosine similarity between the average word vectors from GloVE 2 for 2 http://nlp.stanford.edu/data/glove.6B. the two sentences. Ozates, et al.(2016) proposed incorporating the type information of the dependency relations for sentence similarity calculation in context of sentence summarization for better capturing the syntactic and semantic similarity between two sentences. We use similarity between two sentences computed using their proposed \"Simple Bigram Approximate Kernel\" as an additional feature.", "cite_spans": [ { "start": 159, "end": 179, "text": "Ozates, et al.(2016)", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "Features for Sentence Pair Classification", "sec_num": "3.1" }, { "text": "Overall, the set of \"Extended Similarities\", correspond to a total of 27 features. ", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Features for Sentence Pair Classification", "sec_num": "3.1" }, { "text": "Our deep learning setup for learning sentence pair classification is shown in Figure 3 . We use the features described in the previous section for training our classifier. The BERT with POS embeddings are first fed into a BiLSTM layer with an output dimension of 64. Next, we flatten the n \u00d7 64 matrix into a n * 64 vector and run it through a dropout layer with 0.3 dropout rate. Another linear layer is applied to convert the representation into a vector with length 200. From here, we concatenate our similarity features and send them through a linear layer to obtain class probabilities representing the coreferent (label = 1) and non-coreferent (label = 0) classes.", "cite_spans": [], "ref_spans": [ { "start": 78, "end": 86, "text": "Figure 3", "ref_id": "FIGREF2" } ], "eq_spans": [], "section": "Sentence Pair Classifier", "sec_num": "3.2" }, { "text": "Given labels corresponding to each pair of sentences obtained from our classification module, we employ the clustering algorithm from\u00d6rs et al. (2020) for grouping the sentences in the document. This algorithm, similar to hierarchical clustering, creates clusters in a bottom-up fashion using maximum scores instead of the minimum distance to group two points into the same cluster. For us, score of a pair refers to the probability of the sentences being coreferent with. We refer the interested reader to Algorithm 2 in\u00d6rs et al. 2020for the pseudo-code. In contrast, with algorithms such as k-medoids, the algorithm employed in our solution has the advantage of determining a different number of clusters for each article in a flexible manner.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Article-level Clustering", "sec_num": "3.3" }, { "text": "We used the data from ACL-CASE 2021 (H\u00fcrriyetoglu et al., 2021) (Task 1 Subtask 3) for training and testing our models. The dataset comprises of news articles referring to socio-political and crisis event in three languages: English, Portuguese, and Spanish. We refer the interested reader to the overview paper (H\u00fcrriyetoglu et al., 2021) and the task websites 3 for details of this dataset. We summarize the train and test sizes of the dataset in Table 1 . For the train set, we were provided with 596 English news articles, 21 Portuguese articles, and 11 Spanish articles. For each article, only sentences with event mentions are included in the dataset instead of all sentences. The test performance was evaluated using the CoNLL-2012 average F 1 scores obtained by averaging across M U C, B 3 and CEAF e F 1 scores (Pradhan et al., 2012) and was computed on the setup provided by the organizers on Codalab.", "cite_spans": [ { "start": 312, "end": 339, "text": "(H\u00fcrriyetoglu et al., 2021)", "ref_id": "BIBREF5" }, { "start": 820, "end": 842, "text": "(Pradhan et al., 2012)", "ref_id": "BIBREF11" } ], "ref_spans": [ { "start": 449, "end": 456, "text": "Table 1", "ref_id": "TABREF1" } ], "eq_spans": [], "section": "Dataset and Evaluation", "sec_num": "4.1" }, { "text": "Training Datasets: To handle the low number of examples available with Portuguese and Spanish, we create two datasets for training our models: (1) The \"Multilingual train set\" is obtained by simply putting the examples from all languages together whereas (2) the \"English train set\" is obtained by first employing the Google Translate API 4 and translating all available non-English training examples to English and combining with the English training data. The multilingual dataset can be used directly for training language-agnostic models, for example using cross-lingual embeddings (Conneau et al., 2017) and Multilingual BERT.", "cite_spans": [ { "start": 586, "end": 608, "text": "(Conneau et al., 2017)", "ref_id": "BIBREF3" } ], "ref_spans": [], "eq_spans": [], "section": "Experimental Setup", "sec_num": "4.2" }, { "text": "Feature Extraction: We experimented with two BERT implementations from Huggingface (Wolf et al., 2020) .", "cite_spans": [ { "start": 83, "end": 102, "text": "(Wolf et al., 2020)", "ref_id": "BIBREF13" } ], "ref_spans": [], "eq_spans": [], "section": "Experimental Setup", "sec_num": "4.2" }, { "text": "The first model, bert-base-cased, was pretrained on English text and has 12 layers, 768 hidden, 12 heads and 109M parameters. We fine-tuned this model using our \"English train set\".", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experimental Setup", "sec_num": "4.2" }, { "text": "Our second model, bert-base-multilingual-cased, was pretrained on the top 104 languages in Wikipedia and has 12 layers, 768 hidden, 12 heads and 179M parameters. We fine-tuned this model using our \"Multilingual train set\".", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experimental Setup", "sec_num": "4.2" }, { "text": "We used Stanford's Stanza package (Qi et al., 2020) for obtaining POS, NER, and dependency tree tags. The \"Universal POS tags\" (upos scheme) with 17 POS tags and the NER tags referring to PERSON, NORP (Nationalities/religious/political group), FAC (Facility), ORG (Organization), GPE (Countries/cities/states), LOC (Location), PRODUCT, EVENT, WORK OF ART, and LANGUAGE were used in experiments. 5 When constructing the \"Basic Similarities\", all words are lemmatised before we compare their surface-form overlap. For entities, we use token sort ratio 6 score of more than 90% to define a positive overap occurence instead of an exact match to allow for some small discrepancy in NEs (e.g. \"Sohrabuddin Sheikh\" and \"Sohrabuddin Sheikh 's\" refer to the same entity).", "cite_spans": [ { "start": 34, "end": 51, "text": "(Qi et al., 2020)", "ref_id": "BIBREF12" }, { "start": 395, "end": 396, "text": "5", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "Experimental Setup", "sec_num": "4.2" }, { "text": "Classifier Settings: To train our classifier, we used the Adam optimizer with \u03b2 1 = 0.9, \u03b2 2 = 0.999 and a learning rate of 2e \u2212 5 with linear decay. Cross Entropy Loss was used with class weights computed from the training sample. Each iteration was of batch size 16 and all experiments were ran on Tesla V100 SXM2 32GB GPU device. Five-fold cross-validation (5-CV) experiments were used for parameter tuning. We also report macro and micro Adjusted Rand Index (ARI) and F 1 scores in addition to ConLL F 1 since they were used for selecting the top-3 runs for the test set in line with the measures employed in the previous rounds of the competition (H\u00fcrriyetoglu et al., 2019 (H\u00fcrriyetoglu et al., , 2020 . Since the test labels were not released and evaluation is performed on the competition setup, only CoNLL F 1 scores are reported for the test data. Other details, such as hyperparameter settings and run times, are included in Appendix A.1. Table 2 reports the average scores for our 5-CV setup across the five scoring metrics (CoNLL F 1 , Macro ARI, Micro ARI, Macro F 1 and Micro F 1 ). Table 3 reports the CoNLL F 1 score on the test data for the winning system and our models at the ACL-CASE 2021 Shared Task 1 Subtask 3 across the three languages -English, Portuguese, and Spanish.", "cite_spans": [ { "start": 652, "end": 678, "text": "(H\u00fcrriyetoglu et al., 2019", "ref_id": "BIBREF6" }, { "start": 679, "end": 707, "text": "(H\u00fcrriyetoglu et al., , 2020", "ref_id": "BIBREF7" } ], "ref_spans": [ { "start": 950, "end": 957, "text": "Table 2", "ref_id": "TABREF3" }, { "start": 1098, "end": 1105, "text": "Table 3", "ref_id": "TABREF5" } ], "eq_spans": [], "section": "Experimental Setup", "sec_num": "4.2" }, { "text": "Based on CV experiments, our best model for all scoring measures is the English BERT model with all features included, achieving 84.92% CoNLL F 1 score. The same model also performed the best on the English test set with 81.20% CoNLL F 1 score and was ranked 2 nd among fellow competitors on this shared task.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Results and Analysis", "sec_num": "4.3" }, { "text": "For non-English test sets, our best performing model is the Multilingual BERT model with all features excluding \"Extended similarities\". This model achieved 93.03% CoNLL F 1 score and ranked 1 st for Portuguese. For Spanish, we ob-tained a CoNLL F 1 score of 83.15% and ranked 2 nd among competitors.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Results and Analysis", "sec_num": "4.3" }, { "text": "For the English test set, the BERT model performs better than the Multilingual BERT model on average (79.19% versus 78.01%). Additionally, because the train/validation splits are predominantly comprised of English articles (596/628 = 94.90%), the fluctuations in performance on validation splits largely tally with the fluctuations in performance on the English portion of the data. Therefore, unsurprisingly, BERT (English) performs better than Multilingual BERT for English data.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "BERT versus Multilingual BERT", "sec_num": "4.3.1" }, { "text": "For non-English test sets, we obtained best performance using the Multilingual BERT model. We hypothesize that the translation of non-English examples to English might have caused some loss of inherent signals present in other languages that are useful for the ESCI task. These signals are possibly better harnessed by retaining the language and using language-specific Stanza taggers along with Multilingual BERT.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "BERT versus Multilingual BERT", "sec_num": "4.3.1" }, { "text": "Overall, we find that combining BERT embeddings, POS embeddings and basic similarity features achieve the best validation performance across all measures. We observe that \"Extended similarities\" do not show uniform improvement in performance in multilingual settings. Arguably, there is redundancy among our lexical similarity features and semantic similarities were not found to improve performance on the ESCI task. However, considering the small-scale of our working datasets, these features need further study.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "BERT versus Multilingual BERT", "sec_num": "4.3.1" }, { "text": "We presented a two-step solution for the ESCI task of ACL-CASE 2021. Our solution based on In this paper, we focused on within-document coreference sentences. It is common for coreference resolution tasks to also focus on crossdocument settings (i.e. identify coreferent event mentions across multiple documents) (Zeng et al., 2020) as such models can better aid downstream tasks like contradiction detection or identification of \"fake news\". In future, we hope to extend our models to work across documents. Additionally, multiple events might be presented in a sentence. The shared task focuses on hard clustering (i.e. each sentence can only belong to one cluster). However, we believe it is valuable to also investigate cases where the event clusters overlap.", "cite_spans": [ { "start": 313, "end": 332, "text": "(Zeng et al., 2020)", "ref_id": "BIBREF14" } ], "ref_spans": [], "eq_spans": [], "section": "Conclusions and Future Work", "sec_num": "5" }, { "text": "Our code and supplementary materials can be found on Github at https://github.com/NUS-IDS/ EventSentenceCoref", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "https://emw.ku.edu.tr/case-2021/, https://github.com/emerging-welfare/ case-2021-shared-task", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "https://pypi.org/project/ google-trans-new 5 At present, NER models are only available for Spanish and English in Stanza. 6 https://github.com/seatgeek/ fuzzywuzzy", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null } ], "back_matter": [ { "text": "This research is supported by the National Research Foundation, Singapore under its Industry Alignment Fund -Pre-positioning (IAF-PP) Funding Initiative. Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not reflect the views of National Research Foundation, Singapore.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Acknowledgments", "sec_num": null }, { "text": "A.1 Replication Checklist\u2022 Hyperparameters: Apart from hyperparameters mentioned in Section 4.2, our BERT models take the default configuration from Huggingface (Wolf et al., 2020 ).\u2022 Time taken: For 5 folds over 10 epochs each, our code takes on average 5hours : 27minutes : 48seconds to train, validate and predict. For a single run over 10 epochs, our code takes on average 43minutes : 49seconds to train and predict.", "cite_spans": [ { "start": 161, "end": 179, "text": "(Wolf et al., 2020", "ref_id": "BIBREF13" } ], "ref_spans": [], "eq_spans": [], "section": "A Appendix", "sec_num": null } ], "bib_entries": { "BIBREF0": { "ref_id": "b0", "title": "Revisiting joint modeling of cross-document entity and event coreference resolution", "authors": [ { "first": "Shany", "middle": [], "last": "Barhom", "suffix": "" }, { "first": "Vered", "middle": [], "last": "Shwartz", "suffix": "" }, { "first": "Alon", "middle": [], "last": "Eirew", "suffix": "" }, { "first": "Michael", "middle": [], "last": "Bugert", "suffix": "" } ], "year": 2019, "venue": "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics", "volume": "", "issue": "", "pages": "4179--4189", "other_ids": { "DOI": [ "10.18653/v1/P19-1409" ] }, "num": null, "urls": [], "raw_text": "Shany Barhom, Vered Shwartz, Alon Eirew, Michael Bugert, Nils Reimers, and Ido Dagan. 2019. Re- visiting joint modeling of cross-document entity and event coreference resolution. In Proceedings of the 57th Annual Meeting of the Association for Com- putational Linguistics, pages 4179-4189, Florence, Italy. Association for Computational Linguistics.", "links": null }, "BIBREF1": { "ref_id": "b1", "title": "Unsupervised event coreference resolution with rich linguistic features", "authors": [ { "first": "Cosmin", "middle": [], "last": "Bejan", "suffix": "" }, { "first": "Sanda", "middle": [], "last": "Harabagiu", "suffix": "" } ], "year": 2010, "venue": "Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics", "volume": "", "issue": "", "pages": "1412--1422", "other_ids": {}, "num": null, "urls": [], "raw_text": "Cosmin Bejan and Sanda Harabagiu. 2010. Unsuper- vised event coreference resolution with rich linguis- tic features. In Proceedings of the 48th Annual Meet- ing of the Association for Computational Linguis- tics, pages 1412-1422, Uppsala, Sweden. Associa- tion for Computational Linguistics.", "links": null }, "BIBREF2": { "ref_id": "b2", "title": "Event coreference resolution by iteratively unfolding inter-dependencies among events", "authors": [ { "first": "Prafulla", "middle": [], "last": "Kumar Choubey", "suffix": "" }, { "first": "Ruihong", "middle": [], "last": "Huang", "suffix": "" } ], "year": 2017, "venue": "Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing", "volume": "", "issue": "", "pages": "2124--2133", "other_ids": { "DOI": [ "10.18653/v1/D17-1226" ] }, "num": null, "urls": [], "raw_text": "Prafulla Kumar Choubey and Ruihong Huang. 2017. Event coreference resolution by iteratively unfold- ing inter-dependencies among events. In Proceed- ings of the 2017 Conference on Empirical Methods in Natural Language Processing, pages 2124-2133, Copenhagen, Denmark. Association for Computa- tional Linguistics.", "links": null }, "BIBREF3": { "ref_id": "b3", "title": "Word translation without parallel data", "authors": [ { "first": "Alexis", "middle": [], "last": "Conneau", "suffix": "" }, { "first": "Guillaume", "middle": [], "last": "Lample", "suffix": "" }, { "first": "Marc'aurelio", "middle": [], "last": "Ranzato", "suffix": "" }, { "first": "Ludovic", "middle": [], "last": "Denoyer", "suffix": "" }, { "first": "Herv\u00e9", "middle": [], "last": "J\u00e9gou", "suffix": "" } ], "year": 2017, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Alexis Conneau, Guillaume Lample, Marc'Aurelio Ranzato, Ludovic Denoyer, and Herv\u00e9 J\u00e9gou. 2017. Word translation without parallel data. CoRR.", "links": null }, "BIBREF4": { "ref_id": "b4", "title": "BERT: Pre-training of deep bidirectional transformers for language understanding", "authors": [ { "first": "Jacob", "middle": [], "last": "Devlin", "suffix": "" }, { "first": "Ming-Wei", "middle": [], "last": "Chang", "suffix": "" }, { "first": "Kenton", "middle": [], "last": "Lee", "suffix": "" }, { "first": "Kristina", "middle": [], "last": "Toutanova", "suffix": "" } ], "year": 2019, "venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", "volume": "1", "issue": "", "pages": "4171--4186", "other_ids": { "DOI": [ "10.18653/v1/N19-1423" ] }, "num": null, "urls": [], "raw_text": "Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language under- standing. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171-4186, Minneapolis, Minnesota. Associ- ation for Computational Linguistics.", "links": null }, "BIBREF5": { "ref_id": "b5", "title": "Multilingual protest news detectionshared task 1, case 2021", "authors": [ { "first": "Ali", "middle": [], "last": "H\u00fcrriyetoglu", "suffix": "" }, { "first": "Osman", "middle": [], "last": "Mutlu", "suffix": "" }, { "first": "Erdem", "middle": [], "last": "Farhana Ferdousi Liza", "suffix": "" }, { "first": "Ritesh", "middle": [], "last": "Y\u00f6r\u00fck", "suffix": "" }, { "first": "Shyam", "middle": [], "last": "Kumar", "suffix": "" }, { "first": "", "middle": [], "last": "Ratan", "suffix": "" } ], "year": 2021, "venue": "Proceedings of the 4th Workshop on Challenges and Applications of Automated Extraction of Socio-political Events from Text (CASE 2021), online", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Ali H\u00fcrriyetoglu, Osman Mutlu, Farhana Ferdousi Liza, Erdem Y\u00f6r\u00fck, Ritesh Kumar, and Shyam Ratan. 2021. Multilingual protest news detection - shared task 1, case 2021. In Proceedings of the 4th Workshop on Challenges and Applications of Auto- mated Extraction of Socio-political Events from Text (CASE 2021), online. Association for Computational Linguistics (ACL).", "links": null }, "BIBREF6": { "ref_id": "b6", "title": "A task set proposal for automatic protest information collection across multiple countries", "authors": [ { "first": "Ali", "middle": [], "last": "H\u00fcrriyetoglu", "suffix": "" }, { "first": "Erdem", "middle": [], "last": "Y\u00f6r\u00fck", "suffix": "" }, { "first": "Deniz", "middle": [], "last": "Y\u00fcret", "suffix": "" }, { "first": "Burak", "middle": [], "last": "Agr\u0131 Yoltar", "suffix": "" }, { "first": "F\u0131rat", "middle": [], "last": "G\u00fcrel", "suffix": "" }, { "first": "Osman", "middle": [], "last": "Duru\u015fan", "suffix": "" }, { "first": "", "middle": [], "last": "Mutlu", "suffix": "" } ], "year": 2019, "venue": "Advances in Information Retrieval", "volume": "", "issue": "", "pages": "316--323", "other_ids": {}, "num": null, "urls": [], "raw_text": "Ali H\u00fcrriyetoglu, Erdem Y\u00f6r\u00fck, Deniz Y\u00fcret, \u00c7 agr\u0131 Yoltar, Burak G\u00fcrel, F\u0131rat Duru\u015fan, and Osman Mutlu. 2019. A task set proposal for automatic protest information collection across multiple coun- tries. In Advances in Information Retrieval, pages 316-323, Cham. Springer International Publishing.", "links": null }, "BIBREF7": { "ref_id": "b7", "title": "Automated extraction of socio-political events from news (AESPEN): Workshop and shared task report", "authors": [ { "first": "Ali", "middle": [], "last": "H\u00fcrriyetoglu", "suffix": "" }, { "first": "Vanni", "middle": [], "last": "Zavarella", "suffix": "" }, { "first": "Hristo", "middle": [], "last": "Tanev", "suffix": "" }, { "first": "Erdem", "middle": [], "last": "Y\u00f6r\u00fck", "suffix": "" }, { "first": "Ali", "middle": [], "last": "Safaya", "suffix": "" }, { "first": "Osman", "middle": [], "last": "Mutlu", "suffix": "" } ], "year": 2020, "venue": "Proceedings of the Workshop on Automated Extraction of Socio-political Events from News 2020", "volume": "", "issue": "", "pages": "1--6", "other_ids": {}, "num": null, "urls": [], "raw_text": "Ali H\u00fcrriyetoglu, Vanni Zavarella, Hristo Tanev, Er- dem Y\u00f6r\u00fck, Ali Safaya, and Osman Mutlu. 2020. Automated extraction of socio-political events from news (AESPEN): Workshop and shared task report. In Proceedings of the Workshop on Automated Ex- traction of Socio-political Events from News 2020, pages 1-6, Marseille, France. European Language Resources Association (ELRA).", "links": null }, "BIBREF8": { "ref_id": "b8", "title": "Event linking with sentential features from convolutional neural networks", "authors": [ { "first": "Sebastian", "middle": [], "last": "Krause", "suffix": "" }, { "first": "Feiyu", "middle": [], "last": "Xu", "suffix": "" }, { "first": "Hans", "middle": [], "last": "Uszkoreit", "suffix": "" }, { "first": "Dirk", "middle": [], "last": "Weissenborn", "suffix": "" } ], "year": 2016, "venue": "Proceedings of The 20th SIGNLL Conference on Computational Natural Language Learning", "volume": "", "issue": "", "pages": "239--249", "other_ids": { "DOI": [ "10.18653/v1/K16-1024" ] }, "num": null, "urls": [], "raw_text": "Sebastian Krause, Feiyu Xu, Hans Uszkoreit, and Dirk Weissenborn. 2016. Event linking with sentential features from convolutional neural networks. In Proceedings of The 20th SIGNLL Conference on Computational Natural Language Learning, pages 239-249, Berlin, Germany. Association for Compu- tational Linguistics.", "links": null }, "BIBREF9": { "ref_id": "b9", "title": "Event clustering within news articles", "authors": [ { "first": "Faik", "middle": [], "last": "Kerem\u00f6rs", "suffix": "" }, { "first": "S\u00fcveyda", "middle": [], "last": "Yeniterzi", "suffix": "" }, { "first": "Reyyan", "middle": [], "last": "Yeniterzi", "suffix": "" } ], "year": 2020, "venue": "Proceedings of the Workshop on Automated Extraction of Socio-political Events from News 2020", "volume": "", "issue": "", "pages": "63--68", "other_ids": {}, "num": null, "urls": [], "raw_text": "Faik Kerem\u00d6rs, S\u00fcveyda Yeniterzi, and Reyyan Yen- iterzi. 2020. Event clustering within news articles. In Proceedings of the Workshop on Automated Ex- traction of Socio-political Events from News 2020, pages 63-68, Marseille, France. European Language Resources Association (ELRA).", "links": null }, "BIBREF10": { "ref_id": "b10", "title": "Sentence similarity based on dependency tree kernels for multi-document summarization", "authors": [ { "first": "Arzucan\u00f6zg\u00fcr", "middle": [], "last": "Bet\u00fcl\u00f6zate\u015f", "suffix": "" }, { "first": "Dragomir", "middle": [], "last": "Radev", "suffix": "" } ], "year": 2016, "venue": "Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC'16)", "volume": "", "issue": "", "pages": "2833--2838", "other_ids": {}, "num": null, "urls": [], "raw_text": "\u015e aziye Bet\u00fcl\u00d6zate\u015f, Arzucan\u00d6zg\u00fcr, and Dragomir Radev. 2016. Sentence similarity based on de- pendency tree kernels for multi-document sum- marization. In Proceedings of the Tenth Inter- national Conference on Language Resources and Evaluation (LREC'16), pages 2833-2838, Portoro\u017e, Slovenia. European Language Resources Associa- tion (ELRA).", "links": null }, "BIBREF11": { "ref_id": "b11", "title": "CoNLL-2012 shared task: Modeling multilingual unrestricted coreference in OntoNotes", "authors": [ { "first": "Alessandro", "middle": [], "last": "Sameer Pradhan", "suffix": "" }, { "first": "Nianwen", "middle": [], "last": "Moschitti", "suffix": "" }, { "first": "Olga", "middle": [], "last": "Xue", "suffix": "" }, { "first": "Yuchen", "middle": [], "last": "Uryupina", "suffix": "" }, { "first": "", "middle": [], "last": "Zhang", "suffix": "" } ], "year": 2012, "venue": "Joint Conference on EMNLP and CoNLL -Shared Task", "volume": "", "issue": "", "pages": "1--40", "other_ids": {}, "num": null, "urls": [], "raw_text": "Sameer Pradhan, Alessandro Moschitti, Nianwen Xue, Olga Uryupina, and Yuchen Zhang. 2012. CoNLL- 2012 shared task: Modeling multilingual unre- stricted coreference in OntoNotes. In Joint Confer- ence on EMNLP and CoNLL -Shared Task, pages 1-40, Jeju Island, Korea. Association for Computa- tional Linguistics.", "links": null }, "BIBREF12": { "ref_id": "b12", "title": "Stanza: A Python natural language processing toolkit for many human languages", "authors": [ { "first": "Peng", "middle": [], "last": "Qi", "suffix": "" }, { "first": "Yuhao", "middle": [], "last": "Zhang", "suffix": "" }, { "first": "Yuhui", "middle": [], "last": "Zhang", "suffix": "" }, { "first": "Jason", "middle": [], "last": "Bolton", "suffix": "" }, { "first": "Christopher", "middle": [ "D" ], "last": "Manning", "suffix": "" } ], "year": 2020, "venue": "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: System Demonstrations", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Peng Qi, Yuhao Zhang, Yuhui Zhang, Jason Bolton, and Christopher D. Manning. 2020. Stanza: A Python natural language processing toolkit for many human languages. In Proceedings of the 58th An- nual Meeting of the Association for Computational Linguistics: System Demonstrations.", "links": null }, "BIBREF13": { "ref_id": "b13", "title": "Transformers: State-of-the-art natural language processing", "authors": [ { "first": "Thomas", "middle": [], "last": "Wolf", "suffix": "" }, { "first": "Lysandre", "middle": [], "last": "Debut", "suffix": "" }, { "first": "Victor", "middle": [], "last": "Sanh", "suffix": "" }, { "first": "Julien", "middle": [], "last": "Chaumond", "suffix": "" }, { "first": "Clement", "middle": [], "last": "Delangue", "suffix": "" }, { "first": "Anthony", "middle": [], "last": "Moi", "suffix": "" }, { "first": "Pierric", "middle": [], "last": "Cistac", "suffix": "" }, { "first": "Tim", "middle": [], "last": "Rault", "suffix": "" }, { "first": "Remi", "middle": [], "last": "Louf", "suffix": "" }, { "first": "Morgan", "middle": [], "last": "Funtowicz", "suffix": "" }, { "first": "Joe", "middle": [], "last": "Davison", "suffix": "" }, { "first": "Sam", "middle": [], "last": "Shleifer", "suffix": "" }, { "first": "Clara", "middle": [], "last": "Patrick Von Platen", "suffix": "" }, { "first": "Yacine", "middle": [], "last": "Ma", "suffix": "" }, { "first": "Julien", "middle": [], "last": "Jernite", "suffix": "" }, { "first": "Canwen", "middle": [], "last": "Plu", "suffix": "" }, { "first": "Teven", "middle": [ "Le" ], "last": "Xu", "suffix": "" }, { "first": "Sylvain", "middle": [], "last": "Scao", "suffix": "" }, { "first": "Mariama", "middle": [], "last": "Gugger", "suffix": "" }, { "first": "", "middle": [], "last": "Drame", "suffix": "" } ], "year": 2020, "venue": "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations", "volume": "", "issue": "", "pages": "38--45", "other_ids": { "DOI": [ "10.18653/v1/2020.emnlp-demos.6" ] }, "num": null, "urls": [], "raw_text": "Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pier- ric Cistac, Tim Rault, Remi Louf, Morgan Funtow- icz, Joe Davison, Sam Shleifer, Patrick von Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander Rush. 2020. Trans- formers: State-of-the-art natural language process- ing. In Proceedings of the 2020 Conference on Em- pirical Methods in Natural Language Processing: System Demonstrations, pages 38-45, Online. Asso- ciation for Computational Linguistics.", "links": null }, "BIBREF14": { "ref_id": "b14", "title": "Event coreference resolution with their paraphrases and argument-aware embeddings", "authors": [ { "first": "Yutao", "middle": [], "last": "Zeng", "suffix": "" }, { "first": "Xiaolong", "middle": [], "last": "Jin", "suffix": "" }, { "first": "Saiping", "middle": [], "last": "Guan", "suffix": "" }, { "first": "Jiafeng", "middle": [], "last": "Guo", "suffix": "" }, { "first": "Xueqi", "middle": [], "last": "Cheng", "suffix": "" } ], "year": 2020, "venue": "Proceedings of the 28th International Conference on Computational Linguistics", "volume": "", "issue": "", "pages": "3084--3094", "other_ids": {}, "num": null, "urls": [], "raw_text": "Yutao Zeng, Xiaolong Jin, Saiping Guan, Jiafeng Guo, and Xueqi Cheng. 2020. Event coreference reso- lution with their paraphrases and argument-aware embeddings. In Proceedings of the 28th Inter- national Conference on Computational Linguistics, pages 3084-3094, Barcelona, Spain (Online). Inter- national Committee on Computational Linguistics.", "links": null }, "BIBREF15": { "ref_id": "b15", "title": "Neural coreference resolution with deep biaffine attention by joint mention detection and mention clustering", "authors": [ { "first": "Rui", "middle": [], "last": "Zhang", "suffix": "" }, { "first": "C\u00edcero", "middle": [], "last": "Nogueira", "suffix": "" }, { "first": "Michihiro", "middle": [], "last": "Santos", "suffix": "" }, { "first": "Bing", "middle": [], "last": "Yasunaga", "suffix": "" }, { "first": "Dragomir", "middle": [], "last": "Xiang", "suffix": "" }, { "first": "", "middle": [], "last": "Radev", "suffix": "" } ], "year": 2018, "venue": "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics", "volume": "", "issue": "", "pages": "102--107", "other_ids": { "DOI": [ "10.18653/v1/P18-2017" ] }, "num": null, "urls": [], "raw_text": "Rui Zhang, C\u00edcero Nogueira dos Santos, Michihiro Yasunaga, Bing Xiang, and Dragomir Radev. 2018. Neural coreference resolution with deep biaffine at- tention by joint mention detection and mention clus- tering. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Vol- ume 2: Short Papers), pages 102-107, Melbourne, Australia. Association for Computational Linguis- tics.", "links": null } }, "ref_entries": { "FIGREF0": { "num": null, "uris": null, "text": "Example English article from training dataset from ACL-CASE 2021. The sentences 1, 2, 7 with event mentions as well as the target clustering {[2], [1,7]} are highlighted.", "type_str": "figure" }, "FIGREF1": { "num": null, "uris": null, "text": "Overall model pipeline. Notes. (1) Convert article into sentence pairs for binary classification, and(2)Taking predicted probabilities to perform article level clustering of sentences.", "type_str": "figure" }, "FIGREF2": { "num": null, "uris": null, "text": "Overview of the sentence pair classification model. BERT embeddings, POS embeddings and similarity features are used to train a BiLSTM-based deep learning model.", "type_str": "figure" }, "TABREF1": { "html": null, "text": "Number of observations at different unit levels for train and test set", "num": null, "content": "", "type_str": "table" }, "TABREF3": { "html": null, "text": "Evaluation results over validation sets from 5 folds. Notes. Scores are reported in percentages (%) and averaged across the folds. Best score per column is bolded.", "num": null, "content": "
", "type_str": "table" }, "TABREF5": { "html": null, "text": "Evaluation results over test sets submitted to Codalab. Notes. Scores are reported in percentages (%). Our best score per column is bolded. sentence pair classification effectively harnesses BERT and POS embeddings and combines them with linguistic similarity features for accurate sentence coreference resolution across languages. Indeed, our models ranked 2 nd , 1 st and 2 nd obtaining CoNLL F 1 scores of 81.20%, 93.03%, 83.15% for the English, Portuguese and Spanish test sets, respectively, in the competition.", "num": null, "content": "
", "type_str": "table" } } } }