ACL-OCL / Base_JSON /prefixC /json /case /2021.case-1.4.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "2021",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T13:13:55.245681Z"
},
"title": "PROTEST-ER: Retraining BERT for Protest Event Extraction",
"authors": [
{
"first": "Tommaso",
"middle": [],
"last": "Caselli",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "University of Groningen \u2020 Ko\u00e7 University Symanto Research",
"location": {}
},
"email": "t.caselli@rug.nl"
},
{
"first": "Osman",
"middle": [],
"last": "Mutlu",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "University of Groningen \u2020 Ko\u00e7 University Symanto Research",
"location": {}
},
"email": "omutlu@ku.edu.tr"
},
{
"first": "Angelo",
"middle": [],
"last": "Basile",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "University of Groningen \u2020 Ko\u00e7 University Symanto Research",
"location": {}
},
"email": "angelo.basile@symanto.com"
},
{
"first": "Ali",
"middle": [],
"last": "H\u00fcrriyeto\u011dlu",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "University of Groningen \u2020 Ko\u00e7 University Symanto Research",
"location": {}
},
"email": "ahurriyetoglu@ku.edu.tr"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "We analyze the effect of further pre-training BERT with different domain specific data as an unsupervised domain adaptation strategy for event extraction. Portability of event extraction models is particularly challenging, with large performance drops affecting data on the same text genres (e.g., news). We present PROTEST-ER, a retrained BERT model for protest event extraction. PROTEST-ER outperforms a corresponding generic BERT on outof-domain data of 8.1 points. Our best performing models reach 51.91-46.39 F1 across both domains.",
"pdf_parse": {
"paper_id": "2021",
"_pdf_hash": "",
"abstract": [
{
"text": "We analyze the effect of further pre-training BERT with different domain specific data as an unsupervised domain adaptation strategy for event extraction. Portability of event extraction models is particularly challenging, with large performance drops affecting data on the same text genres (e.g., news). We present PROTEST-ER, a retrained BERT model for protest event extraction. PROTEST-ER outperforms a corresponding generic BERT on outof-domain data of 8.1 points. Our best performing models reach 51.91-46.39 F1 across both domains.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "Events, i.e., things that happen in the world or states that hold true, play a central role in human lives. It is not a simplification to claim that our lives are nothing but a constant sequence of events. Nevertheless not all events are equally relevant, especially when the focus of attention and analysis moves away from individuals and touches upon societies. In this broader context, socio-political events are of particular interest since they directly impact and affect the lives of multiple individuals at the same time. Different actors (e.g., governments, multilateral organizations, NGOs, social movements) have various interests in collecting information and conducting analyses on this type of events. This, however, is a challenging task. The increasing availability and amount of data, thanks to the growth of the Web, calls for the development of automatic solutions based on Natural Language Processing (NLP).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction and Problem Statement",
"sec_num": "1"
},
{
"text": "Besides the good level of maturity reached by NLP systems in many areas, numerous challenges are still pending. Portability of systems, i.e., the reuse of previously trained systems for a specific task on different datasets, is one of them and it is far from being solved (Daum\u00e9 III, 2007; Plank and van Noord, 2011; Axelrod et al., 2011; Ganin and Lempitsky, 2015; Alam et al., 2018; Xie et al., 2018; Zhao et al., 2019; Ben-David et al., 2020) . As such, portability is a domain adaptation problem. Following Ramponi and Plank (2020) , we consider a domain to be a variety where each corpus, or dataset, can be described as a multidimensional region including notions such as topics, genres, writing styles, years of publication, socio-demographic aspects, annotation bias, among other unknown factors. Every dataset belonging to a different variety poses a domain adaptation challenge.",
"cite_spans": [
{
"start": 272,
"end": 289,
"text": "(Daum\u00e9 III, 2007;",
"ref_id": "BIBREF6"
},
{
"start": 290,
"end": 316,
"text": "Plank and van Noord, 2011;",
"ref_id": "BIBREF22"
},
{
"start": 317,
"end": 338,
"text": "Axelrod et al., 2011;",
"ref_id": "BIBREF3"
},
{
"start": 339,
"end": 365,
"text": "Ganin and Lempitsky, 2015;",
"ref_id": "BIBREF8"
},
{
"start": 366,
"end": 384,
"text": "Alam et al., 2018;",
"ref_id": "BIBREF2"
},
{
"start": 385,
"end": 402,
"text": "Xie et al., 2018;",
"ref_id": "BIBREF28"
},
{
"start": 403,
"end": 421,
"text": "Zhao et al., 2019;",
"ref_id": "BIBREF30"
},
{
"start": 422,
"end": 445,
"text": "Ben-David et al., 2020)",
"ref_id": "BIBREF4"
},
{
"start": 511,
"end": 535,
"text": "Ramponi and Plank (2020)",
"ref_id": "BIBREF23"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction and Problem Statement",
"sec_num": "1"
},
{
"text": "Unsupervised domain adaptation has a long tradition in NLP (Blitzer et al., 2006; McClosky et al., 2006; Moore and Lewis, 2010; Ganin et al., 2016; Ruder and Plank, 2017; Guo et al., 2018; Miller, 2019; Nishida et al., 2020) . The availability of large pre-trained transformer-based language models (TLMs), e.g., BERT (Devlin et al., 2019) , has inspired a new trend in domain adaptation, namely domain adaptive retraining (DAR) Han and Eisenstein, 2019; Rietzler et al., 2020; Gururangan et al., 2020) . The idea behind DAR is as simple as effective: first, additional textual material matching the target domain is selected, then the masked language modeling (MLM) objective is used to further train an existing TLMs. The outcome is a new TLM whose representations are shifted to better suit the target domain. Fine-tuning domain adapted TLMs results in improved performance.",
"cite_spans": [
{
"start": 59,
"end": 81,
"text": "(Blitzer et al., 2006;",
"ref_id": "BIBREF5"
},
{
"start": 82,
"end": 104,
"text": "McClosky et al., 2006;",
"ref_id": "BIBREF17"
},
{
"start": 105,
"end": 127,
"text": "Moore and Lewis, 2010;",
"ref_id": "BIBREF19"
},
{
"start": 128,
"end": 147,
"text": "Ganin et al., 2016;",
"ref_id": "BIBREF9"
},
{
"start": 148,
"end": 170,
"text": "Ruder and Plank, 2017;",
"ref_id": "BIBREF25"
},
{
"start": 171,
"end": 188,
"text": "Guo et al., 2018;",
"ref_id": "BIBREF10"
},
{
"start": 189,
"end": 202,
"text": "Miller, 2019;",
"ref_id": "BIBREF18"
},
{
"start": 203,
"end": 224,
"text": "Nishida et al., 2020)",
"ref_id": "BIBREF21"
},
{
"start": 318,
"end": 339,
"text": "(Devlin et al., 2019)",
"ref_id": "BIBREF7"
},
{
"start": 429,
"end": 454,
"text": "Han and Eisenstein, 2019;",
"ref_id": "BIBREF12"
},
{
"start": 455,
"end": 477,
"text": "Rietzler et al., 2020;",
"ref_id": "BIBREF24"
},
{
"start": 478,
"end": 502,
"text": "Gururangan et al., 2020)",
"ref_id": "BIBREF11"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction and Problem Statement",
"sec_num": "1"
},
{
"text": "This contribution applies this approach to develop a portable system for protest event extraction. Our unsupervised domain adaptation setting investigates two related aspects. The first concerns the impact of the data used to adapt a generic TLM to a target domain (i.e., protest events). The second targets the portability in a zero-shot scenario of a domain-adapted TLMs across protest event datasets. Our experimental results provide additional evidence that further pretraining TLM on domain-related data is a \"cheap\" and successful method in single-source single-target unsupervised domain adaptation settings. Furthermore, we show that fine-tuned retrained TLMs results in models with a better portability.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction and Problem Statement",
"sec_num": "1"
},
{
"text": "We focus on the protest event detection task following the 2019 CLEF ProtestNews Lab (H\u00fcrriyetoglu et al., 2019) . 1 Protest events are identified as politically motivated collective actions which lay outside the official mechanisms of political participation of the country in which the action takes place.",
"cite_spans": [
{
"start": 85,
"end": 112,
"text": "(H\u00fcrriyetoglu et al., 2019)",
"ref_id": "BIBREF14"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Task and Data",
"sec_num": "2"
},
{
"text": "The lab is organised around three nonoverlapping subtasks: (a.) document classification; (b.) sentence classification; and (c.) event extraction. Tasks (a.) and (b.) are text classification tasks, requiring systems to distinguish whether a document/sentence is referring to a protest event. The event extraction task is a sequence tagging problem requiring systems to identify event triggers and their corresponding arguments, similarly to other event extraction tasks, e.g., ACE (Linguistic Data Consortium, 2005) .",
"cite_spans": [
{
"start": 476,
"end": 514,
"text": "ACE (Linguistic Data Consortium, 2005)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Task and Data",
"sec_num": "2"
},
{
"text": "The lab is designed to challenge models' portability in an unsupervised setting: systems receive a training and development data belonging to one variety and are asked to test both against a dataset from the same variety and a different one. We report in Table 1 the distribution of the markables (event triggers and arguments) for event extraction across the two varieties. We refer to the same variety (or source) distributions as India and to the different variety (or target) as China. The data are good examples of differences across factors characterising language varieties. For instance, although they belong to the same text genre (news articles), they describe protest events from two countries that have historical and cultural differences concerning what is worth protesting (e.g., caste protests are specific to India) and the type of protests (e.g., riots vs. petitions). Differences in the political systems entail differences in the actors of the protest events which is mirrored in the named entities describing person or organization names. Language is a further challenge. Both datasets are in English but they present dialectal and stylistic differences.",
"cite_spans": [],
"ref_spans": [
{
"start": 255,
"end": 262,
"text": "Table 1",
"ref_id": "TABREF1"
}
],
"eq_spans": [],
"section": "Task and Data",
"sec_num": "2"
},
{
"text": "We quantified differences and similarities by comparing the training data (India train ) against the two test ones (India test and China test ) using the Jensen-Shannon (J-S) divergence and the out-of-vocabulary rate (OOV) that previous work has shown to be particularly useful for this purpose (Ruder and Plank, 2017) . The figures in Table 2 better show how these data distributions occupy different regions in the variety space, with India test being closer to the training data than China test . Tackling these similarities and differences is at the heart of our domain adaptation problem for event extraction. A further challenge is posed by the limited amount of training material. A comparison against the training portion of ACE shows that Protest-News has 5 times less triggers and 4 times less arguments. 2 Unlike ACE, event triggers are not further classified into subtypes. However, seven argument types are annotated, namely participant, organiser, target, etime (event time), place, fname (facility name), and loc (location). The role set is inspired by ACE Attack and Demonstrate event types but they are more fine-grained. The markables are encoded in a BIO scheme (Beginning, Inside, Outside), resulting in different alphabets for triggers (e.g. B-trigger, I-trigger and O) and each of the arguments (e.g. O, B-organiser, I-organiser, B-etime, I-etime, etc.).",
"cite_spans": [
{
"start": 295,
"end": 318,
"text": "(Ruder and Plank, 2017)",
"ref_id": "BIBREF25"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Task and Data",
"sec_num": "2"
},
{
"text": "We applied DAR to English BERT base-uncased to fill a gap in language variety between BERT, trained on the BooksCorpus and Wikipedia, and the ProtestNews's data.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Continue Pre-training to Adapt",
"sec_num": "3"
},
{
"text": "We collected two sets of domain related data from the TREC Washington Post Corpus version Table 5 : J-S (Similarity) and OOV (Diversity) between the DAR datasets WPC-Gen and WPC-EV and the and test data distributions for the event extraction task.",
"cite_spans": [],
"ref_spans": [
{
"start": 90,
"end": 97,
"text": "Table 5",
"ref_id": null
}
],
"eq_spans": [],
"section": "Continue Pre-training to Adapt",
"sec_num": "3"
},
{
"text": "We apply each data collection separately BERT base-uncased by further training for 100 epochs using the MLM objective. The outcomes are two pre-trained language models: NEWS-BERT and PROTEST-ER. The differences between the models are assumed to be minimal but yet relevant to assess the impact of the data used for DAR. To further support this claim we report in Table 5 an analysis of the similarities and differences of 3 https://trec.nist.gov/data/wapost/ the DAR data materials against the India and China test data. As the figures show, the DAR datasets are equally different from the protest event extraction ones. Furthermore, we did not modify BERT original vocabulary by introducing new tokens. More details on the retraining parameters are reported in the Appendix A.1.",
"cite_spans": [],
"ref_spans": [
{
"start": 363,
"end": 370,
"text": "Table 5",
"ref_id": null
}
],
"eq_spans": [],
"section": "Continue Pre-training to Adapt",
"sec_num": "3"
},
{
"text": "Event extraction is framed as a token-level classification task. We adopt a joint strategy where triggers' and arguments' extent and labels are predicted at once (Nguyen et al., 2016) . We used India test to identify the best model (NEWS-BERT vs. PROTEST-ER) and system's input granularity. With respect to this latter point, we investigate whether processing data at document or sentence level could benefit the TLMs as a strategy to deal with limited training materials. We compare each configuration against a generic BERT counterpart. We fine-tune each model by training all the parameters simultaneously. All models are evaluated using the official script from the ProtestNews Lab. Triggers and arguments are correctly identified only if both the extent and the label are correct. We apply to China only the best model and input format.",
"cite_spans": [
{
"start": 162,
"end": 183,
"text": "(Nguyen et al., 2016)",
"ref_id": "BIBREF20"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Experiments and Results",
"sec_num": "4"
},
{
"text": "India data Results for India are illustrated in Table 3. In general, PROTEST-ER obtains better results than BERT and NEWS-BERT. Sentence qualifies as the best input format for PROTEST-ER, while document works best for NEWS-BERT and",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experiments and Results",
"sec_num": "4"
},
{
"text": "The language variety of the data distributions used for DAR has a big impact on the performance of fine-tuned systems, with NEWS-BERT being the worst model. The extra training should have made this model more suited for working with news articles than the corresponding generic BERT. This indicates that selection of suitable data is an essential step for successfully applying DAR.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "BERT.",
"sec_num": null
},
{
"text": "Globally, the results show that DAR has a positive effect on Precision, especially when sentences are used as input for fine tuning the models. Positive effects on Recall can only be observed for PROTEST-ER.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "BERT.",
"sec_num": null
},
{
"text": "With the exclusion of NEWS-BERT, the systems achieve satisfying results for the trigger component. Argument detection, as expected, is more challenging, with no model reaching an F1-score above 50%. PROTEST-ER always performs better, especially when processing the data at sentence level. In numerical terms, PROTEST-ER provides an average gain of 11.74 points. 4 We observe a relationship between argument type frequency in the training data and models's performance where the most frequent arguments, i.e., participant (26.43%), organizer (18.31%), and place (14.45%), obtain the best results. However, PROTEST-ER improves performances also on the least frequent argument types, i.e., loc (6.49%) and fname (5.85) of, respectively, 12.00 and 5.38 points on average, when compared to BERT. Table 4 . We applied only PROTEST-ER keeping the distinction between document vs. sentence input. Although using sentences as input leads to the best results for India, we also observe that the results of the document input models are competitive, leaving open questions whether such a way of processing the input could be an effective strategy for model portability for event extraction. The results clearly indicate that PROTEST-ER is a competitive and pretty robust system. Interestingly, we observe that on the China data, the best results are obtained when processing data at document level.",
"cite_spans": [
{
"start": 362,
"end": 363,
"text": "4",
"ref_id": null
}
],
"ref_spans": [
{
"start": 791,
"end": 798,
"text": "Table 4",
"ref_id": "TABREF6"
}
],
"eq_spans": [],
"section": "BERT.",
"sec_num": null
},
{
"text": "Looking at the portability for the event components, it clearly appears that arguments are more difficult than triggers. Indeed, the absolute F1score of the best models for triggers is in the same range of that for India. When focusing on the arguments, the drops in performances severely affect all argument types, except for fname. We also observe that the biggest drops are registered in those arguments that are most likely to express domain specific properties. For instance, the absolute F1score difference between the best models for India and China for place is 39.79 points, 36.29 for organizer, and 27.11 for etime. On the contrary, only a drop of 9.84 points is observed for participant, suggesting that ways of indicating those who take part to a protest event (e.g. protesters, or rioters) are closer than expected.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "China data Results for China are reported in",
"sec_num": null
},
{
"text": "Our results indicate that DAR is an effective strategy for unsupervised domain adaptation. However, we show that not every data distribution matching a potential target domain has the same impact. In our case, we measure improvements only when using data that more directly target the content of the task, i.e., protest events, possibly supplementing limitations in training materials. We have gathered interesting cues that processing data at document level can actually be an effective strategy also for a sequence labeling task with small training data. We think that this approach allows the TLMs to gain from processing longer sequences and acquire better knowledge. However, more experiments on different tasks (e.g., NER) and with different training sizes are needed to test this hypothesis.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion and Conclusions",
"sec_num": "5"
},
{
"text": "A further positive aspect of DAR is that it requires less training material to boost system's performance, pointing to new directions for few-shot learning. We projected the learning curves of BERT and PROTEST-ER using increasing steps of the training data. PROTEST-ER achieves an overall F1-score \u223c30% with only 10% of the training data, while BERT needs minimally 30% to achieve comparable performances (see Appendix A.3).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion and Conclusions",
"sec_num": "5"
},
{
"text": "Disappointingly, PROTEST-ER falls way back the best model that participated in Protest-News. Skitalinskaya et al. (2019) propose a Bi-LSTM-CRF architecture using FLAIR contextualized word embeddings (Akbik et al., 2018) . They also adopt a joint strategy for trigger and argument prediction. PROTEST-ER obtains a better Precision only on China for the overall evaluation and for trigger. Quite surprisingly, on India it is BERT that achieves better results on trigger, although the model appears to be quite unstable, as shown by the standard deviation. At this stage, it is still unclear whether these disappointing performances are due to the retraining (i.e., need to extend the number of documents used) or the small training corpus.",
"cite_spans": [
{
"start": 93,
"end": 120,
"text": "Skitalinskaya et al. (2019)",
"ref_id": "BIBREF26"
},
{
"start": 199,
"end": 219,
"text": "(Akbik et al., 2018)",
"ref_id": "BIBREF1"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion and Conclusions",
"sec_num": "5"
},
{
"text": "Future work will focus on two aspects. First, we will further investigate the impact of the size of the training data when using TLMs. This will require to experiment with different datasets and tasks. Secondly, we will explore solutions for multilingual extensions of PROTEST-ER.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion and Conclusions",
"sec_num": "5"
},
{
"text": "Preprocessing The unlabeled corpora of (protest related) news articles from the TREC Washington Post version 3 are minimally preprocessed prior to the language model retraining phase. We use the full text, including the title, of each news article. Document Creation Times are removed. We perform sentence splitting using spaCy (Honnibal et al., 2020) .",
"cite_spans": [
{
"start": 328,
"end": 351,
"text": "(Honnibal et al., 2020)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "A.1 BERT-NEWS/PROTEST-ER Further Training",
"sec_num": null
},
{
"text": "Training details We further train the English BERT base-uncased for 100 epochs. We use a batch size of 64 through gradient accumulation. Other hyperparameters are illustrated in Table 6 . Our TLM implementation uses the HuggingFace library (Wolf et al., 2020) . The pretrainig experiment was performed on a single Nvidia V100 GPU and took 8 days. A.2 BERT/PROTEST-ER Fine-tuning Table 7 shows the values of the hyperparameters used for fine-tuning BERT and PROTEST-ER. We used Tensorflow (Abadi et al., 2016) for the implementation and the Huggingface library (Wolf et al., 2020) for implementing the BERT embeddings and loading the data. We used the CRF implementation available from the Tensorflow Addons package. The models are trained for a maximum of 100 epochs, using a constant learning rate of 2e-5; if the validation loss does not improve for 5 consecutive epochs, training is stopped. The best model is selected on the basis of the validation loss. We manually experimented with the learning rates 1e-5, 2e-5, 3e-5. No other hyperparameter optimization was performed. We used the original train, validation, and test splits of the event extraction task of the 2019 CLEF ProtestNews Lab.",
"cite_spans": [
{
"start": 240,
"end": 259,
"text": "(Wolf et al., 2020)",
"ref_id": "BIBREF27"
},
{
"start": 488,
"end": 508,
"text": "(Abadi et al., 2016)",
"ref_id": "BIBREF0"
},
{
"start": 560,
"end": 579,
"text": "(Wolf et al., 2020)",
"ref_id": "BIBREF27"
}
],
"ref_spans": [
{
"start": 178,
"end": 185,
"text": "Table 6",
"ref_id": "TABREF8"
},
{
"start": 379,
"end": 386,
"text": "Table 7",
"ref_id": "TABREF10"
}
],
"eq_spans": [],
"section": "A.1 BERT-NEWS/PROTEST-ER Further Training",
"sec_num": null
},
{
"text": "We conducted all the experiments using the Google Colaboratory platform. The time required to run all the experiments on the free plan of Colaboratory is approximately 20 hours. Figure 1 graphically illustrates the base architecture. ",
"cite_spans": [],
"ref_spans": [
{
"start": 178,
"end": 186,
"text": "Figure 1",
"ref_id": "FIGREF0"
}
],
"eq_spans": [],
"section": "A.1 BERT-NEWS/PROTEST-ER Further Training",
"sec_num": null
},
{
"text": "In the following graphs we plot the learning curves of the BERT and PROTEST-ER model on the India and China dataset. In both cases, we observe that PROTEST-ER obtains competitive scores just using 10% of the training data, suggesting that the TLM's representations are already shifted towards the protest domain. To obtain the same results, the generic BERT models need minimally 30% of the training data, when using documents as input, and 70% of the training, when using sentences. ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A.3 BERT/PROTEST-ER Learning Curves",
"sec_num": null
},
{
"text": "https://emw.ku.edu.tr/ clef-protestnews-2019/",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "The training portion of ACE has 4,312 triggers and 7,811 arguments.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "This figure has been obtained by grouping the scores of all models using the retrained version, regardless of the input format.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [
{
"text": "The authors from Koc University were funded by the European Research Council (ERC) Starting Grant 714868 awarded to Dr. Erdem Y\u00f6r\u00fck for his project Emerging Welfare.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Acknowledgments",
"sec_num": null
}
],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "Tensorflow: A system for large-scale machine learning",
"authors": [
{
"first": "Mart\u00edn",
"middle": [],
"last": "Abadi",
"suffix": ""
},
{
"first": "Paul",
"middle": [],
"last": "Barham",
"suffix": ""
},
{
"first": "Jianmin",
"middle": [],
"last": "Chen",
"suffix": ""
},
{
"first": "Zhifeng",
"middle": [],
"last": "Chen",
"suffix": ""
},
{
"first": "Andy",
"middle": [],
"last": "Davis",
"suffix": ""
},
{
"first": "Jeffrey",
"middle": [],
"last": "Dean",
"suffix": ""
},
{
"first": "Matthieu",
"middle": [],
"last": "Devin",
"suffix": ""
},
{
"first": "Sanjay",
"middle": [],
"last": "Ghemawat",
"suffix": ""
},
{
"first": "Geoffrey",
"middle": [],
"last": "Irving",
"suffix": ""
},
{
"first": "Michael",
"middle": [],
"last": "Isard",
"suffix": ""
},
{
"first": "Manjunath",
"middle": [],
"last": "Kudlur",
"suffix": ""
},
{
"first": "Josh",
"middle": [],
"last": "Levenberg",
"suffix": ""
},
{
"first": "Rajat",
"middle": [],
"last": "Monga",
"suffix": ""
},
{
"first": "Sherry",
"middle": [],
"last": "Moore",
"suffix": ""
},
{
"first": "Derek",
"middle": [
"G"
],
"last": "Murray",
"suffix": ""
},
{
"first": "Benoit",
"middle": [],
"last": "Steiner",
"suffix": ""
},
{
"first": "Paul",
"middle": [],
"last": "Tucker",
"suffix": ""
},
{
"first": "Vijay",
"middle": [],
"last": "Vasudevan",
"suffix": ""
},
{
"first": "Pete",
"middle": [],
"last": "Warden",
"suffix": ""
},
{
"first": "Martin",
"middle": [],
"last": "Wicke",
"suffix": ""
},
{
"first": "Yuan",
"middle": [],
"last": "Yu",
"suffix": ""
},
{
"first": "Xiaoqiang",
"middle": [],
"last": "Zheng",
"suffix": ""
}
],
"year": 2016,
"venue": "12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16)",
"volume": "",
"issue": "",
"pages": "265--283",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Mart\u00edn Abadi, Paul Barham, Jianmin Chen, Zhifeng Chen, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Geoffrey Irving, Michael Isard, Manjunath Kudlur, Josh Levenberg, Rajat Monga, Sherry Moore, Derek G. Murray, Benoit Steiner, Paul Tucker, Vijay Vasudevan, Pete Warden, Martin Wicke, Yuan Yu, and Xiaoqiang Zheng. 2016. Ten- sorflow: A system for large-scale machine learning. In 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16), pages 265- 283, Savannah, GA. USENIX Association.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Contextual string embeddings for sequence labeling",
"authors": [
{
"first": "Alan",
"middle": [],
"last": "Akbik",
"suffix": ""
},
{
"first": "Duncan",
"middle": [],
"last": "Blythe",
"suffix": ""
},
{
"first": "Roland",
"middle": [],
"last": "Vollgraf",
"suffix": ""
}
],
"year": 2018,
"venue": "Proceedings of the 27th International Conference on Computational Linguistics",
"volume": "",
"issue": "",
"pages": "1638--1649",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Alan Akbik, Duncan Blythe, and Roland Vollgraf. 2018. Contextual string embeddings for sequence labeling. In Proceedings of the 27th International Conference on Computational Linguistics, pages 1638-1649, Santa Fe, New Mexico, USA. Associ- ation for Computational Linguistics.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "Domain adaptation with adversarial training and graph embeddings",
"authors": [
{
"first": "Firoj",
"middle": [],
"last": "Alam",
"suffix": ""
},
{
"first": "Shafiq",
"middle": [],
"last": "Joty",
"suffix": ""
},
{
"first": "Muhammad",
"middle": [],
"last": "Imran",
"suffix": ""
}
],
"year": 2018,
"venue": "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics",
"volume": "1",
"issue": "",
"pages": "1077--1087",
"other_ids": {
"DOI": [
"10.18653/v1/P18-1099"
]
},
"num": null,
"urls": [],
"raw_text": "Firoj Alam, Shafiq Joty, and Muhammad Imran. 2018. Domain adaptation with adversarial training and graph embeddings. In Proceedings of the 56th An- nual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1077- 1087, Melbourne, Australia. Association for Compu- tational Linguistics.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Domain adaptation via pseudo in-domain data selection",
"authors": [
{
"first": "Amittai",
"middle": [],
"last": "Axelrod",
"suffix": ""
},
{
"first": "Xiaodong",
"middle": [],
"last": "He",
"suffix": ""
},
{
"first": "Jianfeng",
"middle": [],
"last": "Gao",
"suffix": ""
}
],
"year": 2011,
"venue": "Proceedings of the 2011 Conference on Empirical Methods in Natural Language Processing",
"volume": "",
"issue": "",
"pages": "355--362",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Amittai Axelrod, Xiaodong He, and Jianfeng Gao. 2011. Domain adaptation via pseudo in-domain data selection. In Proceedings of the 2011 Conference on Empirical Methods in Natural Language Processing, pages 355-362, Edinburgh, Scotland, UK. Associa- tion for Computational Linguistics.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "PERL: Pivot-based domain adaptation for pre-trained deep contextualized embedding models",
"authors": [
{
"first": "Eyal",
"middle": [],
"last": "Ben-David",
"suffix": ""
},
{
"first": "Carmel",
"middle": [],
"last": "Rabinovitz",
"suffix": ""
},
{
"first": "Roi",
"middle": [],
"last": "Reichart",
"suffix": ""
}
],
"year": 2020,
"venue": "Transactions of the Association for Computational Linguistics",
"volume": "8",
"issue": "",
"pages": "504--521",
"other_ids": {
"DOI": [
"10.1162/tacl_a_00328"
]
},
"num": null,
"urls": [],
"raw_text": "Eyal Ben-David, Carmel Rabinovitz, and Roi Reichart. 2020. PERL: Pivot-based domain adaptation for pre-trained deep contextualized embedding models. Transactions of the Association for Computational Linguistics, 8:504-521.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "Domain adaptation with structural correspondence learning",
"authors": [
{
"first": "John",
"middle": [],
"last": "Blitzer",
"suffix": ""
},
{
"first": "Ryan",
"middle": [],
"last": "Mcdonald",
"suffix": ""
},
{
"first": "Fernando",
"middle": [],
"last": "Pereira",
"suffix": ""
}
],
"year": 2006,
"venue": "Proceedings of the 2006 conference on empirical methods in natural language processing",
"volume": "",
"issue": "",
"pages": "120--128",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "John Blitzer, Ryan McDonald, and Fernando Pereira. 2006. Domain adaptation with structural correspon- dence learning. In Proceedings of the 2006 confer- ence on empirical methods in natural language pro- cessing, pages 120-128. Association for Computa- tional Linguistics.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "Frustratingly easy domain adaptation",
"authors": [
{
"first": "Hal",
"middle": [],
"last": "Daum\u00e9",
"suffix": ""
},
{
"first": "Iii",
"middle": [],
"last": "",
"suffix": ""
}
],
"year": 2007,
"venue": "Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics",
"volume": "",
"issue": "",
"pages": "256--263",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Hal Daum\u00e9 III. 2007. Frustratingly easy domain adap- tation. In Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics, pages 256-263, Prague, Czech Republic. Association for Computational Linguistics.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "BERT: Pre-training of deep bidirectional transformers for language understanding",
"authors": [
{
"first": "Jacob",
"middle": [],
"last": "Devlin",
"suffix": ""
},
{
"first": "Ming-Wei",
"middle": [],
"last": "Chang",
"suffix": ""
},
{
"first": "Kenton",
"middle": [],
"last": "Lee",
"suffix": ""
},
{
"first": "Kristina",
"middle": [],
"last": "Toutanova",
"suffix": ""
}
],
"year": 2019,
"venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies",
"volume": "1",
"issue": "",
"pages": "4171--4186",
"other_ids": {
"DOI": [
"10.18653/v1/N19-1423"
]
},
"num": null,
"urls": [],
"raw_text": "Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language under- standing. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171-4186, Minneapolis, Minnesota. Associ- ation for Computational Linguistics.",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "Unsupervised domain adaptation by backpropagation",
"authors": [
{
"first": "Yaroslav",
"middle": [],
"last": "Ganin",
"suffix": ""
},
{
"first": "Victor",
"middle": [],
"last": "Lempitsky",
"suffix": ""
}
],
"year": 2015,
"venue": "International conference on machine learning",
"volume": "",
"issue": "",
"pages": "1180--1189",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Yaroslav Ganin and Victor Lempitsky. 2015. Unsuper- vised domain adaptation by backpropagation. In In- ternational conference on machine learning, pages 1180-1189. PMLR.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "Domain-adversarial training of neural networks",
"authors": [
{
"first": "Yaroslav",
"middle": [],
"last": "Ganin",
"suffix": ""
},
{
"first": "Evgeniya",
"middle": [],
"last": "Ustinova",
"suffix": ""
},
{
"first": "Hana",
"middle": [],
"last": "Ajakan",
"suffix": ""
},
{
"first": "Pascal",
"middle": [],
"last": "Germain",
"suffix": ""
},
{
"first": "Hugo",
"middle": [],
"last": "Larochelle",
"suffix": ""
},
{
"first": "Fran\u00e7ois",
"middle": [],
"last": "Laviolette",
"suffix": ""
},
{
"first": "Mario",
"middle": [],
"last": "Marchand",
"suffix": ""
},
{
"first": "Victor",
"middle": [],
"last": "Lempitsky",
"suffix": ""
}
],
"year": 2016,
"venue": "The Journal of Machine Learning Research",
"volume": "17",
"issue": "1",
"pages": "2096--2030",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Yaroslav Ganin, Evgeniya Ustinova, Hana Ajakan, Pascal Germain, Hugo Larochelle, Fran\u00e7ois Lavi- olette, Mario Marchand, and Victor Lempitsky. 2016. Domain-adversarial training of neural net- works. The Journal of Machine Learning Research, 17(1):2096-2030.",
"links": null
},
"BIBREF10": {
"ref_id": "b10",
"title": "Multi-source domain adaptation with mixture of experts",
"authors": [
{
"first": "Jiang",
"middle": [],
"last": "Guo",
"suffix": ""
},
{
"first": "Darsh",
"middle": [],
"last": "Shah",
"suffix": ""
},
{
"first": "Regina",
"middle": [],
"last": "Barzilay",
"suffix": ""
}
],
"year": 2018,
"venue": "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing",
"volume": "",
"issue": "",
"pages": "4694--4703",
"other_ids": {
"DOI": [
"10.18653/v1/D18-1498"
]
},
"num": null,
"urls": [],
"raw_text": "Jiang Guo, Darsh Shah, and Regina Barzilay. 2018. Multi-source domain adaptation with mixture of ex- perts. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 4694-4703, Brussels, Belgium. Association for Computational Linguistics.",
"links": null
},
"BIBREF11": {
"ref_id": "b11",
"title": "Don't stop pretraining: Adapt language models to domains and tasks",
"authors": [
{
"first": "Ana",
"middle": [],
"last": "Suchin Gururangan",
"suffix": ""
},
{
"first": "Swabha",
"middle": [],
"last": "Marasovi\u0107",
"suffix": ""
},
{
"first": "Kyle",
"middle": [],
"last": "Swayamdipta",
"suffix": ""
},
{
"first": "Iz",
"middle": [],
"last": "Lo",
"suffix": ""
},
{
"first": "Doug",
"middle": [],
"last": "Beltagy",
"suffix": ""
},
{
"first": "Noah",
"middle": [
"A"
],
"last": "Downey",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Smith",
"suffix": ""
}
],
"year": 2020,
"venue": "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics",
"volume": "",
"issue": "",
"pages": "8342--8360",
"other_ids": {
"DOI": [
"10.18653/v1/2020.acl-main.740"
]
},
"num": null,
"urls": [],
"raw_text": "Suchin Gururangan, Ana Marasovi\u0107, Swabha Swayamdipta, Kyle Lo, Iz Beltagy, Doug Downey, and Noah A. Smith. 2020. Don't stop pretraining: Adapt language models to domains and tasks. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 8342-8360, Online. Association for Computational Linguistics.",
"links": null
},
"BIBREF12": {
"ref_id": "b12",
"title": "Unsupervised domain adaptation of contextualized embeddings for sequence labeling",
"authors": [
{
"first": "Xiaochuang",
"middle": [],
"last": "Han",
"suffix": ""
},
{
"first": "Jacob",
"middle": [],
"last": "Eisenstein",
"suffix": ""
}
],
"year": 2019,
"venue": "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)",
"volume": "",
"issue": "",
"pages": "4238--4248",
"other_ids": {
"DOI": [
"10.18653/v1/D19-1433"
]
},
"num": null,
"urls": [],
"raw_text": "Xiaochuang Han and Jacob Eisenstein. 2019. Unsu- pervised domain adaptation of contextualized em- beddings for sequence labeling. In Proceedings of the 2019 Conference on Empirical Methods in Nat- ural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 4238-4248, Hong Kong, China. Association for Computational Linguistics.",
"links": null
},
"BIBREF13": {
"ref_id": "b13",
"title": "spaCy: Industrial-strength Natural Language Processing in Python",
"authors": [],
"year": null,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {
"DOI": [
"10.5281/zenodo.1212303"
]
},
"num": null,
"urls": [],
"raw_text": "Matthew Honnibal, Ines Montani, Sofie Van Lan- deghem, and Adriane Boyd. 2020. spaCy: Industrial-strength Natural Language Processing in Python.",
"links": null
},
"BIBREF14": {
"ref_id": "b14",
"title": "Overview of clef 2019 lab protestnews: Extracting protests from news in a cross-context setting",
"authors": [
{
"first": "Ali",
"middle": [],
"last": "H\u00fcrriyetoglu",
"suffix": ""
},
{
"first": "Erdem",
"middle": [],
"last": "Y\u00f6r\u00fck",
"suffix": ""
},
{
"first": "Deniz",
"middle": [],
"last": "Y\u00fcret",
"suffix": ""
},
{
"first": "Burak",
"middle": [],
"last": "Agr\u0131 Yoltar",
"suffix": ""
},
{
"first": "F\u0131rat",
"middle": [],
"last": "G\u00fcrel",
"suffix": ""
},
{
"first": "Osman",
"middle": [],
"last": "Duru\u015fan",
"suffix": ""
},
{
"first": "Arda",
"middle": [],
"last": "Mutlu",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Akdemir",
"suffix": ""
}
],
"year": 2019,
"venue": "Experimental IR Meets Multilinguality, Multimodality, and Interaction",
"volume": "",
"issue": "",
"pages": "425--432",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Ali H\u00fcrriyetoglu, Erdem Y\u00f6r\u00fck, Deniz Y\u00fcret, \u00c7 agr\u0131 Yoltar, Burak G\u00fcrel, F\u0131rat Duru\u015fan, Osman Mutlu, and Arda Akdemir. 2019. Overview of clef 2019 lab protestnews: Extracting protests from news in a cross-context setting. In Experimental IR Meets Multilinguality, Multimodality, and Interac- tion, pages 425-432, Cham. Springer International Publishing.",
"links": null
},
"BIBREF15": {
"ref_id": "b15",
"title": "Cross-Context News Corpus for Protest Event-Related Knowledge Base Construction",
"authors": [
{
"first": "Ali",
"middle": [],
"last": "H\u00fcrriyetoglu",
"suffix": ""
},
{
"first": "Erdem",
"middle": [],
"last": "Y\u00f6r\u00fck",
"suffix": ""
},
{
"first": "Osman",
"middle": [],
"last": "Mutlu",
"suffix": ""
},
{
"first": "F\u0131rat",
"middle": [],
"last": "Duru\u015fan",
"suffix": ""
},
{
"first": "Deniz",
"middle": [],
"last": "Agr\u0131 Yoltar",
"suffix": ""
},
{
"first": "Burak",
"middle": [],
"last": "Y\u00fcret",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "G\u00fcrel",
"suffix": ""
}
],
"year": 2021,
"venue": "Data Intelligence",
"volume": "",
"issue": "",
"pages": "1--28",
"other_ids": {
"DOI": [
"10.1162/dint_a_00092"
]
},
"num": null,
"urls": [],
"raw_text": "Ali H\u00fcrriyetoglu, Erdem Y\u00f6r\u00fck, Osman Mutlu, F\u0131rat Duru\u015fan, \u00c7 agr\u0131 Yoltar, Deniz Y\u00fcret, and Burak G\u00fcrel. 2021. Cross-Context News Corpus for Protest Event-Related Knowledge Base Construc- tion. Data Intelligence, pages 1-28.",
"links": null
},
"BIBREF16": {
"ref_id": "b16",
"title": "ACE (Automatic Content Extraction) English Annotation Guidelines for Events",
"authors": [],
"year": 2005,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Linguistic Data Consortium. 2005. ACE (Automatic Content Extraction) English Annotation Guidelines for Events, 5.4.3 2005.07.01 edition.",
"links": null
},
"BIBREF17": {
"ref_id": "b17",
"title": "Reranking and self-training for parser adaptation",
"authors": [
{
"first": "David",
"middle": [],
"last": "Mcclosky",
"suffix": ""
},
{
"first": "Eugene",
"middle": [],
"last": "Charniak",
"suffix": ""
},
{
"first": "Mark",
"middle": [],
"last": "Johnson",
"suffix": ""
}
],
"year": 2006,
"venue": "Proceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics",
"volume": "",
"issue": "",
"pages": "337--344",
"other_ids": {
"DOI": [
"10.3115/1220175.1220218"
]
},
"num": null,
"urls": [],
"raw_text": "David McClosky, Eugene Charniak, and Mark Johnson. 2006. Reranking and self-training for parser adapta- tion. In Proceedings of the 21st International Con- ference on Computational Linguistics and 44th An- nual Meeting of the Association for Computational Linguistics, pages 337-344, Sydney, Australia. As- sociation for Computational Linguistics.",
"links": null
},
"BIBREF18": {
"ref_id": "b18",
"title": "Simplified neural unsupervised domain adaptation",
"authors": [
{
"first": "Timothy",
"middle": [],
"last": "Miller",
"suffix": ""
}
],
"year": 2019,
"venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies",
"volume": "1",
"issue": "",
"pages": "414--419",
"other_ids": {
"DOI": [
"10.18653/v1/N19-1039"
]
},
"num": null,
"urls": [],
"raw_text": "Timothy Miller. 2019. Simplified neural unsupervised domain adaptation. In Proceedings of the 2019 Con- ference of the North American Chapter of the Asso- ciation for Computational Linguistics: Human Lan- guage Technologies, Volume 1 (Long and Short Pa- pers), pages 414-419, Minneapolis, Minnesota. As- sociation for Computational Linguistics.",
"links": null
},
"BIBREF19": {
"ref_id": "b19",
"title": "Intelligent selection of language model training data",
"authors": [
{
"first": "C",
"middle": [],
"last": "Robert",
"suffix": ""
},
{
"first": "William",
"middle": [],
"last": "Moore",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Lewis",
"suffix": ""
}
],
"year": 2010,
"venue": "Proceedings of the ACL 2010 Conference Short Papers",
"volume": "",
"issue": "",
"pages": "220--224",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Robert C. Moore and William Lewis. 2010. Intelligent selection of language model training data. In Pro- ceedings of the ACL 2010 Conference Short Papers, pages 220-224, Uppsala, Sweden. Association for Computational Linguistics.",
"links": null
},
"BIBREF20": {
"ref_id": "b20",
"title": "Joint event extraction via recurrent neural networks",
"authors": [
{
"first": "Kyunghyun",
"middle": [],
"last": "Thien Huu Nguyen",
"suffix": ""
},
{
"first": "Ralph",
"middle": [],
"last": "Cho",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Grishman",
"suffix": ""
}
],
"year": 2016,
"venue": "Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies",
"volume": "",
"issue": "",
"pages": "300--309",
"other_ids": {
"DOI": [
"10.18653/v1/N16-1034"
]
},
"num": null,
"urls": [],
"raw_text": "Thien Huu Nguyen, Kyunghyun Cho, and Ralph Gr- ishman. 2016. Joint event extraction via recurrent neural networks. In Proceedings of the 2016 Con- ference of the North American Chapter of the As- sociation for Computational Linguistics: Human Language Technologies, pages 300-309, San Diego, California. Association for Computational Linguis- tics.",
"links": null
},
"BIBREF21": {
"ref_id": "b21",
"title": "Unsupervised domain adaptation of language models for reading comprehension",
"authors": [
{
"first": "Kosuke",
"middle": [],
"last": "Nishida",
"suffix": ""
},
{
"first": "Kyosuke",
"middle": [],
"last": "Nishida",
"suffix": ""
},
{
"first": "Itsumi",
"middle": [],
"last": "Saito",
"suffix": ""
},
{
"first": "Hisako",
"middle": [],
"last": "Asano",
"suffix": ""
},
{
"first": "Junji",
"middle": [],
"last": "Tomita",
"suffix": ""
}
],
"year": 2020,
"venue": "Proceedings of the 12th Language Resources and Evaluation Conference",
"volume": "",
"issue": "",
"pages": "5392--5399",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Kosuke Nishida, Kyosuke Nishida, Itsumi Saito, Hisako Asano, and Junji Tomita. 2020. Unsuper- vised domain adaptation of language models for reading comprehension. In Proceedings of the 12th Language Resources and Evaluation Conference, pages 5392-5399, Marseille, France. European Lan- guage Resources Association.",
"links": null
},
"BIBREF22": {
"ref_id": "b22",
"title": "Effective measures of domain similarity for parsing",
"authors": [
{
"first": "Barbara",
"middle": [],
"last": "Plank",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Gertjan Van Noord",
"suffix": ""
}
],
"year": 2011,
"venue": "Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies",
"volume": "",
"issue": "",
"pages": "1566--1576",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Barbara Plank and Gertjan van Noord. 2011. Effec- tive measures of domain similarity for parsing. In Proceedings of the 49th Annual Meeting of the As- sociation for Computational Linguistics: Human Language Technologies, pages 1566-1576, Portland, Oregon, USA. Association for Computational Lin- guistics.",
"links": null
},
"BIBREF23": {
"ref_id": "b23",
"title": "Neural unsupervised domain adaptation in NLP-A survey",
"authors": [
{
"first": "Alan",
"middle": [],
"last": "Ramponi",
"suffix": ""
},
{
"first": "Barbara",
"middle": [],
"last": "Plank",
"suffix": ""
}
],
"year": 2020,
"venue": "Proceedings of the 28th International Conference on Computational Linguistics",
"volume": "",
"issue": "",
"pages": "6838--6855",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Alan Ramponi and Barbara Plank. 2020. Neural unsu- pervised domain adaptation in NLP-A survey. In Proceedings of the 28th International Conference on Computational Linguistics, pages 6838-6855, Barcelona, Spain (Online). International Committee on Computational Linguistics.",
"links": null
},
"BIBREF24": {
"ref_id": "b24",
"title": "Adapt or get left behind: Domain adaptation through BERT language model finetuning for aspect-target sentiment classification",
"authors": [
{
"first": "Alexander",
"middle": [],
"last": "Rietzler",
"suffix": ""
},
{
"first": "Sebastian",
"middle": [],
"last": "Stabinger",
"suffix": ""
},
{
"first": "Paul",
"middle": [],
"last": "Opitz",
"suffix": ""
},
{
"first": "Stefan",
"middle": [],
"last": "Engl",
"suffix": ""
}
],
"year": 2020,
"venue": "Proceedings of the 12th Language Resources and Evaluation Conference",
"volume": "",
"issue": "",
"pages": "4933--4941",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Alexander Rietzler, Sebastian Stabinger, Paul Opitz, and Stefan Engl. 2020. Adapt or get left behind: Domain adaptation through BERT language model finetuning for aspect-target sentiment classification. In Proceedings of the 12th Language Resources and Evaluation Conference, pages 4933-4941, Mar- seille, France. European Language Resources Asso- ciation.",
"links": null
},
"BIBREF25": {
"ref_id": "b25",
"title": "Learning to select data for transfer learning with bayesian optimization",
"authors": [
{
"first": "Sebastian",
"middle": [],
"last": "Ruder",
"suffix": ""
},
{
"first": "Barbara",
"middle": [],
"last": "Plank",
"suffix": ""
}
],
"year": 2017,
"venue": "Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing",
"volume": "",
"issue": "",
"pages": "372--382",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Sebastian Ruder and Barbara Plank. 2017. Learning to select data for transfer learning with bayesian opti- mization. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pages 372-382, Copenhagen, Denmark. Association for Computational Linguistics.",
"links": null
},
"BIBREF26": {
"ref_id": "b26",
"title": "Clef protestnews lab 2019: Contextualized word embeddings for event sentence detection and event extraction",
"authors": [
{
"first": "Gabriella",
"middle": [],
"last": "Skitalinskaya",
"suffix": ""
},
{
"first": "Jonas",
"middle": [],
"last": "Klaff",
"suffix": ""
},
{
"first": "Maximilian",
"middle": [],
"last": "Splieth\u00f6ver",
"suffix": ""
}
],
"year": 2019,
"venue": "CLEF (Working Notes)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Gabriella Skitalinskaya, Jonas Klaff, and Maximilian Splieth\u00f6ver. 2019. Clef protestnews lab 2019: Con- textualized word embeddings for event sentence de- tection and event extraction. In CLEF (Working Notes).",
"links": null
},
"BIBREF27": {
"ref_id": "b27",
"title": "Transformers: State-of-the-art natural language processing",
"authors": [
{
"first": "Thomas",
"middle": [],
"last": "Wolf",
"suffix": ""
},
{
"first": "Lysandre",
"middle": [],
"last": "Debut",
"suffix": ""
},
{
"first": "Victor",
"middle": [],
"last": "Sanh",
"suffix": ""
},
{
"first": "Julien",
"middle": [],
"last": "Chaumond",
"suffix": ""
},
{
"first": "Clement",
"middle": [],
"last": "Delangue",
"suffix": ""
},
{
"first": "Anthony",
"middle": [],
"last": "Moi",
"suffix": ""
},
{
"first": "Pierric",
"middle": [],
"last": "Cistac",
"suffix": ""
},
{
"first": "Tim",
"middle": [],
"last": "Rault",
"suffix": ""
},
{
"first": "Remi",
"middle": [],
"last": "Louf",
"suffix": ""
},
{
"first": "Morgan",
"middle": [],
"last": "Funtowicz",
"suffix": ""
},
{
"first": "Joe",
"middle": [],
"last": "Davison",
"suffix": ""
},
{
"first": "Sam",
"middle": [],
"last": "Shleifer",
"suffix": ""
},
{
"first": "Clara",
"middle": [],
"last": "Patrick Von Platen",
"suffix": ""
},
{
"first": "Yacine",
"middle": [],
"last": "Ma",
"suffix": ""
},
{
"first": "Julien",
"middle": [],
"last": "Jernite",
"suffix": ""
},
{
"first": "Canwen",
"middle": [],
"last": "Plu",
"suffix": ""
},
{
"first": "Teven",
"middle": [
"Le"
],
"last": "Xu",
"suffix": ""
},
{
"first": "Sylvain",
"middle": [],
"last": "Scao",
"suffix": ""
},
{
"first": "Mariama",
"middle": [],
"last": "Gugger",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Drame",
"suffix": ""
}
],
"year": 2020,
"venue": "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations",
"volume": "",
"issue": "",
"pages": "38--45",
"other_ids": {
"DOI": [
"10.18653/v1/2020.emnlp-demos.6"
]
},
"num": null,
"urls": [],
"raw_text": "Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pier- ric Cistac, Tim Rault, Remi Louf, Morgan Funtow- icz, Joe Davison, Sam Shleifer, Patrick von Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander Rush. 2020. Trans- formers: State-of-the-art natural language process- ing. In Proceedings of the 2020 Conference on Em- pirical Methods in Natural Language Processing: System Demonstrations, pages 38-45, Online. Asso- ciation for Computational Linguistics.",
"links": null
},
"BIBREF28": {
"ref_id": "b28",
"title": "Learning semantic representations for unsupervised domain adaptation",
"authors": [
{
"first": "Shaoan",
"middle": [],
"last": "Xie",
"suffix": ""
},
{
"first": "Zibin",
"middle": [],
"last": "Zheng",
"suffix": ""
},
{
"first": "Liang",
"middle": [],
"last": "Chen",
"suffix": ""
},
{
"first": "Chuan",
"middle": [],
"last": "Chen",
"suffix": ""
}
],
"year": 2018,
"venue": "International Conference on Machine Learning",
"volume": "",
"issue": "",
"pages": "5423--5432",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Shaoan Xie, Zibin Zheng, Liang Chen, and Chuan Chen. 2018. Learning semantic representations for unsupervised domain adaptation. In International Conference on Machine Learning, pages 5423- 5432.",
"links": null
},
"BIBREF29": {
"ref_id": "b29",
"title": "BERT post-training for review reading comprehension and aspect-based sentiment analysis",
"authors": [
{
"first": "Hu",
"middle": [],
"last": "Xu",
"suffix": ""
},
{
"first": "Bing",
"middle": [],
"last": "Liu",
"suffix": ""
},
{
"first": "Lei",
"middle": [],
"last": "Shu",
"suffix": ""
},
{
"first": "Philip",
"middle": [],
"last": "Yu",
"suffix": ""
}
],
"year": 2019,
"venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies",
"volume": "1",
"issue": "",
"pages": "2324--2335",
"other_ids": {
"DOI": [
"10.18653/v1/N19-1242"
]
},
"num": null,
"urls": [],
"raw_text": "Hu Xu, Bing Liu, Lei Shu, and Philip Yu. 2019. BERT post-training for review reading comprehension and aspect-based sentiment analysis. In Proceedings of the 2019 Conference of the North American Chap- ter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 2324-2335, Minneapolis, Minnesota. Association for Computational Linguis- tics.",
"links": null
},
"BIBREF30": {
"ref_id": "b30",
"title": "Multi-source domain adaptation for semantic segmentation",
"authors": [
{
"first": "Sicheng",
"middle": [],
"last": "Zhao",
"suffix": ""
},
{
"first": "Bo",
"middle": [],
"last": "Li",
"suffix": ""
},
{
"first": "Xiangyu",
"middle": [],
"last": "Yue",
"suffix": ""
},
{
"first": "Yang",
"middle": [],
"last": "Gu",
"suffix": ""
},
{
"first": "Pengfei",
"middle": [],
"last": "Xu",
"suffix": ""
},
{
"first": "Runbo",
"middle": [],
"last": "Hu",
"suffix": ""
},
{
"first": "Hua",
"middle": [],
"last": "Chai",
"suffix": ""
},
{
"first": "Kurt",
"middle": [],
"last": "Keutzer",
"suffix": ""
}
],
"year": 2019,
"venue": "Advances in Neural Information Processing Systems",
"volume": "32",
"issue": "",
"pages": "7287--7300",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Sicheng Zhao, Bo Li, Xiangyu Yue, Yang Gu, Pengfei Xu, Runbo Hu, Hua Chai, and Kurt Keutzer. 2019. Multi-source domain adaptation for semantic seg- mentation. In Advances in Neural Information Pro- cessing Systems, volume 32, pages 7287-7300. Cur- ran Associates, Inc.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"num": null,
"uris": null,
"type_str": "figure",
"text": "The base model architecture for the token classifier."
},
"FIGREF1": {
"num": null,
"uris": null,
"type_str": "figure",
"text": "Learning curve for event extraction (triggers and arguments) for BERT and PROTEST-ER models on India and China, according to different portions (percentages) of the training materials (input granularity: sentence). Input data are randomly selected."
},
"FIGREF2": {
"num": null,
"uris": null,
"type_str": "figure",
"text": "Learning curve for event extraction (triggers and arguments) for BERT and PROTEST-ER models on India and China, according to different portions (percentages) of the training materials (input granularity: document). Input data are randomly selected."
},
"TABREF1": {
"type_str": "table",
"num": null,
"text": "Distribution of event triggers and arguments. India is source. China is target.",
"content": "<table/>",
"html": null
},
"TABREF3": {
"type_str": "table",
"num": null,
"text": "",
"content": "<table/>",
"html": null
},
"TABREF4": {
"type_str": "table",
"num": null,
"text": "4.20 42.68 4.98 46.23 1.98 78.97 4.32 63.72 4.76 70.25 1.87 31.50 17.54 29.61 16.09 29.94 16.49 NEWS-BERT Document 36.11 3.77 33.63 7.79 34.18 3.48 69.96 5.18 52.00 10.32 58.87 5.41 22.61 4.69 20.96 9.62 19.96 6.95 PROTEST-ER Document 54.56 3.18 48.47 3.69 51.11 0.87 70.48 1.35 67.90 3.51 69.08 1.24 37.59 20.28 40.20 17.91 37.86 18.42 BERT Sentence 32.85 6.27 25.18 6.61 27.41 4.19 80.01 5.98 29.30 13.03 41.16 12.81 18.95 15.46 22.79 17.38 19.74 15.43 NEWS-BERT Sentence 52.86 8.83 10.76 1.94 17.67 2.32 92.92 1.84 9.83 3.08 18.24 5.90 29.47 6.16 10.15 1.12 14.46 0.85 PROTEST-ER Sentence 49.91 1.99 54.13 0.63 51.91 0.97 77.63 1.41 68.93 1.75 72.99 0.80 39.82 17.61 46.13 17.86 41.98 17.26",
"content": "<table><tr><td>Model</td><td>Input Format</td><td>Overall</td><td/><td/><td>Triggers</td><td/><td/><td>Arguments</td></tr><tr><td/><td>P</td><td>R</td><td>F1</td><td>P</td><td>R</td><td>F1</td><td>P</td><td>R</td><td>F1</td></tr><tr><td colspan=\"2\">BERT 51.52 Best CLEF 2019 Sentence Document 66.20</td><td>55.67</td><td>60.48</td><td>79.79</td><td>69.77</td><td>74.44</td><td>56.55</td><td>48.66</td><td>51.54</td></tr></table>",
"html": null
},
"TABREF5": {
"type_str": "table",
"num": null,
"text": "India data (source). Results for TLM are averaged over five runs. Standard deviation is reported in subscript. Best results correspond to the best system in the 2019 CLEF ProtestNews Lab tasks. Best scores are in bold. Second best scores are in italics. .01 36.53 2.76 46.39 1.02 74.07 4.74 69.30 5.66 71.23 1.05 42.70 18.68 20.11 14.83 25.19 14.71 PROTEST-ER Sentence 52.62 5.34 39.18 3.25 44.62 1.97 74.08 3.20 64.86 7.44 68.73 2.75 39.06 16.03 23.56 11.99 27.02 11.81",
"content": "<table><tr><td>Model</td><td>Input Format</td><td>Overall</td><td/><td/><td>Triggers</td><td/><td/><td>Arguments</td></tr><tr><td/><td>P</td><td>R</td><td>F1</td><td>P</td><td>R</td><td>F1</td><td>P</td><td>R</td><td>F1</td></tr><tr><td colspan=\"2\">PROTEST-ER 64.48 5Best CLEF 2019 Sentence Document 62.65</td><td>46.24</td><td>53.21</td><td>77.27</td><td>70.83</td><td>73.91</td><td>49.64</td><td>33.57</td><td>39.56</td></tr></table>",
"html": null
},
"TABREF6": {
"type_str": "table",
"num": null,
"text": "",
"content": "<table><tr><td/><td>J-S</td><td/><td>OOV</td></tr><tr><td colspan=\"3\">\u2193DAR / Test\u2192 India China</td><td>India China</td></tr><tr><td>WPC-Gen</td><td>0.583</td><td colspan=\"2\">0.594 12.17% 4.38%</td></tr><tr><td>WPC-Ev</td><td>0.562</td><td colspan=\"2\">0.569 11.61% 4.46%</td></tr></table>",
"html": null
},
"TABREF8": {
"type_str": "table",
"num": null,
"text": "Hyperparameter configuration used for generating PROTEST-ER.",
"content": "<table/>",
"html": null
},
"TABREF10": {
"type_str": "table",
"num": null,
"text": "Hyperparameter configuration used for task finetuning.",
"content": "<table/>",
"html": null
}
}
}
}