| { |
| "paper_id": "2021", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T13:14:01.559497Z" |
| }, |
| "title": "CASE 2021 Task 2: Socio-political Fine-grained Event Classification using Fine-tuned RoBERTa Document Embeddings", |
| "authors": [ |
| { |
| "first": "Samantha", |
| "middle": [], |
| "last": "Kent", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "samantha.kent@fkie.fraunhofer.de" |
| }, |
| { |
| "first": "Theresa", |
| "middle": [], |
| "last": "Krumbiegel", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "theresa.krumbiegel@fkie.fraunhofer.de" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "We present our submission to Task 2 of the Socio-political and Crisis Events Detection Shared Task at the CASE @ ACL-IJCNLP 2021 workshop. The task at hand aims at the fine-grained classification of socio-political events. Our best model was a fine-tuned RoBERTa transformer model using document embeddings. The corpus consisted of a balanced selection of sub-events extracted from the ACLED event dataset. We achieved a macro F-score of 0.923 and a micro F-score of 0.932 during our preliminary experiments on a held-out test set. The same model also performed best on the shared task test data (weighted F-score = 0.83). To analyze the results we calculated the topic compactness of the commonly misclassified events and conducted an error analysis.", |
| "pdf_parse": { |
| "paper_id": "2021", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "We present our submission to Task 2 of the Socio-political and Crisis Events Detection Shared Task at the CASE @ ACL-IJCNLP 2021 workshop. The task at hand aims at the fine-grained classification of socio-political events. Our best model was a fine-tuned RoBERTa transformer model using document embeddings. The corpus consisted of a balanced selection of sub-events extracted from the ACLED event dataset. We achieved a macro F-score of 0.923 and a micro F-score of 0.932 during our preliminary experiments on a held-out test set. The same model also performed best on the shared task test data (weighted F-score = 0.83). To analyze the results we calculated the topic compactness of the commonly misclassified events and conducted an error analysis.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Event detection and classification as Natural Language Processing (NLP) tasks can be used to analyze data gathered in the information space. The findings of this analysis can then be connected to events in the physical world and contribute to situational awareness, particularly when they are related to socio-political events. The sheer amount of data that is generated and stored in the information space every day, means that strategies need to be developed to be able to efficiently and effectively process this data. Given the large amounts of data, deep learning strategies are often preferred. However, time and computational constraints may play a role in deciding how to extract and analyze data.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Task 2 in the Socio-political and Crisis Events Detection Shared Task at the CASE @ ACL-IJCNLP 2021 workshop aims at the fine-grained classification of events (Haneczok et al., 2021) . The task is based on data extracted from the Armed Conflict Location & Event Data (ACLED) database (Raleigh et al., 2010) . It consist of socio-political events that have been annotated based on the ACLED event taxonomy, and includes 6 event types and 25 event subtypes. The aim of this task is to label event snippets using a model trained on data from the ACLED dataset, in order to see how robust models are when presented with data that is not directly covered by ACLED or contains unseen event classes. The results presented in this paper pertain only to subtask 1, where the task is the classification of 25 different event subtypes with ACLED-compliant labels. In other words, all the classes are seen classes from the ACLED dataset. The second and third subtasks are zero-short learning tasks that contain unseen classes.", |
| "cite_spans": [ |
| { |
| "start": 159, |
| "end": 182, |
| "text": "(Haneczok et al., 2021)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 284, |
| "end": 306, |
| "text": "(Raleigh et al., 2010)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "This paper proceeds by first describing the data collection process in section 3. Section 4 contains the system description and the following section contains the experimental results. Section 6 provides an overview of the results based on the test data provided by the organizers. Finally, in section 7 the error analysis provides an insight into the system results and the data.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Previous research in event detection and classification shows that there are numerous approaches to solve the problem of detecting events in texts. Xiang and Wang (2019) give a coherent overview of suitable strategies, starting with earlier approaches like pattern matching, and describing methods of machine learning as well as deep learning. There have been a number of shared tasks that have taken place in previous years that contribute to research conducted in this area. Specifically, the shared tasks CLEF 2019 Protest News (H\u00fcrriyetoglu et al., 2019) , AESPEN 2020 (H\u00fcrriyetoglu et al., 2020) , and CASE 2021 (task 1) (H\u00fcrriyetoglu et al., 2021) focus on event detection at both the sentence and document level, as well as event co-reference resolution.", |
| "cite_spans": [ |
| { |
| "start": 531, |
| "end": 558, |
| "text": "(H\u00fcrriyetoglu et al., 2019)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 573, |
| "end": 600, |
| "text": "(H\u00fcrriyetoglu et al., 2020)", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 626, |
| "end": 653, |
| "text": "(H\u00fcrriyetoglu et al., 2021)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Currently, not much research has been conducted that further analyzes event data once the events have been identified. There are a handful of studies across different domains. Peng et al. (2019) achieve state of the art results detecting and classifying social event data with a Pairwise Popularity Graph Convolutional Network (PP-GCN) with an external knowledge base. Nugent et al. (2017) compare different supervised classification methods for detecting a range of different events, and achieve good results with Support Vector Machines (SVM) and Convolutional Neural Networks (CNN). A benchmark corpus for fine-grained political event classification was created by the organizers of this task and an initial exploration and classification of the data is reported on in and . The findings reported that BERT transformer models achieved a micro F1 of (0.943-0.949) and a macro F1 of (0.860-0.889). More simple TF-IDF-weighted character n-gram models also achieved good results. A large dataset of 600,000 annotated ACLED event snippets was used as training data.", |
| "cite_spans": [ |
| { |
| "start": 176, |
| "end": 194, |
| "text": "Peng et al. (2019)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 369, |
| "end": 389, |
| "text": "Nugent et al. (2017)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Due to copyright reasons, the data used in this paper was collected directly from the ACLED website. 1 To create the corpus, all data from each available region was downloaded and then filtered using the following steps.", |
| "cite_spans": [ |
| { |
| "start": 101, |
| "end": 102, |
| "text": "1", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data collection", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Firstly, all events with less than 25 tokens and more than 1000 tokens were removed. The next step was to balance the corpus based on the 25 different fine-grained event classes. Originally, the largest class in the corpus consisted of 36.69% of the events, compared to the smallest with 0.001%. To create a more balanced version of the corpus, we extracted a sample of events per class, with the smallest classes being fully represented and extracting only a percentage of the largest classes. Note that it was not possible to fully balance all of the classes as there was only a very small amount of data for classes such as CHEM WEAP. A random sample of this balanced corpus was then split into a train (n=94000), development (n=9000), and test (n=2500) corpus, which also all contain the balanced class distribution. We observed that randomizing the order of events was crucial, to avoid 1 https://acleddata.com/data-export-tool/ introducing a bias based on the different ACLED regions. Figure 2 illustrates the distribution of the corpus. A more detailed table can be found in appendix A.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 991, |
| "end": 999, |
| "text": "Figure 2", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Data collection", |
| "sec_num": "3" |
| }, |
| { |
| "text": "In a further step, we created three different versions of the original corpus. The first version, referred to as ACLED N, contains the original text from the ACLED download, an example of which can be found below.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data collection", |
| "sec_num": "3" |
| }, |
| { |
| "text": "{text: CPI(M) activists attacked a BJP rally in Hrishyamukh on 18 January 2018., subtype: FORCE_AGAINST_PROTEST} Based on the results presented by , where the BERT transformer model performed slightly better on the corpus with less preprocessing, we decided to include a version with little to no pre-processing. In ACLED L, we replaced all locations from the text using the Flair Named Entity Recognition (NER) tagger (Akbik et al., 2018) with the generic token 'LOC'. The third version, ACLED T, contains a pre-processed version of the original text, but without any time stamps. All dates and times were removed from the text and replaced with 'TIME'. These two alternative versions of the corpus were created to analyse whether or not the information specific to one particular event or set of events would be transferable to the classification of other events.", |
| "cite_spans": [ |
| { |
| "start": 419, |
| "end": 439, |
| "text": "(Akbik et al., 2018)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data collection", |
| "sec_num": "3" |
| }, |
| { |
| "text": "We submitted five system runs for evaluation. The systems differ slightly from each other, either in the model or the way the used data was pre-processed. The general approach for all submitted systems was to use fine-tuned pre-trained transformer document embeddings. All experiments were conducted using the Flair framework (Akbik et al., 2019) .", |
| "cite_spans": [ |
| { |
| "start": 326, |
| "end": 346, |
| "text": "(Akbik et al., 2019)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System Description", |
| "sec_num": "4" |
| }, |
| { |
| "text": "For system 1, we fine-tuned the RoBERTa base model (Liu et al., 2019) , and trained the embeddings using a learning rate of 3e-5, a batch size of 16. Based on our experiments, we trained the model for 2 epochs, because we found that the model overfits if we trained for more than 2 epochs. After each epoch the training data was shuffled and this was also done in the subsequent systems. Additionally, we assigned weights to the different event classes. This was done to smooth out any remaining differences in class sizes. We used the ACLED L version of the corpus as text input.", |
| "cite_spans": [ |
| { |
| "start": 51, |
| "end": 69, |
| "text": "(Liu et al., 2019)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System 1 -RoBERTa ACLED L", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "System 2 again uses the RoBERTa base model (Liu et al., 2019) and the previously mentioned parameters for learning rate (3e-5), batch size (16) and number of epochs (2). The difference to system 1 is, that the text that was used during the fine-tuning of the model was not pre-processed. This means that the text snippets that were obtained from the ACLED (Donnay et al., 2019) database were fed into the system in their original state and, therefore, all information included in the text was kept.", |
| "cite_spans": [ |
| { |
| "start": 43, |
| "end": 61, |
| "text": "(Liu et al., 2019)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 356, |
| "end": 377, |
| "text": "(Donnay et al., 2019)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System 2 -RoBERTa ACLED N", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "For system 3, we used the pre-trained BERT basecased model (Devlin et al., 2019) along with a learning rate of 3e-5, a batch size of 16 and 2 epochs for training. As in system 1, we used the ACLED L corpus.", |
| "cite_spans": [ |
| { |
| "start": 59, |
| "end": 80, |
| "text": "(Devlin et al., 2019)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System 3 -BERT ACLED L", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "System 4 used the same settings as system 3, meaning, the pre-trained BERT base-cased model (Devlin et al., 2019), a learning rate of 3e-5, a batch size of 16 and 2 epochs for training. The input data for system 4 consisted of the original text from ACLED N.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System 4 -BERT ACLED N", |
| "sec_num": "4.4" |
| }, |
| { |
| "text": "Our last system, system 5, made use of the pretrained BERT base-cased model (Devlin et al., 2019) . The learning rate was set to 3e-5, the batch size to 32. It was trained for 2 epochs. For the text input we used the text from ACLED T, where all time and date stamps have been removed.", |
| "cite_spans": [ |
| { |
| "start": 76, |
| "end": 97, |
| "text": "(Devlin et al., 2019)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System 5 -BERT ACLED T", |
| "sec_num": "4.5" |
| }, |
| { |
| "text": "Preliminary model evaluations on 10 held-out test sets show that each of the systems performed comparatively well. The RoBERTa model with the normal ACLED text as input performed slightly better than the other systems. Table 1 below shows the range of Macro and Micro F1 scores across the 10 test sets. Model performance increased or decreased slightly, depending on the samples in the individual test sets. The results also illustrate that the removal of the location or time mentions in the event snippets, does not greatly influence system performance. Rather, the preliminary tests indicate that the fine-tuned RoBERTa embeddings benefit from the inclusion of the more detailed ACLED specific information.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Preliminary experiments", |
| "sec_num": "5" |
| }, |
| { |
| "text": "An analysis of the results of the individual classes, shows that each of the 25 subtypes achieve f1-scores of over 0. 6 Results Table 7 shows the results of our five system submissions. The systems were tested on a test set provided by the organizers, consisting of 829 samples for subtask 1. We find that System 2, using the RoBERTa base model (Liu et al., 2019) and ACLED N as input, performs best with an average weighted F-score of 0.83, average macro F-score of 0.794 and average micro F-score of 0.829. A more detailed overview can be found in appendix A.", |
| "cite_spans": [ |
| { |
| "start": 345, |
| "end": 363, |
| "text": "(Liu et al., 2019)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 128, |
| "end": 135, |
| "text": "Table 7", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Preliminary experiments", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Additionally, the second model that uses original ACLED text, System 4, achieves the second best result. As was the case in our preliminary experiments, we see that the inclusion of specific location and timestamps in the training data, does not greatly influence the ability of the system to predict the different classes correctly or incorrectly. ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Preliminary experiments", |
| "sec_num": "5" |
| }, |
| { |
| "text": "To get a better insight into the workings of our systems, we conducted an error analysis on the test data provided by the organizers for all five submissions. In order to investigate misclassifications made by the models, we decided to look at the performance of the system with regard to the individual classes.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Error Analysis", |
| "sec_num": "7" |
| }, |
| { |
| "text": "As can be seen in Table 3 , all models score low F-scores for either the class OTHER or the class PROPERTY DISTRUCT, or both. The results obtained for these classes substantially lower the overall average F-scores of the models. All models achieve the highest scores for the classes SUIC BOMB, GRENADE and CHEM WEAP. We looked at the word distribution these classes have in our training data as can be seen in figure 2 and 3.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 18, |
| "end": 25, |
| "text": "Table 3", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Analysis of Word Frequencies", |
| "sec_num": "7.1" |
| }, |
| { |
| "text": "Considering these distributions, it can be stated that a specific vocabulary, as can be found in the class SUIC BOMB, is advantageous for a correct classification, while a heterogeneous vocabulary, as can be found in the class OTHER, is disadvantageous. One can tell that while the by far most frequently occurring word in texts regarding the event type SUIC BOMB, namely \"suicide\", is clearly indicative for the given class, the most frequently used words in connection with the event type OTHER, namely \"activity\", \"violent\", \"area\" and \"force\" are rather generic. Furthermore, they can also be found frequently in a number of texts connected to other classes (e.g., \"area\": AIR STRIKE, CHANGE TO GROUP ACT, \"force\": NON STATE ACTOR OVERTAKES TER, NON VIOL TERRIT TRANSFER). This does not hold true for the word \"suicide\".", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Analysis of Word Frequencies", |
| "sec_num": "7.1" |
| }, |
| { |
| "text": "Looking further at the errors, we see that 65 samples of the test data were classified incorrectly by all five models. This makes up between 37% and 45% of errors for the respective systems. It is noticeable that all five models frequently predict the class MOB VIOL for sentences that are gold labeled as PROPERTY DISTRUCT (between 5 and 9 times for the respective systems). No other two classes are confused this often, and to investigate further we analysed these two classes with regard to their topic compactness. We calculated the topic distances of the sentences in comparison to the topic centroids per class in the training data. Figures 4 and 5 show the results of the topic compactness analysis. We see that both classes, MOB VIOL and PROP-ERTY DISTRUCT, are quite compact. There are some outliers, but most of the document vectors are clustered close to each other and the topic centroid. However, if we combine the classes into one topic and again analyse the distribution of document vectors to the topic centroid, we find that there are also very few outliers, as can be seen in figure 6 . This means that the examples for MOB VIOL and PROPERTY DISTRUCT in our training data are similar to each other, which may explain why our models consistently confuse these two classes with regard to the test data provided by the organizers.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 639, |
| "end": 648, |
| "text": "Figures 4", |
| "ref_id": "FIGREF2" |
| }, |
| { |
| "start": 1094, |
| "end": 1102, |
| "text": "figure 6", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Frequent Errors", |
| "sec_num": "7.2" |
| }, |
| { |
| "text": "Looking at the test samples, we further find that due to the large number of classes, it is also difficult for human annotators to distinguish between the different classes in some cases. An example for this is the following:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Frequent Errors", |
| "sec_num": "7.2" |
| }, |
| { |
| "text": "{text: Police said two groups from different communities in Chhabra town of Rajasthan's Baran district pelted stones on each other and torched vehicles parked around after putting six shops afire, guess: MOB_VIOL, gold: PROPERTY_DISTRUCT} All our models consistently predict the event class MOB VIOL for this example, the gold standard annotation is, however, PROP-ERTY DISTRUCT. It can be argued that the given example actually includes both event classes, with the first part of the sentence, \"Police said two groups from different communities in Chhabra town of Rajasthan's Baran district pelted stones on each other\" being an instance of MOB VIOL, while the second part, \"and torched vehicles parked around after putting six shops afire\", belongs to the class PROPERTY DISTRUCT. Test instances like this pose a challenge for the models.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Frequent Errors", |
| "sec_num": "7.2" |
| }, |
| { |
| "text": "In this study we proposed the use of fine-tuned RoBERTa transformer document embeddings for the fine-grained classification of socio-political events. We balanced the corpus to ensure that the 25 subtypes were represented as equally as possible. Compared to the results that were achieved during the preliminary experiments, we observed a drop in performance on the test set provided by the organizers. However, compared to the baseline figures provided by the organizers in , we achieved very similar results with less training data. This suggests that balancing the training data data had a positive effect on model performance.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "8" |
| }, |
| { |
| "text": "Our analysis of the results of both different test sets, the set created for preliminary experiments and the set provided by the organizers for system evaluation, show that there is definitely a difference in performance in the various classes. It also highlighted the issue of events that could be classed as more than one different subtype, and the challenge that these events pose for fine-grained classification. Depending on the given use case, parts of our system could already be implemented in a real world setting in order to analyze the flow of data in the information space and achieve situational awareness in the physical world, as clear cut classes like CHEM WEAP and GRENADE are identified reliably. In a military setting, for example, these classes are far more relevant than occurrences of PROPERTY DISTRUCT.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "8" |
| }, |
| { |
| "text": "In future work, it would be interesting to evaluate if the use of more training data, while still trying to obtain a more even distribution of classes, would further increase performance. Particularly, it raises the question if more training data would increase performance for the classes that currently do not perform as well. A more thorough class analysis, which would contribute to understanding why there seem to be systematic errors in specific classes, could provide insight into answering this question. ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "8" |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "This research has been supported through funding from Philip Morris Impact as part of the Fraud Information Fusion Intelligence Project.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| }, |
| { |
| "text": "Precision ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Class", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Flair: An easy-to-use framework for state-of-the-art nlp", |
| "authors": [ |
| { |
| "first": "Alan", |
| "middle": [], |
| "last": "Akbik", |
| "suffix": "" |
| }, |
| { |
| "first": "Tanja", |
| "middle": [], |
| "last": "Bergmann", |
| "suffix": "" |
| }, |
| { |
| "first": "Duncan", |
| "middle": [], |
| "last": "Blythe", |
| "suffix": "" |
| }, |
| { |
| "first": "Kashif", |
| "middle": [], |
| "last": "Rasul", |
| "suffix": "" |
| }, |
| { |
| "first": "Stefan", |
| "middle": [], |
| "last": "Schweter", |
| "suffix": "" |
| }, |
| { |
| "first": "Roland", |
| "middle": [], |
| "last": "Vollgraf", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "NAACL 2019, 2019 Annual Conference of the North American Chapter of the Association for Computational Linguistics (Demonstrations)", |
| "volume": "", |
| "issue": "", |
| "pages": "54--59", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alan Akbik, Tanja Bergmann, Duncan Blythe, Kashif Rasul, Stefan Schweter, and Roland Vollgraf. 2019. Flair: An easy-to-use framework for state-of-the-art nlp. In NAACL 2019, 2019 Annual Conference of the North American Chapter of the Association for Computational Linguistics (Demonstrations), pages 54-59.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Contextual string embeddings for sequence labeling", |
| "authors": [ |
| { |
| "first": "Alan", |
| "middle": [], |
| "last": "Akbik", |
| "suffix": "" |
| }, |
| { |
| "first": "Duncan", |
| "middle": [], |
| "last": "Blythe", |
| "suffix": "" |
| }, |
| { |
| "first": "Roland", |
| "middle": [], |
| "last": "Vollgraf", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "COLING 2018, 27th International Conference on Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "1638--1649", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alan Akbik, Duncan Blythe, and Roland Vollgraf. 2018. Contextual string embeddings for sequence labeling. In COLING 2018, 27th International Con- ference on Computational Linguistics, pages 1638- 1649.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "BERT: Pre-training of deep bidirectional transformers for language understanding", |
| "authors": [ |
| { |
| "first": "Jacob", |
| "middle": [], |
| "last": "Devlin", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming-Wei", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "Kenton", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Kristina", |
| "middle": [], |
| "last": "Toutanova", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "1", |
| "issue": "", |
| "pages": "4171--4186", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/N19-1423" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language under- standing. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171-4186, Minneapolis, Minnesota. Associ- ation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Integrating conflict event data", |
| "authors": [ |
| { |
| "first": "Karsten", |
| "middle": [], |
| "last": "Donnay", |
| "suffix": "" |
| }, |
| { |
| "first": "Eric", |
| "middle": [ |
| "T" |
| ], |
| "last": "Dunford", |
| "suffix": "" |
| }, |
| { |
| "first": "Erin", |
| "middle": [ |
| "C" |
| ], |
| "last": "Mcgrath", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Backer", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [ |
| "E" |
| ], |
| "last": "Cunningham", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Journal of Conflict Resolution", |
| "volume": "63", |
| "issue": "5", |
| "pages": "1337--1364", |
| "other_ids": { |
| "DOI": [ |
| "10.1177/0022002718777050" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Karsten Donnay, Eric T. Dunford, Erin C. McGrath, David Backer, and David E. Cunningham. 2019. In- tegrating conflict event data. Journal of Conflict Res- olution, 63(5):1337-1364.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Fine-grained event classification in news-like text snippets shared task 2, case 2021", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Haneczok", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Jacquet", |
| "suffix": "" |
| }, |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Piskorski", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Stefanovitch", |
| "suffix": "" |
| } |
| ], |
| "year": 2021, |
| "venue": "Proceedings of the Workshop on Challenges and Applications of Automated Text Extraction of Socio-Political Event from Text (CASE 2021), co-located with the Joint Conference of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J Haneczok, G Jacquet, J Piskorski, and N Stefanovitch. 2021. Fine-grained event classification in news-like text snippets shared task 2, case 2021. In Proceed- ings of the Workshop on Challenges and Applica- tions of Automated Text Extraction of Socio-Political Event from Text (CASE 2021), co-located with the Joint Conference of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Lan- guage Processing (ACL-IJCNLP 2021).", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Multilingual protest news detection -shared task 1, case 2021", |
| "authors": [ |
| { |
| "first": "Ali", |
| "middle": [], |
| "last": "H\u00fcrriyetoglu", |
| "suffix": "" |
| }, |
| { |
| "first": "Osman", |
| "middle": [], |
| "last": "Mutlu", |
| "suffix": "" |
| }, |
| { |
| "first": "Erdem", |
| "middle": [], |
| "last": "Y\u00f6r\u00fck", |
| "suffix": "" |
| }, |
| { |
| "first": "Ritesh", |
| "middle": [], |
| "last": "Kumar", |
| "suffix": "" |
| }, |
| { |
| "first": "Shyam", |
| "middle": [], |
| "last": "Ratan", |
| "suffix": "" |
| } |
| ], |
| "year": 2021, |
| "venue": "Proceedings of the Workshop on Challenges and Applications of Automated Extraction of Sociopolitical Events from Text (CASE 2021), online. Association for Computational Linguistics (ACL)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ali H\u00fcrriyetoglu, Osman Mutlu, Erdem Y\u00f6r\u00fck, Ritesh Kumar, and Shyam Ratan. 2021. Multilingual protest news detection -shared task 1, case 2021. In Proceedings of the Workshop on Challenges and Applications of Automated Extraction of Socio- political Events from Text (CASE 2021), online. As- sociation for Computational Linguistics (ACL).", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Overview of clef 2019 lab protestnews: Extracting protests from news in a cross-context setting", |
| "authors": [ |
| { |
| "first": "Ali", |
| "middle": [], |
| "last": "H\u00fcrriyetoglu", |
| "suffix": "" |
| }, |
| { |
| "first": "Erdem", |
| "middle": [], |
| "last": "Y\u00f6r\u00fck", |
| "suffix": "" |
| }, |
| { |
| "first": "Deniz", |
| "middle": [], |
| "last": "Y\u00fcret", |
| "suffix": "" |
| }, |
| { |
| "first": "Burak", |
| "middle": [], |
| "last": "Agr\u0131 Yoltar", |
| "suffix": "" |
| }, |
| { |
| "first": "F\u0131rat", |
| "middle": [], |
| "last": "G\u00fcrel", |
| "suffix": "" |
| }, |
| { |
| "first": "Osman", |
| "middle": [], |
| "last": "Duru\u015fan", |
| "suffix": "" |
| }, |
| { |
| "first": "Arda", |
| "middle": [], |
| "last": "Mutlu", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Akdemir", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Experimental IR Meets Multilinguality, Multimodality, and Interaction", |
| "volume": "", |
| "issue": "", |
| "pages": "425--432", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ali H\u00fcrriyetoglu, Erdem Y\u00f6r\u00fck, Deniz Y\u00fcret, \u00c7 agr\u0131 Yoltar, Burak G\u00fcrel, F\u0131rat Duru\u015fan, Osman Mutlu, and Arda Akdemir. 2019. Overview of clef 2019 lab protestnews: Extracting protests from news in a cross-context setting. In Experimental IR Meets Multilinguality, Multimodality, and Interac- tion, pages 425-432, Cham. Springer International Publishing.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Automated extraction of socio-political events from news (AESPEN): Workshop and shared task report", |
| "authors": [ |
| { |
| "first": "Ali", |
| "middle": [], |
| "last": "H\u00fcrriyetoglu", |
| "suffix": "" |
| }, |
| { |
| "first": "Vanni", |
| "middle": [], |
| "last": "Zavarella", |
| "suffix": "" |
| }, |
| { |
| "first": "Hristo", |
| "middle": [], |
| "last": "Tanev", |
| "suffix": "" |
| }, |
| { |
| "first": "Erdem", |
| "middle": [], |
| "last": "Y\u00f6r\u00fck", |
| "suffix": "" |
| }, |
| { |
| "first": "Ali", |
| "middle": [], |
| "last": "Safaya", |
| "suffix": "" |
| }, |
| { |
| "first": "Osman", |
| "middle": [], |
| "last": "Mutlu", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Proceedings of the Workshop on Automated Extraction of Socio-political Events from News 2020", |
| "volume": "", |
| "issue": "", |
| "pages": "1--6", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ali H\u00fcrriyetoglu, Vanni Zavarella, Hristo Tanev, Er- dem Y\u00f6r\u00fck, Ali Safaya, and Osman Mutlu. 2020. Automated extraction of socio-political events from news (AESPEN): Workshop and shared task report. In Proceedings of the Workshop on Automated Ex- traction of Socio-political Events from News 2020, pages 1-6, Marseille, France. European Language Resources Association (ELRA).", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Roberta: A robustly optimized bert pretraining approach", |
| "authors": [ |
| { |
| "first": "Yinhan", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Myle", |
| "middle": [], |
| "last": "Ott", |
| "suffix": "" |
| }, |
| { |
| "first": "Naman", |
| "middle": [], |
| "last": "Goyal", |
| "suffix": "" |
| }, |
| { |
| "first": "Jingfei", |
| "middle": [], |
| "last": "Du", |
| "suffix": "" |
| }, |
| { |
| "first": "Mandar", |
| "middle": [], |
| "last": "Joshi", |
| "suffix": "" |
| }, |
| { |
| "first": "Danqi", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Omer", |
| "middle": [], |
| "last": "Levy", |
| "suffix": "" |
| }, |
| { |
| "first": "Mike", |
| "middle": [], |
| "last": "Lewis", |
| "suffix": "" |
| }, |
| { |
| "first": "Luke", |
| "middle": [], |
| "last": "Zettlemoyer", |
| "suffix": "" |
| }, |
| { |
| "first": "Veselin", |
| "middle": [], |
| "last": "Stoyanov", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Man- dar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining ap- proach. Cite arxiv:1907.11692.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "A comparison of classification models for natural disaster and critical event detection from news", |
| "authors": [ |
| { |
| "first": "Tim", |
| "middle": [], |
| "last": "Nugent", |
| "suffix": "" |
| }, |
| { |
| "first": "Fabio", |
| "middle": [], |
| "last": "Petroni", |
| "suffix": "" |
| }, |
| { |
| "first": "Natraj", |
| "middle": [], |
| "last": "Raman", |
| "suffix": "" |
| }, |
| { |
| "first": "Lucas", |
| "middle": [], |
| "last": "Carstens", |
| "suffix": "" |
| }, |
| { |
| "first": "Jochen", |
| "middle": [ |
| "L" |
| ], |
| "last": "Leidner", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "2017 IEEE International Conference on Big Data (Big Data)", |
| "volume": "", |
| "issue": "", |
| "pages": "3750--3759", |
| "other_ids": { |
| "DOI": [ |
| "10.1109/BigData.2017.8258374" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tim Nugent, Fabio Petroni, Natraj Raman, Lucas Carstens, and Jochen L. Leidner. 2017. A compar- ison of classification models for natural disaster and critical event detection from news. In 2017 IEEE International Conference on Big Data (Big Data), pages 3750-3759.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Fine-grained event categorization with heterogeneous graph convolutional networks", |
| "authors": [ |
| { |
| "first": "Hao", |
| "middle": [], |
| "last": "Peng", |
| "suffix": "" |
| }, |
| { |
| "first": "Jianxin", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Qiran", |
| "middle": [], |
| "last": "Gong", |
| "suffix": "" |
| }, |
| { |
| "first": "Yangqiu", |
| "middle": [], |
| "last": "Song", |
| "suffix": "" |
| }, |
| { |
| "first": "Yuanxin", |
| "middle": [], |
| "last": "Ning", |
| "suffix": "" |
| }, |
| { |
| "first": "Kunfeng", |
| "middle": [], |
| "last": "Lai", |
| "suffix": "" |
| }, |
| { |
| "first": "Philip", |
| "middle": [ |
| "S" |
| ], |
| "last": "Yu", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19", |
| "volume": "", |
| "issue": "", |
| "pages": "3238--3245", |
| "other_ids": { |
| "DOI": [ |
| "10.24963/ijcai.2019/449" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hao Peng, Jianxin Li, Qiran Gong, Yangqiu Song, Yuanxin Ning, Kunfeng Lai, and Philip S. Yu. 2019. Fine-grained event categorization with hetero- geneous graph convolutional networks. In Proceed- ings of the Twenty-Eighth International Joint Con- ference on Artificial Intelligence, IJCAI-19, pages 3238-3245. International Joint Conferences on Ar- tificial Intelligence Organization.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "New benchmark corpus and models for fine-grained event classification: To BERT or not to BERT?", |
| "authors": [ |
| { |
| "first": "Jakub", |
| "middle": [], |
| "last": "Piskorski", |
| "suffix": "" |
| }, |
| { |
| "first": "Jacek", |
| "middle": [], |
| "last": "Haneczok", |
| "suffix": "" |
| }, |
| { |
| "first": "Guillaume", |
| "middle": [], |
| "last": "Jacquet", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Proceedings of the 28th International Conference on Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "6663--6678", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jakub Piskorski, Jacek Haneczok, and Guillaume Jacquet. 2020. New benchmark corpus and mod- els for fine-grained event classification: To BERT or not to BERT? In Proceedings of the 28th Inter- national Conference on Computational Linguistics, pages 6663-6678, Barcelona, Spain (Online). Inter- national Committee on Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "TF-IDF character N-grams versus word embedding-based models for fine-grained event classification: A preliminary study", |
| "authors": [ |
| { |
| "first": "Jakub", |
| "middle": [], |
| "last": "Piskorski", |
| "suffix": "" |
| }, |
| { |
| "first": "Guillaume", |
| "middle": [], |
| "last": "Jacquet", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Proceedings of the Workshop on Automated Extraction of Socio-political Events from News 2020", |
| "volume": "", |
| "issue": "", |
| "pages": "26--34", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jakub Piskorski and Guillaume Jacquet. 2020. TF-IDF character N-grams versus word embedding-based models for fine-grained event classification: A pre- liminary study. In Proceedings of the Workshop on Automated Extraction of Socio-political Events from News 2020, pages 26-34, Marseille, France. Euro- pean Language Resources Association (ELRA).", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Introducing acled: An armed conflict location and event dataset: Special data feature", |
| "authors": [ |
| { |
| "first": "Clionadh", |
| "middle": [], |
| "last": "Raleigh", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [], |
| "last": "Linke", |
| "suffix": "" |
| }, |
| { |
| "first": "H\u00e5vard", |
| "middle": [], |
| "last": "Hegre", |
| "suffix": "" |
| }, |
| { |
| "first": "Joakim", |
| "middle": [], |
| "last": "Karlsen", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Journal of Peace Research", |
| "volume": "47", |
| "issue": "5", |
| "pages": "651--660", |
| "other_ids": { |
| "DOI": [ |
| "10.1177/0022343310378914" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Clionadh Raleigh, Andrew Linke, H\u00e5vard Hegre, and Joakim Karlsen. 2010. Introducing acled: An armed conflict location and event dataset: Special data fea- ture. Journal of Peace Research, 47(5):651-660.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "A survey of event extraction from text", |
| "authors": [ |
| { |
| "first": "Wei", |
| "middle": [], |
| "last": "Xiang", |
| "suffix": "" |
| }, |
| { |
| "first": "Bang", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "IEEE Access", |
| "volume": "7", |
| "issue": "", |
| "pages": "173111--173137", |
| "other_ids": { |
| "DOI": [ |
| "10.1109/ACCESS.2019.2956831" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wei Xiang and Bang Wang. 2019. A survey of event ex- traction from text. IEEE Access, 7:173111-173137.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "text": "Class distribution", |
| "type_str": "figure", |
| "num": null, |
| "uris": null |
| }, |
| "FIGREF1": { |
| "text": "Top 10 words in the class OTHER Figure 3: Top 10 words in the class SUIC BOMB", |
| "type_str": "figure", |
| "num": null, |
| "uris": null |
| }, |
| "FIGREF2": { |
| "text": "Distribution of document vectors to topic centroid in class MOB VIOL Figure 5: Distribution of document vectors to topic centroid in class PROPERTY DISTRUCT Figure 6: Distribution of document vectors to topic centroid in the combined class PROPERTY DISTRUCT and MOB VIOL", |
| "type_str": "figure", |
| "num": null, |
| "uris": null |
| }, |
| "TABREF0": { |
| "html": null, |
| "content": "<table><tr><td>Macro F1</td><td>Micro F1</td><td>Weighted F1</td></tr><tr><td colspan=\"2\">Table 1: Preliminary Evaluation Results</td><td/></tr><tr><td>VIOL DEMONSTR, with an F-score of 0.814</td><td/><td/></tr><tr><td>and 0.819 respectively. The highest scoring class</td><td/><td/></tr><tr><td>is CHEM WEAP (F-score = 1), however there</td><td/><td/></tr><tr><td>are only two instances present in this test set.</td><td/><td/></tr><tr><td>PEACE PROTEST and ABDUCT DISSAP also</td><td/><td/></tr><tr><td>score highly, achieving an F-score of 0.978 and</td><td/><td/></tr><tr><td>0.975 respectively. A table containing a detailed</td><td/><td/></tr><tr><td>overview of each class can be found in appendix</td><td/><td/></tr><tr><td>A.</td><td/><td/></tr></table>", |
| "text": "800. The two lowest scoring classes are HQ ESTABLISHED and RoBERTa ACLED L 0.887 -0.919 0.917 -0.929 0.917 -0.929 RoBERTa ACLED N 0.894 -0.923 0.916 -0.932 0.918 -0.931 BERT ACLED L 0.868 -0.911 0.913 -0.928 0.913 -0.928 BERT ACLED N 0.869 -0.900 0.907 -0.925 0.907 -0.925 BERT ACLED T 0.889 -0.918 0.913 -0.929 0.913 -0.929", |
| "type_str": "table", |
| "num": null |
| }, |
| "TABREF2": { |
| "html": null, |
| "content": "<table/>", |
| "text": "System Results", |
| "type_str": "table", |
| "num": null |
| }, |
| "TABREF4": { |
| "html": null, |
| "content": "<table/>", |
| "text": "Event Type Error Analysis", |
| "type_str": "table", |
| "num": null |
| }, |
| "TABREF6": { |
| "html": null, |
| "content": "<table><tr><td>System</td><td>Avg Type</td><td colspan=\"3\">Avg Prec Avg Recall Avg F-score</td></tr><tr><td colspan=\"2\">System 1 micro avg</td><td>0.797</td><td>0.797</td><td>0.797</td></tr><tr><td colspan=\"2\">System 1 macro avg</td><td>0.790</td><td>0.778</td><td>0.770</td></tr><tr><td colspan=\"2\">System 1 weighted avg</td><td>0.824</td><td>0.797</td><td>0.799</td></tr><tr><td colspan=\"2\">System 2 micro avg</td><td>0.829</td><td>0.829</td><td>0.829</td></tr><tr><td colspan=\"2\">System 2 macro avg</td><td>0.807</td><td>0.808</td><td>0.794</td></tr><tr><td colspan=\"2\">System 2 weighted avg</td><td>0.851</td><td>0.829</td><td>0.830</td></tr><tr><td colspan=\"2\">System 3 micro avg</td><td>0.808</td><td>0.808</td><td>0.808</td></tr><tr><td colspan=\"2\">System 3 macro avg</td><td>0.787</td><td>0.779</td><td>0.768</td></tr><tr><td colspan=\"2\">System 3 weighted avg</td><td>0.828</td><td>0.808</td><td>0.808</td></tr><tr><td colspan=\"2\">System 4 micro avg</td><td>0.802</td><td>0.802</td><td>0.802</td></tr><tr><td colspan=\"2\">System 4 macro avg</td><td>0.788</td><td>0.789</td><td>0.774</td></tr><tr><td colspan=\"2\">System 4 weighted avg</td><td>0.841</td><td>0.802</td><td>0.812</td></tr><tr><td colspan=\"2\">System 5 micro avg</td><td>0.793</td><td>0.793</td><td>0.793</td></tr><tr><td colspan=\"2\">System 5 macro avg</td><td>0.780</td><td>0.780</td><td>0.766</td></tr><tr><td colspan=\"2\">System 5 weighted avg</td><td>0.817</td><td>0.893</td><td>0.793</td></tr></table>", |
| "text": "Corpus class distribution of the 25 event subtypes.", |
| "type_str": "table", |
| "num": null |
| }, |
| "TABREF7": { |
| "html": null, |
| "content": "<table><tr><td>Class</td><td colspan=\"4\">Precision Recall F1-score Support</td></tr><tr><td>ABDUCT DISSAP</td><td>0.9787</td><td>0.9718</td><td>0.9753</td><td>142</td></tr><tr><td>AGREEMENT</td><td>0.8974</td><td>1.0000</td><td>0.9459</td><td>35</td></tr><tr><td>AIR STRIKE</td><td>1.0000</td><td>0.9333</td><td>0.9655</td><td>75</td></tr><tr><td>ARMED CLASH</td><td>0.9429</td><td>0.8771</td><td>0.9088</td><td>301</td></tr><tr><td>ARREST</td><td>0.9500</td><td>0.9500</td><td>0.9500</td><td>80</td></tr><tr><td>ART MISS ATTACK</td><td>0.9515</td><td>0.9608</td><td>0.9561</td><td>102</td></tr><tr><td>ATTACK</td><td>0.8361</td><td>0.9217</td><td>0.8768</td><td>166</td></tr><tr><td>CHANGE TO GROUP ACT</td><td>0.9716</td><td>0.9648</td><td>0.9682</td><td>142</td></tr><tr><td>CHEM WEAP</td><td>1.0000</td><td>1.0000</td><td>1.0000</td><td>2</td></tr><tr><td>DISR WEAP</td><td>0.9444</td><td>0.9659</td><td>0.9551</td><td>88</td></tr><tr><td>FORCE AGAINST PROTEST</td><td>0.8302</td><td>0.9167</td><td>0.8713</td><td>48</td></tr><tr><td>GOV REGAINS TERIT</td><td>0.8900</td><td>0.8900</td><td>0.8900</td><td>100</td></tr><tr><td>GRENADE</td><td>0.9744</td><td>0.9870</td><td>0.9806</td><td>77</td></tr><tr><td>HQ ESTABLISHED</td><td>0.6875</td><td>1.0000</td><td>0.8148</td><td>11</td></tr><tr><td>MOB VIOL</td><td>0.8598</td><td>0.8846</td><td>0.8720</td><td>104</td></tr><tr><td colspan=\"2\">NON STATE ACTOR OVERTAKES TER 0.8889</td><td>0.9014</td><td>0.8951</td><td>71</td></tr><tr><td>NON VIOL TERRIT TRANSFER</td><td>0.8966</td><td>0.8387</td><td>0.8667</td><td>31</td></tr><tr><td>OTHER</td><td>0.9531</td><td>0.8971</td><td>0.9242</td><td>68</td></tr><tr><td>PEACE PROTEST</td><td>0.9863</td><td>0.9703</td><td>0.9782</td><td>370</td></tr><tr><td>PROPERTY DISTRUCT</td><td>0.9700</td><td>0.9417</td><td>0.9557</td><td>206</td></tr><tr><td>PROTEST WITH INTER</td><td>0.9082</td><td>0.8990</td><td>0.9036</td><td>99</td></tr><tr><td>REM EXPLOS</td><td>0.9107</td><td>0.9273</td><td>0.9189</td><td>55</td></tr><tr><td>SEX VIOL</td><td>0.9630</td><td>0.8966</td><td>0.9286</td><td>29</td></tr><tr><td>SUIC BOMB</td><td>0.9643</td><td>0.9643</td><td>0.9643</td><td>28</td></tr><tr><td>VIOL DEMONSTR</td><td>0.7722</td><td>0.8714</td><td>0.8188</td><td>70</td></tr><tr><td>weighted avg</td><td>0.9338</td><td>0.9312</td><td>0.9318</td><td>2500</td></tr></table>", |
| "text": "Detailed System Results", |
| "type_str": "table", |
| "num": null |
| }, |
| "TABREF8": { |
| "html": null, |
| "content": "<table/>", |
| "text": "", |
| "type_str": "table", |
| "num": null |
| } |
| } |
| } |
| } |