| { |
| "paper_id": "2020", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T09:02:14.751324Z" |
| }, |
| "title": "Enhancing Quality of Corpus Annotation: Construction of the Multi-Layer Corpus Annotation and Simplified Validation of the Corpus Annotation", |
| "authors": [ |
| { |
| "first": "Youngbin", |
| "middle": [], |
| "last": "Noh", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Korea Advanced Institute of Science and Technology", |
| "location": { |
| "country": "South Korea" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Kuntae", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Korea Advanced Institute of Science and Technology", |
| "location": { |
| "country": "South Korea" |
| } |
| }, |
| "email": "kuntaek@kaist.ac.kr" |
| }, |
| { |
| "first": "Minho", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Korea Advanced Institute of Science and Technology", |
| "location": { |
| "country": "South Korea" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Cheolhun", |
| "middle": [], |
| "last": "Heo", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Korea Advanced Institute of Science and Technology", |
| "location": { |
| "country": "South Korea" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Yongbin", |
| "middle": [], |
| "last": "Jeong", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Korea Advanced Institute of Science and Technology", |
| "location": { |
| "country": "South Korea" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Yoosung", |
| "middle": [], |
| "last": "Jeong", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "Younggyun", |
| "middle": [], |
| "last": "Hahm", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Korea Advanced Institute of Science and Technology", |
| "location": { |
| "country": "South Korea" |
| } |
| }, |
| "email": "hahmyg@kaist.ac.kr" |
| }, |
| { |
| "first": "Taehwan", |
| "middle": [], |
| "last": "Oh", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Korea Advanced Institute of Science and Technology", |
| "location": { |
| "country": "South Korea" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Hyonsu", |
| "middle": [], |
| "last": "Choe", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Yonsei University", |
| "location": { |
| "region": "South Korea" |
| } |
| }, |
| "email": "choehyonsu@yonsei.ac.kr" |
| }, |
| { |
| "first": "Seokwon", |
| "middle": [], |
| "last": "Park", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Yonsei University", |
| "location": { |
| "region": "South Korea" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Jin-Dong", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "jdkim@dbcls.rois.ac.jp" |
| }, |
| { |
| "first": "Key-Sun", |
| "middle": [], |
| "last": "Choi", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Korea Advanced Institute of Science and Technology", |
| "location": { |
| "country": "South Korea" |
| } |
| }, |
| "email": "kschoi@kaist.ac.kr" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "In this paper, we construct simultaneously multi-layer corpus annotation of 7 linguistic layers as a gold set. And we design the validation procedure using the gold set and report the results of a validation procedure for other large-scale corpus annotation of 7 linguistic layers. Furthermore, we present a simplified validation method without a gold set using annotation models learned by the gold set. Based on the validation results, the tendency of annotation across the entire corpus is observed macroscopically, and the corpus annotation validation results are analyzed microscopically to verify the validation methodology to address the case study.", |
| "pdf_parse": { |
| "paper_id": "2020", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "In this paper, we construct simultaneously multi-layer corpus annotation of 7 linguistic layers as a gold set. And we design the validation procedure using the gold set and report the results of a validation procedure for other large-scale corpus annotation of 7 linguistic layers. Furthermore, we present a simplified validation method without a gold set using annotation models learned by the gold set. Based on the validation results, the tendency of annotation across the entire corpus is observed macroscopically, and the corpus annotation validation results are analyzed microscopically to verify the validation methodology to address the case study.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "As a resource for natural language processing, the corpus is annotated with additional information for various purposes. To annotate such various information to the raw corpus, corpus annotation project must be designed elaborately considering the requirements of the annotation procedure, annotation units, annotation tools, human annotators, and so on. A reliable annotation design makes the corpus annotation better quality (Finlayson and Erjavec, 2017) . Also, the design suitability of the corpus annotation project needs to be proved empirically, so the design of the corpus annotation project must be revised and supplemented iteratively (Pustejovsky and Stubbs, 2012) .", |
| "cite_spans": [ |
| { |
| "start": 427, |
| "end": 456, |
| "text": "(Finlayson and Erjavec, 2017)", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 645, |
| "end": 675, |
| "text": "(Pustejovsky and Stubbs, 2012)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, we construct 7 linguistic layers 1 (Ide, 1 We construct and evaluate the 7 linguistic layer corpus an-2017) of multi-layered corpus annotation as gold set (210K Eojeol 2 s) to validate large-scale corpus annotation by the 7 layers as evaluation set (3M Eojeols).", |
| "cite_spans": [ |
| { |
| "start": 50, |
| "end": 55, |
| "text": "(Ide,", |
| "ref_id": null |
| }, |
| { |
| "start": 56, |
| "end": 57, |
| "text": "1", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The gold set is annotated on the subset of the raw corpus of the evaluation set. The annotator groups of gold set by each layer, who annotated gold set assisted by auto-labeling are groups of experts separated from the annotator groups of evaluation set.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "We have designed and applied a corpus annotation method that uses the simple inter-annotator agreement to construct the gold sets at the same time under limited time and human resources. To do this, we assigned one annotation unit to two annotators. According to annotation results from two annotators, we conducted the cross-checking process to determine the final annotation result.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, validation of evaluation sets by layers is performed by comparing two corpus annotations (gold set -evaluation set) 3 using the gold set as a criterion. Comparative validation of the two corpus annotations using a gold set can only be performed on a limited part of the evaluation set sharing the same part of raw corpus to be annotated. notation: morphological analysis, lexical sense analysis, named entity analysis, subject anaphora resolution, co-reference resolution, dependency analysis, and semantic roles analysis. The evaluation sets are also constructed by the same layers.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "2 In Korean, the word segment divided by white space is called \"Eojeol\", this is composed of a noun or verb stem combined with a postposition (\"Josa\") or ending (\"Eomi\") that function as inflectional and derivational particles. (Noh et al., 2018) 3 In this project, we are constructed 7 linguistic layers of corpus annotations as gold set to validate 7 linguistic layers of corpus annotation (evaluation set) constructed by other project groups. The evaluation sets after validation can be downloaded at https://corpus.korean.go.kr/. That's why we present an additional method to validate in the range of evaluation set without a gold set. Thus, we constructed another corpus annotation (silver set) to observe the consistency of annotation across the entire evaluation set by annotating an annotation model using the gold set of each layer as training data. The silver set is compared with the evaluation set to observe the tendency of annotation across the entire evaluation set, and a data-driven statistical analysis is performed to evaluate the appropriateness of the validation results.", |
| "cite_spans": [ |
| { |
| "start": 228, |
| "end": 246, |
| "text": "(Noh et al., 2018)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In following section 2, we introduce the related works to the design and annotation of corpus annotation projects and corpus validation. In section 3, we introduce the design and annotation process for the construction of the gold set performed in this study. In section 4, we address the validation process for the format and annotation contents of the corpus annotation with an evaluation set using a gold set and validate the annotation consistency of the evaluation set without a gold set. In sections 5 and 6, we report the validation results in section 4 and issues related to these results.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The corpus annotation project should be modeled according to the goal of the project to reflect appropriate specifications. This is because the corpus annotation, which is a result of corpus annotation, is learning data for the machine learning algorithm to learn specific phenomena that are not only linguistic but also non-linguistic. Therefore, the corpus annotation project design needs to appropriately reflect the characteristics of the phenomena to be learned in the machine learning algorithm. A broadly used framework is the MATTER cycle (Pustejovsky and Stubbs, 2012) . The MATTER cycle is a cyclical process in which a corpus annotation project design ensures that the corpus annotation produces machine learning results that are appropriate for the goal of the corpus annotation project.", |
| "cite_spans": [ |
| { |
| "start": 547, |
| "end": 577, |
| "text": "(Pustejovsky and Stubbs, 2012)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Works", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The OntoNotes (Hovy et al., 2006; Weischedel et al., 2011; Weischedel et al., 2013 ) is a multi-layer corpus annotation constructed over six years of texts in various genres in three languages (English, Chinese, and Arabic). It is a multilingual, multi-layer corpus annotation that annotated the structural information of the text as well as the semantic information to understand the meaning of the context based on a syntactic structure derived from Penn Treebank corpus, and the predicate-argument structure derived from Penn PropBank. It includes annotations such as word sense disambiguation for verbs and nouns, entity names annotation, the ontology of each word, and coreference relations. OntoNotes had tried to secure at least 90% of inter-annotator agreement in each corpus annotation, improving the quality of corpus annotation.", |
| "cite_spans": [ |
| { |
| "start": 14, |
| "end": 33, |
| "text": "(Hovy et al., 2006;", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 34, |
| "end": 58, |
| "text": "Weischedel et al., 2011;", |
| "ref_id": null |
| }, |
| { |
| "start": 59, |
| "end": 82, |
| "text": "Weischedel et al., 2013", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Works", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The OntoNotes 5.0 (more details in (Weischedel et al., 2013) In NLP area, various evaluation and annotation methodologies have been used to enhance and manage corpus quality as a natural language processing resource. As a corpus annotation quality control methods, inter-annotator agreement (Pustejovsky and Stubbs, 2012) has been generally used to control the result of corpus annotation. Checking inter-annotator agreement among annotators is widely used not only for evaluating the results of annotations from an assigned group of annotators, but also for evaluating the quality of data collected from an unspecified number of annotator, such as crowdsourcing methodology (Nowak and R\u00fcger, 2010; Dumitrache, 2015; Dumitrache et al., 2018; Poesio et al., 2019) .", |
| "cite_spans": [ |
| { |
| "start": 35, |
| "end": 60, |
| "text": "(Weischedel et al., 2013)", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 291, |
| "end": 321, |
| "text": "(Pustejovsky and Stubbs, 2012)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 675, |
| "end": 698, |
| "text": "(Nowak and R\u00fcger, 2010;", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 699, |
| "end": 716, |
| "text": "Dumitrache, 2015;", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 717, |
| "end": 741, |
| "text": "Dumitrache et al., 2018;", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 742, |
| "end": 762, |
| "text": "Poesio et al., 2019)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Gold Set (Ours) Evaluation Set", |
| "sec_num": null |
| }, |
| { |
| "text": "In this section, we address the overall procedures of constructing corpus annotation as a gold set. The gold set and the evaluation set share a raw corpus, and the gold set is a corpus annotation of 7 layers constructed by sampling 7% of the raw corpus. Therefore, in this paper, the corpus annotation of 7 layers is simultaneously constructed to build multilayer corpus annotation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Construction of Corpus Annotation", |
| "sec_num": "3" |
| }, |
| { |
| "text": "To establish the corpus annotation guidelines by layers, the existing corpus annotation guidelines are revised and used according to the project purpose 4 . To make sure that the revised guidelines are not ambiguous or lacking in reflecting actual linguistic phenomena, three different annotator groups 5 constructed sample corpus annotation layer by layer on the same range of raw corpus. These sample corpus annotation also assessed the completeness of the corpus annotation guidelines, but were used as an annotation example in the annotation process. Through this process, it is possible to supplement 4 The annotation guidelines referenced in this project refer to the annotation guideline for each layer from the 21st century Sejong project (morphological analysis, lexical sense analysis), and the guidelines made by the Electronics and Telecommunications Research Institute (ETRI; named entities analysis, subject zero anaphora resolution, co-reference resolution, and semantic role analysis) and Telecommunications Technology Association (TTA; dependency analysis). These guidelines do not refer to literature information in this paper, because it also includes non-public materials. For inquiries about these guidelines, please contact NIKL, ETRI, TTA.", |
| "cite_spans": [ |
| { |
| "start": 606, |
| "end": 607, |
| "text": "4", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Annotation Specification", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "5 The annotation results of the three groups -ours, evaluatee groups, and expert group of National Institute of Korean Language (NIKL) annotated on the same range of raw corpus on the seven layers. The disagreement among the corpus annotation results of the three groups was decided by the NIKL expert group and the annotation guidelines were reflected in the results of the decision by NIKL. the process of evaluating the repetitive annotation scheme through this process.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Annotation Specification", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "In this project, the seven annotator groups 6 were recruited by layers, consisting of experts in Korean or linguistics, who can fully understand annotation guideline and apply this to corpus annotation. They studied and analyzed the existing guidelines and constructed sample corpus annotation. And they annotated the gold sets using the revised guidelines reflecting the results of the annotation of the sample corpus annotation. To ensure validation results of corpus annotation between two annotator groups, they were performed corpus annotation independently not only within their group but also from the annotator group of the evaluatee group. Also, they are completely separated from the annotation group of the evaluatee group and are independently annotated to provide conditions for evaluating the results of the annotation by inter-annotator agreement.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Annotation Specification", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "As mentioned above, the annotators of our project annotated independently of each other. For the annotators to be separated, a workbench with a personal workspace is required. We also designed a web-based annotation environment to reduce the time and location constraints of annotation work.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Annotation Environments", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "This project used a web-based workbench. This annotation environment was designed for this project and was developed to reflects the annotation schemes of the seven layers. This workbench could only be accessed by annotators and administrator designated for each layer. Multiple annotators can annotate one annotation unit, and it is possible to grasp the annotator's annotation status on an online browser.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Annotation Environments", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Annotators can use only the set of tag sets defined in the annotation schemes, and the annotation contents can also annotate only those in the annotation schemes. If the annotation guidelines have updated and annotators need to add new annotation content, they can ask the administrator to remove the restrictions for annotation contents. Also, the annotators annotated referring the results of the automatic preprocessing of the raw corpus by layers.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Annotation Environments", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "This workbench is equipped with a function of randomly assigning a defined annotation unit to an annotator. This allows an annotator not to annotate the entire document at once, but to annotate parts so that one document is annotated by multiple annotators.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Annotation Environments", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "In this paper, corpus annotation as a gold set was constructed using a human-machine collaborative annotation method (Figure 2 ). First, an automatic corpus annotation is automatically annotated using NLP Tools 7 to the subset of the raw corpus. This includes the annotation results of the Korean morphemes, dependency parsing, named entity recognition, and semantic role labeling (recognizing predicate and argument), and so on.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 117, |
| "end": 126, |
| "text": "(Figure 2", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Corpus Annotation Procedure", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Next, the annotator in the annotator group manually re-annotates the annotation unit by referring to the annotation result of the NLP tool. One annotation unit is annotated by randomly assigned two annotators in the group. If the annotation results of an annotation unit by two annotators are same, these annotation units are cross-checked by a randomly assigned annotator within the group. Else if the annotation results of an annotation unit by randomly assigned two annotators are different, these annotation units are decided final annotation result by a randomly assigned annotator within the group. Through these processes, an annotation unit is checked by annotators at least two times. After two parallel processes, those results of two-stage are made corpus annotation gold set.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Corpus Annotation Procedure", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "After constructing the gold sets, using these corpora, we validate the evaluation sets. The validation of the evaluation sets validates the format of corpus annotation, the annotation contents, and the annotation consistency. After that, the integration validation is performed to be used the evaluation sets as a multilayer corpus annotation (refer in Figure 1 ).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 353, |
| "end": 361, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Validation of Corpus Annotation", |
| "sec_num": "4" |
| }, |
| { |
| "text": "The format validation is a process of confirming whether the corpus has been constructed by the defined annotation format, and also is a stage of confirming whether the corpus can be used as electronic data. At this stage, corpus annotation is validated about the standard format and data structure. In addition to the corpus format, in this stage, it is checked whether or not a label defined for each layer is used, and whether other content is included in addition to the specified annotation content. When a format error is detected in this stage, the evaluation set does not proceed to the annotation contents validation stage, and correction and supplementation are required.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Format Validation", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "The annotation contents validation performs dataoriented validation that compares the gold set and evaluation set. At this validation, we compare the annotation contents of the two corpora and report the different annotation content to the evaluatee group. The annotation content validation items are selected based on the annotation guidelines for each layer, and the annotation contents defined in the layer are selected as validation items and shared in the evaluatee groups. Based on this report, the evaluatee groups can modify and supplement their corpus annotation. This process is a method of evaluating based on the inter-annotator agreement between annotation groups, and it is judged that the correct annotation is performed when the annotation of the two groups matched.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Annotation Contents Validation", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "The annotation consistency of the evaluation set is evaluated indirectly by confirming that the tendency of the annotation of the gold set is similar to the evaluation set. To do this, we create an automatic annotation model for each layer using a gold set as training data and construct an automatic corpus annotation (silver set) that annotates automatic annotation on a raw corpus in range of without a gold set. By comparing the silver set and evaluation set, the consistency of the annotation content is analyzed to evaluate the annotation consistency.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Consistency Validation for Evaluation Set using Silver Set", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "To validate the annotation consistency of the evaluation set, we were divided into 10 sections to analyze the tendency of the agreement between the silver set and evaluation set. The average of the agreement of corpus annotations between two corpora in 10 sections was averaged, and when the observed agreement rate of each section deviated from the 99% confidence interval (\u03b1 = 0.01) compared to the mean value, the corresponding section was evaluated to have relatively lower annotation consistency than other sections.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Consistency Validation for Evaluation Set using Silver Set", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "As a final stage in constructing a multi-layer corpus annotation of seven layers, it is necessary to check whether the raw corpus of evaluation sets are preserved and whether the annotation schemes have been observed. To make a multi-layer corpus annotation by combining the seven sets, we compared the statistics of the number of documents, paragraphs, sentences, and Eojeols in each corpus, and confirm that the defined ID assignment rules are followed. Table 2 shows the results of annotation agreement between the gold set and the evaluation set. Annotation contents validation of morphological corpus annotation was validated to match the Eojeol segment and morphological label annotation. In Korean Eojeol, morphemes such as stems (Eogan), postposition (josa), ending (Eomi) are combined to form a single Eojeol. Thus, to analyze morphemes, it is necessary to check whether the Eojeol segmentation is same and whether the same label is annotated to the segmented morpheme. Written corpus annotation was relatively consistent with both segmentation and label annotation in spoken corpus annotation. Segmenting concordance was lower than label annotation concordance, indicating that there was a difference in the morpheme semantic analysis of Eojeol between annotators.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 456, |
| "end": 463, |
| "text": "Table 2", |
| "ref_id": "TABREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Integration Validation", |
| "sec_num": "4.4" |
| }, |
| { |
| "text": "When comparing the annotation agreement between written corpus annotation and spoken corpus annotation, the tendency of the written corpus annotation shows a higher annotation agreement than spoken corpus annotation. It is because a spoken raw corpus transcribed public monologues (news), public conversation (broadcast conversation, interview, lecture, and so on), and private conversation recording. In the case of public monologue, it was refined to a level similar to that of the written raw corpus with well-refined text. In the case of private conversations or broadcast interviews, however, many features of spoken language (i.e., speech break, blur, reduction, slang, and so on) appeared, making it difficult for annotators to analyze text. Table 3 shows the results of annotation consistency validation. Annotation consistency validation was performed separately for written and spoken corpus annotation. To validate the consistency of the annotation, some of the measures used to validate the annotation contents for each layer were used, and the consistency was evaluated through the consistency of the indicator. The 99% confidence interval compared to the average value of the annotated agreement of the silver set and evaluation set for each section was shorter than that of the majority of written corpus annotation (Avg. of CI length: Written = 3.603, Spoken = 3.193). In the case of written corpus annotation, sections 1 and 2 of ZA corpus annotation showed a markedly low agreement, affecting the average CI length of ZA increase. The corpus showing the shortest 99% confidence interval is the named entity annotated written corpus with the smallest difference in the annotation content with the silver set of 10 sections (99% CI length = .422 (21.98 \u2264 CI \u2264 22.430), confidence(\u03b1 = 0.01)= 0.202, \u03c3=0.248). Also, one section out of the 99% confidence interval was analyzed, and it was evaluated that the consistency of the annotation of the named entity written corpus annotation was relatively higher than that of the other corpora.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 749, |
| "end": 756, |
| "text": "Table 3", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Annotation Contents Validation", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "Compared to the annotation content validation result (Table 2) , when the gold set and the evaluation set match 80% or more (MP (Written, Spoken), LS (Written, Spoken), NE (Written, Spoken), ZA (Written, Spoken), DP, SR), the length of the 99% confidence interval, except for the ZA written and MP spoken corpus annotation, is all 1.5 or less. There- Table 3 : Results of consistency validation. The bold indicates the agreement rate between the silver set and evaluation set outside the 99% confidence interval. The following models were used for the automatic annotation model for annotation consistency validation: (Ma and Hovy, 2016) -MP (Written), (Joshi et al., 2019) -ZA (Written, Spoken), CR (Written, Spoken), (Straka et al., 2016) -DP (Written), (Lee et al., 2015; Bae et al., 2017) -SR (Written). The annotation models of layers that have no reference was developed and used as a statistical-based learning model. fore, it was evaluated that it showed high annotation consistency.", |
| "cite_spans": [ |
| { |
| "start": 618, |
| "end": 637, |
| "text": "(Ma and Hovy, 2016)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 653, |
| "end": 673, |
| "text": "(Joshi et al., 2019)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 719, |
| "end": 740, |
| "text": "(Straka et al., 2016)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 756, |
| "end": 774, |
| "text": "(Lee et al., 2015;", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 775, |
| "end": 792, |
| "text": "Bae et al., 2017)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 53, |
| "end": 62, |
| "text": "(Table 2)", |
| "ref_id": "TABREF3" |
| }, |
| { |
| "start": 351, |
| "end": 358, |
| "text": "Table 3", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Consistency Validation", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "The annotation consistency validation by creating a learning model using a gold set started from the hypothesis that the balanced composition of the raw corpus represents a language phenomenon. In the case of the written raw corpus, it is constructed only newspaper articles, so there is relatively little bias in language phenomena according to genres or domains of text. Therefore, it can be said that the written corpus represents more representative of the language phenomenon than the spoken corpus composed of public dialogue, public monologue, and private dialogue.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Consistency Validation", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "The quality of the silver set automatically annotated to a well-balanced corpus does not significantly affect the result of the annotation consistency validation. The goal of annotation consistency validation is to verify that the evaluation set shows the annotation characteristics of the annotation model learned by the gold set. As long as it is annotated silver set as a model that properly trains the gold set, it is evaluated that it does not matter if the agreement rate between the silver set and the valuation set is low. However, when comparing the measured value of the agreement divided into 10 sections with the average value, by setting the confidence interval of the average value to 99% (confidence \u03b1 = 0.01), the evaluation standard of annotation consistency for each section was generously set. Also, if the section is further subdivided into 10 or more, more accurate annotation consistency validation will be possible.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Consistency Validation", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "As an example, when the results of semantic role corpus annotation annotation consistency validation were analyzed in detail, each section showed a maximum difference of 0.69 from the average (in Table 3 , 60.933 \u2264 CI \u2264 61.455, 60.78 \u2264 AgreementRate \u2264 61.91). Although it was recorded that it was out of the 99% confidence interval in two sections, the length of the CI was 0.522, which was shorter after the written named entity corpus annotation, and could be evaluated as showing stable annotation consistency. Also, the agreement between the silver set and the evaluation set constructed with the automatic annotation model trained with the gold set is 61.221, but when comparing the sample annotation corpus and the silver set, the consistency was 66.43. It could be judge indirectly as having no significant effect on annotation consistency.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 196, |
| "end": 203, |
| "text": "Table 3", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Consistency Validation", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "A typical inconsistency was mis-annotation on exceptions (Table 4) . In semantic role annotation, it consists of the cases on the adverbs of Korean that share a root with a specific verb, auxiliary verbs that composes a predicate in combination with the main verb, and verbs that are a part of periphrastic constructions or tagmeme equivalents. Some Korean verbs function as a marker of aspect, modal, and negation in predication or used as an element to form multi-word periphrastic construction. In particular, the verbs included in the periphrastic constructions are characterized by: 1) do not affect the content of the proposition, only play grammatical functions, 2) the use is morphologically fixed, and 3) it cannot form a sentential predicate or is not related to the argument structure. E1 is a disagreement caused by the annotator mistaken complementation for relativization, E2 is a disagreement on multi-word periphrastic construction '-e daeha-' (about/on that), E3 is a disagreement on adverbs that share a root with a specific verb, E4 is a disagreement on an auxiliary verb for negation.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 57, |
| "end": 66, |
| "text": "(Table 4)", |
| "ref_id": "TABREF6" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Case study", |
| "sec_num": "5.3" |
| }, |
| { |
| "text": "In this paper, we propose and implement a methodology for constructing language resources for NLP tasks quickly and efficiently for goals of annotation project, but also try to achieve an appropriate level of corpus annotation quality assurance. We designed the constructing process of gold set to consider agreement within annotators when the results from two annotators for one annotation unit match or not, the annotation contents in the annotator group were once again annotated so that the annotation contents were cross-checked and confirmed. This method is a simple and reliable method to check the difference in the subjectivity of the annotator in a short time. In particular, because Korean has the properties of agglutinative language, it has the possibility that a single annotation unit can be interpreted in multiple meanings, so it is necessary to carefully consider the context information surrounding the annotation units. Even though the annotators have annotated deliberately in compliance with the annotation guideline, there are many possibilities for annotation due to semantic diversity, ambiguity, or subjectivity of annotators.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussions and Conclusions", |
| "sec_num": "6" |
| }, |
| { |
| "text": "To construct a gold set in a short time and use it to validate the evaluation set, the quality and authority of the gold set are always important. That is why we designed the process of determining the appropriate annotation minimizing annotation bias while comparing the annotation results within or among annotator groups. Furthermore, we use the gold set as training data of the annotation model to annotate the silver set. If this method is more elaborately, it could be an alternative to evaluating the quality of a corpus annotation when all gold set corresponding to the evaluation set could not be made.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussions and Conclusions", |
| "sec_num": "6" |
| }, |
| { |
| "text": "It is difficult for everyone to interpret identically the same linguistic phenomenon due to environmental or individual aspect. Also, it is hard to say that the gold set is always correct. Therefore, this paper tried to aim to present a method to reduce individual and group bias when constructing corpus annotation. In future research, we try to further generalize the methodology for constructing and validating corpus annotation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussions and Conclusions", |
| "sec_num": "6" |
| }, |
| { |
| "text": "The annotators by layers is a group of experts with a master's degree or Ph.D. of the Korean language or linguistics. They were qualified as annotators by NIKL before corpus annotation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "We were supported by the Exobrain Korean Language Analysis Toolkit v3.0 developed by the Electronics and Telecommunications Research Institute (ETRI) to automatically annotate the raw corpus.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "This work is written based on the results of the 'Corpus Integration Validation'(NIKL 2019-01-61) project of the ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Korean semantic role labeling with highway bilstmcrfs", |
| "authors": [ |
| { |
| "first": "Jangseong", |
| "middle": [], |
| "last": "Bae", |
| "suffix": "" |
| }, |
| { |
| "first": "Changki", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Hyunki", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Annual Conference on Human and Language Technology", |
| "volume": "", |
| "issue": "", |
| "pages": "159--162", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jangseong Bae, Changki Lee, and Hyunki Kim. 2017. Korean semantic role labeling with highway bilstm- crfs. In Annual Conference on Human and Language Technology, pages 159-162. Human and Language Technology.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Crowdsourcing semantic label propagation in relation classification", |
| "authors": [ |
| { |
| "first": "Anca", |
| "middle": [], |
| "last": "Dumitrache", |
| "suffix": "" |
| }, |
| { |
| "first": "Lora", |
| "middle": [], |
| "last": "Aroyo", |
| "suffix": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Welty", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the First Workshop on Fact Extraction and VERification (FEVER)", |
| "volume": "", |
| "issue": "", |
| "pages": "16--21", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Anca Dumitrache, Lora Aroyo, and Chris Welty. 2018. Crowdsourcing semantic label propagation in relation classification. In Proceedings of the First Workshop on Fact Extraction and VERification (FEVER), pages 16-21, Brussels, Belgium, November. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Crowdsourcing disagreement for collecting semantic annotation", |
| "authors": [ |
| { |
| "first": "Anca", |
| "middle": [], |
| "last": "Dumitrache", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "European Semantic Web Conference", |
| "volume": "", |
| "issue": "", |
| "pages": "701--710", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Anca Dumitrache. 2015. Crowdsourcing disagreement for collecting semantic annotation. In European Se- mantic Web Conference, pages 701-710. Springer.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Overview of annotation creation: Processes and tools", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Mark", |
| "suffix": "" |
| }, |
| { |
| "first": "Toma\u017e", |
| "middle": [], |
| "last": "Finlayson", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Erjavec", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Handbook of Linguistic Annotation", |
| "volume": "", |
| "issue": "", |
| "pages": "167--191", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mark A Finlayson and Toma\u017e Erjavec. 2017. Overview of annotation creation: Processes and tools. In Handbook of Linguistic Annotation, pages 167-191. Springer.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Ontonotes: the 90% solution", |
| "authors": [ |
| { |
| "first": "Eduard", |
| "middle": [], |
| "last": "Hovy", |
| "suffix": "" |
| }, |
| { |
| "first": "Mitch", |
| "middle": [], |
| "last": "Marcus", |
| "suffix": "" |
| }, |
| { |
| "first": "Martha", |
| "middle": [], |
| "last": "Palmer", |
| "suffix": "" |
| }, |
| { |
| "first": "Lance", |
| "middle": [], |
| "last": "Ramshaw", |
| "suffix": "" |
| }, |
| { |
| "first": "Ralph", |
| "middle": [], |
| "last": "Weischedel", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the human language technology conference of the NAACL", |
| "volume": "", |
| "issue": "", |
| "pages": "57--60", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Eduard Hovy, Mitch Marcus, Martha Palmer, Lance Ramshaw, and Ralph Weischedel. 2006. Ontonotes: the 90% solution. In Proceedings of the human lan- guage technology conference of the NAACL, Compan- ion Volume: Short Papers, pages 57-60.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Introduction: The handbook of linguistic annotation", |
| "authors": [ |
| { |
| "first": "Nancy", |
| "middle": [], |
| "last": "Ide", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Handbook of Linguistic Annotation", |
| "volume": "", |
| "issue": "", |
| "pages": "1--18", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nancy Ide. 2017. Introduction: The handbook of linguis- tic annotation. In Handbook of Linguistic Annotation, pages 1-18. Springer.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Bert for coreference resolution: Baselines and analysis", |
| "authors": [ |
| { |
| "first": "Mandar", |
| "middle": [], |
| "last": "Joshi", |
| "suffix": "" |
| }, |
| { |
| "first": "Omer", |
| "middle": [], |
| "last": "Levy", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Daniel", |
| "suffix": "" |
| }, |
| { |
| "first": "Luke", |
| "middle": [], |
| "last": "Weld", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Zettlemoyer", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1908.09091" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mandar Joshi, Omer Levy, Daniel S Weld, and Luke Zettlemoyer. 2019. Bert for coreference res- olution: Baselines and analysis. arXiv preprint arXiv:1908.09091.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Korean semantic role labeling using structured svm", |
| "authors": [ |
| { |
| "first": "Changki", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Soojong", |
| "middle": [], |
| "last": "Lim", |
| "suffix": "" |
| }, |
| { |
| "first": "Hyunki", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Journal of KIISE", |
| "volume": "42", |
| "issue": "2", |
| "pages": "220--226", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Changki Lee, Soojong Lim, and Hyunki Kim. 2015. Korean semantic role labeling using structured svm. Journal of KIISE, 42(2):220-226.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "End-to-end sequence labeling via bi-directional lstm-cnns-crf", |
| "authors": [ |
| { |
| "first": "Xuezhe", |
| "middle": [], |
| "last": "Ma", |
| "suffix": "" |
| }, |
| { |
| "first": "Eduard", |
| "middle": [], |
| "last": "Hovy", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1603.01354" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Xuezhe Ma and Eduard Hovy. 2016. End-to-end se- quence labeling via bi-directional lstm-cnns-crf. arXiv preprint arXiv:1603.01354.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Enhancing universal dependencies for korean", |
| "authors": [ |
| { |
| "first": "Youngbin", |
| "middle": [], |
| "last": "Noh", |
| "suffix": "" |
| }, |
| { |
| "first": "Jiyoon", |
| "middle": [], |
| "last": "Han", |
| "suffix": "" |
| }, |
| { |
| "first": "Hansaem", |
| "middle": [], |
| "last": "Tae Hwan Oh", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the Second Workshop on Universal Dependencies", |
| "volume": "", |
| "issue": "", |
| "pages": "108--116", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Youngbin Noh, Jiyoon Han, Tae Hwan Oh, and Hansaem Kim. 2018. Enhancing universal dependencies for ko- rean. In Proceedings of the Second Workshop on Uni- versal Dependencies (UDW 2018), pages 108-116.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "How reliable are annotations via crowdsourcing: a study about interannotator agreement for multi-label image annotation", |
| "authors": [ |
| { |
| "first": "Stefanie", |
| "middle": [], |
| "last": "Nowak", |
| "suffix": "" |
| }, |
| { |
| "first": "Stefan", |
| "middle": [], |
| "last": "R\u00fcger", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the international conference on Multimedia information retrieval", |
| "volume": "", |
| "issue": "", |
| "pages": "557--566", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Stefanie Nowak and Stefan R\u00fcger. 2010. How reliable are annotations via crowdsourcing: a study about inter- annotator agreement for multi-label image annotation. In Proceedings of the international conference on Mul- timedia information retrieval, pages 557-566.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "A crowdsourced corpus of multiple judgments and disagreement on anaphoric interpretation", |
| "authors": [ |
| { |
| "first": "Massimo", |
| "middle": [], |
| "last": "Poesio", |
| "suffix": "" |
| }, |
| { |
| "first": "Jon", |
| "middle": [], |
| "last": "Chamberlain", |
| "suffix": "" |
| }, |
| { |
| "first": "Silviu", |
| "middle": [], |
| "last": "Paun", |
| "suffix": "" |
| }, |
| { |
| "first": "Juntao", |
| "middle": [], |
| "last": "Yu", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexandra", |
| "middle": [], |
| "last": "Uma", |
| "suffix": "" |
| }, |
| { |
| "first": "Udo", |
| "middle": [], |
| "last": "Kruschwitz", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "1", |
| "issue": "", |
| "pages": "1778--1789", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Massimo Poesio, Jon Chamberlain, Silviu Paun, Juntao Yu, Alexandra Uma, and Udo Kruschwitz. 2019. A crowdsourced corpus of multiple judgments and dis- agreement on anaphoric interpretation. In Proceedings of the 2019 Conference of the North American Chap- ter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 1778-1789.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Natural Language Annotation for Machine Learning: A guide to corpus-building for applications", |
| "authors": [ |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Pustejovsky", |
| "suffix": "" |
| }, |
| { |
| "first": "Amber", |
| "middle": [], |
| "last": "Stubbs", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "James Pustejovsky and Amber Stubbs. 2012. Natural Language Annotation for Machine Learning: A guide to corpus-building for applications. \" O'Reilly Media, Inc.\".", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Udpipe: trainable pipeline for processing conll-u files performing tokenization, morphological analysis, pos tagging and parsing", |
| "authors": [ |
| { |
| "first": "Milan", |
| "middle": [], |
| "last": "Straka", |
| "suffix": "" |
| }, |
| { |
| "first": "Jan", |
| "middle": [], |
| "last": "Hajic", |
| "suffix": "" |
| }, |
| { |
| "first": "Jana", |
| "middle": [], |
| "last": "Strakov\u00e1", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC'16)", |
| "volume": "", |
| "issue": "", |
| "pages": "4290--4297", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Milan Straka, Jan Hajic, and Jana Strakov\u00e1. 2016. Ud- pipe: trainable pipeline for processing conll-u files performing tokenization, morphological analysis, pos tagging and parsing. In Proceedings of the Tenth In- ternational Conference on Language Resources and Evaluation (LREC'16), pages 4290-4297.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Ontonotes release 5.0 ldc2013t19. Linguistic Data Consortium", |
| "authors": [ |
| { |
| "first": "Ralph", |
| "middle": [], |
| "last": "Weischedel", |
| "suffix": "" |
| }, |
| { |
| "first": "Martha", |
| "middle": [], |
| "last": "Palmer", |
| "suffix": "" |
| }, |
| { |
| "first": "Mitchell", |
| "middle": [], |
| "last": "Marcus", |
| "suffix": "" |
| }, |
| { |
| "first": "Eduard", |
| "middle": [], |
| "last": "Hovy", |
| "suffix": "" |
| }, |
| { |
| "first": "Sameer", |
| "middle": [], |
| "last": "Pradhan", |
| "suffix": "" |
| }, |
| { |
| "first": "Lance", |
| "middle": [], |
| "last": "Ramshaw", |
| "suffix": "" |
| }, |
| { |
| "first": "Nianwen", |
| "middle": [], |
| "last": "Xue", |
| "suffix": "" |
| }, |
| { |
| "first": "Ann", |
| "middle": [], |
| "last": "Taylor", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeff", |
| "middle": [], |
| "last": "Kaufman", |
| "suffix": "" |
| }, |
| { |
| "first": "Michelle", |
| "middle": [], |
| "last": "Franchini", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ralph Weischedel, Martha Palmer, Mitchell Marcus, Ed- uard Hovy, Sameer Pradhan, Lance Ramshaw, Nian- wen Xue, Ann Taylor, Jeff Kaufman, Michelle Fran- chini, et al. 2013. Ontonotes release 5.0 ldc2013t19. Linguistic Data Consortium, Philadelphia, PA, 23.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "text": "The flow of the corpus annotation and validation process in this paper. Blue-coloureds indicate our process and result, and green-coloureds indicate evaluatee group's process and result.", |
| "num": null, |
| "type_str": "figure", |
| "uris": null |
| }, |
| "FIGREF1": { |
| "text": "The flow of construction procedure. This figure is a detailed representation of the part of the Human-machine collaborative corpus annotation inFigure 1.", |
| "num": null, |
| "type_str": "figure", |
| "uris": null |
| }, |
| "TABREF1": { |
| "text": "Comparison of corpus specification of gold set, evaluation set and The OntoNotes 5.0. Eojeol is a unit of word segmentation of Korean.", |
| "html": null, |
| "num": null, |
| "type_str": "table", |
| "content": "<table/>" |
| }, |
| "TABREF3": { |
| "text": "Results of annotation contents validation.", |
| "html": null, |
| "num": null, |
| "type_str": "table", |
| "content": "<table/>" |
| }, |
| "TABREF4": { |
| "text": "85.87 86.38 86.09 86.51 86.60 86.44 86.33 86.52 86.49 86.298 0.533 MP Spoken Accuracy 75.35 73.75 76.98 79.30 78.28 79.24 77.40 78.99 80.16 79.32 77.877 3.681 LS Written Accuracy 87.64 86.81 87.4 87.45 87.81 87.99 87.47 87.95 87.41 88.14 87.607 0.703 LS Spoken Accuracy 77.72 79.31 79.51 79.86 78.89 79.2 79.66 78.37 79.09 80.75 79.236 1.498 NE Written Accuracy 22.04 22.10 21.99 22.58 22.1 22.18 22.26 21.97 22.37 22.69 22.228 MUC) 51.32 50.08 50.34 49.66 51.13 51.23 51.36 50.91 49.77 51.09 50.689 1.201 CR Spoken F1 (MUC) 36.39 33.93 33.71 38.42 42.07 33.57 38.25 34.20 33.89 32.34 35.", |
| "html": null, |
| "num": null, |
| "type_str": "table", |
| "content": "<table><tr><td>Layer</td><td>Measures</td><td>1</td><td>2</td><td>3</td><td>4</td><td>5</td><td>6</td><td>7</td><td>8</td><td>9</td><td>10</td><td>Avg.</td><td>CI Length</td></tr><tr><td colspan=\"2\">MP Written Accuracy</td><td colspan=\"12\">85.87 0.449</td></tr><tr><td>NE Spoken</td><td>Accuracy</td><td colspan=\"11\">27.45 20.03 20.10 20.54 21.62 20.55 22.54 19.15 22.06 21.20 21.524</td><td>4.206</td></tr><tr><td>ZA Written</td><td>F1</td><td>13.98</td><td colspan=\"3\">8.55 20.84 38.75</td><td colspan=\"7\">29.3 33.35 31.41 33.31 22.82 31.34 26.365</td><td>17.353</td></tr><tr><td>ZA Spoken</td><td>F1</td><td colspan=\"11\">23.70 22.78 22.72 24.01 23.71 24.08 22.80 23.07 22.81 24.02 23.370</td><td>1.058</td></tr><tr><td colspan=\"13\">CR Written F1 (677</td><td>5.524</td></tr><tr><td>DP Written</td><td>UAS</td><td colspan=\"11\">68.37 68.77 68.11 68.68 68.43 69.00 68.92 69.03 69.73 67.18 68.622</td><td>1.224</td></tr><tr><td>SR Spoken</td><td>F1 (AI)</td><td colspan=\"11\">61.91 61.30 61.14 61.34 61.06 61.10 60.78 61.15 61.25 61.18 61.221</td><td>0.522</td></tr></table>" |
| }, |
| "TABREF6": { |
| "text": "SR example of mis-annotation on a sentence which means 'It is not right to say so about what has already passed.'", |
| "html": null, |
| "num": null, |
| "type_str": "table", |
| "content": "<table/>" |
| } |
| } |
| } |
| } |