| { |
| "paper_id": "Y18-1013", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T13:36:38.774067Z" |
| }, |
| "title": "Automatic Evaluation of English-to-Korean and Korean-to-English Neural Machine Translation Systems by Linguistic Test Points", |
| "authors": [ |
| { |
| "first": "Sung-Kwon", |
| "middle": [], |
| "last": "Choi", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "Language Intelligence Research Group Electronics", |
| "institution": "Telecommunications Research Institute", |
| "location": { |
| "settlement": "Daejeon", |
| "country": "Korea" |
| } |
| }, |
| "email": "choisk@etri.re.kr" |
| }, |
| { |
| "first": "Gyu-Hyeun", |
| "middle": [], |
| "last": "Choi", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "Language Intelligence Research Group Electronics", |
| "institution": "Telecommunications Research Institute", |
| "location": { |
| "settlement": "Daejeon", |
| "country": "Korea" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Youngkil", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "Language Intelligence Research Group Electronics", |
| "institution": "Telecommunications Research Institute", |
| "location": { |
| "settlement": "Daejeon", |
| "country": "Korea" |
| } |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "BLEU is the most well-known automatic evaluation technology in assessing the performance of machine translation systems. However, BLEU does not know which parts of the NMT translation results are good or bad. This paper describes the automatic evaluation approach of NMT systems by linguistic test points. This approach allows automatic evaluation of each linguistic test point not shown in BLEU and provides intuitive insight into the strengths and flaws of NMT systems in handling various important linguistic test points. The linguistic test points used for automatic evaluation were 58 and consisted of 630 sentences. We conducted the evaluation of two bidirectional English/Korean NMT systems. BLEUs of English-to-Korean NMT systems were 0.0898 and 0.2081 respectively, and their automatic evaluations by linguistic test points were 58.35% and 77.31%, respectively. BLEUs of Korean-to-English NMT systems were 0.3939 and 0.4512 respectively, and their automatic evaluations by linguistic test points were 33.10% and 40.47%, respectively. This means that the automatic evaluation approach by linguistic test points has similar results as BLEU assessment. According to automatic evaluation by linguistic test points, we know that both English-to-Korean NMT systems and Korean-to-English NMT systems have strengths in polysemy translations, but has flaws in style translations and translations of sentences with complex syntactic structures.", |
| "pdf_parse": { |
| "paper_id": "Y18-1013", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "BLEU is the most well-known automatic evaluation technology in assessing the performance of machine translation systems. However, BLEU does not know which parts of the NMT translation results are good or bad. This paper describes the automatic evaluation approach of NMT systems by linguistic test points. This approach allows automatic evaluation of each linguistic test point not shown in BLEU and provides intuitive insight into the strengths and flaws of NMT systems in handling various important linguistic test points. The linguistic test points used for automatic evaluation were 58 and consisted of 630 sentences. We conducted the evaluation of two bidirectional English/Korean NMT systems. BLEUs of English-to-Korean NMT systems were 0.0898 and 0.2081 respectively, and their automatic evaluations by linguistic test points were 58.35% and 77.31%, respectively. BLEUs of Korean-to-English NMT systems were 0.3939 and 0.4512 respectively, and their automatic evaluations by linguistic test points were 33.10% and 40.47%, respectively. This means that the automatic evaluation approach by linguistic test points has similar results as BLEU assessment. According to automatic evaluation by linguistic test points, we know that both English-to-Korean NMT systems and Korean-to-English NMT systems have strengths in polysemy translations, but has flaws in style translations and translations of sentences with complex syntactic structures.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Currently, the performance of NMT system is rapidly advancing. Its performance proved to be superior to that of SMT (Bojar et. al., 2016) . The NMT is also applied to both English-to-Korean and Korean-to-English machine translation systems for commercial service. The most wellknown machine translation evaluation technique in assessing the performance of NMT systems is BLEU (Papineni et. al., 2002) . It is strength of BLEU that automatic scores for MT output can be provided in cases where there are existing reference translations by calculating similarity between the MT output and the references. Faults of BLEU are that it does not provide insight into the specific nature of problems encountered in the translation output and scores are tied to the particularities of the reference translations (Lommel et. al., 2014) . By BLEU, developers and users cannot identify which part of the NMT translation result is vulnerable.", |
| "cite_spans": [ |
| { |
| "start": 116, |
| "end": 137, |
| "text": "(Bojar et. al., 2016)", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 376, |
| "end": 400, |
| "text": "(Papineni et. al., 2002)", |
| "ref_id": null |
| }, |
| { |
| "start": 803, |
| "end": 825, |
| "text": "(Lommel et. al., 2014)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "We propose an automatic evaluation approach of neural machine translation systems by linguistic test points. Instead of assigning a general score to an NMT system we conduct an automatic evaluation by each linguistic test point not shown in BLEU. This automatic evaluation approach of this paper can give developers an intuitive insight into the strengths and flaws of NMT systems. Also, the automatic evaluation method by linguistic test points, which is not like BLEU, may provide objective evaluation even without reference sentences.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Section 2 describes existing studies related to the automatic evaluation approach by linguistic test points. Section 3 introduces the design of test set including linguistic test points. In Section 4, we explain the results of automatic evaluation by linguistic test points and analyze the strengths and flaws of two bidirectional English/Korean NMT systems.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Approaches for evaluating machine translation systems can be divided into automatic and manual assessment. An automatic evaluation is the automatic scoring of a machine translation result by calculating the similarity between the machine translation result and the reference. BLEU is representative of this automatic evaluation method. It has the strength of objectively assessing the results of machine translation. However, BLEU is dependent on reference sentences and cannot point to translation errors in machine translation results. Manual evaluation method is for human translators to assign scores to the results of machine translations according to the evaluation criteria. It has the strength of being able to assess the results of machine translations precisely, but depends on human evaluators and is costly and timeconsuming.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "In order to identify the strengths and weaknesses of the machine translation system, previous studies have introduced approaches of using linguistic test set for evaluation purposes. The process for constructing a test set including the linguistic test points can be described as following steps: (Balkan et al., 1994) . Most of test sets were composed of language pairs in similar language family (Bentivogli et. al., 2016; Isabelle et. al., 2017) . Among test sets with the language pairs in heterogeneous language family was Koh (2001) , which was related to Korean, and was consisted of structure part and selection part. While these test sets have been manually constructed, Zhou (2008) has introduced how to automatically build the test set by parser and word aligner. Test set for grasping the strengths and faults of NMT systems started in 2016. Bentivogli (2016) used the English-German test set of IWSLT 2015 to compare PBMT with NMT. Isabelle (2017) constructed 108 English-French test sentences and evaluated them according to the Yes/No questions in order to identify strengths and weaknesses of English-to-French NMT systems. Guillou (2016) established a test set to conduct an assessment on the English-to-French machine translation of pronouns 'it' and 'they' in the DiscoMT 2015 shared task. However, the existing evaluation of machine translation system using test set was focused on three linguistic categories: morpho-syntactic divergences, lexico-syntactic divergences, and syntactic divergences. Unlike previous studies, this paper describes a test set and an evaluation method to evaluate various linguistic phenomena.", |
| "cite_spans": [ |
| { |
| "start": 297, |
| "end": 318, |
| "text": "(Balkan et al., 1994)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 398, |
| "end": 424, |
| "text": "(Bentivogli et. al., 2016;", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 425, |
| "end": 448, |
| "text": "Isabelle et. al., 2017)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 528, |
| "end": 538, |
| "text": "Koh (2001)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 854, |
| "end": 871, |
| "text": "Bentivogli (2016)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 1140, |
| "end": 1154, |
| "text": "Guillou (2016)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "\uf06c Design", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "A linguistic test point is a linguistically motivated unit, which is pre-defined in test set for automatic evaluation. We attempted to collect a variety of linguistic test points that can target at identifying of the strengths and weaknesses of the neural machine translation systems. For this purpose, the linguistic test points related to part-of-speeches, syntactic structures, semantic relations, and target word selection were manually collected from the grammar books. The linguistic test points can be divided into the structure part related to source sentence of source language and the selection part related to target words of target language. They are subdivided into depth of 3. Table 1 ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 691, |
| "end": 698, |
| "text": "Table 1", |
| "ref_id": "TABREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Taxonomy of Linguistic Test Points", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Two English-to-Korean NMT systems used in the experiment are well known for their automatic translation commercialization services in Korea. BLEU and the result of automatic evaluation by linguistic test points are as follows: Table 2 , the results of automatic assessment of NMT2 are 16.41 % and 7.37% higher than those of NMT1, respectively. The automatic evaluation by linguistic test points shows the similar results in the BLEU and NIST. That is, BLEU and NIST of NMT2 in English-to-Korean NMT systems were higher 0.1183 and 1.4093 than those of NMT1, respectively. This shows that the automatic evaluation by linguistic test points is similar to the automatic translation metrics. Thus, we can identify the strengths and weaknesses of NMTs by building simple <target words> for the automatic evaluation by linguistic test points without having to create a lot of reference sentences. Table 3 , the automatic evaluation by linguistic test points of NMT2 is better overall than that of NMT1. The biggest difference between NMT2 and NMT1 was ambiguity. Among ambiguities, structural ambiguity was the biggest difference. The linguistic test points in which NMT1 and NMT2 received low ratings were style, phrase and clause, part-of-speech, and sentence. In the NMT2, machine translation of style was weakest, followed by phrase and clause, part-ofspeech, sentence and collocation. In the NMT1, machine translation of style was also weakest, followed by phrase and clause, ambiguity, collocation, sentence, and part-of-speech. Table 4 shows the detailed results of an automatic evaluation by linguistic test points. It is difficult to talk about \"typical\" Americans. <source words> \"typical\" <target words> \"\uc804\ud615\uc801\uc778\", \"\uc77c\ubc18\uc801\uc778\" <NMT1> \"\uc804\ud615\uc801\uc778\" \ubbf8\uad6d\uc778\uc5d0 \ub300\ud574\uc11c \uc774\uc57c\uae30\ud558\ub294 \uac83\uc740 \uc5b4\ub835\uc2b5\ub2c8\ub2e4 // Evaluation: 1 point <NMT2> \uc804\ud615\uc801\uc778 \ubbf8\uad6d\uc778\ub4e4\uc5d0 \ub300\ud574 \ub9d0\ud558\ub294 \uac83\uc740 \uc5b4\ub835\ub2e4. // Evaluation: 0 point", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 227, |
| "end": 234, |
| "text": "Table 2", |
| "ref_id": "TABREF4" |
| }, |
| { |
| "start": 890, |
| "end": 897, |
| "text": "Table 3", |
| "ref_id": "TABREF5" |
| }, |
| { |
| "start": 1528, |
| "end": 1535, |
| "text": "Table 4", |
| "ref_id": "TABREF7" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Automatic Evaluation by Linguistic Test Points of English-to-Korean NMT Systems", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "# of", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Automatic Evaluation by Linguistic Test Points of English-to-Korean NMT Systems", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "Points of Korean-to-English NMT Systems BLEU and NIST of NMT2 in Korean-to-English NMT systems are higher 0.0573 and 0.5611 than those of NMT1, respectively. As the evaluation in English-to-Korean NMT systems, the automatic evaluation by linguistic test points in Korean-to-English NMT shows the similar evaluation conducted by the automatic translation metrics. Table 6 indicate the strengths and flaws of NMT1 and NMT2 identified by subcategories: Table 7 shows the detailed results of an automatic evaluation by linguistic test points. Table 7 indicates that NMT1 needs to correct the linguistic test points such as article, participial construction, parenthesis, structural ambiguity and natural translation. NMT2 needs to deal with the linguistic test points such as article, relative pronoun, participial construction, subjunctive, parenthesis, structural ambiguity and natural translation. Some examples of automatic evaluation about number and natural translation in Korean-to-English NMT Systems are as follows.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 363, |
| "end": 370, |
| "text": "Table 6", |
| "ref_id": "TABREF10" |
| }, |
| { |
| "start": 450, |
| "end": 457, |
| "text": "Table 7", |
| "ref_id": null |
| }, |
| { |
| "start": 539, |
| "end": 546, |
| "text": "Table 7", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Automatic Evaluation by Linguistic Test", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "\uf06c Number <source sentence> \uc9d1 \uacc4\uc57d\uae08\uc73c\ub85c 1,000 \ub9cc\uc6d0\uc744 \uc9c0\ubd88\ud588\uc5b4\uc694. // I paid ten million won down on the house. <source words> 1,000 \ub9cc\uc6d0 <target words> 10 million won, ten million won <NMT1> I paid 10 million won for the house deposit. // Evaluation: 1 point <NMT2> I paid ten million won down on the house. // Evaluation: 1 point \uf06c Natural translation <source sentence> \uadf8 \uc0ac\uace0\ub294 \uadf8\uc758 \ubd80\uc8fc\uc758\ub85c \uc77c\uc5b4\ub0ac\ub2e4. // His carelessness caused the accident. <source words> \ubd80\uc8fc\uc758\ub85c \uc77c\uc5b4\ub0ac <target words> carelessness caused <NMT1> The accident happened due to his carelessness. // Evaluation: 0 point <NMT2> The accident was brought about by his carelessness. // Evaluation: 0 point", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Automatic Evaluation by Linguistic Test", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "In this paper, we introduced the automatic evaluation approach by linguistic test points for NMT systems. In the automatic evaluation by linguistic test points, the scoring method was as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| }, |
| { |
| "text": "1) If a <source sentence> of a linguistic test point is automatically translated by an NMT system and <target words> of a linguistic test point are in the translation results of the NMT system, give 1 point.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| }, |
| { |
| "text": "2) If a <source sentence> of a linguistic test point is automatically translated by an NMT system and <#target words> of a linguistic test point are in the translation results of the NMT system, give 0.5 point.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| }, |
| { |
| "text": "3) If a <source sentence> of a linguistic test point is automatically translated by an NMT system and <target words> of a linguistic test point are not in the translation results of the NMT system, give 0 point.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| }, |
| { |
| "text": "In accordance with the automatic evaluation by linguistic test points, the strengths and weaknesses of two bidirectional English/Korean NMT systems were identified. According to the automatic evaluation by linguistic test points, the English-to-Korean NMT systems need to correct first the subcategories with linguistic test points such as style, phrase and clause, part-of-speech, and sentence. The Korean-to-English NMT systems need to deal with first the subcategories with linguistic test points such as style, ambiguity and phrase/clause.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| }, |
| { |
| "text": "According to the experiments, we could identify the strengths and weaknesses of NMT systems from the automatic evaluation by linguistic test points without having to create a lot of reference sentences.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| }, |
| { |
| "text": "Papineni, K., Roukos, S., Ward, T., and Zhu, W.-J. 2002.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| }, |
| { |
| "text": "Bleu: a method for automatic evaluation of machine translation. In Proceedings of 40th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Philadelphia, Pennsylvania, USA: 311-318.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| }, |
| { |
| "text": "Zhou, M., Wang, B., Liu, S., Li, M., Zhang, D., and Zhao T. 2008 ", |
| "cite_spans": [ |
| { |
| "start": 10, |
| "end": 64, |
| "text": "Wang, B., Liu, S., Li, M., Zhang, D., and Zhao T. 2008", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| }, |
| { |
| "text": "32nd Pacific Asia Conference on Language, Information and Computation Hong Kong, 1-3 December 2018Copyright 2018 by the authors", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "TSNLP -test suites for natural language processing", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Balkan", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Netter", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Arnold", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Meijer", |
| "suffix": "" |
| } |
| ], |
| "year": 1994, |
| "venue": "Proceedings of the Language Engineering Convention", |
| "volume": "", |
| "issue": "", |
| "pages": "17--22", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Balkan, L., Netter, K., Arnold, D. and Meijer, S. 1994. TSNLP -test suites for natural language processing, Proceedings of the Language Engineering Convention, ELSNET, Paris: 17-22.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Neural versus phrase-based machine translation quality: a case study", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Bentivogli", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Bisazza", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Cettolo", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Federico", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "257--267", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bentivogli, L., Bisazza, A., Cettolo, M., and Federico, M. 2016. Neural versus phrase-based machine translation quality: a case study. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, Austin, Texas: 257-267.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Pacific Asia Conference on Language, Information and Computation Hong Kong", |
| "authors": [], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Pacific Asia Conference on Language, Information and Computation Hong Kong, 1-3 December 2018 Copyright 2018 by the authors", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Findings of the 2016 Conference on Machine Translation. ACL 2016 First Conference on Machine Translation (WMT16). The Association for Computational Linguistics", |
| "authors": [ |
| { |
| "first": "O", |
| "middle": [], |
| "last": "Bojar", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Chatterjee", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Federmann", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Graham", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Haddow", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Huck", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [ |
| "J" |
| ], |
| "last": "Yepes", |
| "suffix": "" |
| }, |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| }, |
| { |
| "first": "V", |
| "middle": [], |
| "last": "Logacheva", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Monz", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Negri", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "N\u00e9v\u00e9ol", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Neves", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Popel", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Post", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Rubino", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Scarton", |
| "suffix": "" |
| }, |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Specia", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Turchi", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Verspoor", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Zampieri", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "131--198", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bojar, O., Chatterjee, R., Federmann, C., Graham, Y., Haddow, B., Huck, M., Yepes, A.J., Koehn, P., Logacheva, V., Monz, C., Negri, M., N\u00e9v\u00e9ol, A., Neves, M., Popel, M., Post, M., Rubino, R., Scarton, C., Specia, L., Turchi, M., Verspoor, K., and Zampieri, M. 2016. Findings of the 2016 Conference on Machine Translation. ACL 2016 First Conference on Machine Translation (WMT16). The Association for Computational Linguistics: 131-198.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "PROTEST: A Test Suite for Evaluating Pronouns in Machine Translation", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Guillou", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Hardmeier", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC", |
| "volume": "", |
| "issue": "", |
| "pages": "636--643", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Guillou, L. and Hardmeier, C. 2016. PROTEST: A Test Suite for Evaluating Pronouns in Machine Translation. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC 2016): 636-643.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "A Challenge Set Approach to Evaluating Machine Translation", |
| "authors": [ |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Isabelle", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Cherry", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Foster", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "2486--2496", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Isabelle, P., Cherry, C., and Foster, G. 2017. A Challenge Set Approach to Evaluating Machine Translation. Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pages: 2486-2496", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "A Test Suite for Evaluation of English-to-Korean Machine Translation Systems", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [ |
| "R" |
| ], |
| "last": "Koh", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [ |
| "N" |
| ], |
| "last": "Maeng", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [ |
| "Y" |
| ], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [ |
| "S" |
| ], |
| "last": "Chae", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [ |
| "S" |
| ], |
| "last": "Choi", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Koh, S. R., Maeng, J. N., Lee, J. Y., Chae, Y. S., and Choi, K. S. 2001. A Test Suite for Evaluation of English-to-Korean Machine Translation Systems. MT Summit:", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Using a New Analytic Measure for the Annotation and Analysis of MT Errors on Real Data", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Lommel", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Burchardt", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Popovic", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Harris", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Avramidis", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Uszkoreit", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 17th Annual Conference of the European Association for Machine Translation", |
| "volume": "", |
| "issue": "", |
| "pages": "165--172", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lommel, A., Burchardt, A., Popovic, M., Harris, K., Avramidis, E., and Uszkoreit, H. 2014. Using a New Analytic Measure for the Annotation and Analysis of MT Errors on Real Data. Proceedings of the 17th Annual Conference of the European Association for Machine Translation: 165-172.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "32nd Pacific Asia Conference on Language, Information and Computation Hong Kong", |
| "authors": [], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "32nd Pacific Asia Conference on Language, Information and Computation Hong Kong, 1-3 December 2018 Copyright 2018 by the authors", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "shows the overall construction of the automatic evaluation approach by linguistic test points.", |
| "num": null |
| }, |
| "FIGREF1": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "Automatic Evaluation Approach by Linguistic Test Points In the Figure 1, Sn refers to the automatically translated result of <source sentence> and Wn, Wi, Wj and Wh mean <target words>. The program will determine the existence of a # and then check whether ~ and // are on the <target words>. If <target words> exist in Sn, give 1 point and otherwise 0 point. If <#target words> exists in Sn, give 0.5 point and otherwise 0 point. If <~target words> exists in Sn, give 0 point and otherwise 1 point. If <target words1//target words2> exists in Sn, give 1 point and otherwise 0 point.", |
| "num": null |
| }, |
| "TABREF1": { |
| "html": null, |
| "content": "<table><tr><td>Category</td><td>Subcategory</td><td/><td colspan=\"2\">Test Points</td><td/><td># of test</td><td># of</td></tr><tr><td/><td/><td/><td/><td/><td/><td>points</td><td>sentences</td></tr><tr><td>Structure</td><td>Part-of-speech</td><td colspan=\"4\">Noun, Adverb, Verb, Conjunction, Symbol,</td><td>29</td><td>290</td></tr><tr><td/><td/><td>Number, etc.</td><td/><td/><td/><td/></tr><tr><td/><td>Phrase and Clause</td><td colspan=\"3\">Gerund, Idiomatic expression, etc.</td><td/><td>10</td><td>100</td></tr><tr><td/><td>Sentence</td><td>Negation,</td><td>Passive,</td><td>Mood,</td><td>Ellipsis,</td><td>14</td><td>140</td></tr><tr><td/><td/><td colspan=\"2\">Coordination, Tense, etc.</td><td/><td/><td/></tr><tr><td>Selection</td><td>Ambiguity</td><td colspan=\"5\">Part-of-speech ambiguity, Structural ambiguity 2</td><td>40</td></tr><tr><td/><td>Collocation</td><td>Collocation</td><td/><td/><td/><td>1</td><td>20</td></tr><tr><td/><td>Polysemy</td><td>Polysemy</td><td/><td/><td/><td>1</td><td>20</td></tr><tr><td/><td>Style</td><td colspan=\"2\">Natural translation</td><td/><td/><td>1</td><td>20</td></tr><tr><td>2</td><td>7</td><td/><td/><td>58</td><td/><td/><td>630</td></tr></table>", |
| "type_str": "table", |
| "text": "shows the taxonomy of linguistic test points.Currently, there are 58 linguistic test points. Each test point of the structure part consists of 10 sentences, and each test point of the selection part is composed of 20 sentences. As a result, the total number of sentences is 630. In practice, new linguistic test points can be easily added.", |
| "num": null |
| }, |
| "TABREF2": { |
| "html": null, |
| "content": "<table><tr><td>3.2 Structure of Linguistic Test Points</td><td><#target words1>\u2026<#target wordsn></td></tr><tr><td>A linguistic test point consists of <source sentence> <source words> and <target words>.</td><td>The <#target words> refers to <target words>, which can be regarded as a <target words> among the results of NMT.</td></tr><tr><td/><td>Ex.) <I could not but get angry.> <could not</td></tr><tr><td/><td>but> <\uc9c0 \uc54a\uace0\ub294 \uc788\uc744 \uc218 \uc5c6\uc5c8> <\uc9c0</td></tr><tr><td/><td>\uc54a\uace0 \uc788\uc744 \uc218 \uc5c6\uc5c8> <\uc9c0 \uc54a\uc744 \uc218</td></tr><tr><td/><td>\uc5c6\uc5c8><#\uc218 \ubc16\uc5d0 \uc5c6\uc5c8></td></tr><tr><td/><td>\uf06c <source sentence> <source words></td></tr><tr><td/><td><target words11//target words12> \u2026</td></tr><tr><td/><td><target wordsn1//target wordsn2></td></tr><tr><td>\uf06c <source sentence> <source words></td><td><target words11//target words12> says that</td></tr><tr><td><target words></td><td><target words> corresponding to <source</td></tr><tr><td>The basic structure of a linguistic test point consists of one <source sentence>, one</td><td>words> can be separated into a <target words11> and a <target words12>.</td></tr><tr><td><source words>, and one <target words>.</td><td>Ex.) <None of the books were interesting></td></tr><tr><td>Ex.) <In these 3 years, 2.5 billion pencils were sold> <2.5 billion> <25 \uc5b5></td><td><none of the books> <\uc5b4\ub290 \ucc45\ub3c4//\uc9c0 \uc54a>, <\ucc45\ub4e4 \uc911 \uc5b4\ub290 \uac83\ub3c4//\uc9c0 \uc54a></td></tr><tr><td>\uf06c <source sentence> <source words> <target words1>\u2026<target wordsn></td><td>\uf06c <source sentence> <source words> <~target words></td></tr><tr><td>A test point may consist of two or more <target words> corresponding to a <source words>.</td><td><~target words> means that <target words> corresponding to <source words> should not appear in results of NMT. In other words, '~' refers to a logical negation among logical</td></tr><tr><td>Ex.) <I could not but get angry> <could not</td><td>connectives.</td></tr><tr><td>but> <\uc9c0 \uc54a\uace0\ub294 \uc788\uc744 \uc218 \uc5c6\uc5c8> <\uc9c0 \uc54a\uace0 \uc788\uc744 \uc218 \uc5c6\uc5c8> <\uc9c0 \uc54a\uc744 \uc218 \uc5c6\uc5c8></td><td>Ex.) <It was September 17> <It> <~\uadf8\uac83></td></tr><tr><td>\uf06c <source sentence> <source words></td><td/></tr></table>", |
| "type_str": "table", |
| "text": "Taxonomy of Linguistic Test Points The <source sentence> refers to sentence of source language that will be translated automatically. The <source words> are a check point words of <source sentence> for linguistic test points of the automatic evaluation. The <target words> indicate the reference of <source words>. The <target words> are extracted from the digital translation dictionary. The linguistic test point has a variety of structures, such as:We have established an automatic evaluation program for automatic evaluation of NMT systems by linguistic test points. With the linguistic test point, the automatic evaluation of an MT system is performed with the following steps: \uf06c For each test point, we calculate the number of matched references against the translated sentence of the NMT system. \uf06c The score of a subcategory can be obtained by summing up the scores of all test points of this subcategory. Then the evaluation of an NMT system can be obtained by summing up the scores of all subcategories.", |
| "num": null |
| }, |
| "TABREF4": { |
| "html": null, |
| "content": "<table><tr><td>: Comparison of BLEU with Automatic</td></tr><tr><td>Evaluation by Linguistic Test Points in English-to-</td></tr><tr><td>Korean NMT Systems</td></tr><tr><td>According to</td></tr></table>", |
| "type_str": "table", |
| "text": "", |
| "num": null |
| }, |
| "TABREF5": { |
| "html": null, |
| "content": "<table><tr><td>indicate the strengths</td></tr></table>", |
| "type_str": "table", |
| "text": "", |
| "num": null |
| }, |
| "TABREF6": { |
| "html": null, |
| "content": "<table/>", |
| "type_str": "table", |
| "text": "Automatic Evaluation by Linguistic Test Points related to Subcategories in English-to-Korean NMT Systems According to", |
| "num": null |
| }, |
| "TABREF7": { |
| "html": null, |
| "content": "<table><tr><td>\uf06c Number</td><td/><td/><td/></tr><tr><td colspan=\"4\"><source sentence> They had three and a half</td></tr><tr><td>dollars.</td><td/><td/><td/></tr><tr><td colspan=\"3\"><source words> three and a half dollars</td><td/></tr><tr><td colspan=\"4\"><target words> 3 \ub2ec\ub7ec 50 \uc13c\ud2b8, 3.5 \ub2ec\ub7ec</td></tr><tr><td><NMT1>\uadf8\ub4e4\uc5d0\uac8c\ub294</td><td>3</td><td>\ub2ec\ub7ec</td><td>\ubc18\uc774</td></tr><tr><td colspan=\"3\">\uc788\uc5c8\uc2b5\ub2c8\ub2e4. // Evaluation: 0 point</td><td/></tr><tr><td colspan=\"4\"><NMT2> \uadf8\ub4e4\uc740 3 \ub2ec\ub7ec 50 \uc13c\ud2b8\ub97c \uac00\uc9c0\uace0</td></tr><tr><td colspan=\"3\">\uc788\uc5c8\ub2e4. // Evaluation: 1 point</td><td/></tr><tr><td>\uf06c Symbol</td><td/><td/><td/></tr><tr><td><source sentence></td><td/><td/><td/></tr></table>", |
| "type_str": "table", |
| "text": "indicates that NMT1 needs to correct the linguistic test points such as number, numeric expression, participial construction, gerund, subjunctive, structural ambiguity, and natural translation. NMT2 needs to revise the linguistic test points such as symbol, participial construction and parenthesis. The following are examples of the automatic evaluation by linguistic test points about number and symbol.", |
| "num": null |
| }, |
| "TABREF8": { |
| "html": null, |
| "content": "<table><tr><td colspan=\"5\">indicates the result of automatic evaluation</td></tr><tr><td colspan=\"5\">by linguistic test points of Korean-to-English NMT</td></tr><tr><td>systems:</td><td/><td/><td/><td/></tr><tr><td>#</td><td>of</td><td>NMT1</td><td>NMT2</td><td>NMT2 -</td></tr><tr><td colspan=\"2\">sentences</td><td/><td/><td>NMT1</td></tr><tr><td>Total</td><td>630</td><td colspan=\"2\">33.10% 40.47%</td><td>7.37</td></tr><tr><td>Structure</td><td>530</td><td colspan=\"2\">33.21% 40.94%</td><td>7.73</td></tr><tr><td>S</td><td>100</td><td colspan=\"2\">33.00% 40.00%</td><td>7</td></tr><tr><td>election</td><td/><td/><td/><td/></tr><tr><td>BLEU</td><td/><td>0.3939</td><td>0.4512</td><td>0.0573</td></tr><tr><td>NIST</td><td/><td>7.1893</td><td>7.7504</td><td>0.5611</td></tr></table>", |
| "type_str": "table", |
| "text": "", |
| "num": null |
| }, |
| "TABREF9": { |
| "html": null, |
| "content": "<table><tr><td>: Comparison of BLEU with Automatic</td></tr><tr><td>Evaluation by Linguistic Test Points in Korean-to-</td></tr><tr><td>English NMT Systems</td></tr></table>", |
| "type_str": "table", |
| "text": "", |
| "num": null |
| }, |
| "TABREF10": { |
| "html": null, |
| "content": "<table><tr><td>Automatic Evaluation by Linguistic Test</td></tr><tr><td>Points related to Subcategories in Korean-to-</td></tr><tr><td>English NMT Systems</td></tr><tr><td>According to Table 6, the Korean-to-English</td></tr><tr><td>NMT systems need to correct first the</td></tr><tr><td>subcategories such as style, ambiguity and</td></tr></table>", |
| "type_str": "table", |
| "text": "", |
| "num": null |
| }, |
| "TABREF12": { |
| "html": null, |
| "content": "<table/>", |
| "type_str": "table", |
| "text": "Detailed Automatic Evaluation by Linguistic Test Points in English-to-Korean NMT Systems", |
| "num": null |
| } |
| } |
| } |
| } |