| { |
| "paper_id": "2020", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T03:41:46.909312Z" |
| }, |
| "title": "HW-TSC's Participation at WMT 2020 Quality Estimation Shared Task", |
| "authors": [ |
| { |
| "first": "Minghan", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Huawei Translation Services Center", |
| "location": { |
| "settlement": "Beijing", |
| "country": "China" |
| } |
| }, |
| "email": "wangminghan@huawei.com" |
| }, |
| { |
| "first": "Hao", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Huawei Translation Services Center", |
| "location": { |
| "settlement": "Beijing", |
| "country": "China" |
| } |
| }, |
| "email": "yanghao30@huawei.com" |
| }, |
| { |
| "first": "Hengchao", |
| "middle": [], |
| "last": "Shang", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Huawei Translation Services Center", |
| "location": { |
| "settlement": "Beijing", |
| "country": "China" |
| } |
| }, |
| "email": "shanghengchao@huawei.com" |
| }, |
| { |
| "first": "Daimeng", |
| "middle": [], |
| "last": "Wei", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Huawei Translation Services Center", |
| "location": { |
| "settlement": "Beijing", |
| "country": "China" |
| } |
| }, |
| "email": "weidaimeng@huawei.com" |
| }, |
| { |
| "first": "Jiaxin", |
| "middle": [], |
| "last": "Guo", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Huawei Translation Services Center", |
| "location": { |
| "settlement": "Beijing", |
| "country": "China" |
| } |
| }, |
| "email": "guojiaxin1@huawei.com" |
| }, |
| { |
| "first": "Lizhi", |
| "middle": [], |
| "last": "Lei", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Huawei Translation Services Center", |
| "location": { |
| "settlement": "Beijing", |
| "country": "China" |
| } |
| }, |
| "email": "leilizhi@huawei.com" |
| }, |
| { |
| "first": "Ying", |
| "middle": [], |
| "last": "Qin", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Huawei Translation Services Center", |
| "location": { |
| "settlement": "Beijing", |
| "country": "China" |
| } |
| }, |
| "email": "qinying@huawei.com" |
| }, |
| { |
| "first": "Shimin", |
| "middle": [], |
| "last": "Tao", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Huawei Translation Services Center", |
| "location": { |
| "settlement": "Beijing", |
| "country": "China" |
| } |
| }, |
| "email": "taoshimin@huawei.com" |
| }, |
| { |
| "first": "Shiliang", |
| "middle": [], |
| "last": "Sun", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "East China Normal University", |
| "location": { |
| "settlement": "Shanghai", |
| "country": "China" |
| } |
| }, |
| "email": "slsun@cs.ecnu.edu.cn" |
| }, |
| { |
| "first": "Yimeng", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Huawei Translation Services Center", |
| "location": { |
| "settlement": "Beijing", |
| "country": "China" |
| } |
| }, |
| "email": "chenyimeng@huawei.com" |
| }, |
| { |
| "first": "Liangyou", |
| "middle": [], |
| "last": "Li", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "Huawei Noah's Ark Lab", |
| "institution": "", |
| "location": { |
| "settlement": "Hong Kong", |
| "country": "China" |
| } |
| }, |
| "email": "liliangyou@huawei.com" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "This paper presents our work in the WMT 2020 Word and Sentence-Level Post-editing Effort Quality Estimation (QE) Shared Task. Our system follows standard Predictor-Estimator architecture, with a pre-trained Transformer as the Predictor, and specific classifiers and regressors as Estimators. We integrate Bottleneck Adapter Layers in the Predictor to improve the transfer learning efficiency and prevent from over-fitting. At the same time, we jointly train the word-and sentence-level tasks with a unified model with multitask learning. Pseudo-PE assisted QE (PEAQE) is proposed, resulting in significant improvements on the performance. Our submissions achieve competitive result in word/sentence-level sub-tasks for both of En-De/Zh language pairs.", |
| "pdf_parse": { |
| "paper_id": "2020", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "This paper presents our work in the WMT 2020 Word and Sentence-Level Post-editing Effort Quality Estimation (QE) Shared Task. Our system follows standard Predictor-Estimator architecture, with a pre-trained Transformer as the Predictor, and specific classifiers and regressors as Estimators. We integrate Bottleneck Adapter Layers in the Predictor to improve the transfer learning efficiency and prevent from over-fitting. At the same time, we jointly train the word-and sentence-level tasks with a unified model with multitask learning. Pseudo-PE assisted QE (PEAQE) is proposed, resulting in significant improvements on the performance. Our submissions achieve competitive result in word/sentence-level sub-tasks for both of En-De/Zh language pairs.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Quality Estimation (QE) assesses the translation quality of machine translation (MT) system output when ground truth reference is not available (Specia et al., 2013 (Specia et al., , 2018 . For the word-level QE task, participants are required to tag tokens and gaps of the translation output from an unknown MT system with OK and BAD, as well as tokens in the source with the same tags. The result is measured by Matthews Correlation Coefficient (MCC). For the sentence-level task, participants are required to predict the Human Translation Error Rate (HTER) scores of the machine translation outputs and the result is evaluated in terms of the Pearson's correlation coefficient.", |
| "cite_spans": [ |
| { |
| "start": 144, |
| "end": 164, |
| "text": "(Specia et al., 2013", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 165, |
| "end": 187, |
| "text": "(Specia et al., , 2018", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Our team participates in some of the sub-tasks in two language pairs (En-De and En-Zh) (Specia et al., 2020) . With regard to the En-De track, at word-level, our model achieves the MCC score of 0.5828 on the target side, and 0.5234 on the source side; at sentence-level, our model ranks the first place with a Pearson R score of 0.7583. With regard to the En-Zh track, we only submit the target side of word-level sub-task, and achieves a MCC score of 0.5872.", |
| "cite_spans": [ |
| { |
| "start": 87, |
| "end": 108, |
| "text": "(Specia et al., 2020)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Our system is implemented with fairseq (Ott et al., 2019 ) (for En-De track) and THUMT (Zhang et al., 2017 ) (for En-Zh track). We extend the original Transformer (Vaswani et al., 2017) model to fit the QE task, and leverage transfer learning to finetune the model with the task specific dataset (Dai and Le, 2015; Howard and Ruder, 2018; Radford et al., 2018) . The contributions of our work are as follows:", |
| "cite_spans": [ |
| { |
| "start": 39, |
| "end": 56, |
| "text": "(Ott et al., 2019", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 87, |
| "end": 106, |
| "text": "(Zhang et al., 2017", |
| "ref_id": "BIBREF22" |
| }, |
| { |
| "start": 163, |
| "end": 185, |
| "text": "(Vaswani et al., 2017)", |
| "ref_id": null |
| }, |
| { |
| "start": 296, |
| "end": 314, |
| "text": "(Dai and Le, 2015;", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 315, |
| "end": 338, |
| "text": "Howard and Ruder, 2018;", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 339, |
| "end": 360, |
| "text": "Radford et al., 2018)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 We follow the Predictor-Estimator (Kim and Lee, 2016; Kim et al., 2017; Kepler et al., 2019) architecture and build a unified QE model based on the standard Transformer MT model.", |
| "cite_spans": [ |
| { |
| "start": 36, |
| "end": 55, |
| "text": "(Kim and Lee, 2016;", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 56, |
| "end": 73, |
| "text": "Kim et al., 2017;", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 74, |
| "end": 94, |
| "text": "Kepler et al., 2019)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 Bottleneck Adapter Layers (Houlsby et al., 2019; Yang et al., 2020) are integrated into the model for efficient transfer learning.", |
| "cite_spans": [ |
| { |
| "start": 28, |
| "end": 50, |
| "text": "(Houlsby et al., 2019;", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 51, |
| "end": 69, |
| "text": "Yang et al., 2020)", |
| "ref_id": "BIBREF21" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 We propose the Pseudo-PE assisted QE (PEAQE) method which effectively improve the performance.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The architecture of our model is shown in Figure 1 .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 42, |
| "end": 51, |
| "text": "Figure 1", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "A more detailed description of the word-and sentence-level QE tasks is given in this section.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Task Description", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Word-level QE estimates the translation quality by producing a sequence of tags for both source and target. For target sentences, both tokens and gaps are required to be tagged with OK or BAD, while for source sentences, only tokens are tagged with OK or BAD. This is usually modelled as a sequential labelling problem. The tag of token indicates whether the word is correctly translated or not, while the tag of gap indicates whether one or more words are missing here. The number of total tags for each MT sentence is 2N + 1, where N is the number of tokens in the sentence. The evaluation metrics of the word-level task is the Matthews Correlation Coefficient (MCC), an appropriate measurement for unbalanced labels. MCC is defined as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Word-Level", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "S = T P + F N N P = T P + F P N M CC = T P N \u2212 SP SP (1 \u2212 S)(1 \u2212 P ) ,", |
| "eq_num": "(1)" |
| } |
| ], |
| "section": "Word-Level", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "where T P , T N , F P and F N represent for true positives, true negatives, false positives and false positives respectively; and N is the number of instance (Fonseca et al., 2019) .", |
| "cite_spans": [ |
| { |
| "start": 158, |
| "end": 180, |
| "text": "(Fonseca et al., 2019)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Word-Level", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "The sentence-level QE predicts the Human Translation Error Rate (HTER) (Specia et al., 2018) of given translation outputs. HTER is an edit-distance measure, calculating the ratio between the number of edits (insertions/deletions/replacements) re- quired and the reference translation length, namely HTER = (Insertions + Delations + Replacement) / Reference Words. In the QE task, where references are not available, HTER is roughly an estimation. As HTER is a real value ranging from 0 to 1, it can be modeled as a regression task. The evaluation metrics of the sentence-level task is the Pearson correlation coefficient.", |
| "cite_spans": [ |
| { |
| "start": 71, |
| "end": 92, |
| "text": "(Specia et al., 2018)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentence-Level", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "The dataset contains 7,000 sentences for the training set, 1,000 for the dev and 1,000 for the test. Except from tags and HTER scores (labels), the dataset also provides post-edit (PE) text, as the reference for generating QE labels. Note that this data is also used in the Automatic Post Editing task in WMT 2020. Detailed statistics of the dataset is listed in Table 1 , with some metrics of the source (SRC) and translation (MT). The proportion of BAD tags agains OK tags is imbalanced, especially for Gap tags. Apart from the brief descriptive statistics listed in the table, our in-depth investigation on the provided dataset unveils some interesting findings:", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 363, |
| "end": 370, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "3" |
| }, |
| { |
| "text": "\u2022 Different from the dataset in WMT 2019 QE task (Fonseca et al., 2019) , which is sampled from IT domain, the dataset this year is collected from Wikipedia. Therefore, mixing data from previous years may not help to improve this year's performance.", |
| "cite_spans": [ |
| { |
| "start": 49, |
| "end": 71, |
| "text": "(Fonseca et al., 2019)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "3" |
| }, |
| { |
| "text": "\u2022 The BLEU score (Papineni et al., 2002) for 2020 dataset is significantly lower than that of 2019, indicating much more operations are required to edit the translation outputs into the references. As a result, the distribution of labels for 2020 dataset is changed as well when comparing with that of last year. Unlike a standard translation task, where various data augmentation techniques, such as backtranslation (Sennrich et al., 2016) , are available, QE task can hardly be improved with data augmentation, as it requires unbiased and high-quality PEs to generate tags and HTER scores. Meanwhile, the change of dataset domain makes it impossible to enlarge the dataset by incorporating the dataset of last year. Facing this challenging task, we propose the PEAQE method, which will be further explained in details in the following section.", |
| "cite_spans": [ |
| { |
| "start": 17, |
| "end": 40, |
| "text": "(Papineni et al., 2002)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 417, |
| "end": 440, |
| "text": "(Sennrich et al., 2016)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Our model follows the Predictor-Estimator (Kim et al., 2017; Kepler et al., 2019) architecture. The Predictor is considered as a feature extractor, which can be a pre-trained language model (LM) or a translation model. In our implementation, we use the standard Transformer without the causal mask as the Predictor, which is pre-trained with dataset in news translation task of WMT 2019 En-De and WMT 2020 En-Zh. The Estimator can be task specific classifiers which map the extracted features into the target space, and can be modelled by several fully connected layers. We use a unified QE model to solve both word-and sentence-level tasks by building three classifiers and a regressor to make prediction on SRC tag, MT token tag, MT gap tag and HTER score, respectively.", |
| "cite_spans": [ |
| { |
| "start": 42, |
| "end": 60, |
| "text": "(Kim et al., 2017;", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 61, |
| "end": 81, |
| "text": "Kepler et al., 2019)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 4.1 Unified QE Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We define the encoder and decoder of the Transformer as functions f and g; SRC and MT text as X and Y ; tags of SRC, MT token and MT gap as V x , V y , V g ; and HTER score as V h . The representation H X and H Y are obtained by passing the X and Y into the encoder and decoder respectively:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 4.1 Unified QE Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "H X = f (X) (2) H Y = g(Y, H X ).", |
| "eq_num": "(3)" |
| } |
| ], |
| "section": "Model 4.1 Unified QE Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "For a translation model, we usually append and prepend the special token s to the SRC and TGT text. Here we follow the same rule and thereby the lengths of SRC and MT embeddings are M +1 and N + 1 respectively. Meanwhile, we append and prepend a special label pad to the original label sequence during training, but loss of the padded label is not computed. All predictions are obtained by performing specific transformations \u03c6 \u2022 on the hidden stats:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 4.1 Unified QE Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "V x = \u03c6 x (H X ) (4) V y = \u03c6 y (H Y ) (5) V g = \u03c6 g (H Y ) (6) V h = \u03c6 h (h Y,0 ).", |
| "eq_num": "(7)" |
| } |
| ], |
| "section": "Model 4.1 Unified QE Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Note that the regressor \u03c6 h only takes the embedding of the MT's first token to make predictions, similar to the usage of [CLS] in BERT (Devlin et al., 2018) . For all classification tasks, we use weighted cross entropy as the loss function, and the weight is calculated as follows:", |
| "cite_spans": [ |
| { |
| "start": 136, |
| "end": 157, |
| "text": "(Devlin et al., 2018)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 4.1 Unified QE Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "w i = |C i | |C i |", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 4.1 Unified QE Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": ", which is the inverse of the proportion of the instance with class C i . We use weighted cross entropy because labels are highly imbalanced, and the loss should be manipulated with the weight. For sentence-level QE, we use mean squared error (MSE) as the loss function, which is quite intuitive.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 4.1 Unified QE Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "The model is trained under the multi-task learning framework by summing up the loss of all subtasks with specific weights:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 4.1 Unified QE Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "L =\u03bb h (V h \u2212 V h ) 2 (8) \u2212 \u03c4 \u2208{x,y,g} \u03bb \u03c4 log P (V \u03c4 |X, Y ),", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 4.1 Unified QE Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "where {x, y, g} represents for classification tasks and h represents for regression task, and \u03bb is the weight of loss for a specific task. Although the multi-task setting could improve the overall performance, the evaluation is performed separately, it means we can train models that are optimized for the specific task, which can be achieved by giving larger weight to the loss of that task.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 4.1 Unified QE Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "As mentioned in the previous section, the provided training set is relatively small, make the model to be easily over-fitted if all weights are updated. Therefore, we decide to integrate the Bottleneck Adapter Layers (BAL) (Houlsby et al., 2019) while keeping parameters of original Transformer fixed (Yang et al., 2020) . BAL can be easily implemented with two fullyconnected layers with a non-linear activation, and is embedded into the Transformer with residual connections after the self-attention layer and the FFN layer, respectively.", |
| "cite_spans": [ |
| { |
| "start": 223, |
| "end": 245, |
| "text": "(Houlsby et al., 2019)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 301, |
| "end": 320, |
| "text": "(Yang et al., 2020)", |
| "ref_id": "BIBREF21" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Bottleneck Adapter Layer", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "In the experiment, we find that the bottle with a \"thick\" neck (\"like FFN layers in the Transformer with higher dimension in the middle part\") could further improve the performance without seriously sacrificing training efficiency. More specifically, we tested three neck sizes, i.e. thin, same and thick. The thin and same neck basically reaches 97%-99% of performance compared with training the full Transformer without using BAL, which yields the same result with (Houlsby et al., 2019) . By increasing the parameter size of BALs, we find that the performance also increases linearly, finally reaching the pick of 104% of the baseline performance with the neck to have 2 \u00d7 hidden size.", |
| "cite_spans": [ |
| { |
| "start": 467, |
| "end": 489, |
| "text": "(Houlsby et al., 2019)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Bottleneck Adapter Layer", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "QE tags can be generated with post-edits (PEs) or reference (REF) of MT. In this dataset, PE is provided, and QE tags are generated accordingly, if PE can be directly used to assist the model learning QE tags, the training efficiency will be dramatically increased. Inspired by the Pseudo-PE technique proposed in the (Kepler et al., 2019) , we hope to fully leverage PE, for example, integrating them as part of the network input. However, for the test set, where PEs are not available, we must find an alternative approach. So, we made following assumption:", |
| "cite_spans": [ |
| { |
| "start": 318, |
| "end": 339, |
| "text": "(Kepler et al., 2019)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pseudo-PE Assisted QE", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "\u03b4(MT, REF) \u2248 \u03b4(MT, PE) + \u03b4(PE, REF), (9)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pseudo-PE Assisted QE", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "where \u03b4 is any distance measurement function. In the equation, PE is regarded as an intermediate node between MT and REF. Under such assumption, if we could find any translation that is better than MT, although not as good as PE, the translation can also be used as a substitute of PE, denoted as PE'. we call this method as Pseudo-PE assisted QE (PEAQE). Finding PE' is relatively easy when we could access unconstrained resources. Using an APE system or a robust online translation system to produce better translation outputs are two feasible approaches. After comparing the BLEU scores of the training set between many online translation services and an APE system trained by us, we decide to use Google Translate outputs as the Pseudo-PE. The BLEU score for official MTs and Google MTs in the dev set are 50.9/ 67.8 for En-De, and 22.62/41.77 for En-Zh, indicating that Google MT outputs, with a high quality, could be used as Pseudo-PEs in the testing phase.", |
| "cite_spans": [ |
| { |
| "start": 110, |
| "end": 121, |
| "text": "MT and REF.", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pseudo-PE Assisted QE", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "To leverage PEs, we simply concatenate them with the SRCs and encoded them via an encoder. We find that using the features of SRC text from the encoder could not produce acceptable predictions. Therefore, we decide to concatenate SRCs with MTs again on the decoder side, and use the decoder to extract features for both of them. More formally:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pseudo-PE Assisted QE", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "H [X;Z] = f ([X; Z])", |
| "eq_num": "(10)" |
| } |
| ], |
| "section": "Pseudo-PE Assisted QE", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "H [X;Y ] = g([X; Y ], H [X;Z] ),", |
| "eq_num": "(11)" |
| } |
| ], |
| "section": "Pseudo-PE Assisted QE", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "where Z represents for official PEs (training) or Pseudo-PEs (testing). Finally, the hidden state H [X;Y ] is sliced with the max length of X, and recover back to H X and H Y , which are used as in the original model. Official PE and Pseudo-PE can be used respectively during training and testing to assist the model to make better prediction.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pseudo-PE Assisted QE", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Our experiments of all sub-tasks for En-De and part of sub-tasks for En-Zh trak are performed on the WMT 2020 dataset. The model without Pseudo-PE assistance is considered as the baseline.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiment", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Our models are implemented with fairseq (Ott et al., 2019) and THUMT (Zhang et al., 2017) . The fairseq version mainly deals with En-De tasks thanks to the pre-trained models trained in WMT 2019 news translation task. The En-Zh pre-trained model is implemented with THUMT and is trained in WMT 2020 news translation task by our team. For the En-De model, input and output embeddings are shared, therefore SRC and TGT text can be conveniently concatenated. For the En-Zh model, vocabulary is not shared, when creating the input sequence, we firstly pass tokens of English (SRC) and Chinese (MT and PE) with specific word embedding layer respectively, and than, concatenate the hidden states of them accordingly. The number of parameters of the En-De and En-Zh models are 270M and 353M, respectively. The batch size used for training is 32. We use Adam (Kingma and Ba, 2015) to optimize parameters with learning rate of 1e-4 without any scheduler. Note that when dealing with labels of sub-tokens, for each token, we only assign the first sub-token with the label and subsequent sub-tokens are assigned with the dummy pad labels, which keeps the distribution of labels unchanged. Our QE models are trained on a Nvidia Tesla V100 GPU, and converge within 5 epochs. Table 2 shows the experimental results on the dev and test sets. The performance of the baseline model is relatively poor. By leveraging PEAQE, the model achieves much better performance, demonstrating that integrating PE directly into the QE model could effectively assist the prediction. With PEs, the model can receive stronger supervision signal and is actually learning the procedure done by the tagging script, making the entire learning process easier. However, we clearly understand that the performance of PEAQE strongly depends on the quality of Pesudo-PEs, which becomes another problem that should be solved in the future.", |
| "cite_spans": [ |
| { |
| "start": 40, |
| "end": 58, |
| "text": "(Ott et al., 2019)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 69, |
| "end": 89, |
| "text": "(Zhang et al., 2017)", |
| "ref_id": "BIBREF22" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 1262, |
| "end": 1269, |
| "text": "Table 2", |
| "ref_id": "TABREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Experimental Settings", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "Here is another interesting finding during our experiment. Initially, we also performed experiments with mBERT (Devlin et al., 2018) and XLM (Conneau and Lample, 2019) but not producing desirable results. The reason might be size of the dataset. We find that performing transfer learning with pre-trained NMT model on the limited size QE dataset is more effective than other pre-trained multilingual LMs. We consider that NMT models are naturally fit for MT related tasks because of the learned prior between bilingual text, which might not be captured by multilingual LMs where text in different languages are trained independently.", |
| "cite_spans": [ |
| { |
| "start": 111, |
| "end": 132, |
| "text": "(Devlin et al., 2018)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental Results", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "We present our works for WMT 2020 QE shared task. The experimental results demonstrate that performing transfer learning with a pre-trained NMT model on the QE task is effective. Compared to only using SRC and MT text, we propose PEAQE which could significantly improve the performance of the model. But generating reliable Pseudo-PEs that are compatible with QE tasks remains a problem that would be investigated in our future works.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Crosslingual language model pretraining", |
| "authors": [ |
| { |
| "first": "Alexis", |
| "middle": [], |
| "last": "Conneau", |
| "suffix": "" |
| }, |
| { |
| "first": "Guillaume", |
| "middle": [], |
| "last": "Lample", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems", |
| "volume": "", |
| "issue": "", |
| "pages": "7057--7067", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alexis Conneau and Guillaume Lample. 2019. Cross- lingual language model pretraining. In Advances in Neural Information Processing Systems 32: An- nual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, 8-14 December 2019, Vancouver, BC, Canada, pages 7057-7067.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Semi-supervised sequence learning", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Andrew", |
| "suffix": "" |
| }, |
| { |
| "first": "Quoc V", |
| "middle": [], |
| "last": "Dai", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Le", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Advances in neural information processing systems", |
| "volume": "", |
| "issue": "", |
| "pages": "3079--3087", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Andrew M Dai and Quoc V Le. 2015. Semi-supervised sequence learning. In Advances in neural informa- tion processing systems, pages 3079-3087.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Bert: Pre-training of deep bidirectional transformers for language understanding", |
| "authors": [ |
| { |
| "first": "Jacob", |
| "middle": [], |
| "last": "Devlin", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming-Wei", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "Kenton", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Kristina", |
| "middle": [], |
| "last": "Toutanova", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1810.04805" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understand- ing. arXiv preprint arXiv:1810.04805.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Martins, Mark Fishel, and Christian Federmann. 2019. Findings of the WMT 2019 shared tasks on quality estimation", |
| "authors": [ |
| { |
| "first": "Erick", |
| "middle": [ |
| "R" |
| ], |
| "last": "Fonseca", |
| "suffix": "" |
| }, |
| { |
| "first": "Lisa", |
| "middle": [], |
| "last": "Yankovskaya", |
| "suffix": "" |
| }, |
| { |
| "first": "F", |
| "middle": [ |
| "T" |
| ], |
| "last": "Andr\u00e9", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the Fourth Conference on Machine Translation, WMT 2019", |
| "volume": "3", |
| "issue": "", |
| "pages": "1--10", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/w19-5401" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Erick R. Fonseca, Lisa Yankovskaya, Andr\u00e9 F. T. Mar- tins, Mark Fishel, and Christian Federmann. 2019. Findings of the WMT 2019 shared tasks on quality estimation. In Proceedings of the Fourth Conference on Machine Translation, WMT 2019, Florence, Italy, August 1-2, 2019 -Volume 3: Shared Task Papers, Day 2, pages 1-10.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Parameter-efficient transfer learning for NLP", |
| "authors": [ |
| { |
| "first": "Neil", |
| "middle": [], |
| "last": "Houlsby", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrei", |
| "middle": [], |
| "last": "Giurgiu", |
| "suffix": "" |
| }, |
| { |
| "first": "Stanislaw", |
| "middle": [], |
| "last": "Jastrzebski", |
| "suffix": "" |
| }, |
| { |
| "first": "Bruna", |
| "middle": [], |
| "last": "Morrone", |
| "suffix": "" |
| }, |
| { |
| "first": "Quentin", |
| "middle": [], |
| "last": "De Laroussilhe", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrea", |
| "middle": [], |
| "last": "Gesmundo", |
| "suffix": "" |
| }, |
| { |
| "first": "Mona", |
| "middle": [], |
| "last": "Attariyan", |
| "suffix": "" |
| }, |
| { |
| "first": "Sylvain", |
| "middle": [], |
| "last": "Gelly", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1902.00751" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Neil Houlsby, Andrei Giurgiu, Stanislaw Jastrzebski, Bruna Morrone, Quentin De Laroussilhe, Andrea Gesmundo, Mona Attariyan, and Sylvain Gelly. 2019. Parameter-efficient transfer learning for NLP. arXiv preprint arXiv:1902.00751.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Universal language model fine-tuning for text classification", |
| "authors": [ |
| { |
| "first": "Jeremy", |
| "middle": [], |
| "last": "Howard", |
| "suffix": "" |
| }, |
| { |
| "first": "Sebastian", |
| "middle": [], |
| "last": "Ruder", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1801.06146" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jeremy Howard and Sebastian Ruder. 2018. Univer- sal language model fine-tuning for text classification. arXiv preprint arXiv:1801.06146.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Unbabel's participation in the WMT19 translation quality estimation shared task", |
| "authors": [ |
| { |
| "first": "Fabio", |
| "middle": [], |
| "last": "Kepler", |
| "suffix": "" |
| }, |
| { |
| "first": "Jonay", |
| "middle": [], |
| "last": "Tr\u00e9nous", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcos", |
| "middle": [ |
| "V" |
| ], |
| "last": "Treviso", |
| "suffix": "" |
| }, |
| { |
| "first": "Miguel", |
| "middle": [], |
| "last": "Vera", |
| "suffix": "" |
| }, |
| { |
| "first": "Ant\u00f3nio", |
| "middle": [], |
| "last": "G\u00f3is", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Amin Farajian", |
| "suffix": "" |
| }, |
| { |
| "first": "Ant\u00f3nio", |
| "middle": [ |
| "V" |
| ], |
| "last": "Lopes", |
| "suffix": "" |
| }, |
| { |
| "first": "Andr\u00e9", |
| "middle": [ |
| "F T" |
| ], |
| "last": "Martins", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the Fourth Conference on Machine Translation, WMT 2019", |
| "volume": "3", |
| "issue": "", |
| "pages": "78--84", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/w19-5406" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Fabio Kepler, Jonay Tr\u00e9nous, Marcos V. Treviso, Miguel Vera, Ant\u00f3nio G\u00f3is, M. Amin Farajian, Ant\u00f3nio V. Lopes, and Andr\u00e9 F. T. Martins. 2019. Unbabel's participation in the WMT19 translation quality estimation shared task. In Proceedings of the Fourth Conference on Machine Translation, WMT 2019, Florence, Italy, August 1-2, 2019 -Volume 3: Shared Task Papers, Day 2, pages 78-84.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Recurrent neural network based translation quality estimation", |
| "authors": [ |
| { |
| "first": "Hyun", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| }, |
| { |
| "first": "Jong-Hyeok", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the First Conference on Machine Translation", |
| "volume": "2", |
| "issue": "", |
| "pages": "787--792", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hyun Kim and Jong-Hyeok Lee. 2016. Recurrent neu- ral network based translation quality estimation. In Proceedings of the First Conference on Machine Translation: Volume 2, Shared Task Papers, pages 787-792.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Predictor-estimator using multilevel task learning with stack propagation for neural quality estimation", |
| "authors": [ |
| { |
| "first": "Hyun", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| }, |
| { |
| "first": "Jong-Hyeok", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Seung-Hoon", |
| "middle": [], |
| "last": "Na", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the Second Conference on Machine Translation", |
| "volume": "", |
| "issue": "", |
| "pages": "562--568", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hyun Kim, Jong-Hyeok Lee, and Seung-Hoon Na. 2017. Predictor-estimator using multilevel task learning with stack propagation for neural quality es- timation. In Proceedings of the Second Conference on Machine Translation, pages 562-568.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Adam: A method for stochastic optimization", |
| "authors": [ |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Diederik", |
| "suffix": "" |
| }, |
| { |
| "first": "Jimmy", |
| "middle": [], |
| "last": "Kingma", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ba", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "3rd International Conference on Learning Representations", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Diederik P. Kingma and Jimmy Ba. 2015. Adam: A method for stochastic optimization. In 3rd Inter- national Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "A unified neural network for quality estimation of machine translation", |
| "authors": [ |
| { |
| "first": "Maoxi", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Qingyu", |
| "middle": [], |
| "last": "Xiang", |
| "suffix": "" |
| }, |
| { |
| "first": "Zhiming", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Mingwen", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "IEICE TRANSAC-TIONS on Information and Systems", |
| "volume": "101", |
| "issue": "9", |
| "pages": "2417--2421", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Maoxi Li, Qingyu Xiang, Zhiming Chen, and Mingwen Wang. 2018. A unified neural network for quality es- timation of machine translation. IEICE TRANSAC- TIONS on Information and Systems, 101(9):2417- 2421.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "fairseq: A fast, extensible toolkit for sequence modeling", |
| "authors": [ |
| { |
| "first": "Myle", |
| "middle": [], |
| "last": "Ott", |
| "suffix": "" |
| }, |
| { |
| "first": "Sergey", |
| "middle": [], |
| "last": "Edunov", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexei", |
| "middle": [], |
| "last": "Baevski", |
| "suffix": "" |
| }, |
| { |
| "first": "Angela", |
| "middle": [], |
| "last": "Fan", |
| "suffix": "" |
| }, |
| { |
| "first": "Sam", |
| "middle": [], |
| "last": "Gross", |
| "suffix": "" |
| }, |
| { |
| "first": "Nathan", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Grangier", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Auli", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics (Demonstrations)", |
| "volume": "", |
| "issue": "", |
| "pages": "48--53", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/N19-4009" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Myle Ott, Sergey Edunov, Alexei Baevski, Angela Fan, Sam Gross, Nathan Ng, David Grangier, and Michael Auli. 2019. fairseq: A fast, extensible toolkit for sequence modeling. In Proceedings of the 2019 Conference of the North American Chap- ter of the Association for Computational Linguistics (Demonstrations), pages 48-53, Minneapolis, Min- nesota. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Bleu: a method for automatic evaluation of machine translation", |
| "authors": [ |
| { |
| "first": "Kishore", |
| "middle": [], |
| "last": "Papineni", |
| "suffix": "" |
| }, |
| { |
| "first": "Salim", |
| "middle": [], |
| "last": "Roukos", |
| "suffix": "" |
| }, |
| { |
| "first": "Todd", |
| "middle": [], |
| "last": "Ward", |
| "suffix": "" |
| }, |
| { |
| "first": "Wei-Jing", |
| "middle": [], |
| "last": "Zhu", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "311--318", |
| "other_ids": { |
| "DOI": [ |
| "10.3115/1073083.1073135" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kishore Papineni, Salim Roukos, Todd Ward, and Wei- Jing Zhu. 2002. Bleu: a method for automatic eval- uation of machine translation. In Proceedings of the 40th Annual Meeting of the Association for Compu- tational Linguistics, July 6-12, 2002, Philadelphia, PA, USA, pages 311-318.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Improving language understanding by generative pre-training", |
| "authors": [ |
| { |
| "first": "Alec", |
| "middle": [], |
| "last": "Radford", |
| "suffix": "" |
| }, |
| { |
| "first": "Karthik", |
| "middle": [], |
| "last": "Narasimhan", |
| "suffix": "" |
| }, |
| { |
| "first": "Tim", |
| "middle": [], |
| "last": "Salimans", |
| "suffix": "" |
| }, |
| { |
| "first": "Ilya", |
| "middle": [], |
| "last": "Sutskever", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alec Radford, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever. 2018. Improving language understanding by generative pre-training. URL https://s3-us-west-2. amazonaws. com/openai- assets/researchcovers/languageunsupervised/language understanding paper. pdf.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Improving neural machine translation models with monolingual data", |
| "authors": [ |
| { |
| "first": "Rico", |
| "middle": [], |
| "last": "Sennrich", |
| "suffix": "" |
| }, |
| { |
| "first": "Barry", |
| "middle": [], |
| "last": "Haddow", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexandra", |
| "middle": [], |
| "last": "Birch", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, ACL 2016", |
| "volume": "1", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/p16-1009" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Rico Sennrich, Barry Haddow, and Alexandra Birch. 2016. Improving neural machine translation mod- els with monolingual data. In Proceedings of the 54th Annual Meeting of the Association for Compu- tational Linguistics, ACL 2016, August 7-12, 2016, Berlin, Germany, Volume 1: Long Papers.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Findings of the wmt 2020 shared task on quality estimation", |
| "authors": [ |
| { |
| "first": "Lucia", |
| "middle": [], |
| "last": "Specia", |
| "suffix": "" |
| }, |
| { |
| "first": "Fr\u00e9d\u00e9ric", |
| "middle": [], |
| "last": "Blain", |
| "suffix": "" |
| }, |
| { |
| "first": "Marina", |
| "middle": [], |
| "last": "Fomicheva", |
| "suffix": "" |
| }, |
| { |
| "first": "Erick", |
| "middle": [], |
| "last": "Fonseca", |
| "suffix": "" |
| }, |
| { |
| "first": "Vishrav", |
| "middle": [], |
| "last": "Chaudhary", |
| "suffix": "" |
| }, |
| { |
| "first": "Francisco", |
| "middle": [], |
| "last": "Guzm\u00e1n", |
| "suffix": "" |
| }, |
| { |
| "first": "Andr\u00e9 Ft", |
| "middle": [], |
| "last": "Martins", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Proceedings of the Fifth Conference on Machine Translation: Shared Task Papers", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lucia Specia, Fr\u00e9d\u00e9ric Blain, Marina Fomicheva, Er- ick Fonseca, Vishrav Chaudhary, Francisco Guzm\u00e1n, and Andr\u00e9 FT Martins. 2020. Findings of the wmt 2020 shared task on quality estimation. In Proceed- ings of the Fifth Conference on Machine Translation: Shared Task Papers.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Quality Estimation for Machine Translation", |
| "authors": [ |
| { |
| "first": "Lucia", |
| "middle": [], |
| "last": "Specia", |
| "suffix": "" |
| }, |
| { |
| "first": "Carolina", |
| "middle": [], |
| "last": "Scarton", |
| "suffix": "" |
| }, |
| { |
| "first": "Gustavo", |
| "middle": [ |
| "Henrique" |
| ], |
| "last": "Paetzold", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Synthesis Lectures on Human Language Technologies. Morgan & Claypool Publishers", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "DOI": [ |
| "10.2200/S00854ED1V01Y201805HLT039" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lucia Specia, Carolina Scarton, and Gustavo Henrique Paetzold. 2018. Quality Estimation for Machine Translation. Synthesis Lectures on Human Lan- guage Technologies. Morgan & Claypool Publish- ers.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "QuEst-A translation quality estimation framework", |
| "authors": [ |
| { |
| "first": "Lucia", |
| "middle": [], |
| "last": "Specia", |
| "suffix": "" |
| }, |
| { |
| "first": "Kashif", |
| "middle": [], |
| "last": "Shah", |
| "suffix": "" |
| }, |
| { |
| "first": "Jos\u00e9 Gc De", |
| "middle": [], |
| "last": "Souza", |
| "suffix": "" |
| }, |
| { |
| "first": "Trevor", |
| "middle": [], |
| "last": "Cohn", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics: System Demonstrations", |
| "volume": "", |
| "issue": "", |
| "pages": "79--84", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lucia Specia, Kashif Shah, Jos\u00e9 GC De Souza, and Trevor Cohn. 2013. QuEst-A translation quality es- timation framework. In Proceedings of the 51st An- nual Meeting of the Association for Computational Linguistics: System Demonstrations, pages 79-84.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Attention is all you need", |
| "authors": [ |
| { |
| "first": "Illia", |
| "middle": [], |
| "last": "Kaiser", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Polosukhin", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Advances in neural information processing systems", |
| "volume": "", |
| "issue": "", |
| "pages": "5998--6008", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in neural information pro- cessing systems, pages 5998-6008.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Alibaba submission for WMT18 quality estimation task", |
| "authors": [ |
| { |
| "first": "Jiayi", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Kai", |
| "middle": [], |
| "last": "Fan", |
| "suffix": "" |
| }, |
| { |
| "first": "Bo", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Fengming", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| }, |
| { |
| "first": "Boxing", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Yangbin", |
| "middle": [], |
| "last": "Shi", |
| "suffix": "" |
| }, |
| { |
| "first": "Luo", |
| "middle": [], |
| "last": "Si", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the Third Conference on Machine Translation: Shared Task Papers", |
| "volume": "", |
| "issue": "", |
| "pages": "809--815", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jiayi Wang, Kai Fan, Bo Li, Fengming Zhou, Boxing Chen, Yangbin Shi, and Luo Si. 2018. Alibaba sub- mission for WMT18 quality estimation task. In Pro- ceedings of the Third Conference on Machine Trans- lation: Shared Task Papers, pages 809-815.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Efficient transfer learning for quality estimation with bottleneck adapter layer", |
| "authors": [ |
| { |
| "first": "Hao", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "Minghan", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Ning", |
| "middle": [], |
| "last": "Xie", |
| "suffix": "" |
| }, |
| { |
| "first": "Ying", |
| "middle": [], |
| "last": "Qin", |
| "suffix": "" |
| }, |
| { |
| "first": "Yao", |
| "middle": [], |
| "last": "Deng", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Proceedings of the 22nd Annual Conference of the European Association for Machine Translation", |
| "volume": "2020", |
| "issue": "", |
| "pages": "29--34", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hao Yang, Minghan Wang, Ning Xie, Ying Qin, and Yao Deng. 2020. Efficient transfer learning for qual- ity estimation with bottleneck adapter layer. In Pro- ceedings of the 22nd Annual Conference of the Eu- ropean Association for Machine Translation, EAMT 2020, Lisbon, Portugal, 3 -5 November, 2020, pages 29-34.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "THUMT: an open source toolkit for neural machine translation", |
| "authors": [ |
| { |
| "first": "Jiacheng", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Yanzhuo", |
| "middle": [], |
| "last": "Ding", |
| "suffix": "" |
| }, |
| { |
| "first": "Shiqi", |
| "middle": [], |
| "last": "Shen", |
| "suffix": "" |
| }, |
| { |
| "first": "Yong", |
| "middle": [], |
| "last": "Cheng", |
| "suffix": "" |
| }, |
| { |
| "first": "Maosong", |
| "middle": [], |
| "last": "Sun", |
| "suffix": "" |
| }, |
| { |
| "first": "Huan-Bo", |
| "middle": [], |
| "last": "Luan", |
| "suffix": "" |
| }, |
| { |
| "first": "Yang", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jiacheng Zhang, Yanzhuo Ding, Shiqi Shen, Yong Cheng, Maosong Sun, Huan-Bo Luan, and Yang Liu. 2017. THUMT: an open source toolkit for neural machine translation. CoRR, abs/1706.06415.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF1": { |
| "uris": null, |
| "type_str": "figure", |
| "text": "This figure shows the architecture of our model, where SRC and Pseudo-PE are concatenated as the encoder input, a copy of SRC and MT are concatenated as the decoder input. The output feature are passed through four linear layers to make prediction for four tasks.", |
| "num": null |
| }, |
| "TABREF2": { |
| "text": "The experimental results of our model, where the baseline model is introduced in section 4.1. The evaluation results of the test set are from the official leader-board.", |
| "num": null, |
| "content": "<table/>", |
| "type_str": "table", |
| "html": null |
| } |
| } |
| } |
| } |