ACL-OCL / Base_JSON /prefixD /json /D17 /D17-1013.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "D17-1013",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T16:14:36.702525Z"
},
"title": "Neural Machine Translation with Word Predictions",
"authors": [
{
"first": "Rongxiang",
"middle": [],
"last": "Weng",
"suffix": "",
"affiliation": {
"laboratory": "State Key Laboratory for Novel Software Technology",
"institution": "Nanjing University",
"location": {
"postCode": "210023",
"settlement": "Nanjing",
"country": "China"
}
},
"email": "wengrx@nlp.nju.edu.cn"
},
{
"first": "Shujian",
"middle": [],
"last": "Huang",
"suffix": "",
"affiliation": {
"laboratory": "State Key Laboratory for Novel Software Technology",
"institution": "Nanjing University",
"location": {
"postCode": "210023",
"settlement": "Nanjing",
"country": "China"
}
},
"email": "huangsj@nlp.nju.edu.cn"
},
{
"first": "Zaixiang",
"middle": [],
"last": "Zheng",
"suffix": "",
"affiliation": {
"laboratory": "State Key Laboratory for Novel Software Technology",
"institution": "Nanjing University",
"location": {
"postCode": "210023",
"settlement": "Nanjing",
"country": "China"
}
},
"email": "zhengzx@nlp.nju.edu.cn"
},
{
"first": "Xinyu",
"middle": [],
"last": "Dai",
"suffix": "",
"affiliation": {
"laboratory": "State Key Laboratory for Novel Software Technology",
"institution": "Nanjing University",
"location": {
"postCode": "210023",
"settlement": "Nanjing",
"country": "China"
}
},
"email": ""
},
{
"first": "Jiajun",
"middle": [],
"last": "Chen",
"suffix": "",
"affiliation": {
"laboratory": "State Key Laboratory for Novel Software Technology",
"institution": "Nanjing University",
"location": {
"postCode": "210023",
"settlement": "Nanjing",
"country": "China"
}
},
"email": "chenjj@nlp.nju.edu.cn"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "In the encoder-decoder architecture for neural machine translation (NMT), the hidden states of the recurrent structures in the encoder and decoder carry the crucial information about the sentence.These vectors are generated by parameters which are updated by back-propagation of translation errors through time. We argue that propagating errors through the end-to-end recurrent structures are not a direct way of control the hidden vectors. In this paper, we propose to use word predictions as a mechanism for direct supervision. More specifically, we require these vectors to be able to predict the vocabulary in target sentence. Our simple mechanism ensures better representations in the encoder and decoder without using any extra data or annotation. It is also helpful in reducing the target side vocabulary and improving the decoding efficiency. Experiments on Chinese-English and German-English machine translation tasks show BLEU improvements by 4.53 and 1.3, respectively.",
"pdf_parse": {
"paper_id": "D17-1013",
"_pdf_hash": "",
"abstract": [
{
"text": "In the encoder-decoder architecture for neural machine translation (NMT), the hidden states of the recurrent structures in the encoder and decoder carry the crucial information about the sentence.These vectors are generated by parameters which are updated by back-propagation of translation errors through time. We argue that propagating errors through the end-to-end recurrent structures are not a direct way of control the hidden vectors. In this paper, we propose to use word predictions as a mechanism for direct supervision. More specifically, we require these vectors to be able to predict the vocabulary in target sentence. Our simple mechanism ensures better representations in the encoder and decoder without using any extra data or annotation. It is also helpful in reducing the target side vocabulary and improving the decoding efficiency. Experiments on Chinese-English and German-English machine translation tasks show BLEU improvements by 4.53 and 1.3, respectively.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "The encoder-decoder based neural machine translation (NMT) models have been developing rapidly. propose to encode the source sentence as a fixed-length vector representation, based on which the decoder generates the target sequence, where both the encoder and decoder are recurrent neural networks (RNN) or their variants Chung et al., 2014; . In this framework, the fixedlength vector plays the crucial role of transition-ing the information of the sentence from the source side to the target side.",
"cite_spans": [
{
"start": 322,
"end": 341,
"text": "Chung et al., 2014;",
"ref_id": "BIBREF4"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Later, attention mechanisms are proposed to enhance the source side representations Luong et al., 2015b) . The source side context is computed at each time-step of decoding, based on the attention weights between the source side representations and the current hidden state of the decoder. However, the hidden states in the recurrent decoder still originate from the single fixed-length representation (Luong et al., 2015b) , or the average of the bi-directional representations ). Here we refer to the representation as initial state.",
"cite_spans": [
{
"start": 84,
"end": 104,
"text": "Luong et al., 2015b)",
"ref_id": "BIBREF11"
},
{
"start": 402,
"end": 423,
"text": "(Luong et al., 2015b)",
"ref_id": "BIBREF11"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Interestingly, Britz et al. (2017) find that the value of initial state does not affect the translation performance, and prefer to set the initial state to be a zero vector. On the contrary, we argue that initial state still plays an important role of translation, which is currently neglected. We notice that beside the end-to-end error back propagation for the initial and transition parameters, there is no direct control of the initial state in the current NMT architectures. Due to the large number of parameters, it may be difficult for the NMT system to learn the proper sentence representation as the initial state. Thus, the model is very likely to get stuck in local minimums, making the translation process arbitrary and unstable.",
"cite_spans": [
{
"start": 15,
"end": 34,
"text": "Britz et al. (2017)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "In this paper, we propose to augment the current NMT architecture with a word prediction mechanism. More specifically, we require the initial state of the decoder to be able to predict all the words in the target sentence. In this way, there is a specific objective for learning the initial state. Thus the learnt source side representation will be better constrained. We further extend this idea by applying the word predictions mechanism to all the hidden states of the decoder. So the transition between different decoder states could be controlled as well.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Our mechanism is simple and requires no additional data or annotation. The proposed word predictions mechanism could be used as a training method and brings no extra computing cost during decoding.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Experiments on the Chinese-English and German-English translation tasks show that both the constraining of the initial state and the decoder hidden states bring significant improvement over the baseline systems. Furthermore, using the word prediction mechanism on the initial state as a word predictor to reduce the target side vocabulary could greatly improve the decoding efficiency, without a significant loss on the translation quality.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Many previous works have noticed the problem of training an NMT system with lots of parameters. Some of them prefer to use the dropout technique (Srivastava et al., 2014; Luong et al., 2015b; Meng et al., 2016) . Another possible choice is to ensemble several models with random starting points Jean et al., 2015; Luong and Manning, 2016) . Both techniques could bring more stable and better results. But they are general training techniques of neural networks, which are not specifically targeting the modeling of the translation process like ours. We will make empirical comparison with them in the experiments.",
"cite_spans": [
{
"start": 145,
"end": 170,
"text": "(Srivastava et al., 2014;",
"ref_id": "BIBREF16"
},
{
"start": 171,
"end": 191,
"text": "Luong et al., 2015b;",
"ref_id": "BIBREF11"
},
{
"start": 192,
"end": 210,
"text": "Meng et al., 2016)",
"ref_id": "BIBREF13"
},
{
"start": 295,
"end": 313,
"text": "Jean et al., 2015;",
"ref_id": "BIBREF6"
},
{
"start": 314,
"end": 338,
"text": "Luong and Manning, 2016)",
"ref_id": "BIBREF10"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "The way we add the word prediction is similar to the research of multi-task learning. Dong et al. (2015) propose to share an encoder between different translation tasks. Luong et al. (2015a) propose to jointly learn the translation task for different languages, the parsing task and the image captioning task, with a shared encoder or decoder. Zhang and Zong (2016) propose to use multitask learning for incorporating source side monolingual data. Different from these attempts, our method focuses solely on the current translation task, and does not require any extra data or annotation.",
"cite_spans": [
{
"start": 86,
"end": 104,
"text": "Dong et al. (2015)",
"ref_id": "BIBREF5"
},
{
"start": 170,
"end": 190,
"text": "Luong et al. (2015a)",
"ref_id": "BIBREF9"
},
{
"start": 344,
"end": 365,
"text": "Zhang and Zong (2016)",
"ref_id": "BIBREF22"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "In the other sequence to sequence tasks, Suzuki and Nagata (2017) propose the idea for predicting words by using encoder information. However, the purpose and the way of our mechanism are different from them.",
"cite_spans": [
{
"start": 41,
"end": 65,
"text": "Suzuki and Nagata (2017)",
"ref_id": "BIBREF19"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "The word prediction technique has been applied in the research of both statistical machine transla-tion (SMT) (Bangalore et al., 2007; Mauser et al., 2009; Jeong et al., 2010; Tran et al., 2014) and NMT (Mi et al., 2016; L'Hostis et al., 2016) . In these research, word prediction mechanisms are employed to decide the selection of words or constrain the target vocabulary, while in this paper, we use word prediction as a control mechanism for neural model training.",
"cite_spans": [
{
"start": 110,
"end": 134,
"text": "(Bangalore et al., 2007;",
"ref_id": "BIBREF1"
},
{
"start": 135,
"end": 155,
"text": "Mauser et al., 2009;",
"ref_id": "BIBREF12"
},
{
"start": 156,
"end": 175,
"text": "Jeong et al., 2010;",
"ref_id": "BIBREF7"
},
{
"start": 176,
"end": 194,
"text": "Tran et al., 2014)",
"ref_id": "BIBREF20"
},
{
"start": 203,
"end": 220,
"text": "(Mi et al., 2016;",
"ref_id": "BIBREF14"
},
{
"start": 221,
"end": 243,
"text": "L'Hostis et al., 2016)",
"ref_id": "BIBREF8"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "We present a popular NMT framework with the encoder-decoder architecture and the attention networks (Luong et al., 2015b) , based on which we propose our word prediction mechanism.",
"cite_spans": [
{
"start": 100,
"end": 121,
"text": "(Luong et al., 2015b)",
"ref_id": "BIBREF11"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "Denote a source-target sentence pair as {x, y} from the training set, where x is the source word",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "sequence (x 1 , x 2 , \u2022 \u2022 \u2022 , x |x| )",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "and y is the target word sequence (y 1 , y 2 , \u2022 \u2022 \u2022 , y |y| ), |x| and |y| are the length of x and y, respectively.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "In the encoding stage, a bi-directional recurrent neural network is used to encode x into a sequence of vectors",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "(h 1 , h 2 , \u2022 \u2022 \u2022 , h |x| ). For each x i , the representation h i is: h i = [ \u2212 \u2192 h i ; \u2190 \u2212 h i ]",
"eq_num": "(1)"
}
],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "where [\u2022; \u2022] denotes the concatenation of column vectors; \u2212 \u2192 h i and \u2190 \u2212 h i denote the hidden vectors for the word x i in the forward and backward RNNs, respectively.",
"cite_spans": [
{
"start": 6,
"end": 12,
"text": "[\u2022; \u2022]",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "The gated recurrent unit (GRU) is used as the recurrent unit in each RNN, which is shown to have promising results in speech recognition and machine translation . Formally, the hidden state h i at time step i of the forward RNN encoder is defined by the GRU function g\u2212 \u2192 e (\u2022, \u2022), as follows:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "\u2212 \u2192 h i = g\u2212 \u2192 e ( \u2212 \u2192 h i\u22121 , emb x i ) (2) = (1 \u2212 \u2212 \u2192 z i ) \u2299 \u2212 \u2192 h i\u22121 + \u2212 \u2192 z i \u2299 \u2212 \u2192 h \u2032 i \u2212 \u2192 z i = \u03c3( \u2212 \u2192 W z [emb x i ; \u2212 \u2192 h i\u22121 ]) (3) \u2212 \u2192 h \u2032 i = tanh( \u2212 \u2192 W[emb x i ; ( \u2212 \u2192 r i \u2299 \u2212 \u2192 h i\u22121 )]) (4) \u2212 \u2192 r i = \u03c3( \u2212 \u2192 W r [emb x i ; \u2212 \u2192 h i\u22121 ])",
"eq_num": "(5)"
}
],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "where \u2299 denotes element-wise product between vectors and emb x i is the word embedding of the x i . tanh(\u2022) and \u03c3(\u2022) are the tanh and sigmoid transformation functions that can be applied elementwise on vectors, respectively. For simplicity, we",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "... ... h 1 h 1 h 2 h 2 h 3 h 3 h i h i ... ... ... Encoder Decoder ... ... ... S 1 y 1 S 2 y 2 S 3 y 3 S j y j x 1 x 2 x 3 x i ... ... y 1 y 2 y j Initial State",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "Figure 1: The NMT model with word prediction for the initial state.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "omit the bias term in each network layer. The backward RNN encoder is defined likewise.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "In the decoding stage, the decoder starts with the initial state s 0 , which is the average of source representations .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "s 0 = \u03c3(W s 1 |x| |x| \u2211 i=1 h i )",
"eq_num": "(6)"
}
],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "At each time step j, the decoder maximizes the conditional probability of generating the jth target word, which is defined as:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "P (y j |y <j , x) = f d (t d ([emb y j\u22121 ; s j ; c j ])) (7) f d (u) = softmax(W f u) (8) t d (v) = tanh(W t v)",
"eq_num": "(9)"
}
],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "where s j is the decoder's hidden state, which is computed by another GRU (as in Equation 2):",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "s j = g d (s j\u22121 , [emb y j\u22121 ; c j ])",
"eq_num": "(10)"
}
],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "and the context vector c j is from the attention mechanism (Luong et al., 2015b) :",
"cite_spans": [
{
"start": 59,
"end": 80,
"text": "(Luong et al., 2015b)",
"ref_id": "BIBREF11"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "c j = |x| \u2211 i=1 a ji h i (11) a ji = exp(e ji ) \u2211 |x| k=1 exp(e jk )",
"eq_num": "(12)"
}
],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "e ji = tanh(W att d [s j\u22121 ; h i ]).",
"eq_num": "(13)"
}
],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "4 NMT with Word Predictions",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Notations and Backgrounds",
"sec_num": "3"
},
{
"text": "The decoder starts the generation of target sentence from the initial state s 0 (Equation 6) generated by the encoder. Currently, the update for the encoder only happens when a translation error occurs in the decoder. The error is propagated through multiple time steps in the recurrent structure until it reaches the encoder. As there are hundreds of millions of parameters in the NMT system, it is hard for the model to learn the exact representation of source sentences. As a result, the values of initial state may not be exact during the translation process, leading to poor translation performances. We propose word prediction as a mechanism to control the values of initial state. The intuition is that since the initial state is responsible for the translation of whole target sentence, it should at least contain information of each word in the target sentence. Thus, we optimize the initial state by making prediction for all target words. For simplicity, we assume each target word is independent of each other.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Word Prediction for the Initial State",
"sec_num": "4.1"
},
{
"text": "Here the word prediction mechanism is a simpler sub-task of translation, where the order of words is not considered. The prediction task could be trained jointly with the translation task in a multi-task learning way (Luong et al., 2015a; Dong et al., 2015; Zhang and Zong, 2016) , where both tasks share the same encoder. In other words, word prediction for the initial state could be interpreted as an improvement for the encoder. We denote this mechanism as WP E .",
"cite_spans": [
{
"start": 217,
"end": 238,
"text": "(Luong et al., 2015a;",
"ref_id": "BIBREF9"
},
{
"start": 239,
"end": 257,
"text": "Dong et al., 2015;",
"ref_id": "BIBREF5"
},
{
"start": 258,
"end": 279,
"text": "Zhang and Zong, 2016)",
"ref_id": "BIBREF22"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Word Prediction for the Initial State",
"sec_num": "4.1"
},
{
"text": "As shown in Figure 1 , a prediction network is added to the initial state. We define the conditional probability of WP E as follows:",
"cite_spans": [],
"ref_spans": [
{
"start": 12,
"end": 20,
"text": "Figure 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "Word Prediction for the Initial State",
"sec_num": "4.1"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "P WP E (y|x) = |y| \u220f j=1 P WP E (y j |x)",
"eq_num": "(14)"
}
],
"section": "Word Prediction for the Initial State",
"sec_num": "4.1"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "P WP E (y j |x) = f p (t p ([s 0 ; c p ]))",
"eq_num": "(15)"
}
],
"section": "Word Prediction for the Initial State",
"sec_num": "4.1"
},
{
"text": "where f p (\u2022) and t p (\u2022) are the softmax layer and non-linear layer as defined in Equation 8-9, with different parameters; c p is defined similar as the",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Word Prediction for the Initial State",
"sec_num": "4.1"
},
{
"text": "... ... h 1 h 1 h 2 h 2 h 3 h 3 h i h i ... ... ... Encoder Decoder ... ... ... S 1 y 1 ... ... y 1 y 2 y j S 2 y 2 ... ... y 2 y 3 y j S 3 y 3 ... ... y 3 y 4 y j S j y j ... y j x 1 x 2 x 3 x i Initial State",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Word Prediction for the Initial State",
"sec_num": "4.1"
},
{
"text": "Figure 2: The NMT model with word predictions for the decoder's hidden states.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Word Prediction for the Initial State",
"sec_num": "4.1"
},
{
"text": "attention network, so the source side information could be enhanced.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Word Prediction for the Initial State",
"sec_num": "4.1"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "c p = |x| \u2211 i=1 a i h i (16) a i = exp(e i ) \u2211 |x| k=1 exp(e k )",
"eq_num": "(17)"
}
],
"section": "Word Prediction for the Initial State",
"sec_num": "4.1"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "e i = tanh(W attp [s 0 , h i ]).",
"eq_num": "(18)"
}
],
"section": "Word Prediction for the Initial State",
"sec_num": "4.1"
},
{
"text": "Similar intuition is also applied for the decoder. Because the hidden states of the decoder are responsible for the translation of target words, they should be able to predict the target words as well. The only difference is that we remove the already generated words from the prediction task. So each hidden state in the decoder is required to predict the target words which remain untranslated. For the first state s 1 of the decoder, the prediction task is similar with the task for the initial state. Since then, the prediction is no longer a separate training task, but integrated into each time step of the training process. We denote this mechanism as WP D .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Word Predictions for Decoder's Hidden States",
"sec_num": "4.2"
},
{
"text": "As shown in Figure 2 , for each time step j in the decoder, the hidden state s j is used for the prediction of (y j , y j+1 , \u2022 \u2022 \u2022 , y |y| ). The conditional probability of WP D is defined as:",
"cite_spans": [],
"ref_spans": [
{
"start": 12,
"end": 20,
"text": "Figure 2",
"ref_id": null
}
],
"eq_spans": [],
"section": "Word Predictions for Decoder's Hidden States",
"sec_num": "4.2"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "P WP D (y j , y j+1 , \u2022 \u2022 \u2022 , y |y| |y <j , x) (19) = |y| \u220f k=j P WP D (y k |y <j , x) P WP D (y k |y <j , x) =f d (p(t d ([emb y j\u22121 ; s j ; c j ])))",
"eq_num": "(20)"
}
],
"section": "Word Predictions for Decoder's Hidden States",
"sec_num": "4.2"
},
{
"text": "where is another non-linear transformation layer, which prepares the current state for the prediction:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Word Predictions for Decoder's Hidden States",
"sec_num": "4.2"
},
{
"text": "f",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Word Predictions for Decoder's Hidden States",
"sec_num": "4.2"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "p(u) = tanh(W p u).",
"eq_num": "(21)"
}
],
"section": "Word Predictions for Decoder's Hidden States",
"sec_num": "4.2"
},
{
"text": "NMT models optimize the networks by maximizing the likelihood of the target translation y given source sentence x, denoted by L T .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Training",
"sec_num": "4.3"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "L T = 1 |y| |y| \u2211 j=1 log P (y j |y <j , x)",
"eq_num": "(22)"
}
],
"section": "Training",
"sec_num": "4.3"
},
{
"text": "where P (y j |y <j , x) is defined in Equation 7.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Training",
"sec_num": "4.3"
},
{
"text": "To optimize the word prediction mechanism, we propose to add extra likelihood functions L WP E and L WP D into the training procedure.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Training",
"sec_num": "4.3"
},
{
"text": "For the WP E , we directly optimize the likelihood of translation and word prediction:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Training",
"sec_num": "4.3"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "L 1 = L T + L WP E (23) L WP E = log P WP E",
"eq_num": "(24)"
}
],
"section": "Training",
"sec_num": "4.3"
},
{
"text": "where P WP E is defined in Equation 14.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Training",
"sec_num": "4.3"
},
{
"text": "For the WP D , we optimize the likelihood as:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Training",
"sec_num": "4.3"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "L 2 = L T + L WP D (25) L WP D = |y| \u2211 j=1 1 |y| \u2212 j + 1 log P WP D",
"eq_num": "(26)"
}
],
"section": "Training",
"sec_num": "4.3"
},
{
"text": "where P WP D is defined in Equation 19; the coefficient of the logarithm is used to calculate the average probability of each prediction. The two mechanisms could also work together, so that both the encoder and the decoder could be improved:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Training",
"sec_num": "4.3"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "L 3 = L T + L WP E + L WP D .",
"eq_num": "(27)"
}
],
"section": "Training",
"sec_num": "4.3"
},
{
"text": "The previously proposed word prediction mechanism could be used only as a extra training objective, which will not be computed during the translation. Thus the computational complexity of our models for translation stays exactly the same. On the other hand, using a smaller and specific vocabulary for each sentence or batch will improve translation efficiency. If the vocabulary is accurate enough, there is also a chance to improve the translation quality (Jean et al., 2015; Mi et al., 2016; L'Hostis et al., 2016) . Our word prediction mechanism WP E provides a natural solution for generating a possible set of target words at sentence level. The prediction could be made from the initial state s 0 , without using extra resources such as word dictionaries, extracted phrases or frequent word lists, as in Mi et al. (2016) .",
"cite_spans": [
{
"start": 458,
"end": 477,
"text": "(Jean et al., 2015;",
"ref_id": "BIBREF6"
},
{
"start": 478,
"end": 494,
"text": "Mi et al., 2016;",
"ref_id": "BIBREF14"
},
{
"start": 495,
"end": 517,
"text": "L'Hostis et al., 2016)",
"ref_id": "BIBREF8"
},
{
"start": 811,
"end": 827,
"text": "Mi et al. (2016)",
"ref_id": "BIBREF14"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Making Use of the Word Predictor",
"sec_num": "4.4"
},
{
"text": "We perform experiments on the Chinese-English (CH-EN) and German-English (DE-EN) machine translation tasks. For the CH-EN, the training data consists of about 8 million sentence pairs 1 . We use NIST MT02 as our validation set, and the NIST MT03, MT04 and MT05 as our test sets. These sets have 878, 919, 1597 and 1082 source sentences, respectively, with 4 references for each sentence. For the DE-EN, the experiments trained on the standard benchmark WMT14, and it has about 4.5 million sentence pairs. We use newstest 2013 (NST13) as validation set, and newstest 2014(NST14) as test set. These sets have 3000 and 2737 source sentences, respectively, with 1 reference for each sentence. Sentences were encoded using byte-pair encoding (BPE) (Britz et al., 2017) .",
"cite_spans": [
{
"start": 743,
"end": 763,
"text": "(Britz et al., 2017)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Data",
"sec_num": "5.1"
},
{
"text": "We implement a baseline system with the bidirectional encoder and the attention mechanism (Luong et al., 2015b) as described in Section 3, denoted as baseNMT. Then our proposed word prediction mechanism on initial state and hidden states of decoder are implemented on the baseNMT system, denoted as WP E and WP D , respectively. We denote the system use both techniques as WP ED . We implement systems with variable-sized vocabulary following (Mi et al., 2016) . For comparison, we also implement systems with dropout (with dropout rate 0.5 on the output layer) and ensemble (ensemble of 4 systems at the output layer) techniques.",
"cite_spans": [
{
"start": 90,
"end": 111,
"text": "(Luong et al., 2015b)",
"ref_id": "BIBREF11"
},
{
"start": 443,
"end": 460,
"text": "(Mi et al., 2016)",
"ref_id": "BIBREF14"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Systems and Techniques",
"sec_num": "5.2"
},
{
"text": "Both our CH-EN and DE-EN experiments are implemented on the open source toolkit dl4mt 2 , with most default parameter settings kept the same. We train the NMT systems with the sentences of length up to 50 words. The source and target vocabularies are limited to the most frequent 30K words for both Chinese and English, respectively, with the out-of-vocabulary words mapped to a special token UNK.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Implementation Details",
"sec_num": "5.3"
},
{
"text": "The dimension of word embedding is set to 512 and the size of the hidden layer is 1024. The recurrent weight matrices are initialized as random orthogonal matrices, and all the bias vectors as zero. Other parameters are initialized by sampling from the Gaussian distribution N (0, 0.01).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Implementation Details",
"sec_num": "5.3"
},
{
"text": "We use the mini-batch stochastic gradient descent (SGD) approach to update the parameters, with a batch size of 32. The learning rate is controlled by AdaDelta (Zeiler, 2012).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Implementation Details",
"sec_num": "5.3"
},
{
"text": "For efficient training of our system, we adopt a simple pre-train strategy. Firstly, the baseNMT system is trained. The training results are used as the initial parameters for pre-training our proposed models with word predictions.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Implementation Details",
"sec_num": "5.3"
},
{
"text": "For decoding during test time, we simply decode until the end-of-sentence symbol eos occurs, using a beam search with a beam width of 5.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Implementation Details",
"sec_num": "5.3"
},
{
"text": "To see the effect of word predictions in translation, we evaluate these systems in case-insensitive IBM-BLEU (Papineni et al., 2002) on both CH-EN and DE-EN tasks.",
"cite_spans": [
{
"start": 109,
"end": 132,
"text": "(Papineni et al., 2002)",
"ref_id": "BIBREF15"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Translation Experiments",
"sec_num": "5.4"
},
{
"text": "The detailed results are show in the BLEU, but the improvement is smaller than on the baseNMT. On the DE-EN experiments, the phenomenon of experiments is similar to CH-EN experiments. The baseNMT system improves 0.94 through dropout method and 0.9 BLEU through ensemble method. The dropout technique also does not work on WP ED and the ensemble technique improves 1.79 BLEU. These comparisons suggests that our system already learns better and stable values for the parameters, enjoying some of the benefits of general training techniques like dropout and ensemble. Compared to dropout and ensemble, our method WP ED achieves the highest improvement against the baseline system on both CH-EN and DE-EN experiments. Along with ensemble method, the improvement could be up to 5.79 BLEU and 1.79 BLEU respectively.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Translation Experiments",
"sec_num": "5.4"
},
{
"text": "Since we include an explicit word prediction mechanism during the training of NMT systems, we also evaluate the prediction performance on the CH-EN experiments to see how the training is improved.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Word Prediction Experiments",
"sec_num": "5.5"
},
{
"text": "For each sentence in the test set, we use the initial state of the given model to make prediction about the possible words. We denote the set of top n words as T n , the set of words in all the references top-n baseNMT WP E Prec. Recall Prec . Recall top-10 45% 17% 73% 30% top-20 33% 21% 63% 43% top-50 21% 30% 41% 55% top-100 14% 39% 28% 68% top-1k 2% 67% 4% 89% top-5k 0.7% 84% 0.9% 95% top-10k 0.4% 90% 0.5% 97% Table 5 : Comparison between baseNMT and WP E in precision and recall for the different prediction size on the CH-EN experiments.",
"cite_spans": [],
"ref_spans": [
{
"start": 242,
"end": 370,
"text": ". Recall top-10 45% 17% 73% 30% top-20 33% 21% 63% 43% top-50 21% 30% 41% 55% top-100 14% 39% 28% 68% top-1k",
"ref_id": "TABREF1"
},
{
"start": 436,
"end": 443,
"text": "Table 5",
"ref_id": null
}
],
"eq_spans": [],
"section": "Word Prediction Experiments",
"sec_num": "5.5"
},
{
"text": "as R. We define the precision, recall of the word prediction as follows:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Word Prediction Experiments",
"sec_num": "5.5"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "precision = |T n \u2229 R| |T n | * 100% (28) recall = |T n \u2229 R| |R| * 100%",
"eq_num": "(29)"
}
],
"section": "Word Prediction Experiments",
"sec_num": "5.5"
},
{
"text": "We compare the prediction performance of baseNMT and WP E . WP ED has similar prediction results with WP E , so we omit its results. As shown in Table 5 , baseNMT system has a relatively lower prediction precision, for example, 45% in top 10 prediction. With an explicit training, the WP E could achieve a much higher precision in all conditions. Specifically, the precision reaches 73% in top 10. This indicates that the initial state in WP E contains more specific information about the prediction of the target words, which may be a step towards better semantic representation, and leads to better translation quality.",
"cite_spans": [],
"ref_spans": [
{
"start": 145,
"end": 152,
"text": "Table 5",
"ref_id": null
}
],
"eq_spans": [],
"section": "Word Prediction Experiments",
"sec_num": "5.5"
},
{
"text": "Because the total words in the references are limited (around 50), the precision goes down, as expected, when a larger prediction set is considered. On the other hand, the recall of WP E is also much higher than baseNMT. When given 1k predictions, WP E could successfully predict 89% of the words in the reference. The recall goes up to 95% with 5k predictions, which is only 1/6 of the current vocabulary.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Word Prediction Experiments",
"sec_num": "5.5"
},
{
"text": "To analyze the process of word prediction, we draw the attention heatmap (Equation 16) between the initial state s 0 and the bi-directional representation of each source side word h i for an example sentence. As shown in Figure 3 , both examples show that the initial states have a very strong attention with all the content words in the source sentence. The blank cells are mostly functions words or high frequent tokens such as \"\u7684 ('s)\", \"\u662f (is)\", \"\u800c (and)\", \"\u5b83 (it)\", comma and period. This indicates that the initial state successfully encodes information about most of the content words in the source sentence, which contributes for a high prediction performance and leads to better translation.",
"cite_spans": [],
"ref_spans": [
{
"start": 221,
"end": 229,
"text": "Figure 3",
"ref_id": "FIGREF1"
}
],
"eq_spans": [],
"section": "Word Prediction Experiments",
"sec_num": "5.5"
},
{
"text": "To make use of the word prediction, we conduct experiments using the predicted vocabulary, with different vocabulary size (1k to 10k) on the CH-EN experiments, denoted as WP E -V and WP ED -V. The comparison is made in both translation quality and decoding time. As all our models with fixed vocabulary size have exactly the same number of parameters for decoding (extra mechanism is used only for training), we only plot the decoding time of the WP ED for comparison. Figure 4 and 5 show the results.",
"cite_spans": [],
"ref_spans": [
{
"start": 469,
"end": 477,
"text": "Figure 4",
"ref_id": null
}
],
"eq_spans": [],
"section": "Improving Decoding Efficiency",
"sec_num": "5.6"
},
{
"text": "When we start the experiments with top 1k vocabulary (1/30 of the baseline settings), the translation quality of both WP E -V and WP ED -V are already higher than the baseNMT; while their decoding time is less than 1/3 of an NMT system with 30k vocabulary. When the size of vocabulary increases, the translation quality improves as well. With a 6k predicted vocabulary (1/5 of the baseline settings), the decoding time is about 60% of a full- (The horizontal line shows the decoding time for the systems with fixed vocabulary. ) vocabulary system; the performances of both systems with variable size vocabulary are comparable their corresponding fixed-vocabulary systems, which is higher than the baseNMT by 2.53 and 4.53 BLEU, respectively.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Improving Decoding Efficiency",
"sec_num": "5.6"
},
{
"text": "Although the comparison may not be fair enough due to the language pair and training conditions, the above relative improvements (e.g. WP ED -V v.s. baseNMT) is much higher than previous research of manipulating the vocabularies (Jean et al., 2015; Mi et al., 2016; L'Hostis et al., 2016) . This is because our mechanism is not only about reducing the vocabulary itself for each sentence or batch, it also brings improvement to the overall translation model. Please note that un-like these research, we keep the target vocabulary to be 30k in all our experiments, because we are not focusing on increasing the vocabulary size in this paper. It will be interesting to combine our mechanism with larger vocabulary to further enhance the translation performance. Again, our mechanism requires no extra annotation, dictionary, alignment or separate discriminative predictor, etc.",
"cite_spans": [
{
"start": 229,
"end": 248,
"text": "(Jean et al., 2015;",
"ref_id": "BIBREF6"
},
{
"start": 249,
"end": 265,
"text": "Mi et al., 2016;",
"ref_id": "BIBREF14"
},
{
"start": 266,
"end": 288,
"text": "L'Hostis et al., 2016)",
"ref_id": "BIBREF8"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Improving Decoding Efficiency",
"sec_num": "5.6"
},
{
"text": "We also analyze real-case translations to see the difference between different systems (Table 6) .",
"cite_spans": [],
"ref_spans": [
{
"start": 87,
"end": 96,
"text": "(Table 6)",
"ref_id": null
}
],
"eq_spans": [],
"section": "Translation Analysis",
"sec_num": "5.7"
},
{
"text": "It is easy to see that the baseNMT system misses the translations of several important words, such as \"advertising\", \"1.5\", which are marked with underline in the reference. It also wrongly translates the company name \"time warner inc.\" as the redundant information \"internet company\"; \"america online\" as \"us line\".",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Translation Analysis",
"sec_num": "5.7"
},
{
"text": "The results of dropout or ensemble show improvement compared to the baseNMT. But they still make mistakes about the translation of \"online\" and the company name \"time warner inc.\".",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Translation Analysis",
"sec_num": "5.7"
},
{
"text": "With WP ED , most of these errors no longer exist, because we force the encoder and decoder to carry the exact information during translation.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Translation Analysis",
"sec_num": "5.7"
},
{
"text": "The encoder-decoder architecture provides a general paradigm for learning machine translation from the source language to the target language. However, due to the large amount of parameters and relatively small training data set, the end-toend learning of an NMT model may not be able to learn the best solution. We argue that at least part of the problem is caused by the long error backpropagation pipeline of the recurrent structures in multiple time steps, which provides no direct control of the information carried by the hidden states in both the encoder and decoder.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusions",
"sec_num": "6"
},
{
"text": "Instead of looking for other annotated data, we notice that the words in the target language sentence could be viewed as a natural annotation. We propose to use the word prediction mechanism to enhance the initial state generated by the encoder and extend the mechanism to control the hidden states of decoder as well. Experiments show promising results on the Chinese-English and German-English translation tasks. As a byproduct, the word predictor could be used to improve the efficiency of decoding, which may be source \u65f6\u4ee3\u534e\u7eb3\u516c\u53f8\u7684\u7f51\u7edc\u516c\u53f8\u7f8e\u56fd\u7ebf\u4e0a\u8bf4, \u5b83\u9884\u671f\u4e8c \u25cb \u25cb \u4e8c\u5e74\u7684\u5e7f\u544a\u4e0e\u5546\u4e1a\u9500\u552e\u5c06\u7531 \u4e8c \u25cb \u25cb \u4e00\u5e74\u7684\u4e8c\u5341\u4e03\u4ebf\u7f8e\u5143\u51cf\u5c11\u5230\u5341\u4e94\u4ebf\u7f8e\u5143\u3002 reference america online , the internet arm of time warner conglomerate , said it expects advertising and commerce revenue to decline from us $ 2.7 billion in 2001 to us $ 1.5 in 2002 . baseNMT in the us line , the internet company 's internet company said on the internet that it expected that the business sales in 2002 would fall from $ UNK billion to $ UNK billion in 2001 .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusions",
"sec_num": "6"
},
{
"text": "baseNMT +dropout on the united states line , UNK 's internet company said on the internet that it expects to reduce the annual advertising and commercial sales from $ UNK billion in 2001 to $ 1.5 billion .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusions",
"sec_num": "6"
},
{
"text": "baseNMT +ensemble in the us line , the internet company 's internet company said that it expected that the advertising and commercial sales volume for 2002 would be reduced from us $ UNK billion to us $ 1.5 billion in 2001 .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusions",
"sec_num": "6"
},
{
"text": "the internet company of time warner inc. , the us online , said that it expects that the advertising and commercial sales in 2002 will decrease from $ UNK billion in 2001 to us $ 1.5 billion . Table 6 : Comparisons of different systems in translating the same example sentence, which from CH-EN test sets. (\"source\" indicates the source sentence; \"reference\" indicates the human translation; the translation results are indicated by their system names, including our best \"WP ED \" systems. The underline words in the reference are missed in the baseNMT output; the bold font indicates improvements over the baseNMT system; and the italic font indicates remaining translation errors.) crucial for large scale applications.",
"cite_spans": [],
"ref_spans": [
{
"start": 193,
"end": 200,
"text": "Table 6",
"ref_id": null
}
],
"eq_spans": [],
"section": "WP ED",
"sec_num": null
},
{
"text": "Our attempts demonstrate that the learning of the large scale neural network systems is still not good enough. In the future, it might be helpful to analyze the benefits of jointly learning other related tasks together with machine translation, to provide further control of the learning process. It is interesting to demonstrate the effectiveness of the proposed mechanism on other sequence to sequence learning tasks as well.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "WP ED",
"sec_num": null
},
{
"text": "includes LDC2002E18, LDC2003E07, LDC2003E14, LDC2004E12, LDC2004T08, LDC2005T06, LDC2005T10, LDC2006E26 and LDC2007T09",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "https://github.com/nyu-dl/dl4mt-tutorial",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [
{
"text": "We would like to thank the anonymous reviewers for their insightful comments. Shujian Huang is the corresponding author. This work is supported by the National Science Foundation of China (No. 61672277, 61472183) , the Jiangsu Provincial Research Foundation for Basic Research (No. BK20170074).",
"cite_spans": [],
"ref_spans": [
{
"start": 188,
"end": 212,
"text": "(No. 61672277, 61472183)",
"ref_id": null
}
],
"eq_spans": [],
"section": "Acknowledgments",
"sec_num": null
}
],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "Neural machine translation by jointly learning to align and translate",
"authors": [
{
"first": "Dzmitry",
"middle": [],
"last": "Bahdanau",
"suffix": ""
},
{
"first": "Kyunghyun",
"middle": [],
"last": "Cho",
"suffix": ""
},
{
"first": "Yoshua",
"middle": [],
"last": "Bengio",
"suffix": ""
}
],
"year": 2014,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2014. Neural machine translation by jointly learning to align and translate. CoRR abs/1409.0473. http://arxiv.org/abs/1409.0473.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Statistical machine translation through global lexical selection and sentence reconstruction",
"authors": [
{
"first": "Srinivas",
"middle": [],
"last": "Bangalore",
"suffix": ""
},
{
"first": "Patrick",
"middle": [],
"last": "Haffner",
"suffix": ""
},
{
"first": "Stephan",
"middle": [],
"last": "Kanthak",
"suffix": ""
}
],
"year": 2007,
"venue": "Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics. Association for Computational Linguistics",
"volume": "",
"issue": "",
"pages": "152--159",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Srinivas Bangalore, Patrick Haffner, and Stephan Kan- thak. 2007. Statistical machine translation through global lexical selection and sentence reconstruction. In Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics. Associa- tion for Computational Linguistics, pages 152-159. http://aclweb.org/anthology/P07-1020.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "Massive Exploration of Neural Machine Translation Architectures",
"authors": [
{
"first": "Denny",
"middle": [],
"last": "Britz",
"suffix": ""
},
{
"first": "Anna",
"middle": [],
"last": "Goldie",
"suffix": ""
},
{
"first": "Minh-Thang",
"middle": [],
"last": "Luong",
"suffix": ""
},
{
"first": "Quoc",
"middle": [],
"last": "Le",
"suffix": ""
}
],
"year": 2017,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Denny Britz, Anna Goldie, Minh-Thang Luong, and Quoc Le. 2017. Massive Exploration of Neural Ma- chine Translation Architectures. ArXiv e-prints .",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Learning phrase representations using rnn encoder-decoder for statistical machine translation",
"authors": [
{
"first": "Kyunghyun",
"middle": [],
"last": "Cho",
"suffix": ""
},
{
"first": "Bart",
"middle": [],
"last": "Van Merrienboer",
"suffix": ""
},
{
"first": "Caglar",
"middle": [],
"last": "Gulcehre",
"suffix": ""
},
{
"first": "Dzmitry",
"middle": [],
"last": "Bahdanau",
"suffix": ""
},
{
"first": "Fethi",
"middle": [],
"last": "Bougares",
"suffix": ""
},
{
"first": "Holger",
"middle": [],
"last": "Schwenk",
"suffix": ""
},
{
"first": "Yoshua",
"middle": [],
"last": "Bengio",
"suffix": ""
}
],
"year": 2014,
"venue": "Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)",
"volume": "",
"issue": "",
"pages": "1724--1734",
"other_ids": {
"DOI": [
"10.3115/v1/D14-1179"
]
},
"num": null,
"urls": [],
"raw_text": "Kyunghyun Cho, Bart van Merrienboer, Caglar Gul- cehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. 2014. Learning phrase representations using rnn encoder-decoder for statistical machine translation. In Proceedings of the 2014 Conference on Empirical Methods in Nat- ural Language Processing (EMNLP). Association for Computational Linguistics, pages 1724-1734. https://doi.org/10.3115/v1/D14-1179.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "Empirical evaluation of gated recurrent neural networks on sequence modeling",
"authors": [
{
"first": "Junyoung",
"middle": [],
"last": "Chung",
"suffix": ""
},
{
"first": "\u00c7aglar",
"middle": [],
"last": "G\u00fcl\u00e7ehre",
"suffix": ""
},
{
"first": "Kyunghyun",
"middle": [],
"last": "Cho",
"suffix": ""
},
{
"first": "Yoshua",
"middle": [],
"last": "Bengio",
"suffix": ""
}
],
"year": 2014,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Junyoung Chung, \u00c7aglar G\u00fcl\u00e7ehre, KyungHyun Cho, and Yoshua Bengio. 2014. Empirical evaluation of gated recurrent neural networks on sequence modeling. CoRR abs/1412.3555.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "Multi-task learning for multiple language translation",
"authors": [
{
"first": "Daxiang",
"middle": [],
"last": "Dong",
"suffix": ""
},
{
"first": "Hua",
"middle": [],
"last": "Wu",
"suffix": ""
},
{
"first": "Wei",
"middle": [],
"last": "He",
"suffix": ""
},
{
"first": "Dianhai",
"middle": [],
"last": "Yu",
"suffix": ""
},
{
"first": "Haifeng",
"middle": [],
"last": "Wang",
"suffix": ""
}
],
"year": 2015,
"venue": "Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing",
"volume": "1",
"issue": "",
"pages": "1723--1732",
"other_ids": {
"DOI": [
"10.3115/v1/P15-1166"
]
},
"num": null,
"urls": [],
"raw_text": "Daxiang Dong, Hua Wu, Wei He, Dianhai Yu, and Haifeng Wang. 2015. Multi-task learning for multiple language translation. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th Inter- national Joint Conference on Natural Language Processing (Volume 1: Long Papers). Association for Computational Linguistics, pages 1723-1732. https://doi.org/10.3115/v1/P15-1166.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "On using very large target vocabulary for neural machine translation",
"authors": [
{
"first": "S\u00e9bastien",
"middle": [],
"last": "Jean",
"suffix": ""
},
{
"first": "Kyunghyun",
"middle": [],
"last": "Cho",
"suffix": ""
},
{
"first": "Roland",
"middle": [],
"last": "Memisevic",
"suffix": ""
},
{
"first": "Yoshua",
"middle": [],
"last": "Bengio",
"suffix": ""
}
],
"year": 2015,
"venue": "Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing",
"volume": "1",
"issue": "",
"pages": "1--10",
"other_ids": {
"DOI": [
"10.3115/v1/P15-1001"
]
},
"num": null,
"urls": [],
"raw_text": "S\u00e9bastien Jean, Kyunghyun Cho, Roland Memisevic, and Yoshua Bengio. 2015. On using very large target vocabulary for neural machine translation. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Lan- guage Processing (Volume 1: Long Papers). Asso- ciation for Computational Linguistics, pages 1-10. https://doi.org/10.3115/v1/P15-1001.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "A discriminative lexicon model for complex morphology",
"authors": [
{
"first": "Minwoo",
"middle": [],
"last": "Jeong",
"suffix": ""
},
{
"first": "Kristina",
"middle": [],
"last": "Toutanova",
"suffix": ""
},
{
"first": "Hisami",
"middle": [],
"last": "Suzuki",
"suffix": ""
},
{
"first": "Chris",
"middle": [],
"last": "Quirk",
"suffix": ""
}
],
"year": 2010,
"venue": "Proceedings of the Ninth Conference of the Association for Machine Translation in the Americas",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Minwoo Jeong, Kristina Toutanova, Hisami Suzuki, and Chris Quirk. 2010. A discriminative lexicon model for complex morphology. In Proceedings of the Ninth Conference of the Association for Machine Translation in the Americas (AMTA 2010).",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "Vocabulary selection strategies for neural machine translation",
"authors": [
{
"first": "L'",
"middle": [],
"last": "Gurvan",
"suffix": ""
},
{
"first": "David",
"middle": [],
"last": "Hostis",
"suffix": ""
},
{
"first": "Michael",
"middle": [],
"last": "Grangier",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Auli",
"suffix": ""
}
],
"year": 2016,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Gurvan L'Hostis, David Grangier, and Michael Auli. 2016. Vocabulary selection strategies for neu- ral machine translation. CoRR abs/1610.00072. http://arxiv.org/abs/1610.00072.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "Multi-task sequence to sequence learning",
"authors": [
{
"first": "Minh-Thang",
"middle": [],
"last": "Luong",
"suffix": ""
},
{
"first": "Quoc",
"middle": [
"V"
],
"last": "Le",
"suffix": ""
},
{
"first": "Ilya",
"middle": [],
"last": "Sutskever",
"suffix": ""
},
{
"first": "Oriol",
"middle": [],
"last": "Vinyals",
"suffix": ""
},
{
"first": "Lukasz",
"middle": [],
"last": "Kaiser",
"suffix": ""
}
],
"year": 2015,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Minh-Thang Luong, Quoc V. Le, Ilya Sutskever, Oriol Vinyals, and Lukasz Kaiser. 2015a. Multi-task se- quence to sequence learning. CoRR abs/1511.06114. http://arxiv.org/abs/1511.06114.",
"links": null
},
"BIBREF10": {
"ref_id": "b10",
"title": "Achieving open vocabulary neural machine translation with hybrid wordcharacter models",
"authors": [
{
"first": "Minh-Thang",
"middle": [],
"last": "Luong",
"suffix": ""
},
{
"first": "Christopher",
"middle": [
"D"
],
"last": "Manning",
"suffix": ""
}
],
"year": 2016,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Minh-Thang Luong and Christopher D. Man- ning. 2016. Achieving open vocabulary neural machine translation with hybrid word- character models. CoRR abs/1604.00788.",
"links": null
},
"BIBREF11": {
"ref_id": "b11",
"title": "Effective approaches to attentionbased neural machine translation",
"authors": [
{
"first": "Minh-Thang",
"middle": [],
"last": "Luong",
"suffix": ""
},
{
"first": "Hieu",
"middle": [],
"last": "Pham",
"suffix": ""
},
{
"first": "Christopher",
"middle": [
"D"
],
"last": "Manning",
"suffix": ""
}
],
"year": 2015,
"venue": "Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing",
"volume": "",
"issue": "",
"pages": "1412--1421",
"other_ids": {
"DOI": [
"10.18653/v1/D15-1166"
]
},
"num": null,
"urls": [],
"raw_text": "Minh-Thang Luong, Hieu Pham, and Christopher D. Manning. 2015b. Effective approaches to attention- based neural machine translation. In Proceed- ings of the 2015 Conference on Empirical Meth- ods in Natural Language Processing. Association for Computational Linguistics, pages 1412-1421. https://doi.org/10.18653/v1/D15-1166.",
"links": null
},
"BIBREF12": {
"ref_id": "b12",
"title": "Extending statistical machine translation with discriminative and trigger-based lexicon models",
"authors": [
{
"first": "Arne",
"middle": [],
"last": "Mauser",
"suffix": ""
},
{
"first": "Sa\u0161a",
"middle": [],
"last": "Hasan",
"suffix": ""
},
{
"first": "Hermann",
"middle": [],
"last": "Ney",
"suffix": ""
}
],
"year": 2009,
"venue": "Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics",
"volume": "",
"issue": "",
"pages": "210--218",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Arne Mauser, Sa\u0161a Hasan, and Hermann Ney. 2009. Extending statistical machine translation with dis- criminative and trigger-based lexicon models. In Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing. Associa- tion for Computational Linguistics, pages 210-218. http://aclweb.org/anthology/D09-1022.",
"links": null
},
"BIBREF13": {
"ref_id": "b13",
"title": "Interactive attention for neural machine translation",
"authors": [
{
"first": "Fandong",
"middle": [],
"last": "Meng",
"suffix": ""
},
{
"first": "Zhengdong",
"middle": [],
"last": "Lu",
"suffix": ""
},
{
"first": "Hang",
"middle": [],
"last": "Li",
"suffix": ""
},
{
"first": "Qun",
"middle": [],
"last": "Liu",
"suffix": ""
}
],
"year": 2016,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Fandong Meng, Zhengdong Lu, Hang Li, and Qun Liu. 2016. Interactive attention for neu- ral machine translation. CoRR abs/1610.05011. http://arxiv.org/abs/1610.05011.",
"links": null
},
"BIBREF14": {
"ref_id": "b14",
"title": "Vocabulary manipulation for neural machine translation",
"authors": [
{
"first": "Haitao",
"middle": [],
"last": "Mi",
"suffix": ""
},
{
"first": "Zhiguo",
"middle": [],
"last": "Wang",
"suffix": ""
},
{
"first": "Abe",
"middle": [],
"last": "Ittycheriah",
"suffix": ""
}
],
"year": 2016,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Haitao Mi, Zhiguo Wang, and Abe Ittycheriah. 2016. Vocabulary manipulation for neural machine translation. CoRR abs/1605.03209.",
"links": null
},
"BIBREF15": {
"ref_id": "b15",
"title": "Bleu: A method for automatic evaluation of machine translation",
"authors": [
{
"first": "Kishore",
"middle": [],
"last": "Papineni",
"suffix": ""
},
{
"first": "Salim",
"middle": [],
"last": "Roukos",
"suffix": ""
},
{
"first": "Todd",
"middle": [],
"last": "Ward",
"suffix": ""
},
{
"first": "Wei-Jing",
"middle": [],
"last": "Zhu",
"suffix": ""
}
],
"year": 2002,
"venue": "Proceedings of the 40th Annual Meeting on Association for Computational Linguistics",
"volume": "",
"issue": "",
"pages": "311--318",
"other_ids": {
"DOI": [
"10.3115/1073083.1073135"
]
},
"num": null,
"urls": [],
"raw_text": "Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. Bleu: A method for automatic evaluation of machine translation. In Proceedings of the 40th Annual Meeting on Association for Computational Linguistics. Association for Computational Linguistics, Stroudsburg, PA, USA, ACL '02, pages 311- 318. https://doi.org/10.3115/1073083.1073135.",
"links": null
},
"BIBREF16": {
"ref_id": "b16",
"title": "Dropout: A simple way to prevent neural networks from overfitting",
"authors": [
{
"first": "Nitish",
"middle": [],
"last": "Srivastava",
"suffix": ""
},
{
"first": "Geoffrey",
"middle": [],
"last": "Hinton",
"suffix": ""
},
{
"first": "Alex",
"middle": [],
"last": "Krizhevsky",
"suffix": ""
},
{
"first": "Ilya",
"middle": [],
"last": "Sutskever",
"suffix": ""
},
{
"first": "Ruslan",
"middle": [],
"last": "Salakhutdinov",
"suffix": ""
}
],
"year": 2014,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. 2014. Dropout: A simple way to prevent neural networks from overfit- ting.",
"links": null
},
"BIBREF18": {
"ref_id": "b18",
"title": "Sequence to sequence learning with neural networks",
"authors": [
{
"first": "Ilya",
"middle": [],
"last": "Sutskever",
"suffix": ""
},
{
"first": "Oriol",
"middle": [],
"last": "Vinyals",
"suffix": ""
},
{
"first": "V",
"middle": [],
"last": "Quoc",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Le",
"suffix": ""
}
],
"year": 2014,
"venue": "Advances in Neural Information Processing Systems 27",
"volume": "",
"issue": "",
"pages": "3104--3112",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Ilya Sutskever, Oriol Vinyals, and Quoc V. Le. 2014. Sequence to sequence learning with neural net- works. In Z. Ghahramani, M. Welling, C. Cortes, N. D. Lawrence, and K. Q. Weinberger, editors, Advances in Neural Information Processing Sys- tems 27, Curran Associates, Inc., pages 3104- 3112. http://papers.nips.cc/paper/5346-sequence- to-sequence-learning-with-neural-networks.pdf.",
"links": null
},
"BIBREF19": {
"ref_id": "b19",
"title": "Rnnbased encoder-decoder approach with word frequency estimation",
"authors": [
{
"first": "Jun",
"middle": [],
"last": "Suzuki",
"suffix": ""
},
{
"first": "Masaaki",
"middle": [],
"last": "Nagata",
"suffix": ""
}
],
"year": 2017,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Jun Suzuki and Masaaki Nagata. 2017. Rnn- based encoder-decoder approach with word frequency estimation. CoRR abs/1701.00138.",
"links": null
},
"BIBREF20": {
"ref_id": "b20",
"title": "Word translation prediction for morphologically rich languages with bilingual neural networks",
"authors": [
{
"first": "Ke",
"middle": [],
"last": "Tran",
"suffix": ""
},
{
"first": "Arianna",
"middle": [],
"last": "Bisazza",
"suffix": ""
},
{
"first": "Christof",
"middle": [],
"last": "Monz",
"suffix": ""
}
],
"year": 2014,
"venue": "Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)",
"volume": "",
"issue": "",
"pages": "1676--1688",
"other_ids": {
"DOI": [
"10.3115/v1/D14-1175"
]
},
"num": null,
"urls": [],
"raw_text": "Ke Tran, Arianna Bisazza, and Christof Monz. 2014. Word translation prediction for morphologically rich languages with bilingual neural networks. In Pro- ceedings of the 2014 Conference on Empirical Meth- ods in Natural Language Processing (EMNLP). Association for Computational Linguistics, pages 1676-1688. https://doi.org/10.3115/v1/D14-1175.",
"links": null
},
"BIBREF21": {
"ref_id": "b21",
"title": "ADADELTA: an adaptive learning rate method",
"authors": [
{
"first": "D",
"middle": [],
"last": "Matthew",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Zeiler",
"suffix": ""
}
],
"year": 2012,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Matthew D. Zeiler. 2012. ADADELTA: an adap- tive learning rate method. CoRR abs/1212.5701. http://arxiv.org/abs/1212.5701.",
"links": null
},
"BIBREF22": {
"ref_id": "b22",
"title": "Exploiting source-side monolingual data in neural machine translation",
"authors": [
{
"first": "Jiajun",
"middle": [],
"last": "Zhang",
"suffix": ""
},
{
"first": "Chengqing",
"middle": [],
"last": "Zong",
"suffix": ""
}
],
"year": 2016,
"venue": "Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing",
"volume": "",
"issue": "",
"pages": "1535--1545",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Jiajun Zhang and Chengqing Zong. 2016. Ex- ploiting source-side monolingual data in neural machine translation. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, pages 1535-1545. http://aclweb.org/anthology/D16-1160.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"type_str": "figure",
"num": null,
"text": "d (\u2022) and t d (\u2022) are the softmax layer and non-linear layer as defined in Equation 8-9; p(\u2022)",
"uris": null
},
"FIGREF1": {
"type_str": "figure",
"num": null,
"text": "Two examples of the attention heatmap between the initial state s 0 and the bi-directional representation of each source side word h i from the CH-EN test sets. (The English translation of each source word is annotated in the parentheses after it. )",
"uris": null
},
"FIGREF2": {
"type_str": "figure",
"num": null,
"text": "BLEU scores with different vocabulary sizes for each sentence on the CH-EN experiments. (The performance of baseNMT, WP E , WP D , WP ED are plotted as horizontal lines for comparison.) Decoding time with different vocabulary sizes for each sentence on the CH-EN experiments.",
"uris": null
},
"TABREF0": {
"type_str": "table",
"text": "Case-insensitive 4-gram BLEU scores of baseNMT, WP E , WP D , WP ED systems on the CH-EN experiments. (The \"IMP\" column presents the improvement of test average compared to the baseNMT. )",
"num": null,
"content": "<table><tr><td>and Table 2. Compared to the baseNMT sys-</td></tr><tr><td>tem, all of our models achieve significant improve-</td></tr><tr><td>ments. On the CH-EN experiments, simply adding</td></tr><tr><td>word predictions to the initial state (WP E ) already</td></tr><tr><td>brings considerable improvements. The average</td></tr><tr><td>improvement on test set is 2.53 BLEU, showing</td></tr><tr><td>that constraining the initial state does lead to a</td></tr><tr><td>higher translation quality. Adding word predic-</td></tr></table>",
"html": null
},
"TABREF1": {
"type_str": "table",
"text": "",
"num": null,
"content": "<table><tr><td colspan=\"3\">: Case-insensitive 4-gram BLEU scores of</td></tr><tr><td colspan=\"3\">baseNMT, WP E , WP D , WP ED systems on the DE-</td></tr><tr><td>EN experiments.</td><td/><td/></tr><tr><td>Models</td><td>Test</td><td>IMP</td></tr><tr><td>baseNMT</td><td>34.86</td><td>\u2212</td></tr><tr><td>WP ED</td><td colspan=\"2\">39.49 +4.53</td></tr><tr><td>baseNMT-dropout</td><td colspan=\"2\">37.02 +2.06</td></tr><tr><td>WP ED -dropout</td><td colspan=\"2\">39.25 +4.29</td></tr><tr><td colspan=\"3\">baseNMT-ensemble(4) 37.71 +2.75</td></tr><tr><td>WP ED -ensemble(4)</td><td colspan=\"2\">40.75 +5.79</td></tr></table>",
"html": null
},
"TABREF2": {
"type_str": "table",
"text": "",
"num": null,
"content": "<table><tr><td>: Average case-insensitive 4-gram BLEU</td></tr><tr><td>scores on the CH-EN experiments for baseNMT</td></tr><tr><td>and WP ED systems, with the dropout and ensemble</td></tr><tr><td>techniques.</td></tr><tr><td>tions to the hidden states in the decoder (WP D )</td></tr><tr><td>leads to further improvements against baseNMT</td></tr><tr><td>(4.15 BLEU), because WP D adds constraints to</td></tr><tr><td>the state transitions through different time steps</td></tr><tr><td>in the decoder. Using both techniques improves</td></tr><tr><td>the baseline by 4.53 BLEU. On the DE-EN ex-</td></tr><tr><td>periments, the improvement of WP E model is 0.41</td></tr><tr><td>BLEU and WP D model is 0.86 BLEU on test set.</td></tr><tr><td>When use both techniques, the WP ED improves on</td></tr><tr><td>the test set is 1.3 BLEU.</td></tr><tr><td>We compare our models with systems using</td></tr><tr><td>dropout and ensemble techniques. The results</td></tr><tr><td>show in Table 3 and 4. On the CH-EN experi-</td></tr><tr><td>ments, the dropout method successfully improves</td></tr><tr><td>the baseNMT system by 2.06 BLEU. However,</td></tr><tr><td>it does not work on our WP ED system. The en-</td></tr><tr><td>semble technique improves the baseNMT system</td></tr><tr><td>by 2.75 BLEU. It still improves WP ED by 1.26</td></tr></table>",
"html": null
},
"TABREF3": {
"type_str": "table",
"text": "",
"num": null,
"content": "<table><tr><td>: Case-insensitive 4-gram BLEU scores on</td></tr><tr><td>the DE-EN experiments for baseNMT and WP ED</td></tr><tr><td>systems, with the dropout and ensemble tech-</td></tr><tr><td>niques.</td></tr></table>",
"html": null
}
}
}
}