| { |
| "paper_id": "2020", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T13:18:11.485507Z" |
| }, |
| "title": "On the Linguistic Representational Power of Neural Machine Translation Models", |
| "authors": [ |
| { |
| "first": "Yonatan", |
| "middle": [], |
| "last": "Belinkov", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "belinkov@mit.edu" |
| }, |
| { |
| "first": "John", |
| "middle": [ |
| "F" |
| ], |
| "last": "Paulson", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "Nadir", |
| "middle": [], |
| "last": "Durrani", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "ndurrani@qf.org.qa" |
| }, |
| { |
| "first": "Fahim", |
| "middle": [], |
| "last": "Dalvi", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "Hassan", |
| "middle": [], |
| "last": "Sajjad", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "hsajjad@qf.org.qa" |
| }, |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Glass", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "glass@mit.edu" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Despite the recent success of deep neural networks in natural language processing and other spheres of artificial intelligence, their interpretability remains a challenge. We analyze the representations learned by neural machine translation (NMT) models at various levels of granularity and evaluate their quality through relevant extrinsic properties. In particular, we seek answers to the following questions: (i) How accurately is word structure captured within the learned representations, which is an important aspect in translating morphologically rich languages? (ii) Do the representations capture long-range dependencies, and effectively handle syntactically divergent languages? (iii) Do the representations capture lexical semantics? We conduct a thorough investigation along several parameters: (i) Which layers in the architecture capture each of these linguistic phenomena; (ii) How does the choice of translation unit (word, character, or subword unit) impact the linguistic properties captured by the underlying representations? (iii) Do the encoder and decoder learn differently and independently? (iv) Do the representations learned by multilingual NMT models capture the same amount of linguistic information as their bilingual counterparts? Our data-driven, quantitative evaluation illuminates important aspects in NMT models and their ability to capture various linguistic phenomena. We show that deep NMT models trained in an end-to-end fashion, without being provided any direct supervision during the training process, learn a non-trivial amount of linguistic information. Notable findings include the following observations: (i) Word morphology and part-of-speech information are captured at the lower layers of the model; (ii) In contrast, lexical semantics or non-local syntactic and semantic dependencies are better represented at the higher layers of the model; (iii) Representations learned using characters are more informed about word-morphology compared to those learned using subword units; and (iv) Representations learned by multilingual models are richer compared to bilingual models.", |
| "pdf_parse": { |
| "paper_id": "2020", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Despite the recent success of deep neural networks in natural language processing and other spheres of artificial intelligence, their interpretability remains a challenge. We analyze the representations learned by neural machine translation (NMT) models at various levels of granularity and evaluate their quality through relevant extrinsic properties. In particular, we seek answers to the following questions: (i) How accurately is word structure captured within the learned representations, which is an important aspect in translating morphologically rich languages? (ii) Do the representations capture long-range dependencies, and effectively handle syntactically divergent languages? (iii) Do the representations capture lexical semantics? We conduct a thorough investigation along several parameters: (i) Which layers in the architecture capture each of these linguistic phenomena; (ii) How does the choice of translation unit (word, character, or subword unit) impact the linguistic properties captured by the underlying representations? (iii) Do the encoder and decoder learn differently and independently? (iv) Do the representations learned by multilingual NMT models capture the same amount of linguistic information as their bilingual counterparts? Our data-driven, quantitative evaluation illuminates important aspects in NMT models and their ability to capture various linguistic phenomena. We show that deep NMT models trained in an end-to-end fashion, without being provided any direct supervision during the training process, learn a non-trivial amount of linguistic information. Notable findings include the following observations: (i) Word morphology and part-of-speech information are captured at the lower layers of the model; (ii) In contrast, lexical semantics or non-local syntactic and semantic dependencies are better represented at the higher layers of the model; (iii) Representations learned using characters are more informed about word-morphology compared to those learned using subword units; and (iv) Representations learned by multilingual models are richer compared to bilingual models.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Deep neural networks have quickly become the predominant approach to most tasks in artificial intelligence, including machine translation (MT). Compared with their traditional counterparts, these models are trained in an end-to-end fashion, providing a simple yet elegant mechanism. This simplicity, however, comes at the price of opaqueness. Unlike traditional systems that contain specialized modules carrying specific sub-tasks, neural MT (NMT) systems train one large network, optimized toward the overall task. For example, non-neural statistical MT systems have sub-components to handle fluency (Heafield 2011) , lexical generation (Koehn, Och, and Marcu 2003) , word reordering (Galley and Manning 2008; Durrani, Schmid, and Fraser 2011) , rich morphology , and a smorgasbord of features (Chiang, Knight, and Wang 2009) for modeling different phenomena. Neural MT systems, on the other hand, contain a single model based on an encoder-decoder mechanism (Sutskever, Vinyals, and Le 2014) with attention (Bahdanau, Cho, and Bengio 2014) . Despite its simplicity, neural MT surpassed non-neural statistical MT within a few years of its emergence. Human evaluation and error analysis revealed that the improvements were obtained through more fluent outputs (Toral and S\u00e1nchez-Cartagena 2017) and better handling of morphology and non-local dependencies (Bentivogli et al. 2016) . However, it is not clear what the role of different components in the network is, what kind of information is learned during the training process, and how different components interact. Consequently, MT systems trained using neural networks are often thought of as a \"black-box\"-that is, they map inputs to outputs, but the internal machinery is opaque and difficult to interpret. Gaining a better understanding of these systems is necessary for improving the design choices and performance. In current practice, their development is often limited to a trial-and-error process, without gaining a real understanding of what the system has learned. We aim to increase model transparency by analyzing the representations learned by NMT models at different levels of granularity in light of various linguistic phenomena-at morphological, syntactic, and semantic levels-that are considered important for the task of machine translation and for learning complex natural language processing (NLP) problems. We thus strive for post-hoc decomposability, in the sense of Lipton (2016) . That is, we analyze models after they have been trained, to uncover what linguistic phenomena are captured within the underlying representations. More specifically, we aim to address the following questions in this article:", |
| "cite_spans": [ |
| { |
| "start": 601, |
| "end": 616, |
| "text": "(Heafield 2011)", |
| "ref_id": "BIBREF45" |
| }, |
| { |
| "start": 638, |
| "end": 666, |
| "text": "(Koehn, Och, and Marcu 2003)", |
| "ref_id": "BIBREF61" |
| }, |
| { |
| "start": 685, |
| "end": 710, |
| "text": "(Galley and Manning 2008;", |
| "ref_id": "BIBREF41" |
| }, |
| { |
| "start": 711, |
| "end": 744, |
| "text": "Durrani, Schmid, and Fraser 2011)", |
| "ref_id": "BIBREF37" |
| }, |
| { |
| "start": 795, |
| "end": 826, |
| "text": "(Chiang, Knight, and Wang 2009)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 960, |
| "end": 993, |
| "text": "(Sutskever, Vinyals, and Le 2014)", |
| "ref_id": "BIBREF110" |
| }, |
| { |
| "start": 1009, |
| "end": 1041, |
| "text": "(Bahdanau, Cho, and Bengio 2014)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 1260, |
| "end": 1294, |
| "text": "(Toral and S\u00e1nchez-Cartagena 2017)", |
| "ref_id": "BIBREF112" |
| }, |
| { |
| "start": 1356, |
| "end": 1380, |
| "text": "(Bentivogli et al. 2016)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 2444, |
| "end": 2457, |
| "text": "Lipton (2016)", |
| "ref_id": "BIBREF68" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "\u2022", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "What linguistic information is captured in deep learning models?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "-Do the NMT representations capture word morphology? -Do the NMT models, being trained on flat sequences of words, still acquire structural information? -Do the NMT models learn informative semantic representations?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "\u2022 Is the language information well distributed across the network or are designated parts (different layers, encoder vs. decoder) more focused on a particular linguistic property?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "\u2022 What impact does the choice of translation unit (characters, subword units, or words) have on the learned representations in terms of different linguistic phenomena?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "\u2022 How does translating into different target languages affect the representations on the (encoder) source-side?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "\u2022 How do the representations acquired by multilingual models compare with those acquired by bilingual models?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "To this end, we follow a simple and effective procedure with three steps: (i) train an NMT system; (ii) use the trained model to generate feature representations for source/target language words; and (iii) train a classifier using the generated features to make predictions for a relevant auxiliary task. We then evaluate the quality of the trained classifier on the given task as a proxy to the quality of the trained NMT model. In this way, we obtain a quantitative measure of how well the original NMT system learns features that are relevant to the given task. This procedure has become common for analyzing various neural NLP models (Belinkov and Glass 2019) . In this work, we analyze NMT representations through several linguistic annotation tasks: part-ofspeech (POS) tagging and morphological tagging for morphological knowledge; combinatory categorial grammar (CCG) supertagging and syntactic dependency labeling for syntactic knowledge; and lexical semantic tagging and semantic dependency labeling for semantic knowledge.", |
| "cite_spans": [ |
| { |
| "start": 638, |
| "end": 663, |
| "text": "(Belinkov and Glass 2019)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "We experiment with several languages with varying degrees of morphological richness and syntactic divergence (compared to English): French, German, Czech, Russian, Arabic, and Hebrew. Our analyses reveal interesting insights such as:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "\u2022", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "NMT models trained in an end-to-end fashion learn a non-trivial amount of linguistic information without being provided with direct supervision during the initial training process.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "\u2022 Linguistic information tends to be organized in a modular manner, whereby different parts of the neural network generate representations with varying amounts and types of linguistic properties.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "\u2022 A hierarchy of language representations emerges in networks trained on the complex tasks studied in this article. The lower layers of the network focus on local, low-level linguistic properties (morphology, POS, local relations), whereas higher layers are more concerned with global, high level properties (lexical semantics, long-range relations).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "\u2022 Character-based representations are better for learning morphology, especially for unknown and low-frequency input words. In contrast, representations learned using subword units are better for handling syntactic and semantic dependencies.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "\u2022 The target language impacts the kind of information learned by the MT system. For example, translating into morphologically poorer languages leads to better source-side word representations. This effect is especially apparent in smaller data regimes.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "\u2022 Representations learned by multilingual NMT models are richer in terms of learning different linguistic phenomena and benefit from shared learning. 1", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "This article is organized into the following sections. Section 2 provides an account of the related work. Section 3 describes the linguistic properties and the representative tasks used to carry out the analysis study. Section 4 describes the methodology taken for analyzing the NMT representations. Section 5 describes data, annotations, and experimental details. Sections 6, 7, and 8 provide empirical results and analysis to evaluate the quality of NMT representations with respect to morphology, syntax, and semantics, respectively, and Section 9 does the same for the multilingual NMT models. Section 10 sheds light on the overall patterns that arise from the experimental results from several angles. Section 11 concludes the article. An open-source implementation of our analysis code is available through the NeuroX toolkit (Dalvi et al. 2019b ).", |
| "cite_spans": [ |
| { |
| "start": 832, |
| "end": 851, |
| "text": "(Dalvi et al. 2019b", |
| "ref_id": "BIBREF33" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "The work related to this article can be divided into several groups:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2." |
| }, |
| { |
| "text": "The first group of related work aims at demystifying what information is learned within the neural network black-box. One line of work visualizes hidden unit activations in recurrent neural networks (RNNs) that are trained for a given task (Elman 1991; Karpathy, Johnson, and Li 2015; K\u00e1d\u00e1r, Chrupa\u0142a, and Alishahi 2017) . Although such visualizations illuminate the inner workings of the network, they are often qualitative in nature and somewhat anecdotal. Other work aims to evaluate systems on specific linguistic phenomena represented in so-called challenge sets. Prominent examples include older work on MT evaluation (King and Falkedal 1990) , as well as more recent evaluations via contrastive translation pairs (Burlot and Yvon 2017; Rios Gonzales, Mascarell, and Sennrich 2017; Bawden et al. 2018) . The latter line of work constructs minimal pairs of translations that differ by a known linguistic property, and evaluates whether the MT system assigns a higher score to the correct translation. The challenge set evaluation may produce informative results on the quality of the overall model for some linguistic property, but it does not directly assess the learned representations.", |
| "cite_spans": [ |
| { |
| "start": 240, |
| "end": 252, |
| "text": "(Elman 1991;", |
| "ref_id": "BIBREF38" |
| }, |
| { |
| "start": 253, |
| "end": 284, |
| "text": "Karpathy, Johnson, and Li 2015;", |
| "ref_id": "BIBREF53" |
| }, |
| { |
| "start": 285, |
| "end": 320, |
| "text": "K\u00e1d\u00e1r, Chrupa\u0142a, and Alishahi 2017)", |
| "ref_id": "BIBREF52" |
| }, |
| { |
| "start": 624, |
| "end": 648, |
| "text": "(King and Falkedal 1990)", |
| "ref_id": "BIBREF56" |
| }, |
| { |
| "start": 720, |
| "end": 742, |
| "text": "(Burlot and Yvon 2017;", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 743, |
| "end": 787, |
| "text": "Rios Gonzales, Mascarell, and Sennrich 2017;", |
| "ref_id": "BIBREF101" |
| }, |
| { |
| "start": 788, |
| "end": 807, |
| "text": "Bawden et al. 2018)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Analysis of Neural Networks", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "A different approach tries to provide a quantitative analysis by correlating parts of the neural network with linguistic properties, for example, by training a classifier to predict a feature of interest (Adi et al. 2017; Hupkes, Veldhoen, and Zuidema 2017; Conneau et al. 2018) . Such an analysis has been conducted on word embeddings (K\u00f6hn 2015; Qian, Qiu, and Huang 2016b) , sentence embeddings (Adi et al. 2017; Ganesh, Gupta, and Varma 2017; Conneau et al. 2018) , and RNN states (Qian, Qiu, and Huang 2016a; Wu and King 2016; Wang, Chung, and Lee 2017) . The language properties mainly analyzed are morphological (Qian, Qiu, and Huang 2016b; Vylomova et al. 2016; Belinkov et al. 2017a; Dalvi et al. 2017) , semantic (Qian, Qiu, and Huang 2016b; Belinkov et al. 2017b) , and syntactic (Tran, Bisazza, and Monz 2018; K\u00f6hn 2015; Conneau et al. 2018) . Recent studies carried a more fine-grained neuron-level analysis for NMT and LM (Bau et al. 2019a; Dalvi et al. 2019a; Lakretz et al. 2019) . In contrast to all of this work, we focus on the representations learned in neural machine translation in light of various linguistic properties (morphological, syntactic, and semantic) and phenomena such as handling low frequency words. Our work is most similar to Shi, Padhi, and Knight (2016) and Vylomova et al. (2016) . The former used hidden vectors from a neural MT encoder to predict syntactic properties on the English source side, whereas we study multiple language properties in different languages. Vylomova et al. (2016) analyzed different representations for morphologically rich languages in MT, but they did not directly measure the quality of the learned representations. Surveying the work on analyzing neural networks in NLP is beyond the scope of the present paper. We have highlighted here several of the more relevant studies and refer to Belinkov and Glass (2019) for a recent survey on the topic.", |
| "cite_spans": [ |
| { |
| "start": 204, |
| "end": 221, |
| "text": "(Adi et al. 2017;", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 222, |
| "end": 257, |
| "text": "Hupkes, Veldhoen, and Zuidema 2017;", |
| "ref_id": "BIBREF49" |
| }, |
| { |
| "start": 258, |
| "end": 278, |
| "text": "Conneau et al. 2018)", |
| "ref_id": "BIBREF29" |
| }, |
| { |
| "start": 336, |
| "end": 347, |
| "text": "(K\u00f6hn 2015;", |
| "ref_id": "BIBREF62" |
| }, |
| { |
| "start": 348, |
| "end": 375, |
| "text": "Qian, Qiu, and Huang 2016b)", |
| "ref_id": "BIBREF96" |
| }, |
| { |
| "start": 398, |
| "end": 415, |
| "text": "(Adi et al. 2017;", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 416, |
| "end": 446, |
| "text": "Ganesh, Gupta, and Varma 2017;", |
| "ref_id": "BIBREF42" |
| }, |
| { |
| "start": 447, |
| "end": 467, |
| "text": "Conneau et al. 2018)", |
| "ref_id": "BIBREF29" |
| }, |
| { |
| "start": 485, |
| "end": 513, |
| "text": "(Qian, Qiu, and Huang 2016a;", |
| "ref_id": "BIBREF94" |
| }, |
| { |
| "start": 514, |
| "end": 531, |
| "text": "Wu and King 2016;", |
| "ref_id": "BIBREF121" |
| }, |
| { |
| "start": 532, |
| "end": 558, |
| "text": "Wang, Chung, and Lee 2017)", |
| "ref_id": "BIBREF114" |
| }, |
| { |
| "start": 619, |
| "end": 647, |
| "text": "(Qian, Qiu, and Huang 2016b;", |
| "ref_id": "BIBREF96" |
| }, |
| { |
| "start": 648, |
| "end": 669, |
| "text": "Vylomova et al. 2016;", |
| "ref_id": "BIBREF114" |
| }, |
| { |
| "start": 670, |
| "end": 692, |
| "text": "Belinkov et al. 2017a;", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 693, |
| "end": 711, |
| "text": "Dalvi et al. 2017)", |
| "ref_id": "BIBREF32" |
| }, |
| { |
| "start": 723, |
| "end": 751, |
| "text": "(Qian, Qiu, and Huang 2016b;", |
| "ref_id": "BIBREF96" |
| }, |
| { |
| "start": 752, |
| "end": 774, |
| "text": "Belinkov et al. 2017b)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 791, |
| "end": 821, |
| "text": "(Tran, Bisazza, and Monz 2018;", |
| "ref_id": "BIBREF113" |
| }, |
| { |
| "start": 822, |
| "end": 832, |
| "text": "K\u00f6hn 2015;", |
| "ref_id": "BIBREF62" |
| }, |
| { |
| "start": 833, |
| "end": 853, |
| "text": "Conneau et al. 2018)", |
| "ref_id": "BIBREF29" |
| }, |
| { |
| "start": 936, |
| "end": 954, |
| "text": "(Bau et al. 2019a;", |
| "ref_id": null |
| }, |
| { |
| "start": 955, |
| "end": 974, |
| "text": "Dalvi et al. 2019a;", |
| "ref_id": null |
| }, |
| { |
| "start": 975, |
| "end": 995, |
| "text": "Lakretz et al. 2019)", |
| "ref_id": "BIBREF64" |
| }, |
| { |
| "start": 1264, |
| "end": 1293, |
| "text": "Shi, Padhi, and Knight (2016)", |
| "ref_id": "BIBREF105" |
| }, |
| { |
| "start": 1298, |
| "end": 1320, |
| "text": "Vylomova et al. (2016)", |
| "ref_id": "BIBREF114" |
| }, |
| { |
| "start": 1509, |
| "end": 1531, |
| "text": "Vylomova et al. (2016)", |
| "ref_id": "BIBREF114" |
| }, |
| { |
| "start": 1859, |
| "end": 1884, |
| "text": "Belinkov and Glass (2019)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Analysis of Neural Networks", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "One of the major challenges in training NMT systems is handling less frequent and out-of-vocabulary words. To address this issue, researchers have resorted to using subword units for training the neural network models. Luong and Manning (2016) trained a hybrid system that integrates character-level representation within a wordbased framework. Ling et al. (2015) used a bidirectional long short-term memory network (LSTM; Hochreiter and Schmidhuber 1997) to compose word embeddings from the character embeddings. Costa-juss\u00e0 and Fonollosa (2016) and Renduchintala et al. (2018) combined convolutional and highway layers to replace the standard lookupbased word representations in NMT systems with character-aware representations. 2 Sennrich, Haddow, and Birch (2016) used byte-pair encoding (BPE), a data-compression algorithm, to segment words into smaller units. A variant of this method known as a wordpiece model is used by Google (Wu et al. 2016a) . Shapiro and Duh (2018) used a similar convolutional architecture on top of BPE. Chung, Cho, and Bengio (2016) used a combination of BPE-based encoder and character-based decoder to improve translation quality. Motivated by their findings, Lee, Cho, and Hofmann (2017) explored using fully character representations (with no word boundaries) on both the source and target sides. As BPE segmentation is not linguistically motivated, an alternative to using morpheme-based segmentation has been explored in Bradbury and Socher (2016) . It is important to address what using different translation units (word, BPE, morpheme, character) entails. Sennrich (2017) performed a comparative evaluation of character-and BPE-based systems on carefully crafted synthetic tests and found that character-based models are effective in handling unknown words, but perform worse in capturing longdistance dependencies. Our work contributes to this body of research by analyzing how models based on different units capture various linguistic properties. We analyze the representations obtained by training systems on word, character, and BPE-based units.", |
| "cite_spans": [ |
| { |
| "start": 219, |
| "end": 243, |
| "text": "Luong and Manning (2016)", |
| "ref_id": "BIBREF71" |
| }, |
| { |
| "start": 345, |
| "end": 363, |
| "text": "Ling et al. (2015)", |
| "ref_id": "BIBREF67" |
| }, |
| { |
| "start": 551, |
| "end": 578, |
| "text": "Renduchintala et al. (2018)", |
| "ref_id": "BIBREF100" |
| }, |
| { |
| "start": 936, |
| "end": 953, |
| "text": "(Wu et al. 2016a)", |
| "ref_id": "BIBREF119" |
| }, |
| { |
| "start": 956, |
| "end": 978, |
| "text": "Shapiro and Duh (2018)", |
| "ref_id": "BIBREF103" |
| }, |
| { |
| "start": 1036, |
| "end": 1065, |
| "text": "Chung, Cho, and Bengio (2016)", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 1195, |
| "end": 1223, |
| "text": "Lee, Cho, and Hofmann (2017)", |
| "ref_id": "BIBREF65" |
| }, |
| { |
| "start": 1460, |
| "end": 1486, |
| "text": "Bradbury and Socher (2016)", |
| "ref_id": "BIBREF18" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Subword Units", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "In this section, we describe the linguistic phenomena for which we analyze NMT representations. We focus on linguistic properties that are considered important for the task of machine translation, and that we believe are intrinsically learned in the model to effectively perform the complex task of translation. We consider properties from the realms of morphology, syntax, and semantics. In each case, we describe linguistic properties of interest and define relevant classification tasks that aim to capture them (see Table 1 for sequence labeling tasks).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 520, |
| "end": 527, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Linguistic Properties", |
| "sec_num": "3." |
| }, |
| { |
| "text": "Modeling the structure of words and their relationship to other words in the sentence is a fundamental task in any NLP application. Languages vary in the way they encode information within words. Some languages exhibit grammatical relations such as subject/object/predicate or gender agreement by only changing the word form, whereas others achieve the same through word order or addition of particles. Morphology (aka word structure), poses an exigent problem in machine translation and is at the heart of dealing with the challenge of data-sparsity. Although English is limited in morphology, other languages such as Czech, Arabic, and Russian have highly inflected morphology. This entails that for each lemma many possible word variants could exist, thus causing an out-of-vocabulary word problem. For example, Huck et al. (2017) found only one morphological variant of the Czech word \"\u010d\u00ebska\" (plural of English \"kneecap\") in a corpus of 50K parallel sentences. It required 50M sentences, a size of parallel corpus Table 1 Example sentence with different word-level annotations. The CCG supertags are taken from Nadejde et al. (2017) . POS and semantic tags are our own annotation, as well as the German translation and its morphological tags. only available for a handful of language pairs, for them to observe all possible variants of the word. Even if such a data set is available, the computational complexity requires NMT systems to limit the vocabulary size. It is therefore important for an MT system to model word-structure with the available data and vocabulary size limitation. In traditional statistical machine translation, this is often addressed by splitting tokens in morphologically rich languages into constituents in a preprocessing step, using word segmentation in Arabic (Pasha et al. 2014; Abdelali et al. 2016) or compound splitting in German (Koehn and Knight 2003) . Previous work also explored generative morphological models, known as Factored Translation Models, that explicitly integrate additional linguistic markup at the word level to learn morphology . In NMT training, using subword units such as byte-pair encoding (Sennrich, Haddow, and Birch 2016) has become a de facto standard in training competition grade systems (Pinnis et al. 2017; . A few have tried morpheme-based segmentation (Bradbury and Socher 2016) , and several even used character-based systems (Chung, Cho, and Bengio 2016; Lee, Cho, and Hofmann 2017) to achieve similar performance as the BPE-segmented systems. Table 2 shows an example of each representation unit. BPE splits words into symbols (a symbol is a sequence of characters) and then iteratively replaces the most frequent sequences of symbols with a new merged symbol. In essence, frequent character n-gram sequences merge to form one symbol. The number of merge operations is controlled by a hyper-parameter OP, which directly affects the granularity of segmentation: a high value of OP means coarse segmentation and a low value means fine-grained segmentation. Note that although BPE and Morfessor (unsupervised morpheme-based segmentation) segment words at a similar level of granularity, the segmentation generated by Morfessor (Smit et al. 2014 ) is linguistically motivated. For example, it splits the gerund verb shooting into the base verb shoot and the suffix ing. In comparison, the BPE segmentation sho + oting has no linguistic justification. At the extreme, the fully character-level units treat each word as a sequence of characters.", |
| "cite_spans": [ |
| { |
| "start": 815, |
| "end": 833, |
| "text": "Huck et al. (2017)", |
| "ref_id": null |
| }, |
| { |
| "start": 1116, |
| "end": 1137, |
| "text": "Nadejde et al. (2017)", |
| "ref_id": null |
| }, |
| { |
| "start": 1795, |
| "end": 1814, |
| "text": "(Pasha et al. 2014;", |
| "ref_id": "BIBREF89" |
| }, |
| { |
| "start": 1815, |
| "end": 1836, |
| "text": "Abdelali et al. 2016)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 1869, |
| "end": 1892, |
| "text": "(Koehn and Knight 2003)", |
| "ref_id": "BIBREF60" |
| }, |
| { |
| "start": 2257, |
| "end": 2277, |
| "text": "(Pinnis et al. 2017;", |
| "ref_id": "BIBREF93" |
| }, |
| { |
| "start": 2325, |
| "end": 2351, |
| "text": "(Bradbury and Socher 2016)", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 2400, |
| "end": 2429, |
| "text": "(Chung, Cho, and Bengio 2016;", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 2430, |
| "end": 2457, |
| "text": "Lee, Cho, and Hofmann 2017)", |
| "ref_id": "BIBREF65" |
| }, |
| { |
| "start": 3200, |
| "end": 3217, |
| "text": "(Smit et al. 2014", |
| "ref_id": "BIBREF107" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 1019, |
| "end": 1026, |
| "text": "Table 1", |
| "ref_id": null |
| }, |
| { |
| "start": 2519, |
| "end": 2526, |
| "text": "Table 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Morphology", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Tagging tasks. In this paper, we study how effective neural MT representations are in learning word morphology and what different translation units offer in this regard. To answer such questions, we focus on the tasks of POS and full morphological tagging, which is the identification of all pertinent morphological features for every word (see Table 1 ). For example, the morphological tag vvfin.3.sg.pres.ind for the word \"empf\u00e4ngt\" (English 'receives') marks that it is a finite verb, third person, singular gender, present tense, and indicative mood.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 345, |
| "end": 352, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Morphology", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Example sentence with different segmentations: words, BPE subwords (Sennrich, Haddow, and Birch 2016), Morfessor-based subwords (Smit et al. 2014) , and characters. Notice that BPE subwords do not necessarily conform to morphemes (\"shooting\" \u2192 \"sho@@\" and \"oting\"), while Morfessor tends to have a more morphological segmentation (\"shoot@@\", \"ing\"). \"@@\" indicates a split subword unit and \" \" marks a word boundary.", |
| "cite_spans": [ |
| { |
| "start": 128, |
| "end": 146, |
| "text": "(Smit et al. 2014)", |
| "ref_id": "BIBREF107" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Table 2", |
| "sec_num": null |
| }, |
| { |
| "text": "Professor admits to shooting his girlfriend BPE Professor admits to sho@@ oting his gir@@ l@@ friend", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Words", |
| "sec_num": null |
| }, |
| { |
| "text": "Professor admit@@ s to shoot@@ ing his girl@@ friend ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Morfessor", |
| "sec_num": null |
| }, |
| { |
| "text": "Linguistic theories argue that words are hierarchically organized in syntactic constituents referred to as syntactic trees. It is therefore natural to think that translation models should be based on trees rather than a flat sequence representation of sentences. For more than a decade of research in machine translation, a tremendous amount of effort has been put into syntax-based machine translation (Yamada and Knight (2002) ; Chiang (2005) , Galley et al. (2006) , Zhang et al. (2007) , Shen, Xu, and Weischedel (2010) ; Neubig and Duh (2014)), with notable success in languages such as Chinese and German, which are syntactically divergent compared to English. However, the sequence-to-sequence NMT systems were able to surpass the performance of the stateof-the-art syntax-based systems in recent MT competitions (Bojar et al. 2016) . The LSTM-based RNN model with the help of the attention mechanism is able to handle long-distance dependencies. There have also been recent attempts to integrate syntax into NMT (Eriguchi, Hashimoto, and Tsuruoka 2016; Stahlberg et al. 2016; Aharoni and Goldberg 2017; Chen et al. 2017; Wu et al. 2017 ), but sequence-to-sequence NMT models without explicit syntax are the state of the art at the moment (Pinnis et al. 2017; .", |
| "cite_spans": [ |
| { |
| "start": 403, |
| "end": 428, |
| "text": "(Yamada and Knight (2002)", |
| "ref_id": "BIBREF124" |
| }, |
| { |
| "start": 431, |
| "end": 444, |
| "text": "Chiang (2005)", |
| "ref_id": "BIBREF23" |
| }, |
| { |
| "start": 447, |
| "end": 467, |
| "text": "Galley et al. (2006)", |
| "ref_id": "BIBREF40" |
| }, |
| { |
| "start": 470, |
| "end": 489, |
| "text": "Zhang et al. (2007)", |
| "ref_id": "BIBREF126" |
| }, |
| { |
| "start": 492, |
| "end": 523, |
| "text": "Shen, Xu, and Weischedel (2010)", |
| "ref_id": "BIBREF104" |
| }, |
| { |
| "start": 820, |
| "end": 839, |
| "text": "(Bojar et al. 2016)", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 1020, |
| "end": 1060, |
| "text": "(Eriguchi, Hashimoto, and Tsuruoka 2016;", |
| "ref_id": "BIBREF39" |
| }, |
| { |
| "start": 1061, |
| "end": 1083, |
| "text": "Stahlberg et al. 2016;", |
| "ref_id": "BIBREF108" |
| }, |
| { |
| "start": 1084, |
| "end": 1110, |
| "text": "Aharoni and Goldberg 2017;", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 1111, |
| "end": 1128, |
| "text": "Chen et al. 2017;", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 1129, |
| "end": 1143, |
| "text": "Wu et al. 2017", |
| "ref_id": "BIBREF117" |
| }, |
| { |
| "start": 1246, |
| "end": 1266, |
| "text": "(Pinnis et al. 2017;", |
| "ref_id": "BIBREF93" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Syntax", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Tagging tasks.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Syntax", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "In this paper, we analyze whether NMT models trained on flat sequences acquire structural syntactic information. To answer this, we use two tagging tasks. First, we use CCG supertagging, which captures global syntactic information locally at the word level by assigning a label to each word annotating its syntactic role in the sentence. The process is almost equivalent to parsing (Bangalore and Joshi 1999). For example, the syntactic tag PP/NP (in Table 1 ) can be thought of as a function that takes a noun phrase on the right (\"the capital of USA\") and returns a prepositional phrase (\"in the capital of USA\"). 3 Second, we use syntactic dependency labeling, the task of assigning a type to each arc in a syntactic dependency tree. In dependency grammar, sentence structure is represented by a labeled directed graph whose vertices are words and whose edges are relations, or dependencies, between the words (Mel\u010duk 1988; Nivre 2005) . A dependency is a directed bi-lexical relation between a head and its dependent, or modifier. Dependency structures are attractive to study for three main reasons. First, dependency formalisms have become increasingly popular in NLP in recent years, and much work has been devoted to developing large annotated data sets for these formalisms. The Universal Dependencies data set (Nivre et al. 2017 ) that is used in this article has been especially influential. Second, there is a fairly rich history of using dependency structures in machine translation, although much work has focused on using constituency structures (Williams et al. 2016) . Third, as dependencies are bi-lexical relations between words, it is straightforward to obtain representations for them from an NMT model. This makes them amenable to the general methodology followed in this paper. Figure 1a shows an example sentence with syntactic dependencies.", |
| "cite_spans": [ |
| { |
| "start": 616, |
| "end": 617, |
| "text": "3", |
| "ref_id": null |
| }, |
| { |
| "start": 913, |
| "end": 926, |
| "text": "(Mel\u010duk 1988;", |
| "ref_id": null |
| }, |
| { |
| "start": 927, |
| "end": 938, |
| "text": "Nivre 2005)", |
| "ref_id": "BIBREF83" |
| }, |
| { |
| "start": 1320, |
| "end": 1338, |
| "text": "(Nivre et al. 2017", |
| "ref_id": "BIBREF84" |
| }, |
| { |
| "start": 1561, |
| "end": 1583, |
| "text": "(Williams et al. 2016)", |
| "ref_id": "BIBREF116" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 451, |
| "end": 458, |
| "text": "Table 1", |
| "ref_id": null |
| }, |
| { |
| "start": 1801, |
| "end": 1810, |
| "text": "Figure 1a", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Syntax", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "The holy grail in MT has long been to achieve an interlingua-based translation model, where the goal is to capture the meaning of the source sentence and generate a target sentence with the same meaning. It has been believed since the inception of MT", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Semantics", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Example sentence with syntactic and semantic relations. (a) Syntactic relations according to the Universal Dependencies formalism. Here \"Obama\" and \" Netanyahu\" are the subject and object of \"receives\", respectively, obl refers to an oblique relation of the locative modifier, nmod denotes the genitive relation, the prepositions \"in\" and \"of\" are treated as case-marking elements, and \"the\" is a determiner. See https://universaldependencies.org/guidelines.html for detailed definitions. (b) Semantic relations according to the PSD formalism. Here ACT-arg and PAT-arg refer respectively to the originator and affected arguments of \"receives\", LOC in the location, and APP is the thing that \"capital\" belongs to. For detailed definitions, see Cinkov\u00e1 et al. (2004) .", |
| "cite_spans": [ |
| { |
| "start": 743, |
| "end": 764, |
| "text": "Cinkov\u00e1 et al. (2004)", |
| "ref_id": "BIBREF26" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 1", |
| "sec_num": null |
| }, |
| { |
| "text": "that without acquiring such meaning representations it will be impossible to generate human-like translations (Weaver 1955) . Traditional statistical MT systems are weak at capturing meaning representations (e.g., \"who does what to whom-namely, what are the agent, the action, and the patient in the sentence [Jones et al. 2012] ). Although neural MT systems are also trained only on parallel data, without providing any direct supervision of word meaning, they are a continuous space model, and are believed to capture word meaning. Johnson et al. (2017) , for example, found preliminary evidence that the shared architecture in their multilingual NMT systems learns a universal interlingua. There have also been some recent efforts to incorporate such information in NMT systems, either explicitly (Rios Gonzales, Mascarell, and Sennrich 2017) or implicitly (Liu, Lu, and Neubig 2018) .", |
| "cite_spans": [ |
| { |
| "start": 110, |
| "end": 123, |
| "text": "(Weaver 1955)", |
| "ref_id": "BIBREF115" |
| }, |
| { |
| "start": 309, |
| "end": 328, |
| "text": "[Jones et al. 2012]", |
| "ref_id": "BIBREF51" |
| }, |
| { |
| "start": 534, |
| "end": 555, |
| "text": "Johnson et al. (2017)", |
| "ref_id": "BIBREF50" |
| }, |
| { |
| "start": 860, |
| "end": 886, |
| "text": "(Liu, Lu, and Neubig 2018)", |
| "ref_id": "BIBREF69" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 1", |
| "sec_num": null |
| }, |
| { |
| "text": "Tagging task. In this article, we study how semantic information is captured in NMT through two tasks: lexical semantic tagging and semantic dependency labeling. First, we utilize the lexical semantic (SEM) tagging task introduced by Bjerva, Plank, and Bos (2016). It is a sequence labeling task: Given a sentence, the goal is to assign to each word a tag representing a semantic class. This is a good task to use as a starting point for investigating semantics because: (i) tagging words with semantic labels is very simple, compared with building complex relational semantic structures; (ii) it provides a large supervised data set to train on, in contrast to most of the available data sets on word sense disambiguation, lexical substitution, and lexical similarity; and (iii) the proposed SEM tagging task is an abstraction over POS tagging 4 aimed at being language-neutral, and oriented to multilingual semantic parsing, all relevant aspects to machine translation. Table 1 shows an example sentence annotated with SEM tags. The semantic tag ENS describes a present-simple event category.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 972, |
| "end": 979, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Figure 1", |
| "sec_num": null |
| }, |
| { |
| "text": "The second semantic task is semantic dependency labeling, the task of assigning a type to each arc in a semantic dependency graph. Such dependencies are also known as predicate-argument relations, and may be seen as a first step toward semantic structure. They capture different aspects from syntactic relations, as can be noticed by the different graph structure (compare Figure 1b to Figure 1a ). Predicate-argument relations have also been used in many (non-neural) MT systems (Komachi, Matsumoto, and Nagata 2006; Wu et al. 2011; Xiong, Zhang, and Li 2012; Li, Resnik, and Daum\u00e9 III 2013) . Figure 1b shows an example sentence annotated with Prague Semantic Dependencies (PSD), a reduction of the tectogrammatical annotation in the Prague Czech-English Dependency Treebank (Cinkov\u00e1 et al. 2004; Cinkov\u00e1 et al. 2009) , which was made available as part of the Semantic Dependency Parsing shared tasks in SemEval (Oepen et al. 2014 (Oepen et al. , 2015 .", |
| "cite_spans": [ |
| { |
| "start": 480, |
| "end": 517, |
| "text": "(Komachi, Matsumoto, and Nagata 2006;", |
| "ref_id": "BIBREF63" |
| }, |
| { |
| "start": 518, |
| "end": 533, |
| "text": "Wu et al. 2011;", |
| "ref_id": "BIBREF118" |
| }, |
| { |
| "start": 534, |
| "end": 560, |
| "text": "Xiong, Zhang, and Li 2012;", |
| "ref_id": "BIBREF122" |
| }, |
| { |
| "start": 561, |
| "end": 592, |
| "text": "Li, Resnik, and Daum\u00e9 III 2013)", |
| "ref_id": "BIBREF66" |
| }, |
| { |
| "start": 777, |
| "end": 798, |
| "text": "(Cinkov\u00e1 et al. 2004;", |
| "ref_id": "BIBREF26" |
| }, |
| { |
| "start": 799, |
| "end": 819, |
| "text": "Cinkov\u00e1 et al. 2009)", |
| "ref_id": "BIBREF27" |
| }, |
| { |
| "start": 914, |
| "end": 932, |
| "text": "(Oepen et al. 2014", |
| "ref_id": "BIBREF86" |
| }, |
| { |
| "start": 933, |
| "end": 953, |
| "text": "(Oepen et al. , 2015", |
| "ref_id": "BIBREF85" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 373, |
| "end": 382, |
| "text": "Figure 1b", |
| "ref_id": null |
| }, |
| { |
| "start": 386, |
| "end": 395, |
| "text": "Figure 1a", |
| "ref_id": null |
| }, |
| { |
| "start": 595, |
| "end": 604, |
| "text": "Figure 1b", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Figure 1", |
| "sec_num": null |
| }, |
| { |
| "text": "We follow a 3-step process for studying linguistic information learned by the trained neural MT systems. The steps include: (i) training a neural MT system; (ii) using the trained model to generate feature representations for words in a language of interest; and (iii) training a classifier using generated features to make predictions for the different linguistic tagging tasks. The quality of the trained classifier on the given task serves as a proxy to the quality of the generated representations. It thus provides a quantitative measure of how well the original MT system learns features that are relevant to the given task.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Methodology", |
| "sec_num": "4." |
| }, |
| { |
| "text": "In this work, we focus on neural MT systems trained using the sequence-tosequence with attention architecture (Bahdanau, Cho, and Bengio 2014), where an encoder network first encodes the source sentence, followed by an attention mechanism to compute a weighted average of the encoder states that the decoder network uses to generate the target sentence. Both the encoder and the decoder networks are recurrent neural networks in our case. Several other architectures, for example the Transformer models (Vaswani et al. 2017) , have recently been proposed for neural MT. We discuss these briefly in Section 10.5.", |
| "cite_spans": [ |
| { |
| "start": 503, |
| "end": 524, |
| "text": "(Vaswani et al. 2017)", |
| "ref_id": "BIBREF113" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Methodology", |
| "sec_num": "4." |
| }, |
| { |
| "text": "Formally, let s = {s 1 , s 2 , . . . , s N } denote a source sentence, and t = {t 1 , t 2 , . . . , t M } denote a target sentence, where s i and t i are words. We first describe the simple case where we have word-level model and linguistic properties. Later we extend this scenario to subword units and to linguistic properties that involve multiple words.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Methodology", |
| "sec_num": "4." |
| }, |
| { |
| "text": "We first use the encoder (Equation 1) to compute a set of hidden states h = {h 1 , h 2 , . . . , h N }, where h i represents the hidden state for word s i . The encoder is a stacked LSTM with L layers, where the output of layer l \u2212 1 is passed as input to layer l (at each timestep). We then use an attention mechanism to compute a weighted average of these hidden states from the previous decoder state (d i\u22121 ), known as the context vector c i (Equation 2). The context vector is a real valued vector of k dimensions, which is set to be the same as the hidden states in our case. The attention model computes a weight w h i for each hidden state of the encoder, thus giving soft alignment for each target word. The context vector is then used by the decoder (Equation 3), which is also a stacked LSTM, to generate the next word in the target sequence:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Methodology", |
| "sec_num": "4." |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "ENC s i : s i , e i\u22121 \u2192 h i (1 \u2264 i \u2264 N)", |
| "eq_num": "(1)" |
| } |
| ], |
| "section": "Methodology", |
| "sec_num": "4." |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "ATTN i : {h 1 , . . . , h N }, d i\u22121 , t i\u22121 \u2192 c i \u2208 R k (1 \u2264 i \u2264 M) (2) DEC t i : c i , d i\u22121 , t i\u22121 \u2192 t i (1 \u2264 i \u2264 M)", |
| "eq_num": "(3)" |
| } |
| ], |
| "section": "Methodology", |
| "sec_num": "4." |
| }, |
| { |
| "text": "After training the NMT system, we freeze the parameters of the network and use the encoder or the decoder as a feature extractor to generate vectors representing words in the sentence. Let ENC l s i denote the representation of a source word s i at layer l in our stacked LSTM. We use ENC l s i from a particular layer l or concatenate all layer representations to train the external classifier for predicting a linguistic tag for s i . The quality of the representation can be deduced from our ability to train a good classifier. For word representations on the target side, we feed our word of interest t i as the previously predicted word, and extract the representation DEC t i (see Figure 2 for illustration).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 687, |
| "end": 695, |
| "text": "Figure 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Methodology", |
| "sec_num": "4." |
| }, |
| { |
| "text": "Generating representations for dependency labeling. We used dependency structures to evaluate the syntactic and semantic quality of the learned NMT representations (see Sections 3.2 and 3.3 for details). Given two words that are known to participate in a relation, a classifier is trained to predict the relation type. For the relation labeling task, the input to the classifier is a concatenation of encoder representations for two words in a relation, ENC l s i and ENC l s j , where (s i , s j ) is a known dependency pair with head s i and modifier s j . Again, we perform experiments with both representations from", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Methodology", |
| "sec_num": "4." |
| }, |
| { |
| "text": "Illustration of our approach: After training an NMT system on parallel data, we extract activations as features from the encoder/decoder and use these along with the labels to train an external classifier. For morphological tagging, we consider the activations for the word alone, while for syntactic/semantic relations we concatenate the activations for the two words involved in the relation. a particular layer l and the concatenated representation from all layers. Note that this formulation assumes that the order of the dependency is known. This formulation can be seen as a dependency labeling problem, where dependency labels are predicted independently. Although limited in scope, this formulation captures a basic notion of structural relations between words. 5", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 2", |
| "sec_num": null |
| }, |
| { |
| "text": "Generating representations with subword and character units. Previous work on analyzing NMT representations has been limited to the analysis of word representations only, where there is a one-to-one mapping from translation units (words) and their NMT representations (hidden states) to their linguistic annotations (e.g., POS tags). 6 In the case of character-or BPE-based systems, each word is split into multiple translation units, and each unit has its own representation. It is less trivial to define which representations should be evaluated when predicting a linguistic property such as the part-of-speech. In this work, we consider two simple approximations, illustrated in Figure 3:", |
| "cite_spans": [ |
| { |
| "start": 334, |
| "end": 335, |
| "text": "6", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 2", |
| "sec_num": null |
| }, |
| { |
| "text": "Average: For every source (or target) word, average the activation values of all the subwords (or characters) comprising it. In the case of a bidirectional encoder, we concatenate the averages from the forward and backward encoders' activations on the subwords (or characters) that represent the current word. 7", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "1.", |
| "sec_num": null |
| }, |
| { |
| "text": "Last: Consider the activation of the last subword (or character) as the representation of the word. For the bidirectional encoder, concatenate the forward encoder's activation on the last subword unit with the backward encoder's activation on the first subword unit.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2.", |
| "sec_num": null |
| }, |
| { |
| "text": "This formalization allows us to analyze the quality of character-and subword-based representations via prediction tasks, which has not been explored before.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2.", |
| "sec_num": null |
| }, |
| { |
| "text": "We experiment with several languages with varying degrees of morphological richness and syntactic divergence (compared to English): Spanish (es), French (fr), German (de), Czech (cs), Arabic (ar), Russian (ru), and Hebrew (he). We trained NMT systems using data made available by the two popular machine translation campaigns, namely, WMT (Bojar et al. 2017 ) and IWSLT (Cettolo et al. 2016) . The MT models were trained using a concatenation of NEWS, TED, and Europarl training data (\u2248 2.5M sentence pairs). The multilingual systems were trained by simply concatenating data from different", |
| "cite_spans": [ |
| { |
| "start": 339, |
| "end": 357, |
| "text": "(Bojar et al. 2017", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 370, |
| "end": 391, |
| "text": "(Cettolo et al. 2016)", |
| "ref_id": "BIBREF20" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "NMT Training Data", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "Illustration of a bidirectional layer. Representations from the forward and backward layers are concatenated. For the average method, all of the hidden states corresponding to subwords or characters of a given word are averaged together for each layer. For the last method, only the hidden state of the final subword or character is considered.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 3", |
| "sec_num": null |
| }, |
| { |
| "text": "language pairs (a total of \u224810M sentence pairs) and training a shared encoder-decoder pipeline. We used German, French, Spanish, and Czech to/from English to train multilingual systems. Language codes were added as prefixes before each sentence. We used official TED test sets to report translation quality (Papineni et al. 2002) . We also used the fully aligned United Nations corpus (Ziemski, Junczys-Dowmunt, and Pouliquen 2016) for training the models in some of our experiments. It includes six languages: Arabic, Chinese, English, French, Spanish, and Russian. This data set has the benefit of multiple alignment of the several languages, which allows for comparable cross-linguistic analysis, for example, studying the effect of only changing the target language. We used the first 2 million sentences of the training set, using the official training/development/test split.", |
| "cite_spans": [ |
| { |
| "start": 307, |
| "end": 329, |
| "text": "(Papineni et al. 2002)", |
| "ref_id": "BIBREF88" |
| }, |
| { |
| "start": 385, |
| "end": 431, |
| "text": "(Ziemski, Junczys-Dowmunt, and Pouliquen 2016)", |
| "ref_id": "BIBREF128" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 3", |
| "sec_num": null |
| }, |
| { |
| "text": "5.2.1 Preprocessing. We used the standard Moses pipeline for preprocessing the data, which includes tokenization, filtering for length, and true-casing. The systems were trained with a a maximum sentence length of 80 words. For the BPE systems, we used a vocabulary size using 50,000 operations. In the case of multilingual systems, we used 90,000 operations. For the character-based systems, we simply split the words into characters. 8 We used Morfessor (Smit et al. 2014) with default settings to get morpheme-segmented data. The subword (BPE and Morfessor) and character-based systems were trained with a maximum sentence length of 100, 100-120, and 400-550 units, respectively. 9", |
| "cite_spans": [ |
| { |
| "start": 436, |
| "end": 437, |
| "text": "8", |
| "ref_id": null |
| }, |
| { |
| "start": 456, |
| "end": 474, |
| "text": "(Smit et al. 2014)", |
| "ref_id": "BIBREF107" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Neural MT Systems", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "Training. We used the seq2seq-attn implementation (Kim 2016) with the following default settings: word embeddings and LSTM states with 500 dimensions, initialized with default Torch initialization, SGD with an initial learning rate of 1.0 and decay rate of 0.5 (after the ninth epoch), and dropout rate of 0.3. We used 2-4 hidden layers for both the encoder and the decoder. The NMT system was trained for 20 epochs, and the model with the best validation loss was used for generating features for the external classifier. These are the settings that we have generally used for the experiments reported in this article. We will explicitly mention in the individual sections where we digress from the prescribed settings.", |
| "cite_spans": [ |
| { |
| "start": 50, |
| "end": 60, |
| "text": "(Kim 2016)", |
| "ref_id": "BIBREF54" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model", |
| "sec_num": "5.2.2" |
| }, |
| { |
| "text": "The classifier is a logistic regression model whose input is either hidden states in wordbased models, or Last or Average representations in character-and subword-based models. Because we concatenate forward and backward states from all layers, this ends up being 4,000/2,000 dimensions when classifying the encoder/decoder: 500 dimensions \u00d7 4 layers \u00d7 2 directions (1 for decoder). The objective function is categorical cross-entropy, optimized by Adam (Kingma and Ba 2014) . Training is run with shuffled mini-batches of size 512 and stopped after 20 epochs. The choice of classifier is motivated by two considerations. First, the classifier takes features only from the current word (or word-pair), without additional context. The goal is to evaluate how well the word representation itself captures pertinent information, potentially including contextual information through the NMT LSTM encoder or decoder. Second, using a linear classifier enables focusing on the quality of the representations learned by the NMT system, rather than obtaining state-of-the-art prediction performance. In the literature on analyzing neural representations by classification tasks, simple linear classifiers are a popular choice (Belinkov and Glass 2019) . Using a stronger classifier may lead to better overall numbers, but does not typically change the relative quality of different representations (Qian, Qiu, and Huang 2016b; Belinkov 2018, Chapter D.1), which is our main concern in this work.", |
| "cite_spans": [ |
| { |
| "start": 454, |
| "end": 474, |
| "text": "(Kingma and Ba 2014)", |
| "ref_id": "BIBREF57" |
| }, |
| { |
| "start": 1217, |
| "end": 1242, |
| "text": "(Belinkov and Glass 2019)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Classifier Settings", |
| "sec_num": "5.3" |
| }, |
| { |
| "text": "We make use of gold-standard annotations wherever available, but in some cases we have to rely on using automatic taggers to obtain the annotations. In particular, to analyze the representations on the decoder side, we require parallel sentences. 10 It is difficult to obtain gold-standard data with parallel sentences, so we rely on automatic annotation tools. An advantage of using automatic annotations, though, is that we can reduce the effect of domain mismatch and high out-of-vocabulary (OOV) rate in analyzing these representations.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Supervised Data and Annotations", |
| "sec_num": "5.4" |
| }, |
| { |
| "text": "We used Tree-Tagger (Schmid 1994) for annotating Russian and the MADAMIRA tagger (Pasha et al. 2014) for annotating Arabic. For the remaining languages (French, Table 3 Train and test data (number of sentences) used to train MT classifiers to predict different tasks. We used automated tools to annotate data for the morphology tasks and gold annotations for syntactic and semantics tasks. German, Spanish, and Czech) we used RDRPOST (Nguyen et al. 2014 ), a state-of-the-art morphological tagger. For experiments using gold tags, we used the Arabic Treebank for Arabic (with the versions and splits described in the MADAMIRA manual) and the Tiger corpus for German. 11 For semantic tagging, we used the semantic tags from the Groningen Parallel Meaning Bank (Abzianidze et al. 2017) . For syntactic relation labeling we used the Universal Dependencies data set (Nivre et al. 2017) . For CCG supertagging we used the English CCGBank (Hockenmaier and Steedman 2007) . 12 For semantic dependency labeling we used PSD, which is a reduction of the tectogrammatical analysis layer of the Prague Czech-English Dependency Treebank, and is made available as part of the Semantic Dependency Parsing data set (Oepen et al. 2014 (Oepen et al. , 2015 . Most of the PSD dependency labels mark semantic roles of arguments, which are called functors in the Prague dependency treebank. 13 PSD annotations are available in English and Czech. Table 3 provides the amount of data used to train the MT classifiers for different NLP tasks. Table 4 details the number of tags (or labels) in each task across different languages.", |
| "cite_spans": [ |
| { |
| "start": 20, |
| "end": 33, |
| "text": "(Schmid 1994)", |
| "ref_id": "BIBREF102" |
| }, |
| { |
| "start": 81, |
| "end": 100, |
| "text": "(Pasha et al. 2014)", |
| "ref_id": "BIBREF89" |
| }, |
| { |
| "start": 434, |
| "end": 453, |
| "text": "(Nguyen et al. 2014", |
| "ref_id": "BIBREF82" |
| }, |
| { |
| "start": 759, |
| "end": 783, |
| "text": "(Abzianidze et al. 2017)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 862, |
| "end": 881, |
| "text": "(Nivre et al. 2017)", |
| "ref_id": "BIBREF84" |
| }, |
| { |
| "start": 933, |
| "end": 964, |
| "text": "(Hockenmaier and Steedman 2007)", |
| "ref_id": "BIBREF47" |
| }, |
| { |
| "start": 967, |
| "end": 969, |
| "text": "12", |
| "ref_id": null |
| }, |
| { |
| "start": 1199, |
| "end": 1217, |
| "text": "(Oepen et al. 2014", |
| "ref_id": "BIBREF86" |
| }, |
| { |
| "start": 1218, |
| "end": 1238, |
| "text": "(Oepen et al. , 2015", |
| "ref_id": "BIBREF85" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 161, |
| "end": 168, |
| "text": "Table 3", |
| "ref_id": null |
| }, |
| { |
| "start": 1425, |
| "end": 1432, |
| "text": "Table 3", |
| "ref_id": null |
| }, |
| { |
| "start": 1519, |
| "end": 1526, |
| "text": "Table 4", |
| "ref_id": "TABREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Supervised Data and Annotations", |
| "sec_num": "5.4" |
| }, |
| { |
| "text": "In this section, we investigate what kind of morphological information is captured within NMT models, using the tasks of POS and morphological tagging. To probe this, we annotated a subset of the training data (see Table 3 ) using POS or morphological 11 http://www.ims.uni-stuttgart.de/forschung/ressourcen/korpora/tiger.html. 12 There are no available CCG banks for the other languages we experiment with, except for a German CCG bank, which is not publicly available (Hockenmaier 2006) . 13 The main differences between PSD and the original tectogrammatical annotation are the omission of elided elements, such that all nodes are surface tokens; the inclusion of functional and punctuation tokens; ignoring most cases of function word attachments to content words; ignoring coreference links; and ignoring grammatemes (tectogrammatical correlates of morphological categories). As a side effect, these simplifications make it straightforward to generate representations for surface tokens participating in dependency relations under the PSD formalism. See http://sdp.delph-in.net for more information on PSD and refer to Cinkov\u00e1 et al. (2009) for details on the original tectogrammatical annotations. taggers. We then generated features from the trained NMT models and trained a linear classifier using these features to predict the POS or morphological tags. Although our goal is not to surpass state-of-the-art tagging performance, we still wanted to compare against several reference points to assess the quality of the underlying representations. To this end we report several baselines: (i) A simple local majority baseline where each word is assigned its most frequent tag and unknown words are assigned the most frequent global tag. (ii) We annotated the data used to train NMT models using the tools mentioned above and trained char-to-tag models using the same sequence-to-sequence regime we used to train our MT systems. This can be seen as a skyline reference. (iii) To have a closer comparison with our MT classifier, we generate features from the trained char-to-tag models and train a linear classifier using these features. This allows us to exactly compare representations learned for the task of translation versus the representations that are directly optimized toward the task (POS or morphological tagging, for example). Table 5 shows the prediction accuracy of the classifiers trained on the encoderside representations. MT classifiers always outperform the majority baseline, which entails that the representations contain non-trivial linguistic information about language Table 5 POS and morphological tagging results: Comparing classifier trained on char-based NMT representations with several baselines: (i) Local majority baseline (most frequent tag), (ii) Character-to-tag trained using sequence-to-sequence model on the same training data as the MT systems, (iii) Classifier trained on representations extracted from (ii) to match the MT generated representations). NMT systems used here to extract representations are character-based models, trained on translating each language to English (and English to German). The classifier results are substantially above the majority baseline, indicating that NMT representations learn non-trivial amounts of morphological information. morphology. The accuracy is high when the language is morphologically poor (e.g., English) or the task is simpler (fewer tags to predict; see Table 4 ). On the contrary, the accuracy in the case of a morphologically rich language such as Czech is lower.", |
| "cite_spans": [ |
| { |
| "start": 470, |
| "end": 488, |
| "text": "(Hockenmaier 2006)", |
| "ref_id": "BIBREF46" |
| }, |
| { |
| "start": 1123, |
| "end": 1144, |
| "text": "Cinkov\u00e1 et al. (2009)", |
| "ref_id": "BIBREF27" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 215, |
| "end": 222, |
| "text": "Table 3", |
| "ref_id": null |
| }, |
| { |
| "start": 2343, |
| "end": 2350, |
| "text": "Table 5", |
| "ref_id": null |
| }, |
| { |
| "start": 2597, |
| "end": 2604, |
| "text": "Table 5", |
| "ref_id": null |
| }, |
| { |
| "start": 3450, |
| "end": 3457, |
| "text": "Table 4", |
| "ref_id": "TABREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Morphology Results", |
| "sec_num": "6." |
| }, |
| { |
| "text": "The char-to-POS/Morph baselines seems to give much higher numbers compared with ours, but remember that these models are trained on considerably more data (the entire data on which the MT models were trained) and with a more sophisticated bilingual LSTM with attention model, compared with the MT classifier, which is trained on a small subset of neural activations using a simple logistic regression. A much closer skyline reference is the POS/Morph classifiers that are trained on the same data and model architecture as the MT classifier, with the difference that the former is trained on the representations optimized for the task itself whereas the latter is trained on the representations optimized toward the task of machine translation. Therefore, this is still comparing apples to oranges, but provides a more exact reference for the quality of MT representations with respect to learning morphology. We now proceed with answering more specific questions regarding several aspects of the NMT systems: (i) How do the representations trained from different translation units (word vs. character vs. subword units) compare? (ii) How do the representations trained from the encoder and decoder compare? (iii) What kind of information do different layers capture? and (iv) How does the target language affect the learned source language representations?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Morphology Results", |
| "sec_num": "6." |
| }, |
| { |
| "text": "We trained NMT systems with different translation units: word, character, and subword units, of which we tried two, namely, BPE (Sennrich, Haddow, and Birch 2016) and morphological segmentation (Smit et al. 2014) . For subword and character units, we found that the activation of the last subword/character unit of a word performed consistently better than using the average of all activations, so we present the results using the Last method throughout the article (see Table 6 for comparison). Figure 4 summarizes the results of predicting morphology with representations learned by different models. The character-based representations consistently outperformed other representations on all language pairs, while the word-based representations achieved the lowest accuracy. The differences are more significant in the case of languages with relatively complex morphology, notably Czech and Russian. We see a difference of up to 14% in favor of using character-based representations when compared with the word-based representations. The improvement is minimal in the Table 6 Classification accuracy for morphological tags using representations generated by aggregating BPE subword or character representations using either the average or the last LSTM state for each word. Here the representations are obtained by concatenating the encoding layers of NMT models trained on translating each language to English. Using the last hidden state consistently outperforms the average state. ", |
| "cite_spans": [ |
| { |
| "start": 194, |
| "end": 212, |
| "text": "(Smit et al. 2014)", |
| "ref_id": "BIBREF107" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 471, |
| "end": 478, |
| "text": "Table 6", |
| "ref_id": null |
| }, |
| { |
| "start": 496, |
| "end": 504, |
| "text": "Figure 4", |
| "ref_id": null |
| }, |
| { |
| "start": 1070, |
| "end": 1077, |
| "text": "Table 6", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Impact of Translation Unit on Learning Morphology", |
| "sec_num": "6.1" |
| }, |
| { |
| "text": "Morphological classification accuracy with different translation units and language pairs. When comparing encoder (decoder) representations, we train NMT models with different source (target) side translation units-words, BPE subwords, Morfessor subwords, or characters-and hold the target (source) side unit fixed as BPE subwords. case of English (1.2%), which is a morphologically simpler language. Comparing subword units as obtained using Morfessor and BPE, we found Morfessor to provide much better morphological tagging performance, especially in the case of the morphologically richer languages, Czech and Russian. The representations learned from morpheme-segmented units were found helpful in learning language morphology. These findings are also somewhat reflected in the translation quality (see Table 7 ). The character-based segmentation gave higher BLEU scores compared with a BPE-based system in the case of the morphologically rich language Czech, but character-based models performed poorly in the case of German, which requires handling long-distance dependencies. Our results (discussed later in Section 7) show that character-based representations are less effective at handling syntactic dependencies. 6.1.1 Handling Unknown and Low Frequency Words. We further investigated whether the performance difference between various representations is due to the difference in modeling infrequent and OOV words. As Table 8 shows, the morphologically richer languages have higher OOV rates. Figure 5 reveals that the gap between different representations is inversely related to the frequency of the word in the training data: Character-based models perform much better than others on less frequent and OOV words. The ranking of different units in low frequency regimes is consistent with the overall results in Figure 4 -characters perform best, followed by Morfessor subwords, BPE subwords, and words.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 807, |
| "end": 814, |
| "text": "Table 7", |
| "ref_id": null |
| }, |
| { |
| "start": 1428, |
| "end": 1435, |
| "text": "Table 8", |
| "ref_id": null |
| }, |
| { |
| "start": 1503, |
| "end": 1511, |
| "text": "Figure 5", |
| "ref_id": "FIGREF1" |
| }, |
| { |
| "start": 1824, |
| "end": 1832, |
| "text": "Figure 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Figure 4", |
| "sec_num": null |
| }, |
| { |
| "text": "The decoder DEC is a crucial part in an MT system with access to both source-side representations and partially generated target-side representations, which it uses to generate the next target word. We now examine whether the representations learned on the decoder-side possess the same amount of morphological knowledge as the encoder side. To probe this, we flipped the language direction and trained NMT systems with English\u2192{German, Czech, Russian} configurations. Then, we use the trained model to encode a source sentence and generate features for words in the target sentence. These features are used to train a classifier on morphological tagging on the target side. Note that in this case the decoder is given the correct target words one by one, similar to the usual NMT training regime. The right-hand side of Figure 4 shows a similar performance trend as in the case of encoder-side representations, with character units performing the best and word units performing the worst. Again, morphological units performed better than the BPE-based units.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 821, |
| "end": 829, |
| "text": "Figure 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Encoder versus Decoder Representations", |
| "sec_num": "6.2" |
| }, |
| { |
| "text": "Comparing encoder representations with decoder representations, it is interesting to see that in several cases the decoder-side representations performed better than the encoder-side representations, even though they are trained using a unidirectional LSTM only. Because we did not see any notable trends in differences between encoder and decoder side representations, we only present the encoder-side results in the rest of the paper.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Encoder versus Decoder Representations", |
| "sec_num": "6.2" |
| }, |
| { |
| "text": "Modern NMT systems use very deep architectures (Wu et al. 2016b; Zhou et al. 2016) . We are interested in understanding what kind of information different layers capture. Given a trained NMT model with multiple layers, we extract feature representations from the different layers in the encoder. We trained 4-layered models (using (NEW+TED+Europarl data). Figure 6 shows morphological tagging results using representations from different encoder and decoder layers across five language pairs. The general trend shows that representations from the first layer are better than those from the higher layers, for the purpose of capturing morphology. We found this observation to be true in multi-layered decoder as well (see the right side of Figure 6 ). We verified these findings with models trained using 2, 3, and 4 layers. Layer 1 was consistently found to give better accuracy on the task of POS tagging and morphology learning. We also found the pattern to hold for representations trained on other units (e.g., character-based units).", |
| "cite_spans": [ |
| { |
| "start": 47, |
| "end": 64, |
| "text": "(Wu et al. 2016b;", |
| "ref_id": "BIBREF120" |
| }, |
| { |
| "start": 65, |
| "end": 82, |
| "text": "Zhou et al. 2016)", |
| "ref_id": "BIBREF127" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 356, |
| "end": 364, |
| "text": "Figure 6", |
| "ref_id": null |
| }, |
| { |
| "start": 739, |
| "end": 747, |
| "text": "Figure 6", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Effect of Network Depth", |
| "sec_num": "6.3" |
| }, |
| { |
| "text": "Another interesting result to note is that concatenating representations from all the layers gave significantly better results compared to any individual layer (see Combination bars in Figure 6 ). This implies that although much of the information related to morphology is captured at the lower layer, some of it is also distributed to the higher layers. We analyzed individual tags across layers and found that open class categories such as verbs and nouns are distributed across several layers, although the majority of the learning of these phenomena is still done at layer 1. Please refer to Dalvi et al. (2019a) for further information.", |
| "cite_spans": [ |
| { |
| "start": 596, |
| "end": 616, |
| "text": "Dalvi et al. (2019a)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 185, |
| "end": 193, |
| "text": "Figure 6", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Effect of Network Depth", |
| "sec_num": "6.3" |
| }, |
| { |
| "text": "The task of machine translation involves translating from one language into another. While translating from morphologically rich languages is a challenging task, translating into such languages is even harder. How does the target language affect the learned source language representations? Does translating into a morphologically rich language", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Effect of Target Language", |
| "sec_num": "6.4" |
| }, |
| { |
| "text": "Morphological tagging accuracy using representations from layers 1 to 4, taken from encoders and decoders of different language pairs. Here, the NMT models were trained with BPE units. Layer 1 generates the best representations and in most cases there is a gradual decrease with each layer. The combination of representations improves beyond layer 1, indicating that some morphological information is distributed among layers. require more knowledge about source language morphology? To investigate these questions, we trained NMT models by keeping the source side constant and using different target languages. To make a fair comparison, the models are trained on the intersection of the training data based on the source language. Figure 7 shows the results of such an experiment, with models translating from Arabic to several languages: English, Hebrew, German, and Arabic itself. These target languages represent a morphologically poor language (English), a morphologically rich language with similar morphology to the source language (Hebrew), and a morphologically rich language with different morphology (German). As the figure shows, the representations that are learned when translating into English are better for predicting POS or morphology than those learned when translating into German, which are in turn better than those learned when translating into Hebrew.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 733, |
| "end": 741, |
| "text": "Figure 7", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Figure 6", |
| "sec_num": null |
| }, |
| { |
| "text": "How should we interpret these results? English is a morphologically poor language that does not display many of the morphological properties that are found in the Arabic source. In contrast, German and Hebrew have richer morphologies, so one could expect that translating into them would make the model learn more about morphology. However, Arabic representations learned from the Arabic\u2192English model are superior in learning morphology. A possible explanation for this phenomenon is that the Arabic\u2192English model is simply better than the Arabic\u2192Hebrew and Arabic-German models, as hinted by the BLEU scores. The inherent difficulty in translating Arabic to Hebrew/German may affect the ability to learn good representations of word structure or perhaps more data is required in the case of these languages to learn Arabic representations of the same quality. However, it turns out that an Arabic\u2192Arabic autoencoder learns to recreate the test sentences extremely well, even though its word representations are actually inferior for the purpose of POS/morphological tagging (Figure 7) . This implies that a higher BLEU score does not necessarily entail better morphological representations. In other words, a better translation model learns more informative representations, but only when it is actually learning to translate rather than merely memorizing the data as in the autoencoder case. We found these results to be consistent in other language pairs, that is, by changing the source from Arabic to German and Czech and also using character models instead of words (see Section A.2 in the Appendix for more details); however, more through study is required along this direction as Bisazza and Tump (2018) performed a similar experiment on a fine-grained tag level and found contrastive results.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 1076, |
| "end": 1086, |
| "text": "(Figure 7)", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Figure 6", |
| "sec_num": null |
| }, |
| { |
| "text": "To evaluate the NMT representations from a syntactic perspective, we consider two tasks. First, we made use of CCG supertagging, which is assumed to capture syntax at the word level. Second, we used dependency relations between any two words in the sentence for which a dependency edge exists, to investigate how words compose. Specifically, we ask the following questions: (i) Do NMT models acquire structural information while they are being trained on flat sequences of bilingual sentences? (ii) How do representations trained on different translation units (word vs. character vs. subword units) compare with respect to syntax? and (iii) Do higher layers learn better representations for these kinds of properties than lower layers?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Syntax Results", |
| "sec_num": "7." |
| }, |
| { |
| "text": "The analysis carried out previously was chiefly based on lexical properties. To strengthen our analysis, we further used dependency relations that are available for many different language pairs unlike CCG supertags. Here we concatenate the representations of two words in a relation and ask the classifier to predict their syntactic", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Syntax Results", |
| "sec_num": "7." |
| }, |
| { |
| "text": "Effect of target language on representation quality of the Arabic source. POS and morphological tagging accuracy, and BLEU scores, using encoder-side representations from NMT models trained with different target languages. These results are obtained with top-layer representations of 2-layer word-based NMT models. The target language has a small effect on source-side representation quality, with better MT models generating better representations, except in the auto-encoder case (Arabic bar).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 7", |
| "sec_num": null |
| }, |
| { |
| "text": "relation. Table 9 shows that NMT representations are syntax aware. In both tasks (CCG supertagging and syntactic dependency labeling), the classifier accuracy is much higher compared to the local majority baseline, 14 demonstrating that the representations learned during NMT training learn non-trivial amounts of syntactic information. 15 We now proceed to answer the other two questions, namely, the impacts of translation unit and representation depth.", |
| "cite_spans": [ |
| { |
| "start": 337, |
| "end": 339, |
| "text": "15", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 10, |
| "end": 17, |
| "text": "Table 9", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Figure 7", |
| "sec_num": null |
| }, |
| { |
| "text": "Although character-based models are effective at handling unknown and lowfrequency words, they have been found poor at capturing long-distance dependencies. Sennrich (2017) performed an evaluation based on contrastive translation pairs and found the subword-based system better in capturing long-distance dependencies. Here 14 For the syntactic dependency majority baseline, we assume the most frequent label of the arc (head-modifier pair). When the pair is unseen during test, we ignore the head and fall back to using modifier only. It is non-trivial to train sequence-to-sequence models for the dependency tasks, so we only rely on the majority baseline for comparison. 15 We do not have similar baselines for the syntax and semantic tasks as we have for the task of morphology prediction. The reason for this discrepancy is that we used automatic tools for annotating data for POS and morphological tagging, but gold annotated data for syntax and semantic tasks. Whereas state-of-the-art POS and morphological tagging tools are freely available, the same is not true for semantic and CCG tagging. We therefore resorted to use the published numbers as the skyline baseline in this case. For syntactic and semantic dependency labeling tasks an additional complexity is how to train a seq-to-seq baseline with such annotations. Remember that the task involves modeling head and modifier word to predict a dependency relation. In the case of semantic dependency, there could be multiple heads for a modifier word.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Impact of Translation Unit on Learning Syntax", |
| "sec_num": "7.1" |
| }, |
| { |
| "text": "Local majority baseline (most frequent tag/label) and classification accuracy using encoder representations generated by the NMT models, on syntactic tasks. The models are trained on translating each language to English (or German in the English case). The classifier results are far superior to the majority baseline, indicating that NMT representations contain a non-trivial amount of syntactic information. we directly pit the representations trained on different translation units against each other and compare their performance in predicting syntactic properties. Figure 8 shows that representations learned from subword units (BPE and Morfessor) consistently outperform the ones learned from character units in both tasks (CCG and syntactic dependency labeling), reinforcing the results found by Sennrich (2017). Characterbased models, on the other hand, do better than word-based models, which could be attributed to unknown words (in the word-based models). We found subword units, particularly those obtained using a morpheme-based segmentation, to give the best results. This could be because the linguistically motivated subword units are more aligned with the syntactic task than the compression-based BPE segmentation. A possible confound is that character-based models start from a lower linguistic level compared to word or subword models and may require more depth to learn longrange dependencies. To verify this, we trained 3-layered character models for Czechto-English and English-to-German. We extracted feature representations and trained", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 570, |
| "end": 578, |
| "text": "Figure 8", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Table 9", |
| "sec_num": null |
| }, |
| { |
| "text": "Dependency labeling and CCG supertagging accuracy using encoder representations obtained from NMT models trained with different translation units on the source side; the target side is fixed as BPE. Subword units generate better representations than character-or word-based ones.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 8", |
| "sec_num": null |
| }, |
| { |
| "text": "Syntactic and semantic dependency labeling accuracy using representations from layers 1 to 4, taken from encoders of BPE-based NMT models in different language pairs. Higher layers generate better representations for these tasks in most cases, but a combination of all layers works best, indicating that some relevant information is also captured in the lower layers.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 9", |
| "sec_num": null |
| }, |
| { |
| "text": "classifiers to predict syntactic dependency labels. Our results show that using an additional layer does improve the prediction accuracy, giving the same result as subword segmentation (Morfessor) in the case of Czech-to-English, but still worse in the case of English-to-Czech (see Table A .4 in the Appendix for results).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 283, |
| "end": 290, |
| "text": "Table A", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Figure 9", |
| "sec_num": null |
| }, |
| { |
| "text": "We previously found that morphology is predominantly being captured in layer 1 of the NMT models. We now repeat the experiments for syntactic dependencies. Figure 9a shows the results of predicting syntactic dependency labels using representations from different layers in the trained models. We found that representations from layer 4 performed better than representations from lower layers except for the French encoder, where layer 3 performs better. We also repeated this experiment with CCG supertags (see Table A .5 in the supplementary material) and found that higher layers (3 and 4) consistently outperform lower layers and except for English-Czech, the final layer gives the best accuracy in all cases. 16 These results are consistent with the syntactic dependency results. We repeated these experiments with the multi-parallel UN corpus by training English-to-{French, Arabic, Spanish, Russian, and English} bilingual models. Comparing successive layers (for example, comparing layer 2 versus layer 3), in the majority of the cases, the higher layer performed statistically significantly better than the lower one (\u03c1 < 0.01), according to the approximate randomization test (Pad\u00f3 2006) . 17 Similar to the results on morphological tagging, a combination of all layers achieved the best results. See the Combination bar in Figure 9a . This implies that although syntax is mainly learned at higher layers, syntactic information is at least partly distributed across the network.", |
| "cite_spans": [ |
| { |
| "start": 713, |
| "end": 715, |
| "text": "16", |
| "ref_id": null |
| }, |
| { |
| "start": 1185, |
| "end": 1196, |
| "text": "(Pad\u00f3 2006)", |
| "ref_id": "BIBREF87" |
| }, |
| { |
| "start": 1199, |
| "end": 1201, |
| "text": "17", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 156, |
| "end": 165, |
| "text": "Figure 9a", |
| "ref_id": null |
| }, |
| { |
| "start": 511, |
| "end": 518, |
| "text": "Table A", |
| "ref_id": null |
| }, |
| { |
| "start": 1333, |
| "end": 1342, |
| "text": "Figure 9a", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Effect of Network Depth", |
| "sec_num": "7.2" |
| }, |
| { |
| "text": "One possible concern with these results is that they may be appearing because of the stacked RNN layers, and not necessarily due to the translation task. In the extreme case, perhaps even a random computation that is performed in stacked RNN layers would lead to improved performance in higher layers. This may be especially concerning when predicting relation labels, as this requires combining information about two words in the sentence. To verify that the actual translation task is important, we can look at the performance with random models, initialized in the same manner but not trained at all. Figure 10 shows that higher layers in random networks generally generate worse representations. Layer 1 does improve the performance compared to layer 0 (word embeddings without contextual information) showing that some information is captured even in random models. However, after layer 1 the performance degrades drastically, demonstrating that higher layers in random models do not generate informative representations.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 604, |
| "end": 613, |
| "text": "Figure 10", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Effect of Network Depth", |
| "sec_num": "7.2" |
| }, |
| { |
| "text": "The experiment with random weights shows that training the NMT system is important. Does the actual translation task matter? Figure 10 also shows the results using representations from English-to-English models, that is, an autoencoder scenario. There", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 125, |
| "end": 134, |
| "text": "Figure 10", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Effect of Network Depth", |
| "sec_num": "7.2" |
| }, |
| { |
| "text": "Syntactic dependency labeling results with representations from different encoding layers of (word-based) NMT models trained on translating English to other languages (en-to-*, averaged over target languages), compared with an auto-encoder (en-to-en) and to untrained modeled with random weights (en-to-* rand). The MT-trained representations improve with each layer, while the random representations degrade after layer 1. The auto-encoder representations also improve but are below the MT-trained ones. These results show that learning to translate is important for obtaining good representations. is a notable degradation in representation quality when comparing the autoencoder results to those of the machine translation models. For example, the best results for predicting syntactic dependencies with the autoencoder are around 80% at layer 4. In contrast, the same layer in the translation models produces a score of 88%. In general, the representations from the machine translation models are always better than those from the autoencoder, and this gap increases as we go higher in the layers. This trend is similar to the results on morphological and semantic tagging with representations from autoencoders that were reported previously.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 10", |
| "sec_num": null |
| }, |
| { |
| "text": "In this section, we analyze two aspects of how information on syntactic dependencies is captured in different NMT layers: how different types of relations are represented and what the effect of head-modifier distance is. The results in this section are obtained using models trained on the United Nations corpus, as described in Section 5.1.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Analysis", |
| "sec_num": "7.3" |
| }, |
| { |
| "text": "Results of syntactic relation labeling at different relation distances using encoder representations from different layers of (word-based) NMT models trained on translating English to other languages (averaged over target language). Long-distance relations are difficult to capture at all layers, but higher layers mitigate more of the drop in performance.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 12", |
| "sec_num": null |
| }, |
| { |
| "text": "7.3.1 Effect of Relation Type. When are higher-layer representations especially important for syntactic relations? Figure 11 breaks down the performance according to the type of syntactic relations. The figure shows the five relations that benefit most from higher layer representations. 18 The general trend is that the quality of the representation improves with higher layers, with up to 20\u221225% improvement with representations from layer 4 compared with layer 1. The improvement is larger for certain relations: dependent clauses (advcl, ccomp), loose relations (list, parataxis), and other typically long-range dependencies such as conjunctions (conj) and appositions (appos). Core nominal arguments like subject (nsubj) and object (obj) also show consistent improvements with higher layers. Relations that do not benefit much from higher layers are mostly function words (aux, cop, det), which are local relations by nature, and the relation between a conjunct and the conjunction (cc), as opposed to the relation between two conjuncts (conj). These relations are local by nature and also typically less ambiguous. For example, the relation between a conjunction and and a noun is always labeled as cc, whereas a verb and a noun may have a subject or object relation.", |
| "cite_spans": [ |
| { |
| "start": 288, |
| "end": 290, |
| "text": "18", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 115, |
| "end": 124, |
| "text": "Figure 11", |
| "ref_id": "FIGREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Figure 12", |
| "sec_num": null |
| }, |
| { |
| "text": "In order to quantify the notions of global and local relations, let us consider relation distance. Figure 12 shows the representation quality as a function of the distance between the words participating in the relation. Predicting long-distance relations is clearly more difficult than predicting short-distance ones. As the distance between the words in the relation grows, the quality of the representations decreases. When no context is available (layer 0, corresponding to word embeddings), the performance quickly drops with longer distance relations. The drop is more moderate in the hidden layers, but in low layers the effect of relation distance can still be as high as 25%. Higher layers of the network mitigate this effect and bring the decrease down to under 5%. Moreover, every layer is performing better than the previous one at each distance group. This indicates that higher layers are much better at capturing long-distance syntactic information.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 99, |
| "end": 108, |
| "text": "Figure 12", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Effect of Relation Distance.", |
| "sec_num": "7.3.2" |
| }, |
| { |
| "text": "We now study how information on meaning is captured in NMT models in the context of lexical semantic (SEM) tagging and semantic dependency labeling tasks (refer to Section 3.3 for details on the tasks). We study the following specific questions: (i) Do NMT systems learn informative semantic representations? (ii) Can a neural network model learn to map a sequence of subwords or character symbols to a meaning representation? and (iii) What layers in the model learn more about semantic tags and relations?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Semantics Results", |
| "sec_num": "8." |
| }, |
| { |
| "text": "The experiments reported in this section on are conducted mainly on English, as the semantic tagging task and data set are recent developments that were initially only available in English. We also experiment and report results for German, for which a new semantic tagging data set is being developed. However, as the German annotations are very sparse (see Section 5.4), we performed a cross-fold evaluation when reporting results for German. For the semantic dependency labeling, we additionally used Czech data to strengthen the empirical evidence.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Semantics Results", |
| "sec_num": "8." |
| }, |
| { |
| "text": "In this section we only report encoder-side representation as the the analysis of decoder-side representations requires parallel data to generate the hidden representations and no standard tools for annotating the data exist. Table 10 shows the results. The classifier achieves 91.4% on the semantic tagging task and 85% and 80% on the task of semantic labeling for Czech and English, respectively. All results are significantly better than the local majority baseline, suggesting that NMT representations learn substantial semantic information.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 226, |
| "end": 234, |
| "text": "Table 10", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Semantics Results", |
| "sec_num": "8." |
| }, |
| { |
| "text": "Local majority baseline (most frequent tag/label) and classification accuracy using encoder representations generated by the NMT models, on semantic tasks. The models are trained on translating Czech to English (cs column) or English to German (en column). The classifier results are far superior to the majority baseline, indicating that NMT representations contain a non-trivial amount of semantic information. ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Table 10", |
| "sec_num": null |
| }, |
| { |
| "text": "Next we investigate whether the representations learned from characters or subword units can effectively model semantic information. We trained classifiers using the representations generated from different NMT models that were trained using character or subword units (BPE and Morfessor). Figure 13 summarizes the results on the semantic dependency labeling task and the semantic tagging task. In the semantic dependency labeling task, the character-based models perform significantly worse compared with the word-based and subword-based counterparts. We found that using subword-based representations, particularly morpheme-based segmentation, gives better performance in most scenarios. These results are in contrast with morphological tagging results, where character-based representations were consistently and significantly better compared with their subword counterparts. On comparing the prediction results between subword and character-based representations, we found that in many cases, characterbased models failed to predict the label correctly when the head and modifier words are further apart, that is, in the case of long-distance dependencies. However, this was not always true as in some cases character-based models were able to correctly predict the dependency label for a head that was 12 words apart. On semantic tags, subword-based (BPE and Morfessor) representations and character-based representation achieve comparable results for English. However, for German, BPE-based representations performed better than the other representations.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 290, |
| "end": 299, |
| "text": "Figure 13", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Impact of Translation Unit on Learning Semantics", |
| "sec_num": "8.1" |
| }, |
| { |
| "text": "We found the representations learned in the lower encoding layer to perform better on the task of morphological tagging. Here we investigate the quality of representations at different encoding layers, from the perspective of semantic properties.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Effect of Network Depth", |
| "sec_num": "8.2" |
| }, |
| { |
| "text": "Semantic tagging and dependency labeling results using representations of NMT models trained with different translation units on the source side; the target side is always BPE.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 13", |
| "sec_num": null |
| }, |
| { |
| "text": "Semantic tagging accuracy using features from the k-th encoding layer of 4-layered NMT models trained with different target languages. \"En\" column is an English autoencoder. BLEU scores are given for reference. Statistically significant differences from layer 1 are shown at p < 0.001 ( * ) and p < 0.01 ( * * ) . See text for details. Concerning lexical semantic tagging, as Table 11 shows, representations from layers 2 and 3 do not consistently improve performance above layer 1. However, representations from layer 4 lead to small but significant improvement with all target languages, according to the approximate randomization test. 19 We observed a similar pattern in the case of semantic dependency labeling tasks (see Figure 9b) , where higher layers (layer 4 in the case of Czech-English and English-German and layer 3 in the case of English-Czech) gave better accuracy. Intuitively, higher layers have a more global perspective because they have access to higher representations of the word and its context, while lower layers have a more local perspective. Layer 1 has access to context but only through one hidden layer, which may not be sufficient for capturing semantics. It appears that higher representations are necessary for learning even relatively simple lexical semantics and, especially, predicate-argument relations.", |
| "cite_spans": [ |
| { |
| "start": 285, |
| "end": 290, |
| "text": "( * )", |
| "ref_id": null |
| }, |
| { |
| "start": 304, |
| "end": 311, |
| "text": "( * * )", |
| "ref_id": null |
| }, |
| { |
| "start": 639, |
| "end": 641, |
| "text": "19", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 376, |
| "end": 384, |
| "text": "Table 11", |
| "ref_id": null |
| }, |
| { |
| "start": 727, |
| "end": 737, |
| "text": "Figure 9b)", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Table 11", |
| "sec_num": null |
| }, |
| { |
| "text": "In this section, we analyze three aspects of lexical semantic information as represented in the semantic tagging data set. First, we categorize semantic tags into coarse-grained categories and compare the classification quality within and across categories. Second, we perform a qualitative analysis of discourse relations and when they are better represented in different NMT layers. Third, we compare the quality of encoder representations when translating into different target languages. The results in this section are obtained using models trained on the United Nations corpus, as described in Section 5.1. Tag Level Analysis. The SEM tags are grouped in coarse-grained categories such as events, names, time, and logical expressions. Figure 14 shows the change in F 1 score (averaged over target languages) when moving from layer 1 to layer 4 representations. The blue bars describe the differences per coarse tag when directly predicting coarse tags. The red bars show the same differences when predicting fine-grained tags and micro-averaging inside each coarse tag. The former shows the differences between the two layers at distinguishing among coarse tags. The latter gives an idea of the differences when distinguishing between fine-grained tags within a coarse category. The first observation is that in the majority of cases there is an advantage for classifiers trained with layer 4 representations, that is, higher layer representations are better suited for learning the SEM tags, at both coarse and fine-grained levels.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 613, |
| "end": 616, |
| "text": "Tag", |
| "ref_id": null |
| }, |
| { |
| "start": 741, |
| "end": 750, |
| "text": "Figure 14", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Analysis of Lexical Semantics", |
| "sec_num": "8.3" |
| }, |
| { |
| "text": "Considering specific tags, higher layers of the NMT model are especially better at capturing semantic information such as discourse relations (DIS tag: subordinate vs. coordinate vs. apposition relations), semantic properties of nouns (roles vs. concepts, within the ENT tag), events and predicate tense (EVE and TNS tags), logic relations and quantifiers (LOG tag: disjunction, conjunction, implication, existential, universal, etc.) , and comparative constructions (COM tag: equatives, comparatives, and superlatives). These examples represent semantic concepts and relations that require a level of abstraction going beyond the lexeme or word form, and thus might be better represented in higher layers in the deep network.", |
| "cite_spans": [ |
| { |
| "start": 356, |
| "end": 434, |
| "text": "(LOG tag: disjunction, conjunction, implication, existential, universal, etc.)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Semantic", |
| "sec_num": "8.3.1" |
| }, |
| { |
| "text": "Relations. Now we analyze specific cases of disagreement between predictions using representations from layer 1 and layer 4. We focus on discourse relations, as they show the largest improvement when going from layer 1 to layer 4 representations (DIS category in Figure 14) . Intuitively, identifying discourse relations requires a relatively large context so it is expected that higher layers would perform better in this case.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 263, |
| "end": 273, |
| "text": "Figure 14)", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Analyzing Discourse", |
| "sec_num": "8.3.2" |
| }, |
| { |
| "text": "There are three discourse relations in the SEM tags annotation scheme: subordinate (SUB), coordinate (COO), and apposition (APP) relations. For each of those, Figure 15 Figure 14 Difference in semantic tagging F 1 when using representations from layer 4 compared to layer 1, showing F 1 when directly predicting coarse tags (blue) and when predicting fine-grained tags and averaging inside each coarse tag (red). The representations are taken from encoders of (word-based) NMT models trained on translating English to other languages (averaged over target language).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 159, |
| "end": 168, |
| "text": "Figure 15", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Analyzing Discourse", |
| "sec_num": "8.3.2" |
| }, |
| { |
| "text": "(examples 1-9) shows the first three cases in the test set where layer 4 representations correctly predicted the tag but layer 1 representations were wrong. Examples 1-3 have subordinate conjunctions (as, after, because) connecting a main and an embedded clause, which layer 4 is able to correctly predict. Layer 1 mistakes these as attribute tags (REL, IST) that are usually used for prepositions. In examples 4-5, the coordinate conjunction and is used to connect sentences/clauses, which layer 4 correctly tags as COO. Layer 1 wrongly predicts the tag AND, which is used for conjunctions connecting shorter expressions like words (e.g., \"murder and sabotage\" in example 1). Example 6 is probably an annotation error, as and connects the phrases \"lame gait\" and \"wrinkled skin\" and should be tagged as AND. In this case, layer 1 is actually correct. In examples 7-9, layer 4 correctly identifies the comma as introducing an apposition, while layer 1 predicts NIL, a tag for punctuation marks without semantic content (e.g., end-of-sentence period). As expected, in most of these cases identifying the discourse function requires a fairly large context.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Analyzing Discourse", |
| "sec_num": "8.3.2" |
| }, |
| { |
| "text": "Finally, we show in examples 10-12 the first three occurrences of AND in the test set, where layer 1 was correct and layer 4 was wrong. Interestingly, two of these (10, 11) are clear cases of and connecting clauses or sentences, which should have been annotated as COO, and the last (12) is a conjunction of two gerunds. The predictions from layer 4 in these cases thus appear justifiable.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Analyzing Discourse", |
| "sec_num": "8.3.2" |
| }, |
| { |
| "text": "Language. Does translating into different languages make the NMT system learn different source-side representations? We previously found a fairly consistent effect of the target language on the quality of encoder representations for POS and morphological tagging, with differences of \u223c2-3% in accuracy. Here we examine whether such an effect exists in SEM tagging. We trained 4-layered English-to-{Arabic,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Effect of Target", |
| "sec_num": "8.3.3" |
| }, |
| { |
| "text": "Examples of cases of disagreement between layer 1 (L1) and layer 4 (L4) representations when predicting semantic tags. The correct tag is italicized and the relevant word is underlined.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 15", |
| "sec_num": null |
| }, |
| { |
| "text": "Russian, French, Spanish, English} models with the multi-parallel UN corpus. Table 11 shows results using features obtained by training NMT systems on different target languages (the English source remains fixed). There are very small differences with different target languages (\u223c0.5%). While the differences are small, they are mostly statistically significant. For example, at layer 4, all the pairwise comparisons with different target languages are statistically significant (p < 0.001). Again, we note that training an English autoencoder results in much worse representations compared with MT models. In contrast, the autoencoder has excellent sentence recreation capability (96.6 BLEU). This indicates that learning to translate (to any foreign language) is important for obtaining useful representations for semantic tagging, as it is for morphological tagging.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 77, |
| "end": 85, |
| "text": "Table 11", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Figure 15", |
| "sec_num": null |
| }, |
| { |
| "text": "In the previous sections, we analyzed how the layer depth impacts the representations from the perspective of specific syntactic relations (Section 7.3.1) and lexical semantic tags (Section 8.3.1). We found that higher layers tend to better represent properties that are more global, loose, and abstract compared with lower layers. Does the same hold for semantic dependencies? Figure 16 shows the improvement from higher layer representations. The five most improved relations are highlighted. The following are examples from the PSD manual (Cinkov\u00e1 et al. 2004) . We give in parentheses the average distance in words between head and modifier, per relation. Four of the top five relations are looser kinds of relations that syntactically correspond to adjuncts: accompaniment (ACMP; \"He works without his glasses\"; distance 6.93); to whose advantage something happens (BEN; \"He did it for their sake\"; distance 4.40); direction (DIR1; \"He made a step from the wall\"; distance 4.41); and location (LOC; \"a match in a foreign country\"; distance 4.77); The fifth is an addressee argument (ADDR-arg; \"He gave the child a toy.\"; distance 3.69). These relations also have longer distances between head and modifier compared to the overall average distance (3.34 words).", |
| "cite_spans": [ |
| { |
| "start": 542, |
| "end": 563, |
| "text": "(Cinkov\u00e1 et al. 2004)", |
| "ref_id": "BIBREF26" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 378, |
| "end": 387, |
| "text": "Figure 16", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Analysis of Semantic Dependencies", |
| "sec_num": "8.4" |
| }, |
| { |
| "text": "In contrast, the relations that benefit the least (highlighted at the top of Figure 16 ) include a disjunctive that captures the relation between the disjunction \"or\" and a word in a list (DISJ.member; distance 2.22); expressing difference (DIFF; \"The goods were delivered four days later\"; distance 2.54 ); a degree specifier for expressing extent (EXT; e.g., the relation between \"about\" or \"almost\" and a quantity; distance 1.87); a rhematizer that often connects a negation word to its negated verb (RHEM; \"Cray Research did not want to...\", example from the PSD data set; distance 2.17); and linking the clause to the preceding text (PREC; \"Hence, I'm happy\"; distance 7.73). Of these, the first three actually drop in performance when using (some of the) higher layer representations, which may be explained by their more local, tight relation. This also partly accords with the distances, which are below average for the bottom 4. The PREC relation is an exception, with a relatively large distance but almost no benefit from higher layers. This can be explained by its use for words linking the clause to the preceding text, which are limited and easy to memorize (\"However\", \"Nevertheless\", \"Moreover\", etc.). As these cases are assigned the main verb as the PREC head, they may span over large distances, but their closed-class nature enables simple memorization even at low layers. However, relation distance does not explain the entire benefit from higher layers, as some of the most distant relations are not the ones that benefit most from higher layers. Still, similar to the syntactic dependencies, the representations from higher layers benefit more in case of looser, less tight semantic relations.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 77, |
| "end": 86, |
| "text": "Figure 16", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Analysis of Semantic Dependencies", |
| "sec_num": "8.4" |
| }, |
| { |
| "text": "Improvement in accuracy of semantic relation labeling with layers 2/3/4 compared with layer 1, when averaging results obtained with encoder representations from (word-based) NMT models trained on translating English to other languages. The five least/most improved relations are highlighted.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 16", |
| "sec_num": null |
| }, |
| { |
| "text": "Languages share orthography, morphological patterns, and grammar. Learning to translate between several pairs simultaneously can help improve the translation quality of the underlying languages and also enable translation for language pairs with little or no parallel data. Johnson et al. (2017) exploited a remarkably simple idea of appending a language code to each training sentence, in a shared encoder-decoder framework. The models are trained with all multilingual data consisting of multiple language pairs at once. In their projection of a small corpus of 74 triples of semantically identical cross-language (Japanese, Korean, English) phrases to 3D space via t-SNE, they found preliminary evidence that the shared architecture in their multilingual NMT systems learns a universal interlingua. We use this idea with our machinery to investigate how effective the multilingual representations are in learning morphology, syntax, and semantics, compared to their corresponding bilingual models.", |
| "cite_spans": [ |
| { |
| "start": 274, |
| "end": 295, |
| "text": "Johnson et al. (2017)", |
| "ref_id": "BIBREF50" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Comparison Against Multilingual Models", |
| "sec_num": "9." |
| }, |
| { |
| "text": "We trained 4-layer multilingual systems with many-to-one and one-to-many configurations with English on one side (encoder/decoder) and German, Spanish, French, and Czech on the other side. The models were trained with BPE subword units. We trained two versions of the multilingual model, one with the same parameters as the bilingual models and another with word embeddings and LSTM states with 1,024 dimensions (as opposed to the default 500 dimensions). Our goal was to investigate the effect of increasing model parameters on translation quality and representation quality in terms of the understudied linguistic phenomenon. The systems were trained on NEWS+TED+Europarl, approximately 2.5M sentences per language pair, and a total of 10M sentences for training the multilingual models. Figure 17 shows BLEU scores comparing bilingual and multilingual translation systems across different language pairs. We see that the many-to-one multilingual system (i.e., *-to-English) is on par or slightly behind bilingual systems when trained with the same number of parameters as the bilingual models. In contrast, the oneto-many multilingual system (i.e., English-to-*) is significantly worse compared with its bilingual counterparts. The reason for this discrepancy could be that generation is a harder task than encoding, especially when translating into morphologically-rich languages: An average difference of \u22121.35 is observed when translating out of English compared with \u22120.13 when translating into English. The larger multilingual models (with twice as many parameters) restored the baseline performance, in fact showing significant improvements in many cases. We also trained two of the bilingual baselines (Czech \u2194 English) by doubling the parameters. While the large multilingual system gave an improvement of +1.4 (see Table 12 ) over the baseline multilingual system by", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 790, |
| "end": 799, |
| "text": "Figure 17", |
| "ref_id": null |
| }, |
| { |
| "start": 1827, |
| "end": 1835, |
| "text": "Table 12", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Comparison Against Multilingual Models", |
| "sec_num": "9." |
| }, |
| { |
| "text": "Comparing bilingual and multilingual systems in terms of translation quality. The multilingual models are many-to-one (*-to-English) or one-to-many (English-to-*) during training, but run on a specific language pair during testing.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 17", |
| "sec_num": null |
| }, |
| { |
| "text": "Comparing bilingual and multilingual Czech\u2194English models across translation quality and classifying different linguistic properties. bi = bilingual model; bi* = larger bilingual model; multi = multilingual model; multi* = larger multilingual model. Larger models have twice the number of word embedding and LSTM state dimensions. The larger models improve substantially the multilingual systems, but only slightly improve the bilingual systems. Morphology = morphological tagging; Syntax = syntactic dependency labeling; Semantics = semantic dependency labeling. doubling the parameters size, the bilingual system only obtained an improvement of +0.2 by increasing the model size. A similar pattern was observed in the opposite direction where increasing the model size gave a BLEU improvement of +1.9 in the multilingual system, but only +0.3 in the case of the bilingual system. These results show that multilingual systems benefit from other language pairs being trained in tandem and suggest that the underlying representations are richer than the ones trained using bilingual models. We now proceed to analyze these representations in light of the understudied linguistic properties. Figures 18, 19a, and 19b show that the representations learned from multilingual models, despite sharing the encoder and decoder representations among four languages, can still effectively learn the same amount of morphology, syntax, and semantics as learned by their bilingual counterparts. In almost all cases, the multilingual representations are either better or at par with the bilingual models. Using a larger multilingual model (double the parameters size) gave consistent improvements in accuracy, which resonate with the improvement in translation quality.", |
| "cite_spans": [ |
| { |
| "start": 1190, |
| "end": 1201, |
| "text": "Figures 18,", |
| "ref_id": null |
| }, |
| { |
| "start": 1202, |
| "end": 1206, |
| "text": "19a,", |
| "ref_id": null |
| }, |
| { |
| "start": 1207, |
| "end": 1214, |
| "text": "and 19b", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Table 12", |
| "sec_num": null |
| }, |
| { |
| "text": "Focusing on morphology (Figure 18 ), a multilingual encoder generates representations that are better than its bilingual counterpart in most cases, while a multilingual decoder slightly degrades representation quality compared to the bilingual decoder. However, using a larger multilingual model leads to substantial improvements. These results mirror the patterns shown in terms of translation quality (Figure 17 ). In the case of syntactic and semantic dependencies (Figures 19a and 19b) , even the default-sized multilingual model works better than the bilingual model, and the larger multilingual", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 23, |
| "end": 33, |
| "text": "(Figure 18", |
| "ref_id": null |
| }, |
| { |
| "start": 403, |
| "end": 413, |
| "text": "(Figure 17", |
| "ref_id": null |
| }, |
| { |
| "start": 468, |
| "end": 489, |
| "text": "(Figures 19a and 19b)", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Table 12", |
| "sec_num": null |
| }, |
| { |
| "text": "Comparing bilingual and multilingual models in terms of morphological tagging accuracy. Same-size multilingual models benefit the encoder representations compared to bilingual models, but not decoder representations; larger multilingual models benefit also the decoder.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 18", |
| "sec_num": null |
| }, |
| { |
| "text": "Comparing bilingual and multilingual models with syntactic and semantic dependencies. Multilingual models consistently improve the NMT representations. model leads to additional small improvements in syntactic dependencies and in Czech semantic dependencies (the Czech-English bars in Figure 19b ). The larger multilingual model does not improve in the case of English semantic dependencies (English-Czech/German bars in Figure 19b ), even though the corresponding BLEU scores do improve with a larger multilingual model (Figure 17) . 20 In an effort to probe whether increasing the number of parameters in a bilingual model would result in similar performance improvements, Table 12 shows results across different properties in Czech\u2194English language pairs. We consistently found that increasing the model size does not lead to the same improvements as we observed in the case of multilingual models. These results reinforce that multilingual NMT models learn richer representations compared to the bilingual model and benefit from the shared properties across different language pairs.", |
| "cite_spans": [ |
| { |
| "start": 535, |
| "end": 537, |
| "text": "20", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 285, |
| "end": 295, |
| "text": "Figure 19b", |
| "ref_id": null |
| }, |
| { |
| "start": 421, |
| "end": 431, |
| "text": "Figure 19b", |
| "ref_id": null |
| }, |
| { |
| "start": 521, |
| "end": 532, |
| "text": "(Figure 17)", |
| "ref_id": null |
| }, |
| { |
| "start": 675, |
| "end": 683, |
| "text": "Table 12", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Figure 19", |
| "sec_num": null |
| }, |
| { |
| "text": "In this section, we discuss the overall patterns that arise from the experimental results from several angles. First, we discuss how to assess the overall quality of the learned NMT representations with regard to other baselines and upper bounds. Second, we consider NMT representations from the perspective of contextualized word representations and contrast them to recent popular representations. Third, we reflect on the methodological approach taken in this work, and what it may or may not tell us about how the NMT model exploits language representations. Fourth, we briefly discuss the relation of our results to other NMT architectures. Finally, we touch upon the role of analysis work in understanding and improving NMT models.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "10." |
| }, |
| { |
| "text": "The analyses presented in this work shed light on the quality of different language representations in NMT, with a particular focus on comparing various NMT components and design choices (layers, translation units, etc.). Our questions, therefore, have mostly been from a comparative perspective: How good are certain NMT representations compared with others with respect to certain linguistic tasks. But just how good are the NMT representations overall? Answers to this question may depend on the use case. One could, for example, evaluate the utility of NMT representations to improve stateof-the-art performance by plugging them as additional features in some strong model. Indeed, McCann et al. (2017) found this approach to yield state-of-the-art results in several language understanding tasks. It is also important to consider the quality of the NMT representations for the understudied tasks here in comparison to other baselines and competitive systems. Throughout the article, we have compared the results to a majority baseline, arguing that NMT representations obtain substantial improvements. Here we compare, for each linguistic task, the best performing NMT representations with several baselines and upper bounds. We compare with the local majority baseline (most frequent tag/label for each word according to the training data, and the globally most frequent tag/label for words not seen in training) and with a classifier trained on word embeddings that are pre-trained on the source-side of the MT training data. We also train an encoder-decoder on converting text to tags, by automatically annotating the source side of the MT parallel data. Then we use this encoder-decoder to tag the test set of the supervised data and evaluate its quality. Finally, we generate representations from the encoder of this encoder-decoder model and train a classifier on them to predict the tags. This setting aims to mimic our main scenario, except that we generate representations with an encoder-decoder specially trained on the linguistic task that we evaluate, rather than representations generated by an NMT model. Table 5 shows the results. A classifier trained on NMT representations performs far better than the majority baseline, as we have already confirmed. A similar classifier trained on representations from a task-specific encoder-decoder performs even better. This indicates that training on a specific task leads to representations more geared toward that task, as may be expected. In fact, a similar behavior has been noted with other contextual word representations (Liu et al. 2019) . Still, the representations do not contain all available information (or, not all information may be extracted by a simple classifier), as the task-specific encoder-decoder performs better than a classifier trained on its representations.", |
| "cite_spans": [ |
| { |
| "start": 678, |
| "end": 706, |
| "text": "Indeed, McCann et al. (2017)", |
| "ref_id": null |
| }, |
| { |
| "start": 2590, |
| "end": 2607, |
| "text": "(Liu et al. 2019)", |
| "ref_id": "BIBREF70" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 2125, |
| "end": 2132, |
| "text": "Table 5", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Assessing Representation Quality", |
| "sec_num": "10.1" |
| }, |
| { |
| "text": "The representations generated by NMT models may be thought of as contextualized word representations (CWRs), as they capture context via the NMT encoder or decoder. We have already mentioned one work exploiting this idea, known as CoVE (McCann et al. 2017) , which used NMT representations as features in other models to perform various NLP tasks. Other prominent contextualizers include ELMo (Peters et al. 2018a) , which trains two separate, forward and backward LSTM language models (with a character CNN building block) and concatenates their representations across several layers; GPT (Radford et al. 2018) and GPT-2 (Radford et al. 2019) , which use transformer language models based on self-attention (Vaswani et al. 2017); and BERT (Devlin et al. 2019) , which uses a bidirectional transformer model trained on masked language modeling (filling the blanks). All these generate representations that feed into task-specific classifiers, potentially with fine-tuning the contextualizer weights. 21 How do NMT representations compare with CWRs trained from raw text? Directly answering this question is beyond the scope of this work, and is also tricky to perform for two reasons. First, CWRs like ELMo, BERT, and GPT require very large amounts of data, on the order of billions of words, which is far beyond what typical NMT systems are trained on (our largest NMT systems are trained on an order of magnitude fewer words). Second, at present, various CWRs are trained in incomparable settings in terms of data, number of parameters, and infrastructure. It seems that a comparison of common CWRs themselves is necessary before they are compared to NMT representations.", |
| "cite_spans": [ |
| { |
| "start": 236, |
| "end": 256, |
| "text": "(McCann et al. 2017)", |
| "ref_id": "BIBREF74" |
| }, |
| { |
| "start": 393, |
| "end": 414, |
| "text": "(Peters et al. 2018a)", |
| "ref_id": "BIBREF91" |
| }, |
| { |
| "start": 590, |
| "end": 611, |
| "text": "(Radford et al. 2018)", |
| "ref_id": "BIBREF97" |
| }, |
| { |
| "start": 616, |
| "end": 643, |
| "text": "GPT-2 (Radford et al. 2019)", |
| "ref_id": null |
| }, |
| { |
| "start": 740, |
| "end": 760, |
| "text": "(Devlin et al. 2019)", |
| "ref_id": "BIBREF34" |
| }, |
| { |
| "start": 1000, |
| "end": 1002, |
| "text": "21", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Contextualized Word Representations", |
| "sec_num": "10.2" |
| }, |
| { |
| "text": "There is, however, indirect information that may tell us something about how CWRs trained on raw texts behave in comparison to NMT representations. Bowman et al. (2018) compared sentence encoders trained on a variety of tasks, including the CoVE translation representations, and evaluated on language understanding tasks. They found language modeling pre-training to perform best, but cautioned that without fine-tuning, many of the results are not far above trivial baselines. They also found that grammar-related tasks benefit more from such pre-training than meaning-oriented tasks. Peters et al. (2018b) compared the ELMo LSTM with similar systems based on convolutions or transformer-style self-attention. They found that all architectures learn hierarchical representations: The word embedding layer focuses on morphology, low encoding layers focus on local syntax, and high encoding layers carry more semantic information. These results are mostly in line with our findings concerning representation depth, although we have not noticed a clear separation between syntactic and semantic properties. Zhang and Bowman (2018) compared representations from NMT and bidirectional language models on POS tagging and CCG supertagging. They found the language model representations to consistently outperform those from NMT. In other work, we have found that language model representations are of similar quality to NMT ones in terms of POS and morphology, but are behind in terms of semantic tagging (Dalvi et al. 2019a ).", |
| "cite_spans": [ |
| { |
| "start": 148, |
| "end": 168, |
| "text": "Bowman et al. (2018)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 586, |
| "end": 607, |
| "text": "Peters et al. (2018b)", |
| "ref_id": "BIBREF92" |
| }, |
| { |
| "start": 1105, |
| "end": 1128, |
| "text": "Zhang and Bowman (2018)", |
| "ref_id": "BIBREF125" |
| }, |
| { |
| "start": 1499, |
| "end": 1518, |
| "text": "(Dalvi et al. 2019a", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Contextualized Word Representations", |
| "sec_num": "10.2" |
| }, |
| { |
| "text": "Tenney et al. (2019) compared representations from CoVE, ELMo, GPT, and BERT on a number of classiciation tasks, partially overlapping with the ones we study. They found that CWRs trained on raw texts outperform the MT representations of CoVE; however, as noted above these models are all trained in very different setups and cannot be fairly compared. Another interesting finding is that learning a weighted mix of layers works better than any one layer, and also better than concatenating. This again indicates that some layers are better than others for different tasks, consistent with our results. Concerning different tasks, Tenney et al. (2019) found that CWRs are especially helpful (compared with a lexical baseline) with syntactic tasks, such as dependency and constituent labeling, and less helpful with certain semantic tasks like capturing fine-grained semantic attributes and pronoun resolution. They did notice improvements with semantic roles, which are related to our predicate-argument relations, where we also noticed significant improvements at higher layers. Finally, Liu et al. (2019) compared ELMo, GPT, and BERT on various classification tasks in terms of their linguistic knowledge and transferability. They found that a simple classifier trained on top of the (frozen) representations led to state-of-the-art results in many cases, but failed on tasks requiring fine-grained linguistic knowledge like conjunct identification. They observed that the first layer of LSTM-based CWRs performs better than other layers, while in transformer-based models the intermediate layers are the best. Considering different pre-training tasks, higher layer representations were more task-specific (and less general) in LSTM models, but not in transformer models. In our investigation, the top layers of the (LSTM) NMT models were better for syntactic and semantic tasks. One possible explanation for this could be that translation is more aligned with the syntactic and semantic properties than language modeling.", |
| "cite_spans": [ |
| { |
| "start": 631, |
| "end": 651, |
| "text": "Tenney et al. (2019)", |
| "ref_id": "BIBREF111" |
| }, |
| { |
| "start": 1089, |
| "end": 1106, |
| "text": "Liu et al. (2019)", |
| "ref_id": "BIBREF70" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Contextualized Word Representations", |
| "sec_num": "10.2" |
| }, |
| { |
| "text": "The development and analysis of CWRs is still ongoing. At present, NMT representations appear to be weaker than those obtained by contextualizers trained on raw texts, at least when the latter are trained on much larger amounts of data. It remains to be seen whether NMT representations can complement raw-text CWRs in certain scenarios.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Contextualized Word Representations", |
| "sec_num": "10.2" |
| }, |
| { |
| "text": "Our methodological approach evaluates whether various linguistic properties are decodable from learned NMT representations. Our assumption was that the quality of a trained classifier can serve as a proxy to the quality of the original model, for a given task. However, it is not clear whether the NMT model really \"cares\" about the linguistic properties, in the sense that it relies on them for performing the translation tasks. In essence, we only provide correlational evidence, not causal evidence. This is a limitation of much of the work using classification tasks to analyze neural networks, as explained by Belinkov and Glass (2019) . One avenue for addressing this question in causal terms is to define interventions: Change something in the representation and test whether and how it impacts the output translation. Bau et al. (2019b) perform such intervention experiments in NMT. They identify individual neurons that capture certain morphological properties-gender, number, and tense-and modify their activations. They evaluate how such intervention affects the output translations, finding that tense is fairly well modified, but gender and number are not as affected. Following similar ideas may be a fruitful area for further investigation of various linguistic properties and how much NMT systems depend on them when producing a translation.", |
| "cite_spans": [ |
| { |
| "start": 615, |
| "end": 640, |
| "text": "Belinkov and Glass (2019)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 826, |
| "end": 844, |
| "text": "Bau et al. (2019b)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "On the Impact of Language Representation on Translation Output", |
| "sec_num": "10.3" |
| }, |
| { |
| "text": "There are various motivations for work on interpretability and analysis of neural network models in NLP and other domains. There are also questions concerning their necessity. Although this article does not aim to solve this debate, 22 we would like to highlight a few potential benefits of the analysis. First, several of our results may serve as guidelines for improving the quality of NMT systems and their utility for other tasks. The results on using different translation units suggest that their choice may depend on what properties one would like to capture. This may have implications for using MT systems in different languages (morphologically rich vs. poor, free vs. fixed word order) or genre (short, simple sentences vs. long, complex ones). The results on representation depth suggest that using NMT representations for contextualization may benefit from combining layers, maybe with task-wise weighting. One could also imagine performing multi-task learning of MT and other tasks, with auxiliary losses integrated in different layers. The results on multilingual systems indicate that such systems may lead to better representations, but often require greater capacity. Inspecting language representations in a zero-shot MT scenario (Johnson et al. 2017; Arivazhagan et al. 2019 ) may also yield new insights for improving such systems.", |
| "cite_spans": [ |
| { |
| "start": 233, |
| "end": 235, |
| "text": "22", |
| "ref_id": null |
| }, |
| { |
| "start": 1249, |
| "end": 1270, |
| "text": "(Johnson et al. 2017;", |
| "ref_id": "BIBREF50" |
| }, |
| { |
| "start": 1271, |
| "end": 1294, |
| "text": "Arivazhagan et al. 2019", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Why Analyze?", |
| "sec_num": "10.4" |
| }, |
| { |
| "text": "The NMT models analyzed in this work are all based on recurrent LSTM encoderdecoder models with attention. Although this is the first successful NMT architecture, and still a dominant one, it is certainly not the only one. Other sucessful architectures include fully convolutional (Gehring et al. 2017) and fully attentional, transformer encoder-decoder models (Vaswani et al. 2017) . There are also non-autoregressive models, which are promising in terms of efficiency (Gu et al. 2018) . At present, NMT systems based on transformer components appear to be the most successful. Combinations of transformer and recurrent components may also be helpful (Chen et al. 2018) .", |
| "cite_spans": [ |
| { |
| "start": 361, |
| "end": 382, |
| "text": "(Vaswani et al. 2017)", |
| "ref_id": "BIBREF113" |
| }, |
| { |
| "start": 470, |
| "end": 486, |
| "text": "(Gu et al. 2018)", |
| "ref_id": "BIBREF44" |
| }, |
| { |
| "start": 652, |
| "end": 670, |
| "text": "(Chen et al. 2018)", |
| "ref_id": "BIBREF22" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Other NMT Architectures", |
| "sec_num": "10.5" |
| }, |
| { |
| "text": "The generalization of the particular results in this work to other architectures is a question of study. Recent efforts to analyze transformer-based NMT models include attempts to extract syntactic trees from self-attention weights (Mare\u010dek and Rosa 2018; Raganato and Tiedemann 2018) and evaluating representations from the transformer encoder (Raganato and Tiedemann 2018) . The latter found that lower layers tend to focus on POS and shallow syntax, whereas higher layers are more focused on semantic tagging. These results are in line with our findings. However, more work is needed to understand the linguistic representational power of various NMT architectures. We expect the questions themselves, and the methods, to remain an active field of investigation with newer architectures and systems.", |
| "cite_spans": [ |
| { |
| "start": 232, |
| "end": 255, |
| "text": "(Mare\u010dek and Rosa 2018;", |
| "ref_id": "BIBREF73" |
| }, |
| { |
| "start": 256, |
| "end": 284, |
| "text": "Raganato and Tiedemann 2018)", |
| "ref_id": "BIBREF99" |
| }, |
| { |
| "start": 345, |
| "end": 374, |
| "text": "(Raganato and Tiedemann 2018)", |
| "ref_id": "BIBREF99" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Other NMT Architectures", |
| "sec_num": "10.5" |
| }, |
| { |
| "text": "In this article, we presented a comprehensive analysis of the representations learned during NMT training from the perspective of core linguistic phenomena, namely, morphology, syntax, and semantics. We evaluated the representation quality on the tasks of morphological, syntactic, and semantic tagging and using syntactic and semantic dependency labeling. Our results show that the representations learned during neural MT training learn a non-trivial amount of linguistic information. We found that different properties are represented to varying extents in different components of the NMT models. The main insights are:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion and Future Work", |
| "sec_num": "11." |
| }, |
| { |
| "text": "Comparing representations at different layer depths, we found that word morphology is learned at the lower layer in the LSTM encoder-decoder model, whereas non-local linguistic phenomena in syntax and semantics are better represented at the higher layers. For example, we found that higher layers are better at predicting clause-level syntactic dependencies, or second and third semantic arguments, in contrast to short-range dependencies, which do not benefit much from higher layers.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "\u2022 Comparing representations with different translation units, we found that representations learned using characters perform best at capturing word morphology, and therefore provide a more viable option when translating morphologically rich languages such as Czech. They are more robust toward handling unknown and low frequency words.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "\u2022 In contrast, representations learned from subword units are better at capturing syntactic and semantic information that requires learning non-local dependencies. Character-based representations, on the other hand, are poor at handling long-range dependencies and therefore inferior when translating syntactically divergent language pairs such as German-English.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "We found morpheme-segmented units to give better representations than the ones learned using non-linguistic BPE units. The former outperformed the latter in most scenarios, even giving slightly better translation quality.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "We found that multilingual models benefit from the shared properties across different language pairs and learn richer representations compared to the bilingual model.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "Future work can expand the analysis into many directions. For instance, in terms of the studied linguistic properties, moving beyond words and relations to explore phrase and sentence structures could be an interesting frontier to explore. The current study focused on NMT models based on LSTMs. Analyzing other architectures such as Transformers (Vaswani et al. 2017) , which recently set a new state of the art compared to both recurrent and convolutional models (Gehring et al. 2017), would be an exciting direction to pursue.", |
| "cite_spans": [ |
| { |
| "start": 334, |
| "end": 368, |
| "text": "Transformers (Vaswani et al. 2017)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "The character-based models reported in this article were trained using bidirectional LSTM models only. We simply segmented words into characters and marked word boundaries. However, we did try charCNN (Kim et al. 2015 ; Costa-juss\u00e0 and Fonollosa 2016) models in our preliminary experiments. The model is a CNN with a highway network over characters and trains an LSTM on top of it. In our results, we found the charCNN variant to perform poorly (see Table A .1), compared to the simple char-based LSTM model, both in translation quality and comparing classifier accuracy. We therefore left it out and focused on char-based LSTM models.", |
| "cite_spans": [ |
| { |
| "start": 201, |
| "end": 217, |
| "text": "(Kim et al. 2015", |
| "ref_id": "BIBREF55" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 450, |
| "end": 457, |
| "text": "Table A", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "A.1 Character-Based Models", |
| "sec_num": null |
| }, |
| { |
| "text": "Our results showed that the representations that are learned when translating into English are better for predicting POS or morphology than those learned when translating into German, which are in turn better than those learned when translating into Hebrew. The inherent difficulty in translating Arabic to (morphologically rich) Hebrew/German languages may affect the ability to learn good representations of word structure, or perhaps more data are required in the case of these languages to learn Arabic represen- tations of the same quality. We found these results to be consistent in other language pairs, that is, by changing the source from Arabic to German and Czech and when training character-based models instead of word-based models. See Tables A.2 and A.3 for these results.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "A.2 Effect of Target Language", |
| "sec_num": null |
| }, |
| { |
| "text": "In order to probe whether character-based models require additional depth in the network to capture the same amount of information, we carried out further experiments training 3-layered character models for Czech-to-English and English-to-German. We extracted feature representations and trained classifiers to predict syntactic dependency labels. Our results show that using an additional layer does improve the prediction accuracy, giving the same result as subword segmentation (Morfessor) in the case of Czech-to-English, but still worse in the case of English-to-Czech (see Table A .4).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 579, |
| "end": 586, |
| "text": "Table A", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "A.3 Three Layered Character-Based Models", |
| "sec_num": null |
| }, |
| { |
| "text": "James Glass. 2019a. What is one grain of", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "A.3 Three Layered Character-Based Models", |
| "sec_num": null |
| }, |
| { |
| "text": "The learned parameters are implicitly shared by all the language pairs being modeled.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "Character-based systems have been used previously in phrase-based MT for handling morphologically rich(Luong, Nakov, and Kan 2010) and closely related language pairs(Durrani et al. 2010;Nakov and Tiedemann 2012) or for transliterating unknown words(Durrani et al. 2014).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "Refer to Steedman and Baldridge (2011) andClark and Curran (2004) for more information on CCG supertagging.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "For instance, proximal and distal demonstratives (e.g., \"this\" and \"that\") are typically assigned the same POS tag (DT) but receive different SEM tags (PRX and DST, respectively), and proper nouns are disambiguated into several classes such as geo-political entity, location, organization, person, and artifact.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "It is also not unrealistic, as dependency parsers often work in two stages, first predicting an unlabeled dependency tree, and then labeling its edges(McDonald and Nivre 2011;McDonald, Lerman, and Pereira 2006). More complicated formulations can be conceived, from predicting the existence of dependencies independently to solving the full parsing task, but dependency labeling is a simple basic task to begin with. 6 Although we studied representations from a charCNN(Kim et al. 2015) inBelinkov et al. (2017a), the extracted features were still based on word representations produced by the charCNN. As a result, in that work we could not analyze and compare subword and character-based models that do not assume a segmentation into words. 7 One could envision more sophisticated averages, such as weighting via an attention mechanism.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "We also explored charCNN(Kim et al. 2015; Costa-juss\u00e0 and Fonollosa 2016) models in our preliminary experiments, and found the charCNN variant to perform poorly, compared with the simple char-based LSTM model both in translation quality and when comparing classifier accuracy. Therefore, we decided to leave them out for brevity. See Appendix for the results.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "The sentence length was varied across different configurations, to keep the training data sizes the same for all systems. 10 We need source sentences to generate encoder states, which in turn are required for obtaining the decoder states that we want to analyze.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "In their study of NMT and language model representations,Zhang and Bowman (2018) noticed that POS is better represented at layer 1 whereas CCG supertags are sometimes, but not always, better represented at layer 2 (out of 2-layer encoders). 17 See Section 11 in the supplementary information for the detailed results.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "The results shown are with English dependencies using NMT models trained on English to other languages, but the trends are similar for other language pairs.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "These results are obtained using models trained on the United Nations multi-parallel corpus.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "One speculation for this might be that translating into several target languages does not add much semantic information on the source side because this kind of information is more language-agnostic, but at this point there is insufficient evidence for this kind of claim.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "See Peters, Ruder, and Smith (2019) for an evaluation of when it is worthwhile to fine-tune.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "See Belinkov and Glass (2019) and references therein for considerations in the context of NLP.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "This work was funded by the QCRI, HBKU, as part of the collaboration with the MIT, CSAIL. Yonatan Belinkov was also partly supported by the Harvard Mind, Brain, and Behavior Initiative.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| }, |
| { |
| "text": " Table A .5 CCG tagging accuracy using features from the k-th encoding layer of 4-layered English-to-* NMT models trained with different target languages (German, Czech, French, and Spanish ", |
| "cite_spans": [ |
| { |
| "start": 154, |
| "end": 189, |
| "text": "(German, Czech, French, and Spanish", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 1, |
| "end": 8, |
| "text": "Table A", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "annex", |
| "sec_num": null |
| }, |
| { |
| "text": "English-Spanish English-French k 0 1 2 3 4 0 1 2 3 4 0 1 2 3 4", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "English-Arabic", |
| "sec_num": null |
| }, |
| { |
| "text": "Along with the syntactic dependency labeling task, we found higher layers to give better classifier accuracy also in the CCG tagging task. See Table A .5 for the results. Table A .6 shows statistical significance results for syntactic dependency labeling experiments from Section 7.2.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 143, |
| "end": 150, |
| "text": "Table A", |
| "ref_id": null |
| }, |
| { |
| "start": 171, |
| "end": 178, |
| "text": "Table A", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "A.4 Layer-Wise Experiments Using CCG Tags", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Farasa: A fast and furious segmenter for Arabic", |
| "authors": [ |
| { |
| "first": "Kareem", |
| "middle": [], |
| "last": "Abdelali", |
| "suffix": "" |
| }, |
| { |
| "first": "Nadir", |
| "middle": [], |
| "last": "Darwish", |
| "suffix": "" |
| }, |
| { |
| "first": "Hamdy", |
| "middle": [], |
| "last": "Durrani", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Mubarak", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Demonstrations", |
| "volume": "16", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Abdelali, lAhmed, Kareem Darwish, Nadir Durrani, and Hamdy Mubarak. 2016. Farasa: A fast and furious segmenter for Arabic. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Demonstrations, pages 11\u221216, San Diego, CA.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "The Parallel Meaning Bank: Towards a multilingual corpus of translations annotated with compositional meaning representations", |
| "authors": [ |
| { |
| "first": "Lasha", |
| "middle": [], |
| "last": "Abzianidze", |
| "suffix": "" |
| }, |
| { |
| "first": "Johannes", |
| "middle": [], |
| "last": "Bjerva", |
| "suffix": "" |
| }, |
| { |
| "first": "Kilian", |
| "middle": [], |
| "last": "Evang", |
| "suffix": "" |
| }, |
| { |
| "first": "Hessel", |
| "middle": [], |
| "last": "Haagsma", |
| "suffix": "" |
| }, |
| { |
| "first": "Rik", |
| "middle": [], |
| "last": "Van Noord", |
| "suffix": "" |
| }, |
| { |
| "first": "Pierre", |
| "middle": [], |
| "last": "Ludmann", |
| "suffix": "" |
| }, |
| { |
| "first": "Duc-Duy", |
| "middle": [], |
| "last": "Nguyen", |
| "suffix": "" |
| }, |
| { |
| "first": "Johan", |
| "middle": [], |
| "last": "Bos", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 15th Conference of the European Chapter", |
| "volume": "2", |
| "issue": "", |
| "pages": "242--247", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Abzianidze, Lasha, Johannes Bjerva, Kilian Evang, Hessel Haagsma, Rik van Noord, Pierre Ludmann, Duc-Duy Nguyen, and Johan Bos. 2017. The Parallel Meaning Bank: Towards a multilingual corpus of translations annotated with compositional meaning representations. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 2, Short Papers, pages 242-247, Valencia.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Fine-grained analysis of sentence embeddings using auxiliary prediction tasks", |
| "authors": [ |
| { |
| "first": "Yossi", |
| "middle": [], |
| "last": "Adi", |
| "suffix": "" |
| }, |
| { |
| "first": "Einat", |
| "middle": [], |
| "last": "Kermany", |
| "suffix": "" |
| }, |
| { |
| "first": "Yonatan", |
| "middle": [], |
| "last": "Belinkov", |
| "suffix": "" |
| }, |
| { |
| "first": "Ofer", |
| "middle": [], |
| "last": "Lavi", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoav", |
| "middle": [], |
| "last": "Goldberg", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "International Conference on Learning Representations", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Adi, Yossi, Einat Kermany, Yonatan Belinkov, Ofer Lavi, and Yoav Goldberg. 2017. Fine-grained analysis of sentence embeddings using auxiliary prediction tasks. In International Conference on Learning Representations (ICLR).", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Towards string-to-tree neural machine translation", |
| "authors": [ |
| { |
| "first": "Roee", |
| "middle": [], |
| "last": "Aharoni", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoav", |
| "middle": [], |
| "last": "Goldberg", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "2", |
| "issue": "", |
| "pages": "132--140", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Aharoni, Roee and Yoav Goldberg. 2017. Towards string-to-tree neural machine translation. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 132-140.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "The missing ingredient in zero-shot neural machine translation", |
| "authors": [ |
| { |
| "first": "Naveen", |
| "middle": [], |
| "last": "Arivazhagan", |
| "suffix": "" |
| }, |
| { |
| "first": "Ankur", |
| "middle": [], |
| "last": "Bapna", |
| "suffix": "" |
| }, |
| { |
| "first": "Orhan", |
| "middle": [], |
| "last": "Firat", |
| "suffix": "" |
| }, |
| { |
| "first": "Roee", |
| "middle": [], |
| "last": "Aharoni", |
| "suffix": "" |
| }, |
| { |
| "first": "Melvin", |
| "middle": [], |
| "last": "Johnson", |
| "suffix": "" |
| }, |
| { |
| "first": "Wolfgang", |
| "middle": [], |
| "last": "Macherey", |
| "suffix": "" |
| }, |
| { |
| "first": ";", |
| "middle": [], |
| "last": "Bahdanau", |
| "suffix": "" |
| }, |
| { |
| "first": "Kyunghyun", |
| "middle": [], |
| "last": "Dzmitry", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Cho", |
| "suffix": "" |
| }, |
| { |
| "first": ";", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| }, |
| { |
| "first": "Aravind", |
| "middle": [ |
| "K" |
| ], |
| "last": "Srinivas", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Joshi", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "Neural machine translation by jointly learning to align and translate", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1903.07091" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Arivazhagan, Naveen, Ankur Bapna, Orhan Firat, Roee Aharoni, Melvin Johnson, and Wolfgang Macherey. 2019. The missing ingredient in zero-shot neural machine translation. arXiv preprint arXiv:1903.07091. Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio. 2014. Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473. Bangalore, Srinivas and Aravind K. Joshi. 1999. Supertagging: An approach to almost parsing. Computational Linguistics, 25(2).", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Nadir Durrani, Fahim Dalvi, and James Glass. 2019a. Identifying and controlling important neurons in neural machine translation. International Conference on Learning Representations (ICLR", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Bau", |
| "suffix": "" |
| }, |
| { |
| "first": "Yonatan", |
| "middle": [], |
| "last": "Anthony", |
| "suffix": "" |
| }, |
| { |
| "first": "Hassan", |
| "middle": [], |
| "last": "Belinkov", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Sajjad", |
| "suffix": "" |
| } |
| ], |
| "year": null, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bau, D. Anthony, Yonatan Belinkov, Hassan Sajjad, Nadir Durrani, Fahim Dalvi, and James Glass. 2019a. Identifying and controlling important neurons in neural machine translation. International Conference on Learning Representations (ICLR).", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Identifying and controlling important neurons in neural machine translation", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Bau", |
| "suffix": "" |
| }, |
| { |
| "first": "Yonatan", |
| "middle": [], |
| "last": "Anthony", |
| "suffix": "" |
| }, |
| { |
| "first": "Hassan", |
| "middle": [], |
| "last": "Belinkov", |
| "suffix": "" |
| }, |
| { |
| "first": "Nadir", |
| "middle": [], |
| "last": "Sajjad", |
| "suffix": "" |
| }, |
| { |
| "first": "Fahim", |
| "middle": [], |
| "last": "Durrani", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Dalvi", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Glass", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "International Conference on Learning Representations (ICLR", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bau, D. Anthony, Yonatan Belinkov, Hassan Sajjad, Nadir Durrani, Fahim Dalvi, and James Glass. 2019b. Identifying and controlling important neurons in neural machine translation. In International Conference on Learning Representations (ICLR).", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Evaluating discourse phenomena in neural machine translation", |
| "authors": [ |
| { |
| "first": "Rachel", |
| "middle": [], |
| "last": "Bawden", |
| "suffix": "" |
| }, |
| { |
| "first": "Rico", |
| "middle": [], |
| "last": "Sennrich", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexandra", |
| "middle": [], |
| "last": "Birch", |
| "suffix": "" |
| }, |
| { |
| "first": "Barry", |
| "middle": [], |
| "last": "Haddow", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "1", |
| "issue": "", |
| "pages": "1304--1313", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bawden, Rachel, Rico Sennrich, Alexandra Birch, and Barry Haddow. 2018. Evaluating discourse phenomena in neural machine translation. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 1304-1313.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "On Internal Language Representations in Deep Learning: An Analysis of Machine Translation and Speech Recognition", |
| "authors": [ |
| { |
| "first": "Yonatan", |
| "middle": [], |
| "last": "Belinkov", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Belinkov, Yonatan. 2018. On Internal Language Representations in Deep Learning: An Analysis of Machine Translation and Speech Recognition. Ph.D. thesis, Massachusetts Institute of Technology.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "What do neural machine translation models learn about morphology?", |
| "authors": [ |
| { |
| "first": "Yonatan", |
| "middle": [], |
| "last": "Belinkov", |
| "suffix": "" |
| }, |
| { |
| "first": "Nadir", |
| "middle": [], |
| "last": "Durrani", |
| "suffix": "" |
| }, |
| { |
| "first": "Fahim", |
| "middle": [], |
| "last": "Dalvi", |
| "suffix": "" |
| }, |
| { |
| "first": "Hassan", |
| "middle": [], |
| "last": "Sajjad", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Glass", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (ACL)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Belinkov, Yonatan, Nadir Durrani, Fahim Dalvi, Hassan Sajjad, and James Glass. 2017a. What do neural machine translation models learn about morphology? In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (ACL), Vancouver.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Analysis methods in neural language processing: A survey", |
| "authors": [ |
| { |
| "first": "Yonatan", |
| "middle": [], |
| "last": "Belinkov", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Glass", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Transactions of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Belinkov, Yonatan and James Glass. 2019. Analysis methods in neural language processing: A survey. Transactions of the Association for Computational Linguistics (TACL).", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Evaluating layers of representation in neural machine translation on part-of-speech and semantic tagging tasks", |
| "authors": [ |
| { |
| "first": "Yonatan", |
| "middle": [], |
| "last": "Belinkov", |
| "suffix": "" |
| }, |
| { |
| "first": "Llu\u00eds", |
| "middle": [], |
| "last": "M\u00e0rquez", |
| "suffix": "" |
| }, |
| { |
| "first": "Hassan", |
| "middle": [], |
| "last": "Sajjad", |
| "suffix": "" |
| }, |
| { |
| "first": "Nadir", |
| "middle": [], |
| "last": "Durrani", |
| "suffix": "" |
| }, |
| { |
| "first": "Fahim", |
| "middle": [], |
| "last": "Dalvi", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Glass", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 8th International Joint Conference on Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Belinkov, Yonatan, Llu\u00eds M\u00e0rquez, Hassan Sajjad, Nadir Durrani, Fahim Dalvi, and James Glass. 2017b. Evaluating layers of representation in neural machine translation on part-of-speech and semantic tagging tasks. In Proceedings of the 8th International Joint Conference on Natural Language Processing (IJCNLP).", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Neural versus phrase-based machine translation quality: A case study", |
| "authors": [ |
| { |
| "first": "Luisa", |
| "middle": [], |
| "last": "Bentivogli", |
| "suffix": "" |
| }, |
| { |
| "first": "Arianna", |
| "middle": [], |
| "last": "Bisazza", |
| "suffix": "" |
| }, |
| { |
| "first": "Mauro", |
| "middle": [], |
| "last": "Cettolo", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcello", |
| "middle": [], |
| "last": "Federico", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "257--267", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bentivogli, Luisa, Arianna Bisazza, Mauro Cettolo, and Marcello Federico. 2016. Neural versus phrase-based machine translation quality: A case study. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 257-267, Austin, TX.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "The lazy encoder: A fine-grained analysis of the role of morphology in neural machine translation", |
| "authors": [ |
| { |
| "first": "Arianna", |
| "middle": [], |
| "last": "Bisazza", |
| "suffix": "" |
| }, |
| { |
| "first": "Clara", |
| "middle": [], |
| "last": "Tump", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "237--265", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bisazza, Arianna and Clara Tump. 2018. The lazy encoder: A fine-grained analysis of the role of morphology in neural machine translation. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 2871-2876, Association for Computational Linguistics, pages 237-265. Brussels.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Semantic tagging with deep residual networks", |
| "authors": [ |
| { |
| "first": "Johannes", |
| "middle": [], |
| "last": "Bjerva", |
| "suffix": "" |
| }, |
| { |
| "first": "Barbara", |
| "middle": [], |
| "last": "Plank", |
| "suffix": "" |
| }, |
| { |
| "first": "Johan", |
| "middle": [], |
| "last": "Bos", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers", |
| "volume": "", |
| "issue": "", |
| "pages": "3531--3541", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bjerva, Johannes, Barbara Plank, and Johan Bos. 2016. Semantic tagging with deep residual networks. In Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, pages 3531-3541, Osaka.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Findings of the 2017 Conference on machine translation (WMT17)", |
| "authors": [ |
| { |
| "first": "Ond\u0159ej", |
| "middle": [], |
| "last": "Bojar", |
| "suffix": "" |
| }, |
| { |
| "first": "Rajen", |
| "middle": [], |
| "last": "Chatterjee", |
| "suffix": "" |
| }, |
| { |
| "first": "Christian", |
| "middle": [], |
| "last": "Federmann", |
| "suffix": "" |
| }, |
| { |
| "first": "Yvette", |
| "middle": [], |
| "last": "Graham", |
| "suffix": "" |
| }, |
| { |
| "first": "Barry", |
| "middle": [], |
| "last": "Haddow", |
| "suffix": "" |
| }, |
| { |
| "first": "Shujian", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "" |
| }, |
| { |
| "first": "Matthias", |
| "middle": [], |
| "last": "Huck", |
| "suffix": "" |
| }, |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| }, |
| { |
| "first": "Qun", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Varvara", |
| "middle": [], |
| "last": "Logacheva", |
| "suffix": "" |
| }, |
| { |
| "first": "Christof", |
| "middle": [], |
| "last": "Monz", |
| "suffix": "" |
| }, |
| { |
| "first": "Matteo", |
| "middle": [], |
| "last": "Negri", |
| "suffix": "" |
| }, |
| { |
| "first": "Matt", |
| "middle": [], |
| "last": "Post", |
| "suffix": "" |
| }, |
| { |
| "first": "Raphael", |
| "middle": [], |
| "last": "Rubino", |
| "suffix": "" |
| }, |
| { |
| "first": "Lucia", |
| "middle": [], |
| "last": "Specia", |
| "suffix": "" |
| }, |
| { |
| "first": "Marco", |
| "middle": [], |
| "last": "Turchi", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the Second Conference on Machine Translation", |
| "volume": "2", |
| "issue": "", |
| "pages": "169--214", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bojar, Ond\u0159ej, Rajen Chatterjee, Christian Federmann, Yvette Graham, Barry Haddow, Shujian Huang, Matthias Huck, Philipp Koehn, Qun Liu, Varvara Logacheva, Christof Monz, Matteo Negri, Matt Post, Raphael Rubino, Lucia Specia, and Marco Turchi. 2017. Findings of the 2017 Conference on machine translation (WMT17). In Proceedings of the Second Conference on Machine Translation, Volume 2: Shared Task Papers, pages 169-214, Copenhagen.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Findings of the 2016 conference on machine translation", |
| "authors": [ |
| { |
| "first": "Ond\u0159ej", |
| "middle": [], |
| "last": "Bojar", |
| "suffix": "" |
| }, |
| { |
| "first": "Rajen", |
| "middle": [], |
| "last": "Chatterjee", |
| "suffix": "" |
| }, |
| { |
| "first": "Christian", |
| "middle": [], |
| "last": "Federmann", |
| "suffix": "" |
| }, |
| { |
| "first": "Yvette", |
| "middle": [], |
| "last": "Graham", |
| "suffix": "" |
| }, |
| { |
| "first": "Barry", |
| "middle": [], |
| "last": "Haddow", |
| "suffix": "" |
| }, |
| { |
| "first": "Matthias", |
| "middle": [], |
| "last": "Huck", |
| "suffix": "" |
| }, |
| { |
| "first": "Antonio", |
| "middle": [ |
| "Jimeno" |
| ], |
| "last": "Yepes", |
| "suffix": "" |
| }, |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| }, |
| { |
| "first": "Varvara", |
| "middle": [], |
| "last": "Logacheva", |
| "suffix": "" |
| }, |
| { |
| "first": "Christof", |
| "middle": [], |
| "last": "Monz", |
| "suffix": "" |
| }, |
| { |
| "first": "Matteo", |
| "middle": [], |
| "last": "Negri", |
| "suffix": "" |
| }, |
| { |
| "first": "Aurelie", |
| "middle": [], |
| "last": "Neveol", |
| "suffix": "" |
| }, |
| { |
| "first": "Mariana", |
| "middle": [], |
| "last": "Neves", |
| "suffix": "" |
| }, |
| { |
| "first": "Martin", |
| "middle": [], |
| "last": "Popel", |
| "suffix": "" |
| }, |
| { |
| "first": "Matt", |
| "middle": [], |
| "last": "Post", |
| "suffix": "" |
| }, |
| { |
| "first": "Raphael", |
| "middle": [], |
| "last": "Rubino", |
| "suffix": "" |
| }, |
| { |
| "first": "Carolina", |
| "middle": [], |
| "last": "Scarton", |
| "suffix": "" |
| }, |
| { |
| "first": "Lucia", |
| "middle": [], |
| "last": "Specia", |
| "suffix": "" |
| }, |
| { |
| "first": "Marco", |
| "middle": [], |
| "last": "Turchi", |
| "suffix": "" |
| }, |
| { |
| "first": "Karin", |
| "middle": [], |
| "last": "Verspoor", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcos", |
| "middle": [], |
| "last": "Zampieri", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the First Conference on Machine Translation", |
| "volume": "", |
| "issue": "", |
| "pages": "131--198", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bojar, Ond\u0159ej, Rajen Chatterjee, Christian Federmann, Yvette Graham, Barry Haddow, Matthias Huck, Antonio Jimeno Yepes, Philipp Koehn, Varvara Logacheva, Christof Monz, Matteo Negri, Aurelie Neveol, Mariana Neves, Martin Popel, Matt Post, Raphael Rubino, Carolina Scarton, Lucia Specia, Marco Turchi, Karin Verspoor, and Marcos Zampieri. 2016. Findings of the 2016 conference on machine translation. In Proceedings of the First Conference on Machine Translation, pages 131-198, Berlin.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Looking for ELMo's friends: Sentence-level pretraining beyond language modeling", |
| "authors": [ |
| { |
| "first": "Samuel", |
| "middle": [ |
| "R" |
| ], |
| "last": "Bowman", |
| "suffix": "" |
| }, |
| { |
| "first": "Ellie", |
| "middle": [], |
| "last": "Pavlick", |
| "suffix": "" |
| }, |
| { |
| "first": "Edouard", |
| "middle": [], |
| "last": "Grave", |
| "suffix": "" |
| }, |
| { |
| "first": "Benjamin", |
| "middle": [], |
| "last": "Van Durme", |
| "suffix": "" |
| }, |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Jan", |
| "middle": [], |
| "last": "Hula", |
| "suffix": "" |
| }, |
| { |
| "first": "Patrick", |
| "middle": [], |
| "last": "Xia", |
| "suffix": "" |
| }, |
| { |
| "first": "Raghavendra", |
| "middle": [], |
| "last": "Pappagari", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [ |
| "Thomas" |
| ], |
| "last": "Mccoy", |
| "suffix": "" |
| }, |
| { |
| "first": "Roma", |
| "middle": [], |
| "last": "Patel", |
| "suffix": "" |
| }, |
| { |
| "first": "Najoung", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| }, |
| { |
| "first": "Ian", |
| "middle": [], |
| "last": "Tenney", |
| "suffix": "" |
| }, |
| { |
| "first": "Yinghui", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "" |
| }, |
| { |
| "first": "Katherin", |
| "middle": [], |
| "last": "Yu", |
| "suffix": "" |
| }, |
| { |
| "first": "Shuning", |
| "middle": [], |
| "last": "Jin", |
| "suffix": "" |
| }, |
| { |
| "first": "Berlin", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bowman, Samuel R., Ellie Pavlick, Edouard Grave, Benjamin Van Durme, Alex Wang, Jan Hula, Patrick Xia, Raghavendra Pappagari, R. Thomas McCoy, Roma Patel, Najoung Kim, Ian Tenney, Yinghui Huang, Katherin Yu, Shuning Jin, and Berlin Chen. 2018. Looking for ELMo's friends: Sentence-level pretraining beyond language modeling. ArXiv:1812.10860.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Metamind neural machine translation system for WMT", |
| "authors": [ |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Bradbury", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the First Conference on Machine Translation", |
| "volume": "", |
| "issue": "", |
| "pages": "264--267", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bradbury, James and Richard Socher. 2016. Metamind neural machine translation system for WMT 2016. In Proceedings of the First Conference on Machine Translation, pages 264-267, Berlin.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Evaluating the morphological competence of machine translation systems", |
| "authors": [ |
| { |
| "first": "Franck", |
| "middle": [], |
| "last": "Burlot", |
| "suffix": "" |
| }, |
| { |
| "first": "Fran\u00e7ois", |
| "middle": [], |
| "last": "Yvon", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the Second Conference on Machine Translation", |
| "volume": "", |
| "issue": "", |
| "pages": "43--55", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Burlot, Franck and Fran\u00e7ois Yvon. 2017. Evaluating the morphological competence of machine translation systems. In Proceedings of the Second Conference on Machine Translation, pages 43-55.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "The IWSLT 2016 evaluation campaign", |
| "authors": [ |
| { |
| "first": "Mauro", |
| "middle": [], |
| "last": "Cettolo", |
| "suffix": "" |
| }, |
| { |
| "first": "Jan", |
| "middle": [], |
| "last": "Niehues", |
| "suffix": "" |
| }, |
| { |
| "first": "Sebastian", |
| "middle": [], |
| "last": "St\u00fcker", |
| "suffix": "" |
| }, |
| { |
| "first": "Bentivogli", |
| "middle": [], |
| "last": "Luisa", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcello", |
| "middle": [], |
| "last": "Federico", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 13th International Workshop on Spoken Language Translation (IWSLT 2016)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Cettolo, Mauro, Jan Niehues, Sebastian St\u00fcker, Bentivogli Luisa, and Marcello Federico. 2016. The IWSLT 2016 evaluation campaign. In Proceedings of the 13th International Workshop on Spoken Language Translation (IWSLT 2016), Seattle, WA.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Neural machine translation with source dependency representation", |
| "authors": [ |
| { |
| "first": "Kehai", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Rui", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Masao", |
| "middle": [], |
| "last": "Utiyama", |
| "suffix": "" |
| }, |
| { |
| "first": "Lemao", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Akihiro", |
| "middle": [], |
| "last": "Tamura", |
| "suffix": "" |
| }, |
| { |
| "first": "Eiichiro", |
| "middle": [], |
| "last": "Sumita", |
| "suffix": "" |
| }, |
| { |
| "first": "Tiejun", |
| "middle": [], |
| "last": "Zhao", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "2846--2852", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chen, Kehai, Rui Wang, Masao Utiyama, Lemao Liu, Akihiro Tamura, Eiichiro Sumita, and Tiejun Zhao. 2017. Neural machine translation with source dependency representation. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pages 2846-2852.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "The best of both worlds: Combining recent advances in neural machine translation", |
| "authors": [ |
| { |
| "first": "Mia", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Orhan", |
| "middle": [], |
| "last": "Xu", |
| "suffix": "" |
| }, |
| { |
| "first": "Ankur", |
| "middle": [], |
| "last": "Firat", |
| "suffix": "" |
| }, |
| { |
| "first": "Melvin", |
| "middle": [], |
| "last": "Bapna", |
| "suffix": "" |
| }, |
| { |
| "first": "Wolfgang", |
| "middle": [], |
| "last": "Johnson", |
| "suffix": "" |
| }, |
| { |
| "first": "George", |
| "middle": [], |
| "last": "Macherey", |
| "suffix": "" |
| }, |
| { |
| "first": "Llion", |
| "middle": [], |
| "last": "Foster", |
| "suffix": "" |
| }, |
| { |
| "first": "Mike", |
| "middle": [], |
| "last": "Jones", |
| "suffix": "" |
| }, |
| { |
| "first": "Noam", |
| "middle": [], |
| "last": "Schuster", |
| "suffix": "" |
| }, |
| { |
| "first": "Niki", |
| "middle": [], |
| "last": "Shazeer", |
| "suffix": "" |
| }, |
| { |
| "first": "Ashish", |
| "middle": [], |
| "last": "Parmar", |
| "suffix": "" |
| }, |
| { |
| "first": "Jakob", |
| "middle": [], |
| "last": "Vaswani", |
| "suffix": "" |
| }, |
| { |
| "first": "Lukasz", |
| "middle": [], |
| "last": "Uszkoreit", |
| "suffix": "" |
| }, |
| { |
| "first": "Zhifeng", |
| "middle": [], |
| "last": "Kaiser", |
| "suffix": "" |
| }, |
| { |
| "first": "Yonghui", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Macduff", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Hughes", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "76--86", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chen, Mia Xu, Orhan Firat, Ankur Bapna, Melvin Johnson, Wolfgang Macherey, George Foster, Llion Jones, Mike Schuster, Noam Shazeer, Niki Parmar, Ashish Vaswani, Jakob Uszkoreit, Lukasz Kaiser, Zhifeng Chen, Yonghui Wu, and Macduff Hughes. 2018. The best of both worlds: Combining recent advances in neural machine translation. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 76-86.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "A hierarchical Phrase-based model for statistical machine translation", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Chiang", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proceedings of the 43rd Annual Meeting on Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "263--270", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chiang, David. 2005. A hierarchical Phrase-based model for statistical machine translation. In Proceedings of the 43rd Annual Meeting on Association for Computational Linguistics, pages 263-270.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "11,001 New features for statistical machine translation", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Chiang", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Knight", |
| "suffix": "" |
| }, |
| { |
| "first": "Wei", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Proceedings of Human Language Technologies: The 2009 Annual Conference of the North American Chapter of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "218--226", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chiang, David, Kevin Knight, and Wei Wang. 2009. 11,001 New features for statistical machine translation. In Proceedings of Human Language Technologies: The 2009 Annual Conference of the North American Chapter of the Association for Computational Linguistics, pages 218-226.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "A character-level decoder without explicit segmentation for neural machine translation", |
| "authors": [ |
| { |
| "first": "Junyoung", |
| "middle": [], |
| "last": "Chung", |
| "suffix": "" |
| }, |
| { |
| "first": "Kyunghyun", |
| "middle": [], |
| "last": "Cho", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "1693--1703", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chung, Junyoung, Kyunghyun Cho, and Yoshua Bengio. 2016. A character-level decoder without explicit segmentation for neural machine translation. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1693-1703, Berlin.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "Annotation of English on the tectogrammatical level: Reference book", |
| "authors": [ |
| { |
| "first": "Silvie", |
| "middle": [], |
| "last": "Cinkov\u00e1", |
| "suffix": "" |
| }, |
| { |
| "first": "Jan", |
| "middle": [], |
| "last": "Haji\u010d", |
| "suffix": "" |
| }, |
| { |
| "first": "Marie", |
| "middle": [], |
| "last": "Mikulov\u00e1", |
| "suffix": "" |
| }, |
| { |
| "first": "Lucie", |
| "middle": [], |
| "last": "Mladov\u00e1", |
| "suffix": "" |
| }, |
| { |
| "first": "Anja", |
| "middle": [], |
| "last": "Nedolu\u017eko", |
| "suffix": "" |
| }, |
| { |
| "first": "Petr", |
| "middle": [], |
| "last": "Pajas", |
| "suffix": "" |
| }, |
| { |
| "first": "Jarmila", |
| "middle": [], |
| "last": "Panevov\u00e1", |
| "suffix": "" |
| }, |
| { |
| "first": "Ji\u0159\u00ed", |
| "middle": [], |
| "last": "Semeck\u1ef3", |
| "suffix": "" |
| }, |
| { |
| "first": "Jan\u01ce", |
| "middle": [], |
| "last": "Sindlerov\u00e1", |
| "suffix": "" |
| }, |
| { |
| "first": "Josef", |
| "middle": [], |
| "last": "Toman", |
| "suffix": "" |
| }, |
| { |
| "first": "Zde\u0148ka", |
| "middle": [], |
| "last": "Ure\u0161ov\u00e1", |
| "suffix": "" |
| }, |
| { |
| "first": "Zden\u011bk\u017eabokrtsk\u00fd", |
| "middle": [], |
| "last": "", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Cinkov\u00e1, Silvie, Jan Haji\u010d, Marie Mikulov\u00e1, Lucie Mladov\u00e1, Anja Nedolu\u017eko, Petr Pajas, Jarmila Panevov\u00e1, Ji\u0159\u00ed Semeck\u1ef3, Jan\u01ce Sindlerov\u00e1, Josef Toman, Zde\u0148ka Ure\u0161ov\u00e1, and Zden\u011bk\u017dabokrtsk\u00fd. 2004. Annotation of English on the tectogrammatical level: Reference book. Technical report, UFAL/CKL, Prague, Czech Republic.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "Tectogrammatical annotation of the", |
| "authors": [ |
| { |
| "first": "Silvie", |
| "middle": [], |
| "last": "Cinkov\u00e1", |
| "suffix": "" |
| }, |
| { |
| "first": "Josef", |
| "middle": [], |
| "last": "Toman", |
| "suffix": "" |
| }, |
| { |
| "first": "Jan", |
| "middle": [], |
| "last": "Haji\u010d", |
| "suffix": "" |
| }, |
| { |
| "first": "V\u00e1clav", |
| "middle": [], |
| "last": "Krist\u00fdna\u010derm\u00e1kov\u00e1", |
| "suffix": "" |
| }, |
| { |
| "first": "Lucie", |
| "middle": [], |
| "last": "Klime\u0161", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Mladov\u00e1", |
| "suffix": "" |
| }, |
| { |
| "first": "Krist\u00fdna", |
| "middle": [], |
| "last": "Jana\u0161indlerov\u00e1", |
| "suffix": "" |
| }, |
| { |
| "first": "Zden\u011bk\u017eabokrtsk\u00fd", |
| "middle": [], |
| "last": "Tom\u0161\u016f", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Wall Street Journal. The Prague Bulletin of Mathematical Linguistics", |
| "volume": "", |
| "issue": "92", |
| "pages": "85--104", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Cinkov\u00e1, Silvie, Josef Toman, Jan Haji\u010d, Krist\u00fdna\u010cerm\u00e1kov\u00e1, V\u00e1clav Klime\u0161, Lucie Mladov\u00e1, Jana\u0160indlerov\u00e1, Krist\u00fdna Tom\u0161\u016f, and Zden\u011bk\u017dabokrtsk\u00fd. 2009. Tectogrammatical annotation of the Wall Street Journal. The Prague Bulletin of Mathematical Linguistics, (92):85-104.", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "The importance of supertagging for wide-coverage CCG parsing", |
| "authors": [ |
| { |
| "first": "Stephen", |
| "middle": [], |
| "last": "Clark", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [ |
| "R" |
| ], |
| "last": "Curran", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "COLING 2004: Proceedings of the 20th International Conference on Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Clark, Stephen and James R. Curran. 2004. The importance of supertagging for wide-coverage CCG parsing. In COLING 2004: Proceedings of the 20th International Conference on Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "What you can cram into a single vector: Probing sentence embeddings for linguistic properties", |
| "authors": [ |
| { |
| "first": "Alexis", |
| "middle": [], |
| "last": "Conneau", |
| "suffix": "" |
| }, |
| { |
| "first": "German", |
| "middle": [], |
| "last": "Kruszewski", |
| "suffix": "" |
| }, |
| { |
| "first": "Guillaume", |
| "middle": [], |
| "last": "Lample", |
| "suffix": "" |
| }, |
| { |
| "first": "Lo\u00efc", |
| "middle": [], |
| "last": "Barrault", |
| "suffix": "" |
| }, |
| { |
| "first": "Marco", |
| "middle": [], |
| "last": "Baroni", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (ACL)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Conneau, Alexis, German Kruszewski, Guillaume Lample, Lo\u00efc Barrault, and Marco Baroni. 2018. What you can cram into a single vector: Probing sentence embeddings for linguistic properties. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (ACL).", |
| "links": null |
| }, |
| "BIBREF30": { |
| "ref_id": "b30", |
| "title": "Character-based neural machine translation", |
| "authors": [ |
| { |
| "first": "Marta", |
| "middle": [ |
| "R" |
| ], |
| "last": "Costa-Juss\u00e0", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [ |
| "R" |
| ], |
| "last": "Jos\u00e9", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Fonollosa", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "2", |
| "issue": "", |
| "pages": "357--361", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Costa-juss\u00e0, Marta R. and Jos\u00e9 A. R. Fonollosa. 2016. Character-based neural machine translation. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 357-361, Berlin.", |
| "links": null |
| }, |
| "BIBREF31": { |
| "ref_id": "b31", |
| "title": "and sand in the desert? Analyzing individual neurons in deep NLP models", |
| "authors": [ |
| { |
| "first": "Fahim", |
| "middle": [], |
| "last": "Dalvi", |
| "suffix": "" |
| }, |
| { |
| "first": "Nadir", |
| "middle": [], |
| "last": "Durrani", |
| "suffix": "" |
| }, |
| { |
| "first": "Hassan", |
| "middle": [], |
| "last": "Sajjad", |
| "suffix": "" |
| }, |
| { |
| "first": "Yonatan", |
| "middle": [], |
| "last": "Belinkov", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [ |
| "Anthony" |
| ], |
| "last": "Bau", |
| "suffix": "" |
| } |
| ], |
| "year": null, |
| "venue": "Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dalvi, Fahim, Nadir Durrani, Hassan Sajjad, Yonatan Belinkov, D. Anthony Bau, and sand in the desert? Analyzing individual neurons in deep NLP models. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence (AAAI, Oral presentation).", |
| "links": null |
| }, |
| "BIBREF32": { |
| "ref_id": "b32", |
| "title": "Understanding and improving morphological learning in the neural machine translation decoder", |
| "authors": [ |
| { |
| "first": "Fahim", |
| "middle": [], |
| "last": "Dalvi", |
| "suffix": "" |
| }, |
| { |
| "first": "Nadir", |
| "middle": [], |
| "last": "Durrani", |
| "suffix": "" |
| }, |
| { |
| "first": "Hassan", |
| "middle": [], |
| "last": "Sajjad", |
| "suffix": "" |
| }, |
| { |
| "first": "Yonatan", |
| "middle": [], |
| "last": "Belinkov", |
| "suffix": "" |
| }, |
| { |
| "first": "Stephan", |
| "middle": [], |
| "last": "Vogel", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 8th International Joint Conference on Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dalvi, Fahim, Nadir Durrani, Hassan Sajjad, Yonatan Belinkov, and Stephan Vogel. 2017. Understanding and improving morphological learning in the neural machine translation decoder. In Proceedings of the 8th International Joint Conference on Natural Language Processing (IJCNLP).", |
| "links": null |
| }, |
| "BIBREF33": { |
| "ref_id": "b33", |
| "title": "NeuroX: A toolkit for analyzing individual neurons in neural networks", |
| "authors": [ |
| { |
| "first": "Fahim", |
| "middle": [], |
| "last": "Dalvi", |
| "suffix": "" |
| }, |
| { |
| "first": "Avery", |
| "middle": [], |
| "last": "Nortonsmith", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [ |
| "Anthony" |
| ], |
| "last": "Bau", |
| "suffix": "" |
| }, |
| { |
| "first": "Yonatan", |
| "middle": [], |
| "last": "Belinkov", |
| "suffix": "" |
| }, |
| { |
| "first": "Hassan", |
| "middle": [], |
| "last": "Sajjad", |
| "suffix": "" |
| }, |
| { |
| "first": "Nadir", |
| "middle": [], |
| "last": "Durrani", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Glass", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the AAAI Conference on Artificial Intelligence (AAAI)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dalvi, Fahim, Avery Nortonsmith, D. Anthony Bau, Yonatan Belinkov, Hassan Sajjad, Nadir Durrani, and James Glass. 2019b. NeuroX: A toolkit for analyzing individual neurons in neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), Honolulu, HI.", |
| "links": null |
| }, |
| "BIBREF34": { |
| "ref_id": "b34", |
| "title": "BERT: Pre-training of deep bidirectional transformers for language understanding", |
| "authors": [ |
| { |
| "first": "Jacob", |
| "middle": [], |
| "last": "Devlin", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming-Wei", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "Kenton", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Kristina", |
| "middle": [], |
| "last": "Toutanova", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 17th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Devlin, Jacob, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 17th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT).", |
| "links": null |
| }, |
| "BIBREF35": { |
| "ref_id": "b35", |
| "title": "Hindi-to-Urdu machine translation through transliteration", |
| "authors": [ |
| { |
| "first": "Nadir", |
| "middle": [], |
| "last": "Durrani", |
| "suffix": "" |
| }, |
| { |
| "first": "Hassan", |
| "middle": [], |
| "last": "Sajjad", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexander", |
| "middle": [], |
| "last": "Fraser", |
| "suffix": "" |
| }, |
| { |
| "first": "Helmut", |
| "middle": [], |
| "last": "Schmid", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "465--474", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Durrani, Nadir, Hassan Sajjad, Alexander Fraser, and Helmut Schmid. 2010. Hindi-to-Urdu machine translation through transliteration. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics, pages 465-474, Uppsala.", |
| "links": null |
| }, |
| "BIBREF36": { |
| "ref_id": "b36", |
| "title": "Integrating an unsupervised transliteration model into statistical machine translation", |
| "authors": [ |
| { |
| "first": "Nadir", |
| "middle": [], |
| "last": "Durrani", |
| "suffix": "" |
| }, |
| { |
| "first": "Hassan", |
| "middle": [], |
| "last": "Sajjad", |
| "suffix": "" |
| }, |
| { |
| "first": "Hieu", |
| "middle": [], |
| "last": "Hoang", |
| "suffix": "" |
| }, |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 15th Conference of the European Chapter of the ACL (EACL 2014)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Durrani, Nadir, Hassan Sajjad, Hieu Hoang, and Philipp Koehn. 2014. Integrating an unsupervised transliteration model into statistical machine translation. In Proceedings of the 15th Conference of the European Chapter of the ACL (EACL 2014), Gothenburg.", |
| "links": null |
| }, |
| "BIBREF37": { |
| "ref_id": "b37", |
| "title": "A joint sequence translation model with integrated reordering", |
| "authors": [ |
| { |
| "first": "Nadir", |
| "middle": [], |
| "last": "Durrani", |
| "suffix": "" |
| }, |
| { |
| "first": "Helmut", |
| "middle": [], |
| "last": "Schmid", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexander", |
| "middle": [], |
| "last": "Fraser", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the Association for Computational Linguistics: Human Language Technologies (ACL-HLT'11)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Durrani, Nadir, Helmut Schmid, and Alexander Fraser. 2011. A joint sequence translation model with integrated reordering. In Proceedings of the Association for Computational Linguistics: Human Language Technologies (ACL-HLT'11), Portland, OR.", |
| "links": null |
| }, |
| "BIBREF38": { |
| "ref_id": "b38", |
| "title": "Distributed representations, simple recurrent networks, and grammatical structure", |
| "authors": [ |
| { |
| "first": "Jeffrey", |
| "middle": [ |
| "L" |
| ], |
| "last": "Elman", |
| "suffix": "" |
| } |
| ], |
| "year": 1991, |
| "venue": "Machine Learning", |
| "volume": "7", |
| "issue": "", |
| "pages": "195--225", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Elman, Jeffrey L. 1991. Distributed representations, simple recurrent networks, and grammatical structure. Machine Learning, 7(2-3):195-225.", |
| "links": null |
| }, |
| "BIBREF39": { |
| "ref_id": "b39", |
| "title": "Tree-to-sequence attentional neural machine translation", |
| "authors": [ |
| { |
| "first": "Akiko", |
| "middle": [], |
| "last": "Eriguchi", |
| "suffix": "" |
| }, |
| { |
| "first": "Kazuma", |
| "middle": [], |
| "last": "Hashimoto", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshimasa", |
| "middle": [], |
| "last": "Tsuruoka", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "823--833", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Eriguchi, Akiko, Kazuma Hashimoto, and Yoshimasa Tsuruoka. 2016. Tree-to-sequence attentional neural machine translation. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 823-833.", |
| "links": null |
| }, |
| "BIBREF40": { |
| "ref_id": "b40", |
| "title": "Scalable inference and training of context-rich syntactic translation models", |
| "authors": [ |
| { |
| "first": "Michel", |
| "middle": [], |
| "last": "Galley", |
| "suffix": "" |
| }, |
| { |
| "first": "Jonathan", |
| "middle": [], |
| "last": "Graehl", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Knight", |
| "suffix": "" |
| }, |
| { |
| "first": "Daniel", |
| "middle": [], |
| "last": "Marcu", |
| "suffix": "" |
| }, |
| { |
| "first": "Steve", |
| "middle": [], |
| "last": "Deneefe", |
| "suffix": "" |
| }, |
| { |
| "first": "Wei", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Ignacio", |
| "middle": [], |
| "last": "Thayer", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "961--968", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Galley, Michel, Jonathan Graehl, Kevin Knight, Daniel Marcu, Steve DeNeefe, Wei Wang, and Ignacio Thayer. 2006. Scalable inference and training of context-rich syntactic translation models. In Proceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics, pages 961-968.", |
| "links": null |
| }, |
| "BIBREF41": { |
| "ref_id": "b41", |
| "title": "A simple and effective hierarchical phrase reordering model", |
| "authors": [ |
| { |
| "first": "Michel", |
| "middle": [], |
| "last": "Galley", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [ |
| "D" |
| ], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "848--856", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Galley, Michel and Christopher D. Manning. 2008. A simple and effective hierarchical phrase reordering model. In Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing, pages 848-856, Honolulu, HI.", |
| "links": null |
| }, |
| "BIBREF42": { |
| "ref_id": "b42", |
| "title": "Interpretation of semantic tweet representations", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Ganesh", |
| "suffix": "" |
| }, |
| { |
| "first": "Manish", |
| "middle": [], |
| "last": "Gupta", |
| "suffix": "" |
| }, |
| { |
| "first": "Vasudeva", |
| "middle": [], |
| "last": "Varma", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 2017 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining", |
| "volume": "", |
| "issue": "", |
| "pages": "95--102", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ganesh, J., Manish Gupta, and Vasudeva Varma. 2017. Interpretation of semantic tweet representations. In Proceedings of the 2017 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining 2017, pages 95-102, New York, NY. Gehring, Jonas, Michael Auli, David Grangier, Denis Yarats, and Yann N.", |
| "links": null |
| }, |
| "BIBREF43": { |
| "ref_id": "b43", |
| "title": "Convolutional sequence to sequence learning", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Dauphin", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 34th International Conference on Machine Learning", |
| "volume": "70", |
| "issue": "", |
| "pages": "1243--1252", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dauphin. 2017. Convolutional sequence to sequence learning. In Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pages 1243-1252, Sydney.", |
| "links": null |
| }, |
| "BIBREF44": { |
| "ref_id": "b44", |
| "title": "Non-autoregressive neural machine translation", |
| "authors": [ |
| { |
| "first": "Jiatao", |
| "middle": [], |
| "last": "Gu", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Bradbury", |
| "suffix": "" |
| }, |
| { |
| "first": "Caiming", |
| "middle": [], |
| "last": "Xiong", |
| "suffix": "" |
| }, |
| { |
| "first": "O", |
| "middle": [ |
| "K" |
| ], |
| "last": "Victor", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "International Conference on Learning Representations", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Gu, Jiatao, James Bradbury, Caiming Xiong, Victor O. K. Li, and Richard Socher. 2018. Non-autoregressive neural machine translation. In International Conference on Learning Representations.", |
| "links": null |
| }, |
| "BIBREF45": { |
| "ref_id": "b45", |
| "title": "KenLM: Faster and smaller language model queries", |
| "authors": [ |
| { |
| "first": "Kenneth", |
| "middle": [], |
| "last": "Heafield", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the Workshop on Statistical Machine Translation (WMT'11), Edinburgh. Hochreiter, Sepp and J\u00fcrgen Schmidhuber. 1997. Long short-term memory", |
| "volume": "9", |
| "issue": "", |
| "pages": "1735--1780", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Heafield, Kenneth. 2011. KenLM: Faster and smaller language model queries. In Proceedings of the Workshop on Statistical Machine Translation (WMT'11), Edinburgh. Hochreiter, Sepp and J\u00fcrgen Schmidhuber. 1997. Long short-term memory. Neural Computation, 9(8):1735-1780.", |
| "links": null |
| }, |
| "BIBREF46": { |
| "ref_id": "b46", |
| "title": "Creating a CCGbank and a wide-coverage CCG lexicon for German", |
| "authors": [ |
| { |
| "first": "Julia", |
| "middle": [], |
| "last": "Hockenmaier", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the 21st International Conference on Computational Linguistics and the 44th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "505--512", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hockenmaier, Julia. 2006. Creating a CCGbank and a wide-coverage CCG lexicon for German. In Proceedings of the 21st International Conference on Computational Linguistics and the 44th Annual Meeting of the Association for Computational Linguistics, pages 505-512.", |
| "links": null |
| }, |
| "BIBREF47": { |
| "ref_id": "b47", |
| "title": "CCGbank: A corpus of CCG derivations and dependency structures extracted from the Penn Treebank", |
| "authors": [ |
| { |
| "first": "Julia", |
| "middle": [], |
| "last": "Hockenmaier", |
| "suffix": "" |
| }, |
| { |
| "first": "Mark", |
| "middle": [], |
| "last": "Steedman", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Computational Linguistics", |
| "volume": "33", |
| "issue": "3", |
| "pages": "355--396", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hockenmaier, Julia and Mark Steedman. 2007. CCGbank: A corpus of CCG derivations and dependency structures extracted from the Penn Treebank. Computational Linguistics, 33(3):355-396.", |
| "links": null |
| }, |
| "BIBREF49": { |
| "ref_id": "b49", |
| "title": "Visualisation and 'diagnostic classifiers' reveal how recurrent and recursive neural networks process hierarchical structure", |
| "authors": [ |
| { |
| "first": "Dieuwke", |
| "middle": [], |
| "last": "Hupkes", |
| "suffix": "" |
| }, |
| { |
| "first": "Sara", |
| "middle": [], |
| "last": "Veldhoen", |
| "suffix": "" |
| }, |
| { |
| "first": "Willem", |
| "middle": [], |
| "last": "Zuidema", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1711.10203" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hupkes, Dieuwke, Sara Veldhoen, and Willem Zuidema. 2017. Visualisation and 'diagnostic classifiers' reveal how recurrent and recursive neural networks process hierarchical structure. arXiv preprint arXiv:1711.10203.", |
| "links": null |
| }, |
| "BIBREF50": { |
| "ref_id": "b50", |
| "title": "Google's multilingual neural machine translation system: Enabling zero-shot translation", |
| "authors": [ |
| { |
| "first": "Melvin", |
| "middle": [], |
| "last": "Johnson", |
| "suffix": "" |
| }, |
| { |
| "first": "Mike", |
| "middle": [], |
| "last": "Schuster", |
| "suffix": "" |
| }, |
| { |
| "first": "Quoc", |
| "middle": [], |
| "last": "Le", |
| "suffix": "" |
| }, |
| { |
| "first": "Maxim", |
| "middle": [], |
| "last": "Krikun", |
| "suffix": "" |
| }, |
| { |
| "first": "Yonghui", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Zhifeng", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Nikhil", |
| "middle": [], |
| "last": "Thorat", |
| "suffix": "" |
| }, |
| { |
| "first": "Fernanda", |
| "middle": [], |
| "last": "Vi\u00e9gas", |
| "suffix": "" |
| }, |
| { |
| "first": "Martin", |
| "middle": [], |
| "last": "Wattenberg", |
| "suffix": "" |
| }, |
| { |
| "first": "Greg", |
| "middle": [], |
| "last": "Corrado", |
| "suffix": "" |
| }, |
| { |
| "first": "Macduff", |
| "middle": [], |
| "last": "Hughes", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Dean", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Transactions of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "5339--351", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Johnson, Melvin, Mike Schuster, Quoc Le, Maxim Krikun, Yonghui Wu, Zhifeng Chen, Nikhil Thorat, Fernanda Vi\u00e9gas, Martin Wattenberg, Greg Corrado, Macduff Hughes, and Jeffrey Dean. 2017. Google's multilingual neural machine translation system: Enabling zero-shot translation. Transactions of the Association for Computational Linguistics, 5339-351.", |
| "links": null |
| }, |
| "BIBREF51": { |
| "ref_id": "b51", |
| "title": "Semantics-based machine translation with hyperedge replacement grammars", |
| "authors": [ |
| { |
| "first": "Bevan", |
| "middle": [], |
| "last": "Jones", |
| "suffix": "" |
| }, |
| { |
| "first": "Jacob", |
| "middle": [], |
| "last": "Andreas", |
| "suffix": "" |
| }, |
| { |
| "first": "Daniel", |
| "middle": [], |
| "last": "Bauer", |
| "suffix": "" |
| }, |
| { |
| "first": "Karl", |
| "middle": [], |
| "last": "Moritz", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Hermann", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Knight", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of COLING 2012", |
| "volume": "", |
| "issue": "", |
| "pages": "1359--1376", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jones, Bevan, Jacob Andreas, Daniel Bauer, Moritz Karl Hermann, and Kevin Knight. 2012. Semantics-based machine translation with hyperedge replacement grammars. In Proceedings of COLING 2012, pages 1359-1376.", |
| "links": null |
| }, |
| "BIBREF52": { |
| "ref_id": "b52", |
| "title": "Representation of linguistic form and function in recurrent neural networks", |
| "authors": [ |
| { |
| "first": "Akos", |
| "middle": [], |
| "last": "K\u00e1d\u00e1r", |
| "suffix": "" |
| }, |
| { |
| "first": "Grzegorz", |
| "middle": [], |
| "last": "Chrupa\u0142a", |
| "suffix": "" |
| }, |
| { |
| "first": "Afra", |
| "middle": [], |
| "last": "Alishahi", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Computational Linguistics", |
| "volume": "43", |
| "issue": "4", |
| "pages": "761--780", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "K\u00e1d\u00e1r, Akos, Grzegorz Chrupa\u0142a, and Afra Alishahi. 2017. Representation of linguistic form and function in recurrent neural networks. Computational Linguistics, 43(4):761-780.", |
| "links": null |
| }, |
| "BIBREF53": { |
| "ref_id": "b53", |
| "title": "Visualizing and understanding recurrent networks", |
| "authors": [ |
| { |
| "first": "Andrej", |
| "middle": [], |
| "last": "Karpathy", |
| "suffix": "" |
| }, |
| { |
| "first": "Justin", |
| "middle": [], |
| "last": "Johnson", |
| "suffix": "" |
| }, |
| { |
| "first": "Fei-Fei", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1506.02078" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Karpathy, Andrej, Justin Johnson, and Fei-Fei Li. 2015. Visualizing and understanding recurrent networks. arXiv preprint arXiv:1506.02078.", |
| "links": null |
| }, |
| "BIBREF54": { |
| "ref_id": "b54", |
| "title": "Seq2seq-attn", |
| "authors": [ |
| { |
| "first": "Yoon", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kim, Yoon. 2016. Seq2seq-attn. https:// github.com/harvardnlp/seq2seq-attn.", |
| "links": null |
| }, |
| "BIBREF55": { |
| "ref_id": "b55", |
| "title": "Character-aware Neural Language Models", |
| "authors": [ |
| { |
| "first": "Yoon", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| }, |
| { |
| "first": "Yacine", |
| "middle": [], |
| "last": "Jernite", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Sontag", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexander", |
| "middle": [ |
| "M" |
| ], |
| "last": "Rush", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1508.06615" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kim, Yoon, Yacine Jernite, David Sontag, and Alexander M. Rush. 2015. Character-aware Neural Language Models. arXiv preprint arXiv:1508.06615.", |
| "links": null |
| }, |
| "BIBREF56": { |
| "ref_id": "b56", |
| "title": "Using Test Suites in Evaluation of Machine Translation Systems", |
| "authors": [ |
| { |
| "first": "Margaret", |
| "middle": [], |
| "last": "King", |
| "suffix": "" |
| }, |
| { |
| "first": "Kirsten", |
| "middle": [], |
| "last": "Falkedal", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "Papers presented to the 13th International Conference on Computational Linguistics", |
| "volume": "2", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "King, Margaret and Kirsten Falkedal. 1990. Using Test Suites in Evaluation of Machine Translation Systems. In COLING 1990 Volume 2: Papers presented to the 13th International Conference on Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF57": { |
| "ref_id": "b57", |
| "title": "Adam: A method for stochastic optimization", |
| "authors": [ |
| { |
| "first": "Diederik", |
| "middle": [], |
| "last": "Kingma", |
| "suffix": "" |
| }, |
| { |
| "first": "Jimmy", |
| "middle": [], |
| "last": "Ba", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1412.6980" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kingma, Diederik and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980.", |
| "links": null |
| }, |
| "BIBREF58": { |
| "ref_id": "b58", |
| "title": "Factored translation models", |
| "authors": [ |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| }, |
| { |
| "first": "Hieu", |
| "middle": [], |
| "last": "Hoang", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Koehn, Philipp and Hieu Hoang. 2007. Factored translation models. In Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning.", |
| "links": null |
| }, |
| "BIBREF59": { |
| "ref_id": "b59", |
| "title": "Moses: Open source toolkit for statistical machine translation", |
| "authors": [ |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| }, |
| { |
| "first": "Hieu", |
| "middle": [], |
| "last": "Hoang", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexandra", |
| "middle": [], |
| "last": "Birch", |
| "suffix": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Callison-Burch", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcello", |
| "middle": [], |
| "last": "Federico", |
| "suffix": "" |
| }, |
| { |
| "first": "Nicola", |
| "middle": [], |
| "last": "Bertoldi", |
| "suffix": "" |
| }, |
| { |
| "first": "Brooke", |
| "middle": [], |
| "last": "Cowan", |
| "suffix": "" |
| }, |
| { |
| "first": "Wade", |
| "middle": [], |
| "last": "Shen", |
| "suffix": "" |
| }, |
| { |
| "first": "Christine", |
| "middle": [], |
| "last": "Moran", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Zens", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of the 45th annual meeting of the ACL on interactive poster and demonstration sessions", |
| "volume": "", |
| "issue": "", |
| "pages": "177--180", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Koehn, Philipp, Hieu Hoang, Alexandra Birch, Chris Callison-Burch, Marcello Federico, Nicola Bertoldi, Brooke Cowan, Wade Shen, Christine Moran, Richard Zens, et al. 2007. Moses: Open source toolkit for statistical machine translation. In Proceedings of the 45th annual meeting of the ACL on interactive poster and demonstration sessions, pages 177-180.", |
| "links": null |
| }, |
| "BIBREF60": { |
| "ref_id": "b60", |
| "title": "Empirical methods for compound splitting", |
| "authors": [ |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Knight", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "10th Conference of the European Chapter of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "187--194", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Koehn, Philipp and Kevin Knight. 2003. Empirical methods for compound splitting. In 10th Conference of the European Chapter of the Association for Computational Linguistics, pages 187-194.", |
| "links": null |
| }, |
| "BIBREF61": { |
| "ref_id": "b61", |
| "title": "Statistical phrase-based translation", |
| "authors": [ |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| }, |
| { |
| "first": "Franz", |
| "middle": [ |
| "Josef" |
| ], |
| "last": "Och", |
| "suffix": "" |
| }, |
| { |
| "first": "Daniel", |
| "middle": [], |
| "last": "Marcu", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proceedings of the North American Chapter of the Association for Computational Linguistics on Human Language Technology (HLT-NAACL'03)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Koehn, Philipp, Franz Josef Och, and Daniel Marcu. 2003. Statistical phrase-based translation. In Proceedings of the North American Chapter of the Association for Computational Linguistics on Human Language Technology (HLT-NAACL'03), Edmonton.", |
| "links": null |
| }, |
| "BIBREF62": { |
| "ref_id": "b62", |
| "title": "What's in an embedding? Analyzing word embeddings through multilingual evaluation", |
| "authors": [ |
| { |
| "first": "Arne", |
| "middle": [], |
| "last": "K\u00f6hn", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "2067--2073", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "K\u00f6hn, Arne. 2015. What's in an embedding? Analyzing word embeddings through multilingual evaluation. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pages 2067-2073, Lisbon.", |
| "links": null |
| }, |
| "BIBREF63": { |
| "ref_id": "b63", |
| "title": "Phrase reordering for statistical machine translation based on predicate-argument structure", |
| "authors": [ |
| { |
| "first": "Mamoru", |
| "middle": [], |
| "last": "Komachi", |
| "suffix": "" |
| }, |
| { |
| "first": "Yuji", |
| "middle": [], |
| "last": "Matsumoto", |
| "suffix": "" |
| }, |
| { |
| "first": "Masaaki", |
| "middle": [], |
| "last": "Nagata", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "International Workshop on Spoken Language Translation (IWSLT)", |
| "volume": "", |
| "issue": "", |
| "pages": "77--82", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Komachi, Mamoru, Yuji Matsumoto, and Masaaki Nagata. 2006. Phrase reordering for statistical machine translation based on predicate-argument structure. In International Workshop on Spoken Language Translation (IWSLT), pages 77-82.", |
| "links": null |
| }, |
| "BIBREF64": { |
| "ref_id": "b64", |
| "title": "The emergence of number and syntax units in LSTM language models", |
| "authors": [ |
| { |
| "first": "Yair", |
| "middle": [], |
| "last": "Lakretz", |
| "suffix": "" |
| }, |
| { |
| "first": "German", |
| "middle": [], |
| "last": "Kruszewski", |
| "suffix": "" |
| }, |
| { |
| "first": "Theo", |
| "middle": [], |
| "last": "Desbordes", |
| "suffix": "" |
| }, |
| { |
| "first": "Dieuwke", |
| "middle": [], |
| "last": "Hupkes", |
| "suffix": "" |
| }, |
| { |
| "first": "Stanislas", |
| "middle": [], |
| "last": "Dehaene", |
| "suffix": "" |
| }, |
| { |
| "first": "Marco", |
| "middle": [], |
| "last": "Baroni", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 17th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lakretz, Yair, German Kruszewski, Theo Desbordes, Dieuwke Hupkes, Stanislas Dehaene, and Marco Baroni. 2019. The emergence of number and syntax units in LSTM language models. In Proceedings of the 17th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT).", |
| "links": null |
| }, |
| "BIBREF65": { |
| "ref_id": "b65", |
| "title": "Fully character-level neural machine translation without explicit segmentation", |
| "authors": [ |
| { |
| "first": "Jason", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Kyunghyun", |
| "middle": [], |
| "last": "Cho", |
| "suffix": "" |
| }, |
| { |
| "first": "Thomas", |
| "middle": [], |
| "last": "Hofmann", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Transactions of the Association for Computational Linguistics", |
| "volume": "5", |
| "issue": "", |
| "pages": "365--378", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lee, Jason, Kyunghyun Cho, and Thomas Hofmann. 2017. Fully character-level neural machine translation without explicit segmentation. Transactions of the Association for Computational Linguistics, 5:365-378.", |
| "links": null |
| }, |
| "BIBREF66": { |
| "ref_id": "b66", |
| "title": "Modeling syntactic and semantic structures in hierarchical phrase-based translation", |
| "authors": [ |
| { |
| "first": "Junhui", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Philip", |
| "middle": [], |
| "last": "Resnik", |
| "suffix": "" |
| }, |
| { |
| "first": "Hal", |
| "middle": [], |
| "last": "Daum\u00e9", |
| "suffix": "" |
| }, |
| { |
| "first": "Iii", |
| "middle": [], |
| "last": "", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "540--549", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Li, Junhui, Philip Resnik, and Hal Daum\u00e9 III. 2013. Modeling syntactic and semantic structures in hierarchical phrase-based translation. In Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 540-549.", |
| "links": null |
| }, |
| "BIBREF67": { |
| "ref_id": "b67", |
| "title": "Character-based neural machine translation", |
| "authors": [ |
| { |
| "first": "Wang", |
| "middle": [], |
| "last": "Ling", |
| "suffix": "" |
| }, |
| { |
| "first": "Isabel", |
| "middle": [], |
| "last": "Trancoso", |
| "suffix": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Dyer", |
| "suffix": "" |
| }, |
| { |
| "first": "Alan", |
| "middle": [ |
| "W" |
| ], |
| "last": "Black", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ling, Wang, Isabel Trancoso, Chris Dyer, and Alan W. Black. 2015. Character-based neural machine translation. CoRR, abs/1511.04586.", |
| "links": null |
| }, |
| "BIBREF68": { |
| "ref_id": "b68", |
| "title": "The mythos of model interpretability", |
| "authors": [ |
| { |
| "first": "Zachary", |
| "middle": [ |
| "C" |
| ], |
| "last": "Lipton", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "ICML Workshop on Human Interpretability in Machine Learning (WHI)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lipton, Zachary C. 2016. The mythos of model interpretability. In ICML Workshop on Human Interpretability in Machine Learning (WHI).", |
| "links": null |
| }, |
| "BIBREF69": { |
| "ref_id": "b69", |
| "title": "Handling homographs in neural machine translation", |
| "authors": [ |
| { |
| "first": "Frederick", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Han", |
| "middle": [], |
| "last": "Lu", |
| "suffix": "" |
| }, |
| { |
| "first": "Graham", |
| "middle": [], |
| "last": "Neubig", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 16th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Liu, Frederick, Han Lu, and Graham Neubig. 2018. Handling homographs in neural machine translation. In Proceedings of the 16th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies.", |
| "links": null |
| }, |
| "BIBREF70": { |
| "ref_id": "b70", |
| "title": "Linguistic knowledge and transferability of contextual representations", |
| "authors": [ |
| { |
| "first": "Nelson", |
| "middle": [ |
| "F" |
| ], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Matt", |
| "middle": [], |
| "last": "Gardner", |
| "suffix": "" |
| }, |
| { |
| "first": "Yonatan", |
| "middle": [], |
| "last": "Belinkov", |
| "suffix": "" |
| }, |
| { |
| "first": "Matthew", |
| "middle": [], |
| "last": "Peters", |
| "suffix": "" |
| }, |
| { |
| "first": "Noah", |
| "middle": [ |
| "A" |
| ], |
| "last": "Smith", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 17th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Liu, Nelson F., Matt Gardner, Yonatan Belinkov, Matthew Peters, and Noah A. Smith. 2019. Linguistic knowledge and transferability of contextual representations. In Proceedings of the 17th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT).", |
| "links": null |
| }, |
| "BIBREF71": { |
| "ref_id": "b71", |
| "title": "Achieving open vocabulary neural machine translation with hybrid word-character Models", |
| "authors": [ |
| { |
| "first": "Minh-Thang", |
| "middle": [], |
| "last": "Luong", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [ |
| "D" |
| ], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1604.00788" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Luong, Minh-Thang and Christopher D. Manning. 2016. Achieving open vocabulary neural machine translation with hybrid word-character Models. arXiv preprint arXiv:1604.00788.", |
| "links": null |
| }, |
| "BIBREF72": { |
| "ref_id": "b72", |
| "title": "A hybrid morpheme-word representation for machine translation of morphologically rich languages", |
| "authors": [ |
| { |
| "first": "Minh-Thang", |
| "middle": [], |
| "last": "Luong", |
| "suffix": "" |
| }, |
| { |
| "first": "Preslav", |
| "middle": [], |
| "last": "Nakov", |
| "suffix": "" |
| }, |
| { |
| "first": "Min-Yen", |
| "middle": [], |
| "last": "Kan", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "148--157", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Luong, Minh-Thang, Preslav Nakov, and Min-Yen Kan. 2010. A hybrid morpheme-word representation for machine translation of morphologically rich languages. In Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing, pages 148-157.", |
| "links": null |
| }, |
| "BIBREF73": { |
| "ref_id": "b73", |
| "title": "Extracting syntactic trees from transformer encoder self-attentions", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Mare\u010dek", |
| "suffix": "" |
| }, |
| { |
| "first": "Rudolf", |
| "middle": [], |
| "last": "Rosa", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP", |
| "volume": "", |
| "issue": "", |
| "pages": "347--349", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mare\u010dek, David and Rudolf Rosa. 2018. Extracting syntactic trees from transformer encoder self-attentions. In Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, pages 347-349, Brussels.", |
| "links": null |
| }, |
| "BIBREF74": { |
| "ref_id": "b74", |
| "title": "Learned in translation: Contextualized word vectors", |
| "authors": [ |
| { |
| "first": "Bryan", |
| "middle": [], |
| "last": "Mccann", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Bradbury", |
| "suffix": "" |
| }, |
| { |
| "first": "Caiming", |
| "middle": [], |
| "last": "Xiong", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Advances in Neural Information Processing Systems", |
| "volume": "30", |
| "issue": "", |
| "pages": "6294--6305", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "McCann, Bryan, James Bradbury, Caiming Xiong, and Richard Socher. 2017. Learned in translation: Contextualized word vectors. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems 30, Curran Associates, Inc., pages 6294-6305.", |
| "links": null |
| }, |
| "BIBREF75": { |
| "ref_id": "b75", |
| "title": "Multilingual dependency analysis with a two-stage discriminative parser", |
| "authors": [ |
| { |
| "first": "Ryan", |
| "middle": [], |
| "last": "Mcdonald", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Lerman", |
| "suffix": "" |
| }, |
| { |
| "first": "Fernando", |
| "middle": [], |
| "last": "Pereira", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the Tenth Conference on Computational Natural Language Learning (CoNLL-X)", |
| "volume": "", |
| "issue": "", |
| "pages": "216--220", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "McDonald, Ryan, Kevin Lerman, and Fernando Pereira. 2006. Multilingual dependency analysis with a two-stage discriminative parser. In Proceedings of the Tenth Conference on Computational Natural Language Learning (CoNLL-X), pages 216-220.", |
| "links": null |
| }, |
| "BIBREF76": { |
| "ref_id": "b76", |
| "title": "Analyzing and integrating dependency parsers", |
| "authors": [ |
| { |
| "first": "Ryan", |
| "middle": [], |
| "last": "Mcdonald", |
| "suffix": "" |
| }, |
| { |
| "first": "Joakim", |
| "middle": [], |
| "last": "Nivre", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Computational Linguistics", |
| "volume": "37", |
| "issue": "1", |
| "pages": "197--230", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "McDonald, Ryan and Joakim Nivre. 2011. Analyzing and integrating dependency parsers. Computational Linguistics, 37(1):197-230.", |
| "links": null |
| }, |
| "BIBREF77": { |
| "ref_id": "b77", |
| "title": "Dependency Syntax: Theory and Practice", |
| "authors": [ |
| { |
| "first": "Igor", |
| "middle": [], |
| "last": "Mel'\u010duk", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Aleksandrovi\u010d", |
| "suffix": "" |
| } |
| ], |
| "year": 1988, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mel'\u010duk, Igor Aleksandrovi\u010d. 1988. Dependency Syntax: Theory and Practice. SUNY Press.", |
| "links": null |
| }, |
| "BIBREF79": { |
| "ref_id": "b79", |
| "title": "Predicting target language CCG supertags improves neural machine translation", |
| "authors": [ |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Dowmunt", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexandra", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Birch", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the Second Conference on Machine Translation", |
| "volume": "", |
| "issue": "", |
| "pages": "68--79", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dowmunt, Philipp Koehn, and Alexandra Birch. 2017. Predicting target language CCG supertags improves neural machine translation. In Proceedings of the Second Conference on Machine Translation, pages 68-79.", |
| "links": null |
| }, |
| "BIBREF80": { |
| "ref_id": "b80", |
| "title": "Combining word-level and character-level models for machine translation between closely-related languages", |
| "authors": [ |
| { |
| "first": "Preslav", |
| "middle": [], |
| "last": "Nakov", |
| "suffix": "" |
| }, |
| { |
| "first": "J\u00f6rg", |
| "middle": [], |
| "last": "Tiedemann", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "2", |
| "issue": "", |
| "pages": "301--305", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nakov, Preslav and J\u00f6rg Tiedemann. 2012. Combining word-level and character-level models for machine translation between closely-related languages. In Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 301-305, Jeju.", |
| "links": null |
| }, |
| "BIBREF81": { |
| "ref_id": "b81", |
| "title": "On the elements of an accurate tree-to-string machine translation system", |
| "authors": [ |
| { |
| "first": "Graham", |
| "middle": [], |
| "last": "Neubig", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Duh", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics", |
| "volume": "2", |
| "issue": "", |
| "pages": "143--149", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Neubig, Graham and Kevin Duh. 2014. On the elements of an accurate tree-to-string machine translation system. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 143-149.", |
| "links": null |
| }, |
| "BIBREF82": { |
| "ref_id": "b82", |
| "title": "Rdrpostagger: A ripple down rules-based part-of-speech tagger", |
| "authors": [ |
| { |
| "first": "Dat", |
| "middle": [], |
| "last": "Nguyen", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Quoc", |
| "suffix": "" |
| }, |
| { |
| "first": "Dang Duc", |
| "middle": [], |
| "last": "Dai Quoc Nguyen", |
| "suffix": "" |
| }, |
| { |
| "first": "Son", |
| "middle": [ |
| "Bao" |
| ], |
| "last": "Pham", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Pham", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the Demonstrations at the 14th Conference of the European Chapter of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "17--20", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nguyen, Dat Quoc, Dai Quoc Nguyen, Dang Duc Pham, and Son Bao Pham. 2014. Rdrpostagger: A ripple down rules-based part-of-speech tagger. In Proceedings of the Demonstrations at the 14th Conference of the European Chapter of the Association for Computational Linguistics, pages 17-20, Gothenburg.", |
| "links": null |
| }, |
| "BIBREF83": { |
| "ref_id": "b83", |
| "title": "Dependency Grammar and Dependency Parsing", |
| "authors": [ |
| { |
| "first": "Joakim", |
| "middle": [], |
| "last": "Nivre", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nivre, Joakim. 2005. Dependency Grammar and Dependency Parsing., Technical Report MSI 015133, V\u00e4xj\u00f6 University, School of Mathematics and Systems Engineering.", |
| "links": null |
| }, |
| "BIBREF84": { |
| "ref_id": "b84", |
| "title": "Universal Dependencies 2.0. LINDAT/CLARIN digital library at the Institute of Formal and Applied Linguistics", |
| "authors": [ |
| { |
| "first": "Joakim", |
| "middle": [], |
| "last": "Nivre", |
| "suffix": "" |
| }, |
| { |
| "first": "\u017deljko", |
| "middle": [], |
| "last": "Agi\u0107", |
| "suffix": "" |
| }, |
| { |
| "first": "Lars", |
| "middle": [], |
| "last": "Ahrenberg", |
| "suffix": "" |
| }, |
| { |
| "first": "Maria", |
| "middle": [ |
| "Jesus" |
| ], |
| "last": "Aranzabe", |
| "suffix": "" |
| }, |
| { |
| "first": "Masayuki", |
| "middle": [], |
| "last": "Asahara", |
| "suffix": "" |
| }, |
| { |
| "first": "Aitziber", |
| "middle": [], |
| "last": "Atutxa", |
| "suffix": "" |
| }, |
| { |
| "first": "Miguel", |
| "middle": [], |
| "last": "Ballesteros", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nivre, Joakim,\u017deljko Agi\u0107, Lars Ahrenberg, Maria Jesus Aranzabe, Masayuki Asahara, Aitziber Atutxa, Miguel Ballesteros et al. 2017. Universal Dependencies 2.0. LINDAT/CLARIN digital library at the Institute of Formal and Applied Linguistics, Charles University.", |
| "links": null |
| }, |
| "BIBREF85": { |
| "ref_id": "b85", |
| "title": "Semeval 2015 task 18: Broadcoverage semantic dependency parsing", |
| "authors": [ |
| { |
| "first": "Stephan", |
| "middle": [], |
| "last": "Oepen", |
| "suffix": "" |
| }, |
| { |
| "first": "Marco", |
| "middle": [], |
| "last": "Kuhlmann", |
| "suffix": "" |
| }, |
| { |
| "first": "Yusuke", |
| "middle": [], |
| "last": "Miyao", |
| "suffix": "" |
| }, |
| { |
| "first": "Daniel", |
| "middle": [], |
| "last": "Zeman", |
| "suffix": "" |
| }, |
| { |
| "first": "Silvie", |
| "middle": [], |
| "last": "Cinkova", |
| "suffix": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Flickinger", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 9th International Workshop on Semantic Evaluation", |
| "volume": "", |
| "issue": "", |
| "pages": "915--926", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Oepen, Stephan, Marco Kuhlmann, Yusuke Miyao, Daniel Zeman, Silvie Cinkova, Dan Flickinger, Jan Hajic, and Zdenka Uresova. 2015. Semeval 2015 task 18: Broad- coverage semantic dependency parsing. In Proceedings of the 9th International Workshop on Semantic Evaluation (SemEval 2015), pages 915-926.", |
| "links": null |
| }, |
| "BIBREF86": { |
| "ref_id": "b86", |
| "title": "Broad-coverage semantic dependency parsing", |
| "authors": [ |
| { |
| "first": "Stephan", |
| "middle": [], |
| "last": "Oepen", |
| "suffix": "" |
| }, |
| { |
| "first": "Marco", |
| "middle": [], |
| "last": "Kuhlmann", |
| "suffix": "" |
| }, |
| { |
| "first": "Yusuke", |
| "middle": [], |
| "last": "Miyao", |
| "suffix": "" |
| }, |
| { |
| "first": "Daniel", |
| "middle": [], |
| "last": "Zeman", |
| "suffix": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Flickinger", |
| "suffix": "" |
| }, |
| { |
| "first": "Jan", |
| "middle": [], |
| "last": "Hajic", |
| "suffix": "" |
| }, |
| { |
| "first": "Angelina", |
| "middle": [], |
| "last": "Ivanova", |
| "suffix": "" |
| }, |
| { |
| "first": "Yi", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 8th International Workshop on Semantic Evaluation", |
| "volume": "8", |
| "issue": "", |
| "pages": "63--72", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Oepen, Stephan, Marco Kuhlmann, Yusuke Miyao, Daniel Zeman, Dan Flickinger, Jan Hajic, Angelina Ivanova, and Yi Zhang. 2014. Semeval 2014 Task 8: Broad-coverage semantic dependency parsing. In Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval 2014), pages 63-72.", |
| "links": null |
| }, |
| "BIBREF87": { |
| "ref_id": "b87", |
| "title": "User's guide to sigf: Significance testing by approximate randomisation", |
| "authors": [ |
| { |
| "first": "Sebastian", |
| "middle": [], |
| "last": "Pad\u00f3", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Pad\u00f3, Sebastian. 2006. User's guide to sigf: Significance testing by approximate randomisation. https://www.nlpado.de/ sebastian/software/sigf.shtml.", |
| "links": null |
| }, |
| "BIBREF88": { |
| "ref_id": "b88", |
| "title": "BLEU: A method for automatic evaluation of machine translation", |
| "authors": [ |
| { |
| "first": "Kishore", |
| "middle": [], |
| "last": "Papineni", |
| "suffix": "" |
| }, |
| { |
| "first": "Salim", |
| "middle": [], |
| "last": "Roukos", |
| "suffix": "" |
| }, |
| { |
| "first": "Todd", |
| "middle": [], |
| "last": "Ward", |
| "suffix": "" |
| }, |
| { |
| "first": "Wei-Jing", |
| "middle": [], |
| "last": "Zhu", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Papineni, Kishore, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. BLEU: A method for automatic evaluation of machine translation. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF89": { |
| "ref_id": "b89", |
| "title": "MADAMIRA: A fast, comprehensive tool for morphological analysis and disambiguation of Arabic", |
| "authors": [ |
| { |
| "first": "Arfath", |
| "middle": [], |
| "last": "Pasha", |
| "suffix": "" |
| }, |
| { |
| "first": "Mohamed", |
| "middle": [], |
| "last": "Al-Badrashiny", |
| "suffix": "" |
| }, |
| { |
| "first": "Mona", |
| "middle": [], |
| "last": "Diab", |
| "suffix": "" |
| }, |
| { |
| "first": "Ahmed", |
| "middle": [ |
| "El" |
| ], |
| "last": "Kholy", |
| "suffix": "" |
| }, |
| { |
| "first": "Ramy", |
| "middle": [], |
| "last": "Eskander", |
| "suffix": "" |
| }, |
| { |
| "first": "Nizar", |
| "middle": [], |
| "last": "Habash", |
| "suffix": "" |
| }, |
| { |
| "first": "Manoj", |
| "middle": [], |
| "last": "Pooleery", |
| "suffix": "" |
| }, |
| { |
| "first": "Owen", |
| "middle": [], |
| "last": "Rambow", |
| "suffix": "" |
| }, |
| { |
| "first": "Ryan", |
| "middle": [], |
| "last": "Roth", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC'14)", |
| "volume": "", |
| "issue": "", |
| "pages": "1094--1101", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Pasha, Arfath, Mohamed Al-Badrashiny, Mona Diab, Ahmed El Kholy, Ramy Eskander, Nizar Habash, Manoj Pooleery, Owen Rambow, and Ryan Roth. 2014. MADAMIRA: A fast, comprehensive tool for morphological analysis and disambiguation of Arabic. In Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC'14), pages 1094-1101, Reykjavik.", |
| "links": null |
| }, |
| "BIBREF90": { |
| "ref_id": "b90", |
| "title": "To tune or not to tune?", |
| "authors": [ |
| { |
| "first": "Matthew", |
| "middle": [], |
| "last": "Peters", |
| "suffix": "" |
| }, |
| { |
| "first": "Sebastian", |
| "middle": [], |
| "last": "Ruder", |
| "suffix": "" |
| }, |
| { |
| "first": "Noah", |
| "middle": [ |
| "A" |
| ], |
| "last": "Smith", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Adapting pretrained representations to diverse tasks", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1903.05987" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Peters, Matthew, Sebastian Ruder, and Noah A. Smith. 2019. To tune or not to tune? Adapting pretrained representations to diverse tasks. arXiv preprint arXiv:1903.05987.", |
| "links": null |
| }, |
| "BIBREF91": { |
| "ref_id": "b91", |
| "title": "Deep contextualized word representations", |
| "authors": [ |
| { |
| "first": "Matthew", |
| "middle": [ |
| "E" |
| ], |
| "last": "Peters", |
| "suffix": "" |
| }, |
| { |
| "first": "Mark", |
| "middle": [], |
| "last": "Neumann", |
| "suffix": "" |
| }, |
| { |
| "first": "Mohit", |
| "middle": [], |
| "last": "Iyyer", |
| "suffix": "" |
| }, |
| { |
| "first": "Matt", |
| "middle": [], |
| "last": "Gardner", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [], |
| "last": "Clark", |
| "suffix": "" |
| }, |
| { |
| "first": "Kenton", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Luke", |
| "middle": [], |
| "last": "Zettlemoyer", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of NAACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Peters, Matthew E., Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark, Kenton Lee, and Luke Zettlemoyer. 2018a. Deep contextualized word representations. In Proceedings of NAACL.", |
| "links": null |
| }, |
| "BIBREF92": { |
| "ref_id": "b92", |
| "title": "Dissecting contextual word embeddings: Architecture and representation", |
| "authors": [ |
| { |
| "first": "Matthew", |
| "middle": [ |
| "E" |
| ], |
| "last": "Peters", |
| "suffix": "" |
| }, |
| { |
| "first": "Mark", |
| "middle": [], |
| "last": "Neumann", |
| "suffix": "" |
| }, |
| { |
| "first": "Luke", |
| "middle": [], |
| "last": "Zettlemoyer", |
| "suffix": "" |
| }, |
| { |
| "first": "Wen-Tau", |
| "middle": [], |
| "last": "Yih", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of EMNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Peters, Matthew E., Mark Neumann, Luke Zettlemoyer, and Wen-tau Yih. 2018b. Dissecting contextual word embeddings: Architecture and representation. In Proceedings of EMNLP.", |
| "links": null |
| }, |
| "BIBREF93": { |
| "ref_id": "b93", |
| "title": "Tilde's machine translation systems for WMT 2017", |
| "authors": [ |
| { |
| "first": "M\u0101rcis", |
| "middle": [], |
| "last": "Pinnis", |
| "suffix": "" |
| }, |
| { |
| "first": "Rihards", |
| "middle": [], |
| "last": "Kri\u0161lauks", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the Second Conference on Machine Translation", |
| "volume": "2", |
| "issue": "", |
| "pages": "374--381", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Pinnis, M\u0101rcis, Rihards Kri\u0161lauks, Toms Miks, Daiga Deksne, and Valters\u0160ics. 2017. Tilde's machine translation systems for WMT 2017. In Proceedings of the Second Conference on Machine Translation, Volume 2: Shared Task Papers, pages 374-381, Copenhagen.", |
| "links": null |
| }, |
| "BIBREF94": { |
| "ref_id": "b94", |
| "title": "Analyzing linguistic knowledge in sequential model of sentence", |
| "authors": [ |
| { |
| "first": "Peng", |
| "middle": [], |
| "last": "Qian", |
| "suffix": "" |
| }, |
| { |
| "first": "Xipeng", |
| "middle": [], |
| "last": "Qiu", |
| "suffix": "" |
| }, |
| { |
| "first": "Xuanjing", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Qian, Peng, Xipeng Qiu, and Xuanjing Huang. 2016a. Analyzing linguistic knowledge in sequential model of sentence. In Proceedings of the 2016", |
| "links": null |
| }, |
| "BIBREF95": { |
| "ref_id": "b95", |
| "title": "Conference on Empirical Methods in Natural Language Processing", |
| "authors": [], |
| "year": null, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "826--835", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Conference on Empirical Methods in Natural Language Processing, pages 826-835, Austin.", |
| "links": null |
| }, |
| "BIBREF96": { |
| "ref_id": "b96", |
| "title": "Investigating language universal and specific properties in word embeddings", |
| "authors": [ |
| { |
| "first": "Peng", |
| "middle": [], |
| "last": "Qian", |
| "suffix": "" |
| }, |
| { |
| "first": "Xipeng", |
| "middle": [], |
| "last": "Qiu", |
| "suffix": "" |
| }, |
| { |
| "first": "Xuanjing", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "1478--1488", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Qian, Peng, Xipeng Qiu, and Xuanjing Huang. 2016b. Investigating language universal and specific properties in word embeddings. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1478-1488, Berlin.", |
| "links": null |
| }, |
| "BIBREF97": { |
| "ref_id": "b97", |
| "title": "Improving language understanding by generative pre-training", |
| "authors": [ |
| { |
| "first": "Alec", |
| "middle": [], |
| "last": "Radford", |
| "suffix": "" |
| }, |
| { |
| "first": "Karthik", |
| "middle": [], |
| "last": "Narasimhan", |
| "suffix": "" |
| }, |
| { |
| "first": "Tim", |
| "middle": [], |
| "last": "Salimans", |
| "suffix": "" |
| }, |
| { |
| "first": "Ilya", |
| "middle": [], |
| "last": "Sutskever", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Radford, Alec, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever. 2018. Improving language understanding by generative pre-training. Technical report, OpenAI. https://s3-us-west- 2.amazonaws.com/openai- assets/research-covers/language- unsupervised/language understanding paper.pdf.", |
| "links": null |
| }, |
| "BIBREF98": { |
| "ref_id": "b98", |
| "title": "Language models are unsupervised multitask learners", |
| "authors": [ |
| { |
| "first": "Alec", |
| "middle": [], |
| "last": "Radford", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeff", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Rewon", |
| "middle": [], |
| "last": "Child", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Luan", |
| "suffix": "" |
| }, |
| { |
| "first": "Dario", |
| "middle": [], |
| "last": "Amodei", |
| "suffix": "" |
| }, |
| { |
| "first": "Ilya", |
| "middle": [], |
| "last": "Sutskever", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Radford, Alec, Jeff Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language models are unsupervised multitask learners. Technical report, OpenAI. https://d4mucfpksywv. cloudfront.net/better-language- models/language models are unsupervised multitask learners.pdf.", |
| "links": null |
| }, |
| "BIBREF99": { |
| "ref_id": "b99", |
| "title": "An analysis of encoder representations in transformer-based machine translation", |
| "authors": [ |
| { |
| "first": "Alessandro", |
| "middle": [], |
| "last": "Raganato", |
| "suffix": "" |
| }, |
| { |
| "first": "J\u00f6rg", |
| "middle": [], |
| "last": "Tiedemann", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP", |
| "volume": "", |
| "issue": "", |
| "pages": "287--297", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Raganato, Alessandro and J\u00f6rg Tiedemann. 2018. An analysis of encoder representations in transformer-based machine translation. In Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, pages 287-297.", |
| "links": null |
| }, |
| "BIBREF100": { |
| "ref_id": "b100", |
| "title": "Character-aware decoder for neural machine translation", |
| "authors": [ |
| { |
| "first": "Adithya", |
| "middle": [], |
| "last": "Renduchintala", |
| "suffix": "" |
| }, |
| { |
| "first": "Pamela", |
| "middle": [], |
| "last": "Shapiro", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Duh", |
| "suffix": "" |
| }, |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1809.02223" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Renduchintala, Adithya, Pamela Shapiro, Kevin Duh, and Philipp Koehn. 2018. Character-aware decoder for neural machine translation. arXiv preprint arXiv:1809.02223.", |
| "links": null |
| }, |
| "BIBREF101": { |
| "ref_id": "b101", |
| "title": "Improving word sense disambiguation in neural machine translation with sense embeddings", |
| "authors": [ |
| { |
| "first": "Rios", |
| "middle": [], |
| "last": "Gonzales", |
| "suffix": "" |
| }, |
| { |
| "first": "Annette", |
| "middle": [], |
| "last": "", |
| "suffix": "" |
| }, |
| { |
| "first": "Laura", |
| "middle": [], |
| "last": "Mascarell", |
| "suffix": "" |
| }, |
| { |
| "first": "Rico", |
| "middle": [], |
| "last": "Sennrich", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the Second Conference on Machine Translation", |
| "volume": "", |
| "issue": "", |
| "pages": "11--19", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Rios Gonzales, Annette, Laura Mascarell, and Rico Sennrich. 2017. Improving word sense disambiguation in neural machine translation with sense embeddings. In Proceedings of the Second Conference on Machine Translation, pages 11-19.", |
| "links": null |
| }, |
| "BIBREF102": { |
| "ref_id": "b102", |
| "title": "How grammatical is character-level neural machine translation? Assessing MT quality with contrastive translation pairs", |
| "authors": [ |
| { |
| "first": "Helmut", |
| "middle": [], |
| "last": "Schmid", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Kyoto", |
| "suffix": "" |
| }, |
| { |
| "first": "Rico", |
| "middle": [], |
| "last": "Sennrich", |
| "suffix": "" |
| } |
| ], |
| "year": 1994, |
| "venue": "Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics", |
| "volume": "2", |
| "issue": "", |
| "pages": "1715--1725", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Schmid, Helmut. 1994. Part-of-speech tagging with neural networks. In Proceedings of the 15th International Conference on Computational Linguistics (COLING 1994), pages 172-176, Kyoto. Sennrich, Rico. 2017. How grammatical is character-level neural machine translation? Assessing MT quality with contrastive translation pairs. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 2, Short Papers, pages 376-382, Valencia. Sennrich, Rico, Alexandra Birch, Anna Currey, Ulrich Germann, Barry Haddow, Kenneth Heafield, Antonio Valerio Miceli Barone, and Philip Williams. 2017. The University of Edinburgh's neural MT systems for WMT17. In Proceedings of the Second Conference on Machine Translation, Volume 2: Shared Task Papers, pages 389-399, Copenhagen. Sennrich, Rico, Barry Haddow, and Alexandra Birch. 2016. Neural machine translation of rare words with subword units. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1715-1725, Berlin.", |
| "links": null |
| }, |
| "BIBREF103": { |
| "ref_id": "b103", |
| "title": "BPE and CharCNNs for translation of morphology: A cross-lingual comparison and analysis", |
| "authors": [ |
| { |
| "first": "Pamela", |
| "middle": [], |
| "last": "Shapiro", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Duh", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1809.01301" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Shapiro, Pamela and Kevin Duh. 2018. BPE and CharCNNs for translation of morphology: A cross-lingual comparison and analysis. arXiv preprint arXiv:1809.01301.", |
| "links": null |
| }, |
| "BIBREF104": { |
| "ref_id": "b104", |
| "title": "String-to-dependency statistical machine translation", |
| "authors": [ |
| { |
| "first": "Libin", |
| "middle": [], |
| "last": "Shen", |
| "suffix": "" |
| }, |
| { |
| "first": "Jinxi", |
| "middle": [], |
| "last": "Xu", |
| "suffix": "" |
| }, |
| { |
| "first": "Ralph", |
| "middle": [], |
| "last": "Weischedel", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Computational Linguistics", |
| "volume": "36", |
| "issue": "4", |
| "pages": "649--671", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Shen, Libin, Jinxi Xu, and Ralph Weischedel. 2010. String-to-dependency statistical machine translation. Computational Linguistics, 36(4):649-671.", |
| "links": null |
| }, |
| "BIBREF105": { |
| "ref_id": "b105", |
| "title": "Does string-based neural MT learn source syntax?", |
| "authors": [ |
| { |
| "first": "Xing", |
| "middle": [], |
| "last": "Shi", |
| "suffix": "" |
| }, |
| { |
| "first": "Inkit", |
| "middle": [], |
| "last": "Padhi", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Knight", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "1526--1534", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Shi, Xing, Inkit Padhi, and Kevin Knight. 2016. Does string-based neural MT learn source syntax? In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 1526-1534", |
| "links": null |
| }, |
| "BIBREF107": { |
| "ref_id": "b107", |
| "title": "Morfessor 2.0: Toolkit for statistical morphological segmentation", |
| "authors": [ |
| { |
| "first": "Peter", |
| "middle": [], |
| "last": "Smit", |
| "suffix": "" |
| }, |
| { |
| "first": "Sami", |
| "middle": [], |
| "last": "Virpioja", |
| "suffix": "" |
| }, |
| { |
| "first": "Stig-Arne", |
| "middle": [], |
| "last": "Gr\u00f6nroos", |
| "suffix": "" |
| }, |
| { |
| "first": "Mikko", |
| "middle": [], |
| "last": "Kurimo", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the Demonstrations at the 14th Conference of the European Chapter of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "21--24", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Smit, Peter, Sami Virpioja, Stig-Arne Gr\u00f6nroos, and Mikko Kurimo. 2014. Morfessor 2.0: Toolkit for statistical morphological segmentation. In Proceedings of the Demonstrations at the 14th Conference of the European Chapter of the Association for Computational Linguistics, pages 21-24, Gothenburg.", |
| "links": null |
| }, |
| "BIBREF108": { |
| "ref_id": "b108", |
| "title": "Syntactically guided neural machine translation", |
| "authors": [ |
| { |
| "first": "Felix", |
| "middle": [], |
| "last": "Stahlberg", |
| "suffix": "" |
| }, |
| { |
| "first": "Eva", |
| "middle": [], |
| "last": "Hasler", |
| "suffix": "" |
| }, |
| { |
| "first": "Aurelien", |
| "middle": [], |
| "last": "Waite", |
| "suffix": "" |
| }, |
| { |
| "first": "Bill", |
| "middle": [], |
| "last": "Byrne", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "2", |
| "issue": "", |
| "pages": "299--305", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Stahlberg, Felix, Eva Hasler, Aurelien Waite, and Bill Byrne. 2016. Syntactically guided neural machine translation. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 299-305.", |
| "links": null |
| }, |
| "BIBREF110": { |
| "ref_id": "b110", |
| "title": "Sequence to sequence learning with neural networks", |
| "authors": [ |
| { |
| "first": "Ilya", |
| "middle": [], |
| "last": "Sutskever", |
| "suffix": "" |
| }, |
| { |
| "first": "Oriol", |
| "middle": [], |
| "last": "Vinyals", |
| "suffix": "" |
| }, |
| { |
| "first": "V", |
| "middle": [], |
| "last": "Quoc", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Le", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Advances in Neural Information Processing Systems", |
| "volume": "", |
| "issue": "", |
| "pages": "3104--3112", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sutskever, Ilya, Oriol Vinyals, and Quoc V. Le. 2014. Sequence to sequence learning with neural networks. In Z. Ghahramani, M. Welling, C. Cortes, N. D. Lawrence, and K. Q. Weinberger, editors, Advances in Neural Information Processing Systems, pages 3104-3112.", |
| "links": null |
| }, |
| "BIBREF111": { |
| "ref_id": "b111", |
| "title": "What do you learn from context? Probing for sentence structure in contextualized word representations", |
| "authors": [ |
| { |
| "first": "Ian", |
| "middle": [], |
| "last": "Tenney", |
| "suffix": "" |
| }, |
| { |
| "first": "Patrick", |
| "middle": [], |
| "last": "Xia", |
| "suffix": "" |
| }, |
| { |
| "first": "Berlin", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Adam", |
| "middle": [], |
| "last": "Poliak", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [ |
| "Thomas" |
| ], |
| "last": "Mccoy", |
| "suffix": "" |
| }, |
| { |
| "first": "Najoung", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| }, |
| { |
| "first": "Benjamin", |
| "middle": [], |
| "last": "Van Durme", |
| "suffix": "" |
| }, |
| { |
| "first": "Samuel", |
| "middle": [ |
| "R" |
| ], |
| "last": "Bowman", |
| "suffix": "" |
| }, |
| { |
| "first": "Dipanjan", |
| "middle": [], |
| "last": "Das", |
| "suffix": "" |
| }, |
| { |
| "first": "Ellie", |
| "middle": [], |
| "last": "Pavlick", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of ICLR", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tenney, Ian, Patrick Xia, Berlin Chen, Alex Wang, Adam Poliak, R. Thomas McCoy, Najoung Kim, Benjamin Van Durme, Samuel R. Bowman, Dipanjan Das, and Ellie Pavlick. 2019. What do you learn from context? Probing for sentence structure in contextualized word representations. In Proceedings of ICLR.", |
| "links": null |
| }, |
| "BIBREF112": { |
| "ref_id": "b112", |
| "title": "A multifaceted evaluation of neural versus phrase-based machine translation for 9 language directions", |
| "authors": [ |
| { |
| "first": "Antonio", |
| "middle": [], |
| "last": "Toral", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "V\u00edctor", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "S\u00e1nchez-Cartagena", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 15th Conference of the European Chapter", |
| "volume": "1", |
| "issue": "", |
| "pages": "1063--1073", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Toral, Antonio and V\u00edctor M. S\u00e1nchez- Cartagena. 2017. A multifaceted evaluation of neural versus phrase-based machine translation for 9 language directions. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 1, Long Papers, pages 1063-1073, Valencia.", |
| "links": null |
| }, |
| "BIBREF113": { |
| "ref_id": "b113", |
| "title": "Attention is all you need", |
| "authors": [ |
| { |
| "first": "Ke", |
| "middle": [], |
| "last": "Tran", |
| "suffix": "" |
| }, |
| { |
| "first": "Arianna", |
| "middle": [], |
| "last": "Bisazza", |
| "suffix": "" |
| }, |
| { |
| "first": "Christof", |
| "middle": [], |
| "last": "Monz", |
| "suffix": "" |
| }, |
| { |
| "first": ";", |
| "middle": [], |
| "last": "Vaswani", |
| "suffix": "" |
| }, |
| { |
| "first": "Ashish", |
| "middle": [], |
| "last": "", |
| "suffix": "" |
| }, |
| { |
| "first": "Noam", |
| "middle": [], |
| "last": "Shazeer", |
| "suffix": "" |
| }, |
| { |
| "first": "Niki", |
| "middle": [], |
| "last": "Parmar", |
| "suffix": "" |
| }, |
| { |
| "first": "Jakob", |
| "middle": [], |
| "last": "Uszkoreit", |
| "suffix": "" |
| }, |
| { |
| "first": "Llion", |
| "middle": [], |
| "last": "Jones", |
| "suffix": "" |
| }, |
| { |
| "first": "Aidan", |
| "middle": [ |
| "N" |
| ], |
| "last": "Gomez", |
| "suffix": "" |
| }, |
| { |
| "first": "Illia", |
| "middle": [], |
| "last": "Kaiser", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Polosukhin", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "The importance of being recurrent for modeling hierarchical structure", |
| "volume": "30", |
| "issue": "", |
| "pages": "5998--6008", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1803.03585" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tran, Ke, Arianna Bisazza, and Christof Monz. 2018. The importance of being recurrent for modeling hierarchical structure. arXiv preprint arXiv:1803.03585. Vaswani, Ashish, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, \u0141 ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems 30, Curran Associates, Inc., pages 5998-6008.", |
| "links": null |
| }, |
| "BIBREF114": { |
| "ref_id": "b114", |
| "title": "Gate activation signal analysis for gated recurrent neural networks and its correlation with phoneme boundaries", |
| "authors": [ |
| { |
| "first": "Ekaterina", |
| "middle": [], |
| "last": "Vylomova", |
| "suffix": "" |
| }, |
| { |
| "first": "Trevor", |
| "middle": [], |
| "last": "Cohn", |
| "suffix": "" |
| }, |
| { |
| "first": "Xuanli", |
| "middle": [], |
| "last": "He", |
| "suffix": "" |
| }, |
| { |
| "first": "Gholamreza", |
| "middle": [], |
| "last": "Haffari", |
| "suffix": "" |
| }, |
| { |
| "first": ";", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Yu", |
| "middle": [], |
| "last": "Hsuan", |
| "suffix": "" |
| }, |
| { |
| "first": "Cheng-Tao", |
| "middle": [], |
| "last": "Chung", |
| "suffix": "" |
| }, |
| { |
| "first": "Hung-Yi", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1606.04217" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Vylomova, Ekaterina, Trevor Cohn, Xuanli He, and Gholamreza Haffari. 2016. Word representation models for morphologically rich languages in neural machine translation. arXiv preprint arXiv:1606.04217. Wang, Yu Hsuan, Cheng-Tao Chung, and Hung-yi Lee. 2017. Gate activation signal analysis for gated recurrent neural networks and its correlation with phoneme boundaries. In Interspeech 2017.", |
| "links": null |
| }, |
| "BIBREF115": { |
| "ref_id": "b115", |
| "title": "Translation. Machine Translation of Languages", |
| "authors": [ |
| { |
| "first": "Warren", |
| "middle": [], |
| "last": "Weaver", |
| "suffix": "" |
| } |
| ], |
| "year": 1955, |
| "venue": "", |
| "volume": "14", |
| "issue": "", |
| "pages": "15--23", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Weaver, Warren. 1955. Translation. Machine Translation of Languages, 14:15-23.", |
| "links": null |
| }, |
| "BIBREF116": { |
| "ref_id": "b116", |
| "title": "Syntax-based statistical machine translation", |
| "authors": [ |
| { |
| "first": "Philip", |
| "middle": [], |
| "last": "Williams", |
| "suffix": "" |
| }, |
| { |
| "first": "Rico", |
| "middle": [], |
| "last": "Sennrich", |
| "suffix": "" |
| }, |
| { |
| "first": "Matt", |
| "middle": [], |
| "last": "Post", |
| "suffix": "" |
| }, |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Synthesis Lectures on Human Language Technologies", |
| "volume": "9", |
| "issue": "4", |
| "pages": "1--208", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Williams, Philip, Rico Sennrich, Matt Post, and Philipp Koehn. 2016. Syntax-based statistical machine translation. Synthesis Lectures on Human Language Technologies, 9(4):1-208.", |
| "links": null |
| }, |
| "BIBREF117": { |
| "ref_id": "b117", |
| "title": "Sequence-to-dependency neural machine translation", |
| "authors": [ |
| { |
| "first": "Shuangzhi", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Dongdong", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Nan", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "Mu", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "698--707", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wu, Shuangzhi, Dongdong Zhang, Nan Yang, Mu Li, and Ming Zhou. 2017. Sequence-to-dependency neural machine translation. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 698-707.", |
| "links": null |
| }, |
| "BIBREF118": { |
| "ref_id": "b118", |
| "title": "Extracting pre-ordering rules from predicate-argument structures", |
| "authors": [ |
| { |
| "first": "Xianchao", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Katsuhito", |
| "middle": [], |
| "last": "Sudoh", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Duh", |
| "suffix": "" |
| }, |
| { |
| "first": "Hajime", |
| "middle": [], |
| "last": "Tsukada", |
| "suffix": "" |
| }, |
| { |
| "first": "Masaaki", |
| "middle": [], |
| "last": "Nagata", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of 5th International Joint Conference on Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "29--37", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wu, Xianchao, Katsuhito Sudoh, Kevin Duh, Hajime Tsukada, and Masaaki Nagata. 2011. Extracting pre-ordering rules from predicate-argument structures. In Proceedings of 5th International Joint Conference on Natural Language Processing, pages 29-37.", |
| "links": null |
| }, |
| "BIBREF119": { |
| "ref_id": "b119", |
| "title": "Google's neural machine translation system: Bridging the gap between human and machine translation", |
| "authors": [ |
| { |
| "first": "Yonghui", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Mike", |
| "middle": [], |
| "last": "Schuster", |
| "suffix": "" |
| }, |
| { |
| "first": "Zhifeng", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Quoc", |
| "middle": [ |
| "V" |
| ], |
| "last": "Le", |
| "suffix": "" |
| }, |
| { |
| "first": "Mohammad", |
| "middle": [], |
| "last": "Norouzi", |
| "suffix": "" |
| }, |
| { |
| "first": "Wolfgang", |
| "middle": [], |
| "last": "Macherey", |
| "suffix": "" |
| }, |
| { |
| "first": "Maxim", |
| "middle": [], |
| "last": "Krikun", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wu, Yonghui, Mike Schuster, Zhifeng Chen, Quoc V. Le, Mohammad Norouzi, Wolfgang Macherey, Maxim Krikun, et al. 2016a. Google's neural machine translation system: Bridging the gap between human and machine translation. CoRR, abs/1609.08144.", |
| "links": null |
| }, |
| "BIBREF120": { |
| "ref_id": "b120", |
| "title": "Google's neural machine translation system: Bridging the gap between human and machine translation", |
| "authors": [ |
| { |
| "first": "Yonghui", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Mike", |
| "middle": [], |
| "last": "Schuster", |
| "suffix": "" |
| }, |
| { |
| "first": "Zhifeng", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Quoc", |
| "middle": [ |
| "V" |
| ], |
| "last": "Le", |
| "suffix": "" |
| }, |
| { |
| "first": "Mohammad", |
| "middle": [], |
| "last": "Norouzi", |
| "suffix": "" |
| }, |
| { |
| "first": "Wolfgang", |
| "middle": [], |
| "last": "Macherey", |
| "suffix": "" |
| }, |
| { |
| "first": "Maxim", |
| "middle": [], |
| "last": "Krikun", |
| "suffix": "" |
| }, |
| { |
| "first": "Yuan", |
| "middle": [], |
| "last": "Cao", |
| "suffix": "" |
| }, |
| { |
| "first": "Qin", |
| "middle": [], |
| "last": "Gao", |
| "suffix": "" |
| }, |
| { |
| "first": "Klaus", |
| "middle": [], |
| "last": "Macherey", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1609.08144" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wu, Yonghui, Mike Schuster, Zhifeng Chen, Quoc V. Le, Mohammad Norouzi, Wolfgang Macherey, Maxim Krikun, Yuan Cao, Qin Gao, Klaus Macherey, et al. 2016b. Google's neural machine translation system: Bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144.", |
| "links": null |
| }, |
| "BIBREF121": { |
| "ref_id": "b121", |
| "title": "Investigating gated recurrent networks for speech synthesis", |
| "authors": [ |
| { |
| "first": "Zhizheng", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Simon", |
| "middle": [], |
| "last": "King", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)", |
| "volume": "", |
| "issue": "", |
| "pages": "5140--5144", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wu, Zhizheng and Simon King. 2016. Investigating gated recurrent networks for speech synthesis. In 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 5140-5144.", |
| "links": null |
| }, |
| "BIBREF122": { |
| "ref_id": "b122", |
| "title": "Modeling the translation of predicate-argument structure for SMT", |
| "authors": [ |
| { |
| "first": "Deyi", |
| "middle": [], |
| "last": "Xiong", |
| "suffix": "" |
| }, |
| { |
| "first": "Min", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Haizhou", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "902--911", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Xiong, Deyi, Min Zhang, and Haizhou Li. 2012. Modeling the translation of predicate-argument structure for SMT. In Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 902-911.", |
| "links": null |
| }, |
| "BIBREF123": { |
| "ref_id": "b123", |
| "title": "CCG supertagging with a recurrent neural network", |
| "authors": [ |
| { |
| "first": "Wenduan", |
| "middle": [], |
| "last": "Xu", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Auli", |
| "suffix": "" |
| }, |
| { |
| "first": "Stephen", |
| "middle": [], |
| "last": "Clark", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing", |
| "volume": "2", |
| "issue": "", |
| "pages": "250--255", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Xu, Wenduan, Michael Auli, and Stephen Clark. 2015. CCG supertagging with a recurrent neural network. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pages 250-255, Beijing.", |
| "links": null |
| }, |
| "BIBREF124": { |
| "ref_id": "b124", |
| "title": "A decoder for syntax-based statistical MT", |
| "authors": [ |
| { |
| "first": "Kenji", |
| "middle": [], |
| "last": "Yamada", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Knight", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yamada, Kenji and Kevin Knight. 2002. A decoder for syntax-based statistical MT. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF125": { |
| "ref_id": "b125", |
| "title": "Language modeling teaches you more syntax than translation does: Lessons learned through auxiliary task analysis", |
| "authors": [ |
| { |
| "first": "Kelly", |
| "middle": [ |
| "W" |
| ], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Samuel", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Bowman", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of BlackboxNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zhang, Kelly W. and Samuel R. Bowman. 2018. Language modeling teaches you more syntax than translation does: Lessons learned through auxiliary task analysis. In Proceedings of BlackboxNLP.", |
| "links": null |
| }, |
| "BIBREF126": { |
| "ref_id": "b126", |
| "title": "A tree-to-tree alignment-based model for statistical machine translation. MT-Summit-07", |
| "authors": [ |
| { |
| "first": "Min", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Hongfei", |
| "middle": [], |
| "last": "Jiang", |
| "suffix": "" |
| }, |
| { |
| "first": "Aiti", |
| "middle": [], |
| "last": "Aw", |
| "suffix": "" |
| }, |
| { |
| "first": "Jun", |
| "middle": [], |
| "last": "Sun", |
| "suffix": "" |
| }, |
| { |
| "first": "Sheng", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Chew Lim", |
| "middle": [], |
| "last": "Tan", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "535--542", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zhang, Min, Hongfei Jiang, AiTi Aw, Jun Sun, Sheng Li, and Chew Lim Tan. 2007. A tree-to-tree alignment-based model for statistical machine translation. MT-Summit-07, pages 535-542.", |
| "links": null |
| }, |
| "BIBREF127": { |
| "ref_id": "b127", |
| "title": "Deep recurrent models with fast-forward connections for neural machine translation", |
| "authors": [ |
| { |
| "first": "Jie", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| }, |
| { |
| "first": "Ying", |
| "middle": [], |
| "last": "Cao", |
| "suffix": "" |
| }, |
| { |
| "first": "Xuguang", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Peng", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Wei", |
| "middle": [], |
| "last": "Xu", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Transactions of the Association for Computational Linguistics", |
| "volume": "4", |
| "issue": "", |
| "pages": "371--383", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zhou, Jie, Ying Cao, Xuguang Wang, Peng Li, and Wei Xu. 2016. Deep recurrent models with fast-forward connections for neural machine translation. Transactions of the Association for Computational Linguistics, 4:371-383.", |
| "links": null |
| }, |
| "BIBREF128": { |
| "ref_id": "b128", |
| "title": "The United Nations Parallel Corpus v1.0", |
| "authors": [ |
| { |
| "first": "Michal", |
| "middle": [], |
| "last": "Ziemski", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcin", |
| "middle": [], |
| "last": "Junczys-Dowmunt", |
| "suffix": "" |
| }, |
| { |
| "first": "Bruno", |
| "middle": [], |
| "last": "Pouliquen", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC 2016)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ziemski, Michal, Marcin Junczys-Dowmunt, and Bruno Pouliquen. 2016. The United Nations Parallel Corpus v1.0. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC 2016), Paris.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "num": null, |
| "uris": null, |
| "text": "Characters P r o f e s s o r a d m i t s t o s h o o t i n g h i s g i r l f r i e n d", |
| "type_str": "figure" |
| }, |
| "FIGREF1": { |
| "num": null, |
| "uris": null, |
| "text": "Morphological tagging accuracy vs. word frequency for different translation units on the encoder-side. The target side is held fixed as BPE. The representations for training the morphological classifier are obtained from the top layer of the encoder. Character representations perform better than other ones, especially in low-frequency regimes.", |
| "type_str": "figure" |
| }, |
| "FIGREF2": { |
| "num": null, |
| "uris": null, |
| "text": "Improvement in accuracy of syntactic relation labeling with layers 2/3/4 compared with layer 1. The figure shows the five most improved relations when averaging results obtained with encoder representations from (word-based) NMT models trained on translating English to other languages.", |
| "type_str": "figure" |
| }, |
| "TABREF2": { |
| "text": "Number of tags (for word-level tasks) and labels (for relation-level tasks) per task in different languages.", |
| "num": null, |
| "html": null, |
| "content": "<table><tr><td/><td>de</td><td>cs</td><td>ru</td><td>en</td><td>ar</td><td>fr</td><td>es</td></tr><tr><td>POS tags</td><td>54</td><td>-</td><td>-</td><td>42</td><td>42</td><td>33</td><td>-</td></tr><tr><td>Morphological tags</td><td colspan=\"3\">509 1,004 602</td><td>-</td><td colspan=\"3\">1,969 183 212</td></tr><tr><td>Semantic tags</td><td>69</td><td>-</td><td>-</td><td>66</td><td>-</td><td>-</td><td/></tr><tr><td>CCG tags</td><td>-</td><td>-</td><td>-</td><td>1,272</td><td>-</td><td>-</td><td/></tr><tr><td>Syntactic dependency labels</td><td>35</td><td>41</td><td>40</td><td>-</td><td>-</td><td>40</td><td>32</td></tr><tr><td>Semantic dependency labels</td><td>-</td><td>64</td><td>-</td><td>87</td><td>-</td><td>-</td><td>-</td></tr></table>", |
| "type_str": "table" |
| }, |
| "TABREF5": { |
| "text": "BLEU scores across language pairs with different translation units on the source side (the target side is held fixed as BPE). The NMT models are trained on NEWS+TED data.", |
| "num": null, |
| "html": null, |
| "content": "<table><tr><td/><td colspan=\"4\">de-en cs-en ru-en en-de</td></tr><tr><td>word</td><td>34.0</td><td>27.5</td><td>20.9</td><td>29.7</td></tr><tr><td>bpe</td><td>35.6</td><td>28.4</td><td>22.4</td><td>30.2</td></tr><tr><td>morfessor</td><td>35.5</td><td>28.5</td><td>22.5</td><td>29.9</td></tr><tr><td>char</td><td>34.9</td><td>29.0</td><td>21.3</td><td>30.0</td></tr><tr><td/><td/><td/><td/><td>-de</td></tr><tr><td>MT</td><td>3.42</td><td>6.46</td><td>6.86</td><td>0.82</td></tr><tr><td>Classifier</td><td>4.42</td><td>6.13</td><td>6.61</td><td>2.09</td></tr></table>", |
| "type_str": "table" |
| }, |
| "TABREF10": { |
| "text": "BLEU scores and morphological classifier accuracy across language pairs, comparing (encoders of) fully character-based LSTM and charCNN LSTM models. Effect of changing the target language on POS and morphological tagging with classifiers trained on the encoder side of both word-based and character-based (here: charCNN) models. The source language, for which classification is done, is always Arabic. Impact of changing the target language on POS tagging accuracy with classifiers trained on the encoder side. Self = German/Czech in rows 1/2, respectively. Results on syntactic and semantic tagging and labeling with representations obtained from char-based models trained with an extra layer.", |
| "num": null, |
| "html": null, |
| "content": "<table><tr><td>Table A.2</td><td/><td/><td/><td/><td/></tr><tr><td/><td/><td/><td>ar</td><td>he</td><td>de</td><td>en</td></tr><tr><td colspan=\"2\">Word POS</td><td/><td colspan=\"4\">67.21 78.13 78.85 80.21</td></tr><tr><td/><td colspan=\"6\">Morphology 55.63 64.87 65.91 67.18</td></tr><tr><td/><td>BLEU</td><td/><td>80.43</td><td colspan=\"2\">9.51 11.49</td><td>23.8</td></tr><tr><td>Char</td><td>POS</td><td/><td colspan=\"4\">87.72 92.67 93.05 93.63</td></tr><tr><td/><td colspan=\"6\">Morphology 75.21 80.50 80.61 81.49</td></tr><tr><td/><td>BLEU</td><td/><td colspan=\"4\">75.48 11.15 12.86 27.82</td></tr><tr><td>Table A.3</td><td/><td/><td/><td/><td/></tr><tr><td/><td/><td/><td colspan=\"3\">Target language</td></tr><tr><td colspan=\"2\">Source language</td><td colspan=\"4\">English Arabic Self</td></tr><tr><td>German</td><td/><td/><td>93.5</td><td>92.7</td><td>89.3</td></tr><tr><td>Czech</td><td/><td/><td>75.7</td><td>75.2</td><td>71.8</td></tr><tr><td>Table A.4</td><td/><td/><td/><td/><td/></tr><tr><td/><td/><td/><td>cs-en</td><td/><td/><td>en-de</td></tr><tr><td/><td colspan=\"4\">Syn Dep Sem Dep</td><td colspan=\"2\">Syn Dep Sem Dep Sem tags</td></tr><tr><td colspan=\"2\">char (layer 2)</td><td>89.3</td><td>84.3</td><td/><td colspan=\"2\">90.3</td><td>78.9</td><td>92.3</td></tr><tr><td colspan=\"2\">char (layer 3)</td><td>90.2</td><td>85.2</td><td/><td colspan=\"2\">91.1</td><td>79.6</td><td>92.7</td></tr><tr><td colspan=\"2\">best subword</td><td>90.3</td><td>86.3</td><td/><td colspan=\"2\">91.4</td><td>80.4</td><td>93.2</td></tr><tr><td/><td/><td colspan=\"2\">BLEU</td><td/><td colspan=\"2\">Accuracy</td></tr><tr><td/><td/><td colspan=\"2\">de-en cs-en</td><td colspan=\"3\">de-en cs-en</td></tr><tr><td colspan=\"2\">char \u2192 bpe</td><td>34.9</td><td>29.0</td><td/><td>79.3</td><td>81.7</td></tr><tr><td colspan=\"2\">charCNN \u2192 bpe</td><td>32.3</td><td>28.0</td><td/><td>79.0</td><td>79.9</td></tr></table>", |
| "type_str": "table" |
| } |
| } |
| } |
| } |