| { |
| "paper_id": "P14-1013", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T09:04:35.340095Z" |
| }, |
| "title": "Learning Topic Representation for SMT with Neural Networks *", |
| "authors": [ |
| { |
| "first": "Lei", |
| "middle": [], |
| "last": "Cui", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Harbin Institute of Technology", |
| "location": { |
| "settlement": "Harbin", |
| "country": "P.R. China" |
| } |
| }, |
| "email": "leicui@hit.edu.cn" |
| }, |
| { |
| "first": "Dongdong", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Microsoft Research", |
| "location": { |
| "settlement": "Beijing", |
| "country": "P.R. China" |
| } |
| }, |
| "email": "dozhang@microsoft.com" |
| }, |
| { |
| "first": "Shujie", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Microsoft Research", |
| "location": { |
| "settlement": "Beijing", |
| "country": "P.R. China" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Qiming", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Tong University", |
| "location": { |
| "settlement": "Shanghai", |
| "country": "P.R. China" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Mu", |
| "middle": [], |
| "last": "Li", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Microsoft Research", |
| "location": { |
| "settlement": "Beijing", |
| "country": "P.R. China" |
| } |
| }, |
| "email": "muli@microsoft.com" |
| }, |
| { |
| "first": "Ming", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Microsoft Research", |
| "location": { |
| "settlement": "Beijing", |
| "country": "P.R. China" |
| } |
| }, |
| "email": "mingzhou@microsoft.com" |
| }, |
| { |
| "first": "Muyun", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Harbin Institute of Technology", |
| "location": { |
| "settlement": "Harbin", |
| "country": "P.R. China" |
| } |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Statistical Machine Translation (SMT) usually utilizes contextual information to disambiguate translation candidates. However, it is often limited to contexts within sentence boundaries, hence broader topical information cannot be leveraged. In this paper, we propose a novel approach to learning topic representation for parallel data using a neural network architecture, where abundant topical contexts are embedded via topic relevant monolingual data. By associating each translation rule with the topic representation, topic relevant rules are selected according to the distributional similarity with the source text during SMT decoding. Experimental results show that our method significantly improves translation accuracy in the NIST Chinese-to-English translation task compared to a state-of-the-art baseline.", |
| "pdf_parse": { |
| "paper_id": "P14-1013", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Statistical Machine Translation (SMT) usually utilizes contextual information to disambiguate translation candidates. However, it is often limited to contexts within sentence boundaries, hence broader topical information cannot be leveraged. In this paper, we propose a novel approach to learning topic representation for parallel data using a neural network architecture, where abundant topical contexts are embedded via topic relevant monolingual data. By associating each translation rule with the topic representation, topic relevant rules are selected according to the distributional similarity with the source text during SMT decoding. Experimental results show that our method significantly improves translation accuracy in the NIST Chinese-to-English translation task compared to a state-of-the-art baseline.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Making translation decisions is a difficult task in many Statistical Machine Translation (SMT) systems. Current translation modeling approaches usually use context dependent information to disambiguate translation candidates. For example, translation sense disambiguation approaches (Carpuat and Wu, 2005; are proposed for phrase-based SMT systems. Meanwhile, for hierarchical phrase-based or syntax-based SMT systems, there is also much work involving rich contexts to guide rule selection Marton and Resnik, 2008; Xiong et al., 2009) . Although these methods are effective and proven successful in many SMT systems, they only leverage within-sentence contexts which are insufficient in exploring broader information. For example, the word driver often means \"the operator of a motor vehicle\" in common texts. But in the sentence \"Finally, we write the user response to the buffer, i.e., pass it to our driver\", we understand that driver means \"computer program\". In this case, people understand the meaning because of the IT topical context which goes beyond sentence-level analysis and requires more relevant knowledge. Therefore, it is important to leverage topic information to learn smarter translation models and achieve better translation performance.", |
| "cite_spans": [ |
| { |
| "start": 283, |
| "end": 305, |
| "text": "(Carpuat and Wu, 2005;", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 491, |
| "end": 515, |
| "text": "Marton and Resnik, 2008;", |
| "ref_id": "BIBREF37" |
| }, |
| { |
| "start": 516, |
| "end": 535, |
| "text": "Xiong et al., 2009)", |
| "ref_id": "BIBREF50" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Topic modeling is a useful mechanism for discovering and characterizing various semantic concepts embedded in a collection of documents. Attempts on topic-based translation modeling include topic-specific lexicon translation models (Zhao and Xing, 2006; Zhao and Xing, 2007) , topic similarity models for synchronous rules (Xiao et al., 2012) , and document-level translation with topic coherence (Xiong and Zhang, 2013) . In addition, topic-based approaches have been used in domain adaptation for SMT (Tam et al., 2007; Su et al., 2012) , where they view different topics as different domains. One typical property of these approaches in common is that they only utilize parallel data where document boundaries are explicitly given. In this way, the topic of a sentence can be inferred with document-level information using off-the-shelf topic modeling toolkits such as Latent Dirichlet Allocation (LDA) or Hidden Topic Markov Model (HTMM) . Most of them also assume that the input must be in document level. However, this situation does not always happen since there is considerable amount of parallel data which does not have document boundaries. In addition, contemporary SMT systems often works on sentence level rather than document level due to the efficiency. Although we can easily apply LDA at the sentence level, it is quite difficult to infer the topic accurately with only a few words in the sentence. This makes previous approaches inefficient when applied them in real-world commercial SMT systems. Therefore, we need to devise a systematical approach to enriching the sentence and inferring its topic more accurately.", |
| "cite_spans": [ |
| { |
| "start": 232, |
| "end": 253, |
| "text": "(Zhao and Xing, 2006;", |
| "ref_id": "BIBREF51" |
| }, |
| { |
| "start": 254, |
| "end": 274, |
| "text": "Zhao and Xing, 2007)", |
| "ref_id": "BIBREF52" |
| }, |
| { |
| "start": 323, |
| "end": 342, |
| "text": "(Xiao et al., 2012)", |
| "ref_id": "BIBREF48" |
| }, |
| { |
| "start": 397, |
| "end": 420, |
| "text": "(Xiong and Zhang, 2013)", |
| "ref_id": "BIBREF49" |
| }, |
| { |
| "start": 503, |
| "end": 521, |
| "text": "(Tam et al., 2007;", |
| "ref_id": "BIBREF45" |
| }, |
| { |
| "start": 522, |
| "end": 538, |
| "text": "Su et al., 2012)", |
| "ref_id": "BIBREF44" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, we propose a novel approach to learning topic representations for sentences. Since the information within the sentence is insufficient for topic modeling, we first enrich sentence contexts via Information Retrieval (IR) methods using content words in the sentence as queries, so that topic-related monolingual documents can be collected. These topic-related documents are utilized to learn a specific topic representation for each sentence using a neural network based approach. Neural network is an effective technique for learning different levels of data representations. The levels inferred from neural network correspond to distinct levels of concepts, where high-level representations are obtained from low-level bag-ofwords input. It is able to detect correlations among any subset of input features through non-linear transformations, which demonstrates the superiority of eliminating the effect of noisy words which are irrelevant to the topic. Our problem fits well into the neural network framework and we expect that it can further improve inferring the topic representations for sentences.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "To incorporate topic representations as translation knowledge into SMT, our neural network based approach directly optimizes similarities between the source language and target language in a compact topic space. This underlying topic space is learned from sentence-level parallel data in order to share topic information across the source and target languages as much as possible. Additionally, our model can be discriminatively trained with a large number of training instances, without expensive sampling methods such as in LDA or HTMM, thus it is more practicable and scalable. Finally, we associate the learned representation to each bilingual translation rule. Topic-related rules are selected according to distributional similarity with the source text, which helps hypotheses generation in SMT decoding. We integrate topic similarity features in the log-linear model and evaluate the performance on the NIST Chinese-to-English translation task. Experimental results demonstrate that our model significantly improves translation accuracy over a state-of-the-art baseline.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Deep learning is an active topic in recent years which has triumphed in many machine learning research areas. This technique began raising public awareness in the mid-2000s after researchers showed how a multi-layer feed-forward neural network can be effectively trained. The training procedure often involves two phases: a layerwise unsupervised pre-training phase and a supervised fine-tuning phase. For pre-training, Restricted Boltzmann Machine (RBM) , auto-encoding (Bengio et al., 2006) and sparse coding are most frequently used. Unsupervised pre-training trains the network one layer at a time and helps guide the parameters of the layer towards better regions in parameter space . Followed by finetuning in this parameter region, deep learning is able to achieve state-of-the-art performance in various research areas, including breakthrough results on the ImageNet dataset for objective recognition , significant error reduction in speech recognition , etc.", |
| "cite_spans": [ |
| { |
| "start": 471, |
| "end": 492, |
| "text": "(Bengio et al., 2006)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background: Deep Learning", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Deep learning has also been successfully applied in a variety of NLP tasks such as part-ofspeech tagging, chunking, named entity recognition, semantic role labeling , parsing (Socher et al., 2011a) , sentiment analysis (Socher et al., 2011b) , etc. Most NLP research converts a high-dimensional and sparse binary representation into a low-dimensional and real-valued representation. This low-dimensional representation is usually learned from huge amount of monolingual texts in the pre-training phase, and then fine-tuned towards task-specific criterion. Inspired by previous successful research, we first learn sentence representations using topic-related monolingual texts in the pretraining phase, and then optimize the bilingual similarity by leveraging sentence-level parallel data in the fine-tuning phase.", |
| "cite_spans": [ |
| { |
| "start": 175, |
| "end": 197, |
| "text": "(Socher et al., 2011a)", |
| "ref_id": "BIBREF42" |
| }, |
| { |
| "start": 219, |
| "end": 241, |
| "text": "(Socher et al., 2011b)", |
| "ref_id": "BIBREF43" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background: Deep Learning", |
| "sec_num": "2" |
| }, |
| { |
| "text": "In this section, we explain our neural network based topic similarity model in detail, as well as how to incorporate the topic similarity features into SMT decoding procedure. Figure 1 sketches the high-level overview which illustrates how to Figure 1 : Overview of neural network based topic similarity model. learn topic representations using sentence-level parallel data. Given a parallel sentence pair f, e , the first step is to treat f and e as queries, and use IR methods to retrieve relevant documents to enrich contextual information for them. Specifically, the ranking model we used is a Vector Space Model (VSM), where the query and document are converted into tf-idf weighted vectors. The most relevant N documents d f and d e are retrieved and converted to a high-dimensional, bag-of-words input f and e for the representation learning 1 .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 176, |
| "end": 184, |
| "text": "Figure 1", |
| "ref_id": null |
| }, |
| { |
| "start": 243, |
| "end": 251, |
| "text": "Figure 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Topic Similarity Model with Neural Network", |
| "sec_num": "3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "= ( ) = ( ) cos( , )", |
| "eq_num": "( , )" |
| } |
| ], |
| "section": "Topic Similarity Model with Neural Network", |
| "sec_num": "3" |
| }, |
| { |
| "text": "There are two phases in our neural network training process: pre-training and fine-tuning. In the pre-training phase (Section 3.1), we build two neural networks with the same structure but different parameters to learn a low-dimensional representation for sentences in two different languages. Then, in the fine-tuning phase (Section 3.2), our model directly optimizes the similarity of two lowdimensional representations, so that it highly correlates to SMT decoding. Finally, the learned representation is used to calculate similarities which are integrated as features in SMT decoding procedure (Section 3.3).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Similarity Model with Neural Network", |
| "sec_num": "3" |
| }, |
| { |
| "text": "In the pre-training phase, we leverage neural network structures to transform high-dimensional sparse vectors to low-dimensional dense vectors. The topic similarity is calculated on top of the learned dense vectors. This dense representation should preserve the information from the bag-of-words input, meanwhile alleviate data sparse problem. Therefore, we use a specially designed mechanism called auto-encoder to solve this problem. Auto-encoder (Bengio et al., 2006) is one of the basic building blocks of deep learning. Assuming that the input is a n-of-V binary vector x representing the bag-of-words (V is the vocabulary size), an auto-encoder consists of an encoding process g(x) and a decoding process h(g(x)). The objective of the auto-encoder is to minimize the reconstruction error L(h(g(x)), x). Our goal is to learn a low-dimensional vector which can preserve information from the original n-of-V vector.", |
| "cite_spans": [ |
| { |
| "start": 449, |
| "end": 470, |
| "text": "(Bengio et al., 2006)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-training using denoising auto-encoder", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "One problem with auto-encoder is that it treats all words in the same way, making no distinguishment between function words and content words. The representation learned by auto-encoders tends to be influenced by the function words, thereby it is not robust. To alleviate this problem, Vincent et al. (2008) proposed the Denoising Auto-Encoder (DAE), which aims to reconstruct a clean, \"repaired\" input from a corrupted, partially destroyed vector. This is done by corrupting the initial input x to get a partially destroyed versionx. DAE is capable of capturing the global structure of the input while ignoring the noise. In our task, for each sentence, we treat the retrieved N relevant documents as a single large document and convert it to a bag-of-words vector x in Figure 2 . With DAE, the input x is manually corrupted by applying masking noise (randomly mask 1 to 0) and gettingx. Denoising training is considered as \"filling in the blanks\" (Vincent et al., 2010) , which means the masking components can be recovered from the non-corrupted components. For example, in IT related texts, if the word driver is masked, it should be predicted through hidden units in neural networks by active signals such as \"buffer\", \"user response\", etc.", |
| "cite_spans": [ |
| { |
| "start": 286, |
| "end": 307, |
| "text": "Vincent et al. (2008)", |
| "ref_id": "BIBREF46" |
| }, |
| { |
| "start": 949, |
| "end": 971, |
| "text": "(Vincent et al., 2010)", |
| "ref_id": "BIBREF47" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 771, |
| "end": 779, |
| "text": "Figure 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Pre-training using denoising auto-encoder", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "In our case, the encoding process transforms the corrupted inputx into g(x) with two layers: a linear layer connected with a non-linear layer. Assuming that the dimension of the g(x) is L, the linear layer forms a L \u00d7 V matrix W which projects the n-of-V vector to a L-dimensional hidden layer. After the bag-of-words input has been transformed, they are fed into a subsequent layer to model the highly non-linear relations among words:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-training using denoising auto-encoder", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "z = f (Wx + b)", |
| "eq_num": "(1)" |
| } |
| ], |
| "section": "Pre-training using denoising auto-encoder", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "where z is the output of the non-linear layer, b is a", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-training using denoising auto-encoder", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "( ) \u210e( ) \u2112(\u210e , )", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-training using denoising auto-encoder", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Figure 2: Denoising auto-encoder with a bag-ofwords input.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-training using denoising auto-encoder", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "L-length bias vector. f (\u2022) is a non-linear function, where common choices include sigmoid function, hyperbolic function, \"hard\" hyperbolic function, rectifier function, etc. In this work, we use the rectifier function as our non-linear function due to its efficiency and better performance (Glorot et al., 2011) :", |
| "cite_spans": [ |
| { |
| "start": 291, |
| "end": 312, |
| "text": "(Glorot et al., 2011)", |
| "ref_id": "BIBREF27" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-training using denoising auto-encoder", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "rec(x) = x if x > 0 0 otherwise (2)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-training using denoising auto-encoder", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "The decoding process consists of a linear layer and a non-linear layer with similar network structures, but different parameters. It transforms the L-dimensional vector g(x) to a V -dimensional vector h(g(x)). To minimize reconstruction error with respect tox, we define the loss function as the L2-norm of the difference between the uncorrupted input and reconstructed input:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-training using denoising auto-encoder", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "L(h(g(x)), x) = h(g(x)) \u2212 x 2", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-training using denoising auto-encoder", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "(3)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-training using denoising auto-encoder", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Multi-layer neural networks are trained with the standard back-propagation algorithm (Rumelhart et al., 1988) . The gradient of the loss function is calculated and back-propagated to the previous layer to update its parameters. Training neural networks involves many factors such as the learning rate and the length of hidden layers. We will discuss the optimization of these parameters in Section 4.", |
| "cite_spans": [ |
| { |
| "start": 85, |
| "end": 109, |
| "text": "(Rumelhart et al., 1988)", |
| "ref_id": "BIBREF40" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-training using denoising auto-encoder", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "In the fine-tuning phase, we stack another layer on top of the two low-dimensional vectors to maximize the similarity between source and target languages. The similarity scores are integrated into the standard log-linear model for making translation decisions. Since the vectors from DAE are trained using information from monolingual training data independently, these vectors may be in-adequate to measure bilingual topic similarity due to their different topic spaces. Therefore, in this stage, parallel sentence pairs are used to help connecting the vectors from different languages because they express the same topic. In fact, the objective of fine-tuning is to discover a latent topic space which is shared by both languages as much as possible. This shared topic space is particularly useful when the SMT decoder tries to match the source texts and translation candidates in the target language.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Fine-tuning with parallel data", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Given a parallel sentence pair f, e , the DAE learns representations for f and e respectively, as z f = g(f) and z e = g(e) in Figure 1 . We then take two vectors as the input to calculate their similarity. Consequently, the whole neural network can be fine-tuned towards the supervised criteria with the help of parallel data. The similarity score of the representation pair z f , z e is defined as the cosine similarity of the two vectors:", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 127, |
| "end": 135, |
| "text": "Figure 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Fine-tuning with parallel data", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "sim(f, e) = cos(z f , z e ) = z f \u2022 z e z f z e", |
| "eq_num": "(4)" |
| } |
| ], |
| "section": "Fine-tuning with parallel data", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Since a parallel sentence pair should have the same topic, our goal is to maximize the similarity score between the source sentence and target sentence. Inspired by the contrastive estimation method (Smith and Eisner, 2005) , for each parallel sentence pair f, e as a positive instance, we select another sentence pair f , e from the training data and treat f, e as a negative instance. To make the similarity of the positive instance larger than the negative instance by some margin \u03b7, we utilize the following pairwise ranking loss: L(f, e) = max{0, \u03b7 \u2212 sim(f, e) + sim(f, e )}", |
| "cite_spans": [ |
| { |
| "start": 199, |
| "end": 223, |
| "text": "(Smith and Eisner, 2005)", |
| "ref_id": "BIBREF41" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Fine-tuning with parallel data", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "(5) where \u03b7 = 1 2 \u2212 sim(f, f ). The rationale behind this criterion is, the smaller sim(f, f ) is, the more we should penalize negative instances.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Fine-tuning with parallel data", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "To effectively train the model in this task, negative instances must be selected carefully. Since different sentences may have very similar topic distributions, we select negative instances that are dissimilar with the positive instances based on the following criteria:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Fine-tuning with parallel data", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "1. For each positive instance f, e , we select e which contains at least 30% different content words from e.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Fine-tuning with parallel data", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "2. If we cannot find such e , remove f, e from the training instances for network learning.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Fine-tuning with parallel data", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "The model minimizes the pairwise ranking loss across all training instances:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Fine-tuning with parallel data", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "L = f,e L(f, e)", |
| "eq_num": "(6)" |
| } |
| ], |
| "section": "Fine-tuning with parallel data", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "We used standard back-propagation algorithm to further fine-tune the neural network parameters W and b in Equation 1. The learned neural networks are used to obtain sentence topic representations, which will be further leveraged to infer topic representations of bilingual translation rules.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Fine-tuning with parallel data", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "We incorporate the learned topic similarity scores into the standard log-linear framework for SMT. When a synchronous rule \u03b1, \u03b3 is extracted from a sentence pair f, e , a triple instance I = ( \u03b1, \u03b3 , f, e , c) is collected for inferring the topic representation of \u03b1, \u03b3 , where c is the count of rule occurrence. Following , we give a count of one for each phrase pair occurrence and a fractional count for each hierarchical phrase pair. The topic representation of \u03b1, \u03b3 is then calculated as the weighted average:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Integration into SMT decoding", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "z \u03b1 = ( \u03b1,\u03b3 , f,e ,c)\u2208T {c \u00d7 z f } ( \u03b1,\u03b3 , f,e ,c)\u2208T {c} (7) z \u03b3 = ( \u03b1,\u03b3 , f,e ,c)\u2208T {c \u00d7 z e } ( \u03b1,\u03b3 , f,e ,c)\u2208T {c}", |
| "eq_num": "(8)" |
| } |
| ], |
| "section": "Integration into SMT decoding", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "where T denotes all instances for the rule \u03b1, \u03b3 , z \u03b1 and z \u03b3 are the source-side and target-side topic vectors respectively. By measuring the similarity between the source texts and bilingual translation rules, the SMT decoder is able to encourage topic relevant translation candidates and penalize topic irrelevant candidates. Therefore, it helps to train a smarter translation model with the embedded topic information. Given a source sentence s to be translated, we define the similarity as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Integration into SMT decoding", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "Sim(z s , z \u03b1 ) = cos(z s , z \u03b1 )", |
| "eq_num": "(9)" |
| } |
| ], |
| "section": "Integration into SMT decoding", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "Sim(z s , z \u03b3 ) = cos(z s , z \u03b3 )", |
| "eq_num": "(10)" |
| } |
| ], |
| "section": "Integration into SMT decoding", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "where z s is the topic representation of s. The similarity calculated against z \u03b1 or z \u03b3 denotes the source-to-source or the source-to-target similarity.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Integration into SMT decoding", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "We also consider the topic sensitivity estimation since general rules have flatter distributions while topic-specific rules have sharper distributions. A standard entropy metric is used to measure the sensitivity of the source-side of \u03b1, \u03b3 as:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Integration into SMT decoding", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Sen(\u03b1) = \u2212 |z\u03b1| i=1 z \u03b1i \u00d7 log z \u03b1i (11)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Integration into SMT decoding", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "where z \u03b1i is a component in the vector z \u03b1 . The target-side sensitivity Sen(\u03b3) can be calculated in a similar way. The larger the sensitivity is, the more topic-specific the rule manifests. In addition to traditional SMT features, we add new topic-related features into the standard loglinear framework. For the SMT system, the best translation candidate\u00ea is given by:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Integration into SMT decoding", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "e = arg max e P (e|f )", |
| "eq_num": "(12)" |
| } |
| ], |
| "section": "Integration into SMT decoding", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "where the translation probability is given by:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Integration into SMT decoding", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "P (e|f ) \u221d i w i \u2022 log \u03c6 i (f, e) = j w j \u2022 log \u03c6 j (f, e) Standard + k w k \u2022 log \u03c6 k (f, e) Topic related", |
| "eq_num": "(13)" |
| } |
| ], |
| "section": "Integration into SMT decoding", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "where \u03c6 j (f, e) is the standard feature function and w j is the corresponding feature weight. \u03c6 k (f, e) is the topic-related feature function and w k is the feature weight. The detailed feature description is as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Integration into SMT decoding", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Standard features: Translation model, including translation probabilities and lexical weights for both directions (4 features), 5-gram language model (1 feature), word count (1 feature), phrase count (1 feature), NULL penalty (1 feature), number of hierarchical rules used (1 feature).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Integration into SMT decoding", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Topic-related features: rule similarity scores (2 features), rule sensitivity scores (2 features).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Integration into SMT decoding", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "We evaluate the performance of our neural network based topic similarity model on a Chineseto-English machine translation task. In neural network training, a large number of monolingual documents are collected in both source and target languages. The documents are mainly from two domains: news and weblog. We use Chinese and English Gigaword corpus (Version 5) which are mainly from news domain. In addition, we also collect weblog documents with a variety of topics from the web. The total data statistics are presented in Table 1 . These documents are built in the format of inverted index using Lucene 2 , which can be efficiently retrieved by the parallel sentence pairs. The most relevant N documents are collected, where we experiment with N = {1, 5, 10, 20, 50}. Table 1 : Statistics of monolingual data, in numbers of documents and words (main content). \"M\" refers to million and \"B\" refers to billion.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 525, |
| "end": 532, |
| "text": "Table 1", |
| "ref_id": null |
| }, |
| { |
| "start": 771, |
| "end": 778, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Setup", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "We implement a distributed framework to speed up the training process of neural networks. The network is learned with mini-batch asynchronous gradient descent with the adaptive learning rate procedure called AdaGrad . We use 32 model replicas in each iteration during the training. The model parameters are averaged after each iteration and sent to each replica for the next iteration. The vocabulary size for the input layer is 100,000, and we choose different lengths for the hidden layer as L = {100, 300, 600, 1000} in the experiments. In the pre-training phase, all parallel data is fed into two neural networks respectively for DAE training, where network parameters W and b are randomly initialized. In the fine-tuning phase, for each parallel sentence pair, we randomly select other ten sentence pairs which satisfy the criterion as negative instances. These training instances are leveraged to optimize the similarity of two vectors.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Setup", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "In SMT training, an in-house hierarchical phrase-based SMT decoder is implemented for our experiments. The CKY decoding algorithm is used and cube pruning is performed with the same default parameter settings as in . The parallel data we use is released by LDC 3 . In total, the datasets contain nearly 1.1 million sentence pairs. Translation models are trained over the parallel data that is automatically word-aligned using GIZA++ in both directions, and the diaggrow-final heuristic is used to refine symmetric word alignment. An in-house language modeling toolkit is used to train the 5-gram language model with modified Kneser-Ney smoothing . The English monolingual data used for language modeling is the same as in Table 1 . The NIST 2003 dataset is the development data. The testing data consists of NIST 2004 NIST , 2005 NIST , 2006 NIST and 2008 datasets. The evaluation metric for the overall translation quality is caseinsensitive BLEU4 (Papineni et al., 2002) . The reported BLEU scores are averaged over 5 times of running MERT (Och, 2003) . A statistical significance test is performed using the bootstrap resampling method .", |
| "cite_spans": [ |
| { |
| "start": 809, |
| "end": 818, |
| "text": "NIST 2004", |
| "ref_id": null |
| }, |
| { |
| "start": 819, |
| "end": 830, |
| "text": "NIST , 2005", |
| "ref_id": null |
| }, |
| { |
| "start": 831, |
| "end": 842, |
| "text": "NIST , 2006", |
| "ref_id": null |
| }, |
| { |
| "start": 843, |
| "end": 856, |
| "text": "NIST and 2008", |
| "ref_id": null |
| }, |
| { |
| "start": 950, |
| "end": 973, |
| "text": "(Papineni et al., 2002)", |
| "ref_id": "BIBREF39" |
| }, |
| { |
| "start": 1043, |
| "end": 1054, |
| "text": "(Och, 2003)", |
| "ref_id": "BIBREF38" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 722, |
| "end": 730, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Setup", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "The baseline is a re-implementation of the Hiero system . The phrase pairs that appear only once in the parallel data are discarded because most of them are noisy. We also use the fix-discount method in Foster et al. (2006) for phrase table smoothing. This implementation makes the system perform much better and the translation model size is much smaller.", |
| "cite_spans": [ |
| { |
| "start": 203, |
| "end": 223, |
| "text": "Foster et al. (2006)", |
| "ref_id": "BIBREF26" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baseline", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "We compare our method with the LDA-based approach proposed by Xiao et al. (2012) . In (Xiao et al., 2012) , the topic of each sentence pair is exactly the same as the document it belongs to. Since some of our parallel data does not have documentlevel information, we rely on the IR method to retrieve the most relevant document and simulate this approach. The PLDA toolkit (Liu et al., 2011) is used to infer topic distributions, which takes 34.5 hours to finish.", |
| "cite_spans": [ |
| { |
| "start": 62, |
| "end": 80, |
| "text": "Xiao et al. (2012)", |
| "ref_id": "BIBREF48" |
| }, |
| { |
| "start": 86, |
| "end": 105, |
| "text": "(Xiao et al., 2012)", |
| "ref_id": "BIBREF48" |
| }, |
| { |
| "start": 373, |
| "end": 391, |
| "text": "(Liu et al., 2011)", |
| "ref_id": "BIBREF36" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baseline", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "We illustrate the relationship among translation accuracy (BLEU), the number of retrieved documents (N ) and the length of hidden layers (L) on different testing datasets. The results are shown in Figure 3 . The best translation accuracy is achieved when N =10 for most settings. This confirms that enriching the source text with topic-related documents is very useful in determining topic representations, thereby help to guide the synchronous rule selection. However, we find that as N becomes larger in the experiments, e.g. N =50, the translation accuracy drops drastically. As more documents are retrieved, less relevant information Figure 3 : End-to-end translation results (BLEU%) using all standard and topic-related features, with different settings on the number of retrieved documents N and the length of hidden layers L.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 197, |
| "end": 205, |
| "text": "Figure 3", |
| "ref_id": null |
| }, |
| { |
| "start": 638, |
| "end": 646, |
| "text": "Figure 3", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Effect of retrieved documents and length of hidden layers", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "is also used to train the neural networks. Irrelevant documents bring so many unrelated topic words hence degrade neural network learning performance.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Effect of retrieved documents and length of hidden layers", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Another important factor is the length of hidden layers L in the network. In deep learning, this parameter is often empirically tuned with human efforts. As shown in Figure 3 , the translation accuracy is better when L is relatively small. Actually, there is no obvious distinction of the performance when L is less than 600. However, when L equals 1,000, the translation accuracy is inferior to other settings. The main reason is that parameters in the neural networks are too many to be effectively trained. As we know when L=1000, there are a total of 100, 000 \u00d7 1, 000 parameters between the linear and non-linear layers in the network. Limited training data prevents the model from getting close to the global optimum. Therefore, the model is likely to fall in local optima and lead to unacceptable representations.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 166, |
| "end": 174, |
| "text": "Figure 3", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Effect of retrieved documents and length of hidden layers", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "We evaluate the performance of adding new topicrelated features to the log-linear model and compare the translation accuracy with the method in (Xiao et al., 2012) . To make different methods comparable, we set the dimension of topic representation as 100 for all settings. This takes 10 hours in pre-training phase and 22 hours in finetuning phase. Table 2 shows how the accuracy is improved with more features added. The results confirm that topic information is indispensable for SMT since both (Xiao et al., 2012) and our neural network based method significantly outperforms the baseline system. Our method improves 0.86 BLEU points at most and 0.76 BLEU points on average over the baseline. We observe that sourceside similarity is more effective than target-side similarity, but their contributions are cumulative. This proves that bilingually induced topic representation with neural network helps the SMT system disambiguate translation candidates. Furthermore, rule sensitivity features improve SMT performance compared with only using similarity features. Because topic-specific rules usually have a larger sensitivity score, they can beat general rules when they obtain the same similarity score against the input sentence. Finally, when all new features are integrated, the performance is the best, preforming substantially better than (Xiao et al., 2012) with 0.39 BLEU points on average.", |
| "cite_spans": [ |
| { |
| "start": 144, |
| "end": 163, |
| "text": "(Xiao et al., 2012)", |
| "ref_id": "BIBREF48" |
| }, |
| { |
| "start": 498, |
| "end": 517, |
| "text": "(Xiao et al., 2012)", |
| "ref_id": "BIBREF48" |
| }, |
| { |
| "start": 1349, |
| "end": 1368, |
| "text": "(Xiao et al., 2012)", |
| "ref_id": "BIBREF48" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 350, |
| "end": 357, |
| "text": "Table 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Effect of topic related features", |
| "sec_num": "4.4" |
| }, |
| { |
| "text": "It is worth mentioning that the performance of (Xiao et al., 2012) is similar to the settings with N =1 and L=100 in Figure 3 . This is not simply coincidence since we can interpret their approach as a special case in our neural network method: when a parallel sentence pair has NIST 2004 NIST 2005 NIST 2006 NIST 2008 Table 2 : Effectiveness of different features in BLEU% (p < 0.05), with N =10 and L=100. \"Sim\" denotes the rule similarity feature and \"Sen\" denotes rule sensitivity feature. \"Src\" and \"Trg\" means utilizing source-side/target-side rule topic vectors to calculate similarity or sensitivity, respectively. The \"Average\" setting is the averaged result of four datasets.", |
| "cite_spans": [ |
| { |
| "start": 47, |
| "end": 66, |
| "text": "(Xiao et al., 2012)", |
| "ref_id": "BIBREF48" |
| }, |
| { |
| "start": 279, |
| "end": 288, |
| "text": "NIST 2004", |
| "ref_id": null |
| }, |
| { |
| "start": 289, |
| "end": 298, |
| "text": "NIST 2005", |
| "ref_id": null |
| }, |
| { |
| "start": 299, |
| "end": 308, |
| "text": "NIST 2006", |
| "ref_id": null |
| }, |
| { |
| "start": 309, |
| "end": 318, |
| "text": "NIST 2008", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 117, |
| "end": 125, |
| "text": "Figure 3", |
| "ref_id": null |
| }, |
| { |
| "start": 319, |
| "end": 326, |
| "text": "Table 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Effect of topic related features", |
| "sec_num": "4.4" |
| }, |
| { |
| "text": "document-level information, that document will be retrieved for training; otherwise, the most relevant document will be retrieved from the monolingual data. Therefore, our method can be viewed as a more general framework than previous LDAbased approaches.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Settings", |
| "sec_num": null |
| }, |
| { |
| "text": "In this section, we give a case study to explain why our method works. An example of translation rule disambiguation for a sentence from the NIST 2005 dataset is shown in Figure 4 . We find that the topic of this sentence is about \"rescue after a natural disaster\". Under this topic, the Chinese rule \"\u53d1\u9001 X\" should be translated to \"deliver X\" or \"distribute X\". However, the baseline system prefers \"send X\" rather than those two candidates. Although the translation probability of \"send X\" is much higher, it is inappropriate in this context since it is usually used in IT texts. For example, \u53d1\u9001\u90ae\u4ef6, send emails , \u53d1\u9001\u4fe1\u606f, send messages and \u53d1\u9001\u6570\u636e, send data . In contrast, with our neural network based approach, the learned topic distributions of \"deliver X\" or \"distribute X\" are more similar with the input sentence than \"send X\", which is shown in Figure 4 . The similarity scores indicate that \"deliver X\" and \"distribute X\" are more appropriate to translate the sentence. Therefore, adding topic-related features is able to keep the topic consistency and substantially improve the translation accuracy.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 171, |
| "end": 179, |
| "text": "Figure 4", |
| "ref_id": null |
| }, |
| { |
| "start": 849, |
| "end": 857, |
| "text": "Figure 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "4.5" |
| }, |
| { |
| "text": "Topic modeling was first leveraged to improve SMT performance in (Zhao and Xing, 2006; Zhao and Xing, 2007) . They proposed a bilingual topical admixture approach for word alignment and assumed that each word-pair follows a topic-specific model. They reported extensive empirical analysis and improved word alignment accuracy as well as translation quality. Following this work, (Xiao et al., 2012) extended topicspecific lexicon translation models to hierarchical phrase-based translation models, where the topic information of synchronous rules was directly inferred with the help of document-level information. Experiments show that their approach not only achieved better translation performance but also provided a faster decoding speed compared with previous lexicon-based LDA methods.", |
| "cite_spans": [ |
| { |
| "start": 65, |
| "end": 86, |
| "text": "(Zhao and Xing, 2006;", |
| "ref_id": "BIBREF51" |
| }, |
| { |
| "start": 87, |
| "end": 107, |
| "text": "Zhao and Xing, 2007)", |
| "ref_id": "BIBREF52" |
| }, |
| { |
| "start": 379, |
| "end": 398, |
| "text": "(Xiao et al., 2012)", |
| "ref_id": "BIBREF48" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Another direction of approaches leveraged topic modeling techniques for domain adaptation. Tam et al. (2007) used bilingual LSA to learn latent topic distributions across different languages and enforce one-to-one topic correspondence during model training. They incorporated the bilingual topic information into language model adaptation and lexicon translation model adaptation, achieving significant improvements in the large-scale evaluation. (Su et al., 2012) investigated the relationship between out-of-domain bilingual data and in-domain monolingual data via topic mapping using HTMM methods. They estimated phrasetopic distributions in translation model adaptation and generated better translation quality. Recently, proposed using vector space model for adaptation where genre resemblance is leveraged to improve translation accuracy. We also investigated multi-domain adaptation where explicit topic information is used to train domain specific models (Cui et al., 2013) .", |
| "cite_spans": [ |
| { |
| "start": 91, |
| "end": 108, |
| "text": "Tam et al. (2007)", |
| "ref_id": "BIBREF45" |
| }, |
| { |
| "start": 447, |
| "end": 464, |
| "text": "(Su et al., 2012)", |
| "ref_id": "BIBREF44" |
| }, |
| { |
| "start": 963, |
| "end": 981, |
| "text": "(Cui et al., 2013)", |
| "ref_id": "BIBREF23" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Generally, most previous research has leveraged conventional topic modeling techniques such as LDA or HTMM. In our work, a novel neural network based approach is proposed to infer topic representations for parallel data. The advantage of", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "(1) the united nations children's fund has also begun delivering basic medical kits (2) the unicef has also started to distribute basic medical kits (3) the united nations children's fund has also begun distributing basic medical kits (4) the united nations children's fund has begun delivering basic medical kits Baseline the united nations children's fund began to send basic medical kits Ours the united nations children's fund has begun to distribute basic medical kits fectiveness of different features in BLEU% (p < 0.05), with N = 10 and L = 100. \"Sim\" rule similarity feature and \"Sen\" denotes rule sensitivity feature. \"Src\" and \"Trg\" means urce-side/target-side rule topic vectors to calculate similarity or sensitivity, respectively. The setting is the averaged results of four datasets. only using similarity features. Because fic rules usually have a larger sensitivhey can beat general rules when they ame similarity score against the input Finally, when all new features are ine performance is the best, preforming y better than (Xiao et al., 2012) with points on average. resting observation is, the performance al., 2012) is quite similar to the set-= 1 and L = 100 in Figure 3 . This ly coincidence since we can interpret ach as a special case in our neural netod. When a parallel sentence pair has evel information, that document will for training. Otherwise, the most simnt will be obtained from the monolin-Our method can be viewed as a more mework than previous LDA-based apssion riments, us LDA-based method, if a document ns M sentences, all M sentences will ame topic distribution of Doc. Alerent sentences may express slightly plications and the topic will change, tional LDA-based approach does not ic transition into consideration. In conpproach directly learns the topic repwith an abundancy of related docudditional to the original document from sentence is extracted, the IR method es other relevant documents which proementary topic information. Therefore, epresentations learned are more finethus more accurate. ", |
| "cite_spans": [ |
| { |
| "start": 1044, |
| "end": 1063, |
| "text": "(Xiao et al., 2012)", |
| "ref_id": "BIBREF48" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 1186, |
| "end": 1194, |
| "text": "Figure 3", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Ref", |
| "sec_num": null |
| }, |
| { |
| "text": "Rules P (\u03b3|\u03b1) Sim(z s , z \u03b1 ) \u53d1\u9001 X,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ref", |
| "sec_num": null |
| }, |
| { |
| "text": "Topic modeling was first leveraged to improve SMT performance in (Zhao and Xing, 2006; Zhao and Xing, 2007) . They proposed a bilingual topical admixture approach for word alignment and assumed that each word-pair follows a topicspecific model. They reported extensive empirical analysis and improved word alignment accuracy as well as translation quality. Following this work, (Xiao et al., 2012) extended topicspecific lexicon translation models to hierarchical phrase-based translation models, where the topic information of synchronous rules was directly inferred with the help of document-level information. Experiments show that their approach not only achieved better translation performance but also provided a faster decoding speed compared with previous lexicon-based methods. Another direction of approaches leveraged topic modeling techniques for domain adaptation. Tam et al. (2007) used bilingual LSA to learn latent topic distributions across different languages and enforce one-to-one topic correspondence during model training. They incorporated the bilingual topic information into language model adaptation and lexicon translation model adaptation, achieving significant improvements in the large-scale evaluation. (Su et al., 2012) investigated the relationship between out-of-domain bilingual data and in-domain monolingual data via topic mapping us- Figure 4 : An example from the NIST 2005 dataset. We illustrate the normalized topic representations of the source sentence and three ambiguous synchronous rules. Details are explained in Section 4.5.", |
| "cite_spans": [ |
| { |
| "start": 65, |
| "end": 86, |
| "text": "(Zhao and Xing, 2006;", |
| "ref_id": "BIBREF51" |
| }, |
| { |
| "start": 87, |
| "end": 107, |
| "text": "Zhao and Xing, 2007)", |
| "ref_id": "BIBREF52" |
| }, |
| { |
| "start": 378, |
| "end": 397, |
| "text": "(Xiao et al., 2012)", |
| "ref_id": "BIBREF48" |
| }, |
| { |
| "start": 878, |
| "end": 895, |
| "text": "Tam et al. (2007)", |
| "ref_id": "BIBREF45" |
| }, |
| { |
| "start": 1234, |
| "end": 1251, |
| "text": "(Su et al., 2012)", |
| "ref_id": "BIBREF44" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 1372, |
| "end": 1380, |
| "text": "Figure 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "our method is that it is applicable to both sentencelevel and document-level SMT, since we do not place any restrictions on the input. In addition, our method directly maximizes the similarity between parallel sentence pairs, which is ideal for SMT decoding. Compared to document-level topic modeling which uses the topic of a document for all sentences within the document (Xiao et al., 2012) , our contributions are:", |
| "cite_spans": [ |
| { |
| "start": 374, |
| "end": 393, |
| "text": "(Xiao et al., 2012)", |
| "ref_id": "BIBREF48" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "\u2022 We proposed a more general approach to leveraging topic information for SMT by using IR methods to get a collection of related documents, regardless of whether or not document boundaries are explicitly given.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "\u2022 We used neural networks to learn topic representations more accurately, with more practicable and scalable modeling techniques.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "\u2022 We directly optimized bilingual topic similarity in the deep learning framework with the help of sentence-level parallel data, so that the learned representation could be easily used in SMT decoding procedure.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "In this paper, we propose a neural network based approach to learning bilingual topic representation for SMT. We enrich contexts of parallel sentence pairs with topic related monolingual data and obtain a set of documents to represent sentences. These documents are converted to a bagof-words input and fed into neural networks. The learned low-dimensional vector is used to obtain the topic representations of synchronous rules. In SMT decoding, appropriate rules are selected to best match source texts according to their similarity in the topic space. Experimental results show that our approach is promising for SMT systems to learn a better translation model. It is a significant improvement over the state-of-the-art Hiero system, as well as a conventional LDA-based method. In the future research, we will extend our neural network methods to address document-level translation, where topic transition between sentences is a crucial problem to be solved. Since the translation of the current sentence is usually influenced by the topic of previous sentences, we plan to leverage recurrent neural networks to model this phenomenon, where the history translation information is naturally combined in the model.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion and Future Work", |
| "sec_num": "6" |
| }, |
| { |
| "text": "We use f and e to denote the n-of-V vector converted from the retrieved documents.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://lucene.apache.org/ 3 LDC2003E14, LDC2002E18, LDC2003E07, LDC2005T06, LDC2005T10, LDC2005E83, LDC2006E34, LDC2006E85, LDC2006E92, LDC2006E26, LDC2007T09", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "The acknowledgments should go immediately before the references. Do not number the acknowledgments section. Do not include this section when submitting your paper for review.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| }, |
| { |
| "text": "We are grateful to the anonymous reviewers for their insightful comments. We also thank Fei Huang (BBN), Nan Yang, Yajuan Duan, Hong Sun and Duyu Tang for the helpful discussions. This work is supported by the National Natural Science Foundation of China (Granted No. 61272384) ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Greedy layer-wise training of deep networks", |
| "authors": [ |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| }, |
| { |
| "first": "Pascal", |
| "middle": [], |
| "last": "Lamblin", |
| "suffix": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Popovici", |
| "suffix": "" |
| }, |
| { |
| "first": "Hugo", |
| "middle": [], |
| "last": "Larochelle", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Advances in Neural Information Processing Systems 19", |
| "volume": "", |
| "issue": "", |
| "pages": "153--160", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yoshua Bengio, Pascal Lamblin, Dan Popovici, and Hugo Larochelle. 2006. Greedy layer-wise train- ing of deep networks. In B. Sch\u00f6lkopf, J. Platt, and T. Hoffman, editors, Advances in Neural Informa- tion Processing Systems 19, pages 153-160. MIT Press, Cambridge, MA.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Learning deep architectures for ai", |
| "authors": [ |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Trends Mach. Learn", |
| "volume": "2", |
| "issue": "1", |
| "pages": "1--127", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yoshua Bengio. 2009. Learning deep architectures for ai. Found. Trends Mach. Learn., 2(1):1-127, Jan- uary.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Latent dirichlet allocation", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [ |
| "M" |
| ], |
| "last": "Blei", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [ |
| "Y" |
| ], |
| "last": "Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [ |
| "I" |
| ], |
| "last": "Jordan", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "J. Mach. Learn. Res", |
| "volume": "3", |
| "issue": "", |
| "pages": "993--1022", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David M. Blei, Andrew Y. Ng, and Michael I. Jordan. 2003. Latent dirichlet allocation. J. Mach. Learn. Res., 3:993-1022, March.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Contextdependent phrasal translation lexicons for statistical machine translation", |
| "authors": [ |
| { |
| "first": "Marine", |
| "middle": [], |
| "last": "Carpuat", |
| "suffix": "" |
| }, |
| { |
| "first": "Dekai", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of Machine Translation Summit XI", |
| "volume": "", |
| "issue": "", |
| "pages": "73--80", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marine Carpuat and Dekai Wu. 2007. Context- dependent phrasal translation lexicons for statistical machine translation. Proceedings of Machine Trans- lation Summit XI, pages 73-80.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Hierarchical phrase-based translation", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Chiang", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Computational Linguistics", |
| "volume": "33", |
| "issue": "2", |
| "pages": "201--228", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David Chiang. 2007. Hierarchical phrase-based trans- lation. Computational Linguistics, 33(2):201-228.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Natural language processing (almost) from scratch", |
| "authors": [ |
| { |
| "first": "Ronan", |
| "middle": [], |
| "last": "Collobert", |
| "suffix": "" |
| }, |
| { |
| "first": "Jason", |
| "middle": [], |
| "last": "Weston", |
| "suffix": "" |
| }, |
| { |
| "first": "L\u00e9on", |
| "middle": [], |
| "last": "Bottou", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Karlen", |
| "suffix": "" |
| }, |
| { |
| "first": "Koray", |
| "middle": [], |
| "last": "Kavukcuoglu", |
| "suffix": "" |
| }, |
| { |
| "first": "Pavel", |
| "middle": [], |
| "last": "Kuksa", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "J. Mach. Learn. Res", |
| "volume": "12", |
| "issue": "", |
| "pages": "2493--2537", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ronan Collobert, Jason Weston, L\u00e9on Bottou, Michael Karlen, Koray Kavukcuoglu, and Pavel Kuksa. 2011. Natural language processing (almost) from scratch. J. Mach. Learn. Res., 12:2493-2537, November.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Context-dependent pre-trained deep neural networks for large-vocabulary speech recognition", |
| "authors": [ |
| { |
| "first": "G", |
| "middle": [ |
| "E" |
| ], |
| "last": "Dahl", |
| "suffix": "" |
| }, |
| { |
| "first": "Dong", |
| "middle": [], |
| "last": "Yu", |
| "suffix": "" |
| }, |
| { |
| "first": "Li", |
| "middle": [], |
| "last": "Deng", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Acero", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Trans. Audio, Speech and Lang. Proc", |
| "volume": "20", |
| "issue": "1", |
| "pages": "30--42", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "G. E. Dahl, Dong Yu, Li Deng, and A. Acero. 2012. Context-dependent pre-trained deep neural networks for large-vocabulary speech recognition. Trans. Au- dio, Speech and Lang. Proc., 20(1):30-42, January.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Hidden topic markov models", |
| "authors": [ |
| { |
| "first": "Amit", |
| "middle": [], |
| "last": "Gruber", |
| "suffix": "" |
| }, |
| { |
| "first": "Michal", |
| "middle": [], |
| "last": "Rosen-Zvi", |
| "suffix": "" |
| }, |
| { |
| "first": "Yair", |
| "middle": [], |
| "last": "Weiss", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of Artificial Intelligence and Statistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Amit Gruber, Michal Rosen-zvi, and Yair Weiss. 2007. Hidden topic markov models. In In Proceedings of Artificial Intelligence and Statistics.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Improving statistical machine translation using lexicalized rule selection", |
| "authors": [ |
| { |
| "first": "Zhongjun", |
| "middle": [], |
| "last": "He", |
| "suffix": "" |
| }, |
| { |
| "first": "Qun", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Shouxun", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of the 22nd International Conference on Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "321--328", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zhongjun He, Qun Liu, and Shouxun Lin. 2008. Im- proving statistical machine translation using lexical- ized rule selection. In Proceedings of the 22nd In- ternational Conference on Computational Linguis- tics (Coling 2008), pages 321-328, Manchester, UK, August. Coling 2008 Organizing Committee.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "A fast learning algorithm for deep belief nets", |
| "authors": [ |
| { |
| "first": "Geoffrey", |
| "middle": [ |
| "E" |
| ], |
| "last": "Hinton", |
| "suffix": "" |
| }, |
| { |
| "first": "Simon", |
| "middle": [], |
| "last": "Osindero", |
| "suffix": "" |
| }, |
| { |
| "first": "Yee-Whye", |
| "middle": [], |
| "last": "Teh", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Neural Comput", |
| "volume": "18", |
| "issue": "7", |
| "pages": "1527--1554", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Geoffrey E. Hinton, Simon Osindero, and Yee-Whye Teh. 2006. A fast learning algorithm for deep belief nets. Neural Comput., 18(7):1527-1554, July.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Improved backing-off for m-gram language modeling", |
| "authors": [ |
| { |
| "first": "Reinhard", |
| "middle": [], |
| "last": "Kneser", |
| "suffix": "" |
| }, |
| { |
| "first": "Hermann", |
| "middle": [], |
| "last": "Ney", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "Acoustics, Speech, and Signal Processing", |
| "volume": "1", |
| "issue": "", |
| "pages": "181--184", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Reinhard Kneser and Hermann Ney. 1995. Im- proved backing-off for m-gram language modeling. In Acoustics, Speech, and Signal Processing, 1995. ICASSP-95., 1995 International Conference on, vol- ume 1, pages 181-184. IEEE.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Statistical significance tests for machine translation evaluation", |
| "authors": [ |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of EMNLP 2004", |
| "volume": "", |
| "issue": "", |
| "pages": "388--395", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Philipp Koehn. 2004. Statistical significance tests for machine translation evaluation. In Dekang Lin and Dekai Wu, editors, Proceedings of EMNLP 2004, pages 388-395, Barcelona, Spain, July. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Imagenet classification with deep convolutional neural networks", |
| "authors": [ |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Krizhevsky", |
| "suffix": "" |
| }, |
| { |
| "first": "Ilya", |
| "middle": [], |
| "last": "Sutskever", |
| "suffix": "" |
| }, |
| { |
| "first": "Geoff", |
| "middle": [], |
| "last": "Hinton", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Advances in Neural Information Processing Systems", |
| "volume": "25", |
| "issue": "", |
| "pages": "1106--1114", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alex Krizhevsky, Ilya Sutskever, and Geoff Hinton. 2012. Imagenet classification with deep convolu- tional neural networks. In P. Bartlett, F.C.N. Pereira, C.J.C. Burges, L. Bottou, and K.Q. Weinberger, ed- itors, Advances in Neural Information Processing Systems 25, pages 1106-1114.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Efficient sparse coding algorithms", |
| "authors": [ |
| { |
| "first": "Honglak", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexis", |
| "middle": [], |
| "last": "Battle", |
| "suffix": "" |
| }, |
| { |
| "first": "Rajat", |
| "middle": [], |
| "last": "Raina", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [ |
| "Y" |
| ], |
| "last": "Ng", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Advances in Neural Information Processing Systems 19", |
| "volume": "", |
| "issue": "", |
| "pages": "801--808", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Honglak Lee, Alexis Battle, Rajat Raina, and An- drew Y. Ng. 2006. Efficient sparse coding algo- rithms. In B. Sch\u00f6lkopf, J. Platt, and T. Hoffman, editors, Advances in Neural Information Processing Systems 19, pages 801-808. MIT Press, Cambridge, MA.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Maximum entropy based rule selection model for syntax-based statistical machine translation", |
| "authors": [ |
| { |
| "first": "Qun", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Zhongjun", |
| "middle": [], |
| "last": "He", |
| "suffix": "" |
| }, |
| { |
| "first": "Yang", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Shouxun", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "153--160", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Qun Liu, Zhongjun He, Yang Liu, and Shouxun Lin. 2008. Maximum entropy based rule selection model for syntax-based statistical machine translation. In Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing, pages References Yoshua Bengio, Pascal Lamblin, Dan Popovici, and Hugo Larochelle. 2006. Greedy layer-wise train- ing of deep networks. In B. Sch\u00f6lkopf, J. Platt, and T. Hoffman, editors, Advances in Neural Informa- tion Processing Systems 19, pages 153-160. MIT Press, Cambridge, MA.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Learning deep architectures for ai", |
| "authors": [ |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Trends Mach. Learn", |
| "volume": "2", |
| "issue": "1", |
| "pages": "1--127", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yoshua Bengio. 2009. Learning deep architectures for ai. Found. Trends Mach. Learn., 2(1):1-127, Jan- uary.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Latent dirichlet allocation", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [ |
| "M" |
| ], |
| "last": "Blei", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [ |
| "Y" |
| ], |
| "last": "Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [ |
| "I" |
| ], |
| "last": "Jordan", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "J. Mach. Learn. Res", |
| "volume": "3", |
| "issue": "", |
| "pages": "993--1022", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David M. Blei, Andrew Y. Ng, and Michael I. Jordan. 2003. Latent dirichlet allocation. J. Mach. Learn. Res., 3:993-1022, March.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Word sense disambiguation vs. statistical machine translation", |
| "authors": [ |
| { |
| "first": "Marine", |
| "middle": [], |
| "last": "Carpuat", |
| "suffix": "" |
| }, |
| { |
| "first": "Dekai", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics (ACL'05)", |
| "volume": "", |
| "issue": "", |
| "pages": "387--394", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marine Carpuat and Dekai Wu. 2005. Word sense dis- ambiguation vs. statistical machine translation. In Proceedings of the 43rd Annual Meeting of the As- sociation for Computational Linguistics (ACL'05), pages 387-394, Ann Arbor, Michigan, June. Asso- ciation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Contextdependent phrasal translation lexicons for statistical machine translation", |
| "authors": [ |
| { |
| "first": "Marine", |
| "middle": [], |
| "last": "Carpuat", |
| "suffix": "" |
| }, |
| { |
| "first": "Dekai", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of Machine Translation Summit XI", |
| "volume": "", |
| "issue": "", |
| "pages": "73--80", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marine Carpuat and Dekai Wu. 2007. Context- dependent phrasal translation lexicons for statistical machine translation. Proceedings of Machine Trans- lation Summit XI, pages 73-80.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Vector space model for adaptation in statistical machine translation", |
| "authors": [ |
| { |
| "first": "Boxing", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Roland", |
| "middle": [], |
| "last": "Kuhn", |
| "suffix": "" |
| }, |
| { |
| "first": "George", |
| "middle": [], |
| "last": "Foster", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "1285--1293", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Boxing Chen, Roland Kuhn, and George Foster. 2013. Vector space model for adaptation in statistical ma- chine translation. In Proceedings of the 51st An- nual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1285- 1293, Sofia, Bulgaria, August. Association for Com- putational Linguistics.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Hierarchical phrase-based translation", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Chiang", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Computational Linguistics", |
| "volume": "33", |
| "issue": "2", |
| "pages": "201--228", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David Chiang. 2007. Hierarchical phrase-based trans- lation. Computational Linguistics, 33(2):201-228.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Natural language processing (almost) from scratch", |
| "authors": [ |
| { |
| "first": "Ronan", |
| "middle": [], |
| "last": "Collobert", |
| "suffix": "" |
| }, |
| { |
| "first": "Jason", |
| "middle": [], |
| "last": "Weston", |
| "suffix": "" |
| }, |
| { |
| "first": "L\u00e9on", |
| "middle": [], |
| "last": "Bottou", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Karlen", |
| "suffix": "" |
| }, |
| { |
| "first": "Koray", |
| "middle": [], |
| "last": "Kavukcuoglu", |
| "suffix": "" |
| }, |
| { |
| "first": "Pavel", |
| "middle": [], |
| "last": "Kuksa", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "J. Mach. Learn. Res", |
| "volume": "12", |
| "issue": "", |
| "pages": "2493--2537", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ronan Collobert, Jason Weston, L\u00e9on Bottou, Michael Karlen, Koray Kavukcuoglu, and Pavel Kuksa. 2011. Natural language processing (almost) from scratch. J. Mach. Learn. Res., 12:2493-2537, November.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Multi-domain adaptation for SMT using multi-task learning", |
| "authors": [ |
| { |
| "first": "Lei", |
| "middle": [], |
| "last": "Cui", |
| "suffix": "" |
| }, |
| { |
| "first": "Xilun", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Dongdong", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Shujie", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Mu", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "1055--1065", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lei Cui, Xilun Chen, Dongdong Zhang, Shujie Liu, Mu Li, and Ming Zhou. 2013. Multi-domain adap- tation for SMT using multi-task learning. In Pro- ceedings of the 2013 Conference on Empirical Meth- ods in Natural Language Processing, pages 1055- 1065, Seattle, Washington, USA, October. Associa- tion for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Context-dependent pre-trained deep neural networks for large-vocabulary speech recognition", |
| "authors": [ |
| { |
| "first": "George", |
| "middle": [ |
| "E" |
| ], |
| "last": "Dahl", |
| "suffix": "" |
| }, |
| { |
| "first": "Dong", |
| "middle": [], |
| "last": "Yu", |
| "suffix": "" |
| }, |
| { |
| "first": "Li", |
| "middle": [], |
| "last": "Deng", |
| "suffix": "" |
| }, |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Acero", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "IEEE Transactions on Audio, Speech and Language Processing", |
| "volume": "20", |
| "issue": "1", |
| "pages": "30--42", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "George E. Dahl, Dong Yu, Li Deng, and Alex Acero. 2012. Context-dependent pre-trained deep neural networks for large-vocabulary speech recognition. IEEE Transactions on Audio, Speech and Language Processing, 20(1):30-42, January.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "Adaptive subgradient methods for online learning and stochastic optimization", |
| "authors": [ |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Duchi", |
| "suffix": "" |
| }, |
| { |
| "first": "Elad", |
| "middle": [], |
| "last": "Hazan", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoram", |
| "middle": [], |
| "last": "Singer", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "J. Mach. Learn. Res", |
| "volume": "12", |
| "issue": "", |
| "pages": "2121--2159", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "John Duchi, Elad Hazan, and Yoram Singer. 2011. Adaptive subgradient methods for online learning and stochastic optimization. J. Mach. Learn. Res., 12:2121-2159, July.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "Phrasetable smoothing for statistical machine translation", |
| "authors": [ |
| { |
| "first": "George", |
| "middle": [], |
| "last": "Foster", |
| "suffix": "" |
| }, |
| { |
| "first": "Roland", |
| "middle": [], |
| "last": "Kuhn", |
| "suffix": "" |
| }, |
| { |
| "first": "Howard", |
| "middle": [], |
| "last": "Johnson", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the 2006 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "53--61", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "George Foster, Roland Kuhn, and Howard Johnson. 2006. Phrasetable smoothing for statistical machine translation. In Proceedings of the 2006 Conference on Empirical Methods in Natural Language Pro- cessing, pages 53-61, Sydney, Australia, July. As- sociation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "Deep sparse rectifier networks", |
| "authors": [ |
| { |
| "first": "Xavier", |
| "middle": [], |
| "last": "Glorot", |
| "suffix": "" |
| }, |
| { |
| "first": "Antoine", |
| "middle": [], |
| "last": "Bordes", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the 14th International Conference on Artificial Intelligence and Statistics. JMLR W&CP", |
| "volume": "15", |
| "issue": "", |
| "pages": "315--323", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Xavier Glorot, Antoine Bordes, and Yoshua Bengio. 2011. Deep sparse rectifier networks. In Proceed- ings of the 14th International Conference on Arti- ficial Intelligence and Statistics. JMLR W&CP Vol- ume, volume 15, pages 315-323.", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "Hidden topic markov models", |
| "authors": [ |
| { |
| "first": "Amit", |
| "middle": [], |
| "last": "Gruber", |
| "suffix": "" |
| }, |
| { |
| "first": "Michal", |
| "middle": [], |
| "last": "Rosen-Zvi", |
| "suffix": "" |
| }, |
| { |
| "first": "Yair", |
| "middle": [], |
| "last": "Weiss", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of Artificial Intelligence and Statistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Amit Gruber, Michal Rosen-zvi, and Yair Weiss. 2007. Hidden topic markov models. In In Proceedings of Artificial Intelligence and Statistics.", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "Improving statistical machine translation using lexicalized rule selection", |
| "authors": [ |
| { |
| "first": "Zhongjun", |
| "middle": [], |
| "last": "He", |
| "suffix": "" |
| }, |
| { |
| "first": "Qun", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Shouxun", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of the 22nd International Conference on Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "321--328", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zhongjun He, Qun Liu, and Shouxun Lin. 2008. Im- proving statistical machine translation using lexical- ized rule selection. In Proceedings of the 22nd In- ternational Conference on Computational Linguis- tics (Coling 2008), pages 321-328, Manchester, UK, August. Coling 2008 Organizing Committee.", |
| "links": null |
| }, |
| "BIBREF30": { |
| "ref_id": "b30", |
| "title": "A fast learning algorithm for deep belief nets", |
| "authors": [ |
| { |
| "first": "Geoffrey", |
| "middle": [ |
| "E" |
| ], |
| "last": "Hinton", |
| "suffix": "" |
| }, |
| { |
| "first": "Simon", |
| "middle": [], |
| "last": "Osindero", |
| "suffix": "" |
| }, |
| { |
| "first": "Yee-Whye", |
| "middle": [], |
| "last": "Teh", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Neural Comput", |
| "volume": "18", |
| "issue": "7", |
| "pages": "1527--1554", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Geoffrey E. Hinton, Simon Osindero, and Yee-Whye Teh. 2006. A fast learning algorithm for deep belief nets. Neural Comput., 18(7):1527-1554, July.", |
| "links": null |
| }, |
| "BIBREF31": { |
| "ref_id": "b31", |
| "title": "Improved backing-off for m-gram language modeling", |
| "authors": [ |
| { |
| "first": "Reinhard", |
| "middle": [], |
| "last": "Kneser", |
| "suffix": "" |
| }, |
| { |
| "first": "Hermann", |
| "middle": [], |
| "last": "Ney", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "Acoustics, Speech, and Signal Processing", |
| "volume": "1", |
| "issue": "", |
| "pages": "181--184", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Reinhard Kneser and Hermann Ney. 1995. Im- proved backing-off for m-gram language modeling. In Acoustics, Speech, and Signal Processing, 1995. ICASSP-95., 1995 International Conference on, vol- ume 1, pages 181-184. IEEE.", |
| "links": null |
| }, |
| "BIBREF32": { |
| "ref_id": "b32", |
| "title": "Statistical significance tests for machine translation evaluation", |
| "authors": [ |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of EMNLP 2004", |
| "volume": "", |
| "issue": "", |
| "pages": "388--395", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Philipp Koehn. 2004. Statistical significance tests for machine translation evaluation. In Dekang Lin and Dekai Wu, editors, Proceedings of EMNLP 2004, pages 388-395, Barcelona, Spain, July. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF33": { |
| "ref_id": "b33", |
| "title": "Imagenet classification with deep convolutional neural networks", |
| "authors": [ |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Krizhevsky", |
| "suffix": "" |
| }, |
| { |
| "first": "Ilya", |
| "middle": [], |
| "last": "Sutskever", |
| "suffix": "" |
| }, |
| { |
| "first": "Geoff", |
| "middle": [], |
| "last": "Hinton", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Advances in Neural Information Processing Systems", |
| "volume": "25", |
| "issue": "", |
| "pages": "1106--1114", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alex Krizhevsky, Ilya Sutskever, and Geoff Hinton. 2012. Imagenet classification with deep convolu- tional neural networks. In P. Bartlett, F.C.N. Pereira, C.J.C. Burges, L. Bottou, and K.Q. Weinberger, ed- itors, Advances in Neural Information Processing Systems 25, pages 1106-1114.", |
| "links": null |
| }, |
| "BIBREF34": { |
| "ref_id": "b34", |
| "title": "Efficient sparse coding algorithms", |
| "authors": [ |
| { |
| "first": "Honglak", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexis", |
| "middle": [], |
| "last": "Battle", |
| "suffix": "" |
| }, |
| { |
| "first": "Rajat", |
| "middle": [], |
| "last": "Raina", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [ |
| "Y" |
| ], |
| "last": "Ng", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Advances in Neural Information Processing Systems 19", |
| "volume": "", |
| "issue": "", |
| "pages": "801--808", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Honglak Lee, Alexis Battle, Rajat Raina, and An- drew Y. Ng. 2006. Efficient sparse coding algo- rithms. In B. Sch\u00f6lkopf, J. Platt, and T. Hoffman, editors, Advances in Neural Information Processing Systems 19, pages 801-808. MIT Press, Cambridge, MA.", |
| "links": null |
| }, |
| "BIBREF35": { |
| "ref_id": "b35", |
| "title": "Maximum entropy based rule selection model for syntax-based statistical machine translation", |
| "authors": [ |
| { |
| "first": "Qun", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Zhongjun", |
| "middle": [], |
| "last": "He", |
| "suffix": "" |
| }, |
| { |
| "first": "Yang", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Shouxun", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "89--97", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Qun Liu, Zhongjun He, Yang Liu, and Shouxun Lin. 2008. Maximum entropy based rule selection model for syntax-based statistical machine translation. In Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing, pages 89-97, Honolulu, Hawaii, October. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF36": { |
| "ref_id": "b36", |
| "title": "Plda+: Parallel latent dirichlet allocation with data placement and pipeline processing", |
| "authors": [ |
| { |
| "first": "Zhiyuan", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Yuzhou", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Edward", |
| "middle": [ |
| "Y" |
| ], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "Maosong", |
| "middle": [], |
| "last": "Sun", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "ACM Transactions on Intelligent Systems and Technology, special issue on Large Scale Machine Learning. Software", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zhiyuan Liu, Yuzhou Zhang, Edward Y. Chang, and Maosong Sun. 2011. Plda+: Parallel latent dirichlet allocation with data placement and pipeline process- ing. ACM Transactions on Intelligent Systems and Technology, special issue on Large Scale Machine Learning. Software available at http://code. google.com/p/plda.", |
| "links": null |
| }, |
| "BIBREF37": { |
| "ref_id": "b37", |
| "title": "Soft syntactic constraints for hierarchical phrased-based translation", |
| "authors": [ |
| { |
| "first": "Yuval", |
| "middle": [], |
| "last": "Marton", |
| "suffix": "" |
| }, |
| { |
| "first": "Philip", |
| "middle": [], |
| "last": "Resnik", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "1003--1011", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yuval Marton and Philip Resnik. 2008. Soft syntactic constraints for hierarchical phrased-based transla- tion. In Proceedings of ACL-08: HLT, pages 1003- 1011, Columbus, Ohio, June. Association for Com- putational Linguistics.", |
| "links": null |
| }, |
| "BIBREF38": { |
| "ref_id": "b38", |
| "title": "Minimum error rate training in statistical machine translation", |
| "authors": [ |
| { |
| "first": "Franz Josef", |
| "middle": [], |
| "last": "Och", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proceedings of the 41st Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "160--167", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Franz Josef Och. 2003. Minimum error rate train- ing in statistical machine translation. In Proceed- ings of the 41st Annual Meeting of the Association for Computational Linguistics, pages 160-167, Sap- poro, Japan, July. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF39": { |
| "ref_id": "b39", |
| "title": "Bleu: a method for automatic evaluation of machine translation", |
| "authors": [ |
| { |
| "first": "Kishore", |
| "middle": [], |
| "last": "Papineni", |
| "suffix": "" |
| }, |
| { |
| "first": "Salim", |
| "middle": [], |
| "last": "Roukos", |
| "suffix": "" |
| }, |
| { |
| "first": "Todd", |
| "middle": [], |
| "last": "Ward", |
| "suffix": "" |
| }, |
| { |
| "first": "Wei-Jing", |
| "middle": [], |
| "last": "Zhu", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proceedings of 40th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "311--318", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kishore Papineni, Salim Roukos, Todd Ward, and Wei- Jing Zhu. 2002. Bleu: a method for automatic evaluation of machine translation. In Proceedings of 40th Annual Meeting of the Association for Com- putational Linguistics, pages 311-318, Philadelphia, Pennsylvania, USA, July. Association for Computa- tional Linguistics.", |
| "links": null |
| }, |
| "BIBREF40": { |
| "ref_id": "b40", |
| "title": "Neurocomputing: Foundations of research. chapter Learning Representations by Back-propagating Errors", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [ |
| "E" |
| ], |
| "last": "Rumelhart", |
| "suffix": "" |
| }, |
| { |
| "first": "Geoffrey", |
| "middle": [ |
| "E" |
| ], |
| "last": "Hinton", |
| "suffix": "" |
| }, |
| { |
| "first": "Ronald", |
| "middle": [ |
| "J" |
| ], |
| "last": "Williams", |
| "suffix": "" |
| } |
| ], |
| "year": 1988, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "696--699", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David E. Rumelhart, Geoffrey E. Hinton, and Ronald J. Williams. 1988. Neurocomputing: Foundations of research. chapter Learning Representations by Back-propagating Errors, pages 696-699. MIT Press, Cambridge, MA, USA.", |
| "links": null |
| }, |
| "BIBREF41": { |
| "ref_id": "b41", |
| "title": "Contrastive estimation: Training log-linear models on unlabeled data", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Noah", |
| "suffix": "" |
| }, |
| { |
| "first": "Jason", |
| "middle": [], |
| "last": "Smith", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Eisner", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics (ACL'05)", |
| "volume": "", |
| "issue": "", |
| "pages": "354--362", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Noah A. Smith and Jason Eisner. 2005. Contrastive estimation: Training log-linear models on unlabeled data. In Proceedings of the 43rd Annual Meet- ing of the Association for Computational Linguis- tics (ACL'05), pages 354-362, Ann Arbor, Michi- gan, June. Association for Computational Linguis- tics.", |
| "links": null |
| }, |
| "BIBREF42": { |
| "ref_id": "b42", |
| "title": "Parsing Natural Scenes and Natural Language with Recursive Neural Networks", |
| "authors": [ |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| }, |
| { |
| "first": "Cliff", |
| "middle": [ |
| "C" |
| ], |
| "last": "Lin", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [ |
| "Y" |
| ], |
| "last": "Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [ |
| "D" |
| ], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the 26th International Conference on Machine Learning (ICML)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Richard Socher, Cliff C. Lin, Andrew Y. Ng, and Christopher D. Manning. 2011a. Parsing Natural Scenes and Natural Language with Recursive Neural Networks. In Proceedings of the 26th International Conference on Machine Learning (ICML).", |
| "links": null |
| }, |
| "BIBREF43": { |
| "ref_id": "b43", |
| "title": "Semi-supervised recursive autoencoders for predicting sentiment distributions", |
| "authors": [ |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Pennington", |
| "suffix": "" |
| }, |
| { |
| "first": "Eric", |
| "middle": [ |
| "H" |
| ], |
| "last": "Huang", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [ |
| "Y" |
| ], |
| "last": "Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [ |
| "D" |
| ], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the 2011 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "151--161", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Richard Socher, Jeffrey Pennington, Eric H. Huang, Andrew Y. Ng, and Christopher D. Manning. 2011b. Semi-supervised recursive autoencoders for predict- ing sentiment distributions. In Proceedings of the 2011 Conference on Empirical Methods in Natural Language Processing, pages 151-161, Edinburgh, Scotland, UK., July. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF44": { |
| "ref_id": "b44", |
| "title": "Translation model adaptation for statistical machine translation with monolingual topic information", |
| "authors": [ |
| { |
| "first": "Jinsong", |
| "middle": [], |
| "last": "Su", |
| "suffix": "" |
| }, |
| { |
| "first": "Hua", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Haifeng", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Yidong", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Xiaodong", |
| "middle": [], |
| "last": "Shi", |
| "suffix": "" |
| }, |
| { |
| "first": "Huailin", |
| "middle": [], |
| "last": "Dong", |
| "suffix": "" |
| }, |
| { |
| "first": "Qun", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "459--468", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jinsong Su, Hua Wu, Haifeng Wang, Yidong Chen, Xiaodong Shi, Huailin Dong, and Qun Liu. 2012. Translation model adaptation for statistical machine translation with monolingual topic information. In Proceedings of the 50th Annual Meeting of the As- sociation for Computational Linguistics (Volume 1: Long Papers), pages 459-468, Jeju Island, Korea, July. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF45": { |
| "ref_id": "b45", |
| "title": "Bilingual lsa-based adaptation for statistical machine translation", |
| "authors": [ |
| { |
| "first": "Yik-Cheung", |
| "middle": [], |
| "last": "Tam", |
| "suffix": "" |
| }, |
| { |
| "first": "Ian", |
| "middle": [], |
| "last": "Lane", |
| "suffix": "" |
| }, |
| { |
| "first": "Tanja", |
| "middle": [], |
| "last": "Schultz", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Machine Translation", |
| "volume": "21", |
| "issue": "4", |
| "pages": "187--207", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yik-Cheung Tam, Ian Lane, and Tanja Schultz. 2007. Bilingual lsa-based adaptation for statistical ma- chine translation. Machine Translation, 21(4):187- 207, December.", |
| "links": null |
| }, |
| "BIBREF46": { |
| "ref_id": "b46", |
| "title": "Extracting and composing robust features with denoising autoencoders", |
| "authors": [ |
| { |
| "first": "Pascal", |
| "middle": [], |
| "last": "Vincent", |
| "suffix": "" |
| }, |
| { |
| "first": "Hugo", |
| "middle": [], |
| "last": "Larochelle", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| }, |
| { |
| "first": "Pierre-Antoine", |
| "middle": [], |
| "last": "Manzagol", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of the 25th International Conference on Machine Learning, ICML '08", |
| "volume": "", |
| "issue": "", |
| "pages": "1096--1103", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Pascal Vincent, Hugo Larochelle, Yoshua Bengio, and Pierre-Antoine Manzagol. 2008. Extracting and composing robust features with denoising autoen- coders. In Proceedings of the 25th International Conference on Machine Learning, ICML '08, pages 1096-1103, New York, NY, USA. ACM.", |
| "links": null |
| }, |
| "BIBREF47": { |
| "ref_id": "b47", |
| "title": "Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion", |
| "authors": [ |
| { |
| "first": "Pascal", |
| "middle": [], |
| "last": "Vincent", |
| "suffix": "" |
| }, |
| { |
| "first": "Hugo", |
| "middle": [], |
| "last": "Larochelle", |
| "suffix": "" |
| }, |
| { |
| "first": "Isabelle", |
| "middle": [], |
| "last": "Lajoie", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| }, |
| { |
| "first": "Pierre-Antoine", |
| "middle": [], |
| "last": "Manzagol", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "J. Mach. Learn. Res", |
| "volume": "11", |
| "issue": "", |
| "pages": "3371--3408", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Pascal Vincent, Hugo Larochelle, Isabelle Lajoie, Yoshua Bengio, and Pierre-Antoine Manzagol. 2010. Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion. J. Mach. Learn. Res., 11:3371- 3408, December.", |
| "links": null |
| }, |
| "BIBREF48": { |
| "ref_id": "b48", |
| "title": "A topic similarity model for hierarchical phrase-based translation", |
| "authors": [ |
| { |
| "first": "Xinyan", |
| "middle": [], |
| "last": "Xiao", |
| "suffix": "" |
| }, |
| { |
| "first": "Deyi", |
| "middle": [], |
| "last": "Xiong", |
| "suffix": "" |
| }, |
| { |
| "first": "Min", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Qun", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Shouxun", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "750--758", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Xinyan Xiao, Deyi Xiong, Min Zhang, Qun Liu, and Shouxun Lin. 2012. A topic similarity model for hierarchical phrase-based translation. In Proceed- ings of the 50th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Pa- pers), pages 750-758, Jeju Island, Korea, July. As- sociation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF49": { |
| "ref_id": "b49", |
| "title": "A topic-based coherence model for statistical machine translation", |
| "authors": [ |
| { |
| "first": "Deyi", |
| "middle": [], |
| "last": "Xiong", |
| "suffix": "" |
| }, |
| { |
| "first": "Min", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "AAAI", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Deyi Xiong and Min Zhang. 2013. A topic-based co- herence model for statistical machine translation. In AAAI.", |
| "links": null |
| }, |
| "BIBREF50": { |
| "ref_id": "b50", |
| "title": "A syntax-driven bracketing model for phrasebased translation", |
| "authors": [ |
| { |
| "first": "Deyi", |
| "middle": [], |
| "last": "Xiong", |
| "suffix": "" |
| }, |
| { |
| "first": "Min", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Aiti", |
| "middle": [], |
| "last": "Aw", |
| "suffix": "" |
| }, |
| { |
| "first": "Haizhou", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "315--323", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Deyi Xiong, Min Zhang, Aiti Aw, and Haizhou Li. 2009. A syntax-driven bracketing model for phrase- based translation. In Proceedings of the Joint Con- ference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP, pages 315- 323, Suntec, Singapore, August. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF51": { |
| "ref_id": "b51", |
| "title": "Bitam: Bilingual topic admixture models for word alignment", |
| "authors": [ |
| { |
| "first": "Bing", |
| "middle": [], |
| "last": "Zhao", |
| "suffix": "" |
| }, |
| { |
| "first": "Eric", |
| "middle": [ |
| "P" |
| ], |
| "last": "Xing", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the COLING/ACL 2006 Main Conference Poster Sessions", |
| "volume": "", |
| "issue": "", |
| "pages": "969--976", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bing Zhao and Eric P. Xing. 2006. Bitam: Bilingual topic admixture models for word alignment. In Pro- ceedings of the COLING/ACL 2006 Main Confer- ence Poster Sessions, pages 969-976, Sydney, Aus- tralia, July. Association for Computational Linguis- tics.", |
| "links": null |
| }, |
| "BIBREF52": { |
| "ref_id": "b52", |
| "title": "Hm-bitam: Bilingual topic exploration, word alignment, and translation", |
| "authors": [ |
| { |
| "first": "Bing", |
| "middle": [], |
| "last": "Zhao", |
| "suffix": "" |
| }, |
| { |
| "first": "Eric", |
| "middle": [ |
| "P" |
| ], |
| "last": "Xing", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Advances in Neural Information Processing Systems", |
| "volume": "20", |
| "issue": "", |
| "pages": "1689--1696", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bing Zhao and Eric P. Xing. 2007. Hm-bitam: Bilin- gual topic exploration, word alignment, and trans- lation. In J.C. Platt, D. Koller, Y. Singer, and S. Roweis, editors, Advances in Neural Informa- tion Processing Systems 20, pages 1689-1696. MIT Press, Cambridge, MA.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "TABREF4": { |
| "html": null, |
| "text": "", |
| "num": null, |
| "type_str": "table", |
| "content": "<table><tr><td>im(Src)</td><td>42.51</td><td/><td/><td colspan=\"2\">41.55</td><td/><td/><td>38.53</td><td/><td colspan=\"2\">31.57</td><td/><td colspan=\"2\">38.54</td><td/><td/><td/><td/><td/><td/><td/><td/></tr><tr><td>im(Trg)</td><td>42.43</td><td/><td/><td colspan=\"2\">41.48</td><td/><td/><td>38.4</td><td/><td colspan=\"2\">31.49</td><td/><td colspan=\"2\">38.45</td><td/><td/><td/><td/><td/><td/><td/><td/></tr><tr><td>im(Src+Trg)</td><td>42.7</td><td/><td/><td colspan=\"2\">41.66</td><td/><td/><td>38.66</td><td/><td colspan=\"2\">31.66</td><td/><td colspan=\"2\">38.67</td><td/><td/><td/><td/><td/><td/><td/><td/></tr><tr><td>im(Src+Trg)+Sen(Src)</td><td>42.77</td><td/><td/><td colspan=\"2\">41.81</td><td/><td/><td>38.85</td><td/><td colspan=\"2\">31.73</td><td/><td colspan=\"2\">38.79</td><td/><td/><td/><td/><td/><td/><td/><td/></tr><tr><td>im(Src+Trg)+Sen(Trg)</td><td>42.85</td><td/><td/><td colspan=\"2\">41.79</td><td/><td/><td>38.76</td><td/><td>31.7</td><td/><td/><td colspan=\"2\">38.78</td><td/><td/><td/><td/><td/><td/><td/><td/></tr><tr><td>im(Src+Trg)+Sen(Src+Trg)</td><td>42.95</td><td/><td/><td colspan=\"2\">41.97</td><td/><td/><td>38.91</td><td/><td colspan=\"2\">31.88</td><td/><td colspan=\"2\">38.93</td><td/><td/><td/><td/><td/><td/><td/><td/></tr><tr><td colspan=\"6\">\u8054\u5408\u56fd\u513f\u7ae5\u57fa\u91d1\u4f1a\u4e5f\u5f00\u59cb\u53d1\u9001\u57fa\u672c\u533b\u836f\u5305</td><td/><td colspan=\"3\"><\u53d1\u9001 X , deliver X></td><td/><td/><td colspan=\"4\"><\u53d1\u9001 X , distribute X></td><td/><td/><td colspan=\"4\"><\u53d1\u9001 X , send X></td><td/></tr><tr><td>0.1</td><td/><td/><td/><td/><td/><td>0.1</td><td/><td/><td/><td colspan=\"2\">0.1</td><td/><td/><td/><td/><td/><td>0.1</td><td/><td/><td/><td/><td/></tr><tr><td>0.08</td><td/><td/><td/><td/><td/><td>0.08</td><td/><td/><td/><td colspan=\"2\">0.08</td><td/><td/><td/><td/><td/><td>0.08</td><td/><td/><td/><td/><td/></tr><tr><td>0.06</td><td/><td/><td/><td/><td/><td>0.06</td><td/><td/><td/><td colspan=\"2\">0.06</td><td/><td/><td/><td/><td/><td>0.06</td><td/><td/><td/><td/><td/></tr><tr><td>0.04</td><td/><td/><td/><td/><td/><td>0.04</td><td/><td/><td/><td colspan=\"2\">0.04</td><td/><td/><td/><td/><td/><td>0.04</td><td/><td/><td/><td/><td/></tr><tr><td>0.02</td><td/><td/><td/><td/><td/><td>0.02</td><td/><td/><td/><td colspan=\"2\">0.02</td><td/><td/><td/><td/><td/><td>0.02</td><td/><td/><td/><td/><td/></tr><tr><td>0</td><td>20</td><td>40</td><td>60</td><td>80</td><td>100</td><td>0 0</td><td>20</td><td>40</td><td>60</td><td>100</td><td>0 0</td><td>20</td><td>40</td><td>60</td><td>80</td><td>100</td><td>0 0</td><td>20</td><td>40</td><td>60</td><td>80</td><td>100</td></tr></table>" |
| }, |
| "TABREF6": { |
| "html": null, |
| "text": "Development and testing data used in the experiments.", |
| "num": null, |
| "type_str": "table", |
| "content": "<table/>" |
| } |
| } |
| } |
| } |