| { |
| "paper_id": "P12-1048", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T09:29:18.448048Z" |
| }, |
| "title": "Translation Model Adaptation for Statistical Machine Translation with Monolingual Topic Information *", |
| "authors": [ |
| { |
| "first": "Jinsong", |
| "middle": [], |
| "last": "Su", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Xiamen University", |
| "location": { |
| "settlement": "Xiamen", |
| "country": "China" |
| } |
| }, |
| "email": "jssu@xmu.edu.cn" |
| }, |
| { |
| "first": "Hua", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "wuhua@baicu.com" |
| }, |
| { |
| "first": "Haifeng", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "wanghaifeng@baicu.com" |
| }, |
| { |
| "first": "Yidong", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Chinese Academy of Sciences", |
| "location": { |
| "settlement": "Beijing", |
| "country": "China" |
| } |
| }, |
| "email": "ydchen@xmu.edu.cn" |
| }, |
| { |
| "first": "Xiaodong", |
| "middle": [], |
| "last": "Shi", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Chinese Academy of Sciences", |
| "location": { |
| "settlement": "Beijing", |
| "country": "China" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Huailin", |
| "middle": [], |
| "last": "Dong", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "hldong@xmu.edu.cn" |
| }, |
| { |
| "first": "Qun", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Chinese Academy of Sciences", |
| "location": { |
| "settlement": "Beijing", |
| "country": "China" |
| } |
| }, |
| "email": "liuqun@ict.ac.cn" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "To adapt a translation model trained from the data in one domain to another, previous works paid more attention to the studies of parallel corpus while ignoring the in-domain monolingual corpora which can be obtained more easily. In this paper, we propose a novel approach for translation model adaptation by utilizing in-domain monolingual topic information instead of the in-domain bilingual corpora, which incorporates the topic information into translation probability estimation. Our method establishes the relationship between the out-of-domain bilingual corpus and the in-domain monolingual corpora via topic mapping and phrase-topic distribution probability estimation from in-domain monolingual corpora. Experimental result on the NIST Chinese-English translation task shows that our approach significantly outperforms the baseline system.", |
| "pdf_parse": { |
| "paper_id": "P12-1048", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "To adapt a translation model trained from the data in one domain to another, previous works paid more attention to the studies of parallel corpus while ignoring the in-domain monolingual corpora which can be obtained more easily. In this paper, we propose a novel approach for translation model adaptation by utilizing in-domain monolingual topic information instead of the in-domain bilingual corpora, which incorporates the topic information into translation probability estimation. Our method establishes the relationship between the out-of-domain bilingual corpus and the in-domain monolingual corpora via topic mapping and phrase-topic distribution probability estimation from in-domain monolingual corpora. Experimental result on the NIST Chinese-English translation task shows that our approach significantly outperforms the baseline system.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "In recent years, statistical machine translation(SMT) has been rapidly developing with more and more novel translation models being proposed and put into practice (Koehn et al., 2003; Och and Ney, 2004; Galley et al., 2006; Liu et al., 2006; Chiang, 2007; Chiang, 2010) . However, similar to other natural language processing(NLP) tasks, SMT systems often suffer from domain adaptation problem during practical applications. The simple reason is that the underlying statistical models always tend to closely approximate the empirical distributions of the training data, which typically consist of bilingual sentences and monolingual target language sentences. When the translated texts and the training data come from the same domain, SMT systems can achieve good performance, otherwise the translation quality degrades dramatically. Therefore, it is of significant importance to develop translation systems which can be effectively transferred from one domain to another, for example, from newswire to weblog.", |
| "cite_spans": [ |
| { |
| "start": 163, |
| "end": 183, |
| "text": "(Koehn et al., 2003;", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 184, |
| "end": 202, |
| "text": "Och and Ney, 2004;", |
| "ref_id": "BIBREF23" |
| }, |
| { |
| "start": 203, |
| "end": 223, |
| "text": "Galley et al., 2006;", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 224, |
| "end": 241, |
| "text": "Liu et al., 2006;", |
| "ref_id": "BIBREF27" |
| }, |
| { |
| "start": 242, |
| "end": 255, |
| "text": "Chiang, 2007;", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 256, |
| "end": 269, |
| "text": "Chiang, 2010)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "According to adaptation emphases, domain adaptation in SMT can be classified into translation model adaptation and language model adaptation. Here we focus on how to adapt a translation model, which is trained from the large-scale out-of-domain bilingual corpus, for domain-specific translation task, leaving others for future work. In this aspect, previous methods can be divided into two categories: one paid attention to collecting more sentence pairs by information retrieval technology (Hildebrand et al., 2005) or synthesized parallel sentences (Ueffing et al., 2008; Wu et al., 2008; Bertoldi and Federico, 2009; Schwenk and Senellart, 2009) , and the other exploited the full potential of existing parallel corpus in a mixture-modeling (Foster and Kuhn, 2007; Civera and Juan, 2007; Lv et al., 2007) framework. However, these approaches focused on the studies of bilingual corpus synthesis and exploitation while ignoring the monolingual corpora, therefore limiting the potential of further translation quality improvement.", |
| "cite_spans": [ |
| { |
| "start": 491, |
| "end": 516, |
| "text": "(Hildebrand et al., 2005)", |
| "ref_id": "BIBREF20" |
| }, |
| { |
| "start": 551, |
| "end": 573, |
| "text": "(Ueffing et al., 2008;", |
| "ref_id": "BIBREF39" |
| }, |
| { |
| "start": 574, |
| "end": 590, |
| "text": "Wu et al., 2008;", |
| "ref_id": "BIBREF40" |
| }, |
| { |
| "start": 591, |
| "end": 619, |
| "text": "Bertoldi and Federico, 2009;", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 620, |
| "end": 648, |
| "text": "Schwenk and Senellart, 2009)", |
| "ref_id": "BIBREF36" |
| }, |
| { |
| "start": 744, |
| "end": 767, |
| "text": "(Foster and Kuhn, 2007;", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 768, |
| "end": 790, |
| "text": "Civera and Juan, 2007;", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 791, |
| "end": 807, |
| "text": "Lv et al., 2007)", |
| "ref_id": "BIBREF28" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, we propose a novel adaptation method to adapt the translation model for domainspecific translation task by utilizing in-domain monolingual corpora. Our approach is inspired by the recent studies (Zhao and Xing, 2006; Zhao and Xing, 2007; Tam et al., 2007; Gong and Zhou, 2010; Ruiz and Federico, 2011) which have shown that a particular translation always appears in some specific topical contexts, and the topical context information has a great effect on translation selection. For example, \"bank\" often occurs in the sentences related to the economy topic when translated into \"y\u00ednh\u00e1ng\", and occurs in the sentences related to the geography topic when translated to \"h\u00e9\u00e0n\". Therefore, the co-occurrence frequency of the phrases in some specific context can be used to constrain the translation candidates of phrases. In a monolingual corpus, if \"bank\" occurs more often in the sentences related to the economy topic than the ones related to the geography topic, it is more likely that \"bank\" is translated to \"y\u00ednh\u00e1ng\" than to \"h\u00e9\u00e0n\". With the out-of-domain bilingual corpus, we first incorporate the topic information into translation probability estimation, aiming to quantify the effect of the topical context information on translation selection. Then, we rescore all phrase pairs according to the phrasetopic and the word-topic posterior distributions of the additional in-domain monolingual corpora. As compared to the previous works, our method takes advantage of both the in-domain monolingual corpora and the out-of-domain bilingual corpus to incorporate the topic information into our translation model, thus breaking down the corpus barrier for translation quality improvement. The experimental results on the NIST data set demonstrate the effectiveness of our method.", |
| "cite_spans": [ |
| { |
| "start": 210, |
| "end": 231, |
| "text": "(Zhao and Xing, 2006;", |
| "ref_id": "BIBREF44" |
| }, |
| { |
| "start": 232, |
| "end": 252, |
| "text": "Zhao and Xing, 2007;", |
| "ref_id": "BIBREF45" |
| }, |
| { |
| "start": 253, |
| "end": 270, |
| "text": "Tam et al., 2007;", |
| "ref_id": "BIBREF38" |
| }, |
| { |
| "start": 271, |
| "end": 291, |
| "text": "Gong and Zhou, 2010;", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 292, |
| "end": 316, |
| "text": "Ruiz and Federico, 2011)", |
| "ref_id": "BIBREF33" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The reminder of this paper is organized as follows: Section 2 provides a brief description of translation probability estimation. Section 3 introduces the adaptation method which incorporates the topic information into the translation model; Section 4 describes and discusses the experimental results; Section 5 briefly summarizes the recent related work about translation model adaptation. Finally, we end with a conclusion and the future work in Section 6.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The statistical translation model, which contains phrase pairs with bi-directional phrase probabilities and bi-directional lexical probabilities, has a great effect on the performance of SMT system. Phrase probability measures the co-occurrence frequency of a phrase pair, and lexical probability is used to validate the quality of the phrase pair by checking how well its words are translated to each other.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "2" |
| }, |
| { |
| "text": "According to the definition proposed by (Koehn et al., 2003) , given a source sentence f = f J 1 = f 1 , . . . , f j , . . . , f J , a target sentence e = e I 1 = e 1 , . . . , e i , . . . , e I , and its word alignment a which is a subset of the Cartesian product of word positions: a \u2286 (j, i) : j = 1, . . . , J; i = 1, . . . , I, the phrase pair (f ,\u1ebd) is said to be consistent (Och and Ney, 2004) with the alignment if and only if: (1) there must be at least one word inside one phrase aligned to a word inside the other phrase and (2) no words inside one phrase can be aligned to a word outside the other phrase. After all consistent phrase pairs are extracted from training corpus, the phrase probabilities are estimated as relative frequencies (Och and Ney, 2004) :", |
| "cite_spans": [ |
| { |
| "start": 40, |
| "end": 60, |
| "text": "(Koehn et al., 2003)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 751, |
| "end": 770, |
| "text": "(Och and Ney, 2004)", |
| "ref_id": "BIBREF23" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "\u03c6(\u1ebd|f ) = count(f ,\u1ebd) \u1ebd count(f ,\u1ebd )", |
| "eq_num": "(1)" |
| } |
| ], |
| "section": "Background", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Here count(f ,\u1ebd) indicates how often the phrase pair (f ,\u1ebd) occurs in the training corpus.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "2" |
| }, |
| { |
| "text": "To obtain the corresponding lexical weight, we first estimate a lexical translation probability distribution w(e|f ) by relative frequency from the training corpus:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "w(e|f ) = count(f, e) e count(f, e )", |
| "eq_num": "(2)" |
| } |
| ], |
| "section": "Background", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Retaining the alignment\u00e3 between the phrase pair (f ,\u1ebd), the corresponding lexical weight is calculated as", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "2" |
| }, |
| { |
| "text": "p w (\u1ebd|f ,\u00e3) = |\u1ebd| i=1 1 |{j|(j, i) \u2208\u00e3}| \u2200(j,i)\u2208\u00e3 w(e i |f j ) (3)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "2" |
| }, |
| { |
| "text": "However, the above-mentioned method only counts the co-occurrence frequency of bilingual phrases, assuming that the translation probability is independent of the context information. Thus, the statistical model estimated from the training data is not suitable for text translation in different domains, resulting in a significant drop in translation quality.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "2" |
| }, |
| { |
| "text": "In this section, we first briefly review the principle of Hidden Topic Markov Model(HTMM) which is the basis of our method, then describe our approach to translation model adaptation in detail.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Translation Model Adaptation via Monolingual Topic Information", |
| "sec_num": "3" |
| }, |
| { |
| "text": "During the last couple of years, topic models such as Probabilistic Latent Semantic Analysis (Hofmann, 1999) and Latent Dirichlet Allocation model (Blei, 2003) , have drawn more and more attention and been applied successfully in NLP community. Based on the \"bag-of-words\" assumption that the order of words can be ignored, these methods model the text corpus by using a co-occurrence matrix of words and documents, and build generative models to infer the latent aspects or topics. Using these models, the words can be clustered into the derived topics with a probability distribution, and the correlation between words can be automatically captured via topics. However, the \"bag-of-words\" assumption is an unrealistic oversimplification because it ignores the order of words. To remedy this problem, Gruber et al.(2007) propose HTMM, which models the topics of words in the document as a Markov chain. Based on the assumption that all words in the same sentence have the same topic and the successive sentences are more likely to have the same topic, HTM-M incorporates the local dependency between words by Hidden Markov Model for better topic estimation.", |
| "cite_spans": [ |
| { |
| "start": 93, |
| "end": 108, |
| "text": "(Hofmann, 1999)", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 147, |
| "end": 159, |
| "text": "(Blei, 2003)", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 802, |
| "end": 821, |
| "text": "Gruber et al.(2007)", |
| "ref_id": "BIBREF17" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Hidden Topic Markov Model", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "HTMM can also be viewed as a soft clustering tool for words in training corpus. That is, HT-MM can estimate the probability distribution of a topic over words, i.e. the topic-word distribution P (word|topic) during training. Besides, HTMM derives inherent topics in sentences rather than in documents, so we can easily obtain the sentencetopic distribution P (topic|sentence) in training corpus. Adopting maximum likelihood estimation(MLE), this posterior distribution makes it possible to effectively calculate the word-topic distribution P (topic|word) and the phrase-topic distribution P (topic|phrase) both of which are very important in our method.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Hidden Topic Markov Model", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "We utilize the additional in-domain monolingual corpora to adapt the out-of-domain translation model for domain-specific translation task. In detail, we build an adapted translation model in the following steps:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Adapted Phrase Probability Estimation", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "\u2022 Build a topic-specific translation model to quantify the effect of the topic information on the translation probability estimation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Adapted Phrase Probability Estimation", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "\u2022 Estimate the topic posterior distributions of phrases in the in-domain monolingual corpora.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Adapted Phrase Probability Estimation", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "\u2022 Score the phrase pairs according to the predefined topic-specific translation model and the topic posterior distribution of phrases.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Adapted Phrase Probability Estimation", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Formally, we incorporate monolingual topic information into translation probability estimation, and decompose the phrase probability \u03c6(\u1ebd|f ) 1 as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Adapted Phrase Probability Estimation", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "\u03c6(\u1ebd|f ) = t f \u03c6(\u1ebd, t f |f ) = t f \u03c6(\u1ebd|f , t f ) \u2022 P (t f |f )", |
| "eq_num": "(4)" |
| } |
| ], |
| "section": "Adapted Phrase Probability Estimation", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "where \u03c6(\u1ebd|f , t f ) indicates the probability of translatingf into\u1ebd given the source-side topic t f , P (t f |f ) denotes the phrase-topic distribution off .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Adapted Phrase Probability Estimation", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "To compute \u03c6(\u1ebd|f ), we first apply HTMM to respectively train two monolingual topic models with the following corpora: one is the source part of the out-of-domain bilingual corpus C f out , the other is the in-domain monolingual corpus C f in in the source language. Then, we respectively estimate \u03c6(\u1ebd|f , t f ) and P (t f |f ) from these two corpora. To avoid confusion, we further refine \u03c6(\u1ebd|f , t f ) and P (t f |f ) with \u03c6(\u1ebd|f , t f out ) and P (t f in |f ), respectively. Here, t f out is the topic clustered from the corpus C f out , and t f in represents the topic derived from the corpus C f in .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Adapted Phrase Probability Estimation", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "However, the two above-mentioned probabilities can not be directly multiplied in formula (4) because they are related to different topic spaces from different corpora. Besides, their topic dimensions are not assured to be the same. To solve this problem, we introduce the topic mapping probability P (t f out |t f in ) to map the in-domain phrase-topic distribution into the one in the out-domain topic space. To be specific, we obtain the out-of-domain phrase-topic distribution P (t f out |f ) as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Adapted Phrase Probability Estimation", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "P (t f out |f ) = t f in P (t f out |t f in ) \u2022 P (t f in |f ) (5)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Adapted Phrase Probability Estimation", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Thus formula (4) can be further refined as the following formula:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Adapted Phrase Probability Estimation", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "\u03c6(\u1ebd|f ) = t f out t f in \u03c6(\u1ebd|f , t f out ) \u2022P (t f out |t f in ) \u2022 P (t f in |f ) (6)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Adapted Phrase Probability Estimation", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Next we will give detailed descriptions of the calculation methods for the three probability distributions mentioned in formula (6).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Adapted Phrase Probability Estimation", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Probability \u03c6(\u1ebd|f , t f out )", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic-Specific Phrase Translation", |
| "sec_num": "3.2.1" |
| }, |
| { |
| "text": "We follow the common practice (Koehn et al., 2003) to calculate the topic-specific phrase translation probability, and the only difference is that our method takes the topical context information into account when collecting the fractional counts of phrase pairs. With the sentence-topic distribution P (t f out |f ) from the relevant topic model of C f out , the conditional probability \u03c6(\u1ebd|f , t f out ) can be easily obtained by MLE method:", |
| "cite_spans": [ |
| { |
| "start": 30, |
| "end": 50, |
| "text": "(Koehn et al., 2003)", |
| "ref_id": "BIBREF24" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic-Specific Phrase Translation", |
| "sec_num": "3.2.1" |
| }, |
| { |
| "text": "\u03c6(\u1ebd|f , t f out ) = f ,e \u2208Cout count f ,e (f ,\u1ebd) \u2022 P (t f out |f ) \u1ebd f ,e \u2208Cout count f ,e (f ,\u1ebd ) \u2022 P (t f out |f ) (7)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic-Specific Phrase Translation", |
| "sec_num": "3.2.1" |
| }, |
| { |
| "text": "where C out is the out-of-domain bilingual training corpus, and count f ,e (f ,\u1ebd) denotes the number of the phrase pair (f ,\u1ebd) in sentence pair f , e .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic-Specific Phrase Translation", |
| "sec_num": "3.2.1" |
| }, |
| { |
| "text": "P (t f out |t f in )", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Mapping Probability", |
| "sec_num": "3.2.2" |
| }, |
| { |
| "text": "Based on the two monolingual topic models respectively trained from C f in and C f out , we compute the topic mapping probability by using source word f as the pivot variable. Noticing that there are some words occurring in one corpus only, we use the words belonging to both corpora during the mapping procedure. Specifically, we decompose P (t f out |t f in ) as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Mapping Probability", |
| "sec_num": "3.2.2" |
| }, |
| { |
| "text": "P (t f out |t f in ) = f \u2208C f out C f in P (t f out |f ) \u2022 P (f |t f in ) (8)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Mapping Probability", |
| "sec_num": "3.2.2" |
| }, |
| { |
| "text": "Here we first get P (f |t f in ) directly from the topic model related to C f in . Then, considering the sentence-topic distribution P (t f out |f ) from the relevant topic model of C f out , we define the wordtopic distribution P (t f out |f ) as:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Mapping Probability", |
| "sec_num": "3.2.2" |
| }, |
| { |
| "text": "P (t f out |f ) = f \u2208C f out count f (f ) \u2022 P (t f out |f ) t f out f \u2208C f out count f (f ) \u2022 P (t f out |f ) (9)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Mapping Probability", |
| "sec_num": "3.2.2" |
| }, |
| { |
| "text": "where count f (f ) denotes the number of the word f in sentence f .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Mapping Probability", |
| "sec_num": "3.2.2" |
| }, |
| { |
| "text": "A simple way to compute the phrase-topic distribution is to take the fractional counts from C f in and then adopt MLE to obtain relative probability. However, it is infeasible in our model because some phrases occur in C f out while being absent in C f in . To solve this problem, we further compute this posterior distribution by the interpolation of two models:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Phrase-Topic Distribution P (t f in |f )", |
| "sec_num": "3.2.3" |
| }, |
| { |
| "text": "P (t f in |f ) = \u03b8 \u2022 P mle (t f in |f ) + (1 \u2212 \u03b8) \u2022 P word (t f in |f ) (10)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Phrase-Topic Distribution P (t f in |f )", |
| "sec_num": "3.2.3" |
| }, |
| { |
| "text": "where P mle (t f in |f ) indicates the phrase-topic distribution by MLE, P word (t f in |f ) denotes the phrase-topic distribution which is decomposed into the topic posterior distribution at the word level, and \u03b8 is the interpolation weight that can be optimized over the development data.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Phrase-Topic Distribution P (t f in |f )", |
| "sec_num": "3.2.3" |
| }, |
| { |
| "text": "Given the number of the phrasef in sentence f denoted as count f (f ), we compute the in-domain phrase-topic distribution in the following way:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Phrase-Topic Distribution P (t f in |f )", |
| "sec_num": "3.2.3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "P mle (t f in |f ) = f \u2208C f in count f (f ) \u2022 P (t f in |f ) t f in f \u2208C f in count f (f ) \u2022 P (t f in |f )", |
| "eq_num": "(11)" |
| } |
| ], |
| "section": "Phrase-Topic Distribution P (t f in |f )", |
| "sec_num": "3.2.3" |
| }, |
| { |
| "text": "Under the assumption that the topics of all words in the same phrase are independent, we consider two methods to calculate P word (t f in |f ). One is a \"Noisy-OR\" combination method (Zens and Ney, 2004) which has shown good performance in calculating similarities between bags-of-words in different languages. Using this method, P word (t f in |f ) is defined as:", |
| "cite_spans": [ |
| { |
| "start": 183, |
| "end": 203, |
| "text": "(Zens and Ney, 2004)", |
| "ref_id": "BIBREF41" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Phrase-Topic Distribution P (t f in |f )", |
| "sec_num": "3.2.3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "P word (t f in |f ) = 1 \u2212 P word (t f in |f ) \u2248 1 \u2212 f j \u2208f P (t f in |f j ) = 1 \u2212 f j \u2208f (1 \u2212 P (t f in |f j ))", |
| "eq_num": "(12)" |
| } |
| ], |
| "section": "Phrase-Topic Distribution P (t f in |f )", |
| "sec_num": "3.2.3" |
| }, |
| { |
| "text": "where P word (t f in |f ) represents the probability that t f in is not the topic of the phrasef . Similarly, P (t f in |f j ) indicates the probability that t f in is not the topic of the word f j . The other method is an \"Averaging\" combination one. With the assumption that t f in is the topic off if at least one of the words inf belongs to this topic, we derive P word (t f in |f ) as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Phrase-Topic Distribution P (t f in |f )", |
| "sec_num": "3.2.3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "P word (t f in |f ) \u2248 f j \u2208f P (t f in |f j )/|f |", |
| "eq_num": "(13)" |
| } |
| ], |
| "section": "Phrase-Topic Distribution P (t f in |f )", |
| "sec_num": "3.2.3" |
| }, |
| { |
| "text": "where |f | denotes the number of words in phrasef .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Phrase-Topic Distribution P (t f in |f )", |
| "sec_num": "3.2.3" |
| }, |
| { |
| "text": "Now we briefly describe how to estimate the adapted lexical weight for phrase pairs, which can be adjusted in a similar way to the phrase probability. Specifically, adopting our method, each word is considered as one phrase consisting of only one word, so", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Adapted Lexical Probability Estimation", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "w(e|f ) = t f out t f in w(e|f, t f out ) \u2022P (t f out |t f in ) \u2022 P (t f in |f ) (14)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Adapted Lexical Probability Estimation", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Here we obtain w(e|f, t f out ) with a similar approach to \u03c6(\u1ebd|f , t f out ), and calculate P (t f out |t f in ) and P (t f in |f ) by resorting to formulas (8) and (9).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Adapted Lexical Probability Estimation", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "With the adjusted lexical translation probability, we resort to formula (4) to update the lexical weight for the phrase pair (f ,\u1ebd).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Adapted Lexical Probability Estimation", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "We evaluate our method on the Chinese-to-English translation task for the weblog text. After a brief description of the experimental setup, we investigate the effects of various factors on the translation system performance.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiment", |
| "sec_num": "4" |
| }, |
| { |
| "text": "In our experiments, the out-of-domain training corpus comes from the FBIS corpus and the Hansards part of LDC2004T07 corpus (54.6K documents with 1M parallel sentences, 25.2M Chinese words and 29M English words). We use the Chinese Sohu weblog in 2009 1 and the English Blog Authorship corpus 2 (Schler et al., 2006) as the in-domain monolingual corpora in the source language and target language, respectively. To obtain more accurate topic information by HTMM, we firstly filter the noisy blog documents and the ones consisting of short sentences. After filtering, there are totally 85K Chinese blog documents with 2.1M sentences and 277K English blog documents with 4.3M sentences used in our experiments. Then, we sample equal numbers of documents from the in-domain monolingual corpora in the source language and the target language to respectively train two in-domain topic models. The web part of the 2006 NIST MT evaluation test data, consisting of 27 documents with 1048 sentences, is used as the development set, and the weblog part of the 2008 NIST MT test data, including 33 documents with 666 sentences, is our test set.", |
| "cite_spans": [ |
| { |
| "start": 295, |
| "end": 316, |
| "text": "(Schler et al., 2006)", |
| "ref_id": "BIBREF35" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental setup", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "To obtain various topic distributions for the outof-domain training corpus and the in-domain monolingual corpora in the source language and the target language respectively, we use HTMM tool developed by Gruber et al.(2007) to conduct topic model training. During this process, we empirically set the same parameter values for the HTMM training of different corpora: topics = 50, \u03b1 = 1.5, \u03b2 = 1.01, iters = 100. See (Gruber et al., 2007) for the meanings of these parameters. Besides, we set the interpolation weight \u03b8 in formula (10) to 0.5 by observing the results on development set in the additional experiments.", |
| "cite_spans": [ |
| { |
| "start": 204, |
| "end": 223, |
| "text": "Gruber et al.(2007)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 416, |
| "end": 437, |
| "text": "(Gruber et al., 2007)", |
| "ref_id": "BIBREF17" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental setup", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "We choose MOSES, a famous open-source phrase-based machine translation system (Koehn et al., 2007) , as the experimental decoder. GIZA++ (Och and Ney, 2003) and the heuristics \"grow-diag-final-and\" are used to generate a wordaligned corpus, from which we extract bilingual phrases with maximum length 7. We use SRILM Toolkits (Stolcke, 2002) to train two 4-gram language models on the filtered English Blog Authorship corpus and the Xinhua portion of Gigaword corpus, respectively. During decoding, we set the ttable-limit as 20, the stack-size as 100, and perform minimum-error-rate training (Och and Ney, 2003) to tune the feature weights for the log-linear model. The translation quality is evaluated by case-insensitive BLEU-4 metric (Papineni et al., 2002) . Finally, we conduct paired bootstrap sampling (Koehn, 2004) to test the significance in BLEU score differences.", |
| "cite_spans": [ |
| { |
| "start": 78, |
| "end": 98, |
| "text": "(Koehn et al., 2007)", |
| "ref_id": "BIBREF26" |
| }, |
| { |
| "start": 137, |
| "end": 156, |
| "text": "(Och and Ney, 2003)", |
| "ref_id": "BIBREF22" |
| }, |
| { |
| "start": 326, |
| "end": 341, |
| "text": "(Stolcke, 2002)", |
| "ref_id": "BIBREF37" |
| }, |
| { |
| "start": 593, |
| "end": 612, |
| "text": "(Och and Ney, 2003)", |
| "ref_id": "BIBREF22" |
| }, |
| { |
| "start": 738, |
| "end": 761, |
| "text": "(Papineni et al., 2002)", |
| "ref_id": "BIBREF34" |
| }, |
| { |
| "start": 810, |
| "end": 823, |
| "text": "(Koehn, 2004)", |
| "ref_id": "BIBREF25" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental setup", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "Our first experiments investigate the effect of different smoothing methods for the in-domain phrasetopic distribution: \"Noisy-OR\" and \"Averaging\". We build adapted phrase tables with these two methods, and then respectively use them in place of the out-of-domain phrase table to test the system performance. For the purpose of studying the generality of our approach, we carry out comparative experiments on two sizes of in-domain monolingual corpora: 5K and 40K. Table 1 reports the BLEU scores of the translation system under various conditions. Using the out-ofdomain phrase table, the baseline system achieves a BLEU score of 20.22. In the experiments with the small-scale in-domain monolingual corpora, the BLEU scores acquired by two methods are 20.45 and 20.54, achieving absolute improvements of 0.23 and 0.32 on the test set, respectively. In the experiments with the large-scale monolingual in-domain corpora, similar results are obtained, with absolute improvements of 0.54 and 0.89 over the baseline system.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 465, |
| "end": 472, |
| "text": "Table 1", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Effect of Different Smoothing Methods", |
| "sec_num": "4.2.1" |
| }, |
| { |
| "text": "From the above experimental results, we know that both \"Noisy-OR\" and \"Averaging\" combination methods improve the performance over the baseline, and \"Averaging\" method seems to be slightly better. This finding fails to echo the promising results in the previous study (Zens and Ney, 2004) . This is because the \"Noisy-OR\" method involves the multiplication of the word-topic distribution (shown in formula (12)), which leads to much sharper phrase-topic distribution than \"Averaging\" method, and is more likely to introduce bias to the translation probability estimation. Due to this reason, all the following experiments only consider the \"Averaging\"method. Tables In the above experiments, we replace the out-ofdomain phrase table with the adapted phrase table. Here we combine these two phrase tables in a loglinear framework to see if we could obtain further improvement. To offer a clear description, we represent the out-of-domain phrase table and the adapted phrase table with \"OutBP\" and \"AdapBP\", respectively. Table 2 shows the results of experiments using different phrase tables. Applying our adaptation approach, both \"AdapBP\" and \"OutBP + AdapBP\" consistently outperform the baseline, and the lat- ter produces further improvements over the former. Specifically, the BLEU scores of the \"OutBP + AdapBP\" method are 20.70 and 21.20, which obtain 0.48 and 0.98 points higher than the baseline method, and 0.16 and 0.09 points higher than the 'AdapBP\" method. The underlying reason is that the probability distribution of each in-domain sentence often converges on some topics in the \"AdapBP\" method and some translation probabilities are overestimated, which leads to negative effects on the translation quality. By using two tables together, our approach reduces the bias introduced by \"AdapBP\", therefore further improving the translation quality.", |
| "cite_spans": [ |
| { |
| "start": 268, |
| "end": 288, |
| "text": "(Zens and Ney, 2004)", |
| "ref_id": "BIBREF41" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 659, |
| "end": 765, |
| "text": "Tables In the above experiments, we replace the out-ofdomain phrase table with the adapted phrase table.", |
| "ref_id": null |
| }, |
| { |
| "start": 1022, |
| "end": 1029, |
| "text": "Table 2", |
| "ref_id": "TABREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Effect of Different Smoothing Methods", |
| "sec_num": "4.2.1" |
| }, |
| { |
| "text": "Finally, we investigate the effect of in-domain monolingual corpus size on translation quality. In the experiment, we try different sizes of in-domain documents to train different monolingual topic models: from 5K to 80K with an increment of 5K each time. Note that here we only focus on the experiments using the \"OutBP + AdapBP\" method, because this method performs better in the previous experiments. Figure 1 shows the BLEU scores of the translation system on the test set. It can be seen that the more data, the better translation quality when the corpus size is less than 30K. The overall BLEU scores corresponding to the range of great N values are generally higher than the ones corresponding to the range of small N values. For example, the BLEU scores under the condition within the range [25K, 80K] are all higher than the ones within the range [5K, 20K] . When N is set to 55K, the BLEU score of our system is 21.40, with 1.18 gains on the baseline system. This difference is statistically significant at P < 0.01 using the significance test tool developed by Zhang et al.(2004) . For this experimental result, we speculate that with the increment of in-domain monolingual data, the corresponding topic models provide more accurate topic information to improve the translation system. However, this effect weakens when the monolingual corpora continue to increase.", |
| "cite_spans": [ |
| { |
| "start": 856, |
| "end": 865, |
| "text": "[5K, 20K]", |
| "ref_id": null |
| }, |
| { |
| "start": 1072, |
| "end": 1090, |
| "text": "Zhang et al.(2004)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 404, |
| "end": 412, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Effect of In-domain Monolingual Corpus Size", |
| "sec_num": "4.2.3" |
| }, |
| { |
| "text": "Most previous researches about translation model adaptation focused on parallel data collection. For example, Hildebrand et al.(2005) employed information retrieval technology to gather the bilingual sentences, which are similar to the test set, from available in-domain and out-of-domain training data to build an adaptive translation model. With the same motivation, Munteanu and Marcu (2005) extracted in-domain bilingual sentence pairs from comparable corpora. Since large-scale monolingual corpus is easier to obtain than parallel corpus, there have been some studies on how to generate parallel sentences with monolingual sentences. In this respect, Ueffing et al. (2008) explored semisupervised learning to obtain synthetic parallel sentences, and Wu et al. (2008) used an in-domain translation dictionary and monolingual corpora to adapt an out-of-domain translation model for the indomain text.", |
| "cite_spans": [ |
| { |
| "start": 110, |
| "end": 133, |
| "text": "Hildebrand et al.(2005)", |
| "ref_id": "BIBREF20" |
| }, |
| { |
| "start": 369, |
| "end": 394, |
| "text": "Munteanu and Marcu (2005)", |
| "ref_id": null |
| }, |
| { |
| "start": 656, |
| "end": 677, |
| "text": "Ueffing et al. (2008)", |
| "ref_id": "BIBREF39" |
| }, |
| { |
| "start": 755, |
| "end": 771, |
| "text": "Wu et al. (2008)", |
| "ref_id": "BIBREF40" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Differing from the above-mentioned works on the acquirement of bilingual resource, several studies (Foster and Kuhn, 2007; Civera and Juan, 2007; Lv et al., 2007) adopted mixture modeling framework to exploit the full potential of the existing parallel corpus. Under this framework, the training corpus is first divided into different parts, each of which is used to train a sub translation model, then these sub models are used together with different weights during decoding. In addition, discriminative weighting methods were proposed to assign appropriate weights to the sentences from training corpus (Matsoukas et al., 2009) or the phrase pairs of phrase table . Final experimental results show that without using any additional resources, these approaches all improve SMT performance sig-nificantly.", |
| "cite_spans": [ |
| { |
| "start": 99, |
| "end": 122, |
| "text": "(Foster and Kuhn, 2007;", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 123, |
| "end": 145, |
| "text": "Civera and Juan, 2007;", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 146, |
| "end": 162, |
| "text": "Lv et al., 2007)", |
| "ref_id": "BIBREF28" |
| }, |
| { |
| "start": 606, |
| "end": 630, |
| "text": "(Matsoukas et al., 2009)", |
| "ref_id": "BIBREF32" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Our method deals with translation model adaptation by making use of the topical context, so let us take a look at the recent research development on the application of topic models in SMT. Assuming each bilingual sentence constitutes a mixture of hidden topics and each word pair follows a topic-specific bilingual translation model, Xing (2006,2007) presented a bilingual topical admixture formalism to improve word alignment by capturing topic sharing at different levels of linguistic granularity. Tam et al.(2007) proposed a bilingual LSA, which enforces one-to-one topic correspondence and enables latent topic distributions to be efficiently transferred across languages, to crosslingual language modeling and translation lexicon adaptation. Recently, Gong and Zhou (2010) also applied topic modeling into domain adaptation in SMT. Their method employed one additional feature function to capture the topic inherent in the source phrase and help the decoder dynamically choose related target phrases according to the specific topic of the source phrase.", |
| "cite_spans": [ |
| { |
| "start": 334, |
| "end": 350, |
| "text": "Xing (2006,2007)", |
| "ref_id": null |
| }, |
| { |
| "start": 501, |
| "end": 517, |
| "text": "Tam et al.(2007)", |
| "ref_id": "BIBREF38" |
| }, |
| { |
| "start": 758, |
| "end": 778, |
| "text": "Gong and Zhou (2010)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Besides, our approach is also related to contextdependent translation. Recent studies have shown that SMT systems can benefit from the utilization of context information. For example, triggerbased lexicon model (Hasan et al., 2008; Mauser et al., 2009) and context-dependent translation selection (Chan et al., 2007; Carpuat and Wu, 2007; . The former generated triplets to capture long-distance dependencies that go beyond the local context of phrases, and the latter built the classifiers which combine rich context information to better select translation during decoding. With the consideration of various local context features, these approaches all yielded stable improvements on different translation tasks.", |
| "cite_spans": [ |
| { |
| "start": 211, |
| "end": 231, |
| "text": "(Hasan et al., 2008;", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 232, |
| "end": 252, |
| "text": "Mauser et al., 2009)", |
| "ref_id": "BIBREF31" |
| }, |
| { |
| "start": 297, |
| "end": 316, |
| "text": "(Chan et al., 2007;", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 317, |
| "end": 338, |
| "text": "Carpuat and Wu, 2007;", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "As compared to the above-mentioned works, our work has the following differences.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "\u2022 We focus on how to adapt a translation model for domain-specific translation task with the help of additional in-domain monolingual corpora, which are far from full exploitation in the parallel data collection and mixture modeling framework.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "\u2022 In addition to the utilization of in-domain monolingual corpora, our method is different from the previous works (Zhao and Xing, 2006; Zhao and Xing, 2007; Tam et al., 2007; Gong and Zhou, 2010) in the following aspects: (1) we use a different topic model -HTMM which has different assumption from PLSA and LDA;", |
| "cite_spans": [ |
| { |
| "start": 115, |
| "end": 136, |
| "text": "(Zhao and Xing, 2006;", |
| "ref_id": "BIBREF44" |
| }, |
| { |
| "start": 137, |
| "end": 157, |
| "text": "Zhao and Xing, 2007;", |
| "ref_id": "BIBREF45" |
| }, |
| { |
| "start": 158, |
| "end": 175, |
| "text": "Tam et al., 2007;", |
| "ref_id": "BIBREF38" |
| }, |
| { |
| "start": 176, |
| "end": 196, |
| "text": "Gong and Zhou, 2010)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "(2) rather than modeling topic-dependent translation lexicons in the training process, we estimate topic-specific lexical probability by taking account of topical context when extracting word pairs, so our method can also be directly applied to topic-dependent phrase probability modeling.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "(3) Instead of rescoring phrase pairs online, our approach calculate the translation probabilities offline, which brings no additional burden to translation systems and is suitable to translate the texts without the topic distribution information.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "\u2022 Different from trigger-based lexicon model and context-dependent translation selection both of which put emphasis on solving the translation ambiguity by the exploitation of the context information at the sentence level, we adopt the topical context information in our method for the following reasons: (1) the topic information captures the context information beyond the scope of sentence;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "(2) the topical context information is integrated into the posterior probability distribution, avoiding the sparseness of word or POS features; (3) the topical context information allows for more fine-grained distinction of different translations than the genre information of corpus.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "This paper presents a novel method for SMT system adaptation by making use of the monolingual corpora in new domains. Our approach first estimates the translation probabilities from the out-ofdomain bilingual corpus given the topic information, and then rescores the phrase pairs via topic mapping and phrase-topic distribution probability estimation from in-domain monolingual corpora. Experimental results show that our method achieves better performance than the baseline system, without increasing the burden of the translation system. In the future, we will verify our method on oth-er language pairs, for example, Chinese to Japanese. Furthermore, since the in-domain phrase-topic distribution is currently estimated with simple smoothing interpolations, we expect that the translation system could benefit from other sophisticated smoothing methods. Finally, the reasonable estimation of topic number for better translation model adaptation will also become our study emphasis.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion and future work", |
| "sec_num": "6" |
| }, |
| { |
| "text": "Due to the limit of space, we omit the description of the calculation method of the phrase probability \u03c6(f |\u1ebd), which can be adjusted in a similar way to \u03c6(\u1ebd|f ) with the help of in-domain monolingual corpus in the target language.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://blog.sohu.com/ 2 http://u.cs.biu.ac.il/ koppel/BlogCorpus.html", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "The authors were supported by 863 State Key Project (Grant No. 2011AA01A207), National Natural Science Foundation of China (Grant Nos. 61005052 and 61103101), Key Technologies R&D Program of China (Grant No. 2012BAH14F03). We thank the anonymous reviewers for their insightful comments. We are also grateful to Ruiyu Fang and Jinming Hu for their kind help in data processing.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgement", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Unsupervised Language Model Adaptation", |
| "authors": [ |
| { |
| "first": "Michiel", |
| "middle": [], |
| "last": "Bacchiani", |
| "suffix": "" |
| }, |
| { |
| "first": "Brian", |
| "middle": [], |
| "last": "Roark", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proc. of ICAS-SP 2003", |
| "volume": "", |
| "issue": "", |
| "pages": "224--227", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Michiel Bacchiani and Brian Roark. 2003. Unsuper- vised Language Model Adaptation. In Proc. of ICAS- SP 2003, pages 224-227.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Improving Machine Translation Performance by Exploiting Non-Parallel Corpora", |
| "authors": [ |
| { |
| "first": "Michiel", |
| "middle": [], |
| "last": "Bacchiani", |
| "suffix": "" |
| }, |
| { |
| "first": "Brian", |
| "middle": [], |
| "last": "Roark", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "477--504", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Michiel Bacchiani and Brian Roark. 2005. Improving Machine Translation Performance by Exploiting Non- Parallel Corpora. Computational Linguistics, pages 477-504.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Domain Adaptation for Statistical Machine Translation with Monolingual Resources", |
| "authors": [ |
| { |
| "first": "Nicola", |
| "middle": [], |
| "last": "Bertoldi", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcello", |
| "middle": [], |
| "last": "Federico", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Proc. of ACL Workshop", |
| "volume": "", |
| "issue": "", |
| "pages": "182--189", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nicola Bertoldi and Marcello Federico. 2009. Domain Adaptation for Statistical Machine Translation with Monolingual Resources. In Proc. of ACL Workshop 2009, pages 182-189.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Latent Dirichlet Allocation. Journal of Machine Learning", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [ |
| "M" |
| ], |
| "last": "Blei", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "993--1022", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David M. Blei. 2003. Latent Dirichlet Allocation. Jour- nal of Machine Learning, pages 993-1022.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Language Model Adaptation in Machine Translation from Speech", |
| "authors": [ |
| { |
| "first": "Ivan", |
| "middle": [], |
| "last": "Bulyko", |
| "suffix": "" |
| }, |
| { |
| "first": "Spyros", |
| "middle": [], |
| "last": "Matsoukas", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Schwartz", |
| "suffix": "" |
| }, |
| { |
| "first": "Long", |
| "middle": [], |
| "last": "Nguyen", |
| "suffix": "" |
| }, |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Makhoul", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proc. of ICASSP", |
| "volume": "", |
| "issue": "", |
| "pages": "117--120", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ivan Bulyko, Spyros Matsoukas, Richard Schwartz, Long Nguyen and John Makhoul. 2007. Language Model Adaptation in Machine Translation from Speech. In Proc. of ICASSP 2007, pages 117-120.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Improving Statistical Machine Translation Using Word Sense Disambiguation", |
| "authors": [ |
| { |
| "first": "Marine", |
| "middle": [], |
| "last": "Carpuat", |
| "suffix": "" |
| }, |
| { |
| "first": "Dekai", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proc. of EMNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "61--72", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marine Carpuat and Dekai Wu. 2007. Improving Statis- tical Machine Translation Using Word Sense Disam- biguation. In Proc. of EMNLP 2007, pages 61-72.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Word sense disambiguation improves statistical machine translation", |
| "authors": [ |
| { |
| "first": "Yee", |
| "middle": [], |
| "last": "Seng Chan", |
| "suffix": "" |
| }, |
| { |
| "first": "Hwee Tou", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Chiang", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proc. of ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "33--40", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yee Seng Chan, Hwee Tou Ng, and David Chiang. 2006. Word sense disambiguation improves statistical ma- chine translation. In Proc. of ACL 2007, pages 33-40.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Bilingual Sense Similarity for Statistical Machine Translation", |
| "authors": [ |
| { |
| "first": "Boxing", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "George", |
| "middle": [], |
| "last": "Foster", |
| "suffix": "" |
| }, |
| { |
| "first": "Roland", |
| "middle": [], |
| "last": "Kuhn", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proc. of ACL 2010", |
| "volume": "", |
| "issue": "", |
| "pages": "834--843", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Boxing Chen, George Foster and Roland Kuhn. 2010. Bilingual Sense Similarity for Statistical Machine Translation. In Proc. of ACL 2010, pages 834-843.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Hierarchical Phrase-Based Translation", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Chiang", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "201--228", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David Chiang. 2007. Hierarchical Phrase-Based Trans- lation. Computational Linguistics, pages 201-228.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Learning to Translate with Source and Target Syntax", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Chiang", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proc. of ACL 2010", |
| "volume": "", |
| "issue": "", |
| "pages": "1443--1452", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David Chiang. 2010. Learning to Translate with Source and Target Syntax. In Proc. of ACL 2010, pages 1443- 1452.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Domain Adaptation in Statistical Machine Translation with Mixture Modelling", |
| "authors": [ |
| { |
| "first": "Jorge", |
| "middle": [], |
| "last": "Civera", |
| "suffix": "" |
| }, |
| { |
| "first": "Alfons", |
| "middle": [], |
| "last": "Juan", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proc. of the Second Workshop on Statistical Machine Translation", |
| "volume": "", |
| "issue": "", |
| "pages": "177--180", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jorge Civera and Alfons Juan. 2007. Domain Adaptation in Statistical Machine Translation with Mixture Mod- elling. In Proc. of the Second Workshop on Statistical Machine Translation, pages 177-180.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Language Model Adaptation for Statistical Machine Translation Based on Information Retrieval", |
| "authors": [ |
| { |
| "first": "Matthias", |
| "middle": [], |
| "last": "Eck", |
| "suffix": "" |
| }, |
| { |
| "first": "Stephan", |
| "middle": [], |
| "last": "Vogel", |
| "suffix": "" |
| }, |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Waibel", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proc. of Fourth International Conference on Language Resources and Evaluation", |
| "volume": "", |
| "issue": "", |
| "pages": "327--330", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Matthias Eck, Stephan Vogel and Alex Waibel. 2004. Language Model Adaptation for Statistical Machine Translation Based on Information Retrieval. In Proc. of Fourth International Conference on Language Re- sources and Evaluation, pages 327-330.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Low Cost Portability for Statistical Machine Translation Based on N-gram Coverage", |
| "authors": [ |
| { |
| "first": "Matthias", |
| "middle": [], |
| "last": "Eck", |
| "suffix": "" |
| }, |
| { |
| "first": "Stephan", |
| "middle": [], |
| "last": "Vogel", |
| "suffix": "" |
| }, |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Waibel", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proc. of MT Summit 2005", |
| "volume": "", |
| "issue": "", |
| "pages": "227--234", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Matthias Eck, Stephan Vogel and Alex Waibel. 2005. Low Cost Portability for Statistical Machine Transla- tion Based on N-gram Coverage. In Proc. of MT Sum- mit 2005, pages 227-234.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Mixture Model Adaptation for SMT", |
| "authors": [ |
| { |
| "first": "George", |
| "middle": [], |
| "last": "Foster", |
| "suffix": "" |
| }, |
| { |
| "first": "Roland", |
| "middle": [], |
| "last": "Kuhn", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proc. of the Second Workshop on Statistical Machine Translation", |
| "volume": "", |
| "issue": "", |
| "pages": "128--135", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "George Foster and Roland Kuhn. 2007. Mixture Model Adaptation for SMT. In Proc. of the Second Workshop on Statistical Machine Translation, pages 128-135.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Discriminative Instance Weighting for Domain Adaptation in Statistical Machine Translation", |
| "authors": [ |
| { |
| "first": "George", |
| "middle": [], |
| "last": "Foster", |
| "suffix": "" |
| }, |
| { |
| "first": "Cyril", |
| "middle": [], |
| "last": "Goutte", |
| "suffix": "" |
| }, |
| { |
| "first": "Roland", |
| "middle": [], |
| "last": "Kuhn", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proc. of EMNLP 2010", |
| "volume": "", |
| "issue": "", |
| "pages": "451--459", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "George Foster, Cyril Goutte and Roland Kuhn. 2010. Discriminative Instance Weighting for Domain Adap- tation in Statistical Machine Translation. In Proc. of EMNLP 2010, pages 451-459.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Scalable Inference and Training of Context-Rich Syntactic Translation Models", |
| "authors": [ |
| { |
| "first": "Michel", |
| "middle": [], |
| "last": "Galley", |
| "suffix": "" |
| }, |
| { |
| "first": "Jonathan", |
| "middle": [], |
| "last": "Graehl", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Knight", |
| "suffix": "" |
| }, |
| { |
| "first": "Daniel", |
| "middle": [], |
| "last": "Marcu", |
| "suffix": "" |
| }, |
| { |
| "first": "Steve", |
| "middle": [], |
| "last": "Deneefe", |
| "suffix": "" |
| }, |
| { |
| "first": "Wei", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Ignacio", |
| "middle": [], |
| "last": "Thayer", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proc. of ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "961--968", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Michel Galley, Jonathan Graehl, Kevin Knight, Daniel Marcu, Steve DeNeefe, Wei Wang and Ignacio Thay- er. 2006. Scalable Inference and Training of Context- Rich Syntactic Translation Models. In Proc. of ACL 2006, pages 961-968.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Improve SMT with Source-side Topic-Document Distributions", |
| "authors": [ |
| { |
| "first": "Zhengxian", |
| "middle": [], |
| "last": "Gong", |
| "suffix": "" |
| }, |
| { |
| "first": "Guodong", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proc. of MT SUMMIT 2010", |
| "volume": "", |
| "issue": "", |
| "pages": "24--28", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zhengxian Gong and Guodong Zhou. 2010. Improve SMT with Source-side Topic-Document Distributions. In Proc. of MT SUMMIT 2010, pages 24-28.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Hidden Topic Markov Models", |
| "authors": [ |
| { |
| "first": "Amit", |
| "middle": [], |
| "last": "Gruber", |
| "suffix": "" |
| }, |
| { |
| "first": "Michal", |
| "middle": [], |
| "last": "Rosen-Zvi", |
| "suffix": "" |
| }, |
| { |
| "first": "Yair", |
| "middle": [], |
| "last": "Weiss", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Journal of Machine Learning Research", |
| "volume": "", |
| "issue": "", |
| "pages": "163--170", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Amit Gruber, Michal Rosen-Zvi and Yair Weiss. 2007. Hidden Topic Markov Models. In Journal of Machine Learning Research, pages 163-170.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Triplet Lexicon Models for Statistical Machine Translation", |
| "authors": [ |
| { |
| "first": "Sa\u0161a", |
| "middle": [], |
| "last": "Hasan", |
| "suffix": "" |
| }, |
| { |
| "first": "Juri", |
| "middle": [], |
| "last": "Ganitkevitch", |
| "suffix": "" |
| }, |
| { |
| "first": "Hermann", |
| "middle": [], |
| "last": "Ney", |
| "suffix": "" |
| }, |
| { |
| "first": "Jes\u00fas", |
| "middle": [], |
| "last": "Andr\u00e9s-Ferrer", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proc. of EMNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "372--381", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sa\u0161a Hasan, Juri Ganitkevitch, Hermann Ney and Jes\u00fas Andr\u00e9s-Ferrer 2008. Triplet Lexicon Models for S- tatistical Machine Translation. In Proc. of EMNLP 2008, pages 372-381.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Improving Statistical Machine Translation using Lexicalized Rule Selection", |
| "authors": [ |
| { |
| "first": "Zhongjun", |
| "middle": [], |
| "last": "He", |
| "suffix": "" |
| }, |
| { |
| "first": "Qun", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Shouxun", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proc. of COLING", |
| "volume": "", |
| "issue": "", |
| "pages": "321--328", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zhongjun He, Qun Liu and Shouxun Lin. 2008. Improv- ing Statistical Machine Translation using Lexicalized Rule Selection. In Proc. of COLING 2008, pages 321- 328.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Adaptation of the Translation Model for Statistical Machine Translation based on Information Retrieval", |
| "authors": [ |
| { |
| "first": "Almut Silja", |
| "middle": [], |
| "last": "Hildebrand", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proc. of EAMT 2005", |
| "volume": "", |
| "issue": "", |
| "pages": "133--142", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Almut Silja Hildebrand. 2005. Adaptation of the Trans- lation Model for Statistical Machine Translation based on Information Retrieval. In Proc. of EAMT 2005, pages 133-142.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Probabilistic Latent Semantic Indexing", |
| "authors": [ |
| { |
| "first": "Thomas", |
| "middle": [], |
| "last": "Hofmann", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "Proc. of SIGIR 1999", |
| "volume": "", |
| "issue": "", |
| "pages": "50--57", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Thomas Hofmann. 1999. Probabilistic Latent Semantic Indexing. In Proc. of SIGIR 1999, pages 50-57.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "A Systematic Comparison of Various Statistical Alignment Models", |
| "authors": [ |
| { |
| "first": "Joseph", |
| "middle": [], |
| "last": "Franz", |
| "suffix": "" |
| }, |
| { |
| "first": "Hermann", |
| "middle": [], |
| "last": "Och", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ney", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "19--51", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Franz Joseph Och and Hermann Ney. 2003. A Systemat- ic Comparison of Various Statistical Alignment Mod- els. Computational Linguistics, pages 19-51.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "The Alignment Template Approach to Statistical Machine Translation", |
| "authors": [ |
| { |
| "first": "Joseph", |
| "middle": [], |
| "last": "Franz", |
| "suffix": "" |
| }, |
| { |
| "first": "Hermann", |
| "middle": [], |
| "last": "Och", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ney", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "417--449", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Franz Joseph Och and Hermann Ney. 2004. The Align- ment Template Approach to Statistical Machine Trans- lation. Computational Linguistics, pages 417-449.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Statistical phrase-based translation", |
| "authors": [ |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| }, |
| { |
| "first": "Josef", |
| "middle": [], |
| "last": "Franz", |
| "suffix": "" |
| }, |
| { |
| "first": "Daniel", |
| "middle": [], |
| "last": "Och", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Marcu", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proc. of HLT-NAACL 2003", |
| "volume": "", |
| "issue": "", |
| "pages": "127--133", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Philipp Koehn, Franz Josef Och and Daniel Marcu. 2003. Statistical phrase-based translation. In Proc. of HLT- NAACL 2003, pages 127-133.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "Statistical Significance Tests for Machine Translation Evaluation", |
| "authors": [ |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proc. of EMNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "388--395", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Philipp Koehn. 2004. Statistical Significance Tests for Machine Translation Evaluation. In Proc. of EMNLP 2004, pages 388-395.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "Moses: Open source toolkit for statistical machine translation", |
| "authors": [ |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| }, |
| { |
| "first": "Hieu", |
| "middle": [], |
| "last": "Hoang", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexandra", |
| "middle": [], |
| "last": "Birch", |
| "suffix": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Callison-Burch", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcello", |
| "middle": [], |
| "last": "Federico", |
| "suffix": "" |
| }, |
| { |
| "first": "Nicola", |
| "middle": [], |
| "last": "Bertoldi", |
| "suffix": "" |
| }, |
| { |
| "first": "Brooke", |
| "middle": [], |
| "last": "Cowan", |
| "suffix": "" |
| }, |
| { |
| "first": "Wade", |
| "middle": [], |
| "last": "Shen", |
| "suffix": "" |
| }, |
| { |
| "first": "Christine", |
| "middle": [], |
| "last": "Moran", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Zens", |
| "suffix": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Dyer", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proc. of ACL 2007, Demonstration Session", |
| "volume": "", |
| "issue": "", |
| "pages": "177--180", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Philipp Koehn, Hieu Hoang, Alexandra Birch, Chris Callison-Burch, Marcello Federico, Nicola Bertoldi, Brooke Cowan, Wade Shen, Christine Moran, Richard Zens, Chris Dyer, Ondrej Bojar, Alexandra Con- stantin, and Evan Herbst. 2007. Moses: Open source toolkit for statistical machine translation. In Proc. of ACL 2007, Demonstration Session, pages 177-180.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "Treeto-String Alignment Template for Statistical Machine Translation", |
| "authors": [ |
| { |
| "first": "Yang", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Qun", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Shouxun", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proc. of ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "609--616", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yang Liu, Qun Liu and Shouxun Lin. 2006. Tree- to-String Alignment Template for Statistical Machine Translation. In Proc. of ACL 2006, pages 609-616.", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "Improving Statistical Machine Translation Performance by Training Data Selection and Optimization", |
| "authors": [ |
| { |
| "first": "Yajuan", |
| "middle": [], |
| "last": "Lv", |
| "suffix": "" |
| }, |
| { |
| "first": "Jin", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "" |
| }, |
| { |
| "first": "Qun", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proc. of EMNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "343--350", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yajuan Lv, Jin Huang and Qun Liu. 2007. Improv- ing Statistical Machine Translation Performance by Training Data Selection and Optimization. In Proc. of EMNLP 2007, pages 343-350.", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "The RWTH Statistical Machine Translation System for the IWSLT", |
| "authors": [ |
| { |
| "first": "Arne", |
| "middle": [], |
| "last": "Mauser", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Zens", |
| "suffix": "" |
| }, |
| { |
| "first": "Evgeny", |
| "middle": [], |
| "last": "Matusov", |
| "suffix": "" |
| }, |
| { |
| "first": "Sa\u0161a", |
| "middle": [], |
| "last": "Hasan", |
| "suffix": "" |
| }, |
| { |
| "first": "Hermann", |
| "middle": [], |
| "last": "Ney", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Arne Mauser, Richard Zens and Evgeny Matusov, Sa\u0161a Hasan and Hermann Ney. 2006. The RWTH Statisti- cal Machine Translation System for the IWSLT 2006", |
| "links": null |
| }, |
| "BIBREF30": { |
| "ref_id": "b30", |
| "title": "Evaluation", |
| "authors": [], |
| "year": null, |
| "venue": "Proc. of International Workshop on Spoken Language Translation", |
| "volume": "", |
| "issue": "", |
| "pages": "103--110", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Evaluation. In Proc. of International Workshop on Spoken Language Translation, pages 103-110.", |
| "links": null |
| }, |
| "BIBREF31": { |
| "ref_id": "b31", |
| "title": "Extending Statistical Machine Translation with Discriminative and Trigger-Based Lexicon Models", |
| "authors": [ |
| { |
| "first": "Arne", |
| "middle": [], |
| "last": "Mauser", |
| "suffix": "" |
| }, |
| { |
| "first": "Sa\u0161a", |
| "middle": [], |
| "last": "Hasan", |
| "suffix": "" |
| }, |
| { |
| "first": "Hermann", |
| "middle": [], |
| "last": "Ney", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Proc. of ACL 2009", |
| "volume": "", |
| "issue": "", |
| "pages": "210--218", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Arne Mauser, Sa\u0161a Hasan and Hermann Ney 2009. Ex- tending Statistical Machine Translation with Discrimi- native and Trigger-Based Lexicon Models. In Proc. of ACL 2009, pages 210-218.", |
| "links": null |
| }, |
| "BIBREF32": { |
| "ref_id": "b32", |
| "title": "Discriminative Corpus Weight Estimation for Machine Translation", |
| "authors": [ |
| { |
| "first": "Spyros", |
| "middle": [], |
| "last": "Matsoukas", |
| "suffix": "" |
| }, |
| { |
| "first": "I", |
| "middle": [], |
| "last": "Antti-Veikko", |
| "suffix": "" |
| }, |
| { |
| "first": "Bing", |
| "middle": [], |
| "last": "Rosti", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Proc. of EMNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "708--717", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Spyros Matsoukas, Antti-Veikko I. Rosti and Bing Zhang 2009. Discriminative Corpus Weight Estimation for Machine Translation. In Proc. of EMNLP 2009, pages 708-717.", |
| "links": null |
| }, |
| "BIBREF33": { |
| "ref_id": "b33", |
| "title": "Topic Adaptation for Lecture Translation through Bilingual Latent Semantic Models", |
| "authors": [ |
| { |
| "first": "Nick", |
| "middle": [], |
| "last": "Ruiz", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcello", |
| "middle": [], |
| "last": "Federico", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proc. of ACL Workshop", |
| "volume": "", |
| "issue": "", |
| "pages": "294--302", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nick Ruiz and Marcello Federico. 2011. Topic Adapta- tion for Lecture Translation through Bilingual Latent Semantic Models. In Proc. of ACL Workshop 2011, pages 294-302.", |
| "links": null |
| }, |
| "BIBREF34": { |
| "ref_id": "b34", |
| "title": "BLEU: A Method for Automatic Evaluation of Machine Translation", |
| "authors": [ |
| { |
| "first": "Kishore", |
| "middle": [], |
| "last": "Papineni", |
| "suffix": "" |
| }, |
| { |
| "first": "Salim", |
| "middle": [], |
| "last": "Roukos", |
| "suffix": "" |
| }, |
| { |
| "first": "Todd", |
| "middle": [], |
| "last": "Ward", |
| "suffix": "" |
| }, |
| { |
| "first": "Weijing", |
| "middle": [], |
| "last": "Zhu", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proc. of ACL 2002", |
| "volume": "", |
| "issue": "", |
| "pages": "311--318", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kishore Papineni, Salim Roukos, Todd Ward and WeiJing Zhu. 2002. BLEU: A Method for Automatic Evalu- ation of Machine Translation. In Proc. of ACL 2002, pages 311-318.", |
| "links": null |
| }, |
| "BIBREF35": { |
| "ref_id": "b35", |
| "title": "Effects of Age and Gender on Blogging", |
| "authors": [ |
| { |
| "first": "Jonathan", |
| "middle": [], |
| "last": "Schler", |
| "suffix": "" |
| }, |
| { |
| "first": "Moshe", |
| "middle": [], |
| "last": "Koppel", |
| "suffix": "" |
| }, |
| { |
| "first": "Shlomo", |
| "middle": [], |
| "last": "Argamon", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Pennebaker", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proc. of 2006 AAAI Spring Symposium on Computational Approaches for Analyzing Weblogs", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jonathan Schler, Moshe Koppel, Shlomo Argamon and James Pennebaker. 2006. Effects of Age and Gender on Blogging. In Proc. of 2006 AAAI Spring Sympo- sium on Computational Approaches for Analyzing We- blogs.", |
| "links": null |
| }, |
| "BIBREF36": { |
| "ref_id": "b36", |
| "title": "Translation Model Adaptation for an Arabic/french News Translation System by Lightly-supervised Training", |
| "authors": [ |
| { |
| "first": "Holger", |
| "middle": [], |
| "last": "Schwenk", |
| "suffix": "" |
| }, |
| { |
| "first": "Jean", |
| "middle": [], |
| "last": "Senellart", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Proc. of MT Summit XII", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Holger Schwenk and Jean Senellart. 2009. Translation Model Adaptation for an Arabic/french News Transla- tion System by Lightly-supervised Training. In Proc. of MT Summit XII.", |
| "links": null |
| }, |
| "BIBREF37": { |
| "ref_id": "b37", |
| "title": "Srilm -An Extensible Language Modeling Toolkit", |
| "authors": [ |
| { |
| "first": "Andreas", |
| "middle": [], |
| "last": "Stolcke", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proc. of ICSLP 2002", |
| "volume": "", |
| "issue": "", |
| "pages": "901--904", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Andreas Stolcke. 2002. Srilm -An Extensible Language Modeling Toolkit. In Proc. of ICSLP 2002, pages 901- 904.", |
| "links": null |
| }, |
| "BIBREF38": { |
| "ref_id": "b38", |
| "title": "Bilingual LSA-based adaptation for statistical machine translation", |
| "authors": [ |
| { |
| "first": "Yik-Cheung", |
| "middle": [], |
| "last": "Tam", |
| "suffix": "" |
| }, |
| { |
| "first": "Ian", |
| "middle": [ |
| "R" |
| ], |
| "last": "Lane", |
| "suffix": "" |
| }, |
| { |
| "first": "Tanja", |
| "middle": [], |
| "last": "Schultz", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Machine Translation", |
| "volume": "", |
| "issue": "", |
| "pages": "187--207", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yik-Cheung Tam, Ian R. Lane and Tanja Schultz. 2007. Bilingual LSA-based adaptation for statistical machine translation. Machine Translation, pages 187-207.", |
| "links": null |
| }, |
| "BIBREF39": { |
| "ref_id": "b39", |
| "title": "Semi-supervised Model Adaptation for Statistical Machine Translation. Machine Translation", |
| "authors": [ |
| { |
| "first": "Nicola", |
| "middle": [], |
| "last": "Ueffing", |
| "suffix": "" |
| }, |
| { |
| "first": "Gholamreza", |
| "middle": [], |
| "last": "Haffari", |
| "suffix": "" |
| }, |
| { |
| "first": "Anoop", |
| "middle": [], |
| "last": "Sarkar", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "77--94", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nicola Ueffing, Gholamreza Haffari and Anoop Sarkar. 2008. Semi-supervised Model Adaptation for Statisti- cal Machine Translation. Machine Translation, pages 77-94.", |
| "links": null |
| }, |
| "BIBREF40": { |
| "ref_id": "b40", |
| "title": "Domain Adaptation for Statistical Machine Translation with Domain Dictionary and Monolingual Corpora", |
| "authors": [ |
| { |
| "first": "Hua", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Haifeng", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Chengqing", |
| "middle": [], |
| "last": "Zong", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proc. of COLING", |
| "volume": "", |
| "issue": "", |
| "pages": "993--1000", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hua Wu, Haifeng Wang and Chengqing Zong. 2008. Do- main Adaptation for Statistical Machine Translation with Domain Dictionary and Monolingual Corpora. In Proc. of COLING 2008, pages 993-1000.", |
| "links": null |
| }, |
| "BIBREF41": { |
| "ref_id": "b41", |
| "title": "Improvments in phrase-based statistical machine translation", |
| "authors": [ |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Zens", |
| "suffix": "" |
| }, |
| { |
| "first": "Hermann", |
| "middle": [], |
| "last": "Ney", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proc. of NAACL", |
| "volume": "", |
| "issue": "", |
| "pages": "257--264", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Richard Zens and Hermann Ney. 2004. Improvments in phrase-based statistical machine translation. In Proc. of NAACL 2004, pages 257-264.", |
| "links": null |
| }, |
| "BIBREF42": { |
| "ref_id": "b42", |
| "title": "Distributed Language Modeling for N-best List Re-ranking", |
| "authors": [ |
| { |
| "first": "Ying", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Almut", |
| "middle": [], |
| "last": "", |
| "suffix": "" |
| }, |
| { |
| "first": "Silja", |
| "middle": [], |
| "last": "Hildebrand", |
| "suffix": "" |
| }, |
| { |
| "first": "Stephan", |
| "middle": [], |
| "last": "Vogel", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proc. of EMNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "216--223", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ying Zhang, Almut Silja Hildebrand and Stephan Vogel. 2006. Distributed Language Modeling for N-best List Re-ranking. In Proc. of EMNLP 2006, pages 216-223.", |
| "links": null |
| }, |
| "BIBREF43": { |
| "ref_id": "b43", |
| "title": "Language Model Adaptation for Statistical Machine Translation with Structured Query Models", |
| "authors": [ |
| { |
| "first": "Bing", |
| "middle": [], |
| "last": "Zhao", |
| "suffix": "" |
| }, |
| { |
| "first": "Matthias", |
| "middle": [], |
| "last": "Eck", |
| "suffix": "" |
| }, |
| { |
| "first": "Stephan", |
| "middle": [], |
| "last": "Vogel", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proc. of COLING", |
| "volume": "", |
| "issue": "", |
| "pages": "411--417", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bing Zhao, Matthias Eck and Stephan Vogel. 2004. Language Model Adaptation for Statistical Machine Translation with Structured Query Models. In Proc. of COLING 2004, pages 411-417.", |
| "links": null |
| }, |
| "BIBREF44": { |
| "ref_id": "b44", |
| "title": "BiTAM: Bilingual Topic AdMixture Models for Word Alignment", |
| "authors": [ |
| { |
| "first": "Bing", |
| "middle": [], |
| "last": "Zhao", |
| "suffix": "" |
| }, |
| { |
| "first": "Eric", |
| "middle": [ |
| "P" |
| ], |
| "last": "Xing", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proc. of ACL/COLING 2006", |
| "volume": "", |
| "issue": "", |
| "pages": "969--976", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bing Zhao and Eric P. Xing. 2006. BiTAM: Bilingual Topic AdMixture Models for Word Alignment. In Proc. of ACL/COLING 2006, pages 969-976.", |
| "links": null |
| }, |
| "BIBREF45": { |
| "ref_id": "b45", |
| "title": "HM-BiTAM: Bilingual Topic Exploration, Word Alignment, and Translation", |
| "authors": [ |
| { |
| "first": "Bing", |
| "middle": [], |
| "last": "Zhao", |
| "suffix": "" |
| }, |
| { |
| "first": "Eric", |
| "middle": [ |
| "P" |
| ], |
| "last": "Xing", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proc. of NIPS 2007", |
| "volume": "", |
| "issue": "", |
| "pages": "1--8", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bing Zhao and Eric P. Xing. 2007. HM-BiTAM: Bilin- gual Topic Exploration, Word Alignment, and Trans- lation. In Proc. of NIPS 2007, pages 1-8.", |
| "links": null |
| }, |
| "BIBREF46": { |
| "ref_id": "b46", |
| "title": "Maximum Entropy based Rule Selection Model for Syntax-based Statistical Machine Translation", |
| "authors": [ |
| { |
| "first": "Qun", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Zhongjun", |
| "middle": [], |
| "last": "He", |
| "suffix": "" |
| }, |
| { |
| "first": "Yang", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Shouxun", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proc. of EMNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "89--97", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Qun Liu, Zhongjun He, Yang Liu and Shouxun Lin. 2008. Maximum Entropy based Rule Selection Model for Syntax-based Statistical Machine Translation. In Proc. of EMNLP 2008, pages 89-97.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "Effect of in-domain monolingual corpus size on translation quality.", |
| "num": null |
| }, |
| "TABREF1": { |
| "content": "<table/>", |
| "html": null, |
| "type_str": "table", |
| "num": null, |
| "text": "Experimental results using different smoothing methods." |
| }, |
| "TABREF2": { |
| "content": "<table><tr><td/><td>Dev) MT06</td><td>(Tst) MT08</td></tr><tr><td/><td>Web</td><td>Weblog</td></tr><tr><td>Baseline</td><td>30.98</td><td>20.22</td></tr><tr><td>AdapBp (5K)</td><td>31.51</td><td>20.54</td></tr><tr><td>+ OutBp</td><td>31.84</td><td>20.70</td></tr><tr><td>AdapBp (40K)</td><td>31.89</td><td>21.11</td></tr><tr><td>+ OutBp</td><td>32.05</td><td>21.20</td></tr></table>", |
| "html": null, |
| "type_str": "table", |
| "num": null, |
| "text": "" |
| }, |
| "TABREF3": { |
| "content": "<table/>", |
| "html": null, |
| "type_str": "table", |
| "num": null, |
| "text": "Experimental results using different phrase tables. OutBp: the out-of-domain phrase table. AdapBp: the adapted phrase table." |
| } |
| } |
| } |
| } |