ACL-OCL / Base_JSON /prefixE /json /E14 /E14-1001.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "E14-1001",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T10:39:58.030155Z"
},
"title": "Improving Word Alignment Using Linguistic Code Switching Data",
"authors": [
{
"first": "Fei",
"middle": [],
"last": "Huang",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Temple University Computer and Information Sciences",
"location": {
"addrLine": "324 Wachman Hall Philadelphia",
"postCode": "19122",
"region": "PA"
}
},
"email": "fei.huang@temple.edu"
},
{
"first": "Alexander",
"middle": [],
"last": "Yates",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Temple University Computer and Information Sciences",
"location": {
"addrLine": "324 Wachman Hall Philadelphia",
"postCode": "19122",
"region": "PA"
}
},
"email": "yates@temple.edu"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "Linguist Code Switching (LCS) is a situation where two or more languages show up in the context of a single conversation. For example, in English-Chinese code switching, there might be a sentence like \"\u2022 \u201a15\u00a9\u00a8 k \u2021meeting (We will have a meeting in 15 minutes)\". Traditional machine translation (MT) systems treat LCS data as noise, or just as regular sentences. However, if LCS data is processed intelligently, it can provide a useful signal for training word alignment and MT models. Moreover, LCS data is from non-news sources which can enhance the diversity of training data for MT. In this paper, we first extract constraints from this code switching data and then incorporate them into a word alignment model training procedure. We also show that by using the code switching data, we can jointly train a word alignment model and a language model using cotraining. Our techniques for incorporating LCS data improve by 2.64 in BLEU score over a baseline MT system trained using only standard sentence-aligned corpora.",
"pdf_parse": {
"paper_id": "E14-1001",
"_pdf_hash": "",
"abstract": [
{
"text": "Linguist Code Switching (LCS) is a situation where two or more languages show up in the context of a single conversation. For example, in English-Chinese code switching, there might be a sentence like \"\u2022 \u201a15\u00a9\u00a8 k \u2021meeting (We will have a meeting in 15 minutes)\". Traditional machine translation (MT) systems treat LCS data as noise, or just as regular sentences. However, if LCS data is processed intelligently, it can provide a useful signal for training word alignment and MT models. Moreover, LCS data is from non-news sources which can enhance the diversity of training data for MT. In this paper, we first extract constraints from this code switching data and then incorporate them into a word alignment model training procedure. We also show that by using the code switching data, we can jointly train a word alignment model and a language model using cotraining. Our techniques for incorporating LCS data improve by 2.64 in BLEU score over a baseline MT system trained using only standard sentence-aligned corpora.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "Many language users are competent in multiple languages, and they often use elements of multiple languages in conversations with other speakers with competence in the same set of languages. For example, native Mandarin speakers who also speak English might use English words in a Chinese sentence, like \"\\ \u2022 \u00f9 \u2021\u00afK solution\u00ed \u00ba(Do you know the solution to this problem ?)\". This phenomenon of mixing languages within a single utterance is known as Linguistic Code Switching (LCS). Examples of these utterances are common in communities of speakers with a shared competency in multiple languages, such as Web forums for Chinese emigr\u00e9s to the United States. For example, more than 50% of the sentences we collected from a Web forum (MITBBS.com) contains both Chinese and English.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Traditional word alignment models take a sentence-level aligned corpus as input and generate word-level alignments for each pair of parallel sentences.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Automatically-gathered LCS data typically contains no sentence-level alignments, but it still has some advantages for training word alignment models and machine translation (MT) systems which are worth exploring. First, because it contains multiple languages in the same sentence and still has a valid meaning, it will tell the relationship between the words from different languages to some extent. Second, most LCS data is formed during people's daily conversation, and thus it contains a diversity of topics that people care about, such as home furnishings, cars, entertainment, etc, that may not show up in standard parallel corpora. Moreover, LCS data is easily accessible from Web communities, such as MITBBS.com, Sina Weibo, Twitter, etc.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "However, like most unedited natural language text on the Web, LCS data contains symbols like emotions, grammar and spelling mistakes, slang and strongly idiomatic usage, and a variety of other phenomena that are difficult to handle. LCS data with different language pairs may also need special handling. For instance, Sinha and Thakur (2005) focus on words in mixed English and Hindi texts where a single word contains elements from both languages; they propose techniques for translating such words into both pure English and pure Hindi. Our study focuses on Chinese-English LCS, where this is rarely a problem, but for other language pairs, Sinha and Thakur's techniques may be required as preprocessing steps. Primarily, though, LCS data requires special-purpose algorithms to use it for word alignment, since it contains no explicit alignment labels.",
"cite_spans": [
{
"start": 318,
"end": 341,
"text": "Sinha and Thakur (2005)",
"ref_id": "BIBREF19"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "In this paper, we investigate two approaches to using LCS data for machine translation. The first approach focuses exclusively on word alignment, and uses patterns extracted from LCS data to guide the EM training procedure for word alignment over a standard sentence-aligned parallel corpus. We focus on two types of patterns in the LCS data: first, English words are almost never correct translations for any Chinese word in the same LCS utterance. Second, for sentences that are mostly Chinese but with some English words, if we propose substitutes for the English words using a Chinese language model, those substitutes are often good translations of the English words. We incorporate these patterns into EM training via the posterior regularization framework (Ganchev et al., 2010) .",
"cite_spans": [
{
"start": 763,
"end": 785,
"text": "(Ganchev et al., 2010)",
"ref_id": "BIBREF11"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Our second approach treats the alignment and language model as two different and complementary views of the data. We apply the cotraining paradigm for semi-supervised learning to incorporate the LCS data into the training procedures for the alignment model and the language model. From the translation table of the alignment model, the training procedure finds candidate translations of the English words in the LCS data, and uses those to supplement the language model training data. From the language model, the training procedure identifies Chinese words that complete the Chinese sentence with high probability, and it uses the English word paired with these completion words as additional training points for translation probabilities. These models are trained repeatedly until they converge to similar predictions on the LCS data. In combination with a larger phrase-based MT system (Koehn et al., 2003) , these two training procedures yield an MT system that achieves a BLEU score of 31.79 on an English-to-Chinese translation task, an improvement of 2.64 in BLEU score over a baseline MT system trained on only our parallel corpora.",
"cite_spans": [
{
"start": 889,
"end": 909,
"text": "(Koehn et al., 2003)",
"ref_id": "BIBREF13"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "The rest of this paper is organized as follows. The next section presents related work. Section 3 gives an overview of word alignment. Sections 4 and 5 detail our two algorithms. Section 6 presents our experiments and discusses results, and Section 7 concludes and discusses future work.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "There has been a lot of research on LCS from the theoretical and socio-linguistic communities (Nilep, 2006; De Fina, 2007) . Computational research on LCS has studied how to identify the boundaries of an individual language within LCS data, or how to predict when an utterance will switch to another language (Chan et al., 2004; Solorio and Liu, 2008) . Manandise and Gdaniec (2011) analyzed the effect on machine translation quality of LCS of Spanish-English and showed that LCS degrades the performance of the syntactic parser. Sinha and Thakur (2005) translate mixed Hindi and English (Hinglish) to pure Hindi and pure English by using two morphological analyzers from both Hindi and English.",
"cite_spans": [
{
"start": 94,
"end": 107,
"text": "(Nilep, 2006;",
"ref_id": "BIBREF17"
},
{
"start": 108,
"end": 122,
"text": "De Fina, 2007)",
"ref_id": "BIBREF7"
},
{
"start": 309,
"end": 328,
"text": "(Chan et al., 2004;",
"ref_id": "BIBREF6"
},
{
"start": 329,
"end": 351,
"text": "Solorio and Liu, 2008)",
"ref_id": "BIBREF20"
},
{
"start": 354,
"end": 382,
"text": "Manandise and Gdaniec (2011)",
"ref_id": "BIBREF16"
},
{
"start": 530,
"end": 553,
"text": "Sinha and Thakur (2005)",
"ref_id": "BIBREF19"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "The difficulty in their problem is that Hindi and English are often mixed into a single word which uses only the English alphabet; approaches based only on the character set cannot tell these words apart from English words. Our current study is for a language pair (English-Chinese) where the words are easy to tell apart, but for MT using code-switching data for other language pairs (such as Hindi-English), we can leverage some of the techniques from their work to separate the tokens into source and target.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "Like our proposed methods, other researchers have used co-training before for MT (Callison-Burch and Osborne, 2003) . They use target strings in multiple languages as different views on translation. However, in our work, we treat the alignment model and language model as different views of LCS data.",
"cite_spans": [
{
"start": 81,
"end": 115,
"text": "(Callison-Burch and Osborne, 2003)",
"ref_id": "BIBREF4"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "In addition to co-training, various other semisupervised approaches for MT and word alignment have been proposed, but these have relied on sentence alignments among multiple languages, rather than LCS data. Kay (2000) proposes using multiple target documents as a way of informing subsequent machine translations. Kumar et al. (2007) described a technique for word alignment in a multi-parallel sentence-aligned corpus and showed that this technique can be used to obtain higher quality bilingual word alignments. Other work like (Eisele, 2006) took the issue one step further that they used bilingual translation systems which share one or more common pivot languages to build systems which non-parallel corpus is used. Unlike the data in these techniques, LCS data requires no manual alignment effort and is freely available in large quantities.",
"cite_spans": [
{
"start": 207,
"end": 217,
"text": "Kay (2000)",
"ref_id": null
},
{
"start": 314,
"end": 333,
"text": "Kumar et al. (2007)",
"ref_id": "BIBREF15"
},
{
"start": 530,
"end": 544,
"text": "(Eisele, 2006)",
"ref_id": "BIBREF9"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "Another line of research has attempted to improve word alignment models by incorporating manually-labeled word alignments in addition to sentence alignments. Callison-Burch et al. (2004) tried to give a higher weight on manually labeled data compared to the automatic alignments. Fraser and Marcu (2006) used a log-linear model with features from IBM models. They alternated the traditional Expectation Maximization algorithm which is applied on a large parallel corpus with a discriminative step aimed at increasing wordalignment quality on a small, manually wordaligned corpus. Ambati et al. 2010tried to manually correct the alignments which are informative during the unsupervised training and applied them to an active learning model. However, labeled word alignment data is expensive to produce. Our approach is complementary, in that we use mixed data that has no word alignments, but still able to learn constraints on word alignments.",
"cite_spans": [
{
"start": 158,
"end": 186,
"text": "Callison-Burch et al. (2004)",
"ref_id": "BIBREF5"
},
{
"start": 280,
"end": 303,
"text": "Fraser and Marcu (2006)",
"ref_id": "BIBREF10"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "Our techniques make use of posterior regularization (PR) framework (Ganchev et al., 2010), which has previously been used for MT (Graca et al., 2008) , but with very different constraints on EM training and different goals. (Graca et al., 2008) use PR to enforce the constraint that one word should not translate to many words, and that if a word s translates to a word t in one MT system, then a model for translation in the reverse direction should translate t to s. Both of these constraints apply to sentence-aligned training data directly, and complement the constraints that we extract from LCS data.",
"cite_spans": [
{
"start": 129,
"end": 149,
"text": "(Graca et al., 2008)",
"ref_id": "BIBREF12"
},
{
"start": 224,
"end": 244,
"text": "(Graca et al., 2008)",
"ref_id": "BIBREF12"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "Statistical word alignment (Brown et al., 1994) is the task identifying which words are translations of each other in a bilingual sentence corpus. It is primarily used for machine translation. The input to an alignment system is a sentence-level aligned bilingual corpus, which consists of pairs of sentences in two languages. One language is denoted as the target language, and the other language as the source language.",
"cite_spans": [
{
"start": 27,
"end": 47,
"text": "(Brown et al., 1994)",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Statistical Word Alignment",
"sec_num": "3"
},
{
"text": "We now introduce the baseline model for word alignment and how we can incorporate the LCS data to improve the model. IBM Model 1 (Brown et al., 1994) and the HMM alignment model (Vogel et al., 1996) are cascaded to form the baseline model for alignment. These two models have a similar formulation L = P (t, a|s) = P (a) j P (t j |s a j ) with a different distortion probability P (a). s and t denote the source and target sentences. a is the alignment, and a j is the index of the source language word that generates the target language word at position j. The HMM model assumes the alignments have a first-order Markov dependency, so that P (a) = j P (a j |a j \u2212 a j\u22121 ). IBM Model 1 ignores the word position and uses a uniform distribution, so P (a) = j P (a j ) where P (a j ) = 1 |t| , where |t| is the length of t.",
"cite_spans": [
{
"start": 129,
"end": 149,
"text": "(Brown et al., 1994)",
"ref_id": "BIBREF3"
},
{
"start": 178,
"end": 198,
"text": "(Vogel et al., 1996)",
"ref_id": "BIBREF22"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Statistical Word Alignment",
"sec_num": "3"
},
{
"text": "Expectation Maximization (Dempster et al., 1977) is typically used to train the alignment model.",
"cite_spans": [
{
"start": 25,
"end": 48,
"text": "(Dempster et al., 1977)",
"ref_id": "BIBREF8"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Statistical Word Alignment",
"sec_num": "3"
},
{
"text": "It tries to maximize the marginal likelihood of the sentence-level aligned pairs. For the HMM alignment model, the forwardbackward algorithm can be used the optimize the posterior probability of the hidden alignment a.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Statistical Word Alignment",
"sec_num": "3"
},
{
"text": "We observed that most LCS sentences are predominantly in one language, which we call the majority language, with just a small number of words from another language, which we call the minority language. The grammar of each sentence appears to mirror the structure of the majority language. Speakers appear to be substituting primarily content words from the minority language, especially nouns and verbs, without changing the structure of the majority language. In this section, we explain two types of constraints we extract from the LCS data that can be helpful for guiding the training of a word alignment model, and we describe how we incorporate those constraints into a full training procedure.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Learning Constraints for Word Alignments from LCS Data",
"sec_num": "4"
},
{
"text": "After inspecting sentences in our LCS data, we found that the words from the target language occurring in the sentence are highly likely not to be the translation of the remaining source word. Figure 1 shows an example LCS sentence where the speaker has replaced the Chinese word \" \u2021\u00a6\" with the corresponding English word \"request\". In most LCS utterances, the minority language replaces or substitutes for words in the majority language, and thus it does not serve as a translation of any majority-language words in the sentence. If we can enforce that a word alignment model avoids pairing words that appear in the same LCS sentence, we can significantly narrow down the possible choices of the translation candidates during word alignment training. Formally, let t LCS be the set of target (Chinese) words and s LCS be the source (English) words in the same sentence of the LCS data. According to our observation, each s LCS j in s LCS should not be aligned with any word t LCS i in t LCS . We call every target-source word pair",
"cite_spans": [],
"ref_spans": [
{
"start": 193,
"end": 201,
"text": "Figure 1",
"ref_id": "FIGREF0"
}
],
"eq_spans": [],
"section": "Preventing bad alignments",
"sec_num": "4.1"
},
{
"text": "(t LCS i , s LCS j )",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Preventing bad alignments",
"sec_num": "4.1"
},
{
"text": "from LCS data a blocked alignment. For a set of word alignments WA = {(s w , t w )} produced by a word alignment model, define",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Preventing bad alignments",
"sec_num": "4.1"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "\u03c6 BA = (sw,tw)\u2208WA 1[(s w , t w ) \u2208 BA]",
"eq_num": "(1)"
}
],
"section": "Preventing bad alignments",
"sec_num": "4.1"
},
{
"text": "where BA is the set of blocked alignments extracted from the LCS data. We want to minimize \u03c6 BA . Figure 2 shows a graphical illustration of this constraint. ",
"cite_spans": [],
"ref_spans": [
{
"start": 98,
"end": 106,
"text": "Figure 2",
"ref_id": "FIGREF1"
}
],
"eq_spans": [],
"section": "Preventing bad alignments",
"sec_num": "4.1"
},
{
"text": "proposed by a language model",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Encouraging alignments with substitutes",
"sec_num": "4.2"
},
{
"text": "Another perspective of using the LCS data is that if we can find some target word set t similar from the target language which shares similar contexts as the source word s LCS j in the LCS data, then we can encourage s LCS j to be aligned with the each word t similar m in t similar . Figure 3 shows example phrases (\"\u00ac\u00af\u00ef AE ? U\" , \" \u00ac\u00af \u2021 \u00a6 ? U\", \"\u00ac\u00af\u00e1 \u00fd ? U\" etc) that appear in a Chinese language model and which share the same left context and right context as the word \"request.\" Our second objective is to encourage minority language words like \"request\" to align with possible substitutes from the majority language's language model. If we see any of \"\u00ef AE, \u2021 \u00a6, \u00e1 \u00fd\" in the parallel corpus, we should encourage the word \"request\" to be aligned with them. We call this target-source word pair (t similar m , s LCS j ) an encouraged alignment. Formally, we define",
"cite_spans": [],
"ref_spans": [
{
"start": 285,
"end": 294,
"text": "Figure 3",
"ref_id": "FIGREF2"
}
],
"eq_spans": [],
"section": "Encouraging alignments with substitutes",
"sec_num": "4.2"
},
{
"text": "\u03c6 EA = |C| \u2212 (sw,tw)\u2208WA 1[(s w , t w ) \u2208 EA] (2)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Encouraging alignments with substitutes",
"sec_num": "4.2"
},
{
"text": "where |C| is the size of the parallel corpus and EA is the encouraged alignment set. We define this expression in such a way that if the optimization procedure minimizes it, it will increase the number of encouraged alignments. Algorithm 1 shows the algorithm of calculating",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Encouraging alignments with substitutes",
"sec_num": "4.2"
},
{
"text": "t similar . (t LCS l , s LCS j , t LCS r",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Encouraging alignments with substitutes",
"sec_num": "4.2"
},
{
"text": ") is a (target, source, target)word tuple contained in the LCS data. l and r denote the left and right target words to the source word. We use the language model output from the target language. For each pair of contexts t l and t r for the source word, we find the exact match of this pair in the ngram. Then we extract the middle word as the candidates for t similar .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Encouraging alignments with substitutes",
"sec_num": "4.2"
},
{
"text": "Here, we only use 3 grams in our experiments, but it is possible to extend this to 5grams, which might lead to further improvements. The EA constraint Algorithm 1: finding t similar 1: Input: s LCS ,t LCS , language model LM 2: Set t similar ={} 3: Extract the 3 grams (t l , t m , t r ) \u2208 gram 3 from LM 4: set S = {} 5: For j from 1 to size(gram 3 )",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Encouraging alignments with substitutes",
"sec_num": "4.2"
},
{
"text": "if (t j l , t j r ) \u2208 S add t j m into C t j l ,t j r else put (t j l , t j r ) into S set C t j l ,t j r = {} 6: Extract tuple (t LCS l , s LCS j , t LCS r ) if (t LCS l , t LCS r ) \u2208 S add C t LCS l ,t LCS r",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Encouraging alignments with substitutes",
"sec_num": "4.2"
},
{
"text": "into t similar 7: Output: t similar is similar to a bilingual dictionary. However, in the bilingual dictionary, each source word might have several target translations (senses), so it might be ambiguous. The candidate translations used in EA are from language model (3 grams in this paper, but it can be extended to 5 grams), which will always match the contexts. Additionally, the bilingual dictionary contains the standard English/Chinese word pairs. But the LCS data is generated from people.s daily conversation; it reflects usage in a variety of domains, including colloquial and figurative usages that may not appear in a dictionary.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Encouraging alignments with substitutes",
"sec_num": "4.2"
},
{
"text": "We incorporate \u03c6 BA and \u03c6 EA into the EM training procedure for the alignment model using posterior regularization (PR) (Ganchev et al., 2010). Formally, let x be the sentence pairs s and t. During the E step, instead of using the posterior p(a|x) to calculate the expected counts, the PR framework tries to find a distribution q(a) which is close to p(a|x), but which also minimizes the properties \u03c6(a, x):",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Constrained parameter estimation",
"sec_num": "4.3"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "min q,\u03be [KL(q(a)||p(a|x, \u03b8)) + \u03c3||\u03be||]",
"eq_num": "(3)"
}
],
"section": "Constrained parameter estimation",
"sec_num": "4.3"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "s.t. E a\u223cq [\u03c6(a, x)] \u2264 \u03be",
"eq_num": "(4)"
}
],
"section": "Constrained parameter estimation",
"sec_num": "4.3"
},
{
"text": "where KL is the Kullback-Leibler divergence, \u03c3 is a free parameter indicating how important the constraints are compared with the marginal log likelihood and \u03be is a small violation allowed in the optimization. To impose multiple constraints, we define a norm ||\u03be|| A = (\u03be t A\u03be), where A is a diagonal matrix whose diagonal entries A ii are free parameters that provide weights on the different constraints. Since we only have two constraints here from LCS data, A = 1 0 0 \u03b1 where \u03b1 controls the relative importance of the two constraints.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Constrained parameter estimation",
"sec_num": "4.3"
},
{
"text": "To make the optimization task in the E-step more tractable, PR transforms it to a dual problem:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Constrained parameter estimation",
"sec_num": "4.3"
},
{
"text": "max \u03bb\u22650, \u03bb * \u2264\u03c3 \u2212 log a p(a|x, \u03b8) exp{\u2212\u03bb \u2022 \u03c6(a, x)}",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Constrained parameter estimation",
"sec_num": "4.3"
},
{
"text": "where \u2022 * is the dual norm of \u2022 A . The gradient of this dual objective is \u2212E q [\u03c6(a, x)]. A projected subgradient descent algorithm is used to perform the optimization.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Constrained parameter estimation",
"sec_num": "4.3"
},
{
"text": "The above approaches alter the translation and distortion probabilities in the alignment model. However, they leave the language model unchanged. We next investigate a technique that uses LCS data to re-estimate parameters for the language model as well as the alignment model simultaneously. Co-training (Blum and Mitchell, 1998 ) is a semi-supervised learning technique that requires two different views of the data. It assumes that each example can be described using two different feature sets which are conditionally independent. Also, each feature set of the data should be sufficient to make accurate prediction.",
"cite_spans": [
{
"start": 305,
"end": 329,
"text": "(Blum and Mitchell, 1998",
"ref_id": "BIBREF2"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Co-training using the LCS data",
"sec_num": "5"
},
{
"text": "The schema fits perfectly into our problem. We can treat the alignment model and the language model as two different views of the LCS data.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Co-training using the LCS data",
"sec_num": "5"
},
{
"text": "We use the same example \"\u00ac\u00afrequest ?U \u2022{\" to show how co-training works, shown in Figure 4 . From the translation table generated by the alignment model, we can get a set of candidate translations of \"request\", such as \"' \u2021 \u00a6\",\"\u017e\u00a6\",etc. We can find the candidate with the highest probability as the translation. Similarly, from the language model, we can extract all the ngrams containing \" \u00ac\u00af\" and \"?U\" as the left and right words and pick the words in the middle such as \" \u00efAE, \u2021\u00a6, \u00e1\u00fd\" etc as the candidate translations. We can then use the candidate with the highest probability as the translation for \"request\". Thus both models can predict translations for the English (minority language) in this example. Each model's predictions can be used as supplemental training data for the other model.",
"cite_spans": [],
"ref_spans": [
{
"start": 82,
"end": 90,
"text": "Figure 4",
"ref_id": "FIGREF3"
}
],
"eq_spans": [],
"section": "Co-training using the LCS data",
"sec_num": "5"
},
{
"text": "Algorithm 2 shows the co-training algorithm for word alignment. At each iteration, a language model and an alignment model are trained. The language model is trained on a Chinese-only corpus plus a corpus of probabilistic LCS sentences where the source words are replaced with target candidates from the alignment model. The alignment model is retrained using a translation table which is updated according to the output word pairs from the language model output and the LCS data. In order to take the sentence probability into consideration, we modify the language model training procedure: when it counts the number of times each ngram appears, instead of adding 1, it adds the probability from the translation model for ngrams in the LCS data that contain predicted translations.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Co-training using the LCS data",
"sec_num": "5"
},
{
"text": "We evaluated our LCS-driven training algorithms on an English-to-Chinese translation task. We use Moses (Koehn et al., 2003) , a phrasebased translation system that learns from bilingual sentence-aligned corpora as the MT system. We supplement the baseline word alignment model in Moses with our LCS data, constrained training procedure, and co-training algorithm as well as IBM 3 model. Because IBM 3 model is a fertility based model which might also alleviate Algorithm 2: Co-training for word alignment and language modeling 1: Input: parallel data X p , LCS data X LCS , language model training data X l 2: Initialize translation table tb for IBM1 model 3: For iteration from 1 to MAX",
"cite_spans": [
{
"start": 104,
"end": 124,
"text": "(Koehn et al., 2003)",
"ref_id": "BIBREF13"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "6.1"
},
{
"text": "tb \u2190 Train-IBM(X p ) tb \u2190 Train-HMM(X p |tb) 4:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "6.1"
},
{
"text": "For each sentence x i in X LCS : For each source word s j in x i : 1) find the translation t j of s j with with probability p j from tb 2) replace s j with t j and update sentence's probability",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "6.1"
},
{
"text": "p s = p s * p j X new l \u2190 X l \u222a x i 5: LM \u2190 Train-LM(X new l ) 6:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "6.1"
},
{
"text": "Extract the tri-gram gram 3 from LM 7:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "6.1"
},
{
"text": "For each sentence x i in X LCS : run Algorithm 1: finding t similar 8:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "6.1"
},
{
"text": "update tb using (t m , s j ) where t m \u2208 t similar and s j \u2208 x i 9: End For 10: Output: word alignment for X p and LM some of the problems caused by LCS data. To clarify, we use IBM1 model and HMM models in succession for the baseline. We trained the IBM1 model first and used the resulting parameters as the initial parameter values to train HMM model. Parameters for the final MT system are tuned with Minimum Error Rate Training (MERT) (Och, 2003) . The tuning set for MERT is the NIST MT06 data set, which includes 1664 sentences. We test the system on NIST MT02 (878 sentences). To evaluate the word alignment results, we manually aligned 250 sentences from NIST MT02 data set. For simplicity, we only have two types of labels for evaluating word alignments: either two words are aligned together or not. (Previous evaluation metrics also consider a third label for \"possible\" alignments.) Out of the word-aligned data, we use 100 sentences as a development set and the rest as our testing set.",
"cite_spans": [
{
"start": 439,
"end": 450,
"text": "(Och, 2003)",
"ref_id": "BIBREF18"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "6.1"
},
{
"text": "Our MT training corpus contains 2,636,692 sentence pairs from two parallel corpora: Hong Kong News (LDC2004T08) and Chinese English News Magazine Parallel Text (LDC2005T10). We use the Stanford Chinese segmenter to segment the Chinese data.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "6.1"
},
{
"text": "We use a ngram model package called SRILM (Stolcke, 2002) to train the language model. Because our modified ngram counts contain factions, we used Witten-Bell smoothing (Witten and Bell, 1991) which supports fractional counts. The 3-gram language model is trained on the Xinhua section of the Chinese Gigaword corpus (LDC2003T09) as well as the Chinese side of the parallel corpora. We also removed the sentences in MT02 from the Gigaword corpus if there is any to avoid the biases.",
"cite_spans": [
{
"start": 42,
"end": 57,
"text": "(Stolcke, 2002)",
"ref_id": "BIBREF21"
},
{
"start": 169,
"end": 192,
"text": "(Witten and Bell, 1991)",
"ref_id": "BIBREF23"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "6.1"
},
{
"text": "We gather the LCS data from \"MITBBS.com,\" a popular forum for Chinese people living in the United States. This forum is separated by discussion topic, and includes topics such as \"Travel\", \"News\", and \"Living style\". We extract data from 29 different topics. To clean up the LCS data, we get rid of HTML mark-up, and we remove patterns that are commonly repeated in forums, like \"Re:\" (for \"reply\" posts) and \"[= 1]\" (for \"repost\"). We change all English letters written in Chinese font into English font. We stem the English words in both the parallel training data and the LCS data. After the cleaning step, we have 245,470 sentences in the LCS data. 120,922 of them actually contain both Chinese and English in the same sentence. 101,302 of them contain only Chinese, and we add these into the language model training data. We discard the sentences that only contain English.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "6.1"
},
{
"text": "In order to incorporate the two constraints during the Posterior Regularization, we need to tune the parameters \u03c3 which controls the weights between the constraints and the marginal likelihood and \u03b1 which controls the relative importance between two constraints on development data. We varied \u03c3 from 0.1 to 1000 and varied \u03b1 over the set {0.01, 0.1, 1, 10, 100}.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Word Alignment Results",
"sec_num": "6.2"
},
{
"text": "After testing the 25 different combinations of \u03c3 and \u03b1 on the development data, we find that the setting with \u03c3 = 100 and \u03b1 = 0.1 achieves the best performance. During PR training, we trained the model 20 iterations for the dual optimization and 5 iterations for the modified EM. Table 1 shows the word alignment results. We can see that incorporating the LCS data into our alignment model improves the performance. Our best co-training+PR + system outperforms the baseline by 8 points. Figure 5 shows an example of how BA is extracted from LCS data can help the word alignment performance. upper figure shows that alignment by the baseline system. We can see that the word \"badminton\" is aligned incorrectly with word \">\u2122\u017d(Taufik)\" . However, in the LCS data, we see that \" >\u2122 \u017d(Taufik)\" and \"badminton\" appear in the same sentence \">\u2122\u017d badminton x\u00b3 (Taufik plays badminton so well)\" and by adding the blocked constraint into the alignment model, it correctly learns that \" >\u2122\u017d(Taufik)\" should be aligned with something else, and it finds \"Taufik\" at end. Table 2 shows some of the translations of \"badminton\" before and after incorporating the LCS data. We can see that it contains some wrong translations like \"\u00ae \u00a5\u00bf(pingpong room)\",\"> \u2122 \u017d(Taufik)\"etc using baseline model. After using the LCS data as constraints and the cotraining framework, these wrong alignments are eliminated and the translation \"\u2039 \u00a5(another way of expressing badminton)\" get a higher probability. We found that IBM 3 model can also correct this specific case. However, our co-training+PR + system still outperforms it by 6 points. Figure 6 shows an example of how EA is extracted from LCS data can help the word alignment. The solid lines show the alignment by the baseline model and we can see that the word \"compiled\" is not aligned with any Chinese word. After using the LCS data and the language model, we find that \"8 B(compile)\" shows up in the same context \"\u00d6(book) \u00e5 5(up)\"as \"compile\" along with \"C \u00be(staple)\" and \"\u00be(staple)\", therefore \"(compile, 8B)\" will be an encouraged alignment. After adding the EA constraint, the model learns that \"compile\" should be aligned with \"8B\".",
"cite_spans": [],
"ref_spans": [
{
"start": 280,
"end": 287,
"text": "Table 1",
"ref_id": "TABREF1"
},
{
"start": 487,
"end": 495,
"text": "Figure 5",
"ref_id": null
},
{
"start": 1057,
"end": 1064,
"text": "Table 2",
"ref_id": "TABREF4"
},
{
"start": 1607,
"end": 1615,
"text": "Figure 6",
"ref_id": null
}
],
"eq_spans": [],
"section": "Word Alignment Results",
"sec_num": "6.2"
},
{
"text": "In this section, we investigated whether improved alignments can improve MT performance. We Figure 5 : After incorporating the BA constraint from the LCS data, the word \"Taufik(>\u2122\u017d)\" is aligned correctly.",
"cite_spans": [],
"ref_spans": [
{
"start": 92,
"end": 100,
"text": "Figure 5",
"ref_id": null
}
],
"eq_spans": [],
"section": "Phrase-based machine translation",
"sec_num": "6.3"
},
{
"text": "PR+co Figure 6 : After incorporating the EA constraint from the LCS data, the word \"compiled(8B)\" is aligned correctly. use different word alignment models' outputs as the first step for Moses and keep the rest of Moses system the same. We incorporate Moses's eight standard features as well as the lexicalized reordering model. We also use the grow-diag-final and alignment symmetrization heuristic. Table 3 shows the machine translation results. We can see that 3 techniques we proposed for word alignment all improve the machine translation result over the baseline system as well as the IBM 3 model. However, although co-training has a bigger improvement on the word alignment compared with PR + , it actually has a lower BLEU score. This phenomenon shows that the improvement in the word alignment does not necessarily lead to the improvement on machine translation.",
"cite_spans": [],
"ref_spans": [
{
"start": 6,
"end": 14,
"text": "Figure 6",
"ref_id": null
},
{
"start": 401,
"end": 408,
"text": "Table 3",
"ref_id": null
}
],
"eq_spans": [],
"section": "Baseline",
"sec_num": null
},
{
"text": "After combining the co-training and the PR + together, co-training+PR + improved slightly over PR + for MT. Table 3 : Machine translation results. All entries marked with an asterisk are better than the baseline with 95% statistical significance computed using paired bootstrap resampling (Koehn, 2004) .",
"cite_spans": [
{
"start": 289,
"end": 302,
"text": "(Koehn, 2004)",
"ref_id": "BIBREF14"
}
],
"ref_spans": [
{
"start": 108,
"end": 115,
"text": "Table 3",
"ref_id": null
}
],
"eq_spans": [],
"section": "Baseline",
"sec_num": null
},
{
"text": "In this paper, we explored two different ways to use LCS data in a MT system: 1) PR framework to incorporate with Blocked Alignment and Encouraged Alignment constraints. 2) A semisupervised co-training procedure. Both techniques improve the performance of word alignment and MT over the baseline.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion and Future Work",
"sec_num": "7"
},
{
"text": "Our techniques are currently limited to sentences where the LCS data contains very short (usually one word) phrases from a minority language. An important line of investigation for generalizing these approaches is to consider techniques that cover longer phrases in the minority language; this can help add more of the LCS data into training.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion and Future Work",
"sec_num": "7"
}
],
"back_matter": [
{
"text": "This work was supported in part by NSF awards 1065397 and 1218692.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Acknowledgements",
"sec_num": null
}
],
"bib_entries": {
"BIBREF1": {
"ref_id": "b1",
"title": "Active semi-supervised learning for improving word alignment",
"authors": [],
"year": null,
"venue": "Proceedings of the Active Learning for NLP Workshop",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Active semi-supervised learning for improving word alignment. In In Proceedings of the Active Learning for NLP Workshop, NAACL.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "Combining labeled and unlabeled data with co-training",
"authors": [
{
"first": "Avrim",
"middle": [],
"last": "Blum",
"suffix": ""
},
{
"first": "Tom",
"middle": [],
"last": "Mitchell",
"suffix": ""
}
],
"year": 1998,
"venue": "Annual Conference on Computational Learning Theaory",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Avrim Blum and Tom Mitchell. 1998. Combining labeled and unlabeled data with co-training. In Annual Conference on Computational Learning Theaory.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "The mathematics of statistical machine translation: Parameter estimation",
"authors": [
{
"first": "P",
"middle": [
"F"
],
"last": "Brown",
"suffix": ""
},
{
"first": "S",
"middle": [
"Della"
],
"last": "Pietra",
"suffix": ""
},
{
"first": "V",
"middle": [
"J"
],
"last": "Della Pietra",
"suffix": ""
},
{
"first": "R",
"middle": [
"L"
],
"last": "Mercer",
"suffix": ""
}
],
"year": 1994,
"venue": "Computational Linguistics",
"volume": "19",
"issue": "2",
"pages": "263--311",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "P. F. Brown, S. Della Pietra, V. J. Della Pietra, and R. L. Mercer. 1994. The mathematics of statistical machine translation: Parameter estimation. Compu- tational Linguistics, 19(2):263-311.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "Cotraining for statistical machine translation",
"authors": [
{
"first": "Chris",
"middle": [],
"last": "Callison",
"suffix": ""
},
{
"first": "-",
"middle": [],
"last": "Burch",
"suffix": ""
},
{
"first": "Miles",
"middle": [],
"last": "Osborne",
"suffix": ""
}
],
"year": 2003,
"venue": "Proceedings of the 6th Annual CLUK Research Colloquium",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Chris Callison-Burch and Miles Osborne. 2003. Co- training for statistical machine translation. In In Proceedings of the 6th Annual CLUK Research Colloquium.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "Statistical machine translation with word-and sentence-aligned parallel corpora",
"authors": [
{
"first": "Chris",
"middle": [],
"last": "Callison",
"suffix": ""
},
{
"first": "-",
"middle": [],
"last": "Burch",
"suffix": ""
},
{
"first": "David",
"middle": [],
"last": "Talbot",
"suffix": ""
},
{
"first": "Miles",
"middle": [],
"last": "Osborne",
"suffix": ""
}
],
"year": 2004,
"venue": "Proceedings of ACL",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Chris Callison-Burch, David Talbot, and Miles Os- borne. 2004. Statistical machine translation with word-and sentence-aligned parallel corpora. In In Proceedings of ACL.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "Detection of language boundary in codeswitching utterances by bi-phone probabilities",
"authors": [
{
"first": "J",
"middle": [
"Y C"
],
"last": "Chan",
"suffix": ""
},
{
"first": "P",
"middle": [
"C"
],
"last": "Ching",
"suffix": ""
},
{
"first": "H",
"middle": [
"M"
],
"last": "Lee",
"suffix": ""
},
{
"first": "T",
"middle": [],
"last": "Meng",
"suffix": ""
}
],
"year": 2004,
"venue": "Proceedings of the International Symposium on Chinese Spoken Language Processing",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "J. Y. C. Chan, P. C. Ching, and H. M. LEE, T.and Meng. 2004. Detection of language boundary in code- switching utterances by bi-phone probabilities. In In Proceedings of the International Symposium on Chinese Spoken Language Processing.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "Code-switching and the construction of ethnic identity in a community of practice",
"authors": [
{
"first": "Fina",
"middle": [],
"last": "De",
"suffix": ""
}
],
"year": 2007,
"venue": "Language in Society",
"volume": "36",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "A De Fina. 2007. Code-switching and the construction of ethnic identity in a community of practice. In Language in Society, volume 36.",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "Maximum likelihood from incomplete data via the em algorithm",
"authors": [
{
"first": "A",
"middle": [
"P"
],
"last": "Dempster",
"suffix": ""
},
{
"first": "N",
"middle": [
"M"
],
"last": "Laird",
"suffix": ""
},
{
"first": "D",
"middle": [
"B"
],
"last": "Rubin",
"suffix": ""
}
],
"year": 1977,
"venue": "In Royal Statistical Society, Ser",
"volume": "39",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "A. P. Dempster, N. M. Laird, and D. B. Rubin. 1977. Maximum likelihood from incomplete data via the em algorithm. In Royal Statistical Society, Ser, volume 39.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "Parallel corpora and phrase-based statistical machine translation for new language pairs via multiple intermediaries",
"authors": [
{
"first": "Andreas",
"middle": [
"Eisele"
],
"last": "",
"suffix": ""
}
],
"year": 2006,
"venue": "International Conference on Language Resources and Evaluation",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Andreas Eisele. 2006. Parallel corpora and phrase-based statistical machine translation for new language pairs via multiple intermediaries. In International Conference on Language Resources and Evaluation.",
"links": null
},
"BIBREF10": {
"ref_id": "b10",
"title": "Semi-supervised training for statistical word alignment",
"authors": [
{
"first": "Alex",
"middle": [],
"last": "Fraser",
"suffix": ""
},
{
"first": "Daniel",
"middle": [],
"last": "Marcu",
"suffix": ""
}
],
"year": 2006,
"venue": "Proceedings of ACL",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Alex Fraser and Daniel Marcu. 2006. Semi-supervised training for statistical word alignment. In In Proceedings of ACL.",
"links": null
},
"BIBREF11": {
"ref_id": "b11",
"title": "Posterior regularization for structured latent variable models",
"authors": [
{
"first": "J",
"middle": [],
"last": "Kuzman Ganchev",
"suffix": ""
},
{
"first": "Jennifer",
"middle": [],
"last": "Graca",
"suffix": ""
},
{
"first": "Ben",
"middle": [],
"last": "Gillenwater",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Taskar",
"suffix": ""
}
],
"year": 2010,
"venue": "In Journal of Machine Learning Research",
"volume": "11",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Kuzman Ganchev, J. Graca, Jennifer Gillenwater, and Ben Taskar. 2010. Posterior regularization for structured latent variable models. In Journal of Machine Learning Research, volume 11.",
"links": null
},
"BIBREF12": {
"ref_id": "b12",
"title": "Expectation maximization and posterior constraints",
"authors": [
{
"first": "J",
"middle": [],
"last": "Graca",
"suffix": ""
},
{
"first": "K",
"middle": [],
"last": "Ganchev",
"suffix": ""
},
{
"first": "B",
"middle": [],
"last": "Taskar",
"suffix": ""
}
],
"year": 2008,
"venue": "NIPS",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "J. Graca, K. Ganchev, and B. Taskar. 2008. Expectation maximization and posterior constraints. In NIPS.",
"links": null
},
"BIBREF13": {
"ref_id": "b13",
"title": "Statistical phrase-based translation",
"authors": [
{
"first": "Philipp",
"middle": [],
"last": "Koehn",
"suffix": ""
},
{
"first": "Franz",
"middle": [
"Josef"
],
"last": "Och",
"suffix": ""
},
{
"first": "Daniel",
"middle": [],
"last": "Marcu",
"suffix": ""
}
],
"year": 2003,
"venue": "NAACL-HLT",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Philipp Koehn, Franz Josef Och, and Daniel Marcu. 2003. Statistical phrase-based translation. In NAACL-HLT.",
"links": null
},
"BIBREF14": {
"ref_id": "b14",
"title": "Statistical significance tests for machine translation evaluation",
"authors": [
{
"first": "Philipp",
"middle": [],
"last": "Koehn",
"suffix": ""
}
],
"year": 2004,
"venue": "Proceedings of EMNLP",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Philipp Koehn. 2004. Statistical significance tests for machine translation evaluation. In In Proceedings of EMNLP.",
"links": null
},
"BIBREF15": {
"ref_id": "b15",
"title": "Improving word alignment with bridge languages",
"authors": [
{
"first": "Shankar",
"middle": [],
"last": "Kumar",
"suffix": ""
},
{
"first": "Franz",
"middle": [
"Josef"
],
"last": "Och",
"suffix": ""
},
{
"first": "Wolfgang",
"middle": [],
"last": "Macherey",
"suffix": ""
}
],
"year": 2007,
"venue": "EMNLP",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Shankar Kumar, Franz Josef Och, and Wolfgang Macherey. 2007. Improving word alignment with bridge languages. In EMNLP.",
"links": null
},
"BIBREF16": {
"ref_id": "b16",
"title": "Morphology to the rescue redux: Resolving borrowings and code-mixing in machine translation",
"authors": [
{
"first": "Esme",
"middle": [],
"last": "Manandise",
"suffix": ""
},
{
"first": "Claudia",
"middle": [],
"last": "Gdaniec",
"suffix": ""
}
],
"year": 2011,
"venue": "SFCM",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Esme Manandise and Claudia Gdaniec. 2011. Mor- phology to the rescue redux: Resolving borrowings and code-mixing in machine translation. In SFCM.",
"links": null
},
"BIBREF17": {
"ref_id": "b17",
"title": "Code switching in sociocultural linguistics",
"authors": [
{
"first": "C",
"middle": [],
"last": "Nilep",
"suffix": ""
}
],
"year": 2006,
"venue": "Colorado Research in Linguistics",
"volume": "19",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "C. Nilep. 2006. Code switching in sociocultural linguistics. In Colorado Research in Linguistics, volume 19.",
"links": null
},
"BIBREF18": {
"ref_id": "b18",
"title": "Minimum error rate training for statistical machine translation",
"authors": [
{
"first": "Franz Josef",
"middle": [],
"last": "Och",
"suffix": ""
}
],
"year": 2003,
"venue": "ACL",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Franz Josef Och. 2003. Minimum error rate training for statistical machine translation. In ACL.",
"links": null
},
"BIBREF19": {
"ref_id": "b19",
"title": "Machine translation of bi-lingual hindi-english (hinglish) text",
"authors": [
{
"first": "R",
"middle": [
"M K"
],
"last": "Sinha",
"suffix": ""
},
{
"first": "A",
"middle": [],
"last": "Thakur",
"suffix": ""
}
],
"year": 2005,
"venue": "Proceedings of the 10th Conference on Machine Translation",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "R.M.K. Sinha and A. Thakur. 2005. Machine translation of bi-lingual hindi-english (hinglish) text. In In Proceedings of the 10th Conference on Machine Translation.",
"links": null
},
"BIBREF20": {
"ref_id": "b20",
"title": "Learning to predict code-switching points",
"authors": [
{
"first": "T",
"middle": [],
"last": "Solorio",
"suffix": ""
},
{
"first": "Y",
"middle": [],
"last": "Liu",
"suffix": ""
}
],
"year": 2008,
"venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "T. Solorio and Y. Liu. 2008. Learning to predict code-switching points. In In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP).",
"links": null
},
"BIBREF21": {
"ref_id": "b21",
"title": "An extensible language modeling toolkit",
"authors": [
{
"first": "A",
"middle": [],
"last": "Stolcke",
"suffix": ""
}
],
"year": 2002,
"venue": "Proc. Intl. Conf. on Spoken Language Processing",
"volume": "2",
"issue": "",
"pages": "901--904",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "A. Stolcke. 2002. An extensible language modeling toolkit. In Proc. Intl. Conf. on Spoken Language Processing, volume 2, pages 901-904.",
"links": null
},
"BIBREF22": {
"ref_id": "b22",
"title": "Hmmbased word alignment in statistical translation",
"authors": [
{
"first": "S",
"middle": [],
"last": "Vogel",
"suffix": ""
},
{
"first": "H",
"middle": [],
"last": "Ney",
"suffix": ""
},
{
"first": "C",
"middle": [],
"last": "Tillmann",
"suffix": ""
}
],
"year": 1996,
"venue": "Proc.COLING",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "S. Vogel, H. Ney, and C. Tillmann. 1996. Hmm- based word alignment in statistical translation. In In Proc.COLING.",
"links": null
},
"BIBREF23": {
"ref_id": "b23",
"title": "The zerofrequency problem: Estimating the probabil-ities of novel events in adaptive text compression",
"authors": [
{
"first": "H",
"middle": [],
"last": "Ian",
"suffix": ""
},
{
"first": "Timothy",
"middle": [
"C"
],
"last": "Witten",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Bell",
"suffix": ""
}
],
"year": 1991,
"venue": "IEEE Transactions on Information Theory",
"volume": "4",
"issue": "",
"pages": "1085--1094",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Ian H. Witten and Timothy C. Bell. 1991. The zero- frequency problem: Estimating the probabil-ities of novel events in adaptive text compression. In IEEE Transactions on Information Theory, volume 4, pages 1085-1094.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"uris": null,
"num": null,
"type_str": "figure",
"text": "The upper sentence is the original LCS sentence. The bottom ones are its translation in pure Chinese and English.Underlined words are the original words in the LCS sentence."
},
"FIGREF1": {
"uris": null,
"num": null,
"type_str": "figure",
"text": "Illustration of the blocked alignment constraint."
},
"FIGREF2": {
"uris": null,
"num": null,
"type_str": "figure",
"text": "Illustration of the encouraged alignment constraint. The dotted rectangle shows the candidate translations of the English word from the tri-gram output from the language model"
},
"FIGREF3": {
"uris": null,
"num": null,
"type_str": "figure",
"text": "The framework of co-training in word alignment. AM represents alignment model and LM represents language model. Green italic words are the encouraged translation and red italic words are the discouraged translation."
},
"TABREF1": {
"html": null,
"num": null,
"type_str": "table",
"content": "<table><tr><td>The</td></tr></table>",
"text": "Word alignment results (PR + means PR+BA+EA)."
},
"TABREF2": {
"html": null,
"num": null,
"type_str": "table",
"content": "<table><tr><td>Baseline:</td></tr><tr><td>PR+BA:</td></tr><tr><td>Indonesia badminton experts think Taufik's ranking favorable</td></tr></table>",
"text": "\u5370\u5c3c \u7fbd\u6bdb\u7403 \u4e13\u5bb6 \u8ba4\u4e3a \u9676\u83f2\u514b \u7684 \u6392\u540d \u5f88\u6709\u5229 Indonesia badminton experts think Taufik's ranking favorable \u5370\u5c3c \u7fbd\u6bdb\u7403 \u4e13\u5bb6 \u8ba4\u4e3a \u9676\u83f2\u514b \u7684 \u6392\u540d \u5f88\u6709\u5229"
},
"TABREF4": {
"html": null,
"num": null,
"type_str": "table",
"content": "<table><tr><td colspan=\"2\">\u5982\u4f55 \u628a \u4e66 compile \u8d77\u6765\uff1f</td></tr><tr><td colspan=\"2\">(How to compile the book ?)</td></tr><tr><td>Trigrams</td><td/></tr><tr><td colspan=\"2\">\u4e66(book) \u96c6\u7eb3(compile) \u8d77\u6765(up)</td></tr><tr><td colspan=\"2\">\u4e66(book) \u88c5\u8ba2(staple) \u8d77\u6765(up)</td></tr><tr><td>\u4e66(book) \u8ba2(staple)</td><td>\u8d77\u6765</td></tr><tr><td>(up) ...</td><td/></tr><tr><td colspan=\"2\">\u7ecf \u8bc4\u5ba1 \u540e \u7684 \u83b7\u5956 \u4f5c\u54c1 \u5219 \u96c6\u7eb3</td></tr><tr><td colspan=\"2\">Winning entries after the review will be compiled</td></tr><tr><td>Wednesday, October 16, 13</td><td/></tr></table>",
"text": "Translation tables of \"badminton\" before and after incorporation of LCS data."
}
}
}
}