ACL-OCL / Base_JSON /prefixP /json /P09 /P09-1009.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "P09-1009",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T08:52:45.528551Z"
},
"title": "Unsupervised Multilingual Grammar Induction",
"authors": [
{
"first": "Benjamin",
"middle": [],
"last": "Snyder",
"suffix": "",
"affiliation": {
"laboratory": "Artificial Intelligence Laboratory",
"institution": "Massachusetts Institute of Technology",
"location": {}
},
"email": "bsnyder@csail.mit.edu"
},
{
"first": "Tahira",
"middle": [],
"last": "Naseem",
"suffix": "",
"affiliation": {
"laboratory": "Artificial Intelligence Laboratory",
"institution": "Massachusetts Institute of Technology",
"location": {}
},
"email": "tahira@csail.mit.edu"
},
{
"first": "Regina",
"middle": [],
"last": "Barzilay",
"suffix": "",
"affiliation": {
"laboratory": "Artificial Intelligence Laboratory",
"institution": "Massachusetts Institute of Technology",
"location": {}
},
"email": "regina@csail.mit.edu"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "We investigate the task of unsupervised constituency parsing from bilingual parallel corpora. Our goal is to use bilingual cues to learn improved parsing models for each language and to evaluate these models on held-out monolingual test data. We formulate a generative Bayesian model which seeks to explain the observed parallel data through a combination of bilingual and monolingual parameters. To this end, we adapt a formalism known as unordered tree alignment to our probabilistic setting. Using this formalism, our model loosely binds parallel trees while allowing language-specific syntactic structure. We perform inference under this model using Markov Chain Monte Carlo and dynamic programming. Applying this model to three parallel corpora (Korean-English, Urdu-English, and Chinese-English) we find substantial performance gains over the CCM model, a strong monolingual baseline. On average, across a variety of testing scenarios, our model achieves an 8.8 absolute gain in F-measure. 1",
"pdf_parse": {
"paper_id": "P09-1009",
"_pdf_hash": "",
"abstract": [
{
"text": "We investigate the task of unsupervised constituency parsing from bilingual parallel corpora. Our goal is to use bilingual cues to learn improved parsing models for each language and to evaluate these models on held-out monolingual test data. We formulate a generative Bayesian model which seeks to explain the observed parallel data through a combination of bilingual and monolingual parameters. To this end, we adapt a formalism known as unordered tree alignment to our probabilistic setting. Using this formalism, our model loosely binds parallel trees while allowing language-specific syntactic structure. We perform inference under this model using Markov Chain Monte Carlo and dynamic programming. Applying this model to three parallel corpora (Korean-English, Urdu-English, and Chinese-English) we find substantial performance gains over the CCM model, a strong monolingual baseline. On average, across a variety of testing scenarios, our model achieves an 8.8 absolute gain in F-measure. 1",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "In this paper we investigate the task of unsupervised constituency parsing when bilingual parallel text is available. Our goal is to improve parsing performance on monolingual test data for each language by using unsupervised bilingual cues at training time. Multilingual learning has been successful for other linguistic induction tasks such as lexicon acquisition, morphological segmentation, and part-of-speech tagging (Genzel, 2005; Snyder et al., 2009) . We focus here on the unsupervised induction of unlabeled constituency brackets. This task has been extensively studied in a monolingual setting and has proven to be difficult (Charniak and Carroll, 1992; Klein and Manning, 2002) .",
"cite_spans": [
{
"start": 422,
"end": 436,
"text": "(Genzel, 2005;",
"ref_id": "BIBREF8"
},
{
"start": 437,
"end": 457,
"text": "Snyder et al., 2009)",
"ref_id": "BIBREF23"
},
{
"start": 635,
"end": 663,
"text": "(Charniak and Carroll, 1992;",
"ref_id": "BIBREF3"
},
{
"start": 664,
"end": 688,
"text": "Klein and Manning, 2002)",
"ref_id": "BIBREF14"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "The key premise of our approach is that ambiguous syntactic structures in one language may correspond to less uncertain structures in the other language. For instance, the English sentence I saw [the student [from MIT]] exhibits the classic problem of PP-attachment ambiguity. However, its Urdu translation, literally glossed as I [[MIT of ] student] saw, uses a genitive phrase that may only be attached to the adjacent noun phrase. Knowing the correspondence between these sentences should help us resolve the English ambiguity.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "One of the main challenges of unsupervised multilingual learning is to exploit cross-lingual patterns discovered in data, while still allowing a wide range of language-specific idiosyncrasies. To this end, we adapt a formalism known as unordered tree alignment (Jiang et al., 1995) to a probabilistic setting. Under this formalism, any two trees can be embedded in an alignment tree. This alignment tree allows arbitrary parts of the two trees to diverge in structure, permitting language-specific grammatical structure to be preserved. Additionally, a computational advantage of this formalism is that the marginalized probability over all possible alignments for any two trees can be efficiently computed with a dynamic program in linear time.",
"cite_spans": [
{
"start": 261,
"end": 281,
"text": "(Jiang et al., 1995)",
"ref_id": "BIBREF12"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "We formulate a generative Bayesian model which seeks to explain the observed parallel data through a combination of bilingual and monolingual parameters. Our model views each pair of sentences as having been generated as follows: First an alignment tree is drawn. Each node in this alignment tree contains either a solitary monolingual constituent or a pair of coupled bilingual constituents. For each solitary mono-lingual constituent, a sequence of part-of-speech tags is drawn from a language-specific distribution. For each pair of coupled bilingual constituents, a pair of part-of-speech sequences are drawn jointly from a cross-lingual distribution. Word-level alignments are then drawn based on the tree alignment. Finally, parallel sentences are assembled from these generated part-of-speech sequences and word-level alignments.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "To perform inference under this model, we use a Metropolis-Hastings within-Gibbs sampler. We sample pairs of trees and then compute marginalized probabilities over all possible alignments using dynamic programming.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "We test the effectiveness of our bilingual grammar induction model on three corpora of parallel text: English-Korean, English-Urdu and English-Chinese. The model is trained using bilingual data with automatically induced word-level alignments, but is tested on purely monolingual data for each language. In all cases, our model outperforms a state-of-the-art baseline: the Constituent Context Model (CCM) (Klein and Manning, 2002) , sometimes by substantial margins. On average, over all the testing scenarios that we studied, our model achieves an absolute increase in F-measure of 8.8 points, and a 19% reduction in error relative to a theoretical upper bound.",
"cite_spans": [
{
"start": 405,
"end": 430,
"text": "(Klein and Manning, 2002)",
"ref_id": "BIBREF14"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "The unsupervised grammar induction task has been studied extensively, mostly in a monolingual setting (Charniak and Carroll, 1992; Stolcke and Omohundro, 1994; Klein and Manning, 2002; Seginer, 2007) . While PCFGs perform poorly on this task, the CCM model (Klein and Manning, 2002) has achieved large gains in performance and is among the state-of-the-art probabilistic models for unsupervised constituency parsing. We therefore use CCM as our basic model of monolingual syntax.",
"cite_spans": [
{
"start": 102,
"end": 130,
"text": "(Charniak and Carroll, 1992;",
"ref_id": "BIBREF3"
},
{
"start": 131,
"end": 159,
"text": "Stolcke and Omohundro, 1994;",
"ref_id": "BIBREF24"
},
{
"start": 160,
"end": 184,
"text": "Klein and Manning, 2002;",
"ref_id": "BIBREF14"
},
{
"start": 185,
"end": 199,
"text": "Seginer, 2007)",
"ref_id": "BIBREF19"
},
{
"start": 257,
"end": 282,
"text": "(Klein and Manning, 2002)",
"ref_id": "BIBREF14"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "While there has been some previous work on bilingual CFG parsing, it has mainly focused on improving MT systems rather than monolingual parsing accuracy. Research in this direction was pioneered by (Wu, 1997) , who developed Inversion Transduction Grammars to capture crosslingual grammar variations such as phrase reorderings. More general formalisms for the same purpose were later developed (Wu and Wong, 1998; Chiang, 2005; Melamed, 2003; Eisner, 2003; Zhang and Gildea, 2005; Blunsom et al., 2008) . We know of only one study which evaluates these bilingual grammar formalisms on the task of grammar induction itself (Smith and Smith, 2004 ). Both our model and even the monolingual CCM baseline yield far higher performance on the same Korean-English corpus.",
"cite_spans": [
{
"start": 198,
"end": 208,
"text": "(Wu, 1997)",
"ref_id": "BIBREF26"
},
{
"start": 394,
"end": 413,
"text": "(Wu and Wong, 1998;",
"ref_id": "BIBREF25"
},
{
"start": 414,
"end": 427,
"text": "Chiang, 2005;",
"ref_id": "BIBREF4"
},
{
"start": 428,
"end": 442,
"text": "Melamed, 2003;",
"ref_id": "BIBREF17"
},
{
"start": 443,
"end": 456,
"text": "Eisner, 2003;",
"ref_id": "BIBREF6"
},
{
"start": 457,
"end": 480,
"text": "Zhang and Gildea, 2005;",
"ref_id": "BIBREF28"
},
{
"start": 481,
"end": 502,
"text": "Blunsom et al., 2008)",
"ref_id": "BIBREF1"
},
{
"start": 622,
"end": 644,
"text": "(Smith and Smith, 2004",
"ref_id": "BIBREF20"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "Our approach is closer to the unsupervised bilingual parsing model developed by Kuhn (2004) , which aims to improve monolingual performance. Assuming that trees induced over parallel sentences have to exhibit certain structural regularities, Kuhn manually specifies a set of rules for determining when parsing decisions in the two languages are inconsistent with GIZA++ wordlevel alignments. By incorporating these constraints into the EM algorithm he was able to improve performance over a monolingual unsupervised PCFG. Still, the performance falls short of state-of-the-art monolingual models such as the CCM.",
"cite_spans": [
{
"start": 80,
"end": 91,
"text": "Kuhn (2004)",
"ref_id": "BIBREF16"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "More recently, there has been a body of work attempting to improve parsing performance by exploiting syntactically annotated parallel data. In one strand of this work, annotations are assumed only in a resource-rich language and are projected onto a resource-poor language using the parallel data Xi and Hwa, 2005) . In another strand of work, syntactic annotations are assumed on both sides of the parallel data, and a model is trained to exploit the parallel data at test time as well (Smith and Smith, 2004; Burkett and Klein, 2008) . In contrast to this work, our goal is to explore the benefits of multilingual grammar induction in a fully unsupervised setting.",
"cite_spans": [
{
"start": 297,
"end": 314,
"text": "Xi and Hwa, 2005)",
"ref_id": "BIBREF27"
},
{
"start": 487,
"end": 510,
"text": "(Smith and Smith, 2004;",
"ref_id": "BIBREF20"
},
{
"start": 511,
"end": 535,
"text": "Burkett and Klein, 2008)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "We finally note a recent paper which uses parameter tying to improve unsupervised dependency parse induction (Cohen and Smith, 2009) . While the primary performance gains occur when tying related parameters within a language, some additional benefit is observed through bilingual tying, even in the absence of a parallel corpus.",
"cite_spans": [
{
"start": 120,
"end": 132,
"text": "Smith, 2009)",
"ref_id": "BIBREF5"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "We propose an unsupervised Bayesian model for learning bilingual syntactic structure using parallel corpora. Our key premise is that difficult-tolearn syntactic structures of one language may correspond to simpler or less uncertain structures in the other language. We treat the part-of-speech tag sequences of parallel sentences, as well as their",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model",
"sec_num": "3"
},
{
"text": "(i) (ii) (iii)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model",
"sec_num": "3"
},
{
"text": "Figure 1: A pair of trees (i) and two possible alignment trees. In (ii), no empty spaces are inserted, but the order of one of the original tree's siblings has been reversed. In (iii), only two pairs of nodes have been aligned (indicated by arrows) and many empty spaces inserted.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model",
"sec_num": "3"
},
{
"text": "word-level alignments, as observed data. We obtain these word-level alignments using GIZA++ (Och and Ney, 2003) .",
"cite_spans": [
{
"start": 92,
"end": 111,
"text": "(Och and Ney, 2003)",
"ref_id": "BIBREF18"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Model",
"sec_num": "3"
},
{
"text": "Our model seeks to explain this observed data through a generative process whereby two aligned parse trees are produced jointly. Though they are aligned, arbitrary parts of the two trees are permitted to diverge, accommodating languagespecific grammatical structure. In effect, our model loosely binds the two trees: node-to-node alignments need only be used where repeated bilingual patterns can be discovered in the data.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model",
"sec_num": "3"
},
{
"text": "We achieve this loose binding of trees by adapting unordered tree alignment (Jiang et al., 1995) to a probabilistic setting. Under this formalism, any two trees can be aligned using an alignment tree. The alignment tree embeds the original two trees within it: each node is labeled by a pair (x, y), (\u03bb, y), or (x, \u03bb) where x is a node from the first tree, y is a node from the second tree, and \u03bb is an empty space. The individual structure of each tree must be preserved under the embedding with the exception of sibling order (to allow variations in phrase and word order).",
"cite_spans": [
{
"start": 76,
"end": 96,
"text": "(Jiang et al., 1995)",
"ref_id": "BIBREF12"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Tree Alignments",
"sec_num": "3.1"
},
{
"text": "The flexibility of this formalism can be demonstrated by two extreme cases: (1) an alignment between two trees may actually align none of their individual nodes, instead inserting an empty space \u03bb for each of the original two trees' nodes. 2if the original trees are isomorphic to one another, the alignment may match their nodes exactly, without inserting any empty spaces. See Figure 1 for an example.",
"cite_spans": [],
"ref_spans": [
{
"start": 379,
"end": 387,
"text": "Figure 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "Tree Alignments",
"sec_num": "3.1"
},
{
"text": "As our basic model of syntactic structure, we adopt the Constituent-Context Model (CCM) of Klein and Manning (2002) . Under this model, the part-of-speech sequence of each span in a sentence is generated either as a constituent yield -if it is dominated by a node in the treeor otherwise as a distituent yield. For example, in the bracketed sentence [John/NNP [climbed/VB [the/DT tree/NN]]], the sequence VB DT NN is generated as a constituent yield, since it constitutes a complete bracket in the tree. On the other hand, the sequence VB DT is generated as a distituent, since it does not. Besides these yields, the contexts (two surrounding POS tags) of constituents and distituents are generated as well. In this example, the context of the constituent VB DT NN would be (NNP, #), while the context of the distituent VB DT would be (NNP, NN). The CCM model employs separate multinomial distributions over constituents, distituents, constituent contexts, and distituent contexts. While this model is deficienteach observed subsequence of part-of-speech tags is generated many times over -its performance is far higher than that of unsupervised PCFGs. Under our bilingual model, each pair of sentences is assumed to have been generated jointly in the following way: First, an unlabeled alignment tree is drawn uniformly from the set of all such trees. This alignment tree specifies the structure of each of the two individual trees, as well as the pairs of nodes which are aligned and those which are not aligned (i.e. paired with a \u03bb).",
"cite_spans": [
{
"start": 91,
"end": 115,
"text": "Klein and Manning (2002)",
"ref_id": "BIBREF14"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Model overview",
"sec_num": "3.2"
},
{
"text": "For each pair of aligned nodes, a corresponding pair of constituents and contexts are jointly drawn from a bilingual distribution. For unaligned nodes (i.e. nodes paired with a \u03bb in the alignment tree), a single constituent and context are drawn, from language-specific distributions. Distituents and their contexts are also drawn from languagespecific distributions. Finally, word-level alignments are drawn based on the structure of the alignment tree.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model overview",
"sec_num": "3.2"
},
{
"text": "In the next two sections, we describe our model in more formal detail by specifying the parameters and generative process by which sentences are formed.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model overview",
"sec_num": "3.2"
},
{
"text": "Our model employs a number of multinomial distributions: The first four distributions correspond exactly to the parameters of the CCM model. Parameter \u03c9 is a \"coupling parameter\" which measures the compatibility of tree-aligned constituent yield pairs. The final two parameters measure the compatibility of syntactic alignments with the observed lexical GIZA++ alignments. Intuitively, aligned nodes should have a high density of word-level alignments between them, and unaligned nodes should have few lexical alignments.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Parameters",
"sec_num": "3.3"
},
{
"text": "\u2022 \u03c0 C i : over constituent yields of language i, \u2022 \u03c0 D i : over distituent yields of language i, \u2022 \u03c6 C i :",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Parameters",
"sec_num": "3.3"
},
{
"text": "More formally, consider a tree-aligned node pair (n 1 , n 2 ) with corresponding yields (y 1 , y 2 ). We call a word-level alignment good if it aligns a word in y 1 with a word in y 2 . We call a wordlevel alignment bad if it aligns a word in y 1 with a word outside y 2 , or vice versa. The Gizascore for (n 1 , n 2 ) is the number of good word alignments minus the number of bad word alignments. For example, suppose the constituent my long name is node-aligned to its Urdu translation mera lamba naam. If only the word-pairs my/mera and name/naam are aligned, then the Giza-score for this node-alignment would be 2. If however, the English word long were (incorrectly) aligned under GIZA++ to some Urdu word outside the corresponding constituent, then the score would drop to 1. This score could even be negative if the number of bad alignments exceeds those that are good. Distribution Gz pair provides a probability for these scores (up to some fixed absolute value).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Parameters",
"sec_num": "3.3"
},
{
"text": "For an unaligned node n with corresponding yield y, only bad GIZA++ alignments are possible, thus the Giza-score for these nodes will always be zero or negative. Distribution Gz node provides a probability for these scores (down to some fixed value). We want our model to find tree alignments such that both aligned node pairs and unaligned nodes have high Giza-score.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Parameters",
"sec_num": "3.3"
},
{
"text": "Now we describe the stochastic process whereby the observed parallel sentences and their wordlevel alignments are generated, according to our model.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Generative Process",
"sec_num": "3.4"
},
{
"text": "As the first step in the Bayesian generative process, all the multinomial parameters listed in the previous section are drawn from their conjugate priors -Dirichlet distributions of appropriate dimension. Then, each pair of word-aligned parallel sentences is generated through the following process:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Generative Process",
"sec_num": "3.4"
},
{
"text": "1. A pair of binary trees T 1 and T 2 along with an alignment tree A are drawn according to P (T 1 , T 2 , A). A is an alignment tree for T 1 and T 2 if it can be obtained by the following steps: First insert blank nodes (labeled by \u03bb) into T 1 and T 2 . Then permute the order of sibling nodes such that the two resulting trees T 1 and T 2 are identical in structure. Finally, overlay T 1 and T 2 to obtain A. We additionally require that A contain no extraneous nodes -that is no nodes with two blank labels (\u03bb, \u03bb). See Figure 1 for an example.",
"cite_spans": [],
"ref_spans": [
{
"start": 522,
"end": 530,
"text": "Figure 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "Generative Process",
"sec_num": "3.4"
},
{
"text": "We define the distribution P (T 1 , T 2 , A) to be uniform over all pairs of binary trees and their alignments.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Generative Process",
"sec_num": "3.4"
},
{
"text": "2. For each node in A of the form (n 1 , \u03bb) (i.e. nodes in T 1 left unaligned by A), draw (i) a constituent yield according to \u03c0 C 1 , (ii) a constituent context according to \u03c6 C 1 , (iii) a Giza-score according to Gz node .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Generative Process",
"sec_num": "3.4"
},
{
"text": "3. For each node in A of the form (\u03bb, n 2 ) (i.e. nodes in T 2 left unaligned by A), draw (i) a constituent yield according to \u03c0 C 2 , (ii) a constituent context according to \u03c6 C 2 , (iii) a Giza-score according to Gz node .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Generative Process",
"sec_num": "3.4"
},
{
"text": "4. For each node in A of the form (n 1 , n 2 ) (i.e. tree-aligned node pairs), draw (i) a pair of constituent yields (y 1 , y 2 ) according to:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Generative Process",
"sec_num": "3.4"
},
{
"text": "\u03c6 C 1 (y 1 ) \u2022 \u03c6 C 2 (y 2 ) \u2022 \u03c9(y 1 , y 2 ) Z (1)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Generative Process",
"sec_num": "3.4"
},
{
"text": "which is a product of experts combining the language specific context-yield distributions as well as the coupling distribution \u03c9 with normalization constant Z, (ii) a pair of contexts according to the appropriate language-specific parameters, (iii) a Giza-score according to Gz pair .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Generative Process",
"sec_num": "3.4"
},
{
"text": "(for each language i \u2208 {1, 2}), draw a distituent yield according to \u03c0 D i and a distituent context according to \u03c6 D i .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "For each span in T i not dominated by a node",
"sec_num": "5."
},
{
"text": "6. Draw actual word-level alignments consistent with the Giza-scores, according to a uniform distribution.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "For each span in T i not dominated by a node",
"sec_num": "5."
},
{
"text": "In the next section we turn to the problem of inference under this model when only the partof-speech tag sequences of parallel sentences and their word-level alignments are observed.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "For each span in T i not dominated by a node",
"sec_num": "5."
},
{
"text": "Given a corpus of paired part-of-speech tag sequences (s 1 , s 2 ) and their GIZA++ alignments g, we would ideally like to predict the set of tree pairs (T 1 , T 2 ) which have highest probability when conditioned on the observed data: P T 1 , T 2 s 1 , s 2 , g . We could rewrite this by explicitly integrating over the yield, context, coupling, Giza-score parameters as well as the alignment trees. However, since maximizing this integral directly would be intractable, we resort to standard Markov chain sampling techniques. We use Gibbs sampling (Hastings, 1970) to draw trees for each sentence conditioned on those drawn for all other sentences. The samples form a Markov chain which is guaranteed to converge to the true joint distribution over all sentences.",
"cite_spans": [
{
"start": 550,
"end": 566,
"text": "(Hastings, 1970)",
"ref_id": "BIBREF10"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Inference",
"sec_num": "3.5"
},
{
"text": "In the monolingual setting, there is a wellknown tree sampling algorithm (Johnson et al., 2007) . This algorithm proceeds in top-down fashion by sampling individual split points using the marginal probabilities of all possible subtrees. These marginals can be efficiently pre-computed and form the \"inside\" table of the famous Inside-Outside algorithm. However, in our setting, trees come in pairs, and their joint probability crucially depends on their alignment.",
"cite_spans": [
{
"start": 73,
"end": 95,
"text": "(Johnson et al., 2007)",
"ref_id": "BIBREF13"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Inference",
"sec_num": "3.5"
},
{
"text": "For the i th parallel sentence, we wish to jointly sample the pair of trees (T 1 , T 2 ) i together with their alignment A i . To do so directly would involve simultaneously marginalizing over all possible subtrees as well as all possible alignments between such subtrees when sampling upper-level split points. We know of no obvious algorithm for computing this marginal. We instead first sample the pair of trees (T 1 , T 2 ) i from a simpler proposal distribution Q. Our proposal distribution assumes that no nodes of the two trees are aligned and therefore allows us to use the recursive topdown sampling algorithm mentioned above. After a new tree pair T * = (T * 1 , T * 2 ) i is drawn from Q, we accept the pair with the following probability:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Inference",
"sec_num": "3.5"
},
{
"text": "min 1, P (T * |T \u2212i , A \u2212i ) Q(T |T \u2212i , A \u2212i ) P (T |T \u2212i , A \u2212i ) Q(T * |T \u2212i , A \u2212i )",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Inference",
"sec_num": "3.5"
},
{
"text": "where T is the previously sampled tree-pair for sentence i, P is the true model probability, and Q is the probability under the proposal distribution. This use of a tractable proposal distribution and acceptance ratio is known as the Metropolis-Hastings algorithm and it preserves the convergence guarantee of the Gibbs sampler (Hastings, 1970) . To compute the terms P (T * |T \u2212i , A \u2212i ) and P (T |T \u2212i , A \u2212i ) in the acceptance ratio above, we need to marginalize over all possible alignments between tree pairs. Fortunately, for any given pair of trees T 1 and T 2 this marginalization can be computed using a dynamic program in time O(|T 1 ||T 2 |). Here we provide a very brief sketch. For every pair of nodes n 1 \u2208 T 1 , n 2 \u2208 T 2 , a table stores the marginal probability of the subtrees rooted at n 1 and n 2 , respectively. A dynamic program builds this table from the bottom up: For each node pair n 1 , n 2 , we sum the probabilities of all local alignment configurations, each multiplied by the appro-priate marginals already computed in the table for lower-level node pairs. This algorithm is an adaptation of the dynamic program presented in (Jiang et al., 1995) for finding minimum cost alignment trees ( Fig. 5 of that publication) .",
"cite_spans": [
{
"start": 328,
"end": 344,
"text": "(Hastings, 1970)",
"ref_id": "BIBREF10"
},
{
"start": 1158,
"end": 1178,
"text": "(Jiang et al., 1995)",
"ref_id": "BIBREF12"
}
],
"ref_spans": [
{
"start": 1222,
"end": 1249,
"text": "Fig. 5 of that publication)",
"ref_id": null
}
],
"eq_spans": [],
"section": "Inference",
"sec_num": "3.5"
},
{
"text": "Once a pair of trees (T 1 , T 2 ) has been sampled, we can proceed to sample an alignment tree A|T 1 , T 2 . 2 We sample individual alignment decisions from the top down, at each step using the alignment marginals for the remaining subtrees (already computed using the afore-mentioned dynamic program). Once the triple (T 1 , T 2 , A) has been sampled, we move on to the next parallel sentence.",
"cite_spans": [
{
"start": 109,
"end": 110,
"text": "2",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Inference",
"sec_num": "3.5"
},
{
"text": "We avoid directly sampling parameter values, instead using the marginalized closed forms for multinomials with Dirichlet conjugate-priors using counts and hyperparameter pseudo-counts (Gelman et al., 2004) . Note that in the case of yield pairs produced according to Distribution 1 (in step 4 of the generative process) conjugacy is technically broken, since the yield pairs are no longer produced by a single multinomial distribution. Nevertheless, we count the produced yields as if they had been generated separately by each of the distributions involved in the numerator of Distribution 1.",
"cite_spans": [
{
"start": 184,
"end": 205,
"text": "(Gelman et al., 2004)",
"ref_id": "BIBREF7"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Inference",
"sec_num": "3.5"
},
{
"text": "We test our model on three corpora of bilingual parallel sentences: English-Korean, English-Urdu, and English-Chinese. Though the model is trained using parallel data, during testing it has access only to monolingual data. This set-up ensures that we are testing our model's ability to learn better parameters at training time, rather than its ability to exploit parallel data at test time. Following (Klein and Manning, 2002) , we restrict our model to binary trees, though we note that the alignment trees do not follow this restriction.",
"cite_spans": [
{
"start": 401,
"end": 426,
"text": "(Klein and Manning, 2002)",
"ref_id": "BIBREF14"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental setup",
"sec_num": "4"
},
{
"text": "Data The Penn Korean Treebank (Han et al., 2002) consists of 5,083 Korean sentences translated into English for the purposes of language training in a military setting. Both the Korean and English sentences are annotated with syntactic trees. We use the first 4,000 sentences for training and the last 1,083 sentences for testing. We note that in the Korean data, a separate tag is given for each morpheme. We simply concatenate all the morpheme tags given for each word and treat the concatenation as a single tag. This procedure results in 199 different tags. The English-Urdu parallel corpus 3 consists of 4,325 sentences from the first three sections of the Penn Treebank and their Urdu translations annotated at the part-of-speech level. The Urdu side of this corpus does not provide tree annotations so here we can test parse accuracy only on English. We use the remaining sections of the Penn Treebank for English testing. The English-Chinese treebank (Bies et al., 2007) consists of 3,850 Chinese newswire sentences translated into English. Both the English and Chinese sentences are annotated with parse trees. We use the first 4/5 for training and the final 1/5 for testing.",
"cite_spans": [
{
"start": 30,
"end": 48,
"text": "(Han et al., 2002)",
"ref_id": "BIBREF9"
},
{
"start": 959,
"end": 978,
"text": "(Bies et al., 2007)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental setup",
"sec_num": "4"
},
{
"text": "During preprocessing of the corpora we remove all punctuation marks and special symbols, following the setup in previous grammar induction work (Klein and Manning, 2002) . To obtain lexical alignments between the parallel sentences we employ GIZA++ (Och and Ney, 2003) . We use intersection alignments, which are one-to-one alignments produced by taking the intersection of oneto-many alignments in each direction. These oneto-one intersection alignments tend to have higher precision.",
"cite_spans": [
{
"start": 144,
"end": 169,
"text": "(Klein and Manning, 2002)",
"ref_id": "BIBREF14"
},
{
"start": 249,
"end": 268,
"text": "(Och and Ney, 2003)",
"ref_id": "BIBREF18"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental setup",
"sec_num": "4"
},
{
"text": "We initialize the trees by making uniform split decisions recursively from the top down for sentences in both languages. Then for each pair of parallel sentences we randomly sample an initial alignment tree for the two sampled trees.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental setup",
"sec_num": "4"
},
{
"text": "Baseline We implement a Bayesian version of the CCM as a baseline. This model uses the same inference procedure as our bilingual model (Gibbs sampling). In fact, our model reduces to this Bayesian CCM when it is assumed that no nodes between the two parallel trees are ever aligned and when word-level alignments are ignored. We also reimplemented the original EM version of CCM and found virtually no difference in performance when using EM or Gibbs sampling. In both cases our implementation achieves F-measure in the range of 69-70% on WSJ10, broadly in line with the performance reported by Klein and Manning (2002) .",
"cite_spans": [
{
"start": 595,
"end": 619,
"text": "Klein and Manning (2002)",
"ref_id": "BIBREF14"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental setup",
"sec_num": "4"
},
{
"text": "Hyperparameters Klein (2005) reports using smoothing pseudo-counts of 2 for constituent Figure 2 : The F-measure of the CCM baseline (dotted line) and bilingual model (solid line) plotted on the y-axis, as the maximum sentence length in the test set is increased (x-axis). Results are averaged over all training scenarios given in Table 1. yields and contexts and 8 for distituent yields and contexts. In our Bayesian model, these similar smoothing counts occur as the parameters of the Dirichlet priors. For Korean we found that the baseline performed well using these values. However, on our English and Chinese data, we found that somewhat higher smoothing values worked best, so we utilized values of 20 and 80 for constituent and distituent smoothing counts, respectively.",
"cite_spans": [
{
"start": 16,
"end": 28,
"text": "Klein (2005)",
"ref_id": "BIBREF15"
}
],
"ref_spans": [
{
"start": 88,
"end": 96,
"text": "Figure 2",
"ref_id": null
},
{
"start": 331,
"end": 339,
"text": "Table 1.",
"ref_id": null
}
],
"eq_spans": [],
"section": "Experimental setup",
"sec_num": "4"
},
{
"text": "Our model additionally requires hyperparameter values for \u03c9 (the coupling distribution for aligned yields), Gz pair and Gz node (the distributions over Giza-scores for aligned nodes and unaligned nodes, respectively). For \u03c9 we used a symmetric Dirichlet prior with parameter 1. For Gz pair and Gz node , in order to create a strong bias towards high Giza-scores, we used non-symmetric Dirichlet priors. In both cases, we capped the absolute value of the scores at 3, to prevent count sparsity. In the case of Gz pair we gave pseudocounts of 1,000 for negative values and zero, and pseudo-counts of 1,000,000 for positive scores. For Gz node we gave a pseudo-count of 1,000,000 for a score of zero, and 1,000 for all negative scores. This very strong prior bias encodes our intuition that syntactic alignments which respect lexical alignments should be preferred. Our method is not sensitive to these exact values and any reasonably strong bias gave similar results.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental setup",
"sec_num": "4"
},
{
"text": "In all our experiments, we consider the hyperparameters fixed and observed values.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental setup",
"sec_num": "4"
},
{
"text": "Testing and evaluation As mentioned above, we test our model only on monolingual data, where the parallel sentences are not provided to the model. To predict the bracketings of these monolingual test sentences, we take the smoothed counts accumulated in the final round of sampling over the training data and perform a maximum likelihood estimate of the monolingual CCM parameters. These parameters are then used to produce the highest probability bracketing of the test set.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental setup",
"sec_num": "4"
},
{
"text": "To evaluate both our model as well as the baseline, we use (unlabeled) bracket precision, recall, and F-measure (Klein and Manning, 2002) . Following previous work, we include the wholesentence brackets but ignore single-word brackets. We perform experiments on different subsets of training and testing data based on the sentencelength. In particular we experimented with sentence length limits of 10, 20, and 30 for both the training and testing sets. We also report the upper bound on F-measure for binary trees. We average the results over 10 separate sampling runs. Table 1 reports the full results of our experiments. In all testing scenarios the bilingual model outperforms its monolingual counterpart in terms of both precision and recall. On average, the bilingual model gains 10.2 percentage points in precision, 7.7 in recall, and 8.8 in F-measure. The gap between monolingual performance and the binary tree upper bound is reduced by over 19%.",
"cite_spans": [
{
"start": 112,
"end": 137,
"text": "(Klein and Manning, 2002)",
"ref_id": "BIBREF14"
}
],
"ref_spans": [
{
"start": 571,
"end": 578,
"text": "Table 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "Experimental setup",
"sec_num": "4"
},
{
"text": "The extent of the gain varies across pairings. For instance, the smallest improvement is observed for English when trained with Urdu. The Korean-English pairing results in substantial improvements for Korean and quite large improvements for English, for which the absolute gain reaches 28 points in F-measure. In the case of Chinese and English, the gains for English are fairly minimal whereas those for Chinese are quite sub- Table 1 : Unlabeled precision, recall and F-measure for the monolingual baseline and the bilingual model on several test sets. We report results for different combinations of maximum sentence length in both the training and test sets. The right most column, in all cases, contains the maximum F-measure achievable using binary trees. The best performance for each test-length is highlighted in bold.",
"cite_spans": [],
"ref_spans": [
{
"start": 428,
"end": 435,
"text": "Table 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "Results",
"sec_num": "5"
},
{
"text": "stantial. This asymmetry should not be surprising, as Chinese on its own seems to be quite a bit more difficult to parse than English.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Results",
"sec_num": "5"
},
{
"text": "We also investigated the impact of sentence length for both the training and testing sets. For our model, adding sentences of greater length to the training set leads to increases in parse accuracy for short sentences. For the baseline, however, adding this additional training data degrades performance in the case of English paired with Korean. Figure 2 summarizes the performance of our model for different sentence lengths on several of the test-sets. As shown in the figure, the largest improvements tend to occur at longer sentence lengths.",
"cite_spans": [],
"ref_spans": [
{
"start": 347,
"end": 355,
"text": "Figure 2",
"ref_id": null
}
],
"eq_spans": [],
"section": "Results",
"sec_num": "5"
},
{
"text": "We have presented a probabilistic model for bilingual grammar induction which uses raw parallel text to learn tree pairs and their alignments. Our formalism loosely binds the two trees, using bilingual patterns when possible, but allowing substantial language-specific variation. We tested our model on three test sets and showed substantial improvement over a state-of-the-art monolingual baseline. 4 ",
"cite_spans": [
{
"start": 400,
"end": 401,
"text": "4",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": "6"
},
{
"text": "Code and the outputs of our experiments are available at http://groups.csail.mit.edu/rbg/code/multiling induction.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "Sampling the alignment tree is important, as it provides us with counts of aligned constituents for the coupling parameter.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "http://www.crulp.org",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "The authors acknowledge the support of the NSF (CA-REER grant IIS-0448168, grant IIS-0835445, and grant IIS-0835652). Thanks to Amir Globerson and members of the MIT NLP group for their helpful suggestions. Any opinions, findings, or conclusions are those of the authors, and do not necessarily reflect the views of the funding organizations",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [],
"bib_entries": {
"BIBREF1": {
"ref_id": "b1",
"title": "Bayesian synchronous grammar induction",
"authors": [
{
"first": "Phil",
"middle": [],
"last": "Blunsom",
"suffix": ""
},
{
"first": "Trevor",
"middle": [],
"last": "Cohn",
"suffix": ""
},
{
"first": "Miles",
"middle": [],
"last": "Osborne",
"suffix": ""
}
],
"year": 2008,
"venue": "Proceedings of NIPS",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Phil Blunsom, Trevor Cohn, and Miles Osborne. 2008. Bayesian synchronous grammar induction. In Pro- ceedings of NIPS.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "Two languages are better than one (for syntactic parsing)",
"authors": [
{
"first": "David",
"middle": [],
"last": "Burkett",
"suffix": ""
},
{
"first": "Dan",
"middle": [],
"last": "Klein",
"suffix": ""
}
],
"year": 2008,
"venue": "Proceedings of EMNLP",
"volume": "",
"issue": "",
"pages": "877--886",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "David Burkett and Dan Klein. 2008. Two languages are better than one (for syntactic parsing). In Pro- ceedings of EMNLP, pages 877-886.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Two experiments on learning probabilistic dependency grammars from corpora",
"authors": [
{
"first": "Eugene",
"middle": [],
"last": "Charniak",
"suffix": ""
},
{
"first": "Glen",
"middle": [],
"last": "Carroll",
"suffix": ""
}
],
"year": 1992,
"venue": "Proceedings of the AAAI Workshop on Statistically-Based NLP Techniques",
"volume": "",
"issue": "",
"pages": "1--13",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Eugene Charniak and Glen Carroll. 1992. Two exper- iments on learning probabilistic dependency gram- mars from corpora. In Proceedings of the AAAI Workshop on Statistically-Based NLP Techniques, pages 1-13.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "A hierarchical phrase-based model for statistical machine translation",
"authors": [
{
"first": "David",
"middle": [],
"last": "Chiang",
"suffix": ""
}
],
"year": 2005,
"venue": "Proceedings of the ACL",
"volume": "",
"issue": "",
"pages": "263--270",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "David Chiang. 2005. A hierarchical phrase-based model for statistical machine translation. In Pro- ceedings of the ACL, pages 263-270.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "Shared logistic normal distributions for soft parameter tying in unsupervised grammar induction",
"authors": [
{
"first": "B",
"middle": [],
"last": "Shay",
"suffix": ""
},
{
"first": "Noah",
"middle": [
"A"
],
"last": "Cohen",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Smith",
"suffix": ""
}
],
"year": 2009,
"venue": "Proceedings of the NAACL/HLT",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Shay B. Cohen and Noah A. Smith. 2009. Shared lo- gistic normal distributions for soft parameter tying in unsupervised grammar induction. In Proceedings of the NAACL/HLT.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "Learning non-isomorphic tree mappings for machine translation",
"authors": [
{
"first": "Jason",
"middle": [],
"last": "Eisner",
"suffix": ""
}
],
"year": 2003,
"venue": "The Companion Volume to the Proceedings of the ACL",
"volume": "",
"issue": "",
"pages": "205--208",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Jason Eisner. 2003. Learning non-isomorphic tree mappings for machine translation. In The Compan- ion Volume to the Proceedings of the ACL, pages 205-208.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "Bayesian data analysis",
"authors": [
{
"first": "Andrew",
"middle": [],
"last": "Gelman",
"suffix": ""
},
{
"first": "John",
"middle": [
"B"
],
"last": "Carlin",
"suffix": ""
},
{
"first": "Hal",
"middle": [
"S"
],
"last": "Stern",
"suffix": ""
},
{
"first": "Donald",
"middle": [
"B"
],
"last": "Rubin",
"suffix": ""
}
],
"year": 2004,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Andrew Gelman, John B. Carlin, Hal S. Stern, and Donald B. Rubin. 2004. Bayesian data analysis. Chapman and Hall/CRC.",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "Inducing a multilingual dictionary from a parallel multitext in related languages",
"authors": [
{
"first": "Dmitriy",
"middle": [],
"last": "Genzel",
"suffix": ""
}
],
"year": 2005,
"venue": "Proceedings of EMNLP/HLT",
"volume": "",
"issue": "",
"pages": "875--882",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Dmitriy Genzel. 2005. Inducing a multilingual dictio- nary from a parallel multitext in related languages. In Proceedings of EMNLP/HLT, pages 875-882.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "Penn Korean Treebank: Development and evaluation",
"authors": [
{
"first": "C",
"middle": [],
"last": "Han",
"suffix": ""
},
{
"first": "N",
"middle": [
"R"
],
"last": "Han",
"suffix": ""
},
{
"first": "E",
"middle": [
"S"
],
"last": "Ko",
"suffix": ""
},
{
"first": "H",
"middle": [],
"last": "Yi",
"suffix": ""
},
{
"first": "M",
"middle": [],
"last": "Palmer",
"suffix": ""
}
],
"year": 2002,
"venue": "Proc. Pacific Asian Conf. Language and Comp",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "C. Han, N.R. Han, E.S. Ko, H. Yi, and M. Palmer. 2002. Penn Korean Treebank: Development and evaluation. In Proc. Pacific Asian Conf. Language and Comp.",
"links": null
},
"BIBREF10": {
"ref_id": "b10",
"title": "Monte carlo sampling methods using Markov chains and their applications",
"authors": [
{
"first": "W",
"middle": [
"K"
],
"last": "Hastings",
"suffix": ""
}
],
"year": 1970,
"venue": "Biometrika",
"volume": "57",
"issue": "",
"pages": "97--109",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "W. K. Hastings. 1970. Monte carlo sampling meth- ods using Markov chains and their applications. Biometrika, 57:97-109.",
"links": null
},
"BIBREF11": {
"ref_id": "b11",
"title": "Bootstrapping parsers via syntactic projection across parallel texts",
"authors": [
{
"first": "R",
"middle": [],
"last": "Hwa",
"suffix": ""
},
{
"first": "P",
"middle": [],
"last": "Resnik",
"suffix": ""
},
{
"first": "A",
"middle": [],
"last": "Weinberg",
"suffix": ""
},
{
"first": "C",
"middle": [],
"last": "Cabezas",
"suffix": ""
},
{
"first": "O",
"middle": [],
"last": "Kolak",
"suffix": ""
}
],
"year": 2005,
"venue": "Journal of Natural Language Engineering",
"volume": "11",
"issue": "3",
"pages": "311--325",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "R. Hwa, P. Resnik, A. Weinberg, C. Cabezas, and O. Kolak. 2005. Bootstrapping parsers via syntactic projection across parallel texts. Journal of Natural Language Engineering, 11(3):311-325.",
"links": null
},
"BIBREF12": {
"ref_id": "b12",
"title": "Alignment of trees -an alternative to tree edit",
"authors": [
{
"first": "T",
"middle": [],
"last": "Jiang",
"suffix": ""
},
{
"first": "L",
"middle": [],
"last": "Wang",
"suffix": ""
},
{
"first": "K",
"middle": [],
"last": "Zhang",
"suffix": ""
}
],
"year": 1995,
"venue": "Theoretical Computer Science",
"volume": "143",
"issue": "1",
"pages": "137--148",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "T. Jiang, L. Wang, and K. Zhang. 1995. Alignment of trees -an alternative to tree edit. Theoretical Com- puter Science, 143(1):137-148.",
"links": null
},
"BIBREF13": {
"ref_id": "b13",
"title": "Bayesian inference for PCFGs via Markov chain Monte Carlo",
"authors": [
{
"first": "M",
"middle": [],
"last": "Johnson",
"suffix": ""
},
{
"first": "T",
"middle": [],
"last": "Griffiths",
"suffix": ""
},
{
"first": "S",
"middle": [],
"last": "Goldwater",
"suffix": ""
}
],
"year": 2007,
"venue": "Proceedings of the NAACL/HLT",
"volume": "",
"issue": "",
"pages": "139--146",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "M. Johnson, T. Griffiths, and S. Goldwater. 2007. Bayesian inference for PCFGs via Markov chain Monte Carlo. In Proceedings of the NAACL/HLT, pages 139-146.",
"links": null
},
"BIBREF14": {
"ref_id": "b14",
"title": "A generative constituent-context model for improved grammar induction",
"authors": [
{
"first": "Dan",
"middle": [],
"last": "Klein",
"suffix": ""
},
{
"first": "Christopher",
"middle": [
"D"
],
"last": "Manning",
"suffix": ""
}
],
"year": 2002,
"venue": "Proceedings of the ACL",
"volume": "",
"issue": "",
"pages": "128--135",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Dan Klein and Christopher D. Manning. 2002. A generative constituent-context model for improved grammar induction. In Proceedings of the ACL, pages 128-135.",
"links": null
},
"BIBREF15": {
"ref_id": "b15",
"title": "The Unsupervised Learning of Natural Language Structure",
"authors": [
{
"first": "D",
"middle": [],
"last": "Klein",
"suffix": ""
}
],
"year": 2005,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "D. Klein. 2005. The Unsupervised Learning of Natu- ral Language Structure. Ph.D. thesis, Stanford Uni- versity.",
"links": null
},
"BIBREF16": {
"ref_id": "b16",
"title": "Experiments in parallel-text based grammar induction",
"authors": [
{
"first": "Jonas",
"middle": [],
"last": "Kuhn",
"suffix": ""
}
],
"year": 2004,
"venue": "Proceedings of the ACL",
"volume": "",
"issue": "",
"pages": "470--477",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Jonas Kuhn. 2004. Experiments in parallel-text based grammar induction. In Proceedings of the ACL, pages 470-477.",
"links": null
},
"BIBREF17": {
"ref_id": "b17",
"title": "Multitext grammars and synchronous parsers",
"authors": [
{
"first": "I",
"middle": [],
"last": "",
"suffix": ""
},
{
"first": "Dan",
"middle": [],
"last": "Melamed",
"suffix": ""
}
],
"year": 2003,
"venue": "Proceedings of the NAACL/HLT",
"volume": "",
"issue": "",
"pages": "79--86",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "I. Dan Melamed. 2003. Multitext grammars and synchronous parsers. In Proceedings of the NAACL/HLT, pages 79-86.",
"links": null
},
"BIBREF18": {
"ref_id": "b18",
"title": "A systematic comparison of various statistical alignment models",
"authors": [
{
"first": "Josef",
"middle": [],
"last": "Franz",
"suffix": ""
},
{
"first": "Hermann",
"middle": [],
"last": "Och",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Ney",
"suffix": ""
}
],
"year": 2003,
"venue": "Computational Linguistics",
"volume": "29",
"issue": "1",
"pages": "19--51",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Franz Josef Och and Hermann Ney. 2003. A sys- tematic comparison of various statistical alignment models. Computational Linguistics, 29(1):19-51.",
"links": null
},
"BIBREF19": {
"ref_id": "b19",
"title": "Fast unsupervised incremental parsing",
"authors": [
{
"first": "Yoav",
"middle": [],
"last": "Seginer",
"suffix": ""
}
],
"year": 2007,
"venue": "Proceedings of the ACL",
"volume": "",
"issue": "",
"pages": "384--391",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Yoav Seginer. 2007. Fast unsupervised incremental parsing. In Proceedings of the ACL, pages 384-391.",
"links": null
},
"BIBREF20": {
"ref_id": "b20",
"title": "Bilingual parsing with factored estimation: Using English to parse Korean",
"authors": [
{
"first": "David",
"middle": [
"A"
],
"last": "Smith",
"suffix": ""
},
{
"first": "Noah",
"middle": [
"A"
],
"last": "Smith",
"suffix": ""
}
],
"year": 2004,
"venue": "Proceeding of EMNLP",
"volume": "",
"issue": "",
"pages": "49--56",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "David A. Smith and Noah A. Smith. 2004. Bilingual parsing with factored estimation: Using English to parse Korean. In Proceeding of EMNLP, pages 49- 56.",
"links": null
},
"BIBREF21": {
"ref_id": "b21",
"title": "Unsupervised multilingual learning for morphological segmentation",
"authors": [
{
"first": "Benjamin",
"middle": [],
"last": "Snyder",
"suffix": ""
},
{
"first": "Regina",
"middle": [],
"last": "Barzilay",
"suffix": ""
}
],
"year": 2008,
"venue": "Proceedings of the ACL/HLT",
"volume": "",
"issue": "",
"pages": "737--745",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Benjamin Snyder and Regina Barzilay. 2008. Un- supervised multilingual learning for morphological segmentation. In Proceedings of the ACL/HLT, pages 737-745.",
"links": null
},
"BIBREF22": {
"ref_id": "b22",
"title": "Unsupervised multilingual learning for POS tagging",
"authors": [
{
"first": "Benjamin",
"middle": [],
"last": "Snyder",
"suffix": ""
},
{
"first": "Tahira",
"middle": [],
"last": "Naseem",
"suffix": ""
},
{
"first": "Jacob",
"middle": [],
"last": "Eisenstein",
"suffix": ""
},
{
"first": "Regina",
"middle": [],
"last": "Barzilay",
"suffix": ""
}
],
"year": 2008,
"venue": "Proceedings of EMNLP",
"volume": "",
"issue": "",
"pages": "1041--1050",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Benjamin Snyder, Tahira Naseem, Jacob Eisenstein, and Regina Barzilay. 2008. Unsupervised multi- lingual learning for POS tagging. In Proceedings of EMNLP, pages 1041-1050.",
"links": null
},
"BIBREF23": {
"ref_id": "b23",
"title": "Adding more languages improves unsupervised multilingual part-of-speech tagging: A Bayesian non-parametric approach",
"authors": [
{
"first": "Benjamin",
"middle": [],
"last": "Snyder",
"suffix": ""
},
{
"first": "Tahira",
"middle": [],
"last": "Naseem",
"suffix": ""
},
{
"first": "Jacob",
"middle": [],
"last": "Eisenstein",
"suffix": ""
},
{
"first": "Regina",
"middle": [],
"last": "Barzilay",
"suffix": ""
}
],
"year": 2009,
"venue": "Proceedings of the NAACL/HLT",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Benjamin Snyder, Tahira Naseem, Jacob Eisenstein, and Regina Barzilay. 2009. Adding more languages improves unsupervised multilingual part-of-speech tagging: A Bayesian non-parametric approach. In Proceedings of the NAACL/HLT.",
"links": null
},
"BIBREF24": {
"ref_id": "b24",
"title": "Inducing probabilistic grammars by Bayesian model merging",
"authors": [
{
"first": "Andreas",
"middle": [],
"last": "Stolcke",
"suffix": ""
},
{
"first": "Stephen",
"middle": [
"M"
],
"last": "Omohundro",
"suffix": ""
}
],
"year": 1994,
"venue": "Proceedings of ICGI",
"volume": "",
"issue": "",
"pages": "106--118",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Andreas Stolcke and Stephen M. Omohundro. 1994. Inducing probabilistic grammars by Bayesian model merging. In Proceedings of ICGI, pages 106-118.",
"links": null
},
"BIBREF25": {
"ref_id": "b25",
"title": "Machine translation with a stochastic grammatical channel",
"authors": [
{
"first": "Dekai",
"middle": [],
"last": "Wu",
"suffix": ""
},
{
"first": "Hongsing",
"middle": [],
"last": "Wong",
"suffix": ""
}
],
"year": 1998,
"venue": "Proceedings of the ACL/COLING",
"volume": "",
"issue": "",
"pages": "1408--1415",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Dekai Wu and Hongsing Wong. 1998. Machine translation with a stochastic grammatical channel. In Proceedings of the ACL/COLING, pages 1408- 1415.",
"links": null
},
"BIBREF26": {
"ref_id": "b26",
"title": "Stochastic inversion transduction grammars and bilingual parsing of parallel corpora",
"authors": [
{
"first": "Dekai",
"middle": [],
"last": "Wu",
"suffix": ""
}
],
"year": 1997,
"venue": "Computational Linguistics",
"volume": "23",
"issue": "3",
"pages": "377--403",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Dekai Wu. 1997. Stochastic inversion transduction grammars and bilingual parsing of parallel corpora. Computational Linguistics, 23(3):377-403.",
"links": null
},
"BIBREF27": {
"ref_id": "b27",
"title": "A backoff model for bootstrapping resources for non-english languages",
"authors": [
{
"first": "Chenhai",
"middle": [],
"last": "Xi",
"suffix": ""
},
{
"first": "Rebecca",
"middle": [],
"last": "Hwa",
"suffix": ""
}
],
"year": 2005,
"venue": "Proceedings of EMNLP",
"volume": "",
"issue": "",
"pages": "851--858",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Chenhai Xi and Rebecca Hwa. 2005. A backoff model for bootstrapping resources for non-english languages. In Proceedings of EMNLP, pages 851 - 858.",
"links": null
},
"BIBREF28": {
"ref_id": "b28",
"title": "Stochastic lexicalized inversion transduction grammar for alignment",
"authors": [
{
"first": "Hao",
"middle": [],
"last": "Zhang",
"suffix": ""
},
{
"first": "Daniel",
"middle": [],
"last": "Gildea",
"suffix": ""
}
],
"year": 2005,
"venue": "Proceedings of the ACL",
"volume": "",
"issue": "",
"pages": "475--482",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Hao Zhang and Daniel Gildea. 2005. Stochastic lex- icalized inversion transduction grammar for align- ment. In Proceedings of the ACL, pages 475-482.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"type_str": "figure",
"num": null,
"text": "over constituent contexts of language i, \u2022 \u03c6 D i : over distituent contexts of language i, \u2022 \u03c9 : over pairs of constituent yields, one from the first language and the other from the second language, \u2022 Gz pair : over a finite set of integer values {\u2212m, . . . , \u22122, \u22121, 0, 1, 2, . . . , m}, measuring the Giza-score of aligned tree node pairs (see below), \u2022 Gz node : over a finite set of integer values {\u2212m, . . . , \u22122, \u22121, 0}, measuring the Gizascore of unaligned tree nodes (see below).",
"uris": null
}
}
}
}