| { |
| "paper_id": "P09-1019", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T08:54:39.916090Z" |
| }, |
| "title": "Efficient Minimum Error Rate Training and Minimum Bayes-Risk Decoding for Translation Hypergraphs and Lattices", |
| "authors": [ |
| { |
| "first": "Shankar", |
| "middle": [], |
| "last": "Kumar", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Google Inc. 1600 Amphitheatre Pkwy. Mountain View", |
| "location": { |
| "postCode": "94043", |
| "region": "CA", |
| "country": "USA" |
| } |
| }, |
| "email": "shankarkumar@google.com" |
| }, |
| { |
| "first": "Wolfgang", |
| "middle": [], |
| "last": "Macherey", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Google Inc. 1600 Amphitheatre Pkwy. Mountain View", |
| "location": { |
| "postCode": "94043", |
| "region": "CA", |
| "country": "USA" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Dyer", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Maryland", |
| "location": { |
| "postCode": "20742", |
| "settlement": "College Park", |
| "region": "MD", |
| "country": "USA" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Franz", |
| "middle": [], |
| "last": "Och", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Google Inc. 1600 Amphitheatre Pkwy. Mountain View", |
| "location": { |
| "postCode": "94043", |
| "region": "CA", |
| "country": "USA" |
| } |
| }, |
| "email": "och@google.com" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Minimum Error Rate Training (MERT) and Minimum Bayes-Risk (MBR) decoding are used in most current state-of-theart Statistical Machine Translation (SMT) systems. The algorithms were originally developed to work with N-best lists of translations, and recently extended to lattices that encode many more hypotheses than typical N-best lists. We here extend lattice-based MERT and MBR algorithms to work with hypergraphs that encode a vast number of translations produced by MT systems based on Synchronous Context Free Grammars. These algorithms are more efficient than the lattice-based versions presented earlier. We show how MERT can be employed to optimize parameters for MBR decoding. Our experiments show speedups from MERT and MBR as well as performance improvements from MBR decoding on several language pairs.", |
| "pdf_parse": { |
| "paper_id": "P09-1019", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Minimum Error Rate Training (MERT) and Minimum Bayes-Risk (MBR) decoding are used in most current state-of-theart Statistical Machine Translation (SMT) systems. The algorithms were originally developed to work with N-best lists of translations, and recently extended to lattices that encode many more hypotheses than typical N-best lists. We here extend lattice-based MERT and MBR algorithms to work with hypergraphs that encode a vast number of translations produced by MT systems based on Synchronous Context Free Grammars. These algorithms are more efficient than the lattice-based versions presented earlier. We show how MERT can be employed to optimize parameters for MBR decoding. Our experiments show speedups from MERT and MBR as well as performance improvements from MBR decoding on several language pairs.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Statistical Machine Translation (SMT) systems have improved considerably by directly using the error criterion in both training and decoding. By doing so, the system can be optimized for the translation task instead of a criterion such as likelihood that is unrelated to the evaluation metric. Two popular techniques that incorporate the error criterion are Minimum Error Rate Training (MERT) (Och, 2003) and Minimum Bayes-Risk (MBR) decoding (Kumar and Byrne, 2004) . These two techniques were originally developed for N -best lists of translation hypotheses and recently extended to translation lattices Tromble et al., 2008 ) generated by a phrase-based SMT system (Och and Ney, 2004) . Translation lattices contain a significantly higher number of translation alternatives relative to Nbest lists. The extension to lattices reduces the runtimes for both MERT and MBR, and gives performance improvements from MBR decoding.", |
| "cite_spans": [ |
| { |
| "start": 393, |
| "end": 404, |
| "text": "(Och, 2003)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 443, |
| "end": 466, |
| "text": "(Kumar and Byrne, 2004)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 606, |
| "end": 626, |
| "text": "Tromble et al., 2008", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 668, |
| "end": 687, |
| "text": "(Och and Ney, 2004)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "SMT systems based on synchronous context free grammars (SCFG) (Chiang, 2007; Zollmann and Venugopal, 2006; Galley et al., 2006) have recently been shown to give competitive performance relative to phrase-based SMT. For these systems, a hypergraph or packed forest provides a compact representation for encoding a huge number of translation hypotheses (Huang, 2008) .", |
| "cite_spans": [ |
| { |
| "start": 62, |
| "end": 76, |
| "text": "(Chiang, 2007;", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 77, |
| "end": 106, |
| "text": "Zollmann and Venugopal, 2006;", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 107, |
| "end": 127, |
| "text": "Galley et al., 2006)", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 351, |
| "end": 364, |
| "text": "(Huang, 2008)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, we extend MERT and MBR decoding to work on hypergraphs produced by SCFG-based MT systems. We present algorithms that are more efficient relative to the lattice algorithms presented in Tromble et al. (2008) . Lattice MBR decoding uses a linear approximation to the BLEU score (Papineni et al., 2001) ; the weights in this linear loss are set heuristically by assuming that n-gram precisions decay exponentially with n. However, this may not be optimal in practice. We employ MERT to select these weights by optimizing BLEU score on a development set.", |
| "cite_spans": [ |
| { |
| "start": 199, |
| "end": 220, |
| "text": "Tromble et al. (2008)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 290, |
| "end": 313, |
| "text": "(Papineni et al., 2001)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "A related MBR-inspired approach for hypergraphs was developed by Zhang and Gildea (2008) . In this work, hypergraphs were rescored to maximize the expected count of synchronous constituents in the translation. In contrast, our MBR algorithm directly selects the hypothesis in the hypergraph with the maximum expected approximate corpus BLEU score (Tromble et al., 2008) . ", |
| "cite_spans": [ |
| { |
| "start": 65, |
| "end": 88, |
| "text": "Zhang and Gildea (2008)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 347, |
| "end": 369, |
| "text": "(Tromble et al., 2008)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "A translation lattice compactly encodes a large number of hypotheses produced by a phrase-based SMT system. The corresponding representation for an SMT system based on SCFGs (e.g. Chiang (2007) , Zollmann and Venugopal (2006) , Mi et al. (2008) ) is a directed hypergraph or a packed forest (Huang, 2008) .", |
| "cite_spans": [ |
| { |
| "start": 180, |
| "end": 193, |
| "text": "Chiang (2007)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 196, |
| "end": 225, |
| "text": "Zollmann and Venugopal (2006)", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 228, |
| "end": 244, |
| "text": "Mi et al. (2008)", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 291, |
| "end": 304, |
| "text": "(Huang, 2008)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Translation Hypergraphs", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Formally, a hypergraph is a pair H = V, E consisting of a vertex set V and a set of hyperedges E \u2286 V * \u00d7 V. Each hyperedge e \u2208 E connects a head vertex h(e) with a sequence of tail vertices T (e) = {v 1 , ..., v n }. The number of tail vertices is called the arity (|e|) of the hyperedge. If the arity of a hyperedge is zero, h(e) is called a source vertex. The arity of a hypergraph is the maximum arity of its hyperedges. A hyperedge of arity 1 is a regular edge, and a hypergraph of arity 1 is a regular graph (lattice). Each hyperedge is labeled with a rule r e from the SCFG. The number of nonterminals on the right-hand side of r e corresponds with the arity of e. An example without scores is shown in Figure 1 . A path in a translation hypergraph induces a translation hypothesis E along with its sequence of SCFG rules D = r 1 , r 2 , ..., r K which, if applied to the start symbol, derives E. The sequence of SCFG rules induced by a path is also called a derivation tree for E.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 709, |
| "end": 717, |
| "text": "Figure 1", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Translation Hypergraphs", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Given a set of source sentences F S 1 with corresponding reference translations R S 1 , the objective of MERT is to find a parameter set\u03bb M 1 which minimizes an automated evaluation criterion under a linear model:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "\u03bb M 1 = arg min \u03bb M 1 \uf6be S X s=1 Err`Rs,\u00ca(Fs; \u03bb M 1 )\u00b4f E(Fs; \u03bb M 1 ) = arg max E \uf6be S X s=1 \u03bbmhm(E, Fs)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "ff .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "In the context of statistical machine translation, the optimization procedure was first described in Och (2003) for N -best lists and later extended to phrase-lattices in . The algorithm is based on the insight that, under a loglinear model, the cost function of any candidate translation can be represented as a line in the plane if the initial parameter set \u03bb M 1 is shifted along a direction d M", |
| "cite_spans": [ |
| { |
| "start": 101, |
| "end": 111, |
| "text": "Och (2003)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "1 . Let C = {E 1 , ..., E K } denote a set of candidate translations, then computing the best scoring translation hypothesis\u00ca out of C results in the following optimization problem:\u00ca", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "(F ; \u03b3) = arg max E\u2208C n (\u03bb M 1 + \u03b3 \u2022 d M 1 ) \u2022 h M 1 (E, F ) o = arg max E\u2208C \uf6be X m \u03bbmhm(E, F ) | {z } =a(E,F ) + \u03b3 \u2022 X m dmhm(E, F ) | {z } =b(E,F ) ff = arg max E\u2208C\u02d8a (E, F ) + \u03b3 \u2022 b(E, F ) | {z } ( * )H", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "ence, the total score ( * ) for each candidate translation E \u2208 C can be described as a line with \u03b3 as the independent variable. For any particular choice of \u03b3, the decoder seeks that translation which yields the largest score and therefore corresponds to the topmost line segment. If \u03b3 is shifted from \u2212\u221e to +\u221e, other translation hypotheses may at some point constitute the topmost line segments and thus change the decision made by the decoder. The entire sequence of topmost line segments is called upper envelope and provides an exhaustive representation of all possible outcomes that the decoder may yield if \u03b3 is shifted along the chosen direction. Both the translations and their corresponding line segments can efficiently be computed without incorporating any error criterion. Once the envelope has been determined, the translation candidates of its constituent line segments are projected onto their corresponding error counts, thus yielding the exact and unsmoothed error surface for all candidate translations encoded in C. The error surface can now easily be traversed in order to find that\u03b3 under which the new parameter set \u03bb M 1 +\u03b3 \u2022 d M 1 minimizes the global error. In this section, we present an extension of the algorithm described in that allows us to efficiently compute and represent upper envelopes over all candidate translations encoded in hypergraphs. Conceptually, the algorithm works by propagating (initially empty) envelopes from the hypergraph's source nodes bottom-up to its unique root node, thereby expanding the envelopes by applying SCFG rules to the partial candidate translations that are associated with the envelope's constituent line segments. To recombine envelopes, we need two operators: the sum and the maximum over convex polygons. To illustrate which operator is applied when, we transform H = V, E into a regular graph with typed nodes by (1) marking all vertices v \u2208 V with the symbol \u2228 and (2) replacing each hyperedge e \u2208 E, |e| > 1, with a small subgraph consisting of a new vertex v \u2227 (e) whose incoming and outgoing edges connect the same head and tail nodes Algorithm 1 \u2227-operation (Sum) input: associative map a: V \u2192 Env(V), hyperarc e output: Minkowski sum of envelopes over T (e) for (i = 0; i < |T (e)|; ++i) { v = Ti(e); pq.enqueue", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "( v, i, 0 ); } L = \u2205; D = e, \u03b51 \u2022 \u2022 \u2022 \u03b5 |e| while (!pq.empty()) { v, i, j = pq.dequeue(); = A[v][j]; D[i+1] = .D; if (L.empty() \u2228 L.back().x < .x) { if (0 < j) { .y += L.back().y -A[v][j-1].y; .m += L.back().m -A[v][j-1].m; } L.push_back( ); L.back().D = D; } else { L.back().y += .y; L.back().m += .m; L.back().D[i+1] = .D; if (0 < j) { L.back().y -= A[v][j-1].y; L.back().m -= A[v][j-1].m; } } if (++j < A[v].size()) pq.enqueue( v, i, j ); } return L;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "in the transformed graph as were connected by e in the original graph. The unique outgoing edge of v \u2227 (e) is associated with the rule r e ; incoming edges are not linked to any rule. Figure 2 illustrates the transformation for a hyperedge with arity 3. The graph transformation is isomorphic.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 184, |
| "end": 192, |
| "text": "Figure 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "The rules associated with every hyperedge specify how line segments in the envelopes of a hyperedge's tail nodes can be combined. Suppose we have a hyperedge e with rule r e : X \u2192 aX 1 bX 2 c and T (e) = {v 1 , v 2 }. Then we substitute X 1 and X 2 in the rule with candidate translations associated with line segments in envelopes Env(v 1 ) and Env(v 2 ) respectively.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "To derive the algorithm, we consider the general case of a hyperedge e with rule r e : X \u2192 w 1 X 1 w 2 ...w n X n w n+1 . Because the right-hand side of r e has n nonterminals, the arity of e is |e| = n. Let T (e) = {v 1 , ..., v n } denote the tail nodes of e. We now assume that each tail node v i \u2208 T (e) is associated with the upper envelope over all candidate translations that are induced by derivations of the corresponding nonterminal symbol X i . These envelopes shall be de-", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Algorithm 2 \u2228-operation (Max) input: array L[0..K-1] containing line objects output: upper envelope of L Sort(L:m); j = 0; K = size(L); for (i = 0; i < K; ++i) { = L[i]; .x = -\u221e; if (0 < j) { if (L[j-1].m == .m) { if ( .y <= L[j-1].y) continue; --j; } while (0 < j) { .x = ( .y -L[j-1].y)/ (L[j-1].m -.m); if (L[j-1].x < .x) break; --j; } if (0 == j) .x = -\u221e; L[j++] = ; } else L[j++] = ; } L.resize(j); return L; noted by Env(v i ).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "To decompose the problem of computing and propagating the tail envelopes over the hyperedge e to its head node, we now define two operations, one for either node type, to specify how envelopes associated with the tail vertices are propagated to the head vertex.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Nodes of Type \"\u2227\": For a type \u2227 node, the resulting envelope is the Minkowski sum over the envelopes of the incoming edges (Berg et al., 2008) . Since the envelopes of the incoming edges are convex hulls, the Minkowski sum provides an upper bound to the number of line segments that constitute the resulting envelope: the bound is the sum over the number of line segments in the envelopes of the incoming edges, i.e.:", |
| "cite_spans": [ |
| { |
| "start": 123, |
| "end": 142, |
| "text": "(Berg et al., 2008)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Env(v \u2227 (e)) \u2264 v\u2228\u2208T (e) Env(v \u2228 ) .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Algorithm 1 shows the pseudo code for computing the Minkowski sum over multiple envelopes. The line objects used in this algorithm are encoded as 4-tuples, each consisting of the xintercept with 's left-adjacent line stored as .x, the slope .m, the y-intercept .y, and the (partial) derivation tree .D. At the beginning, the leftmost line segment of each envelope is inserted into a priority queue pq. The priority is defined in terms of a line's x-intercept such that lower values imply higher priority. Hence, the priority queue enumerates all line segments from left to right in ascending order of their x-intercepts, which is the order needed to compute the Minkowski sum.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Nodes of Type \"\u2228\": The operation performed", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "= ! = max", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Figure 2: Transformation of a hypergraph into a factor graph and bottom-up propagation of envelopes.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "at nodes of type \"\u2228\" computes the convex hull over the union of the envelopes propagated over the incoming edges. This operation is a \"max\" operation and it is identical to the algorithm described in for phrase lattices. Algorithm 2 contains the pseudo code.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "The complete algorithm then works as follows: Traversing all nodes in H bottom-up in topological order, we proceed for each node v \u2208 V over its incoming hyperedges and combine in each such hyperedge e the envelopes associated with the tail nodes T (e) by computing their sum according to Algorithm 1 (\u2227-operation). For each incoming hyperedge e, the resulting envelope is then expanded by applying the rule r e to its constituent line segments. The envelopes associated with different incoming hyperedges of node v are then combined and reduced according to Algorithm 2 (\u2228-operation). By construction, the envelope at the root node is the convex hull over the line segments of all candidate translations that can be derived from the hypergraph.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "The suggested algorithm has similar properties as the algorithm presented in . In particular, it has the same upper bound on the number of line segments that constitute the envelope at the root node, i.e, the size of this envelope is guaranteed to be no larger than the number of edges in the transformed hypergraph.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Error Rate Training", |
| "sec_num": "3" |
| }, |
| { |
| "text": "We first review Minimum Bayes-Risk (MBR) decoding for statistical MT. An MBR decoder seeks the hypothesis with the least expected loss under a probability model (Bickel and Doksum, 1977 ). If we think of statistical MT as a classifier that maps a source sentence F to a target sentence E, the MBR decoder can be expressed as follows:", |
| "cite_spans": [ |
| { |
| "start": 161, |
| "end": 185, |
| "text": "(Bickel and Doksum, 1977", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Bayes-Risk Decoding", |
| "sec_num": "4" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "E = argmin E \u2208G E\u2208G L(E, E )P (E|F ),", |
| "eq_num": "(1)" |
| } |
| ], |
| "section": "Minimum Bayes-Risk Decoding", |
| "sec_num": "4" |
| }, |
| { |
| "text": "where L(E, E ) is the loss between any two hypotheses E and E , P (E|F ) is the probability model, and G is the space of translations (N -best list, lattice, or a hypergraph). MBR decoding for translation can be performed by reranking an N -best list of hypotheses generated by an MT system (Kumar and Byrne, 2004) . This reranking can be done for any sentencelevel loss function such as BLEU (Papineni et al., 2001) , Word Error Rate, or Position-independent Error Rate.", |
| "cite_spans": [ |
| { |
| "start": 291, |
| "end": 314, |
| "text": "(Kumar and Byrne, 2004)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 393, |
| "end": 416, |
| "text": "(Papineni et al., 2001)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Bayes-Risk Decoding", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Recently, Tromble et al. (2008) extended MBR decoding to translation lattices under an approximate BLEU score. They approximated log(BLEU) score by a linear function of n-gram matches and candidate length. If E and E are the reference and the candidate translations respectively, this linear function is given by:", |
| "cite_spans": [ |
| { |
| "start": 10, |
| "end": 31, |
| "text": "Tromble et al. (2008)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Bayes-Risk Decoding", |
| "sec_num": "4" |
| }, |
| { |
| "text": "G(E, E ) = \u03b8 0 |E | + w \u03b8 |w| # w (E )\u03b4 w (E), (2)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Bayes-Risk Decoding", |
| "sec_num": "4" |
| }, |
| { |
| "text": "where w is an n-gram present in either E or E , and \u03b8 0 , \u03b8 1 , ..., \u03b8 N are weights which are determined empirically, where N is the maximum ngram order.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Bayes-Risk Decoding", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Under such a linear decomposition, the MBR decoder (Equation 1) can be written a\u015d", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Bayes-Risk Decoding", |
| "sec_num": "4" |
| }, |
| { |
| "text": "E = argmax E \u2208G \u03b8 0 |E | + w \u03b8 |w| # w (E )p(w|G), (3)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Bayes-Risk Decoding", |
| "sec_num": "4" |
| }, |
| { |
| "text": "where the posterior probability of an n-gram in the lattice is given by", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Bayes-Risk Decoding", |
| "sec_num": "4" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "p(w|G) = E\u2208G 1 w (E)P (E|F ).", |
| "eq_num": "(4)" |
| } |
| ], |
| "section": "Minimum Bayes-Risk Decoding", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Tromble et al. (2008) implement the MBR decoder using Weighted Finite State Automata (WFSA) operations. First, the set of n-grams is extracted from the lattice. Next, the posterior probability of each n-gram is computed. A new automaton is then created by intersecting each ngram with weight (from Equation 2) to an unweighted lattice. Finally, the MBR hypothesis is extracted as the best path in the automaton. We will refer to this procedure as FSAMBR.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Bayes-Risk Decoding", |
| "sec_num": "4" |
| }, |
| { |
| "text": "The above steps are carried out one n-gram at a time. For a moderately large lattice, there can be several thousands of n-grams and the procedure becomes expensive. We now present an alternate approximate procedure which can avoid this enumeration making the resulting algorithm much faster than FSAMBR.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Minimum Bayes-Risk Decoding", |
| "sec_num": "4" |
| }, |
| { |
| "text": "The key idea behind this new algorithm is to rewrite the n-gram posterior probability (Equation 4) as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Efficient MBR for lattices", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "p(w|G) = E\u2208G e\u2208E f (e, w, E)P (E|F ) (5)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Efficient MBR for lattices", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "where f (e, w, E) is a score assigned to edge e on path E containing n-gram w:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Efficient MBR for lattices", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "f (e, w, E) = \uf8f1 \uf8f2 \uf8f3 1 w \u2208 e, p(e|G) > p(e |G),", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Efficient MBR for lattices", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "e precedes e on E 0 otherwise 6In other words, for each path E, we count the edge that contributes n-gram w and has the highest edge posterior probability relative to its predecessors on the path E; there is exactly one such edge on each lattice path E.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Efficient MBR for lattices", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "We note that f (e, w, E) relies on the full path E which means that it cannot be computed based on local statistics. We therefore approximate the quantity f (e, w, E) with f * (e, w, G) that counts the edge e with n-gram w that has the highest arc posterior probability relative to predecessors in the entire lattice G. f * (e, w, G) can be computed locally, and the n-gram posterior probability based on f * can be determined as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Efficient MBR for lattices", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "p(w|G) = X E\u2208G X e\u2208E f * (e, w, G)P (E|F ) (7) = X e\u2208E 1w\u2208ef * (e, w, G) X E\u2208G 1E(e)P (E|F ) = X e\u2208E 1w\u2208ef * (e, w, G)P (e|G),", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Efficient MBR for lattices", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "where P (e|G) is the posterior probability of a lattice edge. The algorithm to perform Lattice MBR is given in Algorithm 3. For each node t in the lattice, we maintain a quantity Score(w, t) for each n-gram w that lies on a path from the source node to t. Score(w, t) is the highest posterior probability among all edges on the paths that terminate on t and contain n-gram w. The forward pass requires computing the n-grams introduced by each edge; to do this, we propagate n-grams (up to maximum order \u22121) terminating on each node.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Efficient MBR for lattices", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "We next extend the Lattice MBR decoding algorithm (Algorithm 3) to rescore hypergraphs produced by a SCFG based MT system. Algorithm 4 is an extension to the MBR decoder on lattices Algorithm 3 MBR Decoding on Lattices 1: Sort the lattice nodes topologically. 2: Compute backward probabilities of each node. 3: Compute posterior prob. of each n-gram: 4: for each edge e do 5: Compute edge posterior probability P (e|G).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Extension to Hypergraphs", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Compute n-gram posterior probs. P (w|G): 7:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "6:", |
| "sec_num": null |
| }, |
| { |
| "text": "for each n-gram w introduced by e do 8:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "6:", |
| "sec_num": null |
| }, |
| { |
| "text": "Propagate n \u2212 1 gram suffix to he. 9:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "6:", |
| "sec_num": null |
| }, |
| { |
| "text": "if p(e|G) > Score(w, T (e)) then 10:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "6:", |
| "sec_num": null |
| }, |
| { |
| "text": "Update posterior probs. and scores: p(w|G) += p(e|G) \u2212 Score(w, T (e)). Score(w, he) = p(e|G). 11: else 12:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "6:", |
| "sec_num": null |
| }, |
| { |
| "text": "Score(w, he) = Score(w, T (e)).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "6:", |
| "sec_num": null |
| }, |
| { |
| "text": "end if 14:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "13:", |
| "sec_num": null |
| }, |
| { |
| "text": "end for 15: end for 16: Assign scores to edges (given by Equation 3). 17: Find best path in the lattice (Equation 3).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "13:", |
| "sec_num": null |
| }, |
| { |
| "text": "(Algorithm 3). However, there are important differences when computing the n-gram posterior probabilities (", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "13:", |
| "sec_num": null |
| }, |
| { |
| "text": "Step 3). In this inside pass, we now maintain both n-gram prefixes and suffixes (up to the maximum order \u22121) on each hypergraph node. This is necessary because unlike a lattice, new ngrams may be created at subsequent nodes by concatenating words both to the left and the right side of the n-gram. When the arity of the edge is 2, a rule has the general form aX 1 bX 2 c, where X 1 and X 2 are sequences from tail nodes. As a result, we need to consider all new sequences which can be created by the cross-product of the n-grams on the two tail nodes. E.g. if X 1 = {c, cd, d} and X 2 = {f, g}, then a total of six sequences will result. In practice, such a cross-product is not pro-Algorithm 4 MBR Decoding on Hypergraphs 1: Sort the hypergraph nodes topologically. 2: Compute inside probabilities of each node. 3: Compute posterior prob. of each hyperedge P (e|G). 4: Compute posterior prob. of each n-gram: 5: for each hyperedge e do 6:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "13:", |
| "sec_num": null |
| }, |
| { |
| "text": "Merge the n-grams on the tail nodes T (e). If the same n-gram is present on multiple tail nodes, keep the highest score.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "13:", |
| "sec_num": null |
| }, |
| { |
| "text": "Apply the rule on e to the n-grams on T (e).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "7:", |
| "sec_num": null |
| }, |
| { |
| "text": "Propagate n \u2212 1 gram prefixes/suffixes to he. 9:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "8:", |
| "sec_num": null |
| }, |
| { |
| "text": "for each n-gram w introduced by this hyperedge do 10:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "8:", |
| "sec_num": null |
| }, |
| { |
| "text": "if p(e|G) > Score(w, T (e)) then 11: p(w|G) += p(e|G) \u2212 Score(w, T (e)) Score(w, he) = p(e|G) 12: else 13: Score(w, he) = Score(w, T (e)) 14:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "8:", |
| "sec_num": null |
| }, |
| { |
| "text": "end if 15: end for 16: end for 17: Assign scores to hyperedges (Equation 3). 18: Find best path in the hypergraph (Equation 3).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "8:", |
| "sec_num": null |
| }, |
| { |
| "text": "hibitive when the maximum n-gram order in MBR does not exceed the order of the n-gram language model used in creating the hypergraph. In the latter case, we will have a small set of unique prefixes and suffixes on the tail nodes.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "8:", |
| "sec_num": null |
| }, |
| { |
| "text": "Lattice MBR Decoding (Equation 3) assumes a linear form for the gain function (Equation 2). This linear function contains n + 1 parameters", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "MERT for MBR Parameter Optimization", |
| "sec_num": "5" |
| }, |
| { |
| "text": "\u03b8 0 , \u03b8 1 , ..., \u03b8 N ,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "MERT for MBR Parameter Optimization", |
| "sec_num": "5" |
| }, |
| { |
| "text": "where N is the maximum order of the n-grams involved. Tromble et al. (2008) obtained these factors as a function of n-gram precisions derived from multiple training runs. However, this does not guarantee that the resulting linear score (Equation 2) is close to the corpus BLEU. We now describe how MERT can be used to estimate these factors to achieve a better approximation to the corpus BLEU. We recall that MERT selects weights in a linear model to optimize an error criterion (e.g. corpus BLEU) on a training set. The lattice MBR decoder (Equation 3) can be written as a linear model:", |
| "cite_spans": [ |
| { |
| "start": 54, |
| "end": 75, |
| "text": "Tromble et al. (2008)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "MERT for MBR Parameter Optimization", |
| "sec_num": "5" |
| }, |
| { |
| "text": "\u00ca = argmax E \u2208G N i=0 \u03b8 i g i (E , F ), where g 0 (E , F ) = |E | and g i (E , F ) = w:|w|=i # w (E )p(w|G).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "MERT for MBR Parameter Optimization", |
| "sec_num": "5" |
| }, |
| { |
| "text": "The linear approximation to BLEU may not hold in practice for unseen test sets or languagepairs. Therefore, we would like to allow the decoder to backoff to the MAP translation in such cases. To do that, we introduce an additional feature function g N +1 (E, F ) equal to the original decoder cost for this sentence. A weight assignment of 1.0 for this feature function and zeros for the other feature functions would imply that the MAP translation is chosen. We now have a total of N +2 feature functions which we optimize using MERT to obtain highest BLEU score on a training set.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "MERT for MBR Parameter Optimization", |
| "sec_num": "5" |
| }, |
| { |
| "text": "We now describe our experiments to evaluate MERT and MBR on lattices and hypergraphs, and show how MERT can be used to tune MBR parameters.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "6" |
| }, |
| { |
| "text": "We report results on two tasks. The first one is the constrained data track of the NIST Arabicto-English (aren) and Chinese-to-English (zhen) translation task 1 . On this task, the parallel and the 1 http://www.nist.gov/speech/tests/mt Dataset # of sentences aren zhen dev 1797 1664 nist02 1043 878 nist03 663 919 monolingual data included all the allowed training sets for the constrained track. Table 1 reports statistics computed over these data sets. Our development set (dev) consists of the NIST 2005 eval set; we use this set for optimizing MBR parameters. We report results on NIST 2002 and NIST 2003 evaluation sets.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 236, |
| "end": 313, |
| "text": "Dataset # of sentences aren zhen dev 1797 1664 nist02 1043 878 nist03", |
| "ref_id": "TABREF0" |
| }, |
| { |
| "start": 405, |
| "end": 412, |
| "text": "Table 1", |
| "ref_id": "TABREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Translation Tasks", |
| "sec_num": "6.1" |
| }, |
| { |
| "text": "The second task consists of systems for 39 language-pairs with English as the target language and trained on at most 300M word tokens mined from the web and other published sources. The development and test sets for this task are randomly selected sentences from the web, and contain 5000 and 1000 sentences respectively.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Translation Tasks", |
| "sec_num": "6.1" |
| }, |
| { |
| "text": "Our phrase-based statistical MT system is similar to the alignment template system described in (Och and Ney, 2004; Tromble et al., 2008) . Translation is performed using a standard dynamic programming beam-search decoder (Och and Ney, 2004) using two decoding passes. The first decoder pass generates either a lattice or an N -best list. MBR decoding is performed in the second pass.", |
| "cite_spans": [ |
| { |
| "start": 96, |
| "end": 115, |
| "text": "(Och and Ney, 2004;", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 116, |
| "end": 137, |
| "text": "Tromble et al., 2008)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 222, |
| "end": 241, |
| "text": "(Och and Ney, 2004)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "MT System Description", |
| "sec_num": "6.2" |
| }, |
| { |
| "text": "We also train two SCFG-based MT systems: a hierarchical phrase-based SMT (Chiang, 2007) system and a syntax augmented machine translation (SAMT) system using the approach described in Zollmann and Venugopal (2006) . Both systems are built on top of our phrase-based systems. In these systems, the decoder generates an initial hypergraph or an N -best list, which are then rescored using MBR decoding. Table 2 shows runtime experiments for the hypergraph MERT implementation in comparison with the phrase-lattice implementation on both the aren and the zhen system. The first two columns show the average amount of time in msecs that either algorithm requires to compute the upper envelope when applied to phrase lattices. Compared to the algorithm described in which is optimized for phrase lattices, the hypergraph implementation causes a small increase in Table 2 : Average time for computing envelopes.", |
| "cite_spans": [ |
| { |
| "start": 73, |
| "end": 87, |
| "text": "(Chiang, 2007)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 184, |
| "end": 213, |
| "text": "Zollmann and Venugopal (2006)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 401, |
| "end": 408, |
| "text": "Table 2", |
| "ref_id": null |
| }, |
| { |
| "start": 858, |
| "end": 865, |
| "text": "Table 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "MT System Description", |
| "sec_num": "6.2" |
| }, |
| { |
| "text": "running time. This increase is mainly due to the representation of line segments; while the phraselattice implementation stores a single backpointer, the hypergraph version stores a vector of backpointers.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "MERT Results", |
| "sec_num": "6.3" |
| }, |
| { |
| "text": "The last two columns show the average amount of time that is required to compute the upper envelope on hypergraphs. For comparison, we prune hypergraphs to the same density (# of edges per edge on the best path) and achieve identical running times for computing the error surface.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "MERT Results", |
| "sec_num": "6.3" |
| }, |
| { |
| "text": "We first compare the new lattice MBR (Algorithm 3) with MBR decoding on 1000-best lists and FSAMBR (Tromble et al., 2008) on lattices generated by the phrase-based systems; evaluation is done using both BLEU and average run-time per sentence ( Table 3) . Note that N -best MBR uses a sentence BLEU loss function. The new lattice MBR algorithm gives about the same performance as FSAMBR while yielding a 20X speedup.", |
| "cite_spans": [ |
| { |
| "start": 99, |
| "end": 121, |
| "text": "(Tromble et al., 2008)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 244, |
| "end": 252, |
| "text": "Table 3)", |
| "ref_id": "TABREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "MBR Results", |
| "sec_num": "6.4" |
| }, |
| { |
| "text": "We next report the performance of MBR on hypergraphs generated by Hiero/SAMT systems. Table 4 compares Hypergraph MBR (HGMBR) with MAP and MBR decoding on 1000 best lists. On some systems such as the Arabic-English SAMT, the gains from Hypergraph MBR over 1000-best MBR are significant. In other cases, Hypergraph MBR performs at least as well as N -best MBR. In all cases, we observe a 7X speedup in runtime. This shows the usefulness of Hypergraph MBR decoding as an efficient alternative to Nbest MBR.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "MBR Results", |
| "sec_num": "6.4" |
| }, |
| { |
| "text": "We now describe the results by tuning MBR ngram parameters (Equation 2) using MERT. We first compute N + 1 MBR feature functions on each edge of the lattice/hypergraph. We also include the total decoder cost on the edge as as additional feature function. MERT is then performed to optimize the BLEU score on a development set; For MERT, we use 40 random initial parameters as well as parameters computed using corpus based statistics (Tromble et al., 2008 Table 5 shows results for NIST systems. We report results on nist03 set and present three systems for each language pair: phrase-based (pb), hierarchical (hier), and SAMT; Lattice MBR is done for the phrase-based system while HGMBR is used for the other two. We select the MBR scaling factor (Tromble et al., 2008) based on the development set; it is set to 0.1, 0.01, 0.5, 0.2, 0.5 and 1.0 for the aren-phrase, aren-hier, aren-samt, zhen-phrase zhen-hier and zhen-samt systems respectively. For the multi-language case, we train phrase-based systems and perform lattice MBR for all language pairs. We use a scaling factor of 0.7 for all pairs. Additional gains can be obtained by tuning this factor; however, we do not explore that dimension in this paper. In all cases, we prune the lattices/hypergraphs to a density of 30 using forward-backward pruning (Sixtus and Ortmanns, 1999) .", |
| "cite_spans": [ |
| { |
| "start": 434, |
| "end": 455, |
| "text": "(Tromble et al., 2008", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 748, |
| "end": 770, |
| "text": "(Tromble et al., 2008)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 1312, |
| "end": 1339, |
| "text": "(Sixtus and Ortmanns, 1999)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 456, |
| "end": 463, |
| "text": "Table 5", |
| "ref_id": "TABREF6" |
| } |
| ], |
| "eq_spans": [], |
| "section": "MBR Parameter Tuning with MERT", |
| "sec_num": "6.5" |
| }, |
| { |
| "text": "We consider a BLEU score difference to be a) gain if is at least 0.2 points, b) drop if it is at most -0.2 points, and c) no change otherwise. The results are shown in Table 6 . In both tables, the following results are reported: Lattice/HGMBR with default parameters (\u22125, 1.5, 2, 3, 4) computed using corpus statistics (Tromble et al., 2008) , Lattice/HGMBR with parameters derived from MERT both without/with the baseline model cost feature (mert\u2212b, mert+b). For multi-language systems, we only show the # of language-pairs with gains/no-changes/drops for each MBR variant with respect to the MAP translation.", |
| "cite_spans": [ |
| { |
| "start": 320, |
| "end": 342, |
| "text": "(Tromble et al., 2008)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 168, |
| "end": 175, |
| "text": "Table 6", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "MBR Parameter Tuning with MERT", |
| "sec_num": "6.5" |
| }, |
| { |
| "text": "We observed in the NIST systems that MERT resulted in short translations relative to MAP on the unseen test set. To prevent this behavior, we modify the MERT error criterion to include a sentence-level brevity scorer with parameter \u03b1: BLEU+brevity(\u03b1). This brevity scorer penalizes each candidate translation that is shorter than the average length over its reference translations, using a penalty term which is linear in the difference between either length. We tune \u03b1 on the development set so that the brevity score of MBR translation is close to that of the MAP translation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "MBR Parameter Tuning with MERT", |
| "sec_num": "6.5" |
| }, |
| { |
| "text": "In the NIST systems, MERT yields small improvements on top of MBR with default parameters. This is the case for Arabic-English Hiero/SAMT. In all other cases, we see no change or even a slight degradation due to MERT. We hypothesize that the default MBR parameters (Tromble et al., 2008) are well tuned. Therefore there is little gain by additional tuning using MERT.", |
| "cite_spans": [ |
| { |
| "start": 265, |
| "end": 287, |
| "text": "(Tromble et al., 2008)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "MBR Parameter Tuning with MERT", |
| "sec_num": "6.5" |
| }, |
| { |
| "text": "In the multi-language systems, the results show a different trend. We observe that MBR with default parameters results in gains on 18 pairs, no differences on 9 pairs, and losses on 12 pairs. When we optimize MBR features with MERT, the number of language pairs with gains/no changes/drops is 22/5/12. Thus, MERT has a bigger impact here than in the NIST systems. We hypothesize that the default MBR parameters are sub-optimal for some language pairs and that MERT helps to find better parameter settings. In particular, MERT avoids the need for manually tuning these parameters by language pair.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "MBR Parameter Tuning with MERT", |
| "sec_num": "6.5" |
| }, |
| { |
| "text": "Finally, when baseline model costs are added as an extra feature (mert+b), the number of pairs with gains/no changes/drops is 26/8/5. This shows that this feature can allow MBR decoding to backoff to the MAP translation. When MBR does not produce a higher BLEU score relative to MAP on the development set, MERT assigns a higher weight to this feature function. We see such an effect for 4 systems.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "MBR Parameter Tuning with MERT", |
| "sec_num": "6.5" |
| }, |
| { |
| "text": "We have presented efficient algorithms which extend previous work on lattice-based MERT and MBR decoding (Tromble et al., 2008) to work with hypergraphs. Our new MERT algorithm can work with both lattices and hypergraphs. On lattices, it achieves similar run-times as the implementation Table 6 : MBR on Multi-language systems.", |
| "cite_spans": [ |
| { |
| "start": 105, |
| "end": 127, |
| "text": "(Tromble et al., 2008)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 287, |
| "end": 294, |
| "text": "Table 6", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "7" |
| }, |
| { |
| "text": "described in . The new Lattice MBR decoder achieves a 20X speedup relative to either FSAMBR implementation described in Tromble et al. (2008) or MBR on 1000-best lists. The algorithm gives comparable results relative to FSAMBR. On hypergraphs produced by Hierarchical and Syntax Augmented MT systems, our MBR algorithm gives a 7X speedup relative to 1000-best MBR while giving comparable or even better performance.", |
| "cite_spans": [ |
| { |
| "start": 120, |
| "end": 141, |
| "text": "Tromble et al. (2008)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "7" |
| }, |
| { |
| "text": "Lattice MBR decoding is obtained under a linear approximation to BLEU, where the weights are obtained using n-gram precisions derived from development data. This may not be optimal in practice for unseen test sets and language pairs, and the resulting linear loss may be quite different from the corpus level BLEU. In this paper, we have described how MERT can be employed to estimate the weights for the linear loss function to maximize BLEU on a development set. On an experiment with 40 language pairs, we obtain improvements on 26 pairs, no difference on 8 pairs and drops on 5 pairs. This was achieved without any need for manual tuning for each language pair. The baseline model cost feature helps the algorithm effectively back off to the MAP translation in language pairs where MBR features alone would not have helped.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "7" |
| }, |
| { |
| "text": "MERT and MBR decoding are popular techniques for incorporating the final evaluation metric into the development of SMT systems. We believe that our efficient algorithms will make them more widely applicable in both SCFG-based and phrase-based MT systems.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "7" |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Computational Geometry: Algorithms and Applications, chapter 13", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Berg", |
| "suffix": "" |
| }, |
| { |
| "first": "O", |
| "middle": [], |
| "last": "Cheong", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Krefeld", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Overmars", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "290--296", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "M. Berg, O. Cheong, M. Krefeld, and M. Overmars, 2008. Computational Geometry: Algorithms and Applications, chapter 13, pages 290-296. Springer- Verlag, 3rd edition.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Mathematical Statistics: Basic Ideas and Selected topics", |
| "authors": [ |
| { |
| "first": "P", |
| "middle": [ |
| "J" |
| ], |
| "last": "Bickel", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [ |
| "A" |
| ], |
| "last": "Doksum", |
| "suffix": "" |
| } |
| ], |
| "year": 1977, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "P. J. Bickel and K. A. Doksum. 1977. Mathematical Statistics: Basic Ideas and Selected topics. Holden- Day Inc., Oakland, CA, USA.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Hierarchical phrase based translation", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Chiang", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Computational Linguistics", |
| "volume": "33", |
| "issue": "2", |
| "pages": "201--228", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "D. Chiang. 2007. Hierarchical phrase based transla- tion . Computational Linguistics, 33(2):201 -228.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Scalable Inference and Training of Context-Rich Syntactic Translation Models", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Galley", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Graehl", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Knight", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Marcu", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Deneefe", |
| "suffix": "" |
| }, |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "I", |
| "middle": [], |
| "last": "Thayer", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "COLING/ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "M. Galley, J. Graehl, K. Knight, D. Marcu, S. DeNeefe, W. Wang, and I. Thayer. 2006. Scalable Inference and Training of Context-Rich Syntactic Translation Models. . In COLING/ACL, Sydney, Australia.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Advanced Dynamic Programming in Semiring and Hypergraph Frameworks", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "COL-ING", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "L. Huang. 2008. Advanced Dynamic Programming in Semiring and Hypergraph Frameworks. In COL- ING, Manchester, UK.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Minimum Bayes-Risk Decoding for Statistical Machine Translation", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Kumar", |
| "suffix": "" |
| }, |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Byrne", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "HLT-NAACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "S. Kumar and W. Byrne. 2004. Minimum Bayes- Risk Decoding for Statistical Machine Translation. In HLT-NAACL, Boston, MA, USA.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Lattice-based Minimum Error Rate Training for Statistical Machine Translation", |
| "authors": [ |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Macherey", |
| "suffix": "" |
| }, |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Och", |
| "suffix": "" |
| }, |
| { |
| "first": "I", |
| "middle": [], |
| "last": "Thayer", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Uszkoreit", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "EMNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "W. Macherey, F. Och, I. Thayer, and J. Uszkoreit. 2008. Lattice-based Minimum Error Rate Train- ing for Statistical Machine Translation. In EMNLP, Honolulu, Hawaii, USA.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Forest-Based Translation", |
| "authors": [ |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Mi", |
| "suffix": "" |
| }, |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "" |
| }, |
| { |
| "first": "Q", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "H. Mi, L. Huang, and Q. Liu. 2008. Forest-Based Translation. In ACL, Columbus, OH, USA.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "The Alignment Template Approach to Statistical Machine Translation", |
| "authors": [ |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Och", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Ney", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Computational Linguistics", |
| "volume": "30", |
| "issue": "4", |
| "pages": "417--449", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "F. Och and H. Ney. 2004. The Alignment Template Approach to Statistical Machine Translation. Com- putational Linguistics, 30(4):417 -449.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Minimum Error Rate Training in Statistical Machine Translation", |
| "authors": [ |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Och", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "F. Och. 2003. Minimum Error Rate Training in Statis- tical Machine Translation. In ACL, Sapporo, Japan.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Bleu: a Method for Automatic Evaluation of Machine Translation", |
| "authors": [ |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Papineni", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Roukos", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Ward", |
| "suffix": "" |
| }, |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Zhu", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "K. Papineni, S. Roukos, T. Ward, and W. Zhu. 2001. Bleu: a Method for Automatic Evaluation of Ma- chine Translation. Technical Report RC22176 (W0109-022), IBM Research Division.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "High Quality Word Graphs Using Forward-Backward Pruning", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Sixtus", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Ortmanns", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "ICASSP", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "A. Sixtus and S. Ortmanns. 1999. High Quality Word Graphs Using Forward-Backward Pruning. In ICASSP, Phoenix, AZ, USA.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Lattice Minimum Bayes-Risk Decoding for Statistical Machine Translation", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Tromble", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Kumar", |
| "suffix": "" |
| }, |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Och", |
| "suffix": "" |
| }, |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Macherey", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "EMNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "R. Tromble, S. Kumar, F. Och, and W. Macherey. 2008. Lattice Minimum Bayes-Risk Decoding for Statis- tical Machine Translation. In EMNLP, Honolulu, Hawaii.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Efficient Multi-pass Decoding for Synchronous Context Free Grammars", |
| "authors": [ |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Gildea", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "H. Zhang and D. Gildea. 2008. Efficient Multi-pass Decoding for Synchronous Context Free Grammars. In ACL, Columbus, OH, USA.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Syntax Augmented Machine Translation via Chart Parsing", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Zollmann", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Venugopal", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "HLT-NAACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "A. Zollmann and A. Venugopal. 2006. Syntax Aug- mented Machine Translation via Chart Parsing. In HLT-NAACL, New York, NY, USA.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF1": { |
| "text": "An example hypergraph.", |
| "type_str": "figure", |
| "uris": null, |
| "num": null |
| }, |
| "TABREF0": { |
| "text": "Statistics over the NIST dev/test sets.", |
| "type_str": "table", |
| "num": null, |
| "content": "<table/>", |
| "html": null |
| }, |
| "TABREF3": { |
| "text": "Lattice MBR for a phrase-based system.", |
| "type_str": "table", |
| "num": null, |
| "content": "<table><tr><td/><td/><td colspan=\"2\">BLEU (%)</td><td/><td>Avg.</td></tr><tr><td/><td>aren</td><td/><td>zhen</td><td/><td>time</td></tr><tr><td/><td colspan=\"5\">nist03 nist02 nist03 nist02 (ms.)</td></tr><tr><td/><td/><td>Hiero</td><td/><td/><td/></tr><tr><td>MAP</td><td>52.8</td><td>62.9</td><td>41.0</td><td>39.8</td><td>-</td></tr><tr><td>N -best MBR</td><td>53.2</td><td>63.0</td><td>41.0</td><td>40.1</td><td>3.7</td></tr><tr><td>HGMBR</td><td>53.3</td><td>63.1</td><td>41.0</td><td>40.2</td><td>0.5</td></tr><tr><td/><td colspan=\"2\">SAMT</td><td/><td/><td/></tr><tr><td>MAP</td><td>53.4</td><td>63.9</td><td>41.3</td><td>40.3</td><td>-</td></tr><tr><td>N -best MBR</td><td>53.8</td><td>64.3</td><td>41.7</td><td>41.1</td><td>3.7</td></tr><tr><td>HGMBR</td><td>54.0</td><td>64.6</td><td>41.8</td><td>41.1</td><td>0.5</td></tr></table>", |
| "html": null |
| }, |
| "TABREF4": { |
| "text": "Hypergraph MBR for Hiero/SAMT systems.", |
| "type_str": "table", |
| "num": null, |
| "content": "<table/>", |
| "html": null |
| }, |
| "TABREF6": { |
| "text": "MBR Parameter Tuning on NIST systems", |
| "type_str": "table", |
| "num": null, |
| "content": "<table><tr><td colspan=\"4\">MBR wrt. MAP default mert-b mert+b</td></tr><tr><td># of gains</td><td>18</td><td>22</td><td>26</td></tr><tr><td># of no-changes</td><td>9</td><td>5</td><td>8</td></tr><tr><td># of drops</td><td>12</td><td>12</td><td>5</td></tr></table>", |
| "html": null |
| } |
| } |
| } |
| } |