File size: 72,158 Bytes
6fa4bc9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
{
    "paper_id": "P03-1019",
    "header": {
        "generated_with": "S2ORC 1.0.0",
        "date_generated": "2023-01-19T09:13:53.225775Z"
    },
    "title": "A Comparative Study on Reordering Constraints in Statistical Machine Translation",
    "authors": [
        {
            "first": "Richard",
            "middle": [],
            "last": "Zens",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "VI RWTH Aachen -University of Technology",
                "location": {}
            },
            "email": "zens@cs.rwth-aachen.de"
        },
        {
            "first": "Hermann",
            "middle": [],
            "last": "Ney",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "VI RWTH Aachen -University of Technology",
                "location": {}
            },
            "email": "ney@cs.rwth-aachen.de"
        }
    ],
    "year": "",
    "venue": null,
    "identifiers": {},
    "abstract": "In statistical machine translation, the generation of a translation hypothesis is computationally expensive. If arbitrary wordreorderings are permitted, the search problem is NP-hard. On the other hand, if we restrict the possible word-reorderings in an appropriate way, we obtain a polynomial-time search algorithm. In this paper, we compare two different reordering constraints, namely the ITG constraints and the IBM constraints. This comparison includes a theoretical discussion on the permitted number of reorderings for each of these constraints. We show a connection between the ITG constraints and the since 1870 known Schr\u00f6der numbers. We evaluate these constraints on two tasks: the Verbmobil task and the Canadian Hansards task. The evaluation consists of two parts: First, we check how many of the Viterbi alignments of the training corpus satisfy each of these constraints. Second, we restrict the search to each of these constraints and compare the resulting translation hypotheses. The experiments will show that the baseline ITG constraints are not sufficient on the Canadian Hansards task. Therefore, we present an extension to the ITG constraints. These extended ITG constraints increase the alignment coverage from about 87% to 96%.",
    "pdf_parse": {
        "paper_id": "P03-1019",
        "_pdf_hash": "",
        "abstract": [
            {
                "text": "In statistical machine translation, the generation of a translation hypothesis is computationally expensive. If arbitrary wordreorderings are permitted, the search problem is NP-hard. On the other hand, if we restrict the possible word-reorderings in an appropriate way, we obtain a polynomial-time search algorithm. In this paper, we compare two different reordering constraints, namely the ITG constraints and the IBM constraints. This comparison includes a theoretical discussion on the permitted number of reorderings for each of these constraints. We show a connection between the ITG constraints and the since 1870 known Schr\u00f6der numbers. We evaluate these constraints on two tasks: the Verbmobil task and the Canadian Hansards task. The evaluation consists of two parts: First, we check how many of the Viterbi alignments of the training corpus satisfy each of these constraints. Second, we restrict the search to each of these constraints and compare the resulting translation hypotheses. The experiments will show that the baseline ITG constraints are not sufficient on the Canadian Hansards task. Therefore, we present an extension to the ITG constraints. These extended ITG constraints increase the alignment coverage from about 87% to 96%.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Abstract",
                "sec_num": null
            }
        ],
        "body_text": [
            {
                "text": "In statistical machine translation, we are given a source language ('French') sentence f J 1 = f 1 . . . f j . . . f J , which is to be translated into a target language ('English') sentence e I 1 = e 1 . . . e i . . . e I . Among all possible target language sentences, we will choose the sentence with the highest probability:\u00ea The decomposition into two knowledge sources in Eq. 2 is the so-called source-channel approach to statistical machine translation (Brown et al., 1990) . It allows an independent modeling of target language model P r(e I 1 ) and translation model P r(f J 1 |e I 1 ). The target language model describes the well-formedness of the target language sentence. The translation model links the source language sentence to the target language sentence. It can be further decomposed into alignment and lexicon model. The argmax operation denotes the search problem, i.e. the generation of the output sentence in the target language. We have to maximize over all possible target language sentences.",
                "cite_spans": [
                    {
                        "start": 460,
                        "end": 480,
                        "text": "(Brown et al., 1990)",
                        "ref_id": "BIBREF1"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "In this paper, we will focus on the alignment problem, i.e. the mapping between source sentence positions and target sentence positions. As the word order in source and target language may differ, the search algorithm has to allow certain word-reorderings. If arbitrary word-reorderings are allowed, the search problem is NP-hard (Knight, 1999) . Therefore, we have to restrict the possible reorderings in some way to make the search problem feasible. Here, we will discuss two such constraints in detail. The first constraints are based on inversion transduction grammars (ITG) (Wu, 1995; Wu, 1997) . In the following, we will call these the ITG constraints. The second constraints are the IBM constraints (Berger et al., 1996) . In the next section, we will describe these constraints from a theoretical point of view. Then, we will describe the resulting search algorithm and its extension for word graph generation. Afterwards, we will analyze the Viterbi alignments produced during the training of the alignment models. Then, we will compare the translation results when restricting the search to either of these constraints.",
                "cite_spans": [
                    {
                        "start": 330,
                        "end": 344,
                        "text": "(Knight, 1999)",
                        "ref_id": "BIBREF2"
                    },
                    {
                        "start": 579,
                        "end": 589,
                        "text": "(Wu, 1995;",
                        "ref_id": "BIBREF14"
                    },
                    {
                        "start": 590,
                        "end": 599,
                        "text": "Wu, 1997)",
                        "ref_id": "BIBREF16"
                    },
                    {
                        "start": 707,
                        "end": 728,
                        "text": "(Berger et al., 1996)",
                        "ref_id": "BIBREF0"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "In this section, we will discuss the reordering constraints from a theoretical point of view. We will answer the question of how many word-reorderings are permitted for the ITG constraints as well as for the IBM constraints. Since we are only interested in the number of possible reorderings, the specific word identities are of no importance here. Furthermore, we assume a one-to-one correspondence between source and target words. Thus, we are interested in the number of word-reorderings, i.e. permutations, that satisfy the chosen constraints. First, we will consider the ITG constraints. Afterwards, we will describe the IBM constraints.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Theoretical Discussion",
                "sec_num": "2"
            },
            {
                "text": "Let us now consider the ITG constraints. Here, we interpret the input sentence as a sequence of blocks. In the beginning, each position is a block of its own. Then, the permutation process can be seen as follows: we select two consecutive blocks and merge them to a single block by choosing between two options: either keep them in monotone order or invert the order. This idea is illustrated in Fig. 1 . The white boxes represent the two blocks to be merged. Now, we investigate, how many permutations are obtainable with this method. A permutation derived by the above method can be represented as a binary tree where the inner nodes are colored either black or white. At black nodes the resulting sequences of the children are inverted. At white nodes they are kept in monotone order. This representation is equivalent to the parse trees of the simple grammar in (Wu, 1997) .",
                "cite_spans": [
                    {
                        "start": 866,
                        "end": 876,
                        "text": "(Wu, 1997)",
                        "ref_id": "BIBREF16"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 396,
                        "end": 402,
                        "text": "Fig. 1",
                        "ref_id": "FIGREF1"
                    }
                ],
                "eq_spans": [],
                "section": "ITG Constraints",
                "sec_num": "2.1"
            },
            {
                "text": "We observe that a given permutation may be constructed in several ways by the above method. For instance, let us consider the identity permutation of 1, 2, ..., n. Any binary tree with n nodes and all inner nodes colored white (monotone order) is a possible representation of this permutation. To obtain a unique representation, we pose an additional constraint on the binary trees: if the right son of a node is an inner node, it has to be colored with the opposite color. With this constraint, each of these binary trees is unique and equivalent to a parse tree of the 'canonical-form' grammar in (Wu, 1997) .",
                "cite_spans": [
                    {
                        "start": 599,
                        "end": 609,
                        "text": "(Wu, 1997)",
                        "ref_id": "BIBREF16"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ITG Constraints",
                "sec_num": "2.1"
            },
            {
                "text": "In (Shapiro and Stephens, 1991) , it is shown that the number of such binary trees with n nodes is the (n \u2212 1)th large Schr\u00f6der number S n\u22121 . The (small) Schr\u00f6der numbers have been first described in (Schr\u00f6der, 1870) as the number of bracketings of a given sequence (Schr\u00f6der's second problem). The large Schr\u00f6der numbers are just twice the Schr\u00f6der numbers. Schr\u00f6der remarked that the ratio between two consecutive Schr\u00f6der numbers approaches 3 + 2 \u221a 2 = 5.8284... . A second-order recurrence for the large Schr\u00f6der numbers is:",
                "cite_spans": [
                    {
                        "start": 3,
                        "end": 31,
                        "text": "(Shapiro and Stephens, 1991)",
                        "ref_id": "BIBREF9"
                    },
                    {
                        "start": 201,
                        "end": 217,
                        "text": "(Schr\u00f6der, 1870)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ITG Constraints",
                "sec_num": "2.1"
            },
            {
                "text": "(n + 1)S n = 3(2n \u2212 1)S n\u22121 \u2212 (n \u2212 2)S n\u22122 with n \u2265 2 and S 0 = 1, S 1 = 2.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ITG Constraints",
                "sec_num": "2.1"
            },
            {
                "text": "The Schr\u00f6der numbers have many combinatorical interpretations. Here, we will mention only two of them. The first one is another way of viewing at the ITG constraints. The number of permutations of the sequence 1, 2, ..., n, which avoid the subsequences (3, 1, 4, 2) and (2, 4, 1, 3), is the large Schr\u00f6der number S n\u22121 . More details on forbidden subsequences can be found in (West, 1995) . The interesting point is that a search with the ITG constraints cannot generate a word-reordering that contains one of these two subsequences. In (Wu, 1997) , these forbidden subsequences are called 'inside-out' transpositions.",
                "cite_spans": [
                    {
                        "start": 376,
                        "end": 388,
                        "text": "(West, 1995)",
                        "ref_id": null
                    },
                    {
                        "start": 537,
                        "end": 547,
                        "text": "(Wu, 1997)",
                        "ref_id": "BIBREF16"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ITG Constraints",
                "sec_num": "2.1"
            },
            {
                "text": "Another interpretation of the Schr\u00f6der numbers is given in (Knuth, 1973) : The number of permutations that can be sorted with an output-restricted doubleended queue (deque) is exactly the large Schr\u00f6der number. Additionally, Knuth presents an approximation for the large Schr\u00f6der numbers:",
                "cite_spans": [
                    {
                        "start": 59,
                        "end": 72,
                        "text": "(Knuth, 1973)",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ITG Constraints",
                "sec_num": "2.1"
            },
            {
                "text": "S n \u2248 c \u2022 (3 + \u221a 8) n \u2022 n \u2212 3 2 (3)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ITG Constraints",
                "sec_num": "2.1"
            },
            {
                "text": "where c is set to 1 2",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ITG Constraints",
                "sec_num": "2.1"
            },
            {
                "text": "(3 \u221a 2 \u2212 4)/\u03c0. This approximation function confirms the result of Schr\u00f6der, and we obtain S n \u2208 \u0398((3 + \u221a 8) n ), i.e. the Schr\u00f6der numbers grow like (3 + \u221a 8) n \u2248 5.83 n .",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ITG Constraints",
                "sec_num": "2.1"
            },
            {
                "text": "In this section, we will describe the IBM constraints (Berger et al., 1996) . Here, we mark each position in the source sentence either as covered or uncovered.",
                "cite_spans": [
                    {
                        "start": 54,
                        "end": 75,
                        "text": "(Berger et al., 1996)",
                        "ref_id": "BIBREF0"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "IBM Constraints",
                "sec_num": "2.2"
            },
            {
                "text": "In the beginning, all source positions are uncovered. Now, the target sentence is produced from bottom to top. A target position must be aligned to one of the first k uncovered source positions. The IBM constraints are illustrated in Fig. 2 . For most of the target positions there are k permitted source positions. Only towards the end of the sentence this is reduced to the number of remaining uncovered source positions. Let n denote the length of the input sequence and let r n denote the permitted number of permutations with the IBM constraints. Then, we obtain:",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 234,
                        "end": 240,
                        "text": "Fig. 2",
                        "ref_id": "FIGREF2"
                    }
                ],
                "eq_spans": [],
                "section": "IBM Constraints",
                "sec_num": "2.2"
            },
            {
                "text": "r n = k n\u2212k \u2022 k! n > k n! n \u2264 k (4)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "IBM Constraints",
                "sec_num": "2.2"
            },
            {
                "text": "Typically, k is set to 4. In this case, we obtain an asymptotic upper and lower bound of 4 n , i.e. r n \u2208 \u0398(4 n ).",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "IBM Constraints",
                "sec_num": "2.2"
            },
            {
                "text": "In Tab. 1, the ratio of the number of permitted reorderings for the discussed constraints is listed as a function of the sentence length. We see that for longer sentences the ITG constraints allow for more reorderings than the IBM constraints. For sentences of length 10 words, there are about twice as many reorderings for the ITG constraints than for the IBM constraints. This ratio steadily increases. For longer sentences, the ITG constraints allow for much more flexibility than the IBM constraints.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "IBM Constraints",
                "sec_num": "2.2"
            },
            {
                "text": "Now, let us get back to more practical aspects. Reordering constraints are more or less useless, if they do not allow the maximization of Eq. 2 to be performed in an efficient way. Therefore, in this section, we will describe different aspects of the search algorithm for the ITG constraints. First, we will present the dynamic programming equations and the resulting complexity. Then, we will describe pruning techniques to accelerate the search. Finally, we will extend the basic algorithm for the generation of word graphs.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Search",
                "sec_num": "3"
            },
            {
                "text": "The ITG constraints allow for a polynomial-time search algorithm. It is based on the following dynamic programming recursion equations. During the search a table Q j l ,jr,e b ,et is constructed. Here, Q j l ,jr,e b ,et denotes the probability of the best hypothesis translating the source words from position j l (left) to position j r (right) which begins with the target language word e b (bottom) and ends with the word e t (top). This is illustrated in Fig. 3 .",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 458,
                        "end": 464,
                        "text": "Fig. 3",
                        "ref_id": "FIGREF3"
                    }
                ],
                "eq_spans": [],
                "section": "Algorithm",
                "sec_num": "3.1"
            },
            {
                "text": "Here, we initialize this table with monotone translations of IBM Model 4. Therefore, Q 0 j l ,j r ,e b ,e t denotes the probability of the best monotone hypothesis of IBM Model 4. Alternatively, we could use any other single-word based lexicon as well as phrasebased models for this initialization. Our choice is the IBM Model4 to make the results as comparable 19 20 S n\u22121 /r n \u2248 1.0 1.2 1.4 1.7 2.1 2.6 3.4 4.3 5.6 7.4 9.8 13.0 17.4 23.3 31.4 as possible to the search with the IBM constraints. We introduce a new parameter p m (m= monotone), which denotes the probability of a monotone combination of two partial hypotheses. We formulated this equation for a bigram language model, but of course, the same method can also be applied for a trigram language model. The resulting algorithm is similar to the CYK-parsing algorithm. It has a worst-case complexity of O(J 3 \u2022 E 4 ). Here, J is the length of the source sentence and E is the vocabulary size of the target language.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Algorithm",
                "sec_num": "3.1"
            },
            {
                "text": "EQUATION",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [
                    {
                        "start": 0,
                        "end": 8,
                        "text": "EQUATION",
                        "ref_id": "EQREF",
                        "raw_str": "Q j l ,j r ,e b ,e t =",
                        "eq_num": "(5)"
                    }
                ],
                "section": "Algorithm",
                "sec_num": "3.1"
            },
            {
                "text": "Although the described search algorithm has a polynomial-time complexity, even with a bigram language model the search space is very large. A full search is possible but time consuming. The situation gets even worse when a trigram language model is used. Therefore, pruning techniques are obligatory to reduce the translation time.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Pruning",
                "sec_num": "3.2"
            },
            {
                "text": "Pruning is applied to hypotheses that translate the same subsequence f j r j l of the source sentence. We use pruning in the following two ways. The first pruning technique is histogram pruning: we restrict the number of translation hypotheses per sequence f jr j l . For each sequence f jr j l , we keep only a fixed number of translation hypotheses. The second pruning technique is threshold pruning: the idea is to remove all hypotheses that have a low probability relative to the best hypothesis. Therefore, we introduce a threshold pruning parameter q, with 0 \u2264 q \u2264 1. Let Q * j l ,j r denote the maximum probability of all translation hypotheses for f j r j l . Then, we prune a hypothesis iff:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Pruning",
                "sec_num": "3.2"
            },
            {
                "text": "Q j l ,j r ,e b ,e t < q \u2022 Q * j l ,j r",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Pruning",
                "sec_num": "3.2"
            },
            {
                "text": "Applying these pruning techniques the computational costs can be reduced significantly with almost no loss in translation quality.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Pruning",
                "sec_num": "3.2"
            },
            {
                "text": "The generation of word graphs for a bottom-top search with the IBM constraints is described in (Ueffing et al., 2002) . These methods cannot be applied to the CYK-style search for the ITG constraints. Here, the idea for the generation of word graphs is the following: assuming we already have word graphs for the source sequences f k j l and f j r k+1 , then we can construct a word graph for the sequence f jr j l by concatenating the partial word graphs either in monotone or inverted order. Now, we describe this idea in a more formal way. A word graph is a directed acyclic graph (dag) with one start and one end node. The edges are annotated with target language words or phrases. We also allow -transitions. These are edges annotated with the empty word. Additionally, edges may be annotated with probabilities of the language or translation model. Each path from start node to end node represents one translation hypothesis. The probability of this hypothesis is calculated by multiplying the probabilities along the path.",
                "cite_spans": [
                    {
                        "start": 95,
                        "end": 117,
                        "text": "(Ueffing et al., 2002)",
                        "ref_id": "BIBREF10"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Generation of Word Graphs",
                "sec_num": "3.3"
            },
            {
                "text": "During the search, we have to combine two word graphs in either monotone or inverted order. This is done in the following way: we are given two word graphs w 1 and w 2 with start and end nodes (s 1 , g 1 ) and (s 2 , g 2 ), respectively. First, we add an -transition (g 1 , s 2 ) from the end node of the first graph w 1 to the start node of the second graph w 2 and annotate this edge with the probability of a monotone concatenation p m . Second, we create a copy of each of the original word graphs w 1 and w 2 . Then, we add an -transition (g 2 , s 1 ) from the end node of the copied second graph to the start node of the copied first graph. This edge is annotated with the probability of a inverted concatenation 1 \u2212 p m . Now, we have obtained two word graphs: one for a monotone and one for a inverted concatenation. The final word graphs is constructed by merging the two start nodes and the two end nodes, respectively.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Generation of Word Graphs",
                "sec_num": "3.3"
            },
            {
                "text": "Let W (j l , j r ) denote the word graph for the source sequence f jr j l . This graph is constructed from the word graphs of all subsequences of (j l , j r ). Therefore, we assume, these word graphs have already been produced. For all source positions k with j l \u2264 k < j r , we combine the word graphs W (j l , k) and W (k + 1, j r ) as described above. Finally, we merge all start nodes of these graphs as well as all end nodes. Now, we have obtained the word graph W (j l , j r ) for the source sequence f j r j l . As initialization, we use the word graphs of the monotone IBM4 search.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Generation of Word Graphs",
                "sec_num": "3.3"
            },
            {
                "text": "In this section, we will extend the ITG constraints described in Sec. 2.1. This extension will go beyond basic reordering constraints.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Extended ITG constraints",
                "sec_num": "3.4"
            },
            {
                "text": "We already mentioned that the use of consecutive phrases within the ITG approach is straightforward. The only thing we have to change is the initialization of the Q-table. Now, we will extend this idea to phrases that are non-consecutive in the source language. For this purpose, we adopt the view of the ITG constraints as a bilingual grammar as, e.g., in (Wu, 1997) . For the baseline ITG constraints, the resulting grammar is:",
                "cite_spans": [
                    {
                        "start": 357,
                        "end": 367,
                        "text": "(Wu, 1997)",
                        "ref_id": "BIBREF16"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Extended ITG constraints",
                "sec_num": "3.4"
            },
            {
                "text": "A \u2192 [AA] | AA | f /e | f / | /e",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Extended ITG constraints",
                "sec_num": "3.4"
            },
            {
                "text": "Here, [AA] denotes a monotone concatenation and AA denotes an inverted concatenation.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Extended ITG constraints",
                "sec_num": "3.4"
            },
            {
                "text": "Let us now consider the case of a source phrase consisting of two parts f 1 and f 2 . Let e denote the corresponding target phrase. We add the productions A \u2192 [e/f 1 A /f 2 ] | e/f 1 A /f 2 to the grammar. The probabilities of these productions are, dependent on the translation direction, p(e|f 1 , f 2 ) or p(f 1 , f 2 |e) , respectively. Obviously, these productions are not in the normal form of an ITG, but with the method described in (Wu, 1997) , they can be normalized.",
                "cite_spans": [
                    {
                        "start": 441,
                        "end": 451,
                        "text": "(Wu, 1997)",
                        "ref_id": "BIBREF16"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 290,
                        "end": 324,
                        "text": "p(e|f 1 , f 2 ) or p(f 1 , f 2 |e)",
                        "ref_id": "FIGREF1"
                    }
                ],
                "eq_spans": [],
                "section": "Extended ITG constraints",
                "sec_num": "3.4"
            },
            {
                "text": "In the following sections we will present results on two tasks. Therefore, in this section we will show the corpus statistics for each of these tasks.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Corpus Statistics",
                "sec_num": "4"
            },
            {
                "text": "The first task we will present results on is the Verbmobil task (Wahlster, 2000) . The domain of this corpus is appointment scheduling, travel planning, and hotel reservation. It consists of transcriptions of spontaneous speech. Table 2 shows the corpus statistics of this corpus. The training corpus (Train) was used to train the IBM model parameters. The remaining free parameters, i.e. p m and the model scaling factors , were adjusted on the development corpus (Dev). The resulting system was evaluated on the test corpus (Test). ",
                "cite_spans": [
                    {
                        "start": 64,
                        "end": 80,
                        "text": "(Wahlster, 2000)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [
                    {
                        "start": 229,
                        "end": 236,
                        "text": "Table 2",
                        "ref_id": "TABREF2"
                    }
                ],
                "eq_spans": [],
                "section": "Verbmobil",
                "sec_num": "4.1"
            },
            {
                "text": "Additionally, we carried out experiments on the Canadian Hansards task. This task contains the proceedings of the Canadian parliament, which are kept by law in both French and English. About 3 million parallel sentences of this bilingual data have been made available by the Linguistic Data Consortium (LDC). Here, we use a subset of the data containing only sentences with a maximum length of 30 words. Table 3 shows the training and test corpus statistics.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 404,
                        "end": 411,
                        "text": "Table 3",
                        "ref_id": "TABREF3"
                    }
                ],
                "eq_spans": [],
                "section": "Canadian Hansards",
                "sec_num": "4.2"
            },
            {
                "text": "In this section, we will investigate for each of the constraints the coverage of the training corpus alignment. For this purpose, we compute the Viterbi alignment of IBM Model 5 with GIZA++ . This alignment is produced without any restrictions on word-reorderings. Then, we check for every sentence if the alignment satisfies each of the constraints. The ratio of the number of satisfied alignments and the total number of sentences is referred to as coverage. Tab. 4 shows the results for the Verbmobil task and for the Canadian Hansards task. It contains the results for both translation directions German-English (S\u2192T) and English-German (T\u2192S) for the Verbmobil task and French-English (S\u2192T) and English-French (T\u2192S) for the Canadian Hansards task, respectively.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Evaluation in Training",
                "sec_num": "5"
            },
            {
                "text": "For the Verbmobil task, the baseline ITG constraints and the IBM constraints result in a similar coverage. It is about 91% for the German-English translation direction and about 88% for the English-German translation direction. A significantly higher coverage of about 96% is obtained with the extended ITG constraints. Thus with the extended ITG constraints, the coverage increases by about 8% absolute. For the Canadian Hansards task, the baseline ITG constraints yield a worse coverage than the IBM constraints. Especially for the English-French translation direction, the ITG coverage of 73.6% is very low. Again, the extended ITG constraints obtained the best results. Here, the coverage increases from about 87% for the IBM constraints to about 96% for the extended ITG constraints.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Evaluation in Training",
                "sec_num": "5"
            },
            {
                "text": "In our experiments, we use the following error criteria:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Evaluation Criteria",
                "sec_num": "6.1"
            },
            {
                "text": "\u2022 WER (word error rate):",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Evaluation Criteria",
                "sec_num": "6.1"
            },
            {
                "text": "The WER is computed as the minimum number of substitution, insertion and deletion operations that have to be performed to convert the generated sentence into the target sentence.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Evaluation Criteria",
                "sec_num": "6.1"
            },
            {
                "text": "\u2022 PER (position-independent word error rate): A shortcoming of the WER is the fact that it requires a perfect word order. The PER compares the words in the two sentences ignoring the word order.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Evaluation Criteria",
                "sec_num": "6.1"
            },
            {
                "text": "\u2022 mWER (multi-reference word error rate): For each test sentence, not only a single reference translation is used, as for the WER, but a whole set of reference translations. For each translation hypothesis, the WER to the most similar sentence is calculated (Nie\u00dfen et al., 2000) .",
                "cite_spans": [
                    {
                        "start": 258,
                        "end": 279,
                        "text": "(Nie\u00dfen et al., 2000)",
                        "ref_id": "BIBREF4"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Evaluation Criteria",
                "sec_num": "6.1"
            },
            {
                "text": "\u2022 BLEU score: This score measures the precision of unigrams, bigrams, trigrams and fourgrams with respect to a whole set of reference translations with a penalty for too short sentences (Papineni et al., 2001 ). BLEU measures accuracy, i.e. large BLEU scores are better.",
                "cite_spans": [
                    {
                        "start": 186,
                        "end": 208,
                        "text": "(Papineni et al., 2001",
                        "ref_id": "BIBREF7"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Evaluation Criteria",
                "sec_num": "6.1"
            },
            {
                "text": "\u2022 SSER (subjective sentence error rate): For a more detailed analysis, subjective judgments by test persons are necessary. Each translated sentence was judged by a human examiner according to an error scale from 0.0 to 1.0 (Nie\u00dfen et al., 2000) .",
                "cite_spans": [
                    {
                        "start": 223,
                        "end": 244,
                        "text": "(Nie\u00dfen et al., 2000)",
                        "ref_id": "BIBREF4"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Evaluation Criteria",
                "sec_num": "6.1"
            },
            {
                "text": "In this section, we will present the translation results for both the IBM constraints and the baseline ITG constraints. We used a single-word based search with IBM Model 4. The initialization for the ITG constraints was done with monotone IBM Model 4 translations. So, the only difference between the two systems are the reordering constraints. In Tab. 5 the results for the Verbmobil task are shown. We see that the results on this task are similar. The search with the ITG constraints yields slightly lower error rates.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Translation Results",
                "sec_num": "6.2"
            },
            {
                "text": "Some translation examples of the Verbmobil task are shown in Tab. 6. We have to keep in mind, that the Verbmobil task consists of transcriptions of spontaneous speech. Therefore, the source sentences as well as the reference translations may have an unorthodox grammatical structure. In the first example, the German verb-group (\"w\u00fcrde vorschlagen\") is split into two parts. The search with the ITG constraints is able to produce a correct translation. With the IBM constraints, it is not possible to translate this verb-group correctly, because the distance between the two parts is too large (more than four words). As we see in the second example, in German the verb of a subordinate clause is placed at the end (\"\u00fcbernachten\"). The IBM search is not able to perform the necessary long-range reordering, as it is done with the ITG search.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Translation Results",
                "sec_num": "6.2"
            },
            {
                "text": "The ITG constraints were introduced in (Wu, 1995) . The applications were, for instance, the segmentation of Chinese character sequences into Chinese \"words\" and the bracketing of the source sentence into sub-sentential chunks. In (Wu, 1996) the baseline ITG constraints were used for statistical machine translation. The resulting algorithm is similar to the one presented in Sect. 3.1, but here, we use monotone translation hypotheses of the full IBM Model 4 as initialization, whereas in (Wu, 1996) a single-word based lexicon model is used. In (Vilar, 1998) a model similar to Wu's method was considered.",
                "cite_spans": [
                    {
                        "start": 39,
                        "end": 49,
                        "text": "(Wu, 1995)",
                        "ref_id": "BIBREF14"
                    },
                    {
                        "start": 231,
                        "end": 241,
                        "text": "(Wu, 1996)",
                        "ref_id": "BIBREF15"
                    },
                    {
                        "start": 491,
                        "end": 501,
                        "text": "(Wu, 1996)",
                        "ref_id": "BIBREF15"
                    },
                    {
                        "start": 548,
                        "end": 561,
                        "text": "(Vilar, 1998)",
                        "ref_id": "BIBREF11"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Related Work",
                "sec_num": "7"
            },
            {
                "text": "We have described the ITG constraints in detail and compared them to the IBM constraints. We draw the following conclusions: especially for long sentences the ITG constraints allow for higher flexibility in word-reordering than the IBM constraints. Regarding the Viterbi alignment in training, the baseline ITG constraints yield a similar coverage as the IBM constraints on the Verbmobil task. On the Canadian Hansards task the baseline ITG constraints were not sufficient. With the extended ITG constraints the coverage improves significantly on both tasks. On the Canadian Hansards task the coverage increases from about 87% to about 96%.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusions",
                "sec_num": "8"
            },
            {
                "text": "We have presented a polynomial-time search algorithm for statistical machine translation based on the ITG constraints and its extension for the generation of word graphs. We have shown the translation results for the Verbmobil task. On this task, the translation quality of the search with the baseline ITG constraints is already competitive with the results for the IBM constraints. Therefore, we expect the search with the extended ITG constraints to outperform the search with the IBM constraints.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusions",
                "sec_num": "8"
            },
            {
                "text": "Future work will include the automatic extraction of the bilingual grammar as well as the use of this grammar for the translation process. Table 6 : Verbmobil: translation examples. source ja, ich w\u00fcrde den Flug um viertel nach sieben vorschlagen. reference yes, I would suggest the flight at a quarter past seven.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 139,
                        "end": 146,
                        "text": "Table 6",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Conclusions",
                "sec_num": "8"
            },
            {
                "text": "ITG yes, I would suggest the flight at seven fifteen. IBM yes, I would be the flight at quarter to seven suggestion.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusions",
                "sec_num": "8"
            },
            {
                "text": "source ich schlage vor, dass wir in Hannover im Hotel Gr\u00fcnschnabel\u00fcbernachten. reference I suggest to stay at the hotel Gr\u00fcnschnabel in Hanover.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusions",
                "sec_num": "8"
            },
            {
                "text": "ITG I suggest that we stay in Hanover at hotel Gr\u00fcnschnabel. IBM I suggest that we are in Hanover at hotel Gr\u00fcnschnabel stay.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusions",
                "sec_num": "8"
            }
        ],
        "back_matter": [],
        "bib_entries": {
            "BIBREF0": {
                "ref_id": "b0",
                "title": "Language translation apparatus and method of using context-based translation models, United States patent",
                "authors": [
                    {
                        "first": "A",
                        "middle": [
                            "L"
                        ],
                        "last": "Berger",
                        "suffix": ""
                    },
                    {
                        "first": "P",
                        "middle": [
                            "F"
                        ],
                        "last": "Brown",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [
                            "A D"
                        ],
                        "last": "Pietra",
                        "suffix": ""
                    },
                    {
                        "first": "V",
                        "middle": [
                            "J D"
                        ],
                        "last": "Pietra",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [
                            "R"
                        ],
                        "last": "Gillett",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [
                            "S"
                        ],
                        "last": "Kehler",
                        "suffix": ""
                    },
                    {
                        "first": "R",
                        "middle": [
                            "L"
                        ],
                        "last": "Mercer",
                        "suffix": ""
                    }
                ],
                "year": 1996,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "A. L. Berger, P. F. Brown, S. A. D. Pietra, V. J. D. Pietra, J. R. Gillett, A. S. Kehler, and R. L. Mercer. 1996. Language translation apparatus and method of using context-based translation models, United States patent, patent number 5510981, April.",
                "links": null
            },
            "BIBREF1": {
                "ref_id": "b1",
                "title": "A statistical approach to machine translation",
                "authors": [
                    {
                        "first": "P",
                        "middle": [
                            "F"
                        ],
                        "last": "Brown",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Cocke",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [
                            "A"
                        ],
                        "last": "Della Pietra",
                        "suffix": ""
                    },
                    {
                        "first": "V",
                        "middle": [
                            "J"
                        ],
                        "last": "Della Pietra",
                        "suffix": ""
                    },
                    {
                        "first": "F",
                        "middle": [],
                        "last": "Jelinek",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [
                            "D"
                        ],
                        "last": "Lafferty",
                        "suffix": ""
                    },
                    {
                        "first": "R",
                        "middle": [
                            "L"
                        ],
                        "last": "Mercer",
                        "suffix": ""
                    },
                    {
                        "first": "P",
                        "middle": [
                            "S"
                        ],
                        "last": "Roossin",
                        "suffix": ""
                    }
                ],
                "year": 1990,
                "venue": "Computational Linguistics",
                "volume": "16",
                "issue": "2",
                "pages": "79--85",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "P. F. Brown, J. Cocke, S. A. Della Pietra, V. J. Della Pietra, F. Jelinek, J. D. Lafferty, R. L. Mercer, and P. S. Roossin. 1990. A statistical approach to machine translation. Computational Linguistics, 16(2):79-85, June.",
                "links": null
            },
            "BIBREF2": {
                "ref_id": "b2",
                "title": "Decoding complexity in wordreplacement translation models",
                "authors": [
                    {
                        "first": "K",
                        "middle": [],
                        "last": "Knight",
                        "suffix": ""
                    }
                ],
                "year": 1999,
                "venue": "Computational Linguistics",
                "volume": "25",
                "issue": "4",
                "pages": "607--615",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "K. Knight. 1999. Decoding complexity in word- replacement translation models. Computational Lin- guistics, 25(4):607-615, December.",
                "links": null
            },
            "BIBREF3": {
                "ref_id": "b3",
                "title": "The Art of Computer Programming, volume 1 -Fundamental Algorithms",
                "authors": [
                    {
                        "first": "D",
                        "middle": [
                            "E"
                        ],
                        "last": "Knuth",
                        "suffix": ""
                    }
                ],
                "year": 1973,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "D. E. Knuth. 1973. The Art of Computer Program- ming, volume 1 -Fundamental Algorithms. Addison- Wesley, Reading, MA, 2nd edition.",
                "links": null
            },
            "BIBREF4": {
                "ref_id": "b4",
                "title": "An evaluation tool for machine translation: Fast evaluation for MT research",
                "authors": [
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Nie\u00dfen",
                        "suffix": ""
                    },
                    {
                        "first": "F",
                        "middle": [
                            "J"
                        ],
                        "last": "Och",
                        "suffix": ""
                    },
                    {
                        "first": "G",
                        "middle": [],
                        "last": "Leusch",
                        "suffix": ""
                    },
                    {
                        "first": "H",
                        "middle": [],
                        "last": "Ney",
                        "suffix": ""
                    }
                ],
                "year": 2000,
                "venue": "Proc. of the Second Int. Conf. on Language Resources and Evaluation (LREC)",
                "volume": "",
                "issue": "",
                "pages": "39--45",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "S. Nie\u00dfen, F. J. Och, G. Leusch, and H. Ney. 2000. An evaluation tool for machine translation: Fast eval- uation for MT research. In Proc. of the Second Int. Conf. on Language Resources and Evaluation (LREC), pages 39-45, Athens, Greece, May.",
                "links": null
            },
            "BIBREF5": {
                "ref_id": "b5",
                "title": "Improved statistical alignment models",
                "authors": [
                    {
                        "first": "F",
                        "middle": [
                            "J"
                        ],
                        "last": "Och",
                        "suffix": ""
                    },
                    {
                        "first": "H",
                        "middle": [],
                        "last": "Ney",
                        "suffix": ""
                    }
                ],
                "year": 2000,
                "venue": "Proc. of the 38th Annual Meeting of the Association for Computational Linguistics (ACL)",
                "volume": "",
                "issue": "",
                "pages": "440--447",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "F. J. Och and H. Ney. 2000. Improved statistical align- ment models. In Proc. of the 38th Annual Meeting of the Association for Computational Linguistics (ACL), pages 440-447, Hong Kong, October.",
                "links": null
            },
            "BIBREF6": {
                "ref_id": "b6",
                "title": "Discriminative training and maximum entropy models for statistical machine translation",
                "authors": [
                    {
                        "first": "F",
                        "middle": [
                            "J"
                        ],
                        "last": "Och",
                        "suffix": ""
                    },
                    {
                        "first": "H",
                        "middle": [],
                        "last": "Ney",
                        "suffix": ""
                    }
                ],
                "year": 2002,
                "venue": "Proc. of the 40th Annual Meeting of the Association for Computational Linguistics (ACL)",
                "volume": "",
                "issue": "",
                "pages": "295--302",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "F. J. Och and H. Ney. 2002. Discriminative training and maximum entropy models for statistical machine translation. In Proc. of the 40th Annual Meeting of the Association for Computational Linguistics (ACL), pages 295-302, July.",
                "links": null
            },
            "BIBREF7": {
                "ref_id": "b7",
                "title": "Bleu: a method for automatic evaluation of machine translation",
                "authors": [
                    {
                        "first": "K",
                        "middle": [
                            "A"
                        ],
                        "last": "Papineni",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Roukos",
                        "suffix": ""
                    },
                    {
                        "first": "T",
                        "middle": [],
                        "last": "Ward",
                        "suffix": ""
                    },
                    {
                        "first": "W",
                        "middle": [
                            "J"
                        ],
                        "last": "Zhu",
                        "suffix": ""
                    }
                ],
                "year": 2001,
                "venue": "IBM Research Division",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "K. A. Papineni, S. Roukos, T. Ward, and W. J. Zhu. 2001. Bleu: a method for automatic evaluation of machine translation. Technical Report RC22176 (W0109-022), IBM Research Division, Thomas J. Watson Research Center, September.",
                "links": null
            },
            "BIBREF8": {
                "ref_id": "b8",
                "title": "Vier combinatorische Probleme. Zeitschrift f\u00fcr Mathematik und Physik",
                "authors": [
                    {
                        "first": "E",
                        "middle": [],
                        "last": "Schr\u00f6der",
                        "suffix": ""
                    }
                ],
                "year": null,
                "venue": "",
                "volume": "1870",
                "issue": "",
                "pages": "361--376",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "E. Schr\u00f6der. 1870. Vier combinatorische Probleme. Zeitschrift f\u00fcr Mathematik und Physik, 15:361-376.",
                "links": null
            },
            "BIBREF9": {
                "ref_id": "b9",
                "title": "Boostrap percolation, the Schr\u00f6der numbers, and the n-kings problem",
                "authors": [
                    {
                        "first": "L",
                        "middle": [],
                        "last": "Shapiro",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [
                            "B"
                        ],
                        "last": "Stephens",
                        "suffix": ""
                    }
                ],
                "year": 1991,
                "venue": "SIAM Journal on Discrete Mathematics",
                "volume": "4",
                "issue": "2",
                "pages": "275--280",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "L. Shapiro and A. B. Stephens. 1991. Boostrap percola- tion, the Schr\u00f6der numbers, and the n-kings problem. SIAM Journal on Discrete Mathematics, 4(2):275- 280, May.",
                "links": null
            },
            "BIBREF10": {
                "ref_id": "b10",
                "title": "Generation of word graphs in statistical machine translation",
                "authors": [
                    {
                        "first": "N",
                        "middle": [],
                        "last": "Ueffing",
                        "suffix": ""
                    },
                    {
                        "first": "F",
                        "middle": [
                            "J"
                        ],
                        "last": "Och",
                        "suffix": ""
                    },
                    {
                        "first": "H",
                        "middle": [],
                        "last": "Ney",
                        "suffix": ""
                    }
                ],
                "year": 2002,
                "venue": "Proc. Conf. on Empirical Methods for Natural Language Processing",
                "volume": "",
                "issue": "",
                "pages": "156--163",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "N. Ueffing, F. J. Och, and H. Ney. 2002. Generation of word graphs in statistical machine translation. In Proc. Conf. on Empirical Methods for Natural Lan- guage Processing, pages 156-163, Philadelphia, PA, July.",
                "links": null
            },
            "BIBREF11": {
                "ref_id": "b11",
                "title": "Aprendizaje de Transductores Subsecuenciales para su empleo en tareas de Dominio Restringido",
                "authors": [
                    {
                        "first": "J",
                        "middle": [
                            "M"
                        ],
                        "last": "Vilar",
                        "suffix": ""
                    }
                ],
                "year": 1998,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "J. M. Vilar. 1998. Aprendizaje de Transductores Subse- cuenciales para su empleo en tareas de Dominio Re- stringido. Ph.D. thesis, Universidad Politecnica de Va- lencia.",
                "links": null
            },
            "BIBREF12": {
                "ref_id": "b12",
                "title": "Verbmobil: Foundations of speech-to-speech translations",
                "authors": [],
                "year": 2000,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "W. Wahlster, editor. 2000. Verbmobil: Foundations of speech-to-speech translations. Springer Verlag, Berlin, Germany, July.",
                "links": null
            },
            "BIBREF13": {
                "ref_id": "b13",
                "title": "Generating trees and the Catalan and Schr\u00f6der numbers",
                "authors": [],
                "year": 1995,
                "venue": "Discrete Mathematics",
                "volume": "146",
                "issue": "",
                "pages": "247--262",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "J. West. 1995. Generating trees and the Catalan and Schr\u00f6der numbers. Discrete Mathematics, 146:247- 262, November.",
                "links": null
            },
            "BIBREF14": {
                "ref_id": "b14",
                "title": "Stochastic inversion transduction grammars, with application to segmentation, bracketing, and alignment of parallel corpora",
                "authors": [
                    {
                        "first": "D",
                        "middle": [],
                        "last": "Wu",
                        "suffix": ""
                    }
                ],
                "year": 1995,
                "venue": "Proc. of the 14th International Joint Conf. on Artificial Intelligence (IJ-CAI)",
                "volume": "",
                "issue": "",
                "pages": "1328--1334",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "D. Wu. 1995. Stochastic inversion transduction gram- mars, with application to segmentation, bracketing, and alignment of parallel corpora. In Proc. of the 14th International Joint Conf. on Artificial Intelligence (IJ- CAI), pages 1328-1334, Montreal, August.",
                "links": null
            },
            "BIBREF15": {
                "ref_id": "b15",
                "title": "A polynomial-time algorithm for statistical machine translation",
                "authors": [
                    {
                        "first": "D",
                        "middle": [],
                        "last": "Wu",
                        "suffix": ""
                    }
                ],
                "year": 1996,
                "venue": "Proc. of the 34th Annual Conf. of the Association for Computational Linguistics (ACL '96)",
                "volume": "",
                "issue": "",
                "pages": "152--158",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "D. Wu. 1996. A polynomial-time algorithm for statis- tical machine translation. In Proc. of the 34th Annual Conf. of the Association for Computational Linguistics (ACL '96), pages 152-158, Santa Cruz, CA, June.",
                "links": null
            },
            "BIBREF16": {
                "ref_id": "b16",
                "title": "Stochastic inversion transduction grammars and bilingual parsing of parallel corpora",
                "authors": [
                    {
                        "first": "D",
                        "middle": [],
                        "last": "Wu",
                        "suffix": ""
                    }
                ],
                "year": 1997,
                "venue": "Computational Linguistics",
                "volume": "23",
                "issue": "3",
                "pages": "377--403",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "D. Wu. 1997. Stochastic inversion transduction gram- mars and bilingual parsing of parallel corpora. Com- putational Linguistics, 23(3):377-403, September.",
                "links": null
            }
        },
        "ref_entries": {
            "FIGREF1": {
                "num": null,
                "type_str": "figure",
                "text": "Illustration of monotone and inverted concatenation of two consecutive blocks.",
                "uris": null
            },
            "FIGREF2": {
                "num": null,
                "type_str": "figure",
                "text": "Illustration of the IBM constraints.",
                "uris": null
            },
            "FIGREF3": {
                "num": null,
                "type_str": "figure",
                "text": "Illustration of the Q-table.",
                "uris": null
            },
            "TABREF0": {
                "html": null,
                "type_str": "table",
                "content": "<table><tr><td>n</td><td>1 ... 6</td><td>7</td><td>8</td><td>9 10 11 12 13 14 15 16</td><td>17</td><td>18</td></tr></table>",
                "num": null,
                "text": "Ratio of the number of permitted reorderings with the ITG constraints S n\u22121 and the IBM constraints r n for different sentence lengths n."
            },
            "TABREF2": {
                "html": null,
                "type_str": "table",
                "content": "<table><tr><td/><td/><td colspan=\"2\">German English</td></tr><tr><td colspan=\"2\">Train Sentences</td><td>58 073</td></tr><tr><td/><td>Words</td><td colspan=\"2\">519 523 549 921</td></tr><tr><td/><td>Vocabulary</td><td>7 939</td><td>4 672</td></tr><tr><td/><td>Singletons</td><td>3 453</td><td>1 698</td></tr><tr><td/><td>average SL</td><td>8.9</td><td>9.5</td></tr><tr><td>Dev</td><td>Sentences</td><td>276</td></tr><tr><td/><td>Words</td><td>3 159</td><td>3 438</td></tr><tr><td/><td>Trigram PP</td><td>-</td><td>28.1</td></tr><tr><td/><td>average SL</td><td>11.5</td><td>12.5</td></tr><tr><td>Test</td><td>Sentences</td><td>251</td></tr><tr><td/><td>Words</td><td>2 628</td><td>2 871</td></tr><tr><td/><td>Trigram PP</td><td>-</td><td>30.5</td></tr><tr><td/><td>average SL</td><td>10.5</td><td>11.4</td></tr></table>",
                "num": null,
                "text": "Statistics of training and test corpus for the Verbmobil task (PP=perplexity, SL=sentence length)."
            },
            "TABREF3": {
                "html": null,
                "type_str": "table",
                "content": "<table><tr><td/><td/><td colspan=\"2\">French English</td></tr><tr><td colspan=\"2\">Train Sentences</td><td>1.5M</td></tr><tr><td/><td>Words</td><td>24M</td><td>22M</td></tr><tr><td/><td colspan=\"3\">Vocabulary 100 269 78 332</td></tr><tr><td/><td>Singletons</td><td colspan=\"2\">40 199 31 319</td></tr><tr><td/><td>average SL</td><td>16.6</td><td>15.1</td></tr><tr><td>Test</td><td>Sentences</td><td>5432</td></tr><tr><td/><td>Words</td><td colspan=\"2\">97 646 88 773</td></tr><tr><td/><td>Trigram PP</td><td>-</td><td>179.8</td></tr><tr><td/><td>average SL</td><td>18.0</td><td>16.3</td></tr></table>",
                "num": null,
                "text": "Statistics of training and test corpus for the Canadian Hansards task (PP=perplexity, SL=sentence length)."
            },
            "TABREF4": {
                "html": null,
                "type_str": "table",
                "content": "<table><tr><td>coverage [%]</td></tr></table>",
                "num": null,
                "text": "Coverage on the training corpus for alignment constraints for the Verbmobil task (VM) and for the Canadian Hansards task (CH)."
            },
            "TABREF5": {
                "html": null,
                "type_str": "table",
                "content": "<table><tr><td>IBM</td><td>46.2</td><td>33.3</td><td>40.0</td><td>42.5</td><td>40.8</td></tr><tr><td>ITG</td><td>45.6</td><td>33.9</td><td>40.0</td><td>37.1</td><td>42.0</td></tr></table>",
                "num": null,
                "text": "Translation results on the Verbmobil task. type automatic human System WER [%] PER [%] mWER [%] BLEU [%] SSER [%]"
            }
        }
    }
}