File size: 85,772 Bytes
6fa4bc9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
{
    "paper_id": "P09-1008",
    "header": {
        "generated_with": "S2ORC 1.0.0",
        "date_generated": "2023-01-19T08:53:32.691045Z"
    },
    "title": "Topological Field Parsing of German",
    "authors": [
        {
            "first": "Jackie",
            "middle": [],
            "last": "Chi",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "University of Toronto Toronto",
                "location": {
                    "postCode": "M5S 3G4",
                    "region": "ON",
                    "country": "Canada"
                }
            },
            "email": ""
        },
        {
            "first": "Kit",
            "middle": [],
            "last": "Cheung",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "University of Toronto Toronto",
                "location": {
                    "postCode": "M5S 3G4",
                    "region": "ON",
                    "country": "Canada"
                }
            },
            "email": "jcheung@cs.toronto.edu"
        },
        {
            "first": "Gerald",
            "middle": [],
            "last": "Penn",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "University of Toronto Toronto",
                "location": {
                    "postCode": "M5S 3G4",
                    "region": "ON",
                    "country": "Canada"
                }
            },
            "email": "gpenn@cs.toronto.edu"
        }
    ],
    "year": "",
    "venue": null,
    "identifiers": {},
    "abstract": "Freer-word-order languages such as German exhibit linguistic phenomena that present unique challenges to traditional CFG parsing. Such phenomena produce discontinuous constituents, which are not naturally modelled by projective phrase structure trees. In this paper, we examine topological field parsing, a shallow form of parsing which identifies the major sections of a sentence in relation to the clausal main verb and the subordinating heads. We report the results of topological field parsing of German using the unlexicalized, latent variable-based Berkeley parser (Petrov et al., 2006) Without any language-or model-dependent adaptation, we achieve state-of-the-art results on the T\u00fcBa-D/Z corpus, and a modified NE-GRA corpus that has been automatically annotated with topological fields (Becker and Frank, 2002). We also perform a qualitative error analysis of the parser output, and discuss strategies to further improve the parsing results.",
    "pdf_parse": {
        "paper_id": "P09-1008",
        "_pdf_hash": "",
        "abstract": [
            {
                "text": "Freer-word-order languages such as German exhibit linguistic phenomena that present unique challenges to traditional CFG parsing. Such phenomena produce discontinuous constituents, which are not naturally modelled by projective phrase structure trees. In this paper, we examine topological field parsing, a shallow form of parsing which identifies the major sections of a sentence in relation to the clausal main verb and the subordinating heads. We report the results of topological field parsing of German using the unlexicalized, latent variable-based Berkeley parser (Petrov et al., 2006) Without any language-or model-dependent adaptation, we achieve state-of-the-art results on the T\u00fcBa-D/Z corpus, and a modified NE-GRA corpus that has been automatically annotated with topological fields (Becker and Frank, 2002). We also perform a qualitative error analysis of the parser output, and discuss strategies to further improve the parsing results.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Abstract",
                "sec_num": null
            }
        ],
        "body_text": [
            {
                "text": "Freer-word-order languages such as German exhibit linguistic phenomena that present unique challenges to traditional CFG parsing. Topic focus ordering and word order constraints that are sensitive to phenomena other than grammatical function produce discontinuous constituents, which are not naturally modelled by projective (i.e., without crossing branches) phrase structure trees. In this paper, we examine topological field parsing, a shallow form of parsing which identifies the major sections of a sentence in relation to the clausal main verb and subordinating heads, when present. We report the results of parsing German using the unlexicalized, latent variable-based Berkeley parser (Petrov et al., 2006) . Without any languageor model-dependent adaptation, we achieve stateof-the-art results on the T\u00fcBa-D/Z corpus (Telljohann et al., 2004) , with a F 1 -measure of 95.15% using gold POS tags. A further reranking of the parser output based on a constraint involving paired punctuation produces a slight additional performance gain. To facilitate comparison with previous work, we also conducted experiments on a modified NEGRA corpus that has been automatically annotated with topological fields (Becker and Frank, 2002) , and found that the Berkeley parser outperforms the method described in that work. Finally, we perform a qualitative error analysis of the parser output on the T\u00fcBa-D/Z corpus, and discuss strategies to further improve the parsing results.",
                "cite_spans": [
                    {
                        "start": 691,
                        "end": 712,
                        "text": "(Petrov et al., 2006)",
                        "ref_id": "BIBREF13"
                    },
                    {
                        "start": 824,
                        "end": 849,
                        "text": "(Telljohann et al., 2004)",
                        "ref_id": "BIBREF16"
                    },
                    {
                        "start": 1206,
                        "end": 1230,
                        "text": "(Becker and Frank, 2002)",
                        "ref_id": "BIBREF0"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "German syntax and parsing have been studied using a variety of grammar formalisms. Hockenmaier (2006) has translated the German TIGER corpus (Brants et al., 2002) into a CCG-based treebank to model word order variations in German. Foth et al. (2004) consider a version of dependency grammars known as weighted constraint dependency grammars for parsing German sentences. On the NEGRA corpus (Skut et al., 1998) , they achieve an accuracy of 89.0% on parsing dependency edges. In Callmeier (2000) , a platform for efficient HPSG parsing is developed. This parser is later extended by Frank et al. (2003) with a topological field parser for more efficient parsing of German. The system by Rohrer and Forst (2006) produces LFG parses using a manually designed grammar and a stochastic parse disambiguation process. They test on the TIGER corpus and achieve an F 1 -measure of 84.20%. In Dubey and Keller (2003) , PCFG parsing of NE-GRA is improved by using sister-head dependencies, which outperforms standard head lexicalization as well as an unlexicalized model. The best performing model with gold tags achieve an F 1 of 75.60%. Sister-head dependencies are useful in this case because of the flat structure of NEGRA's trees.",
                "cite_spans": [
                    {
                        "start": 83,
                        "end": 101,
                        "text": "Hockenmaier (2006)",
                        "ref_id": "BIBREF7"
                    },
                    {
                        "start": 141,
                        "end": 162,
                        "text": "(Brants et al., 2002)",
                        "ref_id": "BIBREF1"
                    },
                    {
                        "start": 231,
                        "end": 249,
                        "text": "Foth et al. (2004)",
                        "ref_id": "BIBREF4"
                    },
                    {
                        "start": 391,
                        "end": 410,
                        "text": "(Skut et al., 1998)",
                        "ref_id": "BIBREF15"
                    },
                    {
                        "start": 479,
                        "end": 495,
                        "text": "Callmeier (2000)",
                        "ref_id": "BIBREF2"
                    },
                    {
                        "start": 583,
                        "end": 602,
                        "text": "Frank et al. (2003)",
                        "ref_id": "BIBREF5"
                    },
                    {
                        "start": 687,
                        "end": 710,
                        "text": "Rohrer and Forst (2006)",
                        "ref_id": "BIBREF14"
                    },
                    {
                        "start": 884,
                        "end": 907,
                        "text": "Dubey and Keller (2003)",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "In contrast to the deeper approaches to parsing described above, topological field parsing identifies the major sections of a sentence in relation to the clausal main verb and subordinating heads, when present. Like other forms of shallow parsing, topological field parsing is useful as the first stage to further processing and eventual semantic analysis. As mentioned above, the output of a topological field parser is used as a guide to the search space of a HPSG parsing algorithm in Frank et al. (2003) . In Neumann et al. (2000) , topological field parsing is part of a divide-andconquer strategy for shallow analysis of German text with the goal of improving an information extraction system.",
                "cite_spans": [
                    {
                        "start": 488,
                        "end": 507,
                        "text": "Frank et al. (2003)",
                        "ref_id": "BIBREF5"
                    },
                    {
                        "start": 510,
                        "end": 534,
                        "text": "In Neumann et al. (2000)",
                        "ref_id": "BIBREF11"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "Existing work in identifying topological fields can be divided into chunkers, which identify the lowest-level non-recursive topological fields, and parsers, which also identify sentence and clausal structure. Veenstra et al. (2002) compare three approaches to topological field chunking based on finite state transducers, memory-based learning, and PCFGs respectively. It is found that the three techniques perform about equally well, with F 1 of 94.1% using POS tags from the TnT tagger, and 98.4% with gold tags. In Liepert (2003) , a topological field chunker is implemented using a multi-class extension to the canonically two-class support vector machine (SVM) machine learning framework. Parameters to the machine learning algorithm are fine-tuned by a genetic search algorithm, with a resulting F 1 -measure of 92.25%. Training the parameters to SVM does not have a large effect on performance, increasing the F 1 -measure in the test set by only 0.11%.",
                "cite_spans": [
                    {
                        "start": 209,
                        "end": 231,
                        "text": "Veenstra et al. (2002)",
                        "ref_id": "BIBREF19"
                    },
                    {
                        "start": 518,
                        "end": 532,
                        "text": "Liepert (2003)",
                        "ref_id": "BIBREF10"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "The corpus-based, stochastic topological field parser of Becker and Frank (2002) is based on a standard treebank PCFG model, in which rule probabilities are estimated by frequency counts. This model includes several enhancements, which are also found in the Berkeley parser. First, they use parameterized categories, splitting nonterminals according to linguistically based intuitions, such as splitting different clause types (they do not distinguish different clause types as basic categories, unlike T\u00fcBa-D/Z). Second, they take into account punctuation, which may help identify clause boundaries. They also binarize the very flat topological tree structures, and prune rules that only occur once. They test their parser on a version of the NEGRA corpus, which has been annotated with topological fields using a semiautomatic method. Ule (2003) proposes a process termed Directed Treebank Refinement (DTR). The goal of DTR is to refine a corpus to improve parsing performance. DTR is comparable to the idea of latent variable grammars on which the Berkeley parser is based, in that both consider the observed treebank to be less than ideal and both attempt to refine it by splitting and merging nonterminals. In this work, splitting and merging nonterminals are done by considering the nonterminals' contexts (i.e., their parent nodes) and the distribution of their productions. Unlike in the Berkeley parser, splitting and merging are distinct stages, rather than parts of a single iteration. Multiple splits are found first, then multiple rounds of merging are performed. No smoothing is done. As an evaluation, DTR is applied to topological field parsing of the T\u00fcBa-D/Z corpus. We discuss the performance of these topological field parsers in more detail below.",
                "cite_spans": [
                    {
                        "start": 57,
                        "end": 80,
                        "text": "Becker and Frank (2002)",
                        "ref_id": "BIBREF0"
                    },
                    {
                        "start": 837,
                        "end": 847,
                        "text": "Ule (2003)",
                        "ref_id": "BIBREF18"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "All of the topological parsing proposals predate the advent of the Berkeley parser. The experiments of this paper demonstrate that the Berkeley parser outperforms previous methods, many of which are specialized for the task of topological field chunking or parsing.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "Topological fields are high-level linear fields in an enclosing syntactic region, such as a clause (H\u00f6hle, 1983) . These fields may have constraints on the number of words or phrases they contain, and do not necessarily form a semantically coherent constituent. Although it has been argued that a few languages have no word-order constraints whatsoever, most \"free word-order\" languages (even Warlpiri) have at the very least some sort of sentence-or clause-initial topic field followed by a second position that is occupied by clitics, a finite verb or certain complementizers and subordinating conjunctions. In a few Germanic languages, including German, the topology is far richer than that, serving to identify all of the components of the verbal head of a clause, except for some cases of long-distance dependen-cies. Topological fields are useful, because while Germanic word order is relatively free with respect to grammatical functions, the order of the topological fields is strict and unvarying. (Telljohann et al., 2006) .",
                "cite_spans": [
                    {
                        "start": 99,
                        "end": 112,
                        "text": "(H\u00f6hle, 1983)",
                        "ref_id": "BIBREF8"
                    },
                    {
                        "start": 1007,
                        "end": 1032,
                        "text": "(Telljohann et al., 2006)",
                        "ref_id": "BIBREF17"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topological Field Model of German",
                "sec_num": "2"
            },
            {
                "text": "In the German topological field model, clauses belong to one of three types: verb-last (VL), verbsecond (V2), and verb-first (V1), each with a specific sequence of topological fields (Table 1) . VL clauses include finite and non-finite subordinate clauses, V2 sentences are typically declarative sentences and WH-questions in matrix clauses, and V1 sentences include yes-no questions, and certain conditional subordinate clauses. Below, we give brief descriptions of the most common topological fields.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 183,
                        "end": 192,
                        "text": "(Table 1)",
                        "ref_id": "TABREF1"
                    }
                ],
                "eq_spans": [],
                "section": "Topological Field Model of German",
                "sec_num": "2"
            },
            {
                "text": "\u2022 VF (Vorfeld or 'pre-field') is the first constituent in sentences of the V2 type. This is often the topic of the sentence, though as an anonymous reviewer pointed out, this position does not correspond to a single function with respect to information structure. (e.g., the reviewer suggested this case, where VF contains the focus: -Wer kommt zur Party? -Peter kommt zur Party. -Who is coming to the Party? -Peter is coming to the party.)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topological Field Model of German",
                "sec_num": "2"
            },
            {
                "text": "\u2022 LK (Linke Klammer or 'left bracket') is the position for finite verbs in V1 and V2 sentences. It is replaced by a complementizer with the field label C in VL sentences.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topological Field Model of German",
                "sec_num": "2"
            },
            {
                "text": "\u2022 MF (Mittelfeld or 'middle field') is an optional field bounded on the left by LK and on the right by the verbal complex VC or by NF. Most verb arguments, adverbs, and prepositional phrases are found here, unless they have been fronted and put in the VF, or are prosodically heavy and postposed to the NF field.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topological Field Model of German",
                "sec_num": "2"
            },
            {
                "text": "\u2022 VC is the verbal complex field. It includes infinite verbs, as well as finite verbs in VL sentences.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topological Field Model of German",
                "sec_num": "2"
            },
            {
                "text": "\u2022 NF (Nachfeld or 'post-field') contains prosodically heavy elements such as postposed prepositional phrases or relative clauses.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topological Field Model of German",
                "sec_num": "2"
            },
            {
                "text": "\u2022 KOORD 1 (Koordinationsfeld or 'coordination field') is a field for clause-level conjunctions.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topological Field Model of German",
                "sec_num": "2"
            },
            {
                "text": "\u2022 LV (Linksversetzung or 'left dislocation') is used for resumptive constructions involving left dislocation. For a detailed linguistic treatment, see (Frey, 2004) .",
                "cite_spans": [
                    {
                        "start": 151,
                        "end": 163,
                        "text": "(Frey, 2004)",
                        "ref_id": "BIBREF6"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topological Field Model of German",
                "sec_num": "2"
            },
            {
                "text": "Exceptions to the topological field model as described above do exist. For instance, parenthetical constructions exist as a mostly syntactically independent clause inside another sentence. In our corpus, they are attached directly underneath a clausal node without any intervening topological field, as in the following example. In this example, the parenthetical construction is highlighted in bold print. Some clause and topological field labels under the NF field are omitted for clarity.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topological Field Model of German",
                "sec_num": "2"
            },
            {
                "text": "(1) (a) (SIMPX \"(VF Man) (LK mu\u00df) (VC verstehen) \"",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topological Field Model of German",
                "sec_num": "2"
            },
            {
                "text": ", (SIMPX sagte er), \" (NF da\u00df diese Minderheiten seit langer Zeit massiv von den Nazis bedroht werden)). \" (b) Translation: \"One must understand,\" he said, \"that these minorities have been massively threatened by the Nazis for a long time.\"",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topological Field Model of German",
                "sec_num": "2"
            },
            {
                "text": "For our experiments, we used the latent variablebased Berkeley parser (Petrov et al., 2006) . Latent variable parsing assumes that an observed treebank represents a coarse approximation of an underlying, optimally refined grammar which makes more fine-grained distinctions in the syntactic categories. For example, the noun phrase category NP in a treebank could be viewed as a coarse approximation of two noun phrase categories corresponding to subjects and object, NP\u02c6S, and NP\u02c6VP. The Berkeley parser automates the process of finding such distinctions. It starts with a simple binarized X-bar grammar style backbone, and goes through iterations of splitting and merging nonterminals, in order to maximize the likelihood of the training set treebank. In the splitting stage, Figure 1 : \"I could never have done that just for aesthetic reasons.\" Sample T\u00fcBa-D/Z tree, with topological field annotations and edge labels. Topological field layer in bold.",
                "cite_spans": [
                    {
                        "start": 70,
                        "end": 91,
                        "text": "(Petrov et al., 2006)",
                        "ref_id": "BIBREF13"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 777,
                        "end": 785,
                        "text": "Figure 1",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "A Latent Variable Parser",
                "sec_num": "3"
            },
            {
                "text": "an Expectation-Maximization algorithm is used to find a good split for each nonterminal. In the merging stage, categories that have been oversplit are merged together to keep the grammar size tractable and reduce sparsity. Finally, a smoothing stage occurs, where the probabilities of rules for each nonterminal are smoothed toward the probabilities of the other nonterminals split from the same syntactic category.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "A Latent Variable Parser",
                "sec_num": "3"
            },
            {
                "text": "The Berkeley parser has been applied to the T\u00fcBaD/Z corpus in the constituent parsing shared task of the ACL-2008 Workshop on Parsing German (Petrov and Klein, 2008) , achieving an F 1measure of 85.10% and 83.18% with and without gold standard POS tags respectively 2 . We chose the Berkeley parser for topological field parsing because it is known to be robust across languages, and because it is an unlexicalized parser. Lexicalization has been shown to be useful in more general parsing applications due to lexical dependencies in constituent parsing (e.g. (K\u00fcbler et al., 2006; Dubey and Keller, 2003) in the case of German). However, topological fields explain a higher level of structure pertaining to clause-level word order, and we hypothesize that lexicalization is unlikely to be helpful.",
                "cite_spans": [
                    {
                        "start": 141,
                        "end": 165,
                        "text": "(Petrov and Klein, 2008)",
                        "ref_id": "BIBREF12"
                    },
                    {
                        "start": 560,
                        "end": 581,
                        "text": "(K\u00fcbler et al., 2006;",
                        "ref_id": "BIBREF9"
                    },
                    {
                        "start": 582,
                        "end": 605,
                        "text": "Dubey and Keller, 2003)",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "A Latent Variable Parser",
                "sec_num": "3"
            },
            {
                "text": "For our experiments, we primarily used the T\u00fcBa-D/Z (T\u00fcbinger Baumbank des Deutschen / Schriftsprache) corpus, consisting of 26116 sentences (20894 training, 2611 development, 2089 test, with a further 522 sentences held out for future ex-periments) 3 taken from the German newspaper die tageszeitung. The corpus consists of four levels of annotation: clausal, topological, phrasal (other than clausal), and lexical. We define the task of topological field parsing to be recovering the first two levels of annotation, following Ule (2003) .",
                "cite_spans": [
                    {
                        "start": 528,
                        "end": 538,
                        "text": "Ule (2003)",
                        "ref_id": "BIBREF18"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Data",
                "sec_num": "4.1"
            },
            {
                "text": "We also tested the parser on a version of the NE-GRA corpus derived by Becker and Frank (2002) , in which syntax trees have been made projective and topological fields have been automatically added through a series of linguistically informed tree modifications. All internal phrasal structure nodes have also been removed. The corpus consists of 20596 sentences, which we split into subsets of the same size as described by Becker and Frank (2002) 4 . The set of topological fields in this corpus differs slightly from the one used in T\u00fcBa-D/Z, making no distinction between clause types, nor consistently marking field or clause conjunctions. Because of the automatic annotation of topological fields, this corpus contains numerous annotation errors. Becker and Frank (2002) manually corrected their test set and evaluated the automatic annotation process, reporting labelled precision and recall of 93.0% and 93.6% compared to their manual annotations. There are also punctuation-related errors, including missing punctuation, sentences ending in commas, and sentences composed of single punctuation marks. We test on this data in order to provide a better comparison with previous work. Although we could have trained the model in Becker and Frank (2002) ",
                "cite_spans": [
                    {
                        "start": 71,
                        "end": 94,
                        "text": "Becker and Frank (2002)",
                        "ref_id": "BIBREF0"
                    },
                    {
                        "start": 424,
                        "end": 447,
                        "text": "Becker and Frank (2002)",
                        "ref_id": "BIBREF0"
                    },
                    {
                        "start": 448,
                        "end": 449,
                        "text": "4",
                        "ref_id": null
                    },
                    {
                        "start": 752,
                        "end": 775,
                        "text": "Becker and Frank (2002)",
                        "ref_id": "BIBREF0"
                    },
                    {
                        "start": 1234,
                        "end": 1257,
                        "text": "Becker and Frank (2002)",
                        "ref_id": "BIBREF0"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Data",
                "sec_num": "4.1"
            },
            {
                "text": "We first report the results of our experiments on the T\u00fcBa-D/Z corpus. For the T\u00fcBa-D/Z corpus, we trained the Berkeley parser using the default parameter settings. The grammar trainer attempts six iterations of splitting, merging, and smoothing before returning the final grammar. Intermediate grammars after each step are also saved. There were training and test sentences without clausal constituents or topological fields, which were ignored by the parser and by the evaluation. As part of our experiment design, we investigated the effect of providing gold POS tags to the parser, and the effect of incorporating edge labels into the nonterminal labels for training and parsing. In all cases, gold annotations which include gold POS tags were used when training the parser. We report the standard PARSEVAL measures of parser performance in Table 2 , obtained by the evalb program by Satoshi Sekine and Michael Collins. This table shows the results after five iterations of grammar modification, parameterized over whether we provide gold POS tags for parsing, and edge labels for training and parsing. The number of iterations was determined by experiments on the development set. In the evaluation, we do not consider edge labels in determining correctness, but do consider punctuation, as Ule (2003) did. If we ignore punctuation in our evaluation, we obtain an F 1 -measure of 95.42% on the best model (+ Gold tags, -Edge labels).",
                "cite_spans": [
                    {
                        "start": 1296,
                        "end": 1306,
                        "text": "Ule (2003)",
                        "ref_id": "BIBREF18"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 845,
                        "end": 852,
                        "text": "Table 2",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4.2"
            },
            {
                "text": "Whether supplying gold POS tags improves performance depends on whether edge labels are considered in the grammar. Without edge labels, gold POS tags improve performance by almost two points, corresponding to a relative error reduction of 33%. In contrast, performance is negatively affected when edge labels are used and gold POS tags are supplied (i.e., + Gold tags, + Edge labels), making the performance worse than not supplying gold tags. Incorporating edge label information does not appear to improve performance, possibly because it oversplits the initial treebank and interferes with the parser's ability to determine optimal splits for refining the grammar. Table 3 : BF02 = (Becker and Frank, 2002) . Parsing results for topological fields and clausal constituents. Results from Ule (2003) and our results were obtained using different training and test sets. The first row of results of Becker and Frank (2002) are from that paper; the rest were obtained by our own experiments using that parser. All results consider punctuation in evaluation.",
                "cite_spans": [
                    {
                        "start": 685,
                        "end": 709,
                        "text": "(Becker and Frank, 2002)",
                        "ref_id": "BIBREF0"
                    },
                    {
                        "start": 790,
                        "end": 800,
                        "text": "Ule (2003)",
                        "ref_id": "BIBREF18"
                    },
                    {
                        "start": 899,
                        "end": 922,
                        "text": "Becker and Frank (2002)",
                        "ref_id": "BIBREF0"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 668,
                        "end": 675,
                        "text": "Table 3",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4.2"
            },
            {
                "text": "To facilitate a more direct comparison with previous work, we also performed experiments on the modified NEGRA corpus. In this corpus, topological fields are parameterized, meaning that they are labelled with further syntactic and semantic information. For example, VF is split into VF-REL for relative clauses, and VF-TOPIC for those containing topics in a verb-second sentence, among others. All productions in the corpus have also been binarized. Tuning the parameter settings on the development set, we found that parameterized categories, binarization, and including punctuation gave the best F 1 performance. First-order horizontal and zeroth order vertical markoviza-tion after six iterations of splitting, merging, and smoothing gave the best F 1 result of 91.78%. We parsed the corpus with both the Berkeley parser and the best performing model of Becker and Frank (2002) .",
                "cite_spans": [
                    {
                        "start": 857,
                        "end": 880,
                        "text": "Becker and Frank (2002)",
                        "ref_id": "BIBREF0"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4.2"
            },
            {
                "text": "The results of these experiments on the test set for sentences of length 40 or less and for all sentences are shown in Table 3 . We also show other results from previous work for reference. We find that we achieve results that are better than the model in Becker and Frank (2002) on the test set. The difference is statistically significant (p = 0.0029, Wilcoxon signed-rank).",
                "cite_spans": [
                    {
                        "start": 256,
                        "end": 279,
                        "text": "Becker and Frank (2002)",
                        "ref_id": "BIBREF0"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 119,
                        "end": 126,
                        "text": "Table 3",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4.2"
            },
            {
                "text": "The results we obtain using the parser of Becker and Frank (2002) are worse than the results described in that paper. We suggest the following reasons for this discrepancy. While the test set used in the paper was manually corrected for evaluation, we did not correct our test set, because it would be difficult to ensure that we adhered to the same correction guidelines. No details of the correction process were provided in the paper, and descriptive grammars of German provide insufficient guidance on many of the examples in NEGRA on issues such as ellipses, short infinitival clauses, and expanded participial constructions modifying nouns. Also, because we could not obtain the exact sets used for training, development, and testing, we had to recreate the sets by randomly splitting the corpus.",
                "cite_spans": [
                    {
                        "start": 42,
                        "end": 65,
                        "text": "Becker and Frank (2002)",
                        "ref_id": "BIBREF0"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4.2"
            },
            {
                "text": "We now return to the T\u00fcBa-D/Z corpus for a more detailed analysis, and examine the categoryspecific results for our best performing model (+ Gold tags, -Edge labels). Overall, Table 4 shows that the best performing topological field categories are those that have constraints on the type of word that is allowed to fill it (finite verbs in LK, verbs in VC, complementizers and subordinating conjunctions in C). VF, in which only one constituent may appear, also performs relatively well. Topological fields that can contain a variable number of heterogeneous constituents, on the other hand, have poorer F 1 -measure results. MF, which is basically defined relative to the positions of fields on either side of it, is parsed several points below LK, C, and VC in accuracy. NF, which contains different kinds of extraposed elements, is parsed at a substantially worse level.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 176,
                        "end": 183,
                        "text": "Table 4",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Category Specific Results",
                "sec_num": "4.3"
            },
            {
                "text": "Poorly parsed categories tend to occur infre-quently, including LV, which marks a rare resumptive construction; FKOORD, which marks topological field coordination; and the discourse marker DM. The other clause-level constituents (PSIMPX for clauses in paratactic constructions, RSIMPX for relative clauses, and SIMPX for other clauses) also perform below average. Table 4 : Category-specific results using grammar with no edge labels and passing in gold POS tags.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 364,
                        "end": 371,
                        "text": "Table 4",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Category Specific Results",
                "sec_num": "4.3"
            },
            {
                "text": "While experimenting with the development set of T\u00fcBa-D/Z, we noticed that the parser sometimes returns parses, in which paired punctuation (e.g. quotation marks, parentheses, brackets) is not placed in the same clause-a linguistically implausible situation. In these cases, the high-level information provided by the paired punctuation is overridden by the overall likelihood of the parse tree. To rectify this problem, we performed a simple post-hoc reranking of the 50-best parses produced by the best parameter settings (+ Gold tags, -Edge labels), selecting the first parse that places paired punctuation in the same clause, or returning the best parse if none of the 50 parses satisfy the constraint. This procedure improved the F 1measure to 95.24% (LP = 95.39%, LR = 95.09%). Overall, 38 sentences were parsed with paired punctuation in different clauses, of which 16 were reranked. Of the 38 sentences, reranking improved performance in 12 sentences, did not affect performance in 23 sentences (of which 10 already had a perfect parse), and hurt performance in three sentences. A two-tailed sign test suggests that rerank-ing improves performance (p = 0.0352). We discuss below why sentences with paired punctuation in different clauses can have perfect parse results.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Reranking for Paired Punctuation",
                "sec_num": "4.4"
            },
            {
                "text": "To investigate the upper-bound in performance that this form of reranking is able to achieve, we calculated some statistics on our (+ Gold tags, -Edge labels) 50-best list. We found that the average rank of the best scoring parse by F 1 -measure is 2.61, and the perfect parse is present for 1649 of the 2088 sentences at an average rank of 1.90. The oracle F 1 -measure is 98.12%, indicating that a more comprehensive reranking procedure might allow further performance gains.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Reranking for Paired Punctuation",
                "sec_num": "4.4"
            },
            {
                "text": "As a further analysis, we extracted the worst scoring fifty sentences by F 1 -measure from the parsed test set (+ Gold tags, -Edge labels), and compared them against the gold standard trees, noting the cause of the error. We analyze the parses before reranking, to see how frequently the paired punctuation problem described above severely affects a parse. The major mistakes made by the parser are summarized in Table 5 Table 5 : Types and frequency of parser errors in the fifty worst scoring parses by F 1 -measure, using parameters (+ Gold tags, -Edge labels).",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 413,
                        "end": 420,
                        "text": "Table 5",
                        "ref_id": null
                    },
                    {
                        "start": 421,
                        "end": 428,
                        "text": "Table 5",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Qualitative Error Analysis",
                "sec_num": "4.5"
            },
            {
                "text": "Misidentification of Parentheticals Parenthetical constructions do not have any dependencies on the rest of the sentence, and exist as a mostly syntactically independent clause inside another sentence. They can occur at the beginning, end, or in the middle of sentences, and are often set off orthographically by punctuation. The parser has problems identifying parenthetical constructions, often positing a parenthetical construction when that constituent is actually attached to a topological field in a neighbouring clause. The following example shows one such misidentification in bracket notation. Clause internal topological fields are omitted for clarity.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Qualitative Error Analysis",
                "sec_num": "4.5"
            },
            {
                "text": "(2) (a) T\u00fcBa-D/Z: (SIMPX Weder das Ausma\u00df der Sch\u00f6nheit noch der fr\u00fchere oder sp\u00e4tere Zeitpunkt der Geburt macht einen der Zwillinge f\u00fcr eine Mutter mehr oder weniger echt / authentisch /\u00fcberlegen). (b) Parser: (SIMPX Weder das Ausma\u00df der Sch\u00f6nheit noch der fr\u00fchere oder sp\u00e4tere Zeitpunkt der Geburt macht einen der Zwillinge f\u00fcr eine Mutter mehr oder weniger echt) (PARENTHETICAL / authentisch / uberlegen.) (c) Translation: \"Neither the degree of beauty nor the earlier or later time of birth makes one of the twins any more or less real/authentic/superior to a mother.\"",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Qualitative Error Analysis",
                "sec_num": "4.5"
            },
            {
                "text": "We hypothesized earlier that lexicalization is unlikely to give us much improvement in performance, because topological fields work on a domain that is higher than that of lexical dependencies such as subcategorization frames. However, given the locally independent nature of legitimate parentheticals, a limited form of lexicalization or some other form of stronger contextual information might be needed to improve identification performance.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Qualitative Error Analysis",
                "sec_num": "4.5"
            },
            {
                "text": "Coordination Problems The second most common type of error involves field and clause coordinations. This category includes missing or incorrect FKOORD fields, and conjunctions of clauses that are misidentified. In the following example, the conjoined MFs and following NF in the correct parse tree are identified as a single long MF. however, no land and no power has such an interest in good relations with Russia (as Germany), and also no land (has) such experience in dealing with Russia as Germany.\"",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Qualitative Error Analysis",
                "sec_num": "4.5"
            },
            {
                "text": "Other Clause Errors Other clause-level errors include the parser predicting too few or too many clauses, or misidentifying the clause type. Clauses are sometimes confused with NFs, and there is one case of a relative clause being misidentified as a main clause with an intransitive verb, as the finite verb appears at the end of the clause in both cases. Some clause errors are tied to incorrect treatment of elliptical constructions, in which an element that is inferable from context is missing.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Qualitative Error Analysis",
                "sec_num": "4.5"
            },
            {
                "text": "Paired Punctuation Problems with paired punctuation are the fourth most common type of error. Punctuation is often a marker of clause or phrase boundaries. Thus, predicting paired punctuation incorrectly can lead to incorrect parses, as in the following example.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Qualitative Error Analysis",
                "sec_num": "4.5"
            },
            {
                "text": "(4) (a) \" Auch (SIMPX wenn der Krieg heute ein Mobilisierungsfaktor ist) \" , so Pau , \" (SIMPX die Leute sehen , da\u00df man f\u00fcr die Arbeit wieder auf die Stra\u00dfe gehen mu\u00df) . \" (b) Parser: (SIMPX \" (LV Auch (SIMPX wenn der Krieg heute ein Mobilisierungsfaktor ist)) \" , so Pau , \" (SIMPX die Leute sehen , da\u00df man f\u00fcr die Arbeit wieder auf die Stra\u00dfe gehen mu\u00df)) . \" (c) Translation: \"Even if the war is a factor for mobilization,\" said Pau, \"the people see, that one must go to the street for employment again.\"",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Qualitative Error Analysis",
                "sec_num": "4.5"
            },
            {
                "text": "Here, the parser predicts a spurious SIMPX clause spanning the text of the entire sentence, but this causes the second pair of quotation marks to be parsed as belonging to two different clauses. The parser also predicts an incorrect LV field. Using the paired punctuation constraint, our reranking procedure was able to correct these errors.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Qualitative Error Analysis",
                "sec_num": "4.5"
            },
            {
                "text": "Surprisingly, there are cases in which paired punctuation does not belong inside the same clause in the gold parses. These cases are either extended quotations, in which each of the quotation mark pair occurs in a different sentence altogether, or cases where the second of the quotation mark pair must be positioned outside of other sentence-final punctuation due to orthographic conventions. Sentence-final punctuation is typically placed outside a clause in this version of T\u00fcBa-D/Z.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Qualitative Error Analysis",
                "sec_num": "4.5"
            },
            {
                "text": "Other Issues Other incorrect parses generated by the parser include problems with the infrequently occurring topological fields like LV and DM, inability to determine the boundary between MF and NF in clauses without a VC field separating the two, and misidentifying appositive constructions. Another issue is that although the parser output may disagree with the gold standard tree in T\u00fcBa-D/Z, the parser output may be a well-formed topological field parse for the same sentence with a different interpretation, for example because of attachment ambiguity. Each of the authors independently checked the fifty worstscoring parses, and determined whether each parse produced by the Berkeley parser could be a wellformed topological parse. Where there was disagreement, we discussed our judgments until we came to a consensus. Of the fifty parses, we determined that nine, or 18%, could be legitimate parses. Another five, or 10%, differ from the gold standard parse only in the placement of punctuation. Thus, the F 1 -measures we presented above may be underestimating the parser's performance.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Qualitative Error Analysis",
                "sec_num": "4.5"
            },
            {
                "text": "In this paper, we examined applying the latentvariable Berkeley parser to the task of topological field parsing of German, which aims to identify the high-level surface structure of sentences. Without any language or model-dependent adaptation, we obtained results which compare favourably to previous work in topological field parsing. We further examined the results of doing a simple reranking process, constraining the output parse to put paired punctuation in the same clause. This reranking was found to result in a minor performance gain.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusion and Future Work",
                "sec_num": "5"
            },
            {
                "text": "Overall, the parser performs extremely well in identifying the traditional left and right brackets of the topological field model; that is, the fields C, LK, and VC. The parser achieves basically perfect results on these fields in the T\u00fcBa-D/Z corpus, with F 1 -measure scores for each at over 98.5%. These scores are higher than previous work in the simpler task of topological field chunking. The focus of future research should thus be on correctly identifying the infrequently occuring fields and constructions, with parenthetical constructions being a particular concern. Possible avenues of future research include doing a more comprehensive discriminative reranking of the parser output. Incorporating more contextual information might be helpful to identify discourse-related constructions such as parentheses, and the DM and LV topological fields.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusion and Future Work",
                "sec_num": "5"
            },
            {
                "text": "The T\u00fcBa-D/Z corpus distinguishes coordinating and non-coordinating particles, as well as clausal and field coordination. These distinctions need not concern us for this explanation.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "This evaluation considered grammatical functions as well as the syntactic category.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "These are the same splits into training, development, and test sets as in the ACL-08 Parsing German workshop. This corpus does not include sentences of length greater than 40.4 16476 training sentences, 1000 development, 1058 testing, and 2062 as held-out data. We were unable to obtain the exact subsets used byBecker and Frank (2002). We will discuss the ramifications of this on our evaluation procedure.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            }
        ],
        "back_matter": [
            {
                "text": "We are grateful to Markus Becker, Anette Frank, Sandra Kuebler, and Slav Petrov for their invaluable help in gathering the resources necessary for our experiments. This work is supported in part by the Natural Sciences and Engineering Research Council of Canada.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Acknowledgements",
                "sec_num": null
            }
        ],
        "bib_entries": {
            "BIBREF0": {
                "ref_id": "b0",
                "title": "A stochastic topological parser for German",
                "authors": [
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Becker",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Frank",
                        "suffix": ""
                    }
                ],
                "year": 2002,
                "venue": "Proceedings of the 19th International Conference on Computational Linguistics",
                "volume": "",
                "issue": "",
                "pages": "71--77",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "M. Becker and A. Frank. 2002. A stochastic topo- logical parser for German. In Proceedings of the 19th International Conference on Computational Linguistics, pages 71-77.",
                "links": null
            },
            "BIBREF1": {
                "ref_id": "b1",
                "title": "The TIGER Treebank",
                "authors": [
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Brants",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Dipper",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Hansen",
                        "suffix": ""
                    },
                    {
                        "first": "W",
                        "middle": [],
                        "last": "Lezius",
                        "suffix": ""
                    },
                    {
                        "first": "G",
                        "middle": [],
                        "last": "Smith",
                        "suffix": ""
                    }
                ],
                "year": 2002,
                "venue": "Proceedings of the Workshop on Treebanks and Linguistic Theories",
                "volume": "",
                "issue": "",
                "pages": "24--41",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "S. Brants, S. Dipper, S. Hansen, W. Lezius, and G. Smith. 2002. The TIGER Treebank. In Proceed- ings of the Workshop on Treebanks and Linguistic Theories, pages 24-41.",
                "links": null
            },
            "BIBREF2": {
                "ref_id": "b2",
                "title": "PET-a platform for experimentation with efficient HPSG processing techniques",
                "authors": [
                    {
                        "first": "U",
                        "middle": [],
                        "last": "Callmeier",
                        "suffix": ""
                    }
                ],
                "year": 2000,
                "venue": "Natural Language Engineering",
                "volume": "6",
                "issue": "01",
                "pages": "99--107",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "U. Callmeier. 2000. PET-a platform for experimen- tation with efficient HPSG processing techniques. Natural Language Engineering, 6(01):99-107.",
                "links": null
            },
            "BIBREF3": {
                "ref_id": "b3",
                "title": "Probabilistic parsing for German using sister-head dependencies",
                "authors": [
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Dubey",
                        "suffix": ""
                    },
                    {
                        "first": "F",
                        "middle": [],
                        "last": "Keller",
                        "suffix": ""
                    }
                ],
                "year": 2003,
                "venue": "Proceedings of the 41st Annual Meeting of the Association for Computational Linguistics",
                "volume": "",
                "issue": "",
                "pages": "96--103",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "A. Dubey and F. Keller. 2003. Probabilistic parsing for German using sister-head dependencies. In Pro- ceedings of the 41st Annual Meeting of the Associa- tion for Computational Linguistics, pages 96-103.",
                "links": null
            },
            "BIBREF4": {
                "ref_id": "b4",
                "title": "A broad-coverage parser for German based on defeasible constraints",
                "authors": [
                    {
                        "first": "K",
                        "middle": [
                            "A"
                        ],
                        "last": "Foth",
                        "suffix": ""
                    },
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Daum",
                        "suffix": ""
                    },
                    {
                        "first": "W",
                        "middle": [],
                        "last": "Menzel",
                        "suffix": ""
                    }
                ],
                "year": 2004,
                "venue": "Constraint Solving and Language Processing",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "K.A. Foth, M. Daum, and W. Menzel. 2004. A broad-coverage parser for German based on defea- sible constraints. Constraint Solving and Language Processing.",
                "links": null
            },
            "BIBREF5": {
                "ref_id": "b5",
                "title": "Integrated shallow and deep parsing: TopP meets HPSG",
                "authors": [
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Frank",
                        "suffix": ""
                    },
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Becker",
                        "suffix": ""
                    },
                    {
                        "first": "B",
                        "middle": [],
                        "last": "Crysmann",
                        "suffix": ""
                    },
                    {
                        "first": "B",
                        "middle": [],
                        "last": "Kiefer",
                        "suffix": ""
                    },
                    {
                        "first": "U",
                        "middle": [],
                        "last": "Schaefer",
                        "suffix": ""
                    }
                ],
                "year": 2003,
                "venue": "Proceedings of the 41st Annual Meeting of the Association for Computational Linguistics",
                "volume": "",
                "issue": "",
                "pages": "104--111",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "A. Frank, M. Becker, B. Crysmann, B. Kiefer, and U. Schaefer. 2003. Integrated shallow and deep parsing: TopP meets HPSG. In Proceedings of the 41st Annual Meeting of the Association for Compu- tational Linguistics, pages 104-111.",
                "links": null
            },
            "BIBREF6": {
                "ref_id": "b6",
                "title": "Notes on the syntax and the pragmatics of German Left Dislocation",
                "authors": [
                    {
                        "first": "W",
                        "middle": [],
                        "last": "Frey",
                        "suffix": ""
                    }
                ],
                "year": 2004,
                "venue": "The Syntax and Semantics of the Left Periphery",
                "volume": "",
                "issue": "",
                "pages": "203--233",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "W. Frey. 2004. Notes on the syntax and the pragmatics of German Left Dislocation. In H. Lohnstein and S. Trissler, editors, The Syntax and Semantics of the Left Periphery, pages 203-233. Mouton de Gruyter, Berlin.",
                "links": null
            },
            "BIBREF7": {
                "ref_id": "b7",
                "title": "Creating a CCGbank and a Wide-Coverage CCG Lexicon for German",
                "authors": [
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Hockenmaier",
                        "suffix": ""
                    }
                ],
                "year": 2006,
                "venue": "Proceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics",
                "volume": "",
                "issue": "",
                "pages": "505--512",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "J. Hockenmaier. 2006. Creating a CCGbank and a Wide-Coverage CCG Lexicon for German. In Pro- ceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meet- ing of the Association for Computational Linguis- tics, pages 505-512.",
                "links": null
            },
            "BIBREF8": {
                "ref_id": "b8",
                "title": "Topologische Felder",
                "authors": [
                    {
                        "first": "T",
                        "middle": [
                            "N"
                        ],
                        "last": "H\u00f6hle",
                        "suffix": ""
                    }
                ],
                "year": 1983,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "T.N. H\u00f6hle. 1983. Topologische Felder. Ph.D. thesis, K\u00f6ln.",
                "links": null
            },
            "BIBREF9": {
                "ref_id": "b9",
                "title": "Is it really that difficult to parse German?",
                "authors": [
                    {
                        "first": "S",
                        "middle": [],
                        "last": "K\u00fcbler",
                        "suffix": ""
                    },
                    {
                        "first": "E",
                        "middle": [
                            "W"
                        ],
                        "last": "Hinrichs",
                        "suffix": ""
                    },
                    {
                        "first": "W",
                        "middle": [],
                        "last": "Maier",
                        "suffix": ""
                    }
                ],
                "year": 2006,
                "venue": "Proceedings of EMNLP",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "S. K\u00fcbler, E.W. Hinrichs, and W. Maier. 2006. Is it re- ally that difficult to parse German? In Proceedings of EMNLP.",
                "links": null
            },
            "BIBREF10": {
                "ref_id": "b10",
                "title": "Topological Fields Chunking for German with SVM's: Optimizing SVM-parameters with GA's",
                "authors": [
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Liepert",
                        "suffix": ""
                    }
                ],
                "year": 2003,
                "venue": "Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP)",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "M. Liepert. 2003. Topological Fields Chunking for German with SVM's: Optimizing SVM-parameters with GA's. In Proceedings of the International Con- ference on Recent Advances in Natural Language Processing (RANLP), Bulgaria.",
                "links": null
            },
            "BIBREF11": {
                "ref_id": "b11",
                "title": "A Divide-and-Conquer Strategy for Shallow Parsing of German Free Texts",
                "authors": [
                    {
                        "first": "G",
                        "middle": [],
                        "last": "Neumann",
                        "suffix": ""
                    },
                    {
                        "first": "C",
                        "middle": [],
                        "last": "Braun",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Piskorski",
                        "suffix": ""
                    }
                ],
                "year": 2000,
                "venue": "Proceedings of the sixth conference on Applied natural language processing",
                "volume": "",
                "issue": "",
                "pages": "239--246",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "G. Neumann, C. Braun, and J. Piskorski. 2000. A Divide-and-Conquer Strategy for Shallow Parsing of German Free Texts. In Proceedings of the sixth conference on Applied natural language processing, pages 239-246. Morgan Kaufmann Publishers Inc. San Francisco, CA, USA.",
                "links": null
            },
            "BIBREF12": {
                "ref_id": "b12",
                "title": "Parsing German with Latent Variable Grammars",
                "authors": [
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Petrov",
                        "suffix": ""
                    },
                    {
                        "first": "D",
                        "middle": [],
                        "last": "Klein",
                        "suffix": ""
                    }
                ],
                "year": 2008,
                "venue": "Proceedings of the ACL-08: HLT Workshop on Parsing German",
                "volume": "",
                "issue": "",
                "pages": "33--39",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "S. Petrov and D. Klein. 2008. Parsing German with Latent Variable Grammars. In Proceedings of the ACL-08: HLT Workshop on Parsing German (PaGe- 08), pages 33-39.",
                "links": null
            },
            "BIBREF13": {
                "ref_id": "b13",
                "title": "Learning accurate, compact, and interpretable tree annotation",
                "authors": [
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Petrov",
                        "suffix": ""
                    },
                    {
                        "first": "L",
                        "middle": [],
                        "last": "Barrett",
                        "suffix": ""
                    },
                    {
                        "first": "R",
                        "middle": [],
                        "last": "Thibaux",
                        "suffix": ""
                    },
                    {
                        "first": "D",
                        "middle": [],
                        "last": "Klein",
                        "suffix": ""
                    }
                ],
                "year": 2006,
                "venue": "Proceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics",
                "volume": "",
                "issue": "",
                "pages": "433--440",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "S. Petrov, L. Barrett, R. Thibaux, and D. Klein. 2006. Learning accurate, compact, and interpretable tree annotation. In Proceedings of the 21st Interna- tional Conference on Computational Linguistics and 44th Annual Meeting of the Association for Compu- tational Linguistics, pages 433-440, Sydney, Aus- tralia, July. Association for Computational Linguis- tics.",
                "links": null
            },
            "BIBREF14": {
                "ref_id": "b14",
                "title": "Improving coverage and parsing quality of a large-scale LFG for German",
                "authors": [
                    {
                        "first": "C",
                        "middle": [],
                        "last": "Rohrer",
                        "suffix": ""
                    },
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Forst",
                        "suffix": ""
                    }
                ],
                "year": 2006,
                "venue": "Proceedings of the Language Resources and Evaluation Conference (LREC-2006)",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "C. Rohrer and M. Forst. 2006. Improving coverage and parsing quality of a large-scale LFG for Ger- man. In Proceedings of the Language Resources and Evaluation Conference (LREC-2006), Genoa, Italy.",
                "links": null
            },
            "BIBREF15": {
                "ref_id": "b15",
                "title": "A Linguistically Interpreted Corpus of German Newspaper Text",
                "authors": [
                    {
                        "first": "W",
                        "middle": [],
                        "last": "Skut",
                        "suffix": ""
                    },
                    {
                        "first": "T",
                        "middle": [],
                        "last": "Brants",
                        "suffix": ""
                    },
                    {
                        "first": "B",
                        "middle": [],
                        "last": "Krenn",
                        "suffix": ""
                    },
                    {
                        "first": "H",
                        "middle": [],
                        "last": "Uszkoreit",
                        "suffix": ""
                    }
                ],
                "year": 1998,
                "venue": "Proceedings of the ESSLLI Workshop on Recent Advances in Corpus Annotation",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "W. Skut, T. Brants, B. Krenn, and H. Uszkoreit. 1998. A Linguistically Interpreted Corpus of Ger- man Newspaper Text. Proceedings of the ESSLLI Workshop on Recent Advances in Corpus Annota- tion.",
                "links": null
            },
            "BIBREF16": {
                "ref_id": "b16",
                "title": "The T\u00fcBa-D/Z treebank: Annotating German with a context-free backbone",
                "authors": [
                    {
                        "first": "H",
                        "middle": [],
                        "last": "Telljohann",
                        "suffix": ""
                    },
                    {
                        "first": "E",
                        "middle": [],
                        "last": "Hinrichs",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Kubler",
                        "suffix": ""
                    }
                ],
                "year": 2004,
                "venue": "Proceedings of the Fourth International Conference on Language Resources and Evaluation (LREC 2004)",
                "volume": "",
                "issue": "",
                "pages": "2229--2235",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "H. Telljohann, E. Hinrichs, and S. Kubler. 2004. The T\u00fcBa-D/Z treebank: Annotating German with a context-free backbone. In Proceedings of the Fourth International Conference on Language Resources and Evaluation (LREC 2004), pages 2229-2235.",
                "links": null
            },
            "BIBREF17": {
                "ref_id": "b17",
                "title": "Stylebook for the Tubingen Treebank of Written German (T\u00fcBa-D/Z). Seminar fur Sprachwissenschaft",
                "authors": [
                    {
                        "first": "H",
                        "middle": [],
                        "last": "Telljohann",
                        "suffix": ""
                    },
                    {
                        "first": "E",
                        "middle": [
                            "W"
                        ],
                        "last": "Hinrichs",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Kubler",
                        "suffix": ""
                    },
                    {
                        "first": "H",
                        "middle": [],
                        "last": "Zinsmeister",
                        "suffix": ""
                    }
                ],
                "year": 2006,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "H. Telljohann, E.W. Hinrichs, S. Kubler, and H. Zins- meister. 2006. Stylebook for the Tubingen Tree- bank of Written German (T\u00fcBa-D/Z). Seminar fur Sprachwissenschaft, Universitat Tubingen, Tubin- gen, Germany.",
                "links": null
            },
            "BIBREF18": {
                "ref_id": "b18",
                "title": "Directed Treebank Refinement for PCFG Parsing",
                "authors": [
                    {
                        "first": "T",
                        "middle": [],
                        "last": "Ule",
                        "suffix": ""
                    }
                ],
                "year": 2003,
                "venue": "Proceedings of Workshop on Treebanks and Linguistic Theories (TLT) 2003",
                "volume": "",
                "issue": "",
                "pages": "177--188",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "T. Ule. 2003. Directed Treebank Refinement for PCFG Parsing. In Proceedings of Workshop on Treebanks and Linguistic Theories (TLT) 2003, pages 177-188.",
                "links": null
            },
            "BIBREF19": {
                "ref_id": "b19",
                "title": "Topological field chunking for German",
                "authors": [
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Veenstra",
                        "suffix": ""
                    },
                    {
                        "first": "F",
                        "middle": [
                            "H"
                        ],
                        "last": "M\u00fcller",
                        "suffix": ""
                    },
                    {
                        "first": "T",
                        "middle": [],
                        "last": "Ule",
                        "suffix": ""
                    }
                ],
                "year": 2002,
                "venue": "Proceedings of the Sixth Conference on Natural Language Learning",
                "volume": "",
                "issue": "",
                "pages": "56--62",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "J. Veenstra, F.H. M\u00fcller, and T. Ule. 2002. Topolog- ical field chunking for German. In Proceedings of the Sixth Conference on Natural Language Learn- ing, pages 56-62.",
                "links": null
            }
        },
        "ref_entries": {
            "FIGREF0": {
                "num": null,
                "text": "(3)(a) T\u00fcBa-D/Z: Auf dem europ\u00e4ischen Kontinent aber hat (FKOORD (MF kein Land und keine Macht ein derartiges Interesse an guten Beziehungen zu Ru\u00dfland) und (MF auch kein Land solche Erfahrungen im Umgang mit Ru\u00dfland)) (NF wie Deutschland). (b) Parser: Auf dem europ\u00e4ischen Kontinent aber hat (MF kein Land und keine Macht ein derartiges Interesse an guten Beziehungen zu Ru\u00dfland und auch kein Land solche Erfahrungen im Umgang mit Ru\u00dfland wie Deutschland). (c) Translation: \"On the European continent,",
                "uris": null,
                "type_str": "figure"
            },
            "TABREF1": {
                "content": "<table><tr><td>: Topological field model of German.</td></tr><tr><td>Simplified from T\u00fcBa-D/Z corpus's annotation</td></tr><tr><td>schema</td></tr></table>",
                "html": null,
                "type_str": "table",
                "text": "",
                "num": null
            },
            "TABREF2": {
                "content": "<table><tr><td>-</td><td>-</td><td>93.53 93.17 93.35 0.08 94.59</td><td>99.43</td><td>79.50</td></tr><tr><td>+</td><td>-</td><td>95.26 95.04 95.15 0.07 95.35</td><td>99.52</td><td>83.86</td></tr><tr><td>-</td><td>+</td><td>92.38 92.67 92.52 0.11 92.82</td><td>99.19</td><td>77.79</td></tr><tr><td>+</td><td>+</td><td>92.36 92.60 92.48 0.11 92.82</td><td>99.19</td><td>77.64</td></tr><tr><td colspan=\"5\">Table 2: Parsing results for topological fields and clausal constituents on the T\u00fcBa-D/Z corpus.</td></tr><tr><td colspan=\"3\">been a fair comparison, as the parser depends quite</td><td/><td/></tr><tr><td colspan=\"3\">heavily on NEGRA's annotation scheme. For ex-</td><td/><td/></tr><tr><td colspan=\"3\">ample, T\u00fcBa-D/Z does not contain an equiva-</td><td/><td/></tr><tr><td colspan=\"3\">lent of the modified NEGRA's parameterized cat-</td><td/><td/></tr><tr><td colspan=\"3\">egories; there exist edge labels in T\u00fcBaD/Z, but</td><td/><td/></tr><tr><td colspan=\"3\">they are used to mark head-dependency relation-</td><td/><td/></tr><tr><td colspan=\"3\">ships, not subtypes of syntactic categories.</td><td/><td/></tr></table>",
                "html": null,
                "type_str": "table",
                "text": "on the T\u00fcBa-D/Z corpus, it would not have Gold tags Edge labels LP% LR% F 1 % CB CB0% CB \u2264 2% EXACT%",
                "num": null
            },
            "TABREF5": {
                "content": "<table><tr><td>Problem</td><td>Freq.</td></tr><tr><td>Misidentification of Parentheticals</td><td>19</td></tr><tr><td>Coordination problems</td><td>13</td></tr><tr><td>Too few SIMPX</td><td>10</td></tr><tr><td>Paired punctuation problem</td><td>9</td></tr><tr><td>Other clause boundary errors</td><td>7</td></tr><tr><td>Other</td><td>6</td></tr><tr><td>Too many SIMPX</td><td>3</td></tr><tr><td>Clause type misidentification</td><td>2</td></tr><tr><td>MF/NF boundary</td><td>2</td></tr><tr><td>LV</td><td>2</td></tr><tr><td>VF/MF boundary</td><td>2</td></tr></table>",
                "html": null,
                "type_str": "table",
                "text": ".",
                "num": null
            }
        }
    }
}