File size: 76,389 Bytes
6fa4bc9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
{
    "paper_id": "P06-1049",
    "header": {
        "generated_with": "S2ORC 1.0.0",
        "date_generated": "2023-01-19T09:24:28.848602Z"
    },
    "title": "A Bottom-up Approach to Sentence Ordering for Multi-document Summarization",
    "authors": [
        {
            "first": "Danushka",
            "middle": [],
            "last": "Bollegala",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "The University of Tokyo",
                "location": {
                    "addrLine": "7-3-1, Bunkyo-ku",
                    "postCode": "113-8656",
                    "settlement": "Hongo, Tokyo",
                    "country": "Japan"
                }
            },
            "email": "danushka@mi.ci.i.u-tokyo.ac.jp"
        },
        {
            "first": "Naoaki",
            "middle": [],
            "last": "Okazaki",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "The University of Tokyo",
                "location": {
                    "addrLine": "7-3-1, Bunkyo-ku",
                    "postCode": "113-8656",
                    "settlement": "Hongo, Tokyo",
                    "country": "Japan"
                }
            },
            "email": "okazaki@mi.ci.i.u-tokyo.ac.jp"
        }
    ],
    "year": "",
    "venue": null,
    "identifiers": {},
    "abstract": "Ordering information is a difficult but important task for applications generating natural-language text. We present a bottom-up approach to arranging sentences extracted for multi-document summarization. To capture the association and order of two textual segments (eg, sentences), we define four criteria, chronology, topical-closeness, precedence, and succession. These criteria are integrated into a criterion by a supervised learning approach. We repeatedly concatenate two textual segments into one segment based on the criterion until we obtain the overall segment with all sentences arranged. Our experimental results show a significant improvement over existing sentence ordering strategies.",
    "pdf_parse": {
        "paper_id": "P06-1049",
        "_pdf_hash": "",
        "abstract": [
            {
                "text": "Ordering information is a difficult but important task for applications generating natural-language text. We present a bottom-up approach to arranging sentences extracted for multi-document summarization. To capture the association and order of two textual segments (eg, sentences), we define four criteria, chronology, topical-closeness, precedence, and succession. These criteria are integrated into a criterion by a supervised learning approach. We repeatedly concatenate two textual segments into one segment based on the criterion until we obtain the overall segment with all sentences arranged. Our experimental results show a significant improvement over existing sentence ordering strategies.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Abstract",
                "sec_num": null
            }
        ],
        "body_text": [
            {
                "text": "Multi-document summarization (MDS) (Radev and McKeown, 1999) tackles the information overload problem by providing a condensed version of a set of documents. Among a number of sub-tasks involved in MDS, eg, sentence extraction, topic detection, sentence ordering, information extraction, sentence generation, etc., most MDS systems have been based on an extraction method, which identifies important textual segments (eg, sentences or paragraphs) in source documents. It is important for such MDS systems to determine a coherent arrangement of the textual segments extracted from multi-documents in order to reconstruct the text structure for summarization. Ordering information is also essential for other text-generation applications such as Question Answering.",
                "cite_spans": [
                    {
                        "start": 35,
                        "end": 60,
                        "text": "(Radev and McKeown, 1999)",
                        "ref_id": "BIBREF9"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "A summary with improperly ordered sentences confuses the reader and degrades the quality/reliability of the summary itself. Barzilay (2002) has provided empirical evidence that proper order of extracted sentences improves their readability significantly. However, ordering a set of sentences into a coherent text is a nontrivial task. For example, identifying rhetorical relations (Mann and Thompson, 1988) in an ordered text has been a difficult task for computers, whereas our task is even more complicated: to reconstruct such relations from unordered sets of sentences. Source documents for a summary may have been written by different authors, by different writing styles, on different dates, and based on different background knowledge. We cannot expect that a set of extracted sentences from such diverse documents will be coherent on their own.",
                "cite_spans": [
                    {
                        "start": 124,
                        "end": 139,
                        "text": "Barzilay (2002)",
                        "ref_id": "BIBREF1"
                    },
                    {
                        "start": 381,
                        "end": 406,
                        "text": "(Mann and Thompson, 1988)",
                        "ref_id": "BIBREF4"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "Several strategies to determine sentence ordering have been proposed as described in section 2. However, the appropriate way to combine these strategies to achieve more coherent summaries remains unsolved. In this paper, we propose four criteria to capture the association of sentences in the context of multi-document summarization for newspaper articles. These criteria are integrated into one criterion by a supervised learning approach. We also propose a bottom-up approach in arranging sentences, which repeatedly concatenates textual segments until the overall segment with all sentences arranged, is achieved.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "Existing methods for sentence ordering are divided into two approaches: making use of chronological information Lin and Hovy, 2001 ; Barzilay et al., 2002; Okazaki et al., 2004) ; and learning the natural order of sentences from large corpora not necessarily based on chronological information (Lapata, 2003; Barzilay and Lee, 2004) . A newspaper usually disseminates descriptions of novel events that have occurred since the last publication. For this reason, ordering sentences according to their publication date is an effective heuristic for multidocument summarization (Lin and Hovy, 2001; . Barzilay et al. (2002) have proposed an improved version of chronological ordering by first grouping sentences into sub-topics discussed in the source documents and then arranging the sentences in each group chronologically. Okazaki et al. (2004) have proposed an algorithm to improve chronological ordering by resolving the presuppositional information of extracted sentences. They assume that each sentence in newspaper articles is written on the basis that presuppositional information should be transferred to the reader before the sentence is interpreted. The proposed algorithm first arranges sentences in a chronological order and then estimates the presuppositional information for each sentence by using the content of the sentences placed before each sentence in its original article. The evaluation results show that the proposed algorithm improves the chronological ordering significantly. Lapata (2003) has suggested a probabilistic model of text structuring and its application to the sentence ordering. Her method calculates the transition probability from one sentence to the next from a corpus based on the Cartesian product between two sentences defined using the following features: verbs (precedent relationships of verbs in the corpus); nouns (entity-based coherence by keeping track of the nouns); and dependencies (structure of sentences). Although she has not compared her method with chronological ordering, it could be applied to generic domains, not relying on the chronological clue provided by newspaper articles. Barzilay and Lee (2004) have proposed content models to deal with topic transition in domain specific text. The content models are formalized by Hidden Markov Models (HMMs) in which the hidden state corresponds to a topic in the domain of interest (eg, earthquake magnitude or previous earthquake occurrences), and the state transitions capture possible information-presentation orderings. The evaluation results showed that their method outperformed Lapata's approach by a wide margin. They did not compare their method with chronological ordering as an application of multi-document summarization.",
                "cite_spans": [
                    {
                        "start": 112,
                        "end": 130,
                        "text": "Lin and Hovy, 2001",
                        "ref_id": "BIBREF3"
                    },
                    {
                        "start": 133,
                        "end": 155,
                        "text": "Barzilay et al., 2002;",
                        "ref_id": "BIBREF1"
                    },
                    {
                        "start": 156,
                        "end": 177,
                        "text": "Okazaki et al., 2004)",
                        "ref_id": "BIBREF7"
                    },
                    {
                        "start": 294,
                        "end": 308,
                        "text": "(Lapata, 2003;",
                        "ref_id": "BIBREF2"
                    },
                    {
                        "start": 309,
                        "end": 332,
                        "text": "Barzilay and Lee, 2004)",
                        "ref_id": "BIBREF0"
                    },
                    {
                        "start": 574,
                        "end": 594,
                        "text": "(Lin and Hovy, 2001;",
                        "ref_id": "BIBREF3"
                    },
                    {
                        "start": 597,
                        "end": 619,
                        "text": "Barzilay et al. (2002)",
                        "ref_id": "BIBREF1"
                    },
                    {
                        "start": 822,
                        "end": 843,
                        "text": "Okazaki et al. (2004)",
                        "ref_id": "BIBREF7"
                    },
                    {
                        "start": 1499,
                        "end": 1512,
                        "text": "Lapata (2003)",
                        "ref_id": "BIBREF2"
                    },
                    {
                        "start": 2140,
                        "end": 2163,
                        "text": "Barzilay and Lee (2004)",
                        "ref_id": "BIBREF0"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Related Work",
                "sec_num": "2"
            },
            {
                "text": "As described above, several good strategies/heuristics to deal with the sentence ordering problem have been proposed. In order to integrate multiple strategies/heuristics, we have formalized them in a machine learning framework and have considered an algorithm to arrange sentences using the integrated strategy.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Related Work",
                "sec_num": "2"
            },
            {
                "text": "We define notation a b to represent that sentence a precedes sentence b. We use the term segment to describe a sequence of ordered sentences. When segment A consists of sentences a 1 , a 2 , ..., a m in this order, we denote as:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3"
            },
            {
                "text": "A = (a 1 a 2 ... a m ).",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3"
            },
            {
                "text": "(1)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3"
            },
            {
                "text": "The two segments A and B can be ordered either B after A or A after B. We define the notation A B to show that segment A precedes segment B.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3"
            },
            {
                "text": "Let us consider a bottom-up approach in arranging sentences. Starting with a set of segments initialized with a sentence for each, we concatenate two segments, with the strongest association (discussed later) of all possible segment pairs, into one segment. Repeating the concatenating will eventually yield a segment with all sentences arranged. The algorithm is considered as a variation of agglomerative hierarchical clustering with the ordering information retained at each concatenating process.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3"
            },
            {
                "text": "The underlying idea of the algorithm, a bottomup approach to text planning, was proposed by Marcu (1997) . Assuming that the semantic units (sentences) and their rhetorical relations (eg, sentence a is an elaboration of sentence d) are given, he transcribed a text structuring task into the problem of finding the best discourse tree that satisfied the set of rhetorical relations. He stated that global coherence could be achieved by satisfying local coherence constraints in ordering and clustering, thereby ensuring that the resultant discourse tree was well-formed.",
                "cite_spans": [
                    {
                        "start": 92,
                        "end": 104,
                        "text": "Marcu (1997)",
                        "ref_id": "BIBREF5"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3"
            },
            {
                "text": "Unfortunately, identifying the rhetorical relation between two sentences has been a difficult task for computers. However, the bottom-up algorithm for arranging sentences can still be applied only if the direction and strength of the association of the two segments (sentences) are defined. Hence, we introduce a function f (A B) to represent the direction and strength of the association of two segments A and B,",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3"
            },
            {
                "text": "EQUATION",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [
                    {
                        "start": 0,
                        "end": 8,
                        "text": "EQUATION",
                        "ref_id": "EQREF",
                        "raw_str": "f (A B) = p (if A precedes B) 0 (if B precedes A) ,",
                        "eq_num": "(2)"
                    }
                ],
                "section": "Method",
                "sec_num": "3"
            },
            {
                "text": "where ",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3"
            },
            {
                "text": "p (0 \u2264 p \u2264 1) denotes",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3"
            },
            {
                "text": "Then we search for the segment pair with the strongest association. Supposing that f (C D) has the highest value, we concatenate C and D to obtain a new segment,",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3"
            },
            {
                "text": "EQUATION",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [
                    {
                        "start": 0,
                        "end": 8,
                        "text": "EQUATION",
                        "ref_id": "EQREF",
                        "raw_str": "F = (c d).",
                        "eq_num": "(6)"
                    }
                ],
                "section": "Method",
                "sec_num": "3"
            },
            {
                "text": "Finally, comparing f (E F ) and f (F E), we obtain the global sentence ordering,",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3"
            },
            {
                "text": "EQUATION",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [
                    {
                        "start": 0,
                        "end": 8,
                        "text": "EQUATION",
                        "ref_id": "EQREF",
                        "raw_str": "G = (b a c d).",
                        "eq_num": "(7)"
                    }
                ],
                "section": "Method",
                "sec_num": "3"
            },
            {
                "text": "In the above description, we have not defined the association of the two segments. The previous work described in Section 2 has addressed the association of textual segments (sentences) to obtain coherent orderings. We define four criteria to capture the association of two segments: chronology; topical-closeness; precedence; and succession. These criteria are integrated into a function f (A B) by using a machine learning approach. The rest of this section explains the four criteria and an integration method with a Support Vector Machine (SVM) (Vapnik, 1998) classifier.",
                "cite_spans": [
                    {
                        "start": 549,
                        "end": 563,
                        "text": "(Vapnik, 1998)",
                        "ref_id": "BIBREF10"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3"
            },
            {
                "text": "Chronology criterion reflects the chronological ordering (Lin and Hovy, 2001; , which arranges sentences in a chronological order of the publication date. We define the association strength of arranging segments B after A measured by a chronology criterion f chro (A B) in the following formula,",
                "cite_spans": [
                    {
                        "start": 57,
                        "end": 77,
                        "text": "(Lin and Hovy, 2001;",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Chronology criterion",
                "sec_num": "3.1"
            },
            {
                "text": "EQUATION",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [
                    {
                        "start": 0,
                        "end": 8,
                        "text": "EQUATION",
                        "ref_id": "EQREF",
                        "raw_str": "f chro (A B) = \uf8f1 \uf8f4 \uf8f4 \uf8f2 \uf8f4 \uf8f4 \uf8f3 1 T(a m ) < T(b 1 ) 1 [D(a m ) = D(b 1 )] \u2227 [N(a m ) < N(b 1 )] 0.5 [T(a m ) = T(b 1 )] \u2227 [D(a m ) = D(b 1 )] 0 otherwise .",
                        "eq_num": "(8)"
                    }
                ],
                "section": "Chronology criterion",
                "sec_num": "3.1"
            },
            {
                "text": "Here, a m represents the last sentence in segment A; b 1 represents the first sentence in segment B; T (s) is the publication date of the sentence s; D(s) is the unique identifier of the document to which sentence s belongs: and N (s) denotes the line number of sentence s in the original document. The chronological order of arranging segment B after A is determined by the comparison between the last sentence in the segment A and the first sentence in the segment B. The chronology criterion assesses the appropriateness of arranging segment B after A if: sentence a m is published earlier than b 1 ; or sentence a m appears before b 1 in the same article. If sentence a m and b 1 are published on the same day but appear in different articles, the criterion assumes the order to be undefined. If none of the above conditions are satisfied, the criterion estimates that segment B will precede A.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Chronology criterion",
                "sec_num": "3.1"
            },
            {
                "text": "The topical-closeness criterion deals with the association, based on the topical similarity, of two ... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... a 3 a 4 .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... ... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... ... .",
                "cite_spans": [
                    {
                        "start": 100,
                        "end": 514,
                        "text": "... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... a 3 a 4 .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... ....",
                        "ref_id": null
                    },
                    {
                        "start": 515,
                        "end": 716,
                        "text": "... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... ...",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topical-closeness criterion",
                "sec_num": "3.2"
            },
            {
                "text": "a 1 a 2 .",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topical-closeness criterion",
                "sec_num": "3.2"
            },
            {
                "text": "Segment A ? Segment B",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topical-closeness criterion",
                "sec_num": "3.2"
            },
            {
                "text": "Original article for sentence b",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topical-closeness criterion",
                "sec_num": "3.2"
            },
            {
                "text": "Original article for sentence b2",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topical-closeness criterion",
                "sec_num": "3.2"
            },
            {
                "text": "Original article for sentence b3",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topical-closeness criterion",
                "sec_num": "3.2"
            },
            {
                "text": ". ... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... Original article . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... 1 Original article max average max max Figure 2 : Precedence criterion segments. The criterion reflects the ordering strategy proposed by Barzilay et al (2002) , which groups sentences referring to the same topic. To measure the topical closeness of two sentences, we represent each sentence with a vector whose elements correspond to the occurrence 1 of the nouns and verbs in the sentence. We define the topical closeness of two segments A and B as follows,",
                "cite_spans": [
                    {
                        "start": 2,
                        "end": 204,
                        "text": "... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... ....",
                        "ref_id": null
                    },
                    {
                        "start": 222,
                        "end": 563,
                        "text": ". ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... 1",
                        "ref_id": null
                    },
                    {
                        "start": 700,
                        "end": 721,
                        "text": "Barzilay et al (2002)",
                        "ref_id": "BIBREF1"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 601,
                        "end": 609,
                        "text": "Figure 2",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Topical-closeness criterion",
                "sec_num": "3.2"
            },
            {
                "text": "f topic (A B) = 1 |B| b\u2208B max a\u2208A sim(a, b). (9)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "1",
                "sec_num": null
            },
            {
                "text": "Here, sim(a, b) denotes the similarity of sentences a and b, which is calculated by the cosine similarity of two vectors corresponding to the sentences. For sentence b \u2208 B, max a\u2208A sim(a, b) chooses the sentence a \u2208 A most similar to sentence b and yields the similarity. The topical-closeness criterion f topic (A B) assigns a higher value when the topic referred by segment B is the same as segment A.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "1",
                "sec_num": null
            },
            {
                "text": "Let us think of the case where we arrange segment A before B. Each sentence in segment B has the presuppositional information that should be conveyed to a reader in advance. Given sentence b \u2208 B, such presuppositional information may be presented by the sentences appearing before the sentence b in the original article. However, we cannot guarantee whether a sentenceextraction method for multi-document summarization chooses any sentences before b for a summary because the extraction method usually deter- 1 The vector values are represented by boolean values, i.e., 1 if the sentence contains a word, otherwise 0. ... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... a 3 .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .. a3 . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... ... .",
                "cite_spans": [
                    {
                        "start": 509,
                        "end": 510,
                        "text": "1",
                        "ref_id": null
                    },
                    {
                        "start": 618,
                        "end": 822,
                        "text": "... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... a 3 .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... ..",
                        "ref_id": null
                    },
                    {
                        "start": 823,
                        "end": 994,
                        "text": "a3 . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... ...",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Precedence criterion",
                "sec_num": "3.3"
            },
            {
                "text": ".. b b 2 b 3 a 3 a 2 a 1 S a1 S a2 S",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Precedence criterion",
                "sec_num": "3.3"
            },
            {
                "text": "Segment A ? Segment B",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Precedence criterion",
                "sec_num": "3.3"
            },
            {
                "text": "Original article for sentence a1",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Precedence criterion",
                "sec_num": "3.3"
            },
            {
                "text": "Original article for sentence a2",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Precedence criterion",
                "sec_num": "3.3"
            },
            {
                "text": "Original article for sentence a3",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Precedence criterion",
                "sec_num": "3.3"
            },
            {
                "text": "Original article for sentence for sentence max average max max .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... .... ... .. .. .... .. ....... ...... .... .. .. .... .. ....... ...... .... .. .. .... .. ....... ...... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... ... ",
                "cite_spans": [
                    {
                        "start": 63,
                        "end": 162,
                        "text": ".... .. .. .... .. ....... ...... . ... ...... .. .., .... ... .... .... .. .. .... .. ....... ....",
                        "ref_id": null
                    },
                    {
                        "start": 163,
                        "end": 330,
                        "text": "... .. .. .... .. ....... ...... .... .. .. .... .. ....... ...... .... .. .. .... .. ....... ...... .... .. .. .... .. ....... ...... . ... ...... .. .., .... ... ...",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Precedence criterion",
                "sec_num": "3.3"
            },
            {
                "text": "Original article Figure 3 : Succession criterion mines a set of sentences, within the constraint of summary length, that maximizes information coverage and excludes redundant information. Precedence criterion measures the substitutability of the presuppositional information of segment B (eg, the sentences appearing before sentence b) as segment A. This criterion is a formalization of the sentence-ordering algorithm proposed by Okazaki et al, (2004) .",
                "cite_spans": [
                    {
                        "start": 431,
                        "end": 452,
                        "text": "Okazaki et al, (2004)",
                        "ref_id": "BIBREF7"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 17,
                        "end": 25,
                        "text": "Figure 3",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": ". 1 b for sentence",
                "sec_num": null
            },
            {
                "text": "We define the precedence criterion in the following formula,",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": ". 1 b for sentence",
                "sec_num": null
            },
            {
                "text": "f pre (A B) = 1 |B| b\u2208B max a\u2208A,p\u2208P b sim(a, p).",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": ". 1 b for sentence",
                "sec_num": null
            },
            {
                "text": "(10) Here, P b is a set of sentences appearing before sentence b in the original article; and sim(a, b) denotes the cosine similarity of sentences a and b (defined as in the topical-closeness criterion). Figure 2 shows an example of calculating the precedence criterion for arranging segment B after A. We approximate the presuppositional information for sentence b by sentences P b , ie, sentences appearing before the sentence b in the original article. Calculating the similarity among sentences in P b and A by the maximum similarity of the possible sentence combinations, Formula 10 is interpreted as the average similarity of the precedent sentences \u2200P b (b \u2208 B) to the segment A.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 204,
                        "end": 212,
                        "text": "Figure 2",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": ". 1 b for sentence",
                "sec_num": null
            },
            {
                "text": "The idea of succession criterion is the exact opposite of the precedence criterion. The succession criterion assesses the coverage of the succedent information for segment A by arranging segment B (11) Here, S a is a set of sentences appearing after sentence a in the original article; and sim(a, b) denotes the cosine similarity of sentences a and b (defined as in the topical-closeness criterion). Figure 3 shows an example of calculating the succession criterion to arrange segments B after A. The succession criterion measures the substitutability of the succedent information (eg, the sentences appearing after the sentence a \u2208 A) as segment B.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 400,
                        "end": 408,
                        "text": "Figure 3",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Succession criterion",
                "sec_num": "3.4"
            },
            {
                "text": "We integrate the four criteria described above to define the function f (A B) to represent the association direction and strength of the two segments A and B (Formula 2). More specifically, given the two segments A and B, function f (A B) is defined to yield the integrated association strength from four values, f chro (A B), f topic (A B), f pre (A B), and f succ (A B). We formalize the integration task as a binary classification problem and employ a Support Vector Machine (SVM) as the classifier. We conducted a supervised learning as follows.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "SVM classifier to assess the integrated criterion",
                "sec_num": "3.5"
            },
            {
                "text": "We partition a human-ordered extract into pairs each of which consists of two non-overlapping segments. Let us explain the partitioning process taking four human-ordered sentences, a b c d shown in Figure 4 . Firstly, we place the partitioning point just after the first sentence a. , we identify ten pairs in total form the four ordered sentences. In general, this process yields n(n 2 \u22121)/6 pairs from ordered n sentences. From each pair of segments, we generate one positive and one negative training instance as follows.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 198,
                        "end": 206,
                        "text": "Figure 4",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "SVM classifier to assess the integrated criterion",
                "sec_num": "3.5"
            },
            {
                "text": "+1 : [f chro (A B), f topic (A B), f pre (A B), f succ (A B)] \u22121 : [f chro (B A), f topic (B A), f pre (B A), f succ (B A)]",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "SVM classifier to assess the integrated criterion",
                "sec_num": "3.5"
            },
            {
                "text": "Given a pair of two segments A and B arranged in an order A B, we calculate four values, f chro (A B), f topic (A B), f pre (A B), and f succ (A B) to obtain the instance with the four-dimensional vector ( Figure 5) . We label the instance (corresponding to A B) as a positive class (ie, +1). Simultaneously, we obtain another instance with a four-dimensional vector corresponding to B A. We label it as a negative class (ie, \u22121). Accumulating these instances as training data, we obtain a binary classifier by using a Support Vector Machine with a quadratic kernel. The SVM classifier yields the association direction of two segments (eg, A B or B A) with the class information (ie, +1 or \u22121). We assign the association strength of two segments by using the class probability estimate that the instance belongs to a positive (+1) class. When an instance is classified into a negative (\u22121) class, we set the association strength as zero (see the definition of Formula 2).",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 206,
                        "end": 215,
                        "text": "Figure 5)",
                        "ref_id": "FIGREF6"
                    }
                ],
                "eq_spans": [],
                "section": "SVM classifier to assess the integrated criterion",
                "sec_num": "3.5"
            },
            {
                "text": "We evaluated the proposed method by using the 3rd Text Summarization Challenge (TSC-3) corpus 2 . The TSC-3 corpus contains 30 sets of extracts, each of which consists of unordered sentences 3 extracted from Japanese newspaper articles relevant to a topic (query). We arrange the extracts by using different algorithms and evaluate the readability of the ordered extracts by a subjective grading and several metrics. In order to construct training data applicable to the proposed method, we asked two human subjects to arrange the extracts and obtained 30(topics) \u00d7 2(humans) = 60 sets of ordered extracts. Table 1 shows the agreement of the ordered extracts between the two subjects. The correlation is measured by three metrics, Spearman's rank correlation, Kendall's rank correlation, and average continuity (described later). The mean correlation values (0.74 for Spearman's rank correlation and 0.69 for Kendall's rank correlation) indicate a certain level of agreement in sentence orderings made by the two subjects. 8 out of 30 extracts were actually identical.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 607,
                        "end": 614,
                        "text": "Table 1",
                        "ref_id": "TABREF1"
                    }
                ],
                "eq_spans": [],
                "section": "Evaluation",
                "sec_num": "4"
            },
            {
                "text": "We applied the leave-one-out method to the proposed method to produce a set of sentence orderings. In this experiment, the leave-out-out method arranges an extract by using an SVM model trained from the rest of the 29 extracts. Repeating this process 30 times with a different topic for each iteration, we generated a set of 30 extracts for evaluation. In addition to the proposed method, we prepared six sets of sentence orderings produced by different algorithms for comparison. We describe briefly the seven algorithms (including the proposed method): Suceedence ordering (SUC) arranges sentences with the succession criterion defined in Formula 11.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Evaluation",
                "sec_num": "4"
            },
            {
                "text": "The last four algorithms (CHR, TOP, PRE, and SUC) arrange sentences by the corresponding criterion alone, each of which uses the association strength directly to arrange sentences without the integration of other criteria. These orderings are expected to show the performance of each expert independently and their contribution to solving the sentence ordering problem.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Evaluation",
                "sec_num": "4"
            },
            {
                "text": "Evaluating a sentence ordering is a challenging task. Intrinsic evaluation that involves human judges to rank a set of sentence orderings is a necessary approach to this task (Barzilay et al., 2002; Okazaki et al., 2004) . We asked two human judges to rate sentence orderings according to the following criteria. A perfect summary is a text that we cannot improve any further by re-ordering. An acceptable summary is one that makes sense and is unnecessary to revise even though there is some room for improvement in terms of readability. A poor summary is one that loses a thread of the story at some places and requires minor amendment to bring it up to an acceptable level. An unacceptable summary is one that leaves much to be improved and requires overall restructuring rather than partial revision. To avoid any disturbance in rating, we inform the judges that the summaries were made from a same set of extracted sentences and only the ordering of sentences is different. Figure 6 shows the distribution of the subjective grading made by two judges to four sets of orderings, RND, CHR, AGL and HUM. Each set of or- derings has 30(topics) \u00d7 2(judges) = 60 ratings. Most RND orderings are rated as unacceptable. Although CHR and AGL orderings have roughly the same number of perfect orderings (ca. 25%), the AGL algorithm gained more acceptable orderings (47%) than the CHR alghrotihm (30%). This fact shows that integration of CHR experts with other experts worked well by pushing poor ordering to an acceptable level. However, a huge gap between AGL and HUM orderings was also found. The judges rated 28% AGL orderings as perfect while the figure rose as high as 82% for HUM orderings. Kendall's coefficient of concordance (Kendall's W ), which asses the inter-judge agreement of overall ratings, reported a higher agreement between the two judges (W = 0.939).",
                "cite_spans": [
                    {
                        "start": 175,
                        "end": 198,
                        "text": "(Barzilay et al., 2002;",
                        "ref_id": "BIBREF1"
                    },
                    {
                        "start": 199,
                        "end": 220,
                        "text": "Okazaki et al., 2004)",
                        "ref_id": "BIBREF7"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 979,
                        "end": 987,
                        "text": "Figure 6",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Subjective grading",
                "sec_num": "4.1"
            },
            {
                "text": "We also evaluated sentence orderings by reusing two sets of gold-standard orderings made for the training data. In general, subjective grading consumes much time and effort, even though we cannot reproduce the evaluation afterwards. The previous studies (Barzilay et al., 2002; Lapata, 2003) employ rank correlation coefficients such as Spearman's rank correlation and Kendall's rank correlation, assuming a sentence ordering to be a rank. Okazaki et al. (2004) propose a metric that assess continuity of pairwise sentences compared with the gold standard. In addition to Spearman's and Kendall's rank correlation coefficients, we propose an average continuity metric, which extends the idea of the continuity metric to continuous k sentences.",
                "cite_spans": [
                    {
                        "start": 254,
                        "end": 277,
                        "text": "(Barzilay et al., 2002;",
                        "ref_id": "BIBREF1"
                    },
                    {
                        "start": 278,
                        "end": 291,
                        "text": "Lapata, 2003)",
                        "ref_id": "BIBREF2"
                    },
                    {
                        "start": 440,
                        "end": 461,
                        "text": "Okazaki et al. (2004)",
                        "ref_id": "BIBREF7"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Metrics for semi-automatic evaluation",
                "sec_num": "4.2"
            },
            {
                "text": "A text with sentences arranged in proper order does not interrupt a human's reading while moving from one sentence to the next. Hence, the quality of a sentence ordering can be estimated by the number of continuous sentences that are also reproduced in the reference sentence ordering. This is equivalent to measuring a precision of continuous sentences in an ordering against the reference ordering. We define P n to measure the precision of n continuous sentences in an ordering to be evaluated as,",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Metrics for semi-automatic evaluation",
                "sec_num": "4.2"
            },
            {
                "text": "EQUATION",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [
                    {
                        "start": 0,
                        "end": 8,
                        "text": "EQUATION",
                        "ref_id": "EQREF",
                        "raw_str": "P n = m N \u2212 n + 1 .",
                        "eq_num": "(12)"
                    }
                ],
                "section": "Metrics for semi-automatic evaluation",
                "sec_num": "4.2"
            },
            {
                "text": "Here, N is the number of sentences in the reference ordering; n is the length of continuous sentences on which we are evaluating; m is the number of continuous sentences that appear in both the evaluation and reference orderings. In Figure 7 , the precision of 3 continuous sentences P 3 is calculated as:",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 233,
                        "end": 241,
                        "text": "Figure 7",
                        "ref_id": "FIGREF8"
                    }
                ],
                "eq_spans": [],
                "section": "Metrics for semi-automatic evaluation",
                "sec_num": "4.2"
            },
            {
                "text": "EQUATION",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [
                    {
                        "start": 0,
                        "end": 8,
                        "text": "EQUATION",
                        "ref_id": "EQREF",
                        "raw_str": "P 3 = 2 5 \u2212 3 + 1 = 0.67.",
                        "eq_num": "(13)"
                    }
                ],
                "section": "Metrics for semi-automatic evaluation",
                "sec_num": "4.2"
            },
            {
                "text": "The Average Continuity (AC) is defined as the logarithmic average of P n over 2 to k:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Metrics for semi-automatic evaluation",
                "sec_num": "4.2"
            },
            {
                "text": "AC = exp 1 k \u2212 1 k n=2 log(P n + \u03b1) . (14)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Metrics for semi-automatic evaluation",
                "sec_num": "4.2"
            },
            {
                "text": "Here, k is a parameter to control the range of the logarithmic average; and \u03b1 is a small value in case if P n is zero. We set k = 4 (ie, more than five continuous sentences are not included for evaluation) and \u03b1 = 0.01. Average Continuity becomes 0 when evaluation and reference orderings share no continuous sentences and 1 when the two orderings are identical. In Figure 7 , Average Continuity is calculated as 0.63. The underlying idea of Formula 14 was proposed by Papineni et al. (2002) as the BLEU metric for the semi-automatic evaluation of machine-translation systems. The original definition of the BLEU metric is to compare a machine-translated text with its reference translation by using the word n-grams. Table 2 reports the resemblance of orderings produced by six algorithms to the human-made ones with three metrics, Spearman's rank correlation, Kendall's rank correlation, and Average Continuity. The proposed method (AGL) outperforms the rest in all evaluation metrics, although the chronological ordering (CHR) appeared to play the major role. The one-way analysis of variance (ANOVA) verified the effects of different algorithms for sentence orderings with all metrics (p < 0.01). We performed Tukey Honest Significant Differences (HSD) test to compare differences among these algorithms. The Tukey test revealed that AGL was significantly better than the rest. Even though we could not compare our experiment with the probabilistic approach (Lapata, 2003) directly due to the difference of the text corpora, the Kendall coefficient reported higher agreement than Lapata's experiment (Kendall=0.48 with lemmatized nouns and Kendall=0.56 with verb-noun dependencies). Figure 8 shows precision P n with different length values of continuous sentence n for the six methods compared in Table 2 . The number of continuous sentences becomes sparse for a higher value of length n. Therefore, the precision values decrease as the length n increases. Although RND ordering reported some continuous sentences for lower n values, no continuous sentences could be observed for the higher n values. Four criteria described in Section 3 (ie, CHR, TOP, PRE, SUC) produce segments of continuous sentences at all values of n.",
                "cite_spans": [
                    {
                        "start": 469,
                        "end": 491,
                        "text": "Papineni et al. (2002)",
                        "ref_id": "BIBREF8"
                    },
                    {
                        "start": 1462,
                        "end": 1476,
                        "text": "(Lapata, 2003)",
                        "ref_id": "BIBREF2"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 366,
                        "end": 374,
                        "text": "Figure 7",
                        "ref_id": "FIGREF8"
                    },
                    {
                        "start": 718,
                        "end": 725,
                        "text": "Table 2",
                        "ref_id": "TABREF2"
                    },
                    {
                        "start": 1687,
                        "end": 1695,
                        "text": "Figure 8",
                        "ref_id": "FIGREF9"
                    },
                    {
                        "start": 1802,
                        "end": 1809,
                        "text": "Table 2",
                        "ref_id": "TABREF2"
                    }
                ],
                "eq_spans": [],
                "section": "Metrics for semi-automatic evaluation",
                "sec_num": "4.2"
            },
            {
                "text": "We present a bottom-up approach to arrange sentences extracted for multi-document summarization. Our experimental results showed a significant improvement over existing sentence ordering strategies. However, the results also implied that chronological ordering played the major role in arranging sentences. A future direction of this study would be to explore the application of the proposed framework to more generic texts, such as documents without chronological information.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusion",
                "sec_num": "5"
            },
            {
                "text": "http://lr-www.pi.titech.ac.jp/tsc/tsc3-en.html 3 Each extract consists of ca. 15 sentences on average.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            }
        ],
        "back_matter": [
            {
                "text": "We used Mainichi Shinbun and Yomiuri Shinbun newspaper articles, and the TSC-3 test collection.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Acknowledgment",
                "sec_num": null
            }
        ],
        "bib_entries": {
            "BIBREF0": {
                "ref_id": "b0",
                "title": "Catching the drift: Probabilistic content models, with applications to generation and summarization",
                "authors": [
                    {
                        "first": "Regina",
                        "middle": [],
                        "last": "Barzilay",
                        "suffix": ""
                    },
                    {
                        "first": "Lillian",
                        "middle": [],
                        "last": "Lee",
                        "suffix": ""
                    }
                ],
                "year": 2004,
                "venue": "HLT-NAACL 2004: Proceedings of the Main Conference",
                "volume": "",
                "issue": "",
                "pages": "113--120",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Regina Barzilay and Lillian Lee. 2004. Catching the drift: Probabilistic content models, with applications to generation and summarization. In HLT-NAACL 2004: Proceedings of the Main Conference, pages 113-120.",
                "links": null
            },
            "BIBREF1": {
                "ref_id": "b1",
                "title": "Inferring strategies for sentence ordering in multidocument news summarization",
                "authors": [
                    {
                        "first": "Regina",
                        "middle": [],
                        "last": "Barzilay",
                        "suffix": ""
                    },
                    {
                        "first": "Noemie",
                        "middle": [],
                        "last": "Elhadad",
                        "suffix": ""
                    },
                    {
                        "first": "Kathleen",
                        "middle": [],
                        "last": "Mckeown",
                        "suffix": ""
                    }
                ],
                "year": 2002,
                "venue": "Journal of Artificial Intelligence Research",
                "volume": "17",
                "issue": "",
                "pages": "35--55",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Regina Barzilay, Noemie Elhadad, and Kathleen McK- eown. 2002. Inferring strategies for sentence order- ing in multidocument news summarization. Journal of Artificial Intelligence Research, 17:35-55.",
                "links": null
            },
            "BIBREF2": {
                "ref_id": "b2",
                "title": "Probabilistic text structuring: Experiments with sentence ordering",
                "authors": [
                    {
                        "first": "Mirella",
                        "middle": [],
                        "last": "Lapata",
                        "suffix": ""
                    }
                ],
                "year": 2003,
                "venue": "Proceedings of the annual meeting of ACL",
                "volume": "",
                "issue": "",
                "pages": "545--552",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Mirella Lapata. 2003. Probabilistic text structuring: Experiments with sentence ordering. Proceedings of the annual meeting of ACL, 2003., pages 545-552.",
                "links": null
            },
            "BIBREF3": {
                "ref_id": "b3",
                "title": "Neats:a multidocument summarizer",
                "authors": [
                    {
                        "first": "C",
                        "middle": [
                            "Y"
                        ],
                        "last": "Lin",
                        "suffix": ""
                    },
                    {
                        "first": "E",
                        "middle": [],
                        "last": "Hovy",
                        "suffix": ""
                    }
                ],
                "year": 2001,
                "venue": "Proceedings of the Document Understanding Workshop(DUC)",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "C.Y. Lin and E. Hovy. 2001. Neats:a multidocument summarizer. Proceedings of the Document Under- standing Workshop(DUC).",
                "links": null
            },
            "BIBREF4": {
                "ref_id": "b4",
                "title": "Rhetorical structure theory: Toward a functional theory of text organization. Text",
                "authors": [
                    {
                        "first": "W",
                        "middle": [],
                        "last": "Mann",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Thompson",
                        "suffix": ""
                    }
                ],
                "year": 1988,
                "venue": "",
                "volume": "8",
                "issue": "",
                "pages": "243--281",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "W. Mann and S. Thompson. 1988. Rhetorical structure theory: Toward a functional theory of text organiza- tion. Text, 8:243-281.",
                "links": null
            },
            "BIBREF5": {
                "ref_id": "b5",
                "title": "From local to global coherence: A bottom-up approach to text planning",
                "authors": [
                    {
                        "first": "Daniel",
                        "middle": [],
                        "last": "Marcu",
                        "suffix": ""
                    }
                ],
                "year": 1997,
                "venue": "Proceedings of the 14th National Conference on Artificial Intelligence",
                "volume": "",
                "issue": "",
                "pages": "629--635",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Daniel Marcu. 1997. From local to global coherence: A bottom-up approach to text planning. In Proceed- ings of the 14th National Conference on Artificial Intelligence, pages 629-635, Providence, Rhode Is- land.",
                "links": null
            },
            "BIBREF6": {
                "ref_id": "b6",
                "title": "Towards multidocument summarization by reformulation: Progress and prospects",
                "authors": [
                    {
                        "first": "Kathleen",
                        "middle": [],
                        "last": "Mckeown",
                        "suffix": ""
                    },
                    {
                        "first": "Judith",
                        "middle": [],
                        "last": "Klavans",
                        "suffix": ""
                    }
                ],
                "year": 1999,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "453--460",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Kathleen McKeown, Judith Klavans, Vasileios Hatzi- vassiloglou, Regina Barzilay, and Eleazar Eskin. 1999. Towards multidocument summarization by reformulation: Progress and prospects. AAAI/IAAI, pages 453-460.",
                "links": null
            },
            "BIBREF7": {
                "ref_id": "b7",
                "title": "Improving chronological sentence ordering by precedence relation",
                "authors": [
                    {
                        "first": "Naoaki",
                        "middle": [],
                        "last": "Okazaki",
                        "suffix": ""
                    },
                    {
                        "first": "Yutaka",
                        "middle": [],
                        "last": "Matsuo",
                        "suffix": ""
                    },
                    {
                        "first": "Mitsuru",
                        "middle": [],
                        "last": "Ishizuka",
                        "suffix": ""
                    }
                ],
                "year": 2004,
                "venue": "Proceedings of 20th International Conference on Computational Linguistics (COLING 04)",
                "volume": "",
                "issue": "",
                "pages": "750--756",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Naoaki Okazaki, Yutaka Matsuo, and Mitsuru Ishizuka. 2004. Improving chronological sentence ordering by precedence relation. In Proceedings of 20th International Conference on Computational Linguistics (COLING 04), pages 750-756.",
                "links": null
            },
            "BIBREF8": {
                "ref_id": "b8",
                "title": "Bleu:a method for automatic evaluation of machine translation",
                "authors": [
                    {
                        "first": "Kishore",
                        "middle": [],
                        "last": "Papineni",
                        "suffix": ""
                    },
                    {
                        "first": "Salim",
                        "middle": [],
                        "last": "Roukos",
                        "suffix": ""
                    },
                    {
                        "first": "Todd",
                        "middle": [],
                        "last": "Ward",
                        "suffix": ""
                    },
                    {
                        "first": "Wei-Jing",
                        "middle": [],
                        "last": "Zhu",
                        "suffix": ""
                    }
                ],
                "year": 2002,
                "venue": "Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics (ACL)",
                "volume": "",
                "issue": "",
                "pages": "311--318",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Kishore Papineni, Salim Roukos, Todd Ward, and Wei- Jing Zhu. 2002. Bleu:a method for automatic eval- uation of machine translation. Proceedings of the 40th Annual Meeting of the Association for Compu- tational Linguistics (ACL), pages 311-318.",
                "links": null
            },
            "BIBREF9": {
                "ref_id": "b9",
                "title": "Generating natural language summaries from multiple on-line sources",
                "authors": [
                    {
                        "first": "R",
                        "middle": [],
                        "last": "Dragomir",
                        "suffix": ""
                    },
                    {
                        "first": "Kathy",
                        "middle": [],
                        "last": "Radev",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "Mckeown",
                        "suffix": ""
                    }
                ],
                "year": 1999,
                "venue": "Computational Linguistics",
                "volume": "24",
                "issue": "",
                "pages": "469--500",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Dragomir R. Radev and Kathy McKeown. 1999. Generating natural language summaries from mul- tiple on-line sources. Computational Linguistics, 24:469-500.",
                "links": null
            },
            "BIBREF10": {
                "ref_id": "b10",
                "title": "Statistical Learning Theory",
                "authors": [
                    {
                        "first": "V",
                        "middle": [],
                        "last": "Vapnik",
                        "suffix": ""
                    }
                ],
                "year": 1998,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "V. Vapnik. 1998. Statistical Learning Theory. Wiley, Chichester, GB.",
                "links": null
            }
        },
        "ref_entries": {
            "FIGREF0": {
                "num": null,
                "type_str": "figure",
                "uris": null,
                "text": "Arranging four sentences A, B, C, and D with a bottom-up approach."
            },
            "FIGREF1": {
                "num": null,
                "type_str": "figure",
                "uris": null,
                "text": "the association strength of the segments A and B. The association strengths of the two segments with different directions, eg, f (A B) and f (B A), are not always identical in our definition, f (A B) = f (B A). (3) Figure 1 shows the process of arranging four sentences a, b, c, and d. Firstly, we initialize four segments with a sentence for each, A = (a), B = (b), C = (c), D = (d). (4) Suppose that f (B A) has the highest value of all possible pairs, eg, f (A B), f (C D), etc, we concatenate B and A to obtain a new segment, E = (b a)."
            },
            "FIGREF4": {
                "num": null,
                "type_str": "figure",
                "uris": null,
                "text": "Figure 4: Partitioning a human-ordered extract into pairs of segments"
            },
            "FIGREF5": {
                "num": null,
                "type_str": "figure",
                "uris": null,
                "text": "Focusing on sentence a arranged just before the partition point and sentence b arranged just after we identify the pair {(a), (b)} of two segments (a) and (b). Enumerating all possible pairs of two segments facing just before/after the partitioning point, we obtain the following pairs, {(a), (b c)} and {(a), (b c d)}. Similarly, segment"
            },
            "FIGREF6": {
                "num": null,
                "type_str": "figure",
                "uris": null,
                "text": "Two vectors in a training data generated from two ordered segments A Bpairs, {(b), (c)}, {(a b), (c)}, {(b), (c d)}, {(a b), (c d)},are obtained from the partitioning point between sentence b and c. Collecting the segment pairs from the partitioning point between sentences c and d (i.e., {(c), (d)}, {(b c), (d)} and {(a b c), (d)})"
            },
            "FIGREF7": {
                "num": null,
                "type_str": "figure",
                "uris": null,
                "text": "Agglomerative ordering (AGL) is an orderingarranged by the proposed method;Random ordering (RND) is the lowest anchor, in which sentences are arranged randomly;Human-made ordering (HUM) is the highest anchor, in which sentences are arranged by a human subject;Chronological ordering (CHR) arranges sentences with the chronology criterion defined in Formula 8. Sentences are arranged in chronological order of their publication date;Topical-closeness ordering (TOP) arranges sentences with the topical-closeness criterion defined in Formula 9;"
            },
            "FIGREF8": {
                "num": null,
                "type_str": "figure",
                "uris": null,
                "text": "eval = (e a b c d) T ref = (a b c d e) An example of an ordering under evaluation T eval and its reference T ref ."
            },
            "FIGREF9": {
                "num": null,
                "type_str": "figure",
                "uris": null,
                "text": "Precision vs unit of measuring continuity."
            },
            "TABREF0": {
                "content": "<table><tr><td>Original article</td></tr><tr><td>for sentence</td></tr><tr><td>b 1</td></tr><tr><td>..</td></tr></table>",
                "num": null,
                "type_str": "table",
                "text": "Original article.... .. .. .... .. ....... ......",
                "html": null
            },
            "TABREF1": {
                "content": "<table><tr><td>Metric</td><td colspan=\"2\">Mean Std. Dev</td><td>Min</td><td>Max</td></tr><tr><td>Spearman</td><td>0.739</td><td>0.304</td><td>-0.2</td><td>1</td></tr><tr><td>Kendall</td><td>0.694</td><td>0.290</td><td>0</td><td>1</td></tr><tr><td colspan=\"2\">Average Continuity 0.401</td><td>0.404</td><td>0.001</td><td>1</td></tr></table>",
                "num": null,
                "type_str": "table",
                "text": "Correlation between two sets of humanordered extracts",
                "html": null
            },
            "TABREF2": {
                "content": "<table><tr><td colspan=\"2\">Method Spearman</td><td>Kendall</td><td>Average</td></tr><tr><td/><td colspan=\"3\">coefficient coefficient Continuity</td></tr><tr><td>RND</td><td>-0.127</td><td>-0.069</td><td>0.011</td></tr><tr><td>TOP</td><td>0.414</td><td>0.400</td><td>0.197</td></tr><tr><td>PRE</td><td>0.415</td><td>0.428</td><td>0.293</td></tr><tr><td>SUC</td><td>0.473</td><td>0.476</td><td>0.291</td></tr><tr><td>CHR</td><td>0.583</td><td>0.587</td><td>0.356</td></tr><tr><td>AGL</td><td>0.603</td><td>0.612</td><td>0.459</td></tr></table>",
                "num": null,
                "type_str": "table",
                "text": "Comparison with human-made ordering",
                "html": null
            }
        }
    }
}