File size: 98,881 Bytes
6fa4bc9 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244 1245 1246 1247 1248 1249 1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334 1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352 1353 1354 1355 1356 1357 1358 1359 1360 1361 1362 1363 1364 1365 1366 1367 1368 1369 1370 1371 1372 1373 1374 1375 1376 1377 1378 1379 1380 1381 1382 1383 1384 1385 1386 1387 1388 1389 1390 1391 1392 1393 1394 1395 1396 1397 1398 1399 1400 1401 1402 1403 1404 1405 1406 1407 1408 1409 1410 1411 1412 1413 1414 1415 1416 1417 1418 1419 1420 1421 1422 1423 1424 1425 1426 1427 1428 1429 1430 1431 1432 1433 1434 1435 1436 1437 1438 1439 1440 1441 1442 1443 1444 1445 1446 1447 1448 1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467 1468 1469 1470 1471 1472 1473 1474 1475 1476 1477 1478 1479 1480 1481 1482 1483 1484 1485 1486 1487 1488 1489 1490 1491 1492 1493 1494 1495 1496 1497 1498 1499 1500 1501 1502 1503 1504 1505 1506 1507 1508 1509 1510 1511 1512 1513 1514 1515 1516 1517 1518 1519 1520 1521 1522 1523 1524 1525 1526 1527 1528 1529 1530 1531 1532 1533 1534 1535 1536 1537 1538 1539 1540 1541 1542 1543 1544 1545 1546 1547 1548 1549 1550 1551 1552 1553 1554 1555 1556 1557 1558 1559 1560 1561 1562 1563 1564 1565 1566 1567 1568 1569 1570 1571 1572 1573 1574 1575 1576 1577 1578 1579 1580 1581 1582 1583 1584 1585 1586 1587 1588 1589 1590 1591 1592 1593 1594 1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612 1613 1614 1615 1616 1617 1618 1619 1620 1621 1622 1623 1624 1625 1626 1627 1628 1629 1630 1631 1632 1633 1634 1635 1636 1637 1638 1639 1640 1641 1642 1643 1644 1645 1646 1647 1648 1649 1650 1651 1652 1653 1654 1655 1656 1657 1658 1659 1660 1661 1662 1663 1664 1665 1666 1667 1668 1669 1670 1671 1672 1673 1674 1675 1676 1677 1678 1679 1680 1681 1682 1683 1684 1685 1686 1687 1688 1689 1690 1691 1692 1693 1694 1695 1696 1697 1698 1699 1700 1701 1702 1703 1704 1705 1706 1707 1708 1709 1710 1711 1712 1713 1714 1715 1716 1717 1718 1719 1720 1721 1722 1723 1724 1725 1726 1727 1728 1729 1730 1731 1732 1733 1734 1735 1736 1737 1738 1739 1740 1741 1742 1743 1744 1745 1746 1747 1748 1749 1750 1751 1752 1753 1754 1755 1756 1757 1758 1759 1760 1761 1762 1763 1764 1765 1766 1767 1768 1769 1770 1771 1772 1773 1774 1775 1776 1777 1778 1779 1780 1781 1782 1783 1784 1785 1786 1787 1788 1789 1790 1791 1792 1793 1794 1795 1796 1797 1798 1799 1800 1801 1802 1803 1804 1805 1806 1807 1808 1809 1810 1811 1812 1813 1814 1815 1816 1817 1818 1819 1820 1821 1822 1823 1824 1825 1826 1827 1828 1829 1830 1831 1832 1833 1834 1835 1836 1837 1838 1839 1840 1841 1842 1843 1844 1845 1846 1847 1848 1849 1850 1851 1852 1853 1854 1855 1856 1857 1858 1859 1860 1861 1862 1863 1864 1865 1866 1867 1868 1869 1870 1871 1872 1873 1874 1875 1876 1877 1878 1879 1880 1881 1882 1883 1884 1885 1886 1887 1888 1889 1890 1891 1892 | {
"paper_id": "P09-1024",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T08:54:08.348332Z"
},
"title": "Automatically Generating Wikipedia Articles: A Structure-Aware Approach",
"authors": [
{
"first": "Christina",
"middle": [],
"last": "Sauper",
"suffix": "",
"affiliation": {
"laboratory": "Artificial Intelligence Laboratory",
"institution": "Massachusetts Institute of Technology",
"location": {}
},
"email": "csauper@csail.mit.edu"
},
{
"first": "Regina",
"middle": [],
"last": "Barzilay",
"suffix": "",
"affiliation": {
"laboratory": "Artificial Intelligence Laboratory",
"institution": "Massachusetts Institute of Technology",
"location": {}
},
"email": "regina@csail.mit.edu"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "In this paper, we investigate an approach for creating a comprehensive textual overview of a subject composed of information drawn from the Internet. We use the high-level structure of human-authored texts to automatically induce a domainspecific template for the topic structure of a new overview. The algorithmic innovation of our work is a method to learn topicspecific extractors for content selection jointly for the entire template. We augment the standard perceptron algorithm with a global integer linear programming formulation to optimize both local fit of information into each topic and global coherence across the entire overview. The results of our evaluation confirm the benefits of incorporating structural information into the content selection process.",
"pdf_parse": {
"paper_id": "P09-1024",
"_pdf_hash": "",
"abstract": [
{
"text": "In this paper, we investigate an approach for creating a comprehensive textual overview of a subject composed of information drawn from the Internet. We use the high-level structure of human-authored texts to automatically induce a domainspecific template for the topic structure of a new overview. The algorithmic innovation of our work is a method to learn topicspecific extractors for content selection jointly for the entire template. We augment the standard perceptron algorithm with a global integer linear programming formulation to optimize both local fit of information into each topic and global coherence across the entire overview. The results of our evaluation confirm the benefits of incorporating structural information into the content selection process.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "In this paper, we consider the task of automatically creating a multi-paragraph overview article that provides a comprehensive summary of a subject of interest. Examples of such overviews include actor biographies from IMDB and disease synopses from Wikipedia. Producing these texts by hand is a labor-intensive task, especially when relevant information is scattered throughout a wide range of Internet sources. Our goal is to automate this process. We aim to create an overview of a subjecte.g., 3-M Syndrome -by intelligently combining relevant excerpts from across the Internet.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "As a starting point, we can employ methods developed for multi-document summarization. However, our task poses additional technical challenges with respect to content planning. Generating a well-rounded overview article requires proactive strategies to gather relevant material, such as searching the Internet. Moreover, the challenge of maintaining output readability is magnified when creating a longer document that discusses multiple topics.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "In our approach, we explore how the highlevel structure of human-authored documents can be used to produce well-formed comprehensive overview articles. We select relevant material for an article using a domain-specific automatically generated content template. For example, a template for articles about diseases might contain diagnosis, causes, symptoms, and treatment. Our system induces these templates by analyzing patterns in the structure of human-authored documents in the domain of interest. Then, it produces a new article by selecting content from the Internet for each part of this template. An example of our system's output 1 is shown in Figure 1 .",
"cite_spans": [],
"ref_spans": [
{
"start": 651,
"end": 659,
"text": "Figure 1",
"ref_id": "FIGREF0"
}
],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "The algorithmic innovation of our work is a method for learning topic-specific extractors for content selection jointly across the entire template. Learning a single topic-specific extractor can be easily achieved in a standard classification framework. However, the choices for different topics in a template are mutually dependent; for example, in a multi-topic article, there is potential for redundancy across topics. Simultaneously learning content selection for all topics enables us to explicitly model these inter-topic connections.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "We formulate this task as a structured classification problem. We estimate the parameters of our model using the perceptron algorithm augmented with an integer linear programming (ILP) formulation, run over a training set of example articles in the given domain.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "The key features of this structure-aware approach are twofold:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Diagnosis . . . No laboratories offering molecular genetic testing for prenatal diagnosis of 3-M syndrome are listed in the GeneTests Laboratory Directory. However, prenatal testing may be available for families in which the disease-causing mutations have been identified in an affected family member in a research or clinical laboratory.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Causes Three M syndrome is thought to be inherited as an autosomal recessive genetic trait. Human traits, including the classic genetic diseases, are the product of the interaction of two genes, one received from the father and one from the mother. In recessive disorders, the condition does not occur unless an individual inherits the same defective gene for the same trait from each parent. . . .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": ". . . Many of the symptoms and physical features associated with the disorder are apparent at birth (congenital). In some cases, individuals who carry a single copy of the disease gene (heterozygotes) may exhibit mild symptoms associated with Three M syndrome.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Symptoms",
"sec_num": null
},
{
"text": "Treatment . . . Genetic counseling will be of benefit for affected individuals and their families. Family members of affected individuals should also receive regular clinical evaluations to detect any symptoms and physical characteristics that may be potentially associated with Three M syndrome or heterozygosity for the disorder. Other treatment for Three M syndrome is symptomatic and supportive. \u2022 Automatic template creation: Templates are automatically induced from humanauthored documents. This ensures that the overview article will have the breadth expected in a comprehensive summary, with content drawn from a wide variety of Internet sources.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Symptoms",
"sec_num": null
},
{
"text": "\u2022 Joint parameter estimation for content selection: Parameters are learned jointly for all topics in the template. This procedure optimizes both local relevance of information for each topic and global coherence across the entire article.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Symptoms",
"sec_num": null
},
{
"text": "We evaluate our approach by creating articles in two domains: Actors and Diseases. For a data set, we use Wikipedia, which contains articles similar to those we wish to produce in terms of length and breadth. An advantage of this data set is that Wikipedia articles explicitly delineate topical sections, facilitating structural analysis. The results of our evaluation confirm the benefits of structureaware content selection over approaches that do not explicitly model topical structure.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Symptoms",
"sec_num": null
},
{
"text": "Concept-to-text generation and text-to-text generation take very different approaches to content selection. In traditional concept-to-text generation, a content planner provides a detailed template for what information should be included in the output and how this information should be organized (Reiter and Dale, 2000) . In text-to-text generation, such templates for information organization are not available; sentences are selected based on their salience properties (Mani and Maybury, 1999) . While this strategy is robust and portable across domains, output summaries often suffer from coherence and coverage problems.",
"cite_spans": [
{
"start": 297,
"end": 320,
"text": "(Reiter and Dale, 2000)",
"ref_id": "BIBREF24"
},
{
"start": 472,
"end": 496,
"text": "(Mani and Maybury, 1999)",
"ref_id": "BIBREF17"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "In between these two approaches is work on domain-specific text-to-text generation. Instances of these tasks are biography generation in summarization and answering definition requests in question-answering. In contrast to a generic summarizer, these applications aim to characterize the types of information that are essential in a given domain. This characterization varies greatly in granularity. For instance, some approaches coarsely discriminate between biographical and non-biographical information (Zhou et al., 2004; Biadsy et al., 2008) , while others go beyond binary distinction by identifying atomic events -e.g., occupation and marital status -that are typically included in a biography (Weischedel et al., 2004; Filatova and Prager, 2005; Filatova et al., 2006) . Commonly, such templates are specified manually and are hard-coded for a particular domain (Fujii and Ishikawa, 2004; Weischedel et al., 2004) .",
"cite_spans": [
{
"start": 506,
"end": 525,
"text": "(Zhou et al., 2004;",
"ref_id": "BIBREF29"
},
{
"start": 526,
"end": 546,
"text": "Biadsy et al., 2008)",
"ref_id": "BIBREF4"
},
{
"start": 701,
"end": 726,
"text": "(Weischedel et al., 2004;",
"ref_id": "BIBREF26"
},
{
"start": 727,
"end": 753,
"text": "Filatova and Prager, 2005;",
"ref_id": "BIBREF11"
},
{
"start": 754,
"end": 776,
"text": "Filatova et al., 2006)",
"ref_id": "BIBREF12"
},
{
"start": 870,
"end": 896,
"text": "(Fujii and Ishikawa, 2004;",
"ref_id": "BIBREF13"
},
{
"start": 897,
"end": 921,
"text": "Weischedel et al., 2004)",
"ref_id": "BIBREF26"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "Our work is related to these approaches; however, content selection in our work is driven by domain-specific automatically induced templates. As our experiments demonstrate, patterns observed in domain-specific training data provide sufficient constraints for topic organization, which is crucial for a comprehensive text.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "Our work also relates to a large body of recent work that uses Wikipedia material. Instances of this work include information extraction, ontology induction and resource acquisition (Wu and Weld, 2007; Biadsy et al., 2008; Nastase, 2008; Nastase and Strube, 2008) . Our focus is on a different task -generation of new overview articles that follow the structure of Wikipedia articles.",
"cite_spans": [
{
"start": 182,
"end": 201,
"text": "(Wu and Weld, 2007;",
"ref_id": "BIBREF27"
},
{
"start": 202,
"end": 222,
"text": "Biadsy et al., 2008;",
"ref_id": "BIBREF4"
},
{
"start": 223,
"end": 237,
"text": "Nastase, 2008;",
"ref_id": "BIBREF22"
},
{
"start": 238,
"end": 263,
"text": "Nastase and Strube, 2008)",
"ref_id": "BIBREF21"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "The goal of our system is to produce a comprehensive overview article given a title -e.g., Cancer. We assume that relevant information on the subject is available on the Internet but scattered among several pages interspersed with noise.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Method",
"sec_num": "3"
},
{
"text": "We are provided with a training corpus consisting of n documents d 1 . . . d n in the same domain -e.g., Diseases. Each document d i has a title and a set of delineated sections 2 s i1 . . . s im . The number of sections m varies between documents. Each section s ij also has a corresponding heading h ije.g., Treatment.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Method",
"sec_num": "3"
},
{
"text": "Our overview article creation process consists of three parts. First, a preprocessing step creates a template and searches for a number of candidate excerpts from the Internet. Next, parameters must be trained for the content selection algorithm using our training data set. Finally, a complete article may be created by combining a selection of candidate excerpts.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Method",
"sec_num": "3"
},
{
"text": "1. Preprocessing (Section 3.1) Our preprocessing step leverages previous work in topic segmentation and query reformulation to prepare a template and a set of candidate excerpts for content selection. Template generation must occur once per domain, whereas search occurs every time an article is generated in both learning and application.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Method",
"sec_num": "3"
},
{
"text": "(a) Template Induction To create a content template, we cluster all section headings h i1 . . . h im for all documents d i . Each cluster is labeled with the most common heading h ij within the cluster. The largest k clusters are selected to become topics t 1 . . . t k , which form the domain-specific content template. (b) Search For each document that we wish to create, we retrieve from the Internet a set of r excerpts e j1 . . . e jr for each topic t j from the template. We define appropriate search queries using the requested document title and topics t j .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Method",
"sec_num": "3"
},
{
"text": "2. Learning Content Selection (Section 3.2) For each topic t j , we learn the corresponding topic-specific parameters w j to determine the quality of a given excerpt. Using the perceptron framework augmented with an ILP formulation for global optimization, the system is trained to select the best excerpt for each document d i and each topic t j . For training, we assume the best excerpt is the original human-authored text s ij .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Method",
"sec_num": "3"
},
{
"text": "3. Application (Section 3.2) Given the title of a requested document, we select several excerpts from the candidate vectors returned by the search procedure (1b) to create a comprehensive overview article. We perform the decoding procedure jointly using learned parameters w 1 . . . w k and the same ILP formulation for global optimization as in training. The result is a new document with k excerpts, one for each topic.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Method",
"sec_num": "3"
},
{
"text": "Template Induction A content template specifies the topical structure of documents in one domain. For instance, the template for articles about actors consists of four topics t 1 . . . t 4 : biography, early life, career, and personal life. Using this template to create the biography of a new actor will ensure that its information coverage is consistent with existing human-authored documents. We aim to derive these templates by discovering common patterns in the organization of documents in a domain of interest. There has been a sizable amount of research on structure induction ranging from linear segmentation (Hearst, 1994) to content modeling (Barzilay and Lee, 2004) . At the core of these methods is the assumption that fragments of text conveying similar information have similar word distribution patterns. Therefore, often a simple segment clustering across domain texts can identify strong patterns in content structure (Barzilay and Elhadad, 2003) . Clusters containing fragments from many documents are indicative of topics that are essential for a comprehensive summary. Given the simplicity and robustness of this approach, we utilize it for template induction.",
"cite_spans": [
{
"start": 618,
"end": 632,
"text": "(Hearst, 1994)",
"ref_id": "BIBREF15"
},
{
"start": 653,
"end": 677,
"text": "(Barzilay and Lee, 2004)",
"ref_id": "BIBREF2"
},
{
"start": 936,
"end": 964,
"text": "(Barzilay and Elhadad, 2003)",
"ref_id": "BIBREF1"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Preprocessing",
"sec_num": "3.1"
},
{
"text": "We cluster all section headings h i1 . . . h im from all documents d i using a repeated bisectioning algorithm (Zhao et al., 2005) . As a similarity function, we use cosine similarity weighted with TF*IDF. We eliminate any clusters with low internal similarity (i.e., smaller than 0.5), as we assume these are \"miscellaneous\" clusters that will not yield unified topics.",
"cite_spans": [
{
"start": 111,
"end": 130,
"text": "(Zhao et al., 2005)",
"ref_id": "BIBREF28"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Preprocessing",
"sec_num": "3.1"
},
{
"text": "We determine the average number of sections k over all documents in our training set, then select the k largest section clusters as topics. We order these topics as t 1 . . . t k using a majority ordering algorithm (Cohen et al., 1998) . This algorithm finds a total order among clusters that is consistent with a maximal number of pairwise relationships observed in our data set.",
"cite_spans": [
{
"start": 215,
"end": 235,
"text": "(Cohen et al., 1998)",
"ref_id": "BIBREF6"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Preprocessing",
"sec_num": "3.1"
},
{
"text": "Each topic t j is identified by the most frequent heading found within the cluster -e.g., Causes. This set of topics forms the content template for a domain. Search To retrieve relevant excerpts, we must define appropriate search queries for each topic t 1 . . . t k . Query reformulation is an active area of research (Agichtein et al., 2001) . We have experimented with several of these methods for drawing search queries from representative words in the body text of each topic; however, we find that the best performance is provided by deriving queries from a conjunction of the document title and topic -e.g., \"3-M syndrome\" diagnosis.",
"cite_spans": [
{
"start": 319,
"end": 343,
"text": "(Agichtein et al., 2001)",
"ref_id": "BIBREF0"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Preprocessing",
"sec_num": "3.1"
},
{
"text": "Using these queries, we search using Yahoo! and retrieve the first ten result pages for each topic. From each of these pages, we extract all possible excerpts consisting of chunks of text between standardized boundary indicators (such as <p> tags). In our experiments, there are an average of 6 excerpts taken from each page. For each topic t j of each document we wish to create, the total number of excerpts r found on the Internet may differ. We label the excerpts e j1 . . . e jr .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Preprocessing",
"sec_num": "3.1"
},
{
"text": "Our selection model takes the content template t 1 . . . t k and the candidate excerpts e j1 . . . e jr for each topic t j produced in the previous steps. It then selects a series of k excerpts, one from each topic, to create a coherent summary.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Selection Model",
"sec_num": "3.2"
},
{
"text": "One possible approach is to perform individual selections from each set of excerpts e j1 . . . e jr and then combine the results. This strategy is commonly used in multi-document summarization (Barzilay et al., 1999; Goldstein et al., 2000; Radev et al., 2000) , where the combination step eliminates the redundancy across selected excerpts. However, separating the two steps may not be optimal for this task -the balance between coverage and redundancy is harder to achieve when a multi-paragraph summary is generated. In addition, a more discriminative selection strategy is needed when candidate excerpts are drawn directly from the web, as they may be contaminated with noise.",
"cite_spans": [
{
"start": 193,
"end": 216,
"text": "(Barzilay et al., 1999;",
"ref_id": "BIBREF3"
},
{
"start": 217,
"end": 240,
"text": "Goldstein et al., 2000;",
"ref_id": "BIBREF14"
},
{
"start": 241,
"end": 260,
"text": "Radev et al., 2000)",
"ref_id": "BIBREF23"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Selection Model",
"sec_num": "3.2"
},
{
"text": "We propose a novel joint training algorithm that learns selection criteria for all the topics simultaneously. This approach enables us to maximize both local fit and global coherence. We implement this algorithm using the perceptron framework, as it can be easily modified for structured prediction while preserving convergence guarantees (Daum\u00e9 III and Marcu, 2005; Snyder and Barzilay, 2007) .",
"cite_spans": [
{
"start": 339,
"end": 366,
"text": "(Daum\u00e9 III and Marcu, 2005;",
"ref_id": "BIBREF9"
},
{
"start": 367,
"end": 393,
"text": "Snyder and Barzilay, 2007)",
"ref_id": "BIBREF25"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Selection Model",
"sec_num": "3.2"
},
{
"text": "In this section, we first describe the structure and decoding procedure of our model. We then present an algorithm to jointly learn the parameters of all topic models.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Selection Model",
"sec_num": "3.2"
},
{
"text": "The model inputs are as follows:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model Structure",
"sec_num": "3.2.1"
},
{
"text": "\u2022 The title of the desired document \u2022 t 1 . . . t k -topics from the content template \u2022 e j1 . . . e jr -candidate excerpts for each topic t j",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model Structure",
"sec_num": "3.2.1"
},
{
"text": "In addition, we define feature and parameter vectors:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model Structure",
"sec_num": "3.2.1"
},
{
"text": "\u2022 \u03c6(e jl ) -feature vector for the lth candidate excerpt for topic t j \u2022 w 1 . . . w k -parameter vectors, one for each of the topics t 1 . . . t k",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model Structure",
"sec_num": "3.2.1"
},
{
"text": "Our model constructs a new article by following these two steps: Ranking First, we attempt to rank candidate excerpts based on how representative they are of each individual topic. For each topic t j , we induce a ranking of the excerpts e j1 . . . e jr by mapping each excerpt e jl to a score:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model Structure",
"sec_num": "3.2.1"
},
{
"text": "score j (e jl ) = \u03c6(e jl ) \u2022 w j",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model Structure",
"sec_num": "3.2.1"
},
{
"text": "Candidates for each topic are ranked from highest to lowest score. After this procedure, the position l of excerpt e jl within the topic-specific candidate vector is the excerpt's rank. Optimizing the Global Objective To avoid redundancy between topics, we formulate an optimization problem using excerpt rankings to create the final article. Given k topics, we would like to select one excerpt e jl for each topic t j , such that the rank is minimized; that is, score j (e jl ) is high.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model Structure",
"sec_num": "3.2.1"
},
{
"text": "To select the optimal excerpts, we employ integer linear programming (ILP). This framework is commonly used in generation and summarization applications where the selection process is driven by multiple constraints (Marciniak and Strube, 2005; Clarke and Lapata, 2007) .",
"cite_spans": [
{
"start": 215,
"end": 243,
"text": "(Marciniak and Strube, 2005;",
"ref_id": "BIBREF19"
},
{
"start": 244,
"end": 268,
"text": "Clarke and Lapata, 2007)",
"ref_id": "BIBREF5"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Model Structure",
"sec_num": "3.2.1"
},
{
"text": "We represent excerpts included in the output using a set of indicator variables, x jl . For each excerpt e jl , the corresponding indicator variable x jl = 1 if the excerpt is included in the final document, and x jl = 0 otherwise.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model Structure",
"sec_num": "3.2.1"
},
{
"text": "Our objective is to minimize the ranks of the excerpts selected for the final document:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model Structure",
"sec_num": "3.2.1"
},
{
"text": "min k j=1 r l=1 l \u2022 x jl",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model Structure",
"sec_num": "3.2.1"
},
{
"text": "We augment this formulation with two types of constraints. Exclusivity Constraints We want to ensure that exactly one indicator x jl is nonzero for each topic t j . These constraints are formulated as follows:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model Structure",
"sec_num": "3.2.1"
},
{
"text": "r l=1 x jl = 1 \u2200j \u2208 {1 . . . k}",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model Structure",
"sec_num": "3.2.1"
},
{
"text": "We also want to prevent redundancy across topics. We define sim(e jl , e j l ) as the cosine similarity between excerpts e jl from topic t j and e j l from topic t j . We introduce constraints that ensure no pair of excerpts has similarity above 0.5:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Redundancy Constraints",
"sec_num": null
},
{
"text": "(x jl + x j l ) \u2022 sim(e jl , e j l ) \u2264 1 \u2200j, j = 1 . . . k \u2200l, l = 1 . . . r",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Redundancy Constraints",
"sec_num": null
},
{
"text": "If excerpts e jl and e j l have cosine similarity sim(e jl , e j l ) > 0.5, only one excerpt may be selected for the final document -i.e., either x jl or x j l may be 1, but not both. Conversely, if sim(e jl , e j l ) \u2264 0.5, both excerpts may be selected. Solving the ILP Solving an integer linear program is NP-hard (Cormen et al., 1992) ; however, in practice there exist several strategies for solving certain ILPs efficiently. In our study, we employed lp solve, 3 an efficient mixed integer programming solver which implements the Branch-and-Bound algorithm. On a larger scale, there are several alternatives to approximate the ILP results, such as a dynamic programming approximation to the knapsack problem (McDonald, 2007 Features As shown in Table 1 , most of the features we select in our model have been employed in previous work on summarization (Mani and Maybury, 1999) . All features except the SIMS feature are defined for individual excerpts in isolation. For each excerpt e jl , the value of the SIMS feature is the count of excerpts e jl in the same topic t j for which sim(e jl , e jl ) > 0.5. This feature quantifies the degree of repetition within a topic, often indicative of an excerpt's accuracy and relevance.",
"cite_spans": [
{
"start": 309,
"end": 338,
"text": "NP-hard (Cormen et al., 1992)",
"ref_id": null
},
{
"start": 714,
"end": 729,
"text": "(McDonald, 2007",
"ref_id": "BIBREF20"
},
{
"start": 858,
"end": 882,
"text": "(Mani and Maybury, 1999)",
"ref_id": "BIBREF17"
}
],
"ref_spans": [
{
"start": 751,
"end": 758,
"text": "Table 1",
"ref_id": "TABREF1"
}
],
"eq_spans": [],
"section": "Redundancy Constraints",
"sec_num": null
},
{
"text": "Generating Training Data For training, we are given n original documents d 1 . . . d n , a content template consisting of topics t 1 . . . t k , and a set of candidate excerpts e ij1 . . . e ijr for each document d i and topic t j . For each section of each document, we add the gold excerpt s ij to the corresponding vector of candidate excerpts e ij1 . . . e ijr . This excerpt represents the target for our training algorithm. Note that the algorithm does not require annotated ranking data; only knowledge of this \"optimal\" excerpt is required. However, if the excerpts provided in the training data have low quality, noise is introduced into the system. Training Procedure Our algorithm is a modification of the perceptron ranking algorithm (Collins, 2002) , which allows for joint learning across several ranking problems (Daum\u00e9 III and Marcu, 2005; Snyder and Barzilay, 2007) . Pseudocode for this algorithm is provided in Figure 2 . First, we define Rank(e ij1 . . . e ijr , w j ), which ranks all excerpts from the candidate excerpt vector e ij1 . . . e ijr for document d i and topic t j . Excerpts are ordered by score j (e jl ) using the current parameter values. We also define Optimize(e ij1 . . . e ijr ), which finds the optimal selection of excerpts (one per topic) given ranked lists of excerpts e ij1 . . . e ijr for each document d i and topic t j . These functions follow the ranking and optimization procedures described in Section 3.2.1. The algorithm maintains k parameter vectors w 1 . . . w k , one associated with each topic t j desired in the final article. During initialization, all parameter vectors are set to zeros (line 2).",
"cite_spans": [
{
"start": 746,
"end": 761,
"text": "(Collins, 2002)",
"ref_id": "BIBREF7"
},
{
"start": 828,
"end": 855,
"text": "(Daum\u00e9 III and Marcu, 2005;",
"ref_id": "BIBREF9"
},
{
"start": 856,
"end": 882,
"text": "Snyder and Barzilay, 2007)",
"ref_id": "BIBREF25"
}
],
"ref_spans": [
{
"start": 930,
"end": 938,
"text": "Figure 2",
"ref_id": "FIGREF2"
}
],
"eq_spans": [],
"section": "Model Training",
"sec_num": "3.2.2"
},
{
"text": "To learn the optimal parameters, this algorithm iterates over the training set until the parameters converge or a maximum number of iterations is reached (line 3). For each document in the training set (line 4), the following steps occur: First, candidate excerpts for each topic are ranked (lines 5-6). Next, decoding through ILP optimization is performed over all ranked lists of candidate excerpts, selecting one excerpt for each topic (line 7). Finally, the parameters are updated in a joint fashion. For each topic (line 8), if the selected excerpt is not similar enough to the gold excerpt (line 9), the parameters for that topic are updated using a standard perceptron update rule (line 10). When convergence is reached or the maximum iteration count is exceeded, the learned parameter values are returned (line 12).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Model Training",
"sec_num": "3.2.2"
},
{
"text": "The use of ILP during each step of training sets this algorithm apart from previous work. In prior research, ILP was used as a postprocessing step to remove redundancy and make other global decisions about parameters (McDonald, 2007; Marciniak and Strube, 2005; Clarke and Lapata, 2007) . However, in our training, we intertwine the complete decoding procedure with the parameter updates. Our joint learning approach finds per-topic parameter values that are maximally suited for the global decoding procedure for content selection.",
"cite_spans": [
{
"start": 217,
"end": 233,
"text": "(McDonald, 2007;",
"ref_id": "BIBREF20"
},
{
"start": 234,
"end": 261,
"text": "Marciniak and Strube, 2005;",
"ref_id": "BIBREF19"
},
{
"start": 262,
"end": 286,
"text": "Clarke and Lapata, 2007)",
"ref_id": "BIBREF5"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Model Training",
"sec_num": "3.2.2"
},
{
"text": "We evaluate our method by observing the quality of automatically created articles in different domains. We compute the similarity of a large number of articles produced by our system and several baselines to the original human-authored articles using ROUGE, a standard metric for summary quality. In addition, we perform an analysis of edi-Input:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "4"
},
{
"text": "d1 . . . dn: A set of n documents, each containing k sections si1 . . . s ik eij1 . . . eijr: Sets of candidate excerpts for each topic tj and document di Define:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "4"
},
{
"text": "Rank(eij1 . . . eijr, wj):",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "4"
},
{
"text": "As described in Section 3.2.1: Calculates scorej(e ijl ) for all excerpts for document di and topic tj, using parameters wj. Orders the list of excerpts by scorej(e ijl ) from highest to lowest. Optimize(ei11 . . . e ikr ):",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "4"
},
{
"text": "As described in Section 3.2.1: Finds the optimal selection of excerpts to form a final article, given ranked lists of excerpts for each topic t1 . . . t k . Returns a list of k excerpts, one for each topic. \u03c6(e ijl ):",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "4"
},
{
"text": "Returns the feature vector representing excerpt e ijl Initialization: Data For evaluation, we consider two domains: American Film Actors and Diseases. These domains have been commonly used in prior work on summarization (Weischedel et al., 2004; Zhou et al., 2004; Filatova and Prager, 2005; Demner-Fushman and Lin, 2007; Biadsy et al., 2008) . Our text corpus consists of articles drawn from the corresponding categories in Wikipedia. There are 2,150 articles in American Film Actors and 523 articles in Diseases. For each domain, we randomly select 90% of articles for training and test on the remaining 10%. Human-authored articles in both domains contain an average of four topics, and each topic contains an average of 193 words. In order to model the real-world scenario where Wikipedia articles are not always available (as for new or specialized topics), we specifically exclude Wikipedia sources during our search pro- cedure (Section 3.1) for evaluation. Baselines Our first baseline, Search, relies solely on search engine ranking for content selection. Using the article title as a query -e.g., Bacillary Angiomatosis, this method selects the web page that is ranked first by the search engine. From this page we select the first k paragraphs where k is defined in the same way as in our full model. If there are less than k paragraphs on the page, all paragraphs are selected, but no other sources are used. This yields a document of comparable size with the output of our system. Despite its simplicity, this baseline is not naive: extracting material from a single document guarantees that the output is coherent, and a page highly ranked by a search engine may readily contain a comprehensive overview of the subject. Our second baseline, No Template, does not use a template to specify desired topics; therefore, there are no constraints on content selection. Instead, we follow a simplified form of previous work on biography creation, where a classifier is trained to distinguish biographical text (Zhou et al., 2004; Biadsy et al., 2008) .",
"cite_spans": [
{
"start": 220,
"end": 245,
"text": "(Weischedel et al., 2004;",
"ref_id": "BIBREF26"
},
{
"start": 246,
"end": 264,
"text": "Zhou et al., 2004;",
"ref_id": "BIBREF29"
},
{
"start": 265,
"end": 291,
"text": "Filatova and Prager, 2005;",
"ref_id": "BIBREF11"
},
{
"start": 292,
"end": 321,
"text": "Demner-Fushman and Lin, 2007;",
"ref_id": "BIBREF10"
},
{
"start": 322,
"end": 342,
"text": "Biadsy et al., 2008)",
"ref_id": "BIBREF4"
},
{
"start": 2017,
"end": 2036,
"text": "(Zhou et al., 2004;",
"ref_id": "BIBREF29"
},
{
"start": 2037,
"end": 2057,
"text": "Biadsy et al., 2008)",
"ref_id": "BIBREF4"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "4"
},
{
"text": "In this case, we train a classifier to distinguish domain-specific text. Positive training data is drawn from all topics in the given domain corpus. To find negative training data, we perform the search procedure as in our full model (see Section 3.1) using only the article titles as search queries. Any excerpts which have very low similarity to the original articles are used as negative examples. During the decoding procedure, we use the same search procedure. We then classify each excerpt as relevant or irrelevant and select the k non-redundant excerpts with the highest relevance confidence scores.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "4"
},
{
"text": "Our third baseline, Disjoint, uses the ranking perceptron framework as in our full system; however, rather than perform an optimization step during training and decoding, we simply select the highest-ranked excerpt for each topic. This equates to standard linear classification for each section individually.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "4"
},
{
"text": "In addition to these baselines, we compare against an Oracle system. For each topic present in the human-authored article, the Oracle selects the excerpt from our full model's candidate excerpts with the highest cosine similarity to the human-authored text. This excerpt is the optimal automatic selection from the results available, and therefore represents an upper bound on our excerpt selection task. Some articles contain additional topics beyond those in the template; in these cases, the Oracle system produces a longer article than our algorithm. Table 2 shows the average number of excerpts selected and sources used in articles created by our full model and each baseline. Automatic Evaluation To assess the quality of the resulting overview articles, we compare them with the original human-authored articles. We use ROUGE, an evaluation metric employed at the Document Understanding Conferences (DUC), which assumes that proximity to human-authored text is an indicator of summary quality. We use the publicly available ROUGE toolkit (Lin, 2004) to compute recall, precision, and F-score for ROUGE-1. We use the Wilcoxon Signed Rank Test to determine statistical significance. Analysis of Human Edits In addition to our automatic evaluation, we perform a study of reactions to system-produced articles by the general public. To achieve this goal, we insert automatically created articles 4 into Wikipedia itself and examine the feedback of Wikipedia editors. Selection of specific articles is constrained by the need to find topics which are currently of \"stub\" status that have enough information available on the Internet to construct a valid article. After a period of time, we analyzed the edits made to the articles to determine the overall editor reaction. We report results on 15 articles in the Diseases category 5 . Since Wikipedia is a live resource, we do not repeat this procedure for our baseline systems. Adding articles from systems which have previously demonstrated poor quality would be improper, especially in Diseases. Therefore, we present this analysis as an additional observation rather than a rigorous technical study.",
"cite_spans": [
{
"start": 1046,
"end": 1057,
"text": "(Lin, 2004)",
"ref_id": "BIBREF16"
}
],
"ref_spans": [
{
"start": 555,
"end": 562,
"text": "Table 2",
"ref_id": "TABREF3"
}
],
"eq_spans": [],
"section": "Experimental Setup",
"sec_num": "4"
},
{
"text": "Automatic Evaluation The results of this evaluation are shown in Table 3 . Our full model outperforms all of the baselines. By surpassing the Disjoint baseline, we demonstrate the benefits of joint classification. Furthermore, the high performance of both our full model and the Disjoint baseline relative to the other baselines shows the importance of structure-aware content selection. The Oracle system, which represents an upper bound on our system's capabilities, performs well.",
"cite_spans": [],
"ref_spans": [
{
"start": 65,
"end": 72,
"text": "Table 3",
"ref_id": "TABREF5"
}
],
"eq_spans": [],
"section": "Results",
"sec_num": "5"
},
{
"text": "The remaining baselines have different flaws: Articles produced by the No Template baseline tend to focus on a single topic extensively at the expense of breadth, because there are no constraints to ensure diverse topic selection. On the other hand, performance of the Search baseline varies dramatically. This is expected; this baseline relies heavily on both the search engine and individual web pages. The search engine must correctly rank relevant pages, and the web pages must provide the important material first. Analysis of Human Edits The results of our observation of editing patterns are shown in Table 4 . These articles have resided on Wikipedia for a period of time ranging from 5-11 months. All of them have been edited, and no articles were removed due to lack of quality. Moreover, ten automatically created articles have been promoted by human editors from stubs to regular Wikipedia entries based on the quality and coverage of the material. Information was removed in three cases for being irrelevant, one entire section and two smaller pieces. The most common changes were small edits to formatting and introduction of links to other Wikipedia articles in the body text.",
"cite_spans": [],
"ref_spans": [
{
"start": 608,
"end": 616,
"text": "Table 4",
"ref_id": "TABREF7"
}
],
"eq_spans": [],
"section": "Results",
"sec_num": "5"
},
{
"text": "In this paper, we investigated an approach for creating a multi-paragraph overview article by selecting relevant material from the web and organizing it into a single coherent text. Our algorithm yields significant gains over a structure-agnostic approach. Moreover, our results demonstrate the benefits of structured classification, which outperforms independently trained topical classifiers. Overall, the results of our evaluation combined with our analysis of human edits confirm that the proposed method can effectively produce comprehensive overview articles. This work opens several directions for future research. Diseases and American Film Actors exhibit fairly consistent article structures, which are successfully captured by a simple template creation process. However, with categories that exhibit structural variability, more sophisticated statistical approaches may be required to produce accurate templates. Moreover, a promising direction is to consider hierarchical discourse formalisms such as RST (Mann and Thompson, 1988) to supplement our template-based approach.",
"cite_spans": [
{
"start": 1013,
"end": 1042,
"text": "RST (Mann and Thompson, 1988)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": "6"
},
{
"text": "This system output was added to Wikipedia at http:// en.wikipedia.org/wiki/3-M syndrome on June 26, 2008. The page's history provides examples of changes performed by human editors to articles created by our system.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "In data sets where such mark-up is not available, one can employ topical segmentation algorithms as an additional preprocessing step.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "In addition to the summary itself, we also include proper citations to the sources from which the material is extracted.5 We are continually submitting new articles; however, we report results on those that have at least a 6 month history at time of writing.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [
{
"text": "The authors acknowledge the support of the NSF (CA-REER grant IIS-0448168, grant IIS-0835445, and grant IIS-0835652) and NIH (grant V54LM008748). Thanks to Mike Collins, Julia Hirschberg, and members of the MIT NLP group for their helpful suggestions and comments. Any opinions, findings, conclusions, or recommendations expressed in this paper are those of the authors, and do not necessarily reflect the views of the funding organizations.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Acknowledgments",
"sec_num": null
}
],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "Learning search engine specific query transformations for question answering",
"authors": [
{
"first": "Eugene",
"middle": [],
"last": "Agichtein",
"suffix": ""
},
{
"first": "Steve",
"middle": [],
"last": "Lawrence",
"suffix": ""
},
{
"first": "Luis",
"middle": [],
"last": "Gravano",
"suffix": ""
}
],
"year": 2001,
"venue": "Proceedings of WWW",
"volume": "",
"issue": "",
"pages": "169--178",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Eugene Agichtein, Steve Lawrence, and Luis Gravano. 2001. Learning search engine specific query transformations for question answering. In Proceedings of WWW, pages 169- 178.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Sentence alignment for monolingual comparable corpora",
"authors": [
{
"first": "Regina",
"middle": [],
"last": "Barzilay",
"suffix": ""
},
{
"first": "Noemie",
"middle": [],
"last": "Elhadad",
"suffix": ""
}
],
"year": 2003,
"venue": "Proceedings of EMNLP",
"volume": "",
"issue": "",
"pages": "25--32",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Regina Barzilay and Noemie Elhadad. 2003. Sentence align- ment for monolingual comparable corpora. In Proceed- ings of EMNLP, pages 25-32.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "Catching the drift: Probabilistic content models, with applications to generation and summarization",
"authors": [
{
"first": "Regina",
"middle": [],
"last": "Barzilay",
"suffix": ""
},
{
"first": "Lillian",
"middle": [],
"last": "Lee",
"suffix": ""
}
],
"year": 2004,
"venue": "Proceedings of HLT-NAACL",
"volume": "",
"issue": "",
"pages": "113--120",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Regina Barzilay and Lillian Lee. 2004. Catching the drift: Probabilistic content models, with applications to genera- tion and summarization. In Proceedings of HLT-NAACL, pages 113-120.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Information fusion in the context of multidocument summarization",
"authors": [
{
"first": "Regina",
"middle": [],
"last": "Barzilay",
"suffix": ""
},
{
"first": "Kathleen",
"middle": [
"R"
],
"last": "Mckeown",
"suffix": ""
},
{
"first": "Michael",
"middle": [],
"last": "Elhadad",
"suffix": ""
}
],
"year": 1999,
"venue": "Proceedings of ACL",
"volume": "",
"issue": "",
"pages": "550--557",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Regina Barzilay, Kathleen R. McKeown, and Michael El- hadad. 1999. Information fusion in the context of multi- document summarization. In Proceedings of ACL, pages 550-557.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "An unsupervised approach to biography production using wikipedia",
"authors": [
{
"first": "Fadi",
"middle": [],
"last": "Biadsy",
"suffix": ""
},
{
"first": "Julia",
"middle": [],
"last": "Hirschberg",
"suffix": ""
},
{
"first": "Elena",
"middle": [],
"last": "Filatova",
"suffix": ""
}
],
"year": 2008,
"venue": "Proceedings of ACL/HLT",
"volume": "",
"issue": "",
"pages": "807--815",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Fadi Biadsy, Julia Hirschberg, and Elena Filatova. 2008. An unsupervised approach to biography production using wikipedia. In Proceedings of ACL/HLT, pages 807-815.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "Modelling compression with discourse constraints",
"authors": [
{
"first": "James",
"middle": [],
"last": "Clarke",
"suffix": ""
},
{
"first": "Mirella",
"middle": [],
"last": "Lapata",
"suffix": ""
}
],
"year": 2007,
"venue": "Proceedings of EMNLP-CoNLL",
"volume": "",
"issue": "",
"pages": "1--11",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "James Clarke and Mirella Lapata. 2007. Modelling com- pression with discourse constraints. In Proceedings of EMNLP-CoNLL, pages 1-11.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "Learning to order things",
"authors": [
{
"first": "William",
"middle": [
"W"
],
"last": "Cohen",
"suffix": ""
},
{
"first": "Robert",
"middle": [
"E"
],
"last": "Schapire",
"suffix": ""
},
{
"first": "Yoram",
"middle": [],
"last": "Singer",
"suffix": ""
}
],
"year": 1998,
"venue": "Proceedings of NIPS",
"volume": "",
"issue": "",
"pages": "451--457",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "William W. Cohen, Robert E. Schapire, and Yoram Singer. 1998. Learning to order things. In Proceedings of NIPS, pages 451-457.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "Ranking algorithms for named-entity extraction: Boosting and the voted perceptron",
"authors": [
{
"first": "Michael",
"middle": [],
"last": "Collins",
"suffix": ""
}
],
"year": 2002,
"venue": "Proceedings of ACL",
"volume": "",
"issue": "",
"pages": "489--496",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Michael Collins. 2002. Ranking algorithms for named-entity extraction: Boosting and the voted perceptron. In Pro- ceedings of ACL, pages 489-496.",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "Intoduction to Algorithms",
"authors": [
{
"first": "H",
"middle": [],
"last": "Thomas",
"suffix": ""
},
{
"first": "Charles",
"middle": [
"E"
],
"last": "Cormen",
"suffix": ""
},
{
"first": "Ronald",
"middle": [
"L"
],
"last": "Leiserson",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Rivest",
"suffix": ""
}
],
"year": 1992,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Thomas H. Cormen, Charles E. Leiserson, and Ronald L. Rivest. 1992. Intoduction to Algorithms. The MIT Press.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "A large-scale exploration of effective global features for a joint entity detection and tracking model",
"authors": [
{
"first": "Hal",
"middle": [],
"last": "Daum\u00e9",
"suffix": ""
},
{
"first": "Iii",
"middle": [],
"last": "",
"suffix": ""
},
{
"first": "Daniel",
"middle": [],
"last": "Marcu",
"suffix": ""
}
],
"year": 2005,
"venue": "Proceedings of HLT/EMNLP",
"volume": "",
"issue": "",
"pages": "97--104",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Hal Daum\u00e9 III and Daniel Marcu. 2005. A large-scale explo- ration of effective global features for a joint entity detec- tion and tracking model. In Proceedings of HLT/EMNLP, pages 97-104.",
"links": null
},
"BIBREF10": {
"ref_id": "b10",
"title": "Answering clinical questions with knowledge-based and statistical techniques",
"authors": [
{
"first": "Dina",
"middle": [],
"last": "Demner-Fushman",
"suffix": ""
},
{
"first": "Jimmy",
"middle": [],
"last": "Lin",
"suffix": ""
}
],
"year": 2007,
"venue": "Computational Linguistics",
"volume": "33",
"issue": "1",
"pages": "63--103",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Dina Demner-Fushman and Jimmy Lin. 2007. Answer- ing clinical questions with knowledge-based and statisti- cal techniques. Computational Linguistics, 33(1):63-103.",
"links": null
},
"BIBREF11": {
"ref_id": "b11",
"title": "Tell me what you do and I'll tell you what you are: Learning occupationrelated activities for biographies",
"authors": [
{
"first": "Elena",
"middle": [],
"last": "Filatova",
"suffix": ""
},
{
"first": "John",
"middle": [
"M"
],
"last": "Prager",
"suffix": ""
}
],
"year": 2005,
"venue": "Proceedings of HLT/EMNLP",
"volume": "",
"issue": "",
"pages": "113--120",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Elena Filatova and John M. Prager. 2005. Tell me what you do and I'll tell you what you are: Learning occupation- related activities for biographies. In Proceedings of HLT/EMNLP, pages 113-120.",
"links": null
},
"BIBREF12": {
"ref_id": "b12",
"title": "Automatic creation of domain templates",
"authors": [
{
"first": "Elena",
"middle": [],
"last": "Filatova",
"suffix": ""
},
{
"first": "Vasileios",
"middle": [],
"last": "Hatzivassiloglou",
"suffix": ""
},
{
"first": "Kathleen",
"middle": [],
"last": "Mckeown",
"suffix": ""
}
],
"year": 2006,
"venue": "Proceedings of ACL",
"volume": "",
"issue": "",
"pages": "207--214",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Elena Filatova, Vasileios Hatzivassiloglou, and Kathleen McKeown. 2006. Automatic creation of domain tem- plates. In Proceedings of ACL, pages 207-214.",
"links": null
},
"BIBREF13": {
"ref_id": "b13",
"title": "Summarizing encyclopedic term descriptions on the web",
"authors": [
{
"first": "Atsushi",
"middle": [],
"last": "Fujii",
"suffix": ""
},
{
"first": "Tetsuya",
"middle": [],
"last": "Ishikawa",
"suffix": ""
}
],
"year": 2004,
"venue": "Proceedings of COLING",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Atsushi Fujii and Tetsuya Ishikawa. 2004. Summarizing en- cyclopedic term descriptions on the web. In Proceedings of COLING, page 645.",
"links": null
},
"BIBREF14": {
"ref_id": "b14",
"title": "Multi-document summarization by sentence extraction",
"authors": [
{
"first": "Jade",
"middle": [],
"last": "Goldstein",
"suffix": ""
},
{
"first": "Vibhu",
"middle": [],
"last": "Mittal",
"suffix": ""
},
{
"first": "Jaime",
"middle": [],
"last": "Carbonell",
"suffix": ""
},
{
"first": "Mark",
"middle": [],
"last": "Kantrowitz",
"suffix": ""
}
],
"year": 2000,
"venue": "Proceedings of NAACL-ANLP",
"volume": "",
"issue": "",
"pages": "40--48",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Jade Goldstein, Vibhu Mittal, Jaime Carbonell, and Mark Kantrowitz. 2000. Multi-document summarization by sentence extraction. In Proceedings of NAACL-ANLP, pages 40-48.",
"links": null
},
"BIBREF15": {
"ref_id": "b15",
"title": "Multi-paragraph segmentation of expository text",
"authors": [
{
"first": "Marti",
"middle": [
"A"
],
"last": "Hearst",
"suffix": ""
}
],
"year": 1994,
"venue": "Proceedings of ACL",
"volume": "",
"issue": "",
"pages": "9--16",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Marti A. Hearst. 1994. Multi-paragraph segmentation of ex- pository text. In Proceedings of ACL, pages 9-16.",
"links": null
},
"BIBREF16": {
"ref_id": "b16",
"title": "ROUGE: A package for automatic evaluation of summaries",
"authors": [
{
"first": "Chin-Yew",
"middle": [],
"last": "Lin",
"suffix": ""
}
],
"year": 2004,
"venue": "Proceedings of ACL",
"volume": "",
"issue": "",
"pages": "74--81",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Chin-Yew Lin. 2004. ROUGE: A package for automatic evaluation of summaries. In Proceedings of ACL, pages 74-81.",
"links": null
},
"BIBREF17": {
"ref_id": "b17",
"title": "Advances in Automatic Text Summarization",
"authors": [
{
"first": "Inderjeet",
"middle": [],
"last": "Mani",
"suffix": ""
},
{
"first": "Mark",
"middle": [
"T"
],
"last": "Maybury",
"suffix": ""
}
],
"year": 1999,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Inderjeet Mani and Mark T. Maybury. 1999. Advances in Automatic Text Summarization. The MIT Press.",
"links": null
},
"BIBREF18": {
"ref_id": "b18",
"title": "Rhetorical structure theory: Toward a functional theory of text organization",
"authors": [
{
"first": "C",
"middle": [],
"last": "William",
"suffix": ""
},
{
"first": "Sandra",
"middle": [
"A"
],
"last": "Mann",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Thompson",
"suffix": ""
}
],
"year": 1988,
"venue": "Text",
"volume": "8",
"issue": "3",
"pages": "243--281",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "William C. Mann and Sandra A. Thompson. 1988. Rhetor- ical structure theory: Toward a functional theory of text organization. Text, 8(3):243-281.",
"links": null
},
"BIBREF19": {
"ref_id": "b19",
"title": "Beyond the pipeline: Discrete optimization in NLP",
"authors": [
{
"first": "Tomasz",
"middle": [],
"last": "Marciniak",
"suffix": ""
},
{
"first": "Michael",
"middle": [],
"last": "Strube",
"suffix": ""
}
],
"year": 2005,
"venue": "Proceedings of CoNLL",
"volume": "",
"issue": "",
"pages": "136--143",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Tomasz Marciniak and Michael Strube. 2005. Beyond the pipeline: Discrete optimization in NLP. In Proceedings of CoNLL, pages 136-143.",
"links": null
},
"BIBREF20": {
"ref_id": "b20",
"title": "A study of global inference algorithms in multi-document summarization",
"authors": [
{
"first": "Ryan",
"middle": [],
"last": "Mcdonald",
"suffix": ""
}
],
"year": 2007,
"venue": "Proceedings of EICR",
"volume": "",
"issue": "",
"pages": "557--564",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Ryan McDonald. 2007. A study of global inference algo- rithms in multi-document summarization. In Proceedings of EICR, pages 557-564.",
"links": null
},
"BIBREF21": {
"ref_id": "b21",
"title": "Decoding wikipedia categories for knowledge acquisition",
"authors": [
{
"first": "Vivi",
"middle": [],
"last": "Nastase",
"suffix": ""
},
{
"first": "Michael",
"middle": [],
"last": "Strube",
"suffix": ""
}
],
"year": 2008,
"venue": "Proceedings of AAAI",
"volume": "",
"issue": "",
"pages": "1219--1224",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Vivi Nastase and Michael Strube. 2008. Decoding wikipedia categories for knowledge acquisition. In Proceedings of AAAI, pages 1219-1224.",
"links": null
},
"BIBREF22": {
"ref_id": "b22",
"title": "Topic-driven multi-document summarization with encyclopedic knowledge and spreading activation",
"authors": [
{
"first": "Vivi",
"middle": [],
"last": "Nastase",
"suffix": ""
}
],
"year": 2008,
"venue": "Proceedings of EMNLP",
"volume": "",
"issue": "",
"pages": "763--772",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Vivi Nastase. 2008. Topic-driven multi-document summa- rization with encyclopedic knowledge and spreading acti- vation. In Proceedings of EMNLP, pages 763-772.",
"links": null
},
"BIBREF23": {
"ref_id": "b23",
"title": "Centroid-based summarization of multiple documents: sentence extraction, utilitybased evaluation, and user studies",
"authors": [
{
"first": "R",
"middle": [],
"last": "Dragomir",
"suffix": ""
},
{
"first": "Hongyan",
"middle": [],
"last": "Radev",
"suffix": ""
},
{
"first": "Malgorzata",
"middle": [],
"last": "Jing",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Budzikowska",
"suffix": ""
}
],
"year": 2000,
"venue": "Proceedings of ANLP/NAACL",
"volume": "",
"issue": "",
"pages": "21--29",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Dragomir R. Radev, Hongyan Jing, and Malgorzata Budzikowska. 2000. Centroid-based summarization of multiple documents: sentence extraction, utility- based evaluation, and user studies. In Proceedings of ANLP/NAACL, pages 21-29.",
"links": null
},
"BIBREF24": {
"ref_id": "b24",
"title": "Building Natural Language Generation Systems",
"authors": [
{
"first": "Ehud",
"middle": [],
"last": "Reiter",
"suffix": ""
},
{
"first": "Robert",
"middle": [],
"last": "Dale",
"suffix": ""
}
],
"year": 2000,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Ehud Reiter and Robert Dale. 2000. Building Natural Lan- guage Generation Systems. Cambridge University Press, Cambridge.",
"links": null
},
"BIBREF25": {
"ref_id": "b25",
"title": "Multiple aspect ranking using the good grief algorithm",
"authors": [
{
"first": "Benjamin",
"middle": [],
"last": "Snyder",
"suffix": ""
},
{
"first": "Regina",
"middle": [],
"last": "Barzilay",
"suffix": ""
}
],
"year": 2007,
"venue": "Proceedings of HLT-NAACL",
"volume": "",
"issue": "",
"pages": "300--307",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Benjamin Snyder and Regina Barzilay. 2007. Multiple as- pect ranking using the good grief algorithm. In Proceed- ings of HLT-NAACL, pages 300-307.",
"links": null
},
"BIBREF26": {
"ref_id": "b26",
"title": "A hybrid approach to answering biographical questions",
"authors": [
{
"first": "Ralph",
"middle": [
"M"
],
"last": "Weischedel",
"suffix": ""
},
{
"first": "Jinxi",
"middle": [],
"last": "Xu",
"suffix": ""
},
{
"first": "Ana",
"middle": [],
"last": "Licuanan",
"suffix": ""
}
],
"year": 2004,
"venue": "New Directions in Question Answering",
"volume": "",
"issue": "",
"pages": "59--70",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Ralph M. Weischedel, Jinxi Xu, and Ana Licuanan. 2004. A hybrid approach to answering biographical questions. In New Directions in Question Answering, pages 59-70.",
"links": null
},
"BIBREF27": {
"ref_id": "b27",
"title": "Autonomously semantifying wikipedia",
"authors": [
{
"first": "Fei",
"middle": [],
"last": "Wu",
"suffix": ""
},
{
"first": "Daniel",
"middle": [
"S"
],
"last": "Weld",
"suffix": ""
}
],
"year": 2007,
"venue": "Proceedings of CIKM",
"volume": "",
"issue": "",
"pages": "41--50",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Fei Wu and Daniel S. Weld. 2007. Autonomously semanti- fying wikipedia. In Proceedings of CIKM, pages 41-50.",
"links": null
},
"BIBREF28": {
"ref_id": "b28",
"title": "Hierarchical clustering algorithms for document datasets",
"authors": [
{
"first": "Ying",
"middle": [],
"last": "Zhao",
"suffix": ""
},
{
"first": "George",
"middle": [],
"last": "Karypis",
"suffix": ""
},
{
"first": "Usama",
"middle": [],
"last": "Fayyad",
"suffix": ""
}
],
"year": 2005,
"venue": "Data Mining and Knowledge Discovery",
"volume": "10",
"issue": "2",
"pages": "141--168",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Ying Zhao, George Karypis, and Usama Fayyad. 2005. Hierarchical clustering algorithms for document datasets. Data Mining and Knowledge Discovery, 10(2):141-168.",
"links": null
},
"BIBREF29": {
"ref_id": "b29",
"title": "Multidocument biography summarization",
"authors": [
{
"first": "L",
"middle": [],
"last": "Zhou",
"suffix": ""
},
{
"first": "M",
"middle": [],
"last": "Ticrea",
"suffix": ""
},
{
"first": "Eduard",
"middle": [],
"last": "Hovy",
"suffix": ""
}
],
"year": 2004,
"venue": "Proceedings of EMNLP",
"volume": "",
"issue": "",
"pages": "434--441",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "L. Zhou, M. Ticrea, and Eduard Hovy. 2004. Multi- document biography summarization. In Proceedings of EMNLP, pages 434-441.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"type_str": "figure",
"num": null,
"uris": null,
"text": "A fragment from the automatically created article for 3-M Syndrome."
},
"FIGREF1": {
"type_str": "figure",
"num": null,
"uris": null,
"text": "For j = 1 . . . k 2 Set parameters wj = 0 Training: Repeat until convergence or while iter < itermax: 4 For i = 1 . . . n 5 For j = 1 . . . k 6 Rank(eij1 . . . eijr, wj) 7 x1 . . . x k = Optimize(ei11 . . . e ikr ) 8 For j = 1 . . . k If sim(xj, sij) < 0.8 10 wj = wj + \u03c6(sij) \u2212 \u03c6(xi) 11 iter = iter + 1 12 Return parameters w1 . . . w k"
},
"FIGREF2": {
"type_str": "figure",
"num": null,
"uris": null,
"text": "An algorithm for learning several ranking problems with a joint decoding mechanism. tor reaction to system-produced articles submitted to Wikipedia."
},
"TABREF1": {
"num": null,
"text": "Features employed in the ranking model.",
"type_str": "table",
"html": null,
"content": "<table/>"
},
"TABREF3": {
"num": null,
"text": "Average number of excerpts selected and sources used in article creation for test articles.",
"type_str": "table",
"html": null,
"content": "<table/>"
},
"TABREF5": {
"num": null,
"text": "Results of ROUGE-1 evaluation. * Significant with respect to our full model for p \u2264 0.05.",
"type_str": "table",
"html": null,
"content": "<table/>"
},
"TABREF7": {
"num": null,
"text": "Distribution of edits on Wikipedia.",
"type_str": "table",
"html": null,
"content": "<table/>"
}
}
}
} |