ACL-OCL / Base_JSON /prefixJ /json /J11 /J11-1001.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "J11-1001",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T03:06:20.518290Z"
},
"title": "Nouveau-ROUGE: A Novelty Metric for Update Summarization",
"authors": [
{
"first": "John",
"middle": [
"M"
],
"last": "Conroy",
"suffix": "",
"affiliation": {},
"email": "conroy@super.org."
},
{
"first": "Judith",
"middle": [
"D"
],
"last": "Schlesinger",
"suffix": "",
"affiliation": {},
"email": "judith@super.org."
},
{
"first": "Dianne",
"middle": [
"P"
],
"last": "O'leary",
"suffix": "",
"affiliation": {},
"email": "oleary@cs.umd.edu"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "An update summary should provide a fluent summarization of new information on a timeevolving topic, assuming that the reader has already reviewed older documents or summaries. In 2007 and 2008, an annual summarization evaluation included an update summarization task. Several participating systems produced update summaries indistinguishable from humangenerated summaries when measured using ROUGE. However, no machine system performed near human-level performance in manual evaluations such as pyramid and overall responsiveness scoring. We present a metric called Nouveau-ROUGE that improves correlation with manual evaluation metrics and can be used to predict both the pyramid score and overall responsiveness for update summaries. Nouveau-ROUGE can serve as a less expensive surrogate for manual evaluations when comparing existing systems and when developing new ones.",
"pdf_parse": {
"paper_id": "J11-1001",
"_pdf_hash": "",
"abstract": [
{
"text": "An update summary should provide a fluent summarization of new information on a timeevolving topic, assuming that the reader has already reviewed older documents or summaries. In 2007 and 2008, an annual summarization evaluation included an update summarization task. Several participating systems produced update summaries indistinguishable from humangenerated summaries when measured using ROUGE. However, no machine system performed near human-level performance in manual evaluations such as pyramid and overall responsiveness scoring. We present a metric called Nouveau-ROUGE that improves correlation with manual evaluation metrics and can be used to predict both the pyramid score and overall responsiveness for update summaries. Nouveau-ROUGE can serve as a less expensive surrogate for manual evaluations when comparing existing systems and when developing new ones.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "Update summaries focus on what is new relative to a previous body of information. They pose new challenges both to algorithm developers and to evaluation of summaries. In 2007, DUC (Document Understanding Conference) introduced an update summarization task, repeated in 2008 for TAC (Text Analysis Conference). 1 This task consisted of producing a multi-document summary for a set of articles on a single topic, followed by one (2008) or two (2007) multi-document summaries for sets of articles on the same topic published at later dates. The goal was to generate a good first summary, along with update(s) that contained new content and minimized redundancy.",
"cite_spans": [
{
"start": 428,
"end": 434,
"text": "(2008)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "The modifier manual is used to identify evaluations, and the corresponding scores, produced by humans. The modifier automatic is used to identify evaluations, and the corresponding scores, produced by machines. Similarly, human-generated and machine-generated will be used to distinguish between summaries created by humans and those generated by machine systems, respectively. 2 Because we are working with update summarization, there is a minimum of two summaries for a set of documents. The first summary for the document set is called the original (Task A) summary and a later summary is called an update (Task B) summary.",
"cite_spans": [
{
"start": 378,
"end": 379,
"text": "2",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "Several machine summarizing systems produced update summaries that were statistically indistinguishable from human-generated summaries, as measured by the ROUGE metrics, the standard metrics for automatic evaluation of summaries. However, none of these machine systems performed near human levels in overall responsiveness or pyramid evaluation, the currently used manual evaluation metrics.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "We define the metric gap (or gap) as the distance between a prediction of a manual score, based on automatic scores, and the observed manual score.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "The purpose of our work is to investigate and mitigate this metric gap by introducing an automatic evaluation that is a better predictor of manual evaluation. Reducing the metric gap is important for two reasons. First, the gap severely limits the usefulness of automatic evaluation and forces the use of much more expensive manual evaluation for comparing existing systems. More importantly, however, this gap is a severe handicap to research on new update summarization methods because it makes it difficult to evaluate new ideas and compare them with existing methods.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "In Section 2, we analyze the results of the TAC 2008 summarization task, demonstrating the large gap between ROUGE automatic metrics and manual evaluation of update summaries. In Section 3, we modify ROUGE to produce scores that correlate significantly better with manual evaluation. We evaluate our new metric on TAC 2008 data in Section 4, demonstrating its superiority as a predictor of manual evaluations.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "TAC 2008 presented 48 20-document sets, with 10 documents in each of two subsets, A and B. Subset B documents were more recent. Original summaries were generated for the A subsets and update summaries were then produced for the B subsets.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "State-of-the-Art Evaluation of Update Summaries",
"sec_num": "2."
},
{
"text": "In TAC 2008, ROUGE was used for automatic evaluation. ROUGE (Lin and Hovy 2000) compares any summary to any other (typically human-generated) summary using a recall-oriented approach. ROUGE-1 and -2 are based on unigrams and bigrams, respectively; ROUGE-SU4 uses bigrams with a maximum skip distance of 4 between bigrams; ROUGE-BE (Hovy, Lin, and Zhou 2005) is an n-gram approach based on basic elements, computed via parsing or automatic entity recognition. ROUGE-2, ROUGE-SU4, and ROUGE-BE were the official automatic metrics for TAC 2008, used to compare machine-generated summaries to human-generated summaries, and to compare human-generated summaries to each other using a jackknife approach. In addition to the three official metrics, we include ROUGE-1 in our study as it is often competitive with the official metrics.",
"cite_spans": [
{
"start": 60,
"end": 79,
"text": "(Lin and Hovy 2000)",
"ref_id": "BIBREF6"
},
{
"start": 331,
"end": 357,
"text": "(Hovy, Lin, and Zhou 2005)",
"ref_id": "BIBREF5"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "State-of-the-Art Evaluation of Update Summaries",
"sec_num": "2."
},
{
"text": "Three manual evaluation metrics were used in TAC 2008: pyramid, overall responsiveness, and linguistic quality (not considered in our work). The pyramid method (Nenkova and Passonneau 2004) is a content-based metric for which human annotators mark content units in the human-generated summaries. The content units are collected across a set of human-generated summaries for a topic, and a weight is computed based on how many human-generated summaries include this content unit. TAC 2008 also used a manual overall responsiveness score. After evaluating data from (Dang 2007 Conroy and Dang 2008) , NIST decided that this score, which evaluates summary usefulness including linguistic quality, is a reliable and stable manual evaluation.",
"cite_spans": [
{
"start": 564,
"end": 574,
"text": "(Dang 2007",
"ref_id": "BIBREF3"
},
{
"start": 575,
"end": 596,
"text": "Conroy and Dang 2008)",
"ref_id": "BIBREF0"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "State-of-the-Art Evaluation of Update Summaries",
"sec_num": "2."
},
{
"text": "We analyzed the three official TAC 2008 automatic evaluation scores to see how well they predict the manual evaluation metrics of overall responsiveness and pyramid score. Figure 1 shows scatter plots of responsiveness and pyramid scores vs. the three official ROUGE measures for the TAC 2008 update task. Each solid data point represents the average score for a human summarizer over 24 document sets, and a dashed line marks the minimum; each open data point represents the average score for a machine system. We also show robust linear least squares fits to the data as well as the Pearson correlation coefficients between ROUGE-BE, -2, and -SU4 and the manual-evaluation scores. Surprisingly, these correlations are higher for the update task than for the original summarization task (data not shown); nevertheless, the gap between the lines' predictions and the scores for the human-generated summaries is larger. We report correlation coefficients only for the machine-generated summaries. Scores for the human-generated summaries are distributed differently, and correlation for the set of human-generated summaries is often not significant due to the small number of human summarizers.",
"cite_spans": [],
"ref_spans": [
{
"start": 172,
"end": 180,
"text": "Figure 1",
"ref_id": "FIGREF0"
}
],
"eq_spans": [],
"section": "State-of-the-Art Evaluation of Update Summaries",
"sec_num": "2."
},
{
"text": "The correlation coefficients in Figure 1 show that the automatic metrics do well in predicting responsiveness and pyramid scoring for machine-generated summaries. In contrast, the scores for human-generated summaries far exceed the predictions, with a large gap between predicted and actual scores. As may be expected, ROUGE is more highly correlated with the pyramid evaluation, which is a pure content evaluation score, whereas the responsiveness score also reflects linguistic quality.",
"cite_spans": [],
"ref_spans": [
{
"start": 32,
"end": 40,
"text": "Figure 1",
"ref_id": "FIGREF0"
}
],
"eq_spans": [],
"section": "State-of-the-Art Evaluation of Update Summaries",
"sec_num": "2."
},
{
"text": "We more formally define the metric gap to be the absolute value of the difference between a manual evaluation score and our prediction of it based only on automatic evaluation scores. A number of TAC 2008 machine systems performed within statistical confidence of human performance in the automatic evaluation metrics, but no system performed near human performance in the manual evaluations. This has also been observed in previous summarization evaluations (Conroy and Dang 2008) . Progress has been made in closing this metric gap but it persists, especially for update summaries.",
"cite_spans": [
{
"start": 459,
"end": 481,
"text": "(Conroy and Dang 2008)",
"ref_id": "BIBREF0"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Improving Automatic Evaluation-Nouveau-ROUGE",
"sec_num": "3."
},
{
"text": "A good update summary must contain essential information but focus on new information. When a machine-generated update summary is good, it is similar to the human-generated update summaries. This is assessed quite well by a ROUGE score. But the machine-generated update summary should also be different from the human-generated original summaries, and we need an automatic metric to assess this difference, or lack of redundancy. We suggest using a ROUGE score to measure similarity, and thus redundancy, between a given original summary and an update summary: A high ROUGE score indicates high redundancy.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Improving Automatic Evaluation-Nouveau-ROUGE",
"sec_num": "3."
},
{
"text": "To illustrate this, we used the CLASSY algorithm (Conroy, Schlesinger, and O'Leary 2006; Schlesinger, O'Leary, and Conroy 2008) to produce original summaries and update summaries for the TAC 2008 data. We also produced update summaries using a variant, projected-CLASSY, that reduces overlap by using a linear algebra projection of the term-sentence matrix (Conroy, Schlesinger, and O'Leary 2007) of candidate sentences against the matrix for the original (Task A) summary in order to favor new information. Table 1 gives average ROUGE-2 scores and 95% confidence intervals, computed via bootstrapping (Efron and Tibshirani 1993) , over the 48 document sets. Two scores are given: R (BB) 2 compares each CLASSY update (Task B) summary to the human-generated summaries, and R (AB) 2 compares each to the original (Task A) model summaries. Whereas the two variants score comparably using R",
"cite_spans": [
{
"start": 49,
"end": 88,
"text": "(Conroy, Schlesinger, and O'Leary 2006;",
"ref_id": "BIBREF1"
},
{
"start": 89,
"end": 127,
"text": "Schlesinger, O'Leary, and Conroy 2008)",
"ref_id": "BIBREF7"
},
{
"start": 357,
"end": 396,
"text": "(Conroy, Schlesinger, and O'Leary 2007)",
"ref_id": "BIBREF2"
},
{
"start": 602,
"end": 629,
"text": "(Efron and Tibshirani 1993)",
"ref_id": "BIBREF4"
}
],
"ref_spans": [
{
"start": 508,
"end": 515,
"text": "Table 1",
"ref_id": "TABREF0"
}
],
"eq_spans": [],
"section": "Improving Automatic Evaluation-Nouveau-ROUGE",
"sec_num": "3."
},
{
"text": "(BB) 2",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Improving Automatic Evaluation-Nouveau-ROUGE",
"sec_num": "3."
},
{
"text": ", there is a significant difference in the R (AB) 2 metric, as desired. Given this evidence, we propose predicting manual scores for update summaries by using two ROUGE scores, R",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Improving Automatic Evaluation-Nouveau-ROUGE",
"sec_num": "3."
},
{
"text": "\u03b1 i,0 \u03b1 i,1 \u03b1 i,2 \u03b1 i,0 \u03b1 i,1 \u03b1 i,2 R 1 \u2212",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Improving Automatic Evaluation-Nouveau-ROUGE",
"sec_num": "3."
},
{
"text": "(AB) i and R (BB) i (i = 1, 2, SU4, ...)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Improving Automatic Evaluation-Nouveau-ROUGE",
"sec_num": "3."
},
{
"text": ", in a three-parameter model called Nouveau-ROUGE:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Improving Automatic Evaluation-Nouveau-ROUGE",
"sec_num": "3."
},
{
"text": "N i = \u03b1 i,0 + \u03b1 i,1 R (AB) i + \u03b1 i,2 R (BB) i",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Improving Automatic Evaluation-Nouveau-ROUGE",
"sec_num": "3."
},
{
"text": "We determine the \u03b1 parameters (Table 2) using robust linear regression on the TAC 2008 evaluation data so that the Nouveau-ROUGE score N i best predicts the manual scores of responsiveness and pyramid performance.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Improving Automatic Evaluation-Nouveau-ROUGE",
"sec_num": "3."
},
{
"text": "Nouveau-ROUGE could be used by researchers to predict how a new system would compare with the TAC 2008 systems in overall responsiveness and pyramid scoring, a comparison that up to now has been impossible.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Improving Automatic Evaluation-Nouveau-ROUGE",
"sec_num": "3."
},
{
"text": "We evaluate Nouveau-ROUGE using cross validation studies to demonstrate that if manual scores are available for a subset of summaries (in this case, those from 29 machine systems, half of those that participated in TAC 2008), then Nouveau-ROUGE can predict manual scores for the remaining (held-back) summaries.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Evaluating Nouveau-ROUGE",
"sec_num": "4."
},
{
"text": "Correlation scores between automatic and manual scores have traditionally been used as a measure of the effectiveness of automatic evaluation as a surrogate for manual evaluation. Pearson correlation coefficients, shown in Table 3 , were computed for the scores for the held-back subset of summaries. Correlation is indeed higher for the Nouveau-ROUGE scores than for any of the ROUGE scores.",
"cite_spans": [],
"ref_spans": [
{
"start": 223,
"end": 230,
"text": "Table 3",
"ref_id": null
}
],
"eq_spans": [],
"section": "Improved Correlation with Manual Evaluation",
"sec_num": "4.1"
},
{
"text": "Correlation scores for TAC 2008 human evaluations.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Table 3",
"sec_num": null
},
{
"text": "Average Pyramid Score Figure 2 shows that ROUGE-BE and ROUGE-1 predictions of both responsiveness and pyramid scores are inferior to the Nouveau-ROUGE-BE predictions. Plots for N 2 and N SU4 are omitted due to space restrictions, but performance improvement relative to ROUGE is greater than that for N BE and less than that for N 1 .",
"cite_spans": [],
"ref_spans": [
{
"start": 22,
"end": 30,
"text": "Figure 2",
"ref_id": "FIGREF1"
}
],
"eq_spans": [],
"section": "Average Responsiveness Score",
"sec_num": null
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "Metric i = 1 i = 2 i = SU4 i = BE i = 1 i = 2 i = SU4 i = BE R",
"eq_num": "(AB"
}
],
"section": "Average Responsiveness Score",
"sec_num": null
},
{
"text": "Metric R (AB) i R (BB) i N i p-value R (AB) i R (BB) i N i p-",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Average Responsiveness Score",
"sec_num": null
},
{
"text": "To show that our results are not due to a lucky partitioning of the data, we used bootstrapping (Efron and Tibshirani 1993) , a resampling method, which allows us to compute our statistical confidence in the results. This model assumes that observed data (scores for the 58 systems) characterize all data. Given this model, the proper sampling method is to choose subsets with replacement. We chose 58 systems (with replacement) and used half to determine the Nouveau-ROUGE parameters and half to test the model. We repeated this process 1,000 times. Table 4 gives the correlation coefficients (for the tested-half of the data) for all four ROUGE metrics with each of the manual evaluations when comparing the machine-generated summaries with the human-generated summaries. Data in the columns labeled \"p-value\" result from a Mann-Whitney Utest for equal medians of R (BB) i and N i of the distributions of correlations returned by the bootstrapping procedure. Because all p-values are small, we can conclude that the differences between the R (BB) i and N i correlation scores are statistically significant for all variants of ROUGE. Table 5 gives the median gap on the TAC 2008 data for predicting responsiveness and pyramid scores. Recall that the gap is the absolute value of the difference between the manual score and the prediction of it. The median gap is always smaller for Nouveau-ROUGE than for ROUGE; in fact, the gap is smaller on every trial. We used the Wilcox sign test to test the significance of this observation. The null hypothesis is that the differences in the gaps between ROUGE and Nouveau-ROUGE has median 0. The p-values from the Wilcox test indicate that the null hypothesis is true with probability 1 128 \u2248 7.812 \u00d7 10 \u22123 , so it can be safely rejected.",
"cite_spans": [
{
"start": 96,
"end": 123,
"text": "(Efron and Tibshirani 1993)",
"ref_id": "BIBREF4"
}
],
"ref_spans": [
{
"start": 551,
"end": 558,
"text": "Table 4",
"ref_id": "TABREF4"
},
{
"start": 1135,
"end": 1142,
"text": "Table 5",
"ref_id": "TABREF6"
}
],
"eq_spans": [],
"section": "Validation Using Bootstrapping Experiments",
"sec_num": "4.2"
},
{
"text": "Our new metric, Nouveau-ROUGE, includes a measure of novelty for an update summary. We demonstrated that it has higher correlation to manual evaluation of overall responsiveness and to pyramid scores than does ROUGE. The most obvious deficiency in ROUGE is the lack of a linguistic quality measurement, which we take to encompass all language-related issues: lexical, syntactic, and semantic. Therefore, we conjecture that most remaining prediction error in Nouveau-ROUGE is a result of omitting linguistic quality and caution that better prediction would be achieved only for systems of comparable linguistic quality. We believe that responsiveness is an imperfect surrogate for task-based summary evaluation such as that done in SUMMAC (Mani et al. 1999) . We would welcome a return to task-based evaluation, as well as research increasing the reliability and consistency of manual evaluation metrics. Investigation could also quantify the impact of low responsiveness and pyramid scores on the ability to perform a specific task.",
"cite_spans": [
{
"start": 731,
"end": 756,
"text": "SUMMAC (Mani et al. 1999)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": "5."
},
{
"text": "NIST uses \"model\" for human-generated summaries and \"peer\" for machine-generated summaries.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "Mind the gap: Dangers of divorcing evaluations of summary content from linguistic quality",
"authors": [
{
"first": "John",
"middle": [
"M"
],
"last": "Conroy",
"suffix": ""
},
{
"first": "Hoa Trang",
"middle": [],
"last": "Dang",
"suffix": ""
}
],
"year": 2008,
"venue": "Proceedings of the 22nd International Conference on Computational Linguistics",
"volume": "",
"issue": "",
"pages": "145--152",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Conroy, John M. and Hoa Trang Dang. 2008. Mind the gap: Dangers of divorcing evaluations of summary content from linguistic quality. In Proceedings of the 22nd International Conference on Computational Linguistics (Coling 2008), pages 145-152, Manchester.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Topic-focused multi-document summarization using an approximate oracle score",
"authors": [
{
"first": "John",
"middle": [
"M"
],
"last": "Conroy",
"suffix": ""
},
{
"first": "Judith",
"middle": [
"D"
],
"last": "Schlesinger",
"suffix": ""
},
{
"first": "Dianne",
"middle": [
"P"
],
"last": "O'leary",
"suffix": ""
}
],
"year": 2006,
"venue": "Proceedings of the COLING/ACL 2006 Main Conference Poster Sessions",
"volume": "",
"issue": "",
"pages": "152--159",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Conroy, John M., Judith D. Schlesinger, and Dianne P. O'Leary. 2006. Topic-focused multi-document summarization using an approximate oracle score. In Proceedings of the COLING/ACL 2006 Main Conference Poster Sessions, pages 152-159, Sydney.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "CLASSY 2007 at DUC",
"authors": [
{
"first": "John",
"middle": [
"M"
],
"last": "Conroy",
"suffix": ""
},
{
"first": "Judith",
"middle": [
"D"
],
"last": "Schlesinger",
"suffix": ""
},
{
"first": "Dianne",
"middle": [
"P"
],
"last": "O'leary",
"suffix": ""
}
],
"year": 2007,
"venue": "Proceedings of the Seventh Document Understanding Conference (DUC)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Conroy, John M., Judith D. Schlesinger, and Dianne P. O'Leary. 2007. CLASSY 2007 at DUC 2007. In Proceedings of the Seventh Document Understanding Conference (DUC), Rochester, NY. Available at http://duc. nist.gov/pubs.html#2007.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Overview of DUC",
"authors": [
{
"first": "Hoa",
"middle": [],
"last": "Dang",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Trang",
"suffix": ""
}
],
"year": 2005,
"venue": "Proceedings of the Seventh Document Understanding Conference (DUC)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Dang, Hoa Trang. 2007. Overview of DUC 2007. In Proceedings of the Seventh Document Understanding Conference (DUC), Rochester, NY. Available at http://duc.nist.gov/ pubs.html#2005.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "An Introduction to the Bootstrap",
"authors": [
{
"first": "B",
"middle": [],
"last": "Efron",
"suffix": ""
},
{
"first": "R",
"middle": [
"J"
],
"last": "Tibshirani",
"suffix": ""
}
],
"year": 1993,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Efron, B. and R. J. Tibshirani. 1993. An Introduction to the Bootstrap. Chapman & Hall, New York.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "Evaluating DUC 2005 using basic elements",
"authors": [
{
"first": "Eduard",
"middle": [],
"last": "Hovy",
"suffix": ""
},
{
"first": "Chin-Yew",
"middle": [],
"last": "Lin",
"suffix": ""
},
{
"first": "Liang",
"middle": [],
"last": "Zhou",
"suffix": ""
}
],
"year": 2005,
"venue": "Proceedings of the Fifth Document Understanding Conference (DUC)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Hovy, Eduard, Chin-Yew Lin, and Liang Zhou. 2005. Evaluating DUC 2005 using basic elements. In Proceedings of the Fifth Document Understanding Conference (DUC), Vancouver. Available at www-nlpir/ nist.gov/projects/duc/pubs.html.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "The automated acquisition of topic signatures for text summarization",
"authors": [
{
"first": "Chin-Yew",
"middle": [],
"last": "Lin",
"suffix": ""
},
{
"first": "Eduard",
"middle": [],
"last": "Hovy ; Morristown",
"suffix": ""
},
{
"first": "N",
"middle": [
"J"
],
"last": "Mani",
"suffix": ""
},
{
"first": "Therese",
"middle": [],
"last": "Inderjeet",
"suffix": ""
},
{
"first": "David",
"middle": [],
"last": "Firmin",
"suffix": ""
},
{
"first": "Gary",
"middle": [],
"last": "House",
"suffix": ""
},
{
"first": "Beth",
"middle": [],
"last": "Klein",
"suffix": ""
},
{
"first": "Lynette",
"middle": [],
"last": "Sundheim",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Hirschman",
"suffix": ""
}
],
"year": 1999,
"venue": "Proceedings of the Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics",
"volume": "",
"issue": "",
"pages": "145--152",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Lin, Chin-Yew and Eduard Hovy. 2000. The automated acquisition of topic signatures for text summarization. In Proceedings of the 18th Conference on Computational Linguistics, pages 495-501, Morristown, NJ. Mani, Inderjeet, Therese Firmin, David House, Gary Klein, Beth Sundheim, and Lynette Hirschman. 1999. The TIPSTER SUMMAC text summarization evaluation. In Proceedings of EACL'99: Ninth Conference of the European Chapter of the Association for Computational Linguistics, pages 77-85, Bergen. Nenkova, Ani and Rebecca Passonneau. 2004. Evaluating content selection in summarization: The pyramid method. In Proceedings of the Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics, pages 145-152, Boston, MA.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "Arabic/ English multi-document summarization with CLASSY-The past and the future",
"authors": [
{
"first": "Judith",
"middle": [
"D"
],
"last": "Schlesinger",
"suffix": ""
},
{
"first": "P",
"middle": [],
"last": "Dianne",
"suffix": ""
},
{
"first": "John",
"middle": [
"M"
],
"last": "O'leary",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Conroy",
"suffix": ""
}
],
"year": 2008,
"venue": "Lecture Notes in Computer Science",
"volume": "4919",
"issue": "",
"pages": "568--581",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Schlesinger, Judith D., Dianne P. O'Leary, and John M. Conroy. 2008. Arabic/ English multi-document summarization with CLASSY-The past and the future. In Alexander F. Gelbukh, editor, CICLing, volume 4919 of Lecture Notes in Computer Science. Springer, Haifa, pages 568-581.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"uris": null,
"type_str": "figure",
"num": null,
"text": "TAC 2008: The update task (Task B) responsiveness and pyramid scores vs. ROUGE scores; human-generated summaries (solid points) and machine-generated summaries (open points)."
},
"FIGREF1": {
"uris": null,
"type_str": "figure",
"num": null,
"text": "TAC 2008: ROUGE and Nouveau-ROUGE responsiveness and pyramid predictions for subtask B."
},
"TABREF0": {
"content": "<table><tr><td>Variation</td><td>R (BB) 2</td><td>R (AB) 2</td></tr><tr><td colspan=\"3\">projected-CLASSY 0.087 (0.080, 0.094) 0.075 (0.070, 0.079)</td></tr><tr><td>CLASSY</td><td colspan=\"2\">0.089 (0.082, 0.096) 0.083 (0.078, 0.088)</td></tr></table>",
"type_str": "table",
"num": null,
"html": null,
"text": "TAC 2008: Average ROUGE-2 scores and 95% confidence intervals for update summaries produced by two variants of CLASSY."
},
"TABREF1": {
"content": "<table><tr><td>Predicting Responsiveness</td><td>Predicting Pyramid Scores</td></tr></table>",
"type_str": "table",
"num": null,
"html": null,
"text": "TAC 2008: Nouveau-ROUGE \u03b1-parameters."
},
"TABREF4": {
"content": "<table><tr><td>Average Responsiveness Score</td><td>Jackknife Pyramid Score</td></tr></table>",
"type_str": "table",
"num": null,
"html": null,
"text": "TAC 2008: Median Pearson correlation coefficients for automatic vs. manual evaluations."
},
"TABREF6": {
"content": "<table><tr><td/><td colspan=\"3\">Responsiveness Metric Gap</td><td colspan=\"3\">Pyramid Metric Gap</td></tr><tr><td colspan=\"3\">Metric R Gap N Gap</td><td>p-value</td><td colspan=\"3\">R Gap N Gap p-value</td></tr><tr><td>R 1</td><td>2.025</td><td>1.277</td><td>7.8e-03</td><td>0.285</td><td>0.187</td><td>7.8e-03</td></tr><tr><td>R 2</td><td>1.655</td><td>1.518</td><td>7.8e-03</td><td>0.241</td><td>0.197</td><td>7.8e-03</td></tr><tr><td>R SU4 R BE</td><td>1.887 1.591</td><td>1.344 1.547</td><td>7.8e-03 7.8e-03</td><td>0.273 0.229</td><td>0.206 0.206</td><td>7.8e-03 7.8e-03</td></tr></table>",
"type_str": "table",
"num": null,
"html": null,
"text": "Narrowing the TAC 2008 metric gap."
}
}
}
}