| { |
| "paper_id": "D15-1013", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T16:28:37.515195Z" |
| }, |
| "title": "Re-evaluating Automatic Summarization with BLEU and 192 Shades of ROUGE", |
| "authors": [ |
| { |
| "first": "Yvette", |
| "middle": [], |
| "last": "Graham", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "graham.yvette@gmail.com" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "We provide an analysis of current evaluation methodologies applied to summarization metrics and identify the following areas of concern: (1) movement away from evaluation by correlation with human assessment; (2) omission of important components of human assessment from evaluations, in addition to large numbers of metric variants; (3) absence of methods of significance testing improvements over a baseline. We outline an evaluation methodology that overcomes all such challenges, providing the first method of significance testing suitable for evaluation of summarization metrics. Our evaluation reveals for the first time which metric variants significantly outperform others, optimal metric variants distinct from current recommended best variants, as well as machine translation metric BLEU to have performance on-par with ROUGE for the purpose of evaluation of summarization systems. We subsequently replicate a recent large-scale evaluation that relied on, what we now know to be, suboptimal ROUGE variants revealing distinct conclusions about the relative performance of state-of-the-art summarization systems.", |
| "pdf_parse": { |
| "paper_id": "D15-1013", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "We provide an analysis of current evaluation methodologies applied to summarization metrics and identify the following areas of concern: (1) movement away from evaluation by correlation with human assessment; (2) omission of important components of human assessment from evaluations, in addition to large numbers of metric variants; (3) absence of methods of significance testing improvements over a baseline. We outline an evaluation methodology that overcomes all such challenges, providing the first method of significance testing suitable for evaluation of summarization metrics. Our evaluation reveals for the first time which metric variants significantly outperform others, optimal metric variants distinct from current recommended best variants, as well as machine translation metric BLEU to have performance on-par with ROUGE for the purpose of evaluation of summarization systems. We subsequently replicate a recent large-scale evaluation that relied on, what we now know to be, suboptimal ROUGE variants revealing distinct conclusions about the relative performance of state-of-the-art summarization systems.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Automatic metrics of summarization evaluation have their origins in machine translation (MT), with ROUGE (Lin and Hovy, 2003) , the first and still most widely used automatic summarization metric, comprising an adaption of the BLEU score (Papineni et al., 2002) . Automatic evaluation in MT and summarization have much in common, as both involve the automatic comparison of systemgenerated texts with one or more human-generated reference texts, contrasting either system-output translations or peer summaries with human reference translations or model summaries, depending on the task. In both MT and summarization evaluation, any newly proposed automatic metric must be assessed by the degree to which it provides a good substitute of human assessment, and although there are obvious parallels between evaluation of systems in the two areas, when it comes to evaluation of metrics, summarization has diverged considerably from methodologies applied to evaluation of metrics in MT.", |
| "cite_spans": [ |
| { |
| "start": 105, |
| "end": 125, |
| "text": "(Lin and Hovy, 2003)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 238, |
| "end": 261, |
| "text": "(Papineni et al., 2002)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Since the inception of BLEU, evaluation of automatic metrics in MT has been by correlation with human assessment. In contrast in summarization, over the years since the introduction of ROUGE, summarization evaluation has seen a variety of different methodologies applied to evaluation of its metrics. Evaluation of summarization metrics has included, for example, the ability of a metric/significance test combination to distinguish between sets of human and system-generated summaries (Rankel et al., 2011) , or by accuracy of conclusions drawn from metrics when combined with a particular significance test, Wilcoxon ranksum (Owczarzak et al., 2012) .", |
| "cite_spans": [ |
| { |
| "start": 486, |
| "end": 507, |
| "text": "(Rankel et al., 2011)", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 627, |
| "end": 651, |
| "text": "(Owczarzak et al., 2012)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Besides moving away from well-established methods such as correlation with human judgment, previous summarization metric evaluations have been additionally limited by inclusion of only a small proportion of possible metrics and variants. For example, although the most commonly used metric ROUGE has a very large number of possible variants, it is common to include only a small range of those in evaluations. This has the obvious disadvantage that superior variants may exist but remain unidentified due to their omission.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Despite such limitations, however, subsequent evaluations of state-of-the-art summarization systems operate under the assumption that recommended ROUGE variants are optimal and rely on this assumption to draw conclusions about the relative performance of systems (Hong et al., 2014) . This forces us to raise some important questions. Firstly, to what degree was the divergence away from evaluation methodologies still applied to MT metrics today well-founded? For example, were the original methodology, by correlation with human assessment, to be applied, would a distinct variant of ROUGE emerge as superior and subsequently lead to distinct system rankings? Secondly, were all variants of ROUGE to be included in evaluations, would a variant originally omitted from the evaluation emerge as superior and lead to further differences in summarization system rankings? Furthermore, although methods of statistical significance testing are commonly applied to evaluation of summarization systems, attempts to identify significant differences in performance of metrics are extremely rare, and when they have been applied unfortunately have not used an appropriate test.", |
| "cite_spans": [ |
| { |
| "start": 263, |
| "end": 282, |
| "text": "(Hong et al., 2014)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "This motivates our review of past and current methodologies applied to the evaluation of summarization metrics. Since MT evaluation in general has its own imperfections, we do not attempt to indiscriminately impose all MT evaluation methodologies on summarization, but specifically revisit evaluation methodologies applied to one particular area of summarization, evaluation of metrics. Correlations with human assessment reveal an extremely wide range in performance among variants, highlighting the importance of an optimal choice of ROUGE variant in system evaluations. Since distinct variants of ROUGE achieve significantly stronger correlation with human assessment than previous recommended best variants, we subsequently replicate a recent evaluation of state-of-the-art summarization systems revealing distinct conclusions about the relative performance of systems. In addition, we include in the evaluation of metrics, an evaluation of BLEU for the purpose of summarization evaluation, and contrary to common belief, precision-based BLEU is on-par with recall-based ROUGE for evaluation of summarization systems.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "When ROUGE (Lin and Hovy, 2003) was first proposed, the methodology applied to its evaluation, in one respect, was similar to that applied to metrics in MT, as ROUGE variants were evaluated by correlation with a form of human assessment. Where the evaluation methodology diverged from MT, however, was with respect to the precise representation of human assessment that was employed. In MT evaluation of metrics, although experimentation has taken place with regards to methods of elicitation of assessments from human judges (Callison-Burch et al., 2008) , human assessment is always aimed to encapsulate the overall quality of translations. In contrast in summarization, metrics are evaluated by the degree to which metric scores correlate with human coverage scores for summaries, a recall-based formulation of the number of peer summary units that a human assessor believed had the same meaning as model summaries. Substitution of overall quality assessments with a recall-based manual metric, unfortunately has the potential to introduce bias into the evaluation of metrics in favor of recallbased formulations.", |
| "cite_spans": [ |
| { |
| "start": 11, |
| "end": 31, |
| "text": "(Lin and Hovy, 2003)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 526, |
| "end": 555, |
| "text": "(Callison-Burch et al., 2008)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "One dimension of summary quality omitted from human coverage scores is, for example, the order in which the units of a summary are arranged within the summary. Despite unit order quite likely being something of importance to a human assessor, assessment of metrics by correlation with human coverage scores does not in any respect take into account the order in which the units of a summary appear, and evaluation by human coverage scores alone means that a summary with its units scrambled or even reversed in theory receives precisely the same metric score as the original. Given current evaluation methodologies for assessment of metrics, a metric that scores two such summaries differently would be unfairly penalized for it. Furthermore, when the linguistic quality of summaries has been assessed in parallel with annotations used to compute human coverage scores, it has been shown that the two dimensions of quality do not correlate with one another (no significant correlation) (Pitler et al., 2010) , providing evidence that coverage scores alone do not fully represent human judgment of the overall quality of summaries.", |
| "cite_spans": [ |
| { |
| "start": 986, |
| "end": 1007, |
| "text": "(Pitler et al., 2010)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Subsequent summarization metric evaluations depart from correlation with human judgment fur-ther by evaluating metrics according to the ability of a metric/significance test combination to identify a significant difference between the quality of human and system-generated summaries (Rankel et al., 2011) . Unfortunately, the evaluation of metrics with respect to how well they distinguish between high-quality human summaries and all system-generated summaries, does not provide insight into the task of metrics, to score better quality system-generated summaries higher than worse quality system-generated summaries, however. This is in contrast to evaluation of MT metrics by correlation with human judgment, where metrics only receive credit for their ability to appropriately score system-output documents relative to other system-output documents. Since differences in quality levels between pairs of systemgenerated summaries are likely to be far smaller than differences in system and human-generated summaries, the methodology unfortunately sets too low a bar for summarization metrics to meet.", |
| "cite_spans": [ |
| { |
| "start": 283, |
| "end": 304, |
| "text": "(Rankel et al., 2011)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Furthermore, the approach to metric evaluation unfortunately does not work in the long-term, as the performance of summarization systems improves and approaches or achieves the quality of a human, a metric that accurately identifies this achievement would be unfairly penalized for it. Separate from the evaluation of metrics, Rankel et al. (2011) make the highly important recommendation of paired tests for identification of significant differences in performance of summarization systems. Since data used in the evaluation of summarization systems is not independent, paired tests are more appropriate and more powerful. Owczarzak et al. (2012) diverge further from correlation with human judgment for evaluation of metrics by assessing the accuracy of metrics to identify significant differences between pairs of systems when combined with a significance test. Although the approach to evaluation of metrics provides insight into the accuracy of conclusions drawn from metric/test combinations, the evaluation is limited by inclusion of only six variants of ROUGE, fewer than 4% of possible ROUGE variants. Despite such limitations, however, subsequent evaluations relied on recommended ROUGE variants to rank state-of-the-art systems (Hong et al., 2014) .", |
| "cite_spans": [ |
| { |
| "start": 327, |
| "end": 347, |
| "text": "Rankel et al. (2011)", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 624, |
| "end": 647, |
| "text": "Owczarzak et al. (2012)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 1239, |
| "end": 1258, |
| "text": "(Hong et al., 2014)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Although methods of identifying significant differences in performance are commonly applied to the evaluation of systems in summarization, the application of significance tests to the evaluation of summarization metrics is extremely rare, and when attempts have been made, unfortunately appropriate tests have not been applied. Computation of confidence intervals for individual correlation with human coverage scores, for example, unfortunately does not provide insight into whether or not a difference in correlation with human coverage scores is significant.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "When large-scale human evaluation of summarization systems takes place, human evaluation commonly takes the form of annotation of whether or not system-generated summary units express the meaning of model summary units, annotations subsequently used to compute human coverage scores. In addition, an evaluation of the linguistic quality of summaries is commonly carried out. As described in Section 2, when used for the evaluation of metrics, linguistic quality is commonly omitted, however, with metrics only assessed by the degree to which they correlate with human coverage scores. In contrast, we include all available human assessment data for evaluating metrics.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Summarization Metric Evaluation", |
| "sec_num": "3" |
| }, |
| { |
| "text": "In DUC-2004 (Over et al., 2007 , human annotations used to compute summary coverage are carried out by identification of matching peer units (PUs), the units in a peer summary that express content of the corresponding model summary. In addition, an overall coverage estimate (E) is provided by the human annotator, the proportion of the corresponding model summary or collective model units (MUs) expressed overall by a given peer summary. Human coverage scores (CS) are computed by combining Matching PUs with coverage estimates as follows:", |
| "cite_spans": [ |
| { |
| "start": 3, |
| "end": 11, |
| "text": "DUC-2004", |
| "ref_id": null |
| }, |
| { |
| "start": 12, |
| "end": 30, |
| "text": "(Over et al., 2007", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Combining Quality and Coverage", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "CS = |M atching P U s| \u2022 E |M U s|", |
| "eq_num": "(1)" |
| } |
| ], |
| "section": "Combining Quality and Coverage", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "In addition to annotations used to compute human coverage scores, human assessors were asked to rate the linguistic quality of summaries under 7 different criteria, providing ratings from A to E, with A denoting highest and E least quality rating. Figure 1 is a scatter-plot of human coverage scores and corresponding linguistic quality scores Coverage Score (%) Mean Ling. Quality (%) Figure 1 : Scatter-plot of mean linguistic quality and coverage scores for human assessments of summaries in DUC-2004 for all human-assessed summaries from DUC-2004, where, for the purpose of comparison, each of the 7 linguistic quality ratings are converted to a corresponding percentage quality (A= 100%; B= 75%; C= 50%; D= 25%; E= 0%). The location of all points almost exclusively within the upper left corner of the plot in Figure 1 indicates that the linguistic quality of almost all summaries reaches at least as high a level as its corresponding coverage score. This follows the intuition that a summary is unlikely to obtain high coverage without sufficient linguistic quality, while the same cannot be said for the converse, that a high level of linguistic quality necessarily leads to high coverage. More importantly, however, linguistic quality scores provide an additional dimension of human assessment, allowing greater discriminatory power between the quality of summaries than was possible with coverage scores alone. Figure 2 includes linguistic quality and coverage score distributions from DUC-2004 human evaluation, where each distribution is skewed in opposing directions, in addition to the distribution of the average of the two scores for summaries.", |
| "cite_spans": [ |
| { |
| "start": 495, |
| "end": 503, |
| "text": "DUC-2004", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 248, |
| "end": 256, |
| "text": "Figure 1", |
| "ref_id": null |
| }, |
| { |
| "start": 386, |
| "end": 394, |
| "text": "Figure 1", |
| "ref_id": null |
| }, |
| { |
| "start": 815, |
| "end": 823, |
| "text": "Figure 1", |
| "ref_id": null |
| }, |
| { |
| "start": 1420, |
| "end": 1428, |
| "text": "Figure 2", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Combining Quality and Coverage", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "For the purpose of metric evaluation, we combine human coverage and linguistic quality scores using the average of the two scores, and use this as a gold standard human score for evaluation of metrics:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Combining Quality and Coverage", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Human Assessment Score = CS+M LQ ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Combining Quality and Coverage", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "ROUGE includes a large number of distinct variants, including eight choices of n-gram counting method (ROUGE-1; 2; 3; 4; S4; SU4; W; L), binary settings such as word-stemming of summaries and an option to remove or retain stop-words. Additional configurations include the use of precision, recall or f-score to compute individual summary scores. Finally, options for computation of the overall score for a system is by computation of the mean or median of that system's summary score distribution. In total, therefore, when employing ROUGE for the evaluation of summarization systems, there are 192 (8 x 2 x 2 x 3 x 2) possible system-level variants to choose from. The fact that final overall ROUGE scores for systems are comprised of the mean or median of ROUGE scores of individual summaries, is, again, a divergence from MT evaluation, as ngram counts used to compute BLEU scores are computed at the document as opposed to sentencelevel. However, in this respect, ROUGE has a distinct advantage over BLEU, as the fact that ROUGE comprises the mean or median score of individual summary scores makes possible the application of standard methods of significance testing differences in system-level ROUGE scores, while BLEU is restricted to the application of randomized methods (Koehn, 2004; . For this purpose, differences in median ROUGE scores can be tested for statistical significance using, for example, Wilcoxon signed-rank test, while paired t-test can be applied to difference of mean ROUGE scores for systems.", |
| "cite_spans": [ |
| { |
| "start": 1280, |
| "end": 1293, |
| "text": "(Koehn, 2004;", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "ROUGE", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Moses (Koehn et al., 2007) multi-bleu 1 was used to compute BLEU (Papineni et al., 2002) scores for summaries and prepare4rouge 2 applied to summaries before running ROUGE (Lin and Hovy, 2003) . Table 1 shows the Pearson correlation of each variant of ROUGE with human assessment, in addition to BLEU's correlation with the same human assessment of summaries from DUC-2004. Somewhat surprisingly, BLEU MT evaluation metric achieves strongest correlation with human assessment overall, r = 0.797, with performance of ROUGE variants ranging from r = 0.786, just below that of BLEU, to as low as r = 0.293. For many pairs of metrics, differences in correlation with human judgment are small, however, and prior to concluding superiority in performance of one metric over another, significance tests should be applied.", |
| "cite_spans": [ |
| { |
| "start": 6, |
| "end": 26, |
| "text": "(Koehn et al., 2007)", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 65, |
| "end": 88, |
| "text": "(Papineni et al., 2002)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 172, |
| "end": 192, |
| "text": "(Lin and Hovy, 2003)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 195, |
| "end": 202, |
| "text": "Table 1", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Metric Evaluation by Pearson's r", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "In MT, recent work has identified the suitability of Williams significance test (Williams, 1959) for evaluation of automatic MT metrics Graham, 2015) , and, for similar reasons, Williams test is suited to significance testing differences in performance of competing summarization metrics which we detail further below. Williams test has additionally been used in evaluation of systems that automatically assess spoken and written language quality (Yannakoudakis et al., 2011; Yannakoudakis and Briscoe, 2012; Evanini et al., 2013) .", |
| "cite_spans": [ |
| { |
| "start": 80, |
| "end": 96, |
| "text": "(Williams, 1959)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 136, |
| "end": 149, |
| "text": "Graham, 2015)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 447, |
| "end": 475, |
| "text": "(Yannakoudakis et al., 2011;", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 476, |
| "end": 508, |
| "text": "Yannakoudakis and Briscoe, 2012;", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 509, |
| "end": 530, |
| "text": "Evanini et al., 2013)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Metric Significance Testing", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Evaluation of a given summarization metric, M new , by Pearson correlation takes the form of quantifying the correlation, r(M new , H), that exists between metric scores for systems and corresponding human assessment scores, and contrasting this correlation with the correlation for some baseline metric, r(M base , H).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Metric Significance Testing", |
| "sec_num": "4" |
| }, |
| { |
| "text": "One approach to testing for significance that may seem reasonable is to apply a significance test 1 https://github.com/moses-smt/mosesdecoder/ commits/master/scripts/generic/multi-bleu.perl 2 http://kavita-ganesan.com/content/ prepare4rouge-script-prepare-rouge-evaluation separately to the correlation of each metric with human assessment, with the hope that the new metric will achieve a significant correlation where the baseline metric does not. The reasoning here is flawed however: the fact that one correlation is significantly higher than zero (r(M new , H)) and that of another is not, does not necessarily mean that the difference between the two correlations is significant. Instead, a specific test should be applied to the difference in correlations. For this same reason, confidence intervals for individual correlations with human assessment are also not useful.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Metric Significance Testing", |
| "sec_num": "4" |
| }, |
| { |
| "text": "If samples that data are drawn from are independent, and differences in correlations are computed on independent data sets, the Fisher r to z transformation is applied to test for significant differences in correlations. Data used for the evaluation of summarization metrics are not independent, as evaluations comprise three sets of scores for precisely the same set of summaries (corresponding to variables X 1 , X 2 and X 3 below), and subsequently three correlations: Williams test 3 (Williams, 1959) evaluates the significance of a difference in dependent correlations (Steiger, 1980) . It is formulated as follows as a test of whether the population correlation between X 1 and X 3 equals the population correlation between X 2 and X 3 :", |
| "cite_spans": [ |
| { |
| "start": 488, |
| "end": 504, |
| "text": "(Williams, 1959)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 574, |
| "end": 589, |
| "text": "(Steiger, 1980)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Metric Significance Testing", |
| "sec_num": "4" |
| }, |
| { |
| "text": "r(M base , H), r(M", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Metric Significance Testing", |
| "sec_num": "4" |
| }, |
| { |
| "text": "t(n \u2212 3) = (r 13 \u2212 r 23 ) (n \u2212 1)(1 + r 12 ) 2K (n\u22121) (n\u22123) + (r 23 +r 13 ) 2 4 (1 \u2212 r 12 ) 3 ,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Metric Significance Testing", |
| "sec_num": "4" |
| }, |
| { |
| "text": "where r ij is the correlation between X i and X j , n is the size of the population, and: -2004, with (Y) and without (N) stemming, with (Y) and without (N) removal of stop words (RSW), aggregated at the summary level using precision (P), recall (R) or f-score (F), aggregated at the system level by average (A) or median (M) summary score, correlations marked with \u2022 signify a metric/variant whose correlation with human assessment is not significantly weaker than that of any other metric/variant (an optimal variant) according to pairwise Williams significance tests, variants employed in Hong et al. (2014) are in bold.", |
| "cite_spans": [ |
| { |
| "start": 90, |
| "end": 105, |
| "text": "-2004, with (Y)", |
| "ref_id": null |
| }, |
| { |
| "start": 592, |
| "end": 610, |
| "text": "Hong et al. (2014)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Metric Significance Testing", |
| "sec_num": "4" |
| }, |
| { |
| "text": "K = 1 \u2212 r 12 2 \u2212", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Metric Significance Testing", |
| "sec_num": "4" |
| }, |
| { |
| "text": "r BLEU 0.797 \u2022 R-2 Y Y A P 0.786 \u2022 R-3 N N A F 0.785 \u2022 R-2 N Y A P 0.783 \u2022 R-3 N Y A P 0.781 \u2022 R-3 Y N A F 0.779 \u2022 R-3 N N A R 0.777 \u2022 R-4 N N A F 0.771 \u2022 R-3 N N A P 0.771 \u2022 R-3 Y N A R 0.770 \u2022 R-2 N Y A F 0.769 \u2022 R-4 N N A R 0.768 \u2022 R-2 Y Y A F 0.768 \u2022 R-3 Y N A P 0.767 \u2022 R-3 N N M F 0.766 \u2022 R-3 N Y A F 0.764 \u2022 R-3 Y Y A P 0.764 \u2022 R-4 Y N A F 0.763 \u2022 R-4 N N A P 0.762 \u2022 R-4 Y N A R 0.761 \u2022 R-3 N N M P 0.760 \u2022 R-4 Y Y A P 0.759 \u2022 R-2 Y N A P 0.759 \u2022 R-4 N Y A P 0.758 \u2022 R-2 N N A P 0.757 \u2022 R-3 N N M R 0.753 \u2022 R-4 Y N A P 0.752 \u2022 R-3 Y Y A F 0.748 \u2022 R-2 N N A F 0.747 \u2022 R-2 Y N A F 0.747 \u2022 R-3 N Y A R 0.746 \u2022 R-3 Y N M P 0.744 \u2022 R-2 N Y M P 0.743 \u2022 R-3 Y N M F 0.743 \u2022 R-2 N Y A R 0.742 \u2022 R-2 Y Y M P 0.741 \u2022 R-2 N Y M F 0.740 \u2022 R-3 Y N M R 0.739 \u2022 R-2 Y Y A R 0.737 \u2022 R-2 Y Y M F 0.735 \u2022 R-2 N N M P 0.734 \u2022 R-3 Y Y M P 0.733 \u2022 R-3 Y Y A R 0.730 R-4 Y Y A F 0.729 \u2022 R-3 Y Y M F 0.726 \u2022 R-S4 Y N A P 0.725 \u2022 R-SU4 N N A P 0.724 \u2022 R-2 Y N M P 0.724 R-S4 N Y A P 0.724 R-SU4 Y N A P 0.723 \u2022 R-S4 N N A P 0.723 \u2022 R-2 N Y M R 0.722 \u2022 R-4 N Y A F 0.721 \u2022 R-1 N N A P 0.720 \u2022 R-2 N N M F 0.719 \u2022 R-SU4 N Y A P 0.719 R-1 Y N A P 0.714 \u2022 R-2 Y Y M R 0.714 \u2022 R-3 Y Y M R 0.713 \u2022 R-4 Y Y A R 0.712 \u2022 R-S4 Y Y A P 0.711 R-SU4 Y Y A P 0.710 R-2 N N A R 0.710 \u2022 R-W N Y A P 0.709 \u2022 R-2 Y N A R 0.707 \u2022 Metric Stemming RSW Ave./Med P/R/F r R-2 Y N M F 0.706 R-3 N Y M P 0.704 \u2022 R-1 N Y A P 0.704 \u2022 R-4 N N M R 0.703 \u2022 R-L N Y A P 0.700 \u2022 R-W Y Y A P 0.700 \u2022 R-4 N Y A R 0.700 \u2022 R-1 Y N M P 0.699 \u2022 R-S4 N Y M P 0.698 R-1 Y Y A P 0.698 \u2022 R-3 N Y M F 0.697 \u2022 R-W N N A P 0.696 \u2022 R-W Y N A P 0.695 \u2022 R-4 N N M F 0.695 \u2022 R-S4 N Y M F 0.693 R-S4 N Y A F 0.691 R-SU4 N Y M P 0.690 R-1 N N M P 0.690 \u2022 R-2 N N M R 0.689 R-L Y Y A P 0.688 \u2022 R-3 N Y M R 0.687 \u2022 R-S4 N N M P 0.687 R-S4 Y N A F 0.687 R-S4 N N A F 0.687 R-4 N N M P 0.687 \u2022 R-L N N A P 0.686 \u2022 R-SU4 N N M P 0.686 R-L Y N A P 0.683 \u2022 R-W N N M P 0.682 \u2022 R-W Y N M P 0.680 \u2022 R-SU4 Y N M P 0.678 R-SU4 N Y A F 0.678 R-S4 Y Y A F 0.676 R-SU4 N Y M F 0.676 R-SU4 N N A F 0.673 R-1 N Y M P 0.673 R-2 Y N M R 0.672 R-SU4 Y N A F 0.671 R-S4 N Y M R 0.670 R-S4 Y N M P 0.670 R-SU4 Y Y A F 0.668 R-S4 N N M F 0.666 R-W N Y M P 0.664 R-S4 Y Y M P 0.664 R-SU4 Y Y M P 0.663 R-L N N M P 0.661 \u2022 R-SU4 N N M F 0.658 R-1 N Y A F 0.656 R-W Y Y M P 0.656 R-S4 N Y A R 0.656 R-L Y N M P 0.656 \u2022 R-W N Y A F 0.655 R-1 N Y M F 0.653 R-L N Y A F 0.652 R-1 Y Y M P 0.651 R-S4 Y Y M F 0.649 R-1 Y Y A F 0.649 R-SU4 Y Y M F 0.649 R-SU4 N Y M R 0.646 R-L N Y M P 0.645 R-W N Y M F 0.642 R-W Y Y A F 0.642 R-4 Y N M R 0.641 R-S4 Y Y A R 0.641 R-4 Y N M F 0.639 Metric Stemming RSW Ave./Med P/R/F r R-L Y Y A F 0.638 R-1 N N A F 0.637 R-S4 Y N M F 0.634 R-4 Y N M P 0.634 R-1 N N M F 0.634 R-SU4 N Y A R 0.633 R-L Y Y M P 0.633 R-SU4 Y Y M R 0.631 R-1 Y N A F 0.630 R-1 Y Y M F 0.629 R-S4 Y Y M R 0.626 R-S4 N N A R 0.626 R-SU4 Y N M F 0.625 R-S4 Y N A R 0.624 R-L N Y M F 0.623 R-SU4 Y Y A R 0.622 R-1 Y N M F 0.617 R-1 N Y M R 0.615 R-W N Y A R 0.613 R-S4 N N M R 0.611 R-L N Y M R 0.609 R-1 N Y A R 0.604 R-L N Y A R 0.601 R-W N N M F 0.600 R-L N N M F 0.599 R-W Y Y A R 0.598 R-W N Y M R 0.597 R-1 Y Y A R 0.595 R-1 Y Y M R 0.591 R-L N N A F 0.586 R-W Y Y M F 0.586 R-W Y N M F 0.585 R-L Y Y A R 0.583 R-L Y Y M F 0.582 R-L Y Y M R 0.579 R-L Y N A F 0.579 R-W N N A F 0.579 R-SU4 N N M R 0.576 R-W Y N A F 0.576 R-SU4 N N A R 0.574 R-SU4 Y N A R 0.571 R-L Y N M F 0.569 R-W Y Y M R 0.567 R-S4 Y N M R 0.566 R-SU4 Y N M R 0.525 R-1 N N M R 0.488 R-1 Y N M R 0.477 R-W Y N M R 0.477 R-1 N N A R 0.470 R-W N N M R 0.470 R-L N N M R 0.470 R-1 Y N A R 0.459 R-W N N A R 0.456 R-W Y N A R 0.452 R-L Y N M R 0.423 R-L N N A R 0.416 R-L Y N A R 0.406 R-4 Y Y M P 0.307 R-4 Y Y M F 0.302 R-4 N Y M P 0.301 R-4 Y Y M R 0.297 R-4 N Y M F 0.296 R-4 N Y M R 0.293", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Metric Significance Testing", |
| "sec_num": "4" |
| }, |
| { |
| "text": "outperform, as a metric that happens to correlate strongly with a higher number of competing metrics in a given competition would be at an unfair advantage. This increased power also means, somewhat counter-intuitively, it can happen for a pair of competing metrics for which the correlation between metric scores is strong, that a small difference in competing correlations with human assessment is significant, while, for a different pair of metrics with a larger difference in correlation, the difference is not significant, because r(M base , M new ) is weak. For example, in Table 1 the difference in correlation with human assessment of BLEU and that of median ROUGE-L precision with stemming and stop-words retained, 0.141 (0.797 \u2212 0.656), is not significant, while the smaller difference in correlation with human assessment between correlations of BLEU and average ROUGE-3 recall with stemming and stopwords removed, 0.067 (0.797 \u2212 0.73) is significant, since scores of the latter pair of metrics correlate with one another with more strength. As part of this research, we have made available an open-source implementation of statistical tests for evaluation of summarization metrics, at https://github.com/ygraham/nlp-williams.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Metric Significance Testing", |
| "sec_num": "4" |
| }, |
| { |
| "text": "In (Owczarzak et al. (2012) ) used to compare systems in Hong et al. (2014) . Current recommended best variants of ROUGE are shown to be significantly outperformed by several other ROUGE variants.", |
| "cite_spans": [ |
| { |
| "start": 3, |
| "end": 27, |
| "text": "(Owczarzak et al. (2012)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 57, |
| "end": 75, |
| "text": "Hong et al. (2014)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Significance Test Results", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "Although BLEU achieves strongest correlation with human assessment overall, Figure 3 reveals the difference between BLEU's correlation with human assessment and that of the best-performing ROUGE variant as not statistically significant, and since ROUGE holds the distinct advantage over BLEU of facilitating standard methods of significance testing differences in scores for systems, for this reason alone we recommend the use of the best-performing ROUGE variant over BLEU, average ROUGE-2 precision with stemming and stopwords removed. Table 2 shows proportions of optimal ROUGE BLEU R2_ST1_RS1_A_P R3_ST0_RSW0_A_F R2_ST0_RS1_A_P R3_ST0_RS1_A_P R3_ST1_RS0_A_F R3_ST0_RS0_A_R R4_ST0_RS0_A_F R3_ST0_RS0_A_P R4_ST1_RS0_M_R R2_ST1_RS0_M_R R1_ST1_RS0_M_R R1_ST1_RS0_M_R R2_ST1_RS0_M_R R4_ST1_RS0_M_R R3_ST0_RS0_A_P R4_ST0_RS0_A_F R3_ST0_RS0_A_R R3_ST1_RS0_A_F R3_ST0_RS1_A_P R2_ST0_RS1_A_P R3_ST0_RS0_A_F R2_ST1_RS1_A_P BLEU Figure 3 : Pairwise significance test outcomes for BLEU, best-performing ROUGE (rows 2-9), and ROUGE applied in Hong et al. (2014) (bottom 3 rows), with (ST1) and without (ST0) stemming, with (RS1) and without (RS0) removal of stop words, for average (A) or median (M) ROUGE precision (P), recall (R) or f-score (F), colored cells denote significant win for row i metric over column j metric with Williams test. variants that can be attributed to each of ROUGE's configuration options. Contrary to prior belief, the vast majority of optimal ROUGE variants are precision-based, showing that the assumption that recall-based metrics are superior for evaluation of summarization systems to be inaccurate, and a likely presence of bias in favor of recall-based metrics in evaluations by correlation with human coverage scores alone. Furthermore, since there exists a vast number of possible formulations that could potentially be applied to evaluation of summaries that are neither purely precision nor recallbased, evaluation methodologies should avoid reliance on assumptions that either precision or recall is superior and instead base conclusions on empirical evidence where possible.", |
| "cite_spans": [ |
| { |
| "start": 1067, |
| "end": 1085, |
| "text": "Hong et al. (2014)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 1352, |
| "end": 1366, |
| "text": "Williams test.", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 76, |
| "end": 84, |
| "text": "Figure 3", |
| "ref_id": null |
| }, |
| { |
| "start": 538, |
| "end": 934, |
| "text": "Table 2 shows proportions of optimal ROUGE BLEU R2_ST1_RS1_A_P R3_ST0_RSW0_A_F R2_ST0_RS1_A_P R3_ST0_RS1_A_P R3_ST1_RS0_A_F R3_ST0_RS0_A_R R4_ST0_RS0_A_F R3_ST0_RS0_A_P R4_ST1_RS0_M_R R2_ST1_RS0_M_R R1_ST1_RS0_M_R R1_ST1_RS0_M_R R2_ST1_RS0_M_R R4_ST1_RS0_M_R R3_ST0_RS0_A_P R4_ST0_RS0_A_F R3_ST0_RS0_A_R R3_ST1_RS0_A_F R3_ST0_RS1_A_P R2_ST0_RS1_A_P R3_ST0_RS0_A_F", |
| "ref_id": "TABREF1" |
| }, |
| { |
| "start": 955, |
| "end": 963, |
| "text": "Figure 3", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Significance Test Results", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "Since we have established that the variants of ROUGE used to rank state-of-the-art and baseline summarization systems in Hong et al. (2014) have significantly weaker correlations with human assessment than several other ROUGE variants, this motivates our replication of the evaluation. We evaluate systems using the variant of ROUGE that achieves strongest correlation with human assessment, average ROUGE-2 precision with stemming and stop-words removed. Table 3 shows ROUGE scores for summarization systems originally presented in Hong et al. (2014) . System rankings diverge considerably from those of the original evaluation. Notably, the system now taking first place had originally ranked in fourth position.", |
| "cite_spans": [ |
| { |
| "start": 121, |
| "end": 139, |
| "text": "Hong et al. (2014)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 533, |
| "end": 551, |
| "text": "Hong et al. (2014)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 456, |
| "end": 463, |
| "text": "Table 3", |
| "ref_id": "TABREF5" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Summarization System Evaluation", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Since the best variant of ROUGE is based on average ROUGE scores as opposed to median ROUGE scores, a difference of means significance test is appropriate provided the normality assumption of score distributions for systems is not violated. In addition, since data used to evaluate systems are not independent, paired tests are also appropriate (Rankel et al., 2011) . ROUGE score distributions for systems were tested for normality using the Shapiro-Wilk test (Royston, 1982) where score distributions for none of the included systems were shown to be significantly non-normal. Figure 4 shows outcomes of paired t-tests for summary score distributions of each pair of systems, revealing three summarization systems not significantly outperformed by any other as DPP, ICSISUMM and REGSUM. In addition, as expected, all state-of-the-art systems significantly outperform all baseline systems.", |
| "cite_spans": [ |
| { |
| "start": 345, |
| "end": 366, |
| "text": "(Rankel et al., 2011)", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 461, |
| "end": 476, |
| "text": "(Royston, 1982)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 579, |
| "end": 587, |
| "text": "Figure 4", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Summarization System Evaluation", |
| "sec_num": "5" |
| }, |
| { |
| "text": "In order to evaluate metrics by correlation with human assessment, it is necessary to obtain a single human evaluation score per system. For example, in the evaluation of metrics in Section 3, we combined linguistic quality and coverage into a single score using the mean of the two scores. Other combinations are of course possible, but without any additional human evaluation data, it is challenging to identify the combination that best rep- Table 4 : Correlation of top-ten metric variants for each alternate combination of linguistic quality and coverage, \u2022 denotes a metric not significantly outperformed by any other under that particular human evaluation combination, highest correlations highlighted in bold font.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 445, |
| "end": 452, |
| "text": "Table 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Human Assessment Combinations", |
| "sec_num": "6" |
| }, |
| { |
| "text": "resents an overall human assessment for a given summary. One possibility would be to search for optimal weights for combining quality and coverage, but there is a risk with this approach that we will not find the most representative combination but simply the combination that best describes the metrics.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Human Assessment Combinations", |
| "sec_num": "6" |
| }, |
| { |
| "text": "An additional variation of human assessment scores is by combining coverage and quality with a variant of the arithmetic mean, such as the harmonic or geometric mean. Table 4 shows correlations of BLEU and the top ten performing variants of ROUGE when evaluated against the arithmetic (mean), harmonic and geometric mean of quality and coverage scores for summaries. In addition, Table 4 includes correlations of metric scores with coverage alone, as well as linguistic quality scores alone to provide additional insight, although linguistic quality scores alone do not provide a sufficient evaluation of metrics -since it is possible to generate summaries with perfect linguistic quality without inclusion of any relevant content whatsoever.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 167, |
| "end": 174, |
| "text": "Table 4", |
| "ref_id": null |
| }, |
| { |
| "start": 380, |
| "end": 387, |
| "text": "Table 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Human Assessment Combinations", |
| "sec_num": "6" |
| }, |
| { |
| "text": "BLEU MT metric achieves highest correlation across all human evaluation combinations and highest again when evaluated against human coverage scores alone, and BLEU's brevity penalty, that like recall penalizes a system for too short output, is a probable cause of the metric overcom-ing the recall-based bias of an evaluation based on coverage scores alone. In addition, our recommended variant, ave. ROUGE-2 prec. with stemming and stop words removed is not significantly outperformed by BLEU or any other variant of ROUGE for any of the three combined mean human assessment scores.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Human Assessment Combinations", |
| "sec_num": "6" |
| }, |
| { |
| "text": "An analysis of evaluation of summarization metrics was provided with an evaluation of BLEU and 192 variants of ROUGE. Detail of the first suitable summarization metric significance test, Williams test, was provided. Results reveal superior variants of metrics distinct from previously best recommendations. Replication of a recent evaluation of state-of-the-art summarization systems also revealed contrasting conclusions about the relative performance of systems. In addition, BLEU achieves strongest correlation with human assessment overall, but does not significantly outperform the best-performing ROUGE variant.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "7" |
| }, |
| { |
| "text": "Also known as Hotelling-Williams.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "We wish to thank the anonymous reviewers. This research is supported by Science Foundation Ireland through the CNGL Programme (Grant 12/CE/I2267) in the ADAPT Centre (www. adaptcentre.ie) at Trinity College Dublin.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Further meta-evaluation of machine translation", |
| "authors": [ |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Callison", |
| "suffix": "" |
| }, |
| { |
| "first": "-", |
| "middle": [], |
| "last": "Burch", |
| "suffix": "" |
| }, |
| { |
| "first": "Cameron", |
| "middle": [], |
| "last": "Fordyce", |
| "suffix": "" |
| }, |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| }, |
| { |
| "first": "Christof", |
| "middle": [], |
| "last": "Monz", |
| "suffix": "" |
| }, |
| { |
| "first": "Josh", |
| "middle": [], |
| "last": "Schroeder", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proc. 3rd Wkshp. Statistical Machine Translation", |
| "volume": "", |
| "issue": "", |
| "pages": "70--106", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chris Callison-Burch, Cameron Fordyce, Philipp Koehn, Christof Monz, and Josh Schroeder. 2008. Further meta-evaluation of machine translation. In Proc. 3rd Wkshp. Statistical Machine Translation, pages 70-106, Columbus, Ohio. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Prompt-based content scoring for automated spoken language assessment", |
| "authors": [ |
| { |
| "first": "Keelan", |
| "middle": [], |
| "last": "Evanini", |
| "suffix": "" |
| }, |
| { |
| "first": "Shasha", |
| "middle": [], |
| "last": "Xie", |
| "suffix": "" |
| }, |
| { |
| "first": "Klaus", |
| "middle": [], |
| "last": "Zechner", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proc. 2013 Conf. North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "157--162", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Keelan Evanini, Shasha Xie, and Klaus Zechner. 2013. Prompt-based content scoring for automated spoken language assessment. In Proc. 2013 Conf. North American Chapter of the Association for Computa- tional Linguistics: Human Language Technologies, pages 157-162, Atlanta, GA. Association for Com- putational Linguistics.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Testing for significance of increased correlation with human judgment", |
| "authors": [ |
| { |
| "first": "Yvette", |
| "middle": [], |
| "last": "Graham", |
| "suffix": "" |
| }, |
| { |
| "first": "Timothy", |
| "middle": [], |
| "last": "Baldwin", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proc. 2014 Conf. Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "172--176", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yvette Graham and Timothy Baldwin. 2014. Testing for significance of increased correlation with human judgment. In Proc. 2014 Conf. Empirical Methods in Natural Language Processing, pages 172-176, Doha, Qatar. Association for Computational Lin- guistics.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Randomized significance tests in machine translation", |
| "authors": [ |
| { |
| "first": "Yvette", |
| "middle": [], |
| "last": "Graham", |
| "suffix": "" |
| }, |
| { |
| "first": "Nitika", |
| "middle": [], |
| "last": "Mathur", |
| "suffix": "" |
| }, |
| { |
| "first": "Timothy", |
| "middle": [], |
| "last": "Baldwin", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proc. 9th Wkshp. Statistical Machine Translation", |
| "volume": "", |
| "issue": "", |
| "pages": "266--274", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yvette Graham, Nitika Mathur, and Timothy Baldwin. 2014. Randomized significance tests in machine translation. In Proc. 9th Wkshp. Statistical Machine Translation, pages 266-274, Baltimore, MD. Asso- ciation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Accurate evaluation of segment-level machine translation metrics", |
| "authors": [ |
| { |
| "first": "Yvette", |
| "middle": [], |
| "last": "Graham", |
| "suffix": "" |
| }, |
| { |
| "first": "Nitika", |
| "middle": [], |
| "last": "Mathur", |
| "suffix": "" |
| }, |
| { |
| "first": "Timothy", |
| "middle": [], |
| "last": "Baldwin", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proc. 2015 Conf. North American Chapter of the Association for Computational Linguistics -Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "1183--1191", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yvette Graham, Nitika Mathur, and Timothy Baldwin. 2015. Accurate evaluation of segment-level ma- chine translation metrics. In Proc. 2015 Conf. North American Chapter of the Association for Compu- tational Linguistics -Human Language Technolo- gies, pages 1183-1191, Denver, CO. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Improving evaluation of machine translation quality estimation", |
| "authors": [ |
| { |
| "first": "Yvette", |
| "middle": [], |
| "last": "Graham", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proc. Fifty-Third Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "1804--1813", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yvette Graham. 2015. Improving evaluation of ma- chine translation quality estimation. In Proc. Fifty- Third Annual Meeting of the Association for Com- putational Linguistics, pages 1804-1813, Beijing, China. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "A repository of state of the art and competitive baseline summaries for generic news summarization", |
| "authors": [ |
| { |
| "first": "Kai", |
| "middle": [], |
| "last": "Hong", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "John", |
| "suffix": "" |
| }, |
| { |
| "first": "Benoit", |
| "middle": [], |
| "last": "Conroy", |
| "suffix": "" |
| }, |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Favre", |
| "suffix": "" |
| }, |
| { |
| "first": "Hui", |
| "middle": [], |
| "last": "Kulesza", |
| "suffix": "" |
| }, |
| { |
| "first": "Ani", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Nenkova", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proc. 9th edition of the Language Resources and Evaluation Conference", |
| "volume": "", |
| "issue": "", |
| "pages": "1608--1616", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kai Hong, John M Conroy, Benoit Favre, Alex Kulesza, Hui Lin, and Ani Nenkova. 2014. A repos- itory of state of the art and competitive baseline sum- maries for generic news summarization. In Proc. 9th edition of the Language Resources and Evaluation Conference, pages 1608-1616, Reykjavik, Iceland.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Moses: Open source toolkit for statistical machine translation", |
| "authors": [ |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| }, |
| { |
| "first": "Hieu", |
| "middle": [], |
| "last": "Hoang", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexandra", |
| "middle": [], |
| "last": "Birch", |
| "suffix": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Callison-Burch", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcello", |
| "middle": [], |
| "last": "Federico", |
| "suffix": "" |
| }, |
| { |
| "first": "Nicola", |
| "middle": [], |
| "last": "Bertoldi", |
| "suffix": "" |
| }, |
| { |
| "first": "Brooke", |
| "middle": [], |
| "last": "Cowan", |
| "suffix": "" |
| }, |
| { |
| "first": "Wade", |
| "middle": [], |
| "last": "Shen", |
| "suffix": "" |
| }, |
| { |
| "first": "Christine", |
| "middle": [], |
| "last": "Moran", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Zens", |
| "suffix": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Dyer", |
| "suffix": "" |
| }, |
| { |
| "first": "Ondrej", |
| "middle": [], |
| "last": "Bojar", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexandra", |
| "middle": [], |
| "last": "Constantin", |
| "suffix": "" |
| }, |
| { |
| "first": "Evan", |
| "middle": [], |
| "last": "Herbst", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proc. 45th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "177--180", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Philipp Koehn, Hieu Hoang, Alexandra Birch, Chris Callison-Burch, Marcello Federico, Nicola Bertoldi, Brooke Cowan, Wade Shen, Christine Moran, Richard Zens, Chris Dyer, Ondrej Bojar, Alexan- dra Constantin, and Evan Herbst. 2007. Moses: Open source toolkit for statistical machine transla- tion. In Proc. 45th Annual Meeting of the Associa- tion for Computational Linguistics, pages 177-180, Prague, Czech Republic. Association for Computa- tional Linguistics.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Statistical significance tests for machine translation evaluation", |
| "authors": [ |
| { |
| "first": "Philipp", |
| "middle": [], |
| "last": "Koehn", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proc. 2004 Conf. Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "388--395", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Philipp Koehn. 2004. Statistical significance tests for machine translation evaluation. In Proc. 2004 Conf. Empirical Methods in Natural Language Process- ing, pages 388-395, Barcelona, Spain. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Automatic evaluation of summaries using n-gram cooccurrence statistics", |
| "authors": [ |
| { |
| "first": "Chin-Yew", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| }, |
| { |
| "first": "Eduard", |
| "middle": [], |
| "last": "Hovy", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proc. 2003 Conf. North American Chapter of the Association for Computational Linguistics on Human Language Technology", |
| "volume": "", |
| "issue": "", |
| "pages": "71--78", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chin-Yew Lin and Eduard Hovy. 2003. Auto- matic evaluation of summaries using n-gram co- occurrence statistics. In Proc. 2003 Conf. North American Chapter of the Association for Compu- tational Linguistics on Human Language Technol- ogy, pages 71-78, Edmonton, Canada. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Duc in context. Information Processing & Management", |
| "authors": [ |
| { |
| "first": "Paul", |
| "middle": [], |
| "last": "Over", |
| "suffix": "" |
| }, |
| { |
| "first": "Hoa", |
| "middle": [], |
| "last": "Dang", |
| "suffix": "" |
| }, |
| { |
| "first": "Donna", |
| "middle": [], |
| "last": "Harman", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "", |
| "volume": "43", |
| "issue": "", |
| "pages": "1506--1520", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Paul Over, Hoa Dang, and Donna Harman. 2007. Duc in context. Information Processing & Management, 43(6):1506-1520.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "An assessment of the accuracy of automatic evaluation in summarization", |
| "authors": [ |
| { |
| "first": "Karolina", |
| "middle": [], |
| "last": "Owczarzak", |
| "suffix": "" |
| }, |
| { |
| "first": "John", |
| "middle": [ |
| "M" |
| ], |
| "last": "Conroy", |
| "suffix": "" |
| }, |
| { |
| "first": "Hoa", |
| "middle": [ |
| "Trang" |
| ], |
| "last": "Dang", |
| "suffix": "" |
| }, |
| { |
| "first": "Ani", |
| "middle": [], |
| "last": "Nenkova", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proc. Wkshp. on Evaluation Metrics and System Comparison for Automatic Summarization", |
| "volume": "", |
| "issue": "", |
| "pages": "1--9", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Karolina Owczarzak, John M. Conroy, Hoa Trang Dang, and Ani Nenkova. 2012. An assessment of the accuracy of automatic evaluation in summariza- tion. In Proc. Wkshp. on Evaluation Metrics and System Comparison for Automatic Summarization, pages 1-9, Quebec, Canada. Association for Com- putational Linguistics.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "BLEU: A method for automatic evaluation of machine translation", |
| "authors": [ |
| { |
| "first": "Kishore", |
| "middle": [], |
| "last": "Papineni", |
| "suffix": "" |
| }, |
| { |
| "first": "Salim", |
| "middle": [], |
| "last": "Roukos", |
| "suffix": "" |
| }, |
| { |
| "first": "Todd", |
| "middle": [], |
| "last": "Ward", |
| "suffix": "" |
| }, |
| { |
| "first": "Wei-Jing", |
| "middle": [], |
| "last": "Zhu", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proc. 40th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "311--318", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kishore Papineni, Salim Roukos, Todd Ward, and Wei- Jing Zhu. 2002. BLEU: A method for automatic evaluation of machine translation. In Proc. 40th An- nual Meeting of the Association for Computational Linguistics, pages 311-318, Philadelphia, PA. Asso- ciation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Automatic evaluation of linguistic quality in multidocument summarization", |
| "authors": [ |
| { |
| "first": "Emily", |
| "middle": [], |
| "last": "Pitler", |
| "suffix": "" |
| }, |
| { |
| "first": "Annie", |
| "middle": [], |
| "last": "Louis", |
| "suffix": "" |
| }, |
| { |
| "first": "Ani", |
| "middle": [], |
| "last": "Nenkova", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proc. 48th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "544--554", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Emily Pitler, Annie Louis, and Ani Nenkova. 2010. Automatic evaluation of linguistic quality in multi- document summarization. In Proc. 48th Annual Meeting of the Association for Computational Lin- guistics, pages 544-554, Uppsala, Sweden. Associ- ation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Ranking human and machine summarization systems", |
| "authors": [ |
| { |
| "first": "Peter", |
| "middle": [], |
| "last": "Rankel", |
| "suffix": "" |
| }, |
| { |
| "first": "John", |
| "middle": [ |
| "M" |
| ], |
| "last": "Conroy", |
| "suffix": "" |
| }, |
| { |
| "first": "Eric", |
| "middle": [ |
| "V" |
| ], |
| "last": "Slud", |
| "suffix": "" |
| }, |
| { |
| "first": "Dianne", |
| "middle": [ |
| "P" |
| ], |
| "last": "O'leary", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proc. 2011 Conf. Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "467--473", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Peter Rankel, John M. Conroy, Eric V. Slud, and Di- anne P. O'Leary. 2011. Ranking human and ma- chine summarization systems. In Proc. 2011 Conf. Empirical Methods in Natural Language Process- ing, pages 467-473, Edinburgh, Scotland. Associ- ation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Algorithm as 181: The W test for normality", |
| "authors": [ |
| { |
| "first": "Patrick", |
| "middle": [], |
| "last": "Royston", |
| "suffix": "" |
| } |
| ], |
| "year": 1982, |
| "venue": "Applied Statistics", |
| "volume": "31", |
| "issue": "", |
| "pages": "176--180", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Patrick Royston. 1982. Algorithm as 181: The W test for normality. Applied Statistics, 31:176-180.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Tests for comparing elements of a correlation matrix", |
| "authors": [ |
| { |
| "first": "James", |
| "middle": [ |
| "H" |
| ], |
| "last": "Steiger", |
| "suffix": "" |
| } |
| ], |
| "year": 1980, |
| "venue": "Psychological Bulletin", |
| "volume": "87", |
| "issue": "2", |
| "pages": "245--251", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "James H. Steiger. 1980. Tests for comparing ele- ments of a correlation matrix. Psychological Bul- letin, 87(2):245-251.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Regression analysis", |
| "authors": [ |
| { |
| "first": "Evan", |
| "middle": [ |
| "J" |
| ], |
| "last": "Williams", |
| "suffix": "" |
| } |
| ], |
| "year": 1959, |
| "venue": "", |
| "volume": "14", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Evan J. Williams. 1959. Regression analysis, vol- ume 14. Wiley New York.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Modeling coherence in ESOL learner texts", |
| "authors": [ |
| { |
| "first": "Helen", |
| "middle": [], |
| "last": "Yannakoudakis", |
| "suffix": "" |
| }, |
| { |
| "first": "Ted", |
| "middle": [], |
| "last": "Briscoe", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proc. Seventh Wkshp. on Building Educational Applications Using NLP", |
| "volume": "", |
| "issue": "", |
| "pages": "33--43", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Helen Yannakoudakis and Ted Briscoe. 2012. Model- ing coherence in ESOL learner texts. In Proc. Sev- enth Wkshp. on Building Educational Applications Using NLP, pages 33-43, Montreal, Canada. Asso- ciation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "A new dataset and method for automatically grading ESOL texts", |
| "authors": [ |
| { |
| "first": "Helen", |
| "middle": [], |
| "last": "Yannakoudakis", |
| "suffix": "" |
| }, |
| { |
| "first": "Ted", |
| "middle": [], |
| "last": "Briscoe", |
| "suffix": "" |
| }, |
| { |
| "first": "Ben", |
| "middle": [], |
| "last": "Medlock", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proc. 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "180--189", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Helen Yannakoudakis, Ted Briscoe, and Ben Medlock. 2011. A new dataset and method for automatically grading ESOL texts. In Proc. 49th Annual Meet- ing of the Association for Computational Linguis- tics: Human Language Technologies, pages 180- 189, Portland, OR. Association for Computational Linguistics.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF1": { |
| "text": "Combining linguistic quality and coverage scores provided by human assessors inDUC- 2004", |
| "uris": null, |
| "num": null, |
| "type_str": "figure" |
| }, |
| "FIGREF2": { |
| "text": "new , H) and r(M new , M base ). If r(M base , H) and r(M new , H) are both > 0, then the third correlation, between metric scores themselves, r(M base , M new ), must also be > 0. The strength of this correlation, directly between scores of pairs of summarization metrics, should be taken into account using a significance test of the difference in correlation between r(M base , H) and r(M new , H).", |
| "uris": null, |
| "num": null, |
| "type_str": "figure" |
| }, |
| "FIGREF3": { |
| "text": "Summarization system pairwise significance test outcomes (paired t-test) for state-ofthe-art (top 7 rows) and baseline systems (bottom 5 rows) ofHong et al. (2014) evaluated with best-performing ROUGE variant: average ROUGE-2 precision with stemming and stop words removed, colored cells denote a significant greater mean score for row i system over column j system according to paired t-test.", |
| "uris": null, |
| "num": null, |
| "type_str": "figure" |
| }, |
| "TABREF0": { |
| "content": "<table><tr><td>Metric</td><td>Stemming</td><td>RSW</td><td>Ave./Med</td><td>P/R/F</td></tr><tr><td/><td/><td/><td/><td>r 13</td><td>2 \u2212 r 23</td><td>2 + 2r 12 r 13 r 23</td></tr></table>", |
| "num": null, |
| "type_str": "table", |
| "html": null, |
| "text": "Since the power of Williams test increases when the third correlation, r(M base , M new ), between metric scores is stronger, metrics should not be ranked by the number of competing metrics they" |
| }, |
| "TABREF1": { |
| "content": "<table/>", |
| "num": null, |
| "type_str": "table", |
| "html": null, |
| "text": "Pearson correlation (r) of BLEU and 192 variants of ROUGE (R-*) with human assessment in DUC" |
| }, |
| "TABREF2": { |
| "content": "<table/>", |
| "num": null, |
| "type_str": "table", |
| "html": null, |
| "text": "\u2022 identifies variants of ROUGE not significantly outperformed by any other variant. Figure 3 shows pairwise Williams significance test outcomes for BLEU, the top ten ROUGE variants, as well as current recommended ROUGE variants" |
| }, |
| "TABREF4": { |
| "content": "<table><tr><td colspan=\"3\">: Proportions of optimal ROUGE variants</td></tr><tr><td colspan=\"3\">attributed to each ROUGE configuration option</td></tr><tr><td>(%).</td><td/><td/></tr><tr><td/><td>ROUGE</td><td>ROUGE</td></tr><tr><td>System</td><td>Best</td><td>Original</td></tr><tr><td>DPP</td><td>8.498</td><td>9.62</td></tr><tr><td>ICSISumm</td><td>8.317</td><td>9.78</td></tr><tr><td>RegSum</td><td>8.187</td><td>9.75</td></tr><tr><td>Submodular</td><td>8.047</td><td>9.35</td></tr><tr><td>CLASSY11</td><td>7.717</td><td>9.20</td></tr><tr><td>CLASSY04</td><td>7.690</td><td>8.96</td></tr><tr><td>OCCAMS V</td><td>7.643</td><td>9.76</td></tr><tr><td>GreedyKL</td><td>6.918</td><td>8.53</td></tr><tr><td>FreqSum</td><td>6.838</td><td>8.11</td></tr><tr><td>TsSum</td><td>6.671</td><td>8.15</td></tr><tr><td>Centroid</td><td>6.660</td><td>7.97</td></tr><tr><td>LexRank</td><td>6.655</td><td>7.47</td></tr></table>", |
| "num": null, |
| "type_str": "table", |
| "html": null, |
| "text": "" |
| }, |
| "TABREF5": { |
| "content": "<table><tr><td>: Summarization systems originally in-</td></tr><tr><td>cluded in Hong et al. (2014) evaluated with the</td></tr><tr><td>best-performing ROUGE variant (Best): average</td></tr><tr><td>ROUGE-2 precision with stemming and stop words</td></tr><tr><td>removed; and evaluated with original suboptimal</td></tr><tr><td>variant (median ROUGE-2 recall with stemming</td></tr><tr><td>and without removal of stop-words)</td></tr></table>", |
| "num": null, |
| "type_str": "table", |
| "html": null, |
| "text": "" |
| } |
| } |
| } |
| } |