{ "paper_id": "P13-1020", "header": { "generated_with": "S2ORC 1.0.0", "date_generated": "2023-01-19T09:36:46.263530Z" }, "title": "Fast and Robust Compressive Summarization with Dual Decomposition and Multi-Task Learning", "authors": [ { "first": "Miguel", "middle": [ "B" ], "last": "Almeida", "suffix": "", "affiliation": { "laboratory": "", "institution": "Instituto Superior T\u00e9cnico", "location": { "postCode": "1049-001", "settlement": "Lisboa", "country": "Portugal" } }, "email": "" }, { "first": "Andr\u00e9", "middle": [ "F T" ], "last": "Martins", "suffix": "", "affiliation": { "laboratory": "", "institution": "Instituto Superior T\u00e9cnico", "location": { "postCode": "1049-001", "settlement": "Lisboa", "country": "Portugal" } }, "email": "" }, { "first": "Priberam", "middle": [], "last": "Labs", "suffix": "", "affiliation": { "laboratory": "", "institution": "Instituto Superior T\u00e9cnico", "location": { "postCode": "1049-001", "settlement": "Lisboa", "country": "Portugal" } }, "email": "" }, { "first": "Alameda", "middle": [ "D" ], "last": "Afonso Henriques", "suffix": "", "affiliation": { "laboratory": "", "institution": "Instituto Superior T\u00e9cnico", "location": { "postCode": "1049-001", "settlement": "Lisboa", "country": "Portugal" } }, "email": "" }, { "first": "", "middle": [], "last": "Lisboa", "suffix": "", "affiliation": { "laboratory": "", "institution": "Instituto Superior T\u00e9cnico", "location": { "postCode": "1049-001", "settlement": "Lisboa", "country": "Portugal" } }, "email": "" } ], "year": "", "venue": null, "identifiers": {}, "abstract": "We present a dual decomposition framework for multi-document summarization, using a model that jointly extracts and compresses sentences. Compared with previous work based on integer linear programming, our approach does not require external solvers, is significantly faster, and is modular in the three qualities a summary should have: conciseness, informativeness, and grammaticality. In addition, we propose a multi-task learning framework to take advantage of existing data for extractive summarization and sentence compression. Experiments in the TAC-2008 dataset yield the highest published ROUGE scores to date, with runtimes that rival those of extractive summarizers.", "pdf_parse": { "paper_id": "P13-1020", "_pdf_hash": "", "abstract": [ { "text": "We present a dual decomposition framework for multi-document summarization, using a model that jointly extracts and compresses sentences. Compared with previous work based on integer linear programming, our approach does not require external solvers, is significantly faster, and is modular in the three qualities a summary should have: conciseness, informativeness, and grammaticality. In addition, we propose a multi-task learning framework to take advantage of existing data for extractive summarization and sentence compression. Experiments in the TAC-2008 dataset yield the highest published ROUGE scores to date, with runtimes that rival those of extractive summarizers.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Abstract", "sec_num": null } ], "body_text": [ { "text": "Automatic text summarization is a seminal problem in information retrieval and natural language processing (Luhn, 1958; Baxendale, 1958; Edmundson, 1969) . Today, with the overwhelming amount of information available on the Web, the demand for fast, robust, and scalable summarization systems is stronger than ever.", "cite_spans": [ { "start": 107, "end": 119, "text": "(Luhn, 1958;", "ref_id": "BIBREF25" }, { "start": 120, "end": 136, "text": "Baxendale, 1958;", "ref_id": "BIBREF1" }, { "start": 137, "end": 153, "text": "Edmundson, 1969)", "ref_id": "BIBREF12" } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "Up to now, extractive systems have been the most popular in multi-document summarization. These systems produce a summary by extracting a representative set of sentences from the original documents (Kupiec et al., 1995; Carbonell and Goldstein, 1998; Radev et al., 2000; Gillick et al., 2008) . This approach has obvious advantages: it reduces the search space by letting decisions be made for each sentence as a whole (avoiding finegrained text generation), and it ensures a grammatical summary, assuming the original sentences are well-formed. The typical trade-offs in these mod-els (maximizing relevance, and penalizing redundancy) lead to submodular optimization problems (Lin and Bilmes, 2010) , which are NP-hard but approximable through greedy algorithms; learning is possible with standard structured prediction algorithms (Sipos et al., 2012; Lin and Bilmes, 2012) . Probabilistic models have also been proposed to capture the problem structure, such as determinantal point processes (Gillenwater et al., 2012) .", "cite_spans": [ { "start": 198, "end": 219, "text": "(Kupiec et al., 1995;", "ref_id": "BIBREF20" }, { "start": 220, "end": 250, "text": "Carbonell and Goldstein, 1998;", "ref_id": "BIBREF4" }, { "start": 251, "end": 270, "text": "Radev et al., 2000;", "ref_id": "BIBREF34" }, { "start": 271, "end": 292, "text": "Gillick et al., 2008)", "ref_id": "BIBREF17" }, { "start": 677, "end": 699, "text": "(Lin and Bilmes, 2010)", "ref_id": "BIBREF21" }, { "start": 832, "end": 852, "text": "(Sipos et al., 2012;", "ref_id": "BIBREF38" }, { "start": 853, "end": 874, "text": "Lin and Bilmes, 2012)", "ref_id": "BIBREF22" }, { "start": 994, "end": 1020, "text": "(Gillenwater et al., 2012)", "ref_id": "BIBREF16" } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "However, extractive systems are rather limited in the summaries they can produce. Long, partly relevant sentences tend not to appear in the summary, or to block the inclusion of other sentences. This has motivated research in compressive summarization (Lin, 2003; Zajic et al., 2006; Daum\u00e9, 2006) , where summaries are formed by compressed sentences (Knight and Marcu, 2000) , not necessarily extracts. While promising results have been achieved by models that simultaneously extract and compress (Martins and Smith, 2009; Woodsend and Lapata, 2010; Berg-Kirkpatrick et al., 2011) , there are still obstacles that need to be surmounted for these systems to enjoy wide adoption. All approaches above are based on integer linear programming (ILP), suffering from slow runtimes, when compared to extractive systems. For example, Woodsend and Lapata (2012) report 55 seconds on average to produce a summary; Berg-Kirkpatrick et al. (2011) report substantially faster runtimes, but fewer compressions are allowed. Having a compressive summarizer which is both fast and expressive remains an open problem. A second inconvenience of ILP-based approaches is that they do not exploit the modularity of the problem, since the declarative specification required by ILP solvers discards important structural information. For example, such solvers are unable to take advantage of efficient dynamic programming routines for sentence compression (McDonald, 2006) . This paper makes progress in two fronts:", "cite_spans": [ { "start": 252, "end": 263, "text": "(Lin, 2003;", "ref_id": "BIBREF23" }, { "start": 264, "end": 283, "text": "Zajic et al., 2006;", "ref_id": "BIBREF47" }, { "start": 284, "end": 296, "text": "Daum\u00e9, 2006)", "ref_id": "BIBREF10" }, { "start": 350, "end": 374, "text": "(Knight and Marcu, 2000)", "ref_id": "BIBREF18" }, { "start": 497, "end": 522, "text": "(Martins and Smith, 2009;", "ref_id": "BIBREF26" }, { "start": 523, "end": 549, "text": "Woodsend and Lapata, 2010;", "ref_id": "BIBREF42" }, { "start": 550, "end": 580, "text": "Berg-Kirkpatrick et al., 2011)", "ref_id": "BIBREF2" }, { "start": 826, "end": 852, "text": "Woodsend and Lapata (2012)", "ref_id": "BIBREF44" }, { "start": 1431, "end": 1447, "text": "(McDonald, 2006)", "ref_id": "BIBREF30" } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "\u2022 We derive a dual decomposition framework for extractive and compressive summarization ( \u00a72-3). Not only is this framework orders of magnitude more efficient than the ILP-based approaches, it also allows the three well-known metrics of summaries-conciseness, informativeness, and grammaticality-to be treated separately in a modular fashion (see Figure 1 ). We also contribute with a novel knapsack factor, along with a linear-time algorithm for the corresponding dual decomposition subproblem.", "cite_spans": [], "ref_spans": [ { "start": 347, "end": 355, "text": "Figure 1", "ref_id": null } ], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "\u2022 We propose multi-task learning ( \u00a74) as a principled way to train compressive summarizers, using auxiliary data for extractive summarization and sentence compression. To this end, we adapt the framework of Evgeniou and Pontil (2004) and Daum\u00e9 (2007) to train structured predictors that share some of their parts.", "cite_spans": [ { "start": 208, "end": 234, "text": "Evgeniou and Pontil (2004)", "ref_id": "BIBREF13" }, { "start": 239, "end": 251, "text": "Daum\u00e9 (2007)", "ref_id": "BIBREF11" } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "Experiments on TAC data ( \u00a75) yield state-of-theart results, with runtimes similar to that of extractive systems. To our best knowledge, this had never been achieved by compressive summarizers.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "In extractive summarization, we are given a set of sentences D := {s 1 , . . . , s N } belonging to one or more documents, and the goal is to extract a subset S \u2286 D that conveys a good summary of D and whose total number of words does not exceed a prespecified budget B. We use an indicator vector y := y n N n=1 to represent an extractive summary, where y n = 1 if s n \u2208 S, and y n = 0 otherwise. Let L n be the number of words of the nth sentence. By designing a quality score function g : {0, 1} N \u2192 R, this can be cast as a global optimization problem with a knapsack constraint:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Extractive Summarization", "sec_num": "2" }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": "maximize g(y) w.r.t. y \u2208 {0, 1} N s.t. N n=1 L n y n \u2264 B.", "eq_num": "(1)" } ], "section": "Extractive Summarization", "sec_num": "2" }, { "text": "Intuitively, a good summary is one which selects sentences that individually convey \"relevant\" information, while collectively having small \"redundancy.\" This trade-off was explicitly modeled in early works through the notion of maximal marginal relevance (Carbonell and Goldstein, 1998; McDonald, 2007) . An alternative are coverage-based models ( \u00a72.1; Filatova and Hatzivassiloglou, 2004; Yih et al., 2007; Gillick et al., 2008) , which seek a set of sentences that covers as many diverse \"concepts\" as possible; redundancy is automatically penalized since redundant sentences cover fewer concepts. Both models can be framed under the framework of submodular optimization (Lin and Bilmes, 2010) , leading to greedy algorithms that have approximation guarantees. However, extending these models to allow for sentence compression (as will be detailed in \u00a73) breaks the diminishing returns property, making submodular optimization no longer applicable.", "cite_spans": [ { "start": 256, "end": 287, "text": "(Carbonell and Goldstein, 1998;", "ref_id": "BIBREF4" }, { "start": 288, "end": 303, "text": "McDonald, 2007)", "ref_id": "BIBREF31" }, { "start": 355, "end": 391, "text": "Filatova and Hatzivassiloglou, 2004;", "ref_id": "BIBREF14" }, { "start": 392, "end": 409, "text": "Yih et al., 2007;", "ref_id": "BIBREF45" }, { "start": 410, "end": 431, "text": "Gillick et al., 2008)", "ref_id": "BIBREF17" }, { "start": 675, "end": 697, "text": "(Lin and Bilmes, 2010)", "ref_id": "BIBREF21" } ], "ref_spans": [], "eq_spans": [], "section": "Extractive Summarization", "sec_num": "2" }, { "text": "Coverage-based extractive summarization can be formalized as follows. Let C(D) := {c 1 , . . . , c M } be a set of relevant concept types which are present in the original documents D. 1 Let \u03c3 m be a relevance score assigned to the mth concept, and let the set I m \u2286 {1, . . . , N } contain the indices of the sentences in which this concept occurs. Then, the following quality score function is defined:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Coverage-Based Summarization", "sec_num": "2.1" }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": "g(y) = M m=1 \u03c3 m u m (y),", "eq_num": "(2)" } ], "section": "Coverage-Based Summarization", "sec_num": "2.1" }, { "text": "where u m (y) := n\u2208Im y n is a Boolean function that indicates whether the mth concept is present in the summary. Plugging this into Eq. 1, one obtains the following Boolean optimization problem:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Coverage-Based Summarization", "sec_num": "2.1" }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": "maximize M m=1 \u03c3 m u m w.r.t. y \u2208 {0, 1} N , u \u2208 {0, 1} M s.t. u m = n\u2208Im y n , \u2200m \u2208 [M ] N n=1 L n y n \u2264 B,", "eq_num": "(3)" } ], "section": "Coverage-Based Summarization", "sec_num": "2.1" }, { "text": "where we used the notation [M ] := {1, . . . , M }. This can be converted into an ILP and addressed with off-the-shelf solvers (Gillick et al., 2008) . A drawback of this approach is that solving an ILP exactly is NP-hard. Even though existing commercial solvers can solve most instances with a moderate speed, they still exhibit poor worst-case behaviour; this is exacerbated when there is the need to combine an extractive component with other modules, as in compressive summarization ( \u00a73).", "cite_spans": [ { "start": 127, "end": 149, "text": "(Gillick et al., 2008)", "ref_id": "BIBREF17" } ], "ref_spans": [], "eq_spans": [], "section": "Coverage-Based Summarization", "sec_num": "2.1" }, { "text": "We next describe how the problem in Eq. 3 can be addressed with dual decomposition, a class of optimization techniques that tackle the dual of combinatorial problems in a modular, extensible, and parallelizable manner (Komodakis et al., 2007; Rush et al., 2010) . In particular, we employ alternating directions dual decomposition (AD 3 ; Martins et al., 2011a for solving a linear relaxation of Eq. 3. AD 3 resembles the subgradientbased algorithm of Rush et al. (2010) , but it enjoys a faster convergence rate. Both algorithms split the original problem into several components, and then iterate between solving independent local subproblems at each component and adjusting multipliers to promote an agreement. 2 The difference between the two methods is that the AD 3 local subproblems, instead of requiring the computation of a locally optimal configuration, require solving a local quadratic problem. Martins et al. (2011b) provided linear-time solutions for several logic constraints, with applications to syntax and frame-semantic parsing (Das et al., 2012) . We will see that AD 3 can also handle budget and knapsack constraints efficiently. To tackle Eq. 3 with dual decomposition, we split the coverage-based summarizer into the following M + 1 components (one per constraint):", "cite_spans": [ { "start": 218, "end": 242, "text": "(Komodakis et al., 2007;", "ref_id": "BIBREF19" }, { "start": 243, "end": 261, "text": "Rush et al., 2010)", "ref_id": "BIBREF36" }, { "start": 339, "end": 360, "text": "Martins et al., 2011a", "ref_id": "BIBREF27" }, { "start": 452, "end": 470, "text": "Rush et al. (2010)", "ref_id": "BIBREF36" }, { "start": 714, "end": 715, "text": "2", "ref_id": null }, { "start": 907, "end": 929, "text": "Martins et al. (2011b)", "ref_id": "BIBREF28" }, { "start": 1047, "end": 1065, "text": "(Das et al., 2012)", "ref_id": "BIBREF9" } ], "ref_spans": [], "eq_spans": [], "section": "A Dual Decomposition Formulation", "sec_num": "2.2" }, { "text": "1. For each of the M concepts in C(D), one component for imposing the logic constraint in Eq. 3. This corresponds to the OR-WITH-OUTPUT factor described by Martins et al. (2011b) ; the AD 3 subproblem for the mth factor can be solved in time O(|I m |).", "cite_spans": [ { "start": 156, "end": 178, "text": "Martins et al. (2011b)", "ref_id": "BIBREF28" } ], "ref_spans": [], "eq_spans": [], "section": "A Dual Decomposition Formulation", "sec_num": "2.2" }, { "text": "2. Another component for the knapsack constraint. This corresponds to a (novel) KNAP-SACK factor, whose AD 3 subproblem is solvable in time O(N ). The actual algorithm is described in the appendix (Algorithm 1). 3", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Dual Decomposition Formulation", "sec_num": "2.2" }, { "text": "We now turn to compressive summarization, which does not limit the summary sentences to be verbatim extracts from the original documents; in-2 For details about dual decomposition and Lagrangian relaxation, see the recent tutorial by Rush and Collins (2012) .", "cite_spans": [ { "start": 234, "end": 257, "text": "Rush and Collins (2012)", "ref_id": "BIBREF35" } ], "ref_spans": [], "eq_spans": [], "section": "Compressive Summarization", "sec_num": "3" }, { "text": "3 The AD 3 subproblem in this case corresponds to computing an Euclidean projection onto the knapsack polytope (Eq. 11). Others addressed the related, but much harder, integer quadratic knapsack problem (McDonald, 2007) . stead, it allows the extraction of compressed sentences where some words can be deleted.", "cite_spans": [ { "start": 203, "end": 219, "text": "(McDonald, 2007)", "ref_id": "BIBREF31" } ], "ref_spans": [], "eq_spans": [], "section": "Compressive Summarization", "sec_num": "3" }, { "text": "Formally, let us express each sentence of D as a sequence of word tokens, s n := t n, Ln =0 , where t n,0 \u2261 $ is a dummy symbol. We represent a compression of s n as an indicator vector z n := z n, Ln =0 , where z n, = 1 if the th word is included in the compression. By convention, the dummy symbol is included if and only if the remaining compression is non-empty. A compressive summary can then be represented by an indicator vector z which is the concatenation of N such vectors, z = z 1 , . . . , z N ; each position in this indicator vector is indexed by a sentence n \u2208 [N ] and a word position", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Compressive Summarization", "sec_num": "3" }, { "text": "\u2208 {0} \u222a [L n ].", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Compressive Summarization", "sec_num": "3" }, { "text": "Models for compressive summarization were proposed by Martins and Smith (2009) and Berg-Kirkpatrick et al. 2011by combining extraction and compression scores. Here, we follow the latter work, by combining a coverage score function g with sentence-level compression score functions h 1 , . . . , h N . This yields the decoding problem:", "cite_spans": [ { "start": 54, "end": 78, "text": "Martins and Smith (2009)", "ref_id": "BIBREF26" } ], "ref_spans": [], "eq_spans": [], "section": "Compressive Summarization", "sec_num": "3" }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": "maximize g(z) + N n=1 h n (z n ) w.r.t. z n \u2208 {0, 1} Ln , \u2200n \u2208 [N ] s.t. N n=1 Ln =1 z n, \u2264 B.", "eq_num": "(4)" } ], "section": "Compressive Summarization", "sec_num": "3" }, { "text": "We use a coverage function similar to Eq. 2, but taking a compressive summary z as argument:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Coverage Model", "sec_num": "3.1" }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": "g(z) = M m=1 \u03c3 m u m (z),", "eq_num": "(5)" } ], "section": "Coverage Model", "sec_num": "3.1" }, { "text": "where we redefine u m as follows. First, we parametrize each occurrence of the mth concept (assumed to be a k-gram) as a triple n, s , e , where n indexes a sentence, s indexes a start position within the sentence, and e indexes the end position. We denote by T m the set of triples representing all occurrences of the mth concept in the original text, and we associate an indicator variable z n, s: e to each member of this set. We then define u m (z) via the following logic constraints:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Coverage Model", "sec_num": "3.1" }, { "text": "\u2022 A concept type is selected if some of its k-gram tokens are selected:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Coverage Model", "sec_num": "3.1" }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": "u m (y) := n, s, e \u2208Tm z n, s: e .", "eq_num": "(6)" } ], "section": "Coverage Model", "sec_num": "3.1" }, { "text": "\u2022 A k-gram concept token is selected if all its words are selected: Figure 1 : Components of our compressive summarizer. Factors depicted in blue belong to the compression model, and aim to enforce grammaticality. The logic factors in red form the coverage component. Finally, the budget factor, in green, is connected to the word nodes; it ensures that the summary fits the word limit. Shaded circles represent active variables while white circles represent inactive variables.", "cite_spans": [], "ref_spans": [ { "start": 68, "end": 76, "text": "Figure 1", "ref_id": null } ], "eq_spans": [], "section": "Coverage Model", "sec_num": "3.1" }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": "z n, s: e := e = s z n, .", "eq_num": "(7)" } ], "section": "Coverage Model", "sec_num": "3.1" }, { "text": "We set concept scores as \u03c3", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Coverage Model", "sec_num": "3.1" }, { "text": "m := w \u2022 \u03a6 cov (D, c m ), where \u03a6 cov (D, c m )", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Coverage Model", "sec_num": "3.1" }, { "text": "is a vector of features (described in \u00a73.5) and w the corresponding weights.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Coverage Model", "sec_num": "3.1" }, { "text": "For the compression score function, we follow Martins and Smith (2009) and decompose it as a sum of local score functions \u03c1 n, defined on dependency arcs:", "cite_spans": [ { "start": 46, "end": 70, "text": "Martins and Smith (2009)", "ref_id": "BIBREF26" } ], "ref_spans": [], "eq_spans": [], "section": "Compression Model", "sec_num": "3.2" }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": "h n (z n ) := Ln =1 \u03c1 n, (z n, , z n,\u03c0( ) ),", "eq_num": "(8)" } ], "section": "Compression Model", "sec_num": "3.2" }, { "text": "where \u03c0( ) denotes the index of the word which is the parent of the th word in the dependency tree (by convention, the root of the tree is the dummy symbol). To model the event that an arc is \"cut\" by disconnecting a child from its head, we define arc-deletion scores \u03c1 n, (0, 1) :", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Compression Model", "sec_num": "3.2" }, { "text": "= w \u2022 \u03a6 comp (s n , , \u03c0( )),", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Compression Model", "sec_num": "3.2" }, { "text": "where \u03a6 comp is a feature map, which is described in detail in \u00a73.5. We set \u03c1 n, (0, 0) = \u03c1 n, (1, 1) = 0, and \u03c1 n, (1, 0) = \u2212\u221e, to allow only the deletion of entire subtrees. A crucial fact is that one can maximize Eq. 8 efficiently with dynamic programming (using the Viterbi algorithm for trees); the total cost is linear in L n . We will exploit this fact in the dual decomposition framework described next. 4", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Compression Model", "sec_num": "3.2" }, { "text": "In previous work, the optimization problem in Eq. 4 was converted into an ILP and fed to an offthe-shelf solver (Martins and Smith, 2009; Berg-Kirkpatrick et al., 2011; Woodsend and Lapata, 2012) . Here, we employ the AD 3 algorithm, in a similar manner as described in \u00a72, but with an additional component for the sentence compressor, and slight modifications in the other components. We have the following N + M + M m=1 |T m | + 1 components in total, illustrated in Figure 1: 1. For each of the N sentences, one component for the compression model. The AD 3 quadratic subproblem for this factor can be addressed by solving a sequence of linear subproblems, as described by . Each of these subproblems corresponds to maximizing an objective function of the same form as Eq. 8; this can be done in O(L n ) time with dynamic programming, as discussed in \u00a73.2.", "cite_spans": [ { "start": 112, "end": 137, "text": "(Martins and Smith, 2009;", "ref_id": "BIBREF26" }, { "start": 138, "end": 168, "text": "Berg-Kirkpatrick et al., 2011;", "ref_id": "BIBREF2" }, { "start": 169, "end": 195, "text": "Woodsend and Lapata, 2012)", "ref_id": "BIBREF44" } ], "ref_spans": [ { "start": 469, "end": 478, "text": "Figure 1:", "ref_id": null } ], "eq_spans": [], "section": "A Dual Decomposition Formulation", "sec_num": "3.3" }, { "text": "2. For each of the M concept types in C(D), one OR-WITH-OUTPUT factor for the logic constraint in Eq. 6. This is analogous to the one described for the extractive case.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Dual Decomposition Formulation", "sec_num": "3.3" }, { "text": "3. For each k-gram concept token in T m , one AND-WITH-OUTPUT factor that imposes the constraint in Eq. 7. This factor was described by Martins et al. (2011b) and its AD 3 subproblem can be solved in time linear in k.", "cite_spans": [ { "start": 136, "end": 158, "text": "Martins et al. (2011b)", "ref_id": "BIBREF28" } ], "ref_spans": [], "eq_spans": [], "section": "A Dual Decomposition Formulation", "sec_num": "3.3" }, { "text": "4. Another component linked to all the words imposing that at most B words can be selected; this is done via a BUDGET factor, a particular case of KNAPSACK. The runtime of this AD 3 subproblem is linear in the number of words.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Dual Decomposition Formulation", "sec_num": "3.3" }, { "text": "In addition, we found it useful to add a second BUDGET factor limiting the number of sentences that can be selected to a prescribed value K. We set K = 6 in our experiments.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Dual Decomposition Formulation", "sec_num": "3.3" }, { "text": "Recall that the problem in Eq. 4 is NP-hard, and that AD 3 is solving a linear relaxation. While there are ways of wrapping AD 3 in an exact search algorithm (Das et al., 2012) , such strategies work best when the solution of the relaxation has few fractional components, which is typical of parsing and translation problems (Rush et al., 2010; Chang and Collins, 2011) , and attractive networks (Taskar et al., 2004) . Unfortunately, this is not the case in summarization, where concepts \"compete\" with each other for inclusion in the summary, leading to frustrated cycles. We chose instead to adopt a fast and simple rounding procedure for obtaining a summary from a fractional solution.", "cite_spans": [ { "start": 158, "end": 176, "text": "(Das et al., 2012)", "ref_id": "BIBREF9" }, { "start": 325, "end": 344, "text": "(Rush et al., 2010;", "ref_id": "BIBREF36" }, { "start": 345, "end": 369, "text": "Chang and Collins, 2011)", "ref_id": "BIBREF5" }, { "start": 396, "end": 417, "text": "(Taskar et al., 2004)", "ref_id": "BIBREF40" } ], "ref_spans": [], "eq_spans": [], "section": "Rounding Strategy", "sec_num": "3.4" }, { "text": "The procedure works as follows. First, solve the LP relaxation using AD 3 , as described above. This yields a solution z * , where each component lies in the unit interval [0, 1]. If these components are all integer, then we have a certificate that this is the optimal solution. Otherwise, we collect the K sentences with the highest values of z * n,0 (\"posteriors\" on sentences), and seek the feasible summary which is the closest (in Euclidean distance) to z * , while only containing those sentences. This can be computed exactly in time O(B K k=1 L n k ), through dynamic programming. 5", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Rounding Strategy", "sec_num": "3.4" }, { "text": "As Berg-Kirkpatrick et al. (2011), we used stemmed word bigrams as concepts, to which we associate the following concept features (\u03a6 cov ): indicators for document counts, features indicating if each of the words in the bigram is a stopword, the earliest position in a document each concept occurs, as well as two and three-way conjunctions of these features.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Features and Hard Constraints", "sec_num": "3.5" }, { "text": "For the compression model, we include the following arc-deletion features (\u03a6 comp ):", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Features and Hard Constraints", "sec_num": "3.5" }, { "text": "\u2022 the dependency label of the arc being deleted, as well as its conjunction with the part-of-speech tag of the head, of the modifier, and of both;", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Features and Hard Constraints", "sec_num": "3.5" }, { "text": "\u2022 the dependency labels of the arc being deleted and of its parent arc;", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Features and Hard Constraints", "sec_num": "3.5" }, { "text": "\u2022 the modifier tag, if the modifier is a function word modifying a verb ;", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Features and Hard Constraints", "sec_num": "3.5" }, { "text": "\u2022 a feature indicating whether the modifier or any of its descendants is a negation word;", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Features and Hard Constraints", "sec_num": "3.5" }, { "text": "\u2022 indicators of whether the modifier is a temporal word (e.g., Friday) or a preposition pointing to a temporal word (e.g., on Friday).", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Features and Hard Constraints", "sec_num": "3.5" }, { "text": "In addition, we included hard constraints to prevent the deletion of certain arcs, following previous work in sentence compression (Clarke and Lapata, 2008) . We never delete arcs whose dependency label is SUB, OBJ, PMOD, SBAR, VC, or PRD (this makes sure we preserve subjects and objects of verbs, arcs departing from prepositions or complementizers, and that we do not break verb chains or predicative complements); arcs linking to a conjunction word or siblings of such arcs (to prevent inconsistencies in handling coordinative conjunctions); arcs linking verbs to other verbs, to adjectives (e.g., make available), to verb particles (e.g., settle down), to the word that (e.g., said that), or to the word to if it is a leaf (e.g., allowed to come); arcs pointing to negation words, cardinal numbers, or determiners; and arcs connecting two proper nouns or words within quotation marks.", "cite_spans": [ { "start": 131, "end": 156, "text": "(Clarke and Lapata, 2008)", "ref_id": "BIBREF7" } ], "ref_spans": [], "eq_spans": [], "section": "Features and Hard Constraints", "sec_num": "3.5" }, { "text": "We next turn to the problem of learning the model from training data. Prior work in compressive summarization has followed one of two strategies: Martins and Smith (2009) and Woodsend and Lapata (2012) learn the extraction and compression models separately, and then post-combine them, circumventing the lack of fully annotated data. Berg-Kirkpatrick et al. (2011) gathered a small dataset of manually compressed summaries, and trained with full supervision. While the latter approach is statistically more principled, it has the disadvantage of requiring fully annotated data, which is difficult to obtain in large quantities. On the other hand, there is plenty of data containing manually written abstracts (from the DUC and TAC conferences) and user-generated text (from Wikipedia) that may provide useful weak supervision. With this in mind, we put together a multi-task learning framework for compressive summarization (which we name task #1). The goal is to take advantage of existing data for related tasks, such as extractive summarization (task #2), and sentence compression (task #3). The three tasks are instances of structured predictors (Bak\u0131r et al., 2007) , and for all of them we assume featurebased models that decompose over \"parts\":", "cite_spans": [ { "start": 146, "end": 170, "text": "Martins and Smith (2009)", "ref_id": "BIBREF26" }, { "start": 1150, "end": 1159, "text": "(Bak\u0131r et", "ref_id": null }, { "start": 1160, "end": 1170, "text": "al., 2007)", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "Multi-Task Learning", "sec_num": "4" }, { "text": "\u2022 For the compressive summarization task, the parts correspond to concept features ( \u00a73.1) and to arc-deletion features ( \u00a73.2).", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Multi-Task Learning", "sec_num": "4" }, { "text": "\u2022 For the extractive summarization task, there are parts for concept features only.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Multi-Task Learning", "sec_num": "4" }, { "text": "\u2022 For the sentence compression task, the parts correspond to arc-deletion features only. This is summarized in Table 1 . Features for the three tasks are populated into feature vectors \u03a6 1 (x, y), \u03a6 2 (x, y), and \u03a6 3 (x, y), respectively, where x, y denotes a task-specific input-output pair. We assume the feature vectors are all D dimensional, where we place zeros in entries corresponding to parts that are absent. Note that this setting is very general and applies to arbitrary structured prediction problems (not just summarization), the only assumption being that some parts are shared between different tasks. Next, we associate weight vectors v 1 , v 2 , v 3 \u2208 R D to each task, along with a \"shared\" vector w. Each task makes predictions according to the rule:", "cite_spans": [], "ref_spans": [ { "start": 111, "end": 118, "text": "Table 1", "ref_id": "TABREF2" } ], "eq_spans": [], "section": "Multi-Task Learning", "sec_num": "4" }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": "y := arg max y (w + v k ) \u2022 \u03a6 k (x, y),", "eq_num": "(9)" } ], "section": "Multi-Task Learning", "sec_num": "4" }, { "text": "where k \u2208 {1, 2, 3}. This setting is equivalent to the approach of Daum\u00e9 (2007) for domain adaptation, which consists in splitting each feature into task-component features and a shared feature; but here we do not duplicate features explicitly. To learn the weights, we regularize the weight vectors separately, and assume that each task has its own loss function L k , so that the total loss L is a weighted sum", "cite_spans": [ { "start": 67, "end": 79, "text": "Daum\u00e9 (2007)", "ref_id": "BIBREF11" } ], "ref_spans": [], "eq_spans": [], "section": "Multi-Task Learning", "sec_num": "4" }, { "text": "L(w, v 1 , v 2 , v 3 ) := 3 k=1 \u03c3 k L k (w + v k )", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Multi-Task Learning", "sec_num": "4" }, { "text": ". This yields the following objective function to be minimized:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Multi-Task Learning", "sec_num": "4" }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": "F (w, v 1 , v 2 , v 3 ) = \u03bb 2 w 2 + 3 k=1 \u03bb k 2 v k 2 + 1 N 3 k=1 \u03c3 k L k (w + v k ),", "eq_num": "(10)" } ], "section": "Multi-Task Learning", "sec_num": "4" }, { "text": "where \u03bb and the \u03bb k 's are regularization constants, and N is the total number of training instances. 6 In our experiments ( \u00a75), we let the L k 's be structured hinge losses (Taskar et al., 2003; Tsochantaridis et al., 2004) , where the corresponding cost functions are concept recall (for task #2), precision of arc deletions (for task #3), and a combination thereof (for task #1). 7 These losses were normalized, and we set \u03c3 k = N/N k , where N k is the number of training instances for the kth task. This ensures all tasks are weighted evenly. We used the same rationale to set \u03bb = \u03bb 1 = \u03bb 2 = \u03bb 3 , choosing this value through cross-validation in the dev set. We optimize Eq. 10 with stochastic subgradient descent. This leads to update rules of the form", "cite_spans": [ { "start": 102, "end": 103, "text": "6", "ref_id": null }, { "start": 175, "end": 196, "text": "(Taskar et al., 2003;", "ref_id": "BIBREF39" }, { "start": 197, "end": 225, "text": "Tsochantaridis et al., 2004)", "ref_id": "BIBREF41" } ], "ref_spans": [], "eq_spans": [], "section": "Multi-Task Learning", "sec_num": "4" }, { "text": "w \u2190 (1 \u2212 \u03b7 t \u03bb)w \u2212 \u03b7 t \u03c3 k\u2207 L k (w + v k ) v j \u2190 (1 \u2212 \u03b7 t \u03bb j )v j \u2212 \u03b7 t \u03b4 jk \u03c3 k\u2207 L k (w + v k ),", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Multi-Task Learning", "sec_num": "4" }, { "text": "where\u2207L k are stochastic subgradients for the kth task, that take only a single instance into account, and \u03b4 jk = 1 if and only if j = k. Stochastic subgradients can be computed via cost-augmented decoding (see footnote 7).", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Multi-Task Learning", "sec_num": "4" }, { "text": "Interestingly, Eq. 10 subsumes previous approaches to train compressive summarizers. The limit \u03bb \u2192 \u221e (keeping the \u03bb k 's fixed) forces w \u2192 0, decoupling all the tasks. In this limit, inference for task #1 (compressive summarization) is based solely on the model learned from that task's data, recovering the approach of Berg-Kirkpatrick et al. (2011) . In the other extreme, setting \u03c3 1 = 0 simply ignores task #1's training data. As a result, the optimal v 1 will be a vector of zeros; since tasks #2 and #3 have no parts in common, the objective will decouple into a sum of two independent terms 6 Note that, by substituting u k := w + v k and solving for w, the problem in Eq. 10 becomes that of minimizing the sum of the losses with a penalty for the (weighted) variance of the vectors {0, u1, u2, u3}, regularizing the difference towards their average, as in Evgeniou and Pontil (2004) . This is similar to the hierarchical joint learning approach of Finkel and Manning (2010) , except that our goal is to learn a new task (compressive summarization) instead of combining tasks.", "cite_spans": [ { "start": 320, "end": 350, "text": "Berg-Kirkpatrick et al. (2011)", "ref_id": "BIBREF2" }, { "start": 864, "end": 890, "text": "Evgeniou and Pontil (2004)", "ref_id": "BIBREF13" }, { "start": 956, "end": 981, "text": "Finkel and Manning (2010)", "ref_id": "BIBREF15" } ], "ref_spans": [], "eq_spans": [], "section": "Multi-Task Learning", "sec_num": "4" }, { "text": "7 Let Y k denote the output set for the kth task. Given a task-specific cost function \u2206 k : Y k \u00d7 Y k \u2192 R, and letting xt, yt T t=1 be the labeled dataset for this task, the structured hinge loss takes the form", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Multi-Task Learning", "sec_num": "4" }, { "text": "L k (u k ) := t max y \u2208Y k (u k \u2022 (\u03a6 k (xt, y ) \u2212 \u03a6 k (xt, yt)) + \u2206 k (y , yt)).", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Multi-Task Learning", "sec_num": "4" }, { "text": "The inner maximization over y is called the cost-augmented decoding problem: it differs from Eq. 9 by the inclusion of the cost term \u2206 k (y , yt). Our costs decompose over the model's factors, hence any decoder for Eq. 9 can be used for the maximization above: for tasks #1-#2, we solve a relaxation by running AD 3 without rounding, and for task #3 we use dynamic programming; see Table 1. involving v 2 and v 3 , which is equivalent to training the two tasks separately and post-combining the models, as Martins and Smith (2009) did.", "cite_spans": [ { "start": 506, "end": 530, "text": "Martins and Smith (2009)", "ref_id": "BIBREF26" } ], "ref_spans": [ { "start": 382, "end": 390, "text": "Table 1.", "ref_id": "TABREF2" } ], "eq_spans": [], "section": "Multi-Task Learning", "sec_num": "4" }, { "text": "We evaluated our compressive summarizers on data from the Text Analysis Conference (TAC) evaluations. We use the same splits as previous work (Berg-Kirkpatrick et al., 2011; Woodsend and Lapata, 2012) : the non-update portions of TAC-2009 for training and TAC-2008 for testing. In addition, we reserved TAC-2010 as a devset. The test partition contains 48 multi-document summarization problems; each provides 10 related news articles as input, and asks for a summary with up to 100 words, which is evaluated against four manually written abstracts. We ignored all the query information present in the TAC datasets.", "cite_spans": [ { "start": 142, "end": 173, "text": "(Berg-Kirkpatrick et al., 2011;", "ref_id": "BIBREF2" }, { "start": 174, "end": 200, "text": "Woodsend and Lapata, 2012)", "ref_id": "BIBREF44" } ], "ref_spans": [], "eq_spans": [], "section": "Experimental setup", "sec_num": "5.1" }, { "text": "Single-Task Learning. In the single-task experiments, we trained a compressive summarizer on the dataset disclosed by Berg-Kirkpatrick et al. 2011, which contains manual compressive summaries for the TAC-2009 data. We trained a structured SVM with stochastic subgradient descent; the cost-augmented inference problems are relaxed and solved with AD 3 , as described in \u00a73.3. 8 We followed the procedure described in Berg-Kirkpatrick et al. (2011) to reduce the number of candidate sentences: scores were defined for each sentence (the sum of the scores of the concepts they cover), and the best-scored sentences were greedily selected up to a limit of 1,000 words. We then tagged and parsed the selected sentences with TurboParser. 9 Our choice of a dependency parser was motivated by our will for a fast system; in particular, TurboParser attains top accuracies at a rate of 1,200 words per second, keeping parsing times below 1 second for each summarization problem.", "cite_spans": [ { "start": 375, "end": 376, "text": "8", "ref_id": null }, { "start": 416, "end": 446, "text": "Berg-Kirkpatrick et al. (2011)", "ref_id": "BIBREF2" } ], "ref_spans": [], "eq_spans": [], "section": "Experimental setup", "sec_num": "5.1" }, { "text": "Multi-Task Learning. For the multi-task experiments, we also used the dataset of Berg-Kirkpatrick et al. (2011), but we augmented the training data with extractive summarization and sentence compression datasets, to help train the compressive summarizer. For extractive summarization, we used the DUC 2003 and 2004 datasets (a total of 80 multi-document summarization problems). We generated oracle extracts by maximizing bigram recall with respect to the manual abstracts, as described in Berg-Kirkpatrick et al. (2011) . For sentence compression, we adapted the Simple English Wikipedia dataset of Woodsend and Lapata (2011), containing aligned sentences for 15,000 articles from the English and Simple English Wikipedias. We kept only the 4,481 sentence pairs corresponding to deletionbased compressions. Table 2 shows the results. The top rows refer to three strong baselines: the ICSI-1 extractive coverage-based system of Gillick et al. (2008) , which achieved the best ROUGE scores in the TAC-2008 evaluation; the compressive summarizer of Berg-Kirkpatrick et al. 2011, denoted BGK'11; and the multi-aspect compressive summarizer of Woodsend and Lapata (2012) , denoted WL'12. All these systems require ILP solvers. The bottom rows show the results achieved by our implementation of a pure extractive system (similar to the learned extractive summarizer of Berg-Kirkpatrick et al., 2011); a system that postcombines extraction and compression components trained separately, as in Martins and Smith (2009) ; and our compressive summarizer trained as a single task, and in the multi-task setting.", "cite_spans": [ { "start": 490, "end": 520, "text": "Berg-Kirkpatrick et al. (2011)", "ref_id": "BIBREF2" }, { "start": 928, "end": 949, "text": "Gillick et al. (2008)", "ref_id": "BIBREF17" }, { "start": 1140, "end": 1166, "text": "Woodsend and Lapata (2012)", "ref_id": "BIBREF44" }, { "start": 1487, "end": 1511, "text": "Martins and Smith (2009)", "ref_id": "BIBREF26" } ], "ref_spans": [ { "start": 808, "end": 815, "text": "Table 2", "ref_id": null } ], "eq_spans": [], "section": "Experimental setup", "sec_num": "5.1" }, { "text": "The ROUGE and Pyramid scores show that the compressive summarizers (when properly trained) yield considerable benefits in content coverage over extractive systems, confirming the results of Berg-Kirkpatrick et al. (2011) . Comparing the two bottom rows, we see a clear benefit by training in the multi-task setting, with a consistent gain in both coverage and linguistic quality. Our ROUGE-2 score (12.30%) is, to our knowledge, the highest reported on the TAC-2008 dataset, with little harm in grammaticality with respect to an extractive system that preserves the original sentences. Figure 2 shows an example summary.", "cite_spans": [ { "start": 190, "end": 220, "text": "Berg-Kirkpatrick et al. (2011)", "ref_id": "BIBREF2" } ], "ref_spans": [ { "start": 586, "end": 594, "text": "Figure 2", "ref_id": null } ], "eq_spans": [], "section": "Results", "sec_num": "5.2" }, { "text": "We conducted another set of experiments to compare the runtime of our compressive summarizer based on AD 3 with the runtimes achieved by GLPK, the ILP solver used by Berg-Kirkpatrick et al. (2011) . We varied the maximum number of it- Table 2 : Results for compressive summarization. Shown are the ROUGE-2 and ROUGE SU-4 recalls with the default options from the ROUGE toolkit (Lin, 2004) ; Pyramid scores (Nenkova and Passonneau, 2004) ; and linguistic quality scores, scored between 1 (very bad) to 5 (very good). For Pyramid, the evaluation was performed by two annotators, each evaluating half of the problems; scores marked with \u2020 were computed by different annotators and are not directly comparable. Linguistic quality was evaluated by two linguists; we show the average of the reported scores. Table 3 : Runtimes of several decoders on a Intel Core i7 processor @2.8 GHz, with 8GB RAM. For each decoder, we show the average time taken to solve a summarization problem in TAC-2008. The reported runtimes of AD 3 and LP-Relax include the time taken to round the solution ( \u00a73.4), which is 0.029 seconds on average. erations of AD 3 in {200, 1000, 5000}, and clocked the time spent by GLPK to solve the exact ILPs and their relaxations. Table 3 depicts the results. 10 We see that our proposed configuration (AD 3 -1000) is orders of magnitude faster than the ILP solver, and 5 times faster than its relaxed variant, while keeping similar accuracy levels. 11 The gain when the number of iterations in AD 3 is increased to 5000 is small, given that the runtime is more Japan dispatched four military ships to help Russia rescue seven crew members aboard a small submarine trapped on the seabed in the Far East. The Russian Pacific Fleet said the crew had 120 hours of oxygen reserves on board when the submarine submerged at midday Thursday (2300 GMT Wednesday) off the Kamchatka peninsula, the stretch of Far Eastern Russia facing the Bering Sea. The submarine, used in rescue, research and intelligence-gathering missions, became stuck at the bottom of the Bay of Berezovaya off Russia's Far East coast when its propeller was caught in a fishing net. The Russian submarine had been tending an underwater antenna mounted to the sea floor when it became snagged on a wire helping to stabilize a ventilation cable attached to the antenna. Rescue crews lowered a British remote-controlled underwater vehicle to a Russian mini-submarine trapped deep under the Pacific Ocean, hoping to free the vessel and its seven trapped crewmen before their air supply ran out. Figure 2 : Example summary from our compressive system. Removed text is grayed out. than doubled; accuracy starts to suffer, however, if the number of iterations is reduced too much. In practice, we observed that the final rounding procedure was crucial, as only 2 out of the 48 test problems had integral solutions (arguably because of the \"repulsive\" nature of the network, as hinted in \u00a73.4). For comparison, we also report in the bottom row the average runtime of the learned extractive baseline. We can see that our system's runtime is competitive with this baseline. To our knowledge, this is the first time a compressive summarizer achieves such a favorable accuracy/speed tradeoff.", "cite_spans": [ { "start": 166, "end": 196, "text": "Berg-Kirkpatrick et al. (2011)", "ref_id": "BIBREF2" }, { "start": 377, "end": 388, "text": "(Lin, 2004)", "ref_id": "BIBREF24" }, { "start": 406, "end": 436, "text": "(Nenkova and Passonneau, 2004)", "ref_id": "BIBREF32" }, { "start": 1271, "end": 1273, "text": "10", "ref_id": null }, { "start": 1461, "end": 1463, "text": "11", "ref_id": null } ], "ref_spans": [ { "start": 235, "end": 242, "text": "Table 2", "ref_id": null }, { "start": 802, "end": 809, "text": "Table 3", "ref_id": null }, { "start": 1242, "end": 1249, "text": "Table 3", "ref_id": null }, { "start": 2565, "end": 2573, "text": "Figure 2", "ref_id": null } ], "eq_spans": [], "section": "Runtimes", "sec_num": "5.3" }, { "text": "We presented a multi-task learning framework for compressive summarization, leveraging data for related tasks in a principled manner. We decode with AD 3 , a fast and modular dual decomposition algorithm which is orders of magnitude faster than ILP-based approaches. Results show that the state of the art is improved in automatic and manual metrics, with speeds close to extractive systems.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Conclusions", "sec_num": "6" }, { "text": "Our approach is modular and easy to extend. For example, a different compression model could incorporate rewriting rules to enable compressions that go beyond word deletion, as in Cohn and Lapata (2008) . Other aspects may be added as additional components in our dual decomposition framework, such as query information (Schilder and Kondadadi, 2008) , discourse con-straints (Clarke and Lapata, 2007) , or lexical preferences (Woodsend and Lapata, 2012 ). Our multitask approach may be used to jointly learn parameters for these aspects; the dual decomposition algorithm ensures that optimization remains tractable even with many components.", "cite_spans": [ { "start": 180, "end": 202, "text": "Cohn and Lapata (2008)", "ref_id": "BIBREF8" }, { "start": 320, "end": 350, "text": "(Schilder and Kondadadi, 2008)", "ref_id": "BIBREF37" }, { "start": 376, "end": 401, "text": "(Clarke and Lapata, 2007)", "ref_id": "BIBREF6" }, { "start": 427, "end": 453, "text": "(Woodsend and Lapata, 2012", "ref_id": "BIBREF44" } ], "ref_spans": [], "eq_spans": [], "section": "Conclusions", "sec_num": "6" }, { "text": "This section describes a linear-time algorithm (Algorithm 1) for solving the following problem:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Projection Onto Knapsack", "sec_num": null }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": "minimize z \u2212 a 2 w.r.t. z n \u2208 [0, 1], \u2200n \u2208 [N ], s.t. N n=1 L n z n \u2264 B,", "eq_num": "(11)" } ], "section": "A Projection Onto Knapsack", "sec_num": null }, { "text": "where a \u2208 R N and L n \u2265 0, \u2200n \u2208 [N ] . This includes as special cases the problems of projecting onto a budget constraint (L n = 1, \u2200n) and onto the simplex (same, plus B = 1).", "cite_spans": [ { "start": 32, "end": 36, "text": "[N ]", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "A Projection Onto Knapsack", "sec_num": null }, { "text": "Let clip(t) := max{0, min{1, t}}. Algorithm 1 starts by clipping a to the unit interval; if that yields a z satisfying N n=1 L n z n \u2264 B, we are done. Otherwise, the solution of Eq. 11 must satisfy N n=1 L n z n = B. It can be shown from the KKT conditions that the solution is of the form z * n := clip(a n + \u03c4 * L n ) for a constant \u03c4 * lying in a particular interval of split-points (line 11). To seek this constant, we use an algorithm due to Pardalos and Kovoor (1990) which iteratively shrinks this interval. The algorithm requires computing medians as a subroutine, which can be done in linear time (Blum et al., 1973) . The overall complexity in O(N ) (Pardalos and Kovoor, 1990) .", "cite_spans": [ { "start": 460, "end": 473, "text": "Kovoor (1990)", "ref_id": "BIBREF33" }, { "start": 606, "end": 625, "text": "(Blum et al., 1973)", "ref_id": "BIBREF3" }, { "start": 674, "end": 687, "text": "Kovoor, 1990)", "ref_id": "BIBREF33" } ], "ref_spans": [], "eq_spans": [], "section": "A Projection Onto Knapsack", "sec_num": null }, { "text": "Previous work has modeled concepts as events(Filatova and Hatzivassiloglou, 2004), salient words(Lin and Bilmes, 2010), and word bigrams(Gillick et al., 2008). In the sequel, we assume concepts are word k-grams, but our model can handle other representations, such as phrases or predicateargument structures.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "The same framework can be readily adapted to other compression models that are efficiently decodable, such as the semi-Markov model ofMcDonald (2006), which would allow incorporating a language model for the compression.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "Briefly, if we link the roots of the K sentences to a superroot node, the problem above can be transformed into that of finding the best configuration in the resulting binary tree subject to a budget constraint. We omit details for space.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "We use the AD 3 implementation in http://www. ark.cs.cmu.edu/AD3, setting the maximum number of iterations to 200 at training time and 1000 at test time. We extended the code to handle the knapsack and budget factors; the modified code will be part of the next release (AD 3 2.1).9 http://www.ark.cs.cmu.edu/TurboParser", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "Within dual decomposition algorithms, we verified experimentally that AD 3 is substantially faster than the subgradient algorithm, which is consistent with previous findings(Martins et al., 2011b).11 The runtimes obtained with the exact ILP solver seem slower than those reported by Berg-Kirkpatrick et al. (2011). (around 1.5 sec. on average, according to theirFig. 3). We conjecture that this difference is due to the restricted set of subtrees that can be deleted byBerg-Kirkpatrick et al. (2011), which greatly reduces their search space.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null } ], "back_matter": [ { "text": "We thank all reviewers for their insightful comments; Trevor Cohn for helpful discussions about multi-task learning; Taylor Berg-Kirkpatrick for answering questions about their summarizer and for providing code; and Helena Figueira and Pedro Mendes for helping with manual evaluation. This work was partially supported by the EU/FEDER programme, QREN/POR Lisboa (Portugal), under the Discooperio project (contract 2011/18501), and by a FCT grant PTDC/EEI-SII/2312/2012.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Acknowledgments", "sec_num": null }, { "text": "Algorithm 1 Projection Onto Knapsack. 1: input: a := an N n=1 , costs Ln N n=1 , maximum cost B 2: 3: {Try to clip into unit interval:} 4:Return z and stop. 7: end if 8: 9: {Run Pardalos and Kovoor (1990)'s algorithm:} 10: Initialize working set W \u2190 {1, . . . , K} 11: Initialize set of split points:Compute \u03c4 \u2190 Median(P) 15:Set s \u2190 s tight + \u03be\u03c4 + n\u2208W Lnclip(an + \u03c4 Ln) 16:If s \u2264 B, set \u03c4L \u2190 \u03c4 ; if s \u2265 B, set \u03c4R \u2190 \u03c4 17:Reduce set of split points:Define the sets:Update tight-sum:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "annex", "sec_num": null }, { "text": "Update slack-sum:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "21:", "sec_num": null } ], "bib_entries": { "BIBREF0": { "ref_id": "b0", "title": "22: end while 23: Define \u03c4 * \u2190 (B \u2212 N i=1 Liai \u2212 s tight )/\u03be 24: Set zn \u2190 clip(an + \u03c4 * Ln)", "authors": [ { "first": "G", "middle": [], "last": "Bak\u0131r", "suffix": "" }, { "first": "T", "middle": [], "last": "Hofmann", "suffix": "" }, { "first": "B", "middle": [], "last": "Sch\u00f6lkopf", "suffix": "" }, { "first": "A", "middle": [], "last": "Smola", "suffix": "" }, { "first": "B", "middle": [], "last": "Taskar", "suffix": "" }, { "first": "S", "middle": [], "last": "Vishwanathan", "suffix": "" } ], "year": 2007, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "G. Bak\u0131r, T. Hofmann, B. Sch\u00f6lkopf, A. Smola, B. Taskar, and S. Vishwanathan. 2007. Predicting Structured Data. The MIT Press. 22: end while 23: Define \u03c4 * \u2190 (B \u2212 N i=1 Liai \u2212 s tight )/\u03be 24: Set zn \u2190 clip(an + \u03c4 * Ln), \u2200n \u2208 [N ] 25: output: z := zn N n=1 .", "links": null }, "BIBREF1": { "ref_id": "b1", "title": "Machine-made index for technical literature-an experiment", "authors": [ { "first": "P", "middle": [ "B" ], "last": "Baxendale", "suffix": "" } ], "year": 1958, "venue": "IBM Journal of Research Development", "volume": "2", "issue": "4", "pages": "354--361", "other_ids": {}, "num": null, "urls": [], "raw_text": "P. B. Baxendale. 1958. Machine-made index for tech- nical literature-an experiment. IBM Journal of Re- search Development, 2(4):354-361.", "links": null }, "BIBREF2": { "ref_id": "b2", "title": "Jointly learning to extract and compress", "authors": [ { "first": "Taylor", "middle": [], "last": "Berg-Kirkpatrick", "suffix": "" }, { "first": "Dan", "middle": [], "last": "Gillick", "suffix": "" }, { "first": "Dan", "middle": [], "last": "Klein", "suffix": "" } ], "year": 2011, "venue": "Proc. of Annual Meeting of the Association for Computational Linguistics", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Taylor Berg-Kirkpatrick, Dan Gillick, and Dan Klein. 2011. Jointly learning to extract and compress. In Proc. of Annual Meeting of the Association for Com- putational Linguistics.", "links": null }, "BIBREF3": { "ref_id": "b3", "title": "Time bounds for selection", "authors": [ { "first": "Manuel", "middle": [], "last": "Blum", "suffix": "" }, { "first": "W", "middle": [], "last": "Robert", "suffix": "" }, { "first": "Vaughan", "middle": [], "last": "Floyd", "suffix": "" }, { "first": "", "middle": [], "last": "Pratt", "suffix": "" }, { "first": "L", "middle": [], "last": "Ronald", "suffix": "" }, { "first": "Robert", "middle": [ "E" ], "last": "Rivest", "suffix": "" }, { "first": "", "middle": [], "last": "Tarjan", "suffix": "" } ], "year": 1973, "venue": "Journal of Computer and System Sciences", "volume": "7", "issue": "4", "pages": "448--461", "other_ids": {}, "num": null, "urls": [], "raw_text": "Manuel Blum, Robert W Floyd, Vaughan Pratt, Ronald L Rivest, and Robert E Tarjan. 1973. Time bounds for selection. Journal of Computer and Sys- tem Sciences, 7(4):448-461.", "links": null }, "BIBREF4": { "ref_id": "b4", "title": "The use of MMR, diversity-based reranking for reordering documents and producing summaries", "authors": [ { "first": "J", "middle": [], "last": "Carbonell", "suffix": "" }, { "first": "J", "middle": [], "last": "Goldstein", "suffix": "" } ], "year": 1998, "venue": "SIGIR", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "J. Carbonell and J. Goldstein. 1998. The use of MMR, diversity-based reranking for reordering documents and producing summaries. In SIGIR.", "links": null }, "BIBREF5": { "ref_id": "b5", "title": "Exact decoding of phrase-based translation models through lagrangian relaxation", "authors": [ { "first": "Y.-W", "middle": [], "last": "Chang", "suffix": "" }, { "first": "M", "middle": [], "last": "Collins", "suffix": "" } ], "year": 2011, "venue": "Proc. of Empirical Methods for Natural Language Processing", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Y.-W. Chang and M. Collins. 2011. Exact decoding of phrase-based translation models through lagrangian relaxation. In Proc. of Empirical Methods for Natu- ral Language Processing.", "links": null }, "BIBREF6": { "ref_id": "b6", "title": "Modelling compression with discourse constraints", "authors": [ { "first": "James", "middle": [], "last": "Clarke", "suffix": "" }, { "first": "Mirella", "middle": [], "last": "Lapata", "suffix": "" } ], "year": 2007, "venue": "Proc. of Empirical Methods in Natural Language Processing", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "James Clarke and Mirella Lapata. 2007. Modelling compression with discourse constraints. In Proc. of Empirical Methods in Natural Language Process- ing.", "links": null }, "BIBREF7": { "ref_id": "b7", "title": "Global Inference for Sentence Compression An Integer Linear Programming Approach", "authors": [ { "first": "J", "middle": [], "last": "Clarke", "suffix": "" }, { "first": "M", "middle": [], "last": "Lapata", "suffix": "" } ], "year": 2008, "venue": "Journal of Artificial Intelligence Research", "volume": "31", "issue": "", "pages": "399--429", "other_ids": {}, "num": null, "urls": [], "raw_text": "J. Clarke and M. Lapata. 2008. Global Inference for Sentence Compression An Integer Linear Program- ming Approach. Journal of Artificial Intelligence Research, 31:399-429.", "links": null }, "BIBREF8": { "ref_id": "b8", "title": "Sentence compression beyond word deletion", "authors": [ { "first": "T", "middle": [], "last": "Cohn", "suffix": "" }, { "first": "M", "middle": [], "last": "Lapata", "suffix": "" } ], "year": 2008, "venue": "Proc. COLING", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "T. Cohn and M. Lapata. 2008. Sentence compression beyond word deletion. In Proc. COLING.", "links": null }, "BIBREF9": { "ref_id": "b9", "title": "An Exact Dual Decomposition Algorithm for Shallow Semantic Parsing with Constraints", "authors": [ { "first": "D", "middle": [], "last": "Das", "suffix": "" }, { "first": "A", "middle": [ "F T" ], "last": "Martins", "suffix": "" }, { "first": "N", "middle": [ "A" ], "last": "Smith", "suffix": "" } ], "year": 2012, "venue": "Proc. of First Joint Conference on Lexical and Computational Semantics (*SEM)", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "D. Das, A. F. T. Martins, and N. A. Smith. 2012. An Exact Dual Decomposition Algorithm for Shallow Semantic Parsing with Constraints. In Proc. of First Joint Conference on Lexical and Computational Se- mantics (*SEM).", "links": null }, "BIBREF10": { "ref_id": "b10", "title": "Practical Structured Learning Techniques for Natural Language Processing", "authors": [ { "first": "H", "middle": [], "last": "Daum\u00e9", "suffix": "" } ], "year": 2006, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "H. Daum\u00e9. 2006. Practical Structured Learning Tech- niques for Natural Language Processing. Ph.D. the- sis, University of Southern California.", "links": null }, "BIBREF11": { "ref_id": "b11", "title": "Frustratingly easy domain adaptation", "authors": [ { "first": "H", "middle": [], "last": "Daum\u00e9", "suffix": "" } ], "year": 2007, "venue": "Proc. of Annual Meeting of the Association for Computational Linguistics", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "H. Daum\u00e9. 2007. Frustratingly easy domain adapta- tion. In Proc. of Annual Meeting of the Association for Computational Linguistics.", "links": null }, "BIBREF12": { "ref_id": "b12", "title": "New methods in automatic extracting", "authors": [ { "first": "H", "middle": [ "P" ], "last": "Edmundson", "suffix": "" } ], "year": 1969, "venue": "Journal of the ACM", "volume": "16", "issue": "2", "pages": "264--285", "other_ids": {}, "num": null, "urls": [], "raw_text": "H. P. Edmundson. 1969. New methods in automatic extracting. Journal of the ACM, 16(2):264-285.", "links": null }, "BIBREF13": { "ref_id": "b13", "title": "Regularized multitask learning", "authors": [ { "first": "T", "middle": [], "last": "Evgeniou", "suffix": "" }, { "first": "M", "middle": [], "last": "Pontil", "suffix": "" } ], "year": 2004, "venue": "Proc. of ACM SIGKDD International Conference on Knowledge Discovery and Data Mining", "volume": "", "issue": "", "pages": "109--117", "other_ids": {}, "num": null, "urls": [], "raw_text": "T. Evgeniou and M. Pontil. 2004. Regularized multi- task learning. In Proc. of ACM SIGKDD Inter- national Conference on Knowledge Discovery and Data Mining, pages 109-117. ACM.", "links": null }, "BIBREF14": { "ref_id": "b14", "title": "A formal model for information selection in multisentence text extraction", "authors": [ { "first": "Elena", "middle": [], "last": "Filatova", "suffix": "" }, { "first": "Vasileios", "middle": [], "last": "Hatzivassiloglou", "suffix": "" } ], "year": 2004, "venue": "Proc. of International Conference on Computational Linguistics", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Elena Filatova and Vasileios Hatzivassiloglou. 2004. A formal model for information selection in multi- sentence text extraction. In Proc. of International Conference on Computational Linguistics.", "links": null }, "BIBREF15": { "ref_id": "b15", "title": "Hierarchical joint learning: Improving joint parsing and named entity recognition with non-jointly labeled data", "authors": [ { "first": "J", "middle": [ "R" ], "last": "Finkel", "suffix": "" }, { "first": "C", "middle": [ "D" ], "last": "Manning", "suffix": "" } ], "year": 2010, "venue": "Proc. of Annual Meeting of the Association for Computational Linguistics", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "J.R. Finkel and C.D. Manning. 2010. Hierarchical joint learning: Improving joint parsing and named entity recognition with non-jointly labeled data. In Proc. of Annual Meeting of the Association for Com- putational Linguistics.", "links": null }, "BIBREF16": { "ref_id": "b16", "title": "Discovering diverse and salient threads in document collections", "authors": [ { "first": "J", "middle": [], "last": "Gillenwater", "suffix": "" }, { "first": "A", "middle": [], "last": "Kulesza", "suffix": "" }, { "first": "B", "middle": [], "last": "Taskar", "suffix": "" } ], "year": 2012, "venue": "Proc. of Empirical Methods in Natural Language Processing", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "J. Gillenwater, A. Kulesza, and B. Taskar. 2012. Dis- covering diverse and salient threads in document collections. In Proc. of Empirical Methods in Natu- ral Language Processing.", "links": null }, "BIBREF17": { "ref_id": "b17", "title": "The icsi summarization system at tac", "authors": [ { "first": "Dan", "middle": [], "last": "Gillick", "suffix": "" }, { "first": "Dilek", "middle": [], "last": "Benoit Favre", "suffix": "" }, { "first": "", "middle": [], "last": "Hakkani-Tur", "suffix": "" } ], "year": 2008, "venue": "Proc. of Text Understanding Conference", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Dan Gillick, Benoit Favre, and Dilek Hakkani-Tur. 2008. The icsi summarization system at tac 2008. In Proc. of Text Understanding Conference.", "links": null }, "BIBREF18": { "ref_id": "b18", "title": "Statistics-based summarization-step one: Sentence compression", "authors": [ { "first": "K", "middle": [], "last": "Knight", "suffix": "" }, { "first": "D", "middle": [], "last": "Marcu", "suffix": "" } ], "year": 2000, "venue": "AAAI/IAAI", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "K. Knight and D. Marcu. 2000. Statistics-based summarization-step one: Sentence compression. In AAAI/IAAI.", "links": null }, "BIBREF19": { "ref_id": "b19", "title": "MRF optimization via dual decomposition: Message-passing revisited", "authors": [ { "first": "N", "middle": [], "last": "Komodakis", "suffix": "" }, { "first": "N", "middle": [], "last": "Paragios", "suffix": "" }, { "first": "G", "middle": [], "last": "Tziritas", "suffix": "" } ], "year": 2007, "venue": "Proc. of International Conference on Computer Vision", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "N. Komodakis, N. Paragios, and G. Tziritas. 2007. MRF optimization via dual decomposition: Message-passing revisited. In Proc. of International Conference on Computer Vision.", "links": null }, "BIBREF20": { "ref_id": "b20", "title": "A trainable document summarizer", "authors": [ { "first": "J", "middle": [], "last": "Kupiec", "suffix": "" }, { "first": "J", "middle": [], "last": "Pedersen", "suffix": "" }, { "first": "F", "middle": [], "last": "Chen", "suffix": "" } ], "year": 1995, "venue": "SIGIR", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "J. Kupiec, J. Pedersen, and F. Chen. 1995. A trainable document summarizer. In SIGIR.", "links": null }, "BIBREF21": { "ref_id": "b21", "title": "Multi-document summarization via budgeted maximization of submodular functions", "authors": [ { "first": "H", "middle": [], "last": "Lin", "suffix": "" }, { "first": "J", "middle": [], "last": "Bilmes", "suffix": "" } ], "year": 2010, "venue": "Proc. of Annual Meeting of the North American chapter of the Association for Computational Linguistics", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "H. Lin and J. Bilmes. 2010. Multi-document summa- rization via budgeted maximization of submodular functions. In Proc. of Annual Meeting of the North American chapter of the Association for Computa- tional Linguistics.", "links": null }, "BIBREF22": { "ref_id": "b22", "title": "Learning mixtures of submodular shells with application to document summarization", "authors": [ { "first": "H", "middle": [], "last": "Lin", "suffix": "" }, { "first": "J", "middle": [], "last": "Bilmes", "suffix": "" } ], "year": 2012, "venue": "Proc. of Uncertainty in Artificial Intelligence", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "H. Lin and J. Bilmes. 2012. Learning mixtures of sub- modular shells with application to document sum- marization. In Proc. of Uncertainty in Artificial In- telligence.", "links": null }, "BIBREF23": { "ref_id": "b23", "title": "Improving summarization performance by sentence compression-a pilot study", "authors": [ { "first": "C.-Y.", "middle": [], "last": "Lin", "suffix": "" } ], "year": 2003, "venue": "the Int. Workshop on Inf. Ret. with Asian Languages", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "C.-Y. Lin. 2003. Improving summarization perfor- mance by sentence compression-a pilot study. In the Int. Workshop on Inf. Ret. with Asian Languages.", "links": null }, "BIBREF24": { "ref_id": "b24", "title": "Rouge: A package for automatic evaluation of summaries", "authors": [ { "first": "Chin-Yew", "middle": [], "last": "Lin", "suffix": "" } ], "year": 2004, "venue": "Text Summarization Branches Out: Proceedings of the ACL-04 Workshop", "volume": "", "issue": "", "pages": "74--81", "other_ids": {}, "num": null, "urls": [], "raw_text": "Chin-Yew Lin. 2004. Rouge: A package for auto- matic evaluation of summaries. In Stan Szpakowicz Marie-Francine Moens, editor, Text Summarization Branches Out: Proceedings of the ACL-04 Work- shop, pages 74-81, Barcelona, Spain, July.", "links": null }, "BIBREF25": { "ref_id": "b25", "title": "The automatic creation of literature abstracts", "authors": [ { "first": "H", "middle": [ "P" ], "last": "Luhn", "suffix": "" } ], "year": 1958, "venue": "IBM Journal of Research Development", "volume": "2", "issue": "2", "pages": "159--165", "other_ids": {}, "num": null, "urls": [], "raw_text": "H. P. Luhn. 1958. The automatic creation of literature abstracts. IBM Journal of Research Development, 2(2):159-165.", "links": null }, "BIBREF26": { "ref_id": "b26", "title": "Summarization with a Joint Model for Sentence Extraction and Compression", "authors": [ { "first": "A", "middle": [ "F T" ], "last": "Martins", "suffix": "" }, { "first": "N", "middle": [ "A" ], "last": "Smith", "suffix": "" } ], "year": 2009, "venue": "North American Chapter of the Association for Computational Linguistics: Workshop on Integer Linear Programming for NLP", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "A. F. T. Martins and N. A. Smith. 2009. Summariza- tion with a Joint Model for Sentence Extraction and Compression. In North American Chapter of the As- sociation for Computational Linguistics: Workshop on Integer Linear Programming for NLP.", "links": null }, "BIBREF27": { "ref_id": "b27", "title": "An Augmented Lagrangian Approach to Constrained MAP Inference", "authors": [ { "first": "A", "middle": [ "F T" ], "last": "Martins", "suffix": "" }, { "first": "M", "middle": [ "A T" ], "last": "Figueiredo", "suffix": "" }, { "first": "P", "middle": [ "M Q" ], "last": "Aguiar", "suffix": "" }, { "first": "N", "middle": [ "A" ], "last": "Smith", "suffix": "" }, { "first": "E", "middle": [ "P" ], "last": "Xing", "suffix": "" } ], "year": 2011, "venue": "Proc. of International Conference on Machine Learning", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "A. F. T. Martins, M. A. T. Figueiredo, P. M. Q. Aguiar, N. A. Smith, and E. P. Xing. 2011a. An Aug- mented Lagrangian Approach to Constrained MAP Inference. In Proc. of International Conference on Machine Learning.", "links": null }, "BIBREF28": { "ref_id": "b28", "title": "Dual Decomposition with Many Overlapping Components", "authors": [ { "first": "A", "middle": [ "F T" ], "last": "Martins", "suffix": "" }, { "first": "N", "middle": [ "A" ], "last": "Smith", "suffix": "" }, { "first": "P", "middle": [ "M Q" ], "last": "Aguiar", "suffix": "" }, { "first": "M", "middle": [ "A T" ], "last": "Figueiredo", "suffix": "" } ], "year": 2011, "venue": "Proc. of Empirical Methods for Natural Language Processing", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "A. F. T. Martins, N. A. Smith, P. M. Q. Aguiar, and M. A. T. Figueiredo. 2011b. Dual Decomposition with Many Overlapping Components. In Proc. of Empirical Methods for Natural Language Process- ing.", "links": null }, "BIBREF30": { "ref_id": "b30", "title": "Discriminative sentence compression with soft syntactic constraints", "authors": [ { "first": "R", "middle": [], "last": "Mcdonald", "suffix": "" } ], "year": 2006, "venue": "Proc. of Annual Meeting of the European Chapter of the Association for Computational Linguistics", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "R. McDonald. 2006. Discriminative sentence com- pression with soft syntactic constraints. In Proc. of Annual Meeting of the European Chapter of the As- sociation for Computational Linguistics.", "links": null }, "BIBREF31": { "ref_id": "b31", "title": "A study of global inference algorithms in multi-document summarization", "authors": [ { "first": "R", "middle": [], "last": "Mcdonald", "suffix": "" } ], "year": 2007, "venue": "ECIR", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "R. McDonald. 2007. A study of global inference algo- rithms in multi-document summarization. In ECIR.", "links": null }, "BIBREF32": { "ref_id": "b32", "title": "Evaluating content selection in summarization: The pyramid method", "authors": [ { "first": "A", "middle": [], "last": "Nenkova", "suffix": "" }, { "first": "R", "middle": [], "last": "Passonneau", "suffix": "" } ], "year": 2004, "venue": "Proceedings of NAACL", "volume": "", "issue": "", "pages": "145--152", "other_ids": {}, "num": null, "urls": [], "raw_text": "A. Nenkova and R. Passonneau. 2004. Evaluating content selection in summarization: The pyramid method. In Proceedings of NAACL, pages 145-152.", "links": null }, "BIBREF33": { "ref_id": "b33", "title": "An algorithm for a singly constrained class of quadratic programs subject to upper and lower bounds", "authors": [ { "first": "M", "middle": [], "last": "Panos", "suffix": "" }, { "first": "Naina", "middle": [], "last": "Pardalos", "suffix": "" }, { "first": "", "middle": [], "last": "Kovoor", "suffix": "" } ], "year": 1990, "venue": "Mathematical Programming", "volume": "46", "issue": "1", "pages": "321--328", "other_ids": {}, "num": null, "urls": [], "raw_text": "Panos M. Pardalos and Naina Kovoor. 1990. An al- gorithm for a singly constrained class of quadratic programs subject to upper and lower bounds. Math- ematical Programming, 46(1):321-328.", "links": null }, "BIBREF34": { "ref_id": "b34", "title": "Centroid-based summarization of multiple documents: sentence extraction, utility-based evaluation, and user studies", "authors": [ { "first": "D", "middle": [ "R" ], "last": "Radev", "suffix": "" }, { "first": "H", "middle": [], "last": "Jing", "suffix": "" }, { "first": "M", "middle": [], "last": "Budzikowska", "suffix": "" } ], "year": 2000, "venue": "the NAACL-ANLP Workshop on Automatic Summarization", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "D. R. Radev, H. Jing, and M. Budzikowska. 2000. Centroid-based summarization of multiple docu- ments: sentence extraction, utility-based evaluation, and user studies. In the NAACL-ANLP Workshop on Automatic Summarization.", "links": null }, "BIBREF35": { "ref_id": "b35", "title": "A Tutorial on Dual Decomposition and Lagrangian Relaxation for Inference in Natural Language Processing", "authors": [ { "first": "A", "middle": [ "M" ], "last": "Rush", "suffix": "" }, { "first": "M", "middle": [], "last": "Collins", "suffix": "" } ], "year": 2012, "venue": "Journal of Artificial Intelligence Research", "volume": "45", "issue": "", "pages": "305--362", "other_ids": {}, "num": null, "urls": [], "raw_text": "A.M. Rush and M. Collins. 2012. A Tutorial on Dual Decomposition and Lagrangian Relaxation for In- ference in Natural Language Processing. Journal of Artificial Intelligence Research, 45:305-362.", "links": null }, "BIBREF36": { "ref_id": "b36", "title": "On dual decomposition and linear programming relaxations for natural language processing", "authors": [ { "first": "A", "middle": [], "last": "Rush", "suffix": "" }, { "first": "D", "middle": [], "last": "Sontag", "suffix": "" }, { "first": "M", "middle": [], "last": "Collins", "suffix": "" }, { "first": "T", "middle": [], "last": "Jaakkola", "suffix": "" } ], "year": 2010, "venue": "Proc. of Empirical Methods for Natural Language Processing", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "A. Rush, D. Sontag, M. Collins, and T. Jaakkola. 2010. On dual decomposition and linear programming re- laxations for natural language processing. In Proc. of Empirical Methods for Natural Language Pro- cessing.", "links": null }, "BIBREF37": { "ref_id": "b37", "title": "Fastsum: Fast and accurate query-based multi-document summarization", "authors": [ { "first": "Frank", "middle": [], "last": "Schilder", "suffix": "" }, { "first": "Ravikumar", "middle": [], "last": "Kondadadi", "suffix": "" } ], "year": 2008, "venue": "Proc. of Annual Meeting of the Association for Computational Linguistics", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Frank Schilder and Ravikumar Kondadadi. 2008. Fast- sum: Fast and accurate query-based multi-document summarization. In Proc. of Annual Meeting of the Association for Computational Linguistics.", "links": null }, "BIBREF38": { "ref_id": "b38", "title": "Large-margin learning of submodular summarization models", "authors": [ { "first": "R", "middle": [], "last": "Sipos", "suffix": "" }, { "first": "P", "middle": [], "last": "Shivaswamy", "suffix": "" }, { "first": "T", "middle": [], "last": "Joachims", "suffix": "" } ], "year": 2012, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "R. Sipos, P. Shivaswamy, and T. Joachims. 2012. Large-margin learning of submodular summariza- tion models.", "links": null }, "BIBREF39": { "ref_id": "b39", "title": "Maxmargin Markov networks", "authors": [ { "first": "B", "middle": [], "last": "Taskar", "suffix": "" }, { "first": "C", "middle": [], "last": "Guestrin", "suffix": "" }, { "first": "D", "middle": [], "last": "Koller", "suffix": "" } ], "year": 2003, "venue": "Proc. of Neural Information Processing Systems", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "B. Taskar, C. Guestrin, and D. Koller. 2003. Max- margin Markov networks. In Proc. of Neural Infor- mation Processing Systems.", "links": null }, "BIBREF40": { "ref_id": "b40", "title": "Learning associative Markov networks", "authors": [ { "first": "B", "middle": [], "last": "Taskar", "suffix": "" }, { "first": "V", "middle": [], "last": "Chatalbashev", "suffix": "" }, { "first": "D", "middle": [], "last": "Koller", "suffix": "" } ], "year": 2004, "venue": "Proc. of International Conference of Machine Learning", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "B. Taskar, V. Chatalbashev, and D. Koller. 2004. Learning associative Markov networks. In Proc. of International Conference of Machine Learning.", "links": null }, "BIBREF41": { "ref_id": "b41", "title": "Support vector machine learning for interdependent and structured output spaces", "authors": [ { "first": "I", "middle": [], "last": "Tsochantaridis", "suffix": "" }, { "first": "T", "middle": [], "last": "Hofmann", "suffix": "" }, { "first": "T", "middle": [], "last": "Joachims", "suffix": "" }, { "first": "Y", "middle": [], "last": "Altun", "suffix": "" } ], "year": 2004, "venue": "Proc. of International Conference of Machine Learning", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "I. Tsochantaridis, T. Hofmann, T. Joachims, and Y. Al- tun. 2004. Support vector machine learning for in- terdependent and structured output spaces. In Proc. of International Conference of Machine Learning.", "links": null }, "BIBREF42": { "ref_id": "b42", "title": "Automatic generation of story highlights", "authors": [ { "first": "K", "middle": [], "last": "Woodsend", "suffix": "" }, { "first": "M", "middle": [], "last": "Lapata", "suffix": "" } ], "year": 2010, "venue": "Proc. of Annual Meeting of the Association for Computational Linguistics", "volume": "", "issue": "", "pages": "565--574", "other_ids": {}, "num": null, "urls": [], "raw_text": "K. Woodsend and M. Lapata. 2010. Automatic gener- ation of story highlights. In Proc. of Annual Meet- ing of the Association for Computational Linguis- tics, pages 565-574.", "links": null }, "BIBREF43": { "ref_id": "b43", "title": "Learning to simplify sentences with quasi-synchronous grammar and integer programming", "authors": [ { "first": "Kristian", "middle": [], "last": "Woodsend", "suffix": "" }, { "first": "Mirella", "middle": [], "last": "Lapata", "suffix": "" } ], "year": 2011, "venue": "Proc. of Empirical Methods in Natural Language Processing", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Kristian Woodsend and Mirella Lapata. 2011. Learn- ing to simplify sentences with quasi-synchronous grammar and integer programming. In Proc. of Em- pirical Methods in Natural Language Processing.", "links": null }, "BIBREF44": { "ref_id": "b44", "title": "Multiple aspect summarization using integer linear programming", "authors": [ { "first": "Kristian", "middle": [], "last": "Woodsend", "suffix": "" }, { "first": "Mirella", "middle": [], "last": "Lapata", "suffix": "" } ], "year": 2012, "venue": "Proc. of Empirical Methods in Natural Language Processing", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Kristian Woodsend and Mirella Lapata. 2012. Mul- tiple aspect summarization using integer linear pro- gramming. In Proc. of Empirical Methods in Natu- ral Language Processing.", "links": null }, "BIBREF45": { "ref_id": "b45", "title": "Multi-document summarization by maximizing informative content-words", "authors": [ { "first": "Joshua", "middle": [], "last": "Wen-Tau Yih", "suffix": "" }, { "first": "Lucy", "middle": [], "last": "Goodman", "suffix": "" }, { "first": "Hisami", "middle": [], "last": "Vanderwende", "suffix": "" }, { "first": "", "middle": [], "last": "Suzuki", "suffix": "" } ], "year": 2007, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Wen-tau Yih, Joshua Goodman, Lucy Vanderwende, and Hisami Suzuki. 2007. Multi-document summa- rization by maximizing informative content-words.", "links": null }, "BIBREF46": { "ref_id": "b46", "title": "Proc. of International Joint Conference on Artifical Intelligence", "authors": [], "year": null, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "In Proc. of International Joint Conference on Artifi- cal Intelligence.", "links": null }, "BIBREF47": { "ref_id": "b47", "title": "Sentence compression as a component of a multidocument summarization system", "authors": [ { "first": "D", "middle": [], "last": "Zajic", "suffix": "" }, { "first": "B", "middle": [], "last": "Dorr", "suffix": "" }, { "first": "J", "middle": [], "last": "Lin", "suffix": "" }, { "first": "R", "middle": [], "last": "Schwartz", "suffix": "" } ], "year": 2006, "venue": "the ACL DUC Workshop", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "D. Zajic, B. Dorr, J. Lin, and R. Schwartz. 2006. Sentence compression as a component of a multi- document summarization system. In the ACL DUC Workshop.", "links": null } }, "ref_entries": { "TABREF0": { "text": "Sentences$ The leader of moderate Kashmiri separatists warned Thursday that ... $ Talks with Kashmiri separatists began last year ...", "html": null, "num": null, "content": "
Concept tokens
\"Kashmiri separatists\"
Concept typeBudget
", "type_str": "table" }, "TABREF2": { "text": "Features and decoders used for each task.", "html": null, "num": null, "content": "", "type_str": "table" } } } }