| { |
| "paper_id": "K16-1028", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T07:11:10.324657Z" |
| }, |
| "title": "Abstractive Text Summarization using Sequence-to-sequence RNNs and Beyond", |
| "authors": [ |
| { |
| "first": "Ramesh", |
| "middle": [], |
| "last": "Nallapati", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "nallapati@us.ibm.com" |
| }, |
| { |
| "first": "Bowen", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "IBM Watson", |
| "location": {} |
| }, |
| "email": "zhou@us.ibm.com" |
| }, |
| { |
| "first": "Ibm", |
| "middle": [], |
| "last": "Watson", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "IBM Watson", |
| "location": {} |
| }, |
| "email": "" |
| }, |
| { |
| "first": "\u00c7aglar", |
| "middle": [], |
| "last": "Gul\u00e7ehre", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Universit\u00e9 de Montr\u00e9al", |
| "location": {} |
| }, |
| "email": "gulcehrc@iro.umontreal.ca" |
| }, |
| { |
| "first": "Bing", |
| "middle": [], |
| "last": "Xiang", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "IBM Watson", |
| "location": {} |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "In this work, we model abstractive text summarization using Attentional Encoder-Decoder Recurrent Neural Networks, and show that they achieve state-of-the-art performance on two different corpora. We propose several novel models that address critical problems in summarization that are not adequately modeled by the basic architecture, such as modeling keywords , capturing the hierarchy of sentence-toword structure, and emitting words that are rare or unseen at training time. Our work shows that many of our proposed models contribute to further improvement in performance. We also propose a new dataset consisting of multi-sentence summaries, and establish performance benchmarks for further research.", |
| "pdf_parse": { |
| "paper_id": "K16-1028", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "In this work, we model abstractive text summarization using Attentional Encoder-Decoder Recurrent Neural Networks, and show that they achieve state-of-the-art performance on two different corpora. We propose several novel models that address critical problems in summarization that are not adequately modeled by the basic architecture, such as modeling keywords , capturing the hierarchy of sentence-toword structure, and emitting words that are rare or unseen at training time. Our work shows that many of our proposed models contribute to further improvement in performance. We also propose a new dataset consisting of multi-sentence summaries, and establish performance benchmarks for further research.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Abstractive text summarization is the task of generating a headline or a short summary consisting of a few sentences that captures the salient ideas of an article or a passage. We use the adjective 'abstractive' to denote a summary that is not a mere selection of a few existing passages or sentences extracted from the source, but a compressed paraphrasing of the main contents of the document, potentially using vocabulary unseen in the source document.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "This task can also be naturally cast as mapping an input sequence of words in a source document to a target sequence of words called summary. In the recent past, deep-learning based models that map an input sequence into another output sequence, called sequence-to-sequence models, have been successful in many problems such as machine translation (Bahdanau et al., 2014) , speech recognition (Bahdanau et al., 2015) and video captioning (Venugopalan et al., 2015) . In the framework of sequence-to-sequence models, a very relevant model to our task is the attentional Recurrent Neural Network (RNN) encoderdecoder model proposed in Bahdanau et al. (2014) , which has produced state-of-the-art performance in machine translation (MT) , which is also a natural language task.", |
| "cite_spans": [ |
| { |
| "start": 348, |
| "end": 371, |
| "text": "(Bahdanau et al., 2014)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 393, |
| "end": 416, |
| "text": "(Bahdanau et al., 2015)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 438, |
| "end": 464, |
| "text": "(Venugopalan et al., 2015)", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 633, |
| "end": 655, |
| "text": "Bahdanau et al. (2014)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 729, |
| "end": 733, |
| "text": "(MT)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Despite the similarities, abstractive summarization is a very different problem from MT. Unlike in MT, the target (summary) is typically very short and does not depend very much on the length of the source (document) in summarization. Additionally, a key challenge in summarization is to optimally compress the original document in a lossy manner such that the key concepts in the original document are preserved, whereas in MT, the translation is expected to be loss-less. In translation, there is a strong notion of almost one-to-one wordlevel alignment between source and target, but in summarization, it is less obvious.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "We make the following main contributions in this work: (i) We apply the off-the-shelf attentional encoder-decoder RNN that was originally developed for machine translation to summarization, and show that it already outperforms stateof-the-art systems on two different English corpora. (ii) Motivated by concrete problems in summarization that are not sufficiently addressed by the machine translation based model, we propose novel models and show that they provide additional improvement in performance. (iii) We propose a new dataset for the task of abstractive summarization of a document into multiple sentences and establish benchmarks.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The rest of the paper is organized as follows. In Section 2, we describe each specific problem in abstractive summarization that we aim to solve, and present a novel model that addresses it. Sec-tion 3 contextualizes our models with respect to closely related work on the topic of abstractive text summarization. We present the results of our experiments on three different data sets in Section 4. We also present some qualitative analysis of the output from our models in Section 5 before concluding the paper with remarks on our future direction in Section 6.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this section, we first describe the basic encoderdecoder RNN that serves as our baseline and then propose several novel models for summarization, each addressing a specific weakness in the baseline.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Models", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Our baseline model corresponds to the neural machine translation model used in Bahdanau et al. (2014) . The encoder consists of a bidirectional GRU-RNN (Chung et al., 2014) , while the decoder consists of a uni-directional GRU-RNN with the same hidden-state size as that of the encoder, and an attention mechanism over the source-hidden states and a soft-max layer over target vocabulary to generate words. In the interest of space, we refer the reader to the original paper for a detailed treatment of this model. In addition to the basic model, we also adapted to the summarization problem, the large vocabulary 'trick' (LVT) described in Jean et al. (2014) . In our approach, the decoder-vocabulary of each mini-batch is restricted to words in the source documents of that batch. In addition, the most frequent words in the target dictionary are added until the vocabulary reaches a fixed size. The aim of this technique is to reduce the size of the soft-max layer of the decoder which is the main computational bottleneck. In addition, this technique also speeds up convergence by focusing the modeling effort only on the words that are essential to a given example. This technique is particularly well suited to summarization since a large proportion of the words in the summary come from the source document in any case.", |
| "cite_spans": [ |
| { |
| "start": 79, |
| "end": 101, |
| "text": "Bahdanau et al. (2014)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 152, |
| "end": 172, |
| "text": "(Chung et al., 2014)", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 641, |
| "end": 659, |
| "text": "Jean et al. (2014)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Encoder-Decoder RNN with Attention and Large Vocabulary Trick", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "In summarization, one of the key challenges is to identify the key concepts and key entities in the document, around which the story revolves. In order to accomplish this goal, we may need to go beyond the word-embeddings-based representation of the input document and capture additional linguistic features such as parts-of-speech tags, named-entity tags, and TF and IDF statistics of the words. We therefore create additional look-up based embedding matrices for the vocabulary of each tag-type, similar to the embeddings for words. For continuous features such as TF and IDF, we convert them into categorical values by discretizing them into a fixed number of bins, and use one-hot representations to indicate the bin number they fall into. This allows us to map them into an embeddings matrix like any other tag-type. Finally, for each word in the source document, we simply look-up its embeddings from all of its associated tags and concatenate them into a single long vector, as shown in Fig. 1 . On the target side, we continue to use only word-based embeddings as the representation. ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 994, |
| "end": 1000, |
| "text": "Fig. 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Capturing Keywords using Feature-rich Encoder", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Often-times in summarization, the keywords or named-entities in a test document that are central to the summary may actually be unseen or rare with respect to training data. Since the vocabulary of the decoder is fixed at training time, it cannot emit these unseen words. Instead, a most common way of handling these out-of-vocabulary (OOV) words is to emit an 'UNK' token as a placeholder. However this does not result in legible summaries. In summarization, an intuitive way to handle such OOV words is to simply point to their location in the source document instead. We model this no-tion using our novel switching decoder/pointer architecture which is graphically represented in Figure 2. In this model, the decoder is equipped with a 'switch' that decides between using the generator or a pointer at every time-step. If the switch is turned on, the decoder produces a word from its target vocabulary in the normal fashion. However, if the switch is turned off, the decoder instead generates a pointer to one of the word-positions in the source. The word at the pointer-location is then copied into the summary. The switch is modeled as a sigmoid activation function over a linear layer based on the entire available context at each timestep as shown below.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 684, |
| "end": 690, |
| "text": "Figure", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Modeling Rare/Unseen Words using Switching Generator-Pointer", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "P (s i = 1) = \u03c3(v s \u2022 (W s h h i + W s e E[o i\u22121 ] + W s c c i + b s )), where P (s i = 1)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Modeling Rare/Unseen Words using Switching Generator-Pointer", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "is the probability of the switch turning on at the i th time-step of the decoder, h i is the hidden state, E[o i\u22121 ] is the embedding vector of the emission from the previous time step, c i is the attention-weighted context vector, and W s h , W s e , W s c , b s and v s are the switch parameters. We use attention distribution over word positions in the document as the distribution to sample the pointer from.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Modeling Rare/Unseen Words using Switching Generator-Pointer", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "P a i (j) \u221d exp(v a \u2022 (W a h h i\u22121 + W a e E[o i\u22121 ] + W a c h d j + b a )), p i = arg max j (P a i (j)) for j \u2208 {1, . . . , N d }.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Modeling Rare/Unseen Words using Switching Generator-Pointer", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "In the above equation, p i is the pointer value at i th word-position in the summary, sampled from the attention distribution P a i over the document word-positions j \u2208 {1, . . . , N d }, where P a i (j) is the probability of the i th time-step in the decoder pointing to the j th position in the document, and h d j is the encoder's hidden state at position j. At training time, we provide the model with explicit pointer information whenever the summary word does not exist in the target vocabulary. When the OOV word in summary occurs in multiple document positions, we break the tie in favor of its first occurrence. At training time, we optimize the conditional log-likelihood shown below, with additional regularization penalties.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Modeling Rare/Unseen Words using Switching Generator-Pointer", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "log P (y|x) = i (g i log{P (y i |y \u2212i , x)P (s i )} +(1 \u2212 g i ) log{P (p(i)|y \u2212i , x)(1 \u2212 P (s i ))})", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Modeling Rare/Unseen Words using Switching Generator-Pointer", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "where y and x are the summary and document words respectively, g i is an indicator function that is set to 0 whenever the word at position i in the summary is OOV with respect to the decoder vocabulary. At test time, the model decides automatically at each time-step whether to generate or to point, based on the estimated switch probability P (s i ). We simply use the arg max of the posterior probability of generation or pointing to generate the best output at each time step.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Modeling Rare/Unseen Words using Switching Generator-Pointer", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "The pointer mechanism may be more robust in handling rare words because it uses the encoder's hidden-state representation of rare words to decide which word from the document to point to. Since the hidden state depends on the entire context of the word, the model is able to accurately point to unseen words although they do not appear in the target vocabulary. ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Modeling Rare/Unseen Words using Switching Generator-Pointer", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "In datasets where the source document is very long, in addition to identifying the keywords in the document, it is also important to identify the key sentences from which the summary can be drawn. This model aims to capture this notion of two levels of importance using two bi-directional RNNs on the source side, one at the word level and the other at the sentence level. The attention mechanism operates at both levels simultaneously. The word-level attention is further re-weighted by the corresponding sentence-level attention and renormalized as shown below:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Capturing Hierarchical Document Structure with Hierarchical Attention", |
| "sec_num": "2.4" |
| }, |
| { |
| "text": "P a (j) = P a w (j)P a s (s(j)) N d k=1 P a w (k)P a s (s(k)) ,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Capturing Hierarchical Document Structure with Hierarchical Attention", |
| "sec_num": "2.4" |
| }, |
| { |
| "text": "where P a w (j) is the word-level attention weight at j th position of the source document, and s(j) is the ID of the sentence at j th word position, P a s (l) is the sentence-level attention weight for the l th sentence in the source, N d is the number of words in the source document, and P a (j) is the re-scaled attention at the j th word position. The re-scaled attention is then used to compute the attentionweighted context vector that goes as input to the hidden state of the decoder. Further, we also concatenate additional positional embeddings to the hidden state of the sentence-level RNN to model positional importance of sentences in the document. This architecture therefore models key sentences as well as keywords within those sentences jointly. A graphical representation of this model is displayed in Figure 3 . ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 820, |
| "end": 828, |
| "text": "Figure 3", |
| "ref_id": "FIGREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Capturing Hierarchical Document Structure with Hierarchical Attention", |
| "sec_num": "2.4" |
| }, |
| { |
| "text": "A vast majority of past work in summarization has been extractive, which consists of identifying key sentences or passages in the source document and reproducing them as summary (Neto et al., 2002; Erkan and Radev, 2004; Wong et al., 2008a; Filippova and Altun, 2013; Colmenares et al., 2015; Litvak and Last, 2008; K. Riedhammer and Hakkani-Tur, 2010; Ricardo Ribeiro, 2013) . Humans on the other hand, tend to paraphrase the original story in their own words. As such, human summaries are abstractive in nature and seldom consist of reproduction of original sentences from the document. The task of abstractive summarization has been standardized using the DUC-2003 and DUC-2004 competitions. 2 The data for these tasks consists of news stories from various topics with multiple reference summaries per story generated by humans. The best performing system on the DUC-2004 task, called TOPIARY (Zajic et al., 2004) , used a combination of linguistically motivated compression techniques, and an unsupervised topic detection algorithm that appends keywords extracted from the article onto the compressed output. Some of the other notable work in the task of abstractive summarization includes using traditional phrase-table based machine translation approaches (Banko et al., 2000) , compression using weighted tree-transformation rules (Cohn and Lapata, 2008) and quasi-synchronous grammar approaches (Woodsend et al., 2010) .", |
| "cite_spans": [ |
| { |
| "start": 178, |
| "end": 197, |
| "text": "(Neto et al., 2002;", |
| "ref_id": "BIBREF22" |
| }, |
| { |
| "start": 198, |
| "end": 220, |
| "text": "Erkan and Radev, 2004;", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 221, |
| "end": 240, |
| "text": "Wong et al., 2008a;", |
| "ref_id": "BIBREF27" |
| }, |
| { |
| "start": 241, |
| "end": 267, |
| "text": "Filippova and Altun, 2013;", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 268, |
| "end": 292, |
| "text": "Colmenares et al., 2015;", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 293, |
| "end": 315, |
| "text": "Litvak and Last, 2008;", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 316, |
| "end": 352, |
| "text": "K. Riedhammer and Hakkani-Tur, 2010;", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 353, |
| "end": 375, |
| "text": "Ricardo Ribeiro, 2013)", |
| "ref_id": "BIBREF23" |
| }, |
| { |
| "start": 659, |
| "end": 671, |
| "text": "DUC-2003 and", |
| "ref_id": null |
| }, |
| { |
| "start": 672, |
| "end": 696, |
| "text": "DUC-2004 competitions. 2", |
| "ref_id": null |
| }, |
| { |
| "start": 896, |
| "end": 916, |
| "text": "(Zajic et al., 2004)", |
| "ref_id": "BIBREF30" |
| }, |
| { |
| "start": 1262, |
| "end": 1282, |
| "text": "(Banko et al., 2000)", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 1338, |
| "end": 1361, |
| "text": "(Cohn and Lapata, 2008)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 1403, |
| "end": 1426, |
| "text": "(Woodsend et al., 2010)", |
| "ref_id": "BIBREF29" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "3" |
| }, |
| { |
| "text": "With the emergence of deep learning as a viable alternative for many NLP tasks (Collobert et al., 2011) , researchers have started considering this framework as an attractive, fully data-driven alternative to abstractive summarization. In Rush et al. (2015) , the authors use convolutional models to encode the source, and a context-sensitive attentional feed-forward neural network to generate the summary, producing state-of-the-art results on Gigaword and DUC datasets. In an extension to this work, Chopra et al. (2016) used a similar convolutional model for the encoder, but replaced the decoder with an RNN, producing further improvement in performance on both datasets. In another paper that is closely related to our work, Hu et al. (2015) introduce a large dataset for Chinese short text summarization. They show promising results on their Chinese dataset using an encoder-decoder RNN, but do not report experiments on English corpora. In another very recent work, used RNN based encoder-decoder for extractive summarization of documents.", |
| "cite_spans": [ |
| { |
| "start": 79, |
| "end": 103, |
| "text": "(Collobert et al., 2011)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 239, |
| "end": 257, |
| "text": "Rush et al. (2015)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 731, |
| "end": 747, |
| "text": "Hu et al. (2015)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Our work starts with the same framework as (Hu et al., 2015 ), but we go beyond the stan-dard architecture and propose novel models that address critical problems in summarization. We analyze the similarities and differences of our proposed models with related work on abstractive summarization below. Feature-rich encoder (Sec. 2.2): Linguistic features such as POS tags, and named-entities as well as TF and IDF information were used in many extractive approaches to summarization (Wong et al., 2008b) , but they are novel in the context of deep learning approaches for abstractive summarization, to the best of our knowledge. Switching generator-pointer model (Sec. 2.3): This model combines extractive and abstractive approaches to summarization in a single end-toend framework. Rush et al. (2015) also used a combination of extractive and abstractive approaches, but their extractive model is a separate log-linear classifier with handcrafted features. Pointer networks have also been used earlier for the problem of rare words in the context of machine translation , but the novel addition of switch in our model allows it to strike a balance between when to be faithful to the original source (e.g., for named entities and OOV) and when it is allowed to be creative. We believe such a process arguably mimics how human produces summaries. For a more detailed treatment of this model, and experiments on multiple tasks, please refer to the parallel work published by some of the authors of this work (Gulcehre et al., 2016) . Hierarchical attention model (Sec. 2.4): Previously proposed hierarchical encoder-decoder models use attention only at sentence-level (Li et al., 2015) . The novelty of our approach lies in joint modeling of attention at both sentence and word levels, where the word-level attention is further influenced by sentence-level attention, thus capturing the notion of important sentences and important words within those sentences. Concatenation of positional embeddings with the hidden state at sentence-level is also new.", |
| "cite_spans": [ |
| { |
| "start": 43, |
| "end": 59, |
| "text": "(Hu et al., 2015", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 483, |
| "end": 503, |
| "text": "(Wong et al., 2008b)", |
| "ref_id": "BIBREF28" |
| }, |
| { |
| "start": 783, |
| "end": 801, |
| "text": "Rush et al. (2015)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 1506, |
| "end": 1529, |
| "text": "(Gulcehre et al., 2016)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 1666, |
| "end": 1683, |
| "text": "(Li et al., 2015)", |
| "ref_id": "BIBREF18" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "3" |
| }, |
| { |
| "text": "In this series of experiments 3 , we used the annotated Gigaword corpus as described in Rush et al. (2015) . We used the scripts made available by the authors of this work 4 to preprocess the data, which resulted in about 3.8M training examples. The script also produces about 400K validation and test examples, but we created a randomly sampled subset of 2000 examples each for validation and testing purposes, on which we report our performance. Further, we also acquired the exact test sample used in Rush et al. (2015) to make precise comparison of our models with theirs. We also made small modifications to the script to extract not only the tokenized words, but also systemgenerated parts-of-speech and named-entity tags.", |
| "cite_spans": [ |
| { |
| "start": 88, |
| "end": 106, |
| "text": "Rush et al. (2015)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 504, |
| "end": 522, |
| "text": "Rush et al. (2015)", |
| "ref_id": "BIBREF24" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Gigaword Corpus", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "Training: For all the models we discuss below, we used 200 dimensional word2vec vectors (Mikolov et al., 2013) trained on the same corpus to initialize the model embeddings, but we allowed them to be updated during training. The hidden state dimension of the encoder and decoder was fixed at 400 in all our experiments. When we used only the first sentence of the document as the source, as done in Rush et al. (2015) , the encoder vocabulary size was 119,505 and that of the decoder stood at 68,885. We used Adadelta (Zeiler, 2012) for training, with an initial learning rate of 0.001. We used a batch-size of 50 and randomly shuffled the training data at every epoch, while sorting every 10 batches according to their lengths to speed up training. We did not use any dropout or regularization, but applied gradient clipping. We used early stopping based on the validation set and used the best model on the validation set to report all test performance numbers. For all our models, we employ the large-vocabulary trick, where we restrict the decoder vocabulary size to 2,000 5 , because it cuts down the training time per epoch by nearly three times, and helps this and all subsequent models converge in only 50%-75% of the epochs needed for the model based on full vocabulary. Decoding: At decode-time, we used beam search of size 5 to generate the summary, and limited the size of summary to a maximum of 30 words, since this is the maximum size we noticed in the sampled validation set. We found that the average system summary length from all our models (7.8 to 8.3) agrees very closely with that of the ground truth on the validation set (about 8.7 words), without any specific tuning.", |
| "cite_spans": [ |
| { |
| "start": 88, |
| "end": 110, |
| "text": "(Mikolov et al., 2013)", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 399, |
| "end": 417, |
| "text": "Rush et al. (2015)", |
| "ref_id": "BIBREF24" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Gigaword Corpus", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "We trained all our models on a single Tesla K40 GPU. Most models took about 10 hours per epoch on an average except the hierarchical attention model, which took 12 hours per epoch. All models typically converged within 15 epochs using our early stopping criterion based on the validation cost. The wall-clock training time until convergence therefore varies between 6-8 days depending on the model. Generating summaries at test time is reasonably fast with a throughput of about 20 summaries per second on a single GPU, using a batch size of 1. Evaluation metrics: In Rush et al. (2015) , the authors used full-length version of Rouge recall 6 to evaluate their systems on the Gigaword corpus 7 . However, full-length recall favors longer summaries, so it may not be fair to use this metric to compare two systems that differ in summary lengths. Full-length F1 solves this problem since it can penalize longer summaries. Therefore, we use full-length F1 scores from 1, 2 and L variants of Rouge using the official script to evaluate our systems. However, in the interest of fair comparison with previous work, we also report full-length recall scores where necessary. In addition, we also report the percentage of tokens in the system summary that occur in the source (which we call 'src. copy rate' in Table 1 ). We describe all our experiments and results on the Gigaword corpus below. words-lvt2k-1sent: This is the baseline attentional encoder-decoder model with the large vocabulary trick. This model is trained only on the first sentence from the source document, as done in Rush et al. (2015) . words-lvt2k-2sent: This model is identical to the model above except for the fact that it is trained on the first two sentences from the source. On this corpus, adding the additional sentence in the source does seem to aid performance, as shown in Table 1 . We also tried adding more sentences, but the performance dropped, which is probably because the latter sentences in this corpus are not pertinent to the summary. words-lvt2k-2sent-hieratt: Since we used two sentences from source document, we trained the hierarchical attention model proposed in Sec 2.4. As shown in Table 1 , this model improves perfor-mance compared to its flatter counterpart by learning the relative importance of the first two sentences automatically.", |
| "cite_spans": [ |
| { |
| "start": 568, |
| "end": 586, |
| "text": "Rush et al. (2015)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 1581, |
| "end": 1599, |
| "text": "Rush et al. (2015)", |
| "ref_id": "BIBREF24" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 1303, |
| "end": 1310, |
| "text": "Table 1", |
| "ref_id": "TABREF4" |
| }, |
| { |
| "start": 1850, |
| "end": 1857, |
| "text": "Table 1", |
| "ref_id": "TABREF4" |
| }, |
| { |
| "start": 2176, |
| "end": 2183, |
| "text": "Table 1", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Computational costs:", |
| "sec_num": null |
| }, |
| { |
| "text": "feats-lvt2k-2sent: Here, we still train on the first two sentences, but we exploit the parts-of-speech and named-entity tags in the annotated gigaword corpus as well as TF, IDF values, to augment the input embeddings on the source side as described in Sec 2.2. In total, our embedding vector grew from the original 100 to 155, and produced incremental gains compared to its counterpart words-lvt2k-2sent as shown in Table 1 , demonstrating the utility of syntax based features in this task.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 416, |
| "end": 423, |
| "text": "Table 1", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Computational costs:", |
| "sec_num": null |
| }, |
| { |
| "text": "feats-lvt2k-2sent-ptr: This is the switching generator/pointer model described in Sec. 2.3, but in addition, we also use feature-rich embeddings on the document side as in the above model. Our experiments indicate that the new model is able to achieve the best performance on our test set by all three Rouge variants as shown in Table 1 .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 329, |
| "end": 336, |
| "text": "Table 1", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Computational costs:", |
| "sec_num": null |
| }, |
| { |
| "text": "Comparison with state-of-the-art: (Rush et al., 2015) reported recall-only from full-length version of Rouge, but the authors kindly provided us with their F1 numbers, as well as their test sample. We compared the performance of our model words-lvt2k-1sent with their best system on their sample, on both Recall as well as F1, as displayed in Table 1 . The reason we did not evaluate our best models here is that this test set consisted of only 1 sentence from the source document, and did not include NLP annotations, which are needed in our best models. The table shows that, despite this fact, our model outperforms the state of the art model of Rush et al. (2015) , on both recall and F1, with statistical significance. In addition, our models exhibit better abstractive ability as shown by the src. copy rate metric in the last column of the table.", |
| "cite_spans": [ |
| { |
| "start": 34, |
| "end": 53, |
| "text": "(Rush et al., 2015)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 650, |
| "end": 668, |
| "text": "Rush et al. (2015)", |
| "ref_id": "BIBREF24" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 343, |
| "end": 351, |
| "text": "Table 1", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Computational costs:", |
| "sec_num": null |
| }, |
| { |
| "text": "We believe the bidirectional RNN we used to model the source captures richer contextual information of every word than the bag-of-embeddings representation used by Rush et al. (2015) in their convolutional and attentional encoders, which might explain our superior performance. Further, explicit modeling of important information such as multiple source sentences, word-level linguistic features, using the switch mechanism to point to source words when needed, and hierarchical attention, solve specific problems in summarization, each boosting performance incrementally. (Rush et al., 2015) 6 ABS+ (Rush et al., 2015) 31 (Rush et al., 2015) 8 ABS+ (Rush et al., 2015) 29 respect to the baseline model on its dataset as given by the 95% confidence interval in the official Rouge script. We report statistical significance only for the best performing models. 'src. copy rate' for the reference data on our validation sample is 45%. Please refer to Section 4 for explanation of notation.", |
| "cite_spans": [ |
| { |
| "start": 164, |
| "end": 182, |
| "text": "Rush et al. (2015)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 573, |
| "end": 592, |
| "text": "(Rush et al., 2015)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 600, |
| "end": 619, |
| "text": "(Rush et al., 2015)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 623, |
| "end": 642, |
| "text": "(Rush et al., 2015)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 650, |
| "end": 669, |
| "text": "(Rush et al., 2015)", |
| "ref_id": "BIBREF24" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Computational costs:", |
| "sec_num": null |
| }, |
| { |
| "text": "The DUC corpus 8 comes in two parts: the 2003 corpus consisting of 624 document, summary pairs and the 2004 corpus consisting of 500 pairs. Since these corpora are too small to train large neural networks on, Rush et al. (2015) trained their models on the Gigaword corpus, but combined it with an additional log-linear extractive summarization model with handcrafted features, that is trained on the DUC 2003 corpus. They call the original neural attention model the ABS model, and the combined model ABS+. The latter model is current state-of-the-art since it outperforms all previously published baselines including non-neural network based extractive and abstractive systems, as measured by the official DUC metric of limited-length recall. In these experiments, we use the same metric to evaluate our models too, but we omit reporting numbers from other systems in the interest of space. In our work, we simply run the model trained on Gigaword corpus as it is, without tuning it on the DUC validation set. The only change we made to the decoder is to suppress the model from emitting the end-of-summary tag, and force it to emit exactly 30 words for every summary, since the official evaluation on this corpus is based on limitedlength Rouge recall. On this corpus too, since we have only a single sentence from source and no NLP annotations, we ran just the model words-lvt2k-1sent.", |
| "cite_spans": [ |
| { |
| "start": 209, |
| "end": 227, |
| "text": "Rush et al. (2015)", |
| "ref_id": "BIBREF24" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "DUC Corpus", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "The performance of this model on the test set is compared with ABS and ABS+ models, as well as TOPIARY, the top performing system on DUC-2004 in Table 2 . We note that although our model consistently outperforms ABS+ on all three variants of Rouge, the differences are not statistically significant. However, when the comparison is made with ABS model, which is really the true un-tuned counterpart of our model, the results are indeed statistically significant. We would also like to bring the reader's attention to the concurrently published work of Chopra et al. (2016) where they also used an RNN based decoder for summary generation. While their numbers on Gigaword corpus are slightly better than our best performance on all three Rouge F1 metrics, our performance is marginally higher on DUC-2004 corpus on Rouge-2 and Rouge-L. We believe their work also confirms the effectiveness of RNN-based models for abstractive text summarization.", |
| "cite_spans": [ |
| { |
| "start": 552, |
| "end": 572, |
| "text": "Chopra et al. (2016)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 145, |
| "end": 152, |
| "text": "Table 2", |
| "ref_id": "TABREF6" |
| } |
| ], |
| "eq_spans": [], |
| "section": "DUC Corpus", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "The existing abstractive text summarization corpora including Gigaword and DUC consist of only one sentence in each summary. In this section, we present a new corpus that comprises multisentence summaries. To produce this corpus, we modify an existing corpus that has been used for the task of passage-based question answering (Hermann et al., 2015) . In this work, the authors used the human generated abstractive summary bullets from new-stories in CNN and Daily Mail websites as questions (with one of the entities hidden), and stories as the corresponding passages from which the system is expected to answer the fill-in-the-blank question. The authors released the scripts that crawl, extract and generate pairs of passages and questions from these websites. With a simple modification of the script, we restored all the summary bullets of each story in the original order to obtain a multi-sentence summary, where each bullet is treated as a sentence. In all, this corpus has 286,817 training pairs, 13,368 validation pairs and 11,487 test pairs, as defined by their scripts. The source documents in the training set have 766 words spanning 29.74 sentences on an average while the summaries consist of 53 words and 3.72 sentences. The unique characteristics of this dataset such as long documents, and ordered multi-sentence summaries present interesting challenges, and we hope will attract future researchers to build and test novel models on it.", |
| "cite_spans": [ |
| { |
| "start": 327, |
| "end": 349, |
| "text": "(Hermann et al., 2015)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "CNN/Daily Mail Corpus", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "The dataset is released in two versions: one consisting of actual entity names, and the other, in which entity occurrences are replaced with document-specific integer-ids beginning from 0. Since the vocabulary size is smaller in the anonymized version, we used it in all our experiments below. We limited the source vocabulary size to 150K, and the target vocabulary to 60K, the source and target lengths to at most 800 and 100 words respectively. We used 100-dimensional word2vec embeddings trained on this dataset as input, and we fixed the model hidden state size at 200. We also created explicit pointers in the training data by matching only the anonymized entityids between source and target on similar lines as we did for the OOV words in Gigaword corpus.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "CNN/Daily Mail Corpus", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "We used a single Tesla K-40 GPU to train our models on this dataset as well. While the flat models (words-lvt2k and words-lvt2k-ptr) took under 5 hours per epoch, the hier-archical attention model was very expensive, consuming nearly 12.5 hours per epoch. Convergence of all models is also slower on this dataset compared to Gigaword, taking nearly 35 epochs for all models. Thus, the wall-clock time for training until convergence is about 7 days for the flat models, but nearly 18 days for the hierarchical attention model. Decoding is also slower as well, with a throughput of 2 examples per second for flat models and 1.5 examples per second for the hierarchical attention model, when run on a single GPU with a batch size of 1. Evaluation: We evaluated our models using the full-length Rouge F1 metric that we employed for the Gigaword corpus, but with one notable difference: in both system and gold summaries, we considered each highlight to be a separate sentence. 9 Results: Results from three models we ran on this corpus are displayed in Table 3 . Although this dataset is smaller and more complex than the Gigaword corpus, it is interesting to note that the Rouge numbers are in the same range. However, our switching pointer/generator model as well as the hierarchical attention model described in Sec. 2.4 fail to outperform the baseline attentional decoder, indicating that further research and experimentation needs to be done on this dataset. These results, although preliminary, should serve as a good baseline for future researchers to compare their models against. Table 4 presents a few high quality and poor quality output on the validation set from feats-lvt2k-2sent, one of our best performing models. Even when the model differs from the target summary, its summaries tend to be very meaningful and relevant, a phenomenon not captured by word/phrase matching evaluation metrics such as Rouge. On the other hand, the model sometimes 'misinterprets' the semantics of the text and generates a summary with a comical interpretation as shown in the poor quality examples in the table. Clearly, capturing the 'meaning' of complex sentences remains a weakness of these models.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 1049, |
| "end": 1056, |
| "text": "Table 3", |
| "ref_id": "TABREF8" |
| }, |
| { |
| "start": 1585, |
| "end": 1592, |
| "text": "Table 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Computational costs:", |
| "sec_num": null |
| }, |
| { |
| "text": "Our next example output, presented in Figure 4 , displays the sample output from the switching generator/pointer model on the Gigaword corpus.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 38, |
| "end": 47, |
| "text": "Figure 4", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Qualitative Analysis", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Good quality summary output S: a man charged with the murder last year of a british backpacker confessed to the slaying on the night he was charged with her killing , according to police evidence presented at a court hearing tuesday . ian douglas previte , ## , is charged with murdering caroline stuttle , ## , of yorkshire , england T: man charged with british backpacker 's death confessed to crime police officer claims O: man charged with murdering british backpacker confessed to murder S: following are the leading scorers in the english premier league after saturday 's matches : ## -alan shearer -lrbnewcastle united -rrb-, james beattie . T: leading scorers in english premier league O: english premier league leading scorers S: volume of transactions at the nigerian stock exchange has continued its decline since last week , a nse official said thursday . the latest statistics showed that a total of ##.### million shares valued at ###.### million naira -lrb-about #.### million us dollars -rrb-were traded on wednesday in , deals . T: transactions dip at nigerian stock exchange O: transactions at nigerian stock exchange down Poor quality summary output S: broccoli and broccoli sprouts contain a chemical that kills the bacteria responsible for most stomach cancer , say researchers , confirming the dietary advice that moms have been handing out for years . in laboratory tests the chemical , <unk> , killed helicobacter pylori , a bacteria that causes stomach ulcers and often fatal stomach cancers . T: for release at #### <unk> mom was right broccoli is good for you say cancer researchers O: broccoli sprouts contain deadly bacteria S: norway delivered a diplomatic protest to russia on monday after three norwegian fisheries research expeditions were barred from russian waters . the norwegian research ships were to continue an annual program of charting fish resources shared by the two countries in the barents sea region . T: norway protests russia barring fisheries research ships O: norway grants diplomatic protest to russia S: j.p. morgan chase 's ability to recover from a slew of recent losses rests largely in the hands of two men , who are both looking to restore tarnished reputations and may be considered for the top job someday . geoffrey <unk> , now the co-head of j.p. morgan 's investment bank , left goldman , sachs & co. more than a decade ago after executives say he lost out in a bid to lead that firm . T: # executives to lead j.p. morgan chase on road to recovery O: j.p. morgan chase may be considered for top job Table 4 : Examples of generated summaries from our best model on the validation set of Gigaword corpus. S: source document, T: target summary, O: system output. Although we displayed equal number of good quality and poor quality summaries in the table, the good ones are far more prevalent than the poor ones. It is apparent from the examples that the model learns to use pointers very accurately not only for named entities, but also for multi-word phrases. Despite its accuracy, the performance improvement of the overall model is not significant. We believe the impact of this model may be more pronounced in other settings with a heavier tail distribution of rare words. We intend to carry out more experiments with this model in the future.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 2562, |
| "end": 2569, |
| "text": "Table 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Qualitative Analysis", |
| "sec_num": "5" |
| }, |
| { |
| "text": "On CNN/Daily Mail data, although our models are able to produce good quality multi-sentence summaries, we notice that the same sentence or phrase often gets repeated in the summary. We believe models that incorporate intra-attention such as can fix this problem by encouraging the model to 'remember' the words it has already produced in the past.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Qualitative Analysis", |
| "sec_num": "5" |
| }, |
| { |
| "text": "In this work, we apply the attentional encoderdecoder for the task of abstractive summarization with very promising results, outperforming stateof-the-art results significantly on two different datasets. Each of our proposed novel models addresses a specific problem in abstractive summarization, yielding further improvement in performance. We also propose a new dataset for multisentence summarization and establish benchmark numbers on it. As part of our future work, we plan to focus our efforts on this data and build more robust models for summaries consisting of multiple sentences.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| }, |
| { |
| "text": "Even when the word does not exist in the source vocabulary, the pointer model may still be able to identify the correct position of the word in the source since it takes into account the contextual representation of the corresponding 'UNK' token encoded by the RNN. Once the position is known, the corresponding token from the source document can be displayed in the summary even when it is not part of the training vocabulary either on the source side or the target side.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://duc.nist.gov/", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "We used Kyunghyun Cho's code (https://github. com/kyunghyuncho/dl4mt-material) as the starting point.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "https://github.com/facebook/NAMAS 5 Larger values improved performance only marginally, but at the cost of much slower training.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://www.berouge.com/Pages/default. aspx 7 confirmed from personal communication with the firstauthor of the paper.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://duc.nist.gov/duc2004/tasks.html", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "This was done by modifying the pre-processing script such that each highlight gets its own \"<a>\" tag in the xml file that goes as input to the evaluation script.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Neural machine translation by jointly learning to align and translate", |
| "authors": [ |
| { |
| "first": "Dzmitry", |
| "middle": [], |
| "last": "Bahdanau", |
| "suffix": "" |
| }, |
| { |
| "first": "Kyunghyun", |
| "middle": [], |
| "last": "Cho", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2014. Neural machine translation by jointly learning to align and translate. CoRR, abs/1409.0473.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Dmitriy Serdyuk, Philemon Brakel, and Yoshua Bengio", |
| "authors": [ |
| { |
| "first": "Dzmitry", |
| "middle": [], |
| "last": "Bahdanau", |
| "suffix": "" |
| }, |
| { |
| "first": "Jan", |
| "middle": [], |
| "last": "Chorowski", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dzmitry Bahdanau, Jan Chorowski, Dmitriy Serdyuk, Philemon Brakel, and Yoshua Bengio. 2015.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "End-to-end attention-based large vocabulary speech recognition", |
| "authors": [], |
| "year": null, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "End-to-end attention-based large vocabulary speech recognition. CoRR, abs/1508.04395.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Headline generation based on statistical translation", |
| "authors": [ |
| { |
| "first": "Michele", |
| "middle": [], |
| "last": "Banko", |
| "suffix": "" |
| }, |
| { |
| "first": "O", |
| "middle": [], |
| "last": "Vibhu", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [ |
| "J" |
| ], |
| "last": "Mittal", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Witbrock", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "Proceedings of the 38th Annual Meeting on Association for Computational Linguistics", |
| "volume": "22", |
| "issue": "", |
| "pages": "318--325", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Michele Banko, Vibhu O. Mittal, and Michael J Wit- brock. 2000. Headline generation based on statis- tical translation. In Proceedings of the 38th Annual Meeting on Association for Computational Linguis- tics, 22:318-325.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Neural summarization by extracting sentences and words", |
| "authors": [ |
| { |
| "first": "Jianpeng", |
| "middle": [], |
| "last": "Cheng", |
| "suffix": "" |
| }, |
| { |
| "first": "Mirella", |
| "middle": [], |
| "last": "Lapata", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jianpeng Cheng and Mirella Lapata. 2016. Neural summarization by extracting sentences and words. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Long short-term memory-networks for machine reading", |
| "authors": [ |
| { |
| "first": "Jianpeng", |
| "middle": [], |
| "last": "Cheng", |
| "suffix": "" |
| }, |
| { |
| "first": "Li", |
| "middle": [], |
| "last": "Dong", |
| "suffix": "" |
| }, |
| { |
| "first": "Mirella", |
| "middle": [], |
| "last": "Lapata", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jianpeng Cheng, Li Dong, and Mirella Lapata. 2016. Long short-term memory-networks for machine reading. CoRR, abs/1601.06733.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Abstractive sentence summarization with attentive recurrent neural networks", |
| "authors": [ |
| { |
| "first": "Sumit", |
| "middle": [], |
| "last": "Chopra", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Auli", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexander", |
| "middle": [ |
| "M" |
| ], |
| "last": "Rush", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "HLT-NAACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sumit Chopra, Michael Auli, and Alexander M. Rush. 2016. Abstractive sentence summarization with at- tentive recurrent neural networks. In HLT-NAACL.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Empirical evaluation of gated recurrent neural networks on sequence modeling", |
| "authors": [ |
| { |
| "first": "Junyoung", |
| "middle": [], |
| "last": "Chung", |
| "suffix": "" |
| }, |
| { |
| "first": "\u00c7aglar", |
| "middle": [], |
| "last": "G\u00fcl\u00e7ehre", |
| "suffix": "" |
| }, |
| { |
| "first": "Kyunghyun", |
| "middle": [], |
| "last": "Cho", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Junyoung Chung, \u00c7aglar G\u00fcl\u00e7ehre, KyungHyun Cho, and Yoshua Bengio. 2014. Empirical evaluation of gated recurrent neural networks on sequence model- ing. CoRR, abs/1412.3555.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Sentence compression beyond word deletion", |
| "authors": [ |
| { |
| "first": "Trevor", |
| "middle": [], |
| "last": "Cohn", |
| "suffix": "" |
| }, |
| { |
| "first": "Mirella", |
| "middle": [], |
| "last": "Lapata", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of the 22Nd International Conference on Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "137--144", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Trevor Cohn and Mirella Lapata. 2008. Sentence compression beyond word deletion. In Proceedings of the 22Nd International Conference on Computa- tional Linguistics -Volume 1, pages 137-144.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Natural language processing (almost) from scratch", |
| "authors": [ |
| { |
| "first": "Ronan", |
| "middle": [], |
| "last": "Collobert", |
| "suffix": "" |
| }, |
| { |
| "first": "Jason", |
| "middle": [], |
| "last": "Weston", |
| "suffix": "" |
| }, |
| { |
| "first": "L\u00e9on", |
| "middle": [], |
| "last": "Bottou", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Karlen", |
| "suffix": "" |
| }, |
| { |
| "first": "Koray", |
| "middle": [], |
| "last": "Kavukcuoglu", |
| "suffix": "" |
| }, |
| { |
| "first": "Pavel", |
| "middle": [ |
| "P" |
| ], |
| "last": "Kuksa", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ronan Collobert, Jason Weston, L\u00e9on Bottou, Michael Karlen, Koray Kavukcuoglu, and Pavel P. Kuksa. 2011. Natural language processing (almost) from scratch. CoRR, abs/1103.0398.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Heads: Headline generation as sequence prediction using an abstract feature-rich space", |
| "authors": [ |
| { |
| "first": "Carlos", |
| "middle": [ |
| "A" |
| ], |
| "last": "Colmenares", |
| "suffix": "" |
| }, |
| { |
| "first": "Marina", |
| "middle": [], |
| "last": "Litvak", |
| "suffix": "" |
| }, |
| { |
| "first": "Amin", |
| "middle": [], |
| "last": "Mantrach", |
| "suffix": "" |
| }, |
| { |
| "first": "Fabrizio", |
| "middle": [], |
| "last": "Silvestri", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "133--142", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Carlos A. Colmenares, Marina Litvak, Amin Mantrach, and Fabrizio Silvestri. 2015. Heads: Headline generation as sequence prediction using an abstract feature-rich space. In Proceedings of the 2015 Con- ference of the North American Chapter of the Asso- ciation for Computational Linguistics: Human Lan- guage Technologies, pages 133-142.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Lexrank: Graphbased lexical centrality as salience in text summarization", |
| "authors": [ |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Erkan", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [ |
| "R" |
| ], |
| "last": "Radev", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Journal of Artificial Intelligence Research", |
| "volume": "22", |
| "issue": "", |
| "pages": "457--479", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "G. Erkan and D. R. Radev. 2004. Lexrank: Graph- based lexical centrality as salience in text summa- rization. Journal of Artificial Intelligence Research, 22:457-479.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Overcoming the lack of parallel data in sentence compression", |
| "authors": [ |
| { |
| "first": "Katja", |
| "middle": [], |
| "last": "Filippova", |
| "suffix": "" |
| }, |
| { |
| "first": "Yasemin", |
| "middle": [], |
| "last": "Altun", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "1481--1491", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Katja Filippova and Yasemin Altun. 2013. Overcom- ing the lack of parallel data in sentence compression. In Proceedings of the 2013 Conference on Empiri- cal Methods in Natural Language Processing, pages 1481-1491.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Pointing the unknown words", |
| "authors": [ |
| { |
| "first": "Caglar", |
| "middle": [], |
| "last": "Gulcehre", |
| "suffix": "" |
| }, |
| { |
| "first": "Sungjin", |
| "middle": [], |
| "last": "Ahn", |
| "suffix": "" |
| }, |
| { |
| "first": "Ramesh", |
| "middle": [], |
| "last": "Nallapati", |
| "suffix": "" |
| }, |
| { |
| "first": "Bowen", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Caglar Gulcehre, Sungjin Ahn, Ramesh Nallapati, Bowen Zhou, and Yoshua Bengio. 2016. Pointing the unknown words. In Proceedings of the 54th An- nual Meeting of the Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Teaching machines to read and comprehend", |
| "authors": [ |
| { |
| "first": "Karl", |
| "middle": [], |
| "last": "Moritz Hermann", |
| "suffix": "" |
| }, |
| { |
| "first": "Tom\u00e1s", |
| "middle": [], |
| "last": "Kocisk\u00fd", |
| "suffix": "" |
| }, |
| { |
| "first": "Edward", |
| "middle": [], |
| "last": "Grefenstette", |
| "suffix": "" |
| }, |
| { |
| "first": "Lasse", |
| "middle": [], |
| "last": "Espeholt", |
| "suffix": "" |
| }, |
| { |
| "first": "Will", |
| "middle": [], |
| "last": "Kay", |
| "suffix": "" |
| }, |
| { |
| "first": "Mustafa", |
| "middle": [], |
| "last": "Suleyman", |
| "suffix": "" |
| }, |
| { |
| "first": "Phil", |
| "middle": [], |
| "last": "Blunsom", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Karl Moritz Hermann, Tom\u00e1s Kocisk\u00fd, Edward Grefenstette, Lasse Espeholt, Will Kay, Mustafa Suleyman, and Phil Blunsom. 2015. Teach- ing machines to read and comprehend. CoRR, abs/1506.03340.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Lcsts: A large scale chinese short text summarization dataset", |
| "authors": [ |
| { |
| "first": "Baotian", |
| "middle": [], |
| "last": "Hu", |
| "suffix": "" |
| }, |
| { |
| "first": "Qingcai", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Fangze", |
| "middle": [], |
| "last": "Zhu", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "1967--1972", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Baotian Hu, Qingcai Chen, and Fangze Zhu. 2015. Lc- sts: A large scale chinese short text summarization dataset. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Process- ing, pages 1967-1972, Lisbon, Portugal, September. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "On using very large target vocabulary for neural machine translation. CoRR, abs/1412", |
| "authors": [ |
| { |
| "first": "S\u00e9bastien", |
| "middle": [], |
| "last": "Jean", |
| "suffix": "" |
| }, |
| { |
| "first": "Kyunghyun", |
| "middle": [], |
| "last": "Cho", |
| "suffix": "" |
| }, |
| { |
| "first": "Roland", |
| "middle": [], |
| "last": "Memisevic", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "S\u00e9bastien Jean, Kyunghyun Cho, Roland Memisevic, and Yoshua Bengio. 2014. On using very large target vocabulary for neural machine translation. CoRR, abs/1412.2007.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Long story short \u00e2\u0202\u015e global unsupervised models for keyphrase based meeting summarization", |
| "authors": [ |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Favre", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Riedhammer", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Hakkani-Tur", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Speech Communication", |
| "volume": "", |
| "issue": "", |
| "pages": "801--815", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "B. Favre K. Riedhammer and D. Hakkani-Tur. 2010. Long story short \u00e2\u0202\u015e global unsupervised models for keyphrase based meeting summarization. In Speech Communication, pages 801-815.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "A hierarchical neural autoencoder for paragraphs and documents", |
| "authors": [ |
| { |
| "first": "Jiwei", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Minh-Thang", |
| "middle": [], |
| "last": "Luong", |
| "suffix": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Jurafsky", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jiwei Li, Minh-Thang Luong, and Dan Jurafsky. 2015. A hierarchical neural autoencoder for paragraphs and documents. CoRR, abs/1506.01057.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Graph-based keyword extraction for single-document summarization", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Litvak", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Last", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Coling", |
| "volume": "", |
| "issue": "", |
| "pages": "17--24", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "M. Litvak and M. Last. 2008. Graph-based keyword extraction for single-document summarization. In Coling 2008, pages 17-24.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Addressing the rare word problem in neural machine translation", |
| "authors": [ |
| { |
| "first": "Thang", |
| "middle": [], |
| "last": "Luong", |
| "suffix": "" |
| }, |
| { |
| "first": "Ilya", |
| "middle": [], |
| "last": "Sutskever", |
| "suffix": "" |
| }, |
| { |
| "first": "Quoc", |
| "middle": [ |
| "V" |
| ], |
| "last": "Le", |
| "suffix": "" |
| }, |
| { |
| "first": "Oriol", |
| "middle": [], |
| "last": "Vinyals", |
| "suffix": "" |
| }, |
| { |
| "first": "Wojciech", |
| "middle": [], |
| "last": "Zaremba", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing of the Asian Federation of Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "11--19", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Thang Luong, Ilya Sutskever, Quoc V. Le, Oriol Vinyals, and Wojciech Zaremba. 2015. Addressing the rare word problem in neural machine translation. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Lan- guage Processing of the Asian Federation of Natural Language Processing, pages 11-19.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Distributed representations of words and phrases and their compositionality", |
| "authors": [ |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| }, |
| { |
| "first": "Ilya", |
| "middle": [], |
| "last": "Sutskever", |
| "suffix": "" |
| }, |
| { |
| "first": "Kai", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Greg", |
| "middle": [], |
| "last": "Corrado", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Dean", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg Cor- rado, and Jeffrey Dean. 2013. Distributed represen- tations of words and phrases and their composition- ality. CoRR, abs/1310.4546.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Automatic text summarization using a machine learning approach", |
| "authors": [ |
| { |
| "first": "Joel", |
| "middle": [ |
| "Larocca" |
| ], |
| "last": "Neto", |
| "suffix": "" |
| }, |
| { |
| "first": "Alex", |
| "middle": [ |
| "Alves" |
| ], |
| "last": "Freitas", |
| "suffix": "" |
| }, |
| { |
| "first": "Celso", |
| "middle": [ |
| "A A" |
| ], |
| "last": "Kaestner", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proceedings of the 16th Brazilian Symposium on Artificial Intelligence: Advances in Artificial Intelligence", |
| "volume": "", |
| "issue": "", |
| "pages": "205--215", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Joel Larocca Neto, Alex Alves Freitas, and Celso A. A. Kaestner. 2002. Automatic text summarization us- ing a machine learning approach. In Proceedings of the 16th Brazilian Symposium on Artificial Intel- ligence: Advances in Artificial Intelligence, pages 205-215.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Self reinforcement for important passage retrieval", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Martins De Matos Jo\u00e3\u010do", |
| "suffix": "" |
| }, |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Neto Anatole Gershman Jaime Carbonell Ricardo Ribeiro", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Lu\u00e3 S Marujo", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "36th international ACM SIGIR conference on Research and development in information retrieval", |
| "volume": "", |
| "issue": "", |
| "pages": "845--848", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David Martins de Matos Jo\u00c3\u010do P. Neto Ana- tole Gershman Jaime Carbonell Ricardo Ribeiro, Lu\u00c3 s Marujo. 2013. Self reinforcement for im- portant passage retrieval. In 36th international ACM SIGIR conference on Research and development in information retrieval, pages 845-848.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "A neural attention model for abstractive sentence summarization", |
| "authors": [ |
| { |
| "first": "Alexander", |
| "middle": [ |
| "M" |
| ], |
| "last": "Rush", |
| "suffix": "" |
| }, |
| { |
| "first": "Sumit", |
| "middle": [], |
| "last": "Chopra", |
| "suffix": "" |
| }, |
| { |
| "first": "Jason", |
| "middle": [], |
| "last": "Weston", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alexander M. Rush, Sumit Chopra, and Jason Weston. 2015. A neural attention model for abstractive sen- tence summarization. CoRR, abs/1509.00685.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "Sequence to sequence -video to text", |
| "authors": [ |
| { |
| "first": "Subhashini", |
| "middle": [], |
| "last": "Venugopalan", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcus", |
| "middle": [], |
| "last": "Rohrbach", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeff", |
| "middle": [], |
| "last": "Donahue", |
| "suffix": "" |
| }, |
| { |
| "first": "Raymond", |
| "middle": [ |
| "J" |
| ], |
| "last": "Mooney", |
| "suffix": "" |
| }, |
| { |
| "first": "Trevor", |
| "middle": [], |
| "last": "Darrell", |
| "suffix": "" |
| }, |
| { |
| "first": "Kate", |
| "middle": [], |
| "last": "Saenko", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Subhashini Venugopalan, Marcus Rohrbach, Jeff Don- ahue, Raymond J. Mooney, Trevor Darrell, and Kate Saenko. 2015. Sequence to sequence -video to text. CoRR, abs/1505.00487.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "Pointer Networks. ArXiv e-prints", |
| "authors": [ |
| { |
| "first": "O", |
| "middle": [], |
| "last": "Vinyals", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Fortunato", |
| "suffix": "" |
| }, |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Jaitly", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "O. Vinyals, M. Fortunato, and N. Jaitly. 2015. Pointer Networks. ArXiv e-prints, June.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "Extractive summarization using supervised and semi-supervised learning", |
| "authors": [ |
| { |
| "first": "Kam-Fai", |
| "middle": [], |
| "last": "Wong", |
| "suffix": "" |
| }, |
| { |
| "first": "Mingli", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Wenjie", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of the 22Nd International Conference on Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "985--992", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kam-Fai Wong, Mingli Wu, and Wenjie Li. 2008a. Extractive summarization using supervised and semi-supervised learning. In Proceedings of the 22Nd International Conference on Computational Linguistics -Volume 1, pages 985-992.", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "Extractive summarization using supervised and semi-supervised learning", |
| "authors": [ |
| { |
| "first": "Kam-Fai", |
| "middle": [], |
| "last": "Wong", |
| "suffix": "" |
| }, |
| { |
| "first": "Mingli", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Wenjie", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of the 22nd Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "985--992", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kam-Fai Wong, Mingli Wu, and Wenjie Li. 2008b. Extractive summarization using supervised and semi-supervised learning. In Proceedings of the 22nd Annual Meeting of the Association for Com- putational Linguistics, pages 985-992.", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "Title generation with quasi-synchronous grammar", |
| "authors": [ |
| { |
| "first": "Kristian", |
| "middle": [], |
| "last": "Woodsend", |
| "suffix": "" |
| }, |
| { |
| "first": "Yansong", |
| "middle": [], |
| "last": "Feng", |
| "suffix": "" |
| }, |
| { |
| "first": "Mirella", |
| "middle": [], |
| "last": "Lapata", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing, EMNLP '10", |
| "volume": "", |
| "issue": "", |
| "pages": "513--523", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kristian Woodsend, Yansong Feng, and Mirella Lap- ata. 2010. Title generation with quasi-synchronous grammar. In Proceedings of the 2010 Conference on Empirical Methods in Natural Language Process- ing, EMNLP '10, pages 513-523, Stroudsburg, PA, USA. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF30": { |
| "ref_id": "b30", |
| "title": "Bbn/umd at duc-2004: Topiary", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Zajic", |
| "suffix": "" |
| }, |
| { |
| "first": "Bonnie", |
| "middle": [ |
| "J" |
| ], |
| "last": "Dorr", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Schwartz", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of the North American Chapter of the Association for Computational Linguistics Workshop on Document Understanding", |
| "volume": "", |
| "issue": "", |
| "pages": "112--119", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David Zajic, Bonnie J. Dorr, and Richard Schwartz. 2004. Bbn/umd at duc-2004: Topiary. In Proceed- ings of the North American Chapter of the Asso- ciation for Computational Linguistics Workshop on Document Understanding, pages 112-119.", |
| "links": null |
| }, |
| "BIBREF31": { |
| "ref_id": "b31", |
| "title": "ADADELTA: an adaptive learning rate method", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Matthew", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Zeiler", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Matthew D. Zeiler. 2012. ADADELTA: an adaptive learning rate method. CoRR, abs/1212.5701.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "text": "Feature-rich-encoder: We use one embedding vector each for POS, NER tags and discretized TF and IDF values, which are concatenated together with word-based embeddings as input to the encoder.", |
| "type_str": "figure", |
| "num": null, |
| "uris": null |
| }, |
| "FIGREF1": { |
| "text": "Switching generator/pointer model: When the switch shows 'G', the traditional generator consisting of the softmax layer is used to produce a word, and when it shows 'P', the pointer network is activated to copy the word from one of the source document positions. When the pointer is activated, the embedding from the source is used as input for the next time-step as shown by the arrow from the encoder to the decoder at the bottom.", |
| "type_str": "figure", |
| "num": null, |
| "uris": null |
| }, |
| "FIGREF2": { |
| "text": "Hierarchical encoder with hierarchical attention: the attention weights at the word level, represented by the dashed arrows are re-scaled by the corresponding sentencelevel attention weights, represented by the dotted arrows. The dashed boxes at the bottom of the top layer RNN represent sentence-level positional embeddings concatenated to the corresponding hidden states.", |
| "type_str": "figure", |
| "num": null, |
| "uris": null |
| }, |
| "FIGREF3": { |
| "text": "Sample output from switching generator/pointer networks. An arrow indicates that a pointer to the source position was used to generate the corresponding summary word.", |
| "type_str": "figure", |
| "num": null, |
| "uris": null |
| }, |
| "TABREF1": { |
| "type_str": "table", |
| "text": "Recall on the test set used by", |
| "html": null, |
| "num": null, |
| "content": "<table><tr><td># Model name</td><td colspan=\"4\">Rouge-1 Rouge-2 Rouge-L Src. copy rate (%)</td></tr><tr><td colspan=\"4\">Full length F1 on our internal test set</td><td/></tr><tr><td>1 words-lvt2k-1sent</td><td>34.97</td><td>17.17</td><td>32.70</td><td>75.85</td></tr><tr><td>2 words-lvt2k-2sent</td><td>35.73</td><td>17.38</td><td>33.25</td><td>79.54</td></tr><tr><td>3 words-lvt2k-2sent-hieratt</td><td>36.05</td><td>18.17</td><td>33.52</td><td>78.52</td></tr><tr><td>4 feats-lvt2k-2sent</td><td>35.90</td><td>17.57</td><td>33.38</td><td>78.92</td></tr><tr><td>5 feats-lvt2k-2sent-ptr</td><td>*36.40</td><td>17.77</td><td>*33.71</td><td>78.70</td></tr><tr><td>Full length</td><td/><td/><td/><td/></tr></table>" |
| }, |
| "TABREF4": { |
| "type_str": "table", |
| "text": "Performance comparison of various models. '*' indicates statistical significance of the corresponding model with", |
| "html": null, |
| "num": null, |
| "content": "<table/>" |
| }, |
| "TABREF6": { |
| "type_str": "table", |
| "text": "Evaluation of our models using the limited-length", |
| "html": null, |
| "num": null, |
| "content": "<table><tr><td>Rouge Recall on DUC validation and test sets. Our best</td></tr><tr><td>model, although trained exclusively on the Gigaword corpus,</td></tr><tr><td>consistently outperforms the ABS+ model which is tuned on</td></tr><tr><td>the DUC-2003 validation corpus in addition to being trained</td></tr><tr><td>on the Gigaword corpus.</td></tr></table>" |
| }, |
| "TABREF8": { |
| "type_str": "table", |
| "text": "Performance of various models on CNN/DailyMail test set using full-length Rouge-F1 metric. Bold faced numbers indicate best performing system.", |
| "html": null, |
| "num": null, |
| "content": "<table/>" |
| } |
| } |
| } |
| } |