| { |
| "paper_id": "2020", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T10:23:29.640757Z" |
| }, |
| "title": "Extractive Financial Narrative Summarisation based on DPPs", |
| "authors": [ |
| { |
| "first": "Lei", |
| "middle": [], |
| "last": "Li", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "leili@bupt.edu.cn" |
| }, |
| { |
| "first": "Yafei", |
| "middle": [], |
| "last": "Jiang", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "jiangyafei@bupt.edu.cn" |
| }, |
| { |
| "first": "Yinan", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "We participate in the FNS-Summarisation 2020 shared task to be held at FNP 2020 workshop at COLING 2020. Based on Determinantal Point Processes (DPPs), we build an extractive automatic financial summarisation system for the specific task. In this system, we first analyze the long report data to select the important narrative parts and generate an intermediate document. Next, we build the kernel Matrix L for the intermediate document, which represents the quality of its sentences. On the basis of L, we then can use the DPPs sampling algorithm to choose those sentences with high quality and diversity as the final summary sentences.", |
| "pdf_parse": { |
| "paper_id": "2020", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "We participate in the FNS-Summarisation 2020 shared task to be held at FNP 2020 workshop at COLING 2020. Based on Determinantal Point Processes (DPPs), we build an extractive automatic financial summarisation system for the specific task. In this system, we first analyze the long report data to select the important narrative parts and generate an intermediate document. Next, we build the kernel Matrix L for the intermediate document, which represents the quality of its sentences. On the basis of L, we then can use the DPPs sampling algorithm to choose those sentences with high quality and diversity as the final summary sentences.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "With the development and progress of the economy, the popularity of the financial sector is increasing day by day. There are quite a few new companies gradually emerging and going public, but investors often find it difficult to deal with the long-form annual reports of various companies, because their content may be tedious and redundant, and it is difficult to filter out effective key information by human resources, so an automatic digest system is needed to help investors effectively screen company information.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, we try to solve the first task contained in FNS (Financial Narrative Summarisation) 2020, the dataset we use is the annual reports in the financial field provided by the organizer. As there is much redundant information in the reports, we plan to select the most useful parts first as the intermediate documents to be summarized. For the shared task of summarisation, we implement a Quality-Diversity model (QD) based on Determinantal Point Processes (DPPs) to represent the intermediate document, and merge three kinds of features to rank the sentences.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Li L et al. (2018) focused on exploring the sampling process. They used WMD sentence similarity to construct new kernel matrix used in Determinantal Point Processes (DPPs). Ma S et al. (2018) divided all sentences into three categories (motivations, methods, and conclusions), and then extracted sentences from each cluster based on rules and severe features to form a summary. Debnath D et al. (2018) built a summary generation system using the OpenNMT tool. Zhong M et al. (2019) analyzed the relationship between the quality of extractive automatic summarization and the model network, and the influence of network architecture, knowledge transfer and learning mode on the effect of the summary system is verified through a series of experiments. Liu Y and Lapata M (2019) studied the influence of pre-training language models in automatic summarization tasks, and emphasized the importance of document encoding. In the extractive summarization task, they used multiple layers of BERT as the encoder, and obtained good results on three datasets. with around 80 pages on average, some annual reports could span over more than 250 pages. These UK annual reports are divided into training, testing and validation sets. The training and validation sets include the full text of each annual report along with the gold-standard summaries. On average there are at least 3 gold-standard summaries for each annual report with some reports containing up to 7 goldstandard summaries. For testing dataset, the task participants were given access only to the full texts. Table 1 shows the dataset details.", |
| "cite_spans": [ |
| { |
| "start": 173, |
| "end": 191, |
| "text": "Ma S et al. (2018)", |
| "ref_id": null |
| }, |
| { |
| "start": 378, |
| "end": 401, |
| "text": "Debnath D et al. (2018)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 460, |
| "end": 481, |
| "text": "Zhong M et al. (2019)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 1561, |
| "end": 1568, |
| "text": "Table 1", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Training Validation Testing Report full text 3000 363 500 Gold summaries 9873 1250 - ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data Type", |
| "sec_num": null |
| }, |
| { |
| "text": "In our system, we present an original Quality-Diversity model for extractive automatic summarisation based on DPPs sampling algorithm (Kulesza and Taskar, 2012) . In this model, a single input document can be regarded as a set of sentences, and the process of extracting abstracts can be regarded as sampling a subset of important sentences from that set. To choose the high-quality and diversity summaries with DPPs, we need to construct the kernel matrix L to represent the document. The main steps of summary generation include pre-processing, feature selection and sentence sampling. Pre-processing, feature selection and sentence sampling are as followed.", |
| "cite_spans": [ |
| { |
| "start": 134, |
| "end": 160, |
| "text": "(Kulesza and Taskar, 2012)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System", |
| "sec_num": "4" |
| }, |
| { |
| "text": "In this task the summary requires extraction from different key sections found in the annual reports. Those sections are usually referred to as \"narrative sections\" or \"front-end\" sections and they usually contain textual information and reviews by the firm's management and board of directors. Sections containing financial statements in terms of tables and numbers are usually referred to as \"back-end\" sections and are not supposed to be part of the narrative summaries. Therefore, in the data pre-processing part, there are mainly two steps: the first step is to detect and extract the narrative sections of the annual reports; the second step is to organize the format of the dataset, and clean the narrative sections extracted in the previous step. The final processed data represents the important narrative parts and will be used for subsequent summary generation/extraction.More implementation details of pre-processing are given below. First, we analyze the dataset and try to catch some underlying laws. Comparing the annual reports and gold summaries, we find that each gold summary contains some sentences appearing in the corresponding annual report, and each corresponds to a continuous piece of content. Upon further analysis, we can discover that the gold summary basically corresponds to four sections in the annual report, namely Highlights, At a glance, Chairman's statement and Chief Executive's review. Although not every annual report contains these four sections and the names of the sections may be different, the contents of these sections are similar and are all narrative sections. Therefore, we extract all these sections included in each annual report and use them as a new dataset together with the gold summaries.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-processing", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "Next, we perform further data cleaning and formatting on the new dataset obtained in the previous step. The dataset provided by the organizer is directly converted from pdf files to txt files, thus the data format is confusing and has many useless content such as headers and footers. The reason why they are considered useless is that headers and footers information of each page is basically the same, such as ANNUAL REPORT AND ACCOUNTS XXX, where XXX represents the year. In order to store sentences by lines, we divide the text according to punctuation marks such as periods, exclamation points, etc. In order to delete non-narrative texts such as tables and numbers, we first use regular expressions to filter financial numbers, and then use the open source NLP toolkit to perform dependency syntactic analysis on the data. If a sentence does not contain the subject, predicate and object, We consider the sentence to be unqualified and delete it. Finally, we use the processed data to generate intermediate documents for the subsequent summarisation instead of the original long annual reports.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-processing", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "To represent the document, we build a kernel matrix L from statistical feature method. We use Sentence Length (SL), Sentence Position (SP), Sentence Coverage (SC) as features according to the work of (Li L et al., 2017) to construct matrix L. Sentence Length can be calculated by", |
| "cite_spans": [ |
| { |
| "start": 200, |
| "end": 219, |
| "text": "(Li L et al., 2017)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Selection", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "sl i = exp \u22121 * (len i \u2212 \u00b5) 2 \u03c3 2 (1)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Selection", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "where sl i is the sentence length feature score, \u00b5 and \u03c3 are the mean and standard deviation of sentence length respectively. Sentence Position can be calculated by", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Selection", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "sp i = 1 \u2212 s i |D| (2)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Selection", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "where s i represents the position of the i \u2212 th sentence in the text, |D| represents the total number of sentences in the text, and sp i is the score of the sentence position feature. Sentence Coverage can be calculated by", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Selection", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "SC i = \u03a3 |S| i=1 num S (word i ) n |S| (3)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Selection", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "where sc i represents the score of sentence coverage, word i represents the i \u2212 th word in sentence S, num s (word i ) represents the total number of sentences covered by word i , and |S| represents the total number of words of in the sentence, n represents the total number of sentences in the input sequence. Elements of L can be calculated by", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Selection", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "L ij = q i Sim ij q j (4)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Selection", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "where q i is the quality of a single sentence, Sim ij represents the similarity between sentences. q i can be calculated by the features of SL, SP and SC. We use the Jaccard similarity between sentences to measure the degree of sentence diversity, so Sim ij can be computed as", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Selection", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Sim ij =", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Selection", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "| {word|word \u2208 sen i and word \u2208 sen j } | | {word|word \u2208 sen i or word \u2208 sen j } |", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Selection", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "It can be seen that the matrix L has the function of measuring the quality and diversity of sentences, which is quite important in further sampling process. For the shared task, we select one feature or the sum of multiple features to get the sentence quality q and arrange these methods into three runs.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Selection", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "In our approach, we use discrete DPPs to select sentences, by constructing matrix L, we can apply the DPPs sampling algorithm to extract summaries. To perform DPPs sampling, we first get the eigenvalue \u03bb n and the eigenvector v n of the matrix L, then project all the sentence vectors into a new lowdimensional feature space. In this space, the magnitude of the vector describes the importance and the cosine similarity between two vectors describes the similarity between sentences. When selecting sentences, the number of sentences in the summary is obtained by probability model.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentence Sampling", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Next, we select a sentence with high quality x i , then remove a list of feature vectors v i in the feature space that contributes to its vector modulus length, and perform series of orthogonalization on feature space.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentence Sampling", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Finally we select new elements according to the new sub-feature space re-projection until the end, which can ensure that the element selected again are both high-quality and low-similar to the previous element, thus choose those high-quality and diversity sentences as summary sentences. the details of DPPs can be referred to the work of Kulesza and Taskar(2012) .", |
| "cite_spans": [ |
| { |
| "start": 339, |
| "end": 363, |
| "text": "Kulesza and Taskar(2012)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentence Sampling", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "In the shared task, we submitted a total of three versions of our system. CIST-BUPT-RUN3 uses SC as the feature that represents document, CIST-BUPT-RUN2 uses SL, while CIST-BUPT-RUN1 merges all three features: SC, SL and SP. We tried three versions on the pre-processed training set and table 2 shows the performance from our experiments.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "5" |
| }, |
| { |
| "text": "ROUGE-1/F ROUGE-2/F ROUGE-L/F CIST-BUPT-RUN1 0.305 0.073 0.166 CIST-BUPT-RUN2 0.293 0.067 0.164 CIST-BUPT-RUN3 0.313 0.079 0.167 The result shows that the systems using different features have similar performance, while CIST-BUPT-RUN3 performs the best and the performance of others are slightly lower. As is shown in table 3, CIST-BUPT-RUN3 still performs the best. We find that the ROUGE score on testing set is higher than that on training set, the reason may be that because for each document in the official training set, there are multiple golden abstracts corresponding to them, we choose to merge these summaries for evaluation, so the ROUGE score is a bit lower. But even so, our experimental results can reflect the pros and cons of the three models. The model of DPPs with SC as feature for the kernel matrix shows the best performance among the three models in our system. We can also infer that SC catches more relations between sentences and thus is superior to the features of sentence's length and position.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System", |
| "sec_num": null |
| }, |
| { |
| "text": "System R-L/R R-L/P R-L/F R-1/R R-1/P R-1/F CIST-BUPT-RUN1", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System", |
| "sec_num": null |
| }, |
| { |
| "text": "Actually, there are multiple ways to represent the document, one typical way is to train a Sentence2Vec model using neural network, which may improve the performance of the model. But because of the COVID-19, the servers available are limited, we don't have enough computing resources to train the model. In the future, we'll try other feasible methods of document representation, seeing if they can help improve the quality of summaries.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion and Future Work", |
| "sec_num": "6" |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "CIST@ CLSciSumm-18: Methods for Computational Linguistics Scientific Citation Linkage", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Chi", |
| "middle": [ |
| "J" |
| ], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "84--95", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Li L, Chi J, Chen M, et al. 2018. CIST@ CLSciSumm-18: Methods for Computational Linguistics Scientific Citation Linkage, Facet Classification and Summarization[C]//BIRNDL@ SIGIR. 2018: 84-95.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "NLP-NITMZ@ CLScisumm-18 In: Proceedings of the 3nd Joint Workshop on Bibliometric-enhanced Information Retrieval and Natural Language Processing for Digital Libraries[C]//BIRNDL@ SIGIR", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Debnath", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Achom", |
| "suffix": "" |
| }, |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Pakray", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "164--171", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Debnath D, Achom A, Pakray P. 2018. NLP-NITMZ@ CLScisumm-18 In: Proceedings of the 3nd Joint Workshop on Bibliometric-enhanced Information Retrieval and Natural Language Processing for Digital Li- braries[C]//BIRNDL@ SIGIR. 2018: 164-171.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Searching for Effective Neural Extractive Summarization: What Works and What's Next", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Zhong", |
| "suffix": "" |
| }, |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1907.03491" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zhong M, Liu P, Wang D et al. 2019. Searching for Effective Neural Extractive Summarization: What Works and What's Next. arXiv preprint arXiv:1907.03491.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Text summarization with pretrained encoders", |
| "authors": [ |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Lapata", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1908.08345" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Liu Y and Lapata M. 2019. Text summarization with pretrained encoders. arXiv preprint arXiv:1908.08345.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Determinantal Point Processes for Machine Learning", |
| "authors": [ |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Kulesza", |
| "suffix": "" |
| }, |
| { |
| "first": "Ben", |
| "middle": [], |
| "last": "Taskar", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Foundations and Trends in Machine Learning", |
| "volume": "5", |
| "issue": "23", |
| "pages": "123--286", |
| "other_ids": { |
| "DOI": [ |
| "10.1561/2200000044" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alex Kulesza and Ben Taskar 2012. Determinantal Point Processes for Machine Learning, Foundations and Trends in Machine Learning: Vol. 5: No. 23, pp 123-286. http://dx.doi.org/10.1561/2200000044.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Chinese Computational Linguistics and Natural Language Processing Based on Naturally Annotated Big Data", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Chi", |
| "middle": [ |
| "J" |
| ], |
| "last": "", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "343--354", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Li L, Zhang Y, Chi J et al. 2017. UIDS: A Multilingual Document Summarization Frame-work Based on Summary Diversity and Hierarchical Topics [M] // Li L, Zhang Y, Chi J et al. Chinese Computational Linguistics and Natural Language Processing Based on Naturally Annotated Big Data. Springer, 2017: 2017: 343-354.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "TABREF0": { |
| "type_str": "table", |
| "text": "For the FNS 2020 Shared task we use 3863 UK annual reports for firms listed on The London Stock Exchange (LSE) covering the period between 2002 and 2017. UK annual reports are lengthy documents", |
| "content": "<table><tr><td>This work is licensed under a Creative Commons Attribution 4.0 International License. License details: http:</td></tr><tr><td>//creativecommons.org/licenses/by/4.0/.</td></tr></table>", |
| "html": null, |
| "num": null |
| }, |
| "TABREF1": { |
| "type_str": "table", |
| "text": "", |
| "content": "<table/>", |
| "html": null, |
| "num": null |
| }, |
| "TABREF2": { |
| "type_str": "table", |
| "text": "Results on Official Validation Set", |
| "content": "<table/>", |
| "html": null, |
| "num": null |
| }, |
| "TABREF4": { |
| "type_str": "table", |
| "text": "", |
| "content": "<table/>", |
| "html": null, |
| "num": null |
| } |
| } |
| } |
| } |