| import json |
| import os |
|
|
| import datasets |
| from datasets.tasks import TextClassification |
|
|
| _CITATION = None |
|
|
|
|
| _DESCRIPTION = """ |
| PubMed dataset for summarization. |
| From paper: A Discourse-Aware Attention Model for Abstractive Summarization of Long Documents" by A. Cohan et al. |
| See: https://aclanthology.org/N18-2097.pdf |
| See: https://github.com/armancohan/long-summarization |
| """ |
| _CITATION = """\ |
| @inproceedings{cohan-etal-2018-discourse, |
| title = "A Discourse-Aware Attention Model for Abstractive Summarization of Long Documents", |
| author = "Cohan, Arman and |
| Dernoncourt, Franck and |
| Kim, Doo Soon and |
| Bui, Trung and |
| Kim, Seokhwan and |
| Chang, Walter and |
| Goharian, Nazli", |
| booktitle = "Proceedings of the 2018 Conference of the North {A}merican Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers)", |
| month = jun, |
| year = "2018", |
| address = "New Orleans, Louisiana", |
| publisher = "Association for Computational Linguistics", |
| url = "https://aclanthology.org/N18-2097", |
| doi = "10.18653/v1/N18-2097", |
| pages = "615--621", |
| abstract = "Neural abstractive summarization models have led to promising results in summarizing relatively short documents. We propose the first model for abstractive summarization of single, longer-form documents (e.g., research papers). Our approach consists of a new hierarchical encoder that models the discourse structure of a document, and an attentive discourse-aware decoder to generate the summary. Empirical results on two large-scale datasets of scientific papers show that our model significantly outperforms state-of-the-art models.", |
| } |
| """ |
| _ABSTRACT = "abstract" |
| _ARTICLE = "article" |
|
|
| class PubMedSummarizationConfig(datasets.BuilderConfig): |
| """BuilderConfig for PubMedSummarization.""" |
|
|
| def __init__(self, **kwargs): |
| """BuilderConfig for PubMedSummarization. |
| Args: |
| **kwargs: keyword arguments forwarded to super. |
| """ |
| super(PubMedSummarizationConfig, self).__init__(**kwargs) |
|
|
|
|
| class PubMedSummarizationDataset(datasets.GeneratorBasedBuilder): |
| """PubMedSummarization Dataset.""" |
| |
| _TRAIN_FILE = "train.zip" |
| _VAL_FILE = "val.zip" |
| _TEST_FILE = "test.zip" |
|
|
| BUILDER_CONFIGS = [ |
| PubMedSummarizationConfig( |
| name="section", |
| version=datasets.Version("1.0.0"), |
| description="PubMed dataset for summarization, concat sections", |
| ), |
| PubMedSummarizationConfig( |
| name="document", |
| version=datasets.Version("1.0.0"), |
| description="PubMed dataset for summarization, document", |
| ), |
| ] |
|
|
| DEFAULT_CONFIG_NAME = "section" |
|
|
| def _info(self): |
| |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=datasets.Features( |
| { |
| _ARTICLE: datasets.Value("string"), |
| _ABSTRACT: datasets.Value("string"), |
| |
| } |
| ), |
| supervised_keys=None, |
| homepage="https://github.com/armancohan/long-summarization", |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
|
|
| train_path = os.path.join(dl_manager.download_and_extract(self._TRAIN_FILE), "train.txt") |
| val_path = os.path.join(dl_manager.download_and_extract(self._VAL_FILE), "val.txt") |
| test_path = os.path.join(dl_manager.download_and_extract(self._TEST_FILE), "test.txt") |
| |
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, gen_kwargs={"filepath": train_path} |
| ), |
| datasets.SplitGenerator( |
| name=datasets.Split.VALIDATION, gen_kwargs={"filepath": val_path} |
| ), |
| datasets.SplitGenerator( |
| name=datasets.Split.TEST, gen_kwargs={"filepath": test_path} |
| ), |
| ] |
| |
| def _generate_examples(self, filepath): |
| """Generate PubMedSummarization examples.""" |
| with open(filepath, encoding="utf-8") as f: |
| for id_, row in enumerate(f): |
| data = json.loads(row) |
|
|
| """ |
| 'article_id': str, |
| 'abstract_text': List[str], |
| 'article_text': List[str], |
| 'section_names': List[str], |
| 'sections': List[List[str]] |
| """ |
| if self.config.name == "document": |
| article = [d.strip() for d in data["article_text"]] |
| article = " ".join(article) |
| else: |
| article = [item.strip() for sublist in data["sections"] for item in sublist] |
| article = " \n ".join(article) |
|
|
| abstract = [ab.replace("<S>", "").replace("</S>", "").strip() for ab in data["abstract_text"]] |
| abstract = " \n ".join(abstract) |
| yield id_, {"article": article, "abstract": abstract} |
|
|