| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """WikiAuto dataset for Text Simplification""" |
|
|
|
|
| import json |
|
|
| import datasets |
|
|
|
|
| _CITATION = """\ |
| @inproceedings{acl/JiangMLZX20, |
| author = {Chao Jiang and |
| Mounica Maddela and |
| Wuwei Lan and |
| Yang Zhong and |
| Wei Xu}, |
| editor = {Dan Jurafsky and |
| Joyce Chai and |
| Natalie Schluter and |
| Joel R. Tetreault}, |
| title = {Neural {CRF} Model for Sentence Alignment in Text Simplification}, |
| booktitle = {Proceedings of the 58th Annual Meeting of the Association for Computational |
| Linguistics, {ACL} 2020, Online, July 5-10, 2020}, |
| pages = {7943--7960}, |
| publisher = {Association for Computational Linguistics}, |
| year = {2020}, |
| url = {https://www.aclweb.org/anthology/2020.acl-main.709/} |
| } |
| """ |
|
|
| |
| |
| _DESCRIPTION = """\ |
| WikiAuto provides a set of aligned sentences from English Wikipedia and Simple English Wikipedia |
| as a resource to train sentence simplification systems. The authors first crowd-sourced a set of manual alignments |
| between sentences in a subset of the Simple English Wikipedia and their corresponding versions in English Wikipedia |
| (this corresponds to the `manual` config), then trained a neural CRF system to predict these alignments. |
| The trained model was then applied to the other articles in Simple English Wikipedia with an English counterpart to |
| create a larger corpus of aligned sentences (corresponding to the `auto`, `auto_acl`, `auto_full_no_split`, and `auto_full_with_split` configs here). |
| """ |
|
|
| |
| _LICENSE = "CC-BY-SA 3.0" |
|
|
| |
| |
| |
| _URLs = { |
| "manual": { |
| "train": "https://www.dropbox.com/sh/ohqaw41v48c7e5p/AACdl4UPKtu7CMMa-CJhz4G7a/wiki-manual/train.tsv?dl=1", |
| "dev": "https://github.com/chaojiang06/wiki-auto/raw/master/wiki-manual/dev.tsv", |
| "test": "https://github.com/chaojiang06/wiki-auto/raw/master/wiki-manual/test.tsv", |
| }, |
| "auto_acl": { |
| "normal": "https://github.com/chaojiang06/wiki-auto/raw/master/wiki-auto/ACL2020/train.src", |
| "simple": "https://github.com/chaojiang06/wiki-auto/raw/master/wiki-auto/ACL2020/train.dst", |
| }, |
| "auto_full_no_split": { |
| "normal": "https://github.com/chaojiang06/wiki-auto/raw/master/wiki-auto/GEM2021/full_no_split/train.src", |
| "simple": "https://github.com/chaojiang06/wiki-auto/raw/master/wiki-auto/GEM2021/full_no_split/train.dst", |
| }, |
| "auto_full_with_split": { |
| "normal": "https://github.com/chaojiang06/wiki-auto/raw/master/wiki-auto/GEM2021/full_with_split/train.src", |
| "simple": "https://github.com/chaojiang06/wiki-auto/raw/master/wiki-auto/GEM2021/full_with_split/train.dst", |
| }, |
| "auto": { |
| "part_1": "https://www.dropbox.com/sh/ohqaw41v48c7e5p/AAATBDhU1zpdcT5x5WgO8DMaa/wiki-auto-all-data/wiki-auto-part-1-data.json?dl=1", |
| "part_2": "https://www.dropbox.com/sh/ohqaw41v48c7e5p/AAATgPkjo_tPt9z12vZxJ3MRa/wiki-auto-all-data/wiki-auto-part-2-data.json?dl=1", |
| }, |
| } |
|
|
|
|
| |
| class WikiAuto(datasets.GeneratorBasedBuilder): |
| """WikiAuto dataset for sentence simplification""" |
|
|
| VERSION = datasets.Version("1.0.0") |
|
|
| BUILDER_CONFIGS = [ |
| datasets.BuilderConfig( |
| name="manual", |
| version=VERSION, |
| description="A set of 10K Wikipedia sentence pairs aligned by crowd workers.", |
| ), |
| datasets.BuilderConfig( |
| name="auto_acl", |
| version=VERSION, |
| description="Automatically aligned and filtered sentence pairs used to train the ACL2020 system.", |
| ), |
| datasets.BuilderConfig( |
| name="auto_full_no_split", |
| version=VERSION, |
| description="All automatically aligned sentence pairs without sentence splitting.", |
| ), |
| datasets.BuilderConfig( |
| name="auto_full_with_split", |
| version=VERSION, |
| description="All automatically aligned sentence pairs with sentence splitting.", |
| ), |
| datasets.BuilderConfig( |
| name="auto", version=VERSION, description="A large set of automatically aligned sentence pairs." |
| ), |
| ] |
|
|
| DEFAULT_CONFIG_NAME = "auto" |
|
|
| def _info(self): |
| if self.config.name == "manual": |
| features = datasets.Features( |
| { |
| "alignment_label": datasets.ClassLabel(names=["notAligned", "aligned", "partialAligned"]), |
| "normal_sentence_id": datasets.Value("string"), |
| "simple_sentence_id": datasets.Value("string"), |
| "normal_sentence": datasets.Value("string"), |
| "simple_sentence": datasets.Value("string"), |
| "gleu_score": datasets.Value("float32"), |
| } |
| ) |
| elif ( |
| self.config.name == "auto_acl" |
| or self.config.name == "auto_full_no_split" |
| or self.config.name == "auto_full_with_split" |
| ): |
| features = datasets.Features( |
| { |
| "normal_sentence": datasets.Value("string"), |
| "simple_sentence": datasets.Value("string"), |
| } |
| ) |
| else: |
| features = datasets.Features( |
| { |
| "example_id": datasets.Value("string"), |
| "normal": { |
| "normal_article_id": datasets.Value("int32"), |
| "normal_article_title": datasets.Value("string"), |
| "normal_article_url": datasets.Value("string"), |
| "normal_article_content": datasets.Sequence( |
| { |
| "normal_sentence_id": datasets.Value("string"), |
| "normal_sentence": datasets.Value("string"), |
| } |
| ), |
| }, |
| "simple": { |
| "simple_article_id": datasets.Value("int32"), |
| "simple_article_title": datasets.Value("string"), |
| "simple_article_url": datasets.Value("string"), |
| "simple_article_content": datasets.Sequence( |
| { |
| "simple_sentence_id": datasets.Value("string"), |
| "simple_sentence": datasets.Value("string"), |
| } |
| ), |
| }, |
| "paragraph_alignment": datasets.Sequence( |
| { |
| "normal_paragraph_id": datasets.Value("string"), |
| "simple_paragraph_id": datasets.Value("string"), |
| } |
| ), |
| "sentence_alignment": datasets.Sequence( |
| { |
| "normal_sentence_id": datasets.Value("string"), |
| "simple_sentence_id": datasets.Value("string"), |
| } |
| ), |
| } |
| ) |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=features, |
| supervised_keys=None, |
| homepage="https://github.com/chaojiang06/wiki-auto", |
| license=_LICENSE, |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| my_urls = _URLs[self.config.name] |
| data_dir = dl_manager.download_and_extract(my_urls) |
| if self.config.name in ["manual", "auto"]: |
| return [ |
| datasets.SplitGenerator( |
| name=spl, |
| gen_kwargs={ |
| "filepaths": data_dir, |
| "split": spl, |
| }, |
| ) |
| for spl in data_dir |
| ] |
| else: |
| return [ |
| datasets.SplitGenerator( |
| name="full", |
| gen_kwargs={"filepaths": data_dir, "split": "full"}, |
| ) |
| ] |
|
|
| def _generate_examples(self, filepaths, split): |
| if self.config.name == "manual": |
| keys = [ |
| "alignment_label", |
| "simple_sentence_id", |
| "normal_sentence_id", |
| "simple_sentence", |
| "normal_sentence", |
| "gleu_score", |
| ] |
| with open(filepaths[split], encoding="utf-8") as f: |
| for id_, line in enumerate(f): |
| values = line.strip().split("\t") |
| assert len(values) == 6, f"Not enough fields in ---- {line} --- {values}" |
| yield id_, dict( |
| [(k, val) if k != "gleu_score" else (k, float(val)) for k, val in zip(keys, values)] |
| ) |
| elif ( |
| self.config.name == "auto_acl" |
| or self.config.name == "auto_full_no_split" |
| or self.config.name == "auto_full_with_split" |
| ): |
| with open(filepaths["normal"], encoding="utf-8") as fi: |
| with open(filepaths["simple"], encoding="utf-8") as fo: |
| for id_, (norm_se, simp_se) in enumerate(zip(fi, fo)): |
| yield id_, { |
| "normal_sentence": norm_se, |
| "simple_sentence": simp_se, |
| } |
| else: |
| dataset_dict = json.load(open(filepaths[split], encoding="utf-8")) |
| for id_, (eid, example_dict) in enumerate(dataset_dict.items()): |
| res = { |
| "example_id": eid, |
| "normal": { |
| "normal_article_id": example_dict["normal"]["id"], |
| "normal_article_title": example_dict["normal"]["title"], |
| "normal_article_url": example_dict["normal"]["url"], |
| "normal_article_content": { |
| "normal_sentence_id": [ |
| sen_id for sen_id, sen_txt in example_dict["normal"]["content"].items() |
| ], |
| "normal_sentence": [ |
| sen_txt for sen_id, sen_txt in example_dict["normal"]["content"].items() |
| ], |
| }, |
| }, |
| "simple": { |
| "simple_article_id": example_dict["simple"]["id"], |
| "simple_article_title": example_dict["simple"]["title"], |
| "simple_article_url": example_dict["simple"]["url"], |
| "simple_article_content": { |
| "simple_sentence_id": [ |
| sen_id for sen_id, sen_txt in example_dict["simple"]["content"].items() |
| ], |
| "simple_sentence": [ |
| sen_txt for sen_id, sen_txt in example_dict["simple"]["content"].items() |
| ], |
| }, |
| }, |
| "paragraph_alignment": { |
| "normal_paragraph_id": [ |
| norm_id for simp_id, norm_id in example_dict.get("paragraph_alignment", []) |
| ], |
| "simple_paragraph_id": [ |
| simp_id for simp_id, norm_id in example_dict.get("paragraph_alignment", []) |
| ], |
| }, |
| "sentence_alignment": { |
| "normal_sentence_id": [ |
| norm_id for simp_id, norm_id in example_dict.get("sentence_alignment", []) |
| ], |
| "simple_sentence_id": [ |
| simp_id for simp_id, norm_id in example_dict.get("sentence_alignment", []) |
| ], |
| }, |
| } |
| yield id_, res |
|
|