| | |
| |
|
| | """Mintaka: A Complex, Natural, and Multilingual Dataset for End-to-End Question Answering""" |
| |
|
| | import json |
| | import datasets |
| |
|
| | logger = datasets.logging.get_logger(__name__) |
| |
|
| | _DESCRIPTION = """\ |
| | Mintaka is a complex, natural, and multilingual dataset designed for experimenting with end-to-end |
| | question-answering models. Mintaka is composed of 20,000 question-answer pairs collected in English, |
| | annotated with Wikidata entities, and translated into Arabic, French, German, Hindi, Italian, |
| | Japanese, Portuguese, and Spanish for a total of 180,000 samples. |
| | Mintaka includes 8 types of complex questions, including superlative, intersection, and multi-hop questions, |
| | which were naturally elicited from crowd workers. |
| | """ |
| |
|
| | _CITATION = """\ |
| | @inproceedings{sen-etal-2022-mintaka, |
| | title = "Mintaka: A Complex, Natural, and Multilingual Dataset for End-to-End Question Answering", |
| | author = "Sen, Priyanka and Aji, Alham Fikri and Saffari, Amir", |
| | booktitle = "Proceedings of the 29th International Conference on Computational Linguistics", |
| | month = oct, |
| | year = "2022", |
| | address = "Gyeongju, Republic of Korea", |
| | publisher = "International Committee on Computational Linguistics", |
| | url = "https://aclanthology.org/2022.coling-1.138", |
| | pages = "1604--1619" |
| | } |
| | """ |
| |
|
| | _LICENSE = """\ |
| | Copyright Amazon.com Inc. or its affiliates. |
| | Attribution 4.0 International |
| | """ |
| |
|
| | _TRAIN_URL = "https://raw.githubusercontent.com/amazon-science/mintaka/main/data/mintaka_train.json" |
| | _DEV_URL = "https://raw.githubusercontent.com/amazon-science/mintaka/main/data/mintaka_dev.json" |
| | _TEST_URL = "https://raw.githubusercontent.com/amazon-science/mintaka/main/data/mintaka_test.json" |
| |
|
| |
|
| | _LANGUAGES = ['en', 'ar', 'de', 'ja', 'hi', 'pt', 'es', 'it', 'fr'] |
| |
|
| | _ALL = "all" |
| |
|
| | class Mintaka(datasets.GeneratorBasedBuilder): |
| | """Mintaka: A Complex, Natural, and Multilingual Dataset for End-to-End Question Answering""" |
| |
|
| | BUILDER_CONFIGS = [ |
| | datasets.BuilderConfig( |
| | name = name, |
| | version = datasets.Version("1.0.0"), |
| | description = f"Mintaka: A Complex, Natural, and Multilingual Dataset for End-to-End Question Answering for {name}", |
| | ) for name in _LANGUAGES |
| | ] |
| |
|
| | BUILDER_CONFIGS.append(datasets.BuilderConfig( |
| | name = _ALL, |
| | version = datasets.Version("1.0.0"), |
| | description = f"Mintaka: A Complex, Natural, and Multilingual Dataset for End-to-End Question Answering", |
| | )) |
| |
|
| | DEFAULT_CONFIG_NAME = 'en' |
| |
|
| | def _info(self): |
| | return datasets.DatasetInfo( |
| | description=_DESCRIPTION, |
| | features=datasets.Features( |
| | { |
| | "id": datasets.Value("string"), |
| | "lang": datasets.Value("string"), |
| | "question": datasets.Value("string"), |
| | "answerText": datasets.Value("string"), |
| | "category": datasets.Value("string"), |
| | "complexityType": datasets.Value("string"), |
| | "questionEntity": [{ |
| | "name": datasets.Value("string"), |
| | "entityType": datasets.Value("string"), |
| | "label": datasets.Value("string"), |
| | "mention": datasets.Value("string"), |
| | "span": [datasets.Value("int32")], |
| | }], |
| | "answerEntity": [{ |
| | "name": datasets.Value("string"), |
| | "label": datasets.Value("string"), |
| | }] |
| | }, |
| | ), |
| | supervised_keys=None, |
| | citation=_CITATION, |
| | license=_LICENSE, |
| | ) |
| |
|
| | def _split_generators(self, dl_manager): |
| | return [ |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TRAIN, |
| | gen_kwargs={ |
| | "file": dl_manager.download_and_extract(_TRAIN_URL), |
| | "lang": self.config.name, |
| | } |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split.VALIDATION, |
| | gen_kwargs={ |
| | "file": dl_manager.download_and_extract(_DEV_URL), |
| | "lang": self.config.name, |
| | } |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TEST, |
| | gen_kwargs={ |
| | "file": dl_manager.download_and_extract(_TEST_URL), |
| | "lang": self.config.name, |
| | } |
| | ), |
| | ] |
| |
|
| | def _generate_examples(self, file, lang): |
| | if lang == _ALL: |
| | langs = _LANGUAGES |
| | else: |
| | langs = [lang] |
| |
|
| | key_ = 0 |
| |
|
| | logger.info("⏳ Generating examples from = %s", ", ".join(lang)) |
| |
|
| | with open(file, encoding='utf-8') as json_file: |
| | data = json.load(json_file) |
| | for lang in langs: |
| | for sample in data: |
| | questionEntity = [ |
| | { |
| | "name": str(qe["name"]), |
| | "entityType": qe["entityType"], |
| | "label": qe["label"] if "label" in qe else "", |
| | |
| | "mention": qe["mention"] if lang == "en" else None, |
| | "span": qe["span"] if lang == "en" else [], |
| | } for qe in sample["questionEntity"] |
| | ] |
| |
|
| | answers = [] |
| | if sample['answer']["answerType"] == "entity" and sample['answer']['answer'] is not None: |
| | answers = sample['answer']['answer'] |
| | elif sample['answer']["answerType"] == "numerical" and "supportingEnt" in sample["answer"]: |
| | answers = sample['answer']['supportingEnt'] |
| |
|
| | |
| | def get_label(labels, lang): |
| | if lang in labels: |
| | return labels[lang] |
| | if 'en' in labels: |
| | return labels['en'] |
| | return None |
| |
|
| | answerEntity = [ |
| | { |
| | "name": str(ae["name"]), |
| | "label": get_label(ae["label"], lang), |
| | } for ae in answers |
| | ] |
| |
|
| | yield key_, { |
| | "id": sample["id"], |
| | "lang": lang, |
| | "question": sample["question"] if lang == 'en' else sample['translations'][lang], |
| | "answerText": sample["answer"]["mention"], |
| | "category": sample["category"], |
| | "complexityType": sample["complexityType"], |
| | "questionEntity": questionEntity, |
| | "answerEntity": answerEntity, |
| | } |
| |
|
| | key_ += 1 |
| |
|