| """LogiQA: A Challenge Dataset for Machine Reading Comprehension with Logical Reasoning""" |
|
|
| import re |
| import datasets |
|
|
| logger = datasets.logging.get_logger(__name__) |
|
|
|
|
| _HOMEPAGE = "https://github.com/lgw863/LogiQA-dataset" |
|
|
| _DESCRIPTION = """\ |
| LogiQA is constructed from the logical comprehension problems from \ |
| publically available questions of the National Civil Servants Examination \ |
| of China, which are designed to test the civil servant candidates’ critical \ |
| thinking and problem solving. This dataset includes the English versions only; \ |
| the Chinese versions are available via the homepage/original source.""" |
|
|
| _CITATION = """\ |
| @article{liu2020logiqa, |
| title={Logiqa: A challenge dataset for machine reading comprehension with logical reasoning}, |
| author={Liu, Jian and Cui, Leyang and Liu, Hanmeng and Huang, Dandan and Wang, Yile and Zhang, Yue}, |
| journal={arXiv preprint arXiv:2007.08124}, |
| year={2020} |
| } |
| """ |
|
|
| _URLS = { |
| "en_train": "https://raw.githubusercontent.com/lgw863/LogiQA-dataset/master/Train.txt", |
| "en_test": "https://raw.githubusercontent.com/lgw863/LogiQA-dataset/master/Test.txt", |
| "en_eval": "https://raw.githubusercontent.com/lgw863/LogiQA-dataset/master/Eval.txt", |
| } |
|
|
| def _process_answer(answer): |
| if not any(answer.startswith(x) for x in "ABCD"): |
| return answer |
| else: |
| return answer[3:] |
|
|
| def _process_sentences(text): |
| text = text.replace("\n", "") |
| sents = text.split(".") |
| text = "" |
| for sent in sents: |
| if len(sent) == 0: |
| continue |
| if len(text) == 0: |
| text += sent |
| elif sent[0].isnumeric(): |
| text += "."+sent |
| else: |
| text += ". "+sent |
| text = text.replace(" ", " ") |
| text = text.replace("\\'", "'") |
| while text.endswith(" "): |
| text = text[:-1] |
| if re.match('^[A-Z][\w\s]+[?.!]$', text) is None: |
| text += "." |
| text = text.replace("?.", "?") |
| text = text.replace("!.", "!") |
| text = text.replace("..", ".") |
| return text |
|
|
| class LogiQA(datasets.GeneratorBasedBuilder): |
| def _info(self): |
| features = datasets.Features( |
| { |
| "context": datasets.Value("string"), |
| "query": datasets.Value("string"), |
| "options": datasets.features.Sequence(datasets.Value("string")), |
| "correct_option": datasets.Value("int32") |
| } |
| ) |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=features, |
| homepage=_HOMEPAGE, |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| downloaded_files = dl_manager.download_and_extract(_URLS) |
| return [ |
| datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": downloaded_files["en_train"]}), |
| datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={"filepath": downloaded_files["en_eval"]}), |
| datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"filepath": downloaded_files["en_test"]}), |
| ] |
|
|
| def _generate_examples(self, filepath): |
| logger.info("generating examples from = %s", filepath) |
| with open(filepath, encoding="utf-8") as f: |
| logiqa = f.readlines() |
| logiqa = [_process_sentences(s) for s in logiqa] |
|
|
| for key in range(int(len(logiqa)/8)): |
| row = 8*key |
| correct_answer = logiqa[row+1].replace(".","") |
| context = logiqa[row+2] |
| query = logiqa[row+3] |
| answers = logiqa[row+4:row+8] |
|
|
| yield key, { |
| "context": context, |
| "query": query, |
| "options": [_process_answer(answers[i]) for i in range(4)], |
| "correct_option": "abcd".index(correct_answer) |
| } |
|
|