| | from pathlib import Path |
| |
|
| | import datasets |
| | import pandas as pd |
| |
|
| | from seacrowd.utils import schemas |
| | from seacrowd.utils.configs import SEACrowdConfig |
| | from seacrowd.utils.constants import Licenses, Tasks |
| |
|
| | _CITATION = """ |
| | @inproceedings{10.1145/3628797.3628837, |
| | author = {Nguyen, Duc-Vu and Nguyen, Quoc-Nam}, |
| | title = {Evaluating the Symbol Binding Ability of Large Language Models for Multiple-Choice Questions in Vietnamese General Education}, |
| | year = {2023}, |
| | isbn = {9798400708916}, |
| | publisher = {Association for Computing Machinery}, |
| | address = {New York, NY, USA}, |
| | url = {https://doi.org/10.1145/3628797.3628837}, |
| | doi = {10.1145/3628797.3628837}, |
| | booktitle = {Proceedings of the 12th International Symposium on Information and Communication Technology}, |
| | pages = {379–386}, |
| | numpages = {8}, |
| | keywords = {Analysis of Language Models, Multiple Choice Symbol Binding, Multiple Choice Question Answering, Language Modeling}, |
| | location = {<conf-loc>, <city>Ho Chi Minh</city>, <country>Vietnam</country>, </conf-loc>}, |
| | series = {SOICT '23} |
| | } |
| | """ |
| |
|
| | _DATASETNAME = "vigetext" |
| |
|
| | _DESCRIPTION = """ |
| | The high-quality dataset with structured guidelines for typing LaTeX formulas in Mathematics, Physics, Chemistry, and |
| | Biology. Objective was to cover the entire scope of the Vietnamese General Education Examination spanning from 2017 to 2023. |
| | This comprehensive approach included the challenging examinations of the years 2017 and 2018, which have been significant |
| | for nearly all Vietnamese students in recent years. It is important to highlight that the exact and unquestionably correct |
| | answers have been exclusively obtained from the Vietnamese Ministry of Education. |
| | """ |
| |
|
| | _HOMEPAGE = "https://huggingface.co/datasets/uitnlp/ViGEText_17to23" |
| |
|
| | _LANGUAGES = ["vie"] |
| |
|
| | _LICENSE = Licenses.UNKNOWN.value |
| |
|
| | _LOCAL = False |
| |
|
| | _URLS = { |
| | _DATASETNAME: { |
| | "train": "https://huggingface.co/datasets/uitnlp/ViGEText_17to23/resolve/main/data/train-00000-of-00001.parquet", |
| | "validation": "https://huggingface.co/datasets/uitnlp/ViGEText_17to23/resolve/main/data/validation-00000-of-00001.parquet", |
| | "test": "https://huggingface.co/datasets/uitnlp/ViGEText_17to23/resolve/main/data/test-00000-of-00001.parquet", |
| | } |
| | } |
| |
|
| | _SUPPORTED_TASKS = [Tasks.QUESTION_ANSWERING] |
| |
|
| | _SOURCE_VERSION = "1.0.0" |
| |
|
| | _SEACROWD_VERSION = "2024.06.20" |
| |
|
| |
|
| | class VigetextDataset(datasets.GeneratorBasedBuilder): |
| | """Vigetext is a dataset for evaluating the Symbol Binding Ability of Large Language Models for Multiple-Choice Questions in Vietnamese General Education.""" |
| |
|
| | SOURCE_VERSION = datasets.Version(_SOURCE_VERSION) |
| | SEACROWD_VERSION = datasets.Version(_SEACROWD_VERSION) |
| |
|
| | BUILDER_CONFIGS = [ |
| | SEACrowdConfig( |
| | name=f"{_DATASETNAME}_source", |
| | version=SOURCE_VERSION, |
| | description=f"{_DATASETNAME} source schema", |
| | schema="source", |
| | subset_id=_DATASETNAME, |
| | ), |
| | SEACrowdConfig( |
| | name=f"{_DATASETNAME}_seacrowd_qa", |
| | version=SEACROWD_VERSION, |
| | description=f"{_DATASETNAME} SEACrowd schema", |
| | schema="seacrowd_qa", |
| | subset_id=_DATASETNAME, |
| | ), |
| | ] |
| |
|
| | DEFAULT_CONFIG_NAME = f"{_DATASETNAME}_source" |
| |
|
| | def _info(self) -> datasets.DatasetInfo: |
| | if self.config.schema == "source": |
| | features = datasets.Features( |
| | { |
| | "id": datasets.Value("string"), |
| | "input": datasets.Value("string"), |
| | "target": datasets.Value("string"), |
| | } |
| | ) |
| |
|
| | else: |
| | features = schemas.qa_features |
| |
|
| | return datasets.DatasetInfo( |
| | description=_DESCRIPTION, |
| | features=features, |
| | homepage=_HOMEPAGE, |
| | license=_LICENSE, |
| | citation=_CITATION, |
| | ) |
| |
|
| | def _split_generators(self, dl_manager: datasets.DownloadManager) -> list[datasets.SplitGenerator]: |
| | """Returns SplitGenerators.""" |
| | urls = _URLS[_DATASETNAME] |
| | data_dir = dl_manager.download_and_extract(urls) |
| |
|
| | return [ |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TRAIN, |
| | gen_kwargs={"filepath": data_dir, "split": "train"}, |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split.VALIDATION, |
| | gen_kwargs={"filepath": data_dir, "split": "validation"}, |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TEST, |
| | gen_kwargs={"filepath": data_dir, "split": "test"}, |
| | ), |
| | ] |
| |
|
| | def _generate_examples(self, filepath: Path, split: str) -> tuple[int, dict]: |
| | df = pd.read_parquet(filepath[split]) |
| | data = df.to_dict(orient="records") |
| | for i, item in enumerate(data): |
| | if self.config.schema == "source": |
| | yield i, { |
| | "id": item["id"], |
| | "input": item["input"], |
| | "target": item["target"], |
| | } |
| | else: |
| | question_and_options = item["input"].split("\n") |
| | answer_map = {opt[0]: opt[2:].strip() for opt in question_and_options[1:]} |
| | yield i, { |
| | "id": str(i), |
| | "question_id": item["id"], |
| | "document_id": "", |
| | "question": question_and_options[0], |
| | "type": "multiple_choice", |
| | "choices": [opt[2:].strip() for opt in question_and_options[1:]], |
| | "context": "", |
| | "answer": [answer_map[item["target"]]], |
| | "meta": {} |
| | } |
| |
|