| | import os |
| | from itertools import chain |
| | from pathlib import Path |
| | from typing import Dict, List, Tuple |
| |
|
| | import datasets |
| |
|
| | from seacrowd.utils import schemas |
| | from seacrowd.utils.configs import SEACrowdConfig |
| | from seacrowd.utils.constants import Licenses, Tasks |
| |
|
| | _CITATION = """\ |
| | @inproceedings{widiaputri-etal-5641, |
| | author = {Widiaputri, Ruhiyah Faradishi and Purwarianti, Ayu and Lestari, Dessi Puji and Azizah, Kurniawati and Tanaya, Dipta and Sakti, Sakriani}, |
| | title = {Speech Recognition and Meaning Interpretation: Towards Disambiguation of Structurally Ambiguous Spoken Utterances in Indonesian}, |
| | booktitle = {Proceedings of the EMNLP 2023}, |
| | year = {2023} |
| | } |
| | """ |
| |
|
| | _DATASETNAME = "struct_amb_ind" |
| |
|
| | _DESCRIPTION = """ |
| | This dataset contains the first Indonesian speech dataset for structurally ambiguous utterances and each of transcription and two disambiguation texts. |
| | The structurally ambiguous sentences were adapted from Types 4,5,6, and 10 of Types Of Syntactic Ambiguity in English by [Taha et al., 1983]. |
| | For each chosen type, 100 structurally ambiguous sentences in Indonesian were made by crowdsourcing. |
| | Each Indonesian ambiguous sentence has two possible interpretations, resulting in two disambiguation text outputs for each ambiguous sentence. |
| | Each disambiguation text is made up of two sentences. All of the sentences have been checked by linguists. |
| | """ |
| |
|
| | _HOMEPAGE = "https://github.com/ha3ci-lab/struct_amb_ind" |
| |
|
| | _LICENSE = Licenses.UNKNOWN.value |
| |
|
| | _LOCAL = True |
| | _LANGUAGES = ["ind"] |
| |
|
| | _URL_TEMPLATES = { |
| | "keys": "https://raw.githubusercontent.com/ha3ci-lab/struct_amb_ind/main/keys/train_dev_test_spk_keys/", |
| | "text": "https://raw.githubusercontent.com/ha3ci-lab/struct_amb_ind/main/text/", |
| | } |
| |
|
| | _URLS = { |
| | "split_train": _URL_TEMPLATES["keys"] + "train_spk", |
| | "split_dev": _URL_TEMPLATES["keys"] + "dev_spk", |
| | "split_test": _URL_TEMPLATES["keys"] + "test_spk", |
| | "text_transcript": _URL_TEMPLATES["text"] + "ID_amb_disam_transcript.txt", |
| | } |
| |
|
| | _SUPPORTED_TASKS = [Tasks.SPEECH_RECOGNITION] |
| |
|
| | _SOURCE_VERSION = "1.0.0" |
| |
|
| | _SEACROWD_VERSION = "2024.06.20" |
| |
|
| |
|
| | class StructAmbInd(datasets.GeneratorBasedBuilder): |
| | """ |
| | This dataset contains the first Indonesian speech dataset for structurally ambiguous utterances and each of transcription and two disambiguation texts. |
| | This dataloader does NOT contain the additional training data for as mentioned in the _HOMEPAGE, as it is already implemented in the dataloader "indspeech_news_lvcsr". |
| | """ |
| |
|
| | SOURCE_VERSION = datasets.Version(_SOURCE_VERSION) |
| | SEACROWD_VERSION = datasets.Version(_SEACROWD_VERSION) |
| |
|
| | BUILDER_CONFIGS = [ |
| | SEACrowdConfig( |
| | name=f"{_DATASETNAME}_source", |
| | version=SOURCE_VERSION, |
| | description=f"{_DATASETNAME} source schema", |
| | schema="source", |
| | subset_id=f"{_DATASETNAME}", |
| | ), |
| | SEACrowdConfig( |
| | name=f"{_DATASETNAME}_seacrowd_sptext", |
| | version=SEACROWD_VERSION, |
| | description=f"{_DATASETNAME} SEACrowd schema", |
| | schema="seacrowd_sptext", |
| | subset_id=f"{_DATASETNAME}", |
| | ), |
| | ] |
| |
|
| | DEFAULT_CONFIG_NAME = f"{_DATASETNAME}_source" |
| |
|
| | def _info(self) -> datasets.DatasetInfo: |
| | if self.config.schema == "source": |
| | features = datasets.Features( |
| | { |
| | "id": datasets.Value("string"), |
| | "speaker_id": datasets.Value("string"), |
| | "path": datasets.Value("string"), |
| | "audio": datasets.Audio(sampling_rate=16_000), |
| | "amb_transcript": datasets.Value("string"), |
| | "disam_text": datasets.Value("string"), |
| | } |
| | ) |
| |
|
| | elif self.config.schema == "seacrowd_sptext": |
| | features = schemas.speech_text_features |
| |
|
| | return datasets.DatasetInfo( |
| | description=_DESCRIPTION, |
| | features=features, |
| | homepage=_HOMEPAGE, |
| | license=_LICENSE, |
| | citation=_CITATION, |
| | ) |
| |
|
| | def _split_generators(self, dl_manager: datasets.DownloadManager) -> List[datasets.SplitGenerator]: |
| | |
| | if self.config.data_dir is None: |
| | raise ValueError("This is a local dataset. Please pass the data_dir kwarg to load_dataset.") |
| | else: |
| | data_dir = self.config.data_dir |
| |
|
| | |
| | audio_urls = [data_dir + "/" + f"{gender}{_id:02}.zip" for gender in ["F", "M"] for _id in range(1, 12, 1)] |
| | audio_files_dir = [Path(dl_manager.extract(audio_url)) / audio_url.split("/")[-1][:-4] for audio_url in audio_urls] |
| | |
| | split_train = Path(dl_manager.download(_URLS["split_train"])) |
| | split_dev = Path(dl_manager.download(_URLS["split_dev"])) |
| | split_test = Path(dl_manager.download(_URLS["split_test"])) |
| | text_transcript = Path(dl_manager.download(_URLS["text_transcript"])) |
| |
|
| | return [ |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TRAIN, |
| | gen_kwargs={"split": split_train, "transcript": text_transcript, "audio_files_dir": audio_files_dir}, |
| | ), |
| | datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={"split": split_dev, "transcript": text_transcript, "audio_files_dir": audio_files_dir}), |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TEST, |
| | gen_kwargs={"split": split_test, "transcript": text_transcript, "audio_files_dir": audio_files_dir}, |
| | ), |
| | ] |
| |
|
| | def _generate_examples(self, split: Path, transcript: Path, audio_files_dir: List[Path]) -> Tuple[int, Dict]: |
| | speaker_ids = open(split, "r").readlines() |
| | speaker_ids = [id.replace("\n", "") for id in speaker_ids] |
| | speech_folders = [audio_folder for audio_folder in audio_files_dir if audio_folder.name.split("/")[-1] in speaker_ids] |
| | speech_files = list(chain(*[list(map((str(speech_folder) + "/").__add__, os.listdir(speech_folder))) for speech_folder in speech_folders])) |
| |
|
| | transcript = open(transcript, "r").readlines() |
| | transcript = [sent.replace("\n", "").split("|") for sent in transcript] |
| | transcript_dict = {sent[0]: {"amb_transcript": sent[1], "disam_text": sent[2]} for sent in transcript} |
| |
|
| | for key, aud_file in enumerate(speech_files): |
| | aud_id = aud_file.split("/")[-1][:-4] |
| | aud_info = aud_id.split("_") |
| | if self.config.schema == "source": |
| | row = { |
| | "id": aud_id, |
| | "speaker_id": aud_info[1], |
| | "path": aud_file, |
| | "audio": aud_file, |
| | "amb_transcript": transcript_dict[aud_id]["amb_transcript"], |
| | "disam_text": transcript_dict[aud_id]["disam_text"], |
| | } |
| | yield key, row |
| | elif self.config.schema == "seacrowd_sptext": |
| | row = { |
| | "id": aud_id, |
| | "path": aud_file, |
| | "audio": aud_file, |
| | "text": transcript_dict[aud_id]["amb_transcript"], |
| | "speaker_id": aud_info[1], |
| | "metadata": { |
| | "speaker_age": None, |
| | "speaker_gender": aud_info[1][0], |
| | }, |
| | } |
| | yield key, row |
| | else: |
| | raise NotImplementedError(f"Schema '{self.config.schema}' is not defined.") |
| |
|