| import datasets |
| import collections |
| import gzip |
| import textwrap |
|
|
| _DESCRIPTION = "tba" |
| _URL = "tba" |
| _CITATION = "tba" |
| _LICENSE = "tba" |
|
|
| class VoxLinguaConfig(datasets.BuilderConfig): |
| """VoxLingua107 corpus.""" |
| def __init__( |
| self, |
| features, |
| url, |
| data_url=None, |
| supervised_keys=None, |
| shuffled=False, |
| deduplicated=False, |
| **kwargs, |
| ): |
| super(VoxLinguaConfig, self).__init__(version=datasets.Version("1.9.0", ""), **kwargs) |
| self.features = features |
| self.data_url = data_url |
| self.url = url |
| self.supervised_keys = supervised_keys |
|
|
| def _languages(): |
| """Create the sorted dictionary of language codes, and language names. |
| Returns: |
| The sorted dictionary as an instance of `collections.OrderedDict`. |
| """ |
| langs = { |
| "af":"", |
| "am":"", |
| "ar":"", |
| "as":"", |
| "az":"", |
| "ba":"", |
| "be":"", |
| "bg":"", |
| "bn":"", |
| "bo":"", |
| "br":"", |
| "bs":"", |
| "ca":"", |
| "ceb":"", |
| "cs":"", |
| "cy":"", |
| "da":"", |
| "de":"", |
| "el":"", |
| "en":"", |
| "eo":"", |
| "es":"", |
| "et":"", |
| "eu":"", |
| "fa":"", |
| "fi":"", |
| "fo":"", |
| "fr":"", |
| "gl":"", |
| "gn":"", |
| "gu":"", |
| "gv":"", |
| "ha":"", |
| "haw":"", |
| "hi":"", |
| "hr":"", |
| "ht":"", |
| "hu":"", |
| "hy":"", |
| "ia":"", |
| "id":"", |
| "is":"", |
| "it":"", |
| "iw":"", |
| "ja":"", |
| "jw":"", |
| "ka":"", |
| "kk":"", |
| "km":"", |
| "kn":"", |
| "ko":"", |
| "la":"", |
| "lb":"", |
| "ln":"", |
| "lo":"", |
| "lt":"", |
| "lv":"", |
| "mg":"", |
| "mi":"", |
| "mk":"", |
| "ml":"", |
| "mn":"", |
| "mr":"", |
| "ms":"", |
| "mt":"", |
| "my":"", |
| "ne":"", |
| "nl":"", |
| "nn":"", |
| "no":"", |
| "oc":"", |
| "pa":"", |
| "pl":"", |
| "ps":"", |
| "pt":"", |
| "ro":"", |
| "ru":"", |
| "sa":"", |
| "sco":"", |
| "sd":"", |
| "si":"", |
| "sk":"", |
| "sl":"", |
| "sn":"", |
| "so":"", |
| "sq":"", |
| "sr":"", |
| "su":"", |
| "sv":"", |
| "sw":"", |
| "ta":"", |
| "te":"", |
| "tg":"", |
| "th":"", |
| "tk":"", |
| "tl":"", |
| "tr":"", |
| "tt":"", |
| "uk":"", |
| "ur":"", |
| "uz":"", |
| "vi":"", |
| "war":"", |
| "yi":"", |
| "yo":"", |
| "zh":"" |
| } |
| return collections.OrderedDict(sorted(langs.items())) |
|
|
|
|
| class VoxLingua(datasets.GeneratorBasedBuilder): |
| BUILDER_CONFIGS = [ |
| VoxLinguaConfig( |
| name = language, |
| description=textwrap.dedent( |
| """ tbd """ |
| ), |
| shuffled=False, |
| deduplicated=False, |
| features=datasets.Features( |
| { |
| "file": datasets.Value("string"), |
| "audio": datasets.Audio(sampling_rate=16_000), |
| "label": datasets.ClassLabel( |
| names=[f"{i}" for i in range(107)] |
| ), |
| } |
| ), |
| supervised_keys=("file", "label"), |
| url="http://bark.phon.ioc.ee/voxlingua107/", |
| data_url="http://bark.phon.ioc.ee/voxlingua107/{language}.zip" |
| ) |
| for language in _languages() |
| ] |
|
|
| BUILDER_CONFIG_CLASS = VoxLinguaConfig |
|
|
| def _info(self): |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=self.config.features, |
| supervised_keys=self.config.supervised_keys, |
| homepage=self.config.url, |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| train_data_urls = [self.config.url + f"{key}.zip" for key in _languages().keys()] |
| downloaded_files_train = dl_manager.download(train_data_urls) |
| dev_data_url = [self.config.url + f"dev.zip"] |
| downloaded_files_dev = dl_manager.download(dev_data_url) |
| return [ |
| datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"archive_path": downloaded_files_train}), |
| datasets.SplitGenerator(name=datasets.Split.DEV, gen_kwargs={"archive_path": downloaded_files_dev}), |
| ] |
| |
| def _generate_examples(self, archive_path, split=None): |
| return "" |
|
|
| |