lhoestq HF Staff commited on
Commit
e739376
·
verified ·
1 Parent(s): 86b798b

Delete loading script

Browse files
Files changed (1) hide show
  1. voxpopuli.py +0 -223
voxpopuli.py DELETED
@@ -1,223 +0,0 @@
1
- from collections import defaultdict
2
- import os
3
- import json
4
- import csv
5
-
6
- import datasets
7
-
8
-
9
- _DESCRIPTION = """
10
- A large-scale multilingual speech corpus for representation learning, semi-supervised learning and interpretation.
11
- """
12
-
13
- _CITATION = """
14
- @inproceedings{wang-etal-2021-voxpopuli,
15
- title = "{V}ox{P}opuli: A Large-Scale Multilingual Speech Corpus for Representation Learning,
16
- Semi-Supervised Learning and Interpretation",
17
- author = "Wang, Changhan and
18
- Riviere, Morgane and
19
- Lee, Ann and
20
- Wu, Anne and
21
- Talnikar, Chaitanya and
22
- Haziza, Daniel and
23
- Williamson, Mary and
24
- Pino, Juan and
25
- Dupoux, Emmanuel",
26
- booktitle = "Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics
27
- and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)",
28
- month = aug,
29
- year = "2021",
30
- publisher = "Association for Computational Linguistics",
31
- url = "https://aclanthology.org/2021.acl-long.80",
32
- doi = "10.18653/v1/2021.acl-long.80",
33
- pages = "993--1003",
34
- }
35
- """
36
-
37
- _HOMEPAGE = "https://github.com/facebookresearch/voxpopuli"
38
-
39
- _LICENSE = "CC0, also see https://www.europarl.europa.eu/legal-notice/en/"
40
-
41
- _ASR_LANGUAGES = [
42
- "en", "de", "fr", "es", "pl", "it", "ro", "hu", "cs", "nl", "fi", "hr",
43
- "sk", "sl", "et", "lt"
44
- ]
45
- _ASR_ACCENTED_LANGUAGES = [
46
- "en_accented"
47
- ]
48
-
49
- _LANGUAGES = _ASR_LANGUAGES + _ASR_ACCENTED_LANGUAGES
50
-
51
- _BASE_DATA_DIR = "data/"
52
-
53
- _N_SHARDS_FILE = _BASE_DATA_DIR + "n_files.json"
54
-
55
- _AUDIO_ARCHIVE_PATH = _BASE_DATA_DIR + "{lang}/{split}/{split}_part_{n_shard}.tar.gz"
56
-
57
- _METADATA_PATH = _BASE_DATA_DIR + "{lang}/asr_{split}.tsv"
58
-
59
-
60
- class VoxpopuliConfig(datasets.BuilderConfig):
61
- """BuilderConfig for VoxPopuli."""
62
-
63
- def __init__(self, name, languages="all", **kwargs):
64
- """
65
- Args:
66
- name: `string` or `List[string]`:
67
- name of a config: either one of the supported languages or "multilang" for many languages.
68
- By default, "multilang" config includes all languages, including accented ones.
69
- To specify a custom set of languages, pass them to the `languages` parameter
70
- languages: `List[string]`: if config is "multilang" can be either "all" for all available languages,
71
- excluding accented ones (default), or a custom list of languages.
72
- **kwargs: keyword arguments forwarded to super.
73
- """
74
- if name == "multilang":
75
- self.languages = _ASR_LANGUAGES if languages == "all" else languages
76
- name = "multilang" if languages == "all" else "_".join(languages)
77
- else:
78
- self.languages = [name]
79
-
80
- super().__init__(name=name, **kwargs)
81
-
82
-
83
- class Voxpopuli(datasets.GeneratorBasedBuilder):
84
- """The VoxPopuli dataset."""
85
-
86
- VERSION = datasets.Version("1.3.0") # TODO: version
87
- BUILDER_CONFIGS = [
88
- VoxpopuliConfig(
89
- name=name,
90
- version=datasets.Version("1.3.0"),
91
- )
92
- for name in _LANGUAGES + ["multilang"]
93
- ]
94
- DEFAULT_WRITER_BATCH_SIZE = 256
95
-
96
- def _info(self):
97
- features = datasets.Features(
98
- {
99
- "audio_id": datasets.Value("string"),
100
- "language": datasets.ClassLabel(names=_LANGUAGES),
101
- "audio": datasets.Audio(sampling_rate=16_000),
102
- "raw_text": datasets.Value("string"),
103
- "normalized_text": datasets.Value("string"),
104
- "gender": datasets.Value("string"), # TODO: ClassVar?
105
- "speaker_id": datasets.Value("string"),
106
- "is_gold_transcript": datasets.Value("bool"),
107
- "accent": datasets.Value("string"),
108
- }
109
- )
110
- return datasets.DatasetInfo(
111
- description=_DESCRIPTION,
112
- features=features,
113
- homepage=_HOMEPAGE,
114
- license=_LICENSE,
115
- citation=_CITATION,
116
- )
117
-
118
- def _split_generators(self, dl_manager):
119
- n_shards_path = dl_manager.download_and_extract(_N_SHARDS_FILE)
120
- with open(n_shards_path) as f:
121
- n_shards = json.load(f)
122
-
123
- if self.config.name == "en_accented":
124
- splits = ["test"]
125
- else:
126
- splits = ["train", "dev", "test"]
127
-
128
- audio_urls = defaultdict(dict)
129
- for split in splits:
130
- for lang in self.config.languages:
131
- audio_urls[split][lang] = [
132
- _AUDIO_ARCHIVE_PATH.format(lang=lang, split=split, n_shard=i) for i in range(n_shards[lang][split])
133
- ]
134
-
135
- meta_urls = defaultdict(dict)
136
- for split in splits:
137
- for lang in self.config.languages:
138
- meta_urls[split][lang] = _METADATA_PATH.format(lang=lang, split=split)
139
-
140
- # dl_manager.download_config.num_proc = len(urls)
141
-
142
- meta_paths = dl_manager.download_and_extract(meta_urls)
143
- audio_paths = dl_manager.download(audio_urls)
144
-
145
- local_extracted_audio_paths = (
146
- dl_manager.extract(audio_paths) if not dl_manager.is_streaming else
147
- {
148
- split: {lang: [None] * len(audio_paths[split][lang]) for lang in self.config.languages} for split in splits
149
- }
150
- )
151
- if self.config.name == "en_accented":
152
- return [
153
- datasets.SplitGenerator(
154
- name=datasets.Split.TEST,
155
- gen_kwargs={
156
- "audio_archives": {
157
- lang: [dl_manager.iter_archive(archive) for archive in lang_archives]
158
- for lang, lang_archives in audio_paths["test"].items()
159
- },
160
- "local_extracted_archives_paths": local_extracted_audio_paths["test"],
161
- "metadata_paths": meta_paths["test"],
162
- }
163
- ),
164
- ]
165
-
166
- return [
167
- datasets.SplitGenerator(
168
- name=datasets.Split.TRAIN,
169
- gen_kwargs={
170
- "audio_archives": {
171
- lang: [dl_manager.iter_archive(archive) for archive in lang_archives]
172
- for lang, lang_archives in audio_paths["train"].items()
173
- },
174
- "local_extracted_archives_paths": local_extracted_audio_paths["train"],
175
- "metadata_paths": meta_paths["train"],
176
- }
177
- ),
178
- datasets.SplitGenerator(
179
- name=datasets.Split.VALIDATION,
180
- gen_kwargs={
181
- "audio_archives": {
182
- lang: [dl_manager.iter_archive(archive) for archive in lang_archives]
183
- for lang, lang_archives in audio_paths["dev"].items()
184
- },
185
- "local_extracted_archives_paths": local_extracted_audio_paths["dev"],
186
- "metadata_paths": meta_paths["dev"],
187
- }
188
- ),
189
- datasets.SplitGenerator(
190
- name=datasets.Split.TEST,
191
- gen_kwargs={
192
- "audio_archives": {
193
- lang: [dl_manager.iter_archive(archive) for archive in lang_archives]
194
- for lang, lang_archives in audio_paths["test"].items()
195
- },
196
- "local_extracted_archives_paths": local_extracted_audio_paths["test"],
197
- "metadata_paths": meta_paths["test"],
198
- }
199
- ),
200
- ]
201
-
202
- def _generate_examples(self, audio_archives, local_extracted_archives_paths, metadata_paths):
203
- assert len(metadata_paths) == len(audio_archives) == len(local_extracted_archives_paths)
204
- features = ["raw_text", "normalized_text", "speaker_id", "gender", "is_gold_transcript", "accent"]
205
-
206
- for lang in self.config.languages:
207
- assert len(audio_archives[lang]) == len(local_extracted_archives_paths[lang])
208
-
209
- meta_path = metadata_paths[lang]
210
- with open(meta_path) as f:
211
- metadata = {x["id"]: x for x in csv.DictReader(f, delimiter="\t")}
212
-
213
- for audio_archive, local_extracted_archive_path in zip(audio_archives[lang], local_extracted_archives_paths[lang]):
214
- for audio_filename, audio_file in audio_archive:
215
- audio_id = audio_filename.split(os.sep)[-1].split(".wav")[0]
216
- path = os.path.join(local_extracted_archive_path, audio_filename) if local_extracted_archive_path else audio_filename
217
-
218
- yield audio_id, {
219
- "audio_id": audio_id,
220
- "language": lang,
221
- **{feature: metadata[audio_id][feature] for feature in features},
222
- "audio": {"path": path, "bytes": audio_file.read()},
223
- }