Justyna Krzywdziak commited on
Commit
da6be8b
·
1 Parent(s): 64560d1
Files changed (1) hide show
  1. test.py +165 -166
test.py CHANGED
@@ -3,10 +3,13 @@
3
  """test set"""
4
 
5
 
 
6
  import os
 
7
 
8
  import datasets
9
- from datasets.tasks import AutomaticSpeechRecognition
 
10
 
11
 
12
  _CITATION = """\
@@ -24,9 +27,9 @@ _DESCRIPTION = """\
24
  Lorem ipsum
25
  """
26
 
27
- _URL = "https://huggingface.co/datasets/j-krzywdziak/test/blob/main/dev_ngrams.tsv"
28
- _DL_URL = "https://huggingface.co/datasets/j-krzywdziak/test/blob/main/clips.tar.gz"
29
-
30
 
31
  # _DL_URLS = {
32
  # "clean": {
@@ -90,173 +93,169 @@ class TestASR(datasets.GeneratorBasedBuilder):
90
  "ngram": datasets.Value("string")
91
  }
92
  ),
93
- supervised_keys=("file", "text"),
94
- homepage=_URL,
95
- citation=_CITATION,
96
- task_templates=[AutomaticSpeechRecognition(audio_column="audio", transcription_column="text")],
97
  )
98
 
99
  def _split_generators(self, dl_manager):
100
- archive_path = dl_manager.download(_DL_URLS[self.config.name])
101
  # (Optional) In non-streaming mode, we can extract the archive locally to have actual local audio files:
102
  local_extracted_archive = dl_manager.extract(archive_path) if not dl_manager.is_streaming else {}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
103
 
104
- if self.config.name == "clean":
105
- train_splits = [
106
- datasets.SplitGenerator(
107
- name="train.100",
108
- gen_kwargs={
109
- "local_extracted_archive": local_extracted_archive.get("train.100"),
110
- "files": dl_manager.iter_archive(archive_path["train.100"]),
111
- },
112
- ),
113
- datasets.SplitGenerator(
114
- name="train.360",
115
- gen_kwargs={
116
- "local_extracted_archive": local_extracted_archive.get("train.360"),
117
- "files": dl_manager.iter_archive(archive_path["train.360"]),
118
- },
119
- ),
120
- ]
121
- dev_splits = [
122
- datasets.SplitGenerator(
123
- name=datasets.Split.VALIDATION,
124
- gen_kwargs={
125
- "local_extracted_archive": local_extracted_archive.get("dev"),
126
- "files": dl_manager.iter_archive(archive_path["dev"]),
127
- },
128
- )
129
- ]
130
- test_splits = [
131
- datasets.SplitGenerator(
132
- name=datasets.Split.TEST,
133
- gen_kwargs={
134
- "local_extracted_archive": local_extracted_archive.get("test"),
135
- "files": dl_manager.iter_archive(archive_path["test"]),
136
- },
137
- )
138
- ]
139
- elif self.config.name == "other":
140
- train_splits = [
141
- datasets.SplitGenerator(
142
- name="train.500",
143
- gen_kwargs={
144
- "local_extracted_archive": local_extracted_archive.get("train.500"),
145
- "files": dl_manager.iter_archive(archive_path["train.500"]),
146
- },
147
- )
148
- ]
149
- dev_splits = [
150
- datasets.SplitGenerator(
151
- name=datasets.Split.VALIDATION,
152
- gen_kwargs={
153
- "local_extracted_archive": local_extracted_archive.get("dev"),
154
- "files": dl_manager.iter_archive(archive_path["dev"]),
155
- },
156
- )
157
- ]
158
- test_splits = [
159
- datasets.SplitGenerator(
160
- name=datasets.Split.TEST,
161
- gen_kwargs={
162
- "local_extracted_archive": local_extracted_archive.get("test"),
163
- "files": dl_manager.iter_archive(archive_path["test"]),
164
- },
165
- )
166
- ]
167
- elif self.config.name == "all":
168
- train_splits = [
169
- datasets.SplitGenerator(
170
- name="train.clean.100",
171
- gen_kwargs={
172
- "local_extracted_archive": local_extracted_archive.get("train.clean.100"),
173
- "files": dl_manager.iter_archive(archive_path["train.clean.100"]),
174
- },
175
- ),
176
- datasets.SplitGenerator(
177
- name="train.clean.360",
178
- gen_kwargs={
179
- "local_extracted_archive": local_extracted_archive.get("train.clean.360"),
180
- "files": dl_manager.iter_archive(archive_path["train.clean.360"]),
181
- },
182
- ),
183
- datasets.SplitGenerator(
184
- name="train.other.500",
185
- gen_kwargs={
186
- "local_extracted_archive": local_extracted_archive.get("train.other.500"),
187
- "files": dl_manager.iter_archive(archive_path["train.other.500"]),
188
- },
189
- ),
190
- ]
191
- dev_splits = [
192
- datasets.SplitGenerator(
193
- name="validation.clean",
194
- gen_kwargs={
195
- "local_extracted_archive": local_extracted_archive.get("dev.clean"),
196
- "files": dl_manager.iter_archive(archive_path["dev.clean"]),
197
- },
198
- ),
199
- datasets.SplitGenerator(
200
- name="validation.other",
201
- gen_kwargs={
202
- "local_extracted_archive": local_extracted_archive.get("dev.other"),
203
- "files": dl_manager.iter_archive(archive_path["dev.other"]),
204
- },
205
- ),
206
- ]
207
- test_splits = [
208
- datasets.SplitGenerator(
209
- name="test.clean",
210
- gen_kwargs={
211
- "local_extracted_archive": local_extracted_archive.get("test.clean"),
212
- "files": dl_manager.iter_archive(archive_path["test.clean"]),
213
- },
214
- ),
215
- datasets.SplitGenerator(
216
- name="test.other",
217
- gen_kwargs={
218
- "local_extracted_archive": local_extracted_archive.get("test.other"),
219
- "files": dl_manager.iter_archive(archive_path["test.other"]),
220
- },
221
- ),
222
- ]
223
-
224
- return train_splits + dev_splits + test_splits
225
 
226
- def _generate_examples(self, files, local_extracted_archive):
227
  """Generate examples from a LibriSpeech archive_path."""
228
- key = 0
229
- audio_data = {}
230
- transcripts = []
231
- for path, f in files:
232
- if path.endswith(".flac"):
233
- id_ = path.split("/")[-1][: -len(".flac")]
234
- audio_data[id_] = f.read()
235
- elif path.endswith(".trans.txt"):
236
- for line in f:
237
- if line:
238
- line = line.decode("utf-8").strip()
239
- id_, transcript = line.split(" ", 1)
240
- audio_file = f"{id_}.flac"
241
- speaker_id, chapter_id = [int(el) for el in id_.split("-")[:2]]
242
- audio_file = (
243
- os.path.join(local_extracted_archive, audio_file)
244
- if local_extracted_archive
245
- else audio_file
246
- )
247
- transcripts.append(
248
- {
249
- "id": id_,
250
- "speaker_id": speaker_id,
251
- "chapter_id": chapter_id,
252
- "file": audio_file,
253
- "text": transcript,
254
- }
255
- )
256
- if audio_data and len(audio_data) == len(transcripts):
257
- for transcript in transcripts:
258
- audio = {"path": transcript["file"], "bytes": audio_data[transcript["id"]]}
259
- yield key, {"audio": audio, **transcript}
260
- key += 1
261
- audio_data = {}
262
- transcripts = []
 
3
  """test set"""
4
 
5
 
6
+ import csv
7
  import os
8
+ import json
9
 
10
  import datasets
11
+ from datasets.utils.py_utils import size_str
12
+ from tqdm import tqdm
13
 
14
 
15
  _CITATION = """\
 
27
  Lorem ipsum
28
  """
29
 
30
+ _BASE_URL = "https://huggingface.co/datasets/j-krzywdziak/test/tree/main"
31
+ _AUDIO_URL = _BASE_URL + "dev.tar.gz"
32
+ _TRANSCRIPT_URL = _BASE_URL + "dev.tsv"
33
 
34
  # _DL_URLS = {
35
  # "clean": {
 
93
  "ngram": datasets.Value("string")
94
  }
95
  ),
96
+ supervised_keys=None,
97
+ homepage=_BASE_URL,
98
+ citation=_CITATION
 
99
  )
100
 
101
  def _split_generators(self, dl_manager):
102
+ archive_path = dl_manager.download(_AUDIO_URL)
103
  # (Optional) In non-streaming mode, we can extract the archive locally to have actual local audio files:
104
  local_extracted_archive = dl_manager.extract(archive_path) if not dl_manager.is_streaming else {}
105
+ meta_path = dl_manager.download(_TRANSCRIPT_URL)
106
+ return [datasets.SplitGenerator(
107
+ name=datasets.Split.TRAIN,
108
+ gen_kwargs={
109
+ "audios": local_extracted_archive,
110
+ "infos": meta_path
111
+ }
112
+ )]
113
+ #
114
+ # if self.config.name == "clean":
115
+ # train_splits = [
116
+ # datasets.SplitGenerator(
117
+ # name="train.100",
118
+ # gen_kwargs={
119
+ # "local_extracted_archive": local_extracted_archive.get("train.100"),
120
+ # "files": dl_manager.iter_archive(archive_path["train.100"]),
121
+ # },
122
+ # ),
123
+ # datasets.SplitGenerator(
124
+ # name="train.360",
125
+ # gen_kwargs={
126
+ # "local_extracted_archive": local_extracted_archive.get("train.360"),
127
+ # "files": dl_manager.iter_archive(archive_path["train.360"]),
128
+ # },
129
+ # ),
130
+ # ]
131
+ # dev_splits = [
132
+ # datasets.SplitGenerator(
133
+ # name=datasets.Split.VALIDATION,
134
+ # gen_kwargs={
135
+ # "local_extracted_archive": local_extracted_archive.get("dev"),
136
+ # "files": dl_manager.iter_archive(archive_path["dev"]),
137
+ # },
138
+ # )
139
+ # ]
140
+ # test_splits = [
141
+ # datasets.SplitGenerator(
142
+ # name=datasets.Split.TEST,
143
+ # gen_kwargs={
144
+ # "local_extracted_archive": local_extracted_archive.get("test"),
145
+ # "files": dl_manager.iter_archive(archive_path["test"]),
146
+ # },
147
+ # )
148
+ # ]
149
+ # elif self.config.name == "other":
150
+ # train_splits = [
151
+ # datasets.SplitGenerator(
152
+ # name="train.500",
153
+ # gen_kwargs={
154
+ # "local_extracted_archive": local_extracted_archive.get("train.500"),
155
+ # "files": dl_manager.iter_archive(archive_path["train.500"]),
156
+ # },
157
+ # )
158
+ # ]
159
+ # dev_splits = [
160
+ # datasets.SplitGenerator(
161
+ # name=datasets.Split.VALIDATION,
162
+ # gen_kwargs={
163
+ # "local_extracted_archive": local_extracted_archive.get("dev"),
164
+ # "files": dl_manager.iter_archive(archive_path["dev"]),
165
+ # },
166
+ # )
167
+ # ]
168
+ # test_splits = [
169
+ # datasets.SplitGenerator(
170
+ # name=datasets.Split.TEST,
171
+ # gen_kwargs={
172
+ # "local_extracted_archive": local_extracted_archive.get("test"),
173
+ # "files": dl_manager.iter_archive(archive_path["test"]),
174
+ # },
175
+ # )
176
+ # ]
177
+ # elif self.config.name == "all":
178
+ # train_splits = [
179
+ # datasets.SplitGenerator(
180
+ # name="train.clean.100",
181
+ # gen_kwargs={
182
+ # "local_extracted_archive": local_extracted_archive.get("train.clean.100"),
183
+ # "files": dl_manager.iter_archive(archive_path["train.clean.100"]),
184
+ # },
185
+ # ),
186
+ # datasets.SplitGenerator(
187
+ # name="train.clean.360",
188
+ # gen_kwargs={
189
+ # "local_extracted_archive": local_extracted_archive.get("train.clean.360"),
190
+ # "files": dl_manager.iter_archive(archive_path["train.clean.360"]),
191
+ # },
192
+ # ),
193
+ # datasets.SplitGenerator(
194
+ # name="train.other.500",
195
+ # gen_kwargs={
196
+ # "local_extracted_archive": local_extracted_archive.get("train.other.500"),
197
+ # "files": dl_manager.iter_archive(archive_path["train.other.500"]),
198
+ # },
199
+ # ),
200
+ # ]
201
+ # dev_splits = [
202
+ # datasets.SplitGenerator(
203
+ # name="validation.clean",
204
+ # gen_kwargs={
205
+ # "local_extracted_archive": local_extracted_archive.get("dev.clean"),
206
+ # "files": dl_manager.iter_archive(archive_path["dev.clean"]),
207
+ # },
208
+ # ),
209
+ # datasets.SplitGenerator(
210
+ # name="validation.other",
211
+ # gen_kwargs={
212
+ # "local_extracted_archive": local_extracted_archive.get("dev.other"),
213
+ # "files": dl_manager.iter_archive(archive_path["dev.other"]),
214
+ # },
215
+ # ),
216
+ # ]
217
+ # test_splits = [
218
+ # datasets.SplitGenerator(
219
+ # name="test.clean",
220
+ # gen_kwargs={
221
+ # "local_extracted_archive": local_extracted_archive.get("test.clean"),
222
+ # "files": dl_manager.iter_archive(archive_path["test.clean"]),
223
+ # },
224
+ # ),
225
+ # datasets.SplitGenerator(
226
+ # name="test.other",
227
+ # gen_kwargs={
228
+ # "local_extracted_archive": local_extracted_archive.get("test.other"),
229
+ # "files": dl_manager.iter_archive(archive_path["test.other"]),
230
+ # },
231
+ # ),
232
+ # ]
233
 
234
+ #return train_splits + dev_splits + test_splits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
235
 
236
+ def _generate_examples(self, meta_path, local_extracted_archive):
237
  """Generate examples from a LibriSpeech archive_path."""
238
+ data_fields = list(self._info().features.keys())
239
+ metadata = {}
240
+ with open(meta_path, encoding="utf-8") as f:
241
+ reader = csv.DictReader(f, delimiter="\t", quoting=csv.QUOTE_NONE)
242
+ for row in tqdm(reader, desc="Reading metadata..."):
243
+ if not row["audio_id"].endswith(".mp3"):
244
+ row["audio_id"] += ".mp3"
245
+ for field in data_fields:
246
+ if field not in row:
247
+ row[field] = ""
248
+ metadata[row["path"]] = row
249
+
250
+ for filename, file in local_extracted_archive:
251
+ _, filename = os.path.split(filename)
252
+ if filename in metadata:
253
+ result = dict(metadata[filename])
254
+ # set the audio feature and the path to the extracted file
255
+ path = os.path.join(local_extracted_archive,
256
+ filename) if local_extracted_archive else filename
257
+ result["audio"] = {"path": path, "bytes": file.read()}
258
+ # set path to None if the audio file doesn't exist locally (i.e. in streaming mode)
259
+ result["path"] = path if local_extracted_archive else filename
260
+
261
+ yield path, result