Files changed (1) hide show
  1. nvidia_canary-1b-flash.json +651 -0
nvidia_canary-1b-flash.json ADDED
@@ -0,0 +1,651 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bomFormat": "CycloneDX",
3
+ "specVersion": "1.6",
4
+ "serialNumber": "urn:uuid:63de5424-35dc-4ace-9a3e-a3b6d685f773",
5
+ "version": 1,
6
+ "metadata": {
7
+ "timestamp": "2025-06-05T09:39:42.036430+00:00",
8
+ "component": {
9
+ "type": "machine-learning-model",
10
+ "bom-ref": "nvidia/canary-1b-flash-c1c73397-2c8d-53c2-9d4d-c2069736a1eb",
11
+ "name": "nvidia/canary-1b-flash",
12
+ "externalReferences": [
13
+ {
14
+ "url": "https://huggingface.co/nvidia/canary-1b-flash",
15
+ "type": "documentation"
16
+ }
17
+ ],
18
+ "modelCard": {
19
+ "modelParameters": {
20
+ "task": "automatic-speech-recognition",
21
+ "datasets": [
22
+ {
23
+ "ref": "librispeech_asr-7baf0ed9-b50c-5f93-8c23-49a2b8749c19"
24
+ },
25
+ {
26
+ "ref": "fisher_corpus-a0c6e2c1-e876-5c66-89b2-cb93697b2a1c"
27
+ },
28
+ {
29
+ "ref": "Switchboard-1-b54b0d1d-3005-514e-9668-98d3c19f793f"
30
+ },
31
+ {
32
+ "ref": "WSJ-0-095442e6-ea65-5f6d-b360-432c7a2f501d"
33
+ },
34
+ {
35
+ "ref": "WSJ-1-0ef003e6-350d-50bb-9df7-9491b0c9b0b3"
36
+ },
37
+ {
38
+ "ref": "National-Singapore-Corpus-Part-1-1fbb2914-35aa-5126-9a84-a8b77169254c"
39
+ },
40
+ {
41
+ "ref": "National-Singapore-Corpus-Part-6-4f83cf7f-3026-5a77-ae37-28a73d4abc24"
42
+ },
43
+ {
44
+ "ref": "vctk-d80444bd-bcc6-5c25-8570-061bb96dae38"
45
+ },
46
+ {
47
+ "ref": "voxpopuli-15fb6343-a710-54f9-842b-3a1b43d6a630"
48
+ },
49
+ {
50
+ "ref": "europarl-7e07ffed-425e-5e05-8847-08a1899f0ac1"
51
+ },
52
+ {
53
+ "ref": "multilingual_librispeech-f260ef31-1d5d-54fe-8e61-88c397c0b7ce"
54
+ },
55
+ {
56
+ "ref": "mozilla-foundation/common_voice_8_0-a994a71f-f9f5-5f65-a3fa-51a56293cd8e"
57
+ },
58
+ {
59
+ "ref": "MLCommons/peoples_speech-f88dc766-1de0-51c6-865d-16930ec19be6"
60
+ }
61
+ ]
62
+ },
63
+ "properties": [
64
+ {
65
+ "name": "library_name",
66
+ "value": "nemo"
67
+ }
68
+ ],
69
+ "quantitativeAnalysis": {
70
+ "performanceMetrics": [
71
+ {
72
+ "slice": "dataset: librispeech_asr, split: test, config: other",
73
+ "type": "wer",
74
+ "value": 2.87
75
+ },
76
+ {
77
+ "slice": "dataset: kensho/spgispeech, split: test, config: test",
78
+ "type": "wer",
79
+ "value": 1.95
80
+ },
81
+ {
82
+ "slice": "dataset: mozilla-foundation/common_voice_16_1, split: test, config: en",
83
+ "type": "wer",
84
+ "value": 6.99
85
+ },
86
+ {
87
+ "slice": "dataset: mozilla-foundation/common_voice_16_1, split: test, config: de",
88
+ "type": "wer",
89
+ "value": 4.09
90
+ },
91
+ {
92
+ "slice": "dataset: mozilla-foundation/common_voice_16_1, split: test, config: es",
93
+ "type": "wer",
94
+ "value": 3.62
95
+ },
96
+ {
97
+ "slice": "dataset: mozilla-foundation/common_voice_16_1, split: test, config: fr",
98
+ "type": "wer",
99
+ "value": 6.15
100
+ },
101
+ {
102
+ "slice": "dataset: google/fleurs, split: test, config: en_us",
103
+ "type": "bleu",
104
+ "value": 32.27
105
+ },
106
+ {
107
+ "slice": "dataset: google/fleurs, split: test, config: en_us",
108
+ "type": "bleu",
109
+ "value": 22.6
110
+ },
111
+ {
112
+ "slice": "dataset: google/fleurs, split: test, config: en_us",
113
+ "type": "bleu",
114
+ "value": 41.22
115
+ },
116
+ {
117
+ "slice": "dataset: google/fleurs, split: test, config: de_de",
118
+ "type": "bleu",
119
+ "value": 35.5
120
+ },
121
+ {
122
+ "slice": "dataset: google/fleurs, split: test, config: es_419",
123
+ "type": "bleu",
124
+ "value": 23.32
125
+ },
126
+ {
127
+ "slice": "dataset: google/fleurs, split: test, config: fr_fr",
128
+ "type": "bleu",
129
+ "value": 33.42
130
+ },
131
+ {
132
+ "slice": "dataset: covost2, split: test, config: de_de",
133
+ "type": "bleu",
134
+ "value": 39.33
135
+ },
136
+ {
137
+ "slice": "dataset: covost2, split: test, config: es_419",
138
+ "type": "bleu",
139
+ "value": 41.86
140
+ },
141
+ {
142
+ "slice": "dataset: covost2, split: test, config: fr_fr",
143
+ "type": "bleu",
144
+ "value": 41.43
145
+ }
146
+ ]
147
+ }
148
+ },
149
+ "authors": [
150
+ {
151
+ "name": "nvidia"
152
+ }
153
+ ],
154
+ "licenses": [
155
+ {
156
+ "license": {
157
+ "id": "CC-BY-4.0",
158
+ "url": "https://spdx.org/licenses/CC-BY-4.0.html"
159
+ }
160
+ }
161
+ ],
162
+ "tags": [
163
+ "nemo",
164
+ "automatic-speech-recognition",
165
+ "automatic-speech-translation",
166
+ "speech",
167
+ "audio",
168
+ "Transformer",
169
+ "FastConformer",
170
+ "Conformer",
171
+ "pytorch",
172
+ "NeMo",
173
+ "hf-asr-leaderboard",
174
+ "en",
175
+ "de",
176
+ "es",
177
+ "fr",
178
+ "dataset:librispeech_asr",
179
+ "dataset:fisher_corpus",
180
+ "dataset:Switchboard-1",
181
+ "dataset:WSJ-0",
182
+ "dataset:WSJ-1",
183
+ "dataset:National-Singapore-Corpus-Part-1",
184
+ "dataset:National-Singapore-Corpus-Part-6",
185
+ "dataset:vctk",
186
+ "dataset:voxpopuli",
187
+ "dataset:europarl",
188
+ "dataset:multilingual_librispeech",
189
+ "dataset:mozilla-foundation/common_voice_8_0",
190
+ "dataset:MLCommons/peoples_speech",
191
+ "arxiv:2104.02821",
192
+ "arxiv:2503.05931",
193
+ "arxiv:1706.03762",
194
+ "arxiv:2409.13523",
195
+ "license:cc-by-4.0",
196
+ "model-index",
197
+ "region:us"
198
+ ]
199
+ }
200
+ },
201
+ "components": [
202
+ {
203
+ "type": "data",
204
+ "bom-ref": "librispeech_asr-7baf0ed9-b50c-5f93-8c23-49a2b8749c19",
205
+ "name": "librispeech_asr",
206
+ "data": [
207
+ {
208
+ "type": "dataset",
209
+ "bom-ref": "librispeech_asr-7baf0ed9-b50c-5f93-8c23-49a2b8749c19",
210
+ "name": "librispeech_asr",
211
+ "contents": {
212
+ "url": "https://huggingface.co/datasets/librispeech_asr",
213
+ "properties": [
214
+ {
215
+ "name": "task_categories",
216
+ "value": "automatic-speech-recognition, audio-classification"
217
+ },
218
+ {
219
+ "name": "task_ids",
220
+ "value": "speaker-identification"
221
+ },
222
+ {
223
+ "name": "language",
224
+ "value": "en"
225
+ },
226
+ {
227
+ "name": "size_categories",
228
+ "value": "100K<n<1M"
229
+ },
230
+ {
231
+ "name": "annotations_creators",
232
+ "value": "expert-generated"
233
+ },
234
+ {
235
+ "name": "language_creators",
236
+ "value": "crowdsourced, expert-generated"
237
+ },
238
+ {
239
+ "name": "pretty_name",
240
+ "value": "LibriSpeech"
241
+ },
242
+ {
243
+ "name": "source_datasets",
244
+ "value": "original"
245
+ },
246
+ {
247
+ "name": "paperswithcode_id",
248
+ "value": "librispeech-1"
249
+ },
250
+ {
251
+ "name": "license",
252
+ "value": "cc-by-4.0"
253
+ }
254
+ ]
255
+ },
256
+ "governance": {
257
+ "owners": [
258
+ {
259
+ "organization": {
260
+ "name": "openslr",
261
+ "url": "https://huggingface.co/openslr"
262
+ }
263
+ }
264
+ ]
265
+ },
266
+ "description": "LibriSpeech is a corpus of approximately 1000 hours of read English speech with sampling rate of 16 kHz,\nprepared by Vassil Panayotov with the assistance of Daniel Povey. The data is derived from read\naudiobooks from the LibriVox project, and has been carefully segmented and aligned.87"
267
+ }
268
+ ]
269
+ },
270
+ {
271
+ "type": "data",
272
+ "bom-ref": "fisher_corpus-a0c6e2c1-e876-5c66-89b2-cb93697b2a1c",
273
+ "name": "fisher_corpus",
274
+ "data": [
275
+ {
276
+ "type": "dataset",
277
+ "bom-ref": "fisher_corpus-a0c6e2c1-e876-5c66-89b2-cb93697b2a1c",
278
+ "name": "fisher_corpus"
279
+ }
280
+ ]
281
+ },
282
+ {
283
+ "type": "data",
284
+ "bom-ref": "Switchboard-1-b54b0d1d-3005-514e-9668-98d3c19f793f",
285
+ "name": "Switchboard-1",
286
+ "data": [
287
+ {
288
+ "type": "dataset",
289
+ "bom-ref": "Switchboard-1-b54b0d1d-3005-514e-9668-98d3c19f793f",
290
+ "name": "Switchboard-1"
291
+ }
292
+ ]
293
+ },
294
+ {
295
+ "type": "data",
296
+ "bom-ref": "WSJ-0-095442e6-ea65-5f6d-b360-432c7a2f501d",
297
+ "name": "WSJ-0",
298
+ "data": [
299
+ {
300
+ "type": "dataset",
301
+ "bom-ref": "WSJ-0-095442e6-ea65-5f6d-b360-432c7a2f501d",
302
+ "name": "WSJ-0"
303
+ }
304
+ ]
305
+ },
306
+ {
307
+ "type": "data",
308
+ "bom-ref": "WSJ-1-0ef003e6-350d-50bb-9df7-9491b0c9b0b3",
309
+ "name": "WSJ-1",
310
+ "data": [
311
+ {
312
+ "type": "dataset",
313
+ "bom-ref": "WSJ-1-0ef003e6-350d-50bb-9df7-9491b0c9b0b3",
314
+ "name": "WSJ-1"
315
+ }
316
+ ]
317
+ },
318
+ {
319
+ "type": "data",
320
+ "bom-ref": "National-Singapore-Corpus-Part-1-1fbb2914-35aa-5126-9a84-a8b77169254c",
321
+ "name": "National-Singapore-Corpus-Part-1",
322
+ "data": [
323
+ {
324
+ "type": "dataset",
325
+ "bom-ref": "National-Singapore-Corpus-Part-1-1fbb2914-35aa-5126-9a84-a8b77169254c",
326
+ "name": "National-Singapore-Corpus-Part-1"
327
+ }
328
+ ]
329
+ },
330
+ {
331
+ "type": "data",
332
+ "bom-ref": "National-Singapore-Corpus-Part-6-4f83cf7f-3026-5a77-ae37-28a73d4abc24",
333
+ "name": "National-Singapore-Corpus-Part-6",
334
+ "data": [
335
+ {
336
+ "type": "dataset",
337
+ "bom-ref": "National-Singapore-Corpus-Part-6-4f83cf7f-3026-5a77-ae37-28a73d4abc24",
338
+ "name": "National-Singapore-Corpus-Part-6"
339
+ }
340
+ ]
341
+ },
342
+ {
343
+ "type": "data",
344
+ "bom-ref": "vctk-d80444bd-bcc6-5c25-8570-061bb96dae38",
345
+ "name": "vctk",
346
+ "data": [
347
+ {
348
+ "type": "dataset",
349
+ "bom-ref": "vctk-d80444bd-bcc6-5c25-8570-061bb96dae38",
350
+ "name": "vctk",
351
+ "contents": {
352
+ "url": "https://huggingface.co/datasets/vctk",
353
+ "properties": [
354
+ {
355
+ "name": "task_categories",
356
+ "value": "automatic-speech-recognition, text-to-speech, text-to-audio"
357
+ },
358
+ {
359
+ "name": "task_ids",
360
+ "value": ""
361
+ },
362
+ {
363
+ "name": "language",
364
+ "value": "en"
365
+ },
366
+ {
367
+ "name": "size_categories",
368
+ "value": "10K<n<100K"
369
+ },
370
+ {
371
+ "name": "annotations_creators",
372
+ "value": "expert-generated"
373
+ },
374
+ {
375
+ "name": "language_creators",
376
+ "value": "crowdsourced"
377
+ },
378
+ {
379
+ "name": "pretty_name",
380
+ "value": "VCTK"
381
+ },
382
+ {
383
+ "name": "source_datasets",
384
+ "value": "original"
385
+ },
386
+ {
387
+ "name": "paperswithcode_id",
388
+ "value": "vctk"
389
+ },
390
+ {
391
+ "name": "license",
392
+ "value": "cc-by-4.0"
393
+ }
394
+ ]
395
+ },
396
+ "governance": {
397
+ "owners": [
398
+ {
399
+ "organization": {
400
+ "name": "CSTR-Edinburgh",
401
+ "url": "https://huggingface.co/CSTR-Edinburgh"
402
+ }
403
+ }
404
+ ]
405
+ },
406
+ "description": "The CSTR VCTK Corpus includes speech data uttered by 110 English speakers with various accents."
407
+ }
408
+ ]
409
+ },
410
+ {
411
+ "type": "data",
412
+ "bom-ref": "voxpopuli-15fb6343-a710-54f9-842b-3a1b43d6a630",
413
+ "name": "voxpopuli",
414
+ "data": [
415
+ {
416
+ "type": "dataset",
417
+ "bom-ref": "voxpopuli-15fb6343-a710-54f9-842b-3a1b43d6a630",
418
+ "name": "voxpopuli"
419
+ }
420
+ ]
421
+ },
422
+ {
423
+ "type": "data",
424
+ "bom-ref": "europarl-7e07ffed-425e-5e05-8847-08a1899f0ac1",
425
+ "name": "europarl",
426
+ "data": [
427
+ {
428
+ "type": "dataset",
429
+ "bom-ref": "europarl-7e07ffed-425e-5e05-8847-08a1899f0ac1",
430
+ "name": "europarl"
431
+ }
432
+ ]
433
+ },
434
+ {
435
+ "type": "data",
436
+ "bom-ref": "multilingual_librispeech-f260ef31-1d5d-54fe-8e61-88c397c0b7ce",
437
+ "name": "multilingual_librispeech",
438
+ "data": [
439
+ {
440
+ "type": "dataset",
441
+ "bom-ref": "multilingual_librispeech-f260ef31-1d5d-54fe-8e61-88c397c0b7ce",
442
+ "name": "multilingual_librispeech",
443
+ "contents": {
444
+ "url": "https://huggingface.co/datasets/multilingual_librispeech",
445
+ "properties": [
446
+ {
447
+ "name": "task_categories",
448
+ "value": "automatic-speech-recognition, audio-classification"
449
+ },
450
+ {
451
+ "name": "task_ids",
452
+ "value": "speaker-identification"
453
+ },
454
+ {
455
+ "name": "language",
456
+ "value": "de, es, fr, it, nl, pl, pt"
457
+ },
458
+ {
459
+ "name": "size_categories",
460
+ "value": "100K<n<1M"
461
+ },
462
+ {
463
+ "name": "annotations_creators",
464
+ "value": "expert-generated"
465
+ },
466
+ {
467
+ "name": "language_creators",
468
+ "value": "crowdsourced, expert-generated"
469
+ },
470
+ {
471
+ "name": "pretty_name",
472
+ "value": "MultiLingual LibriSpeech"
473
+ },
474
+ {
475
+ "name": "source_datasets",
476
+ "value": "original"
477
+ },
478
+ {
479
+ "name": "paperswithcode_id",
480
+ "value": "librispeech-1"
481
+ },
482
+ {
483
+ "name": "license",
484
+ "value": "cc-by-4.0"
485
+ }
486
+ ]
487
+ },
488
+ "governance": {
489
+ "owners": [
490
+ {
491
+ "organization": {
492
+ "name": "legacy-datasets",
493
+ "url": "https://huggingface.co/legacy-datasets"
494
+ }
495
+ }
496
+ ]
497
+ },
498
+ "description": "Multilingual LibriSpeech (MLS) dataset is a large multilingual corpus suitable for speech research. The dataset is derived from read audiobooks from LibriVox and consists of 8 languages - English, German, Dutch, Spanish, French, Italian, Portuguese, Polish."
499
+ }
500
+ ]
501
+ },
502
+ {
503
+ "type": "data",
504
+ "bom-ref": "mozilla-foundation/common_voice_8_0-a994a71f-f9f5-5f65-a3fa-51a56293cd8e",
505
+ "name": "mozilla-foundation/common_voice_8_0",
506
+ "data": [
507
+ {
508
+ "type": "dataset",
509
+ "bom-ref": "mozilla-foundation/common_voice_8_0-a994a71f-f9f5-5f65-a3fa-51a56293cd8e",
510
+ "name": "mozilla-foundation/common_voice_8_0",
511
+ "contents": {
512
+ "url": "https://huggingface.co/datasets/mozilla-foundation/common_voice_8_0",
513
+ "properties": [
514
+ {
515
+ "name": "task_categories",
516
+ "value": "automatic-speech-recognition"
517
+ },
518
+ {
519
+ "name": "annotations_creators",
520
+ "value": "crowdsourced"
521
+ },
522
+ {
523
+ "name": "language_creators",
524
+ "value": "crowdsourced"
525
+ },
526
+ {
527
+ "name": "pretty_name",
528
+ "value": "Common Voice Corpus 8.0"
529
+ },
530
+ {
531
+ "name": "source_datasets",
532
+ "value": "extended|common_voice"
533
+ },
534
+ {
535
+ "name": "paperswithcode_id",
536
+ "value": "common-voice"
537
+ },
538
+ {
539
+ "name": "license",
540
+ "value": "cc0-1.0"
541
+ }
542
+ ]
543
+ },
544
+ "governance": {
545
+ "owners": [
546
+ {
547
+ "organization": {
548
+ "name": "mozilla-foundation",
549
+ "url": "https://huggingface.co/mozilla-foundation"
550
+ }
551
+ }
552
+ ]
553
+ },
554
+ "description": "\n\t\n\t\t\n\t\tDataset Card for Common Voice Corpus 8.0\n\t\n\n\n\t\n\t\t\n\t\tDataset Summary\n\t\n\nThe Common Voice dataset consists of a unique MP3 and corresponding text file. \nMany of the 18243 recorded hours in the dataset also include demographic metadata like age, sex, and accent \nthat can help improve the accuracy of speech recognition engines.\nThe dataset currently consists of 14122 validated hours in 87 languages, but more voices and languages are always added. \nTake a look at the Languages page to\u2026 See the full description on the dataset page: https://huggingface.co/datasets/mozilla-foundation/common_voice_8_0."
555
+ }
556
+ ]
557
+ },
558
+ {
559
+ "type": "data",
560
+ "bom-ref": "MLCommons/peoples_speech-f88dc766-1de0-51c6-865d-16930ec19be6",
561
+ "name": "MLCommons/peoples_speech",
562
+ "data": [
563
+ {
564
+ "type": "dataset",
565
+ "bom-ref": "MLCommons/peoples_speech-f88dc766-1de0-51c6-865d-16930ec19be6",
566
+ "name": "MLCommons/peoples_speech",
567
+ "contents": {
568
+ "url": "https://huggingface.co/datasets/MLCommons/peoples_speech",
569
+ "properties": [
570
+ {
571
+ "name": "task_categories",
572
+ "value": "automatic-speech-recognition"
573
+ },
574
+ {
575
+ "name": "task_ids",
576
+ "value": ""
577
+ },
578
+ {
579
+ "name": "language",
580
+ "value": "en"
581
+ },
582
+ {
583
+ "name": "size_categories",
584
+ "value": "1T<n"
585
+ },
586
+ {
587
+ "name": "annotations_creators",
588
+ "value": "crowdsourced, machine-generated"
589
+ },
590
+ {
591
+ "name": "language_creators",
592
+ "value": "crowdsourced, machine-generated"
593
+ },
594
+ {
595
+ "name": "pretty_name",
596
+ "value": "People's Speech"
597
+ },
598
+ {
599
+ "name": "source_datasets",
600
+ "value": "original"
601
+ },
602
+ {
603
+ "name": "configs",
604
+ "value": "Name of the dataset subset: clean {\"split\": \"train\", \"path\": \"clean/train-*\"}, {\"split\": \"validation\", \"path\": \"clean/validation-*\"}, {\"split\": \"test\", \"path\": \"clean/test-*\"}"
605
+ },
606
+ {
607
+ "name": "configs",
608
+ "value": "Name of the dataset subset: clean_sa {\"split\": \"train\", \"path\": \"clean_sa/train-*\"}, {\"split\": \"validation\", \"path\": \"clean_sa/validation-*\"}, {\"split\": \"test\", \"path\": \"clean_sa/test-*\"}"
609
+ },
610
+ {
611
+ "name": "configs",
612
+ "value": "Name of the dataset subset: dirty {\"split\": \"train\", \"path\": \"dirty/train-*\"}, {\"split\": \"validation\", \"path\": \"dirty/validation-*\"}, {\"split\": \"test\", \"path\": \"dirty/test-*\"}"
613
+ },
614
+ {
615
+ "name": "configs",
616
+ "value": "Name of the dataset subset: dirty_sa {\"split\": \"train\", \"path\": \"dirty_sa/train-*\"}, {\"split\": \"validation\", \"path\": \"dirty_sa/validation-*\"}, {\"split\": \"test\", \"path\": \"dirty_sa/test-*\"}"
617
+ },
618
+ {
619
+ "name": "configs",
620
+ "value": "Name of the dataset subset: microset {\"split\": \"train\", \"path\": \"microset/train-*\"}"
621
+ },
622
+ {
623
+ "name": "configs",
624
+ "value": "Name of the dataset subset: test {\"split\": \"test\", \"path\": \"test/test-*\"}"
625
+ },
626
+ {
627
+ "name": "configs",
628
+ "value": "Name of the dataset subset: validation {\"split\": \"validation\", \"path\": \"validation/validation-*\"}"
629
+ },
630
+ {
631
+ "name": "license",
632
+ "value": "cc-by-2.0, cc-by-2.5, cc-by-3.0, cc-by-4.0, cc-by-sa-3.0, cc-by-sa-4.0"
633
+ }
634
+ ]
635
+ },
636
+ "governance": {
637
+ "owners": [
638
+ {
639
+ "organization": {
640
+ "name": "MLCommons",
641
+ "url": "https://huggingface.co/MLCommons"
642
+ }
643
+ }
644
+ ]
645
+ },
646
+ "description": "\n\t\n\t\t\n\t\tDataset Card for People's Speech\n\t\n\n\n\t\n\t\t\n\t\tDataset Summary\n\t\n\nThe People's Speech Dataset is among the world's largest English speech recognition corpus today that is licensed for academic and commercial usage under CC-BY-SA and CC-BY 4.0. It includes 30,000+ hours of transcribed speech in English languages with a diverse set of speakers. This open dataset is large enough to train speech-to-text systems and crucially is available with a permissive license.\n\n\t\n\t\t\n\t\n\t\n\t\tSupported Tasks\u2026 See the full description on the dataset page: https://huggingface.co/datasets/MLCommons/peoples_speech."
647
+ }
648
+ ]
649
+ }
650
+ ]
651
+ }