Delete automatic_speech_recognition_dummy.py

#3
by lhoestq HF Staff - opened
README.md ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ ---
2
+ configs:
3
+ - config_name: default
4
+ data_files:
5
+ - split: test
6
+ path: "*.flac"
7
+ ---
asr_dummy.py DELETED
@@ -1,173 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2021 The TensorFlow Datasets Authors and the HuggingFace Datasets Authors.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
-
16
- # Lint as: python3
17
- """SUPERB: Speech processing Universal PERformance Benchmark."""
18
-
19
-
20
- import glob
21
- import os
22
- import textwrap
23
-
24
- import datasets
25
-
26
-
27
- _CITATION = """\
28
- @article{DBLP:journals/corr/abs-2105-01051,
29
- author = {Shu{-}Wen Yang and
30
- Po{-}Han Chi and
31
- Yung{-}Sung Chuang and
32
- Cheng{-}I Jeff Lai and
33
- Kushal Lakhotia and
34
- Yist Y. Lin and
35
- Andy T. Liu and
36
- Jiatong Shi and
37
- Xuankai Chang and
38
- Guan{-}Ting Lin and
39
- Tzu{-}Hsien Huang and
40
- Wei{-}Cheng Tseng and
41
- Ko{-}tik Lee and
42
- Da{-}Rong Liu and
43
- Zili Huang and
44
- Shuyan Dong and
45
- Shang{-}Wen Li and
46
- Shinji Watanabe and
47
- Abdelrahman Mohamed and
48
- Hung{-}yi Lee},
49
- title = {{SUPERB:} Speech processing Universal PERformance Benchmark},
50
- journal = {CoRR},
51
- volume = {abs/2105.01051},
52
- year = {2021},
53
- url = {https://arxiv.org/abs/2105.01051},
54
- archivePrefix = {arXiv},
55
- eprint = {2105.01051},
56
- timestamp = {Thu, 01 Jul 2021 13:30:22 +0200},
57
- biburl = {https://dblp.org/rec/journals/corr/abs-2105-01051.bib},
58
- bibsource = {dblp computer science bibliography, https://dblp.org}
59
- }
60
- """
61
-
62
- _DESCRIPTION = """\
63
- Self-supervised learning (SSL) has proven vital for advancing research in
64
- natural language processing (NLP) and computer vision (CV). The paradigm
65
- pretrains a shared model on large volumes of unlabeled data and achieves
66
- state-of-the-art (SOTA) for various tasks with minimal adaptation. However, the
67
- speech processing community lacks a similar setup to systematically explore the
68
- paradigm. To bridge this gap, we introduce Speech processing Universal
69
- PERformance Benchmark (SUPERB). SUPERB is a leaderboard to benchmark the
70
- performance of a shared model across a wide range of speech processing tasks
71
- with minimal architecture changes and labeled data. Among multiple usages of the
72
- shared model, we especially focus on extracting the representation learned from
73
- SSL due to its preferable re-usability. We present a simple framework to solve
74
- SUPERB tasks by learning task-specialized lightweight prediction heads on top of
75
- the frozen shared model. Our results demonstrate that the framework is promising
76
- as SSL representations show competitive generalizability and accessibility
77
- across SUPERB tasks. We release SUPERB as a challenge with a leaderboard and a
78
- benchmark toolkit to fuel the research in representation learning and general
79
- speech processing.
80
-
81
- Note that in order to limit the required storage for preparing this dataset, the
82
- audio is stored in the .flac format and is not converted to a float32 array. To
83
- convert, the audio file to a float32 array, please make use of the `.map()`
84
- function as follows:
85
-
86
-
87
- ```python
88
- import soundfile as sf
89
-
90
- def map_to_array(batch):
91
- speech_array, _ = sf.read(batch["file"])
92
- batch["speech"] = speech_array
93
- return batch
94
-
95
- dataset = dataset.map(map_to_array, remove_columns=["file"])
96
- ```
97
- """
98
-
99
-
100
- class AsrDummybConfig(datasets.BuilderConfig):
101
- """BuilderConfig for Superb."""
102
-
103
- def __init__(
104
- self,
105
- data_url,
106
- url,
107
- **kwargs,
108
- ):
109
- super(AsrDummybConfig, self).__init__(
110
- version=datasets.Version("1.9.0", ""), **kwargs
111
- )
112
- self.data_url = data_url
113
- self.url = url
114
-
115
-
116
- class AsrDummy(datasets.GeneratorBasedBuilder):
117
- """Superb dataset."""
118
-
119
- BUILDER_CONFIGS = [
120
- AsrDummybConfig(
121
- name="asr",
122
- description=textwrap.dedent(
123
- """\
124
- ASR transcribes utterances into words. While PR analyzes the
125
- improvement in modeling phonetics, ASR reflects the significance of
126
- the improvement in a real-world scenario. LibriSpeech
127
- train-clean-100/dev-clean/test-clean subsets are used for
128
- training/validation/testing. The evaluation metric is word error
129
- rate (WER)."""
130
- ),
131
- url="http://www.openslr.org/12",
132
- data_url="http://www.openslr.org/resources/12/",
133
- )
134
- ]
135
-
136
- DEFAULT_CONFIG_NAME = "asr"
137
-
138
- def _info(self):
139
- return datasets.DatasetInfo(
140
- description=_DESCRIPTION,
141
- features=datasets.Features(
142
- {
143
- "id": datasets.Value("string"),
144
- "file": datasets.Value("string"),
145
- }
146
- ),
147
- supervised_keys=("file",),
148
- homepage=self.config.url,
149
- citation=_CITATION,
150
- )
151
-
152
- def _split_generators(self, dl_manager):
153
- DL_URLS = [
154
- f"https://huggingface.co/datasets/Narsil/asr_dummy/raw/main/{i}.flac"
155
- for i in range(1, 5)
156
- ]
157
- archive_path = dl_manager.download_and_extract(DL_URLS)
158
- return [
159
- datasets.SplitGenerator(
160
- name=datasets.Split.TEST,
161
- gen_kwargs={"archive_path": archive_path},
162
- ),
163
- ]
164
-
165
- def _generate_examples(self, archive_path):
166
- """Generate examples."""
167
- for i, filename in enumerate(archive_path):
168
- key = str(i)
169
- example = {
170
- "id": key,
171
- "file": filename,
172
- }
173
- yield key, example
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
asr_dummy.py.lock DELETED
File without changes
automatic_speech_recognition_dummy.py DELETED
@@ -1,167 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2021 The TensorFlow Datasets Authors and the HuggingFace Datasets Authors.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
- # Lint as: python3
16
- """SUPERB: Speech processing Universal PERformance Benchmark."""
17
- import glob
18
- import os
19
- import textwrap
20
- import datasets
21
- from datasets.tasks import AutomaticSpeechRecognition
22
-
23
- _CITATION = """\
24
- @article{DBLP:journals/corr/abs-2105-01051,
25
- author = {Shu{-}Wen Yang and
26
- Po{-}Han Chi and
27
- Yung{-}Sung Chuang and
28
- Cheng{-}I Jeff Lai and
29
- Kushal Lakhotia and
30
- Yist Y. Lin and
31
- Andy T. Liu and
32
- Jiatong Shi and
33
- Xuankai Chang and
34
- Guan{-}Ting Lin and
35
- Tzu{-}Hsien Huang and
36
- Wei{-}Cheng Tseng and
37
- Ko{-}tik Lee and
38
- Da{-}Rong Liu and
39
- Zili Huang and
40
- Shuyan Dong and
41
- Shang{-}Wen Li and
42
- Shinji Watanabe and
43
- Abdelrahman Mohamed and
44
- Hung{-}yi Lee},
45
- title = {{SUPERB:} Speech processing Universal PERformance Benchmark},
46
- journal = {CoRR},
47
- volume = {abs/2105.01051},
48
- year = {2021},
49
- url = {https://arxiv.org/abs/2105.01051},
50
- archivePrefix = {arXiv},
51
- eprint = {2105.01051},
52
- timestamp = {Thu, 01 Jul 2021 13:30:22 +0200},
53
- biburl = {https://dblp.org/rec/journals/corr/abs-2105-01051.bib},
54
- bibsource = {dblp computer science bibliography, https://dblp.org}
55
- }
56
- """
57
-
58
- _DESCRIPTION = """\
59
- Self-supervised learning (SSL) has proven vital for advancing research in
60
- natural language processing (NLP) and computer vision (CV). The paradigm
61
- pretrains a shared model on large volumes of unlabeled data and achieves
62
- state-of-the-art (SOTA) for various tasks with minimal adaptation. However, the
63
- speech processing community lacks a similar setup to systematically explore the
64
- paradigm. To bridge this gap, we introduce Speech processing Universal
65
- PERformance Benchmark (SUPERB). SUPERB is a leaderboard to benchmark the
66
- performance of a shared model across a wide range of speech processing tasks
67
- with minimal architecture changes and labeled data. Among multiple usages of the
68
- shared model, we especially focus on extracting the representation learned from
69
- SSL due to its preferable re-usability. We present a simple framework to solve
70
- SUPERB tasks by learning task-specialized lightweight prediction heads on top of
71
- the frozen shared model. Our results demonstrate that the framework is promising
72
- as SSL representations show competitive generalizability and accessibility
73
- across SUPERB tasks. We release SUPERB as a challenge with a leaderboard and a
74
- benchmark toolkit to fuel the research in representation learning and general
75
- speech processing.
76
- Note that in order to limit the required storage for preparing this dataset, the
77
- audio is stored in the .flac format and is not converted to a float32 array. To
78
- convert, the audio file to a float32 array, please make use of the `.map()`
79
- function as follows:
80
- ```python
81
- import soundfile as sf
82
- def map_to_array(batch):
83
- speech_array, _ = sf.read(batch["file"])
84
- batch["speech"] = speech_array
85
- return batch
86
- dataset = dataset.map(map_to_array, remove_columns=["file"])
87
- ```
88
- """
89
-
90
- class AsrDummybConfig(datasets.BuilderConfig):
91
- """BuilderConfig for Superb."""
92
- def __init__(
93
- self,
94
- data_url,
95
- url,
96
- task_templates=None,
97
- **kwargs,
98
- ):
99
- super(AsrDummybConfig, self).__init__(
100
- version=datasets.Version("1.9.0", ""), **kwargs
101
- )
102
- self.data_url = data_url
103
- self.url = url
104
- self.task_templates = task_templates
105
-
106
- class AsrDummy(datasets.GeneratorBasedBuilder):
107
- """Superb dataset."""
108
- BUILDER_CONFIGS = [
109
- AsrDummybConfig(
110
- name="asr",
111
- description=textwrap.dedent(
112
- """\
113
- ASR transcribes utterances into words. While PR analyzes the
114
- improvement in modeling phonetics, ASR reflects the significance of
115
- the improvement in a real-world scenario. LibriSpeech
116
- train-clean-100/dev-clean/test-clean subsets are used for
117
- training/validation/testing. The evaluation metric is word error
118
- rate (WER)."""
119
- ),
120
- url="http://www.openslr.org/12",
121
- data_url="http://www.openslr.org/resources/12/",
122
- task_templates=[
123
- AutomaticSpeechRecognition(
124
- audio_file_path_column="file", transcription_column="text"
125
- )
126
- ],
127
- )
128
- ]
129
-
130
- DEFAULT_CONFIG_NAME = "asr"
131
- def _info(self):
132
- return datasets.DatasetInfo(
133
- description=_DESCRIPTION,
134
- features=datasets.Features(
135
- {
136
- "id": datasets.Value("string"),
137
- "file": datasets.Value("string"),
138
- }
139
- ),
140
- supervised_keys=("file",),
141
- homepage=self.config.url,
142
- citation=_CITATION,
143
- task_templates=self.config.task_templates,
144
- )
145
-
146
- def _split_generators(self, dl_manager):
147
- DL_URLS = [
148
- f"https://huggingface.co/datasets/Narsil/automatic_speech_recognition_dummy/raw/main/{i}.flac"
149
- for i in range(1, 4)
150
- ]
151
- archive_path = dl_manager.download_and_extract(DL_URLS)
152
- return [
153
- datasets.SplitGenerator(
154
- name=datasets.Split.TEST,
155
- gen_kwargs={"archive_path": archive_path},
156
- ),
157
- ]
158
-
159
- def _generate_examples(self, archive_path):
160
- """Generate examples."""
161
- for i, filename in enumerate(archive_path):
162
- key = str(i)
163
- example = {
164
- "id": key,
165
- "file": filename,
166
- }
167
- yield key, example