DelosLogic commited on
Commit
6c55baa
·
verified ·
1 Parent(s): 77620bb

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ tags:
5
+ - whisper
6
+ - speech-recognition
7
+ - trinidadian-creole
8
+ license: mit
9
+ ---
10
+
11
+ # accento-v1.0
12
+
13
+ Accento ASR model for Trinidadian Creole English.
14
+
15
+ ## Usage
16
+
17
+ ```python
18
+ from accento import AccentoTranscriber
19
+
20
+ transcriber = AccentoTranscriber(model_path="models/accento-v1.0")
21
+ result = transcriber.transcribe("audio.wav")
22
+ print(result.text)
23
+ ```
adapter_config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "WhisperForConditionalGeneration",
7
+ "parent_library": "transformers.models.whisper.modeling_whisper"
8
+ },
9
+ "base_model_name_or_path": "openai/whisper-large-v3-turbo",
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": true,
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.2,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.0",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "fc2",
36
+ "k_proj",
37
+ "fc1",
38
+ "out_proj",
39
+ "v_proj",
40
+ "q_proj"
41
+ ],
42
+ "target_parameters": null,
43
+ "task_type": null,
44
+ "trainable_token_indices": null,
45
+ "use_dora": false,
46
+ "use_qalora": false,
47
+ "use_rslora": false
48
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c832869af91a208aa8d149c2b8c57acf971b99624e0154fa4eb09395101e77c3
3
+ size 111475728
preprocessor_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "chunk_length": 30,
3
+ "dither": 0.0,
4
+ "feature_extractor_type": "WhisperFeatureExtractor",
5
+ "feature_size": 128,
6
+ "hop_length": 160,
7
+ "n_fft": 400,
8
+ "n_samples": 480000,
9
+ "nb_max_frames": 3000,
10
+ "padding_side": "right",
11
+ "padding_value": 0.0,
12
+ "processor_class": "WhisperProcessor",
13
+ "return_attention_mask": false,
14
+ "sampling_rate": 16000
15
+ }
trainer_state.json ADDED
@@ -0,0 +1,313 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 300,
3
+ "best_metric": 22.202486678507995,
4
+ "best_model_checkpoint": "models/iterative-large/iteration2\\checkpoint-300",
5
+ "epoch": 15.0,
6
+ "eval_steps": 50,
7
+ "global_step": 300,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.5128205128205128,
14
+ "grad_norm": 2.996086597442627,
15
+ "learning_rate": 4.5000000000000003e-07,
16
+ "loss": 2.3646,
17
+ "step": 10
18
+ },
19
+ {
20
+ "epoch": 1.0,
21
+ "grad_norm": 3.3487439155578613,
22
+ "learning_rate": 9.500000000000001e-07,
23
+ "loss": 2.4107,
24
+ "step": 20
25
+ },
26
+ {
27
+ "epoch": 1.5128205128205128,
28
+ "grad_norm": 3.2800023555755615,
29
+ "learning_rate": 1.45e-06,
30
+ "loss": 2.521,
31
+ "step": 30
32
+ },
33
+ {
34
+ "epoch": 2.0,
35
+ "grad_norm": 3.7190394401550293,
36
+ "learning_rate": 1.9500000000000004e-06,
37
+ "loss": 2.3518,
38
+ "step": 40
39
+ },
40
+ {
41
+ "epoch": 2.5128205128205128,
42
+ "grad_norm": 4.037679195404053,
43
+ "learning_rate": 2.4500000000000003e-06,
44
+ "loss": 2.544,
45
+ "step": 50
46
+ },
47
+ {
48
+ "epoch": 2.5128205128205128,
49
+ "eval_cer": 12.260967379077616,
50
+ "eval_loss": 2.017274856567383,
51
+ "eval_runtime": 54.2097,
52
+ "eval_samples_per_second": 0.332,
53
+ "eval_steps_per_second": 0.166,
54
+ "eval_wer": 22.735346358792185,
55
+ "step": 50
56
+ },
57
+ {
58
+ "epoch": 3.0,
59
+ "grad_norm": 3.2123286724090576,
60
+ "learning_rate": 2.95e-06,
61
+ "loss": 2.3837,
62
+ "step": 60
63
+ },
64
+ {
65
+ "epoch": 3.5128205128205128,
66
+ "grad_norm": 3.1410415172576904,
67
+ "learning_rate": 3.45e-06,
68
+ "loss": 2.3584,
69
+ "step": 70
70
+ },
71
+ {
72
+ "epoch": 4.0,
73
+ "grad_norm": 4.298803329467773,
74
+ "learning_rate": 3.95e-06,
75
+ "loss": 2.3995,
76
+ "step": 80
77
+ },
78
+ {
79
+ "epoch": 4.512820512820513,
80
+ "grad_norm": 4.144355773925781,
81
+ "learning_rate": 4.450000000000001e-06,
82
+ "loss": 2.3549,
83
+ "step": 90
84
+ },
85
+ {
86
+ "epoch": 5.0,
87
+ "grad_norm": 11.266213417053223,
88
+ "learning_rate": 4.95e-06,
89
+ "loss": 2.4564,
90
+ "step": 100
91
+ },
92
+ {
93
+ "epoch": 5.0,
94
+ "eval_cer": 12.29846269216348,
95
+ "eval_loss": 2.0140507221221924,
96
+ "eval_runtime": 27.888,
97
+ "eval_samples_per_second": 0.645,
98
+ "eval_steps_per_second": 0.323,
99
+ "eval_wer": 22.912966252220247,
100
+ "step": 100
101
+ },
102
+ {
103
+ "epoch": 5.512820512820513,
104
+ "grad_norm": 3.633810520172119,
105
+ "learning_rate": 4.9889049115077e-06,
106
+ "loss": 2.276,
107
+ "step": 110
108
+ },
109
+ {
110
+ "epoch": 6.0,
111
+ "grad_norm": 4.385662078857422,
112
+ "learning_rate": 4.9506779365543054e-06,
113
+ "loss": 2.5345,
114
+ "step": 120
115
+ },
116
+ {
117
+ "epoch": 6.512820512820513,
118
+ "grad_norm": 5.4497785568237305,
119
+ "learning_rate": 4.885600821290692e-06,
120
+ "loss": 2.3369,
121
+ "step": 130
122
+ },
123
+ {
124
+ "epoch": 7.0,
125
+ "grad_norm": 8.95763874053955,
126
+ "learning_rate": 4.794386564209953e-06,
127
+ "loss": 2.4669,
128
+ "step": 140
129
+ },
130
+ {
131
+ "epoch": 7.512820512820513,
132
+ "grad_norm": 2.7476675510406494,
133
+ "learning_rate": 4.6780345278004744e-06,
134
+ "loss": 2.4809,
135
+ "step": 150
136
+ },
137
+ {
138
+ "epoch": 7.512820512820513,
139
+ "eval_cer": 12.185976752905887,
140
+ "eval_loss": 2.0147056579589844,
141
+ "eval_runtime": 28.0093,
142
+ "eval_samples_per_second": 0.643,
143
+ "eval_steps_per_second": 0.321,
144
+ "eval_wer": 22.55772646536412,
145
+ "step": 150
146
+ },
147
+ {
148
+ "epoch": 8.0,
149
+ "grad_norm": 3.588670253753662,
150
+ "learning_rate": 4.537819489321385e-06,
151
+ "loss": 2.4501,
152
+ "step": 160
153
+ },
154
+ {
155
+ "epoch": 8.512820512820513,
156
+ "grad_norm": 3.2149131298065186,
157
+ "learning_rate": 4.3752776740761495e-06,
158
+ "loss": 2.2247,
159
+ "step": 170
160
+ },
161
+ {
162
+ "epoch": 9.0,
163
+ "grad_norm": 7.501545429229736,
164
+ "learning_rate": 4.192189924206652e-06,
165
+ "loss": 2.3179,
166
+ "step": 180
167
+ },
168
+ {
169
+ "epoch": 9.512820512820513,
170
+ "grad_norm": 2.922227144241333,
171
+ "learning_rate": 3.9905621874140396e-06,
172
+ "loss": 2.4516,
173
+ "step": 190
174
+ },
175
+ {
176
+ "epoch": 10.0,
177
+ "grad_norm": 4.930537700653076,
178
+ "learning_rate": 3.772603539375929e-06,
179
+ "loss": 2.3659,
180
+ "step": 200
181
+ },
182
+ {
183
+ "epoch": 10.0,
184
+ "eval_cer": 12.260967379077616,
185
+ "eval_loss": 2.013343334197998,
186
+ "eval_runtime": 28.453,
187
+ "eval_samples_per_second": 0.633,
188
+ "eval_steps_per_second": 0.316,
189
+ "eval_wer": 22.735346358792185,
190
+ "step": 200
191
+ },
192
+ {
193
+ "epoch": 10.512820512820513,
194
+ "grad_norm": 3.6379880905151367,
195
+ "learning_rate": 3.5407019806510035e-06,
196
+ "loss": 2.2918,
197
+ "step": 210
198
+ },
199
+ {
200
+ "epoch": 11.0,
201
+ "grad_norm": 3.8149631023406982,
202
+ "learning_rate": 3.2973982732451753e-06,
203
+ "loss": 2.3306,
204
+ "step": 220
205
+ },
206
+ {
207
+ "epoch": 11.512820512820513,
208
+ "grad_norm": 3.5519373416900635,
209
+ "learning_rate": 3.045358103491357e-06,
210
+ "loss": 2.5811,
211
+ "step": 230
212
+ },
213
+ {
214
+ "epoch": 12.0,
215
+ "grad_norm": 7.584492206573486,
216
+ "learning_rate": 2.7873428762321667e-06,
217
+ "loss": 2.3152,
218
+ "step": 240
219
+ },
220
+ {
221
+ "epoch": 12.512820512820513,
222
+ "grad_norm": 4.233317852020264,
223
+ "learning_rate": 2.526179460290615e-06,
224
+ "loss": 2.264,
225
+ "step": 250
226
+ },
227
+ {
228
+ "epoch": 12.512820512820513,
229
+ "eval_cer": 12.29846269216348,
230
+ "eval_loss": 2.0075132846832275,
231
+ "eval_runtime": 28.4486,
232
+ "eval_samples_per_second": 0.633,
233
+ "eval_steps_per_second": 0.316,
234
+ "eval_wer": 22.380106571936057,
235
+ "step": 250
236
+ },
237
+ {
238
+ "epoch": 13.0,
239
+ "grad_norm": 9.289950370788574,
240
+ "learning_rate": 2.2647292167037143e-06,
241
+ "loss": 2.3654,
242
+ "step": 260
243
+ },
244
+ {
245
+ "epoch": 13.512820512820513,
246
+ "grad_norm": 3.141766309738159,
247
+ "learning_rate": 2.0058566490521848e-06,
248
+ "loss": 2.3092,
249
+ "step": 270
250
+ },
251
+ {
252
+ "epoch": 14.0,
253
+ "grad_norm": 6.005232334136963,
254
+ "learning_rate": 1.7523980193597837e-06,
255
+ "loss": 2.3613,
256
+ "step": 280
257
+ },
258
+ {
259
+ "epoch": 14.512820512820513,
260
+ "grad_norm": 3.238589286804199,
261
+ "learning_rate": 1.5071302734130488e-06,
262
+ "loss": 2.2593,
263
+ "step": 290
264
+ },
265
+ {
266
+ "epoch": 15.0,
267
+ "grad_norm": 11.191150665283203,
268
+ "learning_rate": 1.272740615962148e-06,
269
+ "loss": 2.3288,
270
+ "step": 300
271
+ },
272
+ {
273
+ "epoch": 15.0,
274
+ "eval_cer": 12.223472065991752,
275
+ "eval_loss": 2.00502347946167,
276
+ "eval_runtime": 28.2171,
277
+ "eval_samples_per_second": 0.638,
278
+ "eval_steps_per_second": 0.319,
279
+ "eval_wer": 22.202486678507995,
280
+ "step": 300
281
+ }
282
+ ],
283
+ "logging_steps": 10,
284
+ "max_steps": 400,
285
+ "num_input_tokens_seen": 0,
286
+ "num_train_epochs": 20,
287
+ "save_steps": 50,
288
+ "stateful_callbacks": {
289
+ "EarlyStoppingCallback": {
290
+ "args": {
291
+ "early_stopping_patience": 2,
292
+ "early_stopping_threshold": 0.005
293
+ },
294
+ "attributes": {
295
+ "early_stopping_patience_counter": 0
296
+ }
297
+ },
298
+ "TrainerControl": {
299
+ "args": {
300
+ "should_epoch_stop": false,
301
+ "should_evaluate": false,
302
+ "should_log": false,
303
+ "should_save": true,
304
+ "should_training_stop": false
305
+ },
306
+ "attributes": {}
307
+ }
308
+ },
309
+ "total_flos": 4.1397501689856e+18,
310
+ "train_batch_size": 2,
311
+ "trial_name": null,
312
+ "trial_params": null
313
+ }