Ray121381 commited on
Commit
d5b0260
·
1 Parent(s): d4a0599
Files changed (33) hide show
  1. .DS_Store +0 -0
  2. .gitattributes +0 -0
  3. Child-Emotion-Code +0 -1
  4. audio_model_states/.DS_Store +0 -0
  5. audio_model_states/fastspeech2_canton_onnx_1.4.0/fastspeech2_canton.onnx +3 -0
  6. audio_model_states/fastspeech2_canton_onnx_1.4.0/phone_id_map.txt +289 -0
  7. audio_model_states/fastspeech2_canton_onnx_1.4.0/speaker_id_map.txt +20 -0
  8. audio_model_states/hifigan_csmsc_onnx_0.2.0/hifigan_csmsc.onnx +3 -0
  9. audio_model_states/whisper-small-cantonese/.gitattributes +34 -0
  10. audio_model_states/whisper-small-cantonese/README.md +163 -0
  11. audio_model_states/whisper-small-cantonese/added_tokens.json +109 -0
  12. audio_model_states/whisper-small-cantonese/config.json +154 -0
  13. audio_model_states/whisper-small-cantonese/cts/config.json +487 -0
  14. audio_model_states/whisper-small-cantonese/cts/model.bin +3 -0
  15. audio_model_states/whisper-small-cantonese/cts/preprocessor_config.json +14 -0
  16. audio_model_states/whisper-small-cantonese/cts/vocabulary.json +0 -0
  17. audio_model_states/whisper-small-cantonese/generation_config.json +103 -0
  18. audio_model_states/whisper-small-cantonese/ggml-model.bin +3 -0
  19. audio_model_states/whisper-small-cantonese/merges.txt +0 -0
  20. audio_model_states/whisper-small-cantonese/model.safetensors +3 -0
  21. audio_model_states/whisper-small-cantonese/normalizer.json +1742 -0
  22. audio_model_states/whisper-small-cantonese/optimizer.pt +3 -0
  23. audio_model_states/whisper-small-cantonese/preprocessor_config.json +14 -0
  24. audio_model_states/whisper-small-cantonese/pytorch_model.bin +3 -0
  25. audio_model_states/whisper-small-cantonese/rng_state.pth +3 -0
  26. audio_model_states/whisper-small-cantonese/scaler.pt +3 -0
  27. audio_model_states/whisper-small-cantonese/scheduler.pt +3 -0
  28. audio_model_states/whisper-small-cantonese/special_tokens_map.json +139 -0
  29. audio_model_states/whisper-small-cantonese/tokenizer_config.json +981 -0
  30. audio_model_states/whisper-small-cantonese/trainer_state.json +3891 -0
  31. audio_model_states/whisper-small-cantonese/training_args.bin +3 -0
  32. audio_model_states/whisper-small-cantonese/vocab.json +0 -0
  33. lfs.sh +2 -0
.DS_Store CHANGED
File without changes
.gitattributes CHANGED
File without changes
Child-Emotion-Code DELETED
@@ -1 +0,0 @@
1
- Subproject commit 48d9fdda49f14336ee9bd109ea3b952ff216cf41
 
 
audio_model_states/.DS_Store ADDED
Binary file (6.15 kB). View file
 
audio_model_states/fastspeech2_canton_onnx_1.4.0/fastspeech2_canton.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd4e164f9290eb7f4a696568b8209ffc0435c9d3721ec276ae4d72e8e9f90f65
3
+ size 150460283
audio_model_states/fastspeech2_canton_onnx_1.4.0/phone_id_map.txt ADDED
@@ -0,0 +1,289 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <pad> 0
2
+ <unk> 1
3
+ 1 2
4
+ 2 3
5
+ 3 4
6
+ 4 5
7
+ 5 6
8
+ 6 7
9
+ a 8
10
+ a1 9
11
+ a3 10
12
+ aa 11
13
+ aa1 12
14
+ aa2 13
15
+ aa3 14
16
+ aa4 15
17
+ aa5 16
18
+ aa6 17
19
+ aai1 18
20
+ aai2 19
21
+ aai3 20
22
+ aai4 21
23
+ aai5 22
24
+ aai6 23
25
+ aak1 24
26
+ aak3 25
27
+ aak6 26
28
+ aam1 27
29
+ aam2 28
30
+ aam3 29
31
+ aam4 30
32
+ aam5 31
33
+ aam6 32
34
+ aan1 33
35
+ aan2 34
36
+ aan3 35
37
+ aan4 36
38
+ aan5 37
39
+ aan6 38
40
+ aang1 39
41
+ aang2 40
42
+ aang4 41
43
+ aang5 42
44
+ aang6 43
45
+ aap3 44
46
+ aap6 45
47
+ aat1 46
48
+ aat3 47
49
+ aat6 48
50
+ aau1 49
51
+ aau2 50
52
+ aau3 51
53
+ aau4 52
54
+ aau5 53
55
+ aau6 54
56
+ ai 55
57
+ ai1 56
58
+ ai2 57
59
+ ai3 58
60
+ ai4 59
61
+ ai5 60
62
+ ai6 61
63
+ ak 62
64
+ ak1 63
65
+ ak6 64
66
+ am1 65
67
+ am2 66
68
+ am3 67
69
+ am4 68
70
+ am5 69
71
+ am6 70
72
+ an1 71
73
+ an2 72
74
+ an3 73
75
+ an4 74
76
+ an5 75
77
+ an6 76
78
+ ang1 77
79
+ ang2 78
80
+ ang3 79
81
+ ang4 80
82
+ ang6 81
83
+ ap1 82
84
+ ap6 83
85
+ at1 84
86
+ at2 85
87
+ at6 86
88
+ au 87
89
+ au1 88
90
+ au2 89
91
+ au3 90
92
+ au4 91
93
+ au5 92
94
+ au6 93
95
+ b 94
96
+ c 95
97
+ d 96
98
+ e 97
99
+ e1 98
100
+ e2 99
101
+ e3 100
102
+ e4 101
103
+ e5 102
104
+ e6 103
105
+ ei1 104
106
+ ei2 105
107
+ ei3 106
108
+ ei4 107
109
+ ei5 108
110
+ ei6 109
111
+ ek1 110
112
+ ek3 111
113
+ ek6 112
114
+ eng1 113
115
+ eng2 114
116
+ eng3 115
117
+ eng4 116
118
+ eng5 117
119
+ eng6 118
120
+ eoi1 119
121
+ eoi2 120
122
+ eoi3 121
123
+ eoi4 122
124
+ eoi5 123
125
+ eoi6 124
126
+ eon1 125
127
+ eon2 126
128
+ eon3 127
129
+ eon4 128
130
+ eon5 129
131
+ eon6 130
132
+ eot1 131
133
+ eot6 132
134
+ f 133
135
+ g 134
136
+ gw 135
137
+ h 136
138
+ i1 137
139
+ i2 138
140
+ i3 139
141
+ i4 140
142
+ i5 141
143
+ i6 142
144
+ ik1 143
145
+ ik6 144
146
+ im1 145
147
+ im2 146
148
+ im3 147
149
+ im4 148
150
+ im5 149
151
+ im6 150
152
+ in1 151
153
+ in2 152
154
+ in3 153
155
+ in4 154
156
+ in5 155
157
+ in6 156
158
+ ing1 157
159
+ ing2 158
160
+ ing3 159
161
+ ing4 160
162
+ ing5 161
163
+ ing6 162
164
+ ip3 163
165
+ ip6 164
166
+ it1 165
167
+ it3 166
168
+ it6 167
169
+ iu1 168
170
+ iu2 169
171
+ iu3 170
172
+ iu4 171
173
+ iu5 172
174
+ iu6 173
175
+ j 174
176
+ k 175
177
+ k1 176
178
+ kw 177
179
+ l 178
180
+ m 179
181
+ m1 180
182
+ m3 181
183
+ n 182
184
+ n3 183
185
+ ng 184
186
+ o 185
187
+ o1 186
188
+ o2 187
189
+ o3 188
190
+ o4 189
191
+ o5 190
192
+ o6 191
193
+ oe1 192
194
+ oek2 193
195
+ oek3 194
196
+ oek6 195
197
+ oeng1 196
198
+ oeng2 197
199
+ oeng3 198
200
+ oeng4 199
201
+ oeng5 200
202
+ oeng6 201
203
+ oi 202
204
+ oi1 203
205
+ oi2 204
206
+ oi3 205
207
+ oi4 206
208
+ oi6 207
209
+ ok 208
210
+ ok1 209
211
+ ok2 210
212
+ ok3 211
213
+ ok6 212
214
+ on 213
215
+ on1 214
216
+ on2 215
217
+ on3 216
218
+ on4 217
219
+ on6 218
220
+ ong1 219
221
+ ong2 220
222
+ ong3 221
223
+ ong4 222
224
+ ong5 223
225
+ ong6 224
226
+ ot3 225
227
+ ou 226
228
+ ou1 227
229
+ ou2 228
230
+ ou3 229
231
+ ou4 230
232
+ ou5 231
233
+ ou6 232
234
+ p 233
235
+ s 234
236
+ sil 235
237
+ sp 236
238
+ spl 237
239
+ t 238
240
+ t3 239
241
+ u1 240
242
+ u2 241
243
+ u3 242
244
+ u4 243
245
+ u6 244
246
+ ui1 245
247
+ ui2 246
248
+ ui3 247
249
+ ui4 248
250
+ ui5 249
251
+ ui6 250
252
+ uk 251
253
+ uk1 252
254
+ uk6 253
255
+ un1 254
256
+ un2 255
257
+ un3 256
258
+ un4 257
259
+ un5 258
260
+ un6 259
261
+ ung1 260
262
+ ung2 261
263
+ ung3 262
264
+ ung4 263
265
+ ung5 264
266
+ ung6 265
267
+ ut3 266
268
+ ut6 267
269
+ w 268
270
+ yu1 269
271
+ yu2 270
272
+ yu3 271
273
+ yu4 272
274
+ yu5 273
275
+ yu6 274
276
+ yun1 275
277
+ yun2 276
278
+ yun3 277
279
+ yun4 278
280
+ yun5 279
281
+ yun6 280
282
+ yut3 281
283
+ yut6 282
284
+ z 283
285
+ , 284
286
+ . 285
287
+ ? 286
288
+ ! 287
289
+ <eos> 288
audio_model_states/fastspeech2_canton_onnx_1.4.0/speaker_id_map.txt ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ G0001 0
2
+ G0002 1
3
+ G0003 2
4
+ G0004 3
5
+ G0005 4
6
+ G0006 5
7
+ G0007 6
8
+ G0008 7
9
+ G0009 8
10
+ G0010 9
11
+ G0051 10
12
+ G0052 11
13
+ G0053 12
14
+ G0054 13
15
+ G0055 14
16
+ G0067 15
17
+ G0068 16
18
+ G0070 17
19
+ G0071 18
20
+ G0072 19
audio_model_states/hifigan_csmsc_onnx_0.2.0/hifigan_csmsc.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28c95e5e2e12220590d06bd4c6a3dc44391aa77c1e0d40adc422bbc97f76bdbb
3
+ size 51989194
audio_model_states/whisper-small-cantonese/.gitattributes ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tflite filter=lfs diff=lfs merge=lfs -text
29
+ *.tgz filter=lfs diff=lfs merge=lfs -text
30
+ *.wasm filter=lfs diff=lfs merge=lfs -text
31
+ *.xz filter=lfs diff=lfs merge=lfs -text
32
+ *.zip filter=lfs diff=lfs merge=lfs -text
33
+ *.zst filter=lfs diff=lfs merge=lfs -text
34
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
audio_model_states/whisper-small-cantonese/README.md ADDED
@@ -0,0 +1,163 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - zh
4
+ - yue
5
+ license: apache-2.0
6
+ tags:
7
+ - whisper-event
8
+ - generated_from_trainer
9
+ base_model: openai/whisper-small
10
+ datasets:
11
+ - mozilla-foundation/common_voice_16_0
12
+ - mozilla-foundation/common_voice_17_0
13
+ model-index:
14
+ - name: Whisper Small zh-HK - Alvin
15
+ results:
16
+ - task:
17
+ name: Automatic Speech Recognition
18
+ type: automatic-speech-recognition
19
+ dataset:
20
+ name: mozilla-foundation/common_voice_16_0 yue
21
+ type: mozilla-foundation/common_voice_16_0
22
+ config: yue
23
+ split: test
24
+ args: yue
25
+ metrics:
26
+ - name: Normalized CER
27
+ type: cer
28
+ value: 7.93
29
+ ---
30
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
31
+ should probably proofread and complete it, then remove this comment. -->
32
+
33
+ # Whisper Small Cantonese - Alvin
34
+
35
+ This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the Cantonese language. It achieves a 7.93 CER (without punctuations), 9.72 CER (with punctuations) on Common Voice 16.0
36
+
37
+ ## Training and evaluation data
38
+ For training,
39
+ - CantoMap: Winterstein, Grégoire, Tang, Carmen and Lai, Regine (2020) "CantoMap: a Hong Kong Cantonese MapTask Corpus", in Proceedings of The 12th Language Resources and Evaluation Conference, Marseille: European Language Resources Association, p. 2899-2906.
40
+ - Cantonse-ASR: Yu, Tiezheng, Frieske, Rita, Xu, Peng, Cahyawijaya, Samuel, Yiu, Cheuk Tung, Lovenia, Holy, Dai, Wenliang, Barezi, Elham, Chen, Qifeng, Ma, Xiaojuan, Shi, Bertram, Fung, Pascale (2022) "Automatic Speech Recognition Datasets in Cantonese: A Survey and New Dataset", 2022. Link: https://arxiv.org/pdf/2201.02419.pdf
41
+
42
+ |Name|# of Hours|
43
+ |--|--|
44
+ |Common Voice 16.0 zh-HK Train|138|
45
+ |Common Voice 16.0 yue Train|85|
46
+ |Common Voice 17.0 yue Train|178|
47
+ |Cantonese-ASR|72|
48
+ |CantoMap|23|
49
+ |[Pseudo-Labelled YouTube Data](https://huggingface.co/datasets/alvanlii/cantonese-youtube-pseudo-transcription)|438|
50
+
51
+
52
+ For evaluation, Common Voice 16.0 yue Test set is used.
53
+
54
+ ## Results
55
+ - CER (lower is better): 0.0972
56
+ - down from 0.1073, 0.1581 in the previous versions
57
+ - CER (punctuations removed): 0.0793
58
+ - GPU Inference with Fast Attention (example below): 0.055s/sample
59
+ - Note all GPU evaluations are done on RTX 3090 GPU
60
+ - GPU Inference: 0.308s/sample
61
+ - CPU Inference: 2.57s/sample
62
+ - GPU VRAM: ~1.5 GB
63
+
64
+
65
+ ## Using the Model
66
+ ```
67
+ import librosa
68
+
69
+ import torch
70
+ from transformers import WhisperForConditionalGeneration, WhisperProcessor
71
+
72
+ y, sr = librosa.load('audio.mp3', sr=16000)
73
+
74
+ MODEL_NAME = "alvanlii/whisper-small-cantonese"
75
+
76
+ processor = WhisperProcessor.from_pretrained(MODEL_NAME)
77
+ model = WhisperForConditionalGeneration.from_pretrained(MODEL_NAME)
78
+
79
+ processed_in = processor(y, sampling_rate=sr, return_tensors="pt")
80
+ gout = model.generate(
81
+ input_features=processed_in.input_features,
82
+ output_scores=True, return_dict_in_generate=True
83
+ )
84
+ transcription = processor.batch_decode(gout.sequences, skip_special_tokens=True)[0]
85
+ print(transcription)
86
+ ```
87
+ - Alternatively, you can use huggingface pipelines
88
+ ```
89
+ from transformers import pipeline
90
+ MODEL_NAME = "alvanlii/whisper-small-cantonese"
91
+ lang = "zh"
92
+ pipe = pipeline(
93
+ task="automatic-speech-recognition",
94
+ model=MODEL_NAME,
95
+ chunk_length_s=30,
96
+ device=device,
97
+ )
98
+ pipe.model.config.forced_decoder_ids = pipe.tokenizer.get_decoder_prompt_ids(language=lang, task="transcribe")
99
+ text = pipe(file)["text"]
100
+ ```
101
+
102
+ ## Model Speedup
103
+ Just add attn_implementation="sdpa" for Flash Attention.
104
+ ```
105
+ model = AutoModelForSpeechSeq2Seq.from_pretrained(
106
+ "alvanlii/whisper-small-cantonese",
107
+ torch_dtype=torch_dtype,
108
+ low_cpu_mem_usage=True,
109
+ use_safetensors=True,
110
+ attn_implementation="sdpa",
111
+ )
112
+ ```
113
+ Using Flash Attention reduced the amount of time taken per sample from 0.308s to 0.055s.
114
+
115
+ ## Speculative Decoding
116
+ You can use a bigger model, then use `alvanlii/whisper-small-cantonese` to speed up inference with basically no loss in accuracy.
117
+ ```
118
+ model_id = "simonl0909/whisper-large-v2-cantonese"
119
+ model = AutoModelForSpeechSeq2Seq.from_pretrained(
120
+ model_id,
121
+ torch_dtype=torch_dtype,
122
+ low_cpu_mem_usage=True,
123
+ use_safetensors=True,
124
+ attn_implementation="sdpa",
125
+ )
126
+ model.to(device)
127
+
128
+ processor = AutoProcessor.from_pretrained(model_id)
129
+
130
+ assistant_model_id = "alvanlii/whisper-small-cantonese"
131
+
132
+ assistant_model = AutoModelForSpeechSeq2Seq.from_pretrained(
133
+ assistant_model_id,
134
+ torch_dtype=torch_dtype,
135
+ low_cpu_mem_usage=True,
136
+ use_safetensors=True,
137
+ attn_implementation="sdpa",
138
+ )
139
+
140
+ assistant_model.to(device)
141
+ ...
142
+ model.generate(**inputs, use_cache=True, assistant_model=assistant_model)
143
+ ```
144
+ In the original `simonl0909/whisper-large-v2-cantonese` model, it runs at 0.714s/sample for a CER of 7.65. \
145
+ Using speculative decoding with `alvanlii/whisper-small-cantonese`, it runs at 0.137s/sample for a CER of 7.67, which is much faster.
146
+
147
+ ## Whisper.cpp
148
+ Uploaded a GGML bin file for Whisper cpp as of June 2024. You can download the bin file [here](https://huggingface.co/alvanlii/whisper-small-cantonese/blob/main/ggml-model.bin) and try it out [here](https://whisper.ggerganov.com/).
149
+
150
+ ## Whisper CT2
151
+ For use in WhisperX or FasterWhisper, a CT2 file is needed. The converted model is under [here](https://huggingface.co/alvanlii/whisper-small-cantonese/tree/main/cts)
152
+
153
+ ## Training Hyperparameters
154
+ - learning_rate: 5e-5
155
+ - train_batch_size: 25 (on 1 3090 GPU)
156
+ - eval_batch_size: 8
157
+ - gradient_accumulation_steps: 4
158
+ - total_train_batch_size: 25x4=100
159
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
160
+ - lr_scheduler_type: linear
161
+ - lr_scheduler_warmup_steps: 500
162
+ - training_steps: 15000
163
+ - augmentation: None
audio_model_states/whisper-small-cantonese/added_tokens.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<|af|>": 50327,
3
+ "<|am|>": 50334,
4
+ "<|ar|>": 50272,
5
+ "<|as|>": 50350,
6
+ "<|az|>": 50304,
7
+ "<|ba|>": 50355,
8
+ "<|be|>": 50330,
9
+ "<|bg|>": 50292,
10
+ "<|bn|>": 50302,
11
+ "<|bo|>": 50347,
12
+ "<|br|>": 50309,
13
+ "<|bs|>": 50315,
14
+ "<|ca|>": 50270,
15
+ "<|cs|>": 50283,
16
+ "<|cy|>": 50297,
17
+ "<|da|>": 50285,
18
+ "<|de|>": 50261,
19
+ "<|el|>": 50281,
20
+ "<|endoftext|>": 50257,
21
+ "<|en|>": 50259,
22
+ "<|es|>": 50262,
23
+ "<|et|>": 50307,
24
+ "<|eu|>": 50310,
25
+ "<|fa|>": 50300,
26
+ "<|fi|>": 50277,
27
+ "<|fo|>": 50338,
28
+ "<|fr|>": 50265,
29
+ "<|gl|>": 50319,
30
+ "<|gu|>": 50333,
31
+ "<|haw|>": 50352,
32
+ "<|ha|>": 50354,
33
+ "<|hi|>": 50276,
34
+ "<|hr|>": 50291,
35
+ "<|ht|>": 50339,
36
+ "<|hu|>": 50286,
37
+ "<|hy|>": 50312,
38
+ "<|id|>": 50275,
39
+ "<|is|>": 50311,
40
+ "<|it|>": 50274,
41
+ "<|iw|>": 50279,
42
+ "<|ja|>": 50266,
43
+ "<|jw|>": 50356,
44
+ "<|ka|>": 50329,
45
+ "<|kk|>": 50316,
46
+ "<|km|>": 50323,
47
+ "<|kn|>": 50306,
48
+ "<|ko|>": 50264,
49
+ "<|la|>": 50294,
50
+ "<|lb|>": 50345,
51
+ "<|ln|>": 50353,
52
+ "<|lo|>": 50336,
53
+ "<|lt|>": 50293,
54
+ "<|lv|>": 50301,
55
+ "<|mg|>": 50349,
56
+ "<|mi|>": 50295,
57
+ "<|mk|>": 50308,
58
+ "<|ml|>": 50296,
59
+ "<|mn|>": 50314,
60
+ "<|mr|>": 50320,
61
+ "<|ms|>": 50282,
62
+ "<|mt|>": 50343,
63
+ "<|my|>": 50346,
64
+ "<|ne|>": 50313,
65
+ "<|nl|>": 50271,
66
+ "<|nn|>": 50342,
67
+ "<|nocaptions|>": 50362,
68
+ "<|notimestamps|>": 50363,
69
+ "<|no|>": 50288,
70
+ "<|oc|>": 50328,
71
+ "<|pa|>": 50321,
72
+ "<|pl|>": 50269,
73
+ "<|ps|>": 50340,
74
+ "<|pt|>": 50267,
75
+ "<|ro|>": 50284,
76
+ "<|ru|>": 50263,
77
+ "<|sa|>": 50344,
78
+ "<|sd|>": 50332,
79
+ "<|si|>": 50322,
80
+ "<|sk|>": 50298,
81
+ "<|sl|>": 50305,
82
+ "<|sn|>": 50324,
83
+ "<|so|>": 50326,
84
+ "<|sq|>": 50317,
85
+ "<|sr|>": 50303,
86
+ "<|startoflm|>": 50360,
87
+ "<|startofprev|>": 50361,
88
+ "<|startoftranscript|>": 50258,
89
+ "<|su|>": 50357,
90
+ "<|sv|>": 50273,
91
+ "<|sw|>": 50318,
92
+ "<|ta|>": 50287,
93
+ "<|te|>": 50299,
94
+ "<|tg|>": 50331,
95
+ "<|th|>": 50289,
96
+ "<|tk|>": 50341,
97
+ "<|tl|>": 50348,
98
+ "<|transcribe|>": 50359,
99
+ "<|translate|>": 50358,
100
+ "<|tr|>": 50268,
101
+ "<|tt|>": 50351,
102
+ "<|uk|>": 50280,
103
+ "<|ur|>": 50290,
104
+ "<|uz|>": 50337,
105
+ "<|vi|>": 50278,
106
+ "<|yi|>": 50335,
107
+ "<|yo|>": 50325,
108
+ "<|zh|>": 50260
109
+ }
audio_model_states/whisper-small-cantonese/config.json ADDED
@@ -0,0 +1,154 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "alvanlii/whisper-small-cantonese",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "gelu",
5
+ "apply_spec_augment": false,
6
+ "architectures": [
7
+ "WhisperForConditionalGeneration"
8
+ ],
9
+ "attention_dropout": 0.0,
10
+ "begin_suppress_tokens": [
11
+ 220,
12
+ 50257
13
+ ],
14
+ "forced_decoder_ids": [
15
+ [
16
+ 1,
17
+ 50259
18
+ ],
19
+ [
20
+ 2,
21
+ 50359
22
+ ],
23
+ [
24
+ 3,
25
+ 50363
26
+ ]
27
+ ],
28
+ "suppress_tokens": [
29
+ 1,
30
+ 2,
31
+ 7,
32
+ 8,
33
+ 9,
34
+ 10,
35
+ 14,
36
+ 25,
37
+ 26,
38
+ 27,
39
+ 28,
40
+ 29,
41
+ 31,
42
+ 58,
43
+ 59,
44
+ 60,
45
+ 61,
46
+ 62,
47
+ 63,
48
+ 90,
49
+ 91,
50
+ 92,
51
+ 93,
52
+ 359,
53
+ 503,
54
+ 522,
55
+ 542,
56
+ 873,
57
+ 893,
58
+ 902,
59
+ 918,
60
+ 922,
61
+ 931,
62
+ 1350,
63
+ 1853,
64
+ 1982,
65
+ 2460,
66
+ 2627,
67
+ 3246,
68
+ 3253,
69
+ 3268,
70
+ 3536,
71
+ 3846,
72
+ 3961,
73
+ 4183,
74
+ 4667,
75
+ 6585,
76
+ 6647,
77
+ 7273,
78
+ 9061,
79
+ 9383,
80
+ 10428,
81
+ 10929,
82
+ 11938,
83
+ 12033,
84
+ 12331,
85
+ 12562,
86
+ 13793,
87
+ 14157,
88
+ 14635,
89
+ 15265,
90
+ 15618,
91
+ 16553,
92
+ 16604,
93
+ 18362,
94
+ 18956,
95
+ 20075,
96
+ 21675,
97
+ 22520,
98
+ 26130,
99
+ 26161,
100
+ 26435,
101
+ 28279,
102
+ 29464,
103
+ 31650,
104
+ 32302,
105
+ 32470,
106
+ 36865,
107
+ 42863,
108
+ 47425,
109
+ 49870,
110
+ 50254,
111
+ 50258,
112
+ 50360,
113
+ 50361,
114
+ 50362
115
+ ],
116
+ "bos_token_id": 50257,
117
+ "classifier_proj_size": 256,
118
+ "d_model": 768,
119
+ "decoder_attention_heads": 12,
120
+ "decoder_ffn_dim": 3072,
121
+ "decoder_layerdrop": 0.0,
122
+ "decoder_layers": 12,
123
+ "decoder_start_token_id": 50258,
124
+ "dropout": 0.0,
125
+ "encoder_attention_heads": 12,
126
+ "encoder_ffn_dim": 3072,
127
+ "encoder_layerdrop": 0.0,
128
+ "encoder_layers": 12,
129
+ "eos_token_id": 50257,
130
+ "forced_decoder_ids": null,
131
+ "init_std": 0.02,
132
+ "is_encoder_decoder": true,
133
+ "mask_feature_length": 10,
134
+ "mask_feature_min_masks": 0,
135
+ "mask_feature_prob": 0.0,
136
+ "mask_time_length": 10,
137
+ "mask_time_min_masks": 2,
138
+ "mask_time_prob": 0.05,
139
+ "max_length": 448,
140
+ "max_source_positions": 1500,
141
+ "max_target_positions": 448,
142
+ "median_filter_width": 7,
143
+ "model_type": "whisper",
144
+ "num_hidden_layers": 12,
145
+ "num_mel_bins": 80,
146
+ "pad_token_id": 50257,
147
+ "scale_embedding": false,
148
+ "suppress_tokens": [],
149
+ "torch_dtype": "float32",
150
+ "transformers_version": "4.39.3",
151
+ "use_cache": false,
152
+ "use_weighted_layer_sum": false,
153
+ "vocab_size": 51865
154
+ }
audio_model_states/whisper-small-cantonese/cts/config.json ADDED
@@ -0,0 +1,487 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alignment_heads": [
3
+ [
4
+ 6,
5
+ 0
6
+ ],
7
+ [
8
+ 6,
9
+ 1
10
+ ],
11
+ [
12
+ 6,
13
+ 2
14
+ ],
15
+ [
16
+ 6,
17
+ 3
18
+ ],
19
+ [
20
+ 6,
21
+ 4
22
+ ],
23
+ [
24
+ 6,
25
+ 5
26
+ ],
27
+ [
28
+ 6,
29
+ 6
30
+ ],
31
+ [
32
+ 6,
33
+ 7
34
+ ],
35
+ [
36
+ 6,
37
+ 8
38
+ ],
39
+ [
40
+ 6,
41
+ 9
42
+ ],
43
+ [
44
+ 6,
45
+ 10
46
+ ],
47
+ [
48
+ 6,
49
+ 11
50
+ ],
51
+ [
52
+ 7,
53
+ 0
54
+ ],
55
+ [
56
+ 7,
57
+ 1
58
+ ],
59
+ [
60
+ 7,
61
+ 2
62
+ ],
63
+ [
64
+ 7,
65
+ 3
66
+ ],
67
+ [
68
+ 7,
69
+ 4
70
+ ],
71
+ [
72
+ 7,
73
+ 5
74
+ ],
75
+ [
76
+ 7,
77
+ 6
78
+ ],
79
+ [
80
+ 7,
81
+ 7
82
+ ],
83
+ [
84
+ 7,
85
+ 8
86
+ ],
87
+ [
88
+ 7,
89
+ 9
90
+ ],
91
+ [
92
+ 7,
93
+ 10
94
+ ],
95
+ [
96
+ 7,
97
+ 11
98
+ ],
99
+ [
100
+ 8,
101
+ 0
102
+ ],
103
+ [
104
+ 8,
105
+ 1
106
+ ],
107
+ [
108
+ 8,
109
+ 2
110
+ ],
111
+ [
112
+ 8,
113
+ 3
114
+ ],
115
+ [
116
+ 8,
117
+ 4
118
+ ],
119
+ [
120
+ 8,
121
+ 5
122
+ ],
123
+ [
124
+ 8,
125
+ 6
126
+ ],
127
+ [
128
+ 8,
129
+ 7
130
+ ],
131
+ [
132
+ 8,
133
+ 8
134
+ ],
135
+ [
136
+ 8,
137
+ 9
138
+ ],
139
+ [
140
+ 8,
141
+ 10
142
+ ],
143
+ [
144
+ 8,
145
+ 11
146
+ ],
147
+ [
148
+ 9,
149
+ 0
150
+ ],
151
+ [
152
+ 9,
153
+ 1
154
+ ],
155
+ [
156
+ 9,
157
+ 2
158
+ ],
159
+ [
160
+ 9,
161
+ 3
162
+ ],
163
+ [
164
+ 9,
165
+ 4
166
+ ],
167
+ [
168
+ 9,
169
+ 5
170
+ ],
171
+ [
172
+ 9,
173
+ 6
174
+ ],
175
+ [
176
+ 9,
177
+ 7
178
+ ],
179
+ [
180
+ 9,
181
+ 8
182
+ ],
183
+ [
184
+ 9,
185
+ 9
186
+ ],
187
+ [
188
+ 9,
189
+ 10
190
+ ],
191
+ [
192
+ 9,
193
+ 11
194
+ ],
195
+ [
196
+ 10,
197
+ 0
198
+ ],
199
+ [
200
+ 10,
201
+ 1
202
+ ],
203
+ [
204
+ 10,
205
+ 2
206
+ ],
207
+ [
208
+ 10,
209
+ 3
210
+ ],
211
+ [
212
+ 10,
213
+ 4
214
+ ],
215
+ [
216
+ 10,
217
+ 5
218
+ ],
219
+ [
220
+ 10,
221
+ 6
222
+ ],
223
+ [
224
+ 10,
225
+ 7
226
+ ],
227
+ [
228
+ 10,
229
+ 8
230
+ ],
231
+ [
232
+ 10,
233
+ 9
234
+ ],
235
+ [
236
+ 10,
237
+ 10
238
+ ],
239
+ [
240
+ 10,
241
+ 11
242
+ ],
243
+ [
244
+ 11,
245
+ 0
246
+ ],
247
+ [
248
+ 11,
249
+ 1
250
+ ],
251
+ [
252
+ 11,
253
+ 2
254
+ ],
255
+ [
256
+ 11,
257
+ 3
258
+ ],
259
+ [
260
+ 11,
261
+ 4
262
+ ],
263
+ [
264
+ 11,
265
+ 5
266
+ ],
267
+ [
268
+ 11,
269
+ 6
270
+ ],
271
+ [
272
+ 11,
273
+ 7
274
+ ],
275
+ [
276
+ 11,
277
+ 8
278
+ ],
279
+ [
280
+ 11,
281
+ 9
282
+ ],
283
+ [
284
+ 11,
285
+ 10
286
+ ],
287
+ [
288
+ 11,
289
+ 11
290
+ ]
291
+ ],
292
+ "lang_ids": [
293
+ 50259,
294
+ 50260,
295
+ 50261,
296
+ 50262,
297
+ 50263,
298
+ 50264,
299
+ 50265,
300
+ 50266,
301
+ 50267,
302
+ 50268,
303
+ 50269,
304
+ 50270,
305
+ 50271,
306
+ 50272,
307
+ 50273,
308
+ 50274,
309
+ 50275,
310
+ 50276,
311
+ 50277,
312
+ 50278,
313
+ 50279,
314
+ 50280,
315
+ 50281,
316
+ 50282,
317
+ 50283,
318
+ 50284,
319
+ 50285,
320
+ 50286,
321
+ 50287,
322
+ 50288,
323
+ 50289,
324
+ 50290,
325
+ 50291,
326
+ 50292,
327
+ 50293,
328
+ 50294,
329
+ 50295,
330
+ 50296,
331
+ 50297,
332
+ 50298,
333
+ 50299,
334
+ 50300,
335
+ 50301,
336
+ 50302,
337
+ 50303,
338
+ 50304,
339
+ 50305,
340
+ 50306,
341
+ 50307,
342
+ 50308,
343
+ 50309,
344
+ 50310,
345
+ 50311,
346
+ 50312,
347
+ 50313,
348
+ 50314,
349
+ 50315,
350
+ 50316,
351
+ 50317,
352
+ 50318,
353
+ 50319,
354
+ 50320,
355
+ 50321,
356
+ 50322,
357
+ 50323,
358
+ 50324,
359
+ 50325,
360
+ 50326,
361
+ 50327,
362
+ 50328,
363
+ 50329,
364
+ 50330,
365
+ 50331,
366
+ 50332,
367
+ 50333,
368
+ 50334,
369
+ 50335,
370
+ 50336,
371
+ 50337,
372
+ 50338,
373
+ 50339,
374
+ 50340,
375
+ 50341,
376
+ 50342,
377
+ 50343,
378
+ 50344,
379
+ 50345,
380
+ 50346,
381
+ 50347,
382
+ 50348,
383
+ 50349,
384
+ 50350,
385
+ 50351,
386
+ 50352,
387
+ 50353,
388
+ 50354,
389
+ 50355,
390
+ 50356,
391
+ 50357
392
+ ],
393
+ "suppress_ids": [
394
+ 1,
395
+ 2,
396
+ 7,
397
+ 8,
398
+ 9,
399
+ 10,
400
+ 14,
401
+ 25,
402
+ 26,
403
+ 27,
404
+ 28,
405
+ 29,
406
+ 31,
407
+ 58,
408
+ 59,
409
+ 60,
410
+ 61,
411
+ 62,
412
+ 63,
413
+ 90,
414
+ 91,
415
+ 92,
416
+ 93,
417
+ 359,
418
+ 503,
419
+ 522,
420
+ 542,
421
+ 873,
422
+ 893,
423
+ 902,
424
+ 918,
425
+ 922,
426
+ 931,
427
+ 1350,
428
+ 1853,
429
+ 1982,
430
+ 2460,
431
+ 2627,
432
+ 3246,
433
+ 3253,
434
+ 3268,
435
+ 3536,
436
+ 3846,
437
+ 3961,
438
+ 4183,
439
+ 4667,
440
+ 6585,
441
+ 6647,
442
+ 7273,
443
+ 9061,
444
+ 9383,
445
+ 10428,
446
+ 10929,
447
+ 11938,
448
+ 12033,
449
+ 12331,
450
+ 12562,
451
+ 13793,
452
+ 14157,
453
+ 14635,
454
+ 15265,
455
+ 15618,
456
+ 16553,
457
+ 16604,
458
+ 18362,
459
+ 18956,
460
+ 20075,
461
+ 21675,
462
+ 22520,
463
+ 26130,
464
+ 26161,
465
+ 26435,
466
+ 28279,
467
+ 29464,
468
+ 31650,
469
+ 32302,
470
+ 32470,
471
+ 36865,
472
+ 42863,
473
+ 47425,
474
+ 49870,
475
+ 50254,
476
+ 50258,
477
+ 50358,
478
+ 50359,
479
+ 50360,
480
+ 50361,
481
+ 50362
482
+ ],
483
+ "suppress_ids_begin": [
484
+ 220,
485
+ 50257
486
+ ]
487
+ }
audio_model_states/whisper-small-cantonese/cts/model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d343caa7ef827ddb60bc2c1c8c741b29d2bd99d5387aa5c6f96cbd26f6804b34
3
+ size 967072097
audio_model_states/whisper-small-cantonese/cts/preprocessor_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "chunk_length": 30,
3
+ "feature_extractor_type": "WhisperFeatureExtractor",
4
+ "feature_size": 80,
5
+ "hop_length": 160,
6
+ "n_fft": 400,
7
+ "n_samples": 480000,
8
+ "nb_max_frames": 3000,
9
+ "padding_side": "right",
10
+ "padding_value": 0.0,
11
+ "processor_class": "WhisperProcessor",
12
+ "return_attention_mask": false,
13
+ "sampling_rate": 16000
14
+ }
audio_model_states/whisper-small-cantonese/cts/vocabulary.json ADDED
The diff for this file is too large to render. See raw diff
 
audio_model_states/whisper-small-cantonese/generation_config.json ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "begin_suppress_tokens": [
3
+ 220,
4
+ 50257
5
+ ],
6
+ "bos_token_id": 50257,
7
+ "decoder_start_token_id": 50258,
8
+ "eos_token_id": 50257,
9
+ "max_length": 448,
10
+ "pad_token_id": 50257,
11
+ "suppress_tokens": [
12
+ 1,
13
+ 2,
14
+ 7,
15
+ 8,
16
+ 9,
17
+ 10,
18
+ 14,
19
+ 25,
20
+ 26,
21
+ 27,
22
+ 28,
23
+ 29,
24
+ 31,
25
+ 58,
26
+ 59,
27
+ 60,
28
+ 61,
29
+ 62,
30
+ 63,
31
+ 90,
32
+ 91,
33
+ 92,
34
+ 93,
35
+ 359,
36
+ 503,
37
+ 522,
38
+ 542,
39
+ 873,
40
+ 893,
41
+ 902,
42
+ 918,
43
+ 922,
44
+ 931,
45
+ 1350,
46
+ 1853,
47
+ 1982,
48
+ 2460,
49
+ 2627,
50
+ 3246,
51
+ 3253,
52
+ 3268,
53
+ 3536,
54
+ 3846,
55
+ 3961,
56
+ 4183,
57
+ 4667,
58
+ 6585,
59
+ 6647,
60
+ 7273,
61
+ 9061,
62
+ 9383,
63
+ 10428,
64
+ 10929,
65
+ 11938,
66
+ 12033,
67
+ 12331,
68
+ 12562,
69
+ 13793,
70
+ 14157,
71
+ 14635,
72
+ 15265,
73
+ 15618,
74
+ 16553,
75
+ 16604,
76
+ 18362,
77
+ 18956,
78
+ 20075,
79
+ 21675,
80
+ 22520,
81
+ 26130,
82
+ 26161,
83
+ 26435,
84
+ 28279,
85
+ 29464,
86
+ 31650,
87
+ 32302,
88
+ 32470,
89
+ 36865,
90
+ 42863,
91
+ 47425,
92
+ 49870,
93
+ 50254,
94
+ 50258,
95
+ 50358,
96
+ 50359,
97
+ 50360,
98
+ 50361,
99
+ 50362
100
+ ],
101
+ "transformers_version": "4.38.0.dev0",
102
+ "use_cache": false
103
+ }
audio_model_states/whisper-small-cantonese/ggml-model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab6ddf83921733fd96b0bbb4850ae57c330a1049ab35252bf36b1d339835d2bf
3
+ size 487601967
audio_model_states/whisper-small-cantonese/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
audio_model_states/whisper-small-cantonese/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fde842c011b43b2a7b988ad9b9f1a082731bc29c3d58b0cdffa3e71252ef7256
3
+ size 966995080
audio_model_states/whisper-small-cantonese/normalizer.json ADDED
@@ -0,0 +1,1742 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "accessorise": "accessorize",
3
+ "accessorised": "accessorized",
4
+ "accessorises": "accessorizes",
5
+ "accessorising": "accessorizing",
6
+ "acclimatisation": "acclimatization",
7
+ "acclimatise": "acclimatize",
8
+ "acclimatised": "acclimatized",
9
+ "acclimatises": "acclimatizes",
10
+ "acclimatising": "acclimatizing",
11
+ "accoutrements": "accouterments",
12
+ "aeon": "eon",
13
+ "aeons": "eons",
14
+ "aerogramme": "aerogram",
15
+ "aerogrammes": "aerograms",
16
+ "aeroplane": "airplane",
17
+ "aeroplanes": "airplanes",
18
+ "aesthete": "esthete",
19
+ "aesthetes": "esthetes",
20
+ "aesthetic": "esthetic",
21
+ "aesthetically": "esthetically",
22
+ "aesthetics": "esthetics",
23
+ "aetiology": "etiology",
24
+ "ageing": "aging",
25
+ "aggrandisement": "aggrandizement",
26
+ "agonise": "agonize",
27
+ "agonised": "agonized",
28
+ "agonises": "agonizes",
29
+ "agonising": "agonizing",
30
+ "agonisingly": "agonizingly",
31
+ "almanack": "almanac",
32
+ "almanacks": "almanacs",
33
+ "aluminium": "aluminum",
34
+ "amortisable": "amortizable",
35
+ "amortisation": "amortization",
36
+ "amortisations": "amortizations",
37
+ "amortise": "amortize",
38
+ "amortised": "amortized",
39
+ "amortises": "amortizes",
40
+ "amortising": "amortizing",
41
+ "amphitheatre": "amphitheater",
42
+ "amphitheatres": "amphitheaters",
43
+ "anaemia": "anemia",
44
+ "anaemic": "anemic",
45
+ "anaesthesia": "anesthesia",
46
+ "anaesthetic": "anesthetic",
47
+ "anaesthetics": "anesthetics",
48
+ "anaesthetise": "anesthetize",
49
+ "anaesthetised": "anesthetized",
50
+ "anaesthetises": "anesthetizes",
51
+ "anaesthetising": "anesthetizing",
52
+ "anaesthetist": "anesthetist",
53
+ "anaesthetists": "anesthetists",
54
+ "anaesthetize": "anesthetize",
55
+ "anaesthetized": "anesthetized",
56
+ "anaesthetizes": "anesthetizes",
57
+ "anaesthetizing": "anesthetizing",
58
+ "analogue": "analog",
59
+ "analogues": "analogs",
60
+ "analyse": "analyze",
61
+ "analysed": "analyzed",
62
+ "analyses": "analyzes",
63
+ "analysing": "analyzing",
64
+ "anglicise": "anglicize",
65
+ "anglicised": "anglicized",
66
+ "anglicises": "anglicizes",
67
+ "anglicising": "anglicizing",
68
+ "annualised": "annualized",
69
+ "antagonise": "antagonize",
70
+ "antagonised": "antagonized",
71
+ "antagonises": "antagonizes",
72
+ "antagonising": "antagonizing",
73
+ "apologise": "apologize",
74
+ "apologised": "apologized",
75
+ "apologises": "apologizes",
76
+ "apologising": "apologizing",
77
+ "appal": "appall",
78
+ "appals": "appalls",
79
+ "appetiser": "appetizer",
80
+ "appetisers": "appetizers",
81
+ "appetising": "appetizing",
82
+ "appetisingly": "appetizingly",
83
+ "arbour": "arbor",
84
+ "arbours": "arbors",
85
+ "archaeologically": "archeologically",
86
+ "archaeologist": "archeologist",
87
+ "archaeologists": "archeologists",
88
+ "archaeology": "archeology</span>",
89
+ "archeological": "archaeological",
90
+ "ardour": "ardor",
91
+ "armour": "armor",
92
+ "armoured": "armored",
93
+ "armourer": "armorer",
94
+ "armourers": "armorers",
95
+ "armouries": "armories",
96
+ "armoury": "armory",
97
+ "artefact": "artifact",
98
+ "artefacts": "artifacts",
99
+ "authorise": "authorize",
100
+ "authorised": "authorized",
101
+ "authorises": "authorizes",
102
+ "authorising": "authorizing",
103
+ "axe": "ax",
104
+ "backpedalled": "backpedaled",
105
+ "backpedalling": "backpedaling",
106
+ "bannister": "banister",
107
+ "bannisters": "banisters",
108
+ "baptise": "baptize",
109
+ "baptised": "baptized",
110
+ "baptises": "baptizes",
111
+ "baptising": "baptizing",
112
+ "bastardise": "bastardize",
113
+ "bastardised": "bastardized",
114
+ "bastardises": "bastardizes",
115
+ "bastardising": "bastardizing",
116
+ "battleax": "battleaxe",
117
+ "baulk": "balk",
118
+ "baulked": "balked",
119
+ "baulking": "balking",
120
+ "baulks": "balks",
121
+ "bedevilled": "bedeviled",
122
+ "bedevilling": "bedeviling",
123
+ "behaviour": "behavior",
124
+ "behavioural": "behavioral",
125
+ "behaviourism": "behaviorism",
126
+ "behaviourist": "behaviorist",
127
+ "behaviourists": "behaviorists",
128
+ "behaviours": "behaviors",
129
+ "behove": "behoove",
130
+ "behoved": "behooved",
131
+ "behoves": "behooves",
132
+ "bejewelled": "bejeweled",
133
+ "belabour": "belabor",
134
+ "belaboured": "belabored",
135
+ "belabouring": "belaboring",
136
+ "belabours": "belabors",
137
+ "bevelled": "beveled",
138
+ "bevvies": "bevies",
139
+ "bevvy": "bevy",
140
+ "biassed": "biased",
141
+ "biassing": "biasing",
142
+ "bingeing": "binging",
143
+ "bougainvillaea": "bougainvillea",
144
+ "bougainvillaeas": "bougainvilleas",
145
+ "bowdlerise": "bowdlerize",
146
+ "bowdlerised": "bowdlerized",
147
+ "bowdlerises": "bowdlerizes",
148
+ "bowdlerising": "bowdlerizing",
149
+ "breathalyse": "breathalyze",
150
+ "breathalysed": "breathalyzed",
151
+ "breathalyser": "breathalyzer",
152
+ "breathalysers": "breathalyzers",
153
+ "breathalyses": "breathalyzes",
154
+ "breathalysing": "breathalyzing",
155
+ "brutalise": "brutalize",
156
+ "brutalised": "brutalized",
157
+ "brutalises": "brutalizes",
158
+ "brutalising": "brutalizing",
159
+ "busses": "buses",
160
+ "bussing": "busing",
161
+ "caesarean": "cesarean",
162
+ "caesareans": "cesareans",
163
+ "calibre": "caliber",
164
+ "calibres": "calibers",
165
+ "calliper": "caliper",
166
+ "callipers": "calipers",
167
+ "callisthenics": "calisthenics",
168
+ "canalise": "canalize",
169
+ "canalised": "canalized",
170
+ "canalises": "canalizes",
171
+ "canalising": "canalizing",
172
+ "cancelation": "cancellation",
173
+ "cancelations": "cancellations",
174
+ "cancelled": "canceled",
175
+ "cancelling": "canceling",
176
+ "candour": "candor",
177
+ "cannibalise": "cannibalize",
178
+ "cannibalised": "cannibalized",
179
+ "cannibalises": "cannibalizes",
180
+ "cannibalising": "cannibalizing",
181
+ "canonise": "canonize",
182
+ "canonised": "canonized",
183
+ "canonises": "canonizes",
184
+ "canonising": "canonizing",
185
+ "capitalise": "capitalize",
186
+ "capitalised": "capitalized",
187
+ "capitalises": "capitalizes",
188
+ "capitalising": "capitalizing",
189
+ "caramelise": "caramelize",
190
+ "caramelised": "caramelized",
191
+ "caramelises": "caramelizes",
192
+ "caramelising": "caramelizing",
193
+ "carbonise": "carbonize",
194
+ "carbonised": "carbonized",
195
+ "carbonises": "carbonizes",
196
+ "carbonising": "carbonizing",
197
+ "carolled": "caroled",
198
+ "carolling": "caroling",
199
+ "catalogue": "catalog",
200
+ "catalogued": "cataloged",
201
+ "catalogues": "catalogs",
202
+ "cataloguing": "cataloging",
203
+ "catalyse": "catalyze",
204
+ "catalysed": "catalyzed",
205
+ "catalyses": "catalyzes",
206
+ "catalysing": "catalyzing",
207
+ "categorise": "categorize",
208
+ "categorised": "categorized",
209
+ "categorises": "categorizes",
210
+ "categorising": "categorizing",
211
+ "cauterise": "cauterize",
212
+ "cauterised": "cauterized",
213
+ "cauterises": "cauterizes",
214
+ "cauterising": "cauterizing",
215
+ "cavilled": "caviled",
216
+ "cavilling": "caviling",
217
+ "centigramme": "centigram",
218
+ "centigrammes": "centigrams",
219
+ "centilitre": "centiliter",
220
+ "centilitres": "centiliters",
221
+ "centimetre": "centimeter",
222
+ "centimetres": "centimeters",
223
+ "centralise": "centralize",
224
+ "centralised": "centralized",
225
+ "centralises": "centralizes",
226
+ "centralising": "centralizing",
227
+ "centre": "center",
228
+ "centred": "centered",
229
+ "centrefold": "centerfold",
230
+ "centrefolds": "centerfolds",
231
+ "centrepiece": "centerpiece",
232
+ "centrepieces": "centerpieces",
233
+ "centres": "centers",
234
+ "channelled": "channeled",
235
+ "channelling": "channeling",
236
+ "characterise": "characterize",
237
+ "characterised": "characterized",
238
+ "characterises": "characterizes",
239
+ "characterising": "characterizing",
240
+ "cheque": "check",
241
+ "chequebook": "checkbook",
242
+ "chequebooks": "checkbooks",
243
+ "chequered": "checkered",
244
+ "cheques": "checks",
245
+ "chilli": "chili",
246
+ "chimaera": "chimera",
247
+ "chimaeras": "chimeras",
248
+ "chiselled": "chiseled",
249
+ "chiselling": "chiseling",
250
+ "circularise": "circularize",
251
+ "circularised": "circularized",
252
+ "circularises": "circularizes",
253
+ "circularising": "circularizing",
254
+ "civilise": "civilize",
255
+ "civilised": "civilized",
256
+ "civilises": "civilizes",
257
+ "civilising": "civilizing",
258
+ "clamour": "clamor",
259
+ "clamoured": "clamored",
260
+ "clamouring": "clamoring",
261
+ "clamours": "clamors",
262
+ "clangour": "clangor",
263
+ "clarinettist": "clarinetist",
264
+ "clarinettists": "clarinetists",
265
+ "collectivise": "collectivize",
266
+ "collectivised": "collectivized",
267
+ "collectivises": "collectivizes",
268
+ "collectivising": "collectivizing",
269
+ "colonisation": "colonization",
270
+ "colonise": "colonize",
271
+ "colonised": "colonized",
272
+ "coloniser": "colonizer",
273
+ "colonisers": "colonizers",
274
+ "colonises": "colonizes",
275
+ "colonising": "colonizing",
276
+ "colour": "color",
277
+ "colourant": "colorant",
278
+ "colourants": "colorants",
279
+ "coloured": "colored",
280
+ "coloureds": "coloreds",
281
+ "colourful": "colorful",
282
+ "colourfully": "colorfully",
283
+ "colouring": "coloring",
284
+ "colourize": "colorize",
285
+ "colourized": "colorized",
286
+ "colourizes": "colorizes",
287
+ "colourizing": "colorizing",
288
+ "colourless": "colorless",
289
+ "colours": "colors",
290
+ "commercialise": "commercialize",
291
+ "commercialised": "commercialized",
292
+ "commercialises": "commercializes",
293
+ "commercialising": "commercializing",
294
+ "compartmentalise": "compartmentalize",
295
+ "compartmentalised": "compartmentalized",
296
+ "compartmentalises": "compartmentalizes",
297
+ "compartmentalising": "compartmentalizing",
298
+ "computerise": "computerize",
299
+ "computerised": "computerized",
300
+ "computerises": "computerizes",
301
+ "computerising": "computerizing",
302
+ "conceptualise": "conceptualize",
303
+ "conceptualised": "conceptualized",
304
+ "conceptualises": "conceptualizes",
305
+ "conceptualising": "conceptualizing",
306
+ "connexion": "connection",
307
+ "connexions": "connections",
308
+ "contextualise": "contextualize",
309
+ "contextualised": "contextualized",
310
+ "contextualises": "contextualizes",
311
+ "contextualising": "contextualizing",
312
+ "cosier": "cozier",
313
+ "cosies": "cozies",
314
+ "cosiest": "coziest",
315
+ "cosily": "cozily",
316
+ "cosiness": "coziness",
317
+ "cosy": "cozy",
318
+ "councillor": "councilor",
319
+ "councillors": "councilors",
320
+ "counselled": "counseled",
321
+ "counselling": "counseling",
322
+ "counsellor": "counselor",
323
+ "counsellors": "counselors",
324
+ "crenelated": "crenellated",
325
+ "criminalise": "criminalize",
326
+ "criminalised": "criminalized",
327
+ "criminalises": "criminalizes",
328
+ "criminalising": "criminalizing",
329
+ "criticise": "criticize",
330
+ "criticised": "criticized",
331
+ "criticises": "criticizes",
332
+ "criticising": "criticizing",
333
+ "crueller": "crueler",
334
+ "cruellest": "cruelest",
335
+ "crystallisation": "crystallization",
336
+ "crystallise": "crystallize",
337
+ "crystallised": "crystallized",
338
+ "crystallises": "crystallizes",
339
+ "crystallising": "crystallizing",
340
+ "cudgelled": "cudgeled",
341
+ "cudgelling": "cudgeling",
342
+ "customise": "customize",
343
+ "customised": "customized",
344
+ "customises": "customizes",
345
+ "customising": "customizing",
346
+ "cypher": "cipher",
347
+ "cyphers": "ciphers",
348
+ "decentralisation": "decentralization",
349
+ "decentralise": "decentralize",
350
+ "decentralised": "decentralized",
351
+ "decentralises": "decentralizes",
352
+ "decentralising": "decentralizing",
353
+ "decriminalisation": "decriminalization",
354
+ "decriminalise": "decriminalize",
355
+ "decriminalised": "decriminalized",
356
+ "decriminalises": "decriminalizes",
357
+ "decriminalising": "decriminalizing",
358
+ "defence": "defense",
359
+ "defenceless": "defenseless",
360
+ "defences": "defenses",
361
+ "dehumanisation": "dehumanization",
362
+ "dehumanise": "dehumanize",
363
+ "dehumanised": "dehumanized",
364
+ "dehumanises": "dehumanizes",
365
+ "dehumanising": "dehumanizing",
366
+ "demeanour": "demeanor",
367
+ "demilitarisation": "demilitarization",
368
+ "demilitarise": "demilitarize",
369
+ "demilitarised": "demilitarized",
370
+ "demilitarises": "demilitarizes",
371
+ "demilitarising": "demilitarizing",
372
+ "demobilisation": "demobilization",
373
+ "demobilise": "demobilize",
374
+ "demobilised": "demobilized",
375
+ "demobilises": "demobilizes",
376
+ "demobilising": "demobilizing",
377
+ "democratisation": "democratization",
378
+ "democratise": "democratize",
379
+ "democratised": "democratized",
380
+ "democratises": "democratizes",
381
+ "democratising": "democratizing",
382
+ "demonise": "demonize",
383
+ "demonised": "demonized",
384
+ "demonises": "demonizes",
385
+ "demonising": "demonizing",
386
+ "demoralisation": "demoralization",
387
+ "demoralise": "demoralize",
388
+ "demoralised": "demoralized",
389
+ "demoralises": "demoralizes",
390
+ "demoralising": "demoralizing",
391
+ "denationalisation": "denationalization",
392
+ "denationalise": "denationalize",
393
+ "denationalised": "denationalized",
394
+ "denationalises": "denationalizes",
395
+ "denationalising": "denationalizing",
396
+ "deodorise": "deodorize",
397
+ "deodorised": "deodorized",
398
+ "deodorises": "deodorizes",
399
+ "deodorising": "deodorizing",
400
+ "depersonalise": "depersonalize",
401
+ "depersonalised": "depersonalized",
402
+ "depersonalises": "depersonalizes",
403
+ "depersonalising": "depersonalizing",
404
+ "deputise": "deputize",
405
+ "deputised": "deputized",
406
+ "deputises": "deputizes",
407
+ "deputising": "deputizing",
408
+ "desensitisation": "desensitization",
409
+ "desensitise": "desensitize",
410
+ "desensitised": "desensitized",
411
+ "desensitises": "desensitizes",
412
+ "desensitising": "desensitizing",
413
+ "destabilisation": "destabilization",
414
+ "destabilise": "destabilize",
415
+ "destabilised": "destabilized",
416
+ "destabilises": "destabilizes",
417
+ "destabilising": "destabilizing",
418
+ "dialled": "dialed",
419
+ "dialling": "dialing",
420
+ "dialogue": "dialog",
421
+ "dialogues": "dialogs",
422
+ "diarrhoea": "diarrhea",
423
+ "digitise": "digitize",
424
+ "digitised": "digitized",
425
+ "digitises": "digitizes",
426
+ "digitising": "digitizing",
427
+ "disc": "disk",
428
+ "discolour": "discolor",
429
+ "discoloured": "discolored",
430
+ "discolouring": "discoloring",
431
+ "discolours": "discolors",
432
+ "discs": "disks",
433
+ "disembowelled": "disemboweled",
434
+ "disembowelling": "disemboweling",
435
+ "disfavour": "disfavor",
436
+ "dishevelled": "disheveled",
437
+ "dishonour": "dishonor",
438
+ "dishonourable": "dishonorable",
439
+ "dishonourably": "dishonorably",
440
+ "dishonoured": "dishonored",
441
+ "dishonouring": "dishonoring",
442
+ "dishonours": "dishonors",
443
+ "disorganisation": "disorganization",
444
+ "disorganised": "disorganized",
445
+ "distil": "distill",
446
+ "distils": "distills",
447
+ "dramatisation": "dramatization",
448
+ "dramatisations": "dramatizations",
449
+ "dramatise": "dramatize",
450
+ "dramatised": "dramatized",
451
+ "dramatises": "dramatizes",
452
+ "dramatising": "dramatizing",
453
+ "draught": "draft",
454
+ "draughtboard": "draftboard",
455
+ "draughtboards": "draftboards",
456
+ "draughtier": "draftier",
457
+ "draughtiest": "draftiest",
458
+ "draughts": "drafts",
459
+ "draughtsman": "draftsman",
460
+ "draughtsmanship": "draftsmanship",
461
+ "draughtsmen": "draftsmen",
462
+ "draughtswoman": "draftswoman",
463
+ "draughtswomen": "draftswomen",
464
+ "draughty": "drafty",
465
+ "drivelled": "driveled",
466
+ "drivelling": "driveling",
467
+ "duelled": "dueled",
468
+ "duelling": "dueling",
469
+ "economise": "economize",
470
+ "economised": "economized",
471
+ "economises": "economizes",
472
+ "economising": "economizing",
473
+ "editorialise": "editorialize",
474
+ "editorialised": "editorialized",
475
+ "editorialises": "editorializes",
476
+ "editorialising": "editorializing",
477
+ "edoema": "edema",
478
+ "empathise": "empathize",
479
+ "empathised": "empathized",
480
+ "empathises": "empathizes",
481
+ "empathising": "empathizing",
482
+ "emphasise": "emphasize",
483
+ "emphasised": "emphasized",
484
+ "emphasises": "emphasizes",
485
+ "emphasising": "emphasizing",
486
+ "enamelled": "enameled",
487
+ "enamelling": "enameling",
488
+ "enamoured": "enamored",
489
+ "encyclopaedia": "encyclopedia",
490
+ "encyclopaedias": "encyclopedias",
491
+ "encyclopaedic": "encyclopedic",
492
+ "endeavour": "endeavor",
493
+ "endeavoured": "endeavored",
494
+ "endeavouring": "endeavoring",
495
+ "endeavours": "endeavors",
496
+ "energise": "energize",
497
+ "energised": "energized",
498
+ "energises": "energizes",
499
+ "energising": "energizing",
500
+ "enrol": "enroll",
501
+ "enrols": "enrolls",
502
+ "enthral": "enthrall",
503
+ "enthrals": "enthralls",
504
+ "epaulette": "epaulet",
505
+ "epaulettes": "epaulets",
506
+ "epicentre": "epicenter",
507
+ "epicentres": "epicenters",
508
+ "epilogue": "epilog",
509
+ "epilogues": "epilogs",
510
+ "epitomise": "epitomize",
511
+ "epitomised": "epitomized",
512
+ "epitomises": "epitomizes",
513
+ "epitomising": "epitomizing",
514
+ "equalisation": "equalization",
515
+ "equalise": "equalize",
516
+ "equalised": "equalized",
517
+ "equaliser": "equalizer",
518
+ "equalisers": "equalizers",
519
+ "equalises": "equalizes",
520
+ "equalising": "equalizing",
521
+ "eulogise": "eulogize",
522
+ "eulogised": "eulogized",
523
+ "eulogises": "eulogizes",
524
+ "eulogising": "eulogizing",
525
+ "evangelise": "evangelize",
526
+ "evangelised": "evangelized",
527
+ "evangelises": "evangelizes",
528
+ "evangelising": "evangelizing",
529
+ "exorcise": "exorcize",
530
+ "exorcised": "exorcized",
531
+ "exorcises": "exorcizes",
532
+ "exorcising": "exorcizing",
533
+ "extemporisation": "extemporization",
534
+ "extemporise": "extemporize",
535
+ "extemporised": "extemporized",
536
+ "extemporises": "extemporizes",
537
+ "extemporising": "extemporizing",
538
+ "externalisation": "externalization",
539
+ "externalisations": "externalizations",
540
+ "externalise": "externalize",
541
+ "externalised": "externalized",
542
+ "externalises": "externalizes",
543
+ "externalising": "externalizing",
544
+ "factorise": "factorize",
545
+ "factorised": "factorized",
546
+ "factorises": "factorizes",
547
+ "factorising": "factorizing",
548
+ "faecal": "fecal",
549
+ "faeces": "feces",
550
+ "familiarisation": "familiarization",
551
+ "familiarise": "familiarize",
552
+ "familiarised": "familiarized",
553
+ "familiarises": "familiarizes",
554
+ "familiarising": "familiarizing",
555
+ "fantasise": "fantasize",
556
+ "fantasised": "fantasized",
557
+ "fantasises": "fantasizes",
558
+ "fantasising": "fantasizing",
559
+ "favour": "favor",
560
+ "favourable": "favorable",
561
+ "favourably": "favorably",
562
+ "favoured": "favored",
563
+ "favouring": "favoring",
564
+ "favourite": "favorite",
565
+ "favourites": "favorites",
566
+ "favouritism": "favoritism",
567
+ "favours": "favors",
568
+ "feminise": "feminize",
569
+ "feminised": "feminized",
570
+ "feminises": "feminizes",
571
+ "feminising": "feminizing",
572
+ "fertilisation": "fertilization",
573
+ "fertilise": "fertilize",
574
+ "fertilised": "fertilized",
575
+ "fertiliser": "fertilizer",
576
+ "fertilisers": "fertilizers",
577
+ "fertilises": "fertilizes",
578
+ "fertilising": "fertilizing",
579
+ "fervour": "fervor",
580
+ "fibre": "fiber",
581
+ "fibreglass": "fiberglass",
582
+ "fibres": "fibers",
583
+ "fictionalisation": "fictionalization",
584
+ "fictionalisations": "fictionalizations",
585
+ "fictionalise": "fictionalize",
586
+ "fictionalised": "fictionalized",
587
+ "fictionalises": "fictionalizes",
588
+ "fictionalising": "fictionalizing",
589
+ "fillet": "filet",
590
+ "filleted": "fileted",
591
+ "filleting": "fileting",
592
+ "fillets": "filets",
593
+ "finalisation": "finalization",
594
+ "finalise": "finalize",
595
+ "finalised": "finalized",
596
+ "finalises": "finalizes",
597
+ "finalising": "finalizing",
598
+ "flautist": "flutist",
599
+ "flautists": "flutists",
600
+ "flavour": "flavor",
601
+ "flavoured": "flavored",
602
+ "flavouring": "flavoring",
603
+ "flavourings": "flavorings",
604
+ "flavourless": "flavorless",
605
+ "flavours": "flavors",
606
+ "flavoursome": "flavorsome",
607
+ "flyer / flier": "flier / flyer",
608
+ "foetal": "fetal",
609
+ "foetid": "fetid",
610
+ "foetus": "fetus",
611
+ "foetuses": "fetuses",
612
+ "formalisation": "formalization",
613
+ "formalise": "formalize",
614
+ "formalised": "formalized",
615
+ "formalises": "formalizes",
616
+ "formalising": "formalizing",
617
+ "fossilisation": "fossilization",
618
+ "fossilise": "fossilize",
619
+ "fossilised": "fossilized",
620
+ "fossilises": "fossilizes",
621
+ "fossilising": "fossilizing",
622
+ "fraternisation": "fraternization",
623
+ "fraternise": "fraternize",
624
+ "fraternised": "fraternized",
625
+ "fraternises": "fraternizes",
626
+ "fraternising": "fraternizing",
627
+ "fulfil": "fulfill",
628
+ "fulfilment": "fulfillment",
629
+ "fulfils": "fulfills",
630
+ "funnelled": "funneled",
631
+ "funnelling": "funneling",
632
+ "gage": "gauge",
633
+ "gaged": "gauged",
634
+ "gages": "gauges",
635
+ "gaging": "gauging",
636
+ "galvanise": "galvanize",
637
+ "galvanised": "galvanized",
638
+ "galvanises": "galvanizes",
639
+ "galvanising": "galvanizing",
640
+ "gambolled": "gamboled",
641
+ "gambolling": "gamboling",
642
+ "gaol": "jail",
643
+ "gaolbird": "jailbird",
644
+ "gaolbirds": "jailbirds",
645
+ "gaolbreak": "jailbreak",
646
+ "gaolbreaks": "jailbreaks",
647
+ "gaoled": "jailed",
648
+ "gaoler": "jailer",
649
+ "gaolers": "jailers",
650
+ "gaoling": "jailing",
651
+ "gaols": "jails",
652
+ "gasses": "gases",
653
+ "generalisation": "generalization",
654
+ "generalisations": "generalizations",
655
+ "generalise": "generalize",
656
+ "generalised": "generalized",
657
+ "generalises": "generalizes",
658
+ "generalising": "generalizing",
659
+ "ghettoise": "ghettoize",
660
+ "ghettoised": "ghettoized",
661
+ "ghettoises": "ghettoizes",
662
+ "ghettoising": "ghettoizing",
663
+ "gipsies": "gypsies",
664
+ "glamor": "glamour",
665
+ "glamorise": "glamorize",
666
+ "glamorised": "glamorized",
667
+ "glamorises": "glamorizes",
668
+ "glamorising": "glamorizing",
669
+ "globalisation": "globalization",
670
+ "globalise": "globalize",
671
+ "globalised": "globalized",
672
+ "globalises": "globalizes",
673
+ "globalising": "globalizing",
674
+ "glueing": "gluing",
675
+ "goitre": "goiter",
676
+ "goitres": "goiters",
677
+ "gonorrhoea": "gonorrhea",
678
+ "gramme": "gram",
679
+ "grammes": "grams",
680
+ "gravelled": "graveled",
681
+ "grey": "gray",
682
+ "greyed": "grayed",
683
+ "greying": "graying",
684
+ "greyish": "grayish",
685
+ "greyness": "grayness",
686
+ "greys": "grays",
687
+ "grovelled": "groveled",
688
+ "grovelling": "groveling",
689
+ "groyne": "groin",
690
+ "groynes": "groins",
691
+ "gruelling": "grueling",
692
+ "gruellingly": "gruelingly",
693
+ "gryphon": "griffin",
694
+ "gryphons": "griffins",
695
+ "gynaecological": "gynecological",
696
+ "gynaecologist": "gynecologist",
697
+ "gynaecologists": "gynecologists",
698
+ "gynaecology": "gynecology",
699
+ "haematological": "hematological",
700
+ "haematologist": "hematologist",
701
+ "haematologists": "hematologists",
702
+ "haematology": "hematology",
703
+ "haemoglobin": "hemoglobin",
704
+ "haemophilia": "hemophilia",
705
+ "haemophiliac": "hemophiliac",
706
+ "haemophiliacs": "hemophiliacs",
707
+ "haemorrhage": "hemorrhage",
708
+ "haemorrhaged": "hemorrhaged",
709
+ "haemorrhages": "hemorrhages",
710
+ "haemorrhaging": "hemorrhaging",
711
+ "haemorrhoids": "hemorrhoids",
712
+ "harbour": "harbor",
713
+ "harboured": "harbored",
714
+ "harbouring": "harboring",
715
+ "harbours": "harbors",
716
+ "harmonisation": "harmonization",
717
+ "harmonise": "harmonize",
718
+ "harmonised": "harmonized",
719
+ "harmonises": "harmonizes",
720
+ "harmonising": "harmonizing",
721
+ "homoeopath": "homeopath",
722
+ "homoeopathic": "homeopathic",
723
+ "homoeopaths": "homeopaths",
724
+ "homoeopathy": "homeopathy",
725
+ "homogenise": "homogenize",
726
+ "homogenised": "homogenized",
727
+ "homogenises": "homogenizes",
728
+ "homogenising": "homogenizing",
729
+ "honour": "honor",
730
+ "honourable": "honorable",
731
+ "honourably": "honorably",
732
+ "honoured": "honored",
733
+ "honouring": "honoring",
734
+ "honours": "honors",
735
+ "hospitalisation": "hospitalization",
736
+ "hospitalise": "hospitalize",
737
+ "hospitalised": "hospitalized",
738
+ "hospitalises": "hospitalizes",
739
+ "hospitalising": "hospitalizing",
740
+ "humanise": "humanize",
741
+ "humanised": "humanized",
742
+ "humanises": "humanizes",
743
+ "humanising": "humanizing",
744
+ "humour": "humor",
745
+ "humoured": "humored",
746
+ "humouring": "humoring",
747
+ "humourless": "humorless",
748
+ "humours": "humors",
749
+ "hybridise": "hybridize",
750
+ "hybridised": "hybridized",
751
+ "hybridises": "hybridizes",
752
+ "hybridising": "hybridizing",
753
+ "hypnotise": "hypnotize",
754
+ "hypnotised": "hypnotized",
755
+ "hypnotises": "hypnotizes",
756
+ "hypnotising": "hypnotizing",
757
+ "hypothesise": "hypothesize",
758
+ "hypothesised": "hypothesized",
759
+ "hypothesises": "hypothesizes",
760
+ "hypothesising": "hypothesizing",
761
+ "idealisation": "idealization",
762
+ "idealise": "idealize",
763
+ "idealised": "idealized",
764
+ "idealises": "idealizes",
765
+ "idealising": "idealizing",
766
+ "idolise": "idolize",
767
+ "idolised": "idolized",
768
+ "idolises": "idolizes",
769
+ "idolising": "idolizing",
770
+ "immobilisation": "immobilization",
771
+ "immobilise": "immobilize",
772
+ "immobilised": "immobilized",
773
+ "immobiliser": "immobilizer",
774
+ "immobilisers": "immobilizers",
775
+ "immobilises": "immobilizes",
776
+ "immobilising": "immobilizing",
777
+ "immortalise": "immortalize",
778
+ "immortalised": "immortalized",
779
+ "immortalises": "immortalizes",
780
+ "immortalising": "immortalizing",
781
+ "immunisation": "immunization",
782
+ "immunise": "immunize",
783
+ "immunised": "immunized",
784
+ "immunises": "immunizes",
785
+ "immunising": "immunizing",
786
+ "impanelled": "impaneled",
787
+ "impanelling": "impaneling",
788
+ "imperilled": "imperiled",
789
+ "imperilling": "imperiling",
790
+ "individualise": "individualize",
791
+ "individualised": "individualized",
792
+ "individualises": "individualizes",
793
+ "individualising": "individualizing",
794
+ "industrialise": "industrialize",
795
+ "industrialised": "industrialized",
796
+ "industrialises": "industrializes",
797
+ "industrialising": "industrializing",
798
+ "inflexion": "inflection",
799
+ "inflexions": "inflections",
800
+ "initialise": "initialize",
801
+ "initialised": "initialized",
802
+ "initialises": "initializes",
803
+ "initialising": "initializing",
804
+ "initialled": "initialed",
805
+ "initialling": "initialing",
806
+ "instal": "install",
807
+ "instalment": "installment",
808
+ "instalments": "installments",
809
+ "instals": "installs",
810
+ "instil": "instill",
811
+ "instils": "instills",
812
+ "institutionalisation": "institutionalization",
813
+ "institutionalise": "institutionalize",
814
+ "institutionalised": "institutionalized",
815
+ "institutionalises": "institutionalizes",
816
+ "institutionalising": "institutionalizing",
817
+ "intellectualise": "intellectualize",
818
+ "intellectualised": "intellectualized",
819
+ "intellectualises": "intellectualizes",
820
+ "intellectualising": "intellectualizing",
821
+ "internalisation": "internalization",
822
+ "internalise": "internalize",
823
+ "internalised": "internalized",
824
+ "internalises": "internalizes",
825
+ "internalising": "internalizing",
826
+ "internationalisation": "internationalization",
827
+ "internationalise": "internationalize",
828
+ "internationalised": "internationalized",
829
+ "internationalises": "internationalizes",
830
+ "internationalising": "internationalizing",
831
+ "ionisation": "ionization",
832
+ "ionise": "ionize",
833
+ "ionised": "ionized",
834
+ "ioniser": "ionizer",
835
+ "ionisers": "ionizers",
836
+ "ionises": "ionizes",
837
+ "ionising": "ionizing",
838
+ "italicise": "italicize",
839
+ "italicised": "italicized",
840
+ "italicises": "italicizes",
841
+ "italicising": "italicizing",
842
+ "itemise": "itemize",
843
+ "itemised": "itemized",
844
+ "itemises": "itemizes",
845
+ "itemising": "itemizing",
846
+ "jeopardise": "jeopardize",
847
+ "jeopardised": "jeopardized",
848
+ "jeopardises": "jeopardizes",
849
+ "jeopardising": "jeopardizing",
850
+ "jewelled": "jeweled",
851
+ "jeweller": "jeweler",
852
+ "jewellers": "jewelers",
853
+ "jewellery": "jewelry",
854
+ "judgement": "judgment",
855
+ "kilogramme": "kilogram",
856
+ "kilogrammes": "kilograms",
857
+ "kilometre": "kilometer",
858
+ "kilometres": "kilometers",
859
+ "labelled": "labeled",
860
+ "labelling": "labeling",
861
+ "labour": "labor",
862
+ "laboured": "labored",
863
+ "labourer": "laborer",
864
+ "labourers": "laborers",
865
+ "labouring": "laboring",
866
+ "labours": "labors",
867
+ "lacklustre": "lackluster",
868
+ "legalisation": "legalization",
869
+ "legalise": "legalize",
870
+ "legalised": "legalized",
871
+ "legalises": "legalizes",
872
+ "legalising": "legalizing",
873
+ "legitimise": "legitimize",
874
+ "legitimised": "legitimized",
875
+ "legitimises": "legitimizes",
876
+ "legitimising": "legitimizing",
877
+ "leukaemia": "leukemia",
878
+ "levelled": "leveled",
879
+ "leveller": "leveler",
880
+ "levellers": "levelers",
881
+ "levelling": "leveling",
882
+ "libelled": "libeled",
883
+ "libelling": "libeling",
884
+ "libellous": "libelous",
885
+ "liberalisation": "liberalization",
886
+ "liberalise": "liberalize",
887
+ "liberalised": "liberalized",
888
+ "liberalises": "liberalizes",
889
+ "liberalising": "liberalizing",
890
+ "licence": "license",
891
+ "licenced": "licensed",
892
+ "licences": "licenses",
893
+ "licencing": "licensing",
894
+ "likeable": "likable",
895
+ "lionisation": "lionization",
896
+ "lionise": "lionize",
897
+ "lionised": "lionized",
898
+ "lionises": "lionizes",
899
+ "lionising": "lionizing",
900
+ "liquidise": "liquidize",
901
+ "liquidised": "liquidized",
902
+ "liquidiser": "liquidizer",
903
+ "liquidisers": "liquidizers",
904
+ "liquidises": "liquidizes",
905
+ "liquidising": "liquidizing",
906
+ "litre": "liter",
907
+ "litres": "liters",
908
+ "localise": "localize",
909
+ "localised": "localized",
910
+ "localises": "localizes",
911
+ "localising": "localizing",
912
+ "louvre": "louver",
913
+ "louvred": "louvered",
914
+ "louvres": "louvers",
915
+ "lustre": "luster",
916
+ "magnetise": "magnetize",
917
+ "magnetised": "magnetized",
918
+ "magnetises": "magnetizes",
919
+ "magnetising": "magnetizing",
920
+ "manoeuvrability": "maneuverability",
921
+ "manoeuvrable": "maneuverable",
922
+ "manoeuvre": "maneuver",
923
+ "manoeuvred": "maneuvered",
924
+ "manoeuvres": "maneuvers",
925
+ "manoeuvring": "maneuvering",
926
+ "manoeuvrings": "maneuverings",
927
+ "marginalisation": "marginalization",
928
+ "marginalise": "marginalize",
929
+ "marginalised": "marginalized",
930
+ "marginalises": "marginalizes",
931
+ "marginalising": "marginalizing",
932
+ "marshalled": "marshaled",
933
+ "marshalling": "marshaling",
934
+ "marvelled": "marveled",
935
+ "marvelling": "marveling",
936
+ "marvellous": "marvelous",
937
+ "marvellously": "marvelously",
938
+ "materialisation": "materialization",
939
+ "materialise": "materialize",
940
+ "materialised": "materialized",
941
+ "materialises": "materializes",
942
+ "materialising": "materializing",
943
+ "maximisation": "maximization",
944
+ "maximise": "maximize",
945
+ "maximised": "maximized",
946
+ "maximises": "maximizes",
947
+ "maximising": "maximizing",
948
+ "meagre": "meager",
949
+ "mechanisation": "mechanization",
950
+ "mechanise": "mechanize",
951
+ "mechanised": "mechanized",
952
+ "mechanises": "mechanizes",
953
+ "mechanising": "mechanizing",
954
+ "mediaeval": "medieval",
955
+ "memorialise": "memorialize",
956
+ "memorialised": "memorialized",
957
+ "memorialises": "memorializes",
958
+ "memorialising": "memorializing",
959
+ "memorise": "memorize",
960
+ "memorised": "memorized",
961
+ "memorises": "memorizes",
962
+ "memorising": "memorizing",
963
+ "mesmerise": "mesmerize",
964
+ "mesmerised": "mesmerized",
965
+ "mesmerises": "mesmerizes",
966
+ "mesmerising": "mesmerizing",
967
+ "metabolise": "metabolize",
968
+ "metabolised": "metabolized",
969
+ "metabolises": "metabolizes",
970
+ "metabolising": "metabolizing",
971
+ "metre": "meter",
972
+ "metres": "meters",
973
+ "mhm": "hmm",
974
+ "micrometre": "micrometer",
975
+ "micrometres": "micrometers",
976
+ "militarise": "militarize",
977
+ "militarised": "militarized",
978
+ "militarises": "militarizes",
979
+ "militarising": "militarizing",
980
+ "milligramme": "milligram",
981
+ "milligrammes": "milligrams",
982
+ "millilitre": "milliliter",
983
+ "millilitres": "milliliters",
984
+ "millimetre": "millimeter",
985
+ "millimetres": "millimeters",
986
+ "miniaturisation": "miniaturization",
987
+ "miniaturise": "miniaturize",
988
+ "miniaturised": "miniaturized",
989
+ "miniaturises": "miniaturizes",
990
+ "miniaturising": "miniaturizing",
991
+ "minibusses": "minibuses",
992
+ "minimise": "minimize",
993
+ "minimised": "minimized",
994
+ "minimises": "minimizes",
995
+ "minimising": "minimizing",
996
+ "misbehaviour": "misbehavior",
997
+ "misdemeanour": "misdemeanor",
998
+ "misdemeanours": "misdemeanors",
999
+ "misspelt": "misspelled",
1000
+ "mitre": "miter",
1001
+ "mitres": "miters",
1002
+ "mm": "hmm",
1003
+ "mmm": "hmm",
1004
+ "mobilisation": "mobilization",
1005
+ "mobilise": "mobilize",
1006
+ "mobilised": "mobilized",
1007
+ "mobilises": "mobilizes",
1008
+ "mobilising": "mobilizing",
1009
+ "modelled": "modeled",
1010
+ "modeller": "modeler",
1011
+ "modellers": "modelers",
1012
+ "modelling": "modeling",
1013
+ "modernise": "modernize",
1014
+ "modernised": "modernized",
1015
+ "modernises": "modernizes",
1016
+ "modernising": "modernizing",
1017
+ "moisturise": "moisturize",
1018
+ "moisturised": "moisturized",
1019
+ "moisturiser": "moisturizer",
1020
+ "moisturisers": "moisturizers",
1021
+ "moisturises": "moisturizes",
1022
+ "moisturising": "moisturizing",
1023
+ "monologue": "monolog",
1024
+ "monologues": "monologs",
1025
+ "monopolisation": "monopolization",
1026
+ "monopolise": "monopolize",
1027
+ "monopolised": "monopolized",
1028
+ "monopolises": "monopolizes",
1029
+ "monopolising": "monopolizing",
1030
+ "moralise": "moralize",
1031
+ "moralised": "moralized",
1032
+ "moralises": "moralizes",
1033
+ "moralising": "moralizing",
1034
+ "motorised": "motorized",
1035
+ "mould": "mold",
1036
+ "moulded": "molded",
1037
+ "moulder": "molder",
1038
+ "mouldered": "moldered",
1039
+ "mouldering": "moldering",
1040
+ "moulders": "molders",
1041
+ "mouldier": "moldier",
1042
+ "mouldiest": "moldiest",
1043
+ "moulding": "molding",
1044
+ "mouldings": "moldings",
1045
+ "moulds": "molds",
1046
+ "mouldy": "moldy",
1047
+ "moult": "molt",
1048
+ "moulted": "molted",
1049
+ "moulting": "molting",
1050
+ "moults": "molts",
1051
+ "moustache": "mustache",
1052
+ "moustached": "mustached",
1053
+ "moustaches": "mustaches",
1054
+ "moustachioed": "mustachioed",
1055
+ "multicoloured": "multicolored",
1056
+ "nationalisation": "nationalization",
1057
+ "nationalisations": "nationalizations",
1058
+ "nationalise": "nationalize",
1059
+ "nationalised": "nationalized",
1060
+ "nationalises": "nationalizes",
1061
+ "nationalising": "nationalizing",
1062
+ "naturalisation": "naturalization",
1063
+ "naturalise": "naturalize",
1064
+ "naturalised": "naturalized",
1065
+ "naturalises": "naturalizes",
1066
+ "naturalising": "naturalizing",
1067
+ "neighbour": "neighbor",
1068
+ "neighbourhood": "neighborhood",
1069
+ "neighbourhoods": "neighborhoods",
1070
+ "neighbouring": "neighboring",
1071
+ "neighbourliness": "neighborliness",
1072
+ "neighbourly": "neighborly",
1073
+ "neighbours": "neighbors",
1074
+ "neutralisation": "neutralization",
1075
+ "neutralise": "neutralize",
1076
+ "neutralised": "neutralized",
1077
+ "neutralises": "neutralizes",
1078
+ "neutralising": "neutralizing",
1079
+ "normalisation": "normalization",
1080
+ "normalise": "normalize",
1081
+ "normalised": "normalized",
1082
+ "normalises": "normalizes",
1083
+ "normalising": "normalizing",
1084
+ "odour": "odor",
1085
+ "odourless": "odorless",
1086
+ "odours": "odors",
1087
+ "oesophagus": "esophagus",
1088
+ "oesophaguses": "esophaguses",
1089
+ "oestrogen": "estrogen",
1090
+ "offence": "offense",
1091
+ "offences": "offenses",
1092
+ "omelette": "omelet",
1093
+ "omelettes": "omelets",
1094
+ "optimise": "optimize",
1095
+ "optimised": "optimized",
1096
+ "optimises": "optimizes",
1097
+ "optimising": "optimizing",
1098
+ "organisation": "organization",
1099
+ "organisational": "organizational",
1100
+ "organisations": "organizations",
1101
+ "organise": "organize",
1102
+ "organised": "organized",
1103
+ "organiser": "organizer",
1104
+ "organisers": "organizers",
1105
+ "organises": "organizes",
1106
+ "organising": "organizing",
1107
+ "orthopaedic": "orthopedic",
1108
+ "orthopaedics": "orthopedics",
1109
+ "ostracise": "ostracize",
1110
+ "ostracised": "ostracized",
1111
+ "ostracises": "ostracizes",
1112
+ "ostracising": "ostracizing",
1113
+ "outmanoeuvre": "outmaneuver",
1114
+ "outmanoeuvred": "outmaneuvered",
1115
+ "outmanoeuvres": "outmaneuvers",
1116
+ "outmanoeuvring": "outmaneuvering",
1117
+ "overemphasise": "overemphasize",
1118
+ "overemphasised": "overemphasized",
1119
+ "overemphasises": "overemphasizes",
1120
+ "overemphasising": "overemphasizing",
1121
+ "oxidisation": "oxidization",
1122
+ "oxidise": "oxidize",
1123
+ "oxidised": "oxidized",
1124
+ "oxidises": "oxidizes",
1125
+ "oxidising": "oxidizing",
1126
+ "paederast": "pederast",
1127
+ "paederasts": "pederasts",
1128
+ "paediatric": "pediatric",
1129
+ "paediatrician": "pediatrician",
1130
+ "paediatricians": "pediatricians",
1131
+ "paediatrics": "pediatrics",
1132
+ "paedophile": "pedophile",
1133
+ "paedophiles": "pedophiles",
1134
+ "paedophilia": "pedophilia",
1135
+ "palaeolithic": "paleolithic",
1136
+ "palaeontologist": "paleontologist",
1137
+ "palaeontologists": "paleontologists",
1138
+ "palaeontology": "paleontology",
1139
+ "panelled": "paneled",
1140
+ "panelling": "paneling",
1141
+ "panellist": "panelist",
1142
+ "panellists": "panelists",
1143
+ "paralyse": "paralyze",
1144
+ "paralysed": "paralyzed",
1145
+ "paralyses": "paralyzes",
1146
+ "paralysing": "paralyzing",
1147
+ "parcelled": "parceled",
1148
+ "parcelling": "parceling",
1149
+ "parlour": "parlor",
1150
+ "parlours": "parlors",
1151
+ "particularise": "particularize",
1152
+ "particularised": "particularized",
1153
+ "particularises": "particularizes",
1154
+ "particularising": "particularizing",
1155
+ "passivisation": "passivization",
1156
+ "passivise": "passivize",
1157
+ "passivised": "passivized",
1158
+ "passivises": "passivizes",
1159
+ "passivising": "passivizing",
1160
+ "pasteurisation": "pasteurization",
1161
+ "pasteurise": "pasteurize",
1162
+ "pasteurised": "pasteurized",
1163
+ "pasteurises": "pasteurizes",
1164
+ "pasteurising": "pasteurizing",
1165
+ "patronise": "patronize",
1166
+ "patronised": "patronized",
1167
+ "patronises": "patronizes",
1168
+ "patronising": "patronizing",
1169
+ "patronisingly": "patronizingly",
1170
+ "pedalled": "pedaled",
1171
+ "pedalling": "pedaling",
1172
+ "pedestrianisation": "pedestrianization",
1173
+ "pedestrianise": "pedestrianize",
1174
+ "pedestrianised": "pedestrianized",
1175
+ "pedestrianises": "pedestrianizes",
1176
+ "pedestrianising": "pedestrianizing",
1177
+ "penalise": "penalize",
1178
+ "penalised": "penalized",
1179
+ "penalises": "penalizes",
1180
+ "penalising": "penalizing",
1181
+ "pencilled": "penciled",
1182
+ "pencilling": "penciling",
1183
+ "personalise": "personalize",
1184
+ "personalised": "personalized",
1185
+ "personalises": "personalizes",
1186
+ "personalising": "personalizing",
1187
+ "pharmacopoeia": "pharmacopeia",
1188
+ "pharmacopoeias": "pharmacopeias",
1189
+ "philosophise": "philosophize",
1190
+ "philosophised": "philosophized",
1191
+ "philosophises": "philosophizes",
1192
+ "philosophising": "philosophizing",
1193
+ "philtre": "filter",
1194
+ "philtres": "filters",
1195
+ "phoney": "phony",
1196
+ "plagiarise": "plagiarize",
1197
+ "plagiarised": "plagiarized",
1198
+ "plagiarises": "plagiarizes",
1199
+ "plagiarising": "plagiarizing",
1200
+ "plough": "plow",
1201
+ "ploughed": "plowed",
1202
+ "ploughing": "plowing",
1203
+ "ploughman": "plowman",
1204
+ "ploughmen": "plowmen",
1205
+ "ploughs": "plows",
1206
+ "ploughshare": "plowshare",
1207
+ "ploughshares": "plowshares",
1208
+ "polarisation": "polarization",
1209
+ "polarise": "polarize",
1210
+ "polarised": "polarized",
1211
+ "polarises": "polarizes",
1212
+ "polarising": "polarizing",
1213
+ "politicisation": "politicization",
1214
+ "politicise": "politicize",
1215
+ "politicised": "politicized",
1216
+ "politicises": "politicizes",
1217
+ "politicising": "politicizing",
1218
+ "popularisation": "popularization",
1219
+ "popularise": "popularize",
1220
+ "popularised": "popularized",
1221
+ "popularises": "popularizes",
1222
+ "popularising": "popularizing",
1223
+ "pouffe": "pouf",
1224
+ "pouffes": "poufs",
1225
+ "practise": "practice",
1226
+ "practised": "practiced",
1227
+ "practises": "practices",
1228
+ "practising": "practicing",
1229
+ "praesidium": "presidium",
1230
+ "praesidiums": "presidiums",
1231
+ "pressurisation": "pressurization",
1232
+ "pressurise": "pressurize",
1233
+ "pressurised": "pressurized",
1234
+ "pressurises": "pressurizes",
1235
+ "pressurising": "pressurizing",
1236
+ "pretence": "pretense",
1237
+ "pretences": "pretenses",
1238
+ "primaeval": "primeval",
1239
+ "prioritisation": "prioritization",
1240
+ "prioritise": "prioritize",
1241
+ "prioritised": "prioritized",
1242
+ "prioritises": "prioritizes",
1243
+ "prioritising": "prioritizing",
1244
+ "privatisation": "privatization",
1245
+ "privatisations": "privatizations",
1246
+ "privatise": "privatize",
1247
+ "privatised": "privatized",
1248
+ "privatises": "privatizes",
1249
+ "privatising": "privatizing",
1250
+ "professionalisation": "professionalization",
1251
+ "professionalise": "professionalize",
1252
+ "professionalised": "professionalized",
1253
+ "professionalises": "professionalizes",
1254
+ "professionalising": "professionalizing",
1255
+ "programme": "program",
1256
+ "programmes": "programs",
1257
+ "prologue": "prolog",
1258
+ "prologues": "prologs",
1259
+ "propagandise": "propagandize",
1260
+ "propagandised": "propagandized",
1261
+ "propagandises": "propagandizes",
1262
+ "propagandising": "propagandizing",
1263
+ "proselytise": "proselytize",
1264
+ "proselytised": "proselytized",
1265
+ "proselytiser": "proselytizer",
1266
+ "proselytisers": "proselytizers",
1267
+ "proselytises": "proselytizes",
1268
+ "proselytising": "proselytizing",
1269
+ "psychoanalyse": "psychoanalyze",
1270
+ "psychoanalysed": "psychoanalyzed",
1271
+ "psychoanalyses": "psychoanalyzes",
1272
+ "psychoanalysing": "psychoanalyzing",
1273
+ "publicise": "publicize",
1274
+ "publicised": "publicized",
1275
+ "publicises": "publicizes",
1276
+ "publicising": "publicizing",
1277
+ "pulverisation": "pulverization",
1278
+ "pulverise": "pulverize",
1279
+ "pulverised": "pulverized",
1280
+ "pulverises": "pulverizes",
1281
+ "pulverising": "pulverizing",
1282
+ "pummelled": "pummel",
1283
+ "pummelling": "pummeled",
1284
+ "pyjama": "pajama",
1285
+ "pyjamas": "pajamas",
1286
+ "pzazz": "pizzazz",
1287
+ "quarrelled": "quarreled",
1288
+ "quarrelling": "quarreling",
1289
+ "radicalise": "radicalize",
1290
+ "radicalised": "radicalized",
1291
+ "radicalises": "radicalizes",
1292
+ "radicalising": "radicalizing",
1293
+ "rancour": "rancor",
1294
+ "randomise": "randomize",
1295
+ "randomised": "randomized",
1296
+ "randomises": "randomizes",
1297
+ "randomising": "randomizing",
1298
+ "rationalisation": "rationalization",
1299
+ "rationalisations": "rationalizations",
1300
+ "rationalise": "rationalize",
1301
+ "rationalised": "rationalized",
1302
+ "rationalises": "rationalizes",
1303
+ "rationalising": "rationalizing",
1304
+ "ravelled": "raveled",
1305
+ "ravelling": "raveling",
1306
+ "realisable": "realizable",
1307
+ "realisation": "realization",
1308
+ "realisations": "realizations",
1309
+ "realise": "realize",
1310
+ "realised": "realized",
1311
+ "realises": "realizes",
1312
+ "realising": "realizing",
1313
+ "recognisable": "recognizable",
1314
+ "recognisably": "recognizably",
1315
+ "recognisance": "recognizance",
1316
+ "recognise": "recognize",
1317
+ "recognised": "recognized",
1318
+ "recognises": "recognizes",
1319
+ "recognising": "recognizing",
1320
+ "reconnoitre": "reconnoiter",
1321
+ "reconnoitred": "reconnoitered",
1322
+ "reconnoitres": "reconnoiters",
1323
+ "reconnoitring": "reconnoitering",
1324
+ "refuelled": "refueled",
1325
+ "refuelling": "refueling",
1326
+ "regularisation": "regularization",
1327
+ "regularise": "regularize",
1328
+ "regularised": "regularized",
1329
+ "regularises": "regularizes",
1330
+ "regularising": "regularizing",
1331
+ "remodelled": "remodeled",
1332
+ "remodelling": "remodeling",
1333
+ "remould": "remold",
1334
+ "remoulded": "remolded",
1335
+ "remoulding": "remolding",
1336
+ "remoulds": "remolds",
1337
+ "reorganisation": "reorganization",
1338
+ "reorganisations": "reorganizations",
1339
+ "reorganise": "reorganize",
1340
+ "reorganised": "reorganized",
1341
+ "reorganises": "reorganizes",
1342
+ "reorganising": "reorganizing",
1343
+ "revelled": "reveled",
1344
+ "reveller": "reveler",
1345
+ "revellers": "revelers",
1346
+ "revelling": "reveling",
1347
+ "revitalise": "revitalize",
1348
+ "revitalised": "revitalized",
1349
+ "revitalises": "revitalizes",
1350
+ "revitalising": "revitalizing",
1351
+ "revolutionise": "revolutionize",
1352
+ "revolutionised": "revolutionized",
1353
+ "revolutionises": "revolutionizes",
1354
+ "revolutionising": "revolutionizing",
1355
+ "rhapsodise": "rhapsodize",
1356
+ "rhapsodised": "rhapsodized",
1357
+ "rhapsodises": "rhapsodizes",
1358
+ "rhapsodising": "rhapsodizing",
1359
+ "rigour": "rigor",
1360
+ "rigours": "rigors",
1361
+ "ritualised": "ritualized",
1362
+ "rivalled": "rivaled",
1363
+ "rivalling": "rivaling",
1364
+ "romanticise": "romanticize",
1365
+ "romanticised": "romanticized",
1366
+ "romanticises": "romanticizes",
1367
+ "romanticising": "romanticizing",
1368
+ "rumour": "rumor",
1369
+ "rumoured": "rumored",
1370
+ "rumours": "rumors",
1371
+ "sabre": "saber",
1372
+ "sabres": "sabers",
1373
+ "saltpetre": "saltpeter",
1374
+ "sanitise": "sanitize",
1375
+ "sanitised": "sanitized",
1376
+ "sanitises": "sanitizes",
1377
+ "sanitising": "sanitizing",
1378
+ "satirise": "satirize",
1379
+ "satirised": "satirized",
1380
+ "satirises": "satirizes",
1381
+ "satirising": "satirizing",
1382
+ "saviour": "savior",
1383
+ "saviours": "saviors",
1384
+ "savour": "savor",
1385
+ "savoured": "savored",
1386
+ "savouries": "savories",
1387
+ "savouring": "savoring",
1388
+ "savours": "savors",
1389
+ "savoury": "savory",
1390
+ "scandalise": "scandalize",
1391
+ "scandalised": "scandalized",
1392
+ "scandalises": "scandalizes",
1393
+ "scandalising": "scandalizing",
1394
+ "sceptic": "skeptic",
1395
+ "sceptical": "skeptical",
1396
+ "sceptically": "skeptically",
1397
+ "scepticism": "skepticism",
1398
+ "sceptics": "skeptics",
1399
+ "sceptre": "scepter",
1400
+ "sceptres": "scepters",
1401
+ "scrutinise": "scrutinize",
1402
+ "scrutinised": "scrutinized",
1403
+ "scrutinises": "scrutinizes",
1404
+ "scrutinising": "scrutinizing",
1405
+ "secularisation": "secularization",
1406
+ "secularise": "secularize",
1407
+ "secularised": "secularized",
1408
+ "secularises": "secularizes",
1409
+ "secularising": "secularizing",
1410
+ "sensationalise": "sensationalize",
1411
+ "sensationalised": "sensationalized",
1412
+ "sensationalises": "sensationalizes",
1413
+ "sensationalising": "sensationalizing",
1414
+ "sensitise": "sensitize",
1415
+ "sensitised": "sensitized",
1416
+ "sensitises": "sensitizes",
1417
+ "sensitising": "sensitizing",
1418
+ "sentimentalise": "sentimentalize",
1419
+ "sentimentalised": "sentimentalized",
1420
+ "sentimentalises": "sentimentalizes",
1421
+ "sentimentalising": "sentimentalizing",
1422
+ "sepulchre": "sepulcher",
1423
+ "sepulchres": "sepulchers",
1424
+ "serialisation": "serialization",
1425
+ "serialisations": "serializations",
1426
+ "serialise": "serialize",
1427
+ "serialised": "serialized",
1428
+ "serialises": "serializes",
1429
+ "serialising": "serializing",
1430
+ "sermonise": "sermonize",
1431
+ "sermonised": "sermonized",
1432
+ "sermonises": "sermonizes",
1433
+ "sermonising": "sermonizing",
1434
+ "sheikh": "sheik",
1435
+ "shovelled": "shoveled",
1436
+ "shovelling": "shoveling",
1437
+ "shrivelled": "shriveled",
1438
+ "shrivelling": "shriveling",
1439
+ "signalise": "signalize",
1440
+ "signalised": "signalized",
1441
+ "signalises": "signalizes",
1442
+ "signalising": "signalizing",
1443
+ "signalled": "signaled",
1444
+ "signalling": "signaling",
1445
+ "smoulder": "smolder",
1446
+ "smouldered": "smoldered",
1447
+ "smouldering": "smoldering",
1448
+ "smoulders": "smolders",
1449
+ "snivelled": "sniveled",
1450
+ "snivelling": "sniveling",
1451
+ "snorkelled": "snorkeled",
1452
+ "snorkelling": "snorkeling",
1453
+ "snowplough": "snowplow",
1454
+ "snowploughs": "snowplow",
1455
+ "socialisation": "socialization",
1456
+ "socialise": "socialize",
1457
+ "socialised": "socialized",
1458
+ "socialises": "socializes",
1459
+ "socialising": "socializing",
1460
+ "sodomise": "sodomize",
1461
+ "sodomised": "sodomized",
1462
+ "sodomises": "sodomizes",
1463
+ "sodomising": "sodomizing",
1464
+ "solemnise": "solemnize",
1465
+ "solemnised": "solemnized",
1466
+ "solemnises": "solemnizes",
1467
+ "solemnising": "solemnizing",
1468
+ "sombre": "somber",
1469
+ "specialisation": "specialization",
1470
+ "specialisations": "specializations",
1471
+ "specialise": "specialize",
1472
+ "specialised": "specialized",
1473
+ "specialises": "specializes",
1474
+ "specialising": "specializing",
1475
+ "spectre": "specter",
1476
+ "spectres": "specters",
1477
+ "spiralled": "spiraled",
1478
+ "spiralling": "spiraling",
1479
+ "splendour": "splendor",
1480
+ "splendours": "splendors",
1481
+ "squirrelled": "squirreled",
1482
+ "squirrelling": "squirreling",
1483
+ "stabilisation": "stabilization",
1484
+ "stabilise": "stabilize",
1485
+ "stabilised": "stabilized",
1486
+ "stabiliser": "stabilizer",
1487
+ "stabilisers": "stabilizers",
1488
+ "stabilises": "stabilizes",
1489
+ "stabilising": "stabilizing",
1490
+ "standardisation": "standardization",
1491
+ "standardise": "standardize",
1492
+ "standardised": "standardized",
1493
+ "standardises": "standardizes",
1494
+ "standardising": "standardizing",
1495
+ "stencilled": "stenciled",
1496
+ "stencilling": "stenciling",
1497
+ "sterilisation": "sterilization",
1498
+ "sterilisations": "sterilizations",
1499
+ "sterilise": "sterilize",
1500
+ "sterilised": "sterilized",
1501
+ "steriliser": "sterilizer",
1502
+ "sterilisers": "sterilizers",
1503
+ "sterilises": "sterilizes",
1504
+ "sterilising": "sterilizing",
1505
+ "stigmatisation": "stigmatization",
1506
+ "stigmatise": "stigmatize",
1507
+ "stigmatised": "stigmatized",
1508
+ "stigmatises": "stigmatizes",
1509
+ "stigmatising": "stigmatizing",
1510
+ "storey": "story",
1511
+ "storeys": "stories",
1512
+ "subsidisation": "subsidization",
1513
+ "subsidise": "subsidize",
1514
+ "subsidised": "subsidized",
1515
+ "subsidiser": "subsidizer",
1516
+ "subsidisers": "subsidizers",
1517
+ "subsidises": "subsidizes",
1518
+ "subsidising": "subsidizing",
1519
+ "succour": "succor",
1520
+ "succoured": "succored",
1521
+ "succouring": "succoring",
1522
+ "succours": "succors",
1523
+ "sulphate": "sulfate",
1524
+ "sulphates": "sulfates",
1525
+ "sulphide": "sulfide",
1526
+ "sulphides": "sulfides",
1527
+ "sulphur": "sulfur",
1528
+ "sulphurous": "sulfurous",
1529
+ "summarise": "summarize",
1530
+ "summarised": "summarized",
1531
+ "summarises": "summarizes",
1532
+ "summarising": "summarizing",
1533
+ "swivelled": "swiveled",
1534
+ "swivelling": "swiveling",
1535
+ "symbolise": "symbolize",
1536
+ "symbolised": "symbolized",
1537
+ "symbolises": "symbolizes",
1538
+ "symbolising": "symbolizing",
1539
+ "sympathise": "sympathize",
1540
+ "sympathised": "sympathized",
1541
+ "sympathiser": "sympathizer",
1542
+ "sympathisers": "sympathizers",
1543
+ "sympathises": "sympathizes",
1544
+ "sympathising": "sympathizing",
1545
+ "synchronisation": "synchronization",
1546
+ "synchronise": "synchronize",
1547
+ "synchronised": "synchronized",
1548
+ "synchronises": "synchronizes",
1549
+ "synchronising": "synchronizing",
1550
+ "synthesise": "synthesize",
1551
+ "synthesised": "synthesized",
1552
+ "synthesiser": "synthesizer",
1553
+ "synthesisers": "synthesizers",
1554
+ "synthesises": "synthesizes",
1555
+ "synthesising": "synthesizing",
1556
+ "syphon": "siphon",
1557
+ "syphoned": "siphoned",
1558
+ "syphoning": "siphoning",
1559
+ "syphons": "siphons",
1560
+ "systematisation": "systematization",
1561
+ "systematise": "systematize",
1562
+ "systematised": "systematized",
1563
+ "systematises": "systematizes",
1564
+ "systematising": "systematizing",
1565
+ "tantalise": "tantalize",
1566
+ "tantalised": "tantalized",
1567
+ "tantalises": "tantalizes",
1568
+ "tantalising": "tantalizing",
1569
+ "tantalisingly": "tantalizingly",
1570
+ "tasselled": "tasseled",
1571
+ "technicolour": "technicolor",
1572
+ "temporise": "temporize",
1573
+ "temporised": "temporized",
1574
+ "temporises": "temporizes",
1575
+ "temporising": "temporizing",
1576
+ "tenderise": "tenderize",
1577
+ "tenderised": "tenderized",
1578
+ "tenderises": "tenderizes",
1579
+ "tenderising": "tenderizing",
1580
+ "terrorise": "terrorize",
1581
+ "terrorised": "terrorized",
1582
+ "terrorises": "terrorizes",
1583
+ "terrorising": "terrorizing",
1584
+ "theatre": "theater",
1585
+ "theatregoer": "theatergoer",
1586
+ "theatregoers": "theatergoers",
1587
+ "theatres": "theaters",
1588
+ "theorise": "theorize",
1589
+ "theorised": "theorized",
1590
+ "theorises": "theorizes",
1591
+ "theorising": "theorizing",
1592
+ "tonne": "ton",
1593
+ "tonnes": "tons",
1594
+ "towelled": "toweled",
1595
+ "towelling": "toweling",
1596
+ "toxaemia": "toxemia",
1597
+ "tranquillise": "tranquilize",
1598
+ "tranquillised": "tranquilized",
1599
+ "tranquilliser": "tranquilizer",
1600
+ "tranquillisers": "tranquilizers",
1601
+ "tranquillises": "tranquilizes",
1602
+ "tranquillising": "tranquilizing",
1603
+ "tranquillity": "tranquility",
1604
+ "tranquillize": "tranquilize",
1605
+ "tranquillized": "tranquilized",
1606
+ "tranquillizer": "tranquilizer",
1607
+ "tranquillizers": "tranquilizers",
1608
+ "tranquillizes": "tranquilizes",
1609
+ "tranquillizing": "tranquilizing",
1610
+ "tranquilly": "tranquility",
1611
+ "transistorised": "transistorized",
1612
+ "traumatise": "traumatize",
1613
+ "traumatised": "traumatized",
1614
+ "traumatises": "traumatizes",
1615
+ "traumatising": "traumatizing",
1616
+ "travelled": "traveled",
1617
+ "traveller": "traveler",
1618
+ "travellers": "travelers",
1619
+ "travelling": "traveling",
1620
+ "travelog": "travelogue",
1621
+ "travelogs": "travelogues",
1622
+ "trialled": "trialed",
1623
+ "trialling": "trialing",
1624
+ "tricolour": "tricolor",
1625
+ "tricolours": "tricolors",
1626
+ "trivialise": "trivialize",
1627
+ "trivialised": "trivialized",
1628
+ "trivialises": "trivializes",
1629
+ "trivialising": "trivializing",
1630
+ "tumour": "tumor",
1631
+ "tumours": "tumors",
1632
+ "tunnelled": "tunneled",
1633
+ "tunnelling": "tunneling",
1634
+ "tyrannise": "tyrannize",
1635
+ "tyrannised": "tyrannized",
1636
+ "tyrannises": "tyrannizes",
1637
+ "tyrannising": "tyrannizing",
1638
+ "tyre": "tire",
1639
+ "tyres": "tires",
1640
+ "unauthorised": "unauthorized",
1641
+ "uncivilised": "uncivilized",
1642
+ "underutilised": "underutilized",
1643
+ "unequalled": "unequaled",
1644
+ "unfavourable": "unfavorable",
1645
+ "unfavourably": "unfavorably",
1646
+ "unionisation": "unionization",
1647
+ "unionise": "unionize",
1648
+ "unionised": "unionized",
1649
+ "unionises": "unionizes",
1650
+ "unionising": "unionizing",
1651
+ "unorganised": "unorganized",
1652
+ "unravelled": "unraveled",
1653
+ "unravelling": "unraveling",
1654
+ "unrecognisable": "unrecognizable",
1655
+ "unrecognised": "unrecognized",
1656
+ "unrivalled": "unrivaled",
1657
+ "unsavoury": "unsavory",
1658
+ "untrammelled": "untrammeled",
1659
+ "urbanisation": "urbanization",
1660
+ "urbanise": "urbanize",
1661
+ "urbanised": "urbanized",
1662
+ "urbanises": "urbanizes",
1663
+ "urbanising": "urbanizing",
1664
+ "utilisable": "utilizable",
1665
+ "utilisation": "utilization",
1666
+ "utilise": "utilize",
1667
+ "utilised": "utilized",
1668
+ "utilises": "utilizes",
1669
+ "utilising": "utilizing",
1670
+ "valour": "valor",
1671
+ "vandalise": "vandalize",
1672
+ "vandalised": "vandalized",
1673
+ "vandalises": "vandalizes",
1674
+ "vandalising": "vandalizing",
1675
+ "vaporisation": "vaporization",
1676
+ "vaporise": "vaporize",
1677
+ "vaporised": "vaporized",
1678
+ "vaporises": "vaporizes",
1679
+ "vaporising": "vaporizing",
1680
+ "vapour": "vapor",
1681
+ "vapours": "vapors",
1682
+ "verbalise": "verbalize",
1683
+ "verbalised": "verbalized",
1684
+ "verbalises": "verbalizes",
1685
+ "verbalising": "verbalizing",
1686
+ "victimisation": "victimization",
1687
+ "victimise": "victimize",
1688
+ "victimised": "victimized",
1689
+ "victimises": "victimizes",
1690
+ "victimising": "victimizing",
1691
+ "videodisc": "videodisk",
1692
+ "videodiscs": "videodisks",
1693
+ "vigour": "vigor",
1694
+ "visualisation": "visualization",
1695
+ "visualisations": "visualizations",
1696
+ "visualise": "visualize",
1697
+ "visualised": "visualized",
1698
+ "visualises": "visualizes",
1699
+ "visualising": "visualizing",
1700
+ "vocalisation": "vocalization",
1701
+ "vocalisations": "vocalizations",
1702
+ "vocalise": "vocalize",
1703
+ "vocalised": "vocalized",
1704
+ "vocalises": "vocalizes",
1705
+ "vocalising": "vocalizing",
1706
+ "vulcanised": "vulcanized",
1707
+ "vulgarisation": "vulgarization",
1708
+ "vulgarise": "vulgarize",
1709
+ "vulgarised": "vulgarized",
1710
+ "vulgarises": "vulgarizes",
1711
+ "vulgarising": "vulgarizing",
1712
+ "waggon": "wagon",
1713
+ "waggons": "wagons",
1714
+ "watercolour": "watercolor",
1715
+ "watercolours": "watercolors",
1716
+ "weaselled": "weaseled",
1717
+ "weaselling": "weaseling",
1718
+ "westernisation": "westernization",
1719
+ "westernise": "westernize",
1720
+ "westernised": "westernized",
1721
+ "westernises": "westernizes",
1722
+ "westernising": "westernizing",
1723
+ "womanise": "womanize",
1724
+ "womanised": "womanized",
1725
+ "womaniser": "womanizer",
1726
+ "womanisers": "womanizers",
1727
+ "womanises": "womanizes",
1728
+ "womanising": "womanizing",
1729
+ "woollen": "woolen",
1730
+ "woollens": "woolens",
1731
+ "woollies": "woolies",
1732
+ "woolly": "wooly",
1733
+ "worshipped": "worshiped",
1734
+ "worshipper": "worshiper",
1735
+ "worshipping": "worshiping",
1736
+ "yodelled": "yodeled",
1737
+ "yodelling": "yodeling",
1738
+ "yoghourt": "yogurt",
1739
+ "yoghourts": "yogurts",
1740
+ "yoghurt": "yogurt",
1741
+ "yoghurts": "yogurts"
1742
+ }
audio_model_states/whisper-small-cantonese/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52d3d75f188b0e83816ddfe2c081aa5811fef9f12f90bcf95efbb6f22d4cb03a
3
+ size 1925064044
audio_model_states/whisper-small-cantonese/preprocessor_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "chunk_length": 30,
3
+ "feature_extractor_type": "WhisperFeatureExtractor",
4
+ "feature_size": 80,
5
+ "hop_length": 160,
6
+ "n_fft": 400,
7
+ "n_samples": 480000,
8
+ "nb_max_frames": 3000,
9
+ "padding_side": "right",
10
+ "padding_value": 0.0,
11
+ "processor_class": "WhisperProcessor",
12
+ "return_attention_mask": false,
13
+ "sampling_rate": 16000
14
+ }
audio_model_states/whisper-small-cantonese/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9563aaf2c8c73c9368146c47fb8767f7b2fc4ec19fa3d2f43de391939240e744
3
+ size 967102601
audio_model_states/whisper-small-cantonese/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5debd200034ecb7c5a4f41ef2227951c0f473e8fae7113e0066142d213a47d2d
3
+ size 14244
audio_model_states/whisper-small-cantonese/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b691e5fdc050ffec60fabc6f70ec25a5c9c5a415de9ceda0c71eee2c5823fe64
3
+ size 557
audio_model_states/whisper-small-cantonese/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:554d2698384ad39f5da6b204240a25383b7dfb17d0add781017660bb036bc2f3
3
+ size 1064
audio_model_states/whisper-small-cantonese/special_tokens_map.json ADDED
@@ -0,0 +1,139 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|endoftext|>",
4
+ "<|startoftranscript|>",
5
+ "<|en|>",
6
+ "<|zh|>",
7
+ "<|de|>",
8
+ "<|es|>",
9
+ "<|ru|>",
10
+ "<|ko|>",
11
+ "<|fr|>",
12
+ "<|ja|>",
13
+ "<|pt|>",
14
+ "<|tr|>",
15
+ "<|pl|>",
16
+ "<|ca|>",
17
+ "<|nl|>",
18
+ "<|ar|>",
19
+ "<|sv|>",
20
+ "<|it|>",
21
+ "<|id|>",
22
+ "<|hi|>",
23
+ "<|fi|>",
24
+ "<|vi|>",
25
+ "<|iw|>",
26
+ "<|uk|>",
27
+ "<|el|>",
28
+ "<|ms|>",
29
+ "<|cs|>",
30
+ "<|ro|>",
31
+ "<|da|>",
32
+ "<|hu|>",
33
+ "<|ta|>",
34
+ "<|no|>",
35
+ "<|th|>",
36
+ "<|ur|>",
37
+ "<|hr|>",
38
+ "<|bg|>",
39
+ "<|lt|>",
40
+ "<|la|>",
41
+ "<|mi|>",
42
+ "<|ml|>",
43
+ "<|cy|>",
44
+ "<|sk|>",
45
+ "<|te|>",
46
+ "<|fa|>",
47
+ "<|lv|>",
48
+ "<|bn|>",
49
+ "<|sr|>",
50
+ "<|az|>",
51
+ "<|sl|>",
52
+ "<|kn|>",
53
+ "<|et|>",
54
+ "<|mk|>",
55
+ "<|br|>",
56
+ "<|eu|>",
57
+ "<|is|>",
58
+ "<|hy|>",
59
+ "<|ne|>",
60
+ "<|mn|>",
61
+ "<|bs|>",
62
+ "<|kk|>",
63
+ "<|sq|>",
64
+ "<|sw|>",
65
+ "<|gl|>",
66
+ "<|mr|>",
67
+ "<|pa|>",
68
+ "<|si|>",
69
+ "<|km|>",
70
+ "<|sn|>",
71
+ "<|yo|>",
72
+ "<|so|>",
73
+ "<|af|>",
74
+ "<|oc|>",
75
+ "<|ka|>",
76
+ "<|be|>",
77
+ "<|tg|>",
78
+ "<|sd|>",
79
+ "<|gu|>",
80
+ "<|am|>",
81
+ "<|yi|>",
82
+ "<|lo|>",
83
+ "<|uz|>",
84
+ "<|fo|>",
85
+ "<|ht|>",
86
+ "<|ps|>",
87
+ "<|tk|>",
88
+ "<|nn|>",
89
+ "<|mt|>",
90
+ "<|sa|>",
91
+ "<|lb|>",
92
+ "<|my|>",
93
+ "<|bo|>",
94
+ "<|tl|>",
95
+ "<|mg|>",
96
+ "<|as|>",
97
+ "<|tt|>",
98
+ "<|haw|>",
99
+ "<|ln|>",
100
+ "<|ha|>",
101
+ "<|ba|>",
102
+ "<|jw|>",
103
+ "<|su|>",
104
+ "<|translate|>",
105
+ "<|transcribe|>",
106
+ "<|startoflm|>",
107
+ "<|startofprev|>",
108
+ "<|nocaptions|>",
109
+ "<|notimestamps|>"
110
+ ],
111
+ "bos_token": {
112
+ "content": "<|endoftext|>",
113
+ "lstrip": false,
114
+ "normalized": true,
115
+ "rstrip": false,
116
+ "single_word": false
117
+ },
118
+ "eos_token": {
119
+ "content": "<|endoftext|>",
120
+ "lstrip": false,
121
+ "normalized": true,
122
+ "rstrip": false,
123
+ "single_word": false
124
+ },
125
+ "pad_token": {
126
+ "content": "<|endoftext|>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false
131
+ },
132
+ "unk_token": {
133
+ "content": "",
134
+ "lstrip": false,
135
+ "normalized": true,
136
+ "rstrip": false,
137
+ "single_word": false
138
+ }
139
+ }
audio_model_states/whisper-small-cantonese/tokenizer_config.json ADDED
@@ -0,0 +1,981 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "50257": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "50258": {
14
+ "content": "<|startoftranscript|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "50259": {
22
+ "content": "<|en|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "50260": {
30
+ "content": "<|zh|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "50261": {
38
+ "content": "<|de|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "50262": {
46
+ "content": "<|es|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "50263": {
54
+ "content": "<|ru|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "50264": {
62
+ "content": "<|ko|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "50265": {
70
+ "content": "<|fr|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "50266": {
78
+ "content": "<|ja|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "50267": {
86
+ "content": "<|pt|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "50268": {
94
+ "content": "<|tr|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "50269": {
102
+ "content": "<|pl|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "50270": {
110
+ "content": "<|ca|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "50271": {
118
+ "content": "<|nl|>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": true
124
+ },
125
+ "50272": {
126
+ "content": "<|ar|>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": true
132
+ },
133
+ "50273": {
134
+ "content": "<|sv|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": true
140
+ },
141
+ "50274": {
142
+ "content": "<|it|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": true
148
+ },
149
+ "50275": {
150
+ "content": "<|id|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": true
156
+ },
157
+ "50276": {
158
+ "content": "<|hi|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": true
164
+ },
165
+ "50277": {
166
+ "content": "<|fi|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": true
172
+ },
173
+ "50278": {
174
+ "content": "<|vi|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": true
180
+ },
181
+ "50279": {
182
+ "content": "<|iw|>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": true
188
+ },
189
+ "50280": {
190
+ "content": "<|uk|>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": true
196
+ },
197
+ "50281": {
198
+ "content": "<|el|>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": true
204
+ },
205
+ "50282": {
206
+ "content": "<|ms|>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": true
212
+ },
213
+ "50283": {
214
+ "content": "<|cs|>",
215
+ "lstrip": false,
216
+ "normalized": false,
217
+ "rstrip": false,
218
+ "single_word": false,
219
+ "special": true
220
+ },
221
+ "50284": {
222
+ "content": "<|ro|>",
223
+ "lstrip": false,
224
+ "normalized": false,
225
+ "rstrip": false,
226
+ "single_word": false,
227
+ "special": true
228
+ },
229
+ "50285": {
230
+ "content": "<|da|>",
231
+ "lstrip": false,
232
+ "normalized": false,
233
+ "rstrip": false,
234
+ "single_word": false,
235
+ "special": true
236
+ },
237
+ "50286": {
238
+ "content": "<|hu|>",
239
+ "lstrip": false,
240
+ "normalized": false,
241
+ "rstrip": false,
242
+ "single_word": false,
243
+ "special": true
244
+ },
245
+ "50287": {
246
+ "content": "<|ta|>",
247
+ "lstrip": false,
248
+ "normalized": false,
249
+ "rstrip": false,
250
+ "single_word": false,
251
+ "special": true
252
+ },
253
+ "50288": {
254
+ "content": "<|no|>",
255
+ "lstrip": false,
256
+ "normalized": false,
257
+ "rstrip": false,
258
+ "single_word": false,
259
+ "special": true
260
+ },
261
+ "50289": {
262
+ "content": "<|th|>",
263
+ "lstrip": false,
264
+ "normalized": false,
265
+ "rstrip": false,
266
+ "single_word": false,
267
+ "special": true
268
+ },
269
+ "50290": {
270
+ "content": "<|ur|>",
271
+ "lstrip": false,
272
+ "normalized": false,
273
+ "rstrip": false,
274
+ "single_word": false,
275
+ "special": true
276
+ },
277
+ "50291": {
278
+ "content": "<|hr|>",
279
+ "lstrip": false,
280
+ "normalized": false,
281
+ "rstrip": false,
282
+ "single_word": false,
283
+ "special": true
284
+ },
285
+ "50292": {
286
+ "content": "<|bg|>",
287
+ "lstrip": false,
288
+ "normalized": false,
289
+ "rstrip": false,
290
+ "single_word": false,
291
+ "special": true
292
+ },
293
+ "50293": {
294
+ "content": "<|lt|>",
295
+ "lstrip": false,
296
+ "normalized": false,
297
+ "rstrip": false,
298
+ "single_word": false,
299
+ "special": true
300
+ },
301
+ "50294": {
302
+ "content": "<|la|>",
303
+ "lstrip": false,
304
+ "normalized": false,
305
+ "rstrip": false,
306
+ "single_word": false,
307
+ "special": true
308
+ },
309
+ "50295": {
310
+ "content": "<|mi|>",
311
+ "lstrip": false,
312
+ "normalized": false,
313
+ "rstrip": false,
314
+ "single_word": false,
315
+ "special": true
316
+ },
317
+ "50296": {
318
+ "content": "<|ml|>",
319
+ "lstrip": false,
320
+ "normalized": false,
321
+ "rstrip": false,
322
+ "single_word": false,
323
+ "special": true
324
+ },
325
+ "50297": {
326
+ "content": "<|cy|>",
327
+ "lstrip": false,
328
+ "normalized": false,
329
+ "rstrip": false,
330
+ "single_word": false,
331
+ "special": true
332
+ },
333
+ "50298": {
334
+ "content": "<|sk|>",
335
+ "lstrip": false,
336
+ "normalized": false,
337
+ "rstrip": false,
338
+ "single_word": false,
339
+ "special": true
340
+ },
341
+ "50299": {
342
+ "content": "<|te|>",
343
+ "lstrip": false,
344
+ "normalized": false,
345
+ "rstrip": false,
346
+ "single_word": false,
347
+ "special": true
348
+ },
349
+ "50300": {
350
+ "content": "<|fa|>",
351
+ "lstrip": false,
352
+ "normalized": false,
353
+ "rstrip": false,
354
+ "single_word": false,
355
+ "special": true
356
+ },
357
+ "50301": {
358
+ "content": "<|lv|>",
359
+ "lstrip": false,
360
+ "normalized": false,
361
+ "rstrip": false,
362
+ "single_word": false,
363
+ "special": true
364
+ },
365
+ "50302": {
366
+ "content": "<|bn|>",
367
+ "lstrip": false,
368
+ "normalized": false,
369
+ "rstrip": false,
370
+ "single_word": false,
371
+ "special": true
372
+ },
373
+ "50303": {
374
+ "content": "<|sr|>",
375
+ "lstrip": false,
376
+ "normalized": false,
377
+ "rstrip": false,
378
+ "single_word": false,
379
+ "special": true
380
+ },
381
+ "50304": {
382
+ "content": "<|az|>",
383
+ "lstrip": false,
384
+ "normalized": false,
385
+ "rstrip": false,
386
+ "single_word": false,
387
+ "special": true
388
+ },
389
+ "50305": {
390
+ "content": "<|sl|>",
391
+ "lstrip": false,
392
+ "normalized": false,
393
+ "rstrip": false,
394
+ "single_word": false,
395
+ "special": true
396
+ },
397
+ "50306": {
398
+ "content": "<|kn|>",
399
+ "lstrip": false,
400
+ "normalized": false,
401
+ "rstrip": false,
402
+ "single_word": false,
403
+ "special": true
404
+ },
405
+ "50307": {
406
+ "content": "<|et|>",
407
+ "lstrip": false,
408
+ "normalized": false,
409
+ "rstrip": false,
410
+ "single_word": false,
411
+ "special": true
412
+ },
413
+ "50308": {
414
+ "content": "<|mk|>",
415
+ "lstrip": false,
416
+ "normalized": false,
417
+ "rstrip": false,
418
+ "single_word": false,
419
+ "special": true
420
+ },
421
+ "50309": {
422
+ "content": "<|br|>",
423
+ "lstrip": false,
424
+ "normalized": false,
425
+ "rstrip": false,
426
+ "single_word": false,
427
+ "special": true
428
+ },
429
+ "50310": {
430
+ "content": "<|eu|>",
431
+ "lstrip": false,
432
+ "normalized": false,
433
+ "rstrip": false,
434
+ "single_word": false,
435
+ "special": true
436
+ },
437
+ "50311": {
438
+ "content": "<|is|>",
439
+ "lstrip": false,
440
+ "normalized": false,
441
+ "rstrip": false,
442
+ "single_word": false,
443
+ "special": true
444
+ },
445
+ "50312": {
446
+ "content": "<|hy|>",
447
+ "lstrip": false,
448
+ "normalized": false,
449
+ "rstrip": false,
450
+ "single_word": false,
451
+ "special": true
452
+ },
453
+ "50313": {
454
+ "content": "<|ne|>",
455
+ "lstrip": false,
456
+ "normalized": false,
457
+ "rstrip": false,
458
+ "single_word": false,
459
+ "special": true
460
+ },
461
+ "50314": {
462
+ "content": "<|mn|>",
463
+ "lstrip": false,
464
+ "normalized": false,
465
+ "rstrip": false,
466
+ "single_word": false,
467
+ "special": true
468
+ },
469
+ "50315": {
470
+ "content": "<|bs|>",
471
+ "lstrip": false,
472
+ "normalized": false,
473
+ "rstrip": false,
474
+ "single_word": false,
475
+ "special": true
476
+ },
477
+ "50316": {
478
+ "content": "<|kk|>",
479
+ "lstrip": false,
480
+ "normalized": false,
481
+ "rstrip": false,
482
+ "single_word": false,
483
+ "special": true
484
+ },
485
+ "50317": {
486
+ "content": "<|sq|>",
487
+ "lstrip": false,
488
+ "normalized": false,
489
+ "rstrip": false,
490
+ "single_word": false,
491
+ "special": true
492
+ },
493
+ "50318": {
494
+ "content": "<|sw|>",
495
+ "lstrip": false,
496
+ "normalized": false,
497
+ "rstrip": false,
498
+ "single_word": false,
499
+ "special": true
500
+ },
501
+ "50319": {
502
+ "content": "<|gl|>",
503
+ "lstrip": false,
504
+ "normalized": false,
505
+ "rstrip": false,
506
+ "single_word": false,
507
+ "special": true
508
+ },
509
+ "50320": {
510
+ "content": "<|mr|>",
511
+ "lstrip": false,
512
+ "normalized": false,
513
+ "rstrip": false,
514
+ "single_word": false,
515
+ "special": true
516
+ },
517
+ "50321": {
518
+ "content": "<|pa|>",
519
+ "lstrip": false,
520
+ "normalized": false,
521
+ "rstrip": false,
522
+ "single_word": false,
523
+ "special": true
524
+ },
525
+ "50322": {
526
+ "content": "<|si|>",
527
+ "lstrip": false,
528
+ "normalized": false,
529
+ "rstrip": false,
530
+ "single_word": false,
531
+ "special": true
532
+ },
533
+ "50323": {
534
+ "content": "<|km|>",
535
+ "lstrip": false,
536
+ "normalized": false,
537
+ "rstrip": false,
538
+ "single_word": false,
539
+ "special": true
540
+ },
541
+ "50324": {
542
+ "content": "<|sn|>",
543
+ "lstrip": false,
544
+ "normalized": false,
545
+ "rstrip": false,
546
+ "single_word": false,
547
+ "special": true
548
+ },
549
+ "50325": {
550
+ "content": "<|yo|>",
551
+ "lstrip": false,
552
+ "normalized": false,
553
+ "rstrip": false,
554
+ "single_word": false,
555
+ "special": true
556
+ },
557
+ "50326": {
558
+ "content": "<|so|>",
559
+ "lstrip": false,
560
+ "normalized": false,
561
+ "rstrip": false,
562
+ "single_word": false,
563
+ "special": true
564
+ },
565
+ "50327": {
566
+ "content": "<|af|>",
567
+ "lstrip": false,
568
+ "normalized": false,
569
+ "rstrip": false,
570
+ "single_word": false,
571
+ "special": true
572
+ },
573
+ "50328": {
574
+ "content": "<|oc|>",
575
+ "lstrip": false,
576
+ "normalized": false,
577
+ "rstrip": false,
578
+ "single_word": false,
579
+ "special": true
580
+ },
581
+ "50329": {
582
+ "content": "<|ka|>",
583
+ "lstrip": false,
584
+ "normalized": false,
585
+ "rstrip": false,
586
+ "single_word": false,
587
+ "special": true
588
+ },
589
+ "50330": {
590
+ "content": "<|be|>",
591
+ "lstrip": false,
592
+ "normalized": false,
593
+ "rstrip": false,
594
+ "single_word": false,
595
+ "special": true
596
+ },
597
+ "50331": {
598
+ "content": "<|tg|>",
599
+ "lstrip": false,
600
+ "normalized": false,
601
+ "rstrip": false,
602
+ "single_word": false,
603
+ "special": true
604
+ },
605
+ "50332": {
606
+ "content": "<|sd|>",
607
+ "lstrip": false,
608
+ "normalized": false,
609
+ "rstrip": false,
610
+ "single_word": false,
611
+ "special": true
612
+ },
613
+ "50333": {
614
+ "content": "<|gu|>",
615
+ "lstrip": false,
616
+ "normalized": false,
617
+ "rstrip": false,
618
+ "single_word": false,
619
+ "special": true
620
+ },
621
+ "50334": {
622
+ "content": "<|am|>",
623
+ "lstrip": false,
624
+ "normalized": false,
625
+ "rstrip": false,
626
+ "single_word": false,
627
+ "special": true
628
+ },
629
+ "50335": {
630
+ "content": "<|yi|>",
631
+ "lstrip": false,
632
+ "normalized": false,
633
+ "rstrip": false,
634
+ "single_word": false,
635
+ "special": true
636
+ },
637
+ "50336": {
638
+ "content": "<|lo|>",
639
+ "lstrip": false,
640
+ "normalized": false,
641
+ "rstrip": false,
642
+ "single_word": false,
643
+ "special": true
644
+ },
645
+ "50337": {
646
+ "content": "<|uz|>",
647
+ "lstrip": false,
648
+ "normalized": false,
649
+ "rstrip": false,
650
+ "single_word": false,
651
+ "special": true
652
+ },
653
+ "50338": {
654
+ "content": "<|fo|>",
655
+ "lstrip": false,
656
+ "normalized": false,
657
+ "rstrip": false,
658
+ "single_word": false,
659
+ "special": true
660
+ },
661
+ "50339": {
662
+ "content": "<|ht|>",
663
+ "lstrip": false,
664
+ "normalized": false,
665
+ "rstrip": false,
666
+ "single_word": false,
667
+ "special": true
668
+ },
669
+ "50340": {
670
+ "content": "<|ps|>",
671
+ "lstrip": false,
672
+ "normalized": false,
673
+ "rstrip": false,
674
+ "single_word": false,
675
+ "special": true
676
+ },
677
+ "50341": {
678
+ "content": "<|tk|>",
679
+ "lstrip": false,
680
+ "normalized": false,
681
+ "rstrip": false,
682
+ "single_word": false,
683
+ "special": true
684
+ },
685
+ "50342": {
686
+ "content": "<|nn|>",
687
+ "lstrip": false,
688
+ "normalized": false,
689
+ "rstrip": false,
690
+ "single_word": false,
691
+ "special": true
692
+ },
693
+ "50343": {
694
+ "content": "<|mt|>",
695
+ "lstrip": false,
696
+ "normalized": false,
697
+ "rstrip": false,
698
+ "single_word": false,
699
+ "special": true
700
+ },
701
+ "50344": {
702
+ "content": "<|sa|>",
703
+ "lstrip": false,
704
+ "normalized": false,
705
+ "rstrip": false,
706
+ "single_word": false,
707
+ "special": true
708
+ },
709
+ "50345": {
710
+ "content": "<|lb|>",
711
+ "lstrip": false,
712
+ "normalized": false,
713
+ "rstrip": false,
714
+ "single_word": false,
715
+ "special": true
716
+ },
717
+ "50346": {
718
+ "content": "<|my|>",
719
+ "lstrip": false,
720
+ "normalized": false,
721
+ "rstrip": false,
722
+ "single_word": false,
723
+ "special": true
724
+ },
725
+ "50347": {
726
+ "content": "<|bo|>",
727
+ "lstrip": false,
728
+ "normalized": false,
729
+ "rstrip": false,
730
+ "single_word": false,
731
+ "special": true
732
+ },
733
+ "50348": {
734
+ "content": "<|tl|>",
735
+ "lstrip": false,
736
+ "normalized": false,
737
+ "rstrip": false,
738
+ "single_word": false,
739
+ "special": true
740
+ },
741
+ "50349": {
742
+ "content": "<|mg|>",
743
+ "lstrip": false,
744
+ "normalized": false,
745
+ "rstrip": false,
746
+ "single_word": false,
747
+ "special": true
748
+ },
749
+ "50350": {
750
+ "content": "<|as|>",
751
+ "lstrip": false,
752
+ "normalized": false,
753
+ "rstrip": false,
754
+ "single_word": false,
755
+ "special": true
756
+ },
757
+ "50351": {
758
+ "content": "<|tt|>",
759
+ "lstrip": false,
760
+ "normalized": false,
761
+ "rstrip": false,
762
+ "single_word": false,
763
+ "special": true
764
+ },
765
+ "50352": {
766
+ "content": "<|haw|>",
767
+ "lstrip": false,
768
+ "normalized": false,
769
+ "rstrip": false,
770
+ "single_word": false,
771
+ "special": true
772
+ },
773
+ "50353": {
774
+ "content": "<|ln|>",
775
+ "lstrip": false,
776
+ "normalized": false,
777
+ "rstrip": false,
778
+ "single_word": false,
779
+ "special": true
780
+ },
781
+ "50354": {
782
+ "content": "<|ha|>",
783
+ "lstrip": false,
784
+ "normalized": false,
785
+ "rstrip": false,
786
+ "single_word": false,
787
+ "special": true
788
+ },
789
+ "50355": {
790
+ "content": "<|ba|>",
791
+ "lstrip": false,
792
+ "normalized": false,
793
+ "rstrip": false,
794
+ "single_word": false,
795
+ "special": true
796
+ },
797
+ "50356": {
798
+ "content": "<|jw|>",
799
+ "lstrip": false,
800
+ "normalized": false,
801
+ "rstrip": false,
802
+ "single_word": false,
803
+ "special": true
804
+ },
805
+ "50357": {
806
+ "content": "<|su|>",
807
+ "lstrip": false,
808
+ "normalized": false,
809
+ "rstrip": false,
810
+ "single_word": false,
811
+ "special": true
812
+ },
813
+ "50358": {
814
+ "content": "<|translate|>",
815
+ "lstrip": false,
816
+ "normalized": false,
817
+ "rstrip": false,
818
+ "single_word": false,
819
+ "special": true
820
+ },
821
+ "50359": {
822
+ "content": "<|transcribe|>",
823
+ "lstrip": false,
824
+ "normalized": false,
825
+ "rstrip": false,
826
+ "single_word": false,
827
+ "special": true
828
+ },
829
+ "50360": {
830
+ "content": "<|startoflm|>",
831
+ "lstrip": false,
832
+ "normalized": false,
833
+ "rstrip": false,
834
+ "single_word": false,
835
+ "special": true
836
+ },
837
+ "50361": {
838
+ "content": "<|startofprev|>",
839
+ "lstrip": false,
840
+ "normalized": false,
841
+ "rstrip": false,
842
+ "single_word": false,
843
+ "special": true
844
+ },
845
+ "50362": {
846
+ "content": "<|nocaptions|>",
847
+ "lstrip": false,
848
+ "normalized": false,
849
+ "rstrip": false,
850
+ "single_word": false,
851
+ "special": true
852
+ },
853
+ "50363": {
854
+ "content": "<|notimestamps|>",
855
+ "lstrip": false,
856
+ "normalized": false,
857
+ "rstrip": false,
858
+ "single_word": false,
859
+ "special": true
860
+ }
861
+ },
862
+ "additional_special_tokens": [
863
+ "<|endoftext|>",
864
+ "<|startoftranscript|>",
865
+ "<|en|>",
866
+ "<|zh|>",
867
+ "<|de|>",
868
+ "<|es|>",
869
+ "<|ru|>",
870
+ "<|ko|>",
871
+ "<|fr|>",
872
+ "<|ja|>",
873
+ "<|pt|>",
874
+ "<|tr|>",
875
+ "<|pl|>",
876
+ "<|ca|>",
877
+ "<|nl|>",
878
+ "<|ar|>",
879
+ "<|sv|>",
880
+ "<|it|>",
881
+ "<|id|>",
882
+ "<|hi|>",
883
+ "<|fi|>",
884
+ "<|vi|>",
885
+ "<|iw|>",
886
+ "<|uk|>",
887
+ "<|el|>",
888
+ "<|ms|>",
889
+ "<|cs|>",
890
+ "<|ro|>",
891
+ "<|da|>",
892
+ "<|hu|>",
893
+ "<|ta|>",
894
+ "<|no|>",
895
+ "<|th|>",
896
+ "<|ur|>",
897
+ "<|hr|>",
898
+ "<|bg|>",
899
+ "<|lt|>",
900
+ "<|la|>",
901
+ "<|mi|>",
902
+ "<|ml|>",
903
+ "<|cy|>",
904
+ "<|sk|>",
905
+ "<|te|>",
906
+ "<|fa|>",
907
+ "<|lv|>",
908
+ "<|bn|>",
909
+ "<|sr|>",
910
+ "<|az|>",
911
+ "<|sl|>",
912
+ "<|kn|>",
913
+ "<|et|>",
914
+ "<|mk|>",
915
+ "<|br|>",
916
+ "<|eu|>",
917
+ "<|is|>",
918
+ "<|hy|>",
919
+ "<|ne|>",
920
+ "<|mn|>",
921
+ "<|bs|>",
922
+ "<|kk|>",
923
+ "<|sq|>",
924
+ "<|sw|>",
925
+ "<|gl|>",
926
+ "<|mr|>",
927
+ "<|pa|>",
928
+ "<|si|>",
929
+ "<|km|>",
930
+ "<|sn|>",
931
+ "<|yo|>",
932
+ "<|so|>",
933
+ "<|af|>",
934
+ "<|oc|>",
935
+ "<|ka|>",
936
+ "<|be|>",
937
+ "<|tg|>",
938
+ "<|sd|>",
939
+ "<|gu|>",
940
+ "<|am|>",
941
+ "<|yi|>",
942
+ "<|lo|>",
943
+ "<|uz|>",
944
+ "<|fo|>",
945
+ "<|ht|>",
946
+ "<|ps|>",
947
+ "<|tk|>",
948
+ "<|nn|>",
949
+ "<|mt|>",
950
+ "<|sa|>",
951
+ "<|lb|>",
952
+ "<|my|>",
953
+ "<|bo|>",
954
+ "<|tl|>",
955
+ "<|mg|>",
956
+ "<|as|>",
957
+ "<|tt|>",
958
+ "<|haw|>",
959
+ "<|ln|>",
960
+ "<|ha|>",
961
+ "<|ba|>",
962
+ "<|jw|>",
963
+ "<|su|>",
964
+ "<|translate|>",
965
+ "<|transcribe|>",
966
+ "<|startoflm|>",
967
+ "<|startofprev|>",
968
+ "<|nocaptions|>",
969
+ "<|notimestamps|>"
970
+ ],
971
+ "bos_token": "<|endoftext|>",
972
+ "clean_up_tokenization_spaces": true,
973
+ "eos_token": "<|endoftext|>",
974
+ "errors": "replace",
975
+ "model_max_length": 1024,
976
+ "pad_token": "<|endoftext|>",
977
+ "processor_class": "WhisperProcessor",
978
+ "return_attention_mask": false,
979
+ "tokenizer_class": "WhisperTokenizer",
980
+ "unk_token": ""
981
+ }
audio_model_states/whisper-small-cantonese/trainer_state.json ADDED
@@ -0,0 +1,3891 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 10.695329087048833,
3
+ "best_model_checkpoint": "./model_out/checkpoint-10000",
4
+ "epoch": 5.5514433752775725,
5
+ "eval_steps": 500,
6
+ "global_step": 15000,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.01,
13
+ "learning_rate": 2.4000000000000003e-06,
14
+ "loss": 1.2092,
15
+ "step": 25
16
+ },
17
+ {
18
+ "epoch": 0.02,
19
+ "learning_rate": 4.9000000000000005e-06,
20
+ "loss": 0.8939,
21
+ "step": 50
22
+ },
23
+ {
24
+ "epoch": 0.03,
25
+ "learning_rate": 7.4e-06,
26
+ "loss": 0.7337,
27
+ "step": 75
28
+ },
29
+ {
30
+ "epoch": 0.04,
31
+ "learning_rate": 9.900000000000002e-06,
32
+ "loss": 0.6496,
33
+ "step": 100
34
+ },
35
+ {
36
+ "epoch": 0.05,
37
+ "learning_rate": 1.24e-05,
38
+ "loss": 0.5998,
39
+ "step": 125
40
+ },
41
+ {
42
+ "epoch": 0.06,
43
+ "learning_rate": 1.49e-05,
44
+ "loss": 0.5262,
45
+ "step": 150
46
+ },
47
+ {
48
+ "epoch": 0.06,
49
+ "learning_rate": 1.74e-05,
50
+ "loss": 0.501,
51
+ "step": 175
52
+ },
53
+ {
54
+ "epoch": 0.07,
55
+ "learning_rate": 1.9900000000000003e-05,
56
+ "loss": 0.4801,
57
+ "step": 200
58
+ },
59
+ {
60
+ "epoch": 0.08,
61
+ "learning_rate": 2.2400000000000002e-05,
62
+ "loss": 0.4495,
63
+ "step": 225
64
+ },
65
+ {
66
+ "epoch": 0.09,
67
+ "learning_rate": 2.4900000000000002e-05,
68
+ "loss": 0.4372,
69
+ "step": 250
70
+ },
71
+ {
72
+ "epoch": 0.1,
73
+ "learning_rate": 2.7300000000000003e-05,
74
+ "loss": 0.425,
75
+ "step": 275
76
+ },
77
+ {
78
+ "epoch": 0.11,
79
+ "learning_rate": 2.98e-05,
80
+ "loss": 0.4097,
81
+ "step": 300
82
+ },
83
+ {
84
+ "epoch": 0.12,
85
+ "learning_rate": 3.2300000000000006e-05,
86
+ "loss": 0.3997,
87
+ "step": 325
88
+ },
89
+ {
90
+ "epoch": 0.13,
91
+ "learning_rate": 3.48e-05,
92
+ "loss": 0.3828,
93
+ "step": 350
94
+ },
95
+ {
96
+ "epoch": 0.14,
97
+ "learning_rate": 3.73e-05,
98
+ "loss": 0.3817,
99
+ "step": 375
100
+ },
101
+ {
102
+ "epoch": 0.15,
103
+ "learning_rate": 3.9800000000000005e-05,
104
+ "loss": 0.3696,
105
+ "step": 400
106
+ },
107
+ {
108
+ "epoch": 0.16,
109
+ "learning_rate": 4.23e-05,
110
+ "loss": 0.3808,
111
+ "step": 425
112
+ },
113
+ {
114
+ "epoch": 0.17,
115
+ "learning_rate": 4.4800000000000005e-05,
116
+ "loss": 0.3764,
117
+ "step": 450
118
+ },
119
+ {
120
+ "epoch": 0.18,
121
+ "learning_rate": 4.73e-05,
122
+ "loss": 0.3712,
123
+ "step": 475
124
+ },
125
+ {
126
+ "epoch": 0.19,
127
+ "learning_rate": 4.9800000000000004e-05,
128
+ "loss": 0.3416,
129
+ "step": 500
130
+ },
131
+ {
132
+ "epoch": 0.19,
133
+ "eval_cer": 13.553988474370641,
134
+ "eval_loss": 0.2625325620174408,
135
+ "eval_runtime": 367.6787,
136
+ "eval_samples_per_second": 7.02,
137
+ "eval_steps_per_second": 0.878,
138
+ "step": 500
139
+ },
140
+ {
141
+ "epoch": 0.19,
142
+ "learning_rate": 4.992068965517242e-05,
143
+ "loss": 0.3682,
144
+ "step": 525
145
+ },
146
+ {
147
+ "epoch": 0.2,
148
+ "learning_rate": 4.9834482758620695e-05,
149
+ "loss": 0.3716,
150
+ "step": 550
151
+ },
152
+ {
153
+ "epoch": 0.21,
154
+ "learning_rate": 4.9748275862068966e-05,
155
+ "loss": 0.3581,
156
+ "step": 575
157
+ },
158
+ {
159
+ "epoch": 0.22,
160
+ "learning_rate": 4.966206896551724e-05,
161
+ "loss": 0.3542,
162
+ "step": 600
163
+ },
164
+ {
165
+ "epoch": 0.23,
166
+ "learning_rate": 4.9575862068965514e-05,
167
+ "loss": 0.347,
168
+ "step": 625
169
+ },
170
+ {
171
+ "epoch": 0.24,
172
+ "learning_rate": 4.94896551724138e-05,
173
+ "loss": 0.3603,
174
+ "step": 650
175
+ },
176
+ {
177
+ "epoch": 0.25,
178
+ "learning_rate": 4.940344827586207e-05,
179
+ "loss": 0.3568,
180
+ "step": 675
181
+ },
182
+ {
183
+ "epoch": 0.26,
184
+ "learning_rate": 4.931724137931035e-05,
185
+ "loss": 0.3408,
186
+ "step": 700
187
+ },
188
+ {
189
+ "epoch": 0.27,
190
+ "learning_rate": 4.9231034482758625e-05,
191
+ "loss": 0.346,
192
+ "step": 725
193
+ },
194
+ {
195
+ "epoch": 0.28,
196
+ "learning_rate": 4.9144827586206896e-05,
197
+ "loss": 0.3382,
198
+ "step": 750
199
+ },
200
+ {
201
+ "epoch": 0.29,
202
+ "learning_rate": 4.9058620689655174e-05,
203
+ "loss": 0.3422,
204
+ "step": 775
205
+ },
206
+ {
207
+ "epoch": 0.3,
208
+ "learning_rate": 4.897241379310345e-05,
209
+ "loss": 0.3423,
210
+ "step": 800
211
+ },
212
+ {
213
+ "epoch": 0.31,
214
+ "learning_rate": 4.888620689655173e-05,
215
+ "loss": 0.3245,
216
+ "step": 825
217
+ },
218
+ {
219
+ "epoch": 0.31,
220
+ "learning_rate": 4.88e-05,
221
+ "loss": 0.3378,
222
+ "step": 850
223
+ },
224
+ {
225
+ "epoch": 0.32,
226
+ "learning_rate": 4.871379310344828e-05,
227
+ "loss": 0.3388,
228
+ "step": 875
229
+ },
230
+ {
231
+ "epoch": 0.33,
232
+ "learning_rate": 4.8627586206896556e-05,
233
+ "loss": 0.3189,
234
+ "step": 900
235
+ },
236
+ {
237
+ "epoch": 0.34,
238
+ "learning_rate": 4.8541379310344834e-05,
239
+ "loss": 0.3264,
240
+ "step": 925
241
+ },
242
+ {
243
+ "epoch": 0.35,
244
+ "learning_rate": 4.8455172413793105e-05,
245
+ "loss": 0.3231,
246
+ "step": 950
247
+ },
248
+ {
249
+ "epoch": 0.36,
250
+ "learning_rate": 4.836896551724138e-05,
251
+ "loss": 0.3046,
252
+ "step": 975
253
+ },
254
+ {
255
+ "epoch": 0.37,
256
+ "learning_rate": 4.828275862068965e-05,
257
+ "loss": 0.3115,
258
+ "step": 1000
259
+ },
260
+ {
261
+ "epoch": 0.37,
262
+ "eval_cer": 12.473460721868365,
263
+ "eval_loss": 0.2618062496185303,
264
+ "eval_runtime": 363.8751,
265
+ "eval_samples_per_second": 7.093,
266
+ "eval_steps_per_second": 0.888,
267
+ "step": 1000
268
+ },
269
+ {
270
+ "epoch": 0.38,
271
+ "learning_rate": 4.819655172413793e-05,
272
+ "loss": 0.3251,
273
+ "step": 1025
274
+ },
275
+ {
276
+ "epoch": 0.39,
277
+ "learning_rate": 4.811034482758621e-05,
278
+ "loss": 0.3158,
279
+ "step": 1050
280
+ },
281
+ {
282
+ "epoch": 0.4,
283
+ "learning_rate": 4.8024137931034486e-05,
284
+ "loss": 0.3093,
285
+ "step": 1075
286
+ },
287
+ {
288
+ "epoch": 0.41,
289
+ "learning_rate": 4.7937931034482764e-05,
290
+ "loss": 0.3238,
291
+ "step": 1100
292
+ },
293
+ {
294
+ "epoch": 0.42,
295
+ "learning_rate": 4.7851724137931035e-05,
296
+ "loss": 0.3128,
297
+ "step": 1125
298
+ },
299
+ {
300
+ "epoch": 0.43,
301
+ "learning_rate": 4.776551724137931e-05,
302
+ "loss": 0.3012,
303
+ "step": 1150
304
+ },
305
+ {
306
+ "epoch": 0.43,
307
+ "learning_rate": 4.767931034482759e-05,
308
+ "loss": 0.3103,
309
+ "step": 1175
310
+ },
311
+ {
312
+ "epoch": 0.44,
313
+ "learning_rate": 4.759310344827587e-05,
314
+ "loss": 0.314,
315
+ "step": 1200
316
+ },
317
+ {
318
+ "epoch": 0.45,
319
+ "learning_rate": 4.750689655172414e-05,
320
+ "loss": 0.325,
321
+ "step": 1225
322
+ },
323
+ {
324
+ "epoch": 0.46,
325
+ "learning_rate": 4.742068965517242e-05,
326
+ "loss": 0.2923,
327
+ "step": 1250
328
+ },
329
+ {
330
+ "epoch": 0.47,
331
+ "learning_rate": 4.733448275862069e-05,
332
+ "loss": 0.3047,
333
+ "step": 1275
334
+ },
335
+ {
336
+ "epoch": 0.48,
337
+ "learning_rate": 4.7251724137931034e-05,
338
+ "loss": 0.3007,
339
+ "step": 1300
340
+ },
341
+ {
342
+ "epoch": 0.49,
343
+ "learning_rate": 4.716551724137931e-05,
344
+ "loss": 0.3233,
345
+ "step": 1325
346
+ },
347
+ {
348
+ "epoch": 0.5,
349
+ "learning_rate": 4.707931034482759e-05,
350
+ "loss": 0.2983,
351
+ "step": 1350
352
+ },
353
+ {
354
+ "epoch": 0.51,
355
+ "learning_rate": 4.699310344827587e-05,
356
+ "loss": 0.2961,
357
+ "step": 1375
358
+ },
359
+ {
360
+ "epoch": 0.52,
361
+ "learning_rate": 4.690689655172414e-05,
362
+ "loss": 0.2929,
363
+ "step": 1400
364
+ },
365
+ {
366
+ "epoch": 0.53,
367
+ "learning_rate": 4.6820689655172416e-05,
368
+ "loss": 0.2839,
369
+ "step": 1425
370
+ },
371
+ {
372
+ "epoch": 0.54,
373
+ "learning_rate": 4.673448275862069e-05,
374
+ "loss": 0.2973,
375
+ "step": 1450
376
+ },
377
+ {
378
+ "epoch": 0.55,
379
+ "learning_rate": 4.664827586206897e-05,
380
+ "loss": 0.2896,
381
+ "step": 1475
382
+ },
383
+ {
384
+ "epoch": 0.56,
385
+ "learning_rate": 4.656206896551724e-05,
386
+ "loss": 0.289,
387
+ "step": 1500
388
+ },
389
+ {
390
+ "epoch": 0.56,
391
+ "eval_cer": 11.809978768577494,
392
+ "eval_loss": 0.2488066703081131,
393
+ "eval_runtime": 362.9261,
394
+ "eval_samples_per_second": 7.112,
395
+ "eval_steps_per_second": 0.89,
396
+ "step": 1500
397
+ },
398
+ {
399
+ "epoch": 0.56,
400
+ "learning_rate": 4.647586206896552e-05,
401
+ "loss": 0.2872,
402
+ "step": 1525
403
+ },
404
+ {
405
+ "epoch": 0.57,
406
+ "learning_rate": 4.638965517241379e-05,
407
+ "loss": 0.2965,
408
+ "step": 1550
409
+ },
410
+ {
411
+ "epoch": 0.58,
412
+ "learning_rate": 4.630344827586207e-05,
413
+ "loss": 0.3036,
414
+ "step": 1575
415
+ },
416
+ {
417
+ "epoch": 0.59,
418
+ "learning_rate": 4.621724137931035e-05,
419
+ "loss": 0.2891,
420
+ "step": 1600
421
+ },
422
+ {
423
+ "epoch": 0.6,
424
+ "learning_rate": 4.6131034482758624e-05,
425
+ "loss": 0.2896,
426
+ "step": 1625
427
+ },
428
+ {
429
+ "epoch": 0.61,
430
+ "learning_rate": 4.60448275862069e-05,
431
+ "loss": 0.2992,
432
+ "step": 1650
433
+ },
434
+ {
435
+ "epoch": 0.62,
436
+ "learning_rate": 4.595862068965517e-05,
437
+ "loss": 0.2865,
438
+ "step": 1675
439
+ },
440
+ {
441
+ "epoch": 0.63,
442
+ "learning_rate": 4.587241379310345e-05,
443
+ "loss": 0.3007,
444
+ "step": 1700
445
+ },
446
+ {
447
+ "epoch": 0.64,
448
+ "learning_rate": 4.578620689655173e-05,
449
+ "loss": 0.3108,
450
+ "step": 1725
451
+ },
452
+ {
453
+ "epoch": 0.65,
454
+ "learning_rate": 4.5700000000000006e-05,
455
+ "loss": 0.2861,
456
+ "step": 1750
457
+ },
458
+ {
459
+ "epoch": 0.66,
460
+ "learning_rate": 4.561379310344828e-05,
461
+ "loss": 0.2996,
462
+ "step": 1775
463
+ },
464
+ {
465
+ "epoch": 0.67,
466
+ "learning_rate": 4.5527586206896555e-05,
467
+ "loss": 0.295,
468
+ "step": 1800
469
+ },
470
+ {
471
+ "epoch": 0.68,
472
+ "learning_rate": 4.5441379310344826e-05,
473
+ "loss": 0.2848,
474
+ "step": 1825
475
+ },
476
+ {
477
+ "epoch": 0.68,
478
+ "learning_rate": 4.5355172413793104e-05,
479
+ "loss": 0.2997,
480
+ "step": 1850
481
+ },
482
+ {
483
+ "epoch": 0.69,
484
+ "learning_rate": 4.526896551724138e-05,
485
+ "loss": 0.2828,
486
+ "step": 1875
487
+ },
488
+ {
489
+ "epoch": 0.7,
490
+ "learning_rate": 4.518620689655173e-05,
491
+ "loss": 0.2889,
492
+ "step": 1900
493
+ },
494
+ {
495
+ "epoch": 0.71,
496
+ "learning_rate": 4.5100000000000005e-05,
497
+ "loss": 0.2893,
498
+ "step": 1925
499
+ },
500
+ {
501
+ "epoch": 0.72,
502
+ "learning_rate": 4.5013793103448276e-05,
503
+ "loss": 0.276,
504
+ "step": 1950
505
+ },
506
+ {
507
+ "epoch": 0.73,
508
+ "learning_rate": 4.4927586206896554e-05,
509
+ "loss": 0.2784,
510
+ "step": 1975
511
+ },
512
+ {
513
+ "epoch": 0.74,
514
+ "learning_rate": 4.4841379310344825e-05,
515
+ "loss": 0.2826,
516
+ "step": 2000
517
+ },
518
+ {
519
+ "epoch": 0.74,
520
+ "eval_cer": 13.197603882317258,
521
+ "eval_loss": 0.2393285036087036,
522
+ "eval_runtime": 368.8043,
523
+ "eval_samples_per_second": 6.998,
524
+ "eval_steps_per_second": 0.876,
525
+ "step": 2000
526
+ },
527
+ {
528
+ "epoch": 0.75,
529
+ "learning_rate": 4.475517241379311e-05,
530
+ "loss": 0.2707,
531
+ "step": 2025
532
+ },
533
+ {
534
+ "epoch": 0.76,
535
+ "learning_rate": 4.466896551724138e-05,
536
+ "loss": 0.2824,
537
+ "step": 2050
538
+ },
539
+ {
540
+ "epoch": 0.77,
541
+ "learning_rate": 4.458275862068966e-05,
542
+ "loss": 0.296,
543
+ "step": 2075
544
+ },
545
+ {
546
+ "epoch": 0.78,
547
+ "learning_rate": 4.4496551724137936e-05,
548
+ "loss": 0.2786,
549
+ "step": 2100
550
+ },
551
+ {
552
+ "epoch": 0.79,
553
+ "learning_rate": 4.441034482758621e-05,
554
+ "loss": 0.2802,
555
+ "step": 2125
556
+ },
557
+ {
558
+ "epoch": 0.8,
559
+ "learning_rate": 4.4324137931034485e-05,
560
+ "loss": 0.2791,
561
+ "step": 2150
562
+ },
563
+ {
564
+ "epoch": 0.8,
565
+ "learning_rate": 4.423793103448276e-05,
566
+ "loss": 0.2953,
567
+ "step": 2175
568
+ },
569
+ {
570
+ "epoch": 0.81,
571
+ "learning_rate": 4.415172413793104e-05,
572
+ "loss": 0.2807,
573
+ "step": 2200
574
+ },
575
+ {
576
+ "epoch": 0.82,
577
+ "learning_rate": 4.406551724137931e-05,
578
+ "loss": 0.2776,
579
+ "step": 2225
580
+ },
581
+ {
582
+ "epoch": 0.83,
583
+ "learning_rate": 4.397931034482759e-05,
584
+ "loss": 0.2798,
585
+ "step": 2250
586
+ },
587
+ {
588
+ "epoch": 0.84,
589
+ "learning_rate": 4.389310344827586e-05,
590
+ "loss": 0.2804,
591
+ "step": 2275
592
+ },
593
+ {
594
+ "epoch": 0.85,
595
+ "learning_rate": 4.3806896551724144e-05,
596
+ "loss": 0.2892,
597
+ "step": 2300
598
+ },
599
+ {
600
+ "epoch": 0.86,
601
+ "learning_rate": 4.3720689655172415e-05,
602
+ "loss": 0.2773,
603
+ "step": 2325
604
+ },
605
+ {
606
+ "epoch": 0.87,
607
+ "learning_rate": 4.363448275862069e-05,
608
+ "loss": 0.2733,
609
+ "step": 2350
610
+ },
611
+ {
612
+ "epoch": 0.88,
613
+ "learning_rate": 4.3548275862068964e-05,
614
+ "loss": 0.2805,
615
+ "step": 2375
616
+ },
617
+ {
618
+ "epoch": 0.89,
619
+ "learning_rate": 4.346206896551724e-05,
620
+ "loss": 0.274,
621
+ "step": 2400
622
+ },
623
+ {
624
+ "epoch": 0.9,
625
+ "learning_rate": 4.337586206896552e-05,
626
+ "loss": 0.2711,
627
+ "step": 2425
628
+ },
629
+ {
630
+ "epoch": 0.91,
631
+ "learning_rate": 4.32896551724138e-05,
632
+ "loss": 0.2684,
633
+ "step": 2450
634
+ },
635
+ {
636
+ "epoch": 0.92,
637
+ "learning_rate": 4.320344827586207e-05,
638
+ "loss": 0.2731,
639
+ "step": 2475
640
+ },
641
+ {
642
+ "epoch": 0.93,
643
+ "learning_rate": 4.3117241379310346e-05,
644
+ "loss": 0.2689,
645
+ "step": 2500
646
+ },
647
+ {
648
+ "epoch": 0.93,
649
+ "eval_cer": 13.10282074613285,
650
+ "eval_loss": 0.24490514397621155,
651
+ "eval_runtime": 366.5991,
652
+ "eval_samples_per_second": 7.04,
653
+ "eval_steps_per_second": 0.881,
654
+ "step": 2500
655
+ },
656
+ {
657
+ "epoch": 0.93,
658
+ "learning_rate": 4.3031034482758624e-05,
659
+ "loss": 0.2702,
660
+ "step": 2525
661
+ },
662
+ {
663
+ "epoch": 0.94,
664
+ "learning_rate": 4.29448275862069e-05,
665
+ "loss": 0.2623,
666
+ "step": 2550
667
+ },
668
+ {
669
+ "epoch": 0.95,
670
+ "learning_rate": 4.285862068965518e-05,
671
+ "loss": 0.2749,
672
+ "step": 2575
673
+ },
674
+ {
675
+ "epoch": 0.96,
676
+ "learning_rate": 4.277241379310345e-05,
677
+ "loss": 0.2658,
678
+ "step": 2600
679
+ },
680
+ {
681
+ "epoch": 0.97,
682
+ "learning_rate": 4.268620689655173e-05,
683
+ "loss": 0.2637,
684
+ "step": 2625
685
+ },
686
+ {
687
+ "epoch": 0.98,
688
+ "learning_rate": 4.26e-05,
689
+ "loss": 0.2823,
690
+ "step": 2650
691
+ },
692
+ {
693
+ "epoch": 0.99,
694
+ "learning_rate": 4.2513793103448276e-05,
695
+ "loss": 0.2641,
696
+ "step": 2675
697
+ },
698
+ {
699
+ "epoch": 1.0,
700
+ "learning_rate": 4.2427586206896554e-05,
701
+ "loss": 0.2721,
702
+ "step": 2700
703
+ },
704
+ {
705
+ "epoch": 1.01,
706
+ "learning_rate": 4.234137931034483e-05,
707
+ "loss": 0.1882,
708
+ "step": 2725
709
+ },
710
+ {
711
+ "epoch": 1.02,
712
+ "learning_rate": 4.22551724137931e-05,
713
+ "loss": 0.1776,
714
+ "step": 2750
715
+ },
716
+ {
717
+ "epoch": 1.03,
718
+ "learning_rate": 4.216896551724138e-05,
719
+ "loss": 0.1768,
720
+ "step": 2775
721
+ },
722
+ {
723
+ "epoch": 1.04,
724
+ "learning_rate": 4.208275862068966e-05,
725
+ "loss": 0.1714,
726
+ "step": 2800
727
+ },
728
+ {
729
+ "epoch": 1.05,
730
+ "learning_rate": 4.1996551724137936e-05,
731
+ "loss": 0.1771,
732
+ "step": 2825
733
+ },
734
+ {
735
+ "epoch": 1.05,
736
+ "learning_rate": 4.191034482758621e-05,
737
+ "loss": 0.193,
738
+ "step": 2850
739
+ },
740
+ {
741
+ "epoch": 1.06,
742
+ "learning_rate": 4.1824137931034485e-05,
743
+ "loss": 0.1788,
744
+ "step": 2875
745
+ },
746
+ {
747
+ "epoch": 1.07,
748
+ "learning_rate": 4.1737931034482756e-05,
749
+ "loss": 0.1784,
750
+ "step": 2900
751
+ },
752
+ {
753
+ "epoch": 1.08,
754
+ "learning_rate": 4.1651724137931034e-05,
755
+ "loss": 0.1879,
756
+ "step": 2925
757
+ },
758
+ {
759
+ "epoch": 1.09,
760
+ "learning_rate": 4.156551724137932e-05,
761
+ "loss": 0.1825,
762
+ "step": 2950
763
+ },
764
+ {
765
+ "epoch": 1.1,
766
+ "learning_rate": 4.147931034482759e-05,
767
+ "loss": 0.1854,
768
+ "step": 2975
769
+ },
770
+ {
771
+ "epoch": 1.11,
772
+ "learning_rate": 4.139310344827587e-05,
773
+ "loss": 0.1882,
774
+ "step": 3000
775
+ },
776
+ {
777
+ "epoch": 1.11,
778
+ "eval_cer": 11.67728237791932,
779
+ "eval_loss": 0.25736209750175476,
780
+ "eval_runtime": 361.0573,
781
+ "eval_samples_per_second": 7.148,
782
+ "eval_steps_per_second": 0.895,
783
+ "step": 3000
784
+ },
785
+ {
786
+ "epoch": 1.12,
787
+ "learning_rate": 4.130689655172414e-05,
788
+ "loss": 0.1827,
789
+ "step": 3025
790
+ },
791
+ {
792
+ "epoch": 1.13,
793
+ "learning_rate": 4.1220689655172415e-05,
794
+ "loss": 0.1889,
795
+ "step": 3050
796
+ },
797
+ {
798
+ "epoch": 1.14,
799
+ "learning_rate": 4.113448275862069e-05,
800
+ "loss": 0.1938,
801
+ "step": 3075
802
+ },
803
+ {
804
+ "epoch": 1.15,
805
+ "learning_rate": 4.104827586206897e-05,
806
+ "loss": 0.182,
807
+ "step": 3100
808
+ },
809
+ {
810
+ "epoch": 1.16,
811
+ "learning_rate": 4.096206896551724e-05,
812
+ "loss": 0.1884,
813
+ "step": 3125
814
+ },
815
+ {
816
+ "epoch": 1.17,
817
+ "learning_rate": 4.087586206896552e-05,
818
+ "loss": 0.182,
819
+ "step": 3150
820
+ },
821
+ {
822
+ "epoch": 1.18,
823
+ "learning_rate": 4.078965517241379e-05,
824
+ "loss": 0.1795,
825
+ "step": 3175
826
+ },
827
+ {
828
+ "epoch": 1.18,
829
+ "learning_rate": 4.070344827586207e-05,
830
+ "loss": 0.176,
831
+ "step": 3200
832
+ },
833
+ {
834
+ "epoch": 1.19,
835
+ "learning_rate": 4.0617241379310346e-05,
836
+ "loss": 0.1834,
837
+ "step": 3225
838
+ },
839
+ {
840
+ "epoch": 1.2,
841
+ "learning_rate": 4.0531034482758624e-05,
842
+ "loss": 0.1886,
843
+ "step": 3250
844
+ },
845
+ {
846
+ "epoch": 1.21,
847
+ "learning_rate": 4.0444827586206895e-05,
848
+ "loss": 0.1908,
849
+ "step": 3275
850
+ },
851
+ {
852
+ "epoch": 1.22,
853
+ "learning_rate": 4.035862068965517e-05,
854
+ "loss": 0.1813,
855
+ "step": 3300
856
+ },
857
+ {
858
+ "epoch": 1.23,
859
+ "learning_rate": 4.027241379310345e-05,
860
+ "loss": 0.1875,
861
+ "step": 3325
862
+ },
863
+ {
864
+ "epoch": 1.24,
865
+ "learning_rate": 4.018620689655173e-05,
866
+ "loss": 0.1925,
867
+ "step": 3350
868
+ },
869
+ {
870
+ "epoch": 1.25,
871
+ "learning_rate": 4.0100000000000006e-05,
872
+ "loss": 0.1855,
873
+ "step": 3375
874
+ },
875
+ {
876
+ "epoch": 1.26,
877
+ "learning_rate": 4.001379310344828e-05,
878
+ "loss": 0.194,
879
+ "step": 3400
880
+ },
881
+ {
882
+ "epoch": 1.27,
883
+ "learning_rate": 3.9927586206896554e-05,
884
+ "loss": 0.1827,
885
+ "step": 3425
886
+ },
887
+ {
888
+ "epoch": 1.28,
889
+ "learning_rate": 3.9841379310344825e-05,
890
+ "loss": 0.1806,
891
+ "step": 3450
892
+ },
893
+ {
894
+ "epoch": 1.29,
895
+ "learning_rate": 3.975517241379311e-05,
896
+ "loss": 0.1783,
897
+ "step": 3475
898
+ },
899
+ {
900
+ "epoch": 1.3,
901
+ "learning_rate": 3.966896551724138e-05,
902
+ "loss": 0.1889,
903
+ "step": 3500
904
+ },
905
+ {
906
+ "epoch": 1.3,
907
+ "eval_cer": 11.434637549287231,
908
+ "eval_loss": 0.24476666748523712,
909
+ "eval_runtime": 361.1188,
910
+ "eval_samples_per_second": 7.147,
911
+ "eval_steps_per_second": 0.894,
912
+ "step": 3500
913
+ },
914
+ {
915
+ "epoch": 1.3,
916
+ "learning_rate": 3.958275862068966e-05,
917
+ "loss": 0.185,
918
+ "step": 3525
919
+ },
920
+ {
921
+ "epoch": 1.31,
922
+ "learning_rate": 3.949655172413793e-05,
923
+ "loss": 0.1833,
924
+ "step": 3550
925
+ },
926
+ {
927
+ "epoch": 1.32,
928
+ "learning_rate": 3.941034482758621e-05,
929
+ "loss": 0.1749,
930
+ "step": 3575
931
+ },
932
+ {
933
+ "epoch": 1.33,
934
+ "learning_rate": 3.9324137931034485e-05,
935
+ "loss": 0.1906,
936
+ "step": 3600
937
+ },
938
+ {
939
+ "epoch": 1.34,
940
+ "learning_rate": 3.923793103448276e-05,
941
+ "loss": 0.1934,
942
+ "step": 3625
943
+ },
944
+ {
945
+ "epoch": 1.35,
946
+ "learning_rate": 3.9151724137931034e-05,
947
+ "loss": 0.1782,
948
+ "step": 3650
949
+ },
950
+ {
951
+ "epoch": 1.36,
952
+ "learning_rate": 3.906551724137931e-05,
953
+ "loss": 0.19,
954
+ "step": 3675
955
+ },
956
+ {
957
+ "epoch": 1.37,
958
+ "learning_rate": 3.897931034482758e-05,
959
+ "loss": 0.1943,
960
+ "step": 3700
961
+ },
962
+ {
963
+ "epoch": 1.38,
964
+ "learning_rate": 3.889310344827587e-05,
965
+ "loss": 0.1838,
966
+ "step": 3725
967
+ },
968
+ {
969
+ "epoch": 1.39,
970
+ "learning_rate": 3.8806896551724145e-05,
971
+ "loss": 0.2248,
972
+ "step": 3750
973
+ },
974
+ {
975
+ "epoch": 1.4,
976
+ "learning_rate": 3.8720689655172416e-05,
977
+ "loss": 0.1853,
978
+ "step": 3775
979
+ },
980
+ {
981
+ "epoch": 1.41,
982
+ "learning_rate": 3.863448275862069e-05,
983
+ "loss": 0.1867,
984
+ "step": 3800
985
+ },
986
+ {
987
+ "epoch": 1.42,
988
+ "learning_rate": 3.8548275862068964e-05,
989
+ "loss": 0.1801,
990
+ "step": 3825
991
+ },
992
+ {
993
+ "epoch": 1.42,
994
+ "learning_rate": 3.846206896551724e-05,
995
+ "loss": 0.182,
996
+ "step": 3850
997
+ },
998
+ {
999
+ "epoch": 1.43,
1000
+ "learning_rate": 3.837586206896552e-05,
1001
+ "loss": 0.1843,
1002
+ "step": 3875
1003
+ },
1004
+ {
1005
+ "epoch": 1.44,
1006
+ "learning_rate": 3.82896551724138e-05,
1007
+ "loss": 0.184,
1008
+ "step": 3900
1009
+ },
1010
+ {
1011
+ "epoch": 1.45,
1012
+ "learning_rate": 3.820344827586207e-05,
1013
+ "loss": 0.1808,
1014
+ "step": 3925
1015
+ },
1016
+ {
1017
+ "epoch": 1.46,
1018
+ "learning_rate": 3.8117241379310346e-05,
1019
+ "loss": 0.188,
1020
+ "step": 3950
1021
+ },
1022
+ {
1023
+ "epoch": 1.47,
1024
+ "learning_rate": 3.8031034482758624e-05,
1025
+ "loss": 0.1691,
1026
+ "step": 3975
1027
+ },
1028
+ {
1029
+ "epoch": 1.48,
1030
+ "learning_rate": 3.79448275862069e-05,
1031
+ "loss": 0.1923,
1032
+ "step": 4000
1033
+ },
1034
+ {
1035
+ "epoch": 1.48,
1036
+ "eval_cer": 11.8895966029724,
1037
+ "eval_loss": 0.24924282729625702,
1038
+ "eval_runtime": 360.2098,
1039
+ "eval_samples_per_second": 7.165,
1040
+ "eval_steps_per_second": 0.897,
1041
+ "step": 4000
1042
+ },
1043
+ {
1044
+ "epoch": 1.49,
1045
+ "learning_rate": 3.785862068965517e-05,
1046
+ "loss": 0.175,
1047
+ "step": 4025
1048
+ },
1049
+ {
1050
+ "epoch": 1.5,
1051
+ "learning_rate": 3.777241379310345e-05,
1052
+ "loss": 0.1814,
1053
+ "step": 4050
1054
+ },
1055
+ {
1056
+ "epoch": 1.51,
1057
+ "learning_rate": 3.768620689655172e-05,
1058
+ "loss": 0.1808,
1059
+ "step": 4075
1060
+ },
1061
+ {
1062
+ "epoch": 1.52,
1063
+ "learning_rate": 3.76e-05,
1064
+ "loss": 0.1875,
1065
+ "step": 4100
1066
+ },
1067
+ {
1068
+ "epoch": 1.53,
1069
+ "learning_rate": 3.7513793103448284e-05,
1070
+ "loss": 0.1831,
1071
+ "step": 4125
1072
+ },
1073
+ {
1074
+ "epoch": 1.54,
1075
+ "learning_rate": 3.7427586206896555e-05,
1076
+ "loss": 0.184,
1077
+ "step": 4150
1078
+ },
1079
+ {
1080
+ "epoch": 1.55,
1081
+ "learning_rate": 3.734137931034483e-05,
1082
+ "loss": 0.1774,
1083
+ "step": 4175
1084
+ },
1085
+ {
1086
+ "epoch": 1.55,
1087
+ "learning_rate": 3.72551724137931e-05,
1088
+ "loss": 0.1716,
1089
+ "step": 4200
1090
+ },
1091
+ {
1092
+ "epoch": 1.56,
1093
+ "learning_rate": 3.716896551724138e-05,
1094
+ "loss": 0.1845,
1095
+ "step": 4225
1096
+ },
1097
+ {
1098
+ "epoch": 1.57,
1099
+ "learning_rate": 3.708275862068966e-05,
1100
+ "loss": 0.1854,
1101
+ "step": 4250
1102
+ },
1103
+ {
1104
+ "epoch": 1.58,
1105
+ "learning_rate": 3.6996551724137937e-05,
1106
+ "loss": 0.1824,
1107
+ "step": 4275
1108
+ },
1109
+ {
1110
+ "epoch": 1.59,
1111
+ "learning_rate": 3.691034482758621e-05,
1112
+ "loss": 0.1796,
1113
+ "step": 4300
1114
+ },
1115
+ {
1116
+ "epoch": 1.6,
1117
+ "learning_rate": 3.6824137931034485e-05,
1118
+ "loss": 0.1882,
1119
+ "step": 4325
1120
+ },
1121
+ {
1122
+ "epoch": 1.61,
1123
+ "learning_rate": 3.6737931034482756e-05,
1124
+ "loss": 0.1866,
1125
+ "step": 4350
1126
+ },
1127
+ {
1128
+ "epoch": 1.62,
1129
+ "learning_rate": 3.6651724137931034e-05,
1130
+ "loss": 0.1881,
1131
+ "step": 4375
1132
+ },
1133
+ {
1134
+ "epoch": 1.63,
1135
+ "learning_rate": 3.656551724137931e-05,
1136
+ "loss": 0.1822,
1137
+ "step": 4400
1138
+ },
1139
+ {
1140
+ "epoch": 1.64,
1141
+ "learning_rate": 3.647931034482759e-05,
1142
+ "loss": 0.1803,
1143
+ "step": 4425
1144
+ },
1145
+ {
1146
+ "epoch": 1.65,
1147
+ "learning_rate": 3.639310344827586e-05,
1148
+ "loss": 0.1851,
1149
+ "step": 4450
1150
+ },
1151
+ {
1152
+ "epoch": 1.66,
1153
+ "learning_rate": 3.630689655172414e-05,
1154
+ "loss": 0.1825,
1155
+ "step": 4475
1156
+ },
1157
+ {
1158
+ "epoch": 1.67,
1159
+ "learning_rate": 3.6220689655172416e-05,
1160
+ "loss": 0.1886,
1161
+ "step": 4500
1162
+ },
1163
+ {
1164
+ "epoch": 1.67,
1165
+ "eval_cer": 11.75310888686685,
1166
+ "eval_loss": 0.23957248032093048,
1167
+ "eval_runtime": 361.2575,
1168
+ "eval_samples_per_second": 7.144,
1169
+ "eval_steps_per_second": 0.894,
1170
+ "step": 4500
1171
+ },
1172
+ {
1173
+ "epoch": 1.67,
1174
+ "learning_rate": 3.6134482758620694e-05,
1175
+ "loss": 0.1944,
1176
+ "step": 4525
1177
+ },
1178
+ {
1179
+ "epoch": 1.68,
1180
+ "learning_rate": 3.604827586206897e-05,
1181
+ "loss": 0.1896,
1182
+ "step": 4550
1183
+ },
1184
+ {
1185
+ "epoch": 1.69,
1186
+ "learning_rate": 3.596206896551724e-05,
1187
+ "loss": 0.1819,
1188
+ "step": 4575
1189
+ },
1190
+ {
1191
+ "epoch": 1.7,
1192
+ "learning_rate": 3.587586206896552e-05,
1193
+ "loss": 0.1892,
1194
+ "step": 4600
1195
+ },
1196
+ {
1197
+ "epoch": 1.71,
1198
+ "learning_rate": 3.578965517241379e-05,
1199
+ "loss": 0.1914,
1200
+ "step": 4625
1201
+ },
1202
+ {
1203
+ "epoch": 1.72,
1204
+ "learning_rate": 3.5703448275862076e-05,
1205
+ "loss": 0.181,
1206
+ "step": 4650
1207
+ },
1208
+ {
1209
+ "epoch": 1.73,
1210
+ "learning_rate": 3.5617241379310346e-05,
1211
+ "loss": 0.1799,
1212
+ "step": 4675
1213
+ },
1214
+ {
1215
+ "epoch": 1.74,
1216
+ "learning_rate": 3.5531034482758624e-05,
1217
+ "loss": 0.1784,
1218
+ "step": 4700
1219
+ },
1220
+ {
1221
+ "epoch": 1.75,
1222
+ "learning_rate": 3.5444827586206895e-05,
1223
+ "loss": 0.1812,
1224
+ "step": 4725
1225
+ },
1226
+ {
1227
+ "epoch": 1.76,
1228
+ "learning_rate": 3.535862068965517e-05,
1229
+ "loss": 0.1782,
1230
+ "step": 4750
1231
+ },
1232
+ {
1233
+ "epoch": 1.77,
1234
+ "learning_rate": 3.527241379310345e-05,
1235
+ "loss": 0.1783,
1236
+ "step": 4775
1237
+ },
1238
+ {
1239
+ "epoch": 1.78,
1240
+ "learning_rate": 3.518620689655173e-05,
1241
+ "loss": 0.2002,
1242
+ "step": 4800
1243
+ },
1244
+ {
1245
+ "epoch": 1.79,
1246
+ "learning_rate": 3.51e-05,
1247
+ "loss": 0.1948,
1248
+ "step": 4825
1249
+ },
1250
+ {
1251
+ "epoch": 1.79,
1252
+ "learning_rate": 3.501379310344828e-05,
1253
+ "loss": 0.1791,
1254
+ "step": 4850
1255
+ },
1256
+ {
1257
+ "epoch": 1.8,
1258
+ "learning_rate": 3.492758620689655e-05,
1259
+ "loss": 0.2018,
1260
+ "step": 4875
1261
+ },
1262
+ {
1263
+ "epoch": 1.81,
1264
+ "learning_rate": 3.484137931034483e-05,
1265
+ "loss": 0.1828,
1266
+ "step": 4900
1267
+ },
1268
+ {
1269
+ "epoch": 1.82,
1270
+ "learning_rate": 3.475517241379311e-05,
1271
+ "loss": 0.1834,
1272
+ "step": 4925
1273
+ },
1274
+ {
1275
+ "epoch": 1.83,
1276
+ "learning_rate": 3.466896551724138e-05,
1277
+ "loss": 0.1821,
1278
+ "step": 4950
1279
+ },
1280
+ {
1281
+ "epoch": 1.84,
1282
+ "learning_rate": 3.458275862068966e-05,
1283
+ "loss": 0.1864,
1284
+ "step": 4975
1285
+ },
1286
+ {
1287
+ "epoch": 1.85,
1288
+ "learning_rate": 3.449655172413793e-05,
1289
+ "loss": 0.1899,
1290
+ "step": 5000
1291
+ },
1292
+ {
1293
+ "epoch": 1.85,
1294
+ "eval_cer": 12.048832271762208,
1295
+ "eval_loss": 0.23497651517391205,
1296
+ "eval_runtime": 363.6465,
1297
+ "eval_samples_per_second": 7.098,
1298
+ "eval_steps_per_second": 0.888,
1299
+ "step": 5000
1300
+ },
1301
+ {
1302
+ "epoch": 1.86,
1303
+ "learning_rate": 3.441034482758621e-05,
1304
+ "loss": 0.1805,
1305
+ "step": 5025
1306
+ },
1307
+ {
1308
+ "epoch": 1.87,
1309
+ "learning_rate": 3.4324137931034485e-05,
1310
+ "loss": 0.1826,
1311
+ "step": 5050
1312
+ },
1313
+ {
1314
+ "epoch": 1.88,
1315
+ "learning_rate": 3.423793103448276e-05,
1316
+ "loss": 0.1863,
1317
+ "step": 5075
1318
+ },
1319
+ {
1320
+ "epoch": 1.89,
1321
+ "learning_rate": 3.4151724137931034e-05,
1322
+ "loss": 0.1754,
1323
+ "step": 5100
1324
+ },
1325
+ {
1326
+ "epoch": 1.9,
1327
+ "learning_rate": 3.406551724137931e-05,
1328
+ "loss": 0.1801,
1329
+ "step": 5125
1330
+ },
1331
+ {
1332
+ "epoch": 1.91,
1333
+ "learning_rate": 3.397931034482758e-05,
1334
+ "loss": 0.1846,
1335
+ "step": 5150
1336
+ },
1337
+ {
1338
+ "epoch": 1.92,
1339
+ "learning_rate": 3.389310344827587e-05,
1340
+ "loss": 0.1717,
1341
+ "step": 5175
1342
+ },
1343
+ {
1344
+ "epoch": 1.92,
1345
+ "learning_rate": 3.380689655172414e-05,
1346
+ "loss": 0.1754,
1347
+ "step": 5200
1348
+ },
1349
+ {
1350
+ "epoch": 1.93,
1351
+ "learning_rate": 3.3720689655172416e-05,
1352
+ "loss": 0.1819,
1353
+ "step": 5225
1354
+ },
1355
+ {
1356
+ "epoch": 1.94,
1357
+ "learning_rate": 3.363448275862069e-05,
1358
+ "loss": 0.1792,
1359
+ "step": 5250
1360
+ },
1361
+ {
1362
+ "epoch": 1.95,
1363
+ "learning_rate": 3.3548275862068965e-05,
1364
+ "loss": 0.1837,
1365
+ "step": 5275
1366
+ },
1367
+ {
1368
+ "epoch": 1.96,
1369
+ "learning_rate": 3.346206896551724e-05,
1370
+ "loss": 0.188,
1371
+ "step": 5300
1372
+ },
1373
+ {
1374
+ "epoch": 1.97,
1375
+ "learning_rate": 3.337586206896552e-05,
1376
+ "loss": 0.1875,
1377
+ "step": 5325
1378
+ },
1379
+ {
1380
+ "epoch": 1.98,
1381
+ "learning_rate": 3.32896551724138e-05,
1382
+ "loss": 0.1909,
1383
+ "step": 5350
1384
+ },
1385
+ {
1386
+ "epoch": 1.99,
1387
+ "learning_rate": 3.320344827586207e-05,
1388
+ "loss": 0.1714,
1389
+ "step": 5375
1390
+ },
1391
+ {
1392
+ "epoch": 2.0,
1393
+ "learning_rate": 3.311724137931035e-05,
1394
+ "loss": 0.1921,
1395
+ "step": 5400
1396
+ },
1397
+ {
1398
+ "epoch": 2.01,
1399
+ "learning_rate": 3.3031034482758624e-05,
1400
+ "loss": 0.1109,
1401
+ "step": 5425
1402
+ },
1403
+ {
1404
+ "epoch": 2.02,
1405
+ "learning_rate": 3.29448275862069e-05,
1406
+ "loss": 0.0946,
1407
+ "step": 5450
1408
+ },
1409
+ {
1410
+ "epoch": 2.03,
1411
+ "learning_rate": 3.285862068965517e-05,
1412
+ "loss": 0.0967,
1413
+ "step": 5475
1414
+ },
1415
+ {
1416
+ "epoch": 2.04,
1417
+ "learning_rate": 3.277241379310345e-05,
1418
+ "loss": 0.1009,
1419
+ "step": 5500
1420
+ },
1421
+ {
1422
+ "epoch": 2.04,
1423
+ "eval_cer": 11.404306945708221,
1424
+ "eval_loss": 0.24928008019924164,
1425
+ "eval_runtime": 359.254,
1426
+ "eval_samples_per_second": 7.184,
1427
+ "eval_steps_per_second": 0.899,
1428
+ "step": 5500
1429
+ },
1430
+ {
1431
+ "epoch": 2.04,
1432
+ "learning_rate": 3.268620689655172e-05,
1433
+ "loss": 0.0954,
1434
+ "step": 5525
1435
+ },
1436
+ {
1437
+ "epoch": 2.05,
1438
+ "learning_rate": 3.26e-05,
1439
+ "loss": 0.0999,
1440
+ "step": 5550
1441
+ },
1442
+ {
1443
+ "epoch": 2.06,
1444
+ "learning_rate": 3.251379310344828e-05,
1445
+ "loss": 0.0946,
1446
+ "step": 5575
1447
+ },
1448
+ {
1449
+ "epoch": 2.07,
1450
+ "learning_rate": 3.2427586206896555e-05,
1451
+ "loss": 0.0949,
1452
+ "step": 5600
1453
+ },
1454
+ {
1455
+ "epoch": 2.08,
1456
+ "learning_rate": 3.2341379310344826e-05,
1457
+ "loss": 0.1032,
1458
+ "step": 5625
1459
+ },
1460
+ {
1461
+ "epoch": 2.09,
1462
+ "learning_rate": 3.2255172413793104e-05,
1463
+ "loss": 0.1002,
1464
+ "step": 5650
1465
+ },
1466
+ {
1467
+ "epoch": 2.1,
1468
+ "learning_rate": 3.216896551724138e-05,
1469
+ "loss": 0.0973,
1470
+ "step": 5675
1471
+ },
1472
+ {
1473
+ "epoch": 2.11,
1474
+ "learning_rate": 3.208275862068966e-05,
1475
+ "loss": 0.0984,
1476
+ "step": 5700
1477
+ },
1478
+ {
1479
+ "epoch": 2.12,
1480
+ "learning_rate": 3.199655172413794e-05,
1481
+ "loss": 0.1031,
1482
+ "step": 5725
1483
+ },
1484
+ {
1485
+ "epoch": 2.13,
1486
+ "learning_rate": 3.191034482758621e-05,
1487
+ "loss": 0.0968,
1488
+ "step": 5750
1489
+ },
1490
+ {
1491
+ "epoch": 2.14,
1492
+ "learning_rate": 3.1824137931034486e-05,
1493
+ "loss": 0.0983,
1494
+ "step": 5775
1495
+ },
1496
+ {
1497
+ "epoch": 2.15,
1498
+ "learning_rate": 3.173793103448276e-05,
1499
+ "loss": 0.1088,
1500
+ "step": 5800
1501
+ },
1502
+ {
1503
+ "epoch": 2.16,
1504
+ "learning_rate": 3.165172413793104e-05,
1505
+ "loss": 0.0998,
1506
+ "step": 5825
1507
+ },
1508
+ {
1509
+ "epoch": 2.17,
1510
+ "learning_rate": 3.156551724137931e-05,
1511
+ "loss": 0.0991,
1512
+ "step": 5850
1513
+ },
1514
+ {
1515
+ "epoch": 2.17,
1516
+ "learning_rate": 3.147931034482759e-05,
1517
+ "loss": 0.103,
1518
+ "step": 5875
1519
+ },
1520
+ {
1521
+ "epoch": 2.18,
1522
+ "learning_rate": 3.139310344827586e-05,
1523
+ "loss": 0.1046,
1524
+ "step": 5900
1525
+ },
1526
+ {
1527
+ "epoch": 2.19,
1528
+ "learning_rate": 3.130689655172414e-05,
1529
+ "loss": 0.099,
1530
+ "step": 5925
1531
+ },
1532
+ {
1533
+ "epoch": 2.2,
1534
+ "learning_rate": 3.1220689655172416e-05,
1535
+ "loss": 0.1002,
1536
+ "step": 5950
1537
+ },
1538
+ {
1539
+ "epoch": 2.21,
1540
+ "learning_rate": 3.1134482758620694e-05,
1541
+ "loss": 0.1028,
1542
+ "step": 5975
1543
+ },
1544
+ {
1545
+ "epoch": 2.22,
1546
+ "learning_rate": 3.1048275862068965e-05,
1547
+ "loss": 0.1003,
1548
+ "step": 6000
1549
+ },
1550
+ {
1551
+ "epoch": 2.22,
1552
+ "eval_cer": 11.373976342129207,
1553
+ "eval_loss": 0.2572503387928009,
1554
+ "eval_runtime": 359.7119,
1555
+ "eval_samples_per_second": 7.175,
1556
+ "eval_steps_per_second": 0.898,
1557
+ "step": 6000
1558
+ },
1559
+ {
1560
+ "epoch": 2.23,
1561
+ "learning_rate": 3.096206896551724e-05,
1562
+ "loss": 0.0954,
1563
+ "step": 6025
1564
+ },
1565
+ {
1566
+ "epoch": 2.24,
1567
+ "learning_rate": 3.0875862068965514e-05,
1568
+ "loss": 0.0981,
1569
+ "step": 6050
1570
+ },
1571
+ {
1572
+ "epoch": 2.25,
1573
+ "learning_rate": 3.07896551724138e-05,
1574
+ "loss": 0.1006,
1575
+ "step": 6075
1576
+ },
1577
+ {
1578
+ "epoch": 2.26,
1579
+ "learning_rate": 3.070344827586207e-05,
1580
+ "loss": 0.1005,
1581
+ "step": 6100
1582
+ },
1583
+ {
1584
+ "epoch": 2.27,
1585
+ "learning_rate": 3.061724137931035e-05,
1586
+ "loss": 0.1007,
1587
+ "step": 6125
1588
+ },
1589
+ {
1590
+ "epoch": 2.28,
1591
+ "learning_rate": 3.0531034482758625e-05,
1592
+ "loss": 0.1005,
1593
+ "step": 6150
1594
+ },
1595
+ {
1596
+ "epoch": 2.29,
1597
+ "learning_rate": 3.04448275862069e-05,
1598
+ "loss": 0.0968,
1599
+ "step": 6175
1600
+ },
1601
+ {
1602
+ "epoch": 2.29,
1603
+ "learning_rate": 3.0358620689655177e-05,
1604
+ "loss": 0.1034,
1605
+ "step": 6200
1606
+ },
1607
+ {
1608
+ "epoch": 2.3,
1609
+ "learning_rate": 3.0272413793103448e-05,
1610
+ "loss": 0.0982,
1611
+ "step": 6225
1612
+ },
1613
+ {
1614
+ "epoch": 2.31,
1615
+ "learning_rate": 3.018620689655173e-05,
1616
+ "loss": 0.1033,
1617
+ "step": 6250
1618
+ },
1619
+ {
1620
+ "epoch": 2.32,
1621
+ "learning_rate": 3.01e-05,
1622
+ "loss": 0.1031,
1623
+ "step": 6275
1624
+ },
1625
+ {
1626
+ "epoch": 2.33,
1627
+ "learning_rate": 3.0013793103448278e-05,
1628
+ "loss": 0.0985,
1629
+ "step": 6300
1630
+ },
1631
+ {
1632
+ "epoch": 2.34,
1633
+ "learning_rate": 2.9931034482758624e-05,
1634
+ "loss": 0.1405,
1635
+ "step": 6325
1636
+ },
1637
+ {
1638
+ "epoch": 2.35,
1639
+ "learning_rate": 2.9844827586206898e-05,
1640
+ "loss": 0.1059,
1641
+ "step": 6350
1642
+ },
1643
+ {
1644
+ "epoch": 2.36,
1645
+ "learning_rate": 2.9758620689655176e-05,
1646
+ "loss": 0.098,
1647
+ "step": 6375
1648
+ },
1649
+ {
1650
+ "epoch": 2.37,
1651
+ "learning_rate": 2.967241379310345e-05,
1652
+ "loss": 0.1029,
1653
+ "step": 6400
1654
+ },
1655
+ {
1656
+ "epoch": 2.38,
1657
+ "learning_rate": 2.9586206896551728e-05,
1658
+ "loss": 0.1082,
1659
+ "step": 6425
1660
+ },
1661
+ {
1662
+ "epoch": 2.39,
1663
+ "learning_rate": 2.95e-05,
1664
+ "loss": 0.1049,
1665
+ "step": 6450
1666
+ },
1667
+ {
1668
+ "epoch": 2.4,
1669
+ "learning_rate": 2.941379310344828e-05,
1670
+ "loss": 0.0996,
1671
+ "step": 6475
1672
+ },
1673
+ {
1674
+ "epoch": 2.41,
1675
+ "learning_rate": 2.932758620689655e-05,
1676
+ "loss": 0.1031,
1677
+ "step": 6500
1678
+ },
1679
+ {
1680
+ "epoch": 2.41,
1681
+ "eval_cer": 12.954959053685169,
1682
+ "eval_loss": 0.26603972911834717,
1683
+ "eval_runtime": 369.9518,
1684
+ "eval_samples_per_second": 6.977,
1685
+ "eval_steps_per_second": 0.873,
1686
+ "step": 6500
1687
+ },
1688
+ {
1689
+ "epoch": 2.41,
1690
+ "learning_rate": 2.924137931034483e-05,
1691
+ "loss": 0.1062,
1692
+ "step": 6525
1693
+ },
1694
+ {
1695
+ "epoch": 2.42,
1696
+ "learning_rate": 2.9155172413793103e-05,
1697
+ "loss": 0.1057,
1698
+ "step": 6550
1699
+ },
1700
+ {
1701
+ "epoch": 2.43,
1702
+ "learning_rate": 2.906896551724138e-05,
1703
+ "loss": 0.1063,
1704
+ "step": 6575
1705
+ },
1706
+ {
1707
+ "epoch": 2.44,
1708
+ "learning_rate": 2.8982758620689655e-05,
1709
+ "loss": 0.1064,
1710
+ "step": 6600
1711
+ },
1712
+ {
1713
+ "epoch": 2.45,
1714
+ "learning_rate": 2.8896551724137933e-05,
1715
+ "loss": 0.1026,
1716
+ "step": 6625
1717
+ },
1718
+ {
1719
+ "epoch": 2.46,
1720
+ "learning_rate": 2.881034482758621e-05,
1721
+ "loss": 0.1068,
1722
+ "step": 6650
1723
+ },
1724
+ {
1725
+ "epoch": 2.47,
1726
+ "learning_rate": 2.8724137931034485e-05,
1727
+ "loss": 0.1008,
1728
+ "step": 6675
1729
+ },
1730
+ {
1731
+ "epoch": 2.48,
1732
+ "learning_rate": 2.8637931034482763e-05,
1733
+ "loss": 0.1118,
1734
+ "step": 6700
1735
+ },
1736
+ {
1737
+ "epoch": 2.49,
1738
+ "learning_rate": 2.8551724137931034e-05,
1739
+ "loss": 0.1017,
1740
+ "step": 6725
1741
+ },
1742
+ {
1743
+ "epoch": 2.5,
1744
+ "learning_rate": 2.8465517241379315e-05,
1745
+ "loss": 0.1059,
1746
+ "step": 6750
1747
+ },
1748
+ {
1749
+ "epoch": 2.51,
1750
+ "learning_rate": 2.8379310344827586e-05,
1751
+ "loss": 0.1071,
1752
+ "step": 6775
1753
+ },
1754
+ {
1755
+ "epoch": 2.52,
1756
+ "learning_rate": 2.8293103448275863e-05,
1757
+ "loss": 0.1057,
1758
+ "step": 6800
1759
+ },
1760
+ {
1761
+ "epoch": 2.53,
1762
+ "learning_rate": 2.8206896551724138e-05,
1763
+ "loss": 0.1067,
1764
+ "step": 6825
1765
+ },
1766
+ {
1767
+ "epoch": 2.54,
1768
+ "learning_rate": 2.8120689655172415e-05,
1769
+ "loss": 0.1068,
1770
+ "step": 6850
1771
+ },
1772
+ {
1773
+ "epoch": 2.54,
1774
+ "learning_rate": 2.803448275862069e-05,
1775
+ "loss": 0.1175,
1776
+ "step": 6875
1777
+ },
1778
+ {
1779
+ "epoch": 2.55,
1780
+ "learning_rate": 2.7948275862068968e-05,
1781
+ "loss": 0.0992,
1782
+ "step": 6900
1783
+ },
1784
+ {
1785
+ "epoch": 2.56,
1786
+ "learning_rate": 2.7862068965517242e-05,
1787
+ "loss": 0.1085,
1788
+ "step": 6925
1789
+ },
1790
+ {
1791
+ "epoch": 2.57,
1792
+ "learning_rate": 2.777586206896552e-05,
1793
+ "loss": 0.1066,
1794
+ "step": 6950
1795
+ },
1796
+ {
1797
+ "epoch": 2.58,
1798
+ "learning_rate": 2.768965517241379e-05,
1799
+ "loss": 0.1013,
1800
+ "step": 6975
1801
+ },
1802
+ {
1803
+ "epoch": 2.59,
1804
+ "learning_rate": 2.7603448275862072e-05,
1805
+ "loss": 0.1048,
1806
+ "step": 7000
1807
+ },
1808
+ {
1809
+ "epoch": 2.59,
1810
+ "eval_cer": 12.026084319077949,
1811
+ "eval_loss": 0.25693368911743164,
1812
+ "eval_runtime": 363.146,
1813
+ "eval_samples_per_second": 7.107,
1814
+ "eval_steps_per_second": 0.889,
1815
+ "step": 7000
1816
+ },
1817
+ {
1818
+ "epoch": 2.6,
1819
+ "learning_rate": 2.7517241379310343e-05,
1820
+ "loss": 0.1017,
1821
+ "step": 7025
1822
+ },
1823
+ {
1824
+ "epoch": 2.61,
1825
+ "learning_rate": 2.743103448275862e-05,
1826
+ "loss": 0.1024,
1827
+ "step": 7050
1828
+ },
1829
+ {
1830
+ "epoch": 2.62,
1831
+ "learning_rate": 2.73448275862069e-05,
1832
+ "loss": 0.108,
1833
+ "step": 7075
1834
+ },
1835
+ {
1836
+ "epoch": 2.63,
1837
+ "learning_rate": 2.7258620689655173e-05,
1838
+ "loss": 0.0979,
1839
+ "step": 7100
1840
+ },
1841
+ {
1842
+ "epoch": 2.64,
1843
+ "learning_rate": 2.717241379310345e-05,
1844
+ "loss": 0.1076,
1845
+ "step": 7125
1846
+ },
1847
+ {
1848
+ "epoch": 2.65,
1849
+ "learning_rate": 2.7086206896551725e-05,
1850
+ "loss": 0.1024,
1851
+ "step": 7150
1852
+ },
1853
+ {
1854
+ "epoch": 2.66,
1855
+ "learning_rate": 2.7000000000000002e-05,
1856
+ "loss": 0.1037,
1857
+ "step": 7175
1858
+ },
1859
+ {
1860
+ "epoch": 2.66,
1861
+ "learning_rate": 2.6913793103448277e-05,
1862
+ "loss": 0.1063,
1863
+ "step": 7200
1864
+ },
1865
+ {
1866
+ "epoch": 2.67,
1867
+ "learning_rate": 2.6827586206896554e-05,
1868
+ "loss": 0.1006,
1869
+ "step": 7225
1870
+ },
1871
+ {
1872
+ "epoch": 2.68,
1873
+ "learning_rate": 2.674137931034483e-05,
1874
+ "loss": 0.1093,
1875
+ "step": 7250
1876
+ },
1877
+ {
1878
+ "epoch": 2.69,
1879
+ "learning_rate": 2.6655172413793107e-05,
1880
+ "loss": 0.1057,
1881
+ "step": 7275
1882
+ },
1883
+ {
1884
+ "epoch": 2.7,
1885
+ "learning_rate": 2.6568965517241378e-05,
1886
+ "loss": 0.1015,
1887
+ "step": 7300
1888
+ },
1889
+ {
1890
+ "epoch": 2.71,
1891
+ "learning_rate": 2.648275862068966e-05,
1892
+ "loss": 0.1057,
1893
+ "step": 7325
1894
+ },
1895
+ {
1896
+ "epoch": 2.72,
1897
+ "learning_rate": 2.639655172413793e-05,
1898
+ "loss": 0.1067,
1899
+ "step": 7350
1900
+ },
1901
+ {
1902
+ "epoch": 2.73,
1903
+ "learning_rate": 2.6310344827586207e-05,
1904
+ "loss": 0.104,
1905
+ "step": 7375
1906
+ },
1907
+ {
1908
+ "epoch": 2.74,
1909
+ "learning_rate": 2.6224137931034482e-05,
1910
+ "loss": 0.1035,
1911
+ "step": 7400
1912
+ },
1913
+ {
1914
+ "epoch": 2.75,
1915
+ "learning_rate": 2.613793103448276e-05,
1916
+ "loss": 0.1054,
1917
+ "step": 7425
1918
+ },
1919
+ {
1920
+ "epoch": 2.76,
1921
+ "learning_rate": 2.6051724137931037e-05,
1922
+ "loss": 0.1083,
1923
+ "step": 7450
1924
+ },
1925
+ {
1926
+ "epoch": 2.77,
1927
+ "learning_rate": 2.596551724137931e-05,
1928
+ "loss": 0.1081,
1929
+ "step": 7475
1930
+ },
1931
+ {
1932
+ "epoch": 2.78,
1933
+ "learning_rate": 2.587931034482759e-05,
1934
+ "loss": 0.1056,
1935
+ "step": 7500
1936
+ },
1937
+ {
1938
+ "epoch": 2.78,
1939
+ "eval_cer": 11.631786472550804,
1940
+ "eval_loss": 0.26435503363609314,
1941
+ "eval_runtime": 358.991,
1942
+ "eval_samples_per_second": 7.19,
1943
+ "eval_steps_per_second": 0.9,
1944
+ "step": 7500
1945
+ },
1946
+ {
1947
+ "epoch": 2.78,
1948
+ "learning_rate": 2.5793103448275864e-05,
1949
+ "loss": 0.103,
1950
+ "step": 7525
1951
+ },
1952
+ {
1953
+ "epoch": 2.79,
1954
+ "learning_rate": 2.570689655172414e-05,
1955
+ "loss": 0.1068,
1956
+ "step": 7550
1957
+ },
1958
+ {
1959
+ "epoch": 2.8,
1960
+ "learning_rate": 2.5620689655172416e-05,
1961
+ "loss": 0.1,
1962
+ "step": 7575
1963
+ },
1964
+ {
1965
+ "epoch": 2.81,
1966
+ "learning_rate": 2.5534482758620693e-05,
1967
+ "loss": 0.101,
1968
+ "step": 7600
1969
+ },
1970
+ {
1971
+ "epoch": 2.82,
1972
+ "learning_rate": 2.5448275862068964e-05,
1973
+ "loss": 0.1086,
1974
+ "step": 7625
1975
+ },
1976
+ {
1977
+ "epoch": 2.83,
1978
+ "learning_rate": 2.5362068965517246e-05,
1979
+ "loss": 0.1039,
1980
+ "step": 7650
1981
+ },
1982
+ {
1983
+ "epoch": 2.84,
1984
+ "learning_rate": 2.5275862068965516e-05,
1985
+ "loss": 0.1007,
1986
+ "step": 7675
1987
+ },
1988
+ {
1989
+ "epoch": 2.85,
1990
+ "learning_rate": 2.5189655172413794e-05,
1991
+ "loss": 0.1029,
1992
+ "step": 7700
1993
+ },
1994
+ {
1995
+ "epoch": 2.86,
1996
+ "learning_rate": 2.510344827586207e-05,
1997
+ "loss": 0.0989,
1998
+ "step": 7725
1999
+ },
2000
+ {
2001
+ "epoch": 2.87,
2002
+ "learning_rate": 2.5017241379310346e-05,
2003
+ "loss": 0.1042,
2004
+ "step": 7750
2005
+ },
2006
+ {
2007
+ "epoch": 2.88,
2008
+ "learning_rate": 2.493103448275862e-05,
2009
+ "loss": 0.1018,
2010
+ "step": 7775
2011
+ },
2012
+ {
2013
+ "epoch": 2.89,
2014
+ "learning_rate": 2.48448275862069e-05,
2015
+ "loss": 0.1017,
2016
+ "step": 7800
2017
+ },
2018
+ {
2019
+ "epoch": 2.9,
2020
+ "learning_rate": 2.4758620689655173e-05,
2021
+ "loss": 0.1028,
2022
+ "step": 7825
2023
+ },
2024
+ {
2025
+ "epoch": 2.91,
2026
+ "learning_rate": 2.467241379310345e-05,
2027
+ "loss": 0.0994,
2028
+ "step": 7850
2029
+ },
2030
+ {
2031
+ "epoch": 2.91,
2032
+ "learning_rate": 2.4586206896551725e-05,
2033
+ "loss": 0.1049,
2034
+ "step": 7875
2035
+ },
2036
+ {
2037
+ "epoch": 2.92,
2038
+ "learning_rate": 2.45e-05,
2039
+ "loss": 0.1005,
2040
+ "step": 7900
2041
+ },
2042
+ {
2043
+ "epoch": 2.93,
2044
+ "learning_rate": 2.4413793103448277e-05,
2045
+ "loss": 0.1027,
2046
+ "step": 7925
2047
+ },
2048
+ {
2049
+ "epoch": 2.94,
2050
+ "learning_rate": 2.432758620689655e-05,
2051
+ "loss": 0.1096,
2052
+ "step": 7950
2053
+ },
2054
+ {
2055
+ "epoch": 2.95,
2056
+ "learning_rate": 2.424137931034483e-05,
2057
+ "loss": 0.0999,
2058
+ "step": 7975
2059
+ },
2060
+ {
2061
+ "epoch": 2.96,
2062
+ "learning_rate": 2.4155172413793103e-05,
2063
+ "loss": 0.1021,
2064
+ "step": 8000
2065
+ },
2066
+ {
2067
+ "epoch": 2.96,
2068
+ "eval_cer": 11.506672732787383,
2069
+ "eval_loss": 0.2621135413646698,
2070
+ "eval_runtime": 358.5537,
2071
+ "eval_samples_per_second": 7.198,
2072
+ "eval_steps_per_second": 0.901,
2073
+ "step": 8000
2074
+ },
2075
+ {
2076
+ "epoch": 2.97,
2077
+ "learning_rate": 2.4068965517241378e-05,
2078
+ "loss": 0.0984,
2079
+ "step": 8025
2080
+ },
2081
+ {
2082
+ "epoch": 2.98,
2083
+ "learning_rate": 2.398275862068966e-05,
2084
+ "loss": 0.0976,
2085
+ "step": 8050
2086
+ },
2087
+ {
2088
+ "epoch": 2.99,
2089
+ "learning_rate": 2.3896551724137933e-05,
2090
+ "loss": 0.1034,
2091
+ "step": 8075
2092
+ },
2093
+ {
2094
+ "epoch": 3.0,
2095
+ "learning_rate": 2.3810344827586208e-05,
2096
+ "loss": 0.1011,
2097
+ "step": 8100
2098
+ },
2099
+ {
2100
+ "epoch": 3.01,
2101
+ "learning_rate": 2.3724137931034485e-05,
2102
+ "loss": 0.0651,
2103
+ "step": 8125
2104
+ },
2105
+ {
2106
+ "epoch": 3.02,
2107
+ "learning_rate": 2.363793103448276e-05,
2108
+ "loss": 0.0473,
2109
+ "step": 8150
2110
+ },
2111
+ {
2112
+ "epoch": 3.03,
2113
+ "learning_rate": 2.3551724137931037e-05,
2114
+ "loss": 0.0448,
2115
+ "step": 8175
2116
+ },
2117
+ {
2118
+ "epoch": 3.03,
2119
+ "learning_rate": 2.3465517241379312e-05,
2120
+ "loss": 0.0473,
2121
+ "step": 8200
2122
+ },
2123
+ {
2124
+ "epoch": 3.04,
2125
+ "learning_rate": 2.3379310344827586e-05,
2126
+ "loss": 0.0457,
2127
+ "step": 8225
2128
+ },
2129
+ {
2130
+ "epoch": 3.05,
2131
+ "learning_rate": 2.3293103448275864e-05,
2132
+ "loss": 0.0451,
2133
+ "step": 8250
2134
+ },
2135
+ {
2136
+ "epoch": 3.06,
2137
+ "learning_rate": 2.3206896551724138e-05,
2138
+ "loss": 0.0453,
2139
+ "step": 8275
2140
+ },
2141
+ {
2142
+ "epoch": 3.07,
2143
+ "learning_rate": 2.3120689655172416e-05,
2144
+ "loss": 0.0502,
2145
+ "step": 8300
2146
+ },
2147
+ {
2148
+ "epoch": 3.08,
2149
+ "learning_rate": 2.303448275862069e-05,
2150
+ "loss": 0.0471,
2151
+ "step": 8325
2152
+ },
2153
+ {
2154
+ "epoch": 3.09,
2155
+ "learning_rate": 2.2948275862068965e-05,
2156
+ "loss": 0.0476,
2157
+ "step": 8350
2158
+ },
2159
+ {
2160
+ "epoch": 3.1,
2161
+ "learning_rate": 2.2862068965517242e-05,
2162
+ "loss": 0.0473,
2163
+ "step": 8375
2164
+ },
2165
+ {
2166
+ "epoch": 3.11,
2167
+ "learning_rate": 2.2775862068965517e-05,
2168
+ "loss": 0.0476,
2169
+ "step": 8400
2170
+ },
2171
+ {
2172
+ "epoch": 3.12,
2173
+ "learning_rate": 2.2689655172413794e-05,
2174
+ "loss": 0.0466,
2175
+ "step": 8425
2176
+ },
2177
+ {
2178
+ "epoch": 3.13,
2179
+ "learning_rate": 2.2603448275862072e-05,
2180
+ "loss": 0.0478,
2181
+ "step": 8450
2182
+ },
2183
+ {
2184
+ "epoch": 3.14,
2185
+ "learning_rate": 2.2517241379310347e-05,
2186
+ "loss": 0.0464,
2187
+ "step": 8475
2188
+ },
2189
+ {
2190
+ "epoch": 3.15,
2191
+ "learning_rate": 2.2431034482758624e-05,
2192
+ "loss": 0.0468,
2193
+ "step": 8500
2194
+ },
2195
+ {
2196
+ "epoch": 3.15,
2197
+ "eval_cer": 11.332271762208068,
2198
+ "eval_loss": 0.2711898386478424,
2199
+ "eval_runtime": 359.6452,
2200
+ "eval_samples_per_second": 7.177,
2201
+ "eval_steps_per_second": 0.898,
2202
+ "step": 8500
2203
+ },
2204
+ {
2205
+ "epoch": 3.16,
2206
+ "learning_rate": 2.23448275862069e-05,
2207
+ "loss": 0.0468,
2208
+ "step": 8525
2209
+ },
2210
+ {
2211
+ "epoch": 3.16,
2212
+ "learning_rate": 2.2258620689655173e-05,
2213
+ "loss": 0.0526,
2214
+ "step": 8550
2215
+ },
2216
+ {
2217
+ "epoch": 3.17,
2218
+ "learning_rate": 2.217241379310345e-05,
2219
+ "loss": 0.0465,
2220
+ "step": 8575
2221
+ },
2222
+ {
2223
+ "epoch": 3.18,
2224
+ "learning_rate": 2.2086206896551725e-05,
2225
+ "loss": 0.0499,
2226
+ "step": 8600
2227
+ },
2228
+ {
2229
+ "epoch": 3.19,
2230
+ "learning_rate": 2.2000000000000003e-05,
2231
+ "loss": 0.0466,
2232
+ "step": 8625
2233
+ },
2234
+ {
2235
+ "epoch": 3.2,
2236
+ "learning_rate": 2.1913793103448277e-05,
2237
+ "loss": 0.0451,
2238
+ "step": 8650
2239
+ },
2240
+ {
2241
+ "epoch": 3.21,
2242
+ "learning_rate": 2.182758620689655e-05,
2243
+ "loss": 0.047,
2244
+ "step": 8675
2245
+ },
2246
+ {
2247
+ "epoch": 3.22,
2248
+ "learning_rate": 2.174137931034483e-05,
2249
+ "loss": 0.0482,
2250
+ "step": 8700
2251
+ },
2252
+ {
2253
+ "epoch": 3.23,
2254
+ "learning_rate": 2.1655172413793104e-05,
2255
+ "loss": 0.0513,
2256
+ "step": 8725
2257
+ },
2258
+ {
2259
+ "epoch": 3.24,
2260
+ "learning_rate": 2.1568965517241378e-05,
2261
+ "loss": 0.0483,
2262
+ "step": 8750
2263
+ },
2264
+ {
2265
+ "epoch": 3.25,
2266
+ "learning_rate": 2.1482758620689656e-05,
2267
+ "loss": 0.0491,
2268
+ "step": 8775
2269
+ },
2270
+ {
2271
+ "epoch": 3.26,
2272
+ "learning_rate": 2.139655172413793e-05,
2273
+ "loss": 0.0483,
2274
+ "step": 8800
2275
+ },
2276
+ {
2277
+ "epoch": 3.27,
2278
+ "learning_rate": 2.1310344827586208e-05,
2279
+ "loss": 0.0504,
2280
+ "step": 8825
2281
+ },
2282
+ {
2283
+ "epoch": 3.28,
2284
+ "learning_rate": 2.1224137931034486e-05,
2285
+ "loss": 0.0514,
2286
+ "step": 8850
2287
+ },
2288
+ {
2289
+ "epoch": 3.28,
2290
+ "learning_rate": 2.113793103448276e-05,
2291
+ "loss": 0.0484,
2292
+ "step": 8875
2293
+ },
2294
+ {
2295
+ "epoch": 3.29,
2296
+ "learning_rate": 2.1051724137931038e-05,
2297
+ "loss": 0.0474,
2298
+ "step": 8900
2299
+ },
2300
+ {
2301
+ "epoch": 3.3,
2302
+ "learning_rate": 2.0965517241379312e-05,
2303
+ "loss": 0.0462,
2304
+ "step": 8925
2305
+ },
2306
+ {
2307
+ "epoch": 3.31,
2308
+ "learning_rate": 2.0879310344827586e-05,
2309
+ "loss": 0.0501,
2310
+ "step": 8950
2311
+ },
2312
+ {
2313
+ "epoch": 3.32,
2314
+ "learning_rate": 2.0793103448275864e-05,
2315
+ "loss": 0.0487,
2316
+ "step": 8975
2317
+ },
2318
+ {
2319
+ "epoch": 3.33,
2320
+ "learning_rate": 2.070689655172414e-05,
2321
+ "loss": 0.0499,
2322
+ "step": 9000
2323
+ },
2324
+ {
2325
+ "epoch": 3.33,
2326
+ "eval_cer": 11.555959963603275,
2327
+ "eval_loss": 0.28167569637298584,
2328
+ "eval_runtime": 359.0627,
2329
+ "eval_samples_per_second": 7.188,
2330
+ "eval_steps_per_second": 0.9,
2331
+ "step": 9000
2332
+ },
2333
+ {
2334
+ "epoch": 3.34,
2335
+ "learning_rate": 2.0620689655172416e-05,
2336
+ "loss": 0.0496,
2337
+ "step": 9025
2338
+ },
2339
+ {
2340
+ "epoch": 3.35,
2341
+ "learning_rate": 2.053448275862069e-05,
2342
+ "loss": 0.0471,
2343
+ "step": 9050
2344
+ },
2345
+ {
2346
+ "epoch": 3.36,
2347
+ "learning_rate": 2.0448275862068965e-05,
2348
+ "loss": 0.0469,
2349
+ "step": 9075
2350
+ },
2351
+ {
2352
+ "epoch": 3.37,
2353
+ "learning_rate": 2.0362068965517243e-05,
2354
+ "loss": 0.0511,
2355
+ "step": 9100
2356
+ },
2357
+ {
2358
+ "epoch": 3.38,
2359
+ "learning_rate": 2.0275862068965517e-05,
2360
+ "loss": 0.0476,
2361
+ "step": 9125
2362
+ },
2363
+ {
2364
+ "epoch": 3.39,
2365
+ "learning_rate": 2.0189655172413795e-05,
2366
+ "loss": 0.0483,
2367
+ "step": 9150
2368
+ },
2369
+ {
2370
+ "epoch": 3.4,
2371
+ "learning_rate": 2.010344827586207e-05,
2372
+ "loss": 0.0484,
2373
+ "step": 9175
2374
+ },
2375
+ {
2376
+ "epoch": 3.4,
2377
+ "learning_rate": 2.0017241379310343e-05,
2378
+ "loss": 0.0524,
2379
+ "step": 9200
2380
+ },
2381
+ {
2382
+ "epoch": 3.41,
2383
+ "learning_rate": 1.993103448275862e-05,
2384
+ "loss": 0.0469,
2385
+ "step": 9225
2386
+ },
2387
+ {
2388
+ "epoch": 3.42,
2389
+ "learning_rate": 1.98448275862069e-05,
2390
+ "loss": 0.0454,
2391
+ "step": 9250
2392
+ },
2393
+ {
2394
+ "epoch": 3.43,
2395
+ "learning_rate": 1.9758620689655173e-05,
2396
+ "loss": 0.0489,
2397
+ "step": 9275
2398
+ },
2399
+ {
2400
+ "epoch": 3.44,
2401
+ "learning_rate": 1.967241379310345e-05,
2402
+ "loss": 0.0482,
2403
+ "step": 9300
2404
+ },
2405
+ {
2406
+ "epoch": 3.45,
2407
+ "learning_rate": 1.9586206896551725e-05,
2408
+ "loss": 0.0487,
2409
+ "step": 9325
2410
+ },
2411
+ {
2412
+ "epoch": 3.46,
2413
+ "learning_rate": 1.9500000000000003e-05,
2414
+ "loss": 0.048,
2415
+ "step": 9350
2416
+ },
2417
+ {
2418
+ "epoch": 3.47,
2419
+ "learning_rate": 1.9413793103448277e-05,
2420
+ "loss": 0.0477,
2421
+ "step": 9375
2422
+ },
2423
+ {
2424
+ "epoch": 3.48,
2425
+ "learning_rate": 1.9327586206896552e-05,
2426
+ "loss": 0.0511,
2427
+ "step": 9400
2428
+ },
2429
+ {
2430
+ "epoch": 3.49,
2431
+ "learning_rate": 1.924137931034483e-05,
2432
+ "loss": 0.0468,
2433
+ "step": 9425
2434
+ },
2435
+ {
2436
+ "epoch": 3.5,
2437
+ "learning_rate": 1.9155172413793104e-05,
2438
+ "loss": 0.0502,
2439
+ "step": 9450
2440
+ },
2441
+ {
2442
+ "epoch": 3.51,
2443
+ "learning_rate": 1.906896551724138e-05,
2444
+ "loss": 0.0476,
2445
+ "step": 9475
2446
+ },
2447
+ {
2448
+ "epoch": 3.52,
2449
+ "learning_rate": 1.8982758620689656e-05,
2450
+ "loss": 0.0489,
2451
+ "step": 9500
2452
+ },
2453
+ {
2454
+ "epoch": 3.52,
2455
+ "eval_cer": 11.548377312708523,
2456
+ "eval_loss": 0.28109079599380493,
2457
+ "eval_runtime": 359.545,
2458
+ "eval_samples_per_second": 7.179,
2459
+ "eval_steps_per_second": 0.898,
2460
+ "step": 9500
2461
+ },
2462
+ {
2463
+ "epoch": 3.53,
2464
+ "learning_rate": 1.889655172413793e-05,
2465
+ "loss": 0.0463,
2466
+ "step": 9525
2467
+ },
2468
+ {
2469
+ "epoch": 3.53,
2470
+ "learning_rate": 1.8810344827586208e-05,
2471
+ "loss": 0.0457,
2472
+ "step": 9550
2473
+ },
2474
+ {
2475
+ "epoch": 3.54,
2476
+ "learning_rate": 1.8724137931034482e-05,
2477
+ "loss": 0.0463,
2478
+ "step": 9575
2479
+ },
2480
+ {
2481
+ "epoch": 3.55,
2482
+ "learning_rate": 1.863793103448276e-05,
2483
+ "loss": 0.0526,
2484
+ "step": 9600
2485
+ },
2486
+ {
2487
+ "epoch": 3.56,
2488
+ "learning_rate": 1.8551724137931034e-05,
2489
+ "loss": 0.0499,
2490
+ "step": 9625
2491
+ },
2492
+ {
2493
+ "epoch": 3.57,
2494
+ "learning_rate": 1.8465517241379312e-05,
2495
+ "loss": 0.0488,
2496
+ "step": 9650
2497
+ },
2498
+ {
2499
+ "epoch": 3.58,
2500
+ "learning_rate": 1.837931034482759e-05,
2501
+ "loss": 0.0491,
2502
+ "step": 9675
2503
+ },
2504
+ {
2505
+ "epoch": 3.59,
2506
+ "learning_rate": 1.8293103448275864e-05,
2507
+ "loss": 0.046,
2508
+ "step": 9700
2509
+ },
2510
+ {
2511
+ "epoch": 3.6,
2512
+ "learning_rate": 1.820689655172414e-05,
2513
+ "loss": 0.0475,
2514
+ "step": 9725
2515
+ },
2516
+ {
2517
+ "epoch": 3.61,
2518
+ "learning_rate": 1.8120689655172416e-05,
2519
+ "loss": 0.0512,
2520
+ "step": 9750
2521
+ },
2522
+ {
2523
+ "epoch": 3.62,
2524
+ "learning_rate": 1.803448275862069e-05,
2525
+ "loss": 0.0478,
2526
+ "step": 9775
2527
+ },
2528
+ {
2529
+ "epoch": 3.63,
2530
+ "learning_rate": 1.7948275862068965e-05,
2531
+ "loss": 0.0602,
2532
+ "step": 9800
2533
+ },
2534
+ {
2535
+ "epoch": 3.64,
2536
+ "learning_rate": 1.7862068965517243e-05,
2537
+ "loss": 0.0482,
2538
+ "step": 9825
2539
+ },
2540
+ {
2541
+ "epoch": 3.65,
2542
+ "learning_rate": 1.7775862068965517e-05,
2543
+ "loss": 0.0463,
2544
+ "step": 9850
2545
+ },
2546
+ {
2547
+ "epoch": 3.65,
2548
+ "learning_rate": 1.7689655172413795e-05,
2549
+ "loss": 0.0513,
2550
+ "step": 9875
2551
+ },
2552
+ {
2553
+ "epoch": 3.66,
2554
+ "learning_rate": 1.760344827586207e-05,
2555
+ "loss": 0.0495,
2556
+ "step": 9900
2557
+ },
2558
+ {
2559
+ "epoch": 3.67,
2560
+ "learning_rate": 1.7517241379310344e-05,
2561
+ "loss": 0.0547,
2562
+ "step": 9925
2563
+ },
2564
+ {
2565
+ "epoch": 3.68,
2566
+ "learning_rate": 1.743103448275862e-05,
2567
+ "loss": 0.0485,
2568
+ "step": 9950
2569
+ },
2570
+ {
2571
+ "epoch": 3.69,
2572
+ "learning_rate": 1.7344827586206896e-05,
2573
+ "loss": 0.0498,
2574
+ "step": 9975
2575
+ },
2576
+ {
2577
+ "epoch": 3.7,
2578
+ "learning_rate": 1.7258620689655173e-05,
2579
+ "loss": 0.0479,
2580
+ "step": 10000
2581
+ },
2582
+ {
2583
+ "epoch": 3.7,
2584
+ "eval_cer": 10.695329087048833,
2585
+ "eval_loss": 0.27692216634750366,
2586
+ "eval_runtime": 359.4994,
2587
+ "eval_samples_per_second": 7.179,
2588
+ "eval_steps_per_second": 0.898,
2589
+ "step": 10000
2590
+ },
2591
+ {
2592
+ "epoch": 3.71,
2593
+ "learning_rate": 1.7172413793103448e-05,
2594
+ "loss": 0.0456,
2595
+ "step": 10025
2596
+ },
2597
+ {
2598
+ "epoch": 3.72,
2599
+ "learning_rate": 1.7086206896551722e-05,
2600
+ "loss": 0.0512,
2601
+ "step": 10050
2602
+ },
2603
+ {
2604
+ "epoch": 3.73,
2605
+ "learning_rate": 1.7000000000000003e-05,
2606
+ "loss": 0.0487,
2607
+ "step": 10075
2608
+ },
2609
+ {
2610
+ "epoch": 3.74,
2611
+ "learning_rate": 1.6913793103448278e-05,
2612
+ "loss": 0.0463,
2613
+ "step": 10100
2614
+ },
2615
+ {
2616
+ "epoch": 3.75,
2617
+ "learning_rate": 1.6827586206896552e-05,
2618
+ "loss": 0.0625,
2619
+ "step": 10125
2620
+ },
2621
+ {
2622
+ "epoch": 3.76,
2623
+ "learning_rate": 1.674137931034483e-05,
2624
+ "loss": 0.0474,
2625
+ "step": 10150
2626
+ },
2627
+ {
2628
+ "epoch": 3.77,
2629
+ "learning_rate": 1.6655172413793104e-05,
2630
+ "loss": 0.0476,
2631
+ "step": 10175
2632
+ },
2633
+ {
2634
+ "epoch": 3.77,
2635
+ "learning_rate": 1.6568965517241382e-05,
2636
+ "loss": 0.0502,
2637
+ "step": 10200
2638
+ },
2639
+ {
2640
+ "epoch": 3.78,
2641
+ "learning_rate": 1.6482758620689656e-05,
2642
+ "loss": 0.0522,
2643
+ "step": 10225
2644
+ },
2645
+ {
2646
+ "epoch": 3.79,
2647
+ "learning_rate": 1.639655172413793e-05,
2648
+ "loss": 0.0458,
2649
+ "step": 10250
2650
+ },
2651
+ {
2652
+ "epoch": 3.8,
2653
+ "learning_rate": 1.6310344827586208e-05,
2654
+ "loss": 0.0482,
2655
+ "step": 10275
2656
+ },
2657
+ {
2658
+ "epoch": 3.81,
2659
+ "learning_rate": 1.6224137931034483e-05,
2660
+ "loss": 0.0466,
2661
+ "step": 10300
2662
+ },
2663
+ {
2664
+ "epoch": 3.82,
2665
+ "learning_rate": 1.613793103448276e-05,
2666
+ "loss": 0.0466,
2667
+ "step": 10325
2668
+ },
2669
+ {
2670
+ "epoch": 3.83,
2671
+ "learning_rate": 1.6051724137931035e-05,
2672
+ "loss": 0.0451,
2673
+ "step": 10350
2674
+ },
2675
+ {
2676
+ "epoch": 3.84,
2677
+ "learning_rate": 1.596551724137931e-05,
2678
+ "loss": 0.0492,
2679
+ "step": 10375
2680
+ },
2681
+ {
2682
+ "epoch": 3.85,
2683
+ "learning_rate": 1.5879310344827587e-05,
2684
+ "loss": 0.045,
2685
+ "step": 10400
2686
+ },
2687
+ {
2688
+ "epoch": 3.86,
2689
+ "learning_rate": 1.579310344827586e-05,
2690
+ "loss": 0.0457,
2691
+ "step": 10425
2692
+ },
2693
+ {
2694
+ "epoch": 3.87,
2695
+ "learning_rate": 1.570689655172414e-05,
2696
+ "loss": 0.0471,
2697
+ "step": 10450
2698
+ },
2699
+ {
2700
+ "epoch": 3.88,
2701
+ "learning_rate": 1.5620689655172417e-05,
2702
+ "loss": 0.0473,
2703
+ "step": 10475
2704
+ },
2705
+ {
2706
+ "epoch": 3.89,
2707
+ "learning_rate": 1.553448275862069e-05,
2708
+ "loss": 0.049,
2709
+ "step": 10500
2710
+ },
2711
+ {
2712
+ "epoch": 3.89,
2713
+ "eval_cer": 11.51046405823476,
2714
+ "eval_loss": 0.28438711166381836,
2715
+ "eval_runtime": 359.2212,
2716
+ "eval_samples_per_second": 7.185,
2717
+ "eval_steps_per_second": 0.899,
2718
+ "step": 10500
2719
+ },
2720
+ {
2721
+ "epoch": 3.9,
2722
+ "learning_rate": 1.544827586206897e-05,
2723
+ "loss": 0.0458,
2724
+ "step": 10525
2725
+ },
2726
+ {
2727
+ "epoch": 3.9,
2728
+ "learning_rate": 1.5362068965517243e-05,
2729
+ "loss": 0.0456,
2730
+ "step": 10550
2731
+ },
2732
+ {
2733
+ "epoch": 3.91,
2734
+ "learning_rate": 1.5275862068965517e-05,
2735
+ "loss": 0.0474,
2736
+ "step": 10575
2737
+ },
2738
+ {
2739
+ "epoch": 3.92,
2740
+ "learning_rate": 1.5189655172413793e-05,
2741
+ "loss": 0.0447,
2742
+ "step": 10600
2743
+ },
2744
+ {
2745
+ "epoch": 3.93,
2746
+ "learning_rate": 1.510344827586207e-05,
2747
+ "loss": 0.046,
2748
+ "step": 10625
2749
+ },
2750
+ {
2751
+ "epoch": 3.94,
2752
+ "learning_rate": 1.5017241379310346e-05,
2753
+ "loss": 0.0442,
2754
+ "step": 10650
2755
+ },
2756
+ {
2757
+ "epoch": 3.95,
2758
+ "learning_rate": 1.4931034482758622e-05,
2759
+ "loss": 0.0457,
2760
+ "step": 10675
2761
+ },
2762
+ {
2763
+ "epoch": 3.96,
2764
+ "learning_rate": 1.4844827586206898e-05,
2765
+ "loss": 0.0482,
2766
+ "step": 10700
2767
+ },
2768
+ {
2769
+ "epoch": 3.97,
2770
+ "learning_rate": 1.4758620689655172e-05,
2771
+ "loss": 0.0476,
2772
+ "step": 10725
2773
+ },
2774
+ {
2775
+ "epoch": 3.98,
2776
+ "learning_rate": 1.4672413793103448e-05,
2777
+ "loss": 0.0449,
2778
+ "step": 10750
2779
+ },
2780
+ {
2781
+ "epoch": 3.99,
2782
+ "learning_rate": 1.4586206896551724e-05,
2783
+ "loss": 0.0478,
2784
+ "step": 10775
2785
+ },
2786
+ {
2787
+ "epoch": 4.0,
2788
+ "learning_rate": 1.45e-05,
2789
+ "loss": 0.0517,
2790
+ "step": 10800
2791
+ },
2792
+ {
2793
+ "epoch": 4.01,
2794
+ "learning_rate": 1.4413793103448276e-05,
2795
+ "loss": 0.0285,
2796
+ "step": 10825
2797
+ },
2798
+ {
2799
+ "epoch": 4.02,
2800
+ "learning_rate": 1.432758620689655e-05,
2801
+ "loss": 0.0197,
2802
+ "step": 10850
2803
+ },
2804
+ {
2805
+ "epoch": 4.02,
2806
+ "learning_rate": 1.424137931034483e-05,
2807
+ "loss": 0.0183,
2808
+ "step": 10875
2809
+ },
2810
+ {
2811
+ "epoch": 4.03,
2812
+ "learning_rate": 1.4155172413793106e-05,
2813
+ "loss": 0.0186,
2814
+ "step": 10900
2815
+ },
2816
+ {
2817
+ "epoch": 4.04,
2818
+ "learning_rate": 1.406896551724138e-05,
2819
+ "loss": 0.0203,
2820
+ "step": 10925
2821
+ },
2822
+ {
2823
+ "epoch": 4.05,
2824
+ "learning_rate": 1.3982758620689656e-05,
2825
+ "loss": 0.0185,
2826
+ "step": 10950
2827
+ },
2828
+ {
2829
+ "epoch": 4.06,
2830
+ "learning_rate": 1.3896551724137932e-05,
2831
+ "loss": 0.0173,
2832
+ "step": 10975
2833
+ },
2834
+ {
2835
+ "epoch": 4.07,
2836
+ "learning_rate": 1.3810344827586208e-05,
2837
+ "loss": 0.0193,
2838
+ "step": 11000
2839
+ },
2840
+ {
2841
+ "epoch": 4.07,
2842
+ "eval_cer": 11.116166211707613,
2843
+ "eval_loss": 0.29933854937553406,
2844
+ "eval_runtime": 359.5628,
2845
+ "eval_samples_per_second": 7.178,
2846
+ "eval_steps_per_second": 0.898,
2847
+ "step": 11000
2848
+ },
2849
+ {
2850
+ "epoch": 4.08,
2851
+ "learning_rate": 1.3724137931034484e-05,
2852
+ "loss": 0.0176,
2853
+ "step": 11025
2854
+ },
2855
+ {
2856
+ "epoch": 4.09,
2857
+ "learning_rate": 1.3637931034482759e-05,
2858
+ "loss": 0.0171,
2859
+ "step": 11050
2860
+ },
2861
+ {
2862
+ "epoch": 4.1,
2863
+ "learning_rate": 1.3551724137931035e-05,
2864
+ "loss": 0.0173,
2865
+ "step": 11075
2866
+ },
2867
+ {
2868
+ "epoch": 4.11,
2869
+ "learning_rate": 1.3465517241379311e-05,
2870
+ "loss": 0.0185,
2871
+ "step": 11100
2872
+ },
2873
+ {
2874
+ "epoch": 4.12,
2875
+ "learning_rate": 1.3379310344827587e-05,
2876
+ "loss": 0.0161,
2877
+ "step": 11125
2878
+ },
2879
+ {
2880
+ "epoch": 4.13,
2881
+ "learning_rate": 1.3293103448275861e-05,
2882
+ "loss": 0.0186,
2883
+ "step": 11150
2884
+ },
2885
+ {
2886
+ "epoch": 4.14,
2887
+ "learning_rate": 1.3206896551724137e-05,
2888
+ "loss": 0.0191,
2889
+ "step": 11175
2890
+ },
2891
+ {
2892
+ "epoch": 4.15,
2893
+ "learning_rate": 1.3120689655172413e-05,
2894
+ "loss": 0.0177,
2895
+ "step": 11200
2896
+ },
2897
+ {
2898
+ "epoch": 4.15,
2899
+ "learning_rate": 1.303448275862069e-05,
2900
+ "loss": 0.0184,
2901
+ "step": 11225
2902
+ },
2903
+ {
2904
+ "epoch": 4.16,
2905
+ "learning_rate": 1.2948275862068966e-05,
2906
+ "loss": 0.0195,
2907
+ "step": 11250
2908
+ },
2909
+ {
2910
+ "epoch": 4.17,
2911
+ "learning_rate": 1.2862068965517243e-05,
2912
+ "loss": 0.019,
2913
+ "step": 11275
2914
+ },
2915
+ {
2916
+ "epoch": 4.18,
2917
+ "learning_rate": 1.277586206896552e-05,
2918
+ "loss": 0.0183,
2919
+ "step": 11300
2920
+ },
2921
+ {
2922
+ "epoch": 4.19,
2923
+ "learning_rate": 1.2689655172413795e-05,
2924
+ "loss": 0.0188,
2925
+ "step": 11325
2926
+ },
2927
+ {
2928
+ "epoch": 4.2,
2929
+ "learning_rate": 1.260344827586207e-05,
2930
+ "loss": 0.0209,
2931
+ "step": 11350
2932
+ },
2933
+ {
2934
+ "epoch": 4.21,
2935
+ "learning_rate": 1.2517241379310346e-05,
2936
+ "loss": 0.0171,
2937
+ "step": 11375
2938
+ },
2939
+ {
2940
+ "epoch": 4.22,
2941
+ "learning_rate": 1.2431034482758622e-05,
2942
+ "loss": 0.0163,
2943
+ "step": 11400
2944
+ },
2945
+ {
2946
+ "epoch": 4.23,
2947
+ "learning_rate": 1.2344827586206898e-05,
2948
+ "loss": 0.0181,
2949
+ "step": 11425
2950
+ },
2951
+ {
2952
+ "epoch": 4.24,
2953
+ "learning_rate": 1.2258620689655174e-05,
2954
+ "loss": 0.0169,
2955
+ "step": 11450
2956
+ },
2957
+ {
2958
+ "epoch": 4.25,
2959
+ "learning_rate": 1.2172413793103448e-05,
2960
+ "loss": 0.0197,
2961
+ "step": 11475
2962
+ },
2963
+ {
2964
+ "epoch": 4.26,
2965
+ "learning_rate": 1.2086206896551724e-05,
2966
+ "loss": 0.0163,
2967
+ "step": 11500
2968
+ },
2969
+ {
2970
+ "epoch": 4.26,
2971
+ "eval_cer": 11.010009099181074,
2972
+ "eval_loss": 0.3046729266643524,
2973
+ "eval_runtime": 358.1502,
2974
+ "eval_samples_per_second": 7.206,
2975
+ "eval_steps_per_second": 0.902,
2976
+ "step": 11500
2977
+ },
2978
+ {
2979
+ "epoch": 4.27,
2980
+ "learning_rate": 1.2e-05,
2981
+ "loss": 0.0174,
2982
+ "step": 11525
2983
+ },
2984
+ {
2985
+ "epoch": 4.27,
2986
+ "learning_rate": 1.1913793103448276e-05,
2987
+ "loss": 0.0179,
2988
+ "step": 11550
2989
+ },
2990
+ {
2991
+ "epoch": 4.28,
2992
+ "learning_rate": 1.1827586206896552e-05,
2993
+ "loss": 0.0189,
2994
+ "step": 11575
2995
+ },
2996
+ {
2997
+ "epoch": 4.29,
2998
+ "learning_rate": 1.1741379310344828e-05,
2999
+ "loss": 0.0202,
3000
+ "step": 11600
3001
+ },
3002
+ {
3003
+ "epoch": 4.3,
3004
+ "learning_rate": 1.1655172413793104e-05,
3005
+ "loss": 0.017,
3006
+ "step": 11625
3007
+ },
3008
+ {
3009
+ "epoch": 4.31,
3010
+ "learning_rate": 1.156896551724138e-05,
3011
+ "loss": 0.0186,
3012
+ "step": 11650
3013
+ },
3014
+ {
3015
+ "epoch": 4.32,
3016
+ "learning_rate": 1.1482758620689655e-05,
3017
+ "loss": 0.0193,
3018
+ "step": 11675
3019
+ },
3020
+ {
3021
+ "epoch": 4.33,
3022
+ "learning_rate": 1.1396551724137931e-05,
3023
+ "loss": 0.0173,
3024
+ "step": 11700
3025
+ },
3026
+ {
3027
+ "epoch": 4.34,
3028
+ "learning_rate": 1.1310344827586207e-05,
3029
+ "loss": 0.018,
3030
+ "step": 11725
3031
+ },
3032
+ {
3033
+ "epoch": 4.35,
3034
+ "learning_rate": 1.1224137931034483e-05,
3035
+ "loss": 0.0166,
3036
+ "step": 11750
3037
+ },
3038
+ {
3039
+ "epoch": 4.36,
3040
+ "learning_rate": 1.1137931034482759e-05,
3041
+ "loss": 0.0176,
3042
+ "step": 11775
3043
+ },
3044
+ {
3045
+ "epoch": 4.37,
3046
+ "learning_rate": 1.1051724137931035e-05,
3047
+ "loss": 0.0195,
3048
+ "step": 11800
3049
+ },
3050
+ {
3051
+ "epoch": 4.38,
3052
+ "learning_rate": 1.0965517241379311e-05,
3053
+ "loss": 0.0179,
3054
+ "step": 11825
3055
+ },
3056
+ {
3057
+ "epoch": 4.39,
3058
+ "learning_rate": 1.0879310344827587e-05,
3059
+ "loss": 0.0174,
3060
+ "step": 11850
3061
+ },
3062
+ {
3063
+ "epoch": 4.39,
3064
+ "learning_rate": 1.0793103448275863e-05,
3065
+ "loss": 0.0183,
3066
+ "step": 11875
3067
+ },
3068
+ {
3069
+ "epoch": 4.4,
3070
+ "learning_rate": 1.0706896551724138e-05,
3071
+ "loss": 0.0178,
3072
+ "step": 11900
3073
+ },
3074
+ {
3075
+ "epoch": 4.41,
3076
+ "learning_rate": 1.0620689655172414e-05,
3077
+ "loss": 0.0177,
3078
+ "step": 11925
3079
+ },
3080
+ {
3081
+ "epoch": 4.42,
3082
+ "learning_rate": 1.053448275862069e-05,
3083
+ "loss": 0.0186,
3084
+ "step": 11950
3085
+ },
3086
+ {
3087
+ "epoch": 4.43,
3088
+ "learning_rate": 1.0448275862068967e-05,
3089
+ "loss": 0.0174,
3090
+ "step": 11975
3091
+ },
3092
+ {
3093
+ "epoch": 4.44,
3094
+ "learning_rate": 1.0362068965517242e-05,
3095
+ "loss": 0.0175,
3096
+ "step": 12000
3097
+ },
3098
+ {
3099
+ "epoch": 4.44,
3100
+ "eval_cer": 10.911434637549288,
3101
+ "eval_loss": 0.30570268630981445,
3102
+ "eval_runtime": 358.4295,
3103
+ "eval_samples_per_second": 7.201,
3104
+ "eval_steps_per_second": 0.901,
3105
+ "step": 12000
3106
+ },
3107
+ {
3108
+ "epoch": 4.45,
3109
+ "learning_rate": 1.0275862068965518e-05,
3110
+ "loss": 0.019,
3111
+ "step": 12025
3112
+ },
3113
+ {
3114
+ "epoch": 4.46,
3115
+ "learning_rate": 1.0189655172413794e-05,
3116
+ "loss": 0.0179,
3117
+ "step": 12050
3118
+ },
3119
+ {
3120
+ "epoch": 4.47,
3121
+ "learning_rate": 1.010344827586207e-05,
3122
+ "loss": 0.0192,
3123
+ "step": 12075
3124
+ },
3125
+ {
3126
+ "epoch": 4.48,
3127
+ "learning_rate": 1.0017241379310344e-05,
3128
+ "loss": 0.0167,
3129
+ "step": 12100
3130
+ },
3131
+ {
3132
+ "epoch": 4.49,
3133
+ "learning_rate": 9.93103448275862e-06,
3134
+ "loss": 0.0178,
3135
+ "step": 12125
3136
+ },
3137
+ {
3138
+ "epoch": 4.5,
3139
+ "learning_rate": 9.844827586206896e-06,
3140
+ "loss": 0.0209,
3141
+ "step": 12150
3142
+ },
3143
+ {
3144
+ "epoch": 4.51,
3145
+ "learning_rate": 9.758620689655174e-06,
3146
+ "loss": 0.019,
3147
+ "step": 12175
3148
+ },
3149
+ {
3150
+ "epoch": 4.52,
3151
+ "learning_rate": 9.672413793103448e-06,
3152
+ "loss": 0.0169,
3153
+ "step": 12200
3154
+ },
3155
+ {
3156
+ "epoch": 4.52,
3157
+ "learning_rate": 9.586206896551724e-06,
3158
+ "loss": 0.0175,
3159
+ "step": 12225
3160
+ },
3161
+ {
3162
+ "epoch": 4.53,
3163
+ "learning_rate": 9.5e-06,
3164
+ "loss": 0.0166,
3165
+ "step": 12250
3166
+ },
3167
+ {
3168
+ "epoch": 4.54,
3169
+ "learning_rate": 9.413793103448277e-06,
3170
+ "loss": 0.0184,
3171
+ "step": 12275
3172
+ },
3173
+ {
3174
+ "epoch": 4.55,
3175
+ "learning_rate": 9.327586206896553e-06,
3176
+ "loss": 0.0256,
3177
+ "step": 12300
3178
+ },
3179
+ {
3180
+ "epoch": 4.56,
3181
+ "learning_rate": 9.241379310344827e-06,
3182
+ "loss": 0.0177,
3183
+ "step": 12325
3184
+ },
3185
+ {
3186
+ "epoch": 4.57,
3187
+ "learning_rate": 9.155172413793103e-06,
3188
+ "loss": 0.018,
3189
+ "step": 12350
3190
+ },
3191
+ {
3192
+ "epoch": 4.58,
3193
+ "learning_rate": 9.06896551724138e-06,
3194
+ "loss": 0.0187,
3195
+ "step": 12375
3196
+ },
3197
+ {
3198
+ "epoch": 4.59,
3199
+ "learning_rate": 8.982758620689657e-06,
3200
+ "loss": 0.0175,
3201
+ "step": 12400
3202
+ },
3203
+ {
3204
+ "epoch": 4.6,
3205
+ "learning_rate": 8.896551724137931e-06,
3206
+ "loss": 0.0204,
3207
+ "step": 12425
3208
+ },
3209
+ {
3210
+ "epoch": 4.61,
3211
+ "learning_rate": 8.810344827586207e-06,
3212
+ "loss": 0.0177,
3213
+ "step": 12450
3214
+ },
3215
+ {
3216
+ "epoch": 4.62,
3217
+ "learning_rate": 8.724137931034483e-06,
3218
+ "loss": 0.0174,
3219
+ "step": 12475
3220
+ },
3221
+ {
3222
+ "epoch": 4.63,
3223
+ "learning_rate": 8.63793103448276e-06,
3224
+ "loss": 0.0186,
3225
+ "step": 12500
3226
+ },
3227
+ {
3228
+ "epoch": 4.63,
3229
+ "eval_cer": 11.127540188049743,
3230
+ "eval_loss": 0.30721473693847656,
3231
+ "eval_runtime": 359.2661,
3232
+ "eval_samples_per_second": 7.184,
3233
+ "eval_steps_per_second": 0.899,
3234
+ "step": 12500
3235
+ },
3236
+ {
3237
+ "epoch": 4.64,
3238
+ "learning_rate": 8.555172413793104e-06,
3239
+ "loss": 0.0166,
3240
+ "step": 12525
3241
+ },
3242
+ {
3243
+ "epoch": 4.64,
3244
+ "learning_rate": 8.46896551724138e-06,
3245
+ "loss": 0.0196,
3246
+ "step": 12550
3247
+ },
3248
+ {
3249
+ "epoch": 4.65,
3250
+ "learning_rate": 8.382758620689656e-06,
3251
+ "loss": 0.0187,
3252
+ "step": 12575
3253
+ },
3254
+ {
3255
+ "epoch": 4.66,
3256
+ "learning_rate": 8.296551724137932e-06,
3257
+ "loss": 0.0179,
3258
+ "step": 12600
3259
+ },
3260
+ {
3261
+ "epoch": 4.67,
3262
+ "learning_rate": 8.210344827586206e-06,
3263
+ "loss": 0.016,
3264
+ "step": 12625
3265
+ },
3266
+ {
3267
+ "epoch": 4.68,
3268
+ "learning_rate": 8.124137931034482e-06,
3269
+ "loss": 0.022,
3270
+ "step": 12650
3271
+ },
3272
+ {
3273
+ "epoch": 4.69,
3274
+ "learning_rate": 8.03793103448276e-06,
3275
+ "loss": 0.0156,
3276
+ "step": 12675
3277
+ },
3278
+ {
3279
+ "epoch": 4.7,
3280
+ "learning_rate": 7.951724137931036e-06,
3281
+ "loss": 0.0172,
3282
+ "step": 12700
3283
+ },
3284
+ {
3285
+ "epoch": 4.71,
3286
+ "learning_rate": 7.86551724137931e-06,
3287
+ "loss": 0.0188,
3288
+ "step": 12725
3289
+ },
3290
+ {
3291
+ "epoch": 4.72,
3292
+ "learning_rate": 7.779310344827586e-06,
3293
+ "loss": 0.0208,
3294
+ "step": 12750
3295
+ },
3296
+ {
3297
+ "epoch": 4.73,
3298
+ "learning_rate": 7.693103448275862e-06,
3299
+ "loss": 0.0176,
3300
+ "step": 12775
3301
+ },
3302
+ {
3303
+ "epoch": 4.74,
3304
+ "learning_rate": 7.6068965517241385e-06,
3305
+ "loss": 0.018,
3306
+ "step": 12800
3307
+ },
3308
+ {
3309
+ "epoch": 4.75,
3310
+ "learning_rate": 7.520689655172414e-06,
3311
+ "loss": 0.0168,
3312
+ "step": 12825
3313
+ },
3314
+ {
3315
+ "epoch": 4.76,
3316
+ "learning_rate": 7.43448275862069e-06,
3317
+ "loss": 0.0159,
3318
+ "step": 12850
3319
+ },
3320
+ {
3321
+ "epoch": 4.76,
3322
+ "learning_rate": 7.348275862068967e-06,
3323
+ "loss": 0.0166,
3324
+ "step": 12875
3325
+ },
3326
+ {
3327
+ "epoch": 4.77,
3328
+ "learning_rate": 7.262068965517242e-06,
3329
+ "loss": 0.0182,
3330
+ "step": 12900
3331
+ },
3332
+ {
3333
+ "epoch": 4.78,
3334
+ "learning_rate": 7.175862068965518e-06,
3335
+ "loss": 0.0166,
3336
+ "step": 12925
3337
+ },
3338
+ {
3339
+ "epoch": 4.79,
3340
+ "learning_rate": 7.089655172413794e-06,
3341
+ "loss": 0.0168,
3342
+ "step": 12950
3343
+ },
3344
+ {
3345
+ "epoch": 4.8,
3346
+ "learning_rate": 7.003448275862069e-06,
3347
+ "loss": 0.0174,
3348
+ "step": 12975
3349
+ },
3350
+ {
3351
+ "epoch": 4.81,
3352
+ "learning_rate": 6.917241379310345e-06,
3353
+ "loss": 0.0176,
3354
+ "step": 13000
3355
+ },
3356
+ {
3357
+ "epoch": 4.81,
3358
+ "eval_cer": 11.229905975128904,
3359
+ "eval_loss": 0.30905863642692566,
3360
+ "eval_runtime": 359.315,
3361
+ "eval_samples_per_second": 7.183,
3362
+ "eval_steps_per_second": 0.899,
3363
+ "step": 13000
3364
+ },
3365
+ {
3366
+ "epoch": 4.82,
3367
+ "learning_rate": 6.83103448275862e-06,
3368
+ "loss": 0.0175,
3369
+ "step": 13025
3370
+ },
3371
+ {
3372
+ "epoch": 4.83,
3373
+ "learning_rate": 6.744827586206896e-06,
3374
+ "loss": 0.017,
3375
+ "step": 13050
3376
+ },
3377
+ {
3378
+ "epoch": 4.84,
3379
+ "learning_rate": 6.658620689655173e-06,
3380
+ "loss": 0.0158,
3381
+ "step": 13075
3382
+ },
3383
+ {
3384
+ "epoch": 4.85,
3385
+ "learning_rate": 6.572413793103449e-06,
3386
+ "loss": 0.0182,
3387
+ "step": 13100
3388
+ },
3389
+ {
3390
+ "epoch": 4.86,
3391
+ "learning_rate": 6.4862068965517245e-06,
3392
+ "loss": 0.0175,
3393
+ "step": 13125
3394
+ },
3395
+ {
3396
+ "epoch": 4.87,
3397
+ "learning_rate": 6.4000000000000006e-06,
3398
+ "loss": 0.0191,
3399
+ "step": 13150
3400
+ },
3401
+ {
3402
+ "epoch": 4.88,
3403
+ "learning_rate": 6.313793103448276e-06,
3404
+ "loss": 0.0153,
3405
+ "step": 13175
3406
+ },
3407
+ {
3408
+ "epoch": 4.89,
3409
+ "learning_rate": 6.227586206896552e-06,
3410
+ "loss": 0.0197,
3411
+ "step": 13200
3412
+ },
3413
+ {
3414
+ "epoch": 4.89,
3415
+ "learning_rate": 6.141379310344828e-06,
3416
+ "loss": 0.0165,
3417
+ "step": 13225
3418
+ },
3419
+ {
3420
+ "epoch": 4.9,
3421
+ "learning_rate": 6.055172413793104e-06,
3422
+ "loss": 0.0157,
3423
+ "step": 13250
3424
+ },
3425
+ {
3426
+ "epoch": 4.91,
3427
+ "learning_rate": 5.96896551724138e-06,
3428
+ "loss": 0.0156,
3429
+ "step": 13275
3430
+ },
3431
+ {
3432
+ "epoch": 4.92,
3433
+ "learning_rate": 5.882758620689655e-06,
3434
+ "loss": 0.0175,
3435
+ "step": 13300
3436
+ },
3437
+ {
3438
+ "epoch": 4.93,
3439
+ "learning_rate": 5.796551724137932e-06,
3440
+ "loss": 0.0166,
3441
+ "step": 13325
3442
+ },
3443
+ {
3444
+ "epoch": 4.94,
3445
+ "learning_rate": 5.710344827586207e-06,
3446
+ "loss": 0.016,
3447
+ "step": 13350
3448
+ },
3449
+ {
3450
+ "epoch": 4.95,
3451
+ "learning_rate": 5.624137931034483e-06,
3452
+ "loss": 0.0146,
3453
+ "step": 13375
3454
+ },
3455
+ {
3456
+ "epoch": 4.96,
3457
+ "learning_rate": 5.5379310344827585e-06,
3458
+ "loss": 0.0181,
3459
+ "step": 13400
3460
+ },
3461
+ {
3462
+ "epoch": 4.97,
3463
+ "learning_rate": 5.451724137931035e-06,
3464
+ "loss": 0.016,
3465
+ "step": 13425
3466
+ },
3467
+ {
3468
+ "epoch": 4.98,
3469
+ "learning_rate": 5.3655172413793106e-06,
3470
+ "loss": 0.0186,
3471
+ "step": 13450
3472
+ },
3473
+ {
3474
+ "epoch": 4.99,
3475
+ "learning_rate": 5.279310344827587e-06,
3476
+ "loss": 0.0157,
3477
+ "step": 13475
3478
+ },
3479
+ {
3480
+ "epoch": 5.0,
3481
+ "learning_rate": 5.193103448275862e-06,
3482
+ "loss": 0.0173,
3483
+ "step": 13500
3484
+ },
3485
+ {
3486
+ "epoch": 5.0,
3487
+ "eval_cer": 11.082044282681224,
3488
+ "eval_loss": 0.31240326166152954,
3489
+ "eval_runtime": 358.0319,
3490
+ "eval_samples_per_second": 7.209,
3491
+ "eval_steps_per_second": 0.902,
3492
+ "step": 13500
3493
+ },
3494
+ {
3495
+ "epoch": 5.01,
3496
+ "learning_rate": 5.106896551724139e-06,
3497
+ "loss": 0.0115,
3498
+ "step": 13525
3499
+ },
3500
+ {
3501
+ "epoch": 5.01,
3502
+ "learning_rate": 5.020689655172414e-06,
3503
+ "loss": 0.0083,
3504
+ "step": 13550
3505
+ },
3506
+ {
3507
+ "epoch": 5.02,
3508
+ "learning_rate": 4.93448275862069e-06,
3509
+ "loss": 0.0078,
3510
+ "step": 13575
3511
+ },
3512
+ {
3513
+ "epoch": 5.03,
3514
+ "learning_rate": 4.848275862068966e-06,
3515
+ "loss": 0.0081,
3516
+ "step": 13600
3517
+ },
3518
+ {
3519
+ "epoch": 5.04,
3520
+ "learning_rate": 4.762068965517242e-06,
3521
+ "loss": 0.007,
3522
+ "step": 13625
3523
+ },
3524
+ {
3525
+ "epoch": 5.05,
3526
+ "learning_rate": 4.675862068965517e-06,
3527
+ "loss": 0.008,
3528
+ "step": 13650
3529
+ },
3530
+ {
3531
+ "epoch": 5.06,
3532
+ "learning_rate": 4.589655172413793e-06,
3533
+ "loss": 0.0076,
3534
+ "step": 13675
3535
+ },
3536
+ {
3537
+ "epoch": 5.07,
3538
+ "learning_rate": 4.503448275862069e-06,
3539
+ "loss": 0.0093,
3540
+ "step": 13700
3541
+ },
3542
+ {
3543
+ "epoch": 5.08,
3544
+ "learning_rate": 4.417241379310345e-06,
3545
+ "loss": 0.0073,
3546
+ "step": 13725
3547
+ },
3548
+ {
3549
+ "epoch": 5.09,
3550
+ "learning_rate": 4.331034482758621e-06,
3551
+ "loss": 0.0072,
3552
+ "step": 13750
3553
+ },
3554
+ {
3555
+ "epoch": 5.1,
3556
+ "learning_rate": 4.244827586206897e-06,
3557
+ "loss": 0.0079,
3558
+ "step": 13775
3559
+ },
3560
+ {
3561
+ "epoch": 5.11,
3562
+ "learning_rate": 4.158620689655173e-06,
3563
+ "loss": 0.0076,
3564
+ "step": 13800
3565
+ },
3566
+ {
3567
+ "epoch": 5.12,
3568
+ "learning_rate": 4.072413793103449e-06,
3569
+ "loss": 0.0071,
3570
+ "step": 13825
3571
+ },
3572
+ {
3573
+ "epoch": 5.13,
3574
+ "learning_rate": 3.986206896551725e-06,
3575
+ "loss": 0.0071,
3576
+ "step": 13850
3577
+ },
3578
+ {
3579
+ "epoch": 5.14,
3580
+ "learning_rate": 3.9e-06,
3581
+ "loss": 0.0067,
3582
+ "step": 13875
3583
+ },
3584
+ {
3585
+ "epoch": 5.14,
3586
+ "learning_rate": 3.813793103448276e-06,
3587
+ "loss": 0.0077,
3588
+ "step": 13900
3589
+ },
3590
+ {
3591
+ "epoch": 5.15,
3592
+ "learning_rate": 3.7275862068965524e-06,
3593
+ "loss": 0.0089,
3594
+ "step": 13925
3595
+ },
3596
+ {
3597
+ "epoch": 5.16,
3598
+ "learning_rate": 3.641379310344828e-06,
3599
+ "loss": 0.0077,
3600
+ "step": 13950
3601
+ },
3602
+ {
3603
+ "epoch": 5.17,
3604
+ "learning_rate": 3.5551724137931037e-06,
3605
+ "loss": 0.008,
3606
+ "step": 13975
3607
+ },
3608
+ {
3609
+ "epoch": 5.18,
3610
+ "learning_rate": 3.4689655172413793e-06,
3611
+ "loss": 0.007,
3612
+ "step": 14000
3613
+ },
3614
+ {
3615
+ "epoch": 5.18,
3616
+ "eval_cer": 10.862147406733394,
3617
+ "eval_loss": 0.3205486238002777,
3618
+ "eval_runtime": 361.4728,
3619
+ "eval_samples_per_second": 7.14,
3620
+ "eval_steps_per_second": 0.894,
3621
+ "step": 14000
3622
+ },
3623
+ {
3624
+ "epoch": 5.19,
3625
+ "learning_rate": 3.382758620689655e-06,
3626
+ "loss": 0.0088,
3627
+ "step": 14025
3628
+ },
3629
+ {
3630
+ "epoch": 5.2,
3631
+ "learning_rate": 3.2965517241379314e-06,
3632
+ "loss": 0.0072,
3633
+ "step": 14050
3634
+ },
3635
+ {
3636
+ "epoch": 5.21,
3637
+ "learning_rate": 3.210344827586207e-06,
3638
+ "loss": 0.0085,
3639
+ "step": 14075
3640
+ },
3641
+ {
3642
+ "epoch": 5.22,
3643
+ "learning_rate": 3.1241379310344826e-06,
3644
+ "loss": 0.0068,
3645
+ "step": 14100
3646
+ },
3647
+ {
3648
+ "epoch": 5.23,
3649
+ "learning_rate": 3.0379310344827587e-06,
3650
+ "loss": 0.0081,
3651
+ "step": 14125
3652
+ },
3653
+ {
3654
+ "epoch": 5.24,
3655
+ "learning_rate": 2.9517241379310343e-06,
3656
+ "loss": 0.0074,
3657
+ "step": 14150
3658
+ },
3659
+ {
3660
+ "epoch": 5.25,
3661
+ "learning_rate": 2.8655172413793104e-06,
3662
+ "loss": 0.0077,
3663
+ "step": 14175
3664
+ },
3665
+ {
3666
+ "epoch": 5.26,
3667
+ "learning_rate": 2.779310344827586e-06,
3668
+ "loss": 0.0083,
3669
+ "step": 14200
3670
+ },
3671
+ {
3672
+ "epoch": 5.26,
3673
+ "learning_rate": 2.693103448275862e-06,
3674
+ "loss": 0.008,
3675
+ "step": 14225
3676
+ },
3677
+ {
3678
+ "epoch": 5.27,
3679
+ "learning_rate": 2.606896551724138e-06,
3680
+ "loss": 0.0064,
3681
+ "step": 14250
3682
+ },
3683
+ {
3684
+ "epoch": 5.28,
3685
+ "learning_rate": 2.520689655172414e-06,
3686
+ "loss": 0.0076,
3687
+ "step": 14275
3688
+ },
3689
+ {
3690
+ "epoch": 5.29,
3691
+ "learning_rate": 2.4344827586206897e-06,
3692
+ "loss": 0.0066,
3693
+ "step": 14300
3694
+ },
3695
+ {
3696
+ "epoch": 5.3,
3697
+ "learning_rate": 2.3482758620689658e-06,
3698
+ "loss": 0.008,
3699
+ "step": 14325
3700
+ },
3701
+ {
3702
+ "epoch": 5.31,
3703
+ "learning_rate": 2.2620689655172414e-06,
3704
+ "loss": 0.0072,
3705
+ "step": 14350
3706
+ },
3707
+ {
3708
+ "epoch": 5.32,
3709
+ "learning_rate": 2.1758620689655174e-06,
3710
+ "loss": 0.0078,
3711
+ "step": 14375
3712
+ },
3713
+ {
3714
+ "epoch": 5.33,
3715
+ "learning_rate": 2.089655172413793e-06,
3716
+ "loss": 0.0091,
3717
+ "step": 14400
3718
+ },
3719
+ {
3720
+ "epoch": 5.34,
3721
+ "learning_rate": 2.003448275862069e-06,
3722
+ "loss": 0.007,
3723
+ "step": 14425
3724
+ },
3725
+ {
3726
+ "epoch": 5.35,
3727
+ "learning_rate": 1.9172413793103447e-06,
3728
+ "loss": 0.0078,
3729
+ "step": 14450
3730
+ },
3731
+ {
3732
+ "epoch": 5.36,
3733
+ "learning_rate": 1.8310344827586208e-06,
3734
+ "loss": 0.0074,
3735
+ "step": 14475
3736
+ },
3737
+ {
3738
+ "epoch": 5.37,
3739
+ "learning_rate": 1.7448275862068964e-06,
3740
+ "loss": 0.0073,
3741
+ "step": 14500
3742
+ },
3743
+ {
3744
+ "epoch": 5.37,
3745
+ "eval_cer": 10.790112223233242,
3746
+ "eval_loss": 0.320094496011734,
3747
+ "eval_runtime": 359.2372,
3748
+ "eval_samples_per_second": 7.185,
3749
+ "eval_steps_per_second": 0.899,
3750
+ "step": 14500
3751
+ },
3752
+ {
3753
+ "epoch": 5.38,
3754
+ "learning_rate": 1.6586206896551724e-06,
3755
+ "loss": 0.0094,
3756
+ "step": 14525
3757
+ },
3758
+ {
3759
+ "epoch": 5.38,
3760
+ "learning_rate": 1.5724137931034483e-06,
3761
+ "loss": 0.0142,
3762
+ "step": 14550
3763
+ },
3764
+ {
3765
+ "epoch": 5.39,
3766
+ "learning_rate": 1.4862068965517243e-06,
3767
+ "loss": 0.0085,
3768
+ "step": 14575
3769
+ },
3770
+ {
3771
+ "epoch": 5.4,
3772
+ "learning_rate": 1.4034482758620691e-06,
3773
+ "loss": 0.0113,
3774
+ "step": 14600
3775
+ },
3776
+ {
3777
+ "epoch": 5.41,
3778
+ "learning_rate": 1.317241379310345e-06,
3779
+ "loss": 0.0085,
3780
+ "step": 14625
3781
+ },
3782
+ {
3783
+ "epoch": 5.42,
3784
+ "learning_rate": 1.2310344827586208e-06,
3785
+ "loss": 0.0076,
3786
+ "step": 14650
3787
+ },
3788
+ {
3789
+ "epoch": 5.43,
3790
+ "learning_rate": 1.1448275862068966e-06,
3791
+ "loss": 0.0074,
3792
+ "step": 14675
3793
+ },
3794
+ {
3795
+ "epoch": 5.44,
3796
+ "learning_rate": 1.0586206896551725e-06,
3797
+ "loss": 0.0079,
3798
+ "step": 14700
3799
+ },
3800
+ {
3801
+ "epoch": 5.45,
3802
+ "learning_rate": 9.724137931034483e-07,
3803
+ "loss": 0.0067,
3804
+ "step": 14725
3805
+ },
3806
+ {
3807
+ "epoch": 5.46,
3808
+ "learning_rate": 8.862068965517242e-07,
3809
+ "loss": 0.0075,
3810
+ "step": 14750
3811
+ },
3812
+ {
3813
+ "epoch": 5.47,
3814
+ "learning_rate": 8.000000000000001e-07,
3815
+ "loss": 0.0082,
3816
+ "step": 14775
3817
+ },
3818
+ {
3819
+ "epoch": 5.48,
3820
+ "learning_rate": 7.137931034482759e-07,
3821
+ "loss": 0.008,
3822
+ "step": 14800
3823
+ },
3824
+ {
3825
+ "epoch": 5.49,
3826
+ "learning_rate": 6.275862068965517e-07,
3827
+ "loss": 0.0066,
3828
+ "step": 14825
3829
+ },
3830
+ {
3831
+ "epoch": 5.5,
3832
+ "learning_rate": 5.413793103448277e-07,
3833
+ "loss": 0.0077,
3834
+ "step": 14850
3835
+ },
3836
+ {
3837
+ "epoch": 5.51,
3838
+ "learning_rate": 4.551724137931035e-07,
3839
+ "loss": 0.0075,
3840
+ "step": 14875
3841
+ },
3842
+ {
3843
+ "epoch": 5.51,
3844
+ "learning_rate": 3.689655172413793e-07,
3845
+ "loss": 0.0086,
3846
+ "step": 14900
3847
+ },
3848
+ {
3849
+ "epoch": 5.52,
3850
+ "learning_rate": 2.827586206896552e-07,
3851
+ "loss": 0.0067,
3852
+ "step": 14925
3853
+ },
3854
+ {
3855
+ "epoch": 5.53,
3856
+ "learning_rate": 1.9655172413793103e-07,
3857
+ "loss": 0.0075,
3858
+ "step": 14950
3859
+ },
3860
+ {
3861
+ "epoch": 5.54,
3862
+ "learning_rate": 1.1034482758620691e-07,
3863
+ "loss": 0.0067,
3864
+ "step": 14975
3865
+ },
3866
+ {
3867
+ "epoch": 5.55,
3868
+ "learning_rate": 2.413793103448276e-08,
3869
+ "loss": 0.0072,
3870
+ "step": 15000
3871
+ },
3872
+ {
3873
+ "epoch": 5.55,
3874
+ "eval_cer": 10.763572945101608,
3875
+ "eval_loss": 0.3218766450881958,
3876
+ "eval_runtime": 359.0558,
3877
+ "eval_samples_per_second": 7.188,
3878
+ "eval_steps_per_second": 0.9,
3879
+ "step": 15000
3880
+ }
3881
+ ],
3882
+ "logging_steps": 25,
3883
+ "max_steps": 15000,
3884
+ "num_input_tokens_seen": 0,
3885
+ "num_train_epochs": 6,
3886
+ "save_steps": 500,
3887
+ "total_flos": 4.328506848669696e+20,
3888
+ "train_batch_size": 25,
3889
+ "trial_name": null,
3890
+ "trial_params": null
3891
+ }
audio_model_states/whisper-small-cantonese/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35b9840ed01e73c6d3c78f79f6c07a4c43b0f2853aa99d1bb368c2d4b5439c8c
3
+ size 5048
audio_model_states/whisper-small-cantonese/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
lfs.sh ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ git lfs install
2
+ huggingface-cli lfs-enable-largefiles .