kiritan commited on
Commit
2b4955e
·
verified ·
1 Parent(s): 4aae6dc

Training in progress, step 1000, checkpoint

Browse files
last-checkpoint/added_tokens.json CHANGED
@@ -1,1505 +1,1505 @@
1
  {
2
- "<|0.00|>": 50364,
3
- "<|0.02|>": 50365,
4
- "<|0.04|>": 50366,
5
- "<|0.06|>": 50367,
6
- "<|0.08|>": 50368,
7
- "<|0.10|>": 50369,
8
- "<|0.12|>": 50370,
9
- "<|0.14|>": 50371,
10
- "<|0.16|>": 50372,
11
- "<|0.18|>": 50373,
12
- "<|0.20|>": 50374,
13
- "<|0.22|>": 50375,
14
- "<|0.24|>": 50376,
15
- "<|0.26|>": 50377,
16
- "<|0.28|>": 50378,
17
- "<|0.30|>": 50379,
18
- "<|0.32|>": 50380,
19
- "<|0.34|>": 50381,
20
- "<|0.36|>": 50382,
21
- "<|0.38|>": 50383,
22
- "<|0.40|>": 50384,
23
- "<|0.42|>": 50385,
24
- "<|0.44|>": 50386,
25
- "<|0.46|>": 50387,
26
- "<|0.48|>": 50388,
27
- "<|0.50|>": 50389,
28
- "<|0.52|>": 50390,
29
- "<|0.54|>": 50391,
30
- "<|0.56|>": 50392,
31
- "<|0.58|>": 50393,
32
- "<|0.60|>": 50394,
33
- "<|0.62|>": 50395,
34
- "<|0.64|>": 50396,
35
- "<|0.66|>": 50397,
36
- "<|0.68|>": 50398,
37
- "<|0.70|>": 50399,
38
- "<|0.72|>": 50400,
39
- "<|0.74|>": 50401,
40
- "<|0.76|>": 50402,
41
- "<|0.78|>": 50403,
42
- "<|0.80|>": 50404,
43
- "<|0.82|>": 50405,
44
- "<|0.84|>": 50406,
45
- "<|0.86|>": 50407,
46
- "<|0.88|>": 50408,
47
- "<|0.90|>": 50409,
48
- "<|0.92|>": 50410,
49
- "<|0.94|>": 50411,
50
- "<|0.96|>": 50412,
51
- "<|0.98|>": 50413,
52
- "<|1.00|>": 50414,
53
- "<|1.02|>": 50415,
54
- "<|1.04|>": 50416,
55
- "<|1.06|>": 50417,
56
- "<|1.08|>": 50418,
57
- "<|1.10|>": 50419,
58
- "<|1.12|>": 50420,
59
- "<|1.14|>": 50421,
60
- "<|1.16|>": 50422,
61
- "<|1.18|>": 50423,
62
- "<|1.20|>": 50424,
63
- "<|1.22|>": 50425,
64
- "<|1.24|>": 50426,
65
- "<|1.26|>": 50427,
66
- "<|1.28|>": 50428,
67
- "<|1.30|>": 50429,
68
- "<|1.32|>": 50430,
69
- "<|1.34|>": 50431,
70
- "<|1.36|>": 50432,
71
- "<|1.38|>": 50433,
72
- "<|1.40|>": 50434,
73
- "<|1.42|>": 50435,
74
- "<|1.44|>": 50436,
75
- "<|1.46|>": 50437,
76
- "<|1.48|>": 50438,
77
- "<|1.50|>": 50439,
78
- "<|1.52|>": 50440,
79
- "<|1.54|>": 50441,
80
- "<|1.56|>": 50442,
81
- "<|1.58|>": 50443,
82
- "<|1.60|>": 50444,
83
- "<|1.62|>": 50445,
84
- "<|1.64|>": 50446,
85
- "<|1.66|>": 50447,
86
- "<|1.68|>": 50448,
87
- "<|1.70|>": 50449,
88
- "<|1.72|>": 50450,
89
- "<|1.74|>": 50451,
90
- "<|1.76|>": 50452,
91
- "<|1.78|>": 50453,
92
- "<|1.80|>": 50454,
93
- "<|1.82|>": 50455,
94
- "<|1.84|>": 50456,
95
- "<|1.86|>": 50457,
96
- "<|1.88|>": 50458,
97
- "<|1.90|>": 50459,
98
- "<|1.92|>": 50460,
99
- "<|1.94|>": 50461,
100
- "<|1.96|>": 50462,
101
- "<|1.98|>": 50463,
102
- "<|10.00|>": 50864,
103
- "<|10.02|>": 50865,
104
- "<|10.04|>": 50866,
105
- "<|10.06|>": 50867,
106
- "<|10.08|>": 50868,
107
- "<|10.10|>": 50869,
108
- "<|10.12|>": 50870,
109
- "<|10.14|>": 50871,
110
- "<|10.16|>": 50872,
111
- "<|10.18|>": 50873,
112
- "<|10.20|>": 50874,
113
- "<|10.22|>": 50875,
114
- "<|10.24|>": 50876,
115
- "<|10.26|>": 50877,
116
- "<|10.28|>": 50878,
117
- "<|10.30|>": 50879,
118
- "<|10.32|>": 50880,
119
- "<|10.34|>": 50881,
120
- "<|10.36|>": 50882,
121
- "<|10.38|>": 50883,
122
- "<|10.40|>": 50884,
123
- "<|10.42|>": 50885,
124
- "<|10.44|>": 50886,
125
- "<|10.46|>": 50887,
126
- "<|10.48|>": 50888,
127
- "<|10.50|>": 50889,
128
- "<|10.52|>": 50890,
129
- "<|10.54|>": 50891,
130
- "<|10.56|>": 50892,
131
- "<|10.58|>": 50893,
132
- "<|10.60|>": 50894,
133
- "<|10.62|>": 50895,
134
- "<|10.64|>": 50896,
135
- "<|10.66|>": 50897,
136
- "<|10.68|>": 50898,
137
- "<|10.70|>": 50899,
138
- "<|10.72|>": 50900,
139
- "<|10.74|>": 50901,
140
- "<|10.76|>": 50902,
141
- "<|10.78|>": 50903,
142
- "<|10.80|>": 50904,
143
- "<|10.82|>": 50905,
144
- "<|10.84|>": 50906,
145
- "<|10.86|>": 50907,
146
- "<|10.88|>": 50908,
147
- "<|10.90|>": 50909,
148
- "<|10.92|>": 50910,
149
- "<|10.94|>": 50911,
150
- "<|10.96|>": 50912,
151
- "<|10.98|>": 50913,
152
- "<|11.00|>": 50914,
153
- "<|11.02|>": 50915,
154
- "<|11.04|>": 50916,
155
- "<|11.06|>": 50917,
156
- "<|11.08|>": 50918,
157
- "<|11.10|>": 50919,
158
- "<|11.12|>": 50920,
159
- "<|11.14|>": 50921,
160
- "<|11.16|>": 50922,
161
- "<|11.18|>": 50923,
162
- "<|11.20|>": 50924,
163
- "<|11.22|>": 50925,
164
- "<|11.24|>": 50926,
165
- "<|11.26|>": 50927,
166
- "<|11.28|>": 50928,
167
- "<|11.30|>": 50929,
168
- "<|11.32|>": 50930,
169
- "<|11.34|>": 50931,
170
- "<|11.36|>": 50932,
171
- "<|11.38|>": 50933,
172
- "<|11.40|>": 50934,
173
- "<|11.42|>": 50935,
174
- "<|11.44|>": 50936,
175
- "<|11.46|>": 50937,
176
- "<|11.48|>": 50938,
177
- "<|11.50|>": 50939,
178
- "<|11.52|>": 50940,
179
- "<|11.54|>": 50941,
180
- "<|11.56|>": 50942,
181
- "<|11.58|>": 50943,
182
- "<|11.60|>": 50944,
183
- "<|11.62|>": 50945,
184
- "<|11.64|>": 50946,
185
- "<|11.66|>": 50947,
186
- "<|11.68|>": 50948,
187
- "<|11.70|>": 50949,
188
- "<|11.72|>": 50950,
189
- "<|11.74|>": 50951,
190
- "<|11.76|>": 50952,
191
- "<|11.78|>": 50953,
192
- "<|11.80|>": 50954,
193
- "<|11.82|>": 50955,
194
- "<|11.84|>": 50956,
195
- "<|11.86|>": 50957,
196
- "<|11.88|>": 50958,
197
- "<|11.90|>": 50959,
198
- "<|11.92|>": 50960,
199
- "<|11.94|>": 50961,
200
- "<|11.96|>": 50962,
201
- "<|11.98|>": 50963,
202
- "<|12.00|>": 50964,
203
- "<|12.02|>": 50965,
204
- "<|12.04|>": 50966,
205
- "<|12.06|>": 50967,
206
- "<|12.08|>": 50968,
207
- "<|12.10|>": 50969,
208
- "<|12.12|>": 50970,
209
- "<|12.14|>": 50971,
210
- "<|12.16|>": 50972,
211
- "<|12.18|>": 50973,
212
- "<|12.20|>": 50974,
213
- "<|12.22|>": 50975,
214
- "<|12.24|>": 50976,
215
- "<|12.26|>": 50977,
216
- "<|12.28|>": 50978,
217
- "<|12.30|>": 50979,
218
- "<|12.32|>": 50980,
219
- "<|12.34|>": 50981,
220
- "<|12.36|>": 50982,
221
- "<|12.38|>": 50983,
222
- "<|12.40|>": 50984,
223
- "<|12.42|>": 50985,
224
- "<|12.44|>": 50986,
225
- "<|12.46|>": 50987,
226
- "<|12.48|>": 50988,
227
- "<|12.50|>": 50989,
228
- "<|12.52|>": 50990,
229
- "<|12.54|>": 50991,
230
- "<|12.56|>": 50992,
231
- "<|12.58|>": 50993,
232
- "<|12.60|>": 50994,
233
- "<|12.62|>": 50995,
234
- "<|12.64|>": 50996,
235
- "<|12.66|>": 50997,
236
- "<|12.68|>": 50998,
237
- "<|12.70|>": 50999,
238
- "<|12.72|>": 51000,
239
- "<|12.74|>": 51001,
240
- "<|12.76|>": 51002,
241
- "<|12.78|>": 51003,
242
- "<|12.80|>": 51004,
243
- "<|12.82|>": 51005,
244
- "<|12.84|>": 51006,
245
- "<|12.86|>": 51007,
246
- "<|12.88|>": 51008,
247
- "<|12.90|>": 51009,
248
- "<|12.92|>": 51010,
249
- "<|12.94|>": 51011,
250
- "<|12.96|>": 51012,
251
- "<|12.98|>": 51013,
252
- "<|13.00|>": 51014,
253
- "<|13.02|>": 51015,
254
- "<|13.04|>": 51016,
255
- "<|13.06|>": 51017,
256
- "<|13.08|>": 51018,
257
- "<|13.10|>": 51019,
258
- "<|13.12|>": 51020,
259
- "<|13.14|>": 51021,
260
- "<|13.16|>": 51022,
261
- "<|13.18|>": 51023,
262
- "<|13.20|>": 51024,
263
- "<|13.22|>": 51025,
264
- "<|13.24|>": 51026,
265
- "<|13.26|>": 51027,
266
- "<|13.28|>": 51028,
267
- "<|13.30|>": 51029,
268
- "<|13.32|>": 51030,
269
- "<|13.34|>": 51031,
270
- "<|13.36|>": 51032,
271
- "<|13.38|>": 51033,
272
- "<|13.40|>": 51034,
273
- "<|13.42|>": 51035,
274
- "<|13.44|>": 51036,
275
- "<|13.46|>": 51037,
276
- "<|13.48|>": 51038,
277
- "<|13.50|>": 51039,
278
- "<|13.52|>": 51040,
279
- "<|13.54|>": 51041,
280
- "<|13.56|>": 51042,
281
- "<|13.58|>": 51043,
282
- "<|13.60|>": 51044,
283
- "<|13.62|>": 51045,
284
- "<|13.64|>": 51046,
285
- "<|13.66|>": 51047,
286
- "<|13.68|>": 51048,
287
- "<|13.70|>": 51049,
288
- "<|13.72|>": 51050,
289
- "<|13.74|>": 51051,
290
- "<|13.76|>": 51052,
291
- "<|13.78|>": 51053,
292
- "<|13.80|>": 51054,
293
- "<|13.82|>": 51055,
294
- "<|13.84|>": 51056,
295
- "<|13.86|>": 51057,
296
- "<|13.88|>": 51058,
297
- "<|13.90|>": 51059,
298
- "<|13.92|>": 51060,
299
- "<|13.94|>": 51061,
300
- "<|13.96|>": 51062,
301
- "<|13.98|>": 51063,
302
- "<|14.00|>": 51064,
303
- "<|14.02|>": 51065,
304
- "<|14.04|>": 51066,
305
- "<|14.06|>": 51067,
306
- "<|14.08|>": 51068,
307
- "<|14.10|>": 51069,
308
- "<|14.12|>": 51070,
309
- "<|14.14|>": 51071,
310
- "<|14.16|>": 51072,
311
- "<|14.18|>": 51073,
312
- "<|14.20|>": 51074,
313
- "<|14.22|>": 51075,
314
- "<|14.24|>": 51076,
315
- "<|14.26|>": 51077,
316
- "<|14.28|>": 51078,
317
- "<|14.30|>": 51079,
318
- "<|14.32|>": 51080,
319
- "<|14.34|>": 51081,
320
- "<|14.36|>": 51082,
321
- "<|14.38|>": 51083,
322
- "<|14.40|>": 51084,
323
- "<|14.42|>": 51085,
324
- "<|14.44|>": 51086,
325
- "<|14.46|>": 51087,
326
- "<|14.48|>": 51088,
327
- "<|14.50|>": 51089,
328
- "<|14.52|>": 51090,
329
- "<|14.54|>": 51091,
330
- "<|14.56|>": 51092,
331
- "<|14.58|>": 51093,
332
- "<|14.60|>": 51094,
333
- "<|14.62|>": 51095,
334
- "<|14.64|>": 51096,
335
- "<|14.66|>": 51097,
336
- "<|14.68|>": 51098,
337
- "<|14.70|>": 51099,
338
- "<|14.72|>": 51100,
339
- "<|14.74|>": 51101,
340
- "<|14.76|>": 51102,
341
- "<|14.78|>": 51103,
342
- "<|14.80|>": 51104,
343
- "<|14.82|>": 51105,
344
- "<|14.84|>": 51106,
345
- "<|14.86|>": 51107,
346
- "<|14.88|>": 51108,
347
- "<|14.90|>": 51109,
348
- "<|14.92|>": 51110,
349
- "<|14.94|>": 51111,
350
- "<|14.96|>": 51112,
351
- "<|14.98|>": 51113,
352
- "<|15.00|>": 51114,
353
- "<|15.02|>": 51115,
354
- "<|15.04|>": 51116,
355
- "<|15.06|>": 51117,
356
- "<|15.08|>": 51118,
357
- "<|15.10|>": 51119,
358
- "<|15.12|>": 51120,
359
- "<|15.14|>": 51121,
360
- "<|15.16|>": 51122,
361
- "<|15.18|>": 51123,
362
- "<|15.20|>": 51124,
363
- "<|15.22|>": 51125,
364
- "<|15.24|>": 51126,
365
- "<|15.26|>": 51127,
366
- "<|15.28|>": 51128,
367
- "<|15.30|>": 51129,
368
- "<|15.32|>": 51130,
369
- "<|15.34|>": 51131,
370
- "<|15.36|>": 51132,
371
- "<|15.38|>": 51133,
372
- "<|15.40|>": 51134,
373
- "<|15.42|>": 51135,
374
- "<|15.44|>": 51136,
375
- "<|15.46|>": 51137,
376
- "<|15.48|>": 51138,
377
- "<|15.50|>": 51139,
378
- "<|15.52|>": 51140,
379
- "<|15.54|>": 51141,
380
- "<|15.56|>": 51142,
381
- "<|15.58|>": 51143,
382
- "<|15.60|>": 51144,
383
- "<|15.62|>": 51145,
384
- "<|15.64|>": 51146,
385
- "<|15.66|>": 51147,
386
- "<|15.68|>": 51148,
387
- "<|15.70|>": 51149,
388
- "<|15.72|>": 51150,
389
- "<|15.74|>": 51151,
390
- "<|15.76|>": 51152,
391
- "<|15.78|>": 51153,
392
- "<|15.80|>": 51154,
393
- "<|15.82|>": 51155,
394
- "<|15.84|>": 51156,
395
- "<|15.86|>": 51157,
396
- "<|15.88|>": 51158,
397
- "<|15.90|>": 51159,
398
- "<|15.92|>": 51160,
399
- "<|15.94|>": 51161,
400
- "<|15.96|>": 51162,
401
- "<|15.98|>": 51163,
402
- "<|16.00|>": 51164,
403
- "<|16.02|>": 51165,
404
- "<|16.04|>": 51166,
405
- "<|16.06|>": 51167,
406
- "<|16.08|>": 51168,
407
- "<|16.10|>": 51169,
408
- "<|16.12|>": 51170,
409
- "<|16.14|>": 51171,
410
- "<|16.16|>": 51172,
411
- "<|16.18|>": 51173,
412
- "<|16.20|>": 51174,
413
- "<|16.22|>": 51175,
414
- "<|16.24|>": 51176,
415
- "<|16.26|>": 51177,
416
- "<|16.28|>": 51178,
417
- "<|16.30|>": 51179,
418
- "<|16.32|>": 51180,
419
- "<|16.34|>": 51181,
420
- "<|16.36|>": 51182,
421
- "<|16.38|>": 51183,
422
- "<|16.40|>": 51184,
423
- "<|16.42|>": 51185,
424
- "<|16.44|>": 51186,
425
- "<|16.46|>": 51187,
426
- "<|16.48|>": 51188,
427
- "<|16.50|>": 51189,
428
- "<|16.52|>": 51190,
429
- "<|16.54|>": 51191,
430
- "<|16.56|>": 51192,
431
- "<|16.58|>": 51193,
432
- "<|16.60|>": 51194,
433
- "<|16.62|>": 51195,
434
- "<|16.64|>": 51196,
435
- "<|16.66|>": 51197,
436
- "<|16.68|>": 51198,
437
- "<|16.70|>": 51199,
438
- "<|16.72|>": 51200,
439
- "<|16.74|>": 51201,
440
- "<|16.76|>": 51202,
441
- "<|16.78|>": 51203,
442
- "<|16.80|>": 51204,
443
- "<|16.82|>": 51205,
444
- "<|16.84|>": 51206,
445
- "<|16.86|>": 51207,
446
- "<|16.88|>": 51208,
447
- "<|16.90|>": 51209,
448
- "<|16.92|>": 51210,
449
- "<|16.94|>": 51211,
450
- "<|16.96|>": 51212,
451
- "<|16.98|>": 51213,
452
- "<|17.00|>": 51214,
453
- "<|17.02|>": 51215,
454
- "<|17.04|>": 51216,
455
- "<|17.06|>": 51217,
456
- "<|17.08|>": 51218,
457
- "<|17.10|>": 51219,
458
- "<|17.12|>": 51220,
459
- "<|17.14|>": 51221,
460
- "<|17.16|>": 51222,
461
- "<|17.18|>": 51223,
462
- "<|17.20|>": 51224,
463
- "<|17.22|>": 51225,
464
- "<|17.24|>": 51226,
465
- "<|17.26|>": 51227,
466
- "<|17.28|>": 51228,
467
- "<|17.30|>": 51229,
468
- "<|17.32|>": 51230,
469
- "<|17.34|>": 51231,
470
- "<|17.36|>": 51232,
471
- "<|17.38|>": 51233,
472
- "<|17.40|>": 51234,
473
- "<|17.42|>": 51235,
474
- "<|17.44|>": 51236,
475
- "<|17.46|>": 51237,
476
- "<|17.48|>": 51238,
477
- "<|17.50|>": 51239,
478
- "<|17.52|>": 51240,
479
- "<|17.54|>": 51241,
480
- "<|17.56|>": 51242,
481
- "<|17.58|>": 51243,
482
- "<|17.60|>": 51244,
483
- "<|17.62|>": 51245,
484
- "<|17.64|>": 51246,
485
- "<|17.66|>": 51247,
486
- "<|17.68|>": 51248,
487
- "<|17.70|>": 51249,
488
- "<|17.72|>": 51250,
489
- "<|17.74|>": 51251,
490
- "<|17.76|>": 51252,
491
- "<|17.78|>": 51253,
492
- "<|17.80|>": 51254,
493
- "<|17.82|>": 51255,
494
- "<|17.84|>": 51256,
495
- "<|17.86|>": 51257,
496
- "<|17.88|>": 51258,
497
- "<|17.90|>": 51259,
498
- "<|17.92|>": 51260,
499
- "<|17.94|>": 51261,
500
- "<|17.96|>": 51262,
501
- "<|17.98|>": 51263,
502
- "<|18.00|>": 51264,
503
- "<|18.02|>": 51265,
504
- "<|18.04|>": 51266,
505
- "<|18.06|>": 51267,
506
- "<|18.08|>": 51268,
507
- "<|18.10|>": 51269,
508
- "<|18.12|>": 51270,
509
- "<|18.14|>": 51271,
510
- "<|18.16|>": 51272,
511
- "<|18.18|>": 51273,
512
- "<|18.20|>": 51274,
513
- "<|18.22|>": 51275,
514
- "<|18.24|>": 51276,
515
- "<|18.26|>": 51277,
516
- "<|18.28|>": 51278,
517
- "<|18.30|>": 51279,
518
- "<|18.32|>": 51280,
519
- "<|18.34|>": 51281,
520
- "<|18.36|>": 51282,
521
- "<|18.38|>": 51283,
522
- "<|18.40|>": 51284,
523
- "<|18.42|>": 51285,
524
- "<|18.44|>": 51286,
525
- "<|18.46|>": 51287,
526
- "<|18.48|>": 51288,
527
- "<|18.50|>": 51289,
528
- "<|18.52|>": 51290,
529
- "<|18.54|>": 51291,
530
- "<|18.56|>": 51292,
531
- "<|18.58|>": 51293,
532
- "<|18.60|>": 51294,
533
- "<|18.62|>": 51295,
534
- "<|18.64|>": 51296,
535
- "<|18.66|>": 51297,
536
- "<|18.68|>": 51298,
537
- "<|18.70|>": 51299,
538
- "<|18.72|>": 51300,
539
- "<|18.74|>": 51301,
540
- "<|18.76|>": 51302,
541
- "<|18.78|>": 51303,
542
- "<|18.80|>": 51304,
543
- "<|18.82|>": 51305,
544
- "<|18.84|>": 51306,
545
- "<|18.86|>": 51307,
546
- "<|18.88|>": 51308,
547
- "<|18.90|>": 51309,
548
- "<|18.92|>": 51310,
549
- "<|18.94|>": 51311,
550
- "<|18.96|>": 51312,
551
- "<|18.98|>": 51313,
552
- "<|19.00|>": 51314,
553
- "<|19.02|>": 51315,
554
- "<|19.04|>": 51316,
555
- "<|19.06|>": 51317,
556
- "<|19.08|>": 51318,
557
- "<|19.10|>": 51319,
558
- "<|19.12|>": 51320,
559
- "<|19.14|>": 51321,
560
- "<|19.16|>": 51322,
561
- "<|19.18|>": 51323,
562
- "<|19.20|>": 51324,
563
- "<|19.22|>": 51325,
564
- "<|19.24|>": 51326,
565
- "<|19.26|>": 51327,
566
- "<|19.28|>": 51328,
567
- "<|19.30|>": 51329,
568
- "<|19.32|>": 51330,
569
- "<|19.34|>": 51331,
570
- "<|19.36|>": 51332,
571
- "<|19.38|>": 51333,
572
- "<|19.40|>": 51334,
573
- "<|19.42|>": 51335,
574
- "<|19.44|>": 51336,
575
- "<|19.46|>": 51337,
576
- "<|19.48|>": 51338,
577
- "<|19.50|>": 51339,
578
- "<|19.52|>": 51340,
579
- "<|19.54|>": 51341,
580
- "<|19.56|>": 51342,
581
- "<|19.58|>": 51343,
582
- "<|19.60|>": 51344,
583
- "<|19.62|>": 51345,
584
- "<|19.64|>": 51346,
585
- "<|19.66|>": 51347,
586
- "<|19.68|>": 51348,
587
- "<|19.70|>": 51349,
588
- "<|19.72|>": 51350,
589
- "<|19.74|>": 51351,
590
- "<|19.76|>": 51352,
591
- "<|19.78|>": 51353,
592
- "<|19.80|>": 51354,
593
- "<|19.82|>": 51355,
594
- "<|19.84|>": 51356,
595
- "<|19.86|>": 51357,
596
- "<|19.88|>": 51358,
597
- "<|19.90|>": 51359,
598
- "<|19.92|>": 51360,
599
- "<|19.94|>": 51361,
600
- "<|19.96|>": 51362,
601
- "<|19.98|>": 51363,
602
- "<|2.00|>": 50464,
603
- "<|2.02|>": 50465,
604
- "<|2.04|>": 50466,
605
- "<|2.06|>": 50467,
606
- "<|2.08|>": 50468,
607
- "<|2.10|>": 50469,
608
- "<|2.12|>": 50470,
609
- "<|2.14|>": 50471,
610
- "<|2.16|>": 50472,
611
- "<|2.18|>": 50473,
612
- "<|2.20|>": 50474,
613
- "<|2.22|>": 50475,
614
- "<|2.24|>": 50476,
615
- "<|2.26|>": 50477,
616
- "<|2.28|>": 50478,
617
- "<|2.30|>": 50479,
618
- "<|2.32|>": 50480,
619
- "<|2.34|>": 50481,
620
- "<|2.36|>": 50482,
621
- "<|2.38|>": 50483,
622
- "<|2.40|>": 50484,
623
- "<|2.42|>": 50485,
624
- "<|2.44|>": 50486,
625
- "<|2.46|>": 50487,
626
- "<|2.48|>": 50488,
627
- "<|2.50|>": 50489,
628
- "<|2.52|>": 50490,
629
- "<|2.54|>": 50491,
630
- "<|2.56|>": 50492,
631
- "<|2.58|>": 50493,
632
- "<|2.60|>": 50494,
633
- "<|2.62|>": 50495,
634
- "<|2.64|>": 50496,
635
- "<|2.66|>": 50497,
636
- "<|2.68|>": 50498,
637
- "<|2.70|>": 50499,
638
- "<|2.72|>": 50500,
639
- "<|2.74|>": 50501,
640
- "<|2.76|>": 50502,
641
- "<|2.78|>": 50503,
642
- "<|2.80|>": 50504,
643
- "<|2.82|>": 50505,
644
- "<|2.84|>": 50506,
645
- "<|2.86|>": 50507,
646
- "<|2.88|>": 50508,
647
- "<|2.90|>": 50509,
648
- "<|2.92|>": 50510,
649
- "<|2.94|>": 50511,
650
- "<|2.96|>": 50512,
651
- "<|2.98|>": 50513,
652
- "<|20.00|>": 51364,
653
- "<|20.02|>": 51365,
654
- "<|20.04|>": 51366,
655
- "<|20.06|>": 51367,
656
- "<|20.08|>": 51368,
657
- "<|20.10|>": 51369,
658
- "<|20.12|>": 51370,
659
- "<|20.14|>": 51371,
660
- "<|20.16|>": 51372,
661
- "<|20.18|>": 51373,
662
- "<|20.20|>": 51374,
663
- "<|20.22|>": 51375,
664
- "<|20.24|>": 51376,
665
- "<|20.26|>": 51377,
666
- "<|20.28|>": 51378,
667
- "<|20.30|>": 51379,
668
- "<|20.32|>": 51380,
669
- "<|20.34|>": 51381,
670
- "<|20.36|>": 51382,
671
- "<|20.38|>": 51383,
672
- "<|20.40|>": 51384,
673
- "<|20.42|>": 51385,
674
- "<|20.44|>": 51386,
675
- "<|20.46|>": 51387,
676
- "<|20.48|>": 51388,
677
- "<|20.50|>": 51389,
678
- "<|20.52|>": 51390,
679
- "<|20.54|>": 51391,
680
- "<|20.56|>": 51392,
681
- "<|20.58|>": 51393,
682
- "<|20.60|>": 51394,
683
- "<|20.62|>": 51395,
684
- "<|20.64|>": 51396,
685
- "<|20.66|>": 51397,
686
- "<|20.68|>": 51398,
687
- "<|20.70|>": 51399,
688
- "<|20.72|>": 51400,
689
- "<|20.74|>": 51401,
690
- "<|20.76|>": 51402,
691
- "<|20.78|>": 51403,
692
- "<|20.80|>": 51404,
693
- "<|20.82|>": 51405,
694
- "<|20.84|>": 51406,
695
- "<|20.86|>": 51407,
696
- "<|20.88|>": 51408,
697
- "<|20.90|>": 51409,
698
- "<|20.92|>": 51410,
699
- "<|20.94|>": 51411,
700
- "<|20.96|>": 51412,
701
- "<|20.98|>": 51413,
702
- "<|21.00|>": 51414,
703
- "<|21.02|>": 51415,
704
- "<|21.04|>": 51416,
705
- "<|21.06|>": 51417,
706
- "<|21.08|>": 51418,
707
- "<|21.10|>": 51419,
708
- "<|21.12|>": 51420,
709
- "<|21.14|>": 51421,
710
- "<|21.16|>": 51422,
711
- "<|21.18|>": 51423,
712
- "<|21.20|>": 51424,
713
- "<|21.22|>": 51425,
714
- "<|21.24|>": 51426,
715
- "<|21.26|>": 51427,
716
- "<|21.28|>": 51428,
717
- "<|21.30|>": 51429,
718
- "<|21.32|>": 51430,
719
- "<|21.34|>": 51431,
720
- "<|21.36|>": 51432,
721
- "<|21.38|>": 51433,
722
- "<|21.40|>": 51434,
723
- "<|21.42|>": 51435,
724
- "<|21.44|>": 51436,
725
- "<|21.46|>": 51437,
726
- "<|21.48|>": 51438,
727
- "<|21.50|>": 51439,
728
- "<|21.52|>": 51440,
729
- "<|21.54|>": 51441,
730
- "<|21.56|>": 51442,
731
- "<|21.58|>": 51443,
732
- "<|21.60|>": 51444,
733
- "<|21.62|>": 51445,
734
- "<|21.64|>": 51446,
735
- "<|21.66|>": 51447,
736
- "<|21.68|>": 51448,
737
- "<|21.70|>": 51449,
738
- "<|21.72|>": 51450,
739
- "<|21.74|>": 51451,
740
- "<|21.76|>": 51452,
741
- "<|21.78|>": 51453,
742
- "<|21.80|>": 51454,
743
- "<|21.82|>": 51455,
744
- "<|21.84|>": 51456,
745
- "<|21.86|>": 51457,
746
- "<|21.88|>": 51458,
747
- "<|21.90|>": 51459,
748
- "<|21.92|>": 51460,
749
- "<|21.94|>": 51461,
750
- "<|21.96|>": 51462,
751
- "<|21.98|>": 51463,
752
- "<|22.00|>": 51464,
753
- "<|22.02|>": 51465,
754
- "<|22.04|>": 51466,
755
- "<|22.06|>": 51467,
756
- "<|22.08|>": 51468,
757
- "<|22.10|>": 51469,
758
- "<|22.12|>": 51470,
759
- "<|22.14|>": 51471,
760
- "<|22.16|>": 51472,
761
- "<|22.18|>": 51473,
762
- "<|22.20|>": 51474,
763
- "<|22.22|>": 51475,
764
- "<|22.24|>": 51476,
765
- "<|22.26|>": 51477,
766
- "<|22.28|>": 51478,
767
- "<|22.30|>": 51479,
768
- "<|22.32|>": 51480,
769
- "<|22.34|>": 51481,
770
- "<|22.36|>": 51482,
771
- "<|22.38|>": 51483,
772
- "<|22.40|>": 51484,
773
- "<|22.42|>": 51485,
774
- "<|22.44|>": 51486,
775
- "<|22.46|>": 51487,
776
- "<|22.48|>": 51488,
777
- "<|22.50|>": 51489,
778
- "<|22.52|>": 51490,
779
- "<|22.54|>": 51491,
780
- "<|22.56|>": 51492,
781
- "<|22.58|>": 51493,
782
- "<|22.60|>": 51494,
783
- "<|22.62|>": 51495,
784
- "<|22.64|>": 51496,
785
- "<|22.66|>": 51497,
786
- "<|22.68|>": 51498,
787
- "<|22.70|>": 51499,
788
- "<|22.72|>": 51500,
789
- "<|22.74|>": 51501,
790
- "<|22.76|>": 51502,
791
- "<|22.78|>": 51503,
792
- "<|22.80|>": 51504,
793
- "<|22.82|>": 51505,
794
- "<|22.84|>": 51506,
795
- "<|22.86|>": 51507,
796
- "<|22.88|>": 51508,
797
- "<|22.90|>": 51509,
798
- "<|22.92|>": 51510,
799
- "<|22.94|>": 51511,
800
- "<|22.96|>": 51512,
801
- "<|22.98|>": 51513,
802
- "<|23.00|>": 51514,
803
- "<|23.02|>": 51515,
804
- "<|23.04|>": 51516,
805
- "<|23.06|>": 51517,
806
- "<|23.08|>": 51518,
807
- "<|23.10|>": 51519,
808
- "<|23.12|>": 51520,
809
- "<|23.14|>": 51521,
810
- "<|23.16|>": 51522,
811
- "<|23.18|>": 51523,
812
- "<|23.20|>": 51524,
813
- "<|23.22|>": 51525,
814
- "<|23.24|>": 51526,
815
- "<|23.26|>": 51527,
816
- "<|23.28|>": 51528,
817
- "<|23.30|>": 51529,
818
- "<|23.32|>": 51530,
819
- "<|23.34|>": 51531,
820
- "<|23.36|>": 51532,
821
- "<|23.38|>": 51533,
822
- "<|23.40|>": 51534,
823
- "<|23.42|>": 51535,
824
- "<|23.44|>": 51536,
825
- "<|23.46|>": 51537,
826
- "<|23.48|>": 51538,
827
- "<|23.50|>": 51539,
828
- "<|23.52|>": 51540,
829
- "<|23.54|>": 51541,
830
- "<|23.56|>": 51542,
831
- "<|23.58|>": 51543,
832
- "<|23.60|>": 51544,
833
- "<|23.62|>": 51545,
834
- "<|23.64|>": 51546,
835
- "<|23.66|>": 51547,
836
- "<|23.68|>": 51548,
837
- "<|23.70|>": 51549,
838
- "<|23.72|>": 51550,
839
- "<|23.74|>": 51551,
840
- "<|23.76|>": 51552,
841
- "<|23.78|>": 51553,
842
- "<|23.80|>": 51554,
843
- "<|23.82|>": 51555,
844
- "<|23.84|>": 51556,
845
- "<|23.86|>": 51557,
846
- "<|23.88|>": 51558,
847
- "<|23.90|>": 51559,
848
- "<|23.92|>": 51560,
849
- "<|23.94|>": 51561,
850
- "<|23.96|>": 51562,
851
- "<|23.98|>": 51563,
852
- "<|24.00|>": 51564,
853
- "<|24.02|>": 51565,
854
- "<|24.04|>": 51566,
855
- "<|24.06|>": 51567,
856
- "<|24.08|>": 51568,
857
- "<|24.10|>": 51569,
858
- "<|24.12|>": 51570,
859
- "<|24.14|>": 51571,
860
- "<|24.16|>": 51572,
861
- "<|24.18|>": 51573,
862
- "<|24.20|>": 51574,
863
- "<|24.22|>": 51575,
864
- "<|24.24|>": 51576,
865
- "<|24.26|>": 51577,
866
- "<|24.28|>": 51578,
867
- "<|24.30|>": 51579,
868
- "<|24.32|>": 51580,
869
- "<|24.34|>": 51581,
870
- "<|24.36|>": 51582,
871
- "<|24.38|>": 51583,
872
- "<|24.40|>": 51584,
873
- "<|24.42|>": 51585,
874
- "<|24.44|>": 51586,
875
- "<|24.46|>": 51587,
876
- "<|24.48|>": 51588,
877
- "<|24.50|>": 51589,
878
- "<|24.52|>": 51590,
879
- "<|24.54|>": 51591,
880
- "<|24.56|>": 51592,
881
- "<|24.58|>": 51593,
882
- "<|24.60|>": 51594,
883
- "<|24.62|>": 51595,
884
- "<|24.64|>": 51596,
885
- "<|24.66|>": 51597,
886
- "<|24.68|>": 51598,
887
- "<|24.70|>": 51599,
888
- "<|24.72|>": 51600,
889
- "<|24.74|>": 51601,
890
- "<|24.76|>": 51602,
891
- "<|24.78|>": 51603,
892
- "<|24.80|>": 51604,
893
- "<|24.82|>": 51605,
894
- "<|24.84|>": 51606,
895
- "<|24.86|>": 51607,
896
- "<|24.88|>": 51608,
897
- "<|24.90|>": 51609,
898
- "<|24.92|>": 51610,
899
- "<|24.94|>": 51611,
900
- "<|24.96|>": 51612,
901
- "<|24.98|>": 51613,
902
- "<|25.00|>": 51614,
903
- "<|25.02|>": 51615,
904
- "<|25.04|>": 51616,
905
- "<|25.06|>": 51617,
906
- "<|25.08|>": 51618,
907
- "<|25.10|>": 51619,
908
- "<|25.12|>": 51620,
909
- "<|25.14|>": 51621,
910
- "<|25.16|>": 51622,
911
- "<|25.18|>": 51623,
912
- "<|25.20|>": 51624,
913
- "<|25.22|>": 51625,
914
- "<|25.24|>": 51626,
915
- "<|25.26|>": 51627,
916
- "<|25.28|>": 51628,
917
- "<|25.30|>": 51629,
918
- "<|25.32|>": 51630,
919
- "<|25.34|>": 51631,
920
- "<|25.36|>": 51632,
921
- "<|25.38|>": 51633,
922
- "<|25.40|>": 51634,
923
- "<|25.42|>": 51635,
924
- "<|25.44|>": 51636,
925
- "<|25.46|>": 51637,
926
- "<|25.48|>": 51638,
927
- "<|25.50|>": 51639,
928
- "<|25.52|>": 51640,
929
- "<|25.54|>": 51641,
930
- "<|25.56|>": 51642,
931
- "<|25.58|>": 51643,
932
- "<|25.60|>": 51644,
933
- "<|25.62|>": 51645,
934
- "<|25.64|>": 51646,
935
- "<|25.66|>": 51647,
936
- "<|25.68|>": 51648,
937
- "<|25.70|>": 51649,
938
- "<|25.72|>": 51650,
939
- "<|25.74|>": 51651,
940
- "<|25.76|>": 51652,
941
- "<|25.78|>": 51653,
942
- "<|25.80|>": 51654,
943
- "<|25.82|>": 51655,
944
- "<|25.84|>": 51656,
945
- "<|25.86|>": 51657,
946
- "<|25.88|>": 51658,
947
- "<|25.90|>": 51659,
948
- "<|25.92|>": 51660,
949
- "<|25.94|>": 51661,
950
- "<|25.96|>": 51662,
951
- "<|25.98|>": 51663,
952
- "<|26.00|>": 51664,
953
- "<|26.02|>": 51665,
954
- "<|26.04|>": 51666,
955
- "<|26.06|>": 51667,
956
- "<|26.08|>": 51668,
957
- "<|26.10|>": 51669,
958
- "<|26.12|>": 51670,
959
- "<|26.14|>": 51671,
960
- "<|26.16|>": 51672,
961
- "<|26.18|>": 51673,
962
- "<|26.20|>": 51674,
963
- "<|26.22|>": 51675,
964
- "<|26.24|>": 51676,
965
- "<|26.26|>": 51677,
966
- "<|26.28|>": 51678,
967
- "<|26.30|>": 51679,
968
- "<|26.32|>": 51680,
969
- "<|26.34|>": 51681,
970
- "<|26.36|>": 51682,
971
- "<|26.38|>": 51683,
972
- "<|26.40|>": 51684,
973
- "<|26.42|>": 51685,
974
- "<|26.44|>": 51686,
975
- "<|26.46|>": 51687,
976
- "<|26.48|>": 51688,
977
- "<|26.50|>": 51689,
978
- "<|26.52|>": 51690,
979
- "<|26.54|>": 51691,
980
- "<|26.56|>": 51692,
981
- "<|26.58|>": 51693,
982
- "<|26.60|>": 51694,
983
- "<|26.62|>": 51695,
984
- "<|26.64|>": 51696,
985
- "<|26.66|>": 51697,
986
- "<|26.68|>": 51698,
987
- "<|26.70|>": 51699,
988
- "<|26.72|>": 51700,
989
- "<|26.74|>": 51701,
990
- "<|26.76|>": 51702,
991
- "<|26.78|>": 51703,
992
- "<|26.80|>": 51704,
993
- "<|26.82|>": 51705,
994
- "<|26.84|>": 51706,
995
- "<|26.86|>": 51707,
996
- "<|26.88|>": 51708,
997
- "<|26.90|>": 51709,
998
- "<|26.92|>": 51710,
999
- "<|26.94|>": 51711,
1000
- "<|26.96|>": 51712,
1001
- "<|26.98|>": 51713,
1002
- "<|27.00|>": 51714,
1003
- "<|27.02|>": 51715,
1004
- "<|27.04|>": 51716,
1005
- "<|27.06|>": 51717,
1006
- "<|27.08|>": 51718,
1007
- "<|27.10|>": 51719,
1008
- "<|27.12|>": 51720,
1009
- "<|27.14|>": 51721,
1010
- "<|27.16|>": 51722,
1011
- "<|27.18|>": 51723,
1012
- "<|27.20|>": 51724,
1013
- "<|27.22|>": 51725,
1014
- "<|27.24|>": 51726,
1015
- "<|27.26|>": 51727,
1016
- "<|27.28|>": 51728,
1017
- "<|27.30|>": 51729,
1018
- "<|27.32|>": 51730,
1019
- "<|27.34|>": 51731,
1020
- "<|27.36|>": 51732,
1021
- "<|27.38|>": 51733,
1022
- "<|27.40|>": 51734,
1023
- "<|27.42|>": 51735,
1024
- "<|27.44|>": 51736,
1025
- "<|27.46|>": 51737,
1026
- "<|27.48|>": 51738,
1027
- "<|27.50|>": 51739,
1028
- "<|27.52|>": 51740,
1029
- "<|27.54|>": 51741,
1030
- "<|27.56|>": 51742,
1031
- "<|27.58|>": 51743,
1032
- "<|27.60|>": 51744,
1033
- "<|27.62|>": 51745,
1034
- "<|27.64|>": 51746,
1035
- "<|27.66|>": 51747,
1036
- "<|27.68|>": 51748,
1037
- "<|27.70|>": 51749,
1038
- "<|27.72|>": 51750,
1039
- "<|27.74|>": 51751,
1040
- "<|27.76|>": 51752,
1041
- "<|27.78|>": 51753,
1042
- "<|27.80|>": 51754,
1043
- "<|27.82|>": 51755,
1044
- "<|27.84|>": 51756,
1045
- "<|27.86|>": 51757,
1046
- "<|27.88|>": 51758,
1047
- "<|27.90|>": 51759,
1048
- "<|27.92|>": 51760,
1049
- "<|27.94|>": 51761,
1050
- "<|27.96|>": 51762,
1051
- "<|27.98|>": 51763,
1052
- "<|28.00|>": 51764,
1053
- "<|28.02|>": 51765,
1054
- "<|28.04|>": 51766,
1055
- "<|28.06|>": 51767,
1056
- "<|28.08|>": 51768,
1057
- "<|28.10|>": 51769,
1058
- "<|28.12|>": 51770,
1059
- "<|28.14|>": 51771,
1060
- "<|28.16|>": 51772,
1061
- "<|28.18|>": 51773,
1062
- "<|28.20|>": 51774,
1063
- "<|28.22|>": 51775,
1064
- "<|28.24|>": 51776,
1065
- "<|28.26|>": 51777,
1066
- "<|28.28|>": 51778,
1067
- "<|28.30|>": 51779,
1068
- "<|28.32|>": 51780,
1069
- "<|28.34|>": 51781,
1070
- "<|28.36|>": 51782,
1071
- "<|28.38|>": 51783,
1072
- "<|28.40|>": 51784,
1073
- "<|28.42|>": 51785,
1074
- "<|28.44|>": 51786,
1075
- "<|28.46|>": 51787,
1076
- "<|28.48|>": 51788,
1077
- "<|28.50|>": 51789,
1078
- "<|28.52|>": 51790,
1079
- "<|28.54|>": 51791,
1080
- "<|28.56|>": 51792,
1081
- "<|28.58|>": 51793,
1082
- "<|28.60|>": 51794,
1083
- "<|28.62|>": 51795,
1084
- "<|28.64|>": 51796,
1085
- "<|28.66|>": 51797,
1086
- "<|28.68|>": 51798,
1087
- "<|28.70|>": 51799,
1088
- "<|28.72|>": 51800,
1089
- "<|28.74|>": 51801,
1090
- "<|28.76|>": 51802,
1091
- "<|28.78|>": 51803,
1092
- "<|28.80|>": 51804,
1093
- "<|28.82|>": 51805,
1094
- "<|28.84|>": 51806,
1095
- "<|28.86|>": 51807,
1096
- "<|28.88|>": 51808,
1097
- "<|28.90|>": 51809,
1098
- "<|28.92|>": 51810,
1099
- "<|28.94|>": 51811,
1100
- "<|28.96|>": 51812,
1101
- "<|28.98|>": 51813,
1102
- "<|29.00|>": 51814,
1103
- "<|29.02|>": 51815,
1104
- "<|29.04|>": 51816,
1105
- "<|29.06|>": 51817,
1106
- "<|29.08|>": 51818,
1107
- "<|29.10|>": 51819,
1108
- "<|29.12|>": 51820,
1109
- "<|29.14|>": 51821,
1110
- "<|29.16|>": 51822,
1111
- "<|29.18|>": 51823,
1112
- "<|29.20|>": 51824,
1113
- "<|29.22|>": 51825,
1114
- "<|29.24|>": 51826,
1115
- "<|29.26|>": 51827,
1116
- "<|29.28|>": 51828,
1117
- "<|29.30|>": 51829,
1118
- "<|29.32|>": 51830,
1119
- "<|29.34|>": 51831,
1120
- "<|29.36|>": 51832,
1121
- "<|29.38|>": 51833,
1122
- "<|29.40|>": 51834,
1123
- "<|29.42|>": 51835,
1124
- "<|29.44|>": 51836,
1125
- "<|29.46|>": 51837,
1126
- "<|29.48|>": 51838,
1127
- "<|29.50|>": 51839,
1128
- "<|29.52|>": 51840,
1129
- "<|29.54|>": 51841,
1130
- "<|29.56|>": 51842,
1131
- "<|29.58|>": 51843,
1132
- "<|29.60|>": 51844,
1133
- "<|29.62|>": 51845,
1134
- "<|29.64|>": 51846,
1135
- "<|29.66|>": 51847,
1136
- "<|29.68|>": 51848,
1137
- "<|29.70|>": 51849,
1138
- "<|29.72|>": 51850,
1139
- "<|29.74|>": 51851,
1140
- "<|29.76|>": 51852,
1141
- "<|29.78|>": 51853,
1142
- "<|29.80|>": 51854,
1143
- "<|29.82|>": 51855,
1144
- "<|29.84|>": 51856,
1145
- "<|29.86|>": 51857,
1146
- "<|29.88|>": 51858,
1147
- "<|29.90|>": 51859,
1148
- "<|29.92|>": 51860,
1149
- "<|29.94|>": 51861,
1150
- "<|29.96|>": 51862,
1151
- "<|29.98|>": 51863,
1152
- "<|3.00|>": 50514,
1153
- "<|3.02|>": 50515,
1154
- "<|3.04|>": 50516,
1155
- "<|3.06|>": 50517,
1156
- "<|3.08|>": 50518,
1157
- "<|3.10|>": 50519,
1158
- "<|3.12|>": 50520,
1159
- "<|3.14|>": 50521,
1160
- "<|3.16|>": 50522,
1161
- "<|3.18|>": 50523,
1162
- "<|3.20|>": 50524,
1163
- "<|3.22|>": 50525,
1164
- "<|3.24|>": 50526,
1165
- "<|3.26|>": 50527,
1166
- "<|3.28|>": 50528,
1167
- "<|3.30|>": 50529,
1168
- "<|3.32|>": 50530,
1169
- "<|3.34|>": 50531,
1170
- "<|3.36|>": 50532,
1171
- "<|3.38|>": 50533,
1172
- "<|3.40|>": 50534,
1173
- "<|3.42|>": 50535,
1174
- "<|3.44|>": 50536,
1175
- "<|3.46|>": 50537,
1176
- "<|3.48|>": 50538,
1177
- "<|3.50|>": 50539,
1178
- "<|3.52|>": 50540,
1179
- "<|3.54|>": 50541,
1180
- "<|3.56|>": 50542,
1181
- "<|3.58|>": 50543,
1182
- "<|3.60|>": 50544,
1183
- "<|3.62|>": 50545,
1184
- "<|3.64|>": 50546,
1185
- "<|3.66|>": 50547,
1186
- "<|3.68|>": 50548,
1187
- "<|3.70|>": 50549,
1188
- "<|3.72|>": 50550,
1189
- "<|3.74|>": 50551,
1190
- "<|3.76|>": 50552,
1191
- "<|3.78|>": 50553,
1192
- "<|3.80|>": 50554,
1193
- "<|3.82|>": 50555,
1194
- "<|3.84|>": 50556,
1195
- "<|3.86|>": 50557,
1196
- "<|3.88|>": 50558,
1197
- "<|3.90|>": 50559,
1198
- "<|3.92|>": 50560,
1199
- "<|3.94|>": 50561,
1200
- "<|3.96|>": 50562,
1201
- "<|3.98|>": 50563,
1202
- "<|30.00|>": 51864,
1203
- "<|4.00|>": 50564,
1204
- "<|4.02|>": 50565,
1205
- "<|4.04|>": 50566,
1206
- "<|4.06|>": 50567,
1207
- "<|4.08|>": 50568,
1208
- "<|4.10|>": 50569,
1209
- "<|4.12|>": 50570,
1210
- "<|4.14|>": 50571,
1211
- "<|4.16|>": 50572,
1212
- "<|4.18|>": 50573,
1213
- "<|4.20|>": 50574,
1214
- "<|4.22|>": 50575,
1215
- "<|4.24|>": 50576,
1216
- "<|4.26|>": 50577,
1217
- "<|4.28|>": 50578,
1218
- "<|4.30|>": 50579,
1219
- "<|4.32|>": 50580,
1220
- "<|4.34|>": 50581,
1221
- "<|4.36|>": 50582,
1222
- "<|4.38|>": 50583,
1223
- "<|4.40|>": 50584,
1224
- "<|4.42|>": 50585,
1225
- "<|4.44|>": 50586,
1226
- "<|4.46|>": 50587,
1227
- "<|4.48|>": 50588,
1228
- "<|4.50|>": 50589,
1229
- "<|4.52|>": 50590,
1230
- "<|4.54|>": 50591,
1231
- "<|4.56|>": 50592,
1232
- "<|4.58|>": 50593,
1233
- "<|4.60|>": 50594,
1234
- "<|4.62|>": 50595,
1235
- "<|4.64|>": 50596,
1236
- "<|4.66|>": 50597,
1237
- "<|4.68|>": 50598,
1238
- "<|4.70|>": 50599,
1239
- "<|4.72|>": 50600,
1240
- "<|4.74|>": 50601,
1241
- "<|4.76|>": 50602,
1242
- "<|4.78|>": 50603,
1243
- "<|4.80|>": 50604,
1244
- "<|4.82|>": 50605,
1245
- "<|4.84|>": 50606,
1246
- "<|4.86|>": 50607,
1247
- "<|4.88|>": 50608,
1248
- "<|4.90|>": 50609,
1249
- "<|4.92|>": 50610,
1250
- "<|4.94|>": 50611,
1251
- "<|4.96|>": 50612,
1252
- "<|4.98|>": 50613,
1253
- "<|5.00|>": 50614,
1254
- "<|5.02|>": 50615,
1255
- "<|5.04|>": 50616,
1256
- "<|5.06|>": 50617,
1257
- "<|5.08|>": 50618,
1258
- "<|5.10|>": 50619,
1259
- "<|5.12|>": 50620,
1260
- "<|5.14|>": 50621,
1261
- "<|5.16|>": 50622,
1262
- "<|5.18|>": 50623,
1263
- "<|5.20|>": 50624,
1264
- "<|5.22|>": 50625,
1265
- "<|5.24|>": 50626,
1266
- "<|5.26|>": 50627,
1267
- "<|5.28|>": 50628,
1268
- "<|5.30|>": 50629,
1269
- "<|5.32|>": 50630,
1270
- "<|5.34|>": 50631,
1271
- "<|5.36|>": 50632,
1272
- "<|5.38|>": 50633,
1273
- "<|5.40|>": 50634,
1274
- "<|5.42|>": 50635,
1275
- "<|5.44|>": 50636,
1276
- "<|5.46|>": 50637,
1277
- "<|5.48|>": 50638,
1278
- "<|5.50|>": 50639,
1279
- "<|5.52|>": 50640,
1280
- "<|5.54|>": 50641,
1281
- "<|5.56|>": 50642,
1282
- "<|5.58|>": 50643,
1283
- "<|5.60|>": 50644,
1284
- "<|5.62|>": 50645,
1285
- "<|5.64|>": 50646,
1286
- "<|5.66|>": 50647,
1287
- "<|5.68|>": 50648,
1288
- "<|5.70|>": 50649,
1289
- "<|5.72|>": 50650,
1290
- "<|5.74|>": 50651,
1291
- "<|5.76|>": 50652,
1292
- "<|5.78|>": 50653,
1293
- "<|5.80|>": 50654,
1294
- "<|5.82|>": 50655,
1295
- "<|5.84|>": 50656,
1296
- "<|5.86|>": 50657,
1297
- "<|5.88|>": 50658,
1298
- "<|5.90|>": 50659,
1299
- "<|5.92|>": 50660,
1300
- "<|5.94|>": 50661,
1301
- "<|5.96|>": 50662,
1302
- "<|5.98|>": 50663,
1303
- "<|6.00|>": 50664,
1304
- "<|6.02|>": 50665,
1305
- "<|6.04|>": 50666,
1306
- "<|6.06|>": 50667,
1307
- "<|6.08|>": 50668,
1308
- "<|6.10|>": 50669,
1309
- "<|6.12|>": 50670,
1310
- "<|6.14|>": 50671,
1311
- "<|6.16|>": 50672,
1312
- "<|6.18|>": 50673,
1313
- "<|6.20|>": 50674,
1314
- "<|6.22|>": 50675,
1315
- "<|6.24|>": 50676,
1316
- "<|6.26|>": 50677,
1317
- "<|6.28|>": 50678,
1318
- "<|6.30|>": 50679,
1319
- "<|6.32|>": 50680,
1320
- "<|6.34|>": 50681,
1321
- "<|6.36|>": 50682,
1322
- "<|6.38|>": 50683,
1323
- "<|6.40|>": 50684,
1324
- "<|6.42|>": 50685,
1325
- "<|6.44|>": 50686,
1326
- "<|6.46|>": 50687,
1327
- "<|6.48|>": 50688,
1328
- "<|6.50|>": 50689,
1329
- "<|6.52|>": 50690,
1330
- "<|6.54|>": 50691,
1331
- "<|6.56|>": 50692,
1332
- "<|6.58|>": 50693,
1333
- "<|6.60|>": 50694,
1334
- "<|6.62|>": 50695,
1335
- "<|6.64|>": 50696,
1336
- "<|6.66|>": 50697,
1337
- "<|6.68|>": 50698,
1338
- "<|6.70|>": 50699,
1339
- "<|6.72|>": 50700,
1340
- "<|6.74|>": 50701,
1341
- "<|6.76|>": 50702,
1342
- "<|6.78|>": 50703,
1343
- "<|6.80|>": 50704,
1344
- "<|6.82|>": 50705,
1345
- "<|6.84|>": 50706,
1346
- "<|6.86|>": 50707,
1347
- "<|6.88|>": 50708,
1348
- "<|6.90|>": 50709,
1349
- "<|6.92|>": 50710,
1350
- "<|6.94|>": 50711,
1351
- "<|6.96|>": 50712,
1352
- "<|6.98|>": 50713,
1353
- "<|7.00|>": 50714,
1354
- "<|7.02|>": 50715,
1355
- "<|7.04|>": 50716,
1356
- "<|7.06|>": 50717,
1357
- "<|7.08|>": 50718,
1358
- "<|7.10|>": 50719,
1359
- "<|7.12|>": 50720,
1360
- "<|7.14|>": 50721,
1361
- "<|7.16|>": 50722,
1362
- "<|7.18|>": 50723,
1363
- "<|7.20|>": 50724,
1364
- "<|7.22|>": 50725,
1365
- "<|7.24|>": 50726,
1366
- "<|7.26|>": 50727,
1367
- "<|7.28|>": 50728,
1368
- "<|7.30|>": 50729,
1369
- "<|7.32|>": 50730,
1370
- "<|7.34|>": 50731,
1371
- "<|7.36|>": 50732,
1372
- "<|7.38|>": 50733,
1373
- "<|7.40|>": 50734,
1374
- "<|7.42|>": 50735,
1375
- "<|7.44|>": 50736,
1376
- "<|7.46|>": 50737,
1377
- "<|7.48|>": 50738,
1378
- "<|7.50|>": 50739,
1379
- "<|7.52|>": 50740,
1380
- "<|7.54|>": 50741,
1381
- "<|7.56|>": 50742,
1382
- "<|7.58|>": 50743,
1383
- "<|7.60|>": 50744,
1384
- "<|7.62|>": 50745,
1385
- "<|7.64|>": 50746,
1386
- "<|7.66|>": 50747,
1387
- "<|7.68|>": 50748,
1388
- "<|7.70|>": 50749,
1389
- "<|7.72|>": 50750,
1390
- "<|7.74|>": 50751,
1391
- "<|7.76|>": 50752,
1392
- "<|7.78|>": 50753,
1393
- "<|7.80|>": 50754,
1394
- "<|7.82|>": 50755,
1395
- "<|7.84|>": 50756,
1396
- "<|7.86|>": 50757,
1397
- "<|7.88|>": 50758,
1398
- "<|7.90|>": 50759,
1399
- "<|7.92|>": 50760,
1400
- "<|7.94|>": 50761,
1401
- "<|7.96|>": 50762,
1402
- "<|7.98|>": 50763,
1403
- "<|8.00|>": 50764,
1404
- "<|8.02|>": 50765,
1405
- "<|8.04|>": 50766,
1406
- "<|8.06|>": 50767,
1407
- "<|8.08|>": 50768,
1408
- "<|8.10|>": 50769,
1409
- "<|8.12|>": 50770,
1410
- "<|8.14|>": 50771,
1411
- "<|8.16|>": 50772,
1412
- "<|8.18|>": 50773,
1413
- "<|8.20|>": 50774,
1414
- "<|8.22|>": 50775,
1415
- "<|8.24|>": 50776,
1416
- "<|8.26|>": 50777,
1417
- "<|8.28|>": 50778,
1418
- "<|8.30|>": 50779,
1419
- "<|8.32|>": 50780,
1420
- "<|8.34|>": 50781,
1421
- "<|8.36|>": 50782,
1422
- "<|8.38|>": 50783,
1423
- "<|8.40|>": 50784,
1424
- "<|8.42|>": 50785,
1425
- "<|8.44|>": 50786,
1426
- "<|8.46|>": 50787,
1427
- "<|8.48|>": 50788,
1428
- "<|8.50|>": 50789,
1429
- "<|8.52|>": 50790,
1430
- "<|8.54|>": 50791,
1431
- "<|8.56|>": 50792,
1432
- "<|8.58|>": 50793,
1433
- "<|8.60|>": 50794,
1434
- "<|8.62|>": 50795,
1435
- "<|8.64|>": 50796,
1436
- "<|8.66|>": 50797,
1437
- "<|8.68|>": 50798,
1438
- "<|8.70|>": 50799,
1439
- "<|8.72|>": 50800,
1440
- "<|8.74|>": 50801,
1441
- "<|8.76|>": 50802,
1442
- "<|8.78|>": 50803,
1443
- "<|8.80|>": 50804,
1444
- "<|8.82|>": 50805,
1445
- "<|8.84|>": 50806,
1446
- "<|8.86|>": 50807,
1447
- "<|8.88|>": 50808,
1448
- "<|8.90|>": 50809,
1449
- "<|8.92|>": 50810,
1450
- "<|8.94|>": 50811,
1451
- "<|8.96|>": 50812,
1452
- "<|8.98|>": 50813,
1453
- "<|9.00|>": 50814,
1454
- "<|9.02|>": 50815,
1455
- "<|9.04|>": 50816,
1456
- "<|9.06|>": 50817,
1457
- "<|9.08|>": 50818,
1458
- "<|9.10|>": 50819,
1459
- "<|9.12|>": 50820,
1460
- "<|9.14|>": 50821,
1461
- "<|9.16|>": 50822,
1462
- "<|9.18|>": 50823,
1463
- "<|9.20|>": 50824,
1464
- "<|9.22|>": 50825,
1465
- "<|9.24|>": 50826,
1466
- "<|9.26|>": 50827,
1467
- "<|9.28|>": 50828,
1468
- "<|9.30|>": 50829,
1469
- "<|9.32|>": 50830,
1470
- "<|9.34|>": 50831,
1471
- "<|9.36|>": 50832,
1472
- "<|9.38|>": 50833,
1473
- "<|9.40|>": 50834,
1474
- "<|9.42|>": 50835,
1475
- "<|9.44|>": 50836,
1476
- "<|9.46|>": 50837,
1477
- "<|9.48|>": 50838,
1478
- "<|9.50|>": 50839,
1479
- "<|9.52|>": 50840,
1480
- "<|9.54|>": 50841,
1481
- "<|9.56|>": 50842,
1482
- "<|9.58|>": 50843,
1483
- "<|9.60|>": 50844,
1484
- "<|9.62|>": 50845,
1485
- "<|9.64|>": 50846,
1486
- "<|9.66|>": 50847,
1487
- "<|9.68|>": 50848,
1488
- "<|9.70|>": 50849,
1489
- "<|9.72|>": 50850,
1490
- "<|9.74|>": 50851,
1491
- "<|9.76|>": 50852,
1492
- "<|9.78|>": 50853,
1493
- "<|9.80|>": 50854,
1494
- "<|9.82|>": 50855,
1495
- "<|9.84|>": 50856,
1496
- "<|9.86|>": 50857,
1497
- "<|9.88|>": 50858,
1498
- "<|9.90|>": 50859,
1499
- "<|9.92|>": 50860,
1500
- "<|9.94|>": 50861,
1501
- "<|9.96|>": 50862,
1502
- "<|9.98|>": 50863,
1503
  "<|af|>": 50327,
1504
  "<|am|>": 50334,
1505
  "<|ar|>": 50272,
@@ -1518,6 +1518,7 @@
1518
  "<|da|>": 50285,
1519
  "<|de|>": 50261,
1520
  "<|el|>": 50281,
 
1521
  "<|en|>": 50259,
1522
  "<|es|>": 50262,
1523
  "<|et|>": 50307,
@@ -1564,8 +1565,8 @@
1564
  "<|ne|>": 50313,
1565
  "<|nl|>": 50271,
1566
  "<|nn|>": 50342,
1567
- "<|nocaptions|>": 50362,
1568
- "<|notimestamps|>": 50363,
1569
  "<|no|>": 50288,
1570
  "<|oc|>": 50328,
1571
  "<|pa|>": 50321,
@@ -1583,8 +1584,8 @@
1583
  "<|so|>": 50326,
1584
  "<|sq|>": 50317,
1585
  "<|sr|>": 50303,
1586
- "<|startoflm|>": 50360,
1587
- "<|startofprev|>": 50361,
1588
  "<|startoftranscript|>": 50258,
1589
  "<|su|>": 50357,
1590
  "<|sv|>": 50273,
@@ -1595,8 +1596,8 @@
1595
  "<|th|>": 50289,
1596
  "<|tk|>": 50341,
1597
  "<|tl|>": 50348,
1598
- "<|transcribe|>": 50359,
1599
- "<|translate|>": 50358,
1600
  "<|tr|>": 50268,
1601
  "<|tt|>": 50351,
1602
  "<|uk|>": 50280,
@@ -1605,5 +1606,6 @@
1605
  "<|vi|>": 50278,
1606
  "<|yi|>": 50335,
1607
  "<|yo|>": 50325,
 
1608
  "<|zh|>": 50260
1609
  }
 
1
  {
2
+ "<|0.00|>": 50365,
3
+ "<|0.02|>": 50366,
4
+ "<|0.04|>": 50367,
5
+ "<|0.06|>": 50368,
6
+ "<|0.08|>": 50369,
7
+ "<|0.10|>": 50370,
8
+ "<|0.12|>": 50371,
9
+ "<|0.14|>": 50372,
10
+ "<|0.16|>": 50373,
11
+ "<|0.18|>": 50374,
12
+ "<|0.20|>": 50375,
13
+ "<|0.22|>": 50376,
14
+ "<|0.24|>": 50377,
15
+ "<|0.26|>": 50378,
16
+ "<|0.28|>": 50379,
17
+ "<|0.30|>": 50380,
18
+ "<|0.32|>": 50381,
19
+ "<|0.34|>": 50382,
20
+ "<|0.36|>": 50383,
21
+ "<|0.38|>": 50384,
22
+ "<|0.40|>": 50385,
23
+ "<|0.42|>": 50386,
24
+ "<|0.44|>": 50387,
25
+ "<|0.46|>": 50388,
26
+ "<|0.48|>": 50389,
27
+ "<|0.50|>": 50390,
28
+ "<|0.52|>": 50391,
29
+ "<|0.54|>": 50392,
30
+ "<|0.56|>": 50393,
31
+ "<|0.58|>": 50394,
32
+ "<|0.60|>": 50395,
33
+ "<|0.62|>": 50396,
34
+ "<|0.64|>": 50397,
35
+ "<|0.66|>": 50398,
36
+ "<|0.68|>": 50399,
37
+ "<|0.70|>": 50400,
38
+ "<|0.72|>": 50401,
39
+ "<|0.74|>": 50402,
40
+ "<|0.76|>": 50403,
41
+ "<|0.78|>": 50404,
42
+ "<|0.80|>": 50405,
43
+ "<|0.82|>": 50406,
44
+ "<|0.84|>": 50407,
45
+ "<|0.86|>": 50408,
46
+ "<|0.88|>": 50409,
47
+ "<|0.90|>": 50410,
48
+ "<|0.92|>": 50411,
49
+ "<|0.94|>": 50412,
50
+ "<|0.96|>": 50413,
51
+ "<|0.98|>": 50414,
52
+ "<|1.00|>": 50415,
53
+ "<|1.02|>": 50416,
54
+ "<|1.04|>": 50417,
55
+ "<|1.06|>": 50418,
56
+ "<|1.08|>": 50419,
57
+ "<|1.10|>": 50420,
58
+ "<|1.12|>": 50421,
59
+ "<|1.14|>": 50422,
60
+ "<|1.16|>": 50423,
61
+ "<|1.18|>": 50424,
62
+ "<|1.20|>": 50425,
63
+ "<|1.22|>": 50426,
64
+ "<|1.24|>": 50427,
65
+ "<|1.26|>": 50428,
66
+ "<|1.28|>": 50429,
67
+ "<|1.30|>": 50430,
68
+ "<|1.32|>": 50431,
69
+ "<|1.34|>": 50432,
70
+ "<|1.36|>": 50433,
71
+ "<|1.38|>": 50434,
72
+ "<|1.40|>": 50435,
73
+ "<|1.42|>": 50436,
74
+ "<|1.44|>": 50437,
75
+ "<|1.46|>": 50438,
76
+ "<|1.48|>": 50439,
77
+ "<|1.50|>": 50440,
78
+ "<|1.52|>": 50441,
79
+ "<|1.54|>": 50442,
80
+ "<|1.56|>": 50443,
81
+ "<|1.58|>": 50444,
82
+ "<|1.60|>": 50445,
83
+ "<|1.62|>": 50446,
84
+ "<|1.64|>": 50447,
85
+ "<|1.66|>": 50448,
86
+ "<|1.68|>": 50449,
87
+ "<|1.70|>": 50450,
88
+ "<|1.72|>": 50451,
89
+ "<|1.74|>": 50452,
90
+ "<|1.76|>": 50453,
91
+ "<|1.78|>": 50454,
92
+ "<|1.80|>": 50455,
93
+ "<|1.82|>": 50456,
94
+ "<|1.84|>": 50457,
95
+ "<|1.86|>": 50458,
96
+ "<|1.88|>": 50459,
97
+ "<|1.90|>": 50460,
98
+ "<|1.92|>": 50461,
99
+ "<|1.94|>": 50462,
100
+ "<|1.96|>": 50463,
101
+ "<|1.98|>": 50464,
102
+ "<|10.00|>": 50865,
103
+ "<|10.02|>": 50866,
104
+ "<|10.04|>": 50867,
105
+ "<|10.06|>": 50868,
106
+ "<|10.08|>": 50869,
107
+ "<|10.10|>": 50870,
108
+ "<|10.12|>": 50871,
109
+ "<|10.14|>": 50872,
110
+ "<|10.16|>": 50873,
111
+ "<|10.18|>": 50874,
112
+ "<|10.20|>": 50875,
113
+ "<|10.22|>": 50876,
114
+ "<|10.24|>": 50877,
115
+ "<|10.26|>": 50878,
116
+ "<|10.28|>": 50879,
117
+ "<|10.30|>": 50880,
118
+ "<|10.32|>": 50881,
119
+ "<|10.34|>": 50882,
120
+ "<|10.36|>": 50883,
121
+ "<|10.38|>": 50884,
122
+ "<|10.40|>": 50885,
123
+ "<|10.42|>": 50886,
124
+ "<|10.44|>": 50887,
125
+ "<|10.46|>": 50888,
126
+ "<|10.48|>": 50889,
127
+ "<|10.50|>": 50890,
128
+ "<|10.52|>": 50891,
129
+ "<|10.54|>": 50892,
130
+ "<|10.56|>": 50893,
131
+ "<|10.58|>": 50894,
132
+ "<|10.60|>": 50895,
133
+ "<|10.62|>": 50896,
134
+ "<|10.64|>": 50897,
135
+ "<|10.66|>": 50898,
136
+ "<|10.68|>": 50899,
137
+ "<|10.70|>": 50900,
138
+ "<|10.72|>": 50901,
139
+ "<|10.74|>": 50902,
140
+ "<|10.76|>": 50903,
141
+ "<|10.78|>": 50904,
142
+ "<|10.80|>": 50905,
143
+ "<|10.82|>": 50906,
144
+ "<|10.84|>": 50907,
145
+ "<|10.86|>": 50908,
146
+ "<|10.88|>": 50909,
147
+ "<|10.90|>": 50910,
148
+ "<|10.92|>": 50911,
149
+ "<|10.94|>": 50912,
150
+ "<|10.96|>": 50913,
151
+ "<|10.98|>": 50914,
152
+ "<|11.00|>": 50915,
153
+ "<|11.02|>": 50916,
154
+ "<|11.04|>": 50917,
155
+ "<|11.06|>": 50918,
156
+ "<|11.08|>": 50919,
157
+ "<|11.10|>": 50920,
158
+ "<|11.12|>": 50921,
159
+ "<|11.14|>": 50922,
160
+ "<|11.16|>": 50923,
161
+ "<|11.18|>": 50924,
162
+ "<|11.20|>": 50925,
163
+ "<|11.22|>": 50926,
164
+ "<|11.24|>": 50927,
165
+ "<|11.26|>": 50928,
166
+ "<|11.28|>": 50929,
167
+ "<|11.30|>": 50930,
168
+ "<|11.32|>": 50931,
169
+ "<|11.34|>": 50932,
170
+ "<|11.36|>": 50933,
171
+ "<|11.38|>": 50934,
172
+ "<|11.40|>": 50935,
173
+ "<|11.42|>": 50936,
174
+ "<|11.44|>": 50937,
175
+ "<|11.46|>": 50938,
176
+ "<|11.48|>": 50939,
177
+ "<|11.50|>": 50940,
178
+ "<|11.52|>": 50941,
179
+ "<|11.54|>": 50942,
180
+ "<|11.56|>": 50943,
181
+ "<|11.58|>": 50944,
182
+ "<|11.60|>": 50945,
183
+ "<|11.62|>": 50946,
184
+ "<|11.64|>": 50947,
185
+ "<|11.66|>": 50948,
186
+ "<|11.68|>": 50949,
187
+ "<|11.70|>": 50950,
188
+ "<|11.72|>": 50951,
189
+ "<|11.74|>": 50952,
190
+ "<|11.76|>": 50953,
191
+ "<|11.78|>": 50954,
192
+ "<|11.80|>": 50955,
193
+ "<|11.82|>": 50956,
194
+ "<|11.84|>": 50957,
195
+ "<|11.86|>": 50958,
196
+ "<|11.88|>": 50959,
197
+ "<|11.90|>": 50960,
198
+ "<|11.92|>": 50961,
199
+ "<|11.94|>": 50962,
200
+ "<|11.96|>": 50963,
201
+ "<|11.98|>": 50964,
202
+ "<|12.00|>": 50965,
203
+ "<|12.02|>": 50966,
204
+ "<|12.04|>": 50967,
205
+ "<|12.06|>": 50968,
206
+ "<|12.08|>": 50969,
207
+ "<|12.10|>": 50970,
208
+ "<|12.12|>": 50971,
209
+ "<|12.14|>": 50972,
210
+ "<|12.16|>": 50973,
211
+ "<|12.18|>": 50974,
212
+ "<|12.20|>": 50975,
213
+ "<|12.22|>": 50976,
214
+ "<|12.24|>": 50977,
215
+ "<|12.26|>": 50978,
216
+ "<|12.28|>": 50979,
217
+ "<|12.30|>": 50980,
218
+ "<|12.32|>": 50981,
219
+ "<|12.34|>": 50982,
220
+ "<|12.36|>": 50983,
221
+ "<|12.38|>": 50984,
222
+ "<|12.40|>": 50985,
223
+ "<|12.42|>": 50986,
224
+ "<|12.44|>": 50987,
225
+ "<|12.46|>": 50988,
226
+ "<|12.48|>": 50989,
227
+ "<|12.50|>": 50990,
228
+ "<|12.52|>": 50991,
229
+ "<|12.54|>": 50992,
230
+ "<|12.56|>": 50993,
231
+ "<|12.58|>": 50994,
232
+ "<|12.60|>": 50995,
233
+ "<|12.62|>": 50996,
234
+ "<|12.64|>": 50997,
235
+ "<|12.66|>": 50998,
236
+ "<|12.68|>": 50999,
237
+ "<|12.70|>": 51000,
238
+ "<|12.72|>": 51001,
239
+ "<|12.74|>": 51002,
240
+ "<|12.76|>": 51003,
241
+ "<|12.78|>": 51004,
242
+ "<|12.80|>": 51005,
243
+ "<|12.82|>": 51006,
244
+ "<|12.84|>": 51007,
245
+ "<|12.86|>": 51008,
246
+ "<|12.88|>": 51009,
247
+ "<|12.90|>": 51010,
248
+ "<|12.92|>": 51011,
249
+ "<|12.94|>": 51012,
250
+ "<|12.96|>": 51013,
251
+ "<|12.98|>": 51014,
252
+ "<|13.00|>": 51015,
253
+ "<|13.02|>": 51016,
254
+ "<|13.04|>": 51017,
255
+ "<|13.06|>": 51018,
256
+ "<|13.08|>": 51019,
257
+ "<|13.10|>": 51020,
258
+ "<|13.12|>": 51021,
259
+ "<|13.14|>": 51022,
260
+ "<|13.16|>": 51023,
261
+ "<|13.18|>": 51024,
262
+ "<|13.20|>": 51025,
263
+ "<|13.22|>": 51026,
264
+ "<|13.24|>": 51027,
265
+ "<|13.26|>": 51028,
266
+ "<|13.28|>": 51029,
267
+ "<|13.30|>": 51030,
268
+ "<|13.32|>": 51031,
269
+ "<|13.34|>": 51032,
270
+ "<|13.36|>": 51033,
271
+ "<|13.38|>": 51034,
272
+ "<|13.40|>": 51035,
273
+ "<|13.42|>": 51036,
274
+ "<|13.44|>": 51037,
275
+ "<|13.46|>": 51038,
276
+ "<|13.48|>": 51039,
277
+ "<|13.50|>": 51040,
278
+ "<|13.52|>": 51041,
279
+ "<|13.54|>": 51042,
280
+ "<|13.56|>": 51043,
281
+ "<|13.58|>": 51044,
282
+ "<|13.60|>": 51045,
283
+ "<|13.62|>": 51046,
284
+ "<|13.64|>": 51047,
285
+ "<|13.66|>": 51048,
286
+ "<|13.68|>": 51049,
287
+ "<|13.70|>": 51050,
288
+ "<|13.72|>": 51051,
289
+ "<|13.74|>": 51052,
290
+ "<|13.76|>": 51053,
291
+ "<|13.78|>": 51054,
292
+ "<|13.80|>": 51055,
293
+ "<|13.82|>": 51056,
294
+ "<|13.84|>": 51057,
295
+ "<|13.86|>": 51058,
296
+ "<|13.88|>": 51059,
297
+ "<|13.90|>": 51060,
298
+ "<|13.92|>": 51061,
299
+ "<|13.94|>": 51062,
300
+ "<|13.96|>": 51063,
301
+ "<|13.98|>": 51064,
302
+ "<|14.00|>": 51065,
303
+ "<|14.02|>": 51066,
304
+ "<|14.04|>": 51067,
305
+ "<|14.06|>": 51068,
306
+ "<|14.08|>": 51069,
307
+ "<|14.10|>": 51070,
308
+ "<|14.12|>": 51071,
309
+ "<|14.14|>": 51072,
310
+ "<|14.16|>": 51073,
311
+ "<|14.18|>": 51074,
312
+ "<|14.20|>": 51075,
313
+ "<|14.22|>": 51076,
314
+ "<|14.24|>": 51077,
315
+ "<|14.26|>": 51078,
316
+ "<|14.28|>": 51079,
317
+ "<|14.30|>": 51080,
318
+ "<|14.32|>": 51081,
319
+ "<|14.34|>": 51082,
320
+ "<|14.36|>": 51083,
321
+ "<|14.38|>": 51084,
322
+ "<|14.40|>": 51085,
323
+ "<|14.42|>": 51086,
324
+ "<|14.44|>": 51087,
325
+ "<|14.46|>": 51088,
326
+ "<|14.48|>": 51089,
327
+ "<|14.50|>": 51090,
328
+ "<|14.52|>": 51091,
329
+ "<|14.54|>": 51092,
330
+ "<|14.56|>": 51093,
331
+ "<|14.58|>": 51094,
332
+ "<|14.60|>": 51095,
333
+ "<|14.62|>": 51096,
334
+ "<|14.64|>": 51097,
335
+ "<|14.66|>": 51098,
336
+ "<|14.68|>": 51099,
337
+ "<|14.70|>": 51100,
338
+ "<|14.72|>": 51101,
339
+ "<|14.74|>": 51102,
340
+ "<|14.76|>": 51103,
341
+ "<|14.78|>": 51104,
342
+ "<|14.80|>": 51105,
343
+ "<|14.82|>": 51106,
344
+ "<|14.84|>": 51107,
345
+ "<|14.86|>": 51108,
346
+ "<|14.88|>": 51109,
347
+ "<|14.90|>": 51110,
348
+ "<|14.92|>": 51111,
349
+ "<|14.94|>": 51112,
350
+ "<|14.96|>": 51113,
351
+ "<|14.98|>": 51114,
352
+ "<|15.00|>": 51115,
353
+ "<|15.02|>": 51116,
354
+ "<|15.04|>": 51117,
355
+ "<|15.06|>": 51118,
356
+ "<|15.08|>": 51119,
357
+ "<|15.10|>": 51120,
358
+ "<|15.12|>": 51121,
359
+ "<|15.14|>": 51122,
360
+ "<|15.16|>": 51123,
361
+ "<|15.18|>": 51124,
362
+ "<|15.20|>": 51125,
363
+ "<|15.22|>": 51126,
364
+ "<|15.24|>": 51127,
365
+ "<|15.26|>": 51128,
366
+ "<|15.28|>": 51129,
367
+ "<|15.30|>": 51130,
368
+ "<|15.32|>": 51131,
369
+ "<|15.34|>": 51132,
370
+ "<|15.36|>": 51133,
371
+ "<|15.38|>": 51134,
372
+ "<|15.40|>": 51135,
373
+ "<|15.42|>": 51136,
374
+ "<|15.44|>": 51137,
375
+ "<|15.46|>": 51138,
376
+ "<|15.48|>": 51139,
377
+ "<|15.50|>": 51140,
378
+ "<|15.52|>": 51141,
379
+ "<|15.54|>": 51142,
380
+ "<|15.56|>": 51143,
381
+ "<|15.58|>": 51144,
382
+ "<|15.60|>": 51145,
383
+ "<|15.62|>": 51146,
384
+ "<|15.64|>": 51147,
385
+ "<|15.66|>": 51148,
386
+ "<|15.68|>": 51149,
387
+ "<|15.70|>": 51150,
388
+ "<|15.72|>": 51151,
389
+ "<|15.74|>": 51152,
390
+ "<|15.76|>": 51153,
391
+ "<|15.78|>": 51154,
392
+ "<|15.80|>": 51155,
393
+ "<|15.82|>": 51156,
394
+ "<|15.84|>": 51157,
395
+ "<|15.86|>": 51158,
396
+ "<|15.88|>": 51159,
397
+ "<|15.90|>": 51160,
398
+ "<|15.92|>": 51161,
399
+ "<|15.94|>": 51162,
400
+ "<|15.96|>": 51163,
401
+ "<|15.98|>": 51164,
402
+ "<|16.00|>": 51165,
403
+ "<|16.02|>": 51166,
404
+ "<|16.04|>": 51167,
405
+ "<|16.06|>": 51168,
406
+ "<|16.08|>": 51169,
407
+ "<|16.10|>": 51170,
408
+ "<|16.12|>": 51171,
409
+ "<|16.14|>": 51172,
410
+ "<|16.16|>": 51173,
411
+ "<|16.18|>": 51174,
412
+ "<|16.20|>": 51175,
413
+ "<|16.22|>": 51176,
414
+ "<|16.24|>": 51177,
415
+ "<|16.26|>": 51178,
416
+ "<|16.28|>": 51179,
417
+ "<|16.30|>": 51180,
418
+ "<|16.32|>": 51181,
419
+ "<|16.34|>": 51182,
420
+ "<|16.36|>": 51183,
421
+ "<|16.38|>": 51184,
422
+ "<|16.40|>": 51185,
423
+ "<|16.42|>": 51186,
424
+ "<|16.44|>": 51187,
425
+ "<|16.46|>": 51188,
426
+ "<|16.48|>": 51189,
427
+ "<|16.50|>": 51190,
428
+ "<|16.52|>": 51191,
429
+ "<|16.54|>": 51192,
430
+ "<|16.56|>": 51193,
431
+ "<|16.58|>": 51194,
432
+ "<|16.60|>": 51195,
433
+ "<|16.62|>": 51196,
434
+ "<|16.64|>": 51197,
435
+ "<|16.66|>": 51198,
436
+ "<|16.68|>": 51199,
437
+ "<|16.70|>": 51200,
438
+ "<|16.72|>": 51201,
439
+ "<|16.74|>": 51202,
440
+ "<|16.76|>": 51203,
441
+ "<|16.78|>": 51204,
442
+ "<|16.80|>": 51205,
443
+ "<|16.82|>": 51206,
444
+ "<|16.84|>": 51207,
445
+ "<|16.86|>": 51208,
446
+ "<|16.88|>": 51209,
447
+ "<|16.90|>": 51210,
448
+ "<|16.92|>": 51211,
449
+ "<|16.94|>": 51212,
450
+ "<|16.96|>": 51213,
451
+ "<|16.98|>": 51214,
452
+ "<|17.00|>": 51215,
453
+ "<|17.02|>": 51216,
454
+ "<|17.04|>": 51217,
455
+ "<|17.06|>": 51218,
456
+ "<|17.08|>": 51219,
457
+ "<|17.10|>": 51220,
458
+ "<|17.12|>": 51221,
459
+ "<|17.14|>": 51222,
460
+ "<|17.16|>": 51223,
461
+ "<|17.18|>": 51224,
462
+ "<|17.20|>": 51225,
463
+ "<|17.22|>": 51226,
464
+ "<|17.24|>": 51227,
465
+ "<|17.26|>": 51228,
466
+ "<|17.28|>": 51229,
467
+ "<|17.30|>": 51230,
468
+ "<|17.32|>": 51231,
469
+ "<|17.34|>": 51232,
470
+ "<|17.36|>": 51233,
471
+ "<|17.38|>": 51234,
472
+ "<|17.40|>": 51235,
473
+ "<|17.42|>": 51236,
474
+ "<|17.44|>": 51237,
475
+ "<|17.46|>": 51238,
476
+ "<|17.48|>": 51239,
477
+ "<|17.50|>": 51240,
478
+ "<|17.52|>": 51241,
479
+ "<|17.54|>": 51242,
480
+ "<|17.56|>": 51243,
481
+ "<|17.58|>": 51244,
482
+ "<|17.60|>": 51245,
483
+ "<|17.62|>": 51246,
484
+ "<|17.64|>": 51247,
485
+ "<|17.66|>": 51248,
486
+ "<|17.68|>": 51249,
487
+ "<|17.70|>": 51250,
488
+ "<|17.72|>": 51251,
489
+ "<|17.74|>": 51252,
490
+ "<|17.76|>": 51253,
491
+ "<|17.78|>": 51254,
492
+ "<|17.80|>": 51255,
493
+ "<|17.82|>": 51256,
494
+ "<|17.84|>": 51257,
495
+ "<|17.86|>": 51258,
496
+ "<|17.88|>": 51259,
497
+ "<|17.90|>": 51260,
498
+ "<|17.92|>": 51261,
499
+ "<|17.94|>": 51262,
500
+ "<|17.96|>": 51263,
501
+ "<|17.98|>": 51264,
502
+ "<|18.00|>": 51265,
503
+ "<|18.02|>": 51266,
504
+ "<|18.04|>": 51267,
505
+ "<|18.06|>": 51268,
506
+ "<|18.08|>": 51269,
507
+ "<|18.10|>": 51270,
508
+ "<|18.12|>": 51271,
509
+ "<|18.14|>": 51272,
510
+ "<|18.16|>": 51273,
511
+ "<|18.18|>": 51274,
512
+ "<|18.20|>": 51275,
513
+ "<|18.22|>": 51276,
514
+ "<|18.24|>": 51277,
515
+ "<|18.26|>": 51278,
516
+ "<|18.28|>": 51279,
517
+ "<|18.30|>": 51280,
518
+ "<|18.32|>": 51281,
519
+ "<|18.34|>": 51282,
520
+ "<|18.36|>": 51283,
521
+ "<|18.38|>": 51284,
522
+ "<|18.40|>": 51285,
523
+ "<|18.42|>": 51286,
524
+ "<|18.44|>": 51287,
525
+ "<|18.46|>": 51288,
526
+ "<|18.48|>": 51289,
527
+ "<|18.50|>": 51290,
528
+ "<|18.52|>": 51291,
529
+ "<|18.54|>": 51292,
530
+ "<|18.56|>": 51293,
531
+ "<|18.58|>": 51294,
532
+ "<|18.60|>": 51295,
533
+ "<|18.62|>": 51296,
534
+ "<|18.64|>": 51297,
535
+ "<|18.66|>": 51298,
536
+ "<|18.68|>": 51299,
537
+ "<|18.70|>": 51300,
538
+ "<|18.72|>": 51301,
539
+ "<|18.74|>": 51302,
540
+ "<|18.76|>": 51303,
541
+ "<|18.78|>": 51304,
542
+ "<|18.80|>": 51305,
543
+ "<|18.82|>": 51306,
544
+ "<|18.84|>": 51307,
545
+ "<|18.86|>": 51308,
546
+ "<|18.88|>": 51309,
547
+ "<|18.90|>": 51310,
548
+ "<|18.92|>": 51311,
549
+ "<|18.94|>": 51312,
550
+ "<|18.96|>": 51313,
551
+ "<|18.98|>": 51314,
552
+ "<|19.00|>": 51315,
553
+ "<|19.02|>": 51316,
554
+ "<|19.04|>": 51317,
555
+ "<|19.06|>": 51318,
556
+ "<|19.08|>": 51319,
557
+ "<|19.10|>": 51320,
558
+ "<|19.12|>": 51321,
559
+ "<|19.14|>": 51322,
560
+ "<|19.16|>": 51323,
561
+ "<|19.18|>": 51324,
562
+ "<|19.20|>": 51325,
563
+ "<|19.22|>": 51326,
564
+ "<|19.24|>": 51327,
565
+ "<|19.26|>": 51328,
566
+ "<|19.28|>": 51329,
567
+ "<|19.30|>": 51330,
568
+ "<|19.32|>": 51331,
569
+ "<|19.34|>": 51332,
570
+ "<|19.36|>": 51333,
571
+ "<|19.38|>": 51334,
572
+ "<|19.40|>": 51335,
573
+ "<|19.42|>": 51336,
574
+ "<|19.44|>": 51337,
575
+ "<|19.46|>": 51338,
576
+ "<|19.48|>": 51339,
577
+ "<|19.50|>": 51340,
578
+ "<|19.52|>": 51341,
579
+ "<|19.54|>": 51342,
580
+ "<|19.56|>": 51343,
581
+ "<|19.58|>": 51344,
582
+ "<|19.60|>": 51345,
583
+ "<|19.62|>": 51346,
584
+ "<|19.64|>": 51347,
585
+ "<|19.66|>": 51348,
586
+ "<|19.68|>": 51349,
587
+ "<|19.70|>": 51350,
588
+ "<|19.72|>": 51351,
589
+ "<|19.74|>": 51352,
590
+ "<|19.76|>": 51353,
591
+ "<|19.78|>": 51354,
592
+ "<|19.80|>": 51355,
593
+ "<|19.82|>": 51356,
594
+ "<|19.84|>": 51357,
595
+ "<|19.86|>": 51358,
596
+ "<|19.88|>": 51359,
597
+ "<|19.90|>": 51360,
598
+ "<|19.92|>": 51361,
599
+ "<|19.94|>": 51362,
600
+ "<|19.96|>": 51363,
601
+ "<|19.98|>": 51364,
602
+ "<|2.00|>": 50465,
603
+ "<|2.02|>": 50466,
604
+ "<|2.04|>": 50467,
605
+ "<|2.06|>": 50468,
606
+ "<|2.08|>": 50469,
607
+ "<|2.10|>": 50470,
608
+ "<|2.12|>": 50471,
609
+ "<|2.14|>": 50472,
610
+ "<|2.16|>": 50473,
611
+ "<|2.18|>": 50474,
612
+ "<|2.20|>": 50475,
613
+ "<|2.22|>": 50476,
614
+ "<|2.24|>": 50477,
615
+ "<|2.26|>": 50478,
616
+ "<|2.28|>": 50479,
617
+ "<|2.30|>": 50480,
618
+ "<|2.32|>": 50481,
619
+ "<|2.34|>": 50482,
620
+ "<|2.36|>": 50483,
621
+ "<|2.38|>": 50484,
622
+ "<|2.40|>": 50485,
623
+ "<|2.42|>": 50486,
624
+ "<|2.44|>": 50487,
625
+ "<|2.46|>": 50488,
626
+ "<|2.48|>": 50489,
627
+ "<|2.50|>": 50490,
628
+ "<|2.52|>": 50491,
629
+ "<|2.54|>": 50492,
630
+ "<|2.56|>": 50493,
631
+ "<|2.58|>": 50494,
632
+ "<|2.60|>": 50495,
633
+ "<|2.62|>": 50496,
634
+ "<|2.64|>": 50497,
635
+ "<|2.66|>": 50498,
636
+ "<|2.68|>": 50499,
637
+ "<|2.70|>": 50500,
638
+ "<|2.72|>": 50501,
639
+ "<|2.74|>": 50502,
640
+ "<|2.76|>": 50503,
641
+ "<|2.78|>": 50504,
642
+ "<|2.80|>": 50505,
643
+ "<|2.82|>": 50506,
644
+ "<|2.84|>": 50507,
645
+ "<|2.86|>": 50508,
646
+ "<|2.88|>": 50509,
647
+ "<|2.90|>": 50510,
648
+ "<|2.92|>": 50511,
649
+ "<|2.94|>": 50512,
650
+ "<|2.96|>": 50513,
651
+ "<|2.98|>": 50514,
652
+ "<|20.00|>": 51365,
653
+ "<|20.02|>": 51366,
654
+ "<|20.04|>": 51367,
655
+ "<|20.06|>": 51368,
656
+ "<|20.08|>": 51369,
657
+ "<|20.10|>": 51370,
658
+ "<|20.12|>": 51371,
659
+ "<|20.14|>": 51372,
660
+ "<|20.16|>": 51373,
661
+ "<|20.18|>": 51374,
662
+ "<|20.20|>": 51375,
663
+ "<|20.22|>": 51376,
664
+ "<|20.24|>": 51377,
665
+ "<|20.26|>": 51378,
666
+ "<|20.28|>": 51379,
667
+ "<|20.30|>": 51380,
668
+ "<|20.32|>": 51381,
669
+ "<|20.34|>": 51382,
670
+ "<|20.36|>": 51383,
671
+ "<|20.38|>": 51384,
672
+ "<|20.40|>": 51385,
673
+ "<|20.42|>": 51386,
674
+ "<|20.44|>": 51387,
675
+ "<|20.46|>": 51388,
676
+ "<|20.48|>": 51389,
677
+ "<|20.50|>": 51390,
678
+ "<|20.52|>": 51391,
679
+ "<|20.54|>": 51392,
680
+ "<|20.56|>": 51393,
681
+ "<|20.58|>": 51394,
682
+ "<|20.60|>": 51395,
683
+ "<|20.62|>": 51396,
684
+ "<|20.64|>": 51397,
685
+ "<|20.66|>": 51398,
686
+ "<|20.68|>": 51399,
687
+ "<|20.70|>": 51400,
688
+ "<|20.72|>": 51401,
689
+ "<|20.74|>": 51402,
690
+ "<|20.76|>": 51403,
691
+ "<|20.78|>": 51404,
692
+ "<|20.80|>": 51405,
693
+ "<|20.82|>": 51406,
694
+ "<|20.84|>": 51407,
695
+ "<|20.86|>": 51408,
696
+ "<|20.88|>": 51409,
697
+ "<|20.90|>": 51410,
698
+ "<|20.92|>": 51411,
699
+ "<|20.94|>": 51412,
700
+ "<|20.96|>": 51413,
701
+ "<|20.98|>": 51414,
702
+ "<|21.00|>": 51415,
703
+ "<|21.02|>": 51416,
704
+ "<|21.04|>": 51417,
705
+ "<|21.06|>": 51418,
706
+ "<|21.08|>": 51419,
707
+ "<|21.10|>": 51420,
708
+ "<|21.12|>": 51421,
709
+ "<|21.14|>": 51422,
710
+ "<|21.16|>": 51423,
711
+ "<|21.18|>": 51424,
712
+ "<|21.20|>": 51425,
713
+ "<|21.22|>": 51426,
714
+ "<|21.24|>": 51427,
715
+ "<|21.26|>": 51428,
716
+ "<|21.28|>": 51429,
717
+ "<|21.30|>": 51430,
718
+ "<|21.32|>": 51431,
719
+ "<|21.34|>": 51432,
720
+ "<|21.36|>": 51433,
721
+ "<|21.38|>": 51434,
722
+ "<|21.40|>": 51435,
723
+ "<|21.42|>": 51436,
724
+ "<|21.44|>": 51437,
725
+ "<|21.46|>": 51438,
726
+ "<|21.48|>": 51439,
727
+ "<|21.50|>": 51440,
728
+ "<|21.52|>": 51441,
729
+ "<|21.54|>": 51442,
730
+ "<|21.56|>": 51443,
731
+ "<|21.58|>": 51444,
732
+ "<|21.60|>": 51445,
733
+ "<|21.62|>": 51446,
734
+ "<|21.64|>": 51447,
735
+ "<|21.66|>": 51448,
736
+ "<|21.68|>": 51449,
737
+ "<|21.70|>": 51450,
738
+ "<|21.72|>": 51451,
739
+ "<|21.74|>": 51452,
740
+ "<|21.76|>": 51453,
741
+ "<|21.78|>": 51454,
742
+ "<|21.80|>": 51455,
743
+ "<|21.82|>": 51456,
744
+ "<|21.84|>": 51457,
745
+ "<|21.86|>": 51458,
746
+ "<|21.88|>": 51459,
747
+ "<|21.90|>": 51460,
748
+ "<|21.92|>": 51461,
749
+ "<|21.94|>": 51462,
750
+ "<|21.96|>": 51463,
751
+ "<|21.98|>": 51464,
752
+ "<|22.00|>": 51465,
753
+ "<|22.02|>": 51466,
754
+ "<|22.04|>": 51467,
755
+ "<|22.06|>": 51468,
756
+ "<|22.08|>": 51469,
757
+ "<|22.10|>": 51470,
758
+ "<|22.12|>": 51471,
759
+ "<|22.14|>": 51472,
760
+ "<|22.16|>": 51473,
761
+ "<|22.18|>": 51474,
762
+ "<|22.20|>": 51475,
763
+ "<|22.22|>": 51476,
764
+ "<|22.24|>": 51477,
765
+ "<|22.26|>": 51478,
766
+ "<|22.28|>": 51479,
767
+ "<|22.30|>": 51480,
768
+ "<|22.32|>": 51481,
769
+ "<|22.34|>": 51482,
770
+ "<|22.36|>": 51483,
771
+ "<|22.38|>": 51484,
772
+ "<|22.40|>": 51485,
773
+ "<|22.42|>": 51486,
774
+ "<|22.44|>": 51487,
775
+ "<|22.46|>": 51488,
776
+ "<|22.48|>": 51489,
777
+ "<|22.50|>": 51490,
778
+ "<|22.52|>": 51491,
779
+ "<|22.54|>": 51492,
780
+ "<|22.56|>": 51493,
781
+ "<|22.58|>": 51494,
782
+ "<|22.60|>": 51495,
783
+ "<|22.62|>": 51496,
784
+ "<|22.64|>": 51497,
785
+ "<|22.66|>": 51498,
786
+ "<|22.68|>": 51499,
787
+ "<|22.70|>": 51500,
788
+ "<|22.72|>": 51501,
789
+ "<|22.74|>": 51502,
790
+ "<|22.76|>": 51503,
791
+ "<|22.78|>": 51504,
792
+ "<|22.80|>": 51505,
793
+ "<|22.82|>": 51506,
794
+ "<|22.84|>": 51507,
795
+ "<|22.86|>": 51508,
796
+ "<|22.88|>": 51509,
797
+ "<|22.90|>": 51510,
798
+ "<|22.92|>": 51511,
799
+ "<|22.94|>": 51512,
800
+ "<|22.96|>": 51513,
801
+ "<|22.98|>": 51514,
802
+ "<|23.00|>": 51515,
803
+ "<|23.02|>": 51516,
804
+ "<|23.04|>": 51517,
805
+ "<|23.06|>": 51518,
806
+ "<|23.08|>": 51519,
807
+ "<|23.10|>": 51520,
808
+ "<|23.12|>": 51521,
809
+ "<|23.14|>": 51522,
810
+ "<|23.16|>": 51523,
811
+ "<|23.18|>": 51524,
812
+ "<|23.20|>": 51525,
813
+ "<|23.22|>": 51526,
814
+ "<|23.24|>": 51527,
815
+ "<|23.26|>": 51528,
816
+ "<|23.28|>": 51529,
817
+ "<|23.30|>": 51530,
818
+ "<|23.32|>": 51531,
819
+ "<|23.34|>": 51532,
820
+ "<|23.36|>": 51533,
821
+ "<|23.38|>": 51534,
822
+ "<|23.40|>": 51535,
823
+ "<|23.42|>": 51536,
824
+ "<|23.44|>": 51537,
825
+ "<|23.46|>": 51538,
826
+ "<|23.48|>": 51539,
827
+ "<|23.50|>": 51540,
828
+ "<|23.52|>": 51541,
829
+ "<|23.54|>": 51542,
830
+ "<|23.56|>": 51543,
831
+ "<|23.58|>": 51544,
832
+ "<|23.60|>": 51545,
833
+ "<|23.62|>": 51546,
834
+ "<|23.64|>": 51547,
835
+ "<|23.66|>": 51548,
836
+ "<|23.68|>": 51549,
837
+ "<|23.70|>": 51550,
838
+ "<|23.72|>": 51551,
839
+ "<|23.74|>": 51552,
840
+ "<|23.76|>": 51553,
841
+ "<|23.78|>": 51554,
842
+ "<|23.80|>": 51555,
843
+ "<|23.82|>": 51556,
844
+ "<|23.84|>": 51557,
845
+ "<|23.86|>": 51558,
846
+ "<|23.88|>": 51559,
847
+ "<|23.90|>": 51560,
848
+ "<|23.92|>": 51561,
849
+ "<|23.94|>": 51562,
850
+ "<|23.96|>": 51563,
851
+ "<|23.98|>": 51564,
852
+ "<|24.00|>": 51565,
853
+ "<|24.02|>": 51566,
854
+ "<|24.04|>": 51567,
855
+ "<|24.06|>": 51568,
856
+ "<|24.08|>": 51569,
857
+ "<|24.10|>": 51570,
858
+ "<|24.12|>": 51571,
859
+ "<|24.14|>": 51572,
860
+ "<|24.16|>": 51573,
861
+ "<|24.18|>": 51574,
862
+ "<|24.20|>": 51575,
863
+ "<|24.22|>": 51576,
864
+ "<|24.24|>": 51577,
865
+ "<|24.26|>": 51578,
866
+ "<|24.28|>": 51579,
867
+ "<|24.30|>": 51580,
868
+ "<|24.32|>": 51581,
869
+ "<|24.34|>": 51582,
870
+ "<|24.36|>": 51583,
871
+ "<|24.38|>": 51584,
872
+ "<|24.40|>": 51585,
873
+ "<|24.42|>": 51586,
874
+ "<|24.44|>": 51587,
875
+ "<|24.46|>": 51588,
876
+ "<|24.48|>": 51589,
877
+ "<|24.50|>": 51590,
878
+ "<|24.52|>": 51591,
879
+ "<|24.54|>": 51592,
880
+ "<|24.56|>": 51593,
881
+ "<|24.58|>": 51594,
882
+ "<|24.60|>": 51595,
883
+ "<|24.62|>": 51596,
884
+ "<|24.64|>": 51597,
885
+ "<|24.66|>": 51598,
886
+ "<|24.68|>": 51599,
887
+ "<|24.70|>": 51600,
888
+ "<|24.72|>": 51601,
889
+ "<|24.74|>": 51602,
890
+ "<|24.76|>": 51603,
891
+ "<|24.78|>": 51604,
892
+ "<|24.80|>": 51605,
893
+ "<|24.82|>": 51606,
894
+ "<|24.84|>": 51607,
895
+ "<|24.86|>": 51608,
896
+ "<|24.88|>": 51609,
897
+ "<|24.90|>": 51610,
898
+ "<|24.92|>": 51611,
899
+ "<|24.94|>": 51612,
900
+ "<|24.96|>": 51613,
901
+ "<|24.98|>": 51614,
902
+ "<|25.00|>": 51615,
903
+ "<|25.02|>": 51616,
904
+ "<|25.04|>": 51617,
905
+ "<|25.06|>": 51618,
906
+ "<|25.08|>": 51619,
907
+ "<|25.10|>": 51620,
908
+ "<|25.12|>": 51621,
909
+ "<|25.14|>": 51622,
910
+ "<|25.16|>": 51623,
911
+ "<|25.18|>": 51624,
912
+ "<|25.20|>": 51625,
913
+ "<|25.22|>": 51626,
914
+ "<|25.24|>": 51627,
915
+ "<|25.26|>": 51628,
916
+ "<|25.28|>": 51629,
917
+ "<|25.30|>": 51630,
918
+ "<|25.32|>": 51631,
919
+ "<|25.34|>": 51632,
920
+ "<|25.36|>": 51633,
921
+ "<|25.38|>": 51634,
922
+ "<|25.40|>": 51635,
923
+ "<|25.42|>": 51636,
924
+ "<|25.44|>": 51637,
925
+ "<|25.46|>": 51638,
926
+ "<|25.48|>": 51639,
927
+ "<|25.50|>": 51640,
928
+ "<|25.52|>": 51641,
929
+ "<|25.54|>": 51642,
930
+ "<|25.56|>": 51643,
931
+ "<|25.58|>": 51644,
932
+ "<|25.60|>": 51645,
933
+ "<|25.62|>": 51646,
934
+ "<|25.64|>": 51647,
935
+ "<|25.66|>": 51648,
936
+ "<|25.68|>": 51649,
937
+ "<|25.70|>": 51650,
938
+ "<|25.72|>": 51651,
939
+ "<|25.74|>": 51652,
940
+ "<|25.76|>": 51653,
941
+ "<|25.78|>": 51654,
942
+ "<|25.80|>": 51655,
943
+ "<|25.82|>": 51656,
944
+ "<|25.84|>": 51657,
945
+ "<|25.86|>": 51658,
946
+ "<|25.88|>": 51659,
947
+ "<|25.90|>": 51660,
948
+ "<|25.92|>": 51661,
949
+ "<|25.94|>": 51662,
950
+ "<|25.96|>": 51663,
951
+ "<|25.98|>": 51664,
952
+ "<|26.00|>": 51665,
953
+ "<|26.02|>": 51666,
954
+ "<|26.04|>": 51667,
955
+ "<|26.06|>": 51668,
956
+ "<|26.08|>": 51669,
957
+ "<|26.10|>": 51670,
958
+ "<|26.12|>": 51671,
959
+ "<|26.14|>": 51672,
960
+ "<|26.16|>": 51673,
961
+ "<|26.18|>": 51674,
962
+ "<|26.20|>": 51675,
963
+ "<|26.22|>": 51676,
964
+ "<|26.24|>": 51677,
965
+ "<|26.26|>": 51678,
966
+ "<|26.28|>": 51679,
967
+ "<|26.30|>": 51680,
968
+ "<|26.32|>": 51681,
969
+ "<|26.34|>": 51682,
970
+ "<|26.36|>": 51683,
971
+ "<|26.38|>": 51684,
972
+ "<|26.40|>": 51685,
973
+ "<|26.42|>": 51686,
974
+ "<|26.44|>": 51687,
975
+ "<|26.46|>": 51688,
976
+ "<|26.48|>": 51689,
977
+ "<|26.50|>": 51690,
978
+ "<|26.52|>": 51691,
979
+ "<|26.54|>": 51692,
980
+ "<|26.56|>": 51693,
981
+ "<|26.58|>": 51694,
982
+ "<|26.60|>": 51695,
983
+ "<|26.62|>": 51696,
984
+ "<|26.64|>": 51697,
985
+ "<|26.66|>": 51698,
986
+ "<|26.68|>": 51699,
987
+ "<|26.70|>": 51700,
988
+ "<|26.72|>": 51701,
989
+ "<|26.74|>": 51702,
990
+ "<|26.76|>": 51703,
991
+ "<|26.78|>": 51704,
992
+ "<|26.80|>": 51705,
993
+ "<|26.82|>": 51706,
994
+ "<|26.84|>": 51707,
995
+ "<|26.86|>": 51708,
996
+ "<|26.88|>": 51709,
997
+ "<|26.90|>": 51710,
998
+ "<|26.92|>": 51711,
999
+ "<|26.94|>": 51712,
1000
+ "<|26.96|>": 51713,
1001
+ "<|26.98|>": 51714,
1002
+ "<|27.00|>": 51715,
1003
+ "<|27.02|>": 51716,
1004
+ "<|27.04|>": 51717,
1005
+ "<|27.06|>": 51718,
1006
+ "<|27.08|>": 51719,
1007
+ "<|27.10|>": 51720,
1008
+ "<|27.12|>": 51721,
1009
+ "<|27.14|>": 51722,
1010
+ "<|27.16|>": 51723,
1011
+ "<|27.18|>": 51724,
1012
+ "<|27.20|>": 51725,
1013
+ "<|27.22|>": 51726,
1014
+ "<|27.24|>": 51727,
1015
+ "<|27.26|>": 51728,
1016
+ "<|27.28|>": 51729,
1017
+ "<|27.30|>": 51730,
1018
+ "<|27.32|>": 51731,
1019
+ "<|27.34|>": 51732,
1020
+ "<|27.36|>": 51733,
1021
+ "<|27.38|>": 51734,
1022
+ "<|27.40|>": 51735,
1023
+ "<|27.42|>": 51736,
1024
+ "<|27.44|>": 51737,
1025
+ "<|27.46|>": 51738,
1026
+ "<|27.48|>": 51739,
1027
+ "<|27.50|>": 51740,
1028
+ "<|27.52|>": 51741,
1029
+ "<|27.54|>": 51742,
1030
+ "<|27.56|>": 51743,
1031
+ "<|27.58|>": 51744,
1032
+ "<|27.60|>": 51745,
1033
+ "<|27.62|>": 51746,
1034
+ "<|27.64|>": 51747,
1035
+ "<|27.66|>": 51748,
1036
+ "<|27.68|>": 51749,
1037
+ "<|27.70|>": 51750,
1038
+ "<|27.72|>": 51751,
1039
+ "<|27.74|>": 51752,
1040
+ "<|27.76|>": 51753,
1041
+ "<|27.78|>": 51754,
1042
+ "<|27.80|>": 51755,
1043
+ "<|27.82|>": 51756,
1044
+ "<|27.84|>": 51757,
1045
+ "<|27.86|>": 51758,
1046
+ "<|27.88|>": 51759,
1047
+ "<|27.90|>": 51760,
1048
+ "<|27.92|>": 51761,
1049
+ "<|27.94|>": 51762,
1050
+ "<|27.96|>": 51763,
1051
+ "<|27.98|>": 51764,
1052
+ "<|28.00|>": 51765,
1053
+ "<|28.02|>": 51766,
1054
+ "<|28.04|>": 51767,
1055
+ "<|28.06|>": 51768,
1056
+ "<|28.08|>": 51769,
1057
+ "<|28.10|>": 51770,
1058
+ "<|28.12|>": 51771,
1059
+ "<|28.14|>": 51772,
1060
+ "<|28.16|>": 51773,
1061
+ "<|28.18|>": 51774,
1062
+ "<|28.20|>": 51775,
1063
+ "<|28.22|>": 51776,
1064
+ "<|28.24|>": 51777,
1065
+ "<|28.26|>": 51778,
1066
+ "<|28.28|>": 51779,
1067
+ "<|28.30|>": 51780,
1068
+ "<|28.32|>": 51781,
1069
+ "<|28.34|>": 51782,
1070
+ "<|28.36|>": 51783,
1071
+ "<|28.38|>": 51784,
1072
+ "<|28.40|>": 51785,
1073
+ "<|28.42|>": 51786,
1074
+ "<|28.44|>": 51787,
1075
+ "<|28.46|>": 51788,
1076
+ "<|28.48|>": 51789,
1077
+ "<|28.50|>": 51790,
1078
+ "<|28.52|>": 51791,
1079
+ "<|28.54|>": 51792,
1080
+ "<|28.56|>": 51793,
1081
+ "<|28.58|>": 51794,
1082
+ "<|28.60|>": 51795,
1083
+ "<|28.62|>": 51796,
1084
+ "<|28.64|>": 51797,
1085
+ "<|28.66|>": 51798,
1086
+ "<|28.68|>": 51799,
1087
+ "<|28.70|>": 51800,
1088
+ "<|28.72|>": 51801,
1089
+ "<|28.74|>": 51802,
1090
+ "<|28.76|>": 51803,
1091
+ "<|28.78|>": 51804,
1092
+ "<|28.80|>": 51805,
1093
+ "<|28.82|>": 51806,
1094
+ "<|28.84|>": 51807,
1095
+ "<|28.86|>": 51808,
1096
+ "<|28.88|>": 51809,
1097
+ "<|28.90|>": 51810,
1098
+ "<|28.92|>": 51811,
1099
+ "<|28.94|>": 51812,
1100
+ "<|28.96|>": 51813,
1101
+ "<|28.98|>": 51814,
1102
+ "<|29.00|>": 51815,
1103
+ "<|29.02|>": 51816,
1104
+ "<|29.04|>": 51817,
1105
+ "<|29.06|>": 51818,
1106
+ "<|29.08|>": 51819,
1107
+ "<|29.10|>": 51820,
1108
+ "<|29.12|>": 51821,
1109
+ "<|29.14|>": 51822,
1110
+ "<|29.16|>": 51823,
1111
+ "<|29.18|>": 51824,
1112
+ "<|29.20|>": 51825,
1113
+ "<|29.22|>": 51826,
1114
+ "<|29.24|>": 51827,
1115
+ "<|29.26|>": 51828,
1116
+ "<|29.28|>": 51829,
1117
+ "<|29.30|>": 51830,
1118
+ "<|29.32|>": 51831,
1119
+ "<|29.34|>": 51832,
1120
+ "<|29.36|>": 51833,
1121
+ "<|29.38|>": 51834,
1122
+ "<|29.40|>": 51835,
1123
+ "<|29.42|>": 51836,
1124
+ "<|29.44|>": 51837,
1125
+ "<|29.46|>": 51838,
1126
+ "<|29.48|>": 51839,
1127
+ "<|29.50|>": 51840,
1128
+ "<|29.52|>": 51841,
1129
+ "<|29.54|>": 51842,
1130
+ "<|29.56|>": 51843,
1131
+ "<|29.58|>": 51844,
1132
+ "<|29.60|>": 51845,
1133
+ "<|29.62|>": 51846,
1134
+ "<|29.64|>": 51847,
1135
+ "<|29.66|>": 51848,
1136
+ "<|29.68|>": 51849,
1137
+ "<|29.70|>": 51850,
1138
+ "<|29.72|>": 51851,
1139
+ "<|29.74|>": 51852,
1140
+ "<|29.76|>": 51853,
1141
+ "<|29.78|>": 51854,
1142
+ "<|29.80|>": 51855,
1143
+ "<|29.82|>": 51856,
1144
+ "<|29.84|>": 51857,
1145
+ "<|29.86|>": 51858,
1146
+ "<|29.88|>": 51859,
1147
+ "<|29.90|>": 51860,
1148
+ "<|29.92|>": 51861,
1149
+ "<|29.94|>": 51862,
1150
+ "<|29.96|>": 51863,
1151
+ "<|29.98|>": 51864,
1152
+ "<|3.00|>": 50515,
1153
+ "<|3.02|>": 50516,
1154
+ "<|3.04|>": 50517,
1155
+ "<|3.06|>": 50518,
1156
+ "<|3.08|>": 50519,
1157
+ "<|3.10|>": 50520,
1158
+ "<|3.12|>": 50521,
1159
+ "<|3.14|>": 50522,
1160
+ "<|3.16|>": 50523,
1161
+ "<|3.18|>": 50524,
1162
+ "<|3.20|>": 50525,
1163
+ "<|3.22|>": 50526,
1164
+ "<|3.24|>": 50527,
1165
+ "<|3.26|>": 50528,
1166
+ "<|3.28|>": 50529,
1167
+ "<|3.30|>": 50530,
1168
+ "<|3.32|>": 50531,
1169
+ "<|3.34|>": 50532,
1170
+ "<|3.36|>": 50533,
1171
+ "<|3.38|>": 50534,
1172
+ "<|3.40|>": 50535,
1173
+ "<|3.42|>": 50536,
1174
+ "<|3.44|>": 50537,
1175
+ "<|3.46|>": 50538,
1176
+ "<|3.48|>": 50539,
1177
+ "<|3.50|>": 50540,
1178
+ "<|3.52|>": 50541,
1179
+ "<|3.54|>": 50542,
1180
+ "<|3.56|>": 50543,
1181
+ "<|3.58|>": 50544,
1182
+ "<|3.60|>": 50545,
1183
+ "<|3.62|>": 50546,
1184
+ "<|3.64|>": 50547,
1185
+ "<|3.66|>": 50548,
1186
+ "<|3.68|>": 50549,
1187
+ "<|3.70|>": 50550,
1188
+ "<|3.72|>": 50551,
1189
+ "<|3.74|>": 50552,
1190
+ "<|3.76|>": 50553,
1191
+ "<|3.78|>": 50554,
1192
+ "<|3.80|>": 50555,
1193
+ "<|3.82|>": 50556,
1194
+ "<|3.84|>": 50557,
1195
+ "<|3.86|>": 50558,
1196
+ "<|3.88|>": 50559,
1197
+ "<|3.90|>": 50560,
1198
+ "<|3.92|>": 50561,
1199
+ "<|3.94|>": 50562,
1200
+ "<|3.96|>": 50563,
1201
+ "<|3.98|>": 50564,
1202
+ "<|30.00|>": 51865,
1203
+ "<|4.00|>": 50565,
1204
+ "<|4.02|>": 50566,
1205
+ "<|4.04|>": 50567,
1206
+ "<|4.06|>": 50568,
1207
+ "<|4.08|>": 50569,
1208
+ "<|4.10|>": 50570,
1209
+ "<|4.12|>": 50571,
1210
+ "<|4.14|>": 50572,
1211
+ "<|4.16|>": 50573,
1212
+ "<|4.18|>": 50574,
1213
+ "<|4.20|>": 50575,
1214
+ "<|4.22|>": 50576,
1215
+ "<|4.24|>": 50577,
1216
+ "<|4.26|>": 50578,
1217
+ "<|4.28|>": 50579,
1218
+ "<|4.30|>": 50580,
1219
+ "<|4.32|>": 50581,
1220
+ "<|4.34|>": 50582,
1221
+ "<|4.36|>": 50583,
1222
+ "<|4.38|>": 50584,
1223
+ "<|4.40|>": 50585,
1224
+ "<|4.42|>": 50586,
1225
+ "<|4.44|>": 50587,
1226
+ "<|4.46|>": 50588,
1227
+ "<|4.48|>": 50589,
1228
+ "<|4.50|>": 50590,
1229
+ "<|4.52|>": 50591,
1230
+ "<|4.54|>": 50592,
1231
+ "<|4.56|>": 50593,
1232
+ "<|4.58|>": 50594,
1233
+ "<|4.60|>": 50595,
1234
+ "<|4.62|>": 50596,
1235
+ "<|4.64|>": 50597,
1236
+ "<|4.66|>": 50598,
1237
+ "<|4.68|>": 50599,
1238
+ "<|4.70|>": 50600,
1239
+ "<|4.72|>": 50601,
1240
+ "<|4.74|>": 50602,
1241
+ "<|4.76|>": 50603,
1242
+ "<|4.78|>": 50604,
1243
+ "<|4.80|>": 50605,
1244
+ "<|4.82|>": 50606,
1245
+ "<|4.84|>": 50607,
1246
+ "<|4.86|>": 50608,
1247
+ "<|4.88|>": 50609,
1248
+ "<|4.90|>": 50610,
1249
+ "<|4.92|>": 50611,
1250
+ "<|4.94|>": 50612,
1251
+ "<|4.96|>": 50613,
1252
+ "<|4.98|>": 50614,
1253
+ "<|5.00|>": 50615,
1254
+ "<|5.02|>": 50616,
1255
+ "<|5.04|>": 50617,
1256
+ "<|5.06|>": 50618,
1257
+ "<|5.08|>": 50619,
1258
+ "<|5.10|>": 50620,
1259
+ "<|5.12|>": 50621,
1260
+ "<|5.14|>": 50622,
1261
+ "<|5.16|>": 50623,
1262
+ "<|5.18|>": 50624,
1263
+ "<|5.20|>": 50625,
1264
+ "<|5.22|>": 50626,
1265
+ "<|5.24|>": 50627,
1266
+ "<|5.26|>": 50628,
1267
+ "<|5.28|>": 50629,
1268
+ "<|5.30|>": 50630,
1269
+ "<|5.32|>": 50631,
1270
+ "<|5.34|>": 50632,
1271
+ "<|5.36|>": 50633,
1272
+ "<|5.38|>": 50634,
1273
+ "<|5.40|>": 50635,
1274
+ "<|5.42|>": 50636,
1275
+ "<|5.44|>": 50637,
1276
+ "<|5.46|>": 50638,
1277
+ "<|5.48|>": 50639,
1278
+ "<|5.50|>": 50640,
1279
+ "<|5.52|>": 50641,
1280
+ "<|5.54|>": 50642,
1281
+ "<|5.56|>": 50643,
1282
+ "<|5.58|>": 50644,
1283
+ "<|5.60|>": 50645,
1284
+ "<|5.62|>": 50646,
1285
+ "<|5.64|>": 50647,
1286
+ "<|5.66|>": 50648,
1287
+ "<|5.68|>": 50649,
1288
+ "<|5.70|>": 50650,
1289
+ "<|5.72|>": 50651,
1290
+ "<|5.74|>": 50652,
1291
+ "<|5.76|>": 50653,
1292
+ "<|5.78|>": 50654,
1293
+ "<|5.80|>": 50655,
1294
+ "<|5.82|>": 50656,
1295
+ "<|5.84|>": 50657,
1296
+ "<|5.86|>": 50658,
1297
+ "<|5.88|>": 50659,
1298
+ "<|5.90|>": 50660,
1299
+ "<|5.92|>": 50661,
1300
+ "<|5.94|>": 50662,
1301
+ "<|5.96|>": 50663,
1302
+ "<|5.98|>": 50664,
1303
+ "<|6.00|>": 50665,
1304
+ "<|6.02|>": 50666,
1305
+ "<|6.04|>": 50667,
1306
+ "<|6.06|>": 50668,
1307
+ "<|6.08|>": 50669,
1308
+ "<|6.10|>": 50670,
1309
+ "<|6.12|>": 50671,
1310
+ "<|6.14|>": 50672,
1311
+ "<|6.16|>": 50673,
1312
+ "<|6.18|>": 50674,
1313
+ "<|6.20|>": 50675,
1314
+ "<|6.22|>": 50676,
1315
+ "<|6.24|>": 50677,
1316
+ "<|6.26|>": 50678,
1317
+ "<|6.28|>": 50679,
1318
+ "<|6.30|>": 50680,
1319
+ "<|6.32|>": 50681,
1320
+ "<|6.34|>": 50682,
1321
+ "<|6.36|>": 50683,
1322
+ "<|6.38|>": 50684,
1323
+ "<|6.40|>": 50685,
1324
+ "<|6.42|>": 50686,
1325
+ "<|6.44|>": 50687,
1326
+ "<|6.46|>": 50688,
1327
+ "<|6.48|>": 50689,
1328
+ "<|6.50|>": 50690,
1329
+ "<|6.52|>": 50691,
1330
+ "<|6.54|>": 50692,
1331
+ "<|6.56|>": 50693,
1332
+ "<|6.58|>": 50694,
1333
+ "<|6.60|>": 50695,
1334
+ "<|6.62|>": 50696,
1335
+ "<|6.64|>": 50697,
1336
+ "<|6.66|>": 50698,
1337
+ "<|6.68|>": 50699,
1338
+ "<|6.70|>": 50700,
1339
+ "<|6.72|>": 50701,
1340
+ "<|6.74|>": 50702,
1341
+ "<|6.76|>": 50703,
1342
+ "<|6.78|>": 50704,
1343
+ "<|6.80|>": 50705,
1344
+ "<|6.82|>": 50706,
1345
+ "<|6.84|>": 50707,
1346
+ "<|6.86|>": 50708,
1347
+ "<|6.88|>": 50709,
1348
+ "<|6.90|>": 50710,
1349
+ "<|6.92|>": 50711,
1350
+ "<|6.94|>": 50712,
1351
+ "<|6.96|>": 50713,
1352
+ "<|6.98|>": 50714,
1353
+ "<|7.00|>": 50715,
1354
+ "<|7.02|>": 50716,
1355
+ "<|7.04|>": 50717,
1356
+ "<|7.06|>": 50718,
1357
+ "<|7.08|>": 50719,
1358
+ "<|7.10|>": 50720,
1359
+ "<|7.12|>": 50721,
1360
+ "<|7.14|>": 50722,
1361
+ "<|7.16|>": 50723,
1362
+ "<|7.18|>": 50724,
1363
+ "<|7.20|>": 50725,
1364
+ "<|7.22|>": 50726,
1365
+ "<|7.24|>": 50727,
1366
+ "<|7.26|>": 50728,
1367
+ "<|7.28|>": 50729,
1368
+ "<|7.30|>": 50730,
1369
+ "<|7.32|>": 50731,
1370
+ "<|7.34|>": 50732,
1371
+ "<|7.36|>": 50733,
1372
+ "<|7.38|>": 50734,
1373
+ "<|7.40|>": 50735,
1374
+ "<|7.42|>": 50736,
1375
+ "<|7.44|>": 50737,
1376
+ "<|7.46|>": 50738,
1377
+ "<|7.48|>": 50739,
1378
+ "<|7.50|>": 50740,
1379
+ "<|7.52|>": 50741,
1380
+ "<|7.54|>": 50742,
1381
+ "<|7.56|>": 50743,
1382
+ "<|7.58|>": 50744,
1383
+ "<|7.60|>": 50745,
1384
+ "<|7.62|>": 50746,
1385
+ "<|7.64|>": 50747,
1386
+ "<|7.66|>": 50748,
1387
+ "<|7.68|>": 50749,
1388
+ "<|7.70|>": 50750,
1389
+ "<|7.72|>": 50751,
1390
+ "<|7.74|>": 50752,
1391
+ "<|7.76|>": 50753,
1392
+ "<|7.78|>": 50754,
1393
+ "<|7.80|>": 50755,
1394
+ "<|7.82|>": 50756,
1395
+ "<|7.84|>": 50757,
1396
+ "<|7.86|>": 50758,
1397
+ "<|7.88|>": 50759,
1398
+ "<|7.90|>": 50760,
1399
+ "<|7.92|>": 50761,
1400
+ "<|7.94|>": 50762,
1401
+ "<|7.96|>": 50763,
1402
+ "<|7.98|>": 50764,
1403
+ "<|8.00|>": 50765,
1404
+ "<|8.02|>": 50766,
1405
+ "<|8.04|>": 50767,
1406
+ "<|8.06|>": 50768,
1407
+ "<|8.08|>": 50769,
1408
+ "<|8.10|>": 50770,
1409
+ "<|8.12|>": 50771,
1410
+ "<|8.14|>": 50772,
1411
+ "<|8.16|>": 50773,
1412
+ "<|8.18|>": 50774,
1413
+ "<|8.20|>": 50775,
1414
+ "<|8.22|>": 50776,
1415
+ "<|8.24|>": 50777,
1416
+ "<|8.26|>": 50778,
1417
+ "<|8.28|>": 50779,
1418
+ "<|8.30|>": 50780,
1419
+ "<|8.32|>": 50781,
1420
+ "<|8.34|>": 50782,
1421
+ "<|8.36|>": 50783,
1422
+ "<|8.38|>": 50784,
1423
+ "<|8.40|>": 50785,
1424
+ "<|8.42|>": 50786,
1425
+ "<|8.44|>": 50787,
1426
+ "<|8.46|>": 50788,
1427
+ "<|8.48|>": 50789,
1428
+ "<|8.50|>": 50790,
1429
+ "<|8.52|>": 50791,
1430
+ "<|8.54|>": 50792,
1431
+ "<|8.56|>": 50793,
1432
+ "<|8.58|>": 50794,
1433
+ "<|8.60|>": 50795,
1434
+ "<|8.62|>": 50796,
1435
+ "<|8.64|>": 50797,
1436
+ "<|8.66|>": 50798,
1437
+ "<|8.68|>": 50799,
1438
+ "<|8.70|>": 50800,
1439
+ "<|8.72|>": 50801,
1440
+ "<|8.74|>": 50802,
1441
+ "<|8.76|>": 50803,
1442
+ "<|8.78|>": 50804,
1443
+ "<|8.80|>": 50805,
1444
+ "<|8.82|>": 50806,
1445
+ "<|8.84|>": 50807,
1446
+ "<|8.86|>": 50808,
1447
+ "<|8.88|>": 50809,
1448
+ "<|8.90|>": 50810,
1449
+ "<|8.92|>": 50811,
1450
+ "<|8.94|>": 50812,
1451
+ "<|8.96|>": 50813,
1452
+ "<|8.98|>": 50814,
1453
+ "<|9.00|>": 50815,
1454
+ "<|9.02|>": 50816,
1455
+ "<|9.04|>": 50817,
1456
+ "<|9.06|>": 50818,
1457
+ "<|9.08|>": 50819,
1458
+ "<|9.10|>": 50820,
1459
+ "<|9.12|>": 50821,
1460
+ "<|9.14|>": 50822,
1461
+ "<|9.16|>": 50823,
1462
+ "<|9.18|>": 50824,
1463
+ "<|9.20|>": 50825,
1464
+ "<|9.22|>": 50826,
1465
+ "<|9.24|>": 50827,
1466
+ "<|9.26|>": 50828,
1467
+ "<|9.28|>": 50829,
1468
+ "<|9.30|>": 50830,
1469
+ "<|9.32|>": 50831,
1470
+ "<|9.34|>": 50832,
1471
+ "<|9.36|>": 50833,
1472
+ "<|9.38|>": 50834,
1473
+ "<|9.40|>": 50835,
1474
+ "<|9.42|>": 50836,
1475
+ "<|9.44|>": 50837,
1476
+ "<|9.46|>": 50838,
1477
+ "<|9.48|>": 50839,
1478
+ "<|9.50|>": 50840,
1479
+ "<|9.52|>": 50841,
1480
+ "<|9.54|>": 50842,
1481
+ "<|9.56|>": 50843,
1482
+ "<|9.58|>": 50844,
1483
+ "<|9.60|>": 50845,
1484
+ "<|9.62|>": 50846,
1485
+ "<|9.64|>": 50847,
1486
+ "<|9.66|>": 50848,
1487
+ "<|9.68|>": 50849,
1488
+ "<|9.70|>": 50850,
1489
+ "<|9.72|>": 50851,
1490
+ "<|9.74|>": 50852,
1491
+ "<|9.76|>": 50853,
1492
+ "<|9.78|>": 50854,
1493
+ "<|9.80|>": 50855,
1494
+ "<|9.82|>": 50856,
1495
+ "<|9.84|>": 50857,
1496
+ "<|9.86|>": 50858,
1497
+ "<|9.88|>": 50859,
1498
+ "<|9.90|>": 50860,
1499
+ "<|9.92|>": 50861,
1500
+ "<|9.94|>": 50862,
1501
+ "<|9.96|>": 50863,
1502
+ "<|9.98|>": 50864,
1503
  "<|af|>": 50327,
1504
  "<|am|>": 50334,
1505
  "<|ar|>": 50272,
 
1518
  "<|da|>": 50285,
1519
  "<|de|>": 50261,
1520
  "<|el|>": 50281,
1521
+ "<|endoftext|>": 50257,
1522
  "<|en|>": 50259,
1523
  "<|es|>": 50262,
1524
  "<|et|>": 50307,
 
1565
  "<|ne|>": 50313,
1566
  "<|nl|>": 50271,
1567
  "<|nn|>": 50342,
1568
+ "<|nospeech|>": 50363,
1569
+ "<|notimestamps|>": 50364,
1570
  "<|no|>": 50288,
1571
  "<|oc|>": 50328,
1572
  "<|pa|>": 50321,
 
1584
  "<|so|>": 50326,
1585
  "<|sq|>": 50317,
1586
  "<|sr|>": 50303,
1587
+ "<|startoflm|>": 50361,
1588
+ "<|startofprev|>": 50362,
1589
  "<|startoftranscript|>": 50258,
1590
  "<|su|>": 50357,
1591
  "<|sv|>": 50273,
 
1596
  "<|th|>": 50289,
1597
  "<|tk|>": 50341,
1598
  "<|tl|>": 50348,
1599
+ "<|transcribe|>": 50360,
1600
+ "<|translate|>": 50359,
1601
  "<|tr|>": 50268,
1602
  "<|tt|>": 50351,
1603
  "<|uk|>": 50280,
 
1606
  "<|vi|>": 50278,
1607
  "<|yi|>": 50335,
1608
  "<|yo|>": 50325,
1609
+ "<|yue|>": 50358,
1610
  "<|zh|>": 50260
1611
  }
last-checkpoint/config.json CHANGED
@@ -14,7 +14,7 @@
14
  "decoder_attention_heads": 8,
15
  "decoder_ffn_dim": 2048,
16
  "decoder_layerdrop": 0.0,
17
- "decoder_layers": 2,
18
  "decoder_start_token_id": 50258,
19
  "dropout": 0.0,
20
  "encoder_attention_heads": 8,
@@ -29,11 +29,11 @@
29
  ],
30
  [
31
  2,
32
- 50359
33
  ],
34
  [
35
  3,
36
- 50363
37
  ]
38
  ],
39
  "init_std": 0.02,
@@ -53,7 +53,7 @@
53
  "num_mel_bins": 80,
54
  "pad_token_id": 50257,
55
  "scale_embedding": false,
56
- "torch_dtype": "float32",
57
  "transformers_version": "4.48.3",
58
  "use_cache": true,
59
  "use_weighted_layer_sum": false,
 
14
  "decoder_attention_heads": 8,
15
  "decoder_ffn_dim": 2048,
16
  "decoder_layerdrop": 0.0,
17
+ "decoder_layers": 4,
18
  "decoder_start_token_id": 50258,
19
  "dropout": 0.0,
20
  "encoder_attention_heads": 8,
 
29
  ],
30
  [
31
  2,
32
+ 50360
33
  ],
34
  [
35
  3,
36
+ 50364
37
  ]
38
  ],
39
  "init_std": 0.02,
 
53
  "num_mel_bins": 80,
54
  "pad_token_id": 50257,
55
  "scale_embedding": false,
56
+ "torch_dtype": "bfloat16",
57
  "transformers_version": "4.48.3",
58
  "use_cache": true,
59
  "use_weighted_layer_sum": false,
last-checkpoint/global_step1000/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffed1c4b62e71b3ac467dd7d73441e3e9bcaf21927aa3a3a6b59ac5324ff663d
3
+ size 761059696
last-checkpoint/global_step1000/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05d0081bd55bbeb14fd8f3ea82f05c76379648353359ac1e17863ce23dcaedd4
3
+ size 129965712
last-checkpoint/latest ADDED
@@ -0,0 +1 @@
 
 
1
+ global_step1000
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f1172332b4f4cf2ce6af9c26f710924116b9939c47b97977e6e6ac30b8704561
3
- size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8483c37faf3d85258253e75bc0440ac3555e32e5464cd4a298260f708bcc016
3
+ size 181508256
last-checkpoint/preprocessor_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "chunk_length": 30,
3
  "feature_extractor_type": "WhisperFeatureExtractor",
4
- "feature_size": 80,
5
  "hop_length": 160,
6
  "n_fft": 400,
7
  "n_samples": 480000,
 
1
  {
2
  "chunk_length": 30,
3
  "feature_extractor_type": "WhisperFeatureExtractor",
4
+ "feature_size": 128,
5
  "hop_length": 160,
6
  "n_fft": 400,
7
  "n_samples": 480000,
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b83c8e2618548ad0a6cb55d5f9e419ee82c921eaf0cb178e6e4f4700b370bc6
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0da2a816195a08312d7656fb111d92b08e16f2513648f66a8a00f69acac89b1
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa6158b7f4e5c150fe6ade940825c51c282bf26d832e5e702ff0db258a477b79
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbadae7a908ca2ec608dd3ceac8b5aab1986323a21358ba9a060e8f696f7e6fb
3
  size 1064
last-checkpoint/special_tokens_map.json CHANGED
@@ -1,6 +1,5 @@
1
  {
2
  "additional_special_tokens": [
3
- "<|endoftext|>",
4
  "<|startoftranscript|>",
5
  "<|en|>",
6
  "<|zh|>",
@@ -101,11 +100,12 @@
101
  "<|ba|>",
102
  "<|jw|>",
103
  "<|su|>",
 
104
  "<|translate|>",
105
  "<|transcribe|>",
106
  "<|startoflm|>",
107
  "<|startofprev|>",
108
- "<|nocaptions|>",
109
  "<|notimestamps|>"
110
  ],
111
  "bos_token": {
 
1
  {
2
  "additional_special_tokens": [
 
3
  "<|startoftranscript|>",
4
  "<|en|>",
5
  "<|zh|>",
 
100
  "<|ba|>",
101
  "<|jw|>",
102
  "<|su|>",
103
+ "<|yue|>",
104
  "<|translate|>",
105
  "<|transcribe|>",
106
  "<|startoflm|>",
107
  "<|startofprev|>",
108
+ "<|nospeech|>",
109
  "<|notimestamps|>"
110
  ],
111
  "bos_token": {
last-checkpoint/tokenizer_config.json CHANGED
The diff for this file is too large to render. See raw diff
 
last-checkpoint/trainer_state.json CHANGED
@@ -1,2918 +1,308 @@
1
  {
2
- "best_metric": 95.4926921263555,
3
- "best_model_checkpoint": "./iteboshi_temp/checkpoint-4000",
4
- "epoch": 16.50206440957886,
5
  "eval_steps": 1000,
6
- "global_step": 10000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.04128819157720892,
13
- "grad_norm": 48.469512939453125,
14
  "learning_rate": 1.0000000000000002e-06,
15
- "loss": 17.381,
16
  "step": 25
17
  },
18
  {
19
- "epoch": 0.08257638315441784,
20
- "grad_norm": 52.77425003051758,
21
- "learning_rate": 1.9600000000000003e-06,
22
- "loss": 15.9672,
23
  "step": 50
24
  },
25
  {
26
- "epoch": 0.12386457473162675,
27
- "grad_norm": 63.97101974487305,
28
- "learning_rate": 2.96e-06,
29
- "loss": 12.6578,
30
  "step": 75
31
  },
32
  {
33
- "epoch": 0.16515276630883569,
34
- "grad_norm": 26.345178604125977,
35
- "learning_rate": 3.96e-06,
36
- "loss": 8.2932,
37
  "step": 100
38
  },
39
  {
40
- "epoch": 0.20644095788604458,
41
- "grad_norm": 12.366632461547852,
42
- "learning_rate": 4.960000000000001e-06,
43
- "loss": 6.2485,
44
  "step": 125
45
  },
46
  {
47
- "epoch": 0.2477291494632535,
48
- "grad_norm": 9.131484985351562,
49
- "learning_rate": 5.9600000000000005e-06,
50
- "loss": 5.0845,
51
  "step": 150
52
  },
53
  {
54
- "epoch": 0.28901734104046245,
55
- "grad_norm": 8.288640975952148,
56
- "learning_rate": 6.96e-06,
57
- "loss": 4.5216,
58
  "step": 175
59
  },
60
  {
61
- "epoch": 0.33030553261767137,
62
- "grad_norm": 8.107129096984863,
63
- "learning_rate": 7.960000000000002e-06,
64
- "loss": 4.2152,
65
  "step": 200
66
  },
67
  {
68
- "epoch": 0.37159372419488024,
69
- "grad_norm": 8.564615249633789,
70
- "learning_rate": 8.96e-06,
71
- "loss": 4.095,
72
  "step": 225
73
  },
74
  {
75
- "epoch": 0.41288191577208916,
76
- "grad_norm": 8.67183780670166,
77
- "learning_rate": 9.960000000000001e-06,
78
- "loss": 3.9264,
79
  "step": 250
80
  },
81
  {
82
- "epoch": 0.4541701073492981,
83
- "grad_norm": 8.779706954956055,
84
- "learning_rate": 1.0960000000000002e-05,
85
- "loss": 3.7974,
86
  "step": 275
87
  },
88
  {
89
- "epoch": 0.495458298926507,
90
- "grad_norm": 8.62621021270752,
91
- "learning_rate": 1.196e-05,
92
- "loss": 3.7625,
93
  "step": 300
94
  },
95
  {
96
- "epoch": 0.5367464905037159,
97
- "grad_norm": 9.157522201538086,
98
- "learning_rate": 1.2960000000000001e-05,
99
- "loss": 3.6965,
100
  "step": 325
101
  },
102
  {
103
- "epoch": 0.5780346820809249,
104
- "grad_norm": 8.446877479553223,
105
- "learning_rate": 1.396e-05,
106
- "loss": 3.6216,
107
  "step": 350
108
  },
109
  {
110
- "epoch": 0.6193228736581338,
111
- "grad_norm": 8.873468399047852,
112
- "learning_rate": 1.496e-05,
113
- "loss": 3.5455,
114
  "step": 375
115
  },
116
  {
117
- "epoch": 0.6606110652353427,
118
- "grad_norm": 7.768847465515137,
119
- "learning_rate": 1.5960000000000003e-05,
120
- "loss": 3.4952,
121
  "step": 400
122
  },
123
  {
124
- "epoch": 0.7018992568125516,
125
- "grad_norm": 8.365340232849121,
126
- "learning_rate": 1.696e-05,
127
- "loss": 3.414,
128
  "step": 425
129
  },
130
  {
131
- "epoch": 0.7431874483897605,
132
- "grad_norm": 7.865358829498291,
133
- "learning_rate": 1.796e-05,
134
- "loss": 3.3472,
135
  "step": 450
136
  },
137
  {
138
- "epoch": 0.7844756399669695,
139
- "grad_norm": 8.56877613067627,
140
- "learning_rate": 1.896e-05,
141
- "loss": 3.3018,
142
  "step": 475
143
  },
144
  {
145
- "epoch": 0.8257638315441783,
146
- "grad_norm": 9.091480255126953,
147
- "learning_rate": 1.9960000000000002e-05,
148
- "loss": 3.255,
149
  "step": 500
150
  },
151
  {
152
- "epoch": 0.8670520231213873,
153
- "grad_norm": 8.974713325500488,
154
- "learning_rate": 1.9949473684210527e-05,
155
- "loss": 3.1519,
156
  "step": 525
157
  },
158
  {
159
- "epoch": 0.9083402146985962,
160
- "grad_norm": 9.134045600891113,
161
- "learning_rate": 1.989684210526316e-05,
162
- "loss": 3.0334,
163
  "step": 550
164
  },
165
  {
166
- "epoch": 0.9496284062758051,
167
- "grad_norm": 10.075201988220215,
168
- "learning_rate": 1.984421052631579e-05,
169
- "loss": 2.9656,
170
  "step": 575
171
  },
172
  {
173
- "epoch": 0.990916597853014,
174
- "grad_norm": 9.26815128326416,
175
- "learning_rate": 1.9791578947368423e-05,
176
- "loss": 2.8884,
177
  "step": 600
178
  },
179
  {
180
- "epoch": 1.0313790255986788,
181
- "grad_norm": 8.885565757751465,
182
- "learning_rate": 1.9738947368421057e-05,
183
- "loss": 2.7179,
184
  "step": 625
185
  },
186
  {
187
- "epoch": 1.0726672171758878,
188
- "grad_norm": 8.9259672164917,
189
- "learning_rate": 1.9686315789473687e-05,
190
- "loss": 2.6417,
191
  "step": 650
192
  },
193
  {
194
- "epoch": 1.1139554087530965,
195
- "grad_norm": 10.3223295211792,
196
- "learning_rate": 1.9633684210526316e-05,
197
- "loss": 2.5669,
198
  "step": 675
199
  },
200
  {
201
- "epoch": 1.1552436003303055,
202
- "grad_norm": 10.508142471313477,
203
- "learning_rate": 1.958105263157895e-05,
204
- "loss": 2.5026,
205
  "step": 700
206
  },
207
  {
208
- "epoch": 1.1965317919075145,
209
- "grad_norm": 9.78288745880127,
210
- "learning_rate": 1.952842105263158e-05,
211
- "loss": 2.3907,
212
  "step": 725
213
  },
214
  {
215
- "epoch": 1.2378199834847234,
216
- "grad_norm": 9.629091262817383,
217
- "learning_rate": 1.9475789473684213e-05,
218
- "loss": 2.3983,
219
  "step": 750
220
  },
221
  {
222
- "epoch": 1.2791081750619324,
223
- "grad_norm": 9.543179512023926,
224
- "learning_rate": 1.9423157894736843e-05,
225
- "loss": 2.283,
226
  "step": 775
227
  },
228
  {
229
- "epoch": 1.3203963666391412,
230
- "grad_norm": 9.676267623901367,
231
- "learning_rate": 1.9370526315789476e-05,
232
- "loss": 2.2519,
233
  "step": 800
234
  },
235
  {
236
- "epoch": 1.3616845582163501,
237
- "grad_norm": 9.192870140075684,
238
- "learning_rate": 1.9317894736842106e-05,
239
- "loss": 2.1516,
240
  "step": 825
241
  },
242
  {
243
- "epoch": 1.402972749793559,
244
- "grad_norm": 9.859146118164062,
245
- "learning_rate": 1.926526315789474e-05,
246
- "loss": 2.1643,
247
  "step": 850
248
  },
249
  {
250
- "epoch": 1.4442609413707679,
251
- "grad_norm": 9.815247535705566,
252
- "learning_rate": 1.921263157894737e-05,
253
- "loss": 2.0514,
254
  "step": 875
255
  },
256
  {
257
- "epoch": 1.4855491329479769,
258
- "grad_norm": 10.037995338439941,
259
- "learning_rate": 1.916e-05,
260
- "loss": 2.0867,
261
  "step": 900
262
  },
263
  {
264
- "epoch": 1.5268373245251858,
265
- "grad_norm": 9.680682182312012,
266
- "learning_rate": 1.9107368421052633e-05,
267
- "loss": 1.999,
268
  "step": 925
269
  },
270
  {
271
- "epoch": 1.5681255161023948,
272
- "grad_norm": 9.410001754760742,
273
- "learning_rate": 1.9054736842105263e-05,
274
- "loss": 1.954,
275
  "step": 950
276
  },
277
  {
278
- "epoch": 1.6094137076796038,
279
- "grad_norm": 8.89158821105957,
280
- "learning_rate": 1.9002105263157896e-05,
281
- "loss": 1.92,
282
  "step": 975
283
  },
284
  {
285
- "epoch": 1.6507018992568125,
286
- "grad_norm": 9.326930046081543,
287
- "learning_rate": 1.894947368421053e-05,
288
- "loss": 1.8339,
289
  "step": 1000
290
  },
291
  {
292
- "epoch": 1.6507018992568125,
293
- "eval_cer": 93.62053351369322,
294
- "eval_loss": 1.911484956741333,
295
- "eval_runtime": 801.8096,
296
- "eval_samples_per_second": 13.196,
297
- "eval_steps_per_second": 1.1,
298
- "eval_wer": 99.67939651107967,
299
  "step": 1000
300
- },
301
- {
302
- "epoch": 1.6919900908340215,
303
- "grad_norm": 9.614924430847168,
304
- "learning_rate": 1.889684210526316e-05,
305
- "loss": 1.8351,
306
- "step": 1025
307
- },
308
- {
309
- "epoch": 1.7332782824112303,
310
- "grad_norm": 9.912728309631348,
311
- "learning_rate": 1.8844210526315793e-05,
312
- "loss": 1.7924,
313
- "step": 1050
314
- },
315
- {
316
- "epoch": 1.7745664739884393,
317
- "grad_norm": 9.571353912353516,
318
- "learning_rate": 1.8791578947368423e-05,
319
- "loss": 1.7655,
320
- "step": 1075
321
- },
322
- {
323
- "epoch": 1.8158546655656482,
324
- "grad_norm": 9.284981727600098,
325
- "learning_rate": 1.8738947368421056e-05,
326
- "loss": 1.7737,
327
- "step": 1100
328
- },
329
- {
330
- "epoch": 1.8571428571428572,
331
- "grad_norm": 10.073586463928223,
332
- "learning_rate": 1.8686315789473686e-05,
333
- "loss": 1.7343,
334
- "step": 1125
335
- },
336
- {
337
- "epoch": 1.8984310487200662,
338
- "grad_norm": 8.913012504577637,
339
- "learning_rate": 1.8633684210526316e-05,
340
- "loss": 1.6881,
341
- "step": 1150
342
- },
343
- {
344
- "epoch": 1.939719240297275,
345
- "grad_norm": 9.885432243347168,
346
- "learning_rate": 1.858105263157895e-05,
347
- "loss": 1.6909,
348
- "step": 1175
349
- },
350
- {
351
- "epoch": 1.981007431874484,
352
- "grad_norm": 9.40440559387207,
353
- "learning_rate": 1.852842105263158e-05,
354
- "loss": 1.6276,
355
- "step": 1200
356
- },
357
- {
358
- "epoch": 2.0214698596201486,
359
- "grad_norm": 10.476126670837402,
360
- "learning_rate": 1.8475789473684212e-05,
361
- "loss": 1.5264,
362
- "step": 1225
363
- },
364
- {
365
- "epoch": 2.0627580511973576,
366
- "grad_norm": 8.929309844970703,
367
- "learning_rate": 1.8423157894736842e-05,
368
- "loss": 1.4139,
369
- "step": 1250
370
- },
371
- {
372
- "epoch": 2.1040462427745665,
373
- "grad_norm": 9.487678527832031,
374
- "learning_rate": 1.8370526315789476e-05,
375
- "loss": 1.4765,
376
- "step": 1275
377
- },
378
- {
379
- "epoch": 2.1453344343517755,
380
- "grad_norm": 9.428024291992188,
381
- "learning_rate": 1.831789473684211e-05,
382
- "loss": 1.4275,
383
- "step": 1300
384
- },
385
- {
386
- "epoch": 2.1866226259289845,
387
- "grad_norm": 9.300884246826172,
388
- "learning_rate": 1.826526315789474e-05,
389
- "loss": 1.4115,
390
- "step": 1325
391
- },
392
- {
393
- "epoch": 2.227910817506193,
394
- "grad_norm": 9.60734748840332,
395
- "learning_rate": 1.821263157894737e-05,
396
- "loss": 1.4087,
397
- "step": 1350
398
- },
399
- {
400
- "epoch": 2.269199009083402,
401
- "grad_norm": 8.660606384277344,
402
- "learning_rate": 1.8160000000000002e-05,
403
- "loss": 1.3658,
404
- "step": 1375
405
- },
406
- {
407
- "epoch": 2.310487200660611,
408
- "grad_norm": 9.169739723205566,
409
- "learning_rate": 1.8107368421052632e-05,
410
- "loss": 1.3645,
411
- "step": 1400
412
- },
413
- {
414
- "epoch": 2.35177539223782,
415
- "grad_norm": 9.20264720916748,
416
- "learning_rate": 1.8054736842105266e-05,
417
- "loss": 1.328,
418
- "step": 1425
419
- },
420
- {
421
- "epoch": 2.393063583815029,
422
- "grad_norm": 9.625157356262207,
423
- "learning_rate": 1.8002105263157896e-05,
424
- "loss": 1.3095,
425
- "step": 1450
426
- },
427
- {
428
- "epoch": 2.434351775392238,
429
- "grad_norm": 8.743439674377441,
430
- "learning_rate": 1.794947368421053e-05,
431
- "loss": 1.3244,
432
- "step": 1475
433
- },
434
- {
435
- "epoch": 2.475639966969447,
436
- "grad_norm": 9.378725051879883,
437
- "learning_rate": 1.789684210526316e-05,
438
- "loss": 1.2695,
439
- "step": 1500
440
- },
441
- {
442
- "epoch": 2.516928158546656,
443
- "grad_norm": 9.474600791931152,
444
- "learning_rate": 1.7844210526315792e-05,
445
- "loss": 1.3238,
446
- "step": 1525
447
- },
448
- {
449
- "epoch": 2.558216350123865,
450
- "grad_norm": 8.615851402282715,
451
- "learning_rate": 1.7791578947368422e-05,
452
- "loss": 1.2829,
453
- "step": 1550
454
- },
455
- {
456
- "epoch": 2.5995045417010734,
457
- "grad_norm": 8.702414512634277,
458
- "learning_rate": 1.7738947368421052e-05,
459
- "loss": 1.2914,
460
- "step": 1575
461
- },
462
- {
463
- "epoch": 2.6407927332782823,
464
- "grad_norm": 9.374676704406738,
465
- "learning_rate": 1.7686315789473685e-05,
466
- "loss": 1.2769,
467
- "step": 1600
468
- },
469
- {
470
- "epoch": 2.6820809248554913,
471
- "grad_norm": 8.985688209533691,
472
- "learning_rate": 1.7633684210526315e-05,
473
- "loss": 1.25,
474
- "step": 1625
475
- },
476
- {
477
- "epoch": 2.7233691164327003,
478
- "grad_norm": 8.864657402038574,
479
- "learning_rate": 1.758105263157895e-05,
480
- "loss": 1.2028,
481
- "step": 1650
482
- },
483
- {
484
- "epoch": 2.7646573080099093,
485
- "grad_norm": 8.514275550842285,
486
- "learning_rate": 1.7528421052631582e-05,
487
- "loss": 1.2477,
488
- "step": 1675
489
- },
490
- {
491
- "epoch": 2.805945499587118,
492
- "grad_norm": 9.147972106933594,
493
- "learning_rate": 1.7475789473684212e-05,
494
- "loss": 1.2103,
495
- "step": 1700
496
- },
497
- {
498
- "epoch": 2.847233691164327,
499
- "grad_norm": 9.296348571777344,
500
- "learning_rate": 1.7423157894736845e-05,
501
- "loss": 1.2437,
502
- "step": 1725
503
- },
504
- {
505
- "epoch": 2.8885218827415358,
506
- "grad_norm": 9.178862571716309,
507
- "learning_rate": 1.7370526315789475e-05,
508
- "loss": 1.1999,
509
- "step": 1750
510
- },
511
- {
512
- "epoch": 2.9298100743187447,
513
- "grad_norm": 8.8689603805542,
514
- "learning_rate": 1.731789473684211e-05,
515
- "loss": 1.167,
516
- "step": 1775
517
- },
518
- {
519
- "epoch": 2.9710982658959537,
520
- "grad_norm": 8.430618286132812,
521
- "learning_rate": 1.726526315789474e-05,
522
- "loss": 1.1616,
523
- "step": 1800
524
- },
525
- {
526
- "epoch": 3.0115606936416186,
527
- "grad_norm": 8.175687789916992,
528
- "learning_rate": 1.721263157894737e-05,
529
- "loss": 1.1536,
530
- "step": 1825
531
- },
532
- {
533
- "epoch": 3.0528488852188276,
534
- "grad_norm": 8.549736022949219,
535
- "learning_rate": 1.7160000000000002e-05,
536
- "loss": 1.0421,
537
- "step": 1850
538
- },
539
- {
540
- "epoch": 3.094137076796036,
541
- "grad_norm": 8.512212753295898,
542
- "learning_rate": 1.710736842105263e-05,
543
- "loss": 1.0288,
544
- "step": 1875
545
- },
546
- {
547
- "epoch": 3.135425268373245,
548
- "grad_norm": 7.529111385345459,
549
- "learning_rate": 1.7054736842105265e-05,
550
- "loss": 0.9965,
551
- "step": 1900
552
- },
553
- {
554
- "epoch": 3.176713459950454,
555
- "grad_norm": 8.706369400024414,
556
- "learning_rate": 1.7002105263157895e-05,
557
- "loss": 1.0262,
558
- "step": 1925
559
- },
560
- {
561
- "epoch": 3.218001651527663,
562
- "grad_norm": 8.718097686767578,
563
- "learning_rate": 1.6949473684210528e-05,
564
- "loss": 1.0232,
565
- "step": 1950
566
- },
567
- {
568
- "epoch": 3.259289843104872,
569
- "grad_norm": 9.124544143676758,
570
- "learning_rate": 1.689684210526316e-05,
571
- "loss": 1.0104,
572
- "step": 1975
573
- },
574
- {
575
- "epoch": 3.300578034682081,
576
- "grad_norm": 7.853435039520264,
577
- "learning_rate": 1.684421052631579e-05,
578
- "loss": 0.9948,
579
- "step": 2000
580
- },
581
- {
582
- "epoch": 3.300578034682081,
583
- "eval_cer": 59.421319913335545,
584
- "eval_loss": 1.2762852907180786,
585
- "eval_runtime": 738.0908,
586
- "eval_samples_per_second": 14.336,
587
- "eval_steps_per_second": 1.195,
588
- "eval_wer": 97.35030645921735,
589
- "step": 2000
590
- },
591
- {
592
- "epoch": 3.34186622625929,
593
- "grad_norm": 7.776848793029785,
594
- "learning_rate": 1.6791578947368425e-05,
595
- "loss": 0.9834,
596
- "step": 2025
597
- },
598
- {
599
- "epoch": 3.383154417836499,
600
- "grad_norm": 8.306714057922363,
601
- "learning_rate": 1.6738947368421055e-05,
602
- "loss": 1.0036,
603
- "step": 2050
604
- },
605
- {
606
- "epoch": 3.424442609413708,
607
- "grad_norm": 9.474985122680664,
608
- "learning_rate": 1.6686315789473685e-05,
609
- "loss": 1.0534,
610
- "step": 2075
611
- },
612
- {
613
- "epoch": 3.4657308009909165,
614
- "grad_norm": 7.486320972442627,
615
- "learning_rate": 1.6633684210526318e-05,
616
- "loss": 0.9673,
617
- "step": 2100
618
- },
619
- {
620
- "epoch": 3.5070189925681254,
621
- "grad_norm": 8.373710632324219,
622
- "learning_rate": 1.6581052631578948e-05,
623
- "loss": 1.0009,
624
- "step": 2125
625
- },
626
- {
627
- "epoch": 3.5483071841453344,
628
- "grad_norm": 8.073158264160156,
629
- "learning_rate": 1.652842105263158e-05,
630
- "loss": 1.0142,
631
- "step": 2150
632
- },
633
- {
634
- "epoch": 3.5895953757225434,
635
- "grad_norm": 8.070113182067871,
636
- "learning_rate": 1.647578947368421e-05,
637
- "loss": 0.9742,
638
- "step": 2175
639
- },
640
- {
641
- "epoch": 3.6308835672997524,
642
- "grad_norm": 7.891678333282471,
643
- "learning_rate": 1.6423157894736845e-05,
644
- "loss": 1.0163,
645
- "step": 2200
646
- },
647
- {
648
- "epoch": 3.6721717588769613,
649
- "grad_norm": 7.933286666870117,
650
- "learning_rate": 1.6370526315789475e-05,
651
- "loss": 0.9442,
652
- "step": 2225
653
- },
654
- {
655
- "epoch": 3.71345995045417,
656
- "grad_norm": 8.791912078857422,
657
- "learning_rate": 1.6317894736842105e-05,
658
- "loss": 0.9295,
659
- "step": 2250
660
- },
661
- {
662
- "epoch": 3.754748142031379,
663
- "grad_norm": 7.6395583152771,
664
- "learning_rate": 1.6265263157894738e-05,
665
- "loss": 0.9328,
666
- "step": 2275
667
- },
668
- {
669
- "epoch": 3.796036333608588,
670
- "grad_norm": 8.580839157104492,
671
- "learning_rate": 1.6212631578947368e-05,
672
- "loss": 0.9779,
673
- "step": 2300
674
- },
675
- {
676
- "epoch": 3.837324525185797,
677
- "grad_norm": 8.464973449707031,
678
- "learning_rate": 1.616e-05,
679
- "loss": 0.923,
680
- "step": 2325
681
- },
682
- {
683
- "epoch": 3.878612716763006,
684
- "grad_norm": 8.684146881103516,
685
- "learning_rate": 1.6107368421052634e-05,
686
- "loss": 0.9842,
687
- "step": 2350
688
- },
689
- {
690
- "epoch": 3.9199009083402148,
691
- "grad_norm": 8.286194801330566,
692
- "learning_rate": 1.6054736842105264e-05,
693
- "loss": 0.9283,
694
- "step": 2375
695
- },
696
- {
697
- "epoch": 3.9611890999174237,
698
- "grad_norm": 8.621734619140625,
699
- "learning_rate": 1.6002105263157898e-05,
700
- "loss": 0.9424,
701
- "step": 2400
702
- },
703
- {
704
- "epoch": 4.001651527663088,
705
- "grad_norm": 7.654041767120361,
706
- "learning_rate": 1.5949473684210528e-05,
707
- "loss": 0.9083,
708
- "step": 2425
709
- },
710
- {
711
- "epoch": 4.042939719240297,
712
- "grad_norm": 8.66199016571045,
713
- "learning_rate": 1.589684210526316e-05,
714
- "loss": 0.8082,
715
- "step": 2450
716
- },
717
- {
718
- "epoch": 4.084227910817506,
719
- "grad_norm": 7.401526927947998,
720
- "learning_rate": 1.584421052631579e-05,
721
- "loss": 0.7986,
722
- "step": 2475
723
- },
724
- {
725
- "epoch": 4.125516102394715,
726
- "grad_norm": 6.765837669372559,
727
- "learning_rate": 1.579157894736842e-05,
728
- "loss": 0.8111,
729
- "step": 2500
730
- },
731
- {
732
- "epoch": 4.166804293971924,
733
- "grad_norm": 7.28730583190918,
734
- "learning_rate": 1.5738947368421054e-05,
735
- "loss": 0.8089,
736
- "step": 2525
737
- },
738
- {
739
- "epoch": 4.208092485549133,
740
- "grad_norm": 7.018956661224365,
741
- "learning_rate": 1.5686315789473684e-05,
742
- "loss": 0.7773,
743
- "step": 2550
744
- },
745
- {
746
- "epoch": 4.249380677126342,
747
- "grad_norm": 8.369226455688477,
748
- "learning_rate": 1.5633684210526318e-05,
749
- "loss": 0.7895,
750
- "step": 2575
751
- },
752
- {
753
- "epoch": 4.290668868703551,
754
- "grad_norm": 8.268858909606934,
755
- "learning_rate": 1.5581052631578947e-05,
756
- "loss": 0.857,
757
- "step": 2600
758
- },
759
- {
760
- "epoch": 4.33195706028076,
761
- "grad_norm": 7.69369649887085,
762
- "learning_rate": 1.552842105263158e-05,
763
- "loss": 0.807,
764
- "step": 2625
765
- },
766
- {
767
- "epoch": 4.373245251857969,
768
- "grad_norm": 8.12571907043457,
769
- "learning_rate": 1.547578947368421e-05,
770
- "loss": 0.8066,
771
- "step": 2650
772
- },
773
- {
774
- "epoch": 4.414533443435178,
775
- "grad_norm": 9.947578430175781,
776
- "learning_rate": 1.5423157894736844e-05,
777
- "loss": 0.8245,
778
- "step": 2675
779
- },
780
- {
781
- "epoch": 4.455821635012386,
782
- "grad_norm": 7.1731791496276855,
783
- "learning_rate": 1.5370526315789477e-05,
784
- "loss": 0.809,
785
- "step": 2700
786
- },
787
- {
788
- "epoch": 4.497109826589595,
789
- "grad_norm": 8.121570587158203,
790
- "learning_rate": 1.5317894736842107e-05,
791
- "loss": 0.8382,
792
- "step": 2725
793
- },
794
- {
795
- "epoch": 4.538398018166804,
796
- "grad_norm": 8.422553062438965,
797
- "learning_rate": 1.5265263157894737e-05,
798
- "loss": 0.8017,
799
- "step": 2750
800
- },
801
- {
802
- "epoch": 4.579686209744013,
803
- "grad_norm": 9.263009071350098,
804
- "learning_rate": 1.521263157894737e-05,
805
- "loss": 0.805,
806
- "step": 2775
807
- },
808
- {
809
- "epoch": 4.620974401321222,
810
- "grad_norm": 8.761811256408691,
811
- "learning_rate": 1.516e-05,
812
- "loss": 0.7896,
813
- "step": 2800
814
- },
815
- {
816
- "epoch": 4.662262592898431,
817
- "grad_norm": 7.018781661987305,
818
- "learning_rate": 1.5107368421052632e-05,
819
- "loss": 0.764,
820
- "step": 2825
821
- },
822
- {
823
- "epoch": 4.70355078447564,
824
- "grad_norm": 8.710026741027832,
825
- "learning_rate": 1.5054736842105264e-05,
826
- "loss": 0.7673,
827
- "step": 2850
828
- },
829
- {
830
- "epoch": 4.744838976052849,
831
- "grad_norm": 7.182971954345703,
832
- "learning_rate": 1.5002105263157895e-05,
833
- "loss": 0.7649,
834
- "step": 2875
835
- },
836
- {
837
- "epoch": 4.786127167630058,
838
- "grad_norm": 8.922381401062012,
839
- "learning_rate": 1.4949473684210527e-05,
840
- "loss": 0.8047,
841
- "step": 2900
842
- },
843
- {
844
- "epoch": 4.827415359207267,
845
- "grad_norm": 7.456427097320557,
846
- "learning_rate": 1.4896842105263159e-05,
847
- "loss": 0.7793,
848
- "step": 2925
849
- },
850
- {
851
- "epoch": 4.868703550784476,
852
- "grad_norm": 8.16557502746582,
853
- "learning_rate": 1.484421052631579e-05,
854
- "loss": 0.7642,
855
- "step": 2950
856
- },
857
- {
858
- "epoch": 4.909991742361685,
859
- "grad_norm": 7.600188255310059,
860
- "learning_rate": 1.4791578947368422e-05,
861
- "loss": 0.7719,
862
- "step": 2975
863
- },
864
- {
865
- "epoch": 4.951279933938894,
866
- "grad_norm": 7.850863456726074,
867
- "learning_rate": 1.4738947368421055e-05,
868
- "loss": 0.7577,
869
- "step": 3000
870
- },
871
- {
872
- "epoch": 4.951279933938894,
873
- "eval_cer": 53.34678390058808,
874
- "eval_loss": 1.1084543466567993,
875
- "eval_runtime": 754.0137,
876
- "eval_samples_per_second": 14.033,
877
- "eval_steps_per_second": 1.17,
878
- "eval_wer": 96.64309288071664,
879
- "step": 3000
880
- },
881
- {
882
- "epoch": 4.992568125516103,
883
- "grad_norm": 7.915883541107178,
884
- "learning_rate": 1.4686315789473687e-05,
885
- "loss": 0.7496,
886
- "step": 3025
887
- },
888
- {
889
- "epoch": 5.033030553261767,
890
- "grad_norm": 7.240871906280518,
891
- "learning_rate": 1.4633684210526317e-05,
892
- "loss": 0.687,
893
- "step": 3050
894
- },
895
- {
896
- "epoch": 5.074318744838976,
897
- "grad_norm": 7.480823516845703,
898
- "learning_rate": 1.4581052631578949e-05,
899
- "loss": 0.6601,
900
- "step": 3075
901
- },
902
- {
903
- "epoch": 5.115606936416185,
904
- "grad_norm": 7.587548732757568,
905
- "learning_rate": 1.452842105263158e-05,
906
- "loss": 0.6831,
907
- "step": 3100
908
- },
909
- {
910
- "epoch": 5.156895127993394,
911
- "grad_norm": 7.099322319030762,
912
- "learning_rate": 1.4475789473684212e-05,
913
- "loss": 0.6658,
914
- "step": 3125
915
- },
916
- {
917
- "epoch": 5.198183319570603,
918
- "grad_norm": 7.174043655395508,
919
- "learning_rate": 1.4423157894736843e-05,
920
- "loss": 0.6704,
921
- "step": 3150
922
- },
923
- {
924
- "epoch": 5.239471511147812,
925
- "grad_norm": 6.899098873138428,
926
- "learning_rate": 1.4370526315789475e-05,
927
- "loss": 0.65,
928
- "step": 3175
929
- },
930
- {
931
- "epoch": 5.280759702725021,
932
- "grad_norm": 7.519511699676514,
933
- "learning_rate": 1.4317894736842107e-05,
934
- "loss": 0.6665,
935
- "step": 3200
936
- },
937
- {
938
- "epoch": 5.32204789430223,
939
- "grad_norm": 7.408304214477539,
940
- "learning_rate": 1.4265263157894738e-05,
941
- "loss": 0.6398,
942
- "step": 3225
943
- },
944
- {
945
- "epoch": 5.363336085879438,
946
- "grad_norm": 7.133185386657715,
947
- "learning_rate": 1.4212631578947368e-05,
948
- "loss": 0.6802,
949
- "step": 3250
950
- },
951
- {
952
- "epoch": 5.404624277456647,
953
- "grad_norm": 7.851741313934326,
954
- "learning_rate": 1.416e-05,
955
- "loss": 0.6327,
956
- "step": 3275
957
- },
958
- {
959
- "epoch": 5.445912469033856,
960
- "grad_norm": 7.773444175720215,
961
- "learning_rate": 1.4107368421052632e-05,
962
- "loss": 0.6998,
963
- "step": 3300
964
- },
965
- {
966
- "epoch": 5.487200660611065,
967
- "grad_norm": 8.28067398071289,
968
- "learning_rate": 1.4054736842105263e-05,
969
- "loss": 0.6726,
970
- "step": 3325
971
- },
972
- {
973
- "epoch": 5.528488852188274,
974
- "grad_norm": 6.886124134063721,
975
- "learning_rate": 1.4002105263157897e-05,
976
- "loss": 0.6418,
977
- "step": 3350
978
- },
979
- {
980
- "epoch": 5.569777043765483,
981
- "grad_norm": 6.617015361785889,
982
- "learning_rate": 1.3949473684210528e-05,
983
- "loss": 0.6613,
984
- "step": 3375
985
- },
986
- {
987
- "epoch": 5.611065235342692,
988
- "grad_norm": 7.447840690612793,
989
- "learning_rate": 1.389684210526316e-05,
990
- "loss": 0.6608,
991
- "step": 3400
992
- },
993
- {
994
- "epoch": 5.652353426919901,
995
- "grad_norm": 7.151592254638672,
996
- "learning_rate": 1.3844210526315791e-05,
997
- "loss": 0.6409,
998
- "step": 3425
999
- },
1000
- {
1001
- "epoch": 5.69364161849711,
1002
- "grad_norm": 7.587296962738037,
1003
- "learning_rate": 1.3791578947368423e-05,
1004
- "loss": 0.6483,
1005
- "step": 3450
1006
- },
1007
- {
1008
- "epoch": 5.734929810074319,
1009
- "grad_norm": 7.848781585693359,
1010
- "learning_rate": 1.3738947368421055e-05,
1011
- "loss": 0.6638,
1012
- "step": 3475
1013
- },
1014
- {
1015
- "epoch": 5.776218001651528,
1016
- "grad_norm": 7.8602986335754395,
1017
- "learning_rate": 1.3686315789473685e-05,
1018
- "loss": 0.6464,
1019
- "step": 3500
1020
- },
1021
- {
1022
- "epoch": 5.817506193228737,
1023
- "grad_norm": 7.792623043060303,
1024
- "learning_rate": 1.3633684210526316e-05,
1025
- "loss": 0.6641,
1026
- "step": 3525
1027
- },
1028
- {
1029
- "epoch": 5.858794384805946,
1030
- "grad_norm": 7.796040058135986,
1031
- "learning_rate": 1.3581052631578948e-05,
1032
- "loss": 0.6529,
1033
- "step": 3550
1034
- },
1035
- {
1036
- "epoch": 5.900082576383154,
1037
- "grad_norm": 7.527121543884277,
1038
- "learning_rate": 1.352842105263158e-05,
1039
- "loss": 0.6458,
1040
- "step": 3575
1041
- },
1042
- {
1043
- "epoch": 5.941370767960363,
1044
- "grad_norm": 6.546653747558594,
1045
- "learning_rate": 1.3475789473684211e-05,
1046
- "loss": 0.6087,
1047
- "step": 3600
1048
- },
1049
- {
1050
- "epoch": 5.982658959537572,
1051
- "grad_norm": 8.012529373168945,
1052
- "learning_rate": 1.3423157894736843e-05,
1053
- "loss": 0.6298,
1054
- "step": 3625
1055
- },
1056
- {
1057
- "epoch": 6.023121387283237,
1058
- "grad_norm": 7.391585350036621,
1059
- "learning_rate": 1.3370526315789475e-05,
1060
- "loss": 0.5543,
1061
- "step": 3650
1062
- },
1063
- {
1064
- "epoch": 6.064409578860446,
1065
- "grad_norm": 6.664039134979248,
1066
- "learning_rate": 1.3317894736842108e-05,
1067
- "loss": 0.5252,
1068
- "step": 3675
1069
- },
1070
- {
1071
- "epoch": 6.105697770437655,
1072
- "grad_norm": 6.451517105102539,
1073
- "learning_rate": 1.326526315789474e-05,
1074
- "loss": 0.5712,
1075
- "step": 3700
1076
- },
1077
- {
1078
- "epoch": 6.146985962014864,
1079
- "grad_norm": 6.391648292541504,
1080
- "learning_rate": 1.321263157894737e-05,
1081
- "loss": 0.5664,
1082
- "step": 3725
1083
- },
1084
- {
1085
- "epoch": 6.188274153592072,
1086
- "grad_norm": 6.524181842803955,
1087
- "learning_rate": 1.3160000000000001e-05,
1088
- "loss": 0.5514,
1089
- "step": 3750
1090
- },
1091
- {
1092
- "epoch": 6.229562345169281,
1093
- "grad_norm": 6.788370609283447,
1094
- "learning_rate": 1.3107368421052633e-05,
1095
- "loss": 0.5693,
1096
- "step": 3775
1097
- },
1098
- {
1099
- "epoch": 6.27085053674649,
1100
- "grad_norm": 6.802549839019775,
1101
- "learning_rate": 1.3054736842105264e-05,
1102
- "loss": 0.5332,
1103
- "step": 3800
1104
- },
1105
- {
1106
- "epoch": 6.312138728323699,
1107
- "grad_norm": 6.864429473876953,
1108
- "learning_rate": 1.3002105263157896e-05,
1109
- "loss": 0.5582,
1110
- "step": 3825
1111
- },
1112
- {
1113
- "epoch": 6.353426919900908,
1114
- "grad_norm": 7.168697357177734,
1115
- "learning_rate": 1.2949473684210528e-05,
1116
- "loss": 0.5471,
1117
- "step": 3850
1118
- },
1119
- {
1120
- "epoch": 6.394715111478117,
1121
- "grad_norm": 6.500988960266113,
1122
- "learning_rate": 1.289684210526316e-05,
1123
- "loss": 0.5472,
1124
- "step": 3875
1125
- },
1126
- {
1127
- "epoch": 6.436003303055326,
1128
- "grad_norm": 6.378172874450684,
1129
- "learning_rate": 1.2844210526315791e-05,
1130
- "loss": 0.5512,
1131
- "step": 3900
1132
- },
1133
- {
1134
- "epoch": 6.477291494632535,
1135
- "grad_norm": 6.865861415863037,
1136
- "learning_rate": 1.279157894736842e-05,
1137
- "loss": 0.5739,
1138
- "step": 3925
1139
- },
1140
- {
1141
- "epoch": 6.518579686209744,
1142
- "grad_norm": 6.929198741912842,
1143
- "learning_rate": 1.2738947368421052e-05,
1144
- "loss": 0.5654,
1145
- "step": 3950
1146
- },
1147
- {
1148
- "epoch": 6.559867877786953,
1149
- "grad_norm": 6.951797008514404,
1150
- "learning_rate": 1.2686315789473684e-05,
1151
- "loss": 0.5422,
1152
- "step": 3975
1153
- },
1154
- {
1155
- "epoch": 6.601156069364162,
1156
- "grad_norm": 7.957674980163574,
1157
- "learning_rate": 1.2633684210526316e-05,
1158
- "loss": 0.5464,
1159
- "step": 4000
1160
- },
1161
- {
1162
- "epoch": 6.601156069364162,
1163
- "eval_cer": 48.25066202010707,
1164
- "eval_loss": 1.0574802160263062,
1165
- "eval_runtime": 742.515,
1166
- "eval_samples_per_second": 14.25,
1167
- "eval_steps_per_second": 1.188,
1168
- "eval_wer": 95.4926921263555,
1169
- "step": 4000
1170
- },
1171
- {
1172
- "epoch": 6.642444260941371,
1173
- "grad_norm": 7.094240665435791,
1174
- "learning_rate": 1.2581052631578949e-05,
1175
- "loss": 0.5365,
1176
- "step": 4025
1177
- },
1178
- {
1179
- "epoch": 6.68373245251858,
1180
- "grad_norm": 8.255016326904297,
1181
- "learning_rate": 1.252842105263158e-05,
1182
- "loss": 0.5768,
1183
- "step": 4050
1184
- },
1185
- {
1186
- "epoch": 6.725020644095789,
1187
- "grad_norm": 7.434171676635742,
1188
- "learning_rate": 1.2475789473684212e-05,
1189
- "loss": 0.5892,
1190
- "step": 4075
1191
- },
1192
- {
1193
- "epoch": 6.766308835672998,
1194
- "grad_norm": 6.989074230194092,
1195
- "learning_rate": 1.2423157894736844e-05,
1196
- "loss": 0.5494,
1197
- "step": 4100
1198
- },
1199
- {
1200
- "epoch": 6.807597027250207,
1201
- "grad_norm": 6.808770656585693,
1202
- "learning_rate": 1.2370526315789476e-05,
1203
- "loss": 0.5281,
1204
- "step": 4125
1205
- },
1206
- {
1207
- "epoch": 6.848885218827416,
1208
- "grad_norm": 7.470989227294922,
1209
- "learning_rate": 1.2317894736842107e-05,
1210
- "loss": 0.5499,
1211
- "step": 4150
1212
- },
1213
- {
1214
- "epoch": 6.890173410404624,
1215
- "grad_norm": 7.557090759277344,
1216
- "learning_rate": 1.2265263157894737e-05,
1217
- "loss": 0.5413,
1218
- "step": 4175
1219
- },
1220
- {
1221
- "epoch": 6.931461601981833,
1222
- "grad_norm": 7.504119396209717,
1223
- "learning_rate": 1.2212631578947369e-05,
1224
- "loss": 0.513,
1225
- "step": 4200
1226
- },
1227
- {
1228
- "epoch": 6.972749793559042,
1229
- "grad_norm": 6.75087308883667,
1230
- "learning_rate": 1.216e-05,
1231
- "loss": 0.5343,
1232
- "step": 4225
1233
- },
1234
- {
1235
- "epoch": 7.013212221304707,
1236
- "grad_norm": 5.9934515953063965,
1237
- "learning_rate": 1.2107368421052632e-05,
1238
- "loss": 0.4879,
1239
- "step": 4250
1240
- },
1241
- {
1242
- "epoch": 7.054500412881916,
1243
- "grad_norm": 6.634411334991455,
1244
- "learning_rate": 1.2054736842105264e-05,
1245
- "loss": 0.4454,
1246
- "step": 4275
1247
- },
1248
- {
1249
- "epoch": 7.095788604459124,
1250
- "grad_norm": 7.614193916320801,
1251
- "learning_rate": 1.2002105263157895e-05,
1252
- "loss": 0.4392,
1253
- "step": 4300
1254
- },
1255
- {
1256
- "epoch": 7.137076796036333,
1257
- "grad_norm": 5.324835777282715,
1258
- "learning_rate": 1.1949473684210527e-05,
1259
- "loss": 0.456,
1260
- "step": 4325
1261
- },
1262
- {
1263
- "epoch": 7.178364987613542,
1264
- "grad_norm": 6.1424241065979,
1265
- "learning_rate": 1.189684210526316e-05,
1266
- "loss": 0.46,
1267
- "step": 4350
1268
- },
1269
- {
1270
- "epoch": 7.219653179190751,
1271
- "grad_norm": 7.1868767738342285,
1272
- "learning_rate": 1.1844210526315792e-05,
1273
- "loss": 0.4591,
1274
- "step": 4375
1275
- },
1276
- {
1277
- "epoch": 7.26094137076796,
1278
- "grad_norm": 6.41041374206543,
1279
- "learning_rate": 1.1791578947368424e-05,
1280
- "loss": 0.4618,
1281
- "step": 4400
1282
- },
1283
- {
1284
- "epoch": 7.302229562345169,
1285
- "grad_norm": 6.095521450042725,
1286
- "learning_rate": 1.1738947368421054e-05,
1287
- "loss": 0.4692,
1288
- "step": 4425
1289
- },
1290
- {
1291
- "epoch": 7.343517753922378,
1292
- "grad_norm": 6.523583889007568,
1293
- "learning_rate": 1.1686315789473685e-05,
1294
- "loss": 0.4631,
1295
- "step": 4450
1296
- },
1297
- {
1298
- "epoch": 7.384805945499587,
1299
- "grad_norm": 6.192616939544678,
1300
- "learning_rate": 1.1633684210526317e-05,
1301
- "loss": 0.4689,
1302
- "step": 4475
1303
- },
1304
- {
1305
- "epoch": 7.426094137076796,
1306
- "grad_norm": 5.833901882171631,
1307
- "learning_rate": 1.1581052631578948e-05,
1308
- "loss": 0.4472,
1309
- "step": 4500
1310
- },
1311
- {
1312
- "epoch": 7.467382328654005,
1313
- "grad_norm": 7.350675582885742,
1314
- "learning_rate": 1.152842105263158e-05,
1315
- "loss": 0.472,
1316
- "step": 4525
1317
- },
1318
- {
1319
- "epoch": 7.508670520231214,
1320
- "grad_norm": 6.825387477874756,
1321
- "learning_rate": 1.1475789473684212e-05,
1322
- "loss": 0.4773,
1323
- "step": 4550
1324
- },
1325
- {
1326
- "epoch": 7.549958711808423,
1327
- "grad_norm": 7.2338762283325195,
1328
- "learning_rate": 1.1423157894736843e-05,
1329
- "loss": 0.4515,
1330
- "step": 4575
1331
- },
1332
- {
1333
- "epoch": 7.591246903385632,
1334
- "grad_norm": 7.510733604431152,
1335
- "learning_rate": 1.1370526315789473e-05,
1336
- "loss": 0.4676,
1337
- "step": 4600
1338
- },
1339
- {
1340
- "epoch": 7.632535094962841,
1341
- "grad_norm": 6.926485538482666,
1342
- "learning_rate": 1.1317894736842105e-05,
1343
- "loss": 0.4692,
1344
- "step": 4625
1345
- },
1346
- {
1347
- "epoch": 7.673823286540049,
1348
- "grad_norm": 7.200405120849609,
1349
- "learning_rate": 1.1265263157894737e-05,
1350
- "loss": 0.4874,
1351
- "step": 4650
1352
- },
1353
- {
1354
- "epoch": 7.715111478117258,
1355
- "grad_norm": 7.665420055389404,
1356
- "learning_rate": 1.1212631578947368e-05,
1357
- "loss": 0.4515,
1358
- "step": 4675
1359
- },
1360
- {
1361
- "epoch": 7.756399669694467,
1362
- "grad_norm": 6.377760410308838,
1363
- "learning_rate": 1.1160000000000002e-05,
1364
- "loss": 0.4639,
1365
- "step": 4700
1366
- },
1367
- {
1368
- "epoch": 7.797687861271676,
1369
- "grad_norm": 7.134575366973877,
1370
- "learning_rate": 1.1107368421052633e-05,
1371
- "loss": 0.4543,
1372
- "step": 4725
1373
- },
1374
- {
1375
- "epoch": 7.838976052848885,
1376
- "grad_norm": 6.138679027557373,
1377
- "learning_rate": 1.1054736842105265e-05,
1378
- "loss": 0.4276,
1379
- "step": 4750
1380
- },
1381
- {
1382
- "epoch": 7.880264244426094,
1383
- "grad_norm": 6.741207122802734,
1384
- "learning_rate": 1.1002105263157896e-05,
1385
- "loss": 0.4975,
1386
- "step": 4775
1387
- },
1388
- {
1389
- "epoch": 7.921552436003303,
1390
- "grad_norm": 6.514609336853027,
1391
- "learning_rate": 1.0949473684210528e-05,
1392
- "loss": 0.4841,
1393
- "step": 4800
1394
- },
1395
- {
1396
- "epoch": 7.962840627580512,
1397
- "grad_norm": 6.2714009284973145,
1398
- "learning_rate": 1.089684210526316e-05,
1399
- "loss": 0.4658,
1400
- "step": 4825
1401
- },
1402
- {
1403
- "epoch": 8.003303055326176,
1404
- "grad_norm": 7.009129524230957,
1405
- "learning_rate": 1.084421052631579e-05,
1406
- "loss": 0.4919,
1407
- "step": 4850
1408
- },
1409
- {
1410
- "epoch": 8.044591246903385,
1411
- "grad_norm": 6.334277153015137,
1412
- "learning_rate": 1.0791578947368421e-05,
1413
- "loss": 0.3873,
1414
- "step": 4875
1415
- },
1416
- {
1417
- "epoch": 8.085879438480594,
1418
- "grad_norm": 6.403853416442871,
1419
- "learning_rate": 1.0738947368421053e-05,
1420
- "loss": 0.3961,
1421
- "step": 4900
1422
- },
1423
- {
1424
- "epoch": 8.127167630057803,
1425
- "grad_norm": 6.143258571624756,
1426
- "learning_rate": 1.0686315789473685e-05,
1427
- "loss": 0.3699,
1428
- "step": 4925
1429
- },
1430
- {
1431
- "epoch": 8.168455821635012,
1432
- "grad_norm": 6.0696539878845215,
1433
- "learning_rate": 1.0633684210526316e-05,
1434
- "loss": 0.3858,
1435
- "step": 4950
1436
- },
1437
- {
1438
- "epoch": 8.209744013212221,
1439
- "grad_norm": 6.363874435424805,
1440
- "learning_rate": 1.0581052631578948e-05,
1441
- "loss": 0.392,
1442
- "step": 4975
1443
- },
1444
- {
1445
- "epoch": 8.25103220478943,
1446
- "grad_norm": 5.990923881530762,
1447
- "learning_rate": 1.052842105263158e-05,
1448
- "loss": 0.4182,
1449
- "step": 5000
1450
- },
1451
- {
1452
- "epoch": 8.25103220478943,
1453
- "eval_cer": 47.292882280788234,
1454
- "eval_loss": 1.0573757886886597,
1455
- "eval_runtime": 732.0504,
1456
- "eval_samples_per_second": 14.454,
1457
- "eval_steps_per_second": 1.205,
1458
- "eval_wer": 96.23762376237623,
1459
- "step": 5000
1460
- },
1461
- {
1462
- "epoch": 8.29232039636664,
1463
- "grad_norm": 5.979728698730469,
1464
- "learning_rate": 1.047578947368421e-05,
1465
- "loss": 0.38,
1466
- "step": 5025
1467
- },
1468
- {
1469
- "epoch": 8.333608587943848,
1470
- "grad_norm": 6.235442161560059,
1471
- "learning_rate": 1.0423157894736844e-05,
1472
- "loss": 0.3823,
1473
- "step": 5050
1474
- },
1475
- {
1476
- "epoch": 8.374896779521057,
1477
- "grad_norm": 6.0435471534729,
1478
- "learning_rate": 1.0370526315789476e-05,
1479
- "loss": 0.3889,
1480
- "step": 5075
1481
- },
1482
- {
1483
- "epoch": 8.416184971098266,
1484
- "grad_norm": 5.2331624031066895,
1485
- "learning_rate": 1.0317894736842106e-05,
1486
- "loss": 0.4002,
1487
- "step": 5100
1488
- },
1489
- {
1490
- "epoch": 8.457473162675475,
1491
- "grad_norm": 5.645880699157715,
1492
- "learning_rate": 1.0265263157894738e-05,
1493
- "loss": 0.395,
1494
- "step": 5125
1495
- },
1496
- {
1497
- "epoch": 8.498761354252684,
1498
- "grad_norm": 6.611478805541992,
1499
- "learning_rate": 1.021263157894737e-05,
1500
- "loss": 0.3964,
1501
- "step": 5150
1502
- },
1503
- {
1504
- "epoch": 8.540049545829893,
1505
- "grad_norm": 5.404575347900391,
1506
- "learning_rate": 1.0160000000000001e-05,
1507
- "loss": 0.3886,
1508
- "step": 5175
1509
- },
1510
- {
1511
- "epoch": 8.581337737407102,
1512
- "grad_norm": 7.283960819244385,
1513
- "learning_rate": 1.0107368421052633e-05,
1514
- "loss": 0.4193,
1515
- "step": 5200
1516
- },
1517
- {
1518
- "epoch": 8.622625928984311,
1519
- "grad_norm": 6.334190368652344,
1520
- "learning_rate": 1.0054736842105264e-05,
1521
- "loss": 0.4129,
1522
- "step": 5225
1523
- },
1524
- {
1525
- "epoch": 8.66391412056152,
1526
- "grad_norm": 6.392534255981445,
1527
- "learning_rate": 1.0002105263157896e-05,
1528
- "loss": 0.3832,
1529
- "step": 5250
1530
- },
1531
- {
1532
- "epoch": 8.705202312138729,
1533
- "grad_norm": 6.049492359161377,
1534
- "learning_rate": 9.949473684210526e-06,
1535
- "loss": 0.4008,
1536
- "step": 5275
1537
- },
1538
- {
1539
- "epoch": 8.746490503715938,
1540
- "grad_norm": 5.881133556365967,
1541
- "learning_rate": 9.89684210526316e-06,
1542
- "loss": 0.3927,
1543
- "step": 5300
1544
- },
1545
- {
1546
- "epoch": 8.787778695293147,
1547
- "grad_norm": 6.470844745635986,
1548
- "learning_rate": 9.84421052631579e-06,
1549
- "loss": 0.4036,
1550
- "step": 5325
1551
- },
1552
- {
1553
- "epoch": 8.829066886870356,
1554
- "grad_norm": 6.188573837280273,
1555
- "learning_rate": 9.791578947368422e-06,
1556
- "loss": 0.394,
1557
- "step": 5350
1558
- },
1559
- {
1560
- "epoch": 8.870355078447563,
1561
- "grad_norm": 6.691900730133057,
1562
- "learning_rate": 9.738947368421054e-06,
1563
- "loss": 0.399,
1564
- "step": 5375
1565
- },
1566
- {
1567
- "epoch": 8.911643270024772,
1568
- "grad_norm": 6.903393745422363,
1569
- "learning_rate": 9.686315789473684e-06,
1570
- "loss": 0.3999,
1571
- "step": 5400
1572
- },
1573
- {
1574
- "epoch": 8.952931461601981,
1575
- "grad_norm": 6.473091125488281,
1576
- "learning_rate": 9.633684210526316e-06,
1577
- "loss": 0.3833,
1578
- "step": 5425
1579
- },
1580
- {
1581
- "epoch": 8.99421965317919,
1582
- "grad_norm": 6.615601062774658,
1583
- "learning_rate": 9.581052631578947e-06,
1584
- "loss": 0.3961,
1585
- "step": 5450
1586
- },
1587
- {
1588
- "epoch": 9.034682080924856,
1589
- "grad_norm": 5.409454345703125,
1590
- "learning_rate": 9.52842105263158e-06,
1591
- "loss": 0.3184,
1592
- "step": 5475
1593
- },
1594
- {
1595
- "epoch": 9.075970272502065,
1596
- "grad_norm": 4.625083923339844,
1597
- "learning_rate": 9.475789473684212e-06,
1598
- "loss": 0.308,
1599
- "step": 5500
1600
- },
1601
- {
1602
- "epoch": 9.117258464079274,
1603
- "grad_norm": 6.347021579742432,
1604
- "learning_rate": 9.423157894736842e-06,
1605
- "loss": 0.3087,
1606
- "step": 5525
1607
- },
1608
- {
1609
- "epoch": 9.158546655656481,
1610
- "grad_norm": 6.036218166351318,
1611
- "learning_rate": 9.370526315789474e-06,
1612
- "loss": 0.3361,
1613
- "step": 5550
1614
- },
1615
- {
1616
- "epoch": 9.19983484723369,
1617
- "grad_norm": 6.974196910858154,
1618
- "learning_rate": 9.317894736842105e-06,
1619
- "loss": 0.3277,
1620
- "step": 5575
1621
- },
1622
- {
1623
- "epoch": 9.2411230388109,
1624
- "grad_norm": 6.37134313583374,
1625
- "learning_rate": 9.265263157894737e-06,
1626
- "loss": 0.3227,
1627
- "step": 5600
1628
- },
1629
- {
1630
- "epoch": 9.282411230388108,
1631
- "grad_norm": 5.647594928741455,
1632
- "learning_rate": 9.21263157894737e-06,
1633
- "loss": 0.3408,
1634
- "step": 5625
1635
- },
1636
- {
1637
- "epoch": 9.323699421965317,
1638
- "grad_norm": 5.690819263458252,
1639
- "learning_rate": 9.16e-06,
1640
- "loss": 0.3431,
1641
- "step": 5650
1642
- },
1643
- {
1644
- "epoch": 9.364987613542526,
1645
- "grad_norm": 6.394257545471191,
1646
- "learning_rate": 9.107368421052632e-06,
1647
- "loss": 0.3684,
1648
- "step": 5675
1649
- },
1650
- {
1651
- "epoch": 9.406275805119735,
1652
- "grad_norm": 5.489740371704102,
1653
- "learning_rate": 9.054736842105264e-06,
1654
- "loss": 0.3401,
1655
- "step": 5700
1656
- },
1657
- {
1658
- "epoch": 9.447563996696944,
1659
- "grad_norm": 5.952576637268066,
1660
- "learning_rate": 9.002105263157895e-06,
1661
- "loss": 0.3485,
1662
- "step": 5725
1663
- },
1664
- {
1665
- "epoch": 9.488852188274153,
1666
- "grad_norm": 5.653362274169922,
1667
- "learning_rate": 8.949473684210527e-06,
1668
- "loss": 0.349,
1669
- "step": 5750
1670
- },
1671
- {
1672
- "epoch": 9.530140379851362,
1673
- "grad_norm": 6.067948341369629,
1674
- "learning_rate": 8.896842105263159e-06,
1675
- "loss": 0.3629,
1676
- "step": 5775
1677
- },
1678
- {
1679
- "epoch": 9.571428571428571,
1680
- "grad_norm": 5.475063800811768,
1681
- "learning_rate": 8.84421052631579e-06,
1682
- "loss": 0.3207,
1683
- "step": 5800
1684
- },
1685
- {
1686
- "epoch": 9.61271676300578,
1687
- "grad_norm": 6.7783308029174805,
1688
- "learning_rate": 8.791578947368422e-06,
1689
- "loss": 0.3454,
1690
- "step": 5825
1691
- },
1692
- {
1693
- "epoch": 9.65400495458299,
1694
- "grad_norm": 5.636470317840576,
1695
- "learning_rate": 8.738947368421053e-06,
1696
- "loss": 0.3408,
1697
- "step": 5850
1698
- },
1699
- {
1700
- "epoch": 9.695293146160198,
1701
- "grad_norm": 5.290091037750244,
1702
- "learning_rate": 8.686315789473685e-06,
1703
- "loss": 0.3229,
1704
- "step": 5875
1705
- },
1706
- {
1707
- "epoch": 9.736581337737407,
1708
- "grad_norm": 5.795228481292725,
1709
- "learning_rate": 8.633684210526317e-06,
1710
- "loss": 0.36,
1711
- "step": 5900
1712
- },
1713
- {
1714
- "epoch": 9.777869529314616,
1715
- "grad_norm": 5.348466396331787,
1716
- "learning_rate": 8.581052631578948e-06,
1717
- "loss": 0.3358,
1718
- "step": 5925
1719
- },
1720
- {
1721
- "epoch": 9.819157720891825,
1722
- "grad_norm": 5.813995361328125,
1723
- "learning_rate": 8.528421052631578e-06,
1724
- "loss": 0.349,
1725
- "step": 5950
1726
- },
1727
- {
1728
- "epoch": 9.860445912469034,
1729
- "grad_norm": 5.806293964385986,
1730
- "learning_rate": 8.475789473684212e-06,
1731
- "loss": 0.3493,
1732
- "step": 5975
1733
- },
1734
- {
1735
- "epoch": 9.901734104046243,
1736
- "grad_norm": 6.348217487335205,
1737
- "learning_rate": 8.423157894736843e-06,
1738
- "loss": 0.3164,
1739
- "step": 6000
1740
- },
1741
- {
1742
- "epoch": 9.901734104046243,
1743
- "eval_cer": 49.44172503525042,
1744
- "eval_loss": 1.0615819692611694,
1745
- "eval_runtime": 712.2161,
1746
- "eval_samples_per_second": 14.856,
1747
- "eval_steps_per_second": 1.238,
1748
- "eval_wer": 96.38849599245638,
1749
- "step": 6000
1750
- },
1751
- {
1752
- "epoch": 9.943022295623452,
1753
- "grad_norm": 5.729813575744629,
1754
- "learning_rate": 8.370526315789475e-06,
1755
- "loss": 0.3253,
1756
- "step": 6025
1757
- },
1758
- {
1759
- "epoch": 9.984310487200661,
1760
- "grad_norm": 7.191168785095215,
1761
- "learning_rate": 8.317894736842107e-06,
1762
- "loss": 0.3433,
1763
- "step": 6050
1764
- },
1765
- {
1766
- "epoch": 10.024772914946325,
1767
- "grad_norm": 6.767040729522705,
1768
- "learning_rate": 8.265263157894737e-06,
1769
- "loss": 0.2875,
1770
- "step": 6075
1771
- },
1772
- {
1773
- "epoch": 10.066061106523534,
1774
- "grad_norm": 5.048575401306152,
1775
- "learning_rate": 8.212631578947368e-06,
1776
- "loss": 0.2773,
1777
- "step": 6100
1778
- },
1779
- {
1780
- "epoch": 10.107349298100743,
1781
- "grad_norm": 4.836114883422852,
1782
- "learning_rate": 8.16e-06,
1783
- "loss": 0.2889,
1784
- "step": 6125
1785
- },
1786
- {
1787
- "epoch": 10.148637489677952,
1788
- "grad_norm": 4.830049514770508,
1789
- "learning_rate": 8.107368421052633e-06,
1790
- "loss": 0.3004,
1791
- "step": 6150
1792
- },
1793
- {
1794
- "epoch": 10.189925681255161,
1795
- "grad_norm": 5.718193054199219,
1796
- "learning_rate": 8.054736842105265e-06,
1797
- "loss": 0.2989,
1798
- "step": 6175
1799
- },
1800
- {
1801
- "epoch": 10.23121387283237,
1802
- "grad_norm": 7.24569034576416,
1803
- "learning_rate": 8.002105263157895e-06,
1804
- "loss": 0.2712,
1805
- "step": 6200
1806
- },
1807
- {
1808
- "epoch": 10.27250206440958,
1809
- "grad_norm": 6.074246883392334,
1810
- "learning_rate": 7.949473684210526e-06,
1811
- "loss": 0.2899,
1812
- "step": 6225
1813
- },
1814
- {
1815
- "epoch": 10.313790255986788,
1816
- "grad_norm": 5.600971698760986,
1817
- "learning_rate": 7.896842105263158e-06,
1818
- "loss": 0.3,
1819
- "step": 6250
1820
- },
1821
- {
1822
- "epoch": 10.355078447563997,
1823
- "grad_norm": 6.002727508544922,
1824
- "learning_rate": 7.84421052631579e-06,
1825
- "loss": 0.2895,
1826
- "step": 6275
1827
- },
1828
- {
1829
- "epoch": 10.396366639141206,
1830
- "grad_norm": 5.616031169891357,
1831
- "learning_rate": 7.793684210526316e-06,
1832
- "loss": 0.3123,
1833
- "step": 6300
1834
- },
1835
- {
1836
- "epoch": 10.437654830718415,
1837
- "grad_norm": 5.227489948272705,
1838
- "learning_rate": 7.741052631578948e-06,
1839
- "loss": 0.2796,
1840
- "step": 6325
1841
- },
1842
- {
1843
- "epoch": 10.478943022295624,
1844
- "grad_norm": 5.6762261390686035,
1845
- "learning_rate": 7.68842105263158e-06,
1846
- "loss": 0.3061,
1847
- "step": 6350
1848
- },
1849
- {
1850
- "epoch": 10.520231213872833,
1851
- "grad_norm": 4.9320197105407715,
1852
- "learning_rate": 7.635789473684211e-06,
1853
- "loss": 0.2779,
1854
- "step": 6375
1855
- },
1856
- {
1857
- "epoch": 10.561519405450042,
1858
- "grad_norm": 5.518511772155762,
1859
- "learning_rate": 7.583157894736842e-06,
1860
- "loss": 0.2861,
1861
- "step": 6400
1862
- },
1863
- {
1864
- "epoch": 10.602807597027251,
1865
- "grad_norm": 5.470175266265869,
1866
- "learning_rate": 7.5305263157894745e-06,
1867
- "loss": 0.279,
1868
- "step": 6425
1869
- },
1870
- {
1871
- "epoch": 10.64409578860446,
1872
- "grad_norm": 4.248904228210449,
1873
- "learning_rate": 7.477894736842106e-06,
1874
- "loss": 0.3054,
1875
- "step": 6450
1876
- },
1877
- {
1878
- "epoch": 10.685383980181667,
1879
- "grad_norm": 6.3203935623168945,
1880
- "learning_rate": 7.425263157894738e-06,
1881
- "loss": 0.3168,
1882
- "step": 6475
1883
- },
1884
- {
1885
- "epoch": 10.726672171758876,
1886
- "grad_norm": 5.901068687438965,
1887
- "learning_rate": 7.3726315789473694e-06,
1888
- "loss": 0.3103,
1889
- "step": 6500
1890
- },
1891
- {
1892
- "epoch": 10.767960363336085,
1893
- "grad_norm": 6.379432678222656,
1894
- "learning_rate": 7.32e-06,
1895
- "loss": 0.2715,
1896
- "step": 6525
1897
- },
1898
- {
1899
- "epoch": 10.809248554913294,
1900
- "grad_norm": 5.802628040313721,
1901
- "learning_rate": 7.267368421052632e-06,
1902
- "loss": 0.2981,
1903
- "step": 6550
1904
- },
1905
- {
1906
- "epoch": 10.850536746490503,
1907
- "grad_norm": 6.295346260070801,
1908
- "learning_rate": 7.2147368421052635e-06,
1909
- "loss": 0.2676,
1910
- "step": 6575
1911
- },
1912
- {
1913
- "epoch": 10.891824938067712,
1914
- "grad_norm": 5.411930561065674,
1915
- "learning_rate": 7.162105263157896e-06,
1916
- "loss": 0.2809,
1917
- "step": 6600
1918
- },
1919
- {
1920
- "epoch": 10.933113129644921,
1921
- "grad_norm": 5.577861309051514,
1922
- "learning_rate": 7.109473684210528e-06,
1923
- "loss": 0.2473,
1924
- "step": 6625
1925
- },
1926
- {
1927
- "epoch": 10.97440132122213,
1928
- "grad_norm": 5.272998332977295,
1929
- "learning_rate": 7.056842105263158e-06,
1930
- "loss": 0.2931,
1931
- "step": 6650
1932
- },
1933
- {
1934
- "epoch": 11.014863748967795,
1935
- "grad_norm": 5.220921993255615,
1936
- "learning_rate": 7.00421052631579e-06,
1937
- "loss": 0.2772,
1938
- "step": 6675
1939
- },
1940
- {
1941
- "epoch": 11.056151940545003,
1942
- "grad_norm": 4.028062343597412,
1943
- "learning_rate": 6.951578947368422e-06,
1944
- "loss": 0.2511,
1945
- "step": 6700
1946
- },
1947
- {
1948
- "epoch": 11.097440132122212,
1949
- "grad_norm": 4.835925579071045,
1950
- "learning_rate": 6.8989473684210525e-06,
1951
- "loss": 0.2502,
1952
- "step": 6725
1953
- },
1954
- {
1955
- "epoch": 11.138728323699421,
1956
- "grad_norm": 5.920368194580078,
1957
- "learning_rate": 6.846315789473684e-06,
1958
- "loss": 0.2438,
1959
- "step": 6750
1960
- },
1961
- {
1962
- "epoch": 11.18001651527663,
1963
- "grad_norm": 5.114356517791748,
1964
- "learning_rate": 6.793684210526317e-06,
1965
- "loss": 0.2478,
1966
- "step": 6775
1967
- },
1968
- {
1969
- "epoch": 11.22130470685384,
1970
- "grad_norm": 4.871094703674316,
1971
- "learning_rate": 6.741052631578948e-06,
1972
- "loss": 0.2504,
1973
- "step": 6800
1974
- },
1975
- {
1976
- "epoch": 11.262592898431048,
1977
- "grad_norm": 4.20628547668457,
1978
- "learning_rate": 6.68842105263158e-06,
1979
- "loss": 0.2316,
1980
- "step": 6825
1981
- },
1982
- {
1983
- "epoch": 11.303881090008257,
1984
- "grad_norm": 5.95331335067749,
1985
- "learning_rate": 6.635789473684211e-06,
1986
- "loss": 0.2524,
1987
- "step": 6850
1988
- },
1989
- {
1990
- "epoch": 11.345169281585466,
1991
- "grad_norm": 5.462084770202637,
1992
- "learning_rate": 6.583157894736842e-06,
1993
- "loss": 0.2488,
1994
- "step": 6875
1995
- },
1996
- {
1997
- "epoch": 11.386457473162675,
1998
- "grad_norm": 4.980792999267578,
1999
- "learning_rate": 6.530526315789474e-06,
2000
- "loss": 0.2344,
2001
- "step": 6900
2002
- },
2003
- {
2004
- "epoch": 11.427745664739884,
2005
- "grad_norm": 5.869212627410889,
2006
- "learning_rate": 6.477894736842106e-06,
2007
- "loss": 0.2484,
2008
- "step": 6925
2009
- },
2010
- {
2011
- "epoch": 11.469033856317093,
2012
- "grad_norm": 5.4397873878479,
2013
- "learning_rate": 6.425263157894738e-06,
2014
- "loss": 0.2523,
2015
- "step": 6950
2016
- },
2017
- {
2018
- "epoch": 11.510322047894302,
2019
- "grad_norm": 5.890199661254883,
2020
- "learning_rate": 6.372631578947369e-06,
2021
- "loss": 0.2565,
2022
- "step": 6975
2023
- },
2024
- {
2025
- "epoch": 11.551610239471511,
2026
- "grad_norm": 4.83267879486084,
2027
- "learning_rate": 6.3200000000000005e-06,
2028
- "loss": 0.2319,
2029
- "step": 7000
2030
- },
2031
- {
2032
- "epoch": 11.551610239471511,
2033
- "eval_cer": 49.55349466371671,
2034
- "eval_loss": 1.0929207801818848,
2035
- "eval_runtime": 718.5035,
2036
- "eval_samples_per_second": 14.726,
2037
- "eval_steps_per_second": 1.228,
2038
- "eval_wer": 96.25648279113625,
2039
- "step": 7000
2040
- },
2041
- {
2042
- "epoch": 11.59289843104872,
2043
- "grad_norm": 6.01038932800293,
2044
- "learning_rate": 6.267368421052632e-06,
2045
- "loss": 0.2703,
2046
- "step": 7025
2047
- },
2048
- {
2049
- "epoch": 11.63418662262593,
2050
- "grad_norm": 4.515256404876709,
2051
- "learning_rate": 6.214736842105264e-06,
2052
- "loss": 0.2497,
2053
- "step": 7050
2054
- },
2055
- {
2056
- "epoch": 11.675474814203138,
2057
- "grad_norm": 5.914691925048828,
2058
- "learning_rate": 6.1621052631578945e-06,
2059
- "loss": 0.2374,
2060
- "step": 7075
2061
- },
2062
- {
2063
- "epoch": 11.716763005780347,
2064
- "grad_norm": 5.570655822753906,
2065
- "learning_rate": 6.109473684210527e-06,
2066
- "loss": 0.253,
2067
- "step": 7100
2068
- },
2069
- {
2070
- "epoch": 11.758051197357556,
2071
- "grad_norm": 5.9783244132995605,
2072
- "learning_rate": 6.056842105263159e-06,
2073
- "loss": 0.2364,
2074
- "step": 7125
2075
- },
2076
- {
2077
- "epoch": 11.799339388934765,
2078
- "grad_norm": 5.595921516418457,
2079
- "learning_rate": 6.00421052631579e-06,
2080
- "loss": 0.2375,
2081
- "step": 7150
2082
- },
2083
- {
2084
- "epoch": 11.840627580511974,
2085
- "grad_norm": 4.966527938842773,
2086
- "learning_rate": 5.951578947368422e-06,
2087
- "loss": 0.2652,
2088
- "step": 7175
2089
- },
2090
- {
2091
- "epoch": 11.881915772089183,
2092
- "grad_norm": 4.243671894073486,
2093
- "learning_rate": 5.898947368421053e-06,
2094
- "loss": 0.2348,
2095
- "step": 7200
2096
- },
2097
- {
2098
- "epoch": 11.923203963666392,
2099
- "grad_norm": 4.7744669914245605,
2100
- "learning_rate": 5.846315789473684e-06,
2101
- "loss": 0.2461,
2102
- "step": 7225
2103
- },
2104
- {
2105
- "epoch": 11.964492155243601,
2106
- "grad_norm": 4.887286186218262,
2107
- "learning_rate": 5.793684210526316e-06,
2108
- "loss": 0.2466,
2109
- "step": 7250
2110
- },
2111
- {
2112
- "epoch": 12.004954582989265,
2113
- "grad_norm": 4.767704963684082,
2114
- "learning_rate": 5.7410526315789485e-06,
2115
- "loss": 0.2628,
2116
- "step": 7275
2117
- },
2118
- {
2119
- "epoch": 12.046242774566474,
2120
- "grad_norm": 3.731494426727295,
2121
- "learning_rate": 5.68842105263158e-06,
2122
- "loss": 0.2121,
2123
- "step": 7300
2124
- },
2125
- {
2126
- "epoch": 12.087530966143683,
2127
- "grad_norm": 4.329957962036133,
2128
- "learning_rate": 5.635789473684211e-06,
2129
- "loss": 0.208,
2130
- "step": 7325
2131
- },
2132
- {
2133
- "epoch": 12.128819157720892,
2134
- "grad_norm": 4.992996692657471,
2135
- "learning_rate": 5.5831578947368425e-06,
2136
- "loss": 0.2007,
2137
- "step": 7350
2138
- },
2139
- {
2140
- "epoch": 12.170107349298101,
2141
- "grad_norm": 4.756143093109131,
2142
- "learning_rate": 5.530526315789474e-06,
2143
- "loss": 0.2084,
2144
- "step": 7375
2145
- },
2146
- {
2147
- "epoch": 12.21139554087531,
2148
- "grad_norm": 5.0002827644348145,
2149
- "learning_rate": 5.477894736842105e-06,
2150
- "loss": 0.2285,
2151
- "step": 7400
2152
- },
2153
- {
2154
- "epoch": 12.25268373245252,
2155
- "grad_norm": 3.5156126022338867,
2156
- "learning_rate": 5.425263157894737e-06,
2157
- "loss": 0.214,
2158
- "step": 7425
2159
- },
2160
- {
2161
- "epoch": 12.293971924029728,
2162
- "grad_norm": 4.908341884613037,
2163
- "learning_rate": 5.372631578947369e-06,
2164
- "loss": 0.2023,
2165
- "step": 7450
2166
- },
2167
- {
2168
- "epoch": 12.335260115606937,
2169
- "grad_norm": 5.120205879211426,
2170
- "learning_rate": 5.320000000000001e-06,
2171
- "loss": 0.2199,
2172
- "step": 7475
2173
- },
2174
- {
2175
- "epoch": 12.376548307184144,
2176
- "grad_norm": 5.24783992767334,
2177
- "learning_rate": 5.267368421052632e-06,
2178
- "loss": 0.1918,
2179
- "step": 7500
2180
- },
2181
- {
2182
- "epoch": 12.417836498761353,
2183
- "grad_norm": 4.663553714752197,
2184
- "learning_rate": 5.214736842105263e-06,
2185
- "loss": 0.2018,
2186
- "step": 7525
2187
- },
2188
- {
2189
- "epoch": 12.459124690338562,
2190
- "grad_norm": 4.562893867492676,
2191
- "learning_rate": 5.162105263157895e-06,
2192
- "loss": 0.2141,
2193
- "step": 7550
2194
- },
2195
- {
2196
- "epoch": 12.500412881915771,
2197
- "grad_norm": 5.8733062744140625,
2198
- "learning_rate": 5.1094736842105264e-06,
2199
- "loss": 0.2359,
2200
- "step": 7575
2201
- },
2202
- {
2203
- "epoch": 12.54170107349298,
2204
- "grad_norm": 7.215207576751709,
2205
- "learning_rate": 5.056842105263158e-06,
2206
- "loss": 0.2422,
2207
- "step": 7600
2208
- },
2209
- {
2210
- "epoch": 12.58298926507019,
2211
- "grad_norm": 4.633525371551514,
2212
- "learning_rate": 5.0042105263157906e-06,
2213
- "loss": 0.2188,
2214
- "step": 7625
2215
- },
2216
- {
2217
- "epoch": 12.624277456647398,
2218
- "grad_norm": 5.8343305587768555,
2219
- "learning_rate": 4.951578947368421e-06,
2220
- "loss": 0.2169,
2221
- "step": 7650
2222
- },
2223
- {
2224
- "epoch": 12.665565648224607,
2225
- "grad_norm": 3.9599239826202393,
2226
- "learning_rate": 4.898947368421053e-06,
2227
- "loss": 0.2275,
2228
- "step": 7675
2229
- },
2230
- {
2231
- "epoch": 12.706853839801816,
2232
- "grad_norm": 5.128805637359619,
2233
- "learning_rate": 4.846315789473685e-06,
2234
- "loss": 0.2095,
2235
- "step": 7700
2236
- },
2237
- {
2238
- "epoch": 12.748142031379025,
2239
- "grad_norm": 4.6408162117004395,
2240
- "learning_rate": 4.793684210526316e-06,
2241
- "loss": 0.1983,
2242
- "step": 7725
2243
- },
2244
- {
2245
- "epoch": 12.789430222956234,
2246
- "grad_norm": 4.684074878692627,
2247
- "learning_rate": 4.741052631578948e-06,
2248
- "loss": 0.2247,
2249
- "step": 7750
2250
- },
2251
- {
2252
- "epoch": 12.830718414533443,
2253
- "grad_norm": 4.256105422973633,
2254
- "learning_rate": 4.6884210526315795e-06,
2255
- "loss": 0.2164,
2256
- "step": 7775
2257
- },
2258
- {
2259
- "epoch": 12.872006606110652,
2260
- "grad_norm": 4.794992446899414,
2261
- "learning_rate": 4.63578947368421e-06,
2262
- "loss": 0.2133,
2263
- "step": 7800
2264
- },
2265
- {
2266
- "epoch": 12.913294797687861,
2267
- "grad_norm": 6.137476921081543,
2268
- "learning_rate": 4.583157894736843e-06,
2269
- "loss": 0.2434,
2270
- "step": 7825
2271
- },
2272
- {
2273
- "epoch": 12.95458298926507,
2274
- "grad_norm": 4.048770427703857,
2275
- "learning_rate": 4.5305263157894744e-06,
2276
- "loss": 0.2031,
2277
- "step": 7850
2278
- },
2279
- {
2280
- "epoch": 12.995871180842279,
2281
- "grad_norm": 4.465664863586426,
2282
- "learning_rate": 4.477894736842105e-06,
2283
- "loss": 0.2296,
2284
- "step": 7875
2285
- },
2286
- {
2287
- "epoch": 13.036333608587944,
2288
- "grad_norm": 4.5229668617248535,
2289
- "learning_rate": 4.425263157894737e-06,
2290
- "loss": 0.189,
2291
- "step": 7900
2292
- },
2293
- {
2294
- "epoch": 13.077621800165153,
2295
- "grad_norm": 3.7040534019470215,
2296
- "learning_rate": 4.3726315789473685e-06,
2297
- "loss": 0.1951,
2298
- "step": 7925
2299
- },
2300
- {
2301
- "epoch": 13.118909991742361,
2302
- "grad_norm": 4.726925373077393,
2303
- "learning_rate": 4.32e-06,
2304
- "loss": 0.1743,
2305
- "step": 7950
2306
- },
2307
- {
2308
- "epoch": 13.16019818331957,
2309
- "grad_norm": 4.432371616363525,
2310
- "learning_rate": 4.267368421052632e-06,
2311
- "loss": 0.1948,
2312
- "step": 7975
2313
- },
2314
- {
2315
- "epoch": 13.20148637489678,
2316
- "grad_norm": 6.360472679138184,
2317
- "learning_rate": 4.214736842105263e-06,
2318
- "loss": 0.1899,
2319
- "step": 8000
2320
- },
2321
- {
2322
- "epoch": 13.20148637489678,
2323
- "eval_cer": 48.673667075533345,
2324
- "eval_loss": 1.1223334074020386,
2325
- "eval_runtime": 698.1183,
2326
- "eval_samples_per_second": 15.156,
2327
- "eval_steps_per_second": 1.263,
2328
- "eval_wer": 97.27487034417727,
2329
- "step": 8000
2330
- },
2331
- {
2332
- "epoch": 13.242774566473988,
2333
- "grad_norm": 4.355146408081055,
2334
- "learning_rate": 4.162105263157895e-06,
2335
- "loss": 0.207,
2336
- "step": 8025
2337
- },
2338
- {
2339
- "epoch": 13.284062758051197,
2340
- "grad_norm": 4.026797771453857,
2341
- "learning_rate": 4.109473684210527e-06,
2342
- "loss": 0.1656,
2343
- "step": 8050
2344
- },
2345
- {
2346
- "epoch": 13.325350949628406,
2347
- "grad_norm": 4.001533031463623,
2348
- "learning_rate": 4.056842105263158e-06,
2349
- "loss": 0.1899,
2350
- "step": 8075
2351
- },
2352
- {
2353
- "epoch": 13.366639141205615,
2354
- "grad_norm": 3.8727056980133057,
2355
- "learning_rate": 4.00421052631579e-06,
2356
- "loss": 0.1919,
2357
- "step": 8100
2358
- },
2359
- {
2360
- "epoch": 13.407927332782824,
2361
- "grad_norm": 4.644744873046875,
2362
- "learning_rate": 3.951578947368422e-06,
2363
- "loss": 0.1987,
2364
- "step": 8125
2365
- },
2366
- {
2367
- "epoch": 13.449215524360033,
2368
- "grad_norm": 4.7693071365356445,
2369
- "learning_rate": 3.898947368421052e-06,
2370
- "loss": 0.185,
2371
- "step": 8150
2372
- },
2373
- {
2374
- "epoch": 13.490503715937242,
2375
- "grad_norm": 5.065967082977295,
2376
- "learning_rate": 3.846315789473685e-06,
2377
- "loss": 0.1811,
2378
- "step": 8175
2379
- },
2380
- {
2381
- "epoch": 13.531791907514451,
2382
- "grad_norm": 3.682410955429077,
2383
- "learning_rate": 3.793684210526316e-06,
2384
- "loss": 0.2118,
2385
- "step": 8200
2386
- },
2387
- {
2388
- "epoch": 13.57308009909166,
2389
- "grad_norm": 4.680816650390625,
2390
- "learning_rate": 3.7410526315789473e-06,
2391
- "loss": 0.1995,
2392
- "step": 8225
2393
- },
2394
- {
2395
- "epoch": 13.61436829066887,
2396
- "grad_norm": 5.313553333282471,
2397
- "learning_rate": 3.6884210526315794e-06,
2398
- "loss": 0.1871,
2399
- "step": 8250
2400
- },
2401
- {
2402
- "epoch": 13.655656482246078,
2403
- "grad_norm": 4.104048728942871,
2404
- "learning_rate": 3.635789473684211e-06,
2405
- "loss": 0.1861,
2406
- "step": 8275
2407
- },
2408
- {
2409
- "epoch": 13.696944673823287,
2410
- "grad_norm": 4.6280293464660645,
2411
- "learning_rate": 3.5831578947368422e-06,
2412
- "loss": 0.1936,
2413
- "step": 8300
2414
- },
2415
- {
2416
- "epoch": 13.738232865400496,
2417
- "grad_norm": 5.779093265533447,
2418
- "learning_rate": 3.5305263157894743e-06,
2419
- "loss": 0.1899,
2420
- "step": 8325
2421
- },
2422
- {
2423
- "epoch": 13.779521056977705,
2424
- "grad_norm": 3.84360933303833,
2425
- "learning_rate": 3.4778947368421055e-06,
2426
- "loss": 0.1937,
2427
- "step": 8350
2428
- },
2429
- {
2430
- "epoch": 13.820809248554914,
2431
- "grad_norm": 5.995066165924072,
2432
- "learning_rate": 3.425263157894737e-06,
2433
- "loss": 0.1976,
2434
- "step": 8375
2435
- },
2436
- {
2437
- "epoch": 13.862097440132121,
2438
- "grad_norm": 5.045938491821289,
2439
- "learning_rate": 3.3726315789473683e-06,
2440
- "loss": 0.177,
2441
- "step": 8400
2442
- },
2443
- {
2444
- "epoch": 13.90338563170933,
2445
- "grad_norm": 4.3831658363342285,
2446
- "learning_rate": 3.3200000000000004e-06,
2447
- "loss": 0.1831,
2448
- "step": 8425
2449
- },
2450
- {
2451
- "epoch": 13.94467382328654,
2452
- "grad_norm": 5.6045002937316895,
2453
- "learning_rate": 3.267368421052632e-06,
2454
- "loss": 0.177,
2455
- "step": 8450
2456
- },
2457
- {
2458
- "epoch": 13.985962014863748,
2459
- "grad_norm": 5.313602447509766,
2460
- "learning_rate": 3.2147368421052633e-06,
2461
- "loss": 0.2106,
2462
- "step": 8475
2463
- },
2464
- {
2465
- "epoch": 14.026424442609414,
2466
- "grad_norm": 4.277268409729004,
2467
- "learning_rate": 3.1621052631578953e-06,
2468
- "loss": 0.1697,
2469
- "step": 8500
2470
- },
2471
- {
2472
- "epoch": 14.067712634186623,
2473
- "grad_norm": 4.43281364440918,
2474
- "learning_rate": 3.1094736842105265e-06,
2475
- "loss": 0.1655,
2476
- "step": 8525
2477
- },
2478
- {
2479
- "epoch": 14.109000825763832,
2480
- "grad_norm": 5.5066304206848145,
2481
- "learning_rate": 3.056842105263158e-06,
2482
- "loss": 0.1611,
2483
- "step": 8550
2484
- },
2485
- {
2486
- "epoch": 14.15028901734104,
2487
- "grad_norm": 5.53204345703125,
2488
- "learning_rate": 3.0042105263157894e-06,
2489
- "loss": 0.1808,
2490
- "step": 8575
2491
- },
2492
- {
2493
- "epoch": 14.191577208918249,
2494
- "grad_norm": 4.399189472198486,
2495
- "learning_rate": 2.9515789473684214e-06,
2496
- "loss": 0.1487,
2497
- "step": 8600
2498
- },
2499
- {
2500
- "epoch": 14.232865400495458,
2501
- "grad_norm": 4.710522174835205,
2502
- "learning_rate": 2.8989473684210526e-06,
2503
- "loss": 0.1684,
2504
- "step": 8625
2505
- },
2506
- {
2507
- "epoch": 14.274153592072667,
2508
- "grad_norm": 4.319856643676758,
2509
- "learning_rate": 2.8463157894736843e-06,
2510
- "loss": 0.1626,
2511
- "step": 8650
2512
- },
2513
- {
2514
- "epoch": 14.315441783649876,
2515
- "grad_norm": 5.691439628601074,
2516
- "learning_rate": 2.7936842105263163e-06,
2517
- "loss": 0.1734,
2518
- "step": 8675
2519
- },
2520
- {
2521
- "epoch": 14.356729975227084,
2522
- "grad_norm": 5.005030632019043,
2523
- "learning_rate": 2.7410526315789476e-06,
2524
- "loss": 0.1627,
2525
- "step": 8700
2526
- },
2527
- {
2528
- "epoch": 14.398018166804293,
2529
- "grad_norm": 4.130620002746582,
2530
- "learning_rate": 2.688421052631579e-06,
2531
- "loss": 0.1718,
2532
- "step": 8725
2533
- },
2534
- {
2535
- "epoch": 14.439306358381502,
2536
- "grad_norm": 3.437147617340088,
2537
- "learning_rate": 2.635789473684211e-06,
2538
- "loss": 0.1756,
2539
- "step": 8750
2540
- },
2541
- {
2542
- "epoch": 14.480594549958711,
2543
- "grad_norm": 4.4854044914245605,
2544
- "learning_rate": 2.5831578947368425e-06,
2545
- "loss": 0.1671,
2546
- "step": 8775
2547
- },
2548
- {
2549
- "epoch": 14.52188274153592,
2550
- "grad_norm": 5.259097099304199,
2551
- "learning_rate": 2.5305263157894737e-06,
2552
- "loss": 0.1782,
2553
- "step": 8800
2554
- },
2555
- {
2556
- "epoch": 14.56317093311313,
2557
- "grad_norm": 3.7995481491088867,
2558
- "learning_rate": 2.4778947368421053e-06,
2559
- "loss": 0.1684,
2560
- "step": 8825
2561
- },
2562
- {
2563
- "epoch": 14.604459124690338,
2564
- "grad_norm": 4.331813335418701,
2565
- "learning_rate": 2.425263157894737e-06,
2566
- "loss": 0.1757,
2567
- "step": 8850
2568
- },
2569
- {
2570
- "epoch": 14.645747316267547,
2571
- "grad_norm": 5.367282867431641,
2572
- "learning_rate": 2.3726315789473686e-06,
2573
- "loss": 0.1796,
2574
- "step": 8875
2575
- },
2576
- {
2577
- "epoch": 14.687035507844756,
2578
- "grad_norm": 5.606546401977539,
2579
- "learning_rate": 2.3200000000000002e-06,
2580
- "loss": 0.1911,
2581
- "step": 8900
2582
- },
2583
- {
2584
- "epoch": 14.728323699421965,
2585
- "grad_norm": 3.9387214183807373,
2586
- "learning_rate": 2.267368421052632e-06,
2587
- "loss": 0.1666,
2588
- "step": 8925
2589
- },
2590
- {
2591
- "epoch": 14.769611890999174,
2592
- "grad_norm": 4.426852703094482,
2593
- "learning_rate": 2.2147368421052635e-06,
2594
- "loss": 0.1717,
2595
- "step": 8950
2596
- },
2597
- {
2598
- "epoch": 14.810900082576383,
2599
- "grad_norm": 3.4610788822174072,
2600
- "learning_rate": 2.1621052631578947e-06,
2601
- "loss": 0.168,
2602
- "step": 8975
2603
- },
2604
- {
2605
- "epoch": 14.852188274153592,
2606
- "grad_norm": 4.5235676765441895,
2607
- "learning_rate": 2.1094736842105264e-06,
2608
- "loss": 0.1425,
2609
- "step": 9000
2610
- },
2611
- {
2612
- "epoch": 14.852188274153592,
2613
- "eval_cer": 48.44840828585512,
2614
- "eval_loss": 1.142202377319336,
2615
- "eval_runtime": 701.7243,
2616
- "eval_samples_per_second": 15.079,
2617
- "eval_steps_per_second": 1.257,
2618
- "eval_wer": 96.61480433757662,
2619
- "step": 9000
2620
- },
2621
- {
2622
- "epoch": 14.893476465730801,
2623
- "grad_norm": 4.106043815612793,
2624
- "learning_rate": 2.058947368421053e-06,
2625
- "loss": 0.1853,
2626
- "step": 9025
2627
- },
2628
- {
2629
- "epoch": 14.93476465730801,
2630
- "grad_norm": 4.604458808898926,
2631
- "learning_rate": 2.0063157894736843e-06,
2632
- "loss": 0.1725,
2633
- "step": 9050
2634
- },
2635
- {
2636
- "epoch": 14.97605284888522,
2637
- "grad_norm": 5.01749324798584,
2638
- "learning_rate": 1.953684210526316e-06,
2639
- "loss": 0.1566,
2640
- "step": 9075
2641
- },
2642
- {
2643
- "epoch": 15.016515276630884,
2644
- "grad_norm": 3.69610857963562,
2645
- "learning_rate": 1.9010526315789476e-06,
2646
- "loss": 0.1791,
2647
- "step": 9100
2648
- },
2649
- {
2650
- "epoch": 15.057803468208093,
2651
- "grad_norm": 3.566075563430786,
2652
- "learning_rate": 1.848421052631579e-06,
2653
- "loss": 0.1626,
2654
- "step": 9125
2655
- },
2656
- {
2657
- "epoch": 15.099091659785302,
2658
- "grad_norm": 3.8838043212890625,
2659
- "learning_rate": 1.7957894736842108e-06,
2660
- "loss": 0.1507,
2661
- "step": 9150
2662
- },
2663
- {
2664
- "epoch": 15.14037985136251,
2665
- "grad_norm": 4.462072372436523,
2666
- "learning_rate": 1.7431578947368423e-06,
2667
- "loss": 0.1682,
2668
- "step": 9175
2669
- },
2670
- {
2671
- "epoch": 15.18166804293972,
2672
- "grad_norm": 5.543285846710205,
2673
- "learning_rate": 1.6905263157894739e-06,
2674
- "loss": 0.141,
2675
- "step": 9200
2676
- },
2677
- {
2678
- "epoch": 15.222956234516928,
2679
- "grad_norm": 4.328714847564697,
2680
- "learning_rate": 1.6378947368421053e-06,
2681
- "loss": 0.1747,
2682
- "step": 9225
2683
- },
2684
- {
2685
- "epoch": 15.264244426094137,
2686
- "grad_norm": 3.924700975418091,
2687
- "learning_rate": 1.585263157894737e-06,
2688
- "loss": 0.1727,
2689
- "step": 9250
2690
- },
2691
- {
2692
- "epoch": 15.305532617671346,
2693
- "grad_norm": 4.647414207458496,
2694
- "learning_rate": 1.5326315789473686e-06,
2695
- "loss": 0.1512,
2696
- "step": 9275
2697
- },
2698
- {
2699
- "epoch": 15.346820809248555,
2700
- "grad_norm": 3.2526192665100098,
2701
- "learning_rate": 1.48e-06,
2702
- "loss": 0.1521,
2703
- "step": 9300
2704
- },
2705
- {
2706
- "epoch": 15.388109000825764,
2707
- "grad_norm": 3.697683334350586,
2708
- "learning_rate": 1.4273684210526317e-06,
2709
- "loss": 0.1498,
2710
- "step": 9325
2711
- },
2712
- {
2713
- "epoch": 15.429397192402973,
2714
- "grad_norm": 3.709627866744995,
2715
- "learning_rate": 1.374736842105263e-06,
2716
- "loss": 0.1654,
2717
- "step": 9350
2718
- },
2719
- {
2720
- "epoch": 15.470685383980182,
2721
- "grad_norm": 3.7767739295959473,
2722
- "learning_rate": 1.322105263157895e-06,
2723
- "loss": 0.1466,
2724
- "step": 9375
2725
- },
2726
- {
2727
- "epoch": 15.511973575557391,
2728
- "grad_norm": 3.6882431507110596,
2729
- "learning_rate": 1.2694736842105266e-06,
2730
- "loss": 0.1496,
2731
- "step": 9400
2732
- },
2733
- {
2734
- "epoch": 15.5532617671346,
2735
- "grad_norm": 3.9242100715637207,
2736
- "learning_rate": 1.216842105263158e-06,
2737
- "loss": 0.1491,
2738
- "step": 9425
2739
- },
2740
- {
2741
- "epoch": 15.59454995871181,
2742
- "grad_norm": 5.024509429931641,
2743
- "learning_rate": 1.1642105263157896e-06,
2744
- "loss": 0.1549,
2745
- "step": 9450
2746
- },
2747
- {
2748
- "epoch": 15.635838150289018,
2749
- "grad_norm": 4.048088550567627,
2750
- "learning_rate": 1.1115789473684213e-06,
2751
- "loss": 0.1564,
2752
- "step": 9475
2753
- },
2754
- {
2755
- "epoch": 15.677126341866225,
2756
- "grad_norm": 4.535757541656494,
2757
- "learning_rate": 1.0589473684210527e-06,
2758
- "loss": 0.1548,
2759
- "step": 9500
2760
- },
2761
- {
2762
- "epoch": 15.718414533443434,
2763
- "grad_norm": 4.47163200378418,
2764
- "learning_rate": 1.0063157894736843e-06,
2765
- "loss": 0.1521,
2766
- "step": 9525
2767
- },
2768
- {
2769
- "epoch": 15.759702725020643,
2770
- "grad_norm": 3.4826717376708984,
2771
- "learning_rate": 9.536842105263158e-07,
2772
- "loss": 0.151,
2773
- "step": 9550
2774
- },
2775
- {
2776
- "epoch": 15.800990916597852,
2777
- "grad_norm": 3.6962897777557373,
2778
- "learning_rate": 9.010526315789474e-07,
2779
- "loss": 0.1611,
2780
- "step": 9575
2781
- },
2782
- {
2783
- "epoch": 15.842279108175061,
2784
- "grad_norm": 3.6767818927764893,
2785
- "learning_rate": 8.48421052631579e-07,
2786
- "loss": 0.1594,
2787
- "step": 9600
2788
- },
2789
- {
2790
- "epoch": 15.88356729975227,
2791
- "grad_norm": 2.950598955154419,
2792
- "learning_rate": 7.957894736842107e-07,
2793
- "loss": 0.1583,
2794
- "step": 9625
2795
- },
2796
- {
2797
- "epoch": 15.92485549132948,
2798
- "grad_norm": 4.535677909851074,
2799
- "learning_rate": 7.431578947368422e-07,
2800
- "loss": 0.1743,
2801
- "step": 9650
2802
- },
2803
- {
2804
- "epoch": 15.966143682906688,
2805
- "grad_norm": 3.578261613845825,
2806
- "learning_rate": 6.905263157894737e-07,
2807
- "loss": 0.1279,
2808
- "step": 9675
2809
- },
2810
- {
2811
- "epoch": 16.006606110652353,
2812
- "grad_norm": 3.0115976333618164,
2813
- "learning_rate": 6.378947368421053e-07,
2814
- "loss": 0.1442,
2815
- "step": 9700
2816
- },
2817
- {
2818
- "epoch": 16.04789430222956,
2819
- "grad_norm": 4.302647590637207,
2820
- "learning_rate": 5.852631578947369e-07,
2821
- "loss": 0.1524,
2822
- "step": 9725
2823
- },
2824
- {
2825
- "epoch": 16.08918249380677,
2826
- "grad_norm": 3.516667366027832,
2827
- "learning_rate": 5.326315789473684e-07,
2828
- "loss": 0.1431,
2829
- "step": 9750
2830
- },
2831
- {
2832
- "epoch": 16.13047068538398,
2833
- "grad_norm": 3.802489757537842,
2834
- "learning_rate": 4.800000000000001e-07,
2835
- "loss": 0.1607,
2836
- "step": 9775
2837
- },
2838
- {
2839
- "epoch": 16.17175887696119,
2840
- "grad_norm": 2.9078941345214844,
2841
- "learning_rate": 4.273684210526316e-07,
2842
- "loss": 0.134,
2843
- "step": 9800
2844
- },
2845
- {
2846
- "epoch": 16.213047068538398,
2847
- "grad_norm": 3.3042197227478027,
2848
- "learning_rate": 3.7473684210526323e-07,
2849
- "loss": 0.1614,
2850
- "step": 9825
2851
- },
2852
- {
2853
- "epoch": 16.254335260115607,
2854
- "grad_norm": 4.696882724761963,
2855
- "learning_rate": 3.2210526315789476e-07,
2856
- "loss": 0.1379,
2857
- "step": 9850
2858
- },
2859
- {
2860
- "epoch": 16.295623451692816,
2861
- "grad_norm": 4.522494316101074,
2862
- "learning_rate": 2.6947368421052635e-07,
2863
- "loss": 0.1586,
2864
- "step": 9875
2865
- },
2866
- {
2867
- "epoch": 16.336911643270025,
2868
- "grad_norm": 4.257038593292236,
2869
- "learning_rate": 2.168421052631579e-07,
2870
- "loss": 0.1423,
2871
- "step": 9900
2872
- },
2873
- {
2874
- "epoch": 16.378199834847234,
2875
- "grad_norm": 4.159156799316406,
2876
- "learning_rate": 1.642105263157895e-07,
2877
- "loss": 0.1617,
2878
- "step": 9925
2879
- },
2880
- {
2881
- "epoch": 16.419488026424442,
2882
- "grad_norm": 4.338253974914551,
2883
- "learning_rate": 1.1157894736842106e-07,
2884
- "loss": 0.1597,
2885
- "step": 9950
2886
- },
2887
- {
2888
- "epoch": 16.46077621800165,
2889
- "grad_norm": 5.437859058380127,
2890
- "learning_rate": 5.8947368421052637e-08,
2891
- "loss": 0.1629,
2892
- "step": 9975
2893
- },
2894
- {
2895
- "epoch": 16.50206440957886,
2896
- "grad_norm": 4.178699970245361,
2897
- "learning_rate": 6.315789473684211e-09,
2898
- "loss": 0.161,
2899
- "step": 10000
2900
- },
2901
- {
2902
- "epoch": 16.50206440957886,
2903
- "eval_cer": 49.64348354407163,
2904
- "eval_loss": 1.1519368886947632,
2905
- "eval_runtime": 708.8725,
2906
- "eval_samples_per_second": 14.927,
2907
- "eval_steps_per_second": 1.244,
2908
- "eval_wer": 96.77510608203677,
2909
- "step": 10000
2910
  }
2911
  ],
2912
  "logging_steps": 25,
2913
- "max_steps": 10000,
2914
  "num_input_tokens_seen": 0,
2915
- "num_train_epochs": 17,
2916
  "save_steps": 1000,
2917
  "stateful_callbacks": {
2918
  "TrainerControl": {
@@ -2921,13 +311,13 @@
2921
  "should_evaluate": false,
2922
  "should_log": false,
2923
  "should_save": true,
2924
- "should_training_stop": true
2925
  },
2926
  "attributes": {}
2927
  }
2928
  },
2929
- "total_flos": 1.949150849531904e+19,
2930
- "train_batch_size": 12,
2931
  "trial_name": null,
2932
  "trial_params": null
2933
  }
 
1
  {
2
+ "best_metric": 99.20792079207921,
3
+ "best_model_checkpoint": "./iteboshi_temp/checkpoint-1000",
4
+ "epoch": 1.1013215859030836,
5
  "eval_steps": 1000,
6
+ "global_step": 1000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.02753303964757709,
13
+ "grad_norm": 8.593878746032715,
14
  "learning_rate": 1.0000000000000002e-06,
15
+ "loss": 11.2184,
16
  "step": 25
17
  },
18
  {
19
+ "epoch": 0.05506607929515418,
20
+ "grad_norm": 5.720788955688477,
21
+ "learning_rate": 2.0000000000000003e-06,
22
+ "loss": 9.9763,
23
  "step": 50
24
  },
25
  {
26
+ "epoch": 0.08259911894273128,
27
+ "grad_norm": 3.1753361225128174,
28
+ "learning_rate": 3e-06,
29
+ "loss": 7.2986,
30
  "step": 75
31
  },
32
  {
33
+ "epoch": 0.11013215859030837,
34
+ "grad_norm": 1.9015727043151855,
35
+ "learning_rate": 4.000000000000001e-06,
36
+ "loss": 5.6306,
37
  "step": 100
38
  },
39
  {
40
+ "epoch": 0.13766519823788545,
41
+ "grad_norm": 1.7933436632156372,
42
+ "learning_rate": 5e-06,
43
+ "loss": 4.7211,
44
  "step": 125
45
  },
46
  {
47
+ "epoch": 0.16519823788546256,
48
+ "grad_norm": 1.9143211841583252,
49
+ "learning_rate": 6e-06,
50
+ "loss": 4.1871,
51
  "step": 150
52
  },
53
  {
54
+ "epoch": 0.19273127753303965,
55
+ "grad_norm": 1.8120731115341187,
56
+ "learning_rate": 7e-06,
57
+ "loss": 3.728,
58
  "step": 175
59
  },
60
  {
61
+ "epoch": 0.22026431718061673,
62
+ "grad_norm": 1.5976663827896118,
63
+ "learning_rate": 8.000000000000001e-06,
64
+ "loss": 3.3999,
65
  "step": 200
66
  },
67
  {
68
+ "epoch": 0.24779735682819384,
69
+ "grad_norm": 1.4841374158859253,
70
+ "learning_rate": 9e-06,
71
+ "loss": 3.1248,
72
  "step": 225
73
  },
74
  {
75
+ "epoch": 0.2753303964757709,
76
+ "grad_norm": 1.571273922920227,
77
+ "learning_rate": 1e-05,
78
+ "loss": 2.9023,
79
  "step": 250
80
  },
81
  {
82
+ "epoch": 0.30286343612334804,
83
+ "grad_norm": 1.5719149112701416,
84
+ "learning_rate": 1.1000000000000001e-05,
85
+ "loss": 2.6609,
86
  "step": 275
87
  },
88
  {
89
+ "epoch": 0.3303964757709251,
90
+ "grad_norm": 1.4737190008163452,
91
+ "learning_rate": 1.2e-05,
92
+ "loss": 2.4913,
93
  "step": 300
94
  },
95
  {
96
+ "epoch": 0.3579295154185022,
97
+ "grad_norm": 1.5062530040740967,
98
+ "learning_rate": 1.3000000000000001e-05,
99
+ "loss": 2.3478,
100
  "step": 325
101
  },
102
  {
103
+ "epoch": 0.3854625550660793,
104
+ "grad_norm": 1.5584601163864136,
105
+ "learning_rate": 1.4e-05,
106
+ "loss": 2.2405,
107
  "step": 350
108
  },
109
  {
110
+ "epoch": 0.4129955947136564,
111
+ "grad_norm": 1.5639300346374512,
112
+ "learning_rate": 1.5000000000000002e-05,
113
+ "loss": 2.0632,
114
  "step": 375
115
  },
116
  {
117
+ "epoch": 0.44052863436123346,
118
+ "grad_norm": 1.574596881866455,
119
+ "learning_rate": 1.6000000000000003e-05,
120
+ "loss": 1.9272,
121
  "step": 400
122
  },
123
  {
124
+ "epoch": 0.46806167400881055,
125
+ "grad_norm": 1.5253156423568726,
126
+ "learning_rate": 1.7e-05,
127
+ "loss": 1.8261,
128
  "step": 425
129
  },
130
  {
131
+ "epoch": 0.4955947136563877,
132
+ "grad_norm": 1.5775047540664673,
133
+ "learning_rate": 1.8e-05,
134
+ "loss": 1.7746,
135
  "step": 450
136
  },
137
  {
138
+ "epoch": 0.5231277533039648,
139
+ "grad_norm": 1.388596534729004,
140
+ "learning_rate": 1.9e-05,
141
+ "loss": 1.709,
142
  "step": 475
143
  },
144
  {
145
+ "epoch": 0.5506607929515418,
146
+ "grad_norm": 1.2958333492279053,
147
+ "learning_rate": 2e-05,
148
+ "loss": 1.6352,
149
  "step": 500
150
  },
151
  {
152
+ "epoch": 0.5781938325991189,
153
+ "grad_norm": 1.3782161474227905,
154
+ "learning_rate": 1.9974358974358975e-05,
155
+ "loss": 1.5945,
156
  "step": 525
157
  },
158
  {
159
+ "epoch": 0.6057268722466961,
160
+ "grad_norm": 1.3231383562088013,
161
+ "learning_rate": 1.994871794871795e-05,
162
+ "loss": 1.5081,
163
  "step": 550
164
  },
165
  {
166
+ "epoch": 0.6332599118942731,
167
+ "grad_norm": 1.4640569686889648,
168
+ "learning_rate": 1.9923076923076926e-05,
169
+ "loss": 1.4445,
170
  "step": 575
171
  },
172
  {
173
+ "epoch": 0.6607929515418502,
174
+ "grad_norm": 1.2492103576660156,
175
+ "learning_rate": 1.98974358974359e-05,
176
+ "loss": 1.4051,
177
  "step": 600
178
  },
179
  {
180
+ "epoch": 0.6883259911894273,
181
+ "grad_norm": 1.7467623949050903,
182
+ "learning_rate": 1.9871794871794873e-05,
183
+ "loss": 1.3562,
184
  "step": 625
185
  },
186
  {
187
+ "epoch": 0.7158590308370044,
188
+ "grad_norm": 1.420554757118225,
189
+ "learning_rate": 1.9846153846153847e-05,
190
+ "loss": 1.3596,
191
  "step": 650
192
  },
193
  {
194
+ "epoch": 0.7433920704845814,
195
+ "grad_norm": 1.239926815032959,
196
+ "learning_rate": 1.9820512820512824e-05,
197
+ "loss": 1.3058,
198
  "step": 675
199
  },
200
  {
201
+ "epoch": 0.7709251101321586,
202
+ "grad_norm": 1.1721984148025513,
203
+ "learning_rate": 1.9794871794871798e-05,
204
+ "loss": 1.2863,
205
  "step": 700
206
  },
207
  {
208
+ "epoch": 0.7984581497797357,
209
+ "grad_norm": 1.3416831493377686,
210
+ "learning_rate": 1.976923076923077e-05,
211
+ "loss": 1.3276,
212
  "step": 725
213
  },
214
  {
215
+ "epoch": 0.8259911894273128,
216
+ "grad_norm": 1.4568818807601929,
217
+ "learning_rate": 1.9743589743589745e-05,
218
+ "loss": 1.3027,
219
  "step": 750
220
  },
221
  {
222
+ "epoch": 0.8535242290748899,
223
+ "grad_norm": 1.4908838272094727,
224
+ "learning_rate": 1.9717948717948722e-05,
225
+ "loss": 1.2134,
226
  "step": 775
227
  },
228
  {
229
+ "epoch": 0.8810572687224669,
230
+ "grad_norm": 1.3530240058898926,
231
+ "learning_rate": 1.9692307692307696e-05,
232
+ "loss": 1.2212,
233
  "step": 800
234
  },
235
  {
236
+ "epoch": 0.9085903083700441,
237
+ "grad_norm": 1.1275429725646973,
238
+ "learning_rate": 1.9666666666666666e-05,
239
+ "loss": 1.1196,
240
  "step": 825
241
  },
242
  {
243
+ "epoch": 0.9361233480176211,
244
+ "grad_norm": 1.3054031133651733,
245
+ "learning_rate": 1.9641025641025643e-05,
246
+ "loss": 1.14,
247
  "step": 850
248
  },
249
  {
250
+ "epoch": 0.9636563876651982,
251
+ "grad_norm": 1.0485919713974,
252
+ "learning_rate": 1.9615384615384617e-05,
253
+ "loss": 1.1557,
254
  "step": 875
255
  },
256
  {
257
+ "epoch": 0.9911894273127754,
258
+ "grad_norm": 1.1853173971176147,
259
+ "learning_rate": 1.958974358974359e-05,
260
+ "loss": 1.1259,
261
  "step": 900
262
  },
263
  {
264
+ "epoch": 1.0187224669603525,
265
+ "grad_norm": 1.0613934993743896,
266
+ "learning_rate": 1.9564102564102564e-05,
267
+ "loss": 1.0608,
268
  "step": 925
269
  },
270
  {
271
+ "epoch": 1.0462555066079295,
272
+ "grad_norm": 1.2642765045166016,
273
+ "learning_rate": 1.953846153846154e-05,
274
+ "loss": 0.9548,
275
  "step": 950
276
  },
277
  {
278
+ "epoch": 1.0737885462555066,
279
+ "grad_norm": 1.2651335000991821,
280
+ "learning_rate": 1.9512820512820515e-05,
281
+ "loss": 0.9568,
282
  "step": 975
283
  },
284
  {
285
+ "epoch": 1.1013215859030836,
286
+ "grad_norm": 1.0594816207885742,
287
+ "learning_rate": 1.9487179487179488e-05,
288
+ "loss": 0.9609,
289
  "step": 1000
290
  },
291
  {
292
+ "epoch": 1.1013215859030836,
293
+ "eval_cer": 98.0001834168262,
294
+ "eval_loss": 1.1190180778503418,
295
+ "eval_runtime": 1035.5852,
296
+ "eval_samples_per_second": 10.217,
297
+ "eval_steps_per_second": 2.555,
298
+ "eval_wer": 99.20792079207921,
299
  "step": 1000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
300
  }
301
  ],
302
  "logging_steps": 25,
303
+ "max_steps": 20000,
304
  "num_input_tokens_seen": 0,
305
+ "num_train_epochs": 23,
306
  "save_steps": 1000,
307
  "stateful_callbacks": {
308
  "TrainerControl": {
 
311
  "should_evaluate": false,
312
  "should_log": false,
313
  "should_save": true,
314
+ "should_training_stop": false
315
  },
316
  "attributes": {}
317
  }
318
  },
319
+ "total_flos": 1.6881707400350925e+18,
320
+ "train_batch_size": 4,
321
  "trial_name": null,
322
  "trial_params": null
323
  }
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8fe144e7bbecfefafbc5f7ba6e25af714f15b9643283e39c2a95c6ef7dcd775
3
- size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6af773437c2683442b487a292e99dfac19fe476b5315a9fe09845da921d84893
3
+ size 6776
last-checkpoint/vocab.json CHANGED
@@ -314,7 +314,6 @@
314
  ";;": 35746,
315
  "<": 27,
316
  "</": 3433,
317
- "<|endoftext|>": 50257,
318
  "=": 28,
319
  "=\"": 13114,
320
  "=\"#": 34106,
 
314
  ";;": 35746,
315
  "<": 27,
316
  "</": 3433,
 
317
  "=": 28,
318
  "=\"": 13114,
319
  "=\"#": 34106,
last-checkpoint/zero_to_fp32.py ADDED
@@ -0,0 +1,760 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/usr/bin/env python
2
+
3
+ # Copyright (c) Microsoft Corporation.
4
+ # SPDX-License-Identifier: Apache-2.0
5
+
6
+ # DeepSpeed Team
7
+
8
+ # This script extracts fp32 consolidated weights from a zero 1, 2 and 3 DeepSpeed checkpoints. It gets
9
+ # copied into the top level checkpoint dir, so the user can easily do the conversion at any point in
10
+ # the future. Once extracted, the weights don't require DeepSpeed and can be used in any
11
+ # application.
12
+ #
13
+ # example:
14
+ # python zero_to_fp32.py . output_dir/
15
+ # or
16
+ # python zero_to_fp32.py . output_dir/ --safe_serialization
17
+
18
+ import argparse
19
+ import torch
20
+ import glob
21
+ import math
22
+ import os
23
+ import re
24
+ import gc
25
+ import json
26
+ import numpy as np
27
+ from tqdm import tqdm
28
+ from collections import OrderedDict
29
+ from dataclasses import dataclass
30
+
31
+ # while this script doesn't use deepspeed to recover data, since the checkpoints are pickled with
32
+ # DeepSpeed data structures it has to be available in the current python environment.
33
+ from deepspeed.utils import logger
34
+ from deepspeed.checkpoint.constants import (DS_VERSION, OPTIMIZER_STATE_DICT, SINGLE_PARTITION_OF_FP32_GROUPS,
35
+ FP32_FLAT_GROUPS, ZERO_STAGE, PARTITION_COUNT, PARAM_SHAPES, BUFFER_NAMES,
36
+ FROZEN_PARAM_SHAPES, FROZEN_PARAM_FRAGMENTS)
37
+
38
+
39
+ @dataclass
40
+ class zero_model_state:
41
+ buffers: dict()
42
+ param_shapes: dict()
43
+ shared_params: list
44
+ ds_version: int
45
+ frozen_param_shapes: dict()
46
+ frozen_param_fragments: dict()
47
+
48
+
49
+ debug = 0
50
+
51
+ # load to cpu
52
+ device = torch.device('cpu')
53
+
54
+
55
+ def atoi(text):
56
+ return int(text) if text.isdigit() else text
57
+
58
+
59
+ def natural_keys(text):
60
+ '''
61
+ alist.sort(key=natural_keys) sorts in human order
62
+ http://nedbatchelder.com/blog/200712/human_sorting.html
63
+ (See Toothy's implementation in the comments)
64
+ '''
65
+ return [atoi(c) for c in re.split(r'(\d+)', text)]
66
+
67
+
68
+ def get_model_state_file(checkpoint_dir, zero_stage):
69
+ if not os.path.isdir(checkpoint_dir):
70
+ raise FileNotFoundError(f"Directory '{checkpoint_dir}' doesn't exist")
71
+
72
+ # there should be only one file
73
+ if zero_stage <= 2:
74
+ file = os.path.join(checkpoint_dir, "mp_rank_00_model_states.pt")
75
+ elif zero_stage == 3:
76
+ file = os.path.join(checkpoint_dir, "zero_pp_rank_0_mp_rank_00_model_states.pt")
77
+
78
+ if not os.path.exists(file):
79
+ raise FileNotFoundError(f"can't find model states file at '{file}'")
80
+
81
+ return file
82
+
83
+
84
+ def get_checkpoint_files(checkpoint_dir, glob_pattern):
85
+ # XXX: need to test that this simple glob rule works for multi-node setup too
86
+ ckpt_files = sorted(glob.glob(os.path.join(checkpoint_dir, glob_pattern)), key=natural_keys)
87
+
88
+ if len(ckpt_files) == 0:
89
+ raise FileNotFoundError(f"can't find {glob_pattern} files in directory '{checkpoint_dir}'")
90
+
91
+ return ckpt_files
92
+
93
+
94
+ def get_optim_files(checkpoint_dir):
95
+ return get_checkpoint_files(checkpoint_dir, "*_optim_states.pt")
96
+
97
+
98
+ def get_model_state_files(checkpoint_dir):
99
+ return get_checkpoint_files(checkpoint_dir, "*_model_states.pt")
100
+
101
+
102
+ def parse_model_states(files):
103
+ zero_model_states = []
104
+ for file in files:
105
+ state_dict = torch.load(file, map_location=device, weights_only=False)
106
+
107
+ if BUFFER_NAMES not in state_dict:
108
+ raise ValueError(f"{file} is not a model state checkpoint")
109
+ buffer_names = state_dict[BUFFER_NAMES]
110
+ if debug:
111
+ print("Found buffers:", buffer_names)
112
+
113
+ # recover just the buffers while restoring them to fp32 if they were saved in fp16
114
+ buffers = {k: v.float() for k, v in state_dict["module"].items() if k in buffer_names}
115
+ param_shapes = state_dict[PARAM_SHAPES]
116
+
117
+ # collect parameters that are included in param_shapes
118
+ param_names = []
119
+ for s in param_shapes:
120
+ for name in s.keys():
121
+ param_names.append(name)
122
+
123
+ # update with frozen parameters
124
+ frozen_param_shapes = state_dict.get(FROZEN_PARAM_SHAPES, None)
125
+ if frozen_param_shapes is not None:
126
+ if debug:
127
+ print(f"Found frozen_param_shapes: {frozen_param_shapes}")
128
+ param_names += list(frozen_param_shapes.keys())
129
+
130
+ # handle shared params
131
+ shared_params = [[k, v] for k, v in state_dict["shared_params"].items()]
132
+
133
+ ds_version = state_dict.get(DS_VERSION, None)
134
+
135
+ frozen_param_fragments = state_dict.get(FROZEN_PARAM_FRAGMENTS, None)
136
+
137
+ z_model_state = zero_model_state(buffers=buffers,
138
+ param_shapes=param_shapes,
139
+ shared_params=shared_params,
140
+ ds_version=ds_version,
141
+ frozen_param_shapes=frozen_param_shapes,
142
+ frozen_param_fragments=frozen_param_fragments)
143
+ zero_model_states.append(z_model_state)
144
+
145
+ return zero_model_states
146
+
147
+
148
+ def parse_optim_states(files, ds_checkpoint_dir):
149
+ total_files = len(files)
150
+ state_dicts = []
151
+ for f in tqdm(files, desc='Loading checkpoint shards'):
152
+ state_dict = torch.load(f, map_location=device, mmap=True, weights_only=False)
153
+ # immediately discard the potentially huge 2 optimizer states as we only care for fp32 master weights
154
+ # and also handle the case where it was already removed by another helper script
155
+ state_dict["optimizer_state_dict"].pop("optimizer_state_dict", None)
156
+ state_dicts.append(state_dict)
157
+
158
+ if not ZERO_STAGE in state_dicts[0][OPTIMIZER_STATE_DICT]:
159
+ raise ValueError(f"{files[0]} is not a zero checkpoint")
160
+ zero_stage = state_dicts[0][OPTIMIZER_STATE_DICT][ZERO_STAGE]
161
+ world_size = state_dicts[0][OPTIMIZER_STATE_DICT][PARTITION_COUNT]
162
+
163
+ # For ZeRO-2 each param group can have different partition_count as data parallelism for expert
164
+ # parameters can be different from data parallelism for non-expert parameters. So we can just
165
+ # use the max of the partition_count to get the dp world_size.
166
+
167
+ if type(world_size) is list:
168
+ world_size = max(world_size)
169
+
170
+ if world_size != total_files:
171
+ raise ValueError(
172
+ f"Expected {world_size} of '*_optim_states.pt' under '{ds_checkpoint_dir}' but found {total_files} files. "
173
+ "Possibly due to an overwrite of an old checkpoint, or a checkpoint didn't get saved by one or more processes."
174
+ )
175
+
176
+ # the groups are named differently in each stage
177
+ if zero_stage <= 2:
178
+ fp32_groups_key = SINGLE_PARTITION_OF_FP32_GROUPS
179
+ elif zero_stage == 3:
180
+ fp32_groups_key = FP32_FLAT_GROUPS
181
+ else:
182
+ raise ValueError(f"unknown zero stage {zero_stage}")
183
+
184
+ fp32_flat_groups = [state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key] for i in range(len(state_dicts))]
185
+ return zero_stage, world_size, fp32_flat_groups
186
+
187
+
188
+ def _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters):
189
+ """
190
+ Returns fp32 state_dict reconstructed from ds checkpoint
191
+
192
+ Args:
193
+ - ``ds_checkpoint_dir``: path to the deepspeed checkpoint folder (where the optimizer files are)
194
+
195
+ """
196
+ print(f"Processing zero checkpoint '{ds_checkpoint_dir}'")
197
+
198
+ optim_files = get_optim_files(ds_checkpoint_dir)
199
+ zero_stage, world_size, fp32_flat_groups = parse_optim_states(optim_files, ds_checkpoint_dir)
200
+ print(f"Detected checkpoint of type zero stage {zero_stage}, world_size: {world_size}")
201
+
202
+ model_files = get_model_state_files(ds_checkpoint_dir)
203
+
204
+ zero_model_states = parse_model_states(model_files)
205
+ print(f'Parsing checkpoint created by deepspeed=={zero_model_states[0].ds_version}')
206
+
207
+ if zero_stage <= 2:
208
+ return _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states,
209
+ exclude_frozen_parameters)
210
+ elif zero_stage == 3:
211
+ return _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states,
212
+ exclude_frozen_parameters)
213
+
214
+
215
+ def _zero2_merge_frozen_params(state_dict, zero_model_states):
216
+ if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0:
217
+ return
218
+
219
+ frozen_param_shapes = zero_model_states[0].frozen_param_shapes
220
+ frozen_param_fragments = zero_model_states[0].frozen_param_fragments
221
+
222
+ if debug:
223
+ num_elem = sum(s.numel() for s in frozen_param_shapes.values())
224
+ print(f'rank 0: {FROZEN_PARAM_SHAPES}.numel = {num_elem}')
225
+
226
+ wanted_params = len(frozen_param_shapes)
227
+ wanted_numel = sum(s.numel() for s in frozen_param_shapes.values())
228
+ avail_numel = sum([p.numel() for p in frozen_param_fragments.values()])
229
+ print(f'Frozen params: Have {avail_numel} numels to process.')
230
+ print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params')
231
+
232
+ total_params = 0
233
+ total_numel = 0
234
+ for name, shape in frozen_param_shapes.items():
235
+ total_params += 1
236
+ unpartitioned_numel = shape.numel()
237
+ total_numel += unpartitioned_numel
238
+
239
+ state_dict[name] = frozen_param_fragments[name]
240
+
241
+ if debug:
242
+ print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ")
243
+
244
+ print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements")
245
+
246
+
247
+ def _has_callable(obj, fn):
248
+ attr = getattr(obj, fn, None)
249
+ return callable(attr)
250
+
251
+
252
+ def _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states):
253
+ param_shapes = zero_model_states[0].param_shapes
254
+
255
+ # Reconstruction protocol:
256
+ #
257
+ # XXX: document this
258
+
259
+ if debug:
260
+ for i in range(world_size):
261
+ for j in range(len(fp32_flat_groups[0])):
262
+ print(f"{FP32_FLAT_GROUPS}[{i}][{j}].shape={fp32_flat_groups[i][j].shape}")
263
+
264
+ # XXX: memory usage doubles here (zero2)
265
+ num_param_groups = len(fp32_flat_groups[0])
266
+ merged_single_partition_of_fp32_groups = []
267
+ for i in range(num_param_groups):
268
+ merged_partitions = [sd[i] for sd in fp32_flat_groups]
269
+ full_single_fp32_vector = torch.cat(merged_partitions, 0)
270
+ merged_single_partition_of_fp32_groups.append(full_single_fp32_vector)
271
+ avail_numel = sum(
272
+ [full_single_fp32_vector.numel() for full_single_fp32_vector in merged_single_partition_of_fp32_groups])
273
+
274
+ if debug:
275
+ wanted_params = sum([len(shapes) for shapes in param_shapes])
276
+ wanted_numel = sum([sum(shape.numel() for shape in shapes.values()) for shapes in param_shapes])
277
+ # not asserting if there is a mismatch due to possible padding
278
+ print(f"Have {avail_numel} numels to process.")
279
+ print(f"Need {wanted_numel} numels in {wanted_params} params.")
280
+
281
+ # params
282
+ # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support
283
+ # out-of-core computing solution
284
+ total_numel = 0
285
+ total_params = 0
286
+ for shapes, full_single_fp32_vector in zip(param_shapes, merged_single_partition_of_fp32_groups):
287
+ offset = 0
288
+ avail_numel = full_single_fp32_vector.numel()
289
+ for name, shape in shapes.items():
290
+
291
+ unpartitioned_numel = shape.numel() if _has_callable(shape, 'numel') else math.prod(shape)
292
+ total_numel += unpartitioned_numel
293
+ total_params += 1
294
+
295
+ if debug:
296
+ print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ")
297
+ state_dict[name] = full_single_fp32_vector.narrow(0, offset, unpartitioned_numel).view(shape)
298
+ offset += unpartitioned_numel
299
+
300
+ # Z2 started to align to 2*world_size to improve nccl performance. Therefore both offset and
301
+ # avail_numel can differ by anywhere between 0..2*world_size. Due to two unrelated complex
302
+ # paddings performed in the code it's almost impossible to predict the exact numbers w/o the
303
+ # live optimizer object, so we are checking that the numbers are within the right range
304
+ align_to = 2 * world_size
305
+
306
+ def zero2_align(x):
307
+ return align_to * math.ceil(x / align_to)
308
+
309
+ if debug:
310
+ print(f"original offset={offset}, avail_numel={avail_numel}")
311
+
312
+ offset = zero2_align(offset)
313
+ avail_numel = zero2_align(avail_numel)
314
+
315
+ if debug:
316
+ print(f"aligned offset={offset}, avail_numel={avail_numel}")
317
+
318
+ # Sanity check
319
+ if offset != avail_numel:
320
+ raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong")
321
+
322
+ print(f"Reconstructed fp32 state dict with {total_params} params {total_numel} elements")
323
+
324
+
325
+ def _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states,
326
+ exclude_frozen_parameters):
327
+ state_dict = OrderedDict()
328
+
329
+ # buffers
330
+ buffers = zero_model_states[0].buffers
331
+ state_dict.update(buffers)
332
+ if debug:
333
+ print(f"added {len(buffers)} buffers")
334
+
335
+ if not exclude_frozen_parameters:
336
+ _zero2_merge_frozen_params(state_dict, zero_model_states)
337
+
338
+ _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states)
339
+
340
+ # recover shared parameters
341
+ for pair in zero_model_states[0].shared_params:
342
+ if pair[1] in state_dict:
343
+ state_dict[pair[0]] = state_dict[pair[1]]
344
+
345
+ return state_dict
346
+
347
+
348
+ def zero3_partitioned_param_info(unpartitioned_numel, world_size):
349
+ remainder = unpartitioned_numel % world_size
350
+ padding_numel = (world_size - remainder) if remainder else 0
351
+ partitioned_numel = math.ceil(unpartitioned_numel / world_size)
352
+ return partitioned_numel, padding_numel
353
+
354
+
355
+ def _zero3_merge_frozen_params(state_dict, world_size, zero_model_states):
356
+ if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0:
357
+ return
358
+
359
+ if debug:
360
+ for i in range(world_size):
361
+ num_elem = sum(s.numel() for s in zero_model_states[i].frozen_param_fragments.values())
362
+ print(f'rank {i}: {FROZEN_PARAM_SHAPES}.numel = {num_elem}')
363
+
364
+ frozen_param_shapes = zero_model_states[0].frozen_param_shapes
365
+ wanted_params = len(frozen_param_shapes)
366
+ wanted_numel = sum(s.numel() for s in frozen_param_shapes.values())
367
+ avail_numel = sum([p.numel() for p in zero_model_states[0].frozen_param_fragments.values()]) * world_size
368
+ print(f'Frozen params: Have {avail_numel} numels to process.')
369
+ print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params')
370
+
371
+ total_params = 0
372
+ total_numel = 0
373
+ for name, shape in zero_model_states[0].frozen_param_shapes.items():
374
+ total_params += 1
375
+ unpartitioned_numel = shape.numel()
376
+ total_numel += unpartitioned_numel
377
+
378
+ param_frags = tuple(model_state.frozen_param_fragments[name] for model_state in zero_model_states)
379
+ state_dict[name] = torch.cat(param_frags, 0).narrow(0, 0, unpartitioned_numel).view(shape)
380
+
381
+ partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size)
382
+
383
+ if debug:
384
+ print(
385
+ f"Frozen params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}"
386
+ )
387
+
388
+ print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements")
389
+
390
+
391
+ class GatheredTensor:
392
+ """
393
+ A pseudo tensor that collects partitioned weights.
394
+ It is more memory efficient when there are multiple groups.
395
+ """
396
+
397
+ def __init__(self, flat_groups, flat_groups_offset, offset, partitioned_numel, shape):
398
+ self.flat_groups = flat_groups
399
+ self.flat_groups_offset = flat_groups_offset
400
+ self.offset = offset
401
+ self.partitioned_numel = partitioned_numel
402
+ self.shape = shape
403
+ self.dtype = self.flat_groups[0][0].dtype
404
+
405
+ def contiguous(self):
406
+ """
407
+ Merge partitioned weights from flat_groups into a single tensor.
408
+ """
409
+ end_idx = self.offset + self.partitioned_numel
410
+ world_size = len(self.flat_groups)
411
+ pad_flat_param_chunks = []
412
+
413
+ for rank_i in range(world_size):
414
+ # for each rank, we need to collect weights from related group/groups
415
+ flat_groups_at_rank_i = self.flat_groups[rank_i]
416
+ start_group_id = None
417
+ end_group_id = None
418
+ for group_id in range(len(self.flat_groups_offset)):
419
+ if self.flat_groups_offset[group_id] <= self.offset < self.flat_groups_offset[group_id + 1]:
420
+ start_group_id = group_id
421
+ if self.flat_groups_offset[group_id] < end_idx <= self.flat_groups_offset[group_id + 1]:
422
+ end_group_id = group_id
423
+ break
424
+ # collect weights from related group/groups
425
+ for group_id in range(start_group_id, end_group_id + 1):
426
+ flat_tensor = flat_groups_at_rank_i[group_id]
427
+ start_offset = self.offset - self.flat_groups_offset[group_id]
428
+ end_offset = min(end_idx, self.flat_groups_offset[group_id + 1]) - self.flat_groups_offset[group_id]
429
+ pad_flat_param_chunks.append(flat_tensor[start_offset:end_offset])
430
+
431
+ # collect weights from all ranks
432
+ pad_flat_param = torch.cat(pad_flat_param_chunks, dim=0)
433
+ param = pad_flat_param[:self.shape.numel()].view(self.shape).contiguous()
434
+ return param
435
+
436
+
437
+ def _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states):
438
+ param_shapes = zero_model_states[0].param_shapes
439
+ avail_numel = sum([flat_group.numel() for flat_group in fp32_flat_groups[0]]) * world_size
440
+
441
+ # Reconstruction protocol: For zero3 we need to zip the partitions together at boundary of each
442
+ # param, re-consolidating each param, while dealing with padding if any
443
+
444
+ # merge list of dicts, preserving order
445
+ param_shapes = {k: v for d in param_shapes for k, v in d.items()}
446
+
447
+ if debug:
448
+ for i in range(world_size):
449
+ print(f"{FP32_FLAT_GROUPS}[{i}].shape={fp32_flat_groups[i].shape}")
450
+
451
+ wanted_params = len(param_shapes)
452
+ wanted_numel = sum(shape.numel() for shape in param_shapes.values())
453
+ # not asserting if there is a mismatch due to possible padding
454
+ avail_numel = fp32_flat_groups[0].numel() * world_size
455
+ print(f"Trainable params: Have {avail_numel} numels to process.")
456
+ print(f"Trainable params: Need {wanted_numel} numels in {wanted_params} params.")
457
+
458
+ # params
459
+ # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support
460
+ # out-of-core computing solution
461
+ offset = 0
462
+ total_numel = 0
463
+ total_params = 0
464
+ flat_groups_offset = [0] + list(np.cumsum([flat_tensor.numel() for flat_tensor in fp32_flat_groups[0]]))
465
+ for name, shape in tqdm(param_shapes.items(), desc='Gathering sharded weights'):
466
+ unpartitioned_numel = shape.numel()
467
+ total_numel += unpartitioned_numel
468
+ total_params += 1
469
+ partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size)
470
+
471
+ if debug:
472
+ print(
473
+ f"Trainable params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}"
474
+ )
475
+
476
+ # memory efficient tensor
477
+ tensor = GatheredTensor(fp32_flat_groups, flat_groups_offset, offset, partitioned_numel, shape)
478
+ state_dict[name] = tensor
479
+ offset += partitioned_numel
480
+
481
+ offset *= world_size
482
+
483
+ # Sanity check
484
+ if offset != avail_numel:
485
+ raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong")
486
+
487
+ print(f"Reconstructed Trainable fp32 state dict with {total_params} params {total_numel} elements")
488
+
489
+
490
+ def _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states,
491
+ exclude_frozen_parameters):
492
+ state_dict = OrderedDict()
493
+
494
+ # buffers
495
+ buffers = zero_model_states[0].buffers
496
+ state_dict.update(buffers)
497
+ if debug:
498
+ print(f"added {len(buffers)} buffers")
499
+
500
+ if not exclude_frozen_parameters:
501
+ _zero3_merge_frozen_params(state_dict, world_size, zero_model_states)
502
+
503
+ _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states)
504
+
505
+ # recover shared parameters
506
+ for pair in zero_model_states[0].shared_params:
507
+ if pair[1] in state_dict:
508
+ state_dict[pair[0]] = state_dict[pair[1]]
509
+
510
+ return state_dict
511
+
512
+
513
+ def to_torch_tensor(state_dict, return_empty_tensor=False):
514
+ """
515
+ Convert state_dict of GatheredTensor to torch tensor
516
+ """
517
+ torch_state_dict = {}
518
+ converted_tensors = {}
519
+ for name, tensor in state_dict.items():
520
+ tensor_id = id(tensor)
521
+ if tensor_id in converted_tensors: # shared tensors
522
+ shared_tensor = torch_state_dict[converted_tensors[tensor_id]]
523
+ torch_state_dict[name] = shared_tensor
524
+ else:
525
+ converted_tensors[tensor_id] = name
526
+ if return_empty_tensor:
527
+ torch_state_dict[name] = torch.empty(tensor.shape, dtype=tensor.dtype)
528
+ else:
529
+ torch_state_dict[name] = tensor.contiguous()
530
+ return torch_state_dict
531
+
532
+
533
+ def get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir,
534
+ tag=None,
535
+ exclude_frozen_parameters=False,
536
+ lazy_mode=False):
537
+ """
538
+ Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated state_dict that can be loaded with
539
+ ``load_state_dict()`` and used for training without DeepSpeed or shared with others, for example
540
+ via a model hub.
541
+
542
+ Args:
543
+ - ``checkpoint_dir``: path to the desired checkpoint folder
544
+ - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in 'latest' file. e.g., ``global_step14``
545
+ - ``exclude_frozen_parameters``: exclude frozen parameters
546
+ - ``lazy_mode``: get state_dict in lazy mode. It returns a dict of pesduo tensor instead of torch tensor, which is more memory efficient.
547
+ Convert the pesduo tensor to torch tensor by ``.contiguous()``
548
+
549
+ Returns:
550
+ - pytorch ``state_dict``
551
+
552
+ A typical usage might be ::
553
+
554
+ from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint
555
+ # do the training and checkpoint saving
556
+ state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir) # already on cpu
557
+ model = model.cpu() # move to cpu
558
+ model.load_state_dict(state_dict)
559
+ # submit to model hub or save the model to share with others
560
+
561
+ In this example the ``model`` will no longer be usable in the deepspeed context of the same
562
+ application. i.e. you will need to re-initialize the deepspeed engine, since
563
+ ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it.
564
+
565
+ If you want it all done for you, use ``load_state_dict_from_zero_checkpoint`` instead.
566
+
567
+ Note: the above usage may not work if your application doesn't have sufficient free CPU memory.
568
+ You may need to use the offline approach using the ``zero_to_fp32.py`` script that is saved with
569
+ the checkpoint. Or you can load state_dict in lazy mode ::
570
+
571
+ from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint
572
+ state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, lazy_mode=True) # not on cpu
573
+ for name, lazy_tensor in state_dict.item():
574
+ tensor = lazy_tensor.contiguous() # to cpu
575
+ print(name, tensor)
576
+ # del tensor to release memory if it no longer in use
577
+ """
578
+ if tag is None:
579
+ latest_path = os.path.join(checkpoint_dir, 'latest')
580
+ if os.path.isfile(latest_path):
581
+ with open(latest_path, 'r') as fd:
582
+ tag = fd.read().strip()
583
+ else:
584
+ raise ValueError(f"Unable to find 'latest' file at {latest_path}")
585
+
586
+ ds_checkpoint_dir = os.path.join(checkpoint_dir, tag)
587
+
588
+ if not os.path.isdir(ds_checkpoint_dir):
589
+ raise FileNotFoundError(f"Directory '{ds_checkpoint_dir}' doesn't exist")
590
+
591
+ state_dict = _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters)
592
+ if lazy_mode:
593
+ return state_dict
594
+ else:
595
+ return to_torch_tensor(state_dict)
596
+
597
+
598
+ def convert_zero_checkpoint_to_fp32_state_dict(checkpoint_dir,
599
+ output_dir,
600
+ max_shard_size="5GB",
601
+ safe_serialization=False,
602
+ tag=None,
603
+ exclude_frozen_parameters=False):
604
+ """
605
+ Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` file that can be
606
+ loaded with ``torch.load(file)`` + ``load_state_dict()`` and used for training without DeepSpeed.
607
+
608
+ Args:
609
+ - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``)
610
+ - ``output_dir``: directory to the pytorch fp32 state_dict output files
611
+ - ``max_shard_size``: the maximum size for a checkpoint before being sharded, default value is 5GB
612
+ - ``safe_serialization``: whether to save the model using `safetensors` or the traditional PyTorch way (that uses `pickle`).
613
+ - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14``
614
+ - ``exclude_frozen_parameters``: exclude frozen parameters
615
+ """
616
+
617
+ # Dependency pre-check
618
+ if safe_serialization:
619
+ try:
620
+ from safetensors.torch import save_file
621
+ except ImportError:
622
+ print('If you want to use `safe_serialization`, please `pip install safetensors`')
623
+ raise
624
+ if max_shard_size is not None:
625
+ try:
626
+ from huggingface_hub import split_torch_state_dict_into_shards
627
+ except ImportError:
628
+ print('If you want to use `max_shard_size`, please `pip install huggingface_hub`')
629
+ raise
630
+
631
+ # Convert zero checkpoint to state_dict
632
+ state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir,
633
+ tag,
634
+ exclude_frozen_parameters,
635
+ lazy_mode=True)
636
+
637
+ # Shard the model if it is too big.
638
+ weights_name = "model.safetensors" if safe_serialization else "pytorch_model.bin"
639
+ if max_shard_size is not None:
640
+ filename_pattern = weights_name.replace(".bin", "{suffix}.bin").replace(".safetensors", "{suffix}.safetensors")
641
+ # an memory-efficient approach for sharding
642
+ empty_state_dict = to_torch_tensor(state_dict, return_empty_tensor=True)
643
+ state_dict_split = split_torch_state_dict_into_shards(empty_state_dict,
644
+ filename_pattern=filename_pattern,
645
+ max_shard_size=max_shard_size)
646
+ else:
647
+ from collections import namedtuple
648
+ StateDictSplit = namedtuple("StateDictSplit", ["is_sharded", "filename_to_tensors"])
649
+ state_dict_split = StateDictSplit(is_sharded=False,
650
+ filename_to_tensors={weights_name: list(state_dict.keys())})
651
+
652
+ # Save the model by shard
653
+ os.makedirs(output_dir, exist_ok=True)
654
+ filename_to_tensors = state_dict_split.filename_to_tensors.items()
655
+ for shard_file, tensors in tqdm(filename_to_tensors, desc="Saving checkpoint shards"):
656
+ shard_state_dict = {tensor_name: state_dict[tensor_name] for tensor_name in tensors}
657
+ shard_state_dict = to_torch_tensor(shard_state_dict)
658
+ output_path = os.path.join(output_dir, shard_file)
659
+ if safe_serialization:
660
+ save_file(shard_state_dict, output_path, metadata={"format": "pt"})
661
+ else:
662
+ torch.save(shard_state_dict, output_path)
663
+ # release the memory of current shard
664
+ for tensor_name in list(shard_state_dict.keys()):
665
+ del state_dict[tensor_name]
666
+ del shard_state_dict[tensor_name]
667
+ del shard_state_dict
668
+ gc.collect()
669
+
670
+ # Save index if sharded
671
+ if state_dict_split.is_sharded:
672
+ index = {
673
+ "metadata": state_dict_split.metadata,
674
+ "weight_map": state_dict_split.tensor_to_filename,
675
+ }
676
+ save_index_file = "model.safetensors.index.json" if safe_serialization else "pytorch_model.bin.index.json"
677
+ save_index_file = os.path.join(output_dir, save_index_file)
678
+ with open(save_index_file, "w", encoding="utf-8") as f:
679
+ content = json.dumps(index, indent=2, sort_keys=True) + "\n"
680
+ f.write(content)
681
+
682
+
683
+ def load_state_dict_from_zero_checkpoint(model, checkpoint_dir, tag=None):
684
+ """
685
+ 1. Put the provided model to cpu
686
+ 2. Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict``
687
+ 3. Load it into the provided model
688
+
689
+ Args:
690
+ - ``model``: the model object to update
691
+ - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``)
692
+ - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14``
693
+
694
+ Returns:
695
+ - ``model`: modified model
696
+
697
+ Make sure you have plenty of CPU memory available before you call this function. If you don't
698
+ have enough use the ``zero_to_fp32.py`` utility to do the conversion. You will find it
699
+ conveniently placed for you in the checkpoint folder.
700
+
701
+ A typical usage might be ::
702
+
703
+ from deepspeed.utils.zero_to_fp32 import load_state_dict_from_zero_checkpoint
704
+ model = load_state_dict_from_zero_checkpoint(trainer.model, checkpoint_dir)
705
+ # submit to model hub or save the model to share with others
706
+
707
+ Note, that once this was run, the ``model`` will no longer be usable in the deepspeed context
708
+ of the same application. i.e. you will need to re-initialize the deepspeed engine, since
709
+ ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it.
710
+
711
+ """
712
+ logger.info(f"Extracting fp32 weights")
713
+ state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag)
714
+
715
+ logger.info(f"Overwriting model with fp32 weights")
716
+ model = model.cpu()
717
+ model.load_state_dict(state_dict, strict=False)
718
+
719
+ return model
720
+
721
+
722
+ if __name__ == "__main__":
723
+ parser = argparse.ArgumentParser()
724
+ parser.add_argument("checkpoint_dir",
725
+ type=str,
726
+ help="path to the desired checkpoint folder, e.g., path/checkpoint-12")
727
+ parser.add_argument("output_dir",
728
+ type=str,
729
+ help="directory to the pytorch fp32 state_dict output files"
730
+ "(e.g. path/checkpoint-12-output/)")
731
+ parser.add_argument(
732
+ "--max_shard_size",
733
+ type=str,
734
+ default="5GB",
735
+ help="The maximum size for a checkpoint before being sharded. Checkpoints shard will then be each of size"
736
+ "lower than this size. If expressed as a string, needs to be digits followed by a unit (like `5MB`"
737
+ "We default it to 5GB in order for models to be able to run easily on free-tier google colab instances"
738
+ "without CPU OOM issues.")
739
+ parser.add_argument(
740
+ "--safe_serialization",
741
+ default=False,
742
+ action='store_true',
743
+ help="Whether to save the model using `safetensors` or the traditional PyTorch way (that uses `pickle`).")
744
+ parser.add_argument("-t",
745
+ "--tag",
746
+ type=str,
747
+ default=None,
748
+ help="checkpoint tag used as a unique identifier for checkpoint. e.g., global_step1")
749
+ parser.add_argument("--exclude_frozen_parameters", action='store_true', help="exclude frozen parameters")
750
+ parser.add_argument("-d", "--debug", action='store_true', help="enable debug")
751
+ args = parser.parse_args()
752
+
753
+ debug = args.debug
754
+
755
+ convert_zero_checkpoint_to_fp32_state_dict(args.checkpoint_dir,
756
+ args.output_dir,
757
+ max_shard_size=args.max_shard_size,
758
+ safe_serialization=args.safe_serialization,
759
+ tag=args.tag,
760
+ exclude_frozen_parameters=args.exclude_frozen_parameters)