Check commited on
Commit
440a47a
Β·
1 Parent(s): 79865b8

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/trainer_state.json +798 -3
  9. model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629957682.4055254/events.out.tfevents.1629957682.8e89bd551565.924.81 +3 -0
  11. model-bin/finetune/base/log/1629958130.6461391/events.out.tfevents.1629958130.8e89bd551565.924.83 +3 -0
  12. model-bin/finetune/base/log/1629958565.4751377/events.out.tfevents.1629958565.8e89bd551565.924.85 +3 -0
  13. model-bin/finetune/base/log/1629959004.1539824/events.out.tfevents.1629959004.8e89bd551565.924.87 +3 -0
  14. model-bin/finetune/base/log/1629959449.6561632/events.out.tfevents.1629959452.8e89bd551565.924.89 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629957682.8e89bd551565.924.80 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629958130.8e89bd551565.924.82 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629958564.8e89bd551565.924.84 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629959004.8e89bd551565.924.86 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629959449.8e89bd551565.924.88 +3 -0
model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dd888f34151939c973bc1b406e70d1518151a52f0f47a7f3b3e4bf6e20a0814c
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d17ec4c2fabf26148ba9f60beaad001b3799a3176977e6859fa7e6018f2db434
3
  size 722165393
model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:db9d3ab6bd57b00e154b150f52824ff898ace57456f314ffd780af9ef1025982
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ec6551a97db0f833d76aed5735e0b34957ecb5d651053551090b61d2c0caa7a
3
  size 377909911
model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48c031af756610829efb4da17d1a6a0df0c7110ff4a69596b5d680c83878235d
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb79bc39717b4fbbb27b011a7feffc6343508dc1c8a09652858314b9dda02de0
3
  size 14503
model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:70fe0de8b811344c7522c0f25fa011a101a624c1ba8b39523e69af7715d8bd97
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa2098ffd50df5cf9c42261163807b82f181e741115f8e20fd104e5182b93533
3
  size 559
model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:35eb7d62fd453a5a24e6c8356ac861d30761c911bc452442e0b8186f6dd6e9f9
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c63fed4c8116d16f4e14d3f5b432cbccea3336c281a082e52182f2419bb6712
3
  size 623
model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.17637692697401752,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-92342",
4
- "epoch": 751.995983935743,
5
- "global_step": 93337,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -218118,11 +218118,806 @@
218118
  "eval_steps_per_second": 0.694,
218119
  "eval_wer": 0.1765940508069769,
218120
  "step": 93337
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
218121
  }
218122
  ],
218123
  "max_steps": 620000,
218124
  "num_train_epochs": 5000,
218125
- "total_flos": 2.6266383085344024e+20,
218126
  "trial_name": null,
218127
  "trial_params": null
218128
  }
 
1
  {
2
  "best_metric": 0.17637692697401752,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-92342",
4
+ "epoch": 756.995983935743,
5
+ "global_step": 93960,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
218118
  "eval_steps_per_second": 0.694,
218119
  "eval_wer": 0.1765940508069769,
218120
  "step": 93337
218121
+ },
218122
+ {
218123
+ "epoch": 746.02,
218124
+ "learning_rate": 8.520785256410257e-06,
218125
+ "loss": 0.2982,
218126
+ "step": 93340
218127
+ },
218128
+ {
218129
+ "epoch": 746.06,
218130
+ "learning_rate": 8.52070512820513e-06,
218131
+ "loss": 0.2931,
218132
+ "step": 93345
218133
+ },
218134
+ {
218135
+ "epoch": 746.1,
218136
+ "learning_rate": 8.520625e-06,
218137
+ "loss": 0.2853,
218138
+ "step": 93350
218139
+ },
218140
+ {
218141
+ "epoch": 746.14,
218142
+ "learning_rate": 8.520544871794872e-06,
218143
+ "loss": 0.3717,
218144
+ "step": 93355
218145
+ },
218146
+ {
218147
+ "epoch": 746.18,
218148
+ "learning_rate": 8.520464743589746e-06,
218149
+ "loss": 0.8425,
218150
+ "step": 93360
218151
+ },
218152
+ {
218153
+ "epoch": 746.22,
218154
+ "learning_rate": 8.520384615384615e-06,
218155
+ "loss": 0.8917,
218156
+ "step": 93365
218157
+ },
218158
+ {
218159
+ "epoch": 746.26,
218160
+ "learning_rate": 8.520304487179488e-06,
218161
+ "loss": 0.2906,
218162
+ "step": 93370
218163
+ },
218164
+ {
218165
+ "epoch": 746.3,
218166
+ "learning_rate": 8.52022435897436e-06,
218167
+ "loss": 0.3077,
218168
+ "step": 93375
218169
+ },
218170
+ {
218171
+ "epoch": 746.34,
218172
+ "learning_rate": 8.520144230769231e-06,
218173
+ "loss": 0.3911,
218174
+ "step": 93380
218175
+ },
218176
+ {
218177
+ "epoch": 746.38,
218178
+ "learning_rate": 8.520064102564102e-06,
218179
+ "loss": 0.7335,
218180
+ "step": 93385
218181
+ },
218182
+ {
218183
+ "epoch": 746.42,
218184
+ "learning_rate": 8.519983974358975e-06,
218185
+ "loss": 0.8044,
218186
+ "step": 93390
218187
+ },
218188
+ {
218189
+ "epoch": 746.46,
218190
+ "learning_rate": 8.519903846153847e-06,
218191
+ "loss": 0.3038,
218192
+ "step": 93395
218193
+ },
218194
+ {
218195
+ "epoch": 746.5,
218196
+ "learning_rate": 8.519823717948718e-06,
218197
+ "loss": 0.2585,
218198
+ "step": 93400
218199
+ },
218200
+ {
218201
+ "epoch": 746.54,
218202
+ "learning_rate": 8.519743589743591e-06,
218203
+ "loss": 0.4126,
218204
+ "step": 93405
218205
+ },
218206
+ {
218207
+ "epoch": 746.58,
218208
+ "learning_rate": 8.519663461538462e-06,
218209
+ "loss": 0.7608,
218210
+ "step": 93410
218211
+ },
218212
+ {
218213
+ "epoch": 746.62,
218214
+ "learning_rate": 8.519583333333334e-06,
218215
+ "loss": 1.017,
218216
+ "step": 93415
218217
+ },
218218
+ {
218219
+ "epoch": 746.66,
218220
+ "learning_rate": 8.519503205128205e-06,
218221
+ "loss": 0.3271,
218222
+ "step": 93420
218223
+ },
218224
+ {
218225
+ "epoch": 746.7,
218226
+ "learning_rate": 8.519423076923078e-06,
218227
+ "loss": 0.4955,
218228
+ "step": 93425
218229
+ },
218230
+ {
218231
+ "epoch": 746.74,
218232
+ "learning_rate": 8.51934294871795e-06,
218233
+ "loss": 0.4092,
218234
+ "step": 93430
218235
+ },
218236
+ {
218237
+ "epoch": 746.78,
218238
+ "learning_rate": 8.519262820512821e-06,
218239
+ "loss": 0.723,
218240
+ "step": 93435
218241
+ },
218242
+ {
218243
+ "epoch": 746.82,
218244
+ "learning_rate": 8.519182692307692e-06,
218245
+ "loss": 0.8715,
218246
+ "step": 93440
218247
+ },
218248
+ {
218249
+ "epoch": 746.86,
218250
+ "learning_rate": 8.519102564102565e-06,
218251
+ "loss": 0.2983,
218252
+ "step": 93445
218253
+ },
218254
+ {
218255
+ "epoch": 746.9,
218256
+ "learning_rate": 8.519022435897437e-06,
218257
+ "loss": 0.2913,
218258
+ "step": 93450
218259
+ },
218260
+ {
218261
+ "epoch": 746.94,
218262
+ "learning_rate": 8.518942307692308e-06,
218263
+ "loss": 0.384,
218264
+ "step": 93455
218265
+ },
218266
+ {
218267
+ "epoch": 746.98,
218268
+ "learning_rate": 8.518862179487181e-06,
218269
+ "loss": 0.7787,
218270
+ "step": 93460
218271
+ },
218272
+ {
218273
+ "epoch": 747.0,
218274
+ "eval_loss": 0.4655894935131073,
218275
+ "eval_runtime": 39.6595,
218276
+ "eval_samples_per_second": 21.13,
218277
+ "eval_steps_per_second": 0.681,
218278
+ "eval_wer": 0.19279159265837773,
218279
+ "step": 93462
218280
+ },
218281
+ {
218282
+ "epoch": 753.02,
218283
+ "learning_rate": 8.518782051282053e-06,
218284
+ "loss": 0.3159,
218285
+ "step": 93465
218286
+ },
218287
+ {
218288
+ "epoch": 753.06,
218289
+ "learning_rate": 8.518701923076924e-06,
218290
+ "loss": 0.2954,
218291
+ "step": 93470
218292
+ },
218293
+ {
218294
+ "epoch": 753.1,
218295
+ "learning_rate": 8.518621794871795e-06,
218296
+ "loss": 0.32,
218297
+ "step": 93475
218298
+ },
218299
+ {
218300
+ "epoch": 753.14,
218301
+ "learning_rate": 8.518541666666668e-06,
218302
+ "loss": 0.4195,
218303
+ "step": 93480
218304
+ },
218305
+ {
218306
+ "epoch": 753.18,
218307
+ "learning_rate": 8.518461538461538e-06,
218308
+ "loss": 0.709,
218309
+ "step": 93485
218310
+ },
218311
+ {
218312
+ "epoch": 753.22,
218313
+ "learning_rate": 8.518381410256411e-06,
218314
+ "loss": 0.8771,
218315
+ "step": 93490
218316
+ },
218317
+ {
218318
+ "epoch": 753.27,
218319
+ "learning_rate": 8.518301282051282e-06,
218320
+ "loss": 0.2633,
218321
+ "step": 93495
218322
+ },
218323
+ {
218324
+ "epoch": 753.31,
218325
+ "learning_rate": 8.518221153846154e-06,
218326
+ "loss": 0.2711,
218327
+ "step": 93500
218328
+ },
218329
+ {
218330
+ "epoch": 753.35,
218331
+ "learning_rate": 8.518141025641027e-06,
218332
+ "loss": 0.4551,
218333
+ "step": 93505
218334
+ },
218335
+ {
218336
+ "epoch": 753.39,
218337
+ "learning_rate": 8.518060897435898e-06,
218338
+ "loss": 0.7144,
218339
+ "step": 93510
218340
+ },
218341
+ {
218342
+ "epoch": 753.43,
218343
+ "learning_rate": 8.51798076923077e-06,
218344
+ "loss": 0.8337,
218345
+ "step": 93515
218346
+ },
218347
+ {
218348
+ "epoch": 753.47,
218349
+ "learning_rate": 8.517900641025641e-06,
218350
+ "loss": 0.2779,
218351
+ "step": 93520
218352
+ },
218353
+ {
218354
+ "epoch": 753.51,
218355
+ "learning_rate": 8.517820512820514e-06,
218356
+ "loss": 0.4137,
218357
+ "step": 93525
218358
+ },
218359
+ {
218360
+ "epoch": 753.55,
218361
+ "learning_rate": 8.517740384615385e-06,
218362
+ "loss": 0.3697,
218363
+ "step": 93530
218364
+ },
218365
+ {
218366
+ "epoch": 753.59,
218367
+ "learning_rate": 8.517660256410257e-06,
218368
+ "loss": 0.7943,
218369
+ "step": 93535
218370
+ },
218371
+ {
218372
+ "epoch": 753.63,
218373
+ "learning_rate": 8.517580128205128e-06,
218374
+ "loss": 0.8019,
218375
+ "step": 93540
218376
+ },
218377
+ {
218378
+ "epoch": 753.67,
218379
+ "learning_rate": 8.517500000000001e-06,
218380
+ "loss": 0.3453,
218381
+ "step": 93545
218382
+ },
218383
+ {
218384
+ "epoch": 753.71,
218385
+ "learning_rate": 8.517419871794872e-06,
218386
+ "loss": 0.3126,
218387
+ "step": 93550
218388
+ },
218389
+ {
218390
+ "epoch": 753.75,
218391
+ "learning_rate": 8.517339743589744e-06,
218392
+ "loss": 0.3679,
218393
+ "step": 93555
218394
+ },
218395
+ {
218396
+ "epoch": 753.79,
218397
+ "learning_rate": 8.517259615384617e-06,
218398
+ "loss": 0.7963,
218399
+ "step": 93560
218400
+ },
218401
+ {
218402
+ "epoch": 753.83,
218403
+ "learning_rate": 8.517179487179488e-06,
218404
+ "loss": 0.8713,
218405
+ "step": 93565
218406
+ },
218407
+ {
218408
+ "epoch": 753.87,
218409
+ "learning_rate": 8.51709935897436e-06,
218410
+ "loss": 0.2662,
218411
+ "step": 93570
218412
+ },
218413
+ {
218414
+ "epoch": 753.91,
218415
+ "learning_rate": 8.517019230769231e-06,
218416
+ "loss": 0.306,
218417
+ "step": 93575
218418
+ },
218419
+ {
218420
+ "epoch": 753.95,
218421
+ "learning_rate": 8.516939102564104e-06,
218422
+ "loss": 0.3898,
218423
+ "step": 93580
218424
+ },
218425
+ {
218426
+ "epoch": 753.99,
218427
+ "learning_rate": 8.516858974358975e-06,
218428
+ "loss": 0.7886,
218429
+ "step": 93585
218430
+ },
218431
+ {
218432
+ "epoch": 754.0,
218433
+ "eval_loss": 0.33248934149742126,
218434
+ "eval_runtime": 40.6199,
218435
+ "eval_samples_per_second": 20.655,
218436
+ "eval_steps_per_second": 0.665,
218437
+ "eval_wer": 0.19042131796903133,
218438
+ "step": 93586
218439
+ },
218440
+ {
218441
+ "epoch": 748.03,
218442
+ "learning_rate": 8.516778846153847e-06,
218443
+ "loss": 0.3288,
218444
+ "step": 93590
218445
+ },
218446
+ {
218447
+ "epoch": 748.07,
218448
+ "learning_rate": 8.516698717948718e-06,
218449
+ "loss": 0.2822,
218450
+ "step": 93595
218451
+ },
218452
+ {
218453
+ "epoch": 748.11,
218454
+ "learning_rate": 8.516618589743591e-06,
218455
+ "loss": 0.3252,
218456
+ "step": 93600
218457
+ },
218458
+ {
218459
+ "epoch": 748.15,
218460
+ "learning_rate": 8.516538461538462e-06,
218461
+ "loss": 0.3961,
218462
+ "step": 93605
218463
+ },
218464
+ {
218465
+ "epoch": 748.19,
218466
+ "learning_rate": 8.516458333333334e-06,
218467
+ "loss": 0.9088,
218468
+ "step": 93610
218469
+ },
218470
+ {
218471
+ "epoch": 748.23,
218472
+ "learning_rate": 8.516378205128207e-06,
218473
+ "loss": 0.7068,
218474
+ "step": 93615
218475
+ },
218476
+ {
218477
+ "epoch": 748.27,
218478
+ "learning_rate": 8.516298076923078e-06,
218479
+ "loss": 0.2668,
218480
+ "step": 93620
218481
+ },
218482
+ {
218483
+ "epoch": 748.31,
218484
+ "learning_rate": 8.51621794871795e-06,
218485
+ "loss": 0.298,
218486
+ "step": 93625
218487
+ },
218488
+ {
218489
+ "epoch": 748.35,
218490
+ "learning_rate": 8.516137820512821e-06,
218491
+ "loss": 0.415,
218492
+ "step": 93630
218493
+ },
218494
+ {
218495
+ "epoch": 748.39,
218496
+ "learning_rate": 8.516057692307694e-06,
218497
+ "loss": 0.9226,
218498
+ "step": 93635
218499
+ },
218500
+ {
218501
+ "epoch": 748.43,
218502
+ "learning_rate": 8.515977564102564e-06,
218503
+ "loss": 0.69,
218504
+ "step": 93640
218505
+ },
218506
+ {
218507
+ "epoch": 748.47,
218508
+ "learning_rate": 8.515897435897437e-06,
218509
+ "loss": 0.2476,
218510
+ "step": 93645
218511
+ },
218512
+ {
218513
+ "epoch": 748.51,
218514
+ "learning_rate": 8.515817307692308e-06,
218515
+ "loss": 0.325,
218516
+ "step": 93650
218517
+ },
218518
+ {
218519
+ "epoch": 748.55,
218520
+ "learning_rate": 8.51573717948718e-06,
218521
+ "loss": 0.4583,
218522
+ "step": 93655
218523
+ },
218524
+ {
218525
+ "epoch": 748.59,
218526
+ "learning_rate": 8.515657051282052e-06,
218527
+ "loss": 0.8614,
218528
+ "step": 93660
218529
+ },
218530
+ {
218531
+ "epoch": 748.63,
218532
+ "learning_rate": 8.515576923076924e-06,
218533
+ "loss": 0.6535,
218534
+ "step": 93665
218535
+ },
218536
+ {
218537
+ "epoch": 748.67,
218538
+ "learning_rate": 8.515496794871795e-06,
218539
+ "loss": 0.3172,
218540
+ "step": 93670
218541
+ },
218542
+ {
218543
+ "epoch": 748.71,
218544
+ "learning_rate": 8.515416666666667e-06,
218545
+ "loss": 0.314,
218546
+ "step": 93675
218547
+ },
218548
+ {
218549
+ "epoch": 748.75,
218550
+ "learning_rate": 8.51533653846154e-06,
218551
+ "loss": 0.4656,
218552
+ "step": 93680
218553
+ },
218554
+ {
218555
+ "epoch": 748.79,
218556
+ "learning_rate": 8.515256410256411e-06,
218557
+ "loss": 0.9987,
218558
+ "step": 93685
218559
+ },
218560
+ {
218561
+ "epoch": 748.83,
218562
+ "learning_rate": 8.515176282051282e-06,
218563
+ "loss": 0.6576,
218564
+ "step": 93690
218565
+ },
218566
+ {
218567
+ "epoch": 748.87,
218568
+ "learning_rate": 8.515096153846154e-06,
218569
+ "loss": 0.3404,
218570
+ "step": 93695
218571
+ },
218572
+ {
218573
+ "epoch": 748.91,
218574
+ "learning_rate": 8.515016025641027e-06,
218575
+ "loss": 0.2821,
218576
+ "step": 93700
218577
+ },
218578
+ {
218579
+ "epoch": 748.95,
218580
+ "learning_rate": 8.514935897435898e-06,
218581
+ "loss": 0.3785,
218582
+ "step": 93705
218583
+ },
218584
+ {
218585
+ "epoch": 748.99,
218586
+ "learning_rate": 8.51485576923077e-06,
218587
+ "loss": 0.966,
218588
+ "step": 93710
218589
+ },
218590
+ {
218591
+ "epoch": 749.0,
218592
+ "eval_loss": 0.4866238534450531,
218593
+ "eval_runtime": 38.5067,
218594
+ "eval_samples_per_second": 21.788,
218595
+ "eval_steps_per_second": 0.701,
218596
+ "eval_wer": 0.19120631341600902,
218597
+ "step": 93711
218598
+ },
218599
+ {
218600
+ "epoch": 749.03,
218601
+ "learning_rate": 8.514775641025642e-06,
218602
+ "loss": 0.369,
218603
+ "step": 93715
218604
+ },
218605
+ {
218606
+ "epoch": 749.07,
218607
+ "learning_rate": 8.514695512820514e-06,
218608
+ "loss": 0.3605,
218609
+ "step": 93720
218610
+ },
218611
+ {
218612
+ "epoch": 749.11,
218613
+ "learning_rate": 8.514615384615385e-06,
218614
+ "loss": 0.2692,
218615
+ "step": 93725
218616
+ },
218617
+ {
218618
+ "epoch": 749.15,
218619
+ "learning_rate": 8.514535256410257e-06,
218620
+ "loss": 0.4142,
218621
+ "step": 93730
218622
+ },
218623
+ {
218624
+ "epoch": 749.19,
218625
+ "learning_rate": 8.51445512820513e-06,
218626
+ "loss": 0.9357,
218627
+ "step": 93735
218628
+ },
218629
+ {
218630
+ "epoch": 749.23,
218631
+ "learning_rate": 8.514375000000001e-06,
218632
+ "loss": 0.6897,
218633
+ "step": 93740
218634
+ },
218635
+ {
218636
+ "epoch": 749.27,
218637
+ "learning_rate": 8.514294871794872e-06,
218638
+ "loss": 0.3877,
218639
+ "step": 93745
218640
+ },
218641
+ {
218642
+ "epoch": 749.31,
218643
+ "learning_rate": 8.514214743589744e-06,
218644
+ "loss": 0.3677,
218645
+ "step": 93750
218646
+ },
218647
+ {
218648
+ "epoch": 749.35,
218649
+ "learning_rate": 8.514134615384617e-06,
218650
+ "loss": 0.4088,
218651
+ "step": 93755
218652
+ },
218653
+ {
218654
+ "epoch": 749.39,
218655
+ "learning_rate": 8.514054487179488e-06,
218656
+ "loss": 0.8744,
218657
+ "step": 93760
218658
+ },
218659
+ {
218660
+ "epoch": 749.43,
218661
+ "learning_rate": 8.51397435897436e-06,
218662
+ "loss": 0.651,
218663
+ "step": 93765
218664
+ },
218665
+ {
218666
+ "epoch": 749.47,
218667
+ "learning_rate": 8.513894230769232e-06,
218668
+ "loss": 0.2776,
218669
+ "step": 93770
218670
+ },
218671
+ {
218672
+ "epoch": 749.51,
218673
+ "learning_rate": 8.513814102564104e-06,
218674
+ "loss": 0.4864,
218675
+ "step": 93775
218676
+ },
218677
+ {
218678
+ "epoch": 749.55,
218679
+ "learning_rate": 8.513733974358975e-06,
218680
+ "loss": 0.4179,
218681
+ "step": 93780
218682
+ },
218683
+ {
218684
+ "epoch": 749.59,
218685
+ "learning_rate": 8.513653846153847e-06,
218686
+ "loss": 0.9494,
218687
+ "step": 93785
218688
+ },
218689
+ {
218690
+ "epoch": 749.63,
218691
+ "learning_rate": 8.51357371794872e-06,
218692
+ "loss": 0.6507,
218693
+ "step": 93790
218694
+ },
218695
+ {
218696
+ "epoch": 749.67,
218697
+ "learning_rate": 8.51349358974359e-06,
218698
+ "loss": 0.3145,
218699
+ "step": 93795
218700
+ },
218701
+ {
218702
+ "epoch": 749.71,
218703
+ "learning_rate": 8.513413461538462e-06,
218704
+ "loss": 0.3237,
218705
+ "step": 93800
218706
+ },
218707
+ {
218708
+ "epoch": 749.75,
218709
+ "learning_rate": 8.513333333333335e-06,
218710
+ "loss": 0.397,
218711
+ "step": 93805
218712
+ },
218713
+ {
218714
+ "epoch": 749.79,
218715
+ "learning_rate": 8.513253205128205e-06,
218716
+ "loss": 0.9229,
218717
+ "step": 93810
218718
+ },
218719
+ {
218720
+ "epoch": 749.83,
218721
+ "learning_rate": 8.513173076923078e-06,
218722
+ "loss": 0.6785,
218723
+ "step": 93815
218724
+ },
218725
+ {
218726
+ "epoch": 749.87,
218727
+ "learning_rate": 8.51309294871795e-06,
218728
+ "loss": 0.3045,
218729
+ "step": 93820
218730
+ },
218731
+ {
218732
+ "epoch": 749.91,
218733
+ "learning_rate": 8.51301282051282e-06,
218734
+ "loss": 0.362,
218735
+ "step": 93825
218736
+ },
218737
+ {
218738
+ "epoch": 749.95,
218739
+ "learning_rate": 8.512932692307692e-06,
218740
+ "loss": 0.4119,
218741
+ "step": 93830
218742
+ },
218743
+ {
218744
+ "epoch": 749.99,
218745
+ "learning_rate": 8.512852564102565e-06,
218746
+ "loss": 1.0564,
218747
+ "step": 93835
218748
+ },
218749
+ {
218750
+ "epoch": 750.0,
218751
+ "eval_loss": 0.3657403886318207,
218752
+ "eval_runtime": 40.3739,
218753
+ "eval_samples_per_second": 20.756,
218754
+ "eval_steps_per_second": 0.669,
218755
+ "eval_wer": 0.18745001090671126,
218756
+ "step": 93836
218757
+ },
218758
+ {
218759
+ "epoch": 756.03,
218760
+ "learning_rate": 8.512772435897437e-06,
218761
+ "loss": 0.2971,
218762
+ "step": 93840
218763
+ },
218764
+ {
218765
+ "epoch": 756.07,
218766
+ "learning_rate": 8.512692307692308e-06,
218767
+ "loss": 0.3112,
218768
+ "step": 93845
218769
+ },
218770
+ {
218771
+ "epoch": 756.11,
218772
+ "learning_rate": 8.51261217948718e-06,
218773
+ "loss": 0.323,
218774
+ "step": 93850
218775
+ },
218776
+ {
218777
+ "epoch": 756.15,
218778
+ "learning_rate": 8.512532051282052e-06,
218779
+ "loss": 0.3942,
218780
+ "step": 93855
218781
+ },
218782
+ {
218783
+ "epoch": 756.19,
218784
+ "learning_rate": 8.512451923076924e-06,
218785
+ "loss": 0.9728,
218786
+ "step": 93860
218787
+ },
218788
+ {
218789
+ "epoch": 756.23,
218790
+ "learning_rate": 8.512371794871795e-06,
218791
+ "loss": 0.6976,
218792
+ "step": 93865
218793
+ },
218794
+ {
218795
+ "epoch": 756.27,
218796
+ "learning_rate": 8.512291666666668e-06,
218797
+ "loss": 0.3086,
218798
+ "step": 93870
218799
+ },
218800
+ {
218801
+ "epoch": 756.31,
218802
+ "learning_rate": 8.51221153846154e-06,
218803
+ "loss": 0.2799,
218804
+ "step": 93875
218805
+ },
218806
+ {
218807
+ "epoch": 756.35,
218808
+ "learning_rate": 8.512131410256411e-06,
218809
+ "loss": 0.3701,
218810
+ "step": 93880
218811
+ },
218812
+ {
218813
+ "epoch": 756.39,
218814
+ "learning_rate": 8.512051282051282e-06,
218815
+ "loss": 0.9532,
218816
+ "step": 93885
218817
+ },
218818
+ {
218819
+ "epoch": 756.43,
218820
+ "learning_rate": 8.511971153846155e-06,
218821
+ "loss": 0.7129,
218822
+ "step": 93890
218823
+ },
218824
+ {
218825
+ "epoch": 756.47,
218826
+ "learning_rate": 8.511891025641027e-06,
218827
+ "loss": 0.308,
218828
+ "step": 93895
218829
+ },
218830
+ {
218831
+ "epoch": 756.51,
218832
+ "learning_rate": 8.511810897435898e-06,
218833
+ "loss": 0.322,
218834
+ "step": 93900
218835
+ },
218836
+ {
218837
+ "epoch": 756.55,
218838
+ "learning_rate": 8.511730769230771e-06,
218839
+ "loss": 0.4047,
218840
+ "step": 93905
218841
+ },
218842
+ {
218843
+ "epoch": 756.59,
218844
+ "learning_rate": 8.511650641025642e-06,
218845
+ "loss": 0.8819,
218846
+ "step": 93910
218847
+ },
218848
+ {
218849
+ "epoch": 756.63,
218850
+ "learning_rate": 8.511570512820514e-06,
218851
+ "loss": 0.7582,
218852
+ "step": 93915
218853
+ },
218854
+ {
218855
+ "epoch": 756.67,
218856
+ "learning_rate": 8.511490384615385e-06,
218857
+ "loss": 0.286,
218858
+ "step": 93920
218859
+ },
218860
+ {
218861
+ "epoch": 756.71,
218862
+ "learning_rate": 8.511410256410258e-06,
218863
+ "loss": 0.326,
218864
+ "step": 93925
218865
+ },
218866
+ {
218867
+ "epoch": 756.76,
218868
+ "learning_rate": 8.511330128205128e-06,
218869
+ "loss": 0.4234,
218870
+ "step": 93930
218871
+ },
218872
+ {
218873
+ "epoch": 756.8,
218874
+ "learning_rate": 8.511250000000001e-06,
218875
+ "loss": 0.8456,
218876
+ "step": 93935
218877
+ },
218878
+ {
218879
+ "epoch": 756.84,
218880
+ "learning_rate": 8.511169871794872e-06,
218881
+ "loss": 0.6125,
218882
+ "step": 93940
218883
+ },
218884
+ {
218885
+ "epoch": 756.88,
218886
+ "learning_rate": 8.511089743589744e-06,
218887
+ "loss": 0.2815,
218888
+ "step": 93945
218889
+ },
218890
+ {
218891
+ "epoch": 756.92,
218892
+ "learning_rate": 8.511009615384615e-06,
218893
+ "loss": 0.3681,
218894
+ "step": 93950
218895
+ },
218896
+ {
218897
+ "epoch": 756.96,
218898
+ "learning_rate": 8.510929487179488e-06,
218899
+ "loss": 0.4144,
218900
+ "step": 93955
218901
+ },
218902
+ {
218903
+ "epoch": 757.0,
218904
+ "learning_rate": 8.51084935897436e-06,
218905
+ "loss": 1.381,
218906
+ "step": 93960
218907
+ },
218908
+ {
218909
+ "epoch": 757.0,
218910
+ "eval_loss": 0.3931798040866852,
218911
+ "eval_runtime": 39.9111,
218912
+ "eval_samples_per_second": 20.997,
218913
+ "eval_steps_per_second": 0.677,
218914
+ "eval_wer": 0.18971618401097712,
218915
+ "step": 93960
218916
  }
218917
  ],
218918
  "max_steps": 620000,
218919
  "num_train_epochs": 5000,
218920
+ "total_flos": 2.6441517960043604e+20,
218921
  "trial_name": null,
218922
  "trial_params": null
218923
  }
model-bin/finetune/base/{checkpoint-93337 β†’ checkpoint-93960}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629957682.4055254/events.out.tfevents.1629957682.8e89bd551565.924.81 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08d105a78974f4235280dc7827683e6e3a02694328e40455e1f73a5dc7409ca9
3
+ size 4194
model-bin/finetune/base/log/1629958130.6461391/events.out.tfevents.1629958130.8e89bd551565.924.83 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2444dcaf2d538fbf02c4eaaf5596949914d14932ae1c55a8297d6eafb1a16d3
3
+ size 4194
model-bin/finetune/base/log/1629958565.4751377/events.out.tfevents.1629958565.8e89bd551565.924.85 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a20a88ca95d6bc9a281c7be2358a950b88090ce81a72127f6fa436dc3ac5ece
3
+ size 4194
model-bin/finetune/base/log/1629959004.1539824/events.out.tfevents.1629959004.8e89bd551565.924.87 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1332b148c9309be6a77dc7699c5c92d81dfaa642851a89ce97d1b0f76df3ca8c
3
+ size 4194
model-bin/finetune/base/log/1629959449.6561632/events.out.tfevents.1629959452.8e89bd551565.924.89 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:230ecdd4f4f83e9fede175e00b6bd646cba700243f95f1fe38ee3279124b1444
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629957682.8e89bd551565.924.80 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62dc008537e2ba3f1407197e57f85e60c50b420a27db552981753562a26523c8
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629958130.8e89bd551565.924.82 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96c60b3ff922ed4778261d4c34a02d7b87b681251ab2e3f0e56228f0dd414727
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629958564.8e89bd551565.924.84 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5511f356e623b81a41a778588a50378720fdce16512852166215e84262114767
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629959004.8e89bd551565.924.86 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b46e42f14ae69899e0202eccc573e1d0e0f284b7cdc43ed9271ebe7616bb09c7
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629959449.8e89bd551565.924.88 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76ecb0389ff04935739f34be9a4d9d2b1ebd5043d47b450d6bf3730518d4f1bc
3
+ size 8622