Check commited on
Commit
2a31d3b
Β·
1 Parent(s): 5bdf9dc

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629908936.4564571/events.out.tfevents.1629908936.7e498afd5545.7645.85 +3 -0
  11. model-bin/finetune/base/log/1629909414.6742997/events.out.tfevents.1629909414.7e498afd5545.7645.87 +3 -0
  12. model-bin/finetune/base/log/1629909897.4578655/events.out.tfevents.1629909897.7e498afd5545.7645.89 +3 -0
  13. model-bin/finetune/base/log/1629910378.9335363/events.out.tfevents.1629910378.7e498afd5545.7645.91 +3 -0
  14. model-bin/finetune/base/log/1629910855.6983657/events.out.tfevents.1629910855.7e498afd5545.7645.93 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629908936.7e498afd5545.7645.84 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629909414.7e498afd5545.7645.86 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629909897.7e498afd5545.7645.88 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629910378.7e498afd5545.7645.90 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629910855.7e498afd5545.7645.92 +3 -0
model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cee9520ecfcf573b7470abc40c8d1a678f4ce8139f7bc8f0a0df02438bcd739b
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24cad9f28dc11fe9279978fcf3a0bc408a45425a697994c5b4c4738f375d14e2
3
  size 722165393
model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cb3faac825cb095efe79913ef557cfc08dda78f8c31112d90e14c7511260d3ee
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6dbf8329e6b28204af9dc8b010f5c4abbdaea31770e2d6efdbf98e30b385f207
3
  size 377909911
model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f36484815dd77b257aac189d6d35455009b7c2f31308b82f2185abfeba225193
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3185208803e8bbdf570955ae54957beb26a219cefb74153cdf5bac8630a1e1e9
3
  size 14503
model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87e2d1745c681c91159da5acdbfd8bc474c3ecf40e467f65b9d5603d68c91173
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a54e5d0967fa297fe8e8cf400b549017a28943c4e450107074b7aa711703e84
3
  size 559
model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b0e58b6f31048670ccaa6bae7bbd667060549520d587a77664846428f983e350
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:441d6ca4b8122d743fb64fa9296f5c7a8f4ce79ace5f3fd8fc1317da9d4f1aab
3
  size 623
model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1822989349981638,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-79148",
4
- "epoch": 656.995983935743,
5
- "global_step": 81513,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -203073,11 +203073,800 @@
203073
  "eval_steps_per_second": 0.594,
203074
  "eval_wer": 0.1944121915820029,
203075
  "step": 81513
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
203076
  }
203077
  ],
203078
  "max_steps": 620000,
203079
  "num_train_epochs": 5000,
203080
- "total_flos": 2.2939365245267496e+20,
203081
  "trial_name": null,
203082
  "trial_params": null
203083
  }
 
1
  {
2
  "best_metric": 0.1822989349981638,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-79148",
4
+ "epoch": 661.995983935743,
5
+ "global_step": 82134,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
203073
  "eval_steps_per_second": 0.594,
203074
  "eval_wer": 0.1944121915820029,
203075
  "step": 81513
203076
+ },
203077
+ {
203078
+ "epoch": 657.02,
203079
+ "learning_rate": 8.699773828756059e-06,
203080
+ "loss": 0.3692,
203081
+ "step": 81515
203082
+ },
203083
+ {
203084
+ "epoch": 657.06,
203085
+ "learning_rate": 8.699693053311795e-06,
203086
+ "loss": 0.3053,
203087
+ "step": 81520
203088
+ },
203089
+ {
203090
+ "epoch": 657.1,
203091
+ "learning_rate": 8.699612277867529e-06,
203092
+ "loss": 0.3155,
203093
+ "step": 81525
203094
+ },
203095
+ {
203096
+ "epoch": 657.14,
203097
+ "learning_rate": 8.699531502423265e-06,
203098
+ "loss": 0.4045,
203099
+ "step": 81530
203100
+ },
203101
+ {
203102
+ "epoch": 657.18,
203103
+ "learning_rate": 8.699450726979e-06,
203104
+ "loss": 0.6462,
203105
+ "step": 81535
203106
+ },
203107
+ {
203108
+ "epoch": 657.22,
203109
+ "learning_rate": 8.699369951534734e-06,
203110
+ "loss": 0.9745,
203111
+ "step": 81540
203112
+ },
203113
+ {
203114
+ "epoch": 657.26,
203115
+ "learning_rate": 8.69928917609047e-06,
203116
+ "loss": 0.2742,
203117
+ "step": 81545
203118
+ },
203119
+ {
203120
+ "epoch": 657.3,
203121
+ "learning_rate": 8.699208400646204e-06,
203122
+ "loss": 0.2758,
203123
+ "step": 81550
203124
+ },
203125
+ {
203126
+ "epoch": 657.34,
203127
+ "learning_rate": 8.69912762520194e-06,
203128
+ "loss": 0.3884,
203129
+ "step": 81555
203130
+ },
203131
+ {
203132
+ "epoch": 657.38,
203133
+ "learning_rate": 8.699046849757674e-06,
203134
+ "loss": 0.6769,
203135
+ "step": 81560
203136
+ },
203137
+ {
203138
+ "epoch": 657.42,
203139
+ "learning_rate": 8.69896607431341e-06,
203140
+ "loss": 1.1057,
203141
+ "step": 81565
203142
+ },
203143
+ {
203144
+ "epoch": 657.46,
203145
+ "learning_rate": 8.698885298869144e-06,
203146
+ "loss": 0.3067,
203147
+ "step": 81570
203148
+ },
203149
+ {
203150
+ "epoch": 657.5,
203151
+ "learning_rate": 8.69880452342488e-06,
203152
+ "loss": 0.3199,
203153
+ "step": 81575
203154
+ },
203155
+ {
203156
+ "epoch": 657.54,
203157
+ "learning_rate": 8.698723747980614e-06,
203158
+ "loss": 0.3696,
203159
+ "step": 81580
203160
+ },
203161
+ {
203162
+ "epoch": 657.58,
203163
+ "learning_rate": 8.69864297253635e-06,
203164
+ "loss": 0.6531,
203165
+ "step": 81585
203166
+ },
203167
+ {
203168
+ "epoch": 657.62,
203169
+ "learning_rate": 8.698562197092086e-06,
203170
+ "loss": 1.0473,
203171
+ "step": 81590
203172
+ },
203173
+ {
203174
+ "epoch": 657.66,
203175
+ "learning_rate": 8.69848142164782e-06,
203176
+ "loss": 0.321,
203177
+ "step": 81595
203178
+ },
203179
+ {
203180
+ "epoch": 657.7,
203181
+ "learning_rate": 8.698400646203556e-06,
203182
+ "loss": 0.3643,
203183
+ "step": 81600
203184
+ },
203185
+ {
203186
+ "epoch": 657.74,
203187
+ "learning_rate": 8.69831987075929e-06,
203188
+ "loss": 0.3277,
203189
+ "step": 81605
203190
+ },
203191
+ {
203192
+ "epoch": 657.78,
203193
+ "learning_rate": 8.698239095315026e-06,
203194
+ "loss": 0.6669,
203195
+ "step": 81610
203196
+ },
203197
+ {
203198
+ "epoch": 657.82,
203199
+ "learning_rate": 8.69815831987076e-06,
203200
+ "loss": 0.9607,
203201
+ "step": 81615
203202
+ },
203203
+ {
203204
+ "epoch": 657.86,
203205
+ "learning_rate": 8.698077544426496e-06,
203206
+ "loss": 0.2718,
203207
+ "step": 81620
203208
+ },
203209
+ {
203210
+ "epoch": 657.9,
203211
+ "learning_rate": 8.69799676898223e-06,
203212
+ "loss": 0.2763,
203213
+ "step": 81625
203214
+ },
203215
+ {
203216
+ "epoch": 657.94,
203217
+ "learning_rate": 8.697915993537966e-06,
203218
+ "loss": 0.3533,
203219
+ "step": 81630
203220
+ },
203221
+ {
203222
+ "epoch": 657.98,
203223
+ "learning_rate": 8.6978352180937e-06,
203224
+ "loss": 0.6378,
203225
+ "step": 81635
203226
+ },
203227
+ {
203228
+ "epoch": 658.0,
203229
+ "eval_loss": 0.45627066493034363,
203230
+ "eval_runtime": 41.9319,
203231
+ "eval_samples_per_second": 20.056,
203232
+ "eval_steps_per_second": 0.644,
203233
+ "eval_wer": 0.19326638248681455,
203234
+ "step": 81637
203235
+ },
203236
+ {
203237
+ "epoch": 658.02,
203238
+ "learning_rate": 8.697754442649436e-06,
203239
+ "loss": 0.4026,
203240
+ "step": 81640
203241
+ },
203242
+ {
203243
+ "epoch": 658.06,
203244
+ "learning_rate": 8.69767366720517e-06,
203245
+ "loss": 0.2801,
203246
+ "step": 81645
203247
+ },
203248
+ {
203249
+ "epoch": 658.1,
203250
+ "learning_rate": 8.697592891760906e-06,
203251
+ "loss": 0.2926,
203252
+ "step": 81650
203253
+ },
203254
+ {
203255
+ "epoch": 658.14,
203256
+ "learning_rate": 8.697512116316642e-06,
203257
+ "loss": 0.3782,
203258
+ "step": 81655
203259
+ },
203260
+ {
203261
+ "epoch": 658.18,
203262
+ "learning_rate": 8.697431340872376e-06,
203263
+ "loss": 0.7188,
203264
+ "step": 81660
203265
+ },
203266
+ {
203267
+ "epoch": 658.22,
203268
+ "learning_rate": 8.697350565428112e-06,
203269
+ "loss": 0.9318,
203270
+ "step": 81665
203271
+ },
203272
+ {
203273
+ "epoch": 658.27,
203274
+ "learning_rate": 8.697269789983846e-06,
203275
+ "loss": 0.3419,
203276
+ "step": 81670
203277
+ },
203278
+ {
203279
+ "epoch": 658.31,
203280
+ "learning_rate": 8.697189014539582e-06,
203281
+ "loss": 0.2777,
203282
+ "step": 81675
203283
+ },
203284
+ {
203285
+ "epoch": 658.35,
203286
+ "learning_rate": 8.697108239095316e-06,
203287
+ "loss": 0.3644,
203288
+ "step": 81680
203289
+ },
203290
+ {
203291
+ "epoch": 658.39,
203292
+ "learning_rate": 8.697027463651052e-06,
203293
+ "loss": 0.8372,
203294
+ "step": 81685
203295
+ },
203296
+ {
203297
+ "epoch": 658.43,
203298
+ "learning_rate": 8.696946688206786e-06,
203299
+ "loss": 0.9454,
203300
+ "step": 81690
203301
+ },
203302
+ {
203303
+ "epoch": 658.47,
203304
+ "learning_rate": 8.696865912762522e-06,
203305
+ "loss": 0.4416,
203306
+ "step": 81695
203307
+ },
203308
+ {
203309
+ "epoch": 658.51,
203310
+ "learning_rate": 8.696785137318256e-06,
203311
+ "loss": 0.2996,
203312
+ "step": 81700
203313
+ },
203314
+ {
203315
+ "epoch": 658.55,
203316
+ "learning_rate": 8.696704361873992e-06,
203317
+ "loss": 0.3234,
203318
+ "step": 81705
203319
+ },
203320
+ {
203321
+ "epoch": 658.59,
203322
+ "learning_rate": 8.696623586429726e-06,
203323
+ "loss": 0.7701,
203324
+ "step": 81710
203325
+ },
203326
+ {
203327
+ "epoch": 658.63,
203328
+ "learning_rate": 8.696542810985462e-06,
203329
+ "loss": 0.908,
203330
+ "step": 81715
203331
+ },
203332
+ {
203333
+ "epoch": 658.67,
203334
+ "learning_rate": 8.696462035541197e-06,
203335
+ "loss": 0.3623,
203336
+ "step": 81720
203337
+ },
203338
+ {
203339
+ "epoch": 658.71,
203340
+ "learning_rate": 8.696381260096932e-06,
203341
+ "loss": 0.2868,
203342
+ "step": 81725
203343
+ },
203344
+ {
203345
+ "epoch": 658.75,
203346
+ "learning_rate": 8.696300484652667e-06,
203347
+ "loss": 0.3455,
203348
+ "step": 81730
203349
+ },
203350
+ {
203351
+ "epoch": 658.79,
203352
+ "learning_rate": 8.696219709208402e-06,
203353
+ "loss": 0.8271,
203354
+ "step": 81735
203355
+ },
203356
+ {
203357
+ "epoch": 658.83,
203358
+ "learning_rate": 8.696138933764137e-06,
203359
+ "loss": 0.932,
203360
+ "step": 81740
203361
+ },
203362
+ {
203363
+ "epoch": 658.87,
203364
+ "learning_rate": 8.696058158319872e-06,
203365
+ "loss": 0.3335,
203366
+ "step": 81745
203367
+ },
203368
+ {
203369
+ "epoch": 658.91,
203370
+ "learning_rate": 8.695977382875607e-06,
203371
+ "loss": 0.2877,
203372
+ "step": 81750
203373
+ },
203374
+ {
203375
+ "epoch": 658.95,
203376
+ "learning_rate": 8.695896607431341e-06,
203377
+ "loss": 0.4708,
203378
+ "step": 81755
203379
+ },
203380
+ {
203381
+ "epoch": 658.99,
203382
+ "learning_rate": 8.695815831987077e-06,
203383
+ "loss": 0.7955,
203384
+ "step": 81760
203385
+ },
203386
+ {
203387
+ "epoch": 659.0,
203388
+ "eval_loss": 0.3713783025741577,
203389
+ "eval_runtime": 41.2538,
203390
+ "eval_samples_per_second": 20.386,
203391
+ "eval_steps_per_second": 0.654,
203392
+ "eval_wer": 0.18491199883151976,
203393
+ "step": 81761
203394
+ },
203395
+ {
203396
+ "epoch": 659.03,
203397
+ "learning_rate": 8.695735056542811e-06,
203398
+ "loss": 0.3734,
203399
+ "step": 81765
203400
+ },
203401
+ {
203402
+ "epoch": 659.07,
203403
+ "learning_rate": 8.695654281098547e-06,
203404
+ "loss": 0.3216,
203405
+ "step": 81770
203406
+ },
203407
+ {
203408
+ "epoch": 659.11,
203409
+ "learning_rate": 8.695573505654281e-06,
203410
+ "loss": 0.3628,
203411
+ "step": 81775
203412
+ },
203413
+ {
203414
+ "epoch": 659.15,
203415
+ "learning_rate": 8.695492730210017e-06,
203416
+ "loss": 0.4213,
203417
+ "step": 81780
203418
+ },
203419
+ {
203420
+ "epoch": 659.19,
203421
+ "learning_rate": 8.695411954765753e-06,
203422
+ "loss": 0.9702,
203423
+ "step": 81785
203424
+ },
203425
+ {
203426
+ "epoch": 659.23,
203427
+ "learning_rate": 8.695331179321487e-06,
203428
+ "loss": 1.3116,
203429
+ "step": 81790
203430
+ },
203431
+ {
203432
+ "epoch": 659.27,
203433
+ "learning_rate": 8.695250403877223e-06,
203434
+ "loss": 0.2895,
203435
+ "step": 81795
203436
+ },
203437
+ {
203438
+ "epoch": 659.31,
203439
+ "learning_rate": 8.695169628432957e-06,
203440
+ "loss": 0.2964,
203441
+ "step": 81800
203442
+ },
203443
+ {
203444
+ "epoch": 659.35,
203445
+ "learning_rate": 8.695088852988693e-06,
203446
+ "loss": 0.3786,
203447
+ "step": 81805
203448
+ },
203449
+ {
203450
+ "epoch": 659.39,
203451
+ "learning_rate": 8.695008077544427e-06,
203452
+ "loss": 0.8122,
203453
+ "step": 81810
203454
+ },
203455
+ {
203456
+ "epoch": 659.43,
203457
+ "learning_rate": 8.694927302100163e-06,
203458
+ "loss": 0.6819,
203459
+ "step": 81815
203460
+ },
203461
+ {
203462
+ "epoch": 659.47,
203463
+ "learning_rate": 8.694846526655897e-06,
203464
+ "loss": 0.3317,
203465
+ "step": 81820
203466
+ },
203467
+ {
203468
+ "epoch": 659.51,
203469
+ "learning_rate": 8.694765751211633e-06,
203470
+ "loss": 0.3594,
203471
+ "step": 81825
203472
+ },
203473
+ {
203474
+ "epoch": 659.55,
203475
+ "learning_rate": 8.694684975767367e-06,
203476
+ "loss": 0.3832,
203477
+ "step": 81830
203478
+ },
203479
+ {
203480
+ "epoch": 659.59,
203481
+ "learning_rate": 8.694604200323103e-06,
203482
+ "loss": 1.1161,
203483
+ "step": 81835
203484
+ },
203485
+ {
203486
+ "epoch": 659.63,
203487
+ "learning_rate": 8.694523424878837e-06,
203488
+ "loss": 0.6244,
203489
+ "step": 81840
203490
+ },
203491
+ {
203492
+ "epoch": 659.67,
203493
+ "learning_rate": 8.694442649434573e-06,
203494
+ "loss": 0.2728,
203495
+ "step": 81845
203496
+ },
203497
+ {
203498
+ "epoch": 659.71,
203499
+ "learning_rate": 8.694361873990309e-06,
203500
+ "loss": 0.3422,
203501
+ "step": 81850
203502
+ },
203503
+ {
203504
+ "epoch": 659.76,
203505
+ "learning_rate": 8.694281098546043e-06,
203506
+ "loss": 0.4158,
203507
+ "step": 81855
203508
+ },
203509
+ {
203510
+ "epoch": 659.8,
203511
+ "learning_rate": 8.694200323101779e-06,
203512
+ "loss": 0.9222,
203513
+ "step": 81860
203514
+ },
203515
+ {
203516
+ "epoch": 659.84,
203517
+ "learning_rate": 8.694119547657513e-06,
203518
+ "loss": 0.6799,
203519
+ "step": 81865
203520
+ },
203521
+ {
203522
+ "epoch": 659.88,
203523
+ "learning_rate": 8.694038772213249e-06,
203524
+ "loss": 0.3446,
203525
+ "step": 81870
203526
+ },
203527
+ {
203528
+ "epoch": 659.92,
203529
+ "learning_rate": 8.693957996768983e-06,
203530
+ "loss": 0.319,
203531
+ "step": 81875
203532
+ },
203533
+ {
203534
+ "epoch": 659.96,
203535
+ "learning_rate": 8.693877221324719e-06,
203536
+ "loss": 0.3603,
203537
+ "step": 81880
203538
+ },
203539
+ {
203540
+ "epoch": 660.0,
203541
+ "learning_rate": 8.693796445880453e-06,
203542
+ "loss": 1.205,
203543
+ "step": 81885
203544
+ },
203545
+ {
203546
+ "epoch": 660.0,
203547
+ "eval_loss": 0.4129309058189392,
203548
+ "eval_runtime": 44.2238,
203549
+ "eval_samples_per_second": 19.017,
203550
+ "eval_steps_per_second": 0.611,
203551
+ "eval_wer": 0.1928848546958495,
203552
+ "step": 81885
203553
+ },
203554
+ {
203555
+ "epoch": 655.04,
203556
+ "learning_rate": 8.693715670436189e-06,
203557
+ "loss": 0.3022,
203558
+ "step": 81890
203559
+ },
203560
+ {
203561
+ "epoch": 655.08,
203562
+ "learning_rate": 8.693634894991923e-06,
203563
+ "loss": 0.2905,
203564
+ "step": 81895
203565
+ },
203566
+ {
203567
+ "epoch": 655.12,
203568
+ "learning_rate": 8.693554119547659e-06,
203569
+ "loss": 0.3468,
203570
+ "step": 81900
203571
+ },
203572
+ {
203573
+ "epoch": 655.16,
203574
+ "learning_rate": 8.693473344103393e-06,
203575
+ "loss": 0.4777,
203576
+ "step": 81905
203577
+ },
203578
+ {
203579
+ "epoch": 655.2,
203580
+ "learning_rate": 8.693392568659129e-06,
203581
+ "loss": 1.1993,
203582
+ "step": 81910
203583
+ },
203584
+ {
203585
+ "epoch": 655.24,
203586
+ "learning_rate": 8.693311793214864e-06,
203587
+ "loss": 0.287,
203588
+ "step": 81915
203589
+ },
203590
+ {
203591
+ "epoch": 655.28,
203592
+ "learning_rate": 8.693231017770599e-06,
203593
+ "loss": 0.3406,
203594
+ "step": 81920
203595
+ },
203596
+ {
203597
+ "epoch": 655.32,
203598
+ "learning_rate": 8.693150242326334e-06,
203599
+ "loss": 0.3755,
203600
+ "step": 81925
203601
+ },
203602
+ {
203603
+ "epoch": 655.36,
203604
+ "learning_rate": 8.693069466882069e-06,
203605
+ "loss": 0.4703,
203606
+ "step": 81930
203607
+ },
203608
+ {
203609
+ "epoch": 655.4,
203610
+ "learning_rate": 8.692988691437804e-06,
203611
+ "loss": 1.1881,
203612
+ "step": 81935
203613
+ },
203614
+ {
203615
+ "epoch": 655.44,
203616
+ "learning_rate": 8.692907915993539e-06,
203617
+ "loss": 0.3511,
203618
+ "step": 81940
203619
+ },
203620
+ {
203621
+ "epoch": 655.48,
203622
+ "learning_rate": 8.692827140549274e-06,
203623
+ "loss": 0.2856,
203624
+ "step": 81945
203625
+ },
203626
+ {
203627
+ "epoch": 655.52,
203628
+ "learning_rate": 8.692746365105009e-06,
203629
+ "loss": 0.2809,
203630
+ "step": 81950
203631
+ },
203632
+ {
203633
+ "epoch": 655.56,
203634
+ "learning_rate": 8.692665589660744e-06,
203635
+ "loss": 0.4301,
203636
+ "step": 81955
203637
+ },
203638
+ {
203639
+ "epoch": 655.6,
203640
+ "learning_rate": 8.692584814216479e-06,
203641
+ "loss": 1.2094,
203642
+ "step": 81960
203643
+ },
203644
+ {
203645
+ "epoch": 655.64,
203646
+ "learning_rate": 8.692504038772214e-06,
203647
+ "loss": 0.3171,
203648
+ "step": 81965
203649
+ },
203650
+ {
203651
+ "epoch": 655.68,
203652
+ "learning_rate": 8.692423263327948e-06,
203653
+ "loss": 0.3045,
203654
+ "step": 81970
203655
+ },
203656
+ {
203657
+ "epoch": 655.72,
203658
+ "learning_rate": 8.692342487883684e-06,
203659
+ "loss": 0.3573,
203660
+ "step": 81975
203661
+ },
203662
+ {
203663
+ "epoch": 655.76,
203664
+ "learning_rate": 8.692261712439418e-06,
203665
+ "loss": 0.4384,
203666
+ "step": 81980
203667
+ },
203668
+ {
203669
+ "epoch": 655.8,
203670
+ "learning_rate": 8.692180936995154e-06,
203671
+ "loss": 1.3024,
203672
+ "step": 81985
203673
+ },
203674
+ {
203675
+ "epoch": 655.84,
203676
+ "learning_rate": 8.69210016155089e-06,
203677
+ "loss": 0.3106,
203678
+ "step": 81990
203679
+ },
203680
+ {
203681
+ "epoch": 655.88,
203682
+ "learning_rate": 8.692019386106624e-06,
203683
+ "loss": 0.2804,
203684
+ "step": 81995
203685
+ },
203686
+ {
203687
+ "epoch": 655.92,
203688
+ "learning_rate": 8.69193861066236e-06,
203689
+ "loss": 0.4209,
203690
+ "step": 82000
203691
+ },
203692
+ {
203693
+ "epoch": 655.96,
203694
+ "learning_rate": 8.691857835218094e-06,
203695
+ "loss": 0.4513,
203696
+ "step": 82005
203697
+ },
203698
+ {
203699
+ "epoch": 656.0,
203700
+ "learning_rate": 8.69177705977383e-06,
203701
+ "loss": 1.2045,
203702
+ "step": 82010
203703
+ },
203704
+ {
203705
+ "epoch": 656.0,
203706
+ "eval_loss": 0.3971728980541229,
203707
+ "eval_runtime": 43.3663,
203708
+ "eval_samples_per_second": 19.393,
203709
+ "eval_steps_per_second": 0.623,
203710
+ "eval_wer": 0.19588447653429603,
203711
+ "step": 82010
203712
+ },
203713
+ {
203714
+ "epoch": 661.04,
203715
+ "learning_rate": 8.691696284329564e-06,
203716
+ "loss": 0.3776,
203717
+ "step": 82015
203718
+ },
203719
+ {
203720
+ "epoch": 661.08,
203721
+ "learning_rate": 8.6916155088853e-06,
203722
+ "loss": 0.3147,
203723
+ "step": 82020
203724
+ },
203725
+ {
203726
+ "epoch": 661.12,
203727
+ "learning_rate": 8.691534733441034e-06,
203728
+ "loss": 0.3062,
203729
+ "step": 82025
203730
+ },
203731
+ {
203732
+ "epoch": 661.16,
203733
+ "learning_rate": 8.69145395799677e-06,
203734
+ "loss": 0.4539,
203735
+ "step": 82030
203736
+ },
203737
+ {
203738
+ "epoch": 661.2,
203739
+ "learning_rate": 8.691373182552504e-06,
203740
+ "loss": 1.04,
203741
+ "step": 82035
203742
+ },
203743
+ {
203744
+ "epoch": 661.24,
203745
+ "learning_rate": 8.69129240710824e-06,
203746
+ "loss": 0.3421,
203747
+ "step": 82040
203748
+ },
203749
+ {
203750
+ "epoch": 661.28,
203751
+ "learning_rate": 8.691211631663974e-06,
203752
+ "loss": 0.3186,
203753
+ "step": 82045
203754
+ },
203755
+ {
203756
+ "epoch": 661.32,
203757
+ "learning_rate": 8.69113085621971e-06,
203758
+ "loss": 0.3361,
203759
+ "step": 82050
203760
+ },
203761
+ {
203762
+ "epoch": 661.36,
203763
+ "learning_rate": 8.691050080775446e-06,
203764
+ "loss": 0.4341,
203765
+ "step": 82055
203766
+ },
203767
+ {
203768
+ "epoch": 661.4,
203769
+ "learning_rate": 8.69096930533118e-06,
203770
+ "loss": 1.3379,
203771
+ "step": 82060
203772
+ },
203773
+ {
203774
+ "epoch": 661.44,
203775
+ "learning_rate": 8.690888529886916e-06,
203776
+ "loss": 0.3727,
203777
+ "step": 82065
203778
+ },
203779
+ {
203780
+ "epoch": 661.48,
203781
+ "learning_rate": 8.69080775444265e-06,
203782
+ "loss": 0.3282,
203783
+ "step": 82070
203784
+ },
203785
+ {
203786
+ "epoch": 661.52,
203787
+ "learning_rate": 8.690726978998386e-06,
203788
+ "loss": 0.3196,
203789
+ "step": 82075
203790
+ },
203791
+ {
203792
+ "epoch": 661.56,
203793
+ "learning_rate": 8.69064620355412e-06,
203794
+ "loss": 0.4741,
203795
+ "step": 82080
203796
+ },
203797
+ {
203798
+ "epoch": 661.6,
203799
+ "learning_rate": 8.690565428109856e-06,
203800
+ "loss": 1.3915,
203801
+ "step": 82085
203802
+ },
203803
+ {
203804
+ "epoch": 661.64,
203805
+ "learning_rate": 8.69048465266559e-06,
203806
+ "loss": 0.3393,
203807
+ "step": 82090
203808
+ },
203809
+ {
203810
+ "epoch": 661.68,
203811
+ "learning_rate": 8.690403877221326e-06,
203812
+ "loss": 0.28,
203813
+ "step": 82095
203814
+ },
203815
+ {
203816
+ "epoch": 661.72,
203817
+ "learning_rate": 8.69032310177706e-06,
203818
+ "loss": 0.3303,
203819
+ "step": 82100
203820
+ },
203821
+ {
203822
+ "epoch": 661.76,
203823
+ "learning_rate": 8.690242326332796e-06,
203824
+ "loss": 0.6743,
203825
+ "step": 82105
203826
+ },
203827
+ {
203828
+ "epoch": 661.8,
203829
+ "learning_rate": 8.69016155088853e-06,
203830
+ "loss": 1.2618,
203831
+ "step": 82110
203832
+ },
203833
+ {
203834
+ "epoch": 661.84,
203835
+ "learning_rate": 8.690080775444266e-06,
203836
+ "loss": 0.3449,
203837
+ "step": 82115
203838
+ },
203839
+ {
203840
+ "epoch": 661.88,
203841
+ "learning_rate": 8.690000000000002e-06,
203842
+ "loss": 0.2685,
203843
+ "step": 82120
203844
+ },
203845
+ {
203846
+ "epoch": 661.92,
203847
+ "learning_rate": 8.689919224555736e-06,
203848
+ "loss": 0.3878,
203849
+ "step": 82125
203850
+ },
203851
+ {
203852
+ "epoch": 661.96,
203853
+ "learning_rate": 8.689838449111471e-06,
203854
+ "loss": 0.4573,
203855
+ "step": 82130
203856
+ },
203857
+ {
203858
+ "epoch": 662.0,
203859
+ "eval_loss": 0.4005543291568756,
203860
+ "eval_runtime": 41.7634,
203861
+ "eval_samples_per_second": 20.137,
203862
+ "eval_steps_per_second": 0.646,
203863
+ "eval_wer": 0.1882754117390989,
203864
+ "step": 82134
203865
  }
203866
  ],
203867
  "max_steps": 620000,
203868
  "num_train_epochs": 5000,
203869
+ "total_flos": 2.311439623193885e+20,
203870
  "trial_name": null,
203871
  "trial_params": null
203872
  }
model-bin/finetune/base/{checkpoint-81513 β†’ checkpoint-82134}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629908936.4564571/events.out.tfevents.1629908936.7e498afd5545.7645.85 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88043dbfc014775fe34ca52feef56a845d5c72b7f834cd2eedd1f005e7c3513d
3
+ size 4194
model-bin/finetune/base/log/1629909414.6742997/events.out.tfevents.1629909414.7e498afd5545.7645.87 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:539ba4c32aa9ba29fc524630bc75b00c130f9b9dc379ec813c8e151c46813bb0
3
+ size 4194
model-bin/finetune/base/log/1629909897.4578655/events.out.tfevents.1629909897.7e498afd5545.7645.89 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46d079512bd7c680908d7de7849f2a7ba8c53b3461349a64dea7f4d8278b42a4
3
+ size 4194
model-bin/finetune/base/log/1629910378.9335363/events.out.tfevents.1629910378.7e498afd5545.7645.91 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7360042c324e7514187a2a1547a3649d50d5a8ca62f7a386040ae23d995edb64
3
+ size 4194
model-bin/finetune/base/log/1629910855.6983657/events.out.tfevents.1629910855.7e498afd5545.7645.93 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfb2728e096f7c3f4446de6971af6f710e4ddc59901ae230c6e1442fbc4fddad
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629908936.7e498afd5545.7645.84 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ef0885530910541691764e60631a45e999938a1cc3a3cc4a013d99dfb3186db
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629909414.7e498afd5545.7645.86 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30d2a2747b4fa0af5a584e24ce818500b405bbf8d7d908aee5384cf2621ac9d2
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629909897.7e498afd5545.7645.88 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e10c39d49cfbac496ce109b11fbc1ecc98c61886ba2f40ec35b9b1b12f137036
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629910378.7e498afd5545.7645.90 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af8657c949f1bec17cdd19d42d9b4f80924851715f071810fe3ea20f2fb35ed9
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629910855.7e498afd5545.7645.92 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4c2303e1d3c4b46ba59d1a386403959162cae895c54fd8f726e8ab1e8cb0bb8
3
+ size 8462