Check commited on
Commit
6138e4a
Β·
1 Parent(s): 7d06cdc

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/trainer_state.json +793 -4
  9. model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629918685.2041333/events.out.tfevents.1629918685.7e498afd5545.7645.125 +3 -0
  11. model-bin/finetune/base/log/1629919158.470076/events.out.tfevents.1629919158.7e498afd5545.7645.127 +3 -0
  12. model-bin/finetune/base/log/1629919633.5511148/events.out.tfevents.1629919633.7e498afd5545.7645.129 +3 -0
  13. model-bin/finetune/base/log/1629920107.9077082/events.out.tfevents.1629920107.7e498afd5545.7645.131 +3 -0
  14. model-bin/finetune/base/log/1629920581.73902/events.out.tfevents.1629920581.7e498afd5545.7645.133 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629918685.7e498afd5545.7645.124 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629919158.7e498afd5545.7645.126 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629919633.7e498afd5545.7645.128 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629920107.7e498afd5545.7645.130 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629920581.7e498afd5545.7645.132 +3 -0
model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:84d3d4dcc8ff892ba962ed1dad272a1bd4c1a58ba3dd060bdf9a8cc9bad703ce
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35cbe7ab3362f3d9d416def618359265c26b9a47243d5ebfdbce021a9d514ce2
3
  size 722165393
model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:42dbfbaee58c4d5da382b4e7f0fcb067d75a3277d73d7f99a3cf4f8d1c4ed695
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb88936fbbed35267cc18d24cb4addb0e1c5f77ccc0bac83f3e42191852ec657
3
  size 377909911
model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:704c31ddfe3810c1417803352304434a53a78b40b0bf6380385ac47452b7641e
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef9efe43d872c0b7c4c3f9cf869d4e3159ef3fe70063cc12e48d87f2ee4f2a8b
3
  size 14503
model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:424b1cd729b3c4e1fd0997337376e31f6c134ee62c0160776a7e0bd89bd9c008
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42c7556e64554470eb5a108fd2e304f85439dd31cac51b64a3aeafaa85ae7bc8
3
  size 559
model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:93e9242ed52a0bcbcdb87cd3fb121d9c60638c18bdd4ddd660112a5416e3e0b1
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6efdd0e0c70faa9ea3b1397d573a4c203c65597ac8842e40857d44ef64c08d75
3
  size 623
model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1822989349981638,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-79148",
4
- "epoch": 672.0,
5
- "global_step": 84001,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -206241,11 +206241,800 @@
206241
  "eval_steps_per_second": 0.638,
206242
  "eval_wer": 0.19464897572355472,
206243
  "step": 84001
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
206244
  }
206245
  ],
206246
- "max_steps": 625000,
206247
  "num_train_epochs": 5000,
206248
- "total_flos": 2.363888951443245e+20,
206249
  "trial_name": null,
206250
  "trial_params": null
206251
  }
 
1
  {
2
  "best_metric": 0.1822989349981638,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-79148",
4
+ "epoch": 681.995983935743,
5
+ "global_step": 84623,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
206241
  "eval_steps_per_second": 0.638,
206242
  "eval_wer": 0.19464897572355472,
206243
  "step": 84001
206244
+ },
206245
+ {
206246
+ "epoch": 677.03,
206247
+ "learning_rate": 8.659579967689823e-06,
206248
+ "loss": 0.4234,
206249
+ "step": 84005
206250
+ },
206251
+ {
206252
+ "epoch": 677.07,
206253
+ "learning_rate": 8.659499192245559e-06,
206254
+ "loss": 0.3168,
206255
+ "step": 84010
206256
+ },
206257
+ {
206258
+ "epoch": 677.11,
206259
+ "learning_rate": 8.659418416801293e-06,
206260
+ "loss": 0.3753,
206261
+ "step": 84015
206262
+ },
206263
+ {
206264
+ "epoch": 677.15,
206265
+ "learning_rate": 8.659337641357029e-06,
206266
+ "loss": 0.4001,
206267
+ "step": 84020
206268
+ },
206269
+ {
206270
+ "epoch": 677.19,
206271
+ "learning_rate": 8.659256865912763e-06,
206272
+ "loss": 1.0413,
206273
+ "step": 84025
206274
+ },
206275
+ {
206276
+ "epoch": 677.23,
206277
+ "learning_rate": 8.659176090468498e-06,
206278
+ "loss": 0.7222,
206279
+ "step": 84030
206280
+ },
206281
+ {
206282
+ "epoch": 677.27,
206283
+ "learning_rate": 8.659095315024233e-06,
206284
+ "loss": 0.2808,
206285
+ "step": 84035
206286
+ },
206287
+ {
206288
+ "epoch": 677.31,
206289
+ "learning_rate": 8.659014539579968e-06,
206290
+ "loss": 0.3472,
206291
+ "step": 84040
206292
+ },
206293
+ {
206294
+ "epoch": 677.35,
206295
+ "learning_rate": 8.658933764135703e-06,
206296
+ "loss": 0.4189,
206297
+ "step": 84045
206298
+ },
206299
+ {
206300
+ "epoch": 677.39,
206301
+ "learning_rate": 8.658852988691438e-06,
206302
+ "loss": 0.8356,
206303
+ "step": 84050
206304
+ },
206305
+ {
206306
+ "epoch": 677.43,
206307
+ "learning_rate": 8.658772213247174e-06,
206308
+ "loss": 0.6287,
206309
+ "step": 84055
206310
+ },
206311
+ {
206312
+ "epoch": 677.47,
206313
+ "learning_rate": 8.658691437802908e-06,
206314
+ "loss": 0.2809,
206315
+ "step": 84060
206316
+ },
206317
+ {
206318
+ "epoch": 677.51,
206319
+ "learning_rate": 8.658610662358644e-06,
206320
+ "loss": 0.3183,
206321
+ "step": 84065
206322
+ },
206323
+ {
206324
+ "epoch": 677.55,
206325
+ "learning_rate": 8.658529886914378e-06,
206326
+ "loss": 0.3778,
206327
+ "step": 84070
206328
+ },
206329
+ {
206330
+ "epoch": 677.59,
206331
+ "learning_rate": 8.658449111470114e-06,
206332
+ "loss": 0.9175,
206333
+ "step": 84075
206334
+ },
206335
+ {
206336
+ "epoch": 677.63,
206337
+ "learning_rate": 8.658368336025848e-06,
206338
+ "loss": 0.6633,
206339
+ "step": 84080
206340
+ },
206341
+ {
206342
+ "epoch": 677.67,
206343
+ "learning_rate": 8.658287560581584e-06,
206344
+ "loss": 0.2622,
206345
+ "step": 84085
206346
+ },
206347
+ {
206348
+ "epoch": 677.71,
206349
+ "learning_rate": 8.658206785137318e-06,
206350
+ "loss": 0.3968,
206351
+ "step": 84090
206352
+ },
206353
+ {
206354
+ "epoch": 677.76,
206355
+ "learning_rate": 8.658126009693054e-06,
206356
+ "loss": 0.4656,
206357
+ "step": 84095
206358
+ },
206359
+ {
206360
+ "epoch": 677.8,
206361
+ "learning_rate": 8.658045234248788e-06,
206362
+ "loss": 0.9295,
206363
+ "step": 84100
206364
+ },
206365
+ {
206366
+ "epoch": 677.84,
206367
+ "learning_rate": 8.657964458804524e-06,
206368
+ "loss": 0.7699,
206369
+ "step": 84105
206370
+ },
206371
+ {
206372
+ "epoch": 677.88,
206373
+ "learning_rate": 8.657883683360258e-06,
206374
+ "loss": 0.2955,
206375
+ "step": 84110
206376
+ },
206377
+ {
206378
+ "epoch": 677.92,
206379
+ "learning_rate": 8.657802907915994e-06,
206380
+ "loss": 0.3351,
206381
+ "step": 84115
206382
+ },
206383
+ {
206384
+ "epoch": 677.96,
206385
+ "learning_rate": 8.65772213247173e-06,
206386
+ "loss": 0.5043,
206387
+ "step": 84120
206388
+ },
206389
+ {
206390
+ "epoch": 678.0,
206391
+ "learning_rate": 8.657641357027464e-06,
206392
+ "loss": 1.13,
206393
+ "step": 84125
206394
+ },
206395
+ {
206396
+ "epoch": 678.0,
206397
+ "eval_loss": 0.43177497386932373,
206398
+ "eval_runtime": 41.6294,
206399
+ "eval_samples_per_second": 20.202,
206400
+ "eval_steps_per_second": 0.649,
206401
+ "eval_wer": 0.19316737684084623,
206402
+ "step": 84125
206403
+ },
206404
+ {
206405
+ "epoch": 678.04,
206406
+ "learning_rate": 8.6575605815832e-06,
206407
+ "loss": 0.387,
206408
+ "step": 84130
206409
+ },
206410
+ {
206411
+ "epoch": 678.08,
206412
+ "learning_rate": 8.657479806138934e-06,
206413
+ "loss": 0.2739,
206414
+ "step": 84135
206415
+ },
206416
+ {
206417
+ "epoch": 678.12,
206418
+ "learning_rate": 8.65739903069467e-06,
206419
+ "loss": 0.3239,
206420
+ "step": 84140
206421
+ },
206422
+ {
206423
+ "epoch": 678.16,
206424
+ "learning_rate": 8.657318255250404e-06,
206425
+ "loss": 0.4463,
206426
+ "step": 84145
206427
+ },
206428
+ {
206429
+ "epoch": 678.2,
206430
+ "learning_rate": 8.65723747980614e-06,
206431
+ "loss": 1.1074,
206432
+ "step": 84150
206433
+ },
206434
+ {
206435
+ "epoch": 678.24,
206436
+ "learning_rate": 8.657156704361874e-06,
206437
+ "loss": 0.3395,
206438
+ "step": 84155
206439
+ },
206440
+ {
206441
+ "epoch": 678.28,
206442
+ "learning_rate": 8.65707592891761e-06,
206443
+ "loss": 0.372,
206444
+ "step": 84160
206445
+ },
206446
+ {
206447
+ "epoch": 678.32,
206448
+ "learning_rate": 8.656995153473344e-06,
206449
+ "loss": 0.3044,
206450
+ "step": 84165
206451
+ },
206452
+ {
206453
+ "epoch": 678.36,
206454
+ "learning_rate": 8.65691437802908e-06,
206455
+ "loss": 0.4819,
206456
+ "step": 84170
206457
+ },
206458
+ {
206459
+ "epoch": 678.4,
206460
+ "learning_rate": 8.656833602584816e-06,
206461
+ "loss": 1.3375,
206462
+ "step": 84175
206463
+ },
206464
+ {
206465
+ "epoch": 678.44,
206466
+ "learning_rate": 8.65675282714055e-06,
206467
+ "loss": 0.4084,
206468
+ "step": 84180
206469
+ },
206470
+ {
206471
+ "epoch": 678.48,
206472
+ "learning_rate": 8.656672051696286e-06,
206473
+ "loss": 0.3376,
206474
+ "step": 84185
206475
+ },
206476
+ {
206477
+ "epoch": 678.52,
206478
+ "learning_rate": 8.65659127625202e-06,
206479
+ "loss": 0.3932,
206480
+ "step": 84190
206481
+ },
206482
+ {
206483
+ "epoch": 678.56,
206484
+ "learning_rate": 8.656510500807756e-06,
206485
+ "loss": 0.4701,
206486
+ "step": 84195
206487
+ },
206488
+ {
206489
+ "epoch": 678.6,
206490
+ "learning_rate": 8.65642972536349e-06,
206491
+ "loss": 1.2044,
206492
+ "step": 84200
206493
+ },
206494
+ {
206495
+ "epoch": 678.64,
206496
+ "learning_rate": 8.656348949919226e-06,
206497
+ "loss": 0.9007,
206498
+ "step": 84205
206499
+ },
206500
+ {
206501
+ "epoch": 678.68,
206502
+ "learning_rate": 8.65626817447496e-06,
206503
+ "loss": 0.3381,
206504
+ "step": 84210
206505
+ },
206506
+ {
206507
+ "epoch": 678.72,
206508
+ "learning_rate": 8.656187399030696e-06,
206509
+ "loss": 0.3584,
206510
+ "step": 84215
206511
+ },
206512
+ {
206513
+ "epoch": 678.76,
206514
+ "learning_rate": 8.65610662358643e-06,
206515
+ "loss": 0.5029,
206516
+ "step": 84220
206517
+ },
206518
+ {
206519
+ "epoch": 678.8,
206520
+ "learning_rate": 8.656025848142166e-06,
206521
+ "loss": 1.2432,
206522
+ "step": 84225
206523
+ },
206524
+ {
206525
+ "epoch": 678.84,
206526
+ "learning_rate": 8.655945072697901e-06,
206527
+ "loss": 0.3662,
206528
+ "step": 84230
206529
+ },
206530
+ {
206531
+ "epoch": 678.88,
206532
+ "learning_rate": 8.655864297253636e-06,
206533
+ "loss": 0.2631,
206534
+ "step": 84235
206535
+ },
206536
+ {
206537
+ "epoch": 678.92,
206538
+ "learning_rate": 8.655783521809371e-06,
206539
+ "loss": 0.3069,
206540
+ "step": 84240
206541
+ },
206542
+ {
206543
+ "epoch": 678.96,
206544
+ "learning_rate": 8.655702746365105e-06,
206545
+ "loss": 0.4808,
206546
+ "step": 84245
206547
+ },
206548
+ {
206549
+ "epoch": 679.0,
206550
+ "eval_loss": 0.345355361700058,
206551
+ "eval_runtime": 41.7372,
206552
+ "eval_samples_per_second": 20.15,
206553
+ "eval_steps_per_second": 0.647,
206554
+ "eval_wer": 0.1902878488244342,
206555
+ "step": 84249
206556
+ },
206557
+ {
206558
+ "epoch": 673.01,
206559
+ "learning_rate": 8.655621970920841e-06,
206560
+ "loss": 0.4596,
206561
+ "step": 84250
206562
+ },
206563
+ {
206564
+ "epoch": 673.05,
206565
+ "learning_rate": 8.655541195476575e-06,
206566
+ "loss": 0.3423,
206567
+ "step": 84255
206568
+ },
206569
+ {
206570
+ "epoch": 673.09,
206571
+ "learning_rate": 8.655460420032311e-06,
206572
+ "loss": 0.3191,
206573
+ "step": 84260
206574
+ },
206575
+ {
206576
+ "epoch": 673.13,
206577
+ "learning_rate": 8.655379644588045e-06,
206578
+ "loss": 0.2949,
206579
+ "step": 84265
206580
+ },
206581
+ {
206582
+ "epoch": 673.17,
206583
+ "learning_rate": 8.655298869143781e-06,
206584
+ "loss": 0.6105,
206585
+ "step": 84270
206586
+ },
206587
+ {
206588
+ "epoch": 673.21,
206589
+ "learning_rate": 8.655218093699515e-06,
206590
+ "loss": 1.1691,
206591
+ "step": 84275
206592
+ },
206593
+ {
206594
+ "epoch": 673.25,
206595
+ "learning_rate": 8.655137318255251e-06,
206596
+ "loss": 0.3023,
206597
+ "step": 84280
206598
+ },
206599
+ {
206600
+ "epoch": 673.29,
206601
+ "learning_rate": 8.655056542810985e-06,
206602
+ "loss": 0.2756,
206603
+ "step": 84285
206604
+ },
206605
+ {
206606
+ "epoch": 673.33,
206607
+ "learning_rate": 8.654975767366721e-06,
206608
+ "loss": 0.3465,
206609
+ "step": 84290
206610
+ },
206611
+ {
206612
+ "epoch": 673.37,
206613
+ "learning_rate": 8.654894991922457e-06,
206614
+ "loss": 0.6593,
206615
+ "step": 84295
206616
+ },
206617
+ {
206618
+ "epoch": 673.41,
206619
+ "learning_rate": 8.654814216478191e-06,
206620
+ "loss": 1.3793,
206621
+ "step": 84300
206622
+ },
206623
+ {
206624
+ "epoch": 673.45,
206625
+ "learning_rate": 8.654733441033927e-06,
206626
+ "loss": 0.3088,
206627
+ "step": 84305
206628
+ },
206629
+ {
206630
+ "epoch": 673.49,
206631
+ "learning_rate": 8.654652665589661e-06,
206632
+ "loss": 0.3144,
206633
+ "step": 84310
206634
+ },
206635
+ {
206636
+ "epoch": 673.53,
206637
+ "learning_rate": 8.654571890145397e-06,
206638
+ "loss": 0.4673,
206639
+ "step": 84315
206640
+ },
206641
+ {
206642
+ "epoch": 673.57,
206643
+ "learning_rate": 8.654491114701131e-06,
206644
+ "loss": 0.5631,
206645
+ "step": 84320
206646
+ },
206647
+ {
206648
+ "epoch": 673.61,
206649
+ "learning_rate": 8.654410339256867e-06,
206650
+ "loss": 1.1314,
206651
+ "step": 84325
206652
+ },
206653
+ {
206654
+ "epoch": 673.65,
206655
+ "learning_rate": 8.654329563812601e-06,
206656
+ "loss": 0.3437,
206657
+ "step": 84330
206658
+ },
206659
+ {
206660
+ "epoch": 673.69,
206661
+ "learning_rate": 8.654248788368337e-06,
206662
+ "loss": 0.3081,
206663
+ "step": 84335
206664
+ },
206665
+ {
206666
+ "epoch": 673.73,
206667
+ "learning_rate": 8.654168012924071e-06,
206668
+ "loss": 0.3422,
206669
+ "step": 84340
206670
+ },
206671
+ {
206672
+ "epoch": 673.77,
206673
+ "learning_rate": 8.654087237479807e-06,
206674
+ "loss": 0.5191,
206675
+ "step": 84345
206676
+ },
206677
+ {
206678
+ "epoch": 673.81,
206679
+ "learning_rate": 8.654006462035543e-06,
206680
+ "loss": 1.1358,
206681
+ "step": 84350
206682
+ },
206683
+ {
206684
+ "epoch": 673.85,
206685
+ "learning_rate": 8.653925686591277e-06,
206686
+ "loss": 0.3092,
206687
+ "step": 84355
206688
+ },
206689
+ {
206690
+ "epoch": 673.89,
206691
+ "learning_rate": 8.653844911147013e-06,
206692
+ "loss": 0.2932,
206693
+ "step": 84360
206694
+ },
206695
+ {
206696
+ "epoch": 673.93,
206697
+ "learning_rate": 8.653764135702747e-06,
206698
+ "loss": 0.336,
206699
+ "step": 84365
206700
+ },
206701
+ {
206702
+ "epoch": 673.97,
206703
+ "learning_rate": 8.653683360258483e-06,
206704
+ "loss": 0.5495,
206705
+ "step": 84370
206706
+ },
206707
+ {
206708
+ "epoch": 674.0,
206709
+ "eval_loss": 0.3955157697200775,
206710
+ "eval_runtime": 43.0467,
206711
+ "eval_samples_per_second": 19.514,
206712
+ "eval_steps_per_second": 0.627,
206713
+ "eval_wer": 0.18563897991166833,
206714
+ "step": 84374
206715
+ },
206716
+ {
206717
+ "epoch": 674.01,
206718
+ "learning_rate": 8.653602584814217e-06,
206719
+ "loss": 0.3837,
206720
+ "step": 84375
206721
+ },
206722
+ {
206723
+ "epoch": 674.05,
206724
+ "learning_rate": 8.653521809369953e-06,
206725
+ "loss": 0.318,
206726
+ "step": 84380
206727
+ },
206728
+ {
206729
+ "epoch": 674.09,
206730
+ "learning_rate": 8.653441033925687e-06,
206731
+ "loss": 0.2944,
206732
+ "step": 84385
206733
+ },
206734
+ {
206735
+ "epoch": 674.13,
206736
+ "learning_rate": 8.653360258481423e-06,
206737
+ "loss": 0.3259,
206738
+ "step": 84390
206739
+ },
206740
+ {
206741
+ "epoch": 674.17,
206742
+ "learning_rate": 8.653279483037157e-06,
206743
+ "loss": 0.5183,
206744
+ "step": 84395
206745
+ },
206746
+ {
206747
+ "epoch": 674.21,
206748
+ "learning_rate": 8.653198707592893e-06,
206749
+ "loss": 1.1968,
206750
+ "step": 84400
206751
+ },
206752
+ {
206753
+ "epoch": 674.25,
206754
+ "learning_rate": 8.653117932148628e-06,
206755
+ "loss": 0.4829,
206756
+ "step": 84405
206757
+ },
206758
+ {
206759
+ "epoch": 674.29,
206760
+ "learning_rate": 8.653037156704363e-06,
206761
+ "loss": 0.3363,
206762
+ "step": 84410
206763
+ },
206764
+ {
206765
+ "epoch": 674.33,
206766
+ "learning_rate": 8.652956381260098e-06,
206767
+ "loss": 0.3947,
206768
+ "step": 84415
206769
+ },
206770
+ {
206771
+ "epoch": 674.37,
206772
+ "learning_rate": 8.652875605815833e-06,
206773
+ "loss": 0.6282,
206774
+ "step": 84420
206775
+ },
206776
+ {
206777
+ "epoch": 674.41,
206778
+ "learning_rate": 8.652794830371568e-06,
206779
+ "loss": 1.218,
206780
+ "step": 84425
206781
+ },
206782
+ {
206783
+ "epoch": 674.45,
206784
+ "learning_rate": 8.652714054927303e-06,
206785
+ "loss": 0.3229,
206786
+ "step": 84430
206787
+ },
206788
+ {
206789
+ "epoch": 674.49,
206790
+ "learning_rate": 8.652633279483038e-06,
206791
+ "loss": 0.3122,
206792
+ "step": 84435
206793
+ },
206794
+ {
206795
+ "epoch": 674.53,
206796
+ "learning_rate": 8.652552504038773e-06,
206797
+ "loss": 0.3114,
206798
+ "step": 84440
206799
+ },
206800
+ {
206801
+ "epoch": 674.57,
206802
+ "learning_rate": 8.652471728594508e-06,
206803
+ "loss": 0.525,
206804
+ "step": 84445
206805
+ },
206806
+ {
206807
+ "epoch": 674.61,
206808
+ "learning_rate": 8.652390953150243e-06,
206809
+ "loss": 1.1697,
206810
+ "step": 84450
206811
+ },
206812
+ {
206813
+ "epoch": 674.65,
206814
+ "learning_rate": 8.652310177705978e-06,
206815
+ "loss": 0.3382,
206816
+ "step": 84455
206817
+ },
206818
+ {
206819
+ "epoch": 674.69,
206820
+ "learning_rate": 8.652229402261713e-06,
206821
+ "loss": 0.2772,
206822
+ "step": 84460
206823
+ },
206824
+ {
206825
+ "epoch": 674.73,
206826
+ "learning_rate": 8.652148626817448e-06,
206827
+ "loss": 0.4419,
206828
+ "step": 84465
206829
+ },
206830
+ {
206831
+ "epoch": 674.77,
206832
+ "learning_rate": 8.652067851373184e-06,
206833
+ "loss": 0.4908,
206834
+ "step": 84470
206835
+ },
206836
+ {
206837
+ "epoch": 674.81,
206838
+ "learning_rate": 8.651987075928918e-06,
206839
+ "loss": 1.2181,
206840
+ "step": 84475
206841
+ },
206842
+ {
206843
+ "epoch": 674.85,
206844
+ "learning_rate": 8.651906300484654e-06,
206845
+ "loss": 0.3422,
206846
+ "step": 84480
206847
+ },
206848
+ {
206849
+ "epoch": 674.89,
206850
+ "learning_rate": 8.651825525040388e-06,
206851
+ "loss": 0.2761,
206852
+ "step": 84485
206853
+ },
206854
+ {
206855
+ "epoch": 674.93,
206856
+ "learning_rate": 8.651744749596124e-06,
206857
+ "loss": 0.3842,
206858
+ "step": 84490
206859
+ },
206860
+ {
206861
+ "epoch": 674.97,
206862
+ "learning_rate": 8.651663974151858e-06,
206863
+ "loss": 0.595,
206864
+ "step": 84495
206865
+ },
206866
+ {
206867
+ "epoch": 675.0,
206868
+ "eval_loss": 0.3872044086456299,
206869
+ "eval_runtime": 42.1888,
206870
+ "eval_samples_per_second": 19.91,
206871
+ "eval_steps_per_second": 0.64,
206872
+ "eval_wer": 0.1955033508683433,
206873
+ "step": 84499
206874
+ },
206875
+ {
206876
+ "epoch": 681.01,
206877
+ "learning_rate": 8.651583198707594e-06,
206878
+ "loss": 0.3997,
206879
+ "step": 84500
206880
+ },
206881
+ {
206882
+ "epoch": 681.05,
206883
+ "learning_rate": 8.651502423263328e-06,
206884
+ "loss": 0.3117,
206885
+ "step": 84505
206886
+ },
206887
+ {
206888
+ "epoch": 681.09,
206889
+ "learning_rate": 8.651421647819064e-06,
206890
+ "loss": 0.2902,
206891
+ "step": 84510
206892
+ },
206893
+ {
206894
+ "epoch": 681.13,
206895
+ "learning_rate": 8.651340872374798e-06,
206896
+ "loss": 0.3179,
206897
+ "step": 84515
206898
+ },
206899
+ {
206900
+ "epoch": 681.17,
206901
+ "learning_rate": 8.651260096930534e-06,
206902
+ "loss": 0.6692,
206903
+ "step": 84520
206904
+ },
206905
+ {
206906
+ "epoch": 681.21,
206907
+ "learning_rate": 8.65117932148627e-06,
206908
+ "loss": 1.0544,
206909
+ "step": 84525
206910
+ },
206911
+ {
206912
+ "epoch": 681.25,
206913
+ "learning_rate": 8.651098546042004e-06,
206914
+ "loss": 0.3278,
206915
+ "step": 84530
206916
+ },
206917
+ {
206918
+ "epoch": 681.29,
206919
+ "learning_rate": 8.65101777059774e-06,
206920
+ "loss": 0.3176,
206921
+ "step": 84535
206922
+ },
206923
+ {
206924
+ "epoch": 681.33,
206925
+ "learning_rate": 8.650936995153474e-06,
206926
+ "loss": 0.3031,
206927
+ "step": 84540
206928
+ },
206929
+ {
206930
+ "epoch": 681.37,
206931
+ "learning_rate": 8.65085621970921e-06,
206932
+ "loss": 0.5016,
206933
+ "step": 84545
206934
+ },
206935
+ {
206936
+ "epoch": 681.41,
206937
+ "learning_rate": 8.650775444264944e-06,
206938
+ "loss": 1.1741,
206939
+ "step": 84550
206940
+ },
206941
+ {
206942
+ "epoch": 681.45,
206943
+ "learning_rate": 8.65069466882068e-06,
206944
+ "loss": 0.3562,
206945
+ "step": 84555
206946
+ },
206947
+ {
206948
+ "epoch": 681.49,
206949
+ "learning_rate": 8.650613893376414e-06,
206950
+ "loss": 0.3223,
206951
+ "step": 84560
206952
+ },
206953
+ {
206954
+ "epoch": 681.53,
206955
+ "learning_rate": 8.65053311793215e-06,
206956
+ "loss": 0.4556,
206957
+ "step": 84565
206958
+ },
206959
+ {
206960
+ "epoch": 681.57,
206961
+ "learning_rate": 8.650452342487884e-06,
206962
+ "loss": 0.5309,
206963
+ "step": 84570
206964
+ },
206965
+ {
206966
+ "epoch": 681.61,
206967
+ "learning_rate": 8.65037156704362e-06,
206968
+ "loss": 1.1481,
206969
+ "step": 84575
206970
+ },
206971
+ {
206972
+ "epoch": 681.65,
206973
+ "learning_rate": 8.650290791599356e-06,
206974
+ "loss": 0.2845,
206975
+ "step": 84580
206976
+ },
206977
+ {
206978
+ "epoch": 681.69,
206979
+ "learning_rate": 8.65021001615509e-06,
206980
+ "loss": 0.2835,
206981
+ "step": 84585
206982
+ },
206983
+ {
206984
+ "epoch": 681.73,
206985
+ "learning_rate": 8.650129240710826e-06,
206986
+ "loss": 0.3701,
206987
+ "step": 84590
206988
+ },
206989
+ {
206990
+ "epoch": 681.77,
206991
+ "learning_rate": 8.65004846526656e-06,
206992
+ "loss": 0.6443,
206993
+ "step": 84595
206994
+ },
206995
+ {
206996
+ "epoch": 681.81,
206997
+ "learning_rate": 8.649967689822296e-06,
206998
+ "loss": 1.162,
206999
+ "step": 84600
207000
+ },
207001
+ {
207002
+ "epoch": 681.85,
207003
+ "learning_rate": 8.64988691437803e-06,
207004
+ "loss": 0.4954,
207005
+ "step": 84605
207006
+ },
207007
+ {
207008
+ "epoch": 681.89,
207009
+ "learning_rate": 8.649806138933766e-06,
207010
+ "loss": 0.3061,
207011
+ "step": 84610
207012
+ },
207013
+ {
207014
+ "epoch": 681.93,
207015
+ "learning_rate": 8.6497253634895e-06,
207016
+ "loss": 0.3989,
207017
+ "step": 84615
207018
+ },
207019
+ {
207020
+ "epoch": 681.97,
207021
+ "learning_rate": 8.649644588045235e-06,
207022
+ "loss": 0.6317,
207023
+ "step": 84620
207024
+ },
207025
+ {
207026
+ "epoch": 682.0,
207027
+ "eval_loss": 0.39930081367492676,
207028
+ "eval_runtime": 41.9885,
207029
+ "eval_samples_per_second": 20.029,
207030
+ "eval_steps_per_second": 0.643,
207031
+ "eval_wer": 0.1948430329958482,
207032
+ "step": 84623
207033
  }
207034
  ],
207035
+ "max_steps": 620000,
207036
  "num_train_epochs": 5000,
207037
+ "total_flos": 2.381427276537856e+20,
207038
  "trial_name": null,
207039
  "trial_params": null
207040
  }
model-bin/finetune/base/{checkpoint-84001 β†’ checkpoint-84623}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629918685.2041333/events.out.tfevents.1629918685.7e498afd5545.7645.125 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af115ebcaa932720ca5f61eac12c1708c971478f9031fd7c71f6bbdcd38d3114
3
+ size 4194
model-bin/finetune/base/log/1629919158.470076/events.out.tfevents.1629919158.7e498afd5545.7645.127 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:152bd11c3fb723992d5901fdfde6d8724f4bcc1a1bee816c0b0a880eca20636d
3
+ size 4194
model-bin/finetune/base/log/1629919633.5511148/events.out.tfevents.1629919633.7e498afd5545.7645.129 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:def08badfc86d444a8006546440dbddcd86e44438e527d09dd537f3455e44d1e
3
+ size 4194
model-bin/finetune/base/log/1629920107.9077082/events.out.tfevents.1629920107.7e498afd5545.7645.131 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97bbc2bebdf36e1a4f62a81001727c236d3d0196edcb6a6654510774d3b1c8b1
3
+ size 4194
model-bin/finetune/base/log/1629920581.73902/events.out.tfevents.1629920581.7e498afd5545.7645.133 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14b8209ac7cf25e41ee5f9e3c64d3a277beace45ad9b524942a7171ddc5fd0db
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629918685.7e498afd5545.7645.124 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:807ea28568a6a9358dd92912de8f1dc795c701fc2c28be7c9c9ead23f8edbbae
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629919158.7e498afd5545.7645.126 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:332cb27651394d717d50ee48bb0807af978196c81a9a44e788d41c0261d8f4ef
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1629919633.7e498afd5545.7645.128 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34af7bbdde9fae14e9fab7148897e7fed6c4bfce53241968c4a782cce479a2e0
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629920107.7e498afd5545.7645.130 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95abb7889cbebd87b7ee44cf4e87723817c32595a60340dfdea0f5d6477bf475
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629920581.7e498afd5545.7645.132 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bd9bc5b503e356d1ca2146c19f89352640811ad781f5377fb5908a72a9f979c
3
+ size 8622