Check commited on
Commit
ccccc13
Β·
1 Parent(s): 3b30d05

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/rng_state.pth +2 -2
  6. model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/trainer_state.json +798 -3
  9. model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630132562.2337933/events.out.tfevents.1630132562.86bb0ddabf9b.4092.51 +3 -0
  11. model-bin/finetune/base/log/1630132940.836788/events.out.tfevents.1630132940.86bb0ddabf9b.4092.53 +3 -0
  12. model-bin/finetune/base/log/1630133330.209003/events.out.tfevents.1630133330.86bb0ddabf9b.4092.55 +3 -0
  13. model-bin/finetune/base/log/1630133714.3736/events.out.tfevents.1630133714.86bb0ddabf9b.4092.57 +3 -0
  14. model-bin/finetune/base/log/1630134097.5583656/events.out.tfevents.1630134097.86bb0ddabf9b.4092.59 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630132562.86bb0ddabf9b.4092.50 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630132940.86bb0ddabf9b.4092.52 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630133330.86bb0ddabf9b.4092.54 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630133714.86bb0ddabf9b.4092.56 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630134097.86bb0ddabf9b.4092.58 +3 -0
model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:092b700951279c90023bb11c835e0f5f54e15c88589d7fae37929377088a3adf
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51fcbfe601df65783aad9ed90b143196bb9116032626928bb4e87c5585f1c83a
3
  size 722165393
model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c7b657e06bef8e9094828e09219b103fa0568ba25f4c6a79a87c063cf5d83325
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3904af36b18b4d095fb0124e026e85f54816e0865cc1484799d44b454c39030
3
  size 377909911
model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c1c37a1cae1c9c87ffcb6d013494288d4405456b0beb4882c39271505234ed5f
3
- size 14567
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1fa1ccd4d088fa375b9751da87ad6f087ce1ff3bd1852b863df97283f7f7a5c
3
+ size 14503
model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3ebeea6250e483ecd40494e28df8405c19ff072bd9538b8863c97fc25c728f7f
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e6d50a58f036c734ac486a1b8b8de8f2a72641d68799526e48e85c9dc1db370
3
  size 559
model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3eefa8eb85bed9a407060ff02d546b477b03e6766a73c07eae1e5b9ca1444db3
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ecb97a7d6b086ed775dfd796d92ea4da8cacd42ec572a22c69cec1b8c8c780d8
3
  size 623
model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
- "epoch": 967.9960159362549,
5
- "global_step": 121084,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -253419,11 +253419,806 @@
253419
  "eval_steps_per_second": 0.763,
253420
  "eval_wer": 0.18734491315136476,
253421
  "step": 121084
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
253422
  }
253423
  ],
253424
  "max_steps": 625000,
253425
  "num_train_epochs": 5000,
253426
- "total_flos": 3.407525538155879e+20,
253427
  "trial_name": null,
253428
  "trial_params": null
253429
  }
 
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
+ "epoch": 973.0,
5
+ "global_step": 121707,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
253419
  "eval_steps_per_second": 0.763,
253420
  "eval_wer": 0.18734491315136476,
253421
  "step": 121084
253422
+ },
253423
+ {
253424
+ "epoch": 976.01,
253425
+ "learning_rate": 8.076314102564103e-06,
253426
+ "loss": 0.4885,
253427
+ "step": 121085
253428
+ },
253429
+ {
253430
+ "epoch": 976.05,
253431
+ "learning_rate": 8.076233974358975e-06,
253432
+ "loss": 0.2731,
253433
+ "step": 121090
253434
+ },
253435
+ {
253436
+ "epoch": 976.09,
253437
+ "learning_rate": 8.076153846153846e-06,
253438
+ "loss": 0.2863,
253439
+ "step": 121095
253440
+ },
253441
+ {
253442
+ "epoch": 976.13,
253443
+ "learning_rate": 8.076073717948717e-06,
253444
+ "loss": 0.3848,
253445
+ "step": 121100
253446
+ },
253447
+ {
253448
+ "epoch": 976.17,
253449
+ "learning_rate": 8.07599358974359e-06,
253450
+ "loss": 0.5399,
253451
+ "step": 121105
253452
+ },
253453
+ {
253454
+ "epoch": 976.21,
253455
+ "learning_rate": 8.075913461538462e-06,
253456
+ "loss": 1.2772,
253457
+ "step": 121110
253458
+ },
253459
+ {
253460
+ "epoch": 976.25,
253461
+ "learning_rate": 8.075833333333333e-06,
253462
+ "loss": 0.3456,
253463
+ "step": 121115
253464
+ },
253465
+ {
253466
+ "epoch": 976.29,
253467
+ "learning_rate": 8.075753205128206e-06,
253468
+ "loss": 0.2335,
253469
+ "step": 121120
253470
+ },
253471
+ {
253472
+ "epoch": 976.33,
253473
+ "learning_rate": 8.075673076923078e-06,
253474
+ "loss": 0.311,
253475
+ "step": 121125
253476
+ },
253477
+ {
253478
+ "epoch": 976.37,
253479
+ "learning_rate": 8.075592948717949e-06,
253480
+ "loss": 0.5154,
253481
+ "step": 121130
253482
+ },
253483
+ {
253484
+ "epoch": 976.41,
253485
+ "learning_rate": 8.07551282051282e-06,
253486
+ "loss": 1.08,
253487
+ "step": 121135
253488
+ },
253489
+ {
253490
+ "epoch": 976.45,
253491
+ "learning_rate": 8.075432692307693e-06,
253492
+ "loss": 0.3043,
253493
+ "step": 121140
253494
+ },
253495
+ {
253496
+ "epoch": 976.49,
253497
+ "learning_rate": 8.075352564102565e-06,
253498
+ "loss": 0.2545,
253499
+ "step": 121145
253500
+ },
253501
+ {
253502
+ "epoch": 976.53,
253503
+ "learning_rate": 8.075272435897436e-06,
253504
+ "loss": 0.3678,
253505
+ "step": 121150
253506
+ },
253507
+ {
253508
+ "epoch": 976.57,
253509
+ "learning_rate": 8.07519230769231e-06,
253510
+ "loss": 0.4868,
253511
+ "step": 121155
253512
+ },
253513
+ {
253514
+ "epoch": 976.61,
253515
+ "learning_rate": 8.07511217948718e-06,
253516
+ "loss": 1.1736,
253517
+ "step": 121160
253518
+ },
253519
+ {
253520
+ "epoch": 976.65,
253521
+ "learning_rate": 8.075032051282052e-06,
253522
+ "loss": 0.2969,
253523
+ "step": 121165
253524
+ },
253525
+ {
253526
+ "epoch": 976.69,
253527
+ "learning_rate": 8.074951923076923e-06,
253528
+ "loss": 0.3378,
253529
+ "step": 121170
253530
+ },
253531
+ {
253532
+ "epoch": 976.73,
253533
+ "learning_rate": 8.074871794871796e-06,
253534
+ "loss": 0.31,
253535
+ "step": 121175
253536
+ },
253537
+ {
253538
+ "epoch": 976.77,
253539
+ "learning_rate": 8.074791666666668e-06,
253540
+ "loss": 0.521,
253541
+ "step": 121180
253542
+ },
253543
+ {
253544
+ "epoch": 976.81,
253545
+ "learning_rate": 8.074711538461539e-06,
253546
+ "loss": 1.0544,
253547
+ "step": 121185
253548
+ },
253549
+ {
253550
+ "epoch": 976.85,
253551
+ "learning_rate": 8.07463141025641e-06,
253552
+ "loss": 0.291,
253553
+ "step": 121190
253554
+ },
253555
+ {
253556
+ "epoch": 976.9,
253557
+ "learning_rate": 8.074551282051283e-06,
253558
+ "loss": 0.2925,
253559
+ "step": 121195
253560
+ },
253561
+ {
253562
+ "epoch": 976.94,
253563
+ "learning_rate": 8.074471153846155e-06,
253564
+ "loss": 0.3824,
253565
+ "step": 121200
253566
+ },
253567
+ {
253568
+ "epoch": 976.98,
253569
+ "learning_rate": 8.074391025641026e-06,
253570
+ "loss": 0.5301,
253571
+ "step": 121205
253572
+ },
253573
+ {
253574
+ "epoch": 977.0,
253575
+ "eval_loss": 0.4441799521446228,
253576
+ "eval_runtime": 35.4497,
253577
+ "eval_samples_per_second": 23.752,
253578
+ "eval_steps_per_second": 0.762,
253579
+ "eval_wer": 0.17582178361512227,
253580
+ "step": 121208
253581
+ },
253582
+ {
253583
+ "epoch": 969.02,
253584
+ "learning_rate": 8.0743108974359e-06,
253585
+ "loss": 0.3701,
253586
+ "step": 121210
253587
+ },
253588
+ {
253589
+ "epoch": 969.06,
253590
+ "learning_rate": 8.07423076923077e-06,
253591
+ "loss": 0.3003,
253592
+ "step": 121215
253593
+ },
253594
+ {
253595
+ "epoch": 969.1,
253596
+ "learning_rate": 8.074150641025642e-06,
253597
+ "loss": 0.2455,
253598
+ "step": 121220
253599
+ },
253600
+ {
253601
+ "epoch": 969.14,
253602
+ "learning_rate": 8.074070512820513e-06,
253603
+ "loss": 0.311,
253604
+ "step": 121225
253605
+ },
253606
+ {
253607
+ "epoch": 969.18,
253608
+ "learning_rate": 8.073990384615386e-06,
253609
+ "loss": 0.6479,
253610
+ "step": 121230
253611
+ },
253612
+ {
253613
+ "epoch": 969.22,
253614
+ "learning_rate": 8.073910256410256e-06,
253615
+ "loss": 1.1652,
253616
+ "step": 121235
253617
+ },
253618
+ {
253619
+ "epoch": 969.26,
253620
+ "learning_rate": 8.073830128205129e-06,
253621
+ "loss": 0.3719,
253622
+ "step": 121240
253623
+ },
253624
+ {
253625
+ "epoch": 969.3,
253626
+ "learning_rate": 8.07375e-06,
253627
+ "loss": 0.2902,
253628
+ "step": 121245
253629
+ },
253630
+ {
253631
+ "epoch": 969.34,
253632
+ "learning_rate": 8.073669871794872e-06,
253633
+ "loss": 0.3676,
253634
+ "step": 121250
253635
+ },
253636
+ {
253637
+ "epoch": 969.38,
253638
+ "learning_rate": 8.073589743589745e-06,
253639
+ "loss": 0.6999,
253640
+ "step": 121255
253641
+ },
253642
+ {
253643
+ "epoch": 969.42,
253644
+ "learning_rate": 8.073509615384616e-06,
253645
+ "loss": 1.0091,
253646
+ "step": 121260
253647
+ },
253648
+ {
253649
+ "epoch": 969.46,
253650
+ "learning_rate": 8.073429487179488e-06,
253651
+ "loss": 0.3483,
253652
+ "step": 121265
253653
+ },
253654
+ {
253655
+ "epoch": 969.5,
253656
+ "learning_rate": 8.073349358974359e-06,
253657
+ "loss": 0.3236,
253658
+ "step": 121270
253659
+ },
253660
+ {
253661
+ "epoch": 969.54,
253662
+ "learning_rate": 8.073269230769232e-06,
253663
+ "loss": 0.3426,
253664
+ "step": 121275
253665
+ },
253666
+ {
253667
+ "epoch": 969.58,
253668
+ "learning_rate": 8.073189102564103e-06,
253669
+ "loss": 0.5984,
253670
+ "step": 121280
253671
+ },
253672
+ {
253673
+ "epoch": 969.62,
253674
+ "learning_rate": 8.073108974358975e-06,
253675
+ "loss": 0.9732,
253676
+ "step": 121285
253677
+ },
253678
+ {
253679
+ "epoch": 969.66,
253680
+ "learning_rate": 8.073028846153846e-06,
253681
+ "loss": 0.2848,
253682
+ "step": 121290
253683
+ },
253684
+ {
253685
+ "epoch": 969.7,
253686
+ "learning_rate": 8.072948717948719e-06,
253687
+ "loss": 0.2934,
253688
+ "step": 121295
253689
+ },
253690
+ {
253691
+ "epoch": 969.74,
253692
+ "learning_rate": 8.072884615384617e-06,
253693
+ "loss": 1.169,
253694
+ "step": 121300
253695
+ },
253696
+ {
253697
+ "epoch": 969.78,
253698
+ "learning_rate": 8.072804487179488e-06,
253699
+ "loss": 0.548,
253700
+ "step": 121305
253701
+ },
253702
+ {
253703
+ "epoch": 969.82,
253704
+ "learning_rate": 8.07272435897436e-06,
253705
+ "loss": 1.056,
253706
+ "step": 121310
253707
+ },
253708
+ {
253709
+ "epoch": 969.86,
253710
+ "learning_rate": 8.07264423076923e-06,
253711
+ "loss": 0.249,
253712
+ "step": 121315
253713
+ },
253714
+ {
253715
+ "epoch": 969.9,
253716
+ "learning_rate": 8.072564102564104e-06,
253717
+ "loss": 0.3025,
253718
+ "step": 121320
253719
+ },
253720
+ {
253721
+ "epoch": 969.94,
253722
+ "learning_rate": 8.072483974358975e-06,
253723
+ "loss": 0.3114,
253724
+ "step": 121325
253725
+ },
253726
+ {
253727
+ "epoch": 969.98,
253728
+ "learning_rate": 8.072403846153846e-06,
253729
+ "loss": 0.5554,
253730
+ "step": 121330
253731
+ },
253732
+ {
253733
+ "epoch": 970.0,
253734
+ "eval_loss": 0.3604688346385956,
253735
+ "eval_runtime": 35.6725,
253736
+ "eval_samples_per_second": 23.604,
253737
+ "eval_steps_per_second": 0.757,
253738
+ "eval_wer": 0.1775592828224407,
253739
+ "step": 121333
253740
+ },
253741
+ {
253742
+ "epoch": 978.02,
253743
+ "learning_rate": 8.07232371794872e-06,
253744
+ "loss": 0.363,
253745
+ "step": 121335
253746
+ },
253747
+ {
253748
+ "epoch": 978.06,
253749
+ "learning_rate": 8.07224358974359e-06,
253750
+ "loss": 0.3365,
253751
+ "step": 121340
253752
+ },
253753
+ {
253754
+ "epoch": 978.1,
253755
+ "learning_rate": 8.072163461538462e-06,
253756
+ "loss": 0.2485,
253757
+ "step": 121345
253758
+ },
253759
+ {
253760
+ "epoch": 978.14,
253761
+ "learning_rate": 8.072083333333333e-06,
253762
+ "loss": 0.4608,
253763
+ "step": 121350
253764
+ },
253765
+ {
253766
+ "epoch": 978.18,
253767
+ "learning_rate": 8.072003205128207e-06,
253768
+ "loss": 0.6248,
253769
+ "step": 121355
253770
+ },
253771
+ {
253772
+ "epoch": 978.22,
253773
+ "learning_rate": 8.071923076923078e-06,
253774
+ "loss": 1.1636,
253775
+ "step": 121360
253776
+ },
253777
+ {
253778
+ "epoch": 978.26,
253779
+ "learning_rate": 8.07184294871795e-06,
253780
+ "loss": 0.273,
253781
+ "step": 121365
253782
+ },
253783
+ {
253784
+ "epoch": 978.3,
253785
+ "learning_rate": 8.07176282051282e-06,
253786
+ "loss": 0.3089,
253787
+ "step": 121370
253788
+ },
253789
+ {
253790
+ "epoch": 978.34,
253791
+ "learning_rate": 8.071682692307694e-06,
253792
+ "loss": 0.323,
253793
+ "step": 121375
253794
+ },
253795
+ {
253796
+ "epoch": 978.38,
253797
+ "learning_rate": 8.071602564102565e-06,
253798
+ "loss": 0.599,
253799
+ "step": 121380
253800
+ },
253801
+ {
253802
+ "epoch": 978.42,
253803
+ "learning_rate": 8.071522435897436e-06,
253804
+ "loss": 1.039,
253805
+ "step": 121385
253806
+ },
253807
+ {
253808
+ "epoch": 978.46,
253809
+ "learning_rate": 8.07144230769231e-06,
253810
+ "loss": 0.2914,
253811
+ "step": 121390
253812
+ },
253813
+ {
253814
+ "epoch": 978.5,
253815
+ "learning_rate": 8.071362179487179e-06,
253816
+ "loss": 0.246,
253817
+ "step": 121395
253818
+ },
253819
+ {
253820
+ "epoch": 978.54,
253821
+ "learning_rate": 8.071282051282052e-06,
253822
+ "loss": 0.3878,
253823
+ "step": 121400
253824
+ },
253825
+ {
253826
+ "epoch": 978.58,
253827
+ "learning_rate": 8.071201923076924e-06,
253828
+ "loss": 0.6029,
253829
+ "step": 121405
253830
+ },
253831
+ {
253832
+ "epoch": 978.62,
253833
+ "learning_rate": 8.071121794871795e-06,
253834
+ "loss": 0.9838,
253835
+ "step": 121410
253836
+ },
253837
+ {
253838
+ "epoch": 978.66,
253839
+ "learning_rate": 8.071041666666666e-06,
253840
+ "loss": 0.301,
253841
+ "step": 121415
253842
+ },
253843
+ {
253844
+ "epoch": 978.7,
253845
+ "learning_rate": 8.07096153846154e-06,
253846
+ "loss": 0.254,
253847
+ "step": 121420
253848
+ },
253849
+ {
253850
+ "epoch": 978.74,
253851
+ "learning_rate": 8.07088141025641e-06,
253852
+ "loss": 0.3564,
253853
+ "step": 121425
253854
+ },
253855
+ {
253856
+ "epoch": 978.78,
253857
+ "learning_rate": 8.070801282051282e-06,
253858
+ "loss": 0.5654,
253859
+ "step": 121430
253860
+ },
253861
+ {
253862
+ "epoch": 978.82,
253863
+ "learning_rate": 8.070721153846155e-06,
253864
+ "loss": 0.9491,
253865
+ "step": 121435
253866
+ },
253867
+ {
253868
+ "epoch": 978.86,
253869
+ "learning_rate": 8.070641025641026e-06,
253870
+ "loss": 0.2507,
253871
+ "step": 121440
253872
+ },
253873
+ {
253874
+ "epoch": 978.9,
253875
+ "learning_rate": 8.070560897435898e-06,
253876
+ "loss": 0.5248,
253877
+ "step": 121445
253878
+ },
253879
+ {
253880
+ "epoch": 978.94,
253881
+ "learning_rate": 8.070480769230769e-06,
253882
+ "loss": 0.2928,
253883
+ "step": 121450
253884
+ },
253885
+ {
253886
+ "epoch": 978.98,
253887
+ "learning_rate": 8.070400641025642e-06,
253888
+ "loss": 0.6818,
253889
+ "step": 121455
253890
+ },
253891
+ {
253892
+ "epoch": 979.0,
253893
+ "eval_loss": 0.43977269530296326,
253894
+ "eval_runtime": 35.1739,
253895
+ "eval_samples_per_second": 23.938,
253896
+ "eval_steps_per_second": 0.768,
253897
+ "eval_wer": 0.18553898155632303,
253898
+ "step": 121457
253899
+ },
253900
+ {
253901
+ "epoch": 971.02,
253902
+ "learning_rate": 8.070320512820514e-06,
253903
+ "loss": 0.2944,
253904
+ "step": 121460
253905
+ },
253906
+ {
253907
+ "epoch": 971.06,
253908
+ "learning_rate": 8.070240384615385e-06,
253909
+ "loss": 0.2993,
253910
+ "step": 121465
253911
+ },
253912
+ {
253913
+ "epoch": 971.1,
253914
+ "learning_rate": 8.070160256410256e-06,
253915
+ "loss": 0.4033,
253916
+ "step": 121470
253917
+ },
253918
+ {
253919
+ "epoch": 971.14,
253920
+ "learning_rate": 8.07008012820513e-06,
253921
+ "loss": 0.3557,
253922
+ "step": 121475
253923
+ },
253924
+ {
253925
+ "epoch": 971.18,
253926
+ "learning_rate": 8.07e-06,
253927
+ "loss": 0.7841,
253928
+ "step": 121480
253929
+ },
253930
+ {
253931
+ "epoch": 971.22,
253932
+ "learning_rate": 8.069919871794872e-06,
253933
+ "loss": 0.8209,
253934
+ "step": 121485
253935
+ },
253936
+ {
253937
+ "epoch": 971.26,
253938
+ "learning_rate": 8.069839743589745e-06,
253939
+ "loss": 0.297,
253940
+ "step": 121490
253941
+ },
253942
+ {
253943
+ "epoch": 971.3,
253944
+ "learning_rate": 8.069759615384616e-06,
253945
+ "loss": 0.2987,
253946
+ "step": 121495
253947
+ },
253948
+ {
253949
+ "epoch": 971.34,
253950
+ "learning_rate": 8.069679487179488e-06,
253951
+ "loss": 0.3397,
253952
+ "step": 121500
253953
+ },
253954
+ {
253955
+ "epoch": 971.38,
253956
+ "learning_rate": 8.069599358974359e-06,
253957
+ "loss": 0.6693,
253958
+ "step": 121505
253959
+ },
253960
+ {
253961
+ "epoch": 971.42,
253962
+ "learning_rate": 8.069519230769232e-06,
253963
+ "loss": 0.7268,
253964
+ "step": 121510
253965
+ },
253966
+ {
253967
+ "epoch": 971.46,
253968
+ "learning_rate": 8.069439102564104e-06,
253969
+ "loss": 0.2656,
253970
+ "step": 121515
253971
+ },
253972
+ {
253973
+ "epoch": 971.5,
253974
+ "learning_rate": 8.069358974358975e-06,
253975
+ "loss": 0.3631,
253976
+ "step": 121520
253977
+ },
253978
+ {
253979
+ "epoch": 971.54,
253980
+ "learning_rate": 8.069278846153846e-06,
253981
+ "loss": 0.4355,
253982
+ "step": 121525
253983
+ },
253984
+ {
253985
+ "epoch": 971.58,
253986
+ "learning_rate": 8.06919871794872e-06,
253987
+ "loss": 0.6999,
253988
+ "step": 121530
253989
+ },
253990
+ {
253991
+ "epoch": 971.62,
253992
+ "learning_rate": 8.06911858974359e-06,
253993
+ "loss": 0.7744,
253994
+ "step": 121535
253995
+ },
253996
+ {
253997
+ "epoch": 971.66,
253998
+ "learning_rate": 8.069038461538462e-06,
253999
+ "loss": 0.3289,
254000
+ "step": 121540
254001
+ },
254002
+ {
254003
+ "epoch": 971.7,
254004
+ "learning_rate": 8.068958333333335e-06,
254005
+ "loss": 0.2904,
254006
+ "step": 121545
254007
+ },
254008
+ {
254009
+ "epoch": 971.74,
254010
+ "learning_rate": 8.068878205128205e-06,
254011
+ "loss": 0.2956,
254012
+ "step": 121550
254013
+ },
254014
+ {
254015
+ "epoch": 971.78,
254016
+ "learning_rate": 8.068798076923078e-06,
254017
+ "loss": 0.682,
254018
+ "step": 121555
254019
+ },
254020
+ {
254021
+ "epoch": 971.82,
254022
+ "learning_rate": 8.06871794871795e-06,
254023
+ "loss": 0.9232,
254024
+ "step": 121560
254025
+ },
254026
+ {
254027
+ "epoch": 971.86,
254028
+ "learning_rate": 8.06863782051282e-06,
254029
+ "loss": 0.346,
254030
+ "step": 121565
254031
+ },
254032
+ {
254033
+ "epoch": 971.9,
254034
+ "learning_rate": 8.068557692307692e-06,
254035
+ "loss": 0.2865,
254036
+ "step": 121570
254037
+ },
254038
+ {
254039
+ "epoch": 971.94,
254040
+ "learning_rate": 8.068477564102565e-06,
254041
+ "loss": 0.3959,
254042
+ "step": 121575
254043
+ },
254044
+ {
254045
+ "epoch": 971.98,
254046
+ "learning_rate": 8.068397435897436e-06,
254047
+ "loss": 0.7047,
254048
+ "step": 121580
254049
+ },
254050
+ {
254051
+ "epoch": 972.0,
254052
+ "eval_loss": 0.3686648905277252,
254053
+ "eval_runtime": 35.6826,
254054
+ "eval_samples_per_second": 23.597,
254055
+ "eval_steps_per_second": 0.757,
254056
+ "eval_wer": 0.18131747080854813,
254057
+ "step": 121582
254058
+ },
254059
+ {
254060
+ "epoch": 972.02,
254061
+ "learning_rate": 8.068317307692308e-06,
254062
+ "loss": 0.3209,
254063
+ "step": 121585
254064
+ },
254065
+ {
254066
+ "epoch": 972.06,
254067
+ "learning_rate": 8.06823717948718e-06,
254068
+ "loss": 0.2486,
254069
+ "step": 121590
254070
+ },
254071
+ {
254072
+ "epoch": 972.1,
254073
+ "learning_rate": 8.068157051282052e-06,
254074
+ "loss": 0.3214,
254075
+ "step": 121595
254076
+ },
254077
+ {
254078
+ "epoch": 972.14,
254079
+ "learning_rate": 8.068076923076923e-06,
254080
+ "loss": 0.2991,
254081
+ "step": 121600
254082
+ },
254083
+ {
254084
+ "epoch": 972.18,
254085
+ "learning_rate": 8.067996794871795e-06,
254086
+ "loss": 0.6965,
254087
+ "step": 121605
254088
+ },
254089
+ {
254090
+ "epoch": 972.22,
254091
+ "learning_rate": 8.067916666666668e-06,
254092
+ "loss": 0.784,
254093
+ "step": 121610
254094
+ },
254095
+ {
254096
+ "epoch": 972.26,
254097
+ "learning_rate": 8.06783653846154e-06,
254098
+ "loss": 0.3802,
254099
+ "step": 121615
254100
+ },
254101
+ {
254102
+ "epoch": 972.3,
254103
+ "learning_rate": 8.06775641025641e-06,
254104
+ "loss": 0.318,
254105
+ "step": 121620
254106
+ },
254107
+ {
254108
+ "epoch": 972.34,
254109
+ "learning_rate": 8.067676282051282e-06,
254110
+ "loss": 0.4142,
254111
+ "step": 121625
254112
+ },
254113
+ {
254114
+ "epoch": 972.38,
254115
+ "learning_rate": 8.067596153846155e-06,
254116
+ "loss": 0.6518,
254117
+ "step": 121630
254118
+ },
254119
+ {
254120
+ "epoch": 972.42,
254121
+ "learning_rate": 8.067516025641026e-06,
254122
+ "loss": 0.8874,
254123
+ "step": 121635
254124
+ },
254125
+ {
254126
+ "epoch": 972.46,
254127
+ "learning_rate": 8.067435897435898e-06,
254128
+ "loss": 0.3634,
254129
+ "step": 121640
254130
+ },
254131
+ {
254132
+ "epoch": 972.5,
254133
+ "learning_rate": 8.06735576923077e-06,
254134
+ "loss": 0.3147,
254135
+ "step": 121645
254136
+ },
254137
+ {
254138
+ "epoch": 972.54,
254139
+ "learning_rate": 8.067275641025642e-06,
254140
+ "loss": 0.3491,
254141
+ "step": 121650
254142
+ },
254143
+ {
254144
+ "epoch": 972.58,
254145
+ "learning_rate": 8.067195512820513e-06,
254146
+ "loss": 0.7093,
254147
+ "step": 121655
254148
+ },
254149
+ {
254150
+ "epoch": 972.62,
254151
+ "learning_rate": 8.067115384615385e-06,
254152
+ "loss": 0.9415,
254153
+ "step": 121660
254154
+ },
254155
+ {
254156
+ "epoch": 972.66,
254157
+ "learning_rate": 8.067035256410258e-06,
254158
+ "loss": 0.3118,
254159
+ "step": 121665
254160
+ },
254161
+ {
254162
+ "epoch": 972.7,
254163
+ "learning_rate": 8.06695512820513e-06,
254164
+ "loss": 0.2936,
254165
+ "step": 121670
254166
+ },
254167
+ {
254168
+ "epoch": 972.74,
254169
+ "learning_rate": 8.066875e-06,
254170
+ "loss": 0.338,
254171
+ "step": 121675
254172
+ },
254173
+ {
254174
+ "epoch": 972.78,
254175
+ "learning_rate": 8.066794871794874e-06,
254176
+ "loss": 0.6909,
254177
+ "step": 121680
254178
+ },
254179
+ {
254180
+ "epoch": 972.82,
254181
+ "learning_rate": 8.066714743589745e-06,
254182
+ "loss": 0.7566,
254183
+ "step": 121685
254184
+ },
254185
+ {
254186
+ "epoch": 972.86,
254187
+ "learning_rate": 8.066634615384616e-06,
254188
+ "loss": 0.2324,
254189
+ "step": 121690
254190
+ },
254191
+ {
254192
+ "epoch": 972.9,
254193
+ "learning_rate": 8.066554487179488e-06,
254194
+ "loss": 0.309,
254195
+ "step": 121695
254196
+ },
254197
+ {
254198
+ "epoch": 972.94,
254199
+ "learning_rate": 8.06647435897436e-06,
254200
+ "loss": 0.367,
254201
+ "step": 121700
254202
+ },
254203
+ {
254204
+ "epoch": 972.98,
254205
+ "learning_rate": 8.06639423076923e-06,
254206
+ "loss": 0.9295,
254207
+ "step": 121705
254208
+ },
254209
+ {
254210
+ "epoch": 973.0,
254211
+ "eval_loss": 0.3698074519634247,
254212
+ "eval_runtime": 37.0962,
254213
+ "eval_samples_per_second": 22.698,
254214
+ "eval_steps_per_second": 0.728,
254215
+ "eval_wer": 0.18579426172148356,
254216
+ "step": 121707
254217
  }
254218
  ],
254219
  "max_steps": 625000,
254220
  "num_train_epochs": 5000,
254221
+ "total_flos": 3.425064251533899e+20,
254222
  "trial_name": null,
254223
  "trial_params": null
254224
  }
model-bin/finetune/base/{checkpoint-121084 β†’ checkpoint-121707}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630132562.2337933/events.out.tfevents.1630132562.86bb0ddabf9b.4092.51 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:baddbba9e71714999abac43eec83bf67c7533fa1604d8205e55509dda3ec6247
3
+ size 4194
model-bin/finetune/base/log/1630132940.836788/events.out.tfevents.1630132940.86bb0ddabf9b.4092.53 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fa409f40a52697f7113a591f335ed4829d67463862bb2cb043c4eb843e87e91
3
+ size 4194
model-bin/finetune/base/log/1630133330.209003/events.out.tfevents.1630133330.86bb0ddabf9b.4092.55 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70d936ec4eca46cf5a46ce6ea0444453fcb73a523136686f76dba8b6202dbc21
3
+ size 4194
model-bin/finetune/base/log/1630133714.3736/events.out.tfevents.1630133714.86bb0ddabf9b.4092.57 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1476550766e11c54c6b5a6eaa11fa671f5a33d9a5c6feba04057397204dfdf0f
3
+ size 4194
model-bin/finetune/base/log/1630134097.5583656/events.out.tfevents.1630134097.86bb0ddabf9b.4092.59 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:babd24ecf25f909721bcf44a4929f50942d73cbb5c12f82e1d49ad88d6fd3685
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630132562.86bb0ddabf9b.4092.50 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45e83d1153142d32dac1279afb71f50560619d3196cde2e0d55e9daec9cef2da
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630132940.86bb0ddabf9b.4092.52 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d70449ee1a35ac9e05b353210136e33592e5b2e2db9a3037aee364c6f9f7f00
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630133330.86bb0ddabf9b.4092.54 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16cc3ddada4eec5c151b06e9ad33a8320e77e41bceb204f77fc798dc2aa4e2cd
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630133714.86bb0ddabf9b.4092.56 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bebae397e40701784ff5201669ce666cce1c985cecdfb3adb1657096f97ae639
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630134097.86bb0ddabf9b.4092.58 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84f7af9e0a043b7745405261785cfadf7bc068c1fda0fd55408003dfe277d65a
3
+ size 8622