Check commited on
Commit
a9aa119
Β·
1 Parent(s): c4db64e

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629964487.380869/events.out.tfevents.1629964487.8e89bd551565.924.111 +3 -0
  11. model-bin/finetune/base/log/1629964926.187713/events.out.tfevents.1629964926.8e89bd551565.924.113 +3 -0
  12. model-bin/finetune/base/log/1629965370.4697628/events.out.tfevents.1629965370.8e89bd551565.924.115 +3 -0
  13. model-bin/finetune/base/log/1629965797.9083533/events.out.tfevents.1629965797.8e89bd551565.924.117 +3 -0
  14. model-bin/finetune/base/log/1629966225.1204813/events.out.tfevents.1629966225.8e89bd551565.924.119 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629964487.8e89bd551565.924.110 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629964926.8e89bd551565.924.112 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629965368.8e89bd551565.924.114 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629965797.8e89bd551565.924.116 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629966225.8e89bd551565.924.118 +3 -0
model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2659a34e917a75bf44ea9098c88f7ec69a641c2c56021fdcfeeeee6867230351
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f10fcf2ba5b5043370eb75b02ab550fa8de52a8d48b785cd2a7a34ee8cd3764
3
  size 722165393
model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a6ebd6b3f01e881ab9cc0e3de1da27d69fbcade5063e4daa9fc9f58b3cfcce5
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03328ac6ccc2b112de19e4fb255a6b080339e8765bde06155a660382a04b6fd0
3
  size 377909911
model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c6036c6ef3bafb3cadfc12a8274b3ca9f3a75647660af17949bb2098c645a4d1
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:836e20a6bf2106602ba0bcf5f0a522df38ed03fb672c5d7ac728e418b994c5c5
3
  size 14503
model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3ea1dcb8a64d00044fdd03ddccc110577c7eb81ecd8aad34e49e3d94f09fd63b
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1489a51728ff137e38a35bf5e8ede61ad98416177be49f5b464465239ddbc43f
3
  size 559
model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f3535739f660ec3f51f187361d26affc8b8b369b384ed5f372ffd4fc2a923fa
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72dea23aa8c63e88416a4f9fa0ccbab7dee5eb283cd5168ed9449f3c30f52e3b
3
  size 623
model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.17565318086415285,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-94333",
4
- "epoch": 766.995983935743,
5
- "global_step": 95203,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -220491,11 +220491,800 @@
220491
  "eval_steps_per_second": 0.695,
220492
  "eval_wer": 0.18689181453921008,
220493
  "step": 95203
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
220494
  }
220495
  ],
220496
  "max_steps": 620000,
220497
  "num_train_epochs": 5000,
220498
- "total_flos": 2.6790445779490485e+20,
220499
  "trial_name": null,
220500
  "trial_params": null
220501
  }
 
1
  {
2
  "best_metric": 0.17565318086415285,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-94333",
4
+ "epoch": 771.995983935743,
5
+ "global_step": 95824,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
220491
  "eval_steps_per_second": 0.695,
220492
  "eval_wer": 0.18689181453921008,
220493
  "step": 95203
220494
+ },
220495
+ {
220496
+ "epoch": 761.02,
220497
+ "learning_rate": 8.490913461538462e-06,
220498
+ "loss": 0.3735,
220499
+ "step": 95205
220500
+ },
220501
+ {
220502
+ "epoch": 761.06,
220503
+ "learning_rate": 8.490833333333333e-06,
220504
+ "loss": 0.281,
220505
+ "step": 95210
220506
+ },
220507
+ {
220508
+ "epoch": 761.1,
220509
+ "learning_rate": 8.490753205128207e-06,
220510
+ "loss": 0.3138,
220511
+ "step": 95215
220512
+ },
220513
+ {
220514
+ "epoch": 761.14,
220515
+ "learning_rate": 8.490673076923078e-06,
220516
+ "loss": 0.3501,
220517
+ "step": 95220
220518
+ },
220519
+ {
220520
+ "epoch": 761.18,
220521
+ "learning_rate": 8.49059294871795e-06,
220522
+ "loss": 0.674,
220523
+ "step": 95225
220524
+ },
220525
+ {
220526
+ "epoch": 761.22,
220527
+ "learning_rate": 8.49051282051282e-06,
220528
+ "loss": 1.1776,
220529
+ "step": 95230
220530
+ },
220531
+ {
220532
+ "epoch": 761.26,
220533
+ "learning_rate": 8.490432692307694e-06,
220534
+ "loss": 0.2888,
220535
+ "step": 95235
220536
+ },
220537
+ {
220538
+ "epoch": 761.3,
220539
+ "learning_rate": 8.490352564102565e-06,
220540
+ "loss": 0.2533,
220541
+ "step": 95240
220542
+ },
220543
+ {
220544
+ "epoch": 761.34,
220545
+ "learning_rate": 8.490272435897436e-06,
220546
+ "loss": 0.3197,
220547
+ "step": 95245
220548
+ },
220549
+ {
220550
+ "epoch": 761.38,
220551
+ "learning_rate": 8.49019230769231e-06,
220552
+ "loss": 0.6192,
220553
+ "step": 95250
220554
+ },
220555
+ {
220556
+ "epoch": 761.42,
220557
+ "learning_rate": 8.490112179487179e-06,
220558
+ "loss": 1.0574,
220559
+ "step": 95255
220560
+ },
220561
+ {
220562
+ "epoch": 761.46,
220563
+ "learning_rate": 8.490032051282052e-06,
220564
+ "loss": 0.3482,
220565
+ "step": 95260
220566
+ },
220567
+ {
220568
+ "epoch": 761.5,
220569
+ "learning_rate": 8.489951923076924e-06,
220570
+ "loss": 0.3048,
220571
+ "step": 95265
220572
+ },
220573
+ {
220574
+ "epoch": 761.54,
220575
+ "learning_rate": 8.489871794871795e-06,
220576
+ "loss": 0.3254,
220577
+ "step": 95270
220578
+ },
220579
+ {
220580
+ "epoch": 761.58,
220581
+ "learning_rate": 8.489791666666666e-06,
220582
+ "loss": 0.7295,
220583
+ "step": 95275
220584
+ },
220585
+ {
220586
+ "epoch": 761.62,
220587
+ "learning_rate": 8.48971153846154e-06,
220588
+ "loss": 0.9835,
220589
+ "step": 95280
220590
+ },
220591
+ {
220592
+ "epoch": 761.66,
220593
+ "learning_rate": 8.48963141025641e-06,
220594
+ "loss": 0.3579,
220595
+ "step": 95285
220596
+ },
220597
+ {
220598
+ "epoch": 761.7,
220599
+ "learning_rate": 8.489551282051282e-06,
220600
+ "loss": 0.2816,
220601
+ "step": 95290
220602
+ },
220603
+ {
220604
+ "epoch": 761.74,
220605
+ "learning_rate": 8.489471153846155e-06,
220606
+ "loss": 0.345,
220607
+ "step": 95295
220608
+ },
220609
+ {
220610
+ "epoch": 761.78,
220611
+ "learning_rate": 8.489391025641026e-06,
220612
+ "loss": 0.5479,
220613
+ "step": 95300
220614
+ },
220615
+ {
220616
+ "epoch": 761.82,
220617
+ "learning_rate": 8.489310897435898e-06,
220618
+ "loss": 1.0586,
220619
+ "step": 95305
220620
+ },
220621
+ {
220622
+ "epoch": 761.86,
220623
+ "learning_rate": 8.489230769230769e-06,
220624
+ "loss": 0.3759,
220625
+ "step": 95310
220626
+ },
220627
+ {
220628
+ "epoch": 761.9,
220629
+ "learning_rate": 8.489150641025642e-06,
220630
+ "loss": 0.2978,
220631
+ "step": 95315
220632
+ },
220633
+ {
220634
+ "epoch": 761.94,
220635
+ "learning_rate": 8.489070512820514e-06,
220636
+ "loss": 0.4061,
220637
+ "step": 95320
220638
+ },
220639
+ {
220640
+ "epoch": 761.98,
220641
+ "learning_rate": 8.488990384615385e-06,
220642
+ "loss": 0.634,
220643
+ "step": 95325
220644
+ },
220645
+ {
220646
+ "epoch": 762.0,
220647
+ "eval_loss": 0.3908168375492096,
220648
+ "eval_runtime": 40.2724,
220649
+ "eval_samples_per_second": 20.858,
220650
+ "eval_steps_per_second": 0.67,
220651
+ "eval_wer": 0.19241309678349922,
220652
+ "step": 95328
220653
+ },
220654
+ {
220655
+ "epoch": 768.02,
220656
+ "learning_rate": 8.488910256410256e-06,
220657
+ "loss": 0.3328,
220658
+ "step": 95330
220659
+ },
220660
+ {
220661
+ "epoch": 768.06,
220662
+ "learning_rate": 8.48883012820513e-06,
220663
+ "loss": 0.307,
220664
+ "step": 95335
220665
+ },
220666
+ {
220667
+ "epoch": 768.1,
220668
+ "learning_rate": 8.48875e-06,
220669
+ "loss": 0.3208,
220670
+ "step": 95340
220671
+ },
220672
+ {
220673
+ "epoch": 768.14,
220674
+ "learning_rate": 8.488669871794872e-06,
220675
+ "loss": 0.4314,
220676
+ "step": 95345
220677
+ },
220678
+ {
220679
+ "epoch": 768.18,
220680
+ "learning_rate": 8.488589743589745e-06,
220681
+ "loss": 0.5051,
220682
+ "step": 95350
220683
+ },
220684
+ {
220685
+ "epoch": 768.22,
220686
+ "learning_rate": 8.488509615384616e-06,
220687
+ "loss": 0.9168,
220688
+ "step": 95355
220689
+ },
220690
+ {
220691
+ "epoch": 768.26,
220692
+ "learning_rate": 8.488429487179488e-06,
220693
+ "loss": 0.2612,
220694
+ "step": 95360
220695
+ },
220696
+ {
220697
+ "epoch": 768.3,
220698
+ "learning_rate": 8.48834935897436e-06,
220699
+ "loss": 0.2868,
220700
+ "step": 95365
220701
+ },
220702
+ {
220703
+ "epoch": 768.34,
220704
+ "learning_rate": 8.488269230769232e-06,
220705
+ "loss": 0.351,
220706
+ "step": 95370
220707
+ },
220708
+ {
220709
+ "epoch": 768.38,
220710
+ "learning_rate": 8.488189102564104e-06,
220711
+ "loss": 0.7034,
220712
+ "step": 95375
220713
+ },
220714
+ {
220715
+ "epoch": 768.42,
220716
+ "learning_rate": 8.488108974358975e-06,
220717
+ "loss": 1.0855,
220718
+ "step": 95380
220719
+ },
220720
+ {
220721
+ "epoch": 768.46,
220722
+ "learning_rate": 8.488028846153848e-06,
220723
+ "loss": 0.2586,
220724
+ "step": 95385
220725
+ },
220726
+ {
220727
+ "epoch": 768.5,
220728
+ "learning_rate": 8.48794871794872e-06,
220729
+ "loss": 0.2588,
220730
+ "step": 95390
220731
+ },
220732
+ {
220733
+ "epoch": 768.54,
220734
+ "learning_rate": 8.48786858974359e-06,
220735
+ "loss": 0.391,
220736
+ "step": 95395
220737
+ },
220738
+ {
220739
+ "epoch": 768.58,
220740
+ "learning_rate": 8.487788461538462e-06,
220741
+ "loss": 0.6738,
220742
+ "step": 95400
220743
+ },
220744
+ {
220745
+ "epoch": 768.62,
220746
+ "learning_rate": 8.487708333333335e-06,
220747
+ "loss": 1.0473,
220748
+ "step": 95405
220749
+ },
220750
+ {
220751
+ "epoch": 768.66,
220752
+ "learning_rate": 8.487628205128205e-06,
220753
+ "loss": 0.2548,
220754
+ "step": 95410
220755
+ },
220756
+ {
220757
+ "epoch": 768.7,
220758
+ "learning_rate": 8.487548076923078e-06,
220759
+ "loss": 0.3107,
220760
+ "step": 95415
220761
+ },
220762
+ {
220763
+ "epoch": 768.74,
220764
+ "learning_rate": 8.48746794871795e-06,
220765
+ "loss": 0.3829,
220766
+ "step": 95420
220767
+ },
220768
+ {
220769
+ "epoch": 768.78,
220770
+ "learning_rate": 8.48738782051282e-06,
220771
+ "loss": 0.6175,
220772
+ "step": 95425
220773
+ },
220774
+ {
220775
+ "epoch": 768.82,
220776
+ "learning_rate": 8.487307692307692e-06,
220777
+ "loss": 1.1168,
220778
+ "step": 95430
220779
+ },
220780
+ {
220781
+ "epoch": 768.86,
220782
+ "learning_rate": 8.487227564102565e-06,
220783
+ "loss": 0.2852,
220784
+ "step": 95435
220785
+ },
220786
+ {
220787
+ "epoch": 768.9,
220788
+ "learning_rate": 8.487147435897436e-06,
220789
+ "loss": 0.2895,
220790
+ "step": 95440
220791
+ },
220792
+ {
220793
+ "epoch": 768.94,
220794
+ "learning_rate": 8.487067307692308e-06,
220795
+ "loss": 0.3991,
220796
+ "step": 95445
220797
+ },
220798
+ {
220799
+ "epoch": 768.98,
220800
+ "learning_rate": 8.48698717948718e-06,
220801
+ "loss": 0.7499,
220802
+ "step": 95450
220803
+ },
220804
+ {
220805
+ "epoch": 769.0,
220806
+ "eval_loss": 0.33350133895874023,
220807
+ "eval_runtime": 39.8873,
220808
+ "eval_samples_per_second": 21.059,
220809
+ "eval_steps_per_second": 0.677,
220810
+ "eval_wer": 0.18204704020714954,
220811
+ "step": 95452
220812
+ },
220813
+ {
220814
+ "epoch": 769.02,
220815
+ "learning_rate": 8.486907051282052e-06,
220816
+ "loss": 0.3954,
220817
+ "step": 95455
220818
+ },
220819
+ {
220820
+ "epoch": 769.06,
220821
+ "learning_rate": 8.486826923076923e-06,
220822
+ "loss": 0.2815,
220823
+ "step": 95460
220824
+ },
220825
+ {
220826
+ "epoch": 769.1,
220827
+ "learning_rate": 8.486746794871795e-06,
220828
+ "loss": 0.2679,
220829
+ "step": 95465
220830
+ },
220831
+ {
220832
+ "epoch": 769.14,
220833
+ "learning_rate": 8.486666666666668e-06,
220834
+ "loss": 0.3121,
220835
+ "step": 95470
220836
+ },
220837
+ {
220838
+ "epoch": 769.18,
220839
+ "learning_rate": 8.48658653846154e-06,
220840
+ "loss": 0.6884,
220841
+ "step": 95475
220842
+ },
220843
+ {
220844
+ "epoch": 769.22,
220845
+ "learning_rate": 8.48650641025641e-06,
220846
+ "loss": 0.948,
220847
+ "step": 95480
220848
+ },
220849
+ {
220850
+ "epoch": 769.27,
220851
+ "learning_rate": 8.486426282051284e-06,
220852
+ "loss": 0.2971,
220853
+ "step": 95485
220854
+ },
220855
+ {
220856
+ "epoch": 769.31,
220857
+ "learning_rate": 8.486346153846155e-06,
220858
+ "loss": 0.369,
220859
+ "step": 95490
220860
+ },
220861
+ {
220862
+ "epoch": 769.35,
220863
+ "learning_rate": 8.486266025641026e-06,
220864
+ "loss": 0.3503,
220865
+ "step": 95495
220866
+ },
220867
+ {
220868
+ "epoch": 769.39,
220869
+ "learning_rate": 8.486185897435898e-06,
220870
+ "loss": 0.79,
220871
+ "step": 95500
220872
+ },
220873
+ {
220874
+ "epoch": 769.43,
220875
+ "learning_rate": 8.48610576923077e-06,
220876
+ "loss": 0.8587,
220877
+ "step": 95505
220878
+ },
220879
+ {
220880
+ "epoch": 769.47,
220881
+ "learning_rate": 8.486025641025642e-06,
220882
+ "loss": 0.2279,
220883
+ "step": 95510
220884
+ },
220885
+ {
220886
+ "epoch": 769.51,
220887
+ "learning_rate": 8.485945512820513e-06,
220888
+ "loss": 0.2907,
220889
+ "step": 95515
220890
+ },
220891
+ {
220892
+ "epoch": 769.55,
220893
+ "learning_rate": 8.485865384615385e-06,
220894
+ "loss": 0.34,
220895
+ "step": 95520
220896
+ },
220897
+ {
220898
+ "epoch": 769.59,
220899
+ "learning_rate": 8.485785256410258e-06,
220900
+ "loss": 0.7663,
220901
+ "step": 95525
220902
+ },
220903
+ {
220904
+ "epoch": 769.63,
220905
+ "learning_rate": 8.485705128205128e-06,
220906
+ "loss": 0.9137,
220907
+ "step": 95530
220908
+ },
220909
+ {
220910
+ "epoch": 769.67,
220911
+ "learning_rate": 8.485625e-06,
220912
+ "loss": 0.3007,
220913
+ "step": 95535
220914
+ },
220915
+ {
220916
+ "epoch": 769.71,
220917
+ "learning_rate": 8.485544871794874e-06,
220918
+ "loss": 0.3495,
220919
+ "step": 95540
220920
+ },
220921
+ {
220922
+ "epoch": 769.75,
220923
+ "learning_rate": 8.485464743589743e-06,
220924
+ "loss": 0.4196,
220925
+ "step": 95545
220926
+ },
220927
+ {
220928
+ "epoch": 769.79,
220929
+ "learning_rate": 8.485384615384616e-06,
220930
+ "loss": 0.7906,
220931
+ "step": 95550
220932
+ },
220933
+ {
220934
+ "epoch": 769.83,
220935
+ "learning_rate": 8.485304487179488e-06,
220936
+ "loss": 0.9746,
220937
+ "step": 95555
220938
+ },
220939
+ {
220940
+ "epoch": 769.87,
220941
+ "learning_rate": 8.485224358974359e-06,
220942
+ "loss": 0.2873,
220943
+ "step": 95560
220944
+ },
220945
+ {
220946
+ "epoch": 769.91,
220947
+ "learning_rate": 8.48514423076923e-06,
220948
+ "loss": 0.291,
220949
+ "step": 95565
220950
+ },
220951
+ {
220952
+ "epoch": 769.95,
220953
+ "learning_rate": 8.485064102564103e-06,
220954
+ "loss": 0.4418,
220955
+ "step": 95570
220956
+ },
220957
+ {
220958
+ "epoch": 769.99,
220959
+ "learning_rate": 8.484983974358975e-06,
220960
+ "loss": 0.7583,
220961
+ "step": 95575
220962
+ },
220963
+ {
220964
+ "epoch": 770.0,
220965
+ "eval_loss": 0.37961849570274353,
220966
+ "eval_runtime": 38.0095,
220967
+ "eval_samples_per_second": 22.126,
220968
+ "eval_steps_per_second": 0.71,
220969
+ "eval_wer": 0.18243194192377496,
220970
+ "step": 95576
220971
+ },
220972
+ {
220973
+ "epoch": 770.03,
220974
+ "learning_rate": 8.484903846153846e-06,
220975
+ "loss": 0.3532,
220976
+ "step": 95580
220977
+ },
220978
+ {
220979
+ "epoch": 770.07,
220980
+ "learning_rate": 8.48482371794872e-06,
220981
+ "loss": 0.2813,
220982
+ "step": 95585
220983
+ },
220984
+ {
220985
+ "epoch": 770.11,
220986
+ "learning_rate": 8.48474358974359e-06,
220987
+ "loss": 0.2872,
220988
+ "step": 95590
220989
+ },
220990
+ {
220991
+ "epoch": 770.15,
220992
+ "learning_rate": 8.484663461538462e-06,
220993
+ "loss": 0.4089,
220994
+ "step": 95595
220995
+ },
220996
+ {
220997
+ "epoch": 770.19,
220998
+ "learning_rate": 8.484583333333333e-06,
220999
+ "loss": 0.8651,
221000
+ "step": 95600
221001
+ },
221002
+ {
221003
+ "epoch": 770.23,
221004
+ "learning_rate": 8.484503205128206e-06,
221005
+ "loss": 0.834,
221006
+ "step": 95605
221007
+ },
221008
+ {
221009
+ "epoch": 770.27,
221010
+ "learning_rate": 8.484423076923078e-06,
221011
+ "loss": 0.4655,
221012
+ "step": 95610
221013
+ },
221014
+ {
221015
+ "epoch": 770.31,
221016
+ "learning_rate": 8.484342948717949e-06,
221017
+ "loss": 0.3295,
221018
+ "step": 95615
221019
+ },
221020
+ {
221021
+ "epoch": 770.35,
221022
+ "learning_rate": 8.48426282051282e-06,
221023
+ "loss": 0.3905,
221024
+ "step": 95620
221025
+ },
221026
+ {
221027
+ "epoch": 770.39,
221028
+ "learning_rate": 8.484182692307694e-06,
221029
+ "loss": 1.058,
221030
+ "step": 95625
221031
+ },
221032
+ {
221033
+ "epoch": 770.43,
221034
+ "learning_rate": 8.484102564102565e-06,
221035
+ "loss": 0.5864,
221036
+ "step": 95630
221037
+ },
221038
+ {
221039
+ "epoch": 770.47,
221040
+ "learning_rate": 8.484022435897436e-06,
221041
+ "loss": 0.2882,
221042
+ "step": 95635
221043
+ },
221044
+ {
221045
+ "epoch": 770.51,
221046
+ "learning_rate": 8.48394230769231e-06,
221047
+ "loss": 0.3196,
221048
+ "step": 95640
221049
+ },
221050
+ {
221051
+ "epoch": 770.55,
221052
+ "learning_rate": 8.48386217948718e-06,
221053
+ "loss": 0.437,
221054
+ "step": 95645
221055
+ },
221056
+ {
221057
+ "epoch": 770.59,
221058
+ "learning_rate": 8.483782051282052e-06,
221059
+ "loss": 0.9501,
221060
+ "step": 95650
221061
+ },
221062
+ {
221063
+ "epoch": 770.63,
221064
+ "learning_rate": 8.483701923076923e-06,
221065
+ "loss": 0.7306,
221066
+ "step": 95655
221067
+ },
221068
+ {
221069
+ "epoch": 770.67,
221070
+ "learning_rate": 8.483621794871796e-06,
221071
+ "loss": 0.3199,
221072
+ "step": 95660
221073
+ },
221074
+ {
221075
+ "epoch": 770.71,
221076
+ "learning_rate": 8.483541666666668e-06,
221077
+ "loss": 0.2947,
221078
+ "step": 95665
221079
+ },
221080
+ {
221081
+ "epoch": 770.76,
221082
+ "learning_rate": 8.483461538461539e-06,
221083
+ "loss": 0.3722,
221084
+ "step": 95670
221085
+ },
221086
+ {
221087
+ "epoch": 770.8,
221088
+ "learning_rate": 8.48338141025641e-06,
221089
+ "loss": 0.8973,
221090
+ "step": 95675
221091
+ },
221092
+ {
221093
+ "epoch": 770.84,
221094
+ "learning_rate": 8.483301282051284e-06,
221095
+ "loss": 0.6271,
221096
+ "step": 95680
221097
+ },
221098
+ {
221099
+ "epoch": 770.88,
221100
+ "learning_rate": 8.483221153846155e-06,
221101
+ "loss": 0.2649,
221102
+ "step": 95685
221103
+ },
221104
+ {
221105
+ "epoch": 770.92,
221106
+ "learning_rate": 8.483141025641026e-06,
221107
+ "loss": 0.3043,
221108
+ "step": 95690
221109
+ },
221110
+ {
221111
+ "epoch": 770.96,
221112
+ "learning_rate": 8.4830608974359e-06,
221113
+ "loss": 0.4054,
221114
+ "step": 95695
221115
+ },
221116
+ {
221117
+ "epoch": 771.0,
221118
+ "learning_rate": 8.482980769230769e-06,
221119
+ "loss": 1.0171,
221120
+ "step": 95700
221121
+ },
221122
+ {
221123
+ "epoch": 771.0,
221124
+ "eval_loss": 0.3492157757282257,
221125
+ "eval_runtime": 38.1524,
221126
+ "eval_samples_per_second": 22.017,
221127
+ "eval_steps_per_second": 0.708,
221128
+ "eval_wer": 0.18850003653101483,
221129
+ "step": 95700
221130
+ },
221131
+ {
221132
+ "epoch": 771.04,
221133
+ "learning_rate": 8.482900641025642e-06,
221134
+ "loss": 0.361,
221135
+ "step": 95705
221136
+ },
221137
+ {
221138
+ "epoch": 771.08,
221139
+ "learning_rate": 8.482820512820513e-06,
221140
+ "loss": 0.2913,
221141
+ "step": 95710
221142
+ },
221143
+ {
221144
+ "epoch": 771.12,
221145
+ "learning_rate": 8.482740384615385e-06,
221146
+ "loss": 0.3224,
221147
+ "step": 95715
221148
+ },
221149
+ {
221150
+ "epoch": 771.16,
221151
+ "learning_rate": 8.482660256410256e-06,
221152
+ "loss": 0.4519,
221153
+ "step": 95720
221154
+ },
221155
+ {
221156
+ "epoch": 771.2,
221157
+ "learning_rate": 8.48258012820513e-06,
221158
+ "loss": 1.2417,
221159
+ "step": 95725
221160
+ },
221161
+ {
221162
+ "epoch": 771.24,
221163
+ "learning_rate": 8.4825e-06,
221164
+ "loss": 0.3535,
221165
+ "step": 95730
221166
+ },
221167
+ {
221168
+ "epoch": 771.28,
221169
+ "learning_rate": 8.482419871794872e-06,
221170
+ "loss": 0.321,
221171
+ "step": 95735
221172
+ },
221173
+ {
221174
+ "epoch": 771.32,
221175
+ "learning_rate": 8.482339743589745e-06,
221176
+ "loss": 0.3113,
221177
+ "step": 95740
221178
+ },
221179
+ {
221180
+ "epoch": 771.36,
221181
+ "learning_rate": 8.482259615384616e-06,
221182
+ "loss": 0.4148,
221183
+ "step": 95745
221184
+ },
221185
+ {
221186
+ "epoch": 771.4,
221187
+ "learning_rate": 8.482179487179488e-06,
221188
+ "loss": 1.2129,
221189
+ "step": 95750
221190
+ },
221191
+ {
221192
+ "epoch": 771.44,
221193
+ "learning_rate": 8.482099358974359e-06,
221194
+ "loss": 0.3261,
221195
+ "step": 95755
221196
+ },
221197
+ {
221198
+ "epoch": 771.48,
221199
+ "learning_rate": 8.482019230769232e-06,
221200
+ "loss": 0.274,
221201
+ "step": 95760
221202
+ },
221203
+ {
221204
+ "epoch": 771.52,
221205
+ "learning_rate": 8.481939102564103e-06,
221206
+ "loss": 0.2832,
221207
+ "step": 95765
221208
+ },
221209
+ {
221210
+ "epoch": 771.56,
221211
+ "learning_rate": 8.481858974358975e-06,
221212
+ "loss": 0.4462,
221213
+ "step": 95770
221214
+ },
221215
+ {
221216
+ "epoch": 771.6,
221217
+ "learning_rate": 8.481778846153846e-06,
221218
+ "loss": 1.3003,
221219
+ "step": 95775
221220
+ },
221221
+ {
221222
+ "epoch": 771.64,
221223
+ "learning_rate": 8.48169871794872e-06,
221224
+ "loss": 0.3361,
221225
+ "step": 95780
221226
+ },
221227
+ {
221228
+ "epoch": 771.68,
221229
+ "learning_rate": 8.48161858974359e-06,
221230
+ "loss": 0.271,
221231
+ "step": 95785
221232
+ },
221233
+ {
221234
+ "epoch": 771.72,
221235
+ "learning_rate": 8.481538461538462e-06,
221236
+ "loss": 0.3512,
221237
+ "step": 95790
221238
+ },
221239
+ {
221240
+ "epoch": 771.76,
221241
+ "learning_rate": 8.481458333333335e-06,
221242
+ "loss": 0.4285,
221243
+ "step": 95795
221244
+ },
221245
+ {
221246
+ "epoch": 771.8,
221247
+ "learning_rate": 8.481378205128206e-06,
221248
+ "loss": 1.1133,
221249
+ "step": 95800
221250
+ },
221251
+ {
221252
+ "epoch": 771.84,
221253
+ "learning_rate": 8.481298076923078e-06,
221254
+ "loss": 0.2918,
221255
+ "step": 95805
221256
+ },
221257
+ {
221258
+ "epoch": 771.88,
221259
+ "learning_rate": 8.481217948717949e-06,
221260
+ "loss": 0.2749,
221261
+ "step": 95810
221262
+ },
221263
+ {
221264
+ "epoch": 771.92,
221265
+ "learning_rate": 8.481137820512822e-06,
221266
+ "loss": 0.3116,
221267
+ "step": 95815
221268
+ },
221269
+ {
221270
+ "epoch": 771.96,
221271
+ "learning_rate": 8.481057692307693e-06,
221272
+ "loss": 0.5563,
221273
+ "step": 95820
221274
+ },
221275
+ {
221276
+ "epoch": 772.0,
221277
+ "eval_loss": 0.4196493923664093,
221278
+ "eval_runtime": 38.6955,
221279
+ "eval_samples_per_second": 21.708,
221280
+ "eval_steps_per_second": 0.698,
221281
+ "eval_wer": 0.1937148929710035,
221282
+ "step": 95824
221283
  }
221284
  ],
221285
  "max_steps": 620000,
221286
  "num_train_epochs": 5000,
221287
+ "total_flos": 2.6965513860738492e+20,
221288
  "trial_name": null,
221289
  "trial_params": null
221290
  }
model-bin/finetune/base/{checkpoint-95203 β†’ checkpoint-95824}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629964487.380869/events.out.tfevents.1629964487.8e89bd551565.924.111 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:370fb1e5a21e5ed84af42256a50fc1559af0569d214371fd6a007326bca845a2
3
+ size 4194
model-bin/finetune/base/log/1629964926.187713/events.out.tfevents.1629964926.8e89bd551565.924.113 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0365e2dbc6178de478f37251cd475fe7c03da2ee0cec7bd1f97fa17f347101db
3
+ size 4194
model-bin/finetune/base/log/1629965370.4697628/events.out.tfevents.1629965370.8e89bd551565.924.115 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29adf051b9afb640abe92caf8706a078a74468221337af4423ef7c5c82942761
3
+ size 4194
model-bin/finetune/base/log/1629965797.9083533/events.out.tfevents.1629965797.8e89bd551565.924.117 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c38a0d0648d9afa30bd6a184295a5da6120c73907057de82af76d662f892710e
3
+ size 4194
model-bin/finetune/base/log/1629966225.1204813/events.out.tfevents.1629966225.8e89bd551565.924.119 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:feb7fc5ecfbaca7f56f28251f593c5306e8cf68c422b21569b004ef7fe7db5d1
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629964487.8e89bd551565.924.110 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce0236705ca4db610d29576f7ded67f7eb44fe7a2163e4ba0315572861a2c512
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629964926.8e89bd551565.924.112 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4230c306ec3cfc64cc1a2c0d801a88a1af60ecac44b4110d9883f70c88c60fc
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629965368.8e89bd551565.924.114 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55863a509b5da2b4920c1bedd9fb7b552136f69591fa01f51c3ae538e771db1d
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629965797.8e89bd551565.924.116 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:780534551afb252eb6467553112322bda2937c2a339d7e10befc9783e0a65572
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629966225.8e89bd551565.924.118 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1639a881fe5098533b41b3e65cccbe304ed455d6471a834b91c59f17348cad7
3
+ size 8462