Check commited on
Commit
78280ac
Β·
1 Parent(s): cb2f2d4

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/trainer_state.json +799 -4
  9. model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629904128.5656571/events.out.tfevents.1629904128.7e498afd5545.7645.65 +3 -0
  11. model-bin/finetune/base/log/1629904601.0615444/events.out.tfevents.1629904601.7e498afd5545.7645.67 +3 -0
  12. model-bin/finetune/base/log/1629905064.9216402/events.out.tfevents.1629905064.7e498afd5545.7645.69 +3 -0
  13. model-bin/finetune/base/log/1629905527.7759078/events.out.tfevents.1629905527.7e498afd5545.7645.71 +3 -0
  14. model-bin/finetune/base/log/1629905996.4959083/events.out.tfevents.1629905996.7e498afd5545.7645.73 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629904128.7e498afd5545.7645.64 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629904601.7e498afd5545.7645.66 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629905064.7e498afd5545.7645.68 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629905527.7e498afd5545.7645.70 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629905996.7e498afd5545.7645.72 +3 -0
model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2402c050872a1b6c2132f1354c8235dbc30ffdaec42a7c3e2fe40a3fa3597e6e
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:499aed0d9010da5763c53fa596bd3690ab653855d90b29a7179c034cca89ab46
3
  size 722165393
model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9ba89a49bd0701ea0db537127d26ed94e4e5fa6eed33bd9ae9d90248a8d473ba
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2f4660a08d626636627c24d99c2d12de51bf5a22d32bade54a02d377b1c8589
3
  size 377909911
model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ea9d23665821b123a85d6ac5512b6a51253cf13d4f317b3d68b6de999fb06ba7
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1359678ed57ef76ba4a01755e2169ac00e857b4fb73f1475b304a099e28649dc
3
  size 14503
model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2edae0f6397edeecd1aedd6821f5f0301b0061a47a55585acacade7161821b94
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e8a586bc9b63765c3fab357431c101314b42a44cf283ddbb669f9978dbc3e9b
3
  size 559
model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5ca9b1c1ac3a3f74bba9515f59a166eb16c9179459f7d8dfba604bf995fff1f5
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb6dd82815b3396a90d14cd2e98065c9d2df14b6b8fa8441658e29c12220c22b
3
  size 623
model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1822989349981638,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-79148",
4
- "epoch": 642.0,
5
- "global_step": 80269,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -201489,11 +201489,806 @@
201489
  "eval_steps_per_second": 0.626,
201490
  "eval_wer": 0.18905362543335547,
201491
  "step": 80269
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
201492
  }
201493
  ],
201494
- "max_steps": 625000,
201495
  "num_train_epochs": 5000,
201496
- "total_flos": 2.258889246771459e+20,
201497
  "trial_name": null,
201498
  "trial_params": null
201499
  }
 
1
  {
2
  "best_metric": 0.1822989349981638,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-79148",
4
+ "epoch": 651.995983935743,
5
+ "global_step": 80891,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
201489
  "eval_steps_per_second": 0.626,
201490
  "eval_wer": 0.18905362543335547,
201491
  "step": 80269
201492
+ },
201493
+ {
201494
+ "epoch": 642.01,
201495
+ "learning_rate": 8.719886914378029e-06,
201496
+ "loss": 0.4176,
201497
+ "step": 80270
201498
+ },
201499
+ {
201500
+ "epoch": 642.05,
201501
+ "learning_rate": 8.719806138933765e-06,
201502
+ "loss": 0.2807,
201503
+ "step": 80275
201504
+ },
201505
+ {
201506
+ "epoch": 642.09,
201507
+ "learning_rate": 8.719725363489499e-06,
201508
+ "loss": 0.3207,
201509
+ "step": 80280
201510
+ },
201511
+ {
201512
+ "epoch": 642.13,
201513
+ "learning_rate": 8.719644588045235e-06,
201514
+ "loss": 0.369,
201515
+ "step": 80285
201516
+ },
201517
+ {
201518
+ "epoch": 642.17,
201519
+ "learning_rate": 8.719563812600969e-06,
201520
+ "loss": 0.614,
201521
+ "step": 80290
201522
+ },
201523
+ {
201524
+ "epoch": 642.21,
201525
+ "learning_rate": 8.719483037156705e-06,
201526
+ "loss": 1.1049,
201527
+ "step": 80295
201528
+ },
201529
+ {
201530
+ "epoch": 642.25,
201531
+ "learning_rate": 8.719402261712439e-06,
201532
+ "loss": 0.327,
201533
+ "step": 80300
201534
+ },
201535
+ {
201536
+ "epoch": 642.29,
201537
+ "learning_rate": 8.719321486268175e-06,
201538
+ "loss": 0.3335,
201539
+ "step": 80305
201540
+ },
201541
+ {
201542
+ "epoch": 642.33,
201543
+ "learning_rate": 8.71924071082391e-06,
201544
+ "loss": 0.392,
201545
+ "step": 80310
201546
+ },
201547
+ {
201548
+ "epoch": 642.37,
201549
+ "learning_rate": 8.719159935379645e-06,
201550
+ "loss": 0.4432,
201551
+ "step": 80315
201552
+ },
201553
+ {
201554
+ "epoch": 642.41,
201555
+ "learning_rate": 8.71907915993538e-06,
201556
+ "loss": 1.3292,
201557
+ "step": 80320
201558
+ },
201559
+ {
201560
+ "epoch": 642.45,
201561
+ "learning_rate": 8.718998384491115e-06,
201562
+ "loss": 0.3665,
201563
+ "step": 80325
201564
+ },
201565
+ {
201566
+ "epoch": 642.49,
201567
+ "learning_rate": 8.71891760904685e-06,
201568
+ "loss": 0.2565,
201569
+ "step": 80330
201570
+ },
201571
+ {
201572
+ "epoch": 642.53,
201573
+ "learning_rate": 8.718836833602585e-06,
201574
+ "loss": 0.4313,
201575
+ "step": 80335
201576
+ },
201577
+ {
201578
+ "epoch": 642.57,
201579
+ "learning_rate": 8.71875605815832e-06,
201580
+ "loss": 0.5453,
201581
+ "step": 80340
201582
+ },
201583
+ {
201584
+ "epoch": 642.61,
201585
+ "learning_rate": 8.718675282714055e-06,
201586
+ "loss": 0.9556,
201587
+ "step": 80345
201588
+ },
201589
+ {
201590
+ "epoch": 642.65,
201591
+ "learning_rate": 8.71859450726979e-06,
201592
+ "loss": 0.2915,
201593
+ "step": 80350
201594
+ },
201595
+ {
201596
+ "epoch": 642.69,
201597
+ "learning_rate": 8.718513731825525e-06,
201598
+ "loss": 0.3673,
201599
+ "step": 80355
201600
+ },
201601
+ {
201602
+ "epoch": 642.73,
201603
+ "learning_rate": 8.71843295638126e-06,
201604
+ "loss": 0.3407,
201605
+ "step": 80360
201606
+ },
201607
+ {
201608
+ "epoch": 642.77,
201609
+ "learning_rate": 8.718352180936995e-06,
201610
+ "loss": 0.4414,
201611
+ "step": 80365
201612
+ },
201613
+ {
201614
+ "epoch": 642.81,
201615
+ "learning_rate": 8.71827140549273e-06,
201616
+ "loss": 1.264,
201617
+ "step": 80370
201618
+ },
201619
+ {
201620
+ "epoch": 642.85,
201621
+ "learning_rate": 8.718190630048466e-06,
201622
+ "loss": 0.3002,
201623
+ "step": 80375
201624
+ },
201625
+ {
201626
+ "epoch": 642.89,
201627
+ "learning_rate": 8.7181098546042e-06,
201628
+ "loss": 0.3777,
201629
+ "step": 80380
201630
+ },
201631
+ {
201632
+ "epoch": 642.93,
201633
+ "learning_rate": 8.718029079159936e-06,
201634
+ "loss": 0.3856,
201635
+ "step": 80385
201636
+ },
201637
+ {
201638
+ "epoch": 642.97,
201639
+ "learning_rate": 8.71794830371567e-06,
201640
+ "loss": 0.5971,
201641
+ "step": 80390
201642
+ },
201643
+ {
201644
+ "epoch": 643.0,
201645
+ "eval_loss": 0.493080735206604,
201646
+ "eval_runtime": 41.096,
201647
+ "eval_samples_per_second": 20.464,
201648
+ "eval_steps_per_second": 0.657,
201649
+ "eval_wer": 0.19013722756292653,
201650
+ "step": 80394
201651
+ },
201652
+ {
201653
+ "epoch": 643.01,
201654
+ "learning_rate": 8.717867528271406e-06,
201655
+ "loss": 0.4632,
201656
+ "step": 80395
201657
+ },
201658
+ {
201659
+ "epoch": 643.05,
201660
+ "learning_rate": 8.71778675282714e-06,
201661
+ "loss": 0.3292,
201662
+ "step": 80400
201663
+ },
201664
+ {
201665
+ "epoch": 643.09,
201666
+ "learning_rate": 8.717705977382876e-06,
201667
+ "loss": 0.274,
201668
+ "step": 80405
201669
+ },
201670
+ {
201671
+ "epoch": 643.13,
201672
+ "learning_rate": 8.71762520193861e-06,
201673
+ "loss": 0.3407,
201674
+ "step": 80410
201675
+ },
201676
+ {
201677
+ "epoch": 643.17,
201678
+ "learning_rate": 8.717544426494346e-06,
201679
+ "loss": 0.6008,
201680
+ "step": 80415
201681
+ },
201682
+ {
201683
+ "epoch": 643.21,
201684
+ "learning_rate": 8.71746365105008e-06,
201685
+ "loss": 1.0789,
201686
+ "step": 80420
201687
+ },
201688
+ {
201689
+ "epoch": 643.25,
201690
+ "learning_rate": 8.717382875605816e-06,
201691
+ "loss": 0.3242,
201692
+ "step": 80425
201693
+ },
201694
+ {
201695
+ "epoch": 643.29,
201696
+ "learning_rate": 8.717302100161552e-06,
201697
+ "loss": 0.3146,
201698
+ "step": 80430
201699
+ },
201700
+ {
201701
+ "epoch": 643.33,
201702
+ "learning_rate": 8.717221324717286e-06,
201703
+ "loss": 0.3943,
201704
+ "step": 80435
201705
+ },
201706
+ {
201707
+ "epoch": 643.37,
201708
+ "learning_rate": 8.717140549273022e-06,
201709
+ "loss": 0.548,
201710
+ "step": 80440
201711
+ },
201712
+ {
201713
+ "epoch": 643.41,
201714
+ "learning_rate": 8.717059773828756e-06,
201715
+ "loss": 1.0559,
201716
+ "step": 80445
201717
+ },
201718
+ {
201719
+ "epoch": 643.45,
201720
+ "learning_rate": 8.716978998384492e-06,
201721
+ "loss": 0.2766,
201722
+ "step": 80450
201723
+ },
201724
+ {
201725
+ "epoch": 643.49,
201726
+ "learning_rate": 8.716898222940226e-06,
201727
+ "loss": 0.2528,
201728
+ "step": 80455
201729
+ },
201730
+ {
201731
+ "epoch": 643.53,
201732
+ "learning_rate": 8.716817447495962e-06,
201733
+ "loss": 0.3173,
201734
+ "step": 80460
201735
+ },
201736
+ {
201737
+ "epoch": 643.57,
201738
+ "learning_rate": 8.716736672051696e-06,
201739
+ "loss": 0.513,
201740
+ "step": 80465
201741
+ },
201742
+ {
201743
+ "epoch": 643.61,
201744
+ "learning_rate": 8.716655896607432e-06,
201745
+ "loss": 1.274,
201746
+ "step": 80470
201747
+ },
201748
+ {
201749
+ "epoch": 643.65,
201750
+ "learning_rate": 8.716575121163166e-06,
201751
+ "loss": 0.2854,
201752
+ "step": 80475
201753
+ },
201754
+ {
201755
+ "epoch": 643.69,
201756
+ "learning_rate": 8.716494345718902e-06,
201757
+ "loss": 0.3299,
201758
+ "step": 80480
201759
+ },
201760
+ {
201761
+ "epoch": 643.73,
201762
+ "learning_rate": 8.716413570274638e-06,
201763
+ "loss": 0.3727,
201764
+ "step": 80485
201765
+ },
201766
+ {
201767
+ "epoch": 643.77,
201768
+ "learning_rate": 8.716332794830372e-06,
201769
+ "loss": 0.5196,
201770
+ "step": 80490
201771
+ },
201772
+ {
201773
+ "epoch": 643.81,
201774
+ "learning_rate": 8.716252019386108e-06,
201775
+ "loss": 1.0925,
201776
+ "step": 80495
201777
+ },
201778
+ {
201779
+ "epoch": 643.85,
201780
+ "learning_rate": 8.716171243941842e-06,
201781
+ "loss": 0.3021,
201782
+ "step": 80500
201783
+ },
201784
+ {
201785
+ "epoch": 643.89,
201786
+ "learning_rate": 8.716090468497578e-06,
201787
+ "loss": 0.2862,
201788
+ "step": 80505
201789
+ },
201790
+ {
201791
+ "epoch": 643.93,
201792
+ "learning_rate": 8.716009693053312e-06,
201793
+ "loss": 0.3624,
201794
+ "step": 80510
201795
+ },
201796
+ {
201797
+ "epoch": 643.97,
201798
+ "learning_rate": 8.715928917609048e-06,
201799
+ "loss": 0.6027,
201800
+ "step": 80515
201801
+ },
201802
+ {
201803
+ "epoch": 644.0,
201804
+ "eval_loss": 0.38939520716667175,
201805
+ "eval_runtime": 42.8774,
201806
+ "eval_samples_per_second": 19.591,
201807
+ "eval_steps_per_second": 0.63,
201808
+ "eval_wer": 0.19194484760522496,
201809
+ "step": 80519
201810
+ },
201811
+ {
201812
+ "epoch": 649.01,
201813
+ "learning_rate": 8.715848142164782e-06,
201814
+ "loss": 0.4317,
201815
+ "step": 80520
201816
+ },
201817
+ {
201818
+ "epoch": 649.05,
201819
+ "learning_rate": 8.715767366720518e-06,
201820
+ "loss": 0.2902,
201821
+ "step": 80525
201822
+ },
201823
+ {
201824
+ "epoch": 649.09,
201825
+ "learning_rate": 8.715686591276252e-06,
201826
+ "loss": 0.346,
201827
+ "step": 80530
201828
+ },
201829
+ {
201830
+ "epoch": 649.13,
201831
+ "learning_rate": 8.715605815831988e-06,
201832
+ "loss": 0.358,
201833
+ "step": 80535
201834
+ },
201835
+ {
201836
+ "epoch": 649.17,
201837
+ "learning_rate": 8.715525040387722e-06,
201838
+ "loss": 0.6588,
201839
+ "step": 80540
201840
+ },
201841
+ {
201842
+ "epoch": 649.21,
201843
+ "learning_rate": 8.715444264943458e-06,
201844
+ "loss": 1.2163,
201845
+ "step": 80545
201846
+ },
201847
+ {
201848
+ "epoch": 649.25,
201849
+ "learning_rate": 8.715363489499193e-06,
201850
+ "loss": 0.3163,
201851
+ "step": 80550
201852
+ },
201853
+ {
201854
+ "epoch": 649.29,
201855
+ "learning_rate": 8.715282714054928e-06,
201856
+ "loss": 0.3852,
201857
+ "step": 80555
201858
+ },
201859
+ {
201860
+ "epoch": 649.33,
201861
+ "learning_rate": 8.715201938610663e-06,
201862
+ "loss": 0.3614,
201863
+ "step": 80560
201864
+ },
201865
+ {
201866
+ "epoch": 649.37,
201867
+ "learning_rate": 8.715121163166398e-06,
201868
+ "loss": 0.5485,
201869
+ "step": 80565
201870
+ },
201871
+ {
201872
+ "epoch": 649.41,
201873
+ "learning_rate": 8.715040387722133e-06,
201874
+ "loss": 1.2827,
201875
+ "step": 80570
201876
+ },
201877
+ {
201878
+ "epoch": 649.45,
201879
+ "learning_rate": 8.714959612277867e-06,
201880
+ "loss": 0.3936,
201881
+ "step": 80575
201882
+ },
201883
+ {
201884
+ "epoch": 649.49,
201885
+ "learning_rate": 8.714878836833603e-06,
201886
+ "loss": 0.3177,
201887
+ "step": 80580
201888
+ },
201889
+ {
201890
+ "epoch": 649.53,
201891
+ "learning_rate": 8.714798061389337e-06,
201892
+ "loss": 0.3352,
201893
+ "step": 80585
201894
+ },
201895
+ {
201896
+ "epoch": 649.57,
201897
+ "learning_rate": 8.714717285945073e-06,
201898
+ "loss": 0.5401,
201899
+ "step": 80590
201900
+ },
201901
+ {
201902
+ "epoch": 649.61,
201903
+ "learning_rate": 8.714636510500807e-06,
201904
+ "loss": 1.0671,
201905
+ "step": 80595
201906
+ },
201907
+ {
201908
+ "epoch": 649.65,
201909
+ "learning_rate": 8.714555735056543e-06,
201910
+ "loss": 0.3136,
201911
+ "step": 80600
201912
+ },
201913
+ {
201914
+ "epoch": 649.69,
201915
+ "learning_rate": 8.714474959612279e-06,
201916
+ "loss": 0.3726,
201917
+ "step": 80605
201918
+ },
201919
+ {
201920
+ "epoch": 649.73,
201921
+ "learning_rate": 8.714394184168013e-06,
201922
+ "loss": 0.3793,
201923
+ "step": 80610
201924
+ },
201925
+ {
201926
+ "epoch": 649.77,
201927
+ "learning_rate": 8.714313408723749e-06,
201928
+ "loss": 0.4862,
201929
+ "step": 80615
201930
+ },
201931
+ {
201932
+ "epoch": 649.81,
201933
+ "learning_rate": 8.714232633279483e-06,
201934
+ "loss": 1.2835,
201935
+ "step": 80620
201936
+ },
201937
+ {
201938
+ "epoch": 649.85,
201939
+ "learning_rate": 8.714151857835219e-06,
201940
+ "loss": 0.2867,
201941
+ "step": 80625
201942
+ },
201943
+ {
201944
+ "epoch": 649.9,
201945
+ "learning_rate": 8.714071082390953e-06,
201946
+ "loss": 0.2916,
201947
+ "step": 80630
201948
+ },
201949
+ {
201950
+ "epoch": 649.94,
201951
+ "learning_rate": 8.713990306946689e-06,
201952
+ "loss": 0.4137,
201953
+ "step": 80635
201954
+ },
201955
+ {
201956
+ "epoch": 649.98,
201957
+ "learning_rate": 8.713909531502423e-06,
201958
+ "loss": 0.6571,
201959
+ "step": 80640
201960
+ },
201961
+ {
201962
+ "epoch": 650.0,
201963
+ "eval_loss": 0.4564504027366638,
201964
+ "eval_runtime": 41.7225,
201965
+ "eval_samples_per_second": 20.157,
201966
+ "eval_steps_per_second": 0.647,
201967
+ "eval_wer": 0.19225489487753775,
201968
+ "step": 80643
201969
+ },
201970
+ {
201971
+ "epoch": 650.02,
201972
+ "learning_rate": 8.713828756058159e-06,
201973
+ "loss": 0.3961,
201974
+ "step": 80645
201975
+ },
201976
+ {
201977
+ "epoch": 650.06,
201978
+ "learning_rate": 8.713747980613893e-06,
201979
+ "loss": 0.2932,
201980
+ "step": 80650
201981
+ },
201982
+ {
201983
+ "epoch": 650.1,
201984
+ "learning_rate": 8.713667205169629e-06,
201985
+ "loss": 0.4166,
201986
+ "step": 80655
201987
+ },
201988
+ {
201989
+ "epoch": 650.14,
201990
+ "learning_rate": 8.713586429725365e-06,
201991
+ "loss": 0.3511,
201992
+ "step": 80660
201993
+ },
201994
+ {
201995
+ "epoch": 650.18,
201996
+ "learning_rate": 8.713505654281099e-06,
201997
+ "loss": 0.6076,
201998
+ "step": 80665
201999
+ },
202000
+ {
202001
+ "epoch": 650.22,
202002
+ "learning_rate": 8.713424878836835e-06,
202003
+ "loss": 1.0276,
202004
+ "step": 80670
202005
+ },
202006
+ {
202007
+ "epoch": 650.26,
202008
+ "learning_rate": 8.713344103392569e-06,
202009
+ "loss": 0.3696,
202010
+ "step": 80675
202011
+ },
202012
+ {
202013
+ "epoch": 650.3,
202014
+ "learning_rate": 8.713263327948305e-06,
202015
+ "loss": 0.336,
202016
+ "step": 80680
202017
+ },
202018
+ {
202019
+ "epoch": 650.34,
202020
+ "learning_rate": 8.713182552504039e-06,
202021
+ "loss": 0.3982,
202022
+ "step": 80685
202023
+ },
202024
+ {
202025
+ "epoch": 650.38,
202026
+ "learning_rate": 8.713101777059775e-06,
202027
+ "loss": 0.7392,
202028
+ "step": 80690
202029
+ },
202030
+ {
202031
+ "epoch": 650.42,
202032
+ "learning_rate": 8.713021001615509e-06,
202033
+ "loss": 1.1054,
202034
+ "step": 80695
202035
+ },
202036
+ {
202037
+ "epoch": 650.46,
202038
+ "learning_rate": 8.712940226171245e-06,
202039
+ "loss": 0.3,
202040
+ "step": 80700
202041
+ },
202042
+ {
202043
+ "epoch": 650.5,
202044
+ "learning_rate": 8.712859450726979e-06,
202045
+ "loss": 0.3392,
202046
+ "step": 80705
202047
+ },
202048
+ {
202049
+ "epoch": 650.54,
202050
+ "learning_rate": 8.712778675282715e-06,
202051
+ "loss": 0.3815,
202052
+ "step": 80710
202053
+ },
202054
+ {
202055
+ "epoch": 650.58,
202056
+ "learning_rate": 8.712697899838449e-06,
202057
+ "loss": 0.7745,
202058
+ "step": 80715
202059
+ },
202060
+ {
202061
+ "epoch": 650.62,
202062
+ "learning_rate": 8.712617124394185e-06,
202063
+ "loss": 1.0624,
202064
+ "step": 80720
202065
+ },
202066
+ {
202067
+ "epoch": 650.66,
202068
+ "learning_rate": 8.71253634894992e-06,
202069
+ "loss": 0.3261,
202070
+ "step": 80725
202071
+ },
202072
+ {
202073
+ "epoch": 650.7,
202074
+ "learning_rate": 8.712455573505655e-06,
202075
+ "loss": 0.3151,
202076
+ "step": 80730
202077
+ },
202078
+ {
202079
+ "epoch": 650.74,
202080
+ "learning_rate": 8.71237479806139e-06,
202081
+ "loss": 0.35,
202082
+ "step": 80735
202083
+ },
202084
+ {
202085
+ "epoch": 650.78,
202086
+ "learning_rate": 8.712294022617125e-06,
202087
+ "loss": 0.7034,
202088
+ "step": 80740
202089
+ },
202090
+ {
202091
+ "epoch": 650.82,
202092
+ "learning_rate": 8.71221324717286e-06,
202093
+ "loss": 1.1486,
202094
+ "step": 80745
202095
+ },
202096
+ {
202097
+ "epoch": 650.86,
202098
+ "learning_rate": 8.712132471728595e-06,
202099
+ "loss": 0.3333,
202100
+ "step": 80750
202101
+ },
202102
+ {
202103
+ "epoch": 650.9,
202104
+ "learning_rate": 8.71205169628433e-06,
202105
+ "loss": 0.3279,
202106
+ "step": 80755
202107
+ },
202108
+ {
202109
+ "epoch": 650.94,
202110
+ "learning_rate": 8.711970920840065e-06,
202111
+ "loss": 0.4051,
202112
+ "step": 80760
202113
+ },
202114
+ {
202115
+ "epoch": 650.98,
202116
+ "learning_rate": 8.7118901453958e-06,
202117
+ "loss": 0.6548,
202118
+ "step": 80765
202119
+ },
202120
+ {
202121
+ "epoch": 651.0,
202122
+ "eval_loss": 0.37322500348091125,
202123
+ "eval_runtime": 42.5291,
202124
+ "eval_samples_per_second": 19.775,
202125
+ "eval_steps_per_second": 0.635,
202126
+ "eval_wer": 0.18461987877017455,
202127
+ "step": 80767
202128
+ },
202129
+ {
202130
+ "epoch": 651.02,
202131
+ "learning_rate": 8.711809369951535e-06,
202132
+ "loss": 0.3237,
202133
+ "step": 80770
202134
+ },
202135
+ {
202136
+ "epoch": 651.06,
202137
+ "learning_rate": 8.71172859450727e-06,
202138
+ "loss": 0.3577,
202139
+ "step": 80775
202140
+ },
202141
+ {
202142
+ "epoch": 651.1,
202143
+ "learning_rate": 8.711647819063006e-06,
202144
+ "loss": 0.303,
202145
+ "step": 80780
202146
+ },
202147
+ {
202148
+ "epoch": 651.14,
202149
+ "learning_rate": 8.71156704361874e-06,
202150
+ "loss": 0.3532,
202151
+ "step": 80785
202152
+ },
202153
+ {
202154
+ "epoch": 651.18,
202155
+ "learning_rate": 8.711486268174476e-06,
202156
+ "loss": 0.7385,
202157
+ "step": 80790
202158
+ },
202159
+ {
202160
+ "epoch": 651.22,
202161
+ "learning_rate": 8.71140549273021e-06,
202162
+ "loss": 0.8566,
202163
+ "step": 80795
202164
+ },
202165
+ {
202166
+ "epoch": 651.27,
202167
+ "learning_rate": 8.711324717285946e-06,
202168
+ "loss": 0.3091,
202169
+ "step": 80800
202170
+ },
202171
+ {
202172
+ "epoch": 651.31,
202173
+ "learning_rate": 8.71124394184168e-06,
202174
+ "loss": 0.3321,
202175
+ "step": 80805
202176
+ },
202177
+ {
202178
+ "epoch": 651.35,
202179
+ "learning_rate": 8.711163166397416e-06,
202180
+ "loss": 0.3993,
202181
+ "step": 80810
202182
+ },
202183
+ {
202184
+ "epoch": 651.39,
202185
+ "learning_rate": 8.71108239095315e-06,
202186
+ "loss": 0.8393,
202187
+ "step": 80815
202188
+ },
202189
+ {
202190
+ "epoch": 651.43,
202191
+ "learning_rate": 8.711001615508886e-06,
202192
+ "loss": 0.9883,
202193
+ "step": 80820
202194
+ },
202195
+ {
202196
+ "epoch": 651.47,
202197
+ "learning_rate": 8.71092084006462e-06,
202198
+ "loss": 0.3495,
202199
+ "step": 80825
202200
+ },
202201
+ {
202202
+ "epoch": 651.51,
202203
+ "learning_rate": 8.710840064620356e-06,
202204
+ "loss": 0.3111,
202205
+ "step": 80830
202206
+ },
202207
+ {
202208
+ "epoch": 651.55,
202209
+ "learning_rate": 8.710759289176092e-06,
202210
+ "loss": 0.3887,
202211
+ "step": 80835
202212
+ },
202213
+ {
202214
+ "epoch": 651.59,
202215
+ "learning_rate": 8.710678513731826e-06,
202216
+ "loss": 0.6717,
202217
+ "step": 80840
202218
+ },
202219
+ {
202220
+ "epoch": 651.63,
202221
+ "learning_rate": 8.710597738287562e-06,
202222
+ "loss": 0.9654,
202223
+ "step": 80845
202224
+ },
202225
+ {
202226
+ "epoch": 651.67,
202227
+ "learning_rate": 8.710516962843296e-06,
202228
+ "loss": 0.3719,
202229
+ "step": 80850
202230
+ },
202231
+ {
202232
+ "epoch": 651.71,
202233
+ "learning_rate": 8.710436187399032e-06,
202234
+ "loss": 0.2985,
202235
+ "step": 80855
202236
+ },
202237
+ {
202238
+ "epoch": 651.75,
202239
+ "learning_rate": 8.710355411954766e-06,
202240
+ "loss": 0.3987,
202241
+ "step": 80860
202242
+ },
202243
+ {
202244
+ "epoch": 651.79,
202245
+ "learning_rate": 8.710274636510502e-06,
202246
+ "loss": 0.8482,
202247
+ "step": 80865
202248
+ },
202249
+ {
202250
+ "epoch": 651.83,
202251
+ "learning_rate": 8.710193861066236e-06,
202252
+ "loss": 0.8805,
202253
+ "step": 80870
202254
+ },
202255
+ {
202256
+ "epoch": 651.87,
202257
+ "learning_rate": 8.710113085621972e-06,
202258
+ "loss": 0.2904,
202259
+ "step": 80875
202260
+ },
202261
+ {
202262
+ "epoch": 651.91,
202263
+ "learning_rate": 8.710032310177706e-06,
202264
+ "loss": 0.3254,
202265
+ "step": 80880
202266
+ },
202267
+ {
202268
+ "epoch": 651.95,
202269
+ "learning_rate": 8.709951534733442e-06,
202270
+ "loss": 0.4833,
202271
+ "step": 80885
202272
+ },
202273
+ {
202274
+ "epoch": 651.99,
202275
+ "learning_rate": 8.709870759289178e-06,
202276
+ "loss": 0.8421,
202277
+ "step": 80890
202278
+ },
202279
+ {
202280
+ "epoch": 652.0,
202281
+ "eval_loss": 0.4140271544456482,
202282
+ "eval_runtime": 42.7669,
202283
+ "eval_samples_per_second": 19.665,
202284
+ "eval_steps_per_second": 0.631,
202285
+ "eval_wer": 0.19266525144572139,
202286
+ "step": 80891
202287
  }
202288
  ],
202289
+ "max_steps": 620000,
202290
  "num_train_epochs": 5000,
202291
+ "total_flos": 2.276386500512849e+20,
202292
  "trial_name": null,
202293
  "trial_params": null
202294
  }
model-bin/finetune/base/{checkpoint-80269 β†’ checkpoint-80891}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629904128.5656571/events.out.tfevents.1629904128.7e498afd5545.7645.65 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:807060dd41c3169495b1e22f76cdb8737f06b98c46114a0759447886e27f200a
3
+ size 4194
model-bin/finetune/base/log/1629904601.0615444/events.out.tfevents.1629904601.7e498afd5545.7645.67 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bffb35c16cae0781fdd54b961b0ca24b32955827b3d412703ba15425873aac3
3
+ size 4194
model-bin/finetune/base/log/1629905064.9216402/events.out.tfevents.1629905064.7e498afd5545.7645.69 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c318e0ec29741b86863d7d0587626020ffe179765e364ef63f2b913bdc9686b
3
+ size 4194
model-bin/finetune/base/log/1629905527.7759078/events.out.tfevents.1629905527.7e498afd5545.7645.71 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16f11a624cab250697a2b34acc4089cb93135b82a6be8b62b6afd2ecafc3ef72
3
+ size 4194
model-bin/finetune/base/log/1629905996.4959083/events.out.tfevents.1629905996.7e498afd5545.7645.73 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72686a7940c11667371b7801acaac5d86f3ce96ab29da3803f18b703853c2200
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629904128.7e498afd5545.7645.64 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5889f431c237f270bd245751e9d85810688e97f44d112e3715b553bdcbee2d2a
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629904601.7e498afd5545.7645.66 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:175b6cd71599f79a269596e913f046cdbfa74bccad3cd822ae0f774cb5e7a779
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629905064.7e498afd5545.7645.68 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a0354dcb77e0dd9b9dc16a1a824a4ec5b4ebfcf7e697b889efae36b88cf8ffb
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629905527.7e498afd5545.7645.70 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:274058007081c38d153a280cb49c8c59516d7e14d2e8bb9a3a33da7129bc5425
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629905996.7e498afd5545.7645.72 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:185b3e42b764871d1a4fde3c95e21fe832306f6d5af8fcaa06896db11a31e1b8
3
+ size 8622