Check commited on
Commit
dc782e2
Β·
1 Parent(s): abf8b95

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/trainer_state.json +959 -5
  9. model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629801593.3967817/events.out.tfevents.1629801593.c435e1c5ee04.920.111 +3 -0
  11. model-bin/finetune/base/log/1629802228.319492/events.out.tfevents.1629802228.c435e1c5ee04.920.113 +3 -0
  12. model-bin/finetune/base/log/1629802982.729967/events.out.tfevents.1629802982.c435e1c5ee04.920.115 +3 -0
  13. model-bin/finetune/base/log/1629803633.8011563/events.out.tfevents.1629803633.c435e1c5ee04.920.117 +3 -0
  14. model-bin/finetune/base/log/1629804285.3739185/events.out.tfevents.1629804285.c435e1c5ee04.920.119 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629801593.c435e1c5ee04.920.110 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629802228.c435e1c5ee04.920.112 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629802982.c435e1c5ee04.920.114 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629803633.c435e1c5ee04.920.116 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629804285.c435e1c5ee04.920.118 +3 -0
model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:54d742d39c8446eace3eb84f293d54d3c92367bac539b2c6aac2610c3b5cd216
3
  size 722165009
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:353d82c12d6735954a514be77e06ca9d192330f66e138929496fa6ba81bcfc02
3
  size 722165009
model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b5929eb300a5e7bc9be64e52e4b06762401ab913925a2a1b0e17dcaad3e94476
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fce8f21fad01f8d884828fbcc6967cef29c49b990d1f4d92914cee1952071dc3
3
  size 377909911
model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6c1a87fab1e51e42d80278684e7ba62284a23bae14afa60f95e959e29d7d01e1
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0eeb08b6e206cc952258c7cd42ad2583b0dbe57250380cbabcfa01e167250da1
3
  size 14503
model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:789e2be1e3c06a3e2bb8e0d942a5bad1e52db387c9c9778d8d055b91e46a0dc3
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfc3a5c9b694edc68b9a3d30779c3730190586b5fec21d95cd916c9a9982b328
3
  size 559
model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d917b2672d75e0efb76bc510eaa2d3141b6f755a5d5799309de61ece06544b7f
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e510737aa08b522929d4f008035f55ac4839684a993beb48faf03dd344ad37fd
3
  size 623
model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
- "best_metric": 0.18631571186315712,
3
- "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-55501",
4
- "epoch": 488.0,
5
- "global_step": 60604,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -176469,11 +176469,965 @@
176469
  "eval_steps_per_second": 0.692,
176470
  "eval_wer": 0.19932432432432431,
176471
  "step": 60604
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
176472
  }
176473
  ],
176474
  "max_steps": 620000,
176475
  "num_train_epochs": 5000,
176476
- "total_flos": 1.7056102466424242e+20,
176477
  "trial_name": null,
176478
  "trial_params": null
176479
  }
 
1
  {
2
+ "best_metric": 0.18588425381903642,
3
+ "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-60729",
4
+ "epoch": 494.0,
5
+ "global_step": 61350,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
176469
  "eval_steps_per_second": 0.692,
176470
  "eval_wer": 0.19932432432432431,
176471
  "step": 60604
176472
+ },
176473
+ {
176474
+ "epoch": 484.01,
176475
+ "learning_rate": 9.045160256410256e-06,
176476
+ "loss": 0.395,
176477
+ "step": 60605
176478
+ },
176479
+ {
176480
+ "epoch": 484.05,
176481
+ "learning_rate": 9.04508012820513e-06,
176482
+ "loss": 0.3529,
176483
+ "step": 60610
176484
+ },
176485
+ {
176486
+ "epoch": 484.09,
176487
+ "learning_rate": 9.045e-06,
176488
+ "loss": 0.3603,
176489
+ "step": 60615
176490
+ },
176491
+ {
176492
+ "epoch": 484.13,
176493
+ "learning_rate": 9.044919871794872e-06,
176494
+ "loss": 0.3763,
176495
+ "step": 60620
176496
+ },
176497
+ {
176498
+ "epoch": 484.17,
176499
+ "learning_rate": 9.044839743589743e-06,
176500
+ "loss": 0.5586,
176501
+ "step": 60625
176502
+ },
176503
+ {
176504
+ "epoch": 484.21,
176505
+ "learning_rate": 9.044759615384616e-06,
176506
+ "loss": 1.1159,
176507
+ "step": 60630
176508
+ },
176509
+ {
176510
+ "epoch": 484.25,
176511
+ "learning_rate": 9.044679487179488e-06,
176512
+ "loss": 0.3736,
176513
+ "step": 60635
176514
+ },
176515
+ {
176516
+ "epoch": 484.29,
176517
+ "learning_rate": 9.044599358974359e-06,
176518
+ "loss": 0.289,
176519
+ "step": 60640
176520
+ },
176521
+ {
176522
+ "epoch": 484.33,
176523
+ "learning_rate": 9.044519230769232e-06,
176524
+ "loss": 0.3933,
176525
+ "step": 60645
176526
+ },
176527
+ {
176528
+ "epoch": 484.37,
176529
+ "learning_rate": 9.044439102564104e-06,
176530
+ "loss": 0.5288,
176531
+ "step": 60650
176532
+ },
176533
+ {
176534
+ "epoch": 484.41,
176535
+ "learning_rate": 9.044358974358975e-06,
176536
+ "loss": 1.1191,
176537
+ "step": 60655
176538
+ },
176539
+ {
176540
+ "epoch": 484.45,
176541
+ "learning_rate": 9.044278846153846e-06,
176542
+ "loss": 0.3084,
176543
+ "step": 60660
176544
+ },
176545
+ {
176546
+ "epoch": 484.49,
176547
+ "learning_rate": 9.04419871794872e-06,
176548
+ "loss": 0.3426,
176549
+ "step": 60665
176550
+ },
176551
+ {
176552
+ "epoch": 484.53,
176553
+ "learning_rate": 9.04411858974359e-06,
176554
+ "loss": 0.3757,
176555
+ "step": 60670
176556
+ },
176557
+ {
176558
+ "epoch": 484.57,
176559
+ "learning_rate": 9.044038461538462e-06,
176560
+ "loss": 0.5491,
176561
+ "step": 60675
176562
+ },
176563
+ {
176564
+ "epoch": 484.61,
176565
+ "learning_rate": 9.043958333333333e-06,
176566
+ "loss": 1.1088,
176567
+ "step": 60680
176568
+ },
176569
+ {
176570
+ "epoch": 484.65,
176571
+ "learning_rate": 9.043878205128206e-06,
176572
+ "loss": 0.3394,
176573
+ "step": 60685
176574
+ },
176575
+ {
176576
+ "epoch": 484.69,
176577
+ "learning_rate": 9.043798076923078e-06,
176578
+ "loss": 0.3385,
176579
+ "step": 60690
176580
+ },
176581
+ {
176582
+ "epoch": 484.73,
176583
+ "learning_rate": 9.04371794871795e-06,
176584
+ "loss": 0.3283,
176585
+ "step": 60695
176586
+ },
176587
+ {
176588
+ "epoch": 484.77,
176589
+ "learning_rate": 9.043637820512822e-06,
176590
+ "loss": 0.5318,
176591
+ "step": 60700
176592
+ },
176593
+ {
176594
+ "epoch": 484.81,
176595
+ "learning_rate": 9.043557692307694e-06,
176596
+ "loss": 1.11,
176597
+ "step": 60705
176598
+ },
176599
+ {
176600
+ "epoch": 484.85,
176601
+ "learning_rate": 9.043477564102565e-06,
176602
+ "loss": 0.3747,
176603
+ "step": 60710
176604
+ },
176605
+ {
176606
+ "epoch": 484.89,
176607
+ "learning_rate": 9.043397435897436e-06,
176608
+ "loss": 0.3217,
176609
+ "step": 60715
176610
+ },
176611
+ {
176612
+ "epoch": 484.93,
176613
+ "learning_rate": 9.04331730769231e-06,
176614
+ "loss": 0.3629,
176615
+ "step": 60720
176616
+ },
176617
+ {
176618
+ "epoch": 484.97,
176619
+ "learning_rate": 9.043237179487179e-06,
176620
+ "loss": 0.6506,
176621
+ "step": 60725
176622
+ },
176623
+ {
176624
+ "epoch": 485.0,
176625
+ "eval_loss": 0.40527278184890747,
176626
+ "eval_runtime": 38.5522,
176627
+ "eval_samples_per_second": 21.737,
176628
+ "eval_steps_per_second": 0.7,
176629
+ "eval_wer": 0.18588425381903642,
176630
+ "step": 60729
176631
+ },
176632
+ {
176633
+ "epoch": 489.01,
176634
+ "learning_rate": 9.043157051282052e-06,
176635
+ "loss": 0.4361,
176636
+ "step": 60730
176637
+ },
176638
+ {
176639
+ "epoch": 489.05,
176640
+ "learning_rate": 9.043076923076923e-06,
176641
+ "loss": 0.3467,
176642
+ "step": 60735
176643
+ },
176644
+ {
176645
+ "epoch": 489.09,
176646
+ "learning_rate": 9.042996794871795e-06,
176647
+ "loss": 0.2973,
176648
+ "step": 60740
176649
+ },
176650
+ {
176651
+ "epoch": 489.13,
176652
+ "learning_rate": 9.042916666666668e-06,
176653
+ "loss": 0.3757,
176654
+ "step": 60745
176655
+ },
176656
+ {
176657
+ "epoch": 489.17,
176658
+ "learning_rate": 9.04283653846154e-06,
176659
+ "loss": 0.63,
176660
+ "step": 60750
176661
+ },
176662
+ {
176663
+ "epoch": 489.21,
176664
+ "learning_rate": 9.04275641025641e-06,
176665
+ "loss": 1.0862,
176666
+ "step": 60755
176667
+ },
176668
+ {
176669
+ "epoch": 489.25,
176670
+ "learning_rate": 9.042676282051282e-06,
176671
+ "loss": 0.3307,
176672
+ "step": 60760
176673
+ },
176674
+ {
176675
+ "epoch": 489.29,
176676
+ "learning_rate": 9.042596153846155e-06,
176677
+ "loss": 0.3111,
176678
+ "step": 60765
176679
+ },
176680
+ {
176681
+ "epoch": 489.33,
176682
+ "learning_rate": 9.042516025641026e-06,
176683
+ "loss": 0.3204,
176684
+ "step": 60770
176685
+ },
176686
+ {
176687
+ "epoch": 489.37,
176688
+ "learning_rate": 9.042435897435898e-06,
176689
+ "loss": 0.7555,
176690
+ "step": 60775
176691
+ },
176692
+ {
176693
+ "epoch": 489.41,
176694
+ "learning_rate": 9.042355769230769e-06,
176695
+ "loss": 1.2464,
176696
+ "step": 60780
176697
+ },
176698
+ {
176699
+ "epoch": 489.45,
176700
+ "learning_rate": 9.042275641025642e-06,
176701
+ "loss": 0.3736,
176702
+ "step": 60785
176703
+ },
176704
+ {
176705
+ "epoch": 489.49,
176706
+ "learning_rate": 9.042195512820513e-06,
176707
+ "loss": 0.343,
176708
+ "step": 60790
176709
+ },
176710
+ {
176711
+ "epoch": 489.53,
176712
+ "learning_rate": 9.042115384615385e-06,
176713
+ "loss": 0.4481,
176714
+ "step": 60795
176715
+ },
176716
+ {
176717
+ "epoch": 489.57,
176718
+ "learning_rate": 9.042035256410258e-06,
176719
+ "loss": 0.5257,
176720
+ "step": 60800
176721
+ },
176722
+ {
176723
+ "epoch": 489.61,
176724
+ "learning_rate": 9.04195512820513e-06,
176725
+ "loss": 1.198,
176726
+ "step": 60805
176727
+ },
176728
+ {
176729
+ "epoch": 489.65,
176730
+ "learning_rate": 9.041875e-06,
176731
+ "loss": 0.311,
176732
+ "step": 60810
176733
+ },
176734
+ {
176735
+ "epoch": 489.69,
176736
+ "learning_rate": 9.041794871794872e-06,
176737
+ "loss": 0.3372,
176738
+ "step": 60815
176739
+ },
176740
+ {
176741
+ "epoch": 489.73,
176742
+ "learning_rate": 9.041714743589745e-06,
176743
+ "loss": 0.3814,
176744
+ "step": 60820
176745
+ },
176746
+ {
176747
+ "epoch": 489.77,
176748
+ "learning_rate": 9.041634615384616e-06,
176749
+ "loss": 0.5157,
176750
+ "step": 60825
176751
+ },
176752
+ {
176753
+ "epoch": 489.81,
176754
+ "learning_rate": 9.041554487179488e-06,
176755
+ "loss": 1.1918,
176756
+ "step": 60830
176757
+ },
176758
+ {
176759
+ "epoch": 489.85,
176760
+ "learning_rate": 9.041474358974359e-06,
176761
+ "loss": 0.3764,
176762
+ "step": 60835
176763
+ },
176764
+ {
176765
+ "epoch": 489.9,
176766
+ "learning_rate": 9.041394230769232e-06,
176767
+ "loss": 0.3416,
176768
+ "step": 60840
176769
+ },
176770
+ {
176771
+ "epoch": 489.94,
176772
+ "learning_rate": 9.041314102564103e-06,
176773
+ "loss": 0.4076,
176774
+ "step": 60845
176775
+ },
176776
+ {
176777
+ "epoch": 489.98,
176778
+ "learning_rate": 9.041233974358975e-06,
176779
+ "loss": 0.7,
176780
+ "step": 60850
176781
+ },
176782
+ {
176783
+ "epoch": 490.0,
176784
+ "eval_loss": 0.4080774784088135,
176785
+ "eval_runtime": 40.3984,
176786
+ "eval_samples_per_second": 20.743,
176787
+ "eval_steps_per_second": 0.668,
176788
+ "eval_wer": 0.19368847712796758,
176789
+ "step": 60853
176790
+ },
176791
+ {
176792
+ "epoch": 486.02,
176793
+ "learning_rate": 9.041153846153848e-06,
176794
+ "loss": 0.4242,
176795
+ "step": 60855
176796
+ },
176797
+ {
176798
+ "epoch": 486.06,
176799
+ "learning_rate": 9.04107371794872e-06,
176800
+ "loss": 0.3067,
176801
+ "step": 60860
176802
+ },
176803
+ {
176804
+ "epoch": 486.1,
176805
+ "learning_rate": 9.04099358974359e-06,
176806
+ "loss": 0.4052,
176807
+ "step": 60865
176808
+ },
176809
+ {
176810
+ "epoch": 486.14,
176811
+ "learning_rate": 9.040913461538462e-06,
176812
+ "loss": 0.4187,
176813
+ "step": 60870
176814
+ },
176815
+ {
176816
+ "epoch": 486.18,
176817
+ "learning_rate": 9.040833333333335e-06,
176818
+ "loss": 0.6364,
176819
+ "step": 60875
176820
+ },
176821
+ {
176822
+ "epoch": 486.22,
176823
+ "learning_rate": 9.040753205128205e-06,
176824
+ "loss": 1.0251,
176825
+ "step": 60880
176826
+ },
176827
+ {
176828
+ "epoch": 486.26,
176829
+ "learning_rate": 9.040673076923078e-06,
176830
+ "loss": 0.3019,
176831
+ "step": 60885
176832
+ },
176833
+ {
176834
+ "epoch": 486.3,
176835
+ "learning_rate": 9.040592948717949e-06,
176836
+ "loss": 0.366,
176837
+ "step": 60890
176838
+ },
176839
+ {
176840
+ "epoch": 486.34,
176841
+ "learning_rate": 9.04051282051282e-06,
176842
+ "loss": 0.3735,
176843
+ "step": 60895
176844
+ },
176845
+ {
176846
+ "epoch": 486.38,
176847
+ "learning_rate": 9.040432692307693e-06,
176848
+ "loss": 0.6687,
176849
+ "step": 60900
176850
+ },
176851
+ {
176852
+ "epoch": 486.42,
176853
+ "learning_rate": 9.040352564102565e-06,
176854
+ "loss": 1.0873,
176855
+ "step": 60905
176856
+ },
176857
+ {
176858
+ "epoch": 486.46,
176859
+ "learning_rate": 9.040272435897436e-06,
176860
+ "loss": 0.2771,
176861
+ "step": 60910
176862
+ },
176863
+ {
176864
+ "epoch": 486.5,
176865
+ "learning_rate": 9.040192307692308e-06,
176866
+ "loss": 0.3668,
176867
+ "step": 60915
176868
+ },
176869
+ {
176870
+ "epoch": 486.54,
176871
+ "learning_rate": 9.04011217948718e-06,
176872
+ "loss": 0.3356,
176873
+ "step": 60920
176874
+ },
176875
+ {
176876
+ "epoch": 486.58,
176877
+ "learning_rate": 9.040032051282052e-06,
176878
+ "loss": 0.6527,
176879
+ "step": 60925
176880
+ },
176881
+ {
176882
+ "epoch": 486.62,
176883
+ "learning_rate": 9.039951923076923e-06,
176884
+ "loss": 0.9712,
176885
+ "step": 60930
176886
+ },
176887
+ {
176888
+ "epoch": 486.66,
176889
+ "learning_rate": 9.039871794871795e-06,
176890
+ "loss": 0.3367,
176891
+ "step": 60935
176892
+ },
176893
+ {
176894
+ "epoch": 486.7,
176895
+ "learning_rate": 9.039791666666668e-06,
176896
+ "loss": 0.3588,
176897
+ "step": 60940
176898
+ },
176899
+ {
176900
+ "epoch": 486.74,
176901
+ "learning_rate": 9.039711538461539e-06,
176902
+ "loss": 0.3842,
176903
+ "step": 60945
176904
+ },
176905
+ {
176906
+ "epoch": 486.78,
176907
+ "learning_rate": 9.03963141025641e-06,
176908
+ "loss": 0.6641,
176909
+ "step": 60950
176910
+ },
176911
+ {
176912
+ "epoch": 486.82,
176913
+ "learning_rate": 9.039551282051284e-06,
176914
+ "loss": 1.1598,
176915
+ "step": 60955
176916
+ },
176917
+ {
176918
+ "epoch": 486.86,
176919
+ "learning_rate": 9.039471153846155e-06,
176920
+ "loss": 0.3243,
176921
+ "step": 60960
176922
+ },
176923
+ {
176924
+ "epoch": 486.9,
176925
+ "learning_rate": 9.039391025641026e-06,
176926
+ "loss": 0.3483,
176927
+ "step": 60965
176928
+ },
176929
+ {
176930
+ "epoch": 486.94,
176931
+ "learning_rate": 9.039310897435898e-06,
176932
+ "loss": 0.4508,
176933
+ "step": 60970
176934
+ },
176935
+ {
176936
+ "epoch": 486.98,
176937
+ "learning_rate": 9.03923076923077e-06,
176938
+ "loss": 0.9019,
176939
+ "step": 60975
176940
+ },
176941
+ {
176942
+ "epoch": 487.0,
176943
+ "eval_loss": 0.3954157531261444,
176944
+ "eval_runtime": 40.0837,
176945
+ "eval_samples_per_second": 20.906,
176946
+ "eval_steps_per_second": 0.674,
176947
+ "eval_wer": 0.19892081085022606,
176948
+ "step": 60978
176949
+ },
176950
+ {
176951
+ "epoch": 491.02,
176952
+ "learning_rate": 9.039150641025642e-06,
176953
+ "loss": 0.341,
176954
+ "step": 60980
176955
+ },
176956
+ {
176957
+ "epoch": 491.06,
176958
+ "learning_rate": 9.039070512820513e-06,
176959
+ "loss": 0.3259,
176960
+ "step": 60985
176961
+ },
176962
+ {
176963
+ "epoch": 491.1,
176964
+ "learning_rate": 9.038990384615385e-06,
176965
+ "loss": 0.3462,
176966
+ "step": 60990
176967
+ },
176968
+ {
176969
+ "epoch": 491.14,
176970
+ "learning_rate": 9.038910256410258e-06,
176971
+ "loss": 0.3834,
176972
+ "step": 60995
176973
+ },
176974
+ {
176975
+ "epoch": 491.18,
176976
+ "learning_rate": 9.038830128205129e-06,
176977
+ "loss": 0.6381,
176978
+ "step": 61000
176979
+ },
176980
+ {
176981
+ "epoch": 491.22,
176982
+ "learning_rate": 9.03875e-06,
176983
+ "loss": 1.1421,
176984
+ "step": 61005
176985
+ },
176986
+ {
176987
+ "epoch": 491.26,
176988
+ "learning_rate": 9.038669871794874e-06,
176989
+ "loss": 0.3287,
176990
+ "step": 61010
176991
+ },
176992
+ {
176993
+ "epoch": 491.3,
176994
+ "learning_rate": 9.038589743589745e-06,
176995
+ "loss": 0.3466,
176996
+ "step": 61015
176997
+ },
176998
+ {
176999
+ "epoch": 491.34,
177000
+ "learning_rate": 9.038509615384616e-06,
177001
+ "loss": 0.4107,
177002
+ "step": 61020
177003
+ },
177004
+ {
177005
+ "epoch": 491.38,
177006
+ "learning_rate": 9.038429487179488e-06,
177007
+ "loss": 0.651,
177008
+ "step": 61025
177009
+ },
177010
+ {
177011
+ "epoch": 491.42,
177012
+ "learning_rate": 9.03834935897436e-06,
177013
+ "loss": 1.0088,
177014
+ "step": 61030
177015
+ },
177016
+ {
177017
+ "epoch": 491.46,
177018
+ "learning_rate": 9.03826923076923e-06,
177019
+ "loss": 0.3308,
177020
+ "step": 61035
177021
+ },
177022
+ {
177023
+ "epoch": 491.5,
177024
+ "learning_rate": 9.038189102564103e-06,
177025
+ "loss": 0.3197,
177026
+ "step": 61040
177027
+ },
177028
+ {
177029
+ "epoch": 491.54,
177030
+ "learning_rate": 9.038108974358976e-06,
177031
+ "loss": 0.3684,
177032
+ "step": 61045
177033
+ },
177034
+ {
177035
+ "epoch": 491.58,
177036
+ "learning_rate": 9.038028846153846e-06,
177037
+ "loss": 0.6528,
177038
+ "step": 61050
177039
+ },
177040
+ {
177041
+ "epoch": 491.62,
177042
+ "learning_rate": 9.03794871794872e-06,
177043
+ "loss": 1.0219,
177044
+ "step": 61055
177045
+ },
177046
+ {
177047
+ "epoch": 491.66,
177048
+ "learning_rate": 9.03786858974359e-06,
177049
+ "loss": 0.4409,
177050
+ "step": 61060
177051
+ },
177052
+ {
177053
+ "epoch": 491.7,
177054
+ "learning_rate": 9.037788461538462e-06,
177055
+ "loss": 0.3411,
177056
+ "step": 61065
177057
+ },
177058
+ {
177059
+ "epoch": 491.74,
177060
+ "learning_rate": 9.037708333333333e-06,
177061
+ "loss": 0.415,
177062
+ "step": 61070
177063
+ },
177064
+ {
177065
+ "epoch": 491.78,
177066
+ "learning_rate": 9.037628205128206e-06,
177067
+ "loss": 0.6908,
177068
+ "step": 61075
177069
+ },
177070
+ {
177071
+ "epoch": 491.82,
177072
+ "learning_rate": 9.037548076923078e-06,
177073
+ "loss": 1.0197,
177074
+ "step": 61080
177075
+ },
177076
+ {
177077
+ "epoch": 491.86,
177078
+ "learning_rate": 9.037467948717949e-06,
177079
+ "loss": 0.3332,
177080
+ "step": 61085
177081
+ },
177082
+ {
177083
+ "epoch": 491.9,
177084
+ "learning_rate": 9.03738782051282e-06,
177085
+ "loss": 0.3519,
177086
+ "step": 61090
177087
+ },
177088
+ {
177089
+ "epoch": 491.94,
177090
+ "learning_rate": 9.037307692307693e-06,
177091
+ "loss": 0.3885,
177092
+ "step": 61095
177093
+ },
177094
+ {
177095
+ "epoch": 491.98,
177096
+ "learning_rate": 9.037227564102565e-06,
177097
+ "loss": 0.6986,
177098
+ "step": 61100
177099
+ },
177100
+ {
177101
+ "epoch": 492.0,
177102
+ "eval_loss": 0.4045012295246124,
177103
+ "eval_runtime": 38.0364,
177104
+ "eval_samples_per_second": 22.216,
177105
+ "eval_steps_per_second": 0.71,
177106
+ "eval_wer": 0.1972527878295547,
177107
+ "step": 61102
177108
+ },
177109
+ {
177110
+ "epoch": 492.02,
177111
+ "learning_rate": 9.037147435897436e-06,
177112
+ "loss": 0.3484,
177113
+ "step": 61105
177114
+ },
177115
+ {
177116
+ "epoch": 492.06,
177117
+ "learning_rate": 9.03706730769231e-06,
177118
+ "loss": 0.29,
177119
+ "step": 61110
177120
+ },
177121
+ {
177122
+ "epoch": 492.1,
177123
+ "learning_rate": 9.03698717948718e-06,
177124
+ "loss": 0.3342,
177125
+ "step": 61115
177126
+ },
177127
+ {
177128
+ "epoch": 492.14,
177129
+ "learning_rate": 9.036907051282052e-06,
177130
+ "loss": 0.4317,
177131
+ "step": 61120
177132
+ },
177133
+ {
177134
+ "epoch": 492.18,
177135
+ "learning_rate": 9.036826923076923e-06,
177136
+ "loss": 0.6594,
177137
+ "step": 61125
177138
+ },
177139
+ {
177140
+ "epoch": 492.22,
177141
+ "learning_rate": 9.036746794871796e-06,
177142
+ "loss": 1.2607,
177143
+ "step": 61130
177144
+ },
177145
+ {
177146
+ "epoch": 492.27,
177147
+ "learning_rate": 9.036666666666668e-06,
177148
+ "loss": 0.529,
177149
+ "step": 61135
177150
+ },
177151
+ {
177152
+ "epoch": 492.31,
177153
+ "learning_rate": 9.036586538461539e-06,
177154
+ "loss": 0.3166,
177155
+ "step": 61140
177156
+ },
177157
+ {
177158
+ "epoch": 492.35,
177159
+ "learning_rate": 9.036506410256412e-06,
177160
+ "loss": 0.4587,
177161
+ "step": 61145
177162
+ },
177163
+ {
177164
+ "epoch": 492.39,
177165
+ "learning_rate": 9.036426282051283e-06,
177166
+ "loss": 0.7595,
177167
+ "step": 61150
177168
+ },
177169
+ {
177170
+ "epoch": 492.43,
177171
+ "learning_rate": 9.036346153846155e-06,
177172
+ "loss": 0.799,
177173
+ "step": 61155
177174
+ },
177175
+ {
177176
+ "epoch": 492.47,
177177
+ "learning_rate": 9.036266025641026e-06,
177178
+ "loss": 0.2994,
177179
+ "step": 61160
177180
+ },
177181
+ {
177182
+ "epoch": 492.51,
177183
+ "learning_rate": 9.0361858974359e-06,
177184
+ "loss": 0.379,
177185
+ "step": 61165
177186
+ },
177187
+ {
177188
+ "epoch": 492.55,
177189
+ "learning_rate": 9.036105769230769e-06,
177190
+ "loss": 0.4963,
177191
+ "step": 61170
177192
+ },
177193
+ {
177194
+ "epoch": 492.59,
177195
+ "learning_rate": 9.036025641025642e-06,
177196
+ "loss": 0.9508,
177197
+ "step": 61175
177198
+ },
177199
+ {
177200
+ "epoch": 492.63,
177201
+ "learning_rate": 9.035945512820513e-06,
177202
+ "loss": 0.9842,
177203
+ "step": 61180
177204
+ },
177205
+ {
177206
+ "epoch": 492.67,
177207
+ "learning_rate": 9.035865384615385e-06,
177208
+ "loss": 0.3276,
177209
+ "step": 61185
177210
+ },
177211
+ {
177212
+ "epoch": 492.71,
177213
+ "learning_rate": 9.035785256410256e-06,
177214
+ "loss": 0.323,
177215
+ "step": 61190
177216
+ },
177217
+ {
177218
+ "epoch": 492.75,
177219
+ "learning_rate": 9.035705128205129e-06,
177220
+ "loss": 0.3674,
177221
+ "step": 61195
177222
+ },
177223
+ {
177224
+ "epoch": 492.79,
177225
+ "learning_rate": 9.035625e-06,
177226
+ "loss": 0.8951,
177227
+ "step": 61200
177228
+ },
177229
+ {
177230
+ "epoch": 492.83,
177231
+ "learning_rate": 9.035544871794872e-06,
177232
+ "loss": 0.9872,
177233
+ "step": 61205
177234
+ },
177235
+ {
177236
+ "epoch": 492.87,
177237
+ "learning_rate": 9.035464743589745e-06,
177238
+ "loss": 0.3025,
177239
+ "step": 61210
177240
+ },
177241
+ {
177242
+ "epoch": 492.91,
177243
+ "learning_rate": 9.035384615384616e-06,
177244
+ "loss": 0.3818,
177245
+ "step": 61215
177246
+ },
177247
+ {
177248
+ "epoch": 492.95,
177249
+ "learning_rate": 9.035304487179488e-06,
177250
+ "loss": 0.4108,
177251
+ "step": 61220
177252
+ },
177253
+ {
177254
+ "epoch": 492.99,
177255
+ "learning_rate": 9.035224358974359e-06,
177256
+ "loss": 0.8702,
177257
+ "step": 61225
177258
+ },
177259
+ {
177260
+ "epoch": 493.0,
177261
+ "eval_loss": 0.41527244448661804,
177262
+ "eval_runtime": 41.5752,
177263
+ "eval_samples_per_second": 20.325,
177264
+ "eval_steps_per_second": 0.649,
177265
+ "eval_wer": 0.19053151321521927,
177266
+ "step": 61226
177267
+ },
177268
+ {
177269
+ "epoch": 493.03,
177270
+ "learning_rate": 9.035144230769232e-06,
177271
+ "loss": 0.403,
177272
+ "step": 61230
177273
+ },
177274
+ {
177275
+ "epoch": 493.07,
177276
+ "learning_rate": 9.035064102564103e-06,
177277
+ "loss": 0.3474,
177278
+ "step": 61235
177279
+ },
177280
+ {
177281
+ "epoch": 493.11,
177282
+ "learning_rate": 9.034983974358975e-06,
177283
+ "loss": 0.3855,
177284
+ "step": 61240
177285
+ },
177286
+ {
177287
+ "epoch": 493.15,
177288
+ "learning_rate": 9.034903846153848e-06,
177289
+ "loss": 0.4023,
177290
+ "step": 61245
177291
+ },
177292
+ {
177293
+ "epoch": 493.19,
177294
+ "learning_rate": 9.034823717948719e-06,
177295
+ "loss": 1.1158,
177296
+ "step": 61250
177297
+ },
177298
+ {
177299
+ "epoch": 493.23,
177300
+ "learning_rate": 9.03474358974359e-06,
177301
+ "loss": 0.7991,
177302
+ "step": 61255
177303
+ },
177304
+ {
177305
+ "epoch": 493.27,
177306
+ "learning_rate": 9.034663461538462e-06,
177307
+ "loss": 0.3647,
177308
+ "step": 61260
177309
+ },
177310
+ {
177311
+ "epoch": 493.31,
177312
+ "learning_rate": 9.034583333333335e-06,
177313
+ "loss": 0.342,
177314
+ "step": 61265
177315
+ },
177316
+ {
177317
+ "epoch": 493.35,
177318
+ "learning_rate": 9.034503205128206e-06,
177319
+ "loss": 0.4179,
177320
+ "step": 61270
177321
+ },
177322
+ {
177323
+ "epoch": 493.4,
177324
+ "learning_rate": 9.034423076923078e-06,
177325
+ "loss": 0.968,
177326
+ "step": 61275
177327
+ },
177328
+ {
177329
+ "epoch": 493.44,
177330
+ "learning_rate": 9.034342948717949e-06,
177331
+ "loss": 0.6878,
177332
+ "step": 61280
177333
+ },
177334
+ {
177335
+ "epoch": 493.48,
177336
+ "learning_rate": 9.034262820512822e-06,
177337
+ "loss": 0.3062,
177338
+ "step": 61285
177339
+ },
177340
+ {
177341
+ "epoch": 493.52,
177342
+ "learning_rate": 9.034182692307693e-06,
177343
+ "loss": 0.2974,
177344
+ "step": 61290
177345
+ },
177346
+ {
177347
+ "epoch": 493.56,
177348
+ "learning_rate": 9.034102564102565e-06,
177349
+ "loss": 0.4139,
177350
+ "step": 61295
177351
+ },
177352
+ {
177353
+ "epoch": 493.6,
177354
+ "learning_rate": 9.034022435897438e-06,
177355
+ "loss": 0.9779,
177356
+ "step": 61300
177357
+ },
177358
+ {
177359
+ "epoch": 493.64,
177360
+ "learning_rate": 9.033942307692309e-06,
177361
+ "loss": 0.5961,
177362
+ "step": 61305
177363
+ },
177364
+ {
177365
+ "epoch": 493.68,
177366
+ "learning_rate": 9.03386217948718e-06,
177367
+ "loss": 0.3516,
177368
+ "step": 61310
177369
+ },
177370
+ {
177371
+ "epoch": 493.72,
177372
+ "learning_rate": 9.033782051282052e-06,
177373
+ "loss": 0.369,
177374
+ "step": 61315
177375
+ },
177376
+ {
177377
+ "epoch": 493.76,
177378
+ "learning_rate": 9.033701923076925e-06,
177379
+ "loss": 0.3593,
177380
+ "step": 61320
177381
+ },
177382
+ {
177383
+ "epoch": 493.8,
177384
+ "learning_rate": 9.033621794871795e-06,
177385
+ "loss": 1.1128,
177386
+ "step": 61325
177387
+ },
177388
+ {
177389
+ "epoch": 493.84,
177390
+ "learning_rate": 9.033541666666668e-06,
177391
+ "loss": 0.7125,
177392
+ "step": 61330
177393
+ },
177394
+ {
177395
+ "epoch": 493.88,
177396
+ "learning_rate": 9.033461538461539e-06,
177397
+ "loss": 0.2982,
177398
+ "step": 61335
177399
+ },
177400
+ {
177401
+ "epoch": 493.92,
177402
+ "learning_rate": 9.03338141025641e-06,
177403
+ "loss": 0.3721,
177404
+ "step": 61340
177405
+ },
177406
+ {
177407
+ "epoch": 493.96,
177408
+ "learning_rate": 9.033301282051283e-06,
177409
+ "loss": 0.4503,
177410
+ "step": 61345
177411
+ },
177412
+ {
177413
+ "epoch": 494.0,
177414
+ "learning_rate": 9.033221153846155e-06,
177415
+ "loss": 1.2852,
177416
+ "step": 61350
177417
+ },
177418
+ {
177419
+ "epoch": 494.0,
177420
+ "eval_loss": 0.44114968180656433,
177421
+ "eval_runtime": 39.8577,
177422
+ "eval_samples_per_second": 21.175,
177423
+ "eval_steps_per_second": 0.677,
177424
+ "eval_wer": 0.19714999265462024,
177425
+ "step": 61350
177426
  }
177427
  ],
177428
  "max_steps": 620000,
177429
  "num_train_epochs": 5000,
177430
+ "total_flos": 1.726617692128464e+20,
177431
  "trial_name": null,
177432
  "trial_params": null
177433
  }
model-bin/finetune/base/{checkpoint-60604 β†’ checkpoint-61350}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629801593.3967817/events.out.tfevents.1629801593.c435e1c5ee04.920.111 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd7ebb6b9ad8bdcb543b4d0b20493ae5f5c2d74faaa9abb297525498ba670dd2
3
+ size 4194
model-bin/finetune/base/log/1629802228.319492/events.out.tfevents.1629802228.c435e1c5ee04.920.113 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bdebc67ff479d3c7a2dbca1af927fed594257ef4b3272741471a6cbe4ed290e
3
+ size 4194
model-bin/finetune/base/log/1629802982.729967/events.out.tfevents.1629802982.c435e1c5ee04.920.115 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:803609220a3c1c6b285ba8193e913ac1accfacd5683e8618ab6d512c6788b3bb
3
+ size 4194
model-bin/finetune/base/log/1629803633.8011563/events.out.tfevents.1629803633.c435e1c5ee04.920.117 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbf42755bccd1212cd17d5ec34bb0ccb56632d0e7a48f49fa8e1910d9aeee5b4
3
+ size 4194
model-bin/finetune/base/log/1629804285.3739185/events.out.tfevents.1629804285.c435e1c5ee04.920.119 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cdc9f3670826ec32a43a56c0da8c1873056c8b9dac6ba0d35260eaf6cfbe0db
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629801593.c435e1c5ee04.920.110 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6fff149151cfa3dbdc11f79036d4a2096438833e33402587b49702d1d38487d
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629802228.c435e1c5ee04.920.112 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90c340b035d4304c7511ab2db433a6d0437c8906b59f6c9e2777fb8f1b16487c
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629802982.c435e1c5ee04.920.114 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f0aac5616423719fba7b23a3b8af9d3d46a2fa430725705fe9a065c0cd4b5a0
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629803633.c435e1c5ee04.920.116 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b55ce2b849ac588f7a9a4ae4c1f5e343fea811d7d3a475be3d235173366d6579
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629804285.c435e1c5ee04.920.118 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49e5c45499277e77336a08ac9e7a8c75f0b85b40bf898b671ff93389db7c45d6
3
+ size 8622