Check commited on
Commit
8991754
Β·
1 Parent(s): 97ce0c1

"auto-commit"

Browse files
Files changed (17) hide show
  1. model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/trainer_state.json +633 -3
  9. model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630203840.7050338/events.out.tfevents.1630203840.cc93b136ebf5.1086.1 +3 -0
  11. model-bin/finetune/base/log/1630204292.5703435/events.out.tfevents.1630204292.cc93b136ebf5.1086.3 +3 -0
  12. model-bin/finetune/base/log/1630204747.1758041/events.out.tfevents.1630204747.cc93b136ebf5.1086.5 +3 -0
  13. model-bin/finetune/base/log/1630205190.8026295/events.out.tfevents.1630205190.cc93b136ebf5.1086.7 +3 -0
  14. model-bin/finetune/base/log/events.out.tfevents.1630203840.cc93b136ebf5.1086.0 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630204292.cc93b136ebf5.1086.2 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630204747.cc93b136ebf5.1086.4 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630205190.cc93b136ebf5.1086.6 +3 -0
model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c37a6c10720a2f9c5a311eeb46c5ff3fd878b61fdb3d52018c8a201bf14df5c1
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d52be403964896ebde8dbc015ab0b31dd8eda84858da6f91dc257157cd345705
3
  size 722165393
model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0f3315de3d2b3896b78e3b43a47cecce273b45bcf96541a4ab3dd8799683e303
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d16fa8e963ee700b85a69e9291858ab4a52074a28116fa5897842dcac0729d57
3
  size 377909911
model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b86b810236303b881bbf5ca5ba5ed6f0c8e236a139fc8382041637bb625e4446
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df8dc682c31d6a272b5908de4ff759bebd27e6bb8b81ddd6b7a5edc073866bc8
3
  size 14503
model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0fbaa9c93c377bbc85cc6181cc1a8b53221928929faaa4dcbecfe2654907c58c
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69fe6b1a171b3398b251c4f574c3897019a42807059a5531cd17cb9782eff416
3
  size 559
model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b27ebbc04d558d0510257f4a9ccd450abe095bc98ee9bdb09bb4fa64e7a21824
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0417338fc26171bc7c4c31f1905061af395fd857f395bcb53cac4e888fb96a67
3
  size 623
model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1689111747851003,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
4
- "epoch": 1122.0,
5
- "global_step": 139135,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -276390,11 +276390,641 @@
276390
  "eval_steps_per_second": 0.736,
276391
  "eval_wer": 0.17295327903986285,
276392
  "step": 139135
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
276393
  }
276394
  ],
276395
  "max_steps": 620000,
276396
  "num_train_epochs": 5000,
276397
- "total_flos": 3.9151223111877696e+20,
276398
  "trial_name": null,
276399
  "trial_params": null
276400
  }
 
1
  {
2
  "best_metric": 0.1689111747851003,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
4
+ "epoch": 1125.995983935743,
5
+ "global_step": 139632,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
276390
  "eval_steps_per_second": 0.736,
276391
  "eval_wer": 0.17295327903986285,
276392
  "step": 139135
276393
+ },
276394
+ {
276395
+ "epoch": 1122.04,
276396
+ "learning_rate": 7.769224555735058e-06,
276397
+ "loss": 0.2935,
276398
+ "step": 139140
276399
+ },
276400
+ {
276401
+ "epoch": 1122.08,
276402
+ "learning_rate": 7.769143780290792e-06,
276403
+ "loss": 0.2453,
276404
+ "step": 139145
276405
+ },
276406
+ {
276407
+ "epoch": 1122.12,
276408
+ "learning_rate": 7.769063004846528e-06,
276409
+ "loss": 0.3313,
276410
+ "step": 139150
276411
+ },
276412
+ {
276413
+ "epoch": 1122.16,
276414
+ "learning_rate": 7.768982229402262e-06,
276415
+ "loss": 0.4549,
276416
+ "step": 139155
276417
+ },
276418
+ {
276419
+ "epoch": 1122.2,
276420
+ "learning_rate": 7.768901453957998e-06,
276421
+ "loss": 1.0227,
276422
+ "step": 139160
276423
+ },
276424
+ {
276425
+ "epoch": 1122.24,
276426
+ "learning_rate": 7.768820678513732e-06,
276427
+ "loss": 0.3099,
276428
+ "step": 139165
276429
+ },
276430
+ {
276431
+ "epoch": 1122.28,
276432
+ "learning_rate": 7.768739903069468e-06,
276433
+ "loss": 0.2407,
276434
+ "step": 139170
276435
+ },
276436
+ {
276437
+ "epoch": 1122.32,
276438
+ "learning_rate": 7.768659127625202e-06,
276439
+ "loss": 0.3476,
276440
+ "step": 139175
276441
+ },
276442
+ {
276443
+ "epoch": 1122.36,
276444
+ "learning_rate": 7.768578352180938e-06,
276445
+ "loss": 0.4268,
276446
+ "step": 139180
276447
+ },
276448
+ {
276449
+ "epoch": 1122.4,
276450
+ "learning_rate": 7.768497576736672e-06,
276451
+ "loss": 1.1477,
276452
+ "step": 139185
276453
+ },
276454
+ {
276455
+ "epoch": 1122.44,
276456
+ "learning_rate": 7.768416801292408e-06,
276457
+ "loss": 0.318,
276458
+ "step": 139190
276459
+ },
276460
+ {
276461
+ "epoch": 1122.48,
276462
+ "learning_rate": 7.768336025848144e-06,
276463
+ "loss": 0.2953,
276464
+ "step": 139195
276465
+ },
276466
+ {
276467
+ "epoch": 1122.52,
276468
+ "learning_rate": 7.768255250403878e-06,
276469
+ "loss": 0.2743,
276470
+ "step": 139200
276471
+ },
276472
+ {
276473
+ "epoch": 1122.56,
276474
+ "learning_rate": 7.768174474959614e-06,
276475
+ "loss": 0.4145,
276476
+ "step": 139205
276477
+ },
276478
+ {
276479
+ "epoch": 1122.6,
276480
+ "learning_rate": 7.768093699515348e-06,
276481
+ "loss": 1.4303,
276482
+ "step": 139210
276483
+ },
276484
+ {
276485
+ "epoch": 1122.65,
276486
+ "learning_rate": 7.768012924071084e-06,
276487
+ "loss": 0.3557,
276488
+ "step": 139215
276489
+ },
276490
+ {
276491
+ "epoch": 1122.69,
276492
+ "learning_rate": 7.767932148626818e-06,
276493
+ "loss": 0.3282,
276494
+ "step": 139220
276495
+ },
276496
+ {
276497
+ "epoch": 1122.73,
276498
+ "learning_rate": 7.767851373182554e-06,
276499
+ "loss": 0.3062,
276500
+ "step": 139225
276501
+ },
276502
+ {
276503
+ "epoch": 1122.77,
276504
+ "learning_rate": 7.767770597738288e-06,
276505
+ "loss": 0.4282,
276506
+ "step": 139230
276507
+ },
276508
+ {
276509
+ "epoch": 1122.81,
276510
+ "learning_rate": 7.767689822294024e-06,
276511
+ "loss": 1.1902,
276512
+ "step": 139235
276513
+ },
276514
+ {
276515
+ "epoch": 1122.85,
276516
+ "learning_rate": 7.767609046849758e-06,
276517
+ "loss": 0.2787,
276518
+ "step": 139240
276519
+ },
276520
+ {
276521
+ "epoch": 1122.89,
276522
+ "learning_rate": 7.767528271405494e-06,
276523
+ "loss": 0.3426,
276524
+ "step": 139245
276525
+ },
276526
+ {
276527
+ "epoch": 1122.93,
276528
+ "learning_rate": 7.76744749596123e-06,
276529
+ "loss": 0.2978,
276530
+ "step": 139250
276531
+ },
276532
+ {
276533
+ "epoch": 1122.97,
276534
+ "learning_rate": 7.767366720516964e-06,
276535
+ "loss": 0.4585,
276536
+ "step": 139255
276537
+ },
276538
+ {
276539
+ "epoch": 1123.0,
276540
+ "eval_loss": 0.3832680881023407,
276541
+ "eval_runtime": 41.7996,
276542
+ "eval_samples_per_second": 20.0,
276543
+ "eval_steps_per_second": 0.646,
276544
+ "eval_wer": 0.17813765182186234,
276545
+ "step": 139259
276546
+ },
276547
+ {
276548
+ "epoch": 1114.01,
276549
+ "learning_rate": 7.7672859450727e-06,
276550
+ "loss": 0.4918,
276551
+ "step": 139260
276552
+ },
276553
+ {
276554
+ "epoch": 1114.05,
276555
+ "learning_rate": 7.767205169628434e-06,
276556
+ "loss": 0.2842,
276557
+ "step": 139265
276558
+ },
276559
+ {
276560
+ "epoch": 1114.09,
276561
+ "learning_rate": 7.76712439418417e-06,
276562
+ "loss": 0.2855,
276563
+ "step": 139270
276564
+ },
276565
+ {
276566
+ "epoch": 1114.13,
276567
+ "learning_rate": 7.767043618739904e-06,
276568
+ "loss": 0.2647,
276569
+ "step": 139275
276570
+ },
276571
+ {
276572
+ "epoch": 1114.17,
276573
+ "learning_rate": 7.76696284329564e-06,
276574
+ "loss": 0.5487,
276575
+ "step": 139280
276576
+ },
276577
+ {
276578
+ "epoch": 1114.21,
276579
+ "learning_rate": 7.766882067851374e-06,
276580
+ "loss": 1.0039,
276581
+ "step": 139285
276582
+ },
276583
+ {
276584
+ "epoch": 1114.25,
276585
+ "learning_rate": 7.76680129240711e-06,
276586
+ "loss": 0.3161,
276587
+ "step": 139290
276588
+ },
276589
+ {
276590
+ "epoch": 1114.29,
276591
+ "learning_rate": 7.766720516962844e-06,
276592
+ "loss": 0.3176,
276593
+ "step": 139295
276594
+ },
276595
+ {
276596
+ "epoch": 1114.33,
276597
+ "learning_rate": 7.76663974151858e-06,
276598
+ "loss": 0.3167,
276599
+ "step": 139300
276600
+ },
276601
+ {
276602
+ "epoch": 1114.37,
276603
+ "learning_rate": 7.766558966074315e-06,
276604
+ "loss": 0.5792,
276605
+ "step": 139305
276606
+ },
276607
+ {
276608
+ "epoch": 1114.41,
276609
+ "learning_rate": 7.76647819063005e-06,
276610
+ "loss": 0.9895,
276611
+ "step": 139310
276612
+ },
276613
+ {
276614
+ "epoch": 1114.45,
276615
+ "learning_rate": 7.766397415185785e-06,
276616
+ "loss": 0.275,
276617
+ "step": 139315
276618
+ },
276619
+ {
276620
+ "epoch": 1114.49,
276621
+ "learning_rate": 7.76631663974152e-06,
276622
+ "loss": 0.2364,
276623
+ "step": 139320
276624
+ },
276625
+ {
276626
+ "epoch": 1114.53,
276627
+ "learning_rate": 7.766235864297255e-06,
276628
+ "loss": 0.3155,
276629
+ "step": 139325
276630
+ },
276631
+ {
276632
+ "epoch": 1114.57,
276633
+ "learning_rate": 7.76615508885299e-06,
276634
+ "loss": 0.5086,
276635
+ "step": 139330
276636
+ },
276637
+ {
276638
+ "epoch": 1114.61,
276639
+ "learning_rate": 7.766074313408725e-06,
276640
+ "loss": 1.0574,
276641
+ "step": 139335
276642
+ },
276643
+ {
276644
+ "epoch": 1114.65,
276645
+ "learning_rate": 7.76599353796446e-06,
276646
+ "loss": 0.3179,
276647
+ "step": 139340
276648
+ },
276649
+ {
276650
+ "epoch": 1114.69,
276651
+ "learning_rate": 7.765912762520195e-06,
276652
+ "loss": 0.3073,
276653
+ "step": 139345
276654
+ },
276655
+ {
276656
+ "epoch": 1114.73,
276657
+ "learning_rate": 7.76583198707593e-06,
276658
+ "loss": 0.3061,
276659
+ "step": 139350
276660
+ },
276661
+ {
276662
+ "epoch": 1114.77,
276663
+ "learning_rate": 7.765751211631665e-06,
276664
+ "loss": 0.5684,
276665
+ "step": 139355
276666
+ },
276667
+ {
276668
+ "epoch": 1114.81,
276669
+ "learning_rate": 7.7656704361874e-06,
276670
+ "loss": 1.2499,
276671
+ "step": 139360
276672
+ },
276673
+ {
276674
+ "epoch": 1114.85,
276675
+ "learning_rate": 7.765589660743135e-06,
276676
+ "loss": 0.3199,
276677
+ "step": 139365
276678
+ },
276679
+ {
276680
+ "epoch": 1114.89,
276681
+ "learning_rate": 7.765508885298871e-06,
276682
+ "loss": 0.2328,
276683
+ "step": 139370
276684
+ },
276685
+ {
276686
+ "epoch": 1114.93,
276687
+ "learning_rate": 7.765428109854605e-06,
276688
+ "loss": 0.3404,
276689
+ "step": 139375
276690
+ },
276691
+ {
276692
+ "epoch": 1114.97,
276693
+ "learning_rate": 7.765347334410341e-06,
276694
+ "loss": 0.6416,
276695
+ "step": 139380
276696
+ },
276697
+ {
276698
+ "epoch": 1115.0,
276699
+ "eval_loss": 0.39288097620010376,
276700
+ "eval_runtime": 42.1595,
276701
+ "eval_samples_per_second": 19.829,
276702
+ "eval_steps_per_second": 0.64,
276703
+ "eval_wer": 0.18118466898954705,
276704
+ "step": 139384
276705
+ },
276706
+ {
276707
+ "epoch": 1124.01,
276708
+ "learning_rate": 7.765266558966075e-06,
276709
+ "loss": 0.3201,
276710
+ "step": 139385
276711
+ },
276712
+ {
276713
+ "epoch": 1124.05,
276714
+ "learning_rate": 7.76518578352181e-06,
276715
+ "loss": 0.2739,
276716
+ "step": 139390
276717
+ },
276718
+ {
276719
+ "epoch": 1124.09,
276720
+ "learning_rate": 7.765105008077545e-06,
276721
+ "loss": 0.3218,
276722
+ "step": 139395
276723
+ },
276724
+ {
276725
+ "epoch": 1124.13,
276726
+ "learning_rate": 7.76502423263328e-06,
276727
+ "loss": 0.3122,
276728
+ "step": 139400
276729
+ },
276730
+ {
276731
+ "epoch": 1124.17,
276732
+ "learning_rate": 7.764943457189015e-06,
276733
+ "loss": 0.5611,
276734
+ "step": 139405
276735
+ },
276736
+ {
276737
+ "epoch": 1124.21,
276738
+ "learning_rate": 7.76486268174475e-06,
276739
+ "loss": 0.9412,
276740
+ "step": 139410
276741
+ },
276742
+ {
276743
+ "epoch": 1124.25,
276744
+ "learning_rate": 7.764781906300485e-06,
276745
+ "loss": 0.2755,
276746
+ "step": 139415
276747
+ },
276748
+ {
276749
+ "epoch": 1124.29,
276750
+ "learning_rate": 7.76470113085622e-06,
276751
+ "loss": 0.2831,
276752
+ "step": 139420
276753
+ },
276754
+ {
276755
+ "epoch": 1124.33,
276756
+ "learning_rate": 7.764620355411957e-06,
276757
+ "loss": 0.3349,
276758
+ "step": 139425
276759
+ },
276760
+ {
276761
+ "epoch": 1124.37,
276762
+ "learning_rate": 7.76453957996769e-06,
276763
+ "loss": 0.5646,
276764
+ "step": 139430
276765
+ },
276766
+ {
276767
+ "epoch": 1124.41,
276768
+ "learning_rate": 7.764458804523427e-06,
276769
+ "loss": 1.1037,
276770
+ "step": 139435
276771
+ },
276772
+ {
276773
+ "epoch": 1124.45,
276774
+ "learning_rate": 7.76437802907916e-06,
276775
+ "loss": 0.31,
276776
+ "step": 139440
276777
+ },
276778
+ {
276779
+ "epoch": 1124.49,
276780
+ "learning_rate": 7.764297253634897e-06,
276781
+ "loss": 0.2414,
276782
+ "step": 139445
276783
+ },
276784
+ {
276785
+ "epoch": 1124.53,
276786
+ "learning_rate": 7.76421647819063e-06,
276787
+ "loss": 0.3528,
276788
+ "step": 139450
276789
+ },
276790
+ {
276791
+ "epoch": 1124.57,
276792
+ "learning_rate": 7.764135702746367e-06,
276793
+ "loss": 0.5236,
276794
+ "step": 139455
276795
+ },
276796
+ {
276797
+ "epoch": 1124.61,
276798
+ "learning_rate": 7.7640549273021e-06,
276799
+ "loss": 1.0225,
276800
+ "step": 139460
276801
+ },
276802
+ {
276803
+ "epoch": 1124.65,
276804
+ "learning_rate": 7.763974151857836e-06,
276805
+ "loss": 0.2665,
276806
+ "step": 139465
276807
+ },
276808
+ {
276809
+ "epoch": 1124.69,
276810
+ "learning_rate": 7.76389337641357e-06,
276811
+ "loss": 0.2925,
276812
+ "step": 139470
276813
+ },
276814
+ {
276815
+ "epoch": 1124.73,
276816
+ "learning_rate": 7.763812600969306e-06,
276817
+ "loss": 0.3135,
276818
+ "step": 139475
276819
+ },
276820
+ {
276821
+ "epoch": 1124.77,
276822
+ "learning_rate": 7.763731825525042e-06,
276823
+ "loss": 0.5225,
276824
+ "step": 139480
276825
+ },
276826
+ {
276827
+ "epoch": 1124.81,
276828
+ "learning_rate": 7.763651050080776e-06,
276829
+ "loss": 1.0957,
276830
+ "step": 139485
276831
+ },
276832
+ {
276833
+ "epoch": 1124.85,
276834
+ "learning_rate": 7.763570274636512e-06,
276835
+ "loss": 0.251,
276836
+ "step": 139490
276837
+ },
276838
+ {
276839
+ "epoch": 1124.89,
276840
+ "learning_rate": 7.763489499192246e-06,
276841
+ "loss": 0.3002,
276842
+ "step": 139495
276843
+ },
276844
+ {
276845
+ "epoch": 1124.93,
276846
+ "learning_rate": 7.763408723747982e-06,
276847
+ "loss": 0.3116,
276848
+ "step": 139500
276849
+ },
276850
+ {
276851
+ "epoch": 1124.97,
276852
+ "learning_rate": 7.763327948303716e-06,
276853
+ "loss": 0.5954,
276854
+ "step": 139505
276855
+ },
276856
+ {
276857
+ "epoch": 1125.0,
276858
+ "eval_loss": 0.3883971869945526,
276859
+ "eval_runtime": 41.5038,
276860
+ "eval_samples_per_second": 20.143,
276861
+ "eval_steps_per_second": 0.651,
276862
+ "eval_wer": 0.18588165062308784,
276863
+ "step": 139508
276864
+ },
276865
+ {
276866
+ "epoch": 1125.02,
276867
+ "learning_rate": 7.763247172859452e-06,
276868
+ "loss": 0.3039,
276869
+ "step": 139510
276870
+ },
276871
+ {
276872
+ "epoch": 1125.06,
276873
+ "learning_rate": 7.763166397415186e-06,
276874
+ "loss": 0.288,
276875
+ "step": 139515
276876
+ },
276877
+ {
276878
+ "epoch": 1125.1,
276879
+ "learning_rate": 7.763085621970922e-06,
276880
+ "loss": 0.3032,
276881
+ "step": 139520
276882
+ },
276883
+ {
276884
+ "epoch": 1125.14,
276885
+ "learning_rate": 7.763004846526656e-06,
276886
+ "loss": 0.338,
276887
+ "step": 139525
276888
+ },
276889
+ {
276890
+ "epoch": 1125.18,
276891
+ "learning_rate": 7.762924071082392e-06,
276892
+ "loss": 0.5968,
276893
+ "step": 139530
276894
+ },
276895
+ {
276896
+ "epoch": 1125.22,
276897
+ "learning_rate": 7.762843295638126e-06,
276898
+ "loss": 0.9873,
276899
+ "step": 139535
276900
+ },
276901
+ {
276902
+ "epoch": 1125.26,
276903
+ "learning_rate": 7.762762520193862e-06,
276904
+ "loss": 0.2503,
276905
+ "step": 139540
276906
+ },
276907
+ {
276908
+ "epoch": 1125.3,
276909
+ "learning_rate": 7.762681744749598e-06,
276910
+ "loss": 0.2329,
276911
+ "step": 139545
276912
+ },
276913
+ {
276914
+ "epoch": 1125.34,
276915
+ "learning_rate": 7.762600969305332e-06,
276916
+ "loss": 0.3378,
276917
+ "step": 139550
276918
+ },
276919
+ {
276920
+ "epoch": 1125.38,
276921
+ "learning_rate": 7.762520193861068e-06,
276922
+ "loss": 0.5935,
276923
+ "step": 139555
276924
+ },
276925
+ {
276926
+ "epoch": 1125.42,
276927
+ "learning_rate": 7.762439418416802e-06,
276928
+ "loss": 1.0363,
276929
+ "step": 139560
276930
+ },
276931
+ {
276932
+ "epoch": 1125.46,
276933
+ "learning_rate": 7.762358642972538e-06,
276934
+ "loss": 0.2453,
276935
+ "step": 139565
276936
+ },
276937
+ {
276938
+ "epoch": 1125.5,
276939
+ "learning_rate": 7.762277867528272e-06,
276940
+ "loss": 0.2953,
276941
+ "step": 139570
276942
+ },
276943
+ {
276944
+ "epoch": 1125.54,
276945
+ "learning_rate": 7.762197092084008e-06,
276946
+ "loss": 0.3298,
276947
+ "step": 139575
276948
+ },
276949
+ {
276950
+ "epoch": 1125.58,
276951
+ "learning_rate": 7.762116316639742e-06,
276952
+ "loss": 0.666,
276953
+ "step": 139580
276954
+ },
276955
+ {
276956
+ "epoch": 1125.62,
276957
+ "learning_rate": 7.762035541195478e-06,
276958
+ "loss": 1.066,
276959
+ "step": 139585
276960
+ },
276961
+ {
276962
+ "epoch": 1125.66,
276963
+ "learning_rate": 7.761954765751212e-06,
276964
+ "loss": 0.3161,
276965
+ "step": 139590
276966
+ },
276967
+ {
276968
+ "epoch": 1125.7,
276969
+ "learning_rate": 7.761873990306948e-06,
276970
+ "loss": 0.3122,
276971
+ "step": 139595
276972
+ },
276973
+ {
276974
+ "epoch": 1125.74,
276975
+ "learning_rate": 7.761793214862684e-06,
276976
+ "loss": 0.2955,
276977
+ "step": 139600
276978
+ },
276979
+ {
276980
+ "epoch": 1125.78,
276981
+ "learning_rate": 7.761712439418418e-06,
276982
+ "loss": 0.5424,
276983
+ "step": 139605
276984
+ },
276985
+ {
276986
+ "epoch": 1125.82,
276987
+ "learning_rate": 7.761631663974154e-06,
276988
+ "loss": 0.8868,
276989
+ "step": 139610
276990
+ },
276991
+ {
276992
+ "epoch": 1125.86,
276993
+ "learning_rate": 7.761550888529888e-06,
276994
+ "loss": 0.2635,
276995
+ "step": 139615
276996
+ },
276997
+ {
276998
+ "epoch": 1125.9,
276999
+ "learning_rate": 7.761470113085624e-06,
277000
+ "loss": 0.3167,
277001
+ "step": 139620
277002
+ },
277003
+ {
277004
+ "epoch": 1125.94,
277005
+ "learning_rate": 7.761389337641358e-06,
277006
+ "loss": 0.3052,
277007
+ "step": 139625
277008
+ },
277009
+ {
277010
+ "epoch": 1125.98,
277011
+ "learning_rate": 7.761308562197094e-06,
277012
+ "loss": 0.5751,
277013
+ "step": 139630
277014
+ },
277015
+ {
277016
+ "epoch": 1126.0,
277017
+ "eval_loss": 0.41787150502204895,
277018
+ "eval_runtime": 42.2152,
277019
+ "eval_samples_per_second": 19.803,
277020
+ "eval_steps_per_second": 0.64,
277021
+ "eval_wer": 0.18331400405914758,
277022
+ "step": 139632
277023
  }
277024
  ],
277025
  "max_steps": 620000,
277026
  "num_train_epochs": 5000,
277027
+ "total_flos": 3.929101689489278e+20,
277028
  "trial_name": null,
277029
  "trial_params": null
277030
  }
model-bin/finetune/base/{checkpoint-139135 β†’ checkpoint-139632}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630203840.7050338/events.out.tfevents.1630203840.cc93b136ebf5.1086.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:273ad8879055fa1e8a3be1e72b9b855c122061cae57d1da0f9bea0929dbd9f0a
3
+ size 4194
model-bin/finetune/base/log/1630204292.5703435/events.out.tfevents.1630204292.cc93b136ebf5.1086.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ca2e0c0c8f9b92125a834acb99214859eb42d27631539078d418d4d7a59d332
3
+ size 4194
model-bin/finetune/base/log/1630204747.1758041/events.out.tfevents.1630204747.cc93b136ebf5.1086.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bb22e6c31b54936b35de42462c29b9ee7cc4aac67112d63ab6877005aedb2c3
3
+ size 4194
model-bin/finetune/base/log/1630205190.8026295/events.out.tfevents.1630205190.cc93b136ebf5.1086.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:042e5b679dc07a09f4ec878e65f77977578dc02281f0cd0ad44acf9b158a362c
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630203840.cc93b136ebf5.1086.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb14bae35002d5cf692a8356c6ec57ad533583f92066c63b16278303cb7a761d
3
+ size 8470
model-bin/finetune/base/log/events.out.tfevents.1630204292.cc93b136ebf5.1086.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fed4d49e432c5adef9fc96ece4196590b69cbc2ef8caf390a569df99a8b2f7c
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630204747.cc93b136ebf5.1086.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac35332ab8480fac9861525d27699460c902247ca9e0af69aba5ee2b42768ad0
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630205190.cc93b136ebf5.1086.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac28020d3a642b31669b12aba95f397cf2494c95d49171996def28b6601352ed
3
+ size 8622