Check commited on
Commit
ce2f4b5
Β·
1 Parent(s): e2f0da8

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/rng_state.pth +2 -2
  6. model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629739443.170523/events.out.tfevents.1629739443.74272264b15c.932.183 +3 -0
  11. model-bin/finetune/base/log/1629740075.5392973/events.out.tfevents.1629740075.74272264b15c.932.185 +3 -0
  12. model-bin/finetune/base/log/1629740711.1643536/events.out.tfevents.1629740711.74272264b15c.932.187 +3 -0
  13. model-bin/finetune/base/log/1629741346.2360308/events.out.tfevents.1629741346.74272264b15c.932.189 +3 -0
  14. model-bin/finetune/base/log/1629741982.7596643/events.out.tfevents.1629741982.74272264b15c.932.191 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629739443.74272264b15c.932.182 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629740075.74272264b15c.932.184 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629740711.74272264b15c.932.186 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629741346.74272264b15c.932.188 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629741982.74272264b15c.932.190 +3 -0
model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba72d974503ea478c5b1729780f9276d10be80ca6fa1c6cd08db6a45a220a068
3
  size 722165009
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:959ce62729227bed55822bda8043de7724c3e5c158eb50b761917a9abfa7be57
3
  size 722165009
model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:92cb28cec2c0eee45942a06d4108f95570fa53db5e6fd7d5a34fde871d8bdfec
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a461917b442f70797596af1555f318276458e7e17d444d349f44961efe363f1
3
  size 377909911
model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7cf89f648b1dff6e8fb245e4c3a61d5629be972463315d60dd829635437bcb14
3
- size 14567
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e36128eac80a2569e908c8e07343e4c8d0eabf3932ed0ad3ce6ddff419c12340
3
+ size 14503
model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9d3cadb12dc304d2d13a4672789668d30e7ae619e0398705d5c7682eaaebbdf8
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65dadf708547d192dd1a1f85344e17c1a7a89640d06852cd8566bd64ca2db817
3
  size 559
model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f7969b68e4aa6b630303de92298d5193fab81b96e65d08c15e07bcc1987850e1
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6011d3763d3a4d577b6b20622f290fc53b7b5fe40a920457f7d6a696ba3dd22
3
  size 623
model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.19748327029386092,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-46666",
4
- "epoch": 398.99598393574297,
5
- "global_step": 49530,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -162384,11 +162384,800 @@
162384
  "eval_steps_per_second": 0.689,
162385
  "eval_wer": 0.19971912188631827,
162386
  "step": 49530
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
162387
  }
162388
  ],
162389
  "max_steps": 620000,
162390
  "num_train_epochs": 5000,
162391
- "total_flos": 1.3936683466505619e+20,
162392
  "trial_name": null,
162393
  "trial_params": null
162394
  }
 
1
  {
2
  "best_metric": 0.19748327029386092,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-46666",
4
+ "epoch": 403.99598393574297,
5
+ "global_step": 50151,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
162384
  "eval_steps_per_second": 0.689,
162385
  "eval_wer": 0.19971912188631827,
162386
  "step": 49530
162387
+ },
162388
+ {
162389
+ "epoch": 399.04,
162390
+ "learning_rate": 9.222483974358976e-06,
162391
+ "loss": 0.3401,
162392
+ "step": 49535
162393
+ },
162394
+ {
162395
+ "epoch": 399.08,
162396
+ "learning_rate": 9.222403846153847e-06,
162397
+ "loss": 0.2955,
162398
+ "step": 49540
162399
+ },
162400
+ {
162401
+ "epoch": 399.12,
162402
+ "learning_rate": 9.222323717948719e-06,
162403
+ "loss": 0.4354,
162404
+ "step": 49545
162405
+ },
162406
+ {
162407
+ "epoch": 399.16,
162408
+ "learning_rate": 9.22224358974359e-06,
162409
+ "loss": 0.5642,
162410
+ "step": 49550
162411
+ },
162412
+ {
162413
+ "epoch": 399.2,
162414
+ "learning_rate": 9.222163461538463e-06,
162415
+ "loss": 1.2955,
162416
+ "step": 49555
162417
+ },
162418
+ {
162419
+ "epoch": 399.24,
162420
+ "learning_rate": 9.222083333333334e-06,
162421
+ "loss": 0.431,
162422
+ "step": 49560
162423
+ },
162424
+ {
162425
+ "epoch": 399.28,
162426
+ "learning_rate": 9.222003205128206e-06,
162427
+ "loss": 0.3745,
162428
+ "step": 49565
162429
+ },
162430
+ {
162431
+ "epoch": 399.32,
162432
+ "learning_rate": 9.221923076923077e-06,
162433
+ "loss": 0.2961,
162434
+ "step": 49570
162435
+ },
162436
+ {
162437
+ "epoch": 399.36,
162438
+ "learning_rate": 9.22184294871795e-06,
162439
+ "loss": 0.5477,
162440
+ "step": 49575
162441
+ },
162442
+ {
162443
+ "epoch": 399.4,
162444
+ "learning_rate": 9.221762820512821e-06,
162445
+ "loss": 1.3279,
162446
+ "step": 49580
162447
+ },
162448
+ {
162449
+ "epoch": 399.44,
162450
+ "learning_rate": 9.221682692307693e-06,
162451
+ "loss": 0.4807,
162452
+ "step": 49585
162453
+ },
162454
+ {
162455
+ "epoch": 399.48,
162456
+ "learning_rate": 9.221602564102566e-06,
162457
+ "loss": 0.3058,
162458
+ "step": 49590
162459
+ },
162460
+ {
162461
+ "epoch": 399.52,
162462
+ "learning_rate": 9.221522435897437e-06,
162463
+ "loss": 0.4265,
162464
+ "step": 49595
162465
+ },
162466
+ {
162467
+ "epoch": 399.56,
162468
+ "learning_rate": 9.221442307692309e-06,
162469
+ "loss": 0.4743,
162470
+ "step": 49600
162471
+ },
162472
+ {
162473
+ "epoch": 399.6,
162474
+ "learning_rate": 9.22136217948718e-06,
162475
+ "loss": 1.4961,
162476
+ "step": 49605
162477
+ },
162478
+ {
162479
+ "epoch": 399.65,
162480
+ "learning_rate": 9.221282051282053e-06,
162481
+ "loss": 0.4391,
162482
+ "step": 49610
162483
+ },
162484
+ {
162485
+ "epoch": 399.69,
162486
+ "learning_rate": 9.221201923076923e-06,
162487
+ "loss": 0.3347,
162488
+ "step": 49615
162489
+ },
162490
+ {
162491
+ "epoch": 399.73,
162492
+ "learning_rate": 9.221121794871796e-06,
162493
+ "loss": 0.3622,
162494
+ "step": 49620
162495
+ },
162496
+ {
162497
+ "epoch": 399.77,
162498
+ "learning_rate": 9.221041666666667e-06,
162499
+ "loss": 0.5091,
162500
+ "step": 49625
162501
+ },
162502
+ {
162503
+ "epoch": 399.81,
162504
+ "learning_rate": 9.220961538461538e-06,
162505
+ "loss": 1.2901,
162506
+ "step": 49630
162507
+ },
162508
+ {
162509
+ "epoch": 399.85,
162510
+ "learning_rate": 9.220881410256411e-06,
162511
+ "loss": 0.3537,
162512
+ "step": 49635
162513
+ },
162514
+ {
162515
+ "epoch": 399.89,
162516
+ "learning_rate": 9.220801282051283e-06,
162517
+ "loss": 0.4095,
162518
+ "step": 49640
162519
+ },
162520
+ {
162521
+ "epoch": 399.93,
162522
+ "learning_rate": 9.220721153846154e-06,
162523
+ "loss": 0.4307,
162524
+ "step": 49645
162525
+ },
162526
+ {
162527
+ "epoch": 399.97,
162528
+ "learning_rate": 9.220641025641026e-06,
162529
+ "loss": 0.5198,
162530
+ "step": 49650
162531
+ },
162532
+ {
162533
+ "epoch": 400.0,
162534
+ "eval_loss": 0.43548643589019775,
162535
+ "eval_runtime": 40.3971,
162536
+ "eval_samples_per_second": 20.769,
162537
+ "eval_steps_per_second": 0.668,
162538
+ "eval_wer": 0.21171731966312707,
162539
+ "step": 49654
162540
+ },
162541
+ {
162542
+ "epoch": 400.01,
162543
+ "learning_rate": 9.220560897435899e-06,
162544
+ "loss": 0.696,
162545
+ "step": 49655
162546
+ },
162547
+ {
162548
+ "epoch": 400.05,
162549
+ "learning_rate": 9.22048076923077e-06,
162550
+ "loss": 0.3643,
162551
+ "step": 49660
162552
+ },
162553
+ {
162554
+ "epoch": 400.09,
162555
+ "learning_rate": 9.220400641025641e-06,
162556
+ "loss": 0.3304,
162557
+ "step": 49665
162558
+ },
162559
+ {
162560
+ "epoch": 400.13,
162561
+ "learning_rate": 9.220320512820513e-06,
162562
+ "loss": 0.3745,
162563
+ "step": 49670
162564
+ },
162565
+ {
162566
+ "epoch": 400.17,
162567
+ "learning_rate": 9.220240384615386e-06,
162568
+ "loss": 0.5855,
162569
+ "step": 49675
162570
+ },
162571
+ {
162572
+ "epoch": 400.21,
162573
+ "learning_rate": 9.220160256410257e-06,
162574
+ "loss": 1.3573,
162575
+ "step": 49680
162576
+ },
162577
+ {
162578
+ "epoch": 400.25,
162579
+ "learning_rate": 9.220080128205128e-06,
162580
+ "loss": 0.3411,
162581
+ "step": 49685
162582
+ },
162583
+ {
162584
+ "epoch": 400.29,
162585
+ "learning_rate": 9.220000000000002e-06,
162586
+ "loss": 0.4071,
162587
+ "step": 49690
162588
+ },
162589
+ {
162590
+ "epoch": 400.33,
162591
+ "learning_rate": 9.219919871794873e-06,
162592
+ "loss": 0.3437,
162593
+ "step": 49695
162594
+ },
162595
+ {
162596
+ "epoch": 400.37,
162597
+ "learning_rate": 9.219839743589744e-06,
162598
+ "loss": 0.6714,
162599
+ "step": 49700
162600
+ },
162601
+ {
162602
+ "epoch": 400.41,
162603
+ "learning_rate": 9.219759615384616e-06,
162604
+ "loss": 1.2815,
162605
+ "step": 49705
162606
+ },
162607
+ {
162608
+ "epoch": 400.45,
162609
+ "learning_rate": 9.219695512820513e-06,
162610
+ "loss": 1.3999,
162611
+ "step": 49710
162612
+ },
162613
+ {
162614
+ "epoch": 400.49,
162615
+ "learning_rate": 9.219615384615386e-06,
162616
+ "loss": 0.4497,
162617
+ "step": 49715
162618
+ },
162619
+ {
162620
+ "epoch": 400.53,
162621
+ "learning_rate": 9.219535256410257e-06,
162622
+ "loss": 0.4104,
162623
+ "step": 49720
162624
+ },
162625
+ {
162626
+ "epoch": 400.57,
162627
+ "learning_rate": 9.219455128205129e-06,
162628
+ "loss": 0.5813,
162629
+ "step": 49725
162630
+ },
162631
+ {
162632
+ "epoch": 400.61,
162633
+ "learning_rate": 9.219375e-06,
162634
+ "loss": 1.224,
162635
+ "step": 49730
162636
+ },
162637
+ {
162638
+ "epoch": 400.65,
162639
+ "learning_rate": 9.219294871794873e-06,
162640
+ "loss": 0.3109,
162641
+ "step": 49735
162642
+ },
162643
+ {
162644
+ "epoch": 400.69,
162645
+ "learning_rate": 9.219214743589745e-06,
162646
+ "loss": 0.3625,
162647
+ "step": 49740
162648
+ },
162649
+ {
162650
+ "epoch": 400.73,
162651
+ "learning_rate": 9.219134615384616e-06,
162652
+ "loss": 0.3834,
162653
+ "step": 49745
162654
+ },
162655
+ {
162656
+ "epoch": 400.77,
162657
+ "learning_rate": 9.219054487179489e-06,
162658
+ "loss": 0.5324,
162659
+ "step": 49750
162660
+ },
162661
+ {
162662
+ "epoch": 400.81,
162663
+ "learning_rate": 9.21897435897436e-06,
162664
+ "loss": 1.2051,
162665
+ "step": 49755
162666
+ },
162667
+ {
162668
+ "epoch": 400.85,
162669
+ "learning_rate": 9.218894230769232e-06,
162670
+ "loss": 0.3304,
162671
+ "step": 49760
162672
+ },
162673
+ {
162674
+ "epoch": 400.9,
162675
+ "learning_rate": 9.218814102564103e-06,
162676
+ "loss": 0.3533,
162677
+ "step": 49765
162678
+ },
162679
+ {
162680
+ "epoch": 400.94,
162681
+ "learning_rate": 9.218733974358976e-06,
162682
+ "loss": 0.3579,
162683
+ "step": 49770
162684
+ },
162685
+ {
162686
+ "epoch": 400.98,
162687
+ "learning_rate": 9.218653846153847e-06,
162688
+ "loss": 0.7775,
162689
+ "step": 49775
162690
+ },
162691
+ {
162692
+ "epoch": 401.0,
162693
+ "eval_loss": 0.45663779973983765,
162694
+ "eval_runtime": 40.3567,
162695
+ "eval_samples_per_second": 20.79,
162696
+ "eval_steps_per_second": 0.669,
162697
+ "eval_wer": 0.20139291932675565,
162698
+ "step": 49778
162699
+ },
162700
+ {
162701
+ "epoch": 398.02,
162702
+ "learning_rate": 9.218573717948719e-06,
162703
+ "loss": 0.4261,
162704
+ "step": 49780
162705
+ },
162706
+ {
162707
+ "epoch": 398.06,
162708
+ "learning_rate": 9.21849358974359e-06,
162709
+ "loss": 0.3171,
162710
+ "step": 49785
162711
+ },
162712
+ {
162713
+ "epoch": 398.1,
162714
+ "learning_rate": 9.218413461538463e-06,
162715
+ "loss": 0.4089,
162716
+ "step": 49790
162717
+ },
162718
+ {
162719
+ "epoch": 398.14,
162720
+ "learning_rate": 9.218333333333333e-06,
162721
+ "loss": 0.3868,
162722
+ "step": 49795
162723
+ },
162724
+ {
162725
+ "epoch": 398.18,
162726
+ "learning_rate": 9.218253205128206e-06,
162727
+ "loss": 0.6769,
162728
+ "step": 49800
162729
+ },
162730
+ {
162731
+ "epoch": 398.22,
162732
+ "learning_rate": 9.218173076923079e-06,
162733
+ "loss": 1.0966,
162734
+ "step": 49805
162735
+ },
162736
+ {
162737
+ "epoch": 398.25,
162738
+ "learning_rate": 9.218092948717949e-06,
162739
+ "loss": 0.3772,
162740
+ "step": 49810
162741
+ },
162742
+ {
162743
+ "epoch": 398.29,
162744
+ "learning_rate": 9.218012820512822e-06,
162745
+ "loss": 0.3179,
162746
+ "step": 49815
162747
+ },
162748
+ {
162749
+ "epoch": 398.33,
162750
+ "learning_rate": 9.217932692307693e-06,
162751
+ "loss": 0.4016,
162752
+ "step": 49820
162753
+ },
162754
+ {
162755
+ "epoch": 398.37,
162756
+ "learning_rate": 9.217852564102564e-06,
162757
+ "loss": 0.6761,
162758
+ "step": 49825
162759
+ },
162760
+ {
162761
+ "epoch": 398.41,
162762
+ "learning_rate": 9.217772435897436e-06,
162763
+ "loss": 1.0311,
162764
+ "step": 49830
162765
+ },
162766
+ {
162767
+ "epoch": 398.45,
162768
+ "learning_rate": 9.217692307692309e-06,
162769
+ "loss": 0.4023,
162770
+ "step": 49835
162771
+ },
162772
+ {
162773
+ "epoch": 398.49,
162774
+ "learning_rate": 9.21761217948718e-06,
162775
+ "loss": 0.2977,
162776
+ "step": 49840
162777
+ },
162778
+ {
162779
+ "epoch": 398.53,
162780
+ "learning_rate": 9.217532051282052e-06,
162781
+ "loss": 0.3801,
162782
+ "step": 49845
162783
+ },
162784
+ {
162785
+ "epoch": 398.57,
162786
+ "learning_rate": 9.217451923076925e-06,
162787
+ "loss": 0.7346,
162788
+ "step": 49850
162789
+ },
162790
+ {
162791
+ "epoch": 398.61,
162792
+ "learning_rate": 9.217371794871796e-06,
162793
+ "loss": 1.0575,
162794
+ "step": 49855
162795
+ },
162796
+ {
162797
+ "epoch": 398.65,
162798
+ "learning_rate": 9.217291666666667e-06,
162799
+ "loss": 0.289,
162800
+ "step": 49860
162801
+ },
162802
+ {
162803
+ "epoch": 398.69,
162804
+ "learning_rate": 9.217211538461539e-06,
162805
+ "loss": 0.3293,
162806
+ "step": 49865
162807
+ },
162808
+ {
162809
+ "epoch": 398.73,
162810
+ "learning_rate": 9.217131410256412e-06,
162811
+ "loss": 0.4006,
162812
+ "step": 49870
162813
+ },
162814
+ {
162815
+ "epoch": 398.77,
162816
+ "learning_rate": 9.217051282051283e-06,
162817
+ "loss": 0.6412,
162818
+ "step": 49875
162819
+ },
162820
+ {
162821
+ "epoch": 398.81,
162822
+ "learning_rate": 9.216971153846154e-06,
162823
+ "loss": 1.2047,
162824
+ "step": 49880
162825
+ },
162826
+ {
162827
+ "epoch": 398.85,
162828
+ "learning_rate": 9.216891025641026e-06,
162829
+ "loss": 0.4079,
162830
+ "step": 49885
162831
+ },
162832
+ {
162833
+ "epoch": 398.89,
162834
+ "learning_rate": 9.216810897435899e-06,
162835
+ "loss": 0.3128,
162836
+ "step": 49890
162837
+ },
162838
+ {
162839
+ "epoch": 398.93,
162840
+ "learning_rate": 9.21673076923077e-06,
162841
+ "loss": 0.4184,
162842
+ "step": 49895
162843
+ },
162844
+ {
162845
+ "epoch": 398.97,
162846
+ "learning_rate": 9.216650641025642e-06,
162847
+ "loss": 0.7058,
162848
+ "step": 49900
162849
+ },
162850
+ {
162851
+ "epoch": 399.0,
162852
+ "eval_loss": 0.4314434826374054,
162853
+ "eval_runtime": 39.886,
162854
+ "eval_samples_per_second": 21.035,
162855
+ "eval_steps_per_second": 0.677,
162856
+ "eval_wer": 0.20628127946616379,
162857
+ "step": 49903
162858
+ },
162859
+ {
162860
+ "epoch": 402.02,
162861
+ "learning_rate": 9.216570512820515e-06,
162862
+ "loss": 0.3856,
162863
+ "step": 49905
162864
+ },
162865
+ {
162866
+ "epoch": 402.06,
162867
+ "learning_rate": 9.216490384615386e-06,
162868
+ "loss": 0.3282,
162869
+ "step": 49910
162870
+ },
162871
+ {
162872
+ "epoch": 402.1,
162873
+ "learning_rate": 9.216410256410257e-06,
162874
+ "loss": 0.3658,
162875
+ "step": 49915
162876
+ },
162877
+ {
162878
+ "epoch": 402.14,
162879
+ "learning_rate": 9.216330128205129e-06,
162880
+ "loss": 0.4127,
162881
+ "step": 49920
162882
+ },
162883
+ {
162884
+ "epoch": 402.18,
162885
+ "learning_rate": 9.216250000000002e-06,
162886
+ "loss": 0.7727,
162887
+ "step": 49925
162888
+ },
162889
+ {
162890
+ "epoch": 402.22,
162891
+ "learning_rate": 9.216169871794871e-06,
162892
+ "loss": 1.1278,
162893
+ "step": 49930
162894
+ },
162895
+ {
162896
+ "epoch": 402.26,
162897
+ "learning_rate": 9.216089743589744e-06,
162898
+ "loss": 0.3621,
162899
+ "step": 49935
162900
+ },
162901
+ {
162902
+ "epoch": 402.3,
162903
+ "learning_rate": 9.216009615384616e-06,
162904
+ "loss": 0.3462,
162905
+ "step": 49940
162906
+ },
162907
+ {
162908
+ "epoch": 402.34,
162909
+ "learning_rate": 9.215929487179487e-06,
162910
+ "loss": 0.4675,
162911
+ "step": 49945
162912
+ },
162913
+ {
162914
+ "epoch": 402.38,
162915
+ "learning_rate": 9.21584935897436e-06,
162916
+ "loss": 0.794,
162917
+ "step": 49950
162918
+ },
162919
+ {
162920
+ "epoch": 402.42,
162921
+ "learning_rate": 9.215769230769232e-06,
162922
+ "loss": 1.0722,
162923
+ "step": 49955
162924
+ },
162925
+ {
162926
+ "epoch": 402.46,
162927
+ "learning_rate": 9.215689102564103e-06,
162928
+ "loss": 0.344,
162929
+ "step": 49960
162930
+ },
162931
+ {
162932
+ "epoch": 402.5,
162933
+ "learning_rate": 9.215608974358974e-06,
162934
+ "loss": 0.3991,
162935
+ "step": 49965
162936
+ },
162937
+ {
162938
+ "epoch": 402.54,
162939
+ "learning_rate": 9.215528846153847e-06,
162940
+ "loss": 0.3788,
162941
+ "step": 49970
162942
+ },
162943
+ {
162944
+ "epoch": 402.58,
162945
+ "learning_rate": 9.215448717948719e-06,
162946
+ "loss": 0.7787,
162947
+ "step": 49975
162948
+ },
162949
+ {
162950
+ "epoch": 402.62,
162951
+ "learning_rate": 9.21536858974359e-06,
162952
+ "loss": 1.2898,
162953
+ "step": 49980
162954
+ },
162955
+ {
162956
+ "epoch": 402.66,
162957
+ "learning_rate": 9.215288461538461e-06,
162958
+ "loss": 0.3289,
162959
+ "step": 49985
162960
+ },
162961
+ {
162962
+ "epoch": 402.7,
162963
+ "learning_rate": 9.215208333333335e-06,
162964
+ "loss": 0.4183,
162965
+ "step": 49990
162966
+ },
162967
+ {
162968
+ "epoch": 402.74,
162969
+ "learning_rate": 9.215128205128206e-06,
162970
+ "loss": 0.3565,
162971
+ "step": 49995
162972
+ },
162973
+ {
162974
+ "epoch": 402.78,
162975
+ "learning_rate": 9.215048076923077e-06,
162976
+ "loss": 0.6545,
162977
+ "step": 50000
162978
+ },
162979
+ {
162980
+ "epoch": 402.82,
162981
+ "learning_rate": 9.21496794871795e-06,
162982
+ "loss": 1.1079,
162983
+ "step": 50005
162984
+ },
162985
+ {
162986
+ "epoch": 402.86,
162987
+ "learning_rate": 9.214887820512822e-06,
162988
+ "loss": 0.3451,
162989
+ "step": 50010
162990
+ },
162991
+ {
162992
+ "epoch": 402.9,
162993
+ "learning_rate": 9.214807692307693e-06,
162994
+ "loss": 0.3267,
162995
+ "step": 50015
162996
+ },
162997
+ {
162998
+ "epoch": 402.94,
162999
+ "learning_rate": 9.214727564102564e-06,
163000
+ "loss": 0.4015,
163001
+ "step": 50020
163002
+ },
163003
+ {
163004
+ "epoch": 402.98,
163005
+ "learning_rate": 9.214647435897437e-06,
163006
+ "loss": 0.8181,
163007
+ "step": 50025
163008
+ },
163009
+ {
163010
+ "epoch": 403.0,
163011
+ "eval_loss": 0.3694767951965332,
163012
+ "eval_runtime": 40.8543,
163013
+ "eval_samples_per_second": 20.536,
163014
+ "eval_steps_per_second": 0.661,
163015
+ "eval_wer": 0.20543960439240783,
163016
+ "step": 50027
163017
+ },
163018
+ {
163019
+ "epoch": 403.02,
163020
+ "learning_rate": 9.214567307692309e-06,
163021
+ "loss": 0.3587,
163022
+ "step": 50030
163023
+ },
163024
+ {
163025
+ "epoch": 403.06,
163026
+ "learning_rate": 9.21448717948718e-06,
163027
+ "loss": 0.358,
163028
+ "step": 50035
163029
+ },
163030
+ {
163031
+ "epoch": 403.1,
163032
+ "learning_rate": 9.214407051282051e-06,
163033
+ "loss": 0.335,
163034
+ "step": 50040
163035
+ },
163036
+ {
163037
+ "epoch": 403.14,
163038
+ "learning_rate": 9.214326923076925e-06,
163039
+ "loss": 0.3689,
163040
+ "step": 50045
163041
+ },
163042
+ {
163043
+ "epoch": 403.18,
163044
+ "learning_rate": 9.214246794871796e-06,
163045
+ "loss": 0.777,
163046
+ "step": 50050
163047
+ },
163048
+ {
163049
+ "epoch": 403.22,
163050
+ "learning_rate": 9.214166666666667e-06,
163051
+ "loss": 0.8766,
163052
+ "step": 50055
163053
+ },
163054
+ {
163055
+ "epoch": 403.27,
163056
+ "learning_rate": 9.21408653846154e-06,
163057
+ "loss": 0.3625,
163058
+ "step": 50060
163059
+ },
163060
+ {
163061
+ "epoch": 403.31,
163062
+ "learning_rate": 9.214006410256412e-06,
163063
+ "loss": 0.3834,
163064
+ "step": 50065
163065
+ },
163066
+ {
163067
+ "epoch": 403.35,
163068
+ "learning_rate": 9.213926282051283e-06,
163069
+ "loss": 0.4218,
163070
+ "step": 50070
163071
+ },
163072
+ {
163073
+ "epoch": 403.39,
163074
+ "learning_rate": 9.213846153846154e-06,
163075
+ "loss": 1.0129,
163076
+ "step": 50075
163077
+ },
163078
+ {
163079
+ "epoch": 403.43,
163080
+ "learning_rate": 9.213766025641027e-06,
163081
+ "loss": 0.8793,
163082
+ "step": 50080
163083
+ },
163084
+ {
163085
+ "epoch": 403.47,
163086
+ "learning_rate": 9.213685897435897e-06,
163087
+ "loss": 0.4697,
163088
+ "step": 50085
163089
+ },
163090
+ {
163091
+ "epoch": 403.51,
163092
+ "learning_rate": 9.21360576923077e-06,
163093
+ "loss": 0.3166,
163094
+ "step": 50090
163095
+ },
163096
+ {
163097
+ "epoch": 403.55,
163098
+ "learning_rate": 9.213525641025642e-06,
163099
+ "loss": 0.4842,
163100
+ "step": 50095
163101
+ },
163102
+ {
163103
+ "epoch": 403.59,
163104
+ "learning_rate": 9.213445512820513e-06,
163105
+ "loss": 0.7842,
163106
+ "step": 50100
163107
+ },
163108
+ {
163109
+ "epoch": 403.63,
163110
+ "learning_rate": 9.213365384615386e-06,
163111
+ "loss": 0.928,
163112
+ "step": 50105
163113
+ },
163114
+ {
163115
+ "epoch": 403.67,
163116
+ "learning_rate": 9.213285256410257e-06,
163117
+ "loss": 0.3819,
163118
+ "step": 50110
163119
+ },
163120
+ {
163121
+ "epoch": 403.71,
163122
+ "learning_rate": 9.213205128205129e-06,
163123
+ "loss": 0.3871,
163124
+ "step": 50115
163125
+ },
163126
+ {
163127
+ "epoch": 403.75,
163128
+ "learning_rate": 9.213125e-06,
163129
+ "loss": 0.5323,
163130
+ "step": 50120
163131
+ },
163132
+ {
163133
+ "epoch": 403.79,
163134
+ "learning_rate": 9.213044871794873e-06,
163135
+ "loss": 0.825,
163136
+ "step": 50125
163137
+ },
163138
+ {
163139
+ "epoch": 403.83,
163140
+ "learning_rate": 9.212964743589744e-06,
163141
+ "loss": 0.9118,
163142
+ "step": 50130
163143
+ },
163144
+ {
163145
+ "epoch": 403.87,
163146
+ "learning_rate": 9.212884615384616e-06,
163147
+ "loss": 0.3248,
163148
+ "step": 50135
163149
+ },
163150
+ {
163151
+ "epoch": 403.91,
163152
+ "learning_rate": 9.212804487179487e-06,
163153
+ "loss": 0.3549,
163154
+ "step": 50140
163155
+ },
163156
+ {
163157
+ "epoch": 403.95,
163158
+ "learning_rate": 9.21272435897436e-06,
163159
+ "loss": 0.3968,
163160
+ "step": 50145
163161
+ },
163162
+ {
163163
+ "epoch": 403.99,
163164
+ "learning_rate": 9.212644230769232e-06,
163165
+ "loss": 0.9281,
163166
+ "step": 50150
163167
+ },
163168
+ {
163169
+ "epoch": 404.0,
163170
+ "eval_loss": 0.42465662956237793,
163171
+ "eval_runtime": 39.9556,
163172
+ "eval_samples_per_second": 21.023,
163173
+ "eval_steps_per_second": 0.676,
163174
+ "eval_wer": 0.2024080463989428,
163175
+ "step": 50151
163176
  }
163177
  ],
163178
  "max_steps": 620000,
163179
  "num_train_epochs": 5000,
163180
+ "total_flos": 1.4111387032874315e+20,
163181
  "trial_name": null,
163182
  "trial_params": null
163183
  }
model-bin/finetune/base/{checkpoint-49530 β†’ checkpoint-50151}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629739443.170523/events.out.tfevents.1629739443.74272264b15c.932.183 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05614eb8b16fee60b725c40faf15d89c81b1692dcfb630b0e8ccaa59f3002563
3
+ size 4194
model-bin/finetune/base/log/1629740075.5392973/events.out.tfevents.1629740075.74272264b15c.932.185 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db5b2ec03ca5c04c382e539caba6df5c92a095e64987efdde3bd12bbb0082c71
3
+ size 4194
model-bin/finetune/base/log/1629740711.1643536/events.out.tfevents.1629740711.74272264b15c.932.187 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:deafe982df91982bd4bf2777889d67f14e611cd54f84d90eee9aff760da28f17
3
+ size 4194
model-bin/finetune/base/log/1629741346.2360308/events.out.tfevents.1629741346.74272264b15c.932.189 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbfabc8af4985357ab68ce92edbfa6cb43e1508c4077d47849a2f855b9900595
3
+ size 4194
model-bin/finetune/base/log/1629741982.7596643/events.out.tfevents.1629741982.74272264b15c.932.191 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb34af6d054f7518c8dfa444fb93b520f61f46399855ce053b2812d332199977
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629739443.74272264b15c.932.182 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22bf85d81fb13298ebe4bbeeeaa1b178b88d9bc2c95369be72ee972a62d41379
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1629740075.74272264b15c.932.184 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f798284ffd604ed3dd7e039c63332e089a39f40591769b416f24209811e4ac0
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629740711.74272264b15c.932.186 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5e9e0f1617165e3111f1c3da411c90483f4034eb0027e90f2e580f746624c31
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629741346.74272264b15c.932.188 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cac169c2a24ba1edd324a8bdd71713f622fcf8c2cade09a614222c8a70cedb2
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629741982.74272264b15c.932.190 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:603c48cf6fd19d1aabdb928ea1f91fa54c59eb0649b8bea3d8f8b476208258c8
3
+ size 8622