Check commited on
Commit
4215eab
Β·
1 Parent(s): 2694101

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630173460.284948/events.out.tfevents.1630173460.86bb0ddabf9b.4092.251 +3 -0
  11. model-bin/finetune/base/log/1630173847.4052565/events.out.tfevents.1630173847.86bb0ddabf9b.4092.253 +3 -0
  12. model-bin/finetune/base/log/1630174238.155075/events.out.tfevents.1630174238.86bb0ddabf9b.4092.255 +3 -0
  13. model-bin/finetune/base/log/1630174625.3833299/events.out.tfevents.1630174625.86bb0ddabf9b.4092.257 +3 -0
  14. model-bin/finetune/base/log/1630175012.849805/events.out.tfevents.1630175012.86bb0ddabf9b.4092.259 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630173460.86bb0ddabf9b.4092.250 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630173847.86bb0ddabf9b.4092.252 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630174238.86bb0ddabf9b.4092.254 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630174625.86bb0ddabf9b.4092.256 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630175012.86bb0ddabf9b.4092.258 +3 -0
model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0131aea205c7b47df894a41ad960996f174a622cdb2a0e466e6c420dddd66eb3
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d508d8a35a8f9a643497f1992285b1c2f7877eb4215a0840cbe0a3f99384b88
3
  size 722165393
model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:984900319b1571a20bc0eff8f0132123a9e2552f902a51b84b84e449b1d6e8ad
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e2bd640a75fb59f050e69639bbc44f6c152b69f570c1567b4a338b6476efa97
3
  size 377909911
model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:24e5394362e1ac456498e0ab0081d483b216a02a39ead2f59423571ab3011946
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6d3d8ba5c02cd61971126a99f8567c0c592c0bc2b36c2483b358d10fd14ecb4
3
  size 14503
model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:137fd28074ff7a16078afbfd1e5c0ac3943fefd3026a1f7f05234a0ac2c9e1f9
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc49d6fa99b57306d86d74ceb133d6a9314c50a72c07e34fb7acc7abe68fa707
3
  size 559
model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8ec230c409ad1e6eb7cefef319d6d09881c43c2e19a2e87131bedb122fa28933
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8927c9b9c4b0eca7da703c920593d2940ba00dd84b20f83a2fca6543dd58d6cc
3
  size 623
model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1689111747851003,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
4
- "epoch": 1075.995983935743,
5
- "global_step": 133530,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -269259,11 +269259,800 @@
269259
  "eval_steps_per_second": 0.757,
269260
  "eval_wer": 0.17710974284679465,
269261
  "step": 133530
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
269262
  }
269263
  ],
269264
  "max_steps": 620000,
269265
  "num_train_epochs": 5000,
269266
- "total_flos": 3.75745971037642e+20,
269267
  "trial_name": null,
269268
  "trial_params": null
269269
  }
 
1
  {
2
  "best_metric": 0.1689111747851003,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
4
+ "epoch": 1080.995983935743,
5
+ "global_step": 134152,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
269259
  "eval_steps_per_second": 0.757,
269260
  "eval_wer": 0.17710974284679465,
269261
  "step": 133530
269262
+ },
269263
+ {
269264
+ "epoch": 1068.04,
269265
+ "learning_rate": 7.876891025641025e-06,
269266
+ "loss": 0.283,
269267
+ "step": 133535
269268
+ },
269269
+ {
269270
+ "epoch": 1068.08,
269271
+ "learning_rate": 7.876810897435898e-06,
269272
+ "loss": 0.6511,
269273
+ "step": 133540
269274
+ },
269275
+ {
269276
+ "epoch": 1068.12,
269277
+ "learning_rate": 7.876730769230771e-06,
269278
+ "loss": 0.2785,
269279
+ "step": 133545
269280
+ },
269281
+ {
269282
+ "epoch": 1068.16,
269283
+ "learning_rate": 7.876650641025641e-06,
269284
+ "loss": 0.438,
269285
+ "step": 133550
269286
+ },
269287
+ {
269288
+ "epoch": 1068.2,
269289
+ "learning_rate": 7.876570512820514e-06,
269290
+ "loss": 1.1175,
269291
+ "step": 133555
269292
+ },
269293
+ {
269294
+ "epoch": 1068.24,
269295
+ "learning_rate": 7.876490384615385e-06,
269296
+ "loss": 0.2989,
269297
+ "step": 133560
269298
+ },
269299
+ {
269300
+ "epoch": 1068.28,
269301
+ "learning_rate": 7.876410256410257e-06,
269302
+ "loss": 0.2486,
269303
+ "step": 133565
269304
+ },
269305
+ {
269306
+ "epoch": 1068.32,
269307
+ "learning_rate": 7.876330128205128e-06,
269308
+ "loss": 0.31,
269309
+ "step": 133570
269310
+ },
269311
+ {
269312
+ "epoch": 1068.36,
269313
+ "learning_rate": 7.876250000000001e-06,
269314
+ "loss": 0.3822,
269315
+ "step": 133575
269316
+ },
269317
+ {
269318
+ "epoch": 1068.4,
269319
+ "learning_rate": 7.876169871794872e-06,
269320
+ "loss": 1.2008,
269321
+ "step": 133580
269322
+ },
269323
+ {
269324
+ "epoch": 1068.44,
269325
+ "learning_rate": 7.876089743589744e-06,
269326
+ "loss": 0.3558,
269327
+ "step": 133585
269328
+ },
269329
+ {
269330
+ "epoch": 1068.48,
269331
+ "learning_rate": 7.876009615384615e-06,
269332
+ "loss": 0.2611,
269333
+ "step": 133590
269334
+ },
269335
+ {
269336
+ "epoch": 1068.52,
269337
+ "learning_rate": 7.875929487179488e-06,
269338
+ "loss": 0.282,
269339
+ "step": 133595
269340
+ },
269341
+ {
269342
+ "epoch": 1068.56,
269343
+ "learning_rate": 7.87584935897436e-06,
269344
+ "loss": 0.4253,
269345
+ "step": 133600
269346
+ },
269347
+ {
269348
+ "epoch": 1068.6,
269349
+ "learning_rate": 7.875769230769231e-06,
269350
+ "loss": 1.191,
269351
+ "step": 133605
269352
+ },
269353
+ {
269354
+ "epoch": 1068.64,
269355
+ "learning_rate": 7.875689102564104e-06,
269356
+ "loss": 0.3124,
269357
+ "step": 133610
269358
+ },
269359
+ {
269360
+ "epoch": 1068.68,
269361
+ "learning_rate": 7.875608974358975e-06,
269362
+ "loss": 0.3442,
269363
+ "step": 133615
269364
+ },
269365
+ {
269366
+ "epoch": 1068.72,
269367
+ "learning_rate": 7.875528846153847e-06,
269368
+ "loss": 0.2791,
269369
+ "step": 133620
269370
+ },
269371
+ {
269372
+ "epoch": 1068.76,
269373
+ "learning_rate": 7.875448717948718e-06,
269374
+ "loss": 0.4265,
269375
+ "step": 133625
269376
+ },
269377
+ {
269378
+ "epoch": 1068.8,
269379
+ "learning_rate": 7.875368589743591e-06,
269380
+ "loss": 1.2426,
269381
+ "step": 133630
269382
+ },
269383
+ {
269384
+ "epoch": 1068.84,
269385
+ "learning_rate": 7.875288461538462e-06,
269386
+ "loss": 0.2532,
269387
+ "step": 133635
269388
+ },
269389
+ {
269390
+ "epoch": 1068.88,
269391
+ "learning_rate": 7.875208333333334e-06,
269392
+ "loss": 0.2682,
269393
+ "step": 133640
269394
+ },
269395
+ {
269396
+ "epoch": 1068.92,
269397
+ "learning_rate": 7.875128205128207e-06,
269398
+ "loss": 0.3925,
269399
+ "step": 133645
269400
+ },
269401
+ {
269402
+ "epoch": 1068.96,
269403
+ "learning_rate": 7.875048076923078e-06,
269404
+ "loss": 0.4542,
269405
+ "step": 133650
269406
+ },
269407
+ {
269408
+ "epoch": 1069.0,
269409
+ "learning_rate": 7.87496794871795e-06,
269410
+ "loss": 1.3908,
269411
+ "step": 133655
269412
+ },
269413
+ {
269414
+ "epoch": 1069.0,
269415
+ "eval_loss": 0.3515404164791107,
269416
+ "eval_runtime": 35.3248,
269417
+ "eval_samples_per_second": 23.638,
269418
+ "eval_steps_per_second": 0.764,
269419
+ "eval_wer": 0.1784813881061842,
269420
+ "step": 133655
269421
+ },
269422
+ {
269423
+ "epoch": 1077.04,
269424
+ "learning_rate": 7.874887820512821e-06,
269425
+ "loss": 0.2722,
269426
+ "step": 133660
269427
+ },
269428
+ {
269429
+ "epoch": 1077.08,
269430
+ "learning_rate": 7.874807692307694e-06,
269431
+ "loss": 0.2621,
269432
+ "step": 133665
269433
+ },
269434
+ {
269435
+ "epoch": 1077.12,
269436
+ "learning_rate": 7.874727564102564e-06,
269437
+ "loss": 0.285,
269438
+ "step": 133670
269439
+ },
269440
+ {
269441
+ "epoch": 1077.16,
269442
+ "learning_rate": 7.874647435897437e-06,
269443
+ "loss": 0.4751,
269444
+ "step": 133675
269445
+ },
269446
+ {
269447
+ "epoch": 1077.2,
269448
+ "learning_rate": 7.874567307692308e-06,
269449
+ "loss": 1.0416,
269450
+ "step": 133680
269451
+ },
269452
+ {
269453
+ "epoch": 1077.24,
269454
+ "learning_rate": 7.87448717948718e-06,
269455
+ "loss": 0.32,
269456
+ "step": 133685
269457
+ },
269458
+ {
269459
+ "epoch": 1077.28,
269460
+ "learning_rate": 7.87440705128205e-06,
269461
+ "loss": 0.2579,
269462
+ "step": 133690
269463
+ },
269464
+ {
269465
+ "epoch": 1077.32,
269466
+ "learning_rate": 7.874326923076924e-06,
269467
+ "loss": 0.2939,
269468
+ "step": 133695
269469
+ },
269470
+ {
269471
+ "epoch": 1077.36,
269472
+ "learning_rate": 7.874246794871795e-06,
269473
+ "loss": 0.4655,
269474
+ "step": 133700
269475
+ },
269476
+ {
269477
+ "epoch": 1077.4,
269478
+ "learning_rate": 7.874166666666667e-06,
269479
+ "loss": 1.1427,
269480
+ "step": 133705
269481
+ },
269482
+ {
269483
+ "epoch": 1077.44,
269484
+ "learning_rate": 7.87408653846154e-06,
269485
+ "loss": 0.2917,
269486
+ "step": 133710
269487
+ },
269488
+ {
269489
+ "epoch": 1077.48,
269490
+ "learning_rate": 7.874006410256411e-06,
269491
+ "loss": 0.3403,
269492
+ "step": 133715
269493
+ },
269494
+ {
269495
+ "epoch": 1077.52,
269496
+ "learning_rate": 7.873926282051282e-06,
269497
+ "loss": 0.4189,
269498
+ "step": 133720
269499
+ },
269500
+ {
269501
+ "epoch": 1077.56,
269502
+ "learning_rate": 7.873846153846154e-06,
269503
+ "loss": 0.4168,
269504
+ "step": 133725
269505
+ },
269506
+ {
269507
+ "epoch": 1077.6,
269508
+ "learning_rate": 7.873766025641027e-06,
269509
+ "loss": 1.1052,
269510
+ "step": 133730
269511
+ },
269512
+ {
269513
+ "epoch": 1077.64,
269514
+ "learning_rate": 7.873685897435898e-06,
269515
+ "loss": 0.3159,
269516
+ "step": 133735
269517
+ },
269518
+ {
269519
+ "epoch": 1077.68,
269520
+ "learning_rate": 7.87360576923077e-06,
269521
+ "loss": 0.2568,
269522
+ "step": 133740
269523
+ },
269524
+ {
269525
+ "epoch": 1077.72,
269526
+ "learning_rate": 7.873525641025642e-06,
269527
+ "loss": 0.3329,
269528
+ "step": 133745
269529
+ },
269530
+ {
269531
+ "epoch": 1077.76,
269532
+ "learning_rate": 7.873445512820514e-06,
269533
+ "loss": 0.4664,
269534
+ "step": 133750
269535
+ },
269536
+ {
269537
+ "epoch": 1077.8,
269538
+ "learning_rate": 7.873365384615385e-06,
269539
+ "loss": 1.05,
269540
+ "step": 133755
269541
+ },
269542
+ {
269543
+ "epoch": 1077.84,
269544
+ "learning_rate": 7.873285256410257e-06,
269545
+ "loss": 0.3629,
269546
+ "step": 133760
269547
+ },
269548
+ {
269549
+ "epoch": 1077.88,
269550
+ "learning_rate": 7.87320512820513e-06,
269551
+ "loss": 0.3166,
269552
+ "step": 133765
269553
+ },
269554
+ {
269555
+ "epoch": 1077.92,
269556
+ "learning_rate": 7.873125000000001e-06,
269557
+ "loss": 0.3112,
269558
+ "step": 133770
269559
+ },
269560
+ {
269561
+ "epoch": 1077.96,
269562
+ "learning_rate": 7.873044871794872e-06,
269563
+ "loss": 0.4494,
269564
+ "step": 133775
269565
+ },
269566
+ {
269567
+ "epoch": 1078.0,
269568
+ "eval_loss": 0.36935585737228394,
269569
+ "eval_runtime": 35.5506,
269570
+ "eval_samples_per_second": 23.488,
269571
+ "eval_steps_per_second": 0.759,
269572
+ "eval_wer": 0.18314665084192566,
269573
+ "step": 133779
269574
+ },
269575
+ {
269576
+ "epoch": 1078.01,
269577
+ "learning_rate": 7.872964743589744e-06,
269578
+ "loss": 0.3194,
269579
+ "step": 133780
269580
+ },
269581
+ {
269582
+ "epoch": 1078.05,
269583
+ "learning_rate": 7.872884615384617e-06,
269584
+ "loss": 0.2743,
269585
+ "step": 133785
269586
+ },
269587
+ {
269588
+ "epoch": 1078.09,
269589
+ "learning_rate": 7.872804487179488e-06,
269590
+ "loss": 0.249,
269591
+ "step": 133790
269592
+ },
269593
+ {
269594
+ "epoch": 1078.13,
269595
+ "learning_rate": 7.87272435897436e-06,
269596
+ "loss": 0.3227,
269597
+ "step": 133795
269598
+ },
269599
+ {
269600
+ "epoch": 1078.17,
269601
+ "learning_rate": 7.872644230769232e-06,
269602
+ "loss": 0.5637,
269603
+ "step": 133800
269604
+ },
269605
+ {
269606
+ "epoch": 1078.21,
269607
+ "learning_rate": 7.872564102564104e-06,
269608
+ "loss": 1.0401,
269609
+ "step": 133805
269610
+ },
269611
+ {
269612
+ "epoch": 1078.25,
269613
+ "learning_rate": 7.872483974358975e-06,
269614
+ "loss": 0.3121,
269615
+ "step": 133810
269616
+ },
269617
+ {
269618
+ "epoch": 1078.29,
269619
+ "learning_rate": 7.872403846153847e-06,
269620
+ "loss": 0.2549,
269621
+ "step": 133815
269622
+ },
269623
+ {
269624
+ "epoch": 1078.33,
269625
+ "learning_rate": 7.87232371794872e-06,
269626
+ "loss": 0.381,
269627
+ "step": 133820
269628
+ },
269629
+ {
269630
+ "epoch": 1078.37,
269631
+ "learning_rate": 7.87224358974359e-06,
269632
+ "loss": 0.4923,
269633
+ "step": 133825
269634
+ },
269635
+ {
269636
+ "epoch": 1078.41,
269637
+ "learning_rate": 7.872163461538462e-06,
269638
+ "loss": 1.0382,
269639
+ "step": 133830
269640
+ },
269641
+ {
269642
+ "epoch": 1078.45,
269643
+ "learning_rate": 7.872083333333334e-06,
269644
+ "loss": 0.2953,
269645
+ "step": 133835
269646
+ },
269647
+ {
269648
+ "epoch": 1078.49,
269649
+ "learning_rate": 7.872003205128205e-06,
269650
+ "loss": 0.3126,
269651
+ "step": 133840
269652
+ },
269653
+ {
269654
+ "epoch": 1078.53,
269655
+ "learning_rate": 7.871923076923078e-06,
269656
+ "loss": 0.3423,
269657
+ "step": 133845
269658
+ },
269659
+ {
269660
+ "epoch": 1078.57,
269661
+ "learning_rate": 7.87184294871795e-06,
269662
+ "loss": 0.5135,
269663
+ "step": 133850
269664
+ },
269665
+ {
269666
+ "epoch": 1078.61,
269667
+ "learning_rate": 7.87176282051282e-06,
269668
+ "loss": 0.8895,
269669
+ "step": 133855
269670
+ },
269671
+ {
269672
+ "epoch": 1078.65,
269673
+ "learning_rate": 7.871682692307692e-06,
269674
+ "loss": 0.2551,
269675
+ "step": 133860
269676
+ },
269677
+ {
269678
+ "epoch": 1078.69,
269679
+ "learning_rate": 7.871602564102565e-06,
269680
+ "loss": 0.2887,
269681
+ "step": 133865
269682
+ },
269683
+ {
269684
+ "epoch": 1078.73,
269685
+ "learning_rate": 7.871522435897437e-06,
269686
+ "loss": 0.3447,
269687
+ "step": 133870
269688
+ },
269689
+ {
269690
+ "epoch": 1078.77,
269691
+ "learning_rate": 7.871442307692308e-06,
269692
+ "loss": 0.4865,
269693
+ "step": 133875
269694
+ },
269695
+ {
269696
+ "epoch": 1078.81,
269697
+ "learning_rate": 7.87136217948718e-06,
269698
+ "loss": 1.1507,
269699
+ "step": 133880
269700
+ },
269701
+ {
269702
+ "epoch": 1078.85,
269703
+ "learning_rate": 7.871282051282052e-06,
269704
+ "loss": 0.3337,
269705
+ "step": 133885
269706
+ },
269707
+ {
269708
+ "epoch": 1078.89,
269709
+ "learning_rate": 7.871201923076924e-06,
269710
+ "loss": 0.3155,
269711
+ "step": 133890
269712
+ },
269713
+ {
269714
+ "epoch": 1078.93,
269715
+ "learning_rate": 7.871121794871795e-06,
269716
+ "loss": 0.3432,
269717
+ "step": 133895
269718
+ },
269719
+ {
269720
+ "epoch": 1078.97,
269721
+ "learning_rate": 7.871041666666668e-06,
269722
+ "loss": 0.5148,
269723
+ "step": 133900
269724
+ },
269725
+ {
269726
+ "epoch": 1079.0,
269727
+ "eval_loss": 0.41137564182281494,
269728
+ "eval_runtime": 35.9724,
269729
+ "eval_samples_per_second": 23.212,
269730
+ "eval_steps_per_second": 0.751,
269731
+ "eval_wer": 0.17041547277936964,
269732
+ "step": 133903
269733
+ },
269734
+ {
269735
+ "epoch": 1071.02,
269736
+ "learning_rate": 7.87096153846154e-06,
269737
+ "loss": 0.3643,
269738
+ "step": 133905
269739
+ },
269740
+ {
269741
+ "epoch": 1071.06,
269742
+ "learning_rate": 7.870881410256411e-06,
269743
+ "loss": 0.3349,
269744
+ "step": 133910
269745
+ },
269746
+ {
269747
+ "epoch": 1071.1,
269748
+ "learning_rate": 7.870801282051282e-06,
269749
+ "loss": 0.275,
269750
+ "step": 133915
269751
+ },
269752
+ {
269753
+ "epoch": 1071.14,
269754
+ "learning_rate": 7.870721153846155e-06,
269755
+ "loss": 0.303,
269756
+ "step": 133920
269757
+ },
269758
+ {
269759
+ "epoch": 1071.18,
269760
+ "learning_rate": 7.870641025641027e-06,
269761
+ "loss": 0.5914,
269762
+ "step": 133925
269763
+ },
269764
+ {
269765
+ "epoch": 1071.22,
269766
+ "learning_rate": 7.870560897435898e-06,
269767
+ "loss": 1.0265,
269768
+ "step": 133930
269769
+ },
269770
+ {
269771
+ "epoch": 1071.26,
269772
+ "learning_rate": 7.87048076923077e-06,
269773
+ "loss": 0.2939,
269774
+ "step": 133935
269775
+ },
269776
+ {
269777
+ "epoch": 1071.3,
269778
+ "learning_rate": 7.870400641025642e-06,
269779
+ "loss": 0.2298,
269780
+ "step": 133940
269781
+ },
269782
+ {
269783
+ "epoch": 1071.34,
269784
+ "learning_rate": 7.870320512820514e-06,
269785
+ "loss": 0.2893,
269786
+ "step": 133945
269787
+ },
269788
+ {
269789
+ "epoch": 1071.38,
269790
+ "learning_rate": 7.870240384615385e-06,
269791
+ "loss": 0.5808,
269792
+ "step": 133950
269793
+ },
269794
+ {
269795
+ "epoch": 1071.42,
269796
+ "learning_rate": 7.870160256410258e-06,
269797
+ "loss": 0.9685,
269798
+ "step": 133955
269799
+ },
269800
+ {
269801
+ "epoch": 1071.46,
269802
+ "learning_rate": 7.87008012820513e-06,
269803
+ "loss": 0.2692,
269804
+ "step": 133960
269805
+ },
269806
+ {
269807
+ "epoch": 1071.5,
269808
+ "learning_rate": 7.870000000000001e-06,
269809
+ "loss": 0.2342,
269810
+ "step": 133965
269811
+ },
269812
+ {
269813
+ "epoch": 1071.54,
269814
+ "learning_rate": 7.869919871794872e-06,
269815
+ "loss": 0.3344,
269816
+ "step": 133970
269817
+ },
269818
+ {
269819
+ "epoch": 1071.58,
269820
+ "learning_rate": 7.869839743589745e-06,
269821
+ "loss": 0.5782,
269822
+ "step": 133975
269823
+ },
269824
+ {
269825
+ "epoch": 1071.62,
269826
+ "learning_rate": 7.869759615384615e-06,
269827
+ "loss": 0.9415,
269828
+ "step": 133980
269829
+ },
269830
+ {
269831
+ "epoch": 1071.66,
269832
+ "learning_rate": 7.869679487179488e-06,
269833
+ "loss": 0.248,
269834
+ "step": 133985
269835
+ },
269836
+ {
269837
+ "epoch": 1071.7,
269838
+ "learning_rate": 7.86959935897436e-06,
269839
+ "loss": 0.2807,
269840
+ "step": 133990
269841
+ },
269842
+ {
269843
+ "epoch": 1071.74,
269844
+ "learning_rate": 7.86951923076923e-06,
269845
+ "loss": 0.3315,
269846
+ "step": 133995
269847
+ },
269848
+ {
269849
+ "epoch": 1071.78,
269850
+ "learning_rate": 7.869439102564104e-06,
269851
+ "loss": 0.6116,
269852
+ "step": 134000
269853
+ },
269854
+ {
269855
+ "epoch": 1071.82,
269856
+ "learning_rate": 7.869358974358975e-06,
269857
+ "loss": 0.9509,
269858
+ "step": 134005
269859
+ },
269860
+ {
269861
+ "epoch": 1071.86,
269862
+ "learning_rate": 7.869278846153846e-06,
269863
+ "loss": 0.2919,
269864
+ "step": 134010
269865
+ },
269866
+ {
269867
+ "epoch": 1071.9,
269868
+ "learning_rate": 7.869198717948718e-06,
269869
+ "loss": 0.296,
269870
+ "step": 134015
269871
+ },
269872
+ {
269873
+ "epoch": 1071.94,
269874
+ "learning_rate": 7.869118589743591e-06,
269875
+ "loss": 0.3532,
269876
+ "step": 134020
269877
+ },
269878
+ {
269879
+ "epoch": 1071.98,
269880
+ "learning_rate": 7.869038461538462e-06,
269881
+ "loss": 0.5742,
269882
+ "step": 134025
269883
+ },
269884
+ {
269885
+ "epoch": 1072.0,
269886
+ "eval_loss": 0.3423796594142914,
269887
+ "eval_runtime": 35.6088,
269888
+ "eval_samples_per_second": 23.449,
269889
+ "eval_steps_per_second": 0.758,
269890
+ "eval_wer": 0.17097562782428327,
269891
+ "step": 134028
269892
+ },
269893
+ {
269894
+ "epoch": 1080.02,
269895
+ "learning_rate": 7.868958333333334e-06,
269896
+ "loss": 0.2569,
269897
+ "step": 134030
269898
+ },
269899
+ {
269900
+ "epoch": 1080.06,
269901
+ "learning_rate": 7.868878205128205e-06,
269902
+ "loss": 0.3096,
269903
+ "step": 134035
269904
+ },
269905
+ {
269906
+ "epoch": 1080.1,
269907
+ "learning_rate": 7.868798076923078e-06,
269908
+ "loss": 0.2902,
269909
+ "step": 134040
269910
+ },
269911
+ {
269912
+ "epoch": 1080.14,
269913
+ "learning_rate": 7.86871794871795e-06,
269914
+ "loss": 0.3467,
269915
+ "step": 134045
269916
+ },
269917
+ {
269918
+ "epoch": 1080.18,
269919
+ "learning_rate": 7.86863782051282e-06,
269920
+ "loss": 0.5397,
269921
+ "step": 134050
269922
+ },
269923
+ {
269924
+ "epoch": 1080.22,
269925
+ "learning_rate": 7.868557692307694e-06,
269926
+ "loss": 1.0189,
269927
+ "step": 134055
269928
+ },
269929
+ {
269930
+ "epoch": 1080.26,
269931
+ "learning_rate": 7.868477564102565e-06,
269932
+ "loss": 0.2523,
269933
+ "step": 134060
269934
+ },
269935
+ {
269936
+ "epoch": 1080.3,
269937
+ "learning_rate": 7.868397435897437e-06,
269938
+ "loss": 0.2536,
269939
+ "step": 134065
269940
+ },
269941
+ {
269942
+ "epoch": 1080.34,
269943
+ "learning_rate": 7.868317307692308e-06,
269944
+ "loss": 0.3514,
269945
+ "step": 134070
269946
+ },
269947
+ {
269948
+ "epoch": 1080.38,
269949
+ "learning_rate": 7.868237179487181e-06,
269950
+ "loss": 0.7254,
269951
+ "step": 134075
269952
+ },
269953
+ {
269954
+ "epoch": 1080.42,
269955
+ "learning_rate": 7.868157051282052e-06,
269956
+ "loss": 1.1177,
269957
+ "step": 134080
269958
+ },
269959
+ {
269960
+ "epoch": 1080.46,
269961
+ "learning_rate": 7.868076923076924e-06,
269962
+ "loss": 0.2958,
269963
+ "step": 134085
269964
+ },
269965
+ {
269966
+ "epoch": 1080.5,
269967
+ "learning_rate": 7.867996794871795e-06,
269968
+ "loss": 0.2778,
269969
+ "step": 134090
269970
+ },
269971
+ {
269972
+ "epoch": 1080.54,
269973
+ "learning_rate": 7.867916666666668e-06,
269974
+ "loss": 0.3512,
269975
+ "step": 134095
269976
+ },
269977
+ {
269978
+ "epoch": 1080.58,
269979
+ "learning_rate": 7.86783653846154e-06,
269980
+ "loss": 0.555,
269981
+ "step": 134100
269982
+ },
269983
+ {
269984
+ "epoch": 1080.62,
269985
+ "learning_rate": 7.86775641025641e-06,
269986
+ "loss": 1.1554,
269987
+ "step": 134105
269988
+ },
269989
+ {
269990
+ "epoch": 1080.66,
269991
+ "learning_rate": 7.867676282051284e-06,
269992
+ "loss": 0.2772,
269993
+ "step": 134110
269994
+ },
269995
+ {
269996
+ "epoch": 1080.7,
269997
+ "learning_rate": 7.867596153846153e-06,
269998
+ "loss": 0.3057,
269999
+ "step": 134115
270000
+ },
270001
+ {
270002
+ "epoch": 1080.74,
270003
+ "learning_rate": 7.867516025641027e-06,
270004
+ "loss": 0.3324,
270005
+ "step": 134120
270006
+ },
270007
+ {
270008
+ "epoch": 1080.78,
270009
+ "learning_rate": 7.867435897435898e-06,
270010
+ "loss": 0.5588,
270011
+ "step": 134125
270012
+ },
270013
+ {
270014
+ "epoch": 1080.82,
270015
+ "learning_rate": 7.86735576923077e-06,
270016
+ "loss": 1.137,
270017
+ "step": 134130
270018
+ },
270019
+ {
270020
+ "epoch": 1080.86,
270021
+ "learning_rate": 7.86727564102564e-06,
270022
+ "loss": 0.293,
270023
+ "step": 134135
270024
+ },
270025
+ {
270026
+ "epoch": 1080.9,
270027
+ "learning_rate": 7.867195512820514e-06,
270028
+ "loss": 0.3207,
270029
+ "step": 134140
270030
+ },
270031
+ {
270032
+ "epoch": 1080.94,
270033
+ "learning_rate": 7.867115384615385e-06,
270034
+ "loss": 0.3501,
270035
+ "step": 134145
270036
+ },
270037
+ {
270038
+ "epoch": 1080.98,
270039
+ "learning_rate": 7.867035256410256e-06,
270040
+ "loss": 0.767,
270041
+ "step": 134150
270042
+ },
270043
+ {
270044
+ "epoch": 1081.0,
270045
+ "eval_loss": 0.3516015112400055,
270046
+ "eval_runtime": 36.4198,
270047
+ "eval_samples_per_second": 22.955,
270048
+ "eval_steps_per_second": 0.741,
270049
+ "eval_wer": 0.1819801839047687,
270050
+ "step": 134152
270051
  }
270052
  ],
270053
  "max_steps": 620000,
270054
  "num_train_epochs": 5000,
270055
+ "total_flos": 3.774979260580794e+20,
270056
  "trial_name": null,
270057
  "trial_params": null
270058
  }
model-bin/finetune/base/{checkpoint-133530 β†’ checkpoint-134152}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630173460.284948/events.out.tfevents.1630173460.86bb0ddabf9b.4092.251 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0663966003fa02b1812f8ba56338d9555f48211f3ecd478b8106d78daf474d5b
3
+ size 4194
model-bin/finetune/base/log/1630173847.4052565/events.out.tfevents.1630173847.86bb0ddabf9b.4092.253 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cb6219d11e7906000bd3e22599a21df47651f33dcfbca94e516cf096d20a6a6
3
+ size 4194
model-bin/finetune/base/log/1630174238.155075/events.out.tfevents.1630174238.86bb0ddabf9b.4092.255 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:973f26703749c96d6fcdbb097e4c200d3f91a76c8ea263a3c5edbe0eda1642f6
3
+ size 4194
model-bin/finetune/base/log/1630174625.3833299/events.out.tfevents.1630174625.86bb0ddabf9b.4092.257 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:737ac480260f8043607fa3b7050dbcdf16b58f43eb25d387e4ae969f04731901
3
+ size 4194
model-bin/finetune/base/log/1630175012.849805/events.out.tfevents.1630175012.86bb0ddabf9b.4092.259 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4b516171f2fe94023507ec5aecab580ebabc50b6f2dc8f6b69dd303220497e3
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630173460.86bb0ddabf9b.4092.250 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9901c68ba5c8ffdc6da733b931c8ce2c57b123cb5df451387131a533d1cac86e
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630173847.86bb0ddabf9b.4092.252 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f14716706d90ec7f61cdff14c85a16441331bcfb3b998d2083754aa3ae8f48a
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1630174238.86bb0ddabf9b.4092.254 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b2fcb555e8f4671789c82bc10264e8283e02e41a6b60b5998b91e24606e47fd
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630174625.86bb0ddabf9b.4092.256 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78813b6ecc3938169f66820ae6701faff5aaa82a4c2eb6ea883c89ec54649874
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630175012.86bb0ddabf9b.4092.258 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c78bcf475dbf4a8089057e2b72d8e4ec4cc76b0860f78a34b849bf98a3c334b9
3
+ size 8622