Check commited on
Commit
b9edd56
Β·
1 Parent(s): 6696c75

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/trainer_state.json +793 -4
  9. model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630146705.8776062/events.out.tfevents.1630146705.86bb0ddabf9b.4092.121 +3 -0
  11. model-bin/finetune/base/log/1630147090.9692104/events.out.tfevents.1630147090.86bb0ddabf9b.4092.123 +3 -0
  12. model-bin/finetune/base/log/1630147594.250904/events.out.tfevents.1630147594.86bb0ddabf9b.4092.125 +3 -0
  13. model-bin/finetune/base/log/1630147976.392592/events.out.tfevents.1630147976.86bb0ddabf9b.4092.127 +3 -0
  14. model-bin/finetune/base/log/1630148364.6244335/events.out.tfevents.1630148364.86bb0ddabf9b.4092.129 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630146705.86bb0ddabf9b.4092.120 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630147090.86bb0ddabf9b.4092.122 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630147594.86bb0ddabf9b.4092.124 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630147976.86bb0ddabf9b.4092.126 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630148364.86bb0ddabf9b.4092.128 +3 -0
model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f63c3bb2813226bca109201f437ebf67c43cfef3ce4b37fc6773b00451c6afd3
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16b005e9fb69b2e7c668f7d9d1ddcf9502c134575ee756c2b4dd2a09e4f3dfc7
3
  size 722165393
model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f31e81ef246b6d82c092a74a34ec533837f2f19d8bb978eb447cf77e84cd3ccd
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cfc721d81f03e5e16e7e7c9c5142ddddbe2c39f6609dd7fe74fbe15bffd4bf7
3
  size 377909911
model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:16b2eee16926d8653ad01b2f3c1ecdf61861739c3ae6c1de09f83de384c7912f
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:499abeb83f7ca59edfcad7e7d1e42dccc904e7481abcc4da5c71266773800888
3
  size 14503
model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:27dcd6b89e9ca24fa782487c54a3e4aa32dcd9bb3a40752f5c73ade334c653ba
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:683ccffd7adf3cb2b0b65666beebc8e51b2ae5066df9b1b2e36f844178028c85
3
  size 559
model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:38924a63617ddcff0e33da0ada89550be5f4f3a143fab0f47de35854193ff038
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:149ab5492f8d36be18b46436a0461744c2fef1b202cb12894209dd371443a41d
3
  size 623
model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.17162025681719809,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-124947",
4
- "epoch": 1010.995983935743,
5
- "global_step": 125446,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -258972,11 +258972,800 @@
258972
  "eval_steps_per_second": 0.724,
258973
  "eval_wer": 0.1797744687003036,
258974
  "step": 125446
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
258975
  }
258976
  ],
258977
- "max_steps": 620000,
258978
  "num_train_epochs": 5000,
258979
- "total_flos": 3.5301452563316874e+20,
258980
  "trial_name": null,
258981
  "trial_params": null
258982
  }
 
1
  {
2
  "best_metric": 0.17162025681719809,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-124947",
4
+ "epoch": 1008.0,
5
+ "global_step": 126068,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
258972
  "eval_steps_per_second": 0.724,
258973
  "eval_wer": 0.1797744687003036,
258974
  "step": 125446
258975
+ },
258976
+ {
258977
+ "epoch": 1011.03,
258978
+ "learning_rate": 8.006394230769231e-06,
258979
+ "loss": 0.3332,
258980
+ "step": 125450
258981
+ },
258982
+ {
258983
+ "epoch": 1011.07,
258984
+ "learning_rate": 8.006314102564104e-06,
258985
+ "loss": 0.2435,
258986
+ "step": 125455
258987
+ },
258988
+ {
258989
+ "epoch": 1011.11,
258990
+ "learning_rate": 8.006233974358974e-06,
258991
+ "loss": 0.2895,
258992
+ "step": 125460
258993
+ },
258994
+ {
258995
+ "epoch": 1011.15,
258996
+ "learning_rate": 8.006153846153847e-06,
258997
+ "loss": 0.3673,
258998
+ "step": 125465
258999
+ },
259000
+ {
259001
+ "epoch": 1011.19,
259002
+ "learning_rate": 8.00607371794872e-06,
259003
+ "loss": 0.761,
259004
+ "step": 125470
259005
+ },
259006
+ {
259007
+ "epoch": 1011.23,
259008
+ "learning_rate": 8.00599358974359e-06,
259009
+ "loss": 0.7003,
259010
+ "step": 125475
259011
+ },
259012
+ {
259013
+ "epoch": 1011.27,
259014
+ "learning_rate": 8.005913461538463e-06,
259015
+ "loss": 0.2643,
259016
+ "step": 125480
259017
+ },
259018
+ {
259019
+ "epoch": 1011.31,
259020
+ "learning_rate": 8.005833333333334e-06,
259021
+ "loss": 0.3928,
259022
+ "step": 125485
259023
+ },
259024
+ {
259025
+ "epoch": 1011.35,
259026
+ "learning_rate": 8.005753205128205e-06,
259027
+ "loss": 0.3336,
259028
+ "step": 125490
259029
+ },
259030
+ {
259031
+ "epoch": 1011.4,
259032
+ "learning_rate": 8.005673076923077e-06,
259033
+ "loss": 0.8739,
259034
+ "step": 125495
259035
+ },
259036
+ {
259037
+ "epoch": 1011.44,
259038
+ "learning_rate": 8.00559294871795e-06,
259039
+ "loss": 0.625,
259040
+ "step": 125500
259041
+ },
259042
+ {
259043
+ "epoch": 1011.48,
259044
+ "learning_rate": 8.005512820512821e-06,
259045
+ "loss": 0.3288,
259046
+ "step": 125505
259047
+ },
259048
+ {
259049
+ "epoch": 1011.52,
259050
+ "learning_rate": 8.005432692307692e-06,
259051
+ "loss": 0.3602,
259052
+ "step": 125510
259053
+ },
259054
+ {
259055
+ "epoch": 1011.56,
259056
+ "learning_rate": 8.005352564102564e-06,
259057
+ "loss": 0.4349,
259058
+ "step": 125515
259059
+ },
259060
+ {
259061
+ "epoch": 1011.6,
259062
+ "learning_rate": 8.005272435897437e-06,
259063
+ "loss": 0.8922,
259064
+ "step": 125520
259065
+ },
259066
+ {
259067
+ "epoch": 1011.64,
259068
+ "learning_rate": 8.005192307692308e-06,
259069
+ "loss": 0.6377,
259070
+ "step": 125525
259071
+ },
259072
+ {
259073
+ "epoch": 1011.68,
259074
+ "learning_rate": 8.00511217948718e-06,
259075
+ "loss": 0.3145,
259076
+ "step": 125530
259077
+ },
259078
+ {
259079
+ "epoch": 1011.72,
259080
+ "learning_rate": 8.005032051282053e-06,
259081
+ "loss": 0.3291,
259082
+ "step": 125535
259083
+ },
259084
+ {
259085
+ "epoch": 1011.76,
259086
+ "learning_rate": 8.004951923076924e-06,
259087
+ "loss": 0.422,
259088
+ "step": 125540
259089
+ },
259090
+ {
259091
+ "epoch": 1011.8,
259092
+ "learning_rate": 8.004871794871795e-06,
259093
+ "loss": 0.924,
259094
+ "step": 125545
259095
+ },
259096
+ {
259097
+ "epoch": 1011.84,
259098
+ "learning_rate": 8.004791666666667e-06,
259099
+ "loss": 0.7038,
259100
+ "step": 125550
259101
+ },
259102
+ {
259103
+ "epoch": 1011.88,
259104
+ "learning_rate": 8.00471153846154e-06,
259105
+ "loss": 0.2517,
259106
+ "step": 125555
259107
+ },
259108
+ {
259109
+ "epoch": 1011.92,
259110
+ "learning_rate": 8.004631410256411e-06,
259111
+ "loss": 0.273,
259112
+ "step": 125560
259113
+ },
259114
+ {
259115
+ "epoch": 1011.96,
259116
+ "learning_rate": 8.004551282051282e-06,
259117
+ "loss": 0.4353,
259118
+ "step": 125565
259119
+ },
259120
+ {
259121
+ "epoch": 1012.0,
259122
+ "learning_rate": 8.004471153846155e-06,
259123
+ "loss": 1.0249,
259124
+ "step": 125570
259125
+ },
259126
+ {
259127
+ "epoch": 1012.0,
259128
+ "eval_loss": 0.4390537440776825,
259129
+ "eval_runtime": 36.2511,
259130
+ "eval_samples_per_second": 23.144,
259131
+ "eval_steps_per_second": 0.745,
259132
+ "eval_wer": 0.1779642141213156,
259133
+ "step": 125570
259134
+ },
259135
+ {
259136
+ "epoch": 1012.04,
259137
+ "learning_rate": 8.004391025641027e-06,
259138
+ "loss": 0.3656,
259139
+ "step": 125575
259140
+ },
259141
+ {
259142
+ "epoch": 1012.08,
259143
+ "learning_rate": 8.004310897435898e-06,
259144
+ "loss": 0.2624,
259145
+ "step": 125580
259146
+ },
259147
+ {
259148
+ "epoch": 1012.12,
259149
+ "learning_rate": 8.00423076923077e-06,
259150
+ "loss": 0.2921,
259151
+ "step": 125585
259152
+ },
259153
+ {
259154
+ "epoch": 1012.16,
259155
+ "learning_rate": 8.004150641025643e-06,
259156
+ "loss": 0.403,
259157
+ "step": 125590
259158
+ },
259159
+ {
259160
+ "epoch": 1012.2,
259161
+ "learning_rate": 8.004070512820514e-06,
259162
+ "loss": 1.3023,
259163
+ "step": 125595
259164
+ },
259165
+ {
259166
+ "epoch": 1012.24,
259167
+ "learning_rate": 8.003990384615385e-06,
259168
+ "loss": 0.3312,
259169
+ "step": 125600
259170
+ },
259171
+ {
259172
+ "epoch": 1012.28,
259173
+ "learning_rate": 8.003910256410257e-06,
259174
+ "loss": 0.2541,
259175
+ "step": 125605
259176
+ },
259177
+ {
259178
+ "epoch": 1012.32,
259179
+ "learning_rate": 8.00383012820513e-06,
259180
+ "loss": 0.3223,
259181
+ "step": 125610
259182
+ },
259183
+ {
259184
+ "epoch": 1012.36,
259185
+ "learning_rate": 8.00375e-06,
259186
+ "loss": 0.3871,
259187
+ "step": 125615
259188
+ },
259189
+ {
259190
+ "epoch": 1012.4,
259191
+ "learning_rate": 8.003669871794872e-06,
259192
+ "loss": 1.08,
259193
+ "step": 125620
259194
+ },
259195
+ {
259196
+ "epoch": 1012.44,
259197
+ "learning_rate": 8.003589743589745e-06,
259198
+ "loss": 0.3281,
259199
+ "step": 125625
259200
+ },
259201
+ {
259202
+ "epoch": 1012.48,
259203
+ "learning_rate": 8.003509615384615e-06,
259204
+ "loss": 0.2986,
259205
+ "step": 125630
259206
+ },
259207
+ {
259208
+ "epoch": 1012.52,
259209
+ "learning_rate": 8.003429487179488e-06,
259210
+ "loss": 0.2999,
259211
+ "step": 125635
259212
+ },
259213
+ {
259214
+ "epoch": 1012.56,
259215
+ "learning_rate": 8.00334935897436e-06,
259216
+ "loss": 0.3972,
259217
+ "step": 125640
259218
+ },
259219
+ {
259220
+ "epoch": 1012.6,
259221
+ "learning_rate": 8.003269230769231e-06,
259222
+ "loss": 1.2367,
259223
+ "step": 125645
259224
+ },
259225
+ {
259226
+ "epoch": 1012.64,
259227
+ "learning_rate": 8.003189102564102e-06,
259228
+ "loss": 0.314,
259229
+ "step": 125650
259230
+ },
259231
+ {
259232
+ "epoch": 1012.68,
259233
+ "learning_rate": 8.003108974358975e-06,
259234
+ "loss": 0.3581,
259235
+ "step": 125655
259236
+ },
259237
+ {
259238
+ "epoch": 1012.72,
259239
+ "learning_rate": 8.003028846153847e-06,
259240
+ "loss": 0.2869,
259241
+ "step": 125660
259242
+ },
259243
+ {
259244
+ "epoch": 1012.76,
259245
+ "learning_rate": 8.002948717948718e-06,
259246
+ "loss": 0.5337,
259247
+ "step": 125665
259248
+ },
259249
+ {
259250
+ "epoch": 1012.8,
259251
+ "learning_rate": 8.002868589743591e-06,
259252
+ "loss": 1.2025,
259253
+ "step": 125670
259254
+ },
259255
+ {
259256
+ "epoch": 1012.84,
259257
+ "learning_rate": 8.002788461538462e-06,
259258
+ "loss": 0.2866,
259259
+ "step": 125675
259260
+ },
259261
+ {
259262
+ "epoch": 1012.88,
259263
+ "learning_rate": 8.002708333333334e-06,
259264
+ "loss": 0.2661,
259265
+ "step": 125680
259266
+ },
259267
+ {
259268
+ "epoch": 1012.92,
259269
+ "learning_rate": 8.002628205128205e-06,
259270
+ "loss": 0.3406,
259271
+ "step": 125685
259272
+ },
259273
+ {
259274
+ "epoch": 1012.96,
259275
+ "learning_rate": 8.002548076923078e-06,
259276
+ "loss": 0.4782,
259277
+ "step": 125690
259278
+ },
259279
+ {
259280
+ "epoch": 1013.0,
259281
+ "eval_loss": 0.32623913884162903,
259282
+ "eval_runtime": 48.1801,
259283
+ "eval_samples_per_second": 17.414,
259284
+ "eval_steps_per_second": 0.56,
259285
+ "eval_wer": 0.18181818181818182,
259286
+ "step": 125694
259287
+ },
259288
+ {
259289
+ "epoch": 1005.01,
259290
+ "learning_rate": 8.00246794871795e-06,
259291
+ "loss": 0.5473,
259292
+ "step": 125695
259293
+ },
259294
+ {
259295
+ "epoch": 1005.05,
259296
+ "learning_rate": 8.002387820512821e-06,
259297
+ "loss": 0.3088,
259298
+ "step": 125700
259299
+ },
259300
+ {
259301
+ "epoch": 1005.09,
259302
+ "learning_rate": 8.002307692307692e-06,
259303
+ "loss": 0.304,
259304
+ "step": 125705
259305
+ },
259306
+ {
259307
+ "epoch": 1005.13,
259308
+ "learning_rate": 8.002227564102565e-06,
259309
+ "loss": 0.3238,
259310
+ "step": 125710
259311
+ },
259312
+ {
259313
+ "epoch": 1005.17,
259314
+ "learning_rate": 8.002147435897437e-06,
259315
+ "loss": 0.5224,
259316
+ "step": 125715
259317
+ },
259318
+ {
259319
+ "epoch": 1005.21,
259320
+ "learning_rate": 8.002067307692308e-06,
259321
+ "loss": 1.0523,
259322
+ "step": 125720
259323
+ },
259324
+ {
259325
+ "epoch": 1005.25,
259326
+ "learning_rate": 8.001987179487181e-06,
259327
+ "loss": 0.3129,
259328
+ "step": 125725
259329
+ },
259330
+ {
259331
+ "epoch": 1005.29,
259332
+ "learning_rate": 8.001907051282052e-06,
259333
+ "loss": 0.3159,
259334
+ "step": 125730
259335
+ },
259336
+ {
259337
+ "epoch": 1005.33,
259338
+ "learning_rate": 8.001826923076924e-06,
259339
+ "loss": 0.3222,
259340
+ "step": 125735
259341
+ },
259342
+ {
259343
+ "epoch": 1005.37,
259344
+ "learning_rate": 8.001746794871795e-06,
259345
+ "loss": 0.5416,
259346
+ "step": 125740
259347
+ },
259348
+ {
259349
+ "epoch": 1005.41,
259350
+ "learning_rate": 8.001666666666668e-06,
259351
+ "loss": 0.9692,
259352
+ "step": 125745
259353
+ },
259354
+ {
259355
+ "epoch": 1005.45,
259356
+ "learning_rate": 8.001586538461538e-06,
259357
+ "loss": 0.2915,
259358
+ "step": 125750
259359
+ },
259360
+ {
259361
+ "epoch": 1005.49,
259362
+ "learning_rate": 8.001506410256411e-06,
259363
+ "loss": 0.243,
259364
+ "step": 125755
259365
+ },
259366
+ {
259367
+ "epoch": 1005.53,
259368
+ "learning_rate": 8.001426282051282e-06,
259369
+ "loss": 0.3572,
259370
+ "step": 125760
259371
+ },
259372
+ {
259373
+ "epoch": 1005.57,
259374
+ "learning_rate": 8.001346153846154e-06,
259375
+ "loss": 0.5895,
259376
+ "step": 125765
259377
+ },
259378
+ {
259379
+ "epoch": 1005.61,
259380
+ "learning_rate": 8.001266025641027e-06,
259381
+ "loss": 1.1071,
259382
+ "step": 125770
259383
+ },
259384
+ {
259385
+ "epoch": 1005.65,
259386
+ "learning_rate": 8.001185897435898e-06,
259387
+ "loss": 0.2927,
259388
+ "step": 125775
259389
+ },
259390
+ {
259391
+ "epoch": 1005.69,
259392
+ "learning_rate": 8.00110576923077e-06,
259393
+ "loss": 0.336,
259394
+ "step": 125780
259395
+ },
259396
+ {
259397
+ "epoch": 1005.73,
259398
+ "learning_rate": 8.00102564102564e-06,
259399
+ "loss": 0.3185,
259400
+ "step": 125785
259401
+ },
259402
+ {
259403
+ "epoch": 1005.77,
259404
+ "learning_rate": 8.000945512820514e-06,
259405
+ "loss": 0.5286,
259406
+ "step": 125790
259407
+ },
259408
+ {
259409
+ "epoch": 1005.81,
259410
+ "learning_rate": 8.000865384615385e-06,
259411
+ "loss": 1.0564,
259412
+ "step": 125795
259413
+ },
259414
+ {
259415
+ "epoch": 1005.85,
259416
+ "learning_rate": 8.000785256410257e-06,
259417
+ "loss": 0.351,
259418
+ "step": 125800
259419
+ },
259420
+ {
259421
+ "epoch": 1005.89,
259422
+ "learning_rate": 8.000705128205128e-06,
259423
+ "loss": 0.3298,
259424
+ "step": 125805
259425
+ },
259426
+ {
259427
+ "epoch": 1005.93,
259428
+ "learning_rate": 8.000625000000001e-06,
259429
+ "loss": 0.338,
259430
+ "step": 125810
259431
+ },
259432
+ {
259433
+ "epoch": 1005.97,
259434
+ "learning_rate": 8.000544871794872e-06,
259435
+ "loss": 0.488,
259436
+ "step": 125815
259437
+ },
259438
+ {
259439
+ "epoch": 1006.0,
259440
+ "eval_loss": 0.4103972613811493,
259441
+ "eval_runtime": 35.6053,
259442
+ "eval_samples_per_second": 23.62,
259443
+ "eval_steps_per_second": 0.758,
259444
+ "eval_wer": 0.17794615271443284,
259445
+ "step": 125819
259446
+ },
259447
+ {
259448
+ "epoch": 1014.01,
259449
+ "learning_rate": 8.000464743589744e-06,
259450
+ "loss": 0.2939,
259451
+ "step": 125820
259452
+ },
259453
+ {
259454
+ "epoch": 1014.05,
259455
+ "learning_rate": 8.000384615384617e-06,
259456
+ "loss": 0.3032,
259457
+ "step": 125825
259458
+ },
259459
+ {
259460
+ "epoch": 1014.09,
259461
+ "learning_rate": 8.000304487179488e-06,
259462
+ "loss": 0.2653,
259463
+ "step": 125830
259464
+ },
259465
+ {
259466
+ "epoch": 1014.13,
259467
+ "learning_rate": 8.00022435897436e-06,
259468
+ "loss": 0.3178,
259469
+ "step": 125835
259470
+ },
259471
+ {
259472
+ "epoch": 1014.17,
259473
+ "learning_rate": 8.000144230769231e-06,
259474
+ "loss": 0.5021,
259475
+ "step": 125840
259476
+ },
259477
+ {
259478
+ "epoch": 1014.21,
259479
+ "learning_rate": 8.000064102564104e-06,
259480
+ "loss": 1.0234,
259481
+ "step": 125845
259482
+ },
259483
+ {
259484
+ "epoch": 1014.25,
259485
+ "learning_rate": 7.999983974358975e-06,
259486
+ "loss": 0.3039,
259487
+ "step": 125850
259488
+ },
259489
+ {
259490
+ "epoch": 1014.29,
259491
+ "learning_rate": 7.999903846153847e-06,
259492
+ "loss": 0.3366,
259493
+ "step": 125855
259494
+ },
259495
+ {
259496
+ "epoch": 1014.33,
259497
+ "learning_rate": 7.999823717948718e-06,
259498
+ "loss": 0.3143,
259499
+ "step": 125860
259500
+ },
259501
+ {
259502
+ "epoch": 1014.37,
259503
+ "learning_rate": 7.999743589743591e-06,
259504
+ "loss": 0.6841,
259505
+ "step": 125865
259506
+ },
259507
+ {
259508
+ "epoch": 1014.41,
259509
+ "learning_rate": 7.999663461538462e-06,
259510
+ "loss": 1.2289,
259511
+ "step": 125870
259512
+ },
259513
+ {
259514
+ "epoch": 1014.45,
259515
+ "learning_rate": 7.999583333333334e-06,
259516
+ "loss": 0.3278,
259517
+ "step": 125875
259518
+ },
259519
+ {
259520
+ "epoch": 1014.49,
259521
+ "learning_rate": 7.999503205128207e-06,
259522
+ "loss": 0.2565,
259523
+ "step": 125880
259524
+ },
259525
+ {
259526
+ "epoch": 1014.53,
259527
+ "learning_rate": 7.999423076923078e-06,
259528
+ "loss": 0.3343,
259529
+ "step": 125885
259530
+ },
259531
+ {
259532
+ "epoch": 1014.57,
259533
+ "learning_rate": 7.99934294871795e-06,
259534
+ "loss": 0.5523,
259535
+ "step": 125890
259536
+ },
259537
+ {
259538
+ "epoch": 1014.61,
259539
+ "learning_rate": 7.999262820512821e-06,
259540
+ "loss": 1.0618,
259541
+ "step": 125895
259542
+ },
259543
+ {
259544
+ "epoch": 1014.65,
259545
+ "learning_rate": 7.999182692307694e-06,
259546
+ "loss": 0.3569,
259547
+ "step": 125900
259548
+ },
259549
+ {
259550
+ "epoch": 1014.69,
259551
+ "learning_rate": 7.999102564102564e-06,
259552
+ "loss": 0.3683,
259553
+ "step": 125905
259554
+ },
259555
+ {
259556
+ "epoch": 1014.73,
259557
+ "learning_rate": 7.999022435897437e-06,
259558
+ "loss": 0.3664,
259559
+ "step": 125910
259560
+ },
259561
+ {
259562
+ "epoch": 1014.77,
259563
+ "learning_rate": 7.998942307692308e-06,
259564
+ "loss": 0.5853,
259565
+ "step": 125915
259566
+ },
259567
+ {
259568
+ "epoch": 1014.81,
259569
+ "learning_rate": 7.99886217948718e-06,
259570
+ "loss": 1.0467,
259571
+ "step": 125920
259572
+ },
259573
+ {
259574
+ "epoch": 1014.85,
259575
+ "learning_rate": 7.998782051282052e-06,
259576
+ "loss": 0.3293,
259577
+ "step": 125925
259578
+ },
259579
+ {
259580
+ "epoch": 1014.89,
259581
+ "learning_rate": 7.998701923076924e-06,
259582
+ "loss": 0.2739,
259583
+ "step": 125930
259584
+ },
259585
+ {
259586
+ "epoch": 1014.93,
259587
+ "learning_rate": 7.998621794871795e-06,
259588
+ "loss": 0.3778,
259589
+ "step": 125935
259590
+ },
259591
+ {
259592
+ "epoch": 1014.97,
259593
+ "learning_rate": 7.998541666666666e-06,
259594
+ "loss": 0.5238,
259595
+ "step": 125940
259596
+ },
259597
+ {
259598
+ "epoch": 1015.0,
259599
+ "eval_loss": 0.36781224608421326,
259600
+ "eval_runtime": 36.1928,
259601
+ "eval_samples_per_second": 23.237,
259602
+ "eval_steps_per_second": 0.746,
259603
+ "eval_wer": 0.18371311169219276,
259604
+ "step": 125943
259605
+ },
259606
+ {
259607
+ "epoch": 1007.02,
259608
+ "learning_rate": 7.99846153846154e-06,
259609
+ "loss": 0.3979,
259610
+ "step": 125945
259611
+ },
259612
+ {
259613
+ "epoch": 1007.06,
259614
+ "learning_rate": 7.998381410256411e-06,
259615
+ "loss": 0.3103,
259616
+ "step": 125950
259617
+ },
259618
+ {
259619
+ "epoch": 1007.1,
259620
+ "learning_rate": 7.998301282051282e-06,
259621
+ "loss": 0.2781,
259622
+ "step": 125955
259623
+ },
259624
+ {
259625
+ "epoch": 1007.14,
259626
+ "learning_rate": 7.998221153846154e-06,
259627
+ "loss": 0.3187,
259628
+ "step": 125960
259629
+ },
259630
+ {
259631
+ "epoch": 1007.18,
259632
+ "learning_rate": 7.998141025641027e-06,
259633
+ "loss": 0.6326,
259634
+ "step": 125965
259635
+ },
259636
+ {
259637
+ "epoch": 1007.22,
259638
+ "learning_rate": 7.998060897435898e-06,
259639
+ "loss": 1.0426,
259640
+ "step": 125970
259641
+ },
259642
+ {
259643
+ "epoch": 1007.26,
259644
+ "learning_rate": 7.99798076923077e-06,
259645
+ "loss": 0.2652,
259646
+ "step": 125975
259647
+ },
259648
+ {
259649
+ "epoch": 1007.3,
259650
+ "learning_rate": 7.997900641025642e-06,
259651
+ "loss": 0.2786,
259652
+ "step": 125980
259653
+ },
259654
+ {
259655
+ "epoch": 1007.34,
259656
+ "learning_rate": 7.997820512820514e-06,
259657
+ "loss": 0.406,
259658
+ "step": 125985
259659
+ },
259660
+ {
259661
+ "epoch": 1007.38,
259662
+ "learning_rate": 7.997740384615385e-06,
259663
+ "loss": 0.6165,
259664
+ "step": 125990
259665
+ },
259666
+ {
259667
+ "epoch": 1007.42,
259668
+ "learning_rate": 7.997660256410257e-06,
259669
+ "loss": 1.0945,
259670
+ "step": 125995
259671
+ },
259672
+ {
259673
+ "epoch": 1007.46,
259674
+ "learning_rate": 7.99758012820513e-06,
259675
+ "loss": 0.2789,
259676
+ "step": 126000
259677
+ },
259678
+ {
259679
+ "epoch": 1007.5,
259680
+ "learning_rate": 7.997500000000001e-06,
259681
+ "loss": 0.2985,
259682
+ "step": 126005
259683
+ },
259684
+ {
259685
+ "epoch": 1007.54,
259686
+ "learning_rate": 7.997419871794872e-06,
259687
+ "loss": 0.2813,
259688
+ "step": 126010
259689
+ },
259690
+ {
259691
+ "epoch": 1007.58,
259692
+ "learning_rate": 7.997339743589744e-06,
259693
+ "loss": 0.6877,
259694
+ "step": 126015
259695
+ },
259696
+ {
259697
+ "epoch": 1007.62,
259698
+ "learning_rate": 7.997259615384617e-06,
259699
+ "loss": 0.9805,
259700
+ "step": 126020
259701
+ },
259702
+ {
259703
+ "epoch": 1007.66,
259704
+ "learning_rate": 7.997179487179488e-06,
259705
+ "loss": 0.2643,
259706
+ "step": 126025
259707
+ },
259708
+ {
259709
+ "epoch": 1007.7,
259710
+ "learning_rate": 7.99709935897436e-06,
259711
+ "loss": 0.2754,
259712
+ "step": 126030
259713
+ },
259714
+ {
259715
+ "epoch": 1007.74,
259716
+ "learning_rate": 7.997019230769232e-06,
259717
+ "loss": 0.3053,
259718
+ "step": 126035
259719
+ },
259720
+ {
259721
+ "epoch": 1007.78,
259722
+ "learning_rate": 7.996939102564104e-06,
259723
+ "loss": 0.6208,
259724
+ "step": 126040
259725
+ },
259726
+ {
259727
+ "epoch": 1007.82,
259728
+ "learning_rate": 7.996858974358975e-06,
259729
+ "loss": 1.0308,
259730
+ "step": 126045
259731
+ },
259732
+ {
259733
+ "epoch": 1007.86,
259734
+ "learning_rate": 7.996778846153847e-06,
259735
+ "loss": 0.2589,
259736
+ "step": 126050
259737
+ },
259738
+ {
259739
+ "epoch": 1007.9,
259740
+ "learning_rate": 7.99669871794872e-06,
259741
+ "loss": 0.2949,
259742
+ "step": 126055
259743
+ },
259744
+ {
259745
+ "epoch": 1007.94,
259746
+ "learning_rate": 7.99661858974359e-06,
259747
+ "loss": 0.316,
259748
+ "step": 126060
259749
+ },
259750
+ {
259751
+ "epoch": 1007.98,
259752
+ "learning_rate": 7.996538461538462e-06,
259753
+ "loss": 0.5452,
259754
+ "step": 126065
259755
+ },
259756
+ {
259757
+ "epoch": 1008.0,
259758
+ "eval_loss": 0.4000602066516876,
259759
+ "eval_runtime": 36.515,
259760
+ "eval_samples_per_second": 23.032,
259761
+ "eval_steps_per_second": 0.739,
259762
+ "eval_wer": 0.18030087568295786,
259763
+ "step": 126068
259764
  }
259765
  ],
259766
+ "max_steps": 625000,
259767
  "num_train_epochs": 5000,
259768
+ "total_flos": 3.547605377048925e+20,
259769
  "trial_name": null,
259770
  "trial_params": null
259771
  }
model-bin/finetune/base/{checkpoint-125446 β†’ checkpoint-126068}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630146705.8776062/events.out.tfevents.1630146705.86bb0ddabf9b.4092.121 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5022a9069cb186dbe772a5b12206d0f5add6f2140b964a82d4f5988b8716ce8f
3
+ size 4194
model-bin/finetune/base/log/1630147090.9692104/events.out.tfevents.1630147090.86bb0ddabf9b.4092.123 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:982f99c0e956f481e5fc347ef161d31d63e408957501f6c1ec947beadfa90553
3
+ size 4194
model-bin/finetune/base/log/1630147594.250904/events.out.tfevents.1630147594.86bb0ddabf9b.4092.125 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8b93e9700ca21911643d51dac49e768b32cf23e330ef4a59fb1d41200ceb663
3
+ size 4194
model-bin/finetune/base/log/1630147976.392592/events.out.tfevents.1630147976.86bb0ddabf9b.4092.127 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f30bfb48fce824788b39b728a038285ad4abede0cc55779a95d5f92ddccfea8c
3
+ size 4194
model-bin/finetune/base/log/1630148364.6244335/events.out.tfevents.1630148364.86bb0ddabf9b.4092.129 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08ee3ab483c8ec102883e3bd12c6b718989b5c7649f48dc2055817ec21d331e5
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630146705.86bb0ddabf9b.4092.120 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5bbf253060e791d99008d225cae68311e731bfbefaca264082bfef66b0920dd
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630147090.86bb0ddabf9b.4092.122 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67bd7e33213924bc59d9ed9c4eadea66e0729ba0d99643999e8961d3ab56b493
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1630147594.86bb0ddabf9b.4092.124 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5511a402159ad90b2f655ce38138505d72085dc9f06b2370a379b8bde9112476
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630147976.86bb0ddabf9b.4092.126 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfc30050d1853a87afb646d8049f4ae0412bf14a163343bff1e2bab7c2cb7daf
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630148364.86bb0ddabf9b.4092.128 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcb1d439a5450a64819aa55a96fbcc2bc3bd14c2128b8ebea5e7a7d1c13a698c
3
+ size 8622