Check commited on
Commit
825e687
Β·
1 Parent(s): fcc4b9c

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/trainer_state.json +798 -3
  9. model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629843978.3964214/events.out.tfevents.1629843978.c435e1c5ee04.920.241 +3 -0
  11. model-bin/finetune/base/log/1629844632.7215848/events.out.tfevents.1629844632.c435e1c5ee04.920.243 +3 -0
  12. model-bin/finetune/base/log/1629845266.9609022/events.out.tfevents.1629845266.c435e1c5ee04.920.245 +3 -0
  13. model-bin/finetune/base/log/1629845904.6603022/events.out.tfevents.1629845904.c435e1c5ee04.920.247 +3 -0
  14. model-bin/finetune/base/log/1629846554.1512308/events.out.tfevents.1629846554.c435e1c5ee04.920.249 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629843978.c435e1c5ee04.920.240 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629844632.c435e1c5ee04.920.242 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629845266.c435e1c5ee04.920.244 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629845904.c435e1c5ee04.920.246 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629846554.c435e1c5ee04.920.248 +3 -0
model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d36abf3a2c5bf4d3050eeafe51b7abe8283b26d662aa895d1c4b2a29d7a740e5
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f97b4fc6c219db9d3afa4880cdcccd79399ebe284feffb6e2efe98ae3ed6a268
3
  size 722165393
model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c93ff5fcd4e89802ee95abc960093c807c19054952e8b77db72016e9b0895371
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f5e391fa4f938d866a965b45f82eebde6c0b705b2466903a9e82f6e9f4212c2
3
  size 377909911
model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3184289183c73c4a5f506139f7799f696e37e4b880a814a87374ea84bfb05744
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb87bc92aafb41622424ca10817f949491f3f5e4b525515986b7a6fc2e6b3641
3
  size 14503
model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b48d35ca13554d6b818ee40b007b767588033ef3615fae55bbe7980615d8fa2
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cc9397ac385b0f2da90d122ad975747c5aef6a7097f14cfe46a57e0bc037043
3
  size 559
model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eab293b67371458c5eedb98ca7c7368a43de31a3f787717c48d0bf8927ac86f1
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbc1e6d224edc3cba35b1ab2f1a230355e5f6ebaabb30a9e0c807f2453ae39b9
3
  size 623
model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.18525332578545145,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-68444",
4
- "epoch": 549.9960159362549,
5
- "global_step": 68818,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -186921,11 +186921,806 @@
186921
  "eval_steps_per_second": 0.688,
186922
  "eval_wer": 0.1943950177935943,
186923
  "step": 68818
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
186924
  }
186925
  ],
186926
  "max_steps": 625000,
186927
  "num_train_epochs": 5000,
186928
- "total_flos": 1.936554826303671e+20,
186929
  "trial_name": null,
186930
  "trial_params": null
186931
  }
 
1
  {
2
  "best_metric": 0.18525332578545145,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-68444",
4
+ "epoch": 555.0,
5
+ "global_step": 69440,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
186921
  "eval_steps_per_second": 0.688,
186922
  "eval_wer": 0.1943950177935943,
186923
  "step": 68818
186924
+ },
186925
+ {
186926
+ "epoch": 550.02,
186927
+ "learning_rate": 8.913557692307693e-06,
186928
+ "loss": 0.4203,
186929
+ "step": 68820
186930
+ },
186931
+ {
186932
+ "epoch": 550.06,
186933
+ "learning_rate": 8.913477564102564e-06,
186934
+ "loss": 0.3611,
186935
+ "step": 68825
186936
+ },
186937
+ {
186938
+ "epoch": 550.1,
186939
+ "learning_rate": 8.913397435897436e-06,
186940
+ "loss": 0.343,
186941
+ "step": 68830
186942
+ },
186943
+ {
186944
+ "epoch": 550.14,
186945
+ "learning_rate": 8.913317307692309e-06,
186946
+ "loss": 0.3587,
186947
+ "step": 68835
186948
+ },
186949
+ {
186950
+ "epoch": 550.18,
186951
+ "learning_rate": 8.91323717948718e-06,
186952
+ "loss": 0.5963,
186953
+ "step": 68840
186954
+ },
186955
+ {
186956
+ "epoch": 550.22,
186957
+ "learning_rate": 8.913157051282052e-06,
186958
+ "loss": 1.1357,
186959
+ "step": 68845
186960
+ },
186961
+ {
186962
+ "epoch": 550.26,
186963
+ "learning_rate": 8.913076923076925e-06,
186964
+ "loss": 0.2899,
186965
+ "step": 68850
186966
+ },
186967
+ {
186968
+ "epoch": 550.3,
186969
+ "learning_rate": 8.912996794871796e-06,
186970
+ "loss": 0.3763,
186971
+ "step": 68855
186972
+ },
186973
+ {
186974
+ "epoch": 550.34,
186975
+ "learning_rate": 8.912916666666667e-06,
186976
+ "loss": 0.3944,
186977
+ "step": 68860
186978
+ },
186979
+ {
186980
+ "epoch": 550.38,
186981
+ "learning_rate": 8.912836538461539e-06,
186982
+ "loss": 0.7077,
186983
+ "step": 68865
186984
+ },
186985
+ {
186986
+ "epoch": 550.42,
186987
+ "learning_rate": 8.912756410256412e-06,
186988
+ "loss": 1.1302,
186989
+ "step": 68870
186990
+ },
186991
+ {
186992
+ "epoch": 550.46,
186993
+ "learning_rate": 8.912676282051283e-06,
186994
+ "loss": 0.2758,
186995
+ "step": 68875
186996
+ },
186997
+ {
186998
+ "epoch": 550.5,
186999
+ "learning_rate": 8.912596153846155e-06,
187000
+ "loss": 0.3103,
187001
+ "step": 68880
187002
+ },
187003
+ {
187004
+ "epoch": 550.54,
187005
+ "learning_rate": 8.912516025641026e-06,
187006
+ "loss": 0.4578,
187007
+ "step": 68885
187008
+ },
187009
+ {
187010
+ "epoch": 550.58,
187011
+ "learning_rate": 8.912435897435899e-06,
187012
+ "loss": 0.6749,
187013
+ "step": 68890
187014
+ },
187015
+ {
187016
+ "epoch": 550.62,
187017
+ "learning_rate": 8.91235576923077e-06,
187018
+ "loss": 1.0549,
187019
+ "step": 68895
187020
+ },
187021
+ {
187022
+ "epoch": 550.66,
187023
+ "learning_rate": 8.912275641025642e-06,
187024
+ "loss": 0.3558,
187025
+ "step": 68900
187026
+ },
187027
+ {
187028
+ "epoch": 550.7,
187029
+ "learning_rate": 8.912195512820515e-06,
187030
+ "loss": 0.3126,
187031
+ "step": 68905
187032
+ },
187033
+ {
187034
+ "epoch": 550.74,
187035
+ "learning_rate": 8.912115384615384e-06,
187036
+ "loss": 0.4069,
187037
+ "step": 68910
187038
+ },
187039
+ {
187040
+ "epoch": 550.78,
187041
+ "learning_rate": 8.912035256410257e-06,
187042
+ "loss": 0.6448,
187043
+ "step": 68915
187044
+ },
187045
+ {
187046
+ "epoch": 550.82,
187047
+ "learning_rate": 8.911955128205129e-06,
187048
+ "loss": 1.0045,
187049
+ "step": 68920
187050
+ },
187051
+ {
187052
+ "epoch": 550.86,
187053
+ "learning_rate": 8.911875e-06,
187054
+ "loss": 0.3323,
187055
+ "step": 68925
187056
+ },
187057
+ {
187058
+ "epoch": 550.9,
187059
+ "learning_rate": 8.911794871794871e-06,
187060
+ "loss": 0.3236,
187061
+ "step": 68930
187062
+ },
187063
+ {
187064
+ "epoch": 550.94,
187065
+ "learning_rate": 8.911714743589745e-06,
187066
+ "loss": 0.3786,
187067
+ "step": 68935
187068
+ },
187069
+ {
187070
+ "epoch": 550.98,
187071
+ "learning_rate": 8.911634615384616e-06,
187072
+ "loss": 0.6312,
187073
+ "step": 68940
187074
+ },
187075
+ {
187076
+ "epoch": 551.0,
187077
+ "eval_loss": 0.3938581943511963,
187078
+ "eval_runtime": 40.8487,
187079
+ "eval_samples_per_second": 20.515,
187080
+ "eval_steps_per_second": 0.661,
187081
+ "eval_wer": 0.19235865724381626,
187082
+ "step": 68943
187083
+ },
187084
+ {
187085
+ "epoch": 555.02,
187086
+ "learning_rate": 8.911554487179487e-06,
187087
+ "loss": 0.6145,
187088
+ "step": 68945
187089
+ },
187090
+ {
187091
+ "epoch": 555.06,
187092
+ "learning_rate": 8.91147435897436e-06,
187093
+ "loss": 0.3448,
187094
+ "step": 68950
187095
+ },
187096
+ {
187097
+ "epoch": 555.1,
187098
+ "learning_rate": 8.911394230769232e-06,
187099
+ "loss": 0.3279,
187100
+ "step": 68955
187101
+ },
187102
+ {
187103
+ "epoch": 555.14,
187104
+ "learning_rate": 8.911314102564103e-06,
187105
+ "loss": 0.3365,
187106
+ "step": 68960
187107
+ },
187108
+ {
187109
+ "epoch": 555.18,
187110
+ "learning_rate": 8.911233974358974e-06,
187111
+ "loss": 0.7031,
187112
+ "step": 68965
187113
+ },
187114
+ {
187115
+ "epoch": 555.22,
187116
+ "learning_rate": 8.911153846153847e-06,
187117
+ "loss": 1.028,
187118
+ "step": 68970
187119
+ },
187120
+ {
187121
+ "epoch": 555.26,
187122
+ "learning_rate": 8.911073717948719e-06,
187123
+ "loss": 0.3392,
187124
+ "step": 68975
187125
+ },
187126
+ {
187127
+ "epoch": 555.3,
187128
+ "learning_rate": 8.91099358974359e-06,
187129
+ "loss": 0.359,
187130
+ "step": 68980
187131
+ },
187132
+ {
187133
+ "epoch": 555.34,
187134
+ "learning_rate": 8.910913461538462e-06,
187135
+ "loss": 0.3914,
187136
+ "step": 68985
187137
+ },
187138
+ {
187139
+ "epoch": 555.38,
187140
+ "learning_rate": 8.910833333333335e-06,
187141
+ "loss": 0.6924,
187142
+ "step": 68990
187143
+ },
187144
+ {
187145
+ "epoch": 555.42,
187146
+ "learning_rate": 8.910753205128206e-06,
187147
+ "loss": 1.0995,
187148
+ "step": 68995
187149
+ },
187150
+ {
187151
+ "epoch": 555.46,
187152
+ "learning_rate": 8.910673076923077e-06,
187153
+ "loss": 0.3136,
187154
+ "step": 69000
187155
+ },
187156
+ {
187157
+ "epoch": 555.5,
187158
+ "learning_rate": 8.91059294871795e-06,
187159
+ "loss": 0.3497,
187160
+ "step": 69005
187161
+ },
187162
+ {
187163
+ "epoch": 555.54,
187164
+ "learning_rate": 8.910512820512822e-06,
187165
+ "loss": 0.3846,
187166
+ "step": 69010
187167
+ },
187168
+ {
187169
+ "epoch": 555.58,
187170
+ "learning_rate": 8.910432692307693e-06,
187171
+ "loss": 0.6693,
187172
+ "step": 69015
187173
+ },
187174
+ {
187175
+ "epoch": 555.62,
187176
+ "learning_rate": 8.910352564102564e-06,
187177
+ "loss": 1.1715,
187178
+ "step": 69020
187179
+ },
187180
+ {
187181
+ "epoch": 555.66,
187182
+ "learning_rate": 8.910272435897437e-06,
187183
+ "loss": 0.3312,
187184
+ "step": 69025
187185
+ },
187186
+ {
187187
+ "epoch": 555.7,
187188
+ "learning_rate": 8.910192307692309e-06,
187189
+ "loss": 0.373,
187190
+ "step": 69030
187191
+ },
187192
+ {
187193
+ "epoch": 555.74,
187194
+ "learning_rate": 8.91011217948718e-06,
187195
+ "loss": 0.3515,
187196
+ "step": 69035
187197
+ },
187198
+ {
187199
+ "epoch": 555.78,
187200
+ "learning_rate": 8.910032051282052e-06,
187201
+ "loss": 0.6376,
187202
+ "step": 69040
187203
+ },
187204
+ {
187205
+ "epoch": 555.82,
187206
+ "learning_rate": 8.909951923076925e-06,
187207
+ "loss": 1.0635,
187208
+ "step": 69045
187209
+ },
187210
+ {
187211
+ "epoch": 555.86,
187212
+ "learning_rate": 8.909871794871796e-06,
187213
+ "loss": 0.3108,
187214
+ "step": 69050
187215
+ },
187216
+ {
187217
+ "epoch": 555.9,
187218
+ "learning_rate": 8.909791666666667e-06,
187219
+ "loss": 0.3613,
187220
+ "step": 69055
187221
+ },
187222
+ {
187223
+ "epoch": 555.94,
187224
+ "learning_rate": 8.90971153846154e-06,
187225
+ "loss": 0.3515,
187226
+ "step": 69060
187227
+ },
187228
+ {
187229
+ "epoch": 555.98,
187230
+ "learning_rate": 8.90963141025641e-06,
187231
+ "loss": 0.7066,
187232
+ "step": 69065
187233
+ },
187234
+ {
187235
+ "epoch": 556.0,
187236
+ "eval_loss": 0.3735567331314087,
187237
+ "eval_runtime": 39.445,
187238
+ "eval_samples_per_second": 21.27,
187239
+ "eval_steps_per_second": 0.684,
187240
+ "eval_wer": 0.19350029815146094,
187241
+ "step": 69067
187242
+ },
187243
+ {
187244
+ "epoch": 556.02,
187245
+ "learning_rate": 8.909551282051283e-06,
187246
+ "loss": 0.372,
187247
+ "step": 69070
187248
+ },
187249
+ {
187250
+ "epoch": 556.06,
187251
+ "learning_rate": 8.909471153846154e-06,
187252
+ "loss": 0.3013,
187253
+ "step": 69075
187254
+ },
187255
+ {
187256
+ "epoch": 556.1,
187257
+ "learning_rate": 8.909391025641026e-06,
187258
+ "loss": 0.3405,
187259
+ "step": 69080
187260
+ },
187261
+ {
187262
+ "epoch": 556.15,
187263
+ "learning_rate": 8.909310897435897e-06,
187264
+ "loss": 0.4157,
187265
+ "step": 69085
187266
+ },
187267
+ {
187268
+ "epoch": 556.19,
187269
+ "learning_rate": 8.90923076923077e-06,
187270
+ "loss": 0.7011,
187271
+ "step": 69090
187272
+ },
187273
+ {
187274
+ "epoch": 556.23,
187275
+ "learning_rate": 8.909150641025642e-06,
187276
+ "loss": 1.0653,
187277
+ "step": 69095
187278
+ },
187279
+ {
187280
+ "epoch": 556.27,
187281
+ "learning_rate": 8.909070512820513e-06,
187282
+ "loss": 0.3359,
187283
+ "step": 69100
187284
+ },
187285
+ {
187286
+ "epoch": 556.31,
187287
+ "learning_rate": 8.908990384615386e-06,
187288
+ "loss": 0.3593,
187289
+ "step": 69105
187290
+ },
187291
+ {
187292
+ "epoch": 556.35,
187293
+ "learning_rate": 8.908910256410257e-06,
187294
+ "loss": 0.4316,
187295
+ "step": 69110
187296
+ },
187297
+ {
187298
+ "epoch": 556.39,
187299
+ "learning_rate": 8.908830128205129e-06,
187300
+ "loss": 0.748,
187301
+ "step": 69115
187302
+ },
187303
+ {
187304
+ "epoch": 556.43,
187305
+ "learning_rate": 8.90875e-06,
187306
+ "loss": 0.8545,
187307
+ "step": 69120
187308
+ },
187309
+ {
187310
+ "epoch": 556.47,
187311
+ "learning_rate": 8.908669871794873e-06,
187312
+ "loss": 0.3329,
187313
+ "step": 69125
187314
+ },
187315
+ {
187316
+ "epoch": 556.51,
187317
+ "learning_rate": 8.908589743589744e-06,
187318
+ "loss": 0.2987,
187319
+ "step": 69130
187320
+ },
187321
+ {
187322
+ "epoch": 556.55,
187323
+ "learning_rate": 8.908509615384616e-06,
187324
+ "loss": 0.3832,
187325
+ "step": 69135
187326
+ },
187327
+ {
187328
+ "epoch": 556.59,
187329
+ "learning_rate": 8.908429487179487e-06,
187330
+ "loss": 0.8224,
187331
+ "step": 69140
187332
+ },
187333
+ {
187334
+ "epoch": 556.63,
187335
+ "learning_rate": 8.90834935897436e-06,
187336
+ "loss": 0.8839,
187337
+ "step": 69145
187338
+ },
187339
+ {
187340
+ "epoch": 556.67,
187341
+ "learning_rate": 8.908269230769232e-06,
187342
+ "loss": 0.3099,
187343
+ "step": 69150
187344
+ },
187345
+ {
187346
+ "epoch": 556.71,
187347
+ "learning_rate": 8.908189102564103e-06,
187348
+ "loss": 0.2932,
187349
+ "step": 69155
187350
+ },
187351
+ {
187352
+ "epoch": 556.75,
187353
+ "learning_rate": 8.908108974358976e-06,
187354
+ "loss": 0.3511,
187355
+ "step": 69160
187356
+ },
187357
+ {
187358
+ "epoch": 556.79,
187359
+ "learning_rate": 8.908028846153847e-06,
187360
+ "loss": 0.6736,
187361
+ "step": 69165
187362
+ },
187363
+ {
187364
+ "epoch": 556.83,
187365
+ "learning_rate": 8.907948717948719e-06,
187366
+ "loss": 0.8941,
187367
+ "step": 69170
187368
+ },
187369
+ {
187370
+ "epoch": 556.87,
187371
+ "learning_rate": 8.90786858974359e-06,
187372
+ "loss": 0.3152,
187373
+ "step": 69175
187374
+ },
187375
+ {
187376
+ "epoch": 556.91,
187377
+ "learning_rate": 8.907788461538463e-06,
187378
+ "loss": 0.355,
187379
+ "step": 69180
187380
+ },
187381
+ {
187382
+ "epoch": 556.95,
187383
+ "learning_rate": 8.907708333333333e-06,
187384
+ "loss": 0.4686,
187385
+ "step": 69185
187386
+ },
187387
+ {
187388
+ "epoch": 556.99,
187389
+ "learning_rate": 8.907628205128206e-06,
187390
+ "loss": 0.8468,
187391
+ "step": 69190
187392
+ },
187393
+ {
187394
+ "epoch": 557.0,
187395
+ "eval_loss": 0.3936476707458496,
187396
+ "eval_runtime": 39.1151,
187397
+ "eval_samples_per_second": 21.45,
187398
+ "eval_steps_per_second": 0.69,
187399
+ "eval_wer": 0.19881948595337717,
187400
+ "step": 69191
187401
+ },
187402
+ {
187403
+ "epoch": 557.03,
187404
+ "learning_rate": 8.907548076923077e-06,
187405
+ "loss": 0.3576,
187406
+ "step": 69195
187407
+ },
187408
+ {
187409
+ "epoch": 557.07,
187410
+ "learning_rate": 8.907467948717949e-06,
187411
+ "loss": 0.2662,
187412
+ "step": 69200
187413
+ },
187414
+ {
187415
+ "epoch": 557.11,
187416
+ "learning_rate": 8.907387820512822e-06,
187417
+ "loss": 0.3339,
187418
+ "step": 69205
187419
+ },
187420
+ {
187421
+ "epoch": 557.15,
187422
+ "learning_rate": 8.907307692307693e-06,
187423
+ "loss": 0.4627,
187424
+ "step": 69210
187425
+ },
187426
+ {
187427
+ "epoch": 557.19,
187428
+ "learning_rate": 8.907227564102564e-06,
187429
+ "loss": 1.0684,
187430
+ "step": 69215
187431
+ },
187432
+ {
187433
+ "epoch": 557.23,
187434
+ "learning_rate": 8.907147435897436e-06,
187435
+ "loss": 0.7947,
187436
+ "step": 69220
187437
+ },
187438
+ {
187439
+ "epoch": 557.27,
187440
+ "learning_rate": 8.907067307692309e-06,
187441
+ "loss": 0.3128,
187442
+ "step": 69225
187443
+ },
187444
+ {
187445
+ "epoch": 557.31,
187446
+ "learning_rate": 8.90698717948718e-06,
187447
+ "loss": 0.3003,
187448
+ "step": 69230
187449
+ },
187450
+ {
187451
+ "epoch": 557.35,
187452
+ "learning_rate": 8.906907051282051e-06,
187453
+ "loss": 0.4299,
187454
+ "step": 69235
187455
+ },
187456
+ {
187457
+ "epoch": 557.39,
187458
+ "learning_rate": 8.906826923076923e-06,
187459
+ "loss": 1.0473,
187460
+ "step": 69240
187461
+ },
187462
+ {
187463
+ "epoch": 557.43,
187464
+ "learning_rate": 8.906746794871796e-06,
187465
+ "loss": 0.5692,
187466
+ "step": 69245
187467
+ },
187468
+ {
187469
+ "epoch": 557.47,
187470
+ "learning_rate": 8.906666666666667e-06,
187471
+ "loss": 0.2719,
187472
+ "step": 69250
187473
+ },
187474
+ {
187475
+ "epoch": 557.51,
187476
+ "learning_rate": 8.906586538461539e-06,
187477
+ "loss": 0.3171,
187478
+ "step": 69255
187479
+ },
187480
+ {
187481
+ "epoch": 557.55,
187482
+ "learning_rate": 8.906506410256412e-06,
187483
+ "loss": 0.4372,
187484
+ "step": 69260
187485
+ },
187486
+ {
187487
+ "epoch": 557.59,
187488
+ "learning_rate": 8.906426282051283e-06,
187489
+ "loss": 0.9317,
187490
+ "step": 69265
187491
+ },
187492
+ {
187493
+ "epoch": 557.63,
187494
+ "learning_rate": 8.906346153846154e-06,
187495
+ "loss": 0.6513,
187496
+ "step": 69270
187497
+ },
187498
+ {
187499
+ "epoch": 557.67,
187500
+ "learning_rate": 8.906266025641026e-06,
187501
+ "loss": 0.272,
187502
+ "step": 69275
187503
+ },
187504
+ {
187505
+ "epoch": 557.71,
187506
+ "learning_rate": 8.906185897435899e-06,
187507
+ "loss": 0.3266,
187508
+ "step": 69280
187509
+ },
187510
+ {
187511
+ "epoch": 557.76,
187512
+ "learning_rate": 8.90610576923077e-06,
187513
+ "loss": 0.453,
187514
+ "step": 69285
187515
+ },
187516
+ {
187517
+ "epoch": 557.8,
187518
+ "learning_rate": 8.906025641025641e-06,
187519
+ "loss": 0.8466,
187520
+ "step": 69290
187521
+ },
187522
+ {
187523
+ "epoch": 557.84,
187524
+ "learning_rate": 8.905945512820513e-06,
187525
+ "loss": 0.6527,
187526
+ "step": 69295
187527
+ },
187528
+ {
187529
+ "epoch": 557.88,
187530
+ "learning_rate": 8.905865384615386e-06,
187531
+ "loss": 0.3322,
187532
+ "step": 69300
187533
+ },
187534
+ {
187535
+ "epoch": 557.92,
187536
+ "learning_rate": 8.905785256410257e-06,
187537
+ "loss": 0.3628,
187538
+ "step": 69305
187539
+ },
187540
+ {
187541
+ "epoch": 557.96,
187542
+ "learning_rate": 8.905705128205129e-06,
187543
+ "loss": 0.4183,
187544
+ "step": 69310
187545
+ },
187546
+ {
187547
+ "epoch": 558.0,
187548
+ "learning_rate": 8.905625000000002e-06,
187549
+ "loss": 1.0645,
187550
+ "step": 69315
187551
+ },
187552
+ {
187553
+ "epoch": 558.0,
187554
+ "eval_loss": 0.4768179655075073,
187555
+ "eval_runtime": 41.3535,
187556
+ "eval_samples_per_second": 20.289,
187557
+ "eval_steps_per_second": 0.653,
187558
+ "eval_wer": 0.1948916076411247,
187559
+ "step": 69315
187560
+ },
187561
+ {
187562
+ "epoch": 554.04,
187563
+ "learning_rate": 8.905544871794873e-06,
187564
+ "loss": 0.3339,
187565
+ "step": 69320
187566
+ },
187567
+ {
187568
+ "epoch": 554.08,
187569
+ "learning_rate": 8.905464743589744e-06,
187570
+ "loss": 0.2938,
187571
+ "step": 69325
187572
+ },
187573
+ {
187574
+ "epoch": 554.12,
187575
+ "learning_rate": 8.905384615384616e-06,
187576
+ "loss": 0.3485,
187577
+ "step": 69330
187578
+ },
187579
+ {
187580
+ "epoch": 554.16,
187581
+ "learning_rate": 8.905304487179489e-06,
187582
+ "loss": 0.6218,
187583
+ "step": 69335
187584
+ },
187585
+ {
187586
+ "epoch": 554.2,
187587
+ "learning_rate": 8.905224358974358e-06,
187588
+ "loss": 1.2255,
187589
+ "step": 69340
187590
+ },
187591
+ {
187592
+ "epoch": 554.24,
187593
+ "learning_rate": 8.905144230769232e-06,
187594
+ "loss": 0.3385,
187595
+ "step": 69345
187596
+ },
187597
+ {
187598
+ "epoch": 554.28,
187599
+ "learning_rate": 8.905064102564105e-06,
187600
+ "loss": 0.3168,
187601
+ "step": 69350
187602
+ },
187603
+ {
187604
+ "epoch": 554.32,
187605
+ "learning_rate": 8.904983974358974e-06,
187606
+ "loss": 0.3309,
187607
+ "step": 69355
187608
+ },
187609
+ {
187610
+ "epoch": 554.36,
187611
+ "learning_rate": 8.904903846153847e-06,
187612
+ "loss": 0.5233,
187613
+ "step": 69360
187614
+ },
187615
+ {
187616
+ "epoch": 554.4,
187617
+ "learning_rate": 8.904823717948719e-06,
187618
+ "loss": 1.4065,
187619
+ "step": 69365
187620
+ },
187621
+ {
187622
+ "epoch": 554.44,
187623
+ "learning_rate": 8.90474358974359e-06,
187624
+ "loss": 0.3589,
187625
+ "step": 69370
187626
+ },
187627
+ {
187628
+ "epoch": 554.48,
187629
+ "learning_rate": 8.904663461538461e-06,
187630
+ "loss": 0.2869,
187631
+ "step": 69375
187632
+ },
187633
+ {
187634
+ "epoch": 554.52,
187635
+ "learning_rate": 8.904583333333334e-06,
187636
+ "loss": 0.3377,
187637
+ "step": 69380
187638
+ },
187639
+ {
187640
+ "epoch": 554.56,
187641
+ "learning_rate": 8.904503205128206e-06,
187642
+ "loss": 0.535,
187643
+ "step": 69385
187644
+ },
187645
+ {
187646
+ "epoch": 554.6,
187647
+ "learning_rate": 8.904423076923077e-06,
187648
+ "loss": 1.4467,
187649
+ "step": 69390
187650
+ },
187651
+ {
187652
+ "epoch": 554.64,
187653
+ "learning_rate": 8.904342948717948e-06,
187654
+ "loss": 0.5932,
187655
+ "step": 69395
187656
+ },
187657
+ {
187658
+ "epoch": 554.68,
187659
+ "learning_rate": 8.904262820512822e-06,
187660
+ "loss": 0.4281,
187661
+ "step": 69400
187662
+ },
187663
+ {
187664
+ "epoch": 554.72,
187665
+ "learning_rate": 8.904182692307693e-06,
187666
+ "loss": 0.3255,
187667
+ "step": 69405
187668
+ },
187669
+ {
187670
+ "epoch": 554.76,
187671
+ "learning_rate": 8.904102564102564e-06,
187672
+ "loss": 0.5435,
187673
+ "step": 69410
187674
+ },
187675
+ {
187676
+ "epoch": 554.8,
187677
+ "learning_rate": 8.904022435897437e-06,
187678
+ "loss": 1.2416,
187679
+ "step": 69415
187680
+ },
187681
+ {
187682
+ "epoch": 554.84,
187683
+ "learning_rate": 8.903942307692309e-06,
187684
+ "loss": 0.6299,
187685
+ "step": 69420
187686
+ },
187687
+ {
187688
+ "epoch": 554.88,
187689
+ "learning_rate": 8.90386217948718e-06,
187690
+ "loss": 0.3255,
187691
+ "step": 69425
187692
+ },
187693
+ {
187694
+ "epoch": 554.92,
187695
+ "learning_rate": 8.903782051282051e-06,
187696
+ "loss": 0.3651,
187697
+ "step": 69430
187698
+ },
187699
+ {
187700
+ "epoch": 554.96,
187701
+ "learning_rate": 8.903701923076924e-06,
187702
+ "loss": 0.4798,
187703
+ "step": 69435
187704
+ },
187705
+ {
187706
+ "epoch": 555.0,
187707
+ "learning_rate": 8.903621794871796e-06,
187708
+ "loss": 1.206,
187709
+ "step": 69440
187710
+ },
187711
+ {
187712
+ "epoch": 555.0,
187713
+ "eval_loss": 0.34815290570259094,
187714
+ "eval_runtime": 40.671,
187715
+ "eval_samples_per_second": 20.629,
187716
+ "eval_steps_per_second": 0.664,
187717
+ "eval_wer": 0.18589017831870933,
187718
+ "step": 69440
187719
  }
187720
  ],
187721
  "max_steps": 625000,
187722
  "num_train_epochs": 5000,
187723
+ "total_flos": 1.9540963521520874e+20,
187724
  "trial_name": null,
187725
  "trial_params": null
187726
  }
model-bin/finetune/base/{checkpoint-68818 β†’ checkpoint-69440}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629843978.3964214/events.out.tfevents.1629843978.c435e1c5ee04.920.241 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5f924f05c1761d6e622e1fe364132cc7ce7a60ed587f21962642aed9c6dea3c
3
+ size 4194
model-bin/finetune/base/log/1629844632.7215848/events.out.tfevents.1629844632.c435e1c5ee04.920.243 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c66b00ffda2d8831a8a42b631b2a7b9edaed066b9afb6bf70a72fc55207e8b1
3
+ size 4194
model-bin/finetune/base/log/1629845266.9609022/events.out.tfevents.1629845266.c435e1c5ee04.920.245 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2600798a7c11f3f5bc24a7889aca81841c8038630b09e3505897c3e25c760ce
3
+ size 4194
model-bin/finetune/base/log/1629845904.6603022/events.out.tfevents.1629845904.c435e1c5ee04.920.247 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fff7009ac8fdf816d297df477743834bfdefea6c6649b21a6e1c53abff5c3578
3
+ size 4194
model-bin/finetune/base/log/1629846554.1512308/events.out.tfevents.1629846554.c435e1c5ee04.920.249 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36d230724e1655ef4bdeeba22e814367809fafcc7e4d7253811e6d0448d06690
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629843978.c435e1c5ee04.920.240 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05432f54c615f59447fc156df5739ef2da6178de5b7a1694bcdc6a38347fc914
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629844632.c435e1c5ee04.920.242 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0909f46e291559be47da07e3c7249cbbaa3d97aca8858064f634c2eaecc6e0d
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629845266.c435e1c5ee04.920.244 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9465a9c811dc5c646b4caa1afdca954c1dee5fd1f750e1b92640e3e4eb19515a
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629845904.c435e1c5ee04.920.246 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b779365eed4809a3b611dd3419710d7f88c6fd81ca818fc2378f5bd0b67bc4eb
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629846554.c435e1c5ee04.920.248 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:628bd5086daaad6d03bd552a731725168c88677e588962a70ed38ebe2787a540
3
+ size 8622