Check commited on
Commit
0c35cac
Β·
1 Parent(s): 063b9a9

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/rng_state.pth +2 -2
  6. model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/trainer_state.json +799 -4
  9. model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629764971.3905332/events.out.tfevents.1629764971.c435e1c5ee04.920.1 +3 -0
  11. model-bin/finetune/base/log/1629765618.347005/events.out.tfevents.1629765618.c435e1c5ee04.920.3 +3 -0
  12. model-bin/finetune/base/log/1629766288.11667/events.out.tfevents.1629766288.c435e1c5ee04.920.5 +3 -0
  13. model-bin/finetune/base/log/1629766954.1795254/events.out.tfevents.1629766954.c435e1c5ee04.920.7 +3 -0
  14. model-bin/finetune/base/log/1629767616.4417202/events.out.tfevents.1629767616.c435e1c5ee04.920.9 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629764971.c435e1c5ee04.920.0 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629765618.c435e1c5ee04.920.2 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629766288.c435e1c5ee04.920.4 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629766954.c435e1c5ee04.920.6 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629767616.c435e1c5ee04.920.8 +3 -0
model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:54045d565c66c3e5b8018fc3afcc2c41061b9918603d0883df91a655d33fee22
3
  size 722165009
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:175a41a98f5203d0db6aa1fbd6a9ae7b059e2710fa700b5bea11a57ff3220cf0
3
  size 722165009
model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1d25122eb873e0266a53d0c13a8c157bfbcf84c44543a60b0d7a792d97aeb34d
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec4d2a6a7040c89cd57e0b9cb66beef6221d858761995ab0704ac3e0af89527a
3
  size 377909911
model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:303bcbfcafc202ca744ae4ed8003b741b08ac41373010724ddcd8ab1d77d651a
3
- size 14439
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a41625b4db78a0c89b01c36cbddae265eafff10b0d61888fbd70eaf471ffd95f
3
+ size 14567
model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:517b8e9913beb30b2ae33d6974fd6fee1e6f87a69ee2028115af6866d0917bf4
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e3fbcb39f68cfc7ed849162b4e354f6037864a19c979d4bf441c7c266c9efa4
3
  size 559
model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:41f48a461a4300556f9a200eddd4d70732e267633a4abe25cef3f92554b7b8c0
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f43a9c3edd8c53660caa0eb9f925a632bd24b739141d5b8e929015a7219a86ed
3
  size 623
model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.18992848189928482,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-52515",
4
- "epoch": 431.0,
5
- "global_step": 53883,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -167919,11 +167919,806 @@
167919
  "eval_steps_per_second": 0.665,
167920
  "eval_wer": 0.20484629294755877,
167921
  "step": 53883
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
167922
  }
167923
  ],
167924
- "max_steps": 625000,
167925
  "num_train_epochs": 5000,
167926
- "total_flos": 1.5162302896695742e+20,
167927
  "trial_name": null,
167928
  "trial_params": null
167929
  }
 
1
  {
2
  "best_metric": 0.18992848189928482,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-52515",
4
+ "epoch": 438.99598393574297,
5
+ "global_step": 54505,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
167919
  "eval_steps_per_second": 0.665,
167920
  "eval_wer": 0.20484629294755877,
167921
  "step": 53883
167922
+ },
167923
+ {
167924
+ "epoch": 431.02,
167925
+ "learning_rate": 9.152804487179488e-06,
167926
+ "loss": 0.419,
167927
+ "step": 53885
167928
+ },
167929
+ {
167930
+ "epoch": 431.06,
167931
+ "learning_rate": 9.152724358974359e-06,
167932
+ "loss": 0.3243,
167933
+ "step": 53890
167934
+ },
167935
+ {
167936
+ "epoch": 431.1,
167937
+ "learning_rate": 9.152644230769232e-06,
167938
+ "loss": 0.341,
167939
+ "step": 53895
167940
+ },
167941
+ {
167942
+ "epoch": 431.14,
167943
+ "learning_rate": 9.152564102564103e-06,
167944
+ "loss": 0.4508,
167945
+ "step": 53900
167946
+ },
167947
+ {
167948
+ "epoch": 431.18,
167949
+ "learning_rate": 9.152483974358975e-06,
167950
+ "loss": 0.7833,
167951
+ "step": 53905
167952
+ },
167953
+ {
167954
+ "epoch": 431.22,
167955
+ "learning_rate": 9.152403846153846e-06,
167956
+ "loss": 1.0565,
167957
+ "step": 53910
167958
+ },
167959
+ {
167960
+ "epoch": 431.25,
167961
+ "learning_rate": 9.15232371794872e-06,
167962
+ "loss": 0.3407,
167963
+ "step": 53915
167964
+ },
167965
+ {
167966
+ "epoch": 431.29,
167967
+ "learning_rate": 9.15224358974359e-06,
167968
+ "loss": 0.3398,
167969
+ "step": 53920
167970
+ },
167971
+ {
167972
+ "epoch": 431.33,
167973
+ "learning_rate": 9.152163461538462e-06,
167974
+ "loss": 0.3471,
167975
+ "step": 53925
167976
+ },
167977
+ {
167978
+ "epoch": 431.37,
167979
+ "learning_rate": 9.152083333333333e-06,
167980
+ "loss": 0.7131,
167981
+ "step": 53930
167982
+ },
167983
+ {
167984
+ "epoch": 431.41,
167985
+ "learning_rate": 9.152003205128206e-06,
167986
+ "loss": 0.996,
167987
+ "step": 53935
167988
+ },
167989
+ {
167990
+ "epoch": 431.45,
167991
+ "learning_rate": 9.151923076923078e-06,
167992
+ "loss": 0.3824,
167993
+ "step": 53940
167994
+ },
167995
+ {
167996
+ "epoch": 431.49,
167997
+ "learning_rate": 9.151842948717949e-06,
167998
+ "loss": 0.321,
167999
+ "step": 53945
168000
+ },
168001
+ {
168002
+ "epoch": 431.53,
168003
+ "learning_rate": 9.151762820512822e-06,
168004
+ "loss": 0.3952,
168005
+ "step": 53950
168006
+ },
168007
+ {
168008
+ "epoch": 431.57,
168009
+ "learning_rate": 9.151682692307693e-06,
168010
+ "loss": 0.6883,
168011
+ "step": 53955
168012
+ },
168013
+ {
168014
+ "epoch": 431.61,
168015
+ "learning_rate": 9.151618589743591e-06,
168016
+ "loss": 1.5506,
168017
+ "step": 53960
168018
+ },
168019
+ {
168020
+ "epoch": 431.65,
168021
+ "learning_rate": 9.151538461538462e-06,
168022
+ "loss": 0.3653,
168023
+ "step": 53965
168024
+ },
168025
+ {
168026
+ "epoch": 431.69,
168027
+ "learning_rate": 9.151458333333334e-06,
168028
+ "loss": 0.4062,
168029
+ "step": 53970
168030
+ },
168031
+ {
168032
+ "epoch": 431.73,
168033
+ "learning_rate": 9.151378205128207e-06,
168034
+ "loss": 0.3621,
168035
+ "step": 53975
168036
+ },
168037
+ {
168038
+ "epoch": 431.77,
168039
+ "learning_rate": 9.151298076923078e-06,
168040
+ "loss": 0.6849,
168041
+ "step": 53980
168042
+ },
168043
+ {
168044
+ "epoch": 431.81,
168045
+ "learning_rate": 9.15121794871795e-06,
168046
+ "loss": 1.0669,
168047
+ "step": 53985
168048
+ },
168049
+ {
168050
+ "epoch": 431.85,
168051
+ "learning_rate": 9.15113782051282e-06,
168052
+ "loss": 0.3909,
168053
+ "step": 53990
168054
+ },
168055
+ {
168056
+ "epoch": 431.89,
168057
+ "learning_rate": 9.151057692307694e-06,
168058
+ "loss": 0.3625,
168059
+ "step": 53995
168060
+ },
168061
+ {
168062
+ "epoch": 431.93,
168063
+ "learning_rate": 9.150977564102565e-06,
168064
+ "loss": 0.4017,
168065
+ "step": 54000
168066
+ },
168067
+ {
168068
+ "epoch": 431.97,
168069
+ "learning_rate": 9.150897435897437e-06,
168070
+ "loss": 0.7696,
168071
+ "step": 54005
168072
+ },
168073
+ {
168074
+ "epoch": 432.0,
168075
+ "eval_loss": 0.37287962436676025,
168076
+ "eval_runtime": 38.0521,
168077
+ "eval_samples_per_second": 22.049,
168078
+ "eval_steps_per_second": 0.71,
168079
+ "eval_wer": 0.20122172932877608,
168080
+ "step": 54008
168081
+ },
168082
+ {
168083
+ "epoch": 435.02,
168084
+ "learning_rate": 9.150817307692308e-06,
168085
+ "loss": 0.3971,
168086
+ "step": 54010
168087
+ },
168088
+ {
168089
+ "epoch": 435.06,
168090
+ "learning_rate": 9.150737179487181e-06,
168091
+ "loss": 0.3162,
168092
+ "step": 54015
168093
+ },
168094
+ {
168095
+ "epoch": 435.1,
168096
+ "learning_rate": 9.150657051282052e-06,
168097
+ "loss": 0.3087,
168098
+ "step": 54020
168099
+ },
168100
+ {
168101
+ "epoch": 435.14,
168102
+ "learning_rate": 9.150576923076924e-06,
168103
+ "loss": 0.3867,
168104
+ "step": 54025
168105
+ },
168106
+ {
168107
+ "epoch": 435.18,
168108
+ "learning_rate": 9.150496794871797e-06,
168109
+ "loss": 0.6681,
168110
+ "step": 54030
168111
+ },
168112
+ {
168113
+ "epoch": 435.22,
168114
+ "learning_rate": 9.150416666666666e-06,
168115
+ "loss": 1.0952,
168116
+ "step": 54035
168117
+ },
168118
+ {
168119
+ "epoch": 435.26,
168120
+ "learning_rate": 9.15033653846154e-06,
168121
+ "loss": 0.3676,
168122
+ "step": 54040
168123
+ },
168124
+ {
168125
+ "epoch": 435.3,
168126
+ "learning_rate": 9.15025641025641e-06,
168127
+ "loss": 0.3552,
168128
+ "step": 54045
168129
+ },
168130
+ {
168131
+ "epoch": 435.34,
168132
+ "learning_rate": 9.150176282051282e-06,
168133
+ "loss": 0.3858,
168134
+ "step": 54050
168135
+ },
168136
+ {
168137
+ "epoch": 435.38,
168138
+ "learning_rate": 9.150096153846154e-06,
168139
+ "loss": 0.7447,
168140
+ "step": 54055
168141
+ },
168142
+ {
168143
+ "epoch": 435.42,
168144
+ "learning_rate": 9.150016025641027e-06,
168145
+ "loss": 1.1432,
168146
+ "step": 54060
168147
+ },
168148
+ {
168149
+ "epoch": 435.46,
168150
+ "learning_rate": 9.149935897435898e-06,
168151
+ "loss": 0.2896,
168152
+ "step": 54065
168153
+ },
168154
+ {
168155
+ "epoch": 435.5,
168156
+ "learning_rate": 9.14985576923077e-06,
168157
+ "loss": 0.3781,
168158
+ "step": 54070
168159
+ },
168160
+ {
168161
+ "epoch": 435.54,
168162
+ "learning_rate": 9.149775641025642e-06,
168163
+ "loss": 0.3899,
168164
+ "step": 54075
168165
+ },
168166
+ {
168167
+ "epoch": 435.58,
168168
+ "learning_rate": 9.149695512820514e-06,
168169
+ "loss": 0.6563,
168170
+ "step": 54080
168171
+ },
168172
+ {
168173
+ "epoch": 435.62,
168174
+ "learning_rate": 9.149615384615385e-06,
168175
+ "loss": 1.1815,
168176
+ "step": 54085
168177
+ },
168178
+ {
168179
+ "epoch": 435.66,
168180
+ "learning_rate": 9.149535256410256e-06,
168181
+ "loss": 0.4117,
168182
+ "step": 54090
168183
+ },
168184
+ {
168185
+ "epoch": 435.7,
168186
+ "learning_rate": 9.14945512820513e-06,
168187
+ "loss": 0.2979,
168188
+ "step": 54095
168189
+ },
168190
+ {
168191
+ "epoch": 435.74,
168192
+ "learning_rate": 9.149375e-06,
168193
+ "loss": 0.3699,
168194
+ "step": 54100
168195
+ },
168196
+ {
168197
+ "epoch": 435.78,
168198
+ "learning_rate": 9.149294871794872e-06,
168199
+ "loss": 0.6889,
168200
+ "step": 54105
168201
+ },
168202
+ {
168203
+ "epoch": 435.82,
168204
+ "learning_rate": 9.149214743589744e-06,
168205
+ "loss": 1.0284,
168206
+ "step": 54110
168207
+ },
168208
+ {
168209
+ "epoch": 435.86,
168210
+ "learning_rate": 9.149134615384617e-06,
168211
+ "loss": 0.3884,
168212
+ "step": 54115
168213
+ },
168214
+ {
168215
+ "epoch": 435.9,
168216
+ "learning_rate": 9.149054487179488e-06,
168217
+ "loss": 0.3676,
168218
+ "step": 54120
168219
+ },
168220
+ {
168221
+ "epoch": 435.94,
168222
+ "learning_rate": 9.14897435897436e-06,
168223
+ "loss": 0.3872,
168224
+ "step": 54125
168225
+ },
168226
+ {
168227
+ "epoch": 435.98,
168228
+ "learning_rate": 9.148894230769232e-06,
168229
+ "loss": 0.8216,
168230
+ "step": 54130
168231
+ },
168232
+ {
168233
+ "epoch": 436.0,
168234
+ "eval_loss": 0.44797709584236145,
168235
+ "eval_runtime": 40.441,
168236
+ "eval_samples_per_second": 20.771,
168237
+ "eval_steps_per_second": 0.668,
168238
+ "eval_wer": 0.19947140444901254,
168239
+ "step": 54132
168240
+ },
168241
+ {
168242
+ "epoch": 436.02,
168243
+ "learning_rate": 9.148814102564104e-06,
168244
+ "loss": 0.3841,
168245
+ "step": 54135
168246
+ },
168247
+ {
168248
+ "epoch": 436.06,
168249
+ "learning_rate": 9.148733974358975e-06,
168250
+ "loss": 0.364,
168251
+ "step": 54140
168252
+ },
168253
+ {
168254
+ "epoch": 436.1,
168255
+ "learning_rate": 9.148653846153846e-06,
168256
+ "loss": 0.384,
168257
+ "step": 54145
168258
+ },
168259
+ {
168260
+ "epoch": 436.14,
168261
+ "learning_rate": 9.14857371794872e-06,
168262
+ "loss": 0.374,
168263
+ "step": 54150
168264
+ },
168265
+ {
168266
+ "epoch": 436.18,
168267
+ "learning_rate": 9.14849358974359e-06,
168268
+ "loss": 0.8492,
168269
+ "step": 54155
168270
+ },
168271
+ {
168272
+ "epoch": 436.22,
168273
+ "learning_rate": 9.148413461538462e-06,
168274
+ "loss": 0.9104,
168275
+ "step": 54160
168276
+ },
168277
+ {
168278
+ "epoch": 436.27,
168279
+ "learning_rate": 9.148333333333335e-06,
168280
+ "loss": 0.3081,
168281
+ "step": 54165
168282
+ },
168283
+ {
168284
+ "epoch": 436.31,
168285
+ "learning_rate": 9.148253205128207e-06,
168286
+ "loss": 0.3592,
168287
+ "step": 54170
168288
+ },
168289
+ {
168290
+ "epoch": 436.35,
168291
+ "learning_rate": 9.148173076923078e-06,
168292
+ "loss": 0.4217,
168293
+ "step": 54175
168294
+ },
168295
+ {
168296
+ "epoch": 436.39,
168297
+ "learning_rate": 9.14809294871795e-06,
168298
+ "loss": 0.888,
168299
+ "step": 54180
168300
+ },
168301
+ {
168302
+ "epoch": 436.43,
168303
+ "learning_rate": 9.148012820512822e-06,
168304
+ "loss": 1.0466,
168305
+ "step": 54185
168306
+ },
168307
+ {
168308
+ "epoch": 436.47,
168309
+ "learning_rate": 9.147932692307692e-06,
168310
+ "loss": 0.3215,
168311
+ "step": 54190
168312
+ },
168313
+ {
168314
+ "epoch": 436.51,
168315
+ "learning_rate": 9.147852564102565e-06,
168316
+ "loss": 0.3265,
168317
+ "step": 54195
168318
+ },
168319
+ {
168320
+ "epoch": 436.55,
168321
+ "learning_rate": 9.147772435897436e-06,
168322
+ "loss": 0.4732,
168323
+ "step": 54200
168324
+ },
168325
+ {
168326
+ "epoch": 436.59,
168327
+ "learning_rate": 9.147692307692308e-06,
168328
+ "loss": 0.9318,
168329
+ "step": 54205
168330
+ },
168331
+ {
168332
+ "epoch": 436.63,
168333
+ "learning_rate": 9.14761217948718e-06,
168334
+ "loss": 0.8966,
168335
+ "step": 54210
168336
+ },
168337
+ {
168338
+ "epoch": 436.67,
168339
+ "learning_rate": 9.147532051282052e-06,
168340
+ "loss": 0.3186,
168341
+ "step": 54215
168342
+ },
168343
+ {
168344
+ "epoch": 436.71,
168345
+ "learning_rate": 9.147451923076924e-06,
168346
+ "loss": 0.3307,
168347
+ "step": 54220
168348
+ },
168349
+ {
168350
+ "epoch": 436.75,
168351
+ "learning_rate": 9.147371794871795e-06,
168352
+ "loss": 0.4654,
168353
+ "step": 54225
168354
+ },
168355
+ {
168356
+ "epoch": 436.79,
168357
+ "learning_rate": 9.147291666666668e-06,
168358
+ "loss": 0.7942,
168359
+ "step": 54230
168360
+ },
168361
+ {
168362
+ "epoch": 436.83,
168363
+ "learning_rate": 9.14721153846154e-06,
168364
+ "loss": 0.9858,
168365
+ "step": 54235
168366
+ },
168367
+ {
168368
+ "epoch": 436.87,
168369
+ "learning_rate": 9.14713141025641e-06,
168370
+ "loss": 0.3137,
168371
+ "step": 54240
168372
+ },
168373
+ {
168374
+ "epoch": 436.91,
168375
+ "learning_rate": 9.147051282051282e-06,
168376
+ "loss": 0.3034,
168377
+ "step": 54245
168378
+ },
168379
+ {
168380
+ "epoch": 436.95,
168381
+ "learning_rate": 9.146971153846155e-06,
168382
+ "loss": 0.4034,
168383
+ "step": 54250
168384
+ },
168385
+ {
168386
+ "epoch": 436.99,
168387
+ "learning_rate": 9.146891025641026e-06,
168388
+ "loss": 0.9007,
168389
+ "step": 54255
168390
+ },
168391
+ {
168392
+ "epoch": 437.0,
168393
+ "eval_loss": 0.41687604784965515,
168394
+ "eval_runtime": 39.6761,
168395
+ "eval_samples_per_second": 21.171,
168396
+ "eval_steps_per_second": 0.681,
168397
+ "eval_wer": 0.19979597784902361,
168398
+ "step": 54256
168399
+ },
168400
+ {
168401
+ "epoch": 434.03,
168402
+ "learning_rate": 9.146810897435898e-06,
168403
+ "loss": 0.366,
168404
+ "step": 54260
168405
+ },
168406
+ {
168407
+ "epoch": 434.07,
168408
+ "learning_rate": 9.146730769230771e-06,
168409
+ "loss": 0.3047,
168410
+ "step": 54265
168411
+ },
168412
+ {
168413
+ "epoch": 434.11,
168414
+ "learning_rate": 9.146650641025642e-06,
168415
+ "loss": 0.4099,
168416
+ "step": 54270
168417
+ },
168418
+ {
168419
+ "epoch": 434.15,
168420
+ "learning_rate": 9.146570512820514e-06,
168421
+ "loss": 0.405,
168422
+ "step": 54275
168423
+ },
168424
+ {
168425
+ "epoch": 434.19,
168426
+ "learning_rate": 9.146490384615385e-06,
168427
+ "loss": 1.0881,
168428
+ "step": 54280
168429
+ },
168430
+ {
168431
+ "epoch": 434.23,
168432
+ "learning_rate": 9.146410256410258e-06,
168433
+ "loss": 0.6728,
168434
+ "step": 54285
168435
+ },
168436
+ {
168437
+ "epoch": 434.27,
168438
+ "learning_rate": 9.14633012820513e-06,
168439
+ "loss": 0.3143,
168440
+ "step": 54290
168441
+ },
168442
+ {
168443
+ "epoch": 434.31,
168444
+ "learning_rate": 9.14625e-06,
168445
+ "loss": 0.3771,
168446
+ "step": 54295
168447
+ },
168448
+ {
168449
+ "epoch": 434.35,
168450
+ "learning_rate": 9.146169871794872e-06,
168451
+ "loss": 0.4689,
168452
+ "step": 54300
168453
+ },
168454
+ {
168455
+ "epoch": 434.39,
168456
+ "learning_rate": 9.146089743589745e-06,
168457
+ "loss": 1.0307,
168458
+ "step": 54305
168459
+ },
168460
+ {
168461
+ "epoch": 434.43,
168462
+ "learning_rate": 9.146009615384615e-06,
168463
+ "loss": 0.6574,
168464
+ "step": 54310
168465
+ },
168466
+ {
168467
+ "epoch": 434.47,
168468
+ "learning_rate": 9.145929487179488e-06,
168469
+ "loss": 0.3558,
168470
+ "step": 54315
168471
+ },
168472
+ {
168473
+ "epoch": 434.51,
168474
+ "learning_rate": 9.145849358974361e-06,
168475
+ "loss": 0.3287,
168476
+ "step": 54320
168477
+ },
168478
+ {
168479
+ "epoch": 434.55,
168480
+ "learning_rate": 9.14576923076923e-06,
168481
+ "loss": 0.4354,
168482
+ "step": 54325
168483
+ },
168484
+ {
168485
+ "epoch": 434.59,
168486
+ "learning_rate": 9.145689102564104e-06,
168487
+ "loss": 0.8869,
168488
+ "step": 54330
168489
+ },
168490
+ {
168491
+ "epoch": 434.63,
168492
+ "learning_rate": 9.145608974358975e-06,
168493
+ "loss": 0.9042,
168494
+ "step": 54335
168495
+ },
168496
+ {
168497
+ "epoch": 434.67,
168498
+ "learning_rate": 9.145528846153846e-06,
168499
+ "loss": 0.3911,
168500
+ "step": 54340
168501
+ },
168502
+ {
168503
+ "epoch": 434.71,
168504
+ "learning_rate": 9.145448717948718e-06,
168505
+ "loss": 0.342,
168506
+ "step": 54345
168507
+ },
168508
+ {
168509
+ "epoch": 434.75,
168510
+ "learning_rate": 9.14536858974359e-06,
168511
+ "loss": 0.4601,
168512
+ "step": 54350
168513
+ },
168514
+ {
168515
+ "epoch": 434.79,
168516
+ "learning_rate": 9.145288461538462e-06,
168517
+ "loss": 1.0639,
168518
+ "step": 54355
168519
+ },
168520
+ {
168521
+ "epoch": 434.83,
168522
+ "learning_rate": 9.145208333333333e-06,
168523
+ "loss": 0.7023,
168524
+ "step": 54360
168525
+ },
168526
+ {
168527
+ "epoch": 434.87,
168528
+ "learning_rate": 9.145128205128207e-06,
168529
+ "loss": 0.3901,
168530
+ "step": 54365
168531
+ },
168532
+ {
168533
+ "epoch": 434.91,
168534
+ "learning_rate": 9.145048076923078e-06,
168535
+ "loss": 0.3836,
168536
+ "step": 54370
168537
+ },
168538
+ {
168539
+ "epoch": 434.95,
168540
+ "learning_rate": 9.14496794871795e-06,
168541
+ "loss": 0.4315,
168542
+ "step": 54375
168543
+ },
168544
+ {
168545
+ "epoch": 434.99,
168546
+ "learning_rate": 9.14488782051282e-06,
168547
+ "loss": 1.001,
168548
+ "step": 54380
168549
+ },
168550
+ {
168551
+ "epoch": 435.0,
168552
+ "eval_loss": 0.3676057457923889,
168553
+ "eval_runtime": 40.0739,
168554
+ "eval_samples_per_second": 20.961,
168555
+ "eval_steps_per_second": 0.674,
168556
+ "eval_wer": 0.20047993019197208,
168557
+ "step": 54381
168558
+ },
168559
+ {
168560
+ "epoch": 438.03,
168561
+ "learning_rate": 9.144807692307694e-06,
168562
+ "loss": 0.3355,
168563
+ "step": 54385
168564
+ },
168565
+ {
168566
+ "epoch": 438.07,
168567
+ "learning_rate": 9.144727564102565e-06,
168568
+ "loss": 0.3302,
168569
+ "step": 54390
168570
+ },
168571
+ {
168572
+ "epoch": 438.11,
168573
+ "learning_rate": 9.144647435897436e-06,
168574
+ "loss": 0.3472,
168575
+ "step": 54395
168576
+ },
168577
+ {
168578
+ "epoch": 438.15,
168579
+ "learning_rate": 9.144567307692308e-06,
168580
+ "loss": 0.4144,
168581
+ "step": 54400
168582
+ },
168583
+ {
168584
+ "epoch": 438.19,
168585
+ "learning_rate": 9.14448717948718e-06,
168586
+ "loss": 0.8747,
168587
+ "step": 54405
168588
+ },
168589
+ {
168590
+ "epoch": 438.23,
168591
+ "learning_rate": 9.144407051282052e-06,
168592
+ "loss": 0.6965,
168593
+ "step": 54410
168594
+ },
168595
+ {
168596
+ "epoch": 438.27,
168597
+ "learning_rate": 9.144326923076924e-06,
168598
+ "loss": 0.3358,
168599
+ "step": 54415
168600
+ },
168601
+ {
168602
+ "epoch": 438.31,
168603
+ "learning_rate": 9.144246794871797e-06,
168604
+ "loss": 0.4455,
168605
+ "step": 54420
168606
+ },
168607
+ {
168608
+ "epoch": 438.35,
168609
+ "learning_rate": 9.144166666666668e-06,
168610
+ "loss": 0.4321,
168611
+ "step": 54425
168612
+ },
168613
+ {
168614
+ "epoch": 438.39,
168615
+ "learning_rate": 9.14408653846154e-06,
168616
+ "loss": 1.0387,
168617
+ "step": 54430
168618
+ },
168619
+ {
168620
+ "epoch": 438.43,
168621
+ "learning_rate": 9.14400641025641e-06,
168622
+ "loss": 0.7209,
168623
+ "step": 54435
168624
+ },
168625
+ {
168626
+ "epoch": 438.47,
168627
+ "learning_rate": 9.143926282051284e-06,
168628
+ "loss": 0.3056,
168629
+ "step": 54440
168630
+ },
168631
+ {
168632
+ "epoch": 438.51,
168633
+ "learning_rate": 9.143846153846155e-06,
168634
+ "loss": 0.36,
168635
+ "step": 54445
168636
+ },
168637
+ {
168638
+ "epoch": 438.55,
168639
+ "learning_rate": 9.143766025641026e-06,
168640
+ "loss": 0.5089,
168641
+ "step": 54450
168642
+ },
168643
+ {
168644
+ "epoch": 438.59,
168645
+ "learning_rate": 9.143685897435898e-06,
168646
+ "loss": 1.0719,
168647
+ "step": 54455
168648
+ },
168649
+ {
168650
+ "epoch": 438.63,
168651
+ "learning_rate": 9.14360576923077e-06,
168652
+ "loss": 0.7689,
168653
+ "step": 54460
168654
+ },
168655
+ {
168656
+ "epoch": 438.67,
168657
+ "learning_rate": 9.14352564102564e-06,
168658
+ "loss": 0.3371,
168659
+ "step": 54465
168660
+ },
168661
+ {
168662
+ "epoch": 438.71,
168663
+ "learning_rate": 9.143445512820514e-06,
168664
+ "loss": 0.3606,
168665
+ "step": 54470
168666
+ },
168667
+ {
168668
+ "epoch": 438.76,
168669
+ "learning_rate": 9.143365384615387e-06,
168670
+ "loss": 0.4898,
168671
+ "step": 54475
168672
+ },
168673
+ {
168674
+ "epoch": 438.8,
168675
+ "learning_rate": 9.143285256410256e-06,
168676
+ "loss": 0.8839,
168677
+ "step": 54480
168678
+ },
168679
+ {
168680
+ "epoch": 438.84,
168681
+ "learning_rate": 9.14320512820513e-06,
168682
+ "loss": 0.7543,
168683
+ "step": 54485
168684
+ },
168685
+ {
168686
+ "epoch": 438.88,
168687
+ "learning_rate": 9.143125e-06,
168688
+ "loss": 0.3785,
168689
+ "step": 54490
168690
+ },
168691
+ {
168692
+ "epoch": 438.92,
168693
+ "learning_rate": 9.143044871794872e-06,
168694
+ "loss": 0.3476,
168695
+ "step": 54495
168696
+ },
168697
+ {
168698
+ "epoch": 438.96,
168699
+ "learning_rate": 9.142964743589743e-06,
168700
+ "loss": 0.4472,
168701
+ "step": 54500
168702
+ },
168703
+ {
168704
+ "epoch": 439.0,
168705
+ "learning_rate": 9.142884615384616e-06,
168706
+ "loss": 1.1279,
168707
+ "step": 54505
168708
+ },
168709
+ {
168710
+ "epoch": 439.0,
168711
+ "eval_loss": 0.4378458261489868,
168712
+ "eval_runtime": 39.018,
168713
+ "eval_samples_per_second": 21.529,
168714
+ "eval_steps_per_second": 0.692,
168715
+ "eval_wer": 0.19587552664646316,
168716
+ "step": 54505
168717
  }
168718
  ],
168719
+ "max_steps": 620000,
168720
  "num_train_epochs": 5000,
168721
+ "total_flos": 1.5338222874079727e+20,
168722
  "trial_name": null,
168723
  "trial_params": null
168724
  }
model-bin/finetune/base/{checkpoint-53883 β†’ checkpoint-54505}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629764971.3905332/events.out.tfevents.1629764971.c435e1c5ee04.920.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29684233bd0fc4fa53d622430a8437859d228e1ebd196acb25eeeea72e1a8f1c
3
+ size 4194
model-bin/finetune/base/log/1629765618.347005/events.out.tfevents.1629765618.c435e1c5ee04.920.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0549caa70a6959dff71a87146413f73853d214adadf9196b1be635a8986d790d
3
+ size 4194
model-bin/finetune/base/log/1629766288.11667/events.out.tfevents.1629766288.c435e1c5ee04.920.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0cc0b65297bee44aa7440209a58938fbbf58bd65000cd58866a923c9ebdf11a
3
+ size 4194
model-bin/finetune/base/log/1629766954.1795254/events.out.tfevents.1629766954.c435e1c5ee04.920.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da279e002f3b92c5bbaf9d3330dcf1c5dada61a83139b6d77bcff44ebd4c700a
3
+ size 4194
model-bin/finetune/base/log/1629767616.4417202/events.out.tfevents.1629767616.c435e1c5ee04.920.9 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc5cc1ea273768db4c516a02293a473944e45c923b417341b4901a28d7e11f7c
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629764971.c435e1c5ee04.920.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:715a5a3034e8b68c78630095bbb82afa35e33b2999fc887e193a2b427caf914c
3
+ size 8630
model-bin/finetune/base/log/events.out.tfevents.1629765618.c435e1c5ee04.920.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7dc7447d2a605ddad8639d873081d8f1f91dd34fe28727938fdba49261e39f2e
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629766288.c435e1c5ee04.920.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95b4ff4f8f7e5cc21e69b1a1df95a4c09ee29051c93d79b474b3291a209da371
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629766954.c435e1c5ee04.920.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f06cf4074fc3d80ff4404a5ce32673db1e7bc7194839e0f0e7041ce5d46a615c
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629767616.c435e1c5ee04.920.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e530f4e120d4d9a33abdc0271cce83168eca7ef2c15f6b01fe615709673f4362
3
+ size 8622