Check commited on
Commit
c4d5bec
Β·
1 Parent(s): d344c28

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/rng_state.pth +2 -2
  6. model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/trainer_state.json +799 -4
  9. model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630038501.523142/events.out.tfevents.1630038501.52f5c7e305a3.886.41 +3 -0
  11. model-bin/finetune/base/log/1630038959.4217858/events.out.tfevents.1630038959.52f5c7e305a3.886.43 +3 -0
  12. model-bin/finetune/base/log/1630039419.2220693/events.out.tfevents.1630039419.52f5c7e305a3.886.45 +3 -0
  13. model-bin/finetune/base/log/1630039886.0656877/events.out.tfevents.1630039886.52f5c7e305a3.886.47 +3 -0
  14. model-bin/finetune/base/log/1630040347.1132815/events.out.tfevents.1630040347.52f5c7e305a3.886.49 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630038501.52f5c7e305a3.886.40 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630038959.52f5c7e305a3.886.42 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630039419.52f5c7e305a3.886.44 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630039886.52f5c7e305a3.886.46 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630040347.52f5c7e305a3.886.48 +3 -0
model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5ed84f90fce277309fcc079c636a580b0a9951877365de9d22e0a45b325f4959
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:449094c3a4589cfafc88ccf9c5aee1fad6b6fb1bc9fd7a1b26db1c8a887d5b97
3
  size 722165393
model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:39714b9078cb72ab6d02184b330358f312f7c0beda11d9d911841507ed9f4072
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c6db6a0631133291652de05f3704ed2b1bbc9427592a09cb7bb6f9bf638a798
3
  size 377909911
model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:11325d828073bc28dba0e547a02b539378ed66d4d97bcd76ac0fa374d53c78a0
3
- size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8d8bd78bcf53e365811adc8ff3da7658419c1c8f347160280fcdc767485f38f
3
+ size 14567
model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b566898de17294f298a5b112ea4a529d2e3e5586226b104d900242be9b719cae
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22ee433ca4bee95d3224c650687a9b377d4580c10728650c0f178ed2466c52ae
3
  size 559
model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aa4b134d7e3400d11766c99ee8aa673798295021b7bb19c2a048b99ee2bff2af
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c18ec007f24c8f92d68d5a7078c003d85cec6eecdd191adcafcd27465e7379f9
3
  size 623
model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
- "epoch": 878.0,
5
- "global_step": 108894,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -237909,11 +237909,806 @@
237909
  "eval_steps_per_second": 0.657,
237910
  "eval_wer": 0.17886361930496134,
237911
  "step": 108894
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
237912
  }
237913
  ],
237914
- "max_steps": 620000,
237915
  "num_train_epochs": 5000,
237916
- "total_flos": 3.064483777918432e+20,
237917
  "trial_name": null,
237918
  "trial_params": null
237919
  }
 
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
+ "epoch": 876.0,
5
+ "global_step": 109516,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
237909
  "eval_steps_per_second": 0.657,
237910
  "eval_wer": 0.17886361930496134,
237911
  "step": 108894
237912
+ },
237913
+ {
237914
+ "epoch": 878.01,
237915
+ "learning_rate": 8.257641357027465e-06,
237916
+ "loss": 0.4942,
237917
+ "step": 108895
237918
+ },
237919
+ {
237920
+ "epoch": 878.05,
237921
+ "learning_rate": 8.257560581583199e-06,
237922
+ "loss": 0.3136,
237923
+ "step": 108900
237924
+ },
237925
+ {
237926
+ "epoch": 878.09,
237927
+ "learning_rate": 8.257479806138935e-06,
237928
+ "loss": 0.2701,
237929
+ "step": 108905
237930
+ },
237931
+ {
237932
+ "epoch": 878.13,
237933
+ "learning_rate": 8.257399030694669e-06,
237934
+ "loss": 0.3846,
237935
+ "step": 108910
237936
+ },
237937
+ {
237938
+ "epoch": 878.17,
237939
+ "learning_rate": 8.257318255250404e-06,
237940
+ "loss": 0.5586,
237941
+ "step": 108915
237942
+ },
237943
+ {
237944
+ "epoch": 878.21,
237945
+ "learning_rate": 8.257237479806139e-06,
237946
+ "loss": 1.1794,
237947
+ "step": 108920
237948
+ },
237949
+ {
237950
+ "epoch": 878.25,
237951
+ "learning_rate": 8.257156704361874e-06,
237952
+ "loss": 0.2967,
237953
+ "step": 108925
237954
+ },
237955
+ {
237956
+ "epoch": 878.29,
237957
+ "learning_rate": 8.257075928917609e-06,
237958
+ "loss": 0.2592,
237959
+ "step": 108930
237960
+ },
237961
+ {
237962
+ "epoch": 878.33,
237963
+ "learning_rate": 8.256995153473344e-06,
237964
+ "loss": 0.7212,
237965
+ "step": 108935
237966
+ },
237967
+ {
237968
+ "epoch": 878.37,
237969
+ "learning_rate": 8.25691437802908e-06,
237970
+ "loss": 0.595,
237971
+ "step": 108940
237972
+ },
237973
+ {
237974
+ "epoch": 878.41,
237975
+ "learning_rate": 8.256833602584814e-06,
237976
+ "loss": 1.1496,
237977
+ "step": 108945
237978
+ },
237979
+ {
237980
+ "epoch": 878.45,
237981
+ "learning_rate": 8.25675282714055e-06,
237982
+ "loss": 0.2888,
237983
+ "step": 108950
237984
+ },
237985
+ {
237986
+ "epoch": 878.49,
237987
+ "learning_rate": 8.256672051696284e-06,
237988
+ "loss": 0.4588,
237989
+ "step": 108955
237990
+ },
237991
+ {
237992
+ "epoch": 878.53,
237993
+ "learning_rate": 8.25659127625202e-06,
237994
+ "loss": 0.3674,
237995
+ "step": 108960
237996
+ },
237997
+ {
237998
+ "epoch": 878.57,
237999
+ "learning_rate": 8.256510500807754e-06,
238000
+ "loss": 0.6112,
238001
+ "step": 108965
238002
+ },
238003
+ {
238004
+ "epoch": 878.61,
238005
+ "learning_rate": 8.25642972536349e-06,
238006
+ "loss": 1.0556,
238007
+ "step": 108970
238008
+ },
238009
+ {
238010
+ "epoch": 878.65,
238011
+ "learning_rate": 8.256348949919224e-06,
238012
+ "loss": 0.2855,
238013
+ "step": 108975
238014
+ },
238015
+ {
238016
+ "epoch": 878.69,
238017
+ "learning_rate": 8.25626817447496e-06,
238018
+ "loss": 0.2686,
238019
+ "step": 108980
238020
+ },
238021
+ {
238022
+ "epoch": 878.73,
238023
+ "learning_rate": 8.256187399030694e-06,
238024
+ "loss": 0.3335,
238025
+ "step": 108985
238026
+ },
238027
+ {
238028
+ "epoch": 878.77,
238029
+ "learning_rate": 8.25610662358643e-06,
238030
+ "loss": 0.534,
238031
+ "step": 108990
238032
+ },
238033
+ {
238034
+ "epoch": 878.81,
238035
+ "learning_rate": 8.256025848142164e-06,
238036
+ "loss": 1.018,
238037
+ "step": 108995
238038
+ },
238039
+ {
238040
+ "epoch": 878.85,
238041
+ "learning_rate": 8.2559450726979e-06,
238042
+ "loss": 0.3416,
238043
+ "step": 109000
238044
+ },
238045
+ {
238046
+ "epoch": 878.89,
238047
+ "learning_rate": 8.255864297253636e-06,
238048
+ "loss": 0.2572,
238049
+ "step": 109005
238050
+ },
238051
+ {
238052
+ "epoch": 878.93,
238053
+ "learning_rate": 8.25578352180937e-06,
238054
+ "loss": 0.3844,
238055
+ "step": 109010
238056
+ },
238057
+ {
238058
+ "epoch": 878.97,
238059
+ "learning_rate": 8.255702746365106e-06,
238060
+ "loss": 0.5158,
238061
+ "step": 109015
238062
+ },
238063
+ {
238064
+ "epoch": 879.0,
238065
+ "eval_loss": 0.3908926248550415,
238066
+ "eval_runtime": 42.6009,
238067
+ "eval_samples_per_second": 19.647,
238068
+ "eval_steps_per_second": 0.634,
238069
+ "eval_wer": 0.1916660648515924,
238070
+ "step": 109018
238071
+ },
238072
+ {
238073
+ "epoch": 879.02,
238074
+ "learning_rate": 8.25562197092084e-06,
238075
+ "loss": 0.4281,
238076
+ "step": 109020
238077
+ },
238078
+ {
238079
+ "epoch": 879.06,
238080
+ "learning_rate": 8.255541195476576e-06,
238081
+ "loss": 0.2807,
238082
+ "step": 109025
238083
+ },
238084
+ {
238085
+ "epoch": 879.1,
238086
+ "learning_rate": 8.25546042003231e-06,
238087
+ "loss": 0.2722,
238088
+ "step": 109030
238089
+ },
238090
+ {
238091
+ "epoch": 879.14,
238092
+ "learning_rate": 8.255379644588046e-06,
238093
+ "loss": 0.3234,
238094
+ "step": 109035
238095
+ },
238096
+ {
238097
+ "epoch": 879.18,
238098
+ "learning_rate": 8.25529886914378e-06,
238099
+ "loss": 0.6337,
238100
+ "step": 109040
238101
+ },
238102
+ {
238103
+ "epoch": 879.22,
238104
+ "learning_rate": 8.255218093699516e-06,
238105
+ "loss": 0.9453,
238106
+ "step": 109045
238107
+ },
238108
+ {
238109
+ "epoch": 879.26,
238110
+ "learning_rate": 8.25513731825525e-06,
238111
+ "loss": 0.3167,
238112
+ "step": 109050
238113
+ },
238114
+ {
238115
+ "epoch": 879.3,
238116
+ "learning_rate": 8.255056542810986e-06,
238117
+ "loss": 0.3273,
238118
+ "step": 109055
238119
+ },
238120
+ {
238121
+ "epoch": 879.34,
238122
+ "learning_rate": 8.25497576736672e-06,
238123
+ "loss": 0.3399,
238124
+ "step": 109060
238125
+ },
238126
+ {
238127
+ "epoch": 879.38,
238128
+ "learning_rate": 8.254894991922456e-06,
238129
+ "loss": 0.6348,
238130
+ "step": 109065
238131
+ },
238132
+ {
238133
+ "epoch": 879.42,
238134
+ "learning_rate": 8.254814216478192e-06,
238135
+ "loss": 1.0596,
238136
+ "step": 109070
238137
+ },
238138
+ {
238139
+ "epoch": 879.46,
238140
+ "learning_rate": 8.254733441033926e-06,
238141
+ "loss": 0.2522,
238142
+ "step": 109075
238143
+ },
238144
+ {
238145
+ "epoch": 879.5,
238146
+ "learning_rate": 8.254652665589662e-06,
238147
+ "loss": 0.2868,
238148
+ "step": 109080
238149
+ },
238150
+ {
238151
+ "epoch": 879.54,
238152
+ "learning_rate": 8.254571890145396e-06,
238153
+ "loss": 0.3137,
238154
+ "step": 109085
238155
+ },
238156
+ {
238157
+ "epoch": 879.58,
238158
+ "learning_rate": 8.254491114701132e-06,
238159
+ "loss": 0.6219,
238160
+ "step": 109090
238161
+ },
238162
+ {
238163
+ "epoch": 879.62,
238164
+ "learning_rate": 8.254410339256866e-06,
238165
+ "loss": 0.9126,
238166
+ "step": 109095
238167
+ },
238168
+ {
238169
+ "epoch": 879.66,
238170
+ "learning_rate": 8.254329563812602e-06,
238171
+ "loss": 0.3179,
238172
+ "step": 109100
238173
+ },
238174
+ {
238175
+ "epoch": 879.7,
238176
+ "learning_rate": 8.254248788368336e-06,
238177
+ "loss": 0.2617,
238178
+ "step": 109105
238179
+ },
238180
+ {
238181
+ "epoch": 879.74,
238182
+ "learning_rate": 8.254168012924072e-06,
238183
+ "loss": 0.3084,
238184
+ "step": 109110
238185
+ },
238186
+ {
238187
+ "epoch": 879.78,
238188
+ "learning_rate": 8.254087237479806e-06,
238189
+ "loss": 0.5737,
238190
+ "step": 109115
238191
+ },
238192
+ {
238193
+ "epoch": 879.82,
238194
+ "learning_rate": 8.254006462035542e-06,
238195
+ "loss": 1.0034,
238196
+ "step": 109120
238197
+ },
238198
+ {
238199
+ "epoch": 879.86,
238200
+ "learning_rate": 8.253925686591276e-06,
238201
+ "loss": 0.2856,
238202
+ "step": 109125
238203
+ },
238204
+ {
238205
+ "epoch": 879.9,
238206
+ "learning_rate": 8.253844911147011e-06,
238207
+ "loss": 0.2544,
238208
+ "step": 109130
238209
+ },
238210
+ {
238211
+ "epoch": 879.94,
238212
+ "learning_rate": 8.253764135702747e-06,
238213
+ "loss": 0.3562,
238214
+ "step": 109135
238215
+ },
238216
+ {
238217
+ "epoch": 879.98,
238218
+ "learning_rate": 8.253683360258481e-06,
238219
+ "loss": 0.6451,
238220
+ "step": 109140
238221
+ },
238222
+ {
238223
+ "epoch": 880.0,
238224
+ "eval_loss": 0.4516902267932892,
238225
+ "eval_runtime": 42.7375,
238226
+ "eval_samples_per_second": 19.608,
238227
+ "eval_steps_per_second": 0.632,
238228
+ "eval_wer": 0.1786863662960796,
238229
+ "step": 109142
238230
+ },
238231
+ {
238232
+ "epoch": 873.02,
238233
+ "learning_rate": 8.253602584814217e-06,
238234
+ "loss": 0.3657,
238235
+ "step": 109145
238236
+ },
238237
+ {
238238
+ "epoch": 873.06,
238239
+ "learning_rate": 8.253521809369951e-06,
238240
+ "loss": 0.2349,
238241
+ "step": 109150
238242
+ },
238243
+ {
238244
+ "epoch": 873.1,
238245
+ "learning_rate": 8.253441033925687e-06,
238246
+ "loss": 0.3169,
238247
+ "step": 109155
238248
+ },
238249
+ {
238250
+ "epoch": 873.14,
238251
+ "learning_rate": 8.253360258481421e-06,
238252
+ "loss": 0.3745,
238253
+ "step": 109160
238254
+ },
238255
+ {
238256
+ "epoch": 873.18,
238257
+ "learning_rate": 8.253279483037157e-06,
238258
+ "loss": 0.7199,
238259
+ "step": 109165
238260
+ },
238261
+ {
238262
+ "epoch": 873.22,
238263
+ "learning_rate": 8.253198707592891e-06,
238264
+ "loss": 1.048,
238265
+ "step": 109170
238266
+ },
238267
+ {
238268
+ "epoch": 873.26,
238269
+ "learning_rate": 8.253117932148627e-06,
238270
+ "loss": 0.2758,
238271
+ "step": 109175
238272
+ },
238273
+ {
238274
+ "epoch": 873.3,
238275
+ "learning_rate": 8.253037156704361e-06,
238276
+ "loss": 0.2872,
238277
+ "step": 109180
238278
+ },
238279
+ {
238280
+ "epoch": 873.34,
238281
+ "learning_rate": 8.252956381260097e-06,
238282
+ "loss": 0.3778,
238283
+ "step": 109185
238284
+ },
238285
+ {
238286
+ "epoch": 873.38,
238287
+ "learning_rate": 8.252875605815831e-06,
238288
+ "loss": 0.8357,
238289
+ "step": 109190
238290
+ },
238291
+ {
238292
+ "epoch": 873.42,
238293
+ "learning_rate": 8.252794830371567e-06,
238294
+ "loss": 0.9124,
238295
+ "step": 109195
238296
+ },
238297
+ {
238298
+ "epoch": 873.46,
238299
+ "learning_rate": 8.252714054927303e-06,
238300
+ "loss": 0.2622,
238301
+ "step": 109200
238302
+ },
238303
+ {
238304
+ "epoch": 873.5,
238305
+ "learning_rate": 8.252633279483037e-06,
238306
+ "loss": 0.2493,
238307
+ "step": 109205
238308
+ },
238309
+ {
238310
+ "epoch": 873.54,
238311
+ "learning_rate": 8.252552504038773e-06,
238312
+ "loss": 0.3982,
238313
+ "step": 109210
238314
+ },
238315
+ {
238316
+ "epoch": 873.58,
238317
+ "learning_rate": 8.252471728594507e-06,
238318
+ "loss": 0.7649,
238319
+ "step": 109215
238320
+ },
238321
+ {
238322
+ "epoch": 873.62,
238323
+ "learning_rate": 8.252390953150243e-06,
238324
+ "loss": 0.9403,
238325
+ "step": 109220
238326
+ },
238327
+ {
238328
+ "epoch": 873.66,
238329
+ "learning_rate": 8.252310177705977e-06,
238330
+ "loss": 0.2741,
238331
+ "step": 109225
238332
+ },
238333
+ {
238334
+ "epoch": 873.7,
238335
+ "learning_rate": 8.252229402261713e-06,
238336
+ "loss": 0.2522,
238337
+ "step": 109230
238338
+ },
238339
+ {
238340
+ "epoch": 873.74,
238341
+ "learning_rate": 8.252148626817447e-06,
238342
+ "loss": 0.3822,
238343
+ "step": 109235
238344
+ },
238345
+ {
238346
+ "epoch": 873.78,
238347
+ "learning_rate": 8.252067851373183e-06,
238348
+ "loss": 0.7637,
238349
+ "step": 109240
238350
+ },
238351
+ {
238352
+ "epoch": 873.82,
238353
+ "learning_rate": 8.251987075928917e-06,
238354
+ "loss": 0.8275,
238355
+ "step": 109245
238356
+ },
238357
+ {
238358
+ "epoch": 873.86,
238359
+ "learning_rate": 8.251906300484653e-06,
238360
+ "loss": 0.3128,
238361
+ "step": 109250
238362
+ },
238363
+ {
238364
+ "epoch": 873.9,
238365
+ "learning_rate": 8.251825525040389e-06,
238366
+ "loss": 0.3143,
238367
+ "step": 109255
238368
+ },
238369
+ {
238370
+ "epoch": 873.94,
238371
+ "learning_rate": 8.251744749596123e-06,
238372
+ "loss": 0.4088,
238373
+ "step": 109260
238374
+ },
238375
+ {
238376
+ "epoch": 873.98,
238377
+ "learning_rate": 8.251663974151859e-06,
238378
+ "loss": 0.7688,
238379
+ "step": 109265
238380
+ },
238381
+ {
238382
+ "epoch": 874.0,
238383
+ "eval_loss": 0.4127758741378784,
238384
+ "eval_runtime": 42.9711,
238385
+ "eval_samples_per_second": 19.501,
238386
+ "eval_steps_per_second": 0.628,
238387
+ "eval_wer": 0.18809031877213694,
238388
+ "step": 109267
238389
+ },
238390
+ {
238391
+ "epoch": 881.02,
238392
+ "learning_rate": 8.251583198707593e-06,
238393
+ "loss": 0.3105,
238394
+ "step": 109270
238395
+ },
238396
+ {
238397
+ "epoch": 881.06,
238398
+ "learning_rate": 8.251502423263329e-06,
238399
+ "loss": 0.4217,
238400
+ "step": 109275
238401
+ },
238402
+ {
238403
+ "epoch": 881.1,
238404
+ "learning_rate": 8.251421647819063e-06,
238405
+ "loss": 0.2634,
238406
+ "step": 109280
238407
+ },
238408
+ {
238409
+ "epoch": 881.14,
238410
+ "learning_rate": 8.251340872374799e-06,
238411
+ "loss": 0.3549,
238412
+ "step": 109285
238413
+ },
238414
+ {
238415
+ "epoch": 881.18,
238416
+ "learning_rate": 8.251260096930533e-06,
238417
+ "loss": 0.7516,
238418
+ "step": 109290
238419
+ },
238420
+ {
238421
+ "epoch": 881.22,
238422
+ "learning_rate": 8.251179321486269e-06,
238423
+ "loss": 0.8462,
238424
+ "step": 109295
238425
+ },
238426
+ {
238427
+ "epoch": 881.27,
238428
+ "learning_rate": 8.251098546042003e-06,
238429
+ "loss": 0.2915,
238430
+ "step": 109300
238431
+ },
238432
+ {
238433
+ "epoch": 881.31,
238434
+ "learning_rate": 8.251017770597739e-06,
238435
+ "loss": 0.2822,
238436
+ "step": 109305
238437
+ },
238438
+ {
238439
+ "epoch": 881.35,
238440
+ "learning_rate": 8.250936995153474e-06,
238441
+ "loss": 0.3588,
238442
+ "step": 109310
238443
+ },
238444
+ {
238445
+ "epoch": 881.39,
238446
+ "learning_rate": 8.250856219709209e-06,
238447
+ "loss": 0.772,
238448
+ "step": 109315
238449
+ },
238450
+ {
238451
+ "epoch": 881.43,
238452
+ "learning_rate": 8.250775444264944e-06,
238453
+ "loss": 0.8989,
238454
+ "step": 109320
238455
+ },
238456
+ {
238457
+ "epoch": 881.47,
238458
+ "learning_rate": 8.250694668820679e-06,
238459
+ "loss": 0.2993,
238460
+ "step": 109325
238461
+ },
238462
+ {
238463
+ "epoch": 881.51,
238464
+ "learning_rate": 8.250613893376414e-06,
238465
+ "loss": 0.2748,
238466
+ "step": 109330
238467
+ },
238468
+ {
238469
+ "epoch": 881.55,
238470
+ "learning_rate": 8.250533117932149e-06,
238471
+ "loss": 0.3367,
238472
+ "step": 109335
238473
+ },
238474
+ {
238475
+ "epoch": 881.59,
238476
+ "learning_rate": 8.250452342487884e-06,
238477
+ "loss": 0.6681,
238478
+ "step": 109340
238479
+ },
238480
+ {
238481
+ "epoch": 881.63,
238482
+ "learning_rate": 8.250371567043619e-06,
238483
+ "loss": 0.7778,
238484
+ "step": 109345
238485
+ },
238486
+ {
238487
+ "epoch": 881.67,
238488
+ "learning_rate": 8.250290791599354e-06,
238489
+ "loss": 0.2912,
238490
+ "step": 109350
238491
+ },
238492
+ {
238493
+ "epoch": 881.71,
238494
+ "learning_rate": 8.250210016155088e-06,
238495
+ "loss": 0.2896,
238496
+ "step": 109355
238497
+ },
238498
+ {
238499
+ "epoch": 881.75,
238500
+ "learning_rate": 8.250129240710824e-06,
238501
+ "loss": 0.3983,
238502
+ "step": 109360
238503
+ },
238504
+ {
238505
+ "epoch": 881.79,
238506
+ "learning_rate": 8.250048465266558e-06,
238507
+ "loss": 0.762,
238508
+ "step": 109365
238509
+ },
238510
+ {
238511
+ "epoch": 881.83,
238512
+ "learning_rate": 8.249967689822294e-06,
238513
+ "loss": 1.0731,
238514
+ "step": 109370
238515
+ },
238516
+ {
238517
+ "epoch": 881.87,
238518
+ "learning_rate": 8.24988691437803e-06,
238519
+ "loss": 0.2488,
238520
+ "step": 109375
238521
+ },
238522
+ {
238523
+ "epoch": 881.91,
238524
+ "learning_rate": 8.249806138933764e-06,
238525
+ "loss": 0.4398,
238526
+ "step": 109380
238527
+ },
238528
+ {
238529
+ "epoch": 881.95,
238530
+ "learning_rate": 8.2497253634895e-06,
238531
+ "loss": 0.3585,
238532
+ "step": 109385
238533
+ },
238534
+ {
238535
+ "epoch": 881.99,
238536
+ "learning_rate": 8.249644588045234e-06,
238537
+ "loss": 0.8093,
238538
+ "step": 109390
238539
+ },
238540
+ {
238541
+ "epoch": 882.0,
238542
+ "eval_loss": 0.36528652906417847,
238543
+ "eval_runtime": 41.9391,
238544
+ "eval_samples_per_second": 19.981,
238545
+ "eval_steps_per_second": 0.644,
238546
+ "eval_wer": 0.1787843536425767,
238547
+ "step": 109391
238548
+ },
238549
+ {
238550
+ "epoch": 875.03,
238551
+ "learning_rate": 8.24956381260097e-06,
238552
+ "loss": 0.3458,
238553
+ "step": 109395
238554
+ },
238555
+ {
238556
+ "epoch": 875.07,
238557
+ "learning_rate": 8.249483037156704e-06,
238558
+ "loss": 0.2919,
238559
+ "step": 109400
238560
+ },
238561
+ {
238562
+ "epoch": 875.11,
238563
+ "learning_rate": 8.24940226171244e-06,
238564
+ "loss": 0.2769,
238565
+ "step": 109405
238566
+ },
238567
+ {
238568
+ "epoch": 875.15,
238569
+ "learning_rate": 8.249321486268174e-06,
238570
+ "loss": 0.3908,
238571
+ "step": 109410
238572
+ },
238573
+ {
238574
+ "epoch": 875.19,
238575
+ "learning_rate": 8.24924071082391e-06,
238576
+ "loss": 0.9022,
238577
+ "step": 109415
238578
+ },
238579
+ {
238580
+ "epoch": 875.23,
238581
+ "learning_rate": 8.249159935379644e-06,
238582
+ "loss": 0.6427,
238583
+ "step": 109420
238584
+ },
238585
+ {
238586
+ "epoch": 875.27,
238587
+ "learning_rate": 8.24907915993538e-06,
238588
+ "loss": 0.2571,
238589
+ "step": 109425
238590
+ },
238591
+ {
238592
+ "epoch": 875.31,
238593
+ "learning_rate": 8.248998384491116e-06,
238594
+ "loss": 0.3107,
238595
+ "step": 109430
238596
+ },
238597
+ {
238598
+ "epoch": 875.35,
238599
+ "learning_rate": 8.24891760904685e-06,
238600
+ "loss": 0.4479,
238601
+ "step": 109435
238602
+ },
238603
+ {
238604
+ "epoch": 875.39,
238605
+ "learning_rate": 8.248836833602586e-06,
238606
+ "loss": 0.9061,
238607
+ "step": 109440
238608
+ },
238609
+ {
238610
+ "epoch": 875.43,
238611
+ "learning_rate": 8.24875605815832e-06,
238612
+ "loss": 0.5396,
238613
+ "step": 109445
238614
+ },
238615
+ {
238616
+ "epoch": 875.47,
238617
+ "learning_rate": 8.248675282714056e-06,
238618
+ "loss": 0.2692,
238619
+ "step": 109450
238620
+ },
238621
+ {
238622
+ "epoch": 875.51,
238623
+ "learning_rate": 8.24859450726979e-06,
238624
+ "loss": 0.3434,
238625
+ "step": 109455
238626
+ },
238627
+ {
238628
+ "epoch": 875.55,
238629
+ "learning_rate": 8.248513731825526e-06,
238630
+ "loss": 0.3502,
238631
+ "step": 109460
238632
+ },
238633
+ {
238634
+ "epoch": 875.59,
238635
+ "learning_rate": 8.24843295638126e-06,
238636
+ "loss": 0.9187,
238637
+ "step": 109465
238638
+ },
238639
+ {
238640
+ "epoch": 875.63,
238641
+ "learning_rate": 8.248352180936996e-06,
238642
+ "loss": 0.7583,
238643
+ "step": 109470
238644
+ },
238645
+ {
238646
+ "epoch": 875.67,
238647
+ "learning_rate": 8.24827140549273e-06,
238648
+ "loss": 0.2532,
238649
+ "step": 109475
238650
+ },
238651
+ {
238652
+ "epoch": 875.71,
238653
+ "learning_rate": 8.248190630048466e-06,
238654
+ "loss": 0.2976,
238655
+ "step": 109480
238656
+ },
238657
+ {
238658
+ "epoch": 875.75,
238659
+ "learning_rate": 8.248109854604202e-06,
238660
+ "loss": 0.3528,
238661
+ "step": 109485
238662
+ },
238663
+ {
238664
+ "epoch": 875.79,
238665
+ "learning_rate": 8.248029079159936e-06,
238666
+ "loss": 0.9945,
238667
+ "step": 109490
238668
+ },
238669
+ {
238670
+ "epoch": 875.83,
238671
+ "learning_rate": 8.247948303715672e-06,
238672
+ "loss": 0.7041,
238673
+ "step": 109495
238674
+ },
238675
+ {
238676
+ "epoch": 875.87,
238677
+ "learning_rate": 8.247867528271406e-06,
238678
+ "loss": 0.3011,
238679
+ "step": 109500
238680
+ },
238681
+ {
238682
+ "epoch": 875.91,
238683
+ "learning_rate": 8.247786752827141e-06,
238684
+ "loss": 0.3035,
238685
+ "step": 109505
238686
+ },
238687
+ {
238688
+ "epoch": 875.95,
238689
+ "learning_rate": 8.247705977382876e-06,
238690
+ "loss": 0.4184,
238691
+ "step": 109510
238692
+ },
238693
+ {
238694
+ "epoch": 875.99,
238695
+ "learning_rate": 8.247625201938611e-06,
238696
+ "loss": 0.9791,
238697
+ "step": 109515
238698
+ },
238699
+ {
238700
+ "epoch": 876.0,
238701
+ "eval_loss": 0.4340699315071106,
238702
+ "eval_runtime": 41.6362,
238703
+ "eval_samples_per_second": 20.127,
238704
+ "eval_steps_per_second": 0.648,
238705
+ "eval_wer": 0.18675573627650305,
238706
+ "step": 109516
238707
  }
238708
  ],
238709
+ "max_steps": 625000,
238710
  "num_train_epochs": 5000,
238711
+ "total_flos": 3.082023996552078e+20,
238712
  "trial_name": null,
238713
  "trial_params": null
238714
  }
model-bin/finetune/base/{checkpoint-108894 β†’ checkpoint-109516}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630038501.523142/events.out.tfevents.1630038501.52f5c7e305a3.886.41 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c401ed62b484604a23848e19989c75b32520117d9550bddd70b4ae6902f2d01
3
+ size 4194
model-bin/finetune/base/log/1630038959.4217858/events.out.tfevents.1630038959.52f5c7e305a3.886.43 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9babea275954112e093ee713694ed7972e862d7f606a9dbe28299c5046ec1e2
3
+ size 4194
model-bin/finetune/base/log/1630039419.2220693/events.out.tfevents.1630039419.52f5c7e305a3.886.45 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:134c9b3b7493265d59b258af820043544de2279a77cfd089fe5c8af3bfb0ff50
3
+ size 4194
model-bin/finetune/base/log/1630039886.0656877/events.out.tfevents.1630039886.52f5c7e305a3.886.47 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c5fbd6ba4ce85628c95c09a0c3f2fd73354a8dcd2ef4877c286639d2dbf9b73
3
+ size 4194
model-bin/finetune/base/log/1630040347.1132815/events.out.tfevents.1630040347.52f5c7e305a3.886.49 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e7199ab5167f975832a728d4180880640c77ba4410e03f83aec4b85ed6ab745
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630038501.52f5c7e305a3.886.40 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07f45d8b7f1f11c9b5196500c64d74f13d34c269482330d6da6e4254ea9f646e
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630038959.52f5c7e305a3.886.42 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:466b30ec82d7b2d6b26d0774df616401eff4867121ff67c4082db2e7602dedf9
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630039419.52f5c7e305a3.886.44 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50405b6d7beae2d9e44615d56dd9bf3572a9ca311f0f9c514052115386d878e3
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630039886.52f5c7e305a3.886.46 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa3da51855c3b23c2474cd23ddfca3a9e8a1afc730e7c04014131f8e82e6fdeb
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630040347.52f5c7e305a3.886.48 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4263db218c3486300f69f56c5ca2c4db0a06692c6fe89cb1de0d5cd7171f9baa
3
+ size 8622