Baselhany commited on
Commit
703450e
·
verified ·
1 Parent(s): 78c8163

Training in progress, step 15000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:04f91bc09632d0ca44900fee5a57b22786f0ffd1c54b0da4556e61fa28621f2f
3
  size 223144592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88c828e9a5ac22cf6dc1cce8a9e4eba4907f38121d874041ad815e9f2a3c568b
3
  size 223144592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ecc271fe9090035a3a652375c681b827d76b3bd635e7efcdc02a095e29e7c80
3
  size 281574266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6c863eb9dc692a7af5553639cc48b58879fa14bd82e27273107d9691b68e075
3
  size 281574266
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:613bc7490770dcf68fa9d54869bc0b3e8eab8d28558ef327003153a0f6445f20
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:024dc893ba560d2de7dd4d2f0b7ab5729ba9520ec363dd17bb65cf05af476b2b
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ef7c8d45fc5d3a46e66964194a46b9205f7ed9df86dc3ce05b4cb8e8f4d5993
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5adfa8fca6cc5ca715985458253bcf718541f3a76aa3473952114e672d00e7b5
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3fb722111df53fcc783bd102468cba8b679f6233ec8483bf8ae8f9c6d76a52fa
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:580e7db3eccdb01c1237319104eea4cc8a57ea956da0fe94439b4fbfeb184146
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 14000,
3
  "best_metric": 0.18538166814028884,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-14000",
5
- "epoch": 8.201523140011716,
6
  "eval_steps": 1000,
7
- "global_step": 14000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1114,6 +1114,85 @@
1114
  "eval_steps_per_second": 0.428,
1115
  "eval_wer": 0.18538166814028884,
1116
  "step": 14000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1117
  }
1118
  ],
1119
  "logging_steps": 100,
@@ -1133,7 +1212,7 @@
1133
  "attributes": {}
1134
  }
1135
  },
1136
- "total_flos": 1.821019937439744e+19,
1137
  "train_batch_size": 8,
1138
  "trial_name": null,
1139
  "trial_params": null
 
2
  "best_global_step": 14000,
3
  "best_metric": 0.18538166814028884,
4
  "best_model_checkpoint": "./distil-whisper/checkpoint-14000",
5
+ "epoch": 8.787346221441124,
6
  "eval_steps": 1000,
7
+ "global_step": 15000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1114
  "eval_steps_per_second": 0.428,
1115
  "eval_wer": 0.18538166814028884,
1116
  "step": 14000
1117
+ },
1118
+ {
1119
+ "epoch": 8.260105448154658,
1120
+ "grad_norm": 7.585418701171875,
1121
+ "learning_rate": 1.7978273989136995e-05,
1122
+ "loss": 0.8818,
1123
+ "step": 14100
1124
+ },
1125
+ {
1126
+ "epoch": 8.318687756297598,
1127
+ "grad_norm": 9.436836242675781,
1128
+ "learning_rate": 1.7374773687386845e-05,
1129
+ "loss": 0.8864,
1130
+ "step": 14200
1131
+ },
1132
+ {
1133
+ "epoch": 8.37727006444054,
1134
+ "grad_norm": 12.12936019897461,
1135
+ "learning_rate": 1.6771273385636692e-05,
1136
+ "loss": 0.8744,
1137
+ "step": 14300
1138
+ },
1139
+ {
1140
+ "epoch": 8.43585237258348,
1141
+ "grad_norm": 11.584985733032227,
1142
+ "learning_rate": 1.6167773083886542e-05,
1143
+ "loss": 0.8542,
1144
+ "step": 14400
1145
+ },
1146
+ {
1147
+ "epoch": 8.494434680726421,
1148
+ "grad_norm": 7.6883440017700195,
1149
+ "learning_rate": 1.5564272782136393e-05,
1150
+ "loss": 0.8714,
1151
+ "step": 14500
1152
+ },
1153
+ {
1154
+ "epoch": 8.553016988869361,
1155
+ "grad_norm": 13.686609268188477,
1156
+ "learning_rate": 1.496077248038624e-05,
1157
+ "loss": 0.8726,
1158
+ "step": 14600
1159
+ },
1160
+ {
1161
+ "epoch": 8.611599297012303,
1162
+ "grad_norm": 10.200602531433105,
1163
+ "learning_rate": 1.436330718165359e-05,
1164
+ "loss": 0.8839,
1165
+ "step": 14700
1166
+ },
1167
+ {
1168
+ "epoch": 8.670181605155243,
1169
+ "grad_norm": 6.929018020629883,
1170
+ "learning_rate": 1.3759806879903441e-05,
1171
+ "loss": 0.8628,
1172
+ "step": 14800
1173
+ },
1174
+ {
1175
+ "epoch": 8.728763913298184,
1176
+ "grad_norm": 9.72988224029541,
1177
+ "learning_rate": 1.3156306578153291e-05,
1178
+ "loss": 0.8672,
1179
+ "step": 14900
1180
+ },
1181
+ {
1182
+ "epoch": 8.787346221441124,
1183
+ "grad_norm": 7.273561477661133,
1184
+ "learning_rate": 1.255280627640314e-05,
1185
+ "loss": 0.8293,
1186
+ "step": 15000
1187
+ },
1188
+ {
1189
+ "epoch": 8.787346221441124,
1190
+ "eval_loss": 0.08473628014326096,
1191
+ "eval_runtime": 147.7929,
1192
+ "eval_samples_per_second": 3.383,
1193
+ "eval_steps_per_second": 0.426,
1194
+ "eval_wer": 0.19834954317712938,
1195
+ "step": 15000
1196
  }
1197
  ],
1198
  "logging_steps": 100,
 
1212
  "attributes": {}
1213
  }
1214
  },
1215
+ "total_flos": 1.951102081695744e+19,
1216
  "train_batch_size": 8,
1217
  "trial_name": null,
1218
  "trial_params": null