Training in progress, step 15000, checkpoint
Browse files
last-checkpoint/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 223144592
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:88c828e9a5ac22cf6dc1cce8a9e4eba4907f38121d874041ad815e9f2a3c568b
|
| 3 |
size 223144592
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 281574266
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d6c863eb9dc692a7af5553639cc48b58879fa14bd82e27273107d9691b68e075
|
| 3 |
size 281574266
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:024dc893ba560d2de7dd4d2f0b7ab5729ba9520ec363dd17bb65cf05af476b2b
|
| 3 |
size 14244
|
last-checkpoint/scaler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 988
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5adfa8fca6cc5ca715985458253bcf718541f3a76aa3473952114e672d00e7b5
|
| 3 |
size 988
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:580e7db3eccdb01c1237319104eea4cc8a57ea956da0fe94439b4fbfeb184146
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -2,9 +2,9 @@
|
|
| 2 |
"best_global_step": 14000,
|
| 3 |
"best_metric": 0.18538166814028884,
|
| 4 |
"best_model_checkpoint": "./distil-whisper/checkpoint-14000",
|
| 5 |
-
"epoch": 8.
|
| 6 |
"eval_steps": 1000,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -1114,6 +1114,85 @@
|
|
| 1114 |
"eval_steps_per_second": 0.428,
|
| 1115 |
"eval_wer": 0.18538166814028884,
|
| 1116 |
"step": 14000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1117 |
}
|
| 1118 |
],
|
| 1119 |
"logging_steps": 100,
|
|
@@ -1133,7 +1212,7 @@
|
|
| 1133 |
"attributes": {}
|
| 1134 |
}
|
| 1135 |
},
|
| 1136 |
-
"total_flos": 1.
|
| 1137 |
"train_batch_size": 8,
|
| 1138 |
"trial_name": null,
|
| 1139 |
"trial_params": null
|
|
|
|
| 2 |
"best_global_step": 14000,
|
| 3 |
"best_metric": 0.18538166814028884,
|
| 4 |
"best_model_checkpoint": "./distil-whisper/checkpoint-14000",
|
| 5 |
+
"epoch": 8.787346221441124,
|
| 6 |
"eval_steps": 1000,
|
| 7 |
+
"global_step": 15000,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 1114 |
"eval_steps_per_second": 0.428,
|
| 1115 |
"eval_wer": 0.18538166814028884,
|
| 1116 |
"step": 14000
|
| 1117 |
+
},
|
| 1118 |
+
{
|
| 1119 |
+
"epoch": 8.260105448154658,
|
| 1120 |
+
"grad_norm": 7.585418701171875,
|
| 1121 |
+
"learning_rate": 1.7978273989136995e-05,
|
| 1122 |
+
"loss": 0.8818,
|
| 1123 |
+
"step": 14100
|
| 1124 |
+
},
|
| 1125 |
+
{
|
| 1126 |
+
"epoch": 8.318687756297598,
|
| 1127 |
+
"grad_norm": 9.436836242675781,
|
| 1128 |
+
"learning_rate": 1.7374773687386845e-05,
|
| 1129 |
+
"loss": 0.8864,
|
| 1130 |
+
"step": 14200
|
| 1131 |
+
},
|
| 1132 |
+
{
|
| 1133 |
+
"epoch": 8.37727006444054,
|
| 1134 |
+
"grad_norm": 12.12936019897461,
|
| 1135 |
+
"learning_rate": 1.6771273385636692e-05,
|
| 1136 |
+
"loss": 0.8744,
|
| 1137 |
+
"step": 14300
|
| 1138 |
+
},
|
| 1139 |
+
{
|
| 1140 |
+
"epoch": 8.43585237258348,
|
| 1141 |
+
"grad_norm": 11.584985733032227,
|
| 1142 |
+
"learning_rate": 1.6167773083886542e-05,
|
| 1143 |
+
"loss": 0.8542,
|
| 1144 |
+
"step": 14400
|
| 1145 |
+
},
|
| 1146 |
+
{
|
| 1147 |
+
"epoch": 8.494434680726421,
|
| 1148 |
+
"grad_norm": 7.6883440017700195,
|
| 1149 |
+
"learning_rate": 1.5564272782136393e-05,
|
| 1150 |
+
"loss": 0.8714,
|
| 1151 |
+
"step": 14500
|
| 1152 |
+
},
|
| 1153 |
+
{
|
| 1154 |
+
"epoch": 8.553016988869361,
|
| 1155 |
+
"grad_norm": 13.686609268188477,
|
| 1156 |
+
"learning_rate": 1.496077248038624e-05,
|
| 1157 |
+
"loss": 0.8726,
|
| 1158 |
+
"step": 14600
|
| 1159 |
+
},
|
| 1160 |
+
{
|
| 1161 |
+
"epoch": 8.611599297012303,
|
| 1162 |
+
"grad_norm": 10.200602531433105,
|
| 1163 |
+
"learning_rate": 1.436330718165359e-05,
|
| 1164 |
+
"loss": 0.8839,
|
| 1165 |
+
"step": 14700
|
| 1166 |
+
},
|
| 1167 |
+
{
|
| 1168 |
+
"epoch": 8.670181605155243,
|
| 1169 |
+
"grad_norm": 6.929018020629883,
|
| 1170 |
+
"learning_rate": 1.3759806879903441e-05,
|
| 1171 |
+
"loss": 0.8628,
|
| 1172 |
+
"step": 14800
|
| 1173 |
+
},
|
| 1174 |
+
{
|
| 1175 |
+
"epoch": 8.728763913298184,
|
| 1176 |
+
"grad_norm": 9.72988224029541,
|
| 1177 |
+
"learning_rate": 1.3156306578153291e-05,
|
| 1178 |
+
"loss": 0.8672,
|
| 1179 |
+
"step": 14900
|
| 1180 |
+
},
|
| 1181 |
+
{
|
| 1182 |
+
"epoch": 8.787346221441124,
|
| 1183 |
+
"grad_norm": 7.273561477661133,
|
| 1184 |
+
"learning_rate": 1.255280627640314e-05,
|
| 1185 |
+
"loss": 0.8293,
|
| 1186 |
+
"step": 15000
|
| 1187 |
+
},
|
| 1188 |
+
{
|
| 1189 |
+
"epoch": 8.787346221441124,
|
| 1190 |
+
"eval_loss": 0.08473628014326096,
|
| 1191 |
+
"eval_runtime": 147.7929,
|
| 1192 |
+
"eval_samples_per_second": 3.383,
|
| 1193 |
+
"eval_steps_per_second": 0.426,
|
| 1194 |
+
"eval_wer": 0.19834954317712938,
|
| 1195 |
+
"step": 15000
|
| 1196 |
}
|
| 1197 |
],
|
| 1198 |
"logging_steps": 100,
|
|
|
|
| 1212 |
"attributes": {}
|
| 1213 |
}
|
| 1214 |
},
|
| 1215 |
+
"total_flos": 1.951102081695744e+19,
|
| 1216 |
"train_batch_size": 8,
|
| 1217 |
"trial_name": null,
|
| 1218 |
"trial_params": null
|