Training in progress, step 17070, checkpoint
Browse files
last-checkpoint/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 223144592
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cf29b318c3b664e026f2d7eeee5f490c11ed0e0a4f94588f656f978ca6f13931
|
| 3 |
size 223144592
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 281574266
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:514cbfa64013589609334f895930da6afc5abf2d3db2045db7212f033771f964
|
| 3 |
size 281574266
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:470e496359a2cf24c1ca6e38338e3dc7a3a2df79fa39edd6fc9efb7bc5840f7f
|
| 3 |
size 14244
|
last-checkpoint/scaler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 988
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f4c44df5fb2a26fd6b2c1b0589b6668747c62230efcf3a91f32d9c3226a7e389
|
| 3 |
size 988
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:72b8b6c65fad4bd8081d5b89d5e37065a8292222d82e3fc9240f4b2e717eb160
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -2,9 +2,9 @@
|
|
| 2 |
"best_global_step": 14000,
|
| 3 |
"best_metric": 0.18538166814028884,
|
| 4 |
"best_model_checkpoint": "./distil-whisper/checkpoint-14000",
|
| 5 |
-
"epoch":
|
| 6 |
"eval_steps": 1000,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -1365,12 +1365,12 @@
|
|
| 1365 |
"should_evaluate": false,
|
| 1366 |
"should_log": false,
|
| 1367 |
"should_save": true,
|
| 1368 |
-
"should_training_stop":
|
| 1369 |
},
|
| 1370 |
"attributes": {}
|
| 1371 |
}
|
| 1372 |
},
|
| 1373 |
-
"total_flos": 2.
|
| 1374 |
"train_batch_size": 8,
|
| 1375 |
"trial_name": null,
|
| 1376 |
"trial_params": null
|
|
|
|
| 2 |
"best_global_step": 14000,
|
| 3 |
"best_metric": 0.18538166814028884,
|
| 4 |
"best_model_checkpoint": "./distil-whisper/checkpoint-14000",
|
| 5 |
+
"epoch": 10.0,
|
| 6 |
"eval_steps": 1000,
|
| 7 |
+
"global_step": 17070,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 1365 |
"should_evaluate": false,
|
| 1366 |
"should_log": false,
|
| 1367 |
"should_save": true,
|
| 1368 |
+
"should_training_stop": true
|
| 1369 |
},
|
| 1370 |
"attributes": {}
|
| 1371 |
}
|
| 1372 |
},
|
| 1373 |
+
"total_flos": 2.2203395997696e+19,
|
| 1374 |
"train_batch_size": 8,
|
| 1375 |
"trial_name": null,
|
| 1376 |
"trial_params": null
|