Training in progress, step 16512, checkpoint
Browse files- last-checkpoint/adapter_model.safetensors +1 -1
- last-checkpoint/global_step16512/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
- last-checkpoint/global_step16512/mp_rank_00_model_states.pt +3 -0
- last-checkpoint/latest +1 -1
- last-checkpoint/rng_state.pth +1 -1
- last-checkpoint/trainer_state.json +4 -4
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 12017472
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0c52a6dd7e5bc676bd35e2ff08757c004c9a77792d247e5966821ce4cda29d1f
|
| 3 |
size 12017472
|
last-checkpoint/global_step16512/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fecc1fd44558636bf7291a837176c0c5d7117ea5d1c6ded1655d2dd3ac118e40
|
| 3 |
+
size 71982309
|
last-checkpoint/global_step16512/mp_rank_00_model_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:04b5d77557f0687686c909d8ecd0930cb52e3c8d3394f7825febe171c4601f55
|
| 3 |
+
size 146356645
|
last-checkpoint/latest
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
|
|
|
|
| 1 |
+
global_step16512
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14709
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9be458c8cf8debad5d6e2a3bd43cb667b3d77295003bd12c52999f7972324e51
|
| 3 |
size 14709
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -2,9 +2,9 @@
|
|
| 2 |
"best_global_step": 16000,
|
| 3 |
"best_metric": 0.5378558039665222,
|
| 4 |
"best_model_checkpoint": "/root/leap-finetune/outputs/sft/lfm2_350m_marathi_optimized_12ep/checkpoint-16000",
|
| 5 |
-
"epoch":
|
| 6 |
"eval_steps": 250,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -6631,12 +6631,12 @@
|
|
| 6631 |
"should_evaluate": false,
|
| 6632 |
"should_log": false,
|
| 6633 |
"should_save": true,
|
| 6634 |
-
"should_training_stop":
|
| 6635 |
},
|
| 6636 |
"attributes": {}
|
| 6637 |
}
|
| 6638 |
},
|
| 6639 |
-
"total_flos": 9.
|
| 6640 |
"train_batch_size": 4,
|
| 6641 |
"trial_name": null,
|
| 6642 |
"trial_params": null
|
|
|
|
| 2 |
"best_global_step": 16000,
|
| 3 |
"best_metric": 0.5378558039665222,
|
| 4 |
"best_model_checkpoint": "/root/leap-finetune/outputs/sft/lfm2_350m_marathi_optimized_12ep/checkpoint-16000",
|
| 5 |
+
"epoch": 12.0,
|
| 6 |
"eval_steps": 250,
|
| 7 |
+
"global_step": 16512,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 6631 |
"should_evaluate": false,
|
| 6632 |
"should_log": false,
|
| 6633 |
"should_save": true,
|
| 6634 |
+
"should_training_stop": true
|
| 6635 |
},
|
| 6636 |
"attributes": {}
|
| 6637 |
}
|
| 6638 |
},
|
| 6639 |
+
"total_flos": 9.171257878139372e+17,
|
| 6640 |
"train_batch_size": 4,
|
| 6641 |
"trial_name": null,
|
| 6642 |
"trial_params": null
|