aghatage commited on
Commit
d80bf5e
·
verified ·
1 Parent(s): ceb1deb

Training in progress, step 16512, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:84490a0ba5e68272e718e1b587f957c02192d8adb3af51f4ee9f3e16c57d8791
3
  size 12017472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c52a6dd7e5bc676bd35e2ff08757c004c9a77792d247e5966821ce4cda29d1f
3
  size 12017472
last-checkpoint/global_step16512/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fecc1fd44558636bf7291a837176c0c5d7117ea5d1c6ded1655d2dd3ac118e40
3
+ size 71982309
last-checkpoint/global_step16512/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04b5d77557f0687686c909d8ecd0930cb52e3c8d3394f7825febe171c4601f55
3
+ size 146356645
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step16500
 
1
+ global_step16512
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:25f34d42adfb703043e77cccecb301eb61eb2fb980871ddcbacc2771de139a9c
3
  size 14709
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9be458c8cf8debad5d6e2a3bd43cb667b3d77295003bd12c52999f7972324e51
3
  size 14709
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 16000,
3
  "best_metric": 0.5378558039665222,
4
  "best_model_checkpoint": "/root/leap-finetune/outputs/sft/lfm2_350m_marathi_optimized_12ep/checkpoint-16000",
5
- "epoch": 11.991819669151063,
6
  "eval_steps": 250,
7
- "global_step": 16500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -6631,12 +6631,12 @@
6631
  "should_evaluate": false,
6632
  "should_log": false,
6633
  "should_save": true,
6634
- "should_training_stop": false
6635
  },
6636
  "attributes": {}
6637
  }
6638
  },
6639
- "total_flos": 9.164882254873231e+17,
6640
  "train_batch_size": 4,
6641
  "trial_name": null,
6642
  "trial_params": null
 
2
  "best_global_step": 16000,
3
  "best_metric": 0.5378558039665222,
4
  "best_model_checkpoint": "/root/leap-finetune/outputs/sft/lfm2_350m_marathi_optimized_12ep/checkpoint-16000",
5
+ "epoch": 12.0,
6
  "eval_steps": 250,
7
+ "global_step": 16512,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
6631
  "should_evaluate": false,
6632
  "should_log": false,
6633
  "should_save": true,
6634
+ "should_training_stop": true
6635
  },
6636
  "attributes": {}
6637
  }
6638
  },
6639
+ "total_flos": 9.171257878139372e+17,
6640
  "train_batch_size": 4,
6641
  "trial_name": null,
6642
  "trial_params": null