Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +25 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/adapter_model.safetensors +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/global_step1000/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/global_step1000/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/global_step1000/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/global_step1000/mp_rank_00_model_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/rng_state_0.pth +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/rng_state_1.pth +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/rng_state_2.pth +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/scheduler.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/tokenizer.json +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/training_args.bin +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/adapter_model.safetensors +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/global_step1500/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/global_step1500/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/global_step1500/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/global_step1500/mp_rank_00_model_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/rng_state_0.pth +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/rng_state_1.pth +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/rng_state_2.pth +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/scheduler.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/tokenizer.json +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/training_args.bin +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/adapter_model.safetensors +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/global_step1668/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/global_step1668/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/global_step1668/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/global_step1668/mp_rank_00_model_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/rng_state_0.pth +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/rng_state_1.pth +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/rng_state_2.pth +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/scheduler.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/tokenizer.json +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/training_args.bin +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/adapter_model.safetensors +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/global_step500/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/global_step500/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/global_step500/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/global_step500/mp_rank_00_model_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/rng_state_0.pth +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/rng_state_1.pth +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/rng_state_2.pth +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/scheduler.pt +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/tokenizer.json +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/training_args.bin +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/final_model/adapter_model.safetensors +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/final_model/tokenizer.json +3 -0
- D1E/llama-3.2-3B_r64_lr2e-4/final_model/training_args.bin +3 -0
- D1E/llama-3.2-3B_r64_lr3e-4/checkpoint-1000/adapter_model.safetensors +3 -0
- D1E/llama-3.2-3B_r64_lr3e-4/checkpoint-1000/global_step1000/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
.gitattributes
CHANGED
|
@@ -77,3 +77,28 @@ D1E/llama-3.2-3B_r64_lr9e-5/checkpoint-1668/tokenizer.json filter=lfs diff=lfs m
|
|
| 77 |
D1E/qwen2.5-coder-3B_r64_lr1e-4/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 78 |
D1E/llama-3.2-3B_r64_lr9e-5/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 79 |
D1E/llama-3.2-3B_r64_lr9e-5/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 77 |
D1E/qwen2.5-coder-3B_r64_lr1e-4/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 78 |
D1E/llama-3.2-3B_r64_lr9e-5/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 79 |
D1E/llama-3.2-3B_r64_lr9e-5/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 80 |
+
D1E/llama-3.2-3B_r64_lr9e-5/checkpoint-1500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 81 |
+
D1E/qwen2.5-coder-3B_r64_lr2e-4/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 82 |
+
D1E/qwen2.5-coder-3B_r64_lr2e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 83 |
+
D1E/qwen2.5-coder-3B_r64_lr3e-4/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 84 |
+
D1E/qwen2.5-coder-3B_r64_lr2e-4/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 85 |
+
D1E/qwen2.5-coder-3B_r32_lr3e-4/checkpoint-1250/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 86 |
+
D1E/qwen2.5-coder-3B_r32_lr9e-5/checkpoint-1250/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 87 |
+
D1E/qwen2.5-coder-3B_r32_lr9e-5/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 88 |
+
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 89 |
+
D1E/qwen2.5-coder-3B_r32_lr9e-5/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 90 |
+
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 91 |
+
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 92 |
+
D1E/llama-3.2-3B_r64_lr2e-4/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 93 |
+
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 94 |
+
D1E/llama-3.2-3B_r64_lr3e-4/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 95 |
+
D1E/llama-3.2-3B_r64_lr3e-4/checkpoint-1668/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 96 |
+
D1E/llama-3.2-3B_r64_lr3e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 97 |
+
D1E/qwen2.5-coder-3B_r32_lr1e-4/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 98 |
+
D1E/llama-3.2-3B_r64_lr3e-4/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 99 |
+
D1E/llama-3.2-3B_r64_lr3e-4/checkpoint-1500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 100 |
+
D1E/qwen2.5-coder-3B_r32_lr1e-4/checkpoint-1250/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 101 |
+
D1E/qwen2.5-coder-3B_r32_lr1e-4/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 102 |
+
D1E/qwen2.5-coder-3B_r32_lr1e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 103 |
+
D1E/qwen2.5-coder-3B_r32_lr3e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 104 |
+
D1E/qwen2.5-coder-3B_r32_lr3e-4/merged_model_temp/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:883a408f6ffe460fc59efeb4e21df4f6b9a44aba95ff979bdd09b13c254aefc1
|
| 3 |
+
size 194563792
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/global_step1000/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a99be5fc17f28fb7289cd8b11ae7a8532657ef5408cc18650ccd86b5c665f035
|
| 3 |
+
size 389040325
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/global_step1000/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aea4c54143995dffe57f30584759be3c4b2fa38d5ed75495d071fce2207fddf6
|
| 3 |
+
size 389040133
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/global_step1000/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:96d29a37237e892681dd00f70691094804fd1e052a93d76706cdf425108b8da4
|
| 3 |
+
size 389039877
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/global_step1000/mp_rank_00_model_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d0da42d0b68c01114f170cfb6a87ade199b6c2a2ae809e2edc1de7b7fc048639
|
| 3 |
+
size 982818405
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e8d95610b72632821b9c388db5b4dea3f6221acf94df7da76015ac3c6c710e4a
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/rng_state_1.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9e501c8af29227dd1bf508b904b7133d417dbdf53a06a3967ac74c4537f74874
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/rng_state_2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d6fe2a0a4bf71c02988c6d7847fbcea3d99fb82eaca5fc7053cca2b1f734f548
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:161e2a445fe9185d2f97aea60249e41b7e0178d6dcccb3d1dd7a08dc0ed19875
|
| 3 |
+
size 1465
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eeb3c044e815ef0b58ace80266693344f5fe7a3464ba70c210a600e16877e05b
|
| 3 |
+
size 7633
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:20ab8478bb0825f79a59f22ab91c74c69f71b099189a530ff678846e38c5ac5b
|
| 3 |
+
size 194563792
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/global_step1500/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b67c009c0106a66e2d40ce63baa91af074302bdba246d2810bb216a69cb4c197
|
| 3 |
+
size 389040325
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/global_step1500/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d146c45248bddf50ee084b6135582d5602d30324af13e39e03a2d50a522146e0
|
| 3 |
+
size 389040133
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/global_step1500/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4e92a19244c42bf86eb72917f76312fd2cf6ea74a0192de7b48066c777e118d7
|
| 3 |
+
size 389039877
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/global_step1500/mp_rank_00_model_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e2c984a773045dfec15cae16520bf9e15744bba1c78f120af2c1baa6b254b855
|
| 3 |
+
size 982818405
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f690fafe0b5b40523cb3ca62f93a31f9944c871320656ffb3cfc17f196363c7c
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/rng_state_1.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0e6050f6093153f4ac9a02c3b08d1dd143d7ec0393d47644aa9c3e56d27b5a0c
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/rng_state_2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:566be891c5b9318d733297c7ed9a6c0caa70c52f261fbdc0db9b56626409ef5b
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9adaf558cd743e47e5726a60b6788c2921a15ddab340576791ccf465ca6c7621
|
| 3 |
+
size 1465
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eeb3c044e815ef0b58ace80266693344f5fe7a3464ba70c210a600e16877e05b
|
| 3 |
+
size 7633
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:698924e3e9b82a12a271f118e0f74d7a9ff5a1e158bf9a3b767f83c77000faa7
|
| 3 |
+
size 194563792
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/global_step1668/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b5cc7619d86ab2f2fb1255c1a9c7d435aab4cd65df7b11ee542f63dbff5d82ed
|
| 3 |
+
size 389040325
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/global_step1668/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a7d956395490f97457612f55a55a84453129d8a3ffab98988700da46b2c1dc48
|
| 3 |
+
size 389040133
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/global_step1668/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d6caa094ef18c0637d16313c77a7d20cd2e89a670b36dd21c7fdb89206eb5790
|
| 3 |
+
size 389039877
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/global_step1668/mp_rank_00_model_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a0a63e1a8f32bc8d483ca10621d2bf71b90fa02c4965996898d8164cc524e5f6
|
| 3 |
+
size 982818405
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8e8fc61d01f99291e7a3bf6e69f1976b9c5017a228efb6d808fa60ab525b0473
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/rng_state_1.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:270efc9ff172c4f4b2259308e821348fbb8d818827cd601c42bce37b28da9153
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/rng_state_2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bb26a611b69fa284fad930563130b5c863fe1cf01106d28e644d2b6f8b33d282
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:20a46fa9adfbc1b429fe60c4eab3e912770ed6df97c0270d2cea7ccdc8c0ef4c
|
| 3 |
+
size 1465
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eeb3c044e815ef0b58ace80266693344f5fe7a3464ba70c210a600e16877e05b
|
| 3 |
+
size 7633
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:af310780abc7a4910290a1cfdcabba66e6f2b85c01b597f582f7cadf712ff99a
|
| 3 |
+
size 194563792
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/global_step500/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e9056d2cd154c9c48c76d079eafff2761a886e8e02c04174c88883dd99376c9e
|
| 3 |
+
size 389040325
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/global_step500/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6dac13b9e680cc976e67f83a3a4d6f153033ba4b692980b2c442adf05293ac46
|
| 3 |
+
size 389040133
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/global_step500/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9f2a6ed81808ea32342ad3f95574775348ab0341480d8ef4f3d3214f01505923
|
| 3 |
+
size 389039877
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/global_step500/mp_rank_00_model_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:10eac567ce3a197dfdfd4fbd08a682b5875d08b08bf5a3ae92346a38a10ca6c6
|
| 3 |
+
size 982818405
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:34f3e18f53d5d32e9dec83fce82cbb88e3e7362888ef7ee3d41e5181b436dc21
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/rng_state_1.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b5d67532c44c34c14f7789a94ee948f31c8c42bc31739567104054a58b70c2f4
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/rng_state_2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8ff1aac8eab43c69bba7f73454464d17ac8570aab2c39618b55dcf1e95cfffd8
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1efce6bd9ad1d89abb8a98af925a5f0f676d43fd6c0e5a94a1e24b7a0d2c918a
|
| 3 |
+
size 1465
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eeb3c044e815ef0b58ace80266693344f5fe7a3464ba70c210a600e16877e05b
|
| 3 |
+
size 7633
|
D1E/llama-3.2-3B_r64_lr2e-4/final_model/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:698924e3e9b82a12a271f118e0f74d7a9ff5a1e158bf9a3b767f83c77000faa7
|
| 3 |
+
size 194563792
|
D1E/llama-3.2-3B_r64_lr2e-4/final_model/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
D1E/llama-3.2-3B_r64_lr2e-4/final_model/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eeb3c044e815ef0b58ace80266693344f5fe7a3464ba70c210a600e16877e05b
|
| 3 |
+
size 7633
|
D1E/llama-3.2-3B_r64_lr3e-4/checkpoint-1000/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:02950835a83d978d300bc1f6d2d5a90881891795acb2aa5ea11298fb17cbbb35
|
| 3 |
+
size 194563792
|
D1E/llama-3.2-3B_r64_lr3e-4/checkpoint-1000/global_step1000/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8f679109c7443dfc28de891f81139cd8102b01a5bf100afe29c3ad4d9afffb1d
|
| 3 |
+
size 389040325
|