Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +19 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/adapter_model.safetensors +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/global_step1000/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/global_step1000/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/global_step1000/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/global_step1000/mp_rank_00_model_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/rng_state_0.pth +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/rng_state_1.pth +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/rng_state_2.pth +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/scheduler.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/tokenizer.json +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/training_args.bin +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/adapter_model.safetensors +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/global_step1500/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/global_step1500/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/global_step1500/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/global_step1500/mp_rank_00_model_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/rng_state_0.pth +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/rng_state_1.pth +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/rng_state_2.pth +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/scheduler.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/tokenizer.json +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/training_args.bin +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/adapter_model.safetensors +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/global_step1668/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/global_step1668/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/global_step1668/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/global_step1668/mp_rank_00_model_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/rng_state_0.pth +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/rng_state_1.pth +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/rng_state_2.pth +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/scheduler.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/tokenizer.json +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/training_args.bin +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/adapter_model.safetensors +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/global_step500/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/global_step500/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/global_step500/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/global_step500/mp_rank_00_model_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/rng_state_0.pth +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/rng_state_1.pth +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/rng_state_2.pth +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/scheduler.pt +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/tokenizer.json +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/training_args.bin +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/final_model/adapter_model.safetensors +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/final_model/tokenizer.json +3 -0
- D1E/llama-3.2-3B_r64_lr1e-5/final_model/training_args.bin +3 -0
- D1E/llama-3.2-3B_r64_lr9e-5/checkpoint-1000/global_step1000/mp_rank_00_model_states.pt +3 -0
- D1E/llama-3.2-3B_r64_lr9e-5/checkpoint-1500/global_step1500/mp_rank_00_model_states.pt +3 -0
.gitattributes
CHANGED
|
@@ -102,3 +102,22 @@ D1E/qwen2.5-coder-3B_r32_lr1e-4/checkpoint-1000/tokenizer.json filter=lfs diff=l
|
|
| 102 |
D1E/qwen2.5-coder-3B_r32_lr1e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 103 |
D1E/qwen2.5-coder-3B_r32_lr3e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 104 |
D1E/qwen2.5-coder-3B_r32_lr3e-4/merged_model_temp/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 102 |
D1E/qwen2.5-coder-3B_r32_lr1e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 103 |
D1E/qwen2.5-coder-3B_r32_lr3e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 104 |
D1E/qwen2.5-coder-3B_r32_lr3e-4/merged_model_temp/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 105 |
+
D1E/qwen2.5-coder-3B_r32_lr3e-4/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 106 |
+
D1E/qwen2.5-coder-3B_r32_lr3e-4/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 107 |
+
D1E/qwen2.5-coder-3B_r16_lr4e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 108 |
+
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 109 |
+
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 110 |
+
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 111 |
+
D1E/llama-3.2-3B_r64_lr1e-5/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 112 |
+
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 113 |
+
D1E/qwen2.5-coder-3B_r16_lr1e-4/checkpoint-2000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 114 |
+
D1E/qwen2.5-coder-3B_r16_lr1e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 115 |
+
D1E/qwen2.5-coder-3B_r16_lr1e-4/merged_model_temp/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 116 |
+
D1E/qwen2.5-coder-3B_r16_lr2e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 117 |
+
D1E/qwen2.5-coder-3B_r16_lr1e-4/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 118 |
+
D1E/qwen2.5-coder-3B_r16_lr1e-4/checkpoint-1500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 119 |
+
D1E/qwen2.5-coder-3B_r16_lr1e-4/checkpoint-2500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 120 |
+
D1E/qwen2.5-coder-3B_r16_lr1e-4/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 121 |
+
D1E/qwen2.5-coder-3B_r16_lr2e-4/checkpoint-1500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 122 |
+
D1E/qwen2.5-coder-3B_r16_lr2e-4/checkpoint-2500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 123 |
+
D1E/qwen2.5-coder-3B_r16_lr2e-4/merged_model_temp/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:447897fefa3da66a8027493e658107361ca956644a8725e991694b46ca9cadc6
|
| 3 |
+
size 194563792
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/global_step1000/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:56fcb521ab6189329cf577f95e8992ad536347e388e9fd8c27629ad476931a21
|
| 3 |
+
size 389040325
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/global_step1000/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6c55181f1853de2bc7e47695cf95c43d2ab9c2b2ffe04b347a361f35a4c0467c
|
| 3 |
+
size 389040133
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/global_step1000/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ec52c28f3546305f04fd0a32f66c93cd7d7e9d4d93c6db5d7f1f4cf03bb8b4ec
|
| 3 |
+
size 389039877
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/global_step1000/mp_rank_00_model_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aaf7e96ac21c86f3ff68dbceec60271c51520058ba25bada1e841f4110ee0b82
|
| 3 |
+
size 982818405
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e8d95610b72632821b9c388db5b4dea3f6221acf94df7da76015ac3c6c710e4a
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/rng_state_1.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9e501c8af29227dd1bf508b904b7133d417dbdf53a06a3967ac74c4537f74874
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/rng_state_2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d6fe2a0a4bf71c02988c6d7847fbcea3d99fb82eaca5fc7053cca2b1f734f548
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5e5004f7549c9144ff3c81714359dde775b7c1a6bbd0b00e60b1c0f2107b204f
|
| 3 |
+
size 1465
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1000/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5a63a1a98d482e9dd0d27c6848233fc3df991a5cb717ecfacae99d62cbbf84bd
|
| 3 |
+
size 7633
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e9853d164d63501baa404a534fec3f6a251c9dd3709f744a800d0c08576a427f
|
| 3 |
+
size 194563792
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/global_step1500/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3ca082e957db1e693a103317a028b7f96a44d64ef1cc061734324143730f5b29
|
| 3 |
+
size 389040325
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/global_step1500/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:14b3cd6ad5f50a8aac043ceea1032bf0d800873bc57c8accbba8007f43fbb998
|
| 3 |
+
size 389040133
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/global_step1500/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7ddd625edbbfd0378f4e9a50a10dc4f793c30bc259305cf834606d2101a36e73
|
| 3 |
+
size 389039877
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/global_step1500/mp_rank_00_model_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:429225d0bf23b002cfe061846ad2b76f414bfc8ec0eb346be2105c73b9493127
|
| 3 |
+
size 982818405
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f690fafe0b5b40523cb3ca62f93a31f9944c871320656ffb3cfc17f196363c7c
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/rng_state_1.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0e6050f6093153f4ac9a02c3b08d1dd143d7ec0393d47644aa9c3e56d27b5a0c
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/rng_state_2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:566be891c5b9318d733297c7ed9a6c0caa70c52f261fbdc0db9b56626409ef5b
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eef4a14a9b904a5f501f9c9f9feeffc5ca771b20cbe971e0a1b20c4aa30beae5
|
| 3 |
+
size 1465
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1500/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5a63a1a98d482e9dd0d27c6848233fc3df991a5cb717ecfacae99d62cbbf84bd
|
| 3 |
+
size 7633
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ed7561f4a9880987281b8b5b11459fdb394828ece7d499683c025ba548f1e5dc
|
| 3 |
+
size 194563792
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/global_step1668/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:005a20f22a129c4add09eedd3dd7d305f3db5648f8119484fe50e533de692f19
|
| 3 |
+
size 389040325
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/global_step1668/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3c9d858545d2d15e3051e6fedb41d1ff83480e8204d7943e9e6838a1a3fa36a8
|
| 3 |
+
size 389040133
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/global_step1668/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:24d826d8b249f8a5f642c17fe4e7df8b11f94d399126fc77a5f105570078cb97
|
| 3 |
+
size 389039877
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/global_step1668/mp_rank_00_model_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f33ff7523d97753f093bc94915306153e7de1140733d929eef32f768fc3dbb61
|
| 3 |
+
size 982818405
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8e8fc61d01f99291e7a3bf6e69f1976b9c5017a228efb6d808fa60ab525b0473
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/rng_state_1.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:270efc9ff172c4f4b2259308e821348fbb8d818827cd601c42bce37b28da9153
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/rng_state_2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bb26a611b69fa284fad930563130b5c863fe1cf01106d28e644d2b6f8b33d282
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4fd2aae3747027b11f382efbc0b3075ff4f256880186cd52fe21525c371f5ab7
|
| 3 |
+
size 1465
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-1668/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5a63a1a98d482e9dd0d27c6848233fc3df991a5cb717ecfacae99d62cbbf84bd
|
| 3 |
+
size 7633
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6be108275f4305ff7fdc1b5dfc4cbe63278cd5634ae801f4b9af866c2055d32a
|
| 3 |
+
size 194563792
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/global_step500/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0be3f46d633aa60a7aa7b44c711875ca0fcd3e17bdd97712f512adb1d59998cb
|
| 3 |
+
size 389040325
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/global_step500/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1d605aa502d8871d99bc244cbf375408f84633927ee0173c5847454917dce416
|
| 3 |
+
size 389040133
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/global_step500/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8f5b6bb68f86d7f1cb19157ac283e4e3bdd2b617b4bb7c75e3bf78cc379168c2
|
| 3 |
+
size 389039877
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/global_step500/mp_rank_00_model_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7dbe8c43bbfc12fb0f93cf374056952166acbbb5c6ea0e96be3910d794b7d77e
|
| 3 |
+
size 982818405
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:34f3e18f53d5d32e9dec83fce82cbb88e3e7362888ef7ee3d41e5181b436dc21
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/rng_state_1.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b5d67532c44c34c14f7789a94ee948f31c8c42bc31739567104054a58b70c2f4
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/rng_state_2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8ff1aac8eab43c69bba7f73454464d17ac8570aab2c39618b55dcf1e95cfffd8
|
| 3 |
+
size 15173
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c08fc523a040688ef8912ec44f3c9dad8a9f3a8720c3c4dfab823932dbb050a2
|
| 3 |
+
size 1465
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
D1E/llama-3.2-3B_r64_lr1e-5/checkpoint-500/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5a63a1a98d482e9dd0d27c6848233fc3df991a5cb717ecfacae99d62cbbf84bd
|
| 3 |
+
size 7633
|
D1E/llama-3.2-3B_r64_lr1e-5/final_model/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ed7561f4a9880987281b8b5b11459fdb394828ece7d499683c025ba548f1e5dc
|
| 3 |
+
size 194563792
|
D1E/llama-3.2-3B_r64_lr1e-5/final_model/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
D1E/llama-3.2-3B_r64_lr1e-5/final_model/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5a63a1a98d482e9dd0d27c6848233fc3df991a5cb717ecfacae99d62cbbf84bd
|
| 3 |
+
size 7633
|
D1E/llama-3.2-3B_r64_lr9e-5/checkpoint-1000/global_step1000/mp_rank_00_model_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ccfbec075d759c1f33d33a59f4f9f316962116260ad3439980980a4c5fdc8610
|
| 3 |
+
size 982818405
|
D1E/llama-3.2-3B_r64_lr9e-5/checkpoint-1500/global_step1500/mp_rank_00_model_states.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2d7d829e9a2cb2e9e0028b713f882f609ac8ab9955d8b33246f1ac5e77c85b1c
|
| 3 |
+
size 982818405
|