gg676 commited on
Commit
ee8d6f2
·
verified ·
1 Parent(s): 4d46971

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +25 -0
  2. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/adapter_model.safetensors +3 -0
  3. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/global_step1000/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
  4. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/global_step1000/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
  5. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/global_step1000/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
  6. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/global_step1000/mp_rank_00_model_states.pt +3 -0
  7. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/rng_state_0.pth +3 -0
  8. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/rng_state_1.pth +3 -0
  9. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/rng_state_2.pth +3 -0
  10. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/scheduler.pt +3 -0
  11. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/tokenizer.json +3 -0
  12. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/training_args.bin +3 -0
  13. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/adapter_model.safetensors +3 -0
  14. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/global_step1500/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
  15. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/global_step1500/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
  16. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/global_step1500/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
  17. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/global_step1500/mp_rank_00_model_states.pt +3 -0
  18. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/rng_state_0.pth +3 -0
  19. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/rng_state_1.pth +3 -0
  20. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/rng_state_2.pth +3 -0
  21. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/scheduler.pt +3 -0
  22. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/tokenizer.json +3 -0
  23. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/training_args.bin +3 -0
  24. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/adapter_model.safetensors +3 -0
  25. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/global_step1668/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
  26. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/global_step1668/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
  27. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/global_step1668/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
  28. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/global_step1668/mp_rank_00_model_states.pt +3 -0
  29. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/rng_state_0.pth +3 -0
  30. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/rng_state_1.pth +3 -0
  31. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/rng_state_2.pth +3 -0
  32. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/scheduler.pt +3 -0
  33. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/tokenizer.json +3 -0
  34. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/training_args.bin +3 -0
  35. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/adapter_model.safetensors +3 -0
  36. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/global_step500/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
  37. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/global_step500/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
  38. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/global_step500/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
  39. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/global_step500/mp_rank_00_model_states.pt +3 -0
  40. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/rng_state_0.pth +3 -0
  41. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/rng_state_1.pth +3 -0
  42. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/rng_state_2.pth +3 -0
  43. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/scheduler.pt +3 -0
  44. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/tokenizer.json +3 -0
  45. D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/training_args.bin +3 -0
  46. D1E/llama-3.2-3B_r64_lr2e-4/final_model/adapter_model.safetensors +3 -0
  47. D1E/llama-3.2-3B_r64_lr2e-4/final_model/tokenizer.json +3 -0
  48. D1E/llama-3.2-3B_r64_lr2e-4/final_model/training_args.bin +3 -0
  49. D1E/llama-3.2-3B_r64_lr3e-4/checkpoint-1000/adapter_model.safetensors +3 -0
  50. D1E/llama-3.2-3B_r64_lr3e-4/checkpoint-1000/global_step1000/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
.gitattributes CHANGED
@@ -77,3 +77,28 @@ D1E/llama-3.2-3B_r64_lr9e-5/checkpoint-1668/tokenizer.json filter=lfs diff=lfs m
77
  D1E/qwen2.5-coder-3B_r64_lr1e-4/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
78
  D1E/llama-3.2-3B_r64_lr9e-5/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
79
  D1E/llama-3.2-3B_r64_lr9e-5/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
77
  D1E/qwen2.5-coder-3B_r64_lr1e-4/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
78
  D1E/llama-3.2-3B_r64_lr9e-5/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
79
  D1E/llama-3.2-3B_r64_lr9e-5/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
80
+ D1E/llama-3.2-3B_r64_lr9e-5/checkpoint-1500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
81
+ D1E/qwen2.5-coder-3B_r64_lr2e-4/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
82
+ D1E/qwen2.5-coder-3B_r64_lr2e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
83
+ D1E/qwen2.5-coder-3B_r64_lr3e-4/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
84
+ D1E/qwen2.5-coder-3B_r64_lr2e-4/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
85
+ D1E/qwen2.5-coder-3B_r32_lr3e-4/checkpoint-1250/tokenizer.json filter=lfs diff=lfs merge=lfs -text
86
+ D1E/qwen2.5-coder-3B_r32_lr9e-5/checkpoint-1250/tokenizer.json filter=lfs diff=lfs merge=lfs -text
87
+ D1E/qwen2.5-coder-3B_r32_lr9e-5/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
88
+ D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/tokenizer.json filter=lfs diff=lfs merge=lfs -text
89
+ D1E/qwen2.5-coder-3B_r32_lr9e-5/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
90
+ D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
91
+ D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
92
+ D1E/llama-3.2-3B_r64_lr2e-4/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
93
+ D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
94
+ D1E/llama-3.2-3B_r64_lr3e-4/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
95
+ D1E/llama-3.2-3B_r64_lr3e-4/checkpoint-1668/tokenizer.json filter=lfs diff=lfs merge=lfs -text
96
+ D1E/llama-3.2-3B_r64_lr3e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
97
+ D1E/qwen2.5-coder-3B_r32_lr1e-4/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
98
+ D1E/llama-3.2-3B_r64_lr3e-4/final_model/tokenizer.json filter=lfs diff=lfs merge=lfs -text
99
+ D1E/llama-3.2-3B_r64_lr3e-4/checkpoint-1500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
100
+ D1E/qwen2.5-coder-3B_r32_lr1e-4/checkpoint-1250/tokenizer.json filter=lfs diff=lfs merge=lfs -text
101
+ D1E/qwen2.5-coder-3B_r32_lr1e-4/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
102
+ D1E/qwen2.5-coder-3B_r32_lr1e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
103
+ D1E/qwen2.5-coder-3B_r32_lr3e-4/checkpoint-500/tokenizer.json filter=lfs diff=lfs merge=lfs -text
104
+ D1E/qwen2.5-coder-3B_r32_lr3e-4/merged_model_temp/tokenizer.json filter=lfs diff=lfs merge=lfs -text
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:883a408f6ffe460fc59efeb4e21df4f6b9a44aba95ff979bdd09b13c254aefc1
3
+ size 194563792
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/global_step1000/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a99be5fc17f28fb7289cd8b11ae7a8532657ef5408cc18650ccd86b5c665f035
3
+ size 389040325
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/global_step1000/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aea4c54143995dffe57f30584759be3c4b2fa38d5ed75495d071fce2207fddf6
3
+ size 389040133
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/global_step1000/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96d29a37237e892681dd00f70691094804fd1e052a93d76706cdf425108b8da4
3
+ size 389039877
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/global_step1000/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0da42d0b68c01114f170cfb6a87ade199b6c2a2ae809e2edc1de7b7fc048639
3
+ size 982818405
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8d95610b72632821b9c388db5b4dea3f6221acf94df7da76015ac3c6c710e4a
3
+ size 15173
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e501c8af29227dd1bf508b904b7133d417dbdf53a06a3967ac74c4537f74874
3
+ size 15173
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6fe2a0a4bf71c02988c6d7847fbcea3d99fb82eaca5fc7053cca2b1f734f548
3
+ size 15173
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:161e2a445fe9185d2f97aea60249e41b7e0178d6dcccb3d1dd7a08dc0ed19875
3
+ size 1465
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
3
+ size 17209920
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eeb3c044e815ef0b58ace80266693344f5fe7a3464ba70c210a600e16877e05b
3
+ size 7633
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20ab8478bb0825f79a59f22ab91c74c69f71b099189a530ff678846e38c5ac5b
3
+ size 194563792
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/global_step1500/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b67c009c0106a66e2d40ce63baa91af074302bdba246d2810bb216a69cb4c197
3
+ size 389040325
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/global_step1500/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d146c45248bddf50ee084b6135582d5602d30324af13e39e03a2d50a522146e0
3
+ size 389040133
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/global_step1500/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e92a19244c42bf86eb72917f76312fd2cf6ea74a0192de7b48066c777e118d7
3
+ size 389039877
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/global_step1500/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2c984a773045dfec15cae16520bf9e15744bba1c78f120af2c1baa6b254b855
3
+ size 982818405
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f690fafe0b5b40523cb3ca62f93a31f9944c871320656ffb3cfc17f196363c7c
3
+ size 15173
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e6050f6093153f4ac9a02c3b08d1dd143d7ec0393d47644aa9c3e56d27b5a0c
3
+ size 15173
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:566be891c5b9318d733297c7ed9a6c0caa70c52f261fbdc0db9b56626409ef5b
3
+ size 15173
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9adaf558cd743e47e5726a60b6788c2921a15ddab340576791ccf465ca6c7621
3
+ size 1465
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
3
+ size 17209920
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1500/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eeb3c044e815ef0b58ace80266693344f5fe7a3464ba70c210a600e16877e05b
3
+ size 7633
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:698924e3e9b82a12a271f118e0f74d7a9ff5a1e158bf9a3b767f83c77000faa7
3
+ size 194563792
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/global_step1668/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5cc7619d86ab2f2fb1255c1a9c7d435aab4cd65df7b11ee542f63dbff5d82ed
3
+ size 389040325
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/global_step1668/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7d956395490f97457612f55a55a84453129d8a3ffab98988700da46b2c1dc48
3
+ size 389040133
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/global_step1668/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6caa094ef18c0637d16313c77a7d20cd2e89a670b36dd21c7fdb89206eb5790
3
+ size 389039877
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/global_step1668/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0a63e1a8f32bc8d483ca10621d2bf71b90fa02c4965996898d8164cc524e5f6
3
+ size 982818405
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e8fc61d01f99291e7a3bf6e69f1976b9c5017a228efb6d808fa60ab525b0473
3
+ size 15173
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:270efc9ff172c4f4b2259308e821348fbb8d818827cd601c42bce37b28da9153
3
+ size 15173
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb26a611b69fa284fad930563130b5c863fe1cf01106d28e644d2b6f8b33d282
3
+ size 15173
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20a46fa9adfbc1b429fe60c4eab3e912770ed6df97c0270d2cea7ccdc8c0ef4c
3
+ size 1465
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
3
+ size 17209920
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-1668/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eeb3c044e815ef0b58ace80266693344f5fe7a3464ba70c210a600e16877e05b
3
+ size 7633
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af310780abc7a4910290a1cfdcabba66e6f2b85c01b597f582f7cadf712ff99a
3
+ size 194563792
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/global_step500/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9056d2cd154c9c48c76d079eafff2761a886e8e02c04174c88883dd99376c9e
3
+ size 389040325
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/global_step500/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6dac13b9e680cc976e67f83a3a4d6f153033ba4b692980b2c442adf05293ac46
3
+ size 389040133
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/global_step500/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f2a6ed81808ea32342ad3f95574775348ab0341480d8ef4f3d3214f01505923
3
+ size 389039877
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/global_step500/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10eac567ce3a197dfdfd4fbd08a682b5875d08b08bf5a3ae92346a38a10ca6c6
3
+ size 982818405
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34f3e18f53d5d32e9dec83fce82cbb88e3e7362888ef7ee3d41e5181b436dc21
3
+ size 15173
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5d67532c44c34c14f7789a94ee948f31c8c42bc31739567104054a58b70c2f4
3
+ size 15173
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ff1aac8eab43c69bba7f73454464d17ac8570aab2c39618b55dcf1e95cfffd8
3
+ size 15173
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1efce6bd9ad1d89abb8a98af925a5f0f676d43fd6c0e5a94a1e24b7a0d2c918a
3
+ size 1465
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
3
+ size 17209920
D1E/llama-3.2-3B_r64_lr2e-4/checkpoint-500/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eeb3c044e815ef0b58ace80266693344f5fe7a3464ba70c210a600e16877e05b
3
+ size 7633
D1E/llama-3.2-3B_r64_lr2e-4/final_model/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:698924e3e9b82a12a271f118e0f74d7a9ff5a1e158bf9a3b767f83c77000faa7
3
+ size 194563792
D1E/llama-3.2-3B_r64_lr2e-4/final_model/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
3
+ size 17209920
D1E/llama-3.2-3B_r64_lr2e-4/final_model/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eeb3c044e815ef0b58ace80266693344f5fe7a3464ba70c210a600e16877e05b
3
+ size 7633
D1E/llama-3.2-3B_r64_lr3e-4/checkpoint-1000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02950835a83d978d300bc1f6d2d5a90881891795acb2aa5ea11298fb17cbbb35
3
+ size 194563792
D1E/llama-3.2-3B_r64_lr3e-4/checkpoint-1000/global_step1000/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f679109c7443dfc28de891f81139cd8102b01a5bf100afe29c3ad4d9afffb1d
3
+ size 389040325