Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes. See raw diff
- .gitattributes +3 -0
- 9000_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__9000_hf/samples_hellaswag_2026-01-07T03-49-27.696211.jsonl +3 -0
- 9000_hf/harness_eval_8shot/__home__aiops__zhuty__nanotron__checkpoints__9000_hf/samples_gsm8k_2026-01-07T07-00-26.302516.jsonl +3 -0
- 9000_hf/model.safetensors +3 -0
- 9000_hf/tokenizer.json +3 -0
- llama32-1b-nt/model/model/decoder/0/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/0/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/0/pp_block/input_layernorm/model_weight.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/0/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/0/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/0/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/1/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/1/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/1/pp_block/input_layernorm/model_weight.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/1/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/1/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/1/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/10/pp_block/input_layernorm/model_weight.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/10/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/12/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/12/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/12/pp_block/input_layernorm/model_weight.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/12/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/12/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/12/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/13/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/13/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/13/pp_block/input_layernorm/model_weight.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/13/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/13/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/13/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/15/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/15/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/15/pp_block/input_layernorm/model_weight.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/15/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/15/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/15/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/4/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/4/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/4/pp_block/input_layernorm/model_weight.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/4/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/4/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/4/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/5/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/5/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/5/pp_block/input_layernorm/model_weight.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/5/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/5/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/5/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
- llama32-1b-nt/model/model/decoder/7/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
.gitattributes
CHANGED
|
@@ -39,3 +39,6 @@ llama32-1b-hf/harness_eval_8shot/__home__aiops__zhuty__nanotron__checkpoints__ll
|
|
| 39 |
6000_hf/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 40 |
6000_hf/harness_eval_8shot/__home__aiops__zhuty__nanotron__checkpoints__6000_hf/samples_gsm8k_2026-01-07T06-57-48.977156.jsonl filter=lfs diff=lfs merge=lfs -text
|
| 41 |
6000_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__6000_hf/samples_hellaswag_2026-01-07T03-53-49.653886.jsonl filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
| 39 |
6000_hf/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 40 |
6000_hf/harness_eval_8shot/__home__aiops__zhuty__nanotron__checkpoints__6000_hf/samples_gsm8k_2026-01-07T06-57-48.977156.jsonl filter=lfs diff=lfs merge=lfs -text
|
| 41 |
6000_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__6000_hf/samples_hellaswag_2026-01-07T03-53-49.653886.jsonl filter=lfs diff=lfs merge=lfs -text
|
| 42 |
+
9000_hf/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 43 |
+
9000_hf/harness_eval_8shot/__home__aiops__zhuty__nanotron__checkpoints__9000_hf/samples_gsm8k_2026-01-07T07-00-26.302516.jsonl filter=lfs diff=lfs merge=lfs -text
|
| 44 |
+
9000_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__9000_hf/samples_hellaswag_2026-01-07T03-49-27.696211.jsonl filter=lfs diff=lfs merge=lfs -text
|
9000_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__9000_hf/samples_hellaswag_2026-01-07T03-49-27.696211.jsonl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e6c951352270e452953eddca955b7b91d59c444b0472e1b823783718174047f7
|
| 3 |
+
size 42644980
|
9000_hf/harness_eval_8shot/__home__aiops__zhuty__nanotron__checkpoints__9000_hf/samples_gsm8k_2026-01-07T07-00-26.302516.jsonl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1e041e7a4dbb3adb96de2e754281310caaed9c4bfc59a9cfd708b88a5b9cef63
|
| 3 |
+
size 16798564
|
9000_hf/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8c90c55635cf0ceea9a88f2d3fed29e503b1f5641431a3b2a35a2c1b5dd3b71c
|
| 3 |
+
size 2471645608
|
9000_hf/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
llama32-1b-nt/model/model/decoder/0/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fcf8ec32f73c8df7edb3dd7727a9729d834dc2af9f5931b5d1b906df44c33b8d
|
| 3 |
+
size 8388848
|
llama32-1b-nt/model/model/decoder/0/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:01be991ceab7faf9d1a9ef49197c66cb138c5454b9ee1dcd68122fc2090fa2b3
|
| 3 |
+
size 12583280
|
llama32-1b-nt/model/model/decoder/0/pp_block/input_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7f7cea49eb581e94c8639be7b7241025a34e45665fc65ecd156eca82045a5381
|
| 3 |
+
size 4192
|
llama32-1b-nt/model/model/decoder/0/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7f8b68f6b1bbb164111f7cfe2dab91c22c523b3a085ea1130a3151357cb10529
|
| 3 |
+
size 33554672
|
llama32-1b-nt/model/model/decoder/0/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:01317c2dd8d23b2ae73991529701efa622052f5b799b451f8c0edbf5ee87f9dd
|
| 3 |
+
size 67109176
|
llama32-1b-nt/model/model/decoder/0/pp_block/post_attention_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:64db8ea88659d892d2fbe951e02eed87a9ab70629a5d60f0cceba4084d139e9c
|
| 3 |
+
size 4192
|
llama32-1b-nt/model/model/decoder/1/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8f95ae205171c65231fdc326d0883bec0ed4650f1371f79e9a87200e7423cc2b
|
| 3 |
+
size 8388848
|
llama32-1b-nt/model/model/decoder/1/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aebcf09a0af1b43d7c2d085bc097c366b9f74575b6cf56d91c2d717e6787a25d
|
| 3 |
+
size 12583280
|
llama32-1b-nt/model/model/decoder/1/pp_block/input_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f2156f2df1e51d68b84ebaf5b0f12a853201adc0026642fb9b7bbef5c06a4ef7
|
| 3 |
+
size 4192
|
llama32-1b-nt/model/model/decoder/1/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d0d5253ae1cd73f3308344ae2dccaadba5f59efca3deeba29073a29900d00987
|
| 3 |
+
size 33554672
|
llama32-1b-nt/model/model/decoder/1/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:41fb5a628feaf46caae796b955e9b508d2594a5096cf14a1f8fd4e5a4b86e3e6
|
| 3 |
+
size 67109176
|
llama32-1b-nt/model/model/decoder/1/pp_block/post_attention_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:71eb41eefe8ea9f1524dc9a8170e7d50f28deac6c5d216cf74192d125e497567
|
| 3 |
+
size 4192
|
llama32-1b-nt/model/model/decoder/10/pp_block/input_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:104343d2bf3568e998ffb42ca2146c24a82570944233c25d26c28b032ba27dae
|
| 3 |
+
size 4192
|
llama32-1b-nt/model/model/decoder/10/pp_block/post_attention_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3b54bac7e7578f5ee0b3a8ba9ab9684016f538f65d2426eab2c31c0e9210abf6
|
| 3 |
+
size 4192
|
llama32-1b-nt/model/model/decoder/12/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:29901b09c966155daa1cc1b2f0b6d4d9b51fbc8b3d96197688195ed0d7ffe27d
|
| 3 |
+
size 8388848
|
llama32-1b-nt/model/model/decoder/12/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f1cf2c5ebbaf548e77e330958557e4b8df8bad02ccb59042388d48ce44302536
|
| 3 |
+
size 12583280
|
llama32-1b-nt/model/model/decoder/12/pp_block/input_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9468c20253d6e657931981408c6a387e3c4ffef4dd80e94b179d177c34797951
|
| 3 |
+
size 4192
|
llama32-1b-nt/model/model/decoder/12/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3808ce3e24b90cedaf8e4c22991fd2070966e9303dae7fc57a1b96dfa7861833
|
| 3 |
+
size 33554672
|
llama32-1b-nt/model/model/decoder/12/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7befa5837d94145ca09d93f89430d65a827a2f409e98c523818bc3eff704c179
|
| 3 |
+
size 67109176
|
llama32-1b-nt/model/model/decoder/12/pp_block/post_attention_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3aa0a0483c4f9e9663792e40827ed48904751d24c9bd4a8e578d2cd11c42328b
|
| 3 |
+
size 4192
|
llama32-1b-nt/model/model/decoder/13/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5697969101bd979a80b7fb440385af088ae58fe5410c972ca25a1fb2e294d580
|
| 3 |
+
size 8388848
|
llama32-1b-nt/model/model/decoder/13/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:37c4a7a29176a102a8f9cf411a0bc042b5f2f97a1d1c236e624a9d06ef3bdd6c
|
| 3 |
+
size 12583280
|
llama32-1b-nt/model/model/decoder/13/pp_block/input_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1587cc94039a6f2bcbf3a618b690d20e59ea52d4819fc49f900382b7f0e8b858
|
| 3 |
+
size 4192
|
llama32-1b-nt/model/model/decoder/13/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fbf8050b19aaf0c1501454fc16c44dd8c26f93d9859858915322f24ec8d88161
|
| 3 |
+
size 33554672
|
llama32-1b-nt/model/model/decoder/13/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8ae1c6fdaba6f5847700be2e43055312c6ff97d54a3d599850b32cd275be8a5b
|
| 3 |
+
size 67109176
|
llama32-1b-nt/model/model/decoder/13/pp_block/post_attention_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4d96a043ae27e520880693b9ae3c3fd3fe465b49b470e7565ea0176dcb5f9b03
|
| 3 |
+
size 4192
|
llama32-1b-nt/model/model/decoder/15/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b6fde1b2d907fac5bc732d6f456ce1fde86df348350a1801e9572685d6d525dd
|
| 3 |
+
size 8388848
|
llama32-1b-nt/model/model/decoder/15/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bf9902c1a5644e99e22a91171c2cbce04c2118d1ff556472b2dee704656b691f
|
| 3 |
+
size 12583280
|
llama32-1b-nt/model/model/decoder/15/pp_block/input_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:833077da913d352cf4f940f2ae912b3e966605aea5894397118c6b7798f83718
|
| 3 |
+
size 4192
|
llama32-1b-nt/model/model/decoder/15/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c26dcf52688014510235340347063d16642aa2eac1d837d3315388400137791d
|
| 3 |
+
size 33554672
|
llama32-1b-nt/model/model/decoder/15/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1e774bc57792a81b328d939ceae2179458351cc36c1097bc2483786d5bc2d1c1
|
| 3 |
+
size 67109176
|
llama32-1b-nt/model/model/decoder/15/pp_block/post_attention_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dc313cf53ddad1eb093038d84c8c69667fb20816bb13088641e8e16bcd2f098b
|
| 3 |
+
size 4192
|
llama32-1b-nt/model/model/decoder/4/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c122f0da7a1a04389dd9ee28927a481e3078a7d1d78e196e4a7ebe3a4ce3dea
|
| 3 |
+
size 8388848
|
llama32-1b-nt/model/model/decoder/4/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:965547dec7eba0f052e8b90170d43a813a4e60d2e9c9a585c7b79e00390112d5
|
| 3 |
+
size 12583280
|
llama32-1b-nt/model/model/decoder/4/pp_block/input_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:406cdadd5a55d620a9497e290330c744c8958e5c864f655cdff9574413c372ad
|
| 3 |
+
size 4192
|
llama32-1b-nt/model/model/decoder/4/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:23b845a5002815f30a73162044270f8d4160c10f795ee4125f0a2e2bd918efac
|
| 3 |
+
size 33554672
|
llama32-1b-nt/model/model/decoder/4/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dea4deb2b747d546ca4311ca545d482b5f77a1ce7f42e594619389116b9c63f8
|
| 3 |
+
size 67109176
|
llama32-1b-nt/model/model/decoder/4/pp_block/post_attention_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6ff191398cbc507b7e1000dd81d9190fe794ed76296334212414257ed51b6efc
|
| 3 |
+
size 4192
|
llama32-1b-nt/model/model/decoder/5/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:22750304d1b25c6bc07110cdd4e908b2b89315743715f6b0a5cba901fa452c81
|
| 3 |
+
size 8388848
|
llama32-1b-nt/model/model/decoder/5/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:809054c9260ad3a88ef4697fbc36ece2019802c8d1733c93f4f1aef4c699c6c2
|
| 3 |
+
size 12583280
|
llama32-1b-nt/model/model/decoder/5/pp_block/input_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9de58f7e0109dd72c4f3d84ff464b246607d8266965bdb883afc5f3211ed62fc
|
| 3 |
+
size 4192
|
llama32-1b-nt/model/model/decoder/5/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e4a01bfdadfd9b0dc0e5abb2f4fe0b992ea163adcdb32d79defb746f61e88ce5
|
| 3 |
+
size 33554672
|
llama32-1b-nt/model/model/decoder/5/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a030cc63b1ca097d13c5b2eb7392eb16f9b474db4e03facd73e83c00e83226d5
|
| 3 |
+
size 67109176
|
llama32-1b-nt/model/model/decoder/5/pp_block/post_attention_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d4aa2cccdb5a23e5cf1f6bf1e346ee770308c6760971dcbc70b74ce8643228fe
|
| 3 |
+
size 4192
|
llama32-1b-nt/model/model/decoder/7/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0ba6d5c261fd0078a04da858ad1ecbf74a06b76a6caa0fbab109b354383237ee
|
| 3 |
+
size 8388848
|