Add files using upload-large-folder tool
Browse files- llama32-3b-nt/model/model/decoder/0/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/0/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/0/pp_block/input_layernorm/model_weight.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/0/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/0/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/0/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/12/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/12/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/12/pp_block/input_layernorm/model_weight.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/12/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/12/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/12/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/13/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/16/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/16/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/16/pp_block/input_layernorm/model_weight.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/16/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/16/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/16/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/18/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/18/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/18/pp_block/input_layernorm/model_weight.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/18/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/18/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/18/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/5/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/5/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/5/pp_block/input_layernorm/model_weight.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/5/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/5/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
- llama32-3b-nt/model/model/decoder/5/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
llama32-3b-nt/model/model/decoder/0/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f0a08d86c740b2b922f9f542ebcb59d5566342e71b621d489d7df30b94c25c4c
|
| 3 |
+
size 18874608
|
llama32-3b-nt/model/model/decoder/0/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6d723e68653fae50092c13d203af7a6f44b48c2b55a4da4f95d9d2b864c934f5
|
| 3 |
+
size 31457648
|
llama32-3b-nt/model/model/decoder/0/pp_block/input_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e8417452d451195a37952805821bebc0ae3b9e324f20590560bce0353a48ae94
|
| 3 |
+
size 6240
|
llama32-3b-nt/model/model/decoder/0/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1a40a3115764f1384aebba170b2826a4d46e98671e1585898a6f392cc1784cf7
|
| 3 |
+
size 50331888
|
llama32-3b-nt/model/model/decoder/0/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1666afc981a7aca298970cafc62fd8650b11b680974727c0b150759bbe276bd6
|
| 3 |
+
size 100663608
|
llama32-3b-nt/model/model/decoder/0/pp_block/post_attention_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ea8d98c9c512ce732ac5d7fe17ecd945687b562e8b744d163f0ee16ceb5f592a
|
| 3 |
+
size 6240
|
llama32-3b-nt/model/model/decoder/12/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c5fc68bf7fef5b281431abfbbea1b034dccb974738556e0362bba333acebc765
|
| 3 |
+
size 18874608
|
llama32-3b-nt/model/model/decoder/12/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2af246555cc48a418c3bbe3f2d2d1e72bd669d026fac8d90d6d4bdb6a90207f1
|
| 3 |
+
size 31457648
|
llama32-3b-nt/model/model/decoder/12/pp_block/input_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:24e5d8ae35087f1512a38f7e95fc35371e61874b51d90b8a6167e75b22c3480e
|
| 3 |
+
size 6240
|
llama32-3b-nt/model/model/decoder/12/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ae06ea561c95f57edb6a2ca63c2eedb56504409e9b293f6c3a750dcb98028dad
|
| 3 |
+
size 50331888
|
llama32-3b-nt/model/model/decoder/12/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a416a78c9fe07902102a320c7961be66a020acc6deb10fef284d195b8970efab
|
| 3 |
+
size 100663608
|
llama32-3b-nt/model/model/decoder/12/pp_block/post_attention_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:394686430fb4a6e56f6ff1d332ad6fef131dab8c63044605688280f0b0df8bbb
|
| 3 |
+
size 6240
|
llama32-3b-nt/model/model/decoder/13/pp_block/post_attention_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:86602423e05fa789b9fd0f814ca1fdbad07a61b63e755b1a1c3527fe6601f8d2
|
| 3 |
+
size 6240
|
llama32-3b-nt/model/model/decoder/16/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:37da8cf762ec043f457ee1518ccfdd0fc553faf85fcad94ce2484f555360601d
|
| 3 |
+
size 18874608
|
llama32-3b-nt/model/model/decoder/16/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:74ea288c717e202c85d10ac55e8569000b4d19ca8f9d44ff44ef135e487b11b3
|
| 3 |
+
size 31457648
|
llama32-3b-nt/model/model/decoder/16/pp_block/input_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d2b984c74d675021cfb16c2052c79a678f8246fadbb5085773f50b1e1a0a3b9c
|
| 3 |
+
size 6240
|
llama32-3b-nt/model/model/decoder/16/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0c4228b1c5eefbad71dc9dc0448a8833bb5cf7d5033fc4438fd185b6899cf31d
|
| 3 |
+
size 50331888
|
llama32-3b-nt/model/model/decoder/16/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a666a4964ea90b98f06e7cb649139b70cb378ba2b4736fc763dcb73bbbf1387d
|
| 3 |
+
size 100663608
|
llama32-3b-nt/model/model/decoder/16/pp_block/post_attention_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:38d6941b68b02a8d958f21b0ba374a18c3cff6c50a7ddc2fc66271bb06e50db7
|
| 3 |
+
size 6240
|
llama32-3b-nt/model/model/decoder/18/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:02d899a14911e45bf3017879c699b54217db5251c8e48d4c7c675418b1599bee
|
| 3 |
+
size 18874608
|
llama32-3b-nt/model/model/decoder/18/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:15276cff9dbb5dc1a4ce6954dee9e3e108d847cea8a15419ab86483f492ff593
|
| 3 |
+
size 31457648
|
llama32-3b-nt/model/model/decoder/18/pp_block/input_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9716e833852a822aa261cc9473012aacf76183afbb0fe4e181585ea9ac853b34
|
| 3 |
+
size 6240
|
llama32-3b-nt/model/model/decoder/18/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:406adc2f49c092333364cf78e261b692da84a1f212767e4b0dd219974c40c8ac
|
| 3 |
+
size 50331888
|
llama32-3b-nt/model/model/decoder/18/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1dfc036b9bb81f14e243c44c379faa2155b71e060366cbab9749da2e0606056a
|
| 3 |
+
size 100663608
|
llama32-3b-nt/model/model/decoder/18/pp_block/post_attention_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aebc970f61d0c18bcc18bbb36620186c98cc013d83655db0b52a69e019efba41
|
| 3 |
+
size 6240
|
llama32-3b-nt/model/model/decoder/5/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c4f6e9bfd3b0c1e6169f5d50d94bc945a1724a4c863f4d1c75901fb60861393c
|
| 3 |
+
size 18874608
|
llama32-3b-nt/model/model/decoder/5/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0ea00218303c10732873504766ac17c5f9883ed92a143f8784654c12557a51a7
|
| 3 |
+
size 31457648
|
llama32-3b-nt/model/model/decoder/5/pp_block/input_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dce54e48e8deb0cf10917d9271d1726ad376ed3bd865cff51bdea0dbb7135a08
|
| 3 |
+
size 6240
|
llama32-3b-nt/model/model/decoder/5/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b8984ef4f0821eac46d240090fa1403428f66692ce822a2ee91fde23964c112f
|
| 3 |
+
size 50331888
|
llama32-3b-nt/model/model/decoder/5/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a303d3ed8d46687731b63f28d4eaee3da32bc2ab9163a833a7dc6ef91a2add78
|
| 3 |
+
size 100663608
|
llama32-3b-nt/model/model/decoder/5/pp_block/post_attention_layernorm/model_weight.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:25268c73f21d21b8cb972b8669bf83e88ccbb187045daaecf6fc253ad7b2329e
|
| 3 |
+
size 6240
|