tyzhu commited on
Commit
fecead7
·
verified ·
1 Parent(s): 18c0e58

Add files using upload-large-folder tool

Browse files
Files changed (31) hide show
  1. llama32-3b-nt/model/model/decoder/0/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  2. llama32-3b-nt/model/model/decoder/0/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  3. llama32-3b-nt/model/model/decoder/0/pp_block/input_layernorm/model_weight.safetensors +3 -0
  4. llama32-3b-nt/model/model/decoder/0/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  5. llama32-3b-nt/model/model/decoder/0/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  6. llama32-3b-nt/model/model/decoder/0/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
  7. llama32-3b-nt/model/model/decoder/12/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  8. llama32-3b-nt/model/model/decoder/12/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  9. llama32-3b-nt/model/model/decoder/12/pp_block/input_layernorm/model_weight.safetensors +3 -0
  10. llama32-3b-nt/model/model/decoder/12/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  11. llama32-3b-nt/model/model/decoder/12/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  12. llama32-3b-nt/model/model/decoder/12/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
  13. llama32-3b-nt/model/model/decoder/13/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
  14. llama32-3b-nt/model/model/decoder/16/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  15. llama32-3b-nt/model/model/decoder/16/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  16. llama32-3b-nt/model/model/decoder/16/pp_block/input_layernorm/model_weight.safetensors +3 -0
  17. llama32-3b-nt/model/model/decoder/16/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  18. llama32-3b-nt/model/model/decoder/16/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  19. llama32-3b-nt/model/model/decoder/16/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
  20. llama32-3b-nt/model/model/decoder/18/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  21. llama32-3b-nt/model/model/decoder/18/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  22. llama32-3b-nt/model/model/decoder/18/pp_block/input_layernorm/model_weight.safetensors +3 -0
  23. llama32-3b-nt/model/model/decoder/18/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  24. llama32-3b-nt/model/model/decoder/18/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  25. llama32-3b-nt/model/model/decoder/18/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
  26. llama32-3b-nt/model/model/decoder/5/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  27. llama32-3b-nt/model/model/decoder/5/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  28. llama32-3b-nt/model/model/decoder/5/pp_block/input_layernorm/model_weight.safetensors +3 -0
  29. llama32-3b-nt/model/model/decoder/5/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  30. llama32-3b-nt/model/model/decoder/5/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  31. llama32-3b-nt/model/model/decoder/5/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
llama32-3b-nt/model/model/decoder/0/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0a08d86c740b2b922f9f542ebcb59d5566342e71b621d489d7df30b94c25c4c
3
+ size 18874608
llama32-3b-nt/model/model/decoder/0/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d723e68653fae50092c13d203af7a6f44b48c2b55a4da4f95d9d2b864c934f5
3
+ size 31457648
llama32-3b-nt/model/model/decoder/0/pp_block/input_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8417452d451195a37952805821bebc0ae3b9e324f20590560bce0353a48ae94
3
+ size 6240
llama32-3b-nt/model/model/decoder/0/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a40a3115764f1384aebba170b2826a4d46e98671e1585898a6f392cc1784cf7
3
+ size 50331888
llama32-3b-nt/model/model/decoder/0/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1666afc981a7aca298970cafc62fd8650b11b680974727c0b150759bbe276bd6
3
+ size 100663608
llama32-3b-nt/model/model/decoder/0/pp_block/post_attention_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea8d98c9c512ce732ac5d7fe17ecd945687b562e8b744d163f0ee16ceb5f592a
3
+ size 6240
llama32-3b-nt/model/model/decoder/12/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5fc68bf7fef5b281431abfbbea1b034dccb974738556e0362bba333acebc765
3
+ size 18874608
llama32-3b-nt/model/model/decoder/12/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2af246555cc48a418c3bbe3f2d2d1e72bd669d026fac8d90d6d4bdb6a90207f1
3
+ size 31457648
llama32-3b-nt/model/model/decoder/12/pp_block/input_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24e5d8ae35087f1512a38f7e95fc35371e61874b51d90b8a6167e75b22c3480e
3
+ size 6240
llama32-3b-nt/model/model/decoder/12/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae06ea561c95f57edb6a2ca63c2eedb56504409e9b293f6c3a750dcb98028dad
3
+ size 50331888
llama32-3b-nt/model/model/decoder/12/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a416a78c9fe07902102a320c7961be66a020acc6deb10fef284d195b8970efab
3
+ size 100663608
llama32-3b-nt/model/model/decoder/12/pp_block/post_attention_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:394686430fb4a6e56f6ff1d332ad6fef131dab8c63044605688280f0b0df8bbb
3
+ size 6240
llama32-3b-nt/model/model/decoder/13/pp_block/post_attention_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86602423e05fa789b9fd0f814ca1fdbad07a61b63e755b1a1c3527fe6601f8d2
3
+ size 6240
llama32-3b-nt/model/model/decoder/16/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37da8cf762ec043f457ee1518ccfdd0fc553faf85fcad94ce2484f555360601d
3
+ size 18874608
llama32-3b-nt/model/model/decoder/16/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74ea288c717e202c85d10ac55e8569000b4d19ca8f9d44ff44ef135e487b11b3
3
+ size 31457648
llama32-3b-nt/model/model/decoder/16/pp_block/input_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2b984c74d675021cfb16c2052c79a678f8246fadbb5085773f50b1e1a0a3b9c
3
+ size 6240
llama32-3b-nt/model/model/decoder/16/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c4228b1c5eefbad71dc9dc0448a8833bb5cf7d5033fc4438fd185b6899cf31d
3
+ size 50331888
llama32-3b-nt/model/model/decoder/16/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a666a4964ea90b98f06e7cb649139b70cb378ba2b4736fc763dcb73bbbf1387d
3
+ size 100663608
llama32-3b-nt/model/model/decoder/16/pp_block/post_attention_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38d6941b68b02a8d958f21b0ba374a18c3cff6c50a7ddc2fc66271bb06e50db7
3
+ size 6240
llama32-3b-nt/model/model/decoder/18/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02d899a14911e45bf3017879c699b54217db5251c8e48d4c7c675418b1599bee
3
+ size 18874608
llama32-3b-nt/model/model/decoder/18/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15276cff9dbb5dc1a4ce6954dee9e3e108d847cea8a15419ab86483f492ff593
3
+ size 31457648
llama32-3b-nt/model/model/decoder/18/pp_block/input_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9716e833852a822aa261cc9473012aacf76183afbb0fe4e181585ea9ac853b34
3
+ size 6240
llama32-3b-nt/model/model/decoder/18/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:406adc2f49c092333364cf78e261b692da84a1f212767e4b0dd219974c40c8ac
3
+ size 50331888
llama32-3b-nt/model/model/decoder/18/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1dfc036b9bb81f14e243c44c379faa2155b71e060366cbab9749da2e0606056a
3
+ size 100663608
llama32-3b-nt/model/model/decoder/18/pp_block/post_attention_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aebc970f61d0c18bcc18bbb36620186c98cc013d83655db0b52a69e019efba41
3
+ size 6240
llama32-3b-nt/model/model/decoder/5/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4f6e9bfd3b0c1e6169f5d50d94bc945a1724a4c863f4d1c75901fb60861393c
3
+ size 18874608
llama32-3b-nt/model/model/decoder/5/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ea00218303c10732873504766ac17c5f9883ed92a143f8784654c12557a51a7
3
+ size 31457648
llama32-3b-nt/model/model/decoder/5/pp_block/input_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dce54e48e8deb0cf10917d9271d1726ad376ed3bd865cff51bdea0dbb7135a08
3
+ size 6240
llama32-3b-nt/model/model/decoder/5/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8984ef4f0821eac46d240090fa1403428f66692ce822a2ee91fde23964c112f
3
+ size 50331888
llama32-3b-nt/model/model/decoder/5/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a303d3ed8d46687731b63f28d4eaee3da32bc2ab9163a833a7dc6ef91a2add78
3
+ size 100663608
llama32-3b-nt/model/model/decoder/5/pp_block/post_attention_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25268c73f21d21b8cb972b8669bf83e88ccbb187045daaecf6fc253ad7b2329e
3
+ size 6240