tyzhu commited on
Commit
18c0e58
·
verified ·
1 Parent(s): 96faf34

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. llama32-3b-nt/model/model/decoder/1/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  2. llama32-3b-nt/model/model/decoder/1/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  3. llama32-3b-nt/model/model/decoder/1/pp_block/input_layernorm/model_weight.safetensors +3 -0
  4. llama32-3b-nt/model/model/decoder/1/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  5. llama32-3b-nt/model/model/decoder/1/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  6. llama32-3b-nt/model/model/decoder/1/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
  7. llama32-3b-nt/model/model/decoder/10/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  8. llama32-3b-nt/model/model/decoder/10/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  9. llama32-3b-nt/model/model/decoder/10/pp_block/input_layernorm/model_weight.safetensors +3 -0
  10. llama32-3b-nt/model/model/decoder/10/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  11. llama32-3b-nt/model/model/decoder/10/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  12. llama32-3b-nt/model/model/decoder/10/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
  13. llama32-3b-nt/model/model/decoder/11/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  14. llama32-3b-nt/model/model/decoder/11/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  15. llama32-3b-nt/model/model/decoder/11/pp_block/input_layernorm/model_weight.safetensors +3 -0
  16. llama32-3b-nt/model/model/decoder/11/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  17. llama32-3b-nt/model/model/decoder/11/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  18. llama32-3b-nt/model/model/decoder/11/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
  19. llama32-3b-nt/model/model/decoder/13/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  20. llama32-3b-nt/model/model/decoder/13/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  21. llama32-3b-nt/model/model/decoder/13/pp_block/input_layernorm/model_weight.safetensors +3 -0
  22. llama32-3b-nt/model/model/decoder/13/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  23. llama32-3b-nt/model/model/decoder/13/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  24. llama32-3b-nt/model/model/decoder/14/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  25. llama32-3b-nt/model/model/decoder/14/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  26. llama32-3b-nt/model/model/decoder/14/pp_block/input_layernorm/model_weight.safetensors +3 -0
  27. llama32-3b-nt/model/model/decoder/14/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  28. llama32-3b-nt/model/model/decoder/14/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  29. llama32-3b-nt/model/model/decoder/14/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
  30. llama32-3b-nt/model/model/decoder/15/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  31. llama32-3b-nt/model/model/decoder/15/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  32. llama32-3b-nt/model/model/decoder/15/pp_block/input_layernorm/model_weight.safetensors +3 -0
  33. llama32-3b-nt/model/model/decoder/15/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  34. llama32-3b-nt/model/model/decoder/15/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  35. llama32-3b-nt/model/model/decoder/15/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
  36. llama32-3b-nt/model/model/decoder/2/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  37. llama32-3b-nt/model/model/decoder/2/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  38. llama32-3b-nt/model/model/decoder/2/pp_block/input_layernorm/model_weight.safetensors +3 -0
  39. llama32-3b-nt/model/model/decoder/2/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  40. llama32-3b-nt/model/model/decoder/2/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  41. llama32-3b-nt/model/model/decoder/2/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
  42. llama32-3b-nt/model/model/decoder/21/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  43. llama32-3b-nt/model/model/decoder/21/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  44. llama32-3b-nt/model/model/decoder/21/pp_block/input_layernorm/model_weight.safetensors +3 -0
  45. llama32-3b-nt/model/model/decoder/21/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  46. llama32-3b-nt/model/model/decoder/21/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  47. llama32-3b-nt/model/model/decoder/21/pp_block/post_attention_layernorm/model_weight.safetensors +3 -0
  48. llama32-3b-nt/model/model/decoder/22/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  49. llama32-3b-nt/model/model/decoder/22/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors +3 -0
  50. llama32-3b-nt/model/model/decoder/22/pp_block/input_layernorm/model_weight.safetensors +3 -0
llama32-3b-nt/model/model/decoder/1/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db16e9099de41c19d5319d8a33a402012192258703e15e63a8e415cb553f8732
3
+ size 18874608
llama32-3b-nt/model/model/decoder/1/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22af9ee92cbd082e3461f74e20cfec04ef47e7a20fc1cf6980f37732be6b2af5
3
+ size 31457648
llama32-3b-nt/model/model/decoder/1/pp_block/input_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41433a152595f679f8101c88d142a646ade5749b4caf2500ec039ba82126a9c0
3
+ size 6240
llama32-3b-nt/model/model/decoder/1/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1f8810490d8bc9d32403c22a986a8b7e0071791ebc9e223f0d2ad56f8a7f4b3
3
+ size 50331888
llama32-3b-nt/model/model/decoder/1/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbfb38e6e52223bf4386c87034b95de43ae9d31206055a4e904f98bdc8d9db17
3
+ size 100663608
llama32-3b-nt/model/model/decoder/1/pp_block/post_attention_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef44cea674462080889559eabc1c85401fa32037c5c2f78855f2f612c8749537
3
+ size 6240
llama32-3b-nt/model/model/decoder/10/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf7c992a53708bfca93e3c70e3cf3607ba777950f7989e257a8b8f908e7f2ca4
3
+ size 18874608
llama32-3b-nt/model/model/decoder/10/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38aa900a66e15a8d9f84abc54475f4414154d8cc7a6d70f88a375ba5e0b50af8
3
+ size 31457648
llama32-3b-nt/model/model/decoder/10/pp_block/input_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55081996a436305fcbaba5a908a63fb56dc0cc5fbd4bc5a0c9591f199a66c240
3
+ size 6240
llama32-3b-nt/model/model/decoder/10/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45f340b3c11021c94e868f34baa9ae18e0f95fbbd344b3c04b9cc69603ff8b3d
3
+ size 50331888
llama32-3b-nt/model/model/decoder/10/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e47e3418908050e42caa3602463e602d4cd95654fb464722e6bb641b21bcb3ae
3
+ size 100663608
llama32-3b-nt/model/model/decoder/10/pp_block/post_attention_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02a71485c1c0964a0f80262a64ebb15515897a236fd6ab24f2ebb43795687b50
3
+ size 6240
llama32-3b-nt/model/model/decoder/11/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c496273e6b1385e54458664fbd6e32bc1d5594a90c5efbea0264038a578bfd9a
3
+ size 18874608
llama32-3b-nt/model/model/decoder/11/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55c18960728ba2b7215c28c9cd56b9ee280ffe1a5f9507add50b392d1c4171f9
3
+ size 31457648
llama32-3b-nt/model/model/decoder/11/pp_block/input_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2e348c2e7169450cccd3201ed225eb458690cdf1898936f60c00c7c19ed2aa9
3
+ size 6240
llama32-3b-nt/model/model/decoder/11/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:932b245b3bcd5426a934948bba67da94ecfe0122b103f6265df05ea3758dbc9d
3
+ size 50331888
llama32-3b-nt/model/model/decoder/11/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3c9d0f80d178bae74a9a2cbbf985a349926ac4c24835fe186d0712eac98f591
3
+ size 100663608
llama32-3b-nt/model/model/decoder/11/pp_block/post_attention_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb20b23cfaf1d4908380c1dfa4acb6f97c6ed8a22524e9e055cf440dfc562e63
3
+ size 6240
llama32-3b-nt/model/model/decoder/13/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9513177642c6e922756b126d2b0f87b1864f445fdee9d9037cead9e917c0a4b2
3
+ size 18874608
llama32-3b-nt/model/model/decoder/13/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41edb98eaaa7d8a53c9c9dc003410cce9eff1d3cb664a4a41ed12da46607c29d
3
+ size 31457648
llama32-3b-nt/model/model/decoder/13/pp_block/input_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d30865329967efc78c71f8fba839b999755193b496a52fdacd0f1ecc877bc420
3
+ size 6240
llama32-3b-nt/model/model/decoder/13/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6340e99d2edcc61c51a2b9cbd5ed14c0798497fe0ddffa0edf1018b4a11f2cd5
3
+ size 50331888
llama32-3b-nt/model/model/decoder/13/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59e3e706bcee69a20dfbadede18b074430d406a3e5ebda24b9cbbcf8709e3e72
3
+ size 100663608
llama32-3b-nt/model/model/decoder/14/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58e396020981994e81398a35ba994cf1f4cbf1ff4971ca7ade531d6ccfc90d47
3
+ size 18874608
llama32-3b-nt/model/model/decoder/14/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:427284e2ea0e1431c462e58c9bc752d713c64a2b88120ed2d55763a7b1790be2
3
+ size 31457648
llama32-3b-nt/model/model/decoder/14/pp_block/input_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fcda620b24608c9011050ed23b3dc8061cbf7bed90d43f0ada0641f13c526c9
3
+ size 6240
llama32-3b-nt/model/model/decoder/14/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b994975293a53148d3ad99567e38cbb59dd201892a512066b4387b60df13d61b
3
+ size 50331888
llama32-3b-nt/model/model/decoder/14/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83058eebe8d1e42211392c96a593afe567fe72a6719147c87042863bd63c3e11
3
+ size 100663608
llama32-3b-nt/model/model/decoder/14/pp_block/post_attention_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee7a2640d5e64bad5dceb086230aaf46eba07dfe1c372e40d853c6b9ef128926
3
+ size 6240
llama32-3b-nt/model/model/decoder/15/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f19240557a71b697978deb9201399ea3c4f1d62f22a8bf74a8525c25b7853e19
3
+ size 18874608
llama32-3b-nt/model/model/decoder/15/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e3d34ea408fe5655a127ed41685921dcb247862c6a5540d4b5b414093cc5576
3
+ size 31457648
llama32-3b-nt/model/model/decoder/15/pp_block/input_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e0f386c173e7c0450078f381f8e72c64714a07c4de160203d8b0c48f3ad2cc3
3
+ size 6240
llama32-3b-nt/model/model/decoder/15/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1f80d260c74ad45c22fb64744d0c66d69e8ada90b085800ed31ae763b1285af
3
+ size 50331888
llama32-3b-nt/model/model/decoder/15/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3147767a860225127eeaaacea2cbfcb933f6f7f7dec3536a6ee3aaf78834817
3
+ size 100663608
llama32-3b-nt/model/model/decoder/15/pp_block/post_attention_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7785b8612fbbd8e9be04c5266d3a6da598a64c5219a44708b2bd1987f059cd8e
3
+ size 6240
llama32-3b-nt/model/model/decoder/2/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3d45588f92db01a31b66b84c6e1f5cf3a108ef4038878e59b103eb8de4b9ece
3
+ size 18874608
llama32-3b-nt/model/model/decoder/2/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:457cc1ee604196f25afd6ae656075e20dfab77534bc2f0b3d3021b1a58b710ef
3
+ size 31457648
llama32-3b-nt/model/model/decoder/2/pp_block/input_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8f9bba8ee59791fa51daaabdda9dc082f5362358858025fd8feeffe0056b4ea
3
+ size 6240
llama32-3b-nt/model/model/decoder/2/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f50eb53d00010cef84b8d3fa1d75e188a808b1873ff237ed2778ed99fe6a403
3
+ size 50331888
llama32-3b-nt/model/model/decoder/2/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:995387559a6e9303c7cea8839723af3c7d77ecc52ba68090d6d70fc27bb57d3b
3
+ size 100663608
llama32-3b-nt/model/model/decoder/2/pp_block/post_attention_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:162371a4655a9a5aaee70e4602e14893d0d575e31693ad93798c91e833de4193
3
+ size 6240
llama32-3b-nt/model/model/decoder/21/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea29134c4c3a50f37aa427f8aea8d4982d0e0a2cc919db6a4d067a9829d90962
3
+ size 18874608
llama32-3b-nt/model/model/decoder/21/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8d3e13858aee490f72d869c2b73670b463bf9f196d8dd375cd8f8307fc4b365
3
+ size 31457648
llama32-3b-nt/model/model/decoder/21/pp_block/input_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d96a3ae0a2bd5890d89dd6ab37b41a1b5f2a8d7d250a09d6c5af07a10e409a4
3
+ size 6240
llama32-3b-nt/model/model/decoder/21/pp_block/mlp/down_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3c4c01213fd970d16614a5229f6e2351fdb1d728b7fab655fe83847846dd8e6
3
+ size 50331888
llama32-3b-nt/model/model/decoder/21/pp_block/mlp/gate_up_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd9b9fc5efb46436786164a9bf639406353f9e68e1bab142ab3f208189382f68
3
+ size 100663608
llama32-3b-nt/model/model/decoder/21/pp_block/post_attention_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf968b16e453b470da4a2b974150bd8b9c58e9e663cfe176c5b06aef99845a21
3
+ size 6240
llama32-3b-nt/model/model/decoder/22/pp_block/attn/o_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:222a1b305e3b748b445c70816642fef274169f699d8b08ddd24dfb31891dfd13
3
+ size 18874608
llama32-3b-nt/model/model/decoder/22/pp_block/attn/qkv_proj/model_weight_pp-rank-0-of-1_tp-rank-0-of-1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b84cae384cb18edc96ae413b2ac43be8a142da8bba55a9e1ea0d6ba4ec44723
3
+ size 31457648
llama32-3b-nt/model/model/decoder/22/pp_block/input_layernorm/model_weight.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc101a7869059f6cf14fbdb08546137c54b30605a49edd94b54f371fdfbb45dd
3
+ size 6240