NicoNico commited on
Commit
1925cd4
·
verified ·
1 Parent(s): c8b3bb1

Upload folder using huggingface_hub

Browse files
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:069ba6953629bf2af4addaeffe7da3dd17c2d1a4478f1200c9ab0bbb8483649c
3
- size 4986536208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17f9a64b2b8edb4f50a9bdada697fb993d740d95ca2288f3604125b5d8b6f295
3
+ size 4982359320
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed54a413f24b6926a4bae89a455487597ec0e44cddd5aee836b872fd376a8c0f
3
- size 2998584312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b792d2d4ac6fad59547965130ef6ccec69fd14e8950de89682ad8ef6fb39dea
3
+ size 2999498736
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 7984959488
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00002-of-00002.safetensors",
@@ -864,11 +864,11 @@
864
  "model.layers.28.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
865
  "model.layers.28.self_attn.v_proj.zeros": "model-00001-of-00002.safetensors",
866
  "model.layers.29.input_layernorm.weight": "model-00002-of-00002.safetensors",
867
- "model.layers.29.mlp.down_proj.channel_scale": "model-00002-of-00002.safetensors",
868
- "model.layers.29.mlp.down_proj.q_perm": "model-00002-of-00002.safetensors",
869
  "model.layers.29.mlp.down_proj.qweight": "model-00002-of-00002.safetensors",
870
- "model.layers.29.mlp.down_proj.scales": "model-00001-of-00002.safetensors",
871
- "model.layers.29.mlp.down_proj.zeros": "model-00001-of-00002.safetensors",
872
  "model.layers.29.mlp.gate_proj.channel_scale": "model-00001-of-00002.safetensors",
873
  "model.layers.29.mlp.gate_proj.q_perm": "model-00001-of-00002.safetensors",
874
  "model.layers.29.mlp.gate_proj.qweight": "model-00001-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 7981697024
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00002-of-00002.safetensors",
 
864
  "model.layers.28.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
865
  "model.layers.28.self_attn.v_proj.zeros": "model-00001-of-00002.safetensors",
866
  "model.layers.29.input_layernorm.weight": "model-00002-of-00002.safetensors",
867
+ "model.layers.29.mlp.down_proj.channel_scale": "model-00001-of-00002.safetensors",
868
+ "model.layers.29.mlp.down_proj.q_perm": "model-00001-of-00002.safetensors",
869
  "model.layers.29.mlp.down_proj.qweight": "model-00002-of-00002.safetensors",
870
+ "model.layers.29.mlp.down_proj.scales": "model-00002-of-00002.safetensors",
871
+ "model.layers.29.mlp.down_proj.zeros": "model-00002-of-00002.safetensors",
872
  "model.layers.29.mlp.gate_proj.channel_scale": "model-00001-of-00002.safetensors",
873
  "model.layers.29.mlp.gate_proj.q_perm": "model-00001-of-00002.safetensors",
874
  "model.layers.29.mlp.gate_proj.qweight": "model-00001-of-00002.safetensors",