tensorwa commited on
Commit
910fd3e
·
verified ·
1 Parent(s): 8de7c0c

Upload model

Browse files
Files changed (1) hide show
  1. model.safetensors.index.json +6 -6
model.safetensors.index.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "metadata": {
3
- "total_size": 17658814464
 
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00004-of-00004.safetensors",
@@ -12,10 +13,10 @@
12
  "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
13
  "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
14
  "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
15
- "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
16
  "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
17
- "model.layers.1.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
18
  "model.layers.1.input_layernorm.weight": "model-00001-of-00004.safetensors",
 
19
  "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
20
  "model.layers.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
21
  "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
@@ -408,8 +409,8 @@
408
  "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
409
  "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
410
  "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
411
- "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
412
  "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
 
413
  "model.layers.7.input_layernorm.weight": "model-00001-of-00004.safetensors",
414
  "model.layers.7.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
415
  "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
@@ -428,8 +429,8 @@
428
  "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
429
  "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
430
  "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
431
- "model.layers.9.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
432
  "model.layers.9.input_layernorm.weight": "model-00001-of-00004.safetensors",
 
433
  "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
434
  "model.layers.9.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
435
  "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
@@ -440,4 +441,3 @@
440
  "model.norm.weight": "model-00004-of-00004.safetensors"
441
  }
442
  }
443
-
 
1
  {
2
  "metadata": {
3
+ "total_size": 17658814464,
4
+ "path": "model_cache/cilantro9246/k7181kp"
5
  },
6
  "weight_map": {
7
  "lm_head.weight": "model-00004-of-00004.safetensors",
 
13
  "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
14
  "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
15
  "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
 
16
  "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
17
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
18
  "model.layers.1.input_layernorm.weight": "model-00001-of-00004.safetensors",
19
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
20
  "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
21
  "model.layers.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
22
  "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
 
409
  "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
410
  "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
411
  "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
 
412
  "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
413
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
414
  "model.layers.7.input_layernorm.weight": "model-00001-of-00004.safetensors",
415
  "model.layers.7.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
416
  "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
 
429
  "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
430
  "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
431
  "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
 
432
  "model.layers.9.input_layernorm.weight": "model-00001-of-00004.safetensors",
433
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
434
  "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
435
  "model.layers.9.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
436
  "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
 
441
  "model.norm.weight": "model-00004-of-00004.safetensors"
442
  }
443
  }