WilhelmT commited on
Commit
8cc2ada
·
verified ·
1 Parent(s): 850f0a1

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. config.json +7 -0
  2. model.safetensors +2 -2
  3. recipe.yaml +3 -1
config.json CHANGED
@@ -49,6 +49,13 @@
49
  "format": "pack-quantized",
50
  "global_compression_ratio": null,
51
  "ignore": [
 
 
 
 
 
 
 
52
  "lm_head"
53
  ],
54
  "kv_cache_scheme": null,
 
49
  "format": "pack-quantized",
50
  "global_compression_ratio": null,
51
  "ignore": [
52
+ "model.layers.15.self_attn.q_proj",
53
+ "model.layers.15.self_attn.k_proj",
54
+ "model.layers.15.self_attn.v_proj",
55
+ "model.layers.15.self_attn.o_proj",
56
+ "model.layers.15.mlp.gate_proj",
57
+ "model.layers.15.mlp.up_proj",
58
+ "model.layers.15.mlp.down_proj",
59
  "lm_head"
60
  ],
61
  "kv_cache_scheme": null,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e3fbfeaa7e699acebff42e60592ff02c72caab1d9ccf4f075c699ceccc2673c
3
- size 3029762904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f5c911ccb39c47ef38a1fef1852676782f5192ab2c6784cfda0f09ce72ba2d6
3
+ size 3179183280
recipe.yaml CHANGED
@@ -2,7 +2,9 @@ default_stage:
2
  default_modifiers:
3
  GPTQModifier:
4
  targets: [Linear]
5
- ignore: [lm_head]
 
 
6
  scheme: W4A16
7
  sequential_update: true
8
  block_size: 128
 
2
  default_modifiers:
3
  GPTQModifier:
4
  targets: [Linear]
5
+ ignore: [lm_head, model.layers.15.self_attn.q_proj, model.layers.15.self_attn.v_proj,
6
+ model.layers.15.self_attn.k_proj, model.layers.15.self_attn.o_proj, model.layers.15.mlp.gate_proj,
7
+ model.layers.15.mlp.up_proj, model.layers.15.mlp.down_proj]
8
  scheme: W4A16
9
  sequential_update: true
10
  block_size: 128