Ryex commited on
Commit
f6f237e
·
verified ·
1 Parent(s): 42d10e6

Initial upload of Floppa-12B (GPTQ)

Browse files
config.json CHANGED
@@ -44,7 +44,9 @@
44
  },
45
  "format": "pack-quantized",
46
  "global_compression_ratio": null,
47
- "ignore": [],
 
 
48
  "kv_cache_scheme": null,
49
  "quant_method": "compressed-tensors",
50
  "quantization_status": "compressed",
 
44
  },
45
  "format": "pack-quantized",
46
  "global_compression_ratio": null,
47
+ "ignore": [
48
+ "embed_out"
49
+ ],
50
  "kv_cache_scheme": null,
51
  "quant_method": "compressed-tensors",
52
  "quantization_status": "compressed",
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a92b7acbca09d0002deb4e3c8d8718e71c9c963505d3c384c0e9d3a414e576a6
3
  size 4999615544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3bd218eb9562e54366f50146a568dc75e3a20f14cbc219e5d8155de77272d33c
3
  size 4999615544
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d81bba0d0ba6568f027673c5c78ab7d6322cc2817584e97da33f7acf0f0d6010
3
- size 2202264608
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a86448798309d43afa484d8bedfc3a81f857b6b4e6105c2c3126bc9870412171
3
+ size 3102483856
model.safetensors.index.json CHANGED
@@ -1,12 +1,10 @@
1
  {
2
  "metadata": {
3
- "total_parameters": 1800448290,
4
- "total_size": 7201794320
5
  },
6
  "weight_map": {
7
- "embed_out.weight_packed": "model-00002-of-00002.safetensors",
8
- "embed_out.weight_scale": "model-00002-of-00002.safetensors",
9
- "embed_out.weight_shape": "model-00002-of-00002.safetensors",
10
  "gpt_neox.embed_in.weight": "model-00001-of-00002.safetensors",
11
  "gpt_neox.final_layer_norm.bias": "model-00002-of-00002.safetensors",
12
  "gpt_neox.final_layer_norm.weight": "model-00002-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_parameters": 2025503008,
4
+ "total_size": 8102013184
5
  },
6
  "weight_map": {
7
+ "embed_out.weight": "model-00002-of-00002.safetensors",
 
 
8
  "gpt_neox.embed_in.weight": "model-00001-of-00002.safetensors",
9
  "gpt_neox.final_layer_norm.bias": "model-00002-of-00002.safetensors",
10
  "gpt_neox.final_layer_norm.weight": "model-00002-of-00002.safetensors",
recipe.yaml CHANGED
@@ -2,7 +2,7 @@ default_stage:
2
  default_modifiers:
3
  GPTQModifier:
4
  targets: [Linear]
5
- ignore: [lm_head]
6
  scheme: W4A16
7
  block_size: 128
8
  dampening_frac: 0.01
 
2
  default_modifiers:
3
  GPTQModifier:
4
  targets: [Linear]
5
+ ignore: [lm_head, embed_out]
6
  scheme: W4A16
7
  block_size: 128
8
  dampening_frac: 0.01