Initial upload of Floppa-12B (GPTQ)
Browse files- config.json +3 -1
- model-00001-of-00002.safetensors +1 -1
- model-00002-of-00002.safetensors +2 -2
- model.safetensors.index.json +3 -5
- recipe.yaml +1 -1
config.json
CHANGED
|
@@ -44,7 +44,9 @@
|
|
| 44 |
},
|
| 45 |
"format": "pack-quantized",
|
| 46 |
"global_compression_ratio": null,
|
| 47 |
-
"ignore": [
|
|
|
|
|
|
|
| 48 |
"kv_cache_scheme": null,
|
| 49 |
"quant_method": "compressed-tensors",
|
| 50 |
"quantization_status": "compressed",
|
|
|
|
| 44 |
},
|
| 45 |
"format": "pack-quantized",
|
| 46 |
"global_compression_ratio": null,
|
| 47 |
+
"ignore": [
|
| 48 |
+
"embed_out"
|
| 49 |
+
],
|
| 50 |
"kv_cache_scheme": null,
|
| 51 |
"quant_method": "compressed-tensors",
|
| 52 |
"quantization_status": "compressed",
|
model-00001-of-00002.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4999615544
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3bd218eb9562e54366f50146a568dc75e3a20f14cbc219e5d8155de77272d33c
|
| 3 |
size 4999615544
|
model-00002-of-00002.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a86448798309d43afa484d8bedfc3a81f857b6b4e6105c2c3126bc9870412171
|
| 3 |
+
size 3102483856
|
model.safetensors.index.json
CHANGED
|
@@ -1,12 +1,10 @@
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
-
"total_parameters":
|
| 4 |
-
"total_size":
|
| 5 |
},
|
| 6 |
"weight_map": {
|
| 7 |
-
"embed_out.
|
| 8 |
-
"embed_out.weight_scale": "model-00002-of-00002.safetensors",
|
| 9 |
-
"embed_out.weight_shape": "model-00002-of-00002.safetensors",
|
| 10 |
"gpt_neox.embed_in.weight": "model-00001-of-00002.safetensors",
|
| 11 |
"gpt_neox.final_layer_norm.bias": "model-00002-of-00002.safetensors",
|
| 12 |
"gpt_neox.final_layer_norm.weight": "model-00002-of-00002.safetensors",
|
|
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
+
"total_parameters": 2025503008,
|
| 4 |
+
"total_size": 8102013184
|
| 5 |
},
|
| 6 |
"weight_map": {
|
| 7 |
+
"embed_out.weight": "model-00002-of-00002.safetensors",
|
|
|
|
|
|
|
| 8 |
"gpt_neox.embed_in.weight": "model-00001-of-00002.safetensors",
|
| 9 |
"gpt_neox.final_layer_norm.bias": "model-00002-of-00002.safetensors",
|
| 10 |
"gpt_neox.final_layer_norm.weight": "model-00002-of-00002.safetensors",
|
recipe.yaml
CHANGED
|
@@ -2,7 +2,7 @@ default_stage:
|
|
| 2 |
default_modifiers:
|
| 3 |
GPTQModifier:
|
| 4 |
targets: [Linear]
|
| 5 |
-
ignore: [lm_head]
|
| 6 |
scheme: W4A16
|
| 7 |
block_size: 128
|
| 8 |
dampening_frac: 0.01
|
|
|
|
| 2 |
default_modifiers:
|
| 3 |
GPTQModifier:
|
| 4 |
targets: [Linear]
|
| 5 |
+
ignore: [lm_head, embed_out]
|
| 6 |
scheme: W4A16
|
| 7 |
block_size: 128
|
| 8 |
dampening_frac: 0.01
|