Upload quantize_config.json
Browse files- quantize_config.json +3 -3
quantize_config.json
CHANGED
|
@@ -1,13 +1,13 @@
|
|
| 1 |
{
|
| 2 |
-
"bits":
|
| 3 |
"group_size": 128,
|
| 4 |
"damp_percent": 0.01,
|
| 5 |
"desc_act": false,
|
| 6 |
"static_groups": false,
|
| 7 |
"sym": true,
|
| 8 |
"true_sequential": true,
|
| 9 |
-
"model_name_or_path": "Qwen-7B-
|
| 10 |
-
"model_file_base_name": "gptq_model-
|
| 11 |
"is_marlin_format": false,
|
| 12 |
"quant_method": "gptq"
|
| 13 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"bits": 8,
|
| 3 |
"group_size": 128,
|
| 4 |
"damp_percent": 0.01,
|
| 5 |
"desc_act": false,
|
| 6 |
"static_groups": false,
|
| 7 |
"sym": true,
|
| 8 |
"true_sequential": true,
|
| 9 |
+
"model_name_or_path": "Qwen-7B-8bit",
|
| 10 |
+
"model_file_base_name": "gptq_model-8bit-128g",
|
| 11 |
"is_marlin_format": false,
|
| 12 |
"quant_method": "gptq"
|
| 13 |
}
|