| { | |
| "bits": 4, | |
| "data_type": "int", | |
| "group_size": 128, | |
| "sym": true, | |
| "autoround_version": "0.12.0", | |
| "block_name_to_quantize": "model.language_model.layers", | |
| "quant_method": "auto-round", | |
| "packing_format": "auto_round:auto_gptq", | |
| "extra_config": { | |
| "model.language_model.layers.0.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.0.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.1.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.1.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.2.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.2.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.4.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.4.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.5.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.5.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.6.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.6.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.8.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.8.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.9.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.9.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.10.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.10.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.12.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.12.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.13.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.13.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.14.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.14.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.16.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.16.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.17.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.17.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.18.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.18.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.20.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.20.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.21.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.21.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.22.linear_attn.in_proj_b": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| }, | |
| "model.language_model.layers.22.linear_attn.in_proj_a": { | |
| "bits": 16, | |
| "data_type": "fp" | |
| } | |
| } | |
| } |