default_stage: default_modifiers: GPTQModifier: targets: [Linear] ignore: [lm_head, model.layers.15.self_attn.q_proj, model.layers.15.self_attn.v_proj, model.layers.15.self_attn.k_proj, model.layers.15.self_attn.o_proj, model.layers.15.mlp.gate_proj, model.layers.15.mlp.up_proj, model.layers.15.mlp.down_proj] scheme: W4A16 sequential_update: true block_size: 128 dampening_frac: 0.01 offload_hessians: false