{ "fp8": { "model\\.layers\\.\\d+": [ "mlp.down_proj", "mlp.gate_proj", "mlp.up_proj", "self_attn.k_proj", "self_attn.o_proj", "self_attn.q_proj", "self_attn.v_proj" ] }, "bf16": { "model\\.embed_tokens": [], "lm_head": [] } }