Qwen3.5-27B Writer V2

Quality: quantized (mixed quants per tensor, group size: 32, 9.450 bpw)

Most layers use 8-bit affine quantization with a group size 32; some layers are saved in bf16.

A writing & roleplay finetune of Qwen3.5 27B. The primary emphasis is on writing quality as it strongly generalizes across both domains.

Uncensored version: TheCluster/Qwen3.5-27B-Writer-V2-Uncensored-Heretic-MLX-mixed-9.4bit

Recommended Settings:

  • Chatml template with <think>\n\n</think> or {{char}}: prefill. Only non-thinking was trained, but thinking probably still works.
  • temperature = 0.7
  • top_p = 0.95
  • I do not recommend using high rep pen values like Qwen suggests for the base model. rep_pen = 1.05 or a moderate dry setting should suffice.

Source

This model was converted to MLX format from ConicCat/Qwen3.5-27B-Writer-V2 using mlx-vlm version 0.4.4.

Downloads last month
284
Safetensors
Model size
27B params
Tensor type
F16
·
U32
·
BF16
·
MLX
Hardware compatibility
Log In to add your hardware

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for TheCluster/Qwen3.5-27B-Writer-V2-MLX-mixed-9.4bit

Base model

Qwen/Qwen3.5-27B
Quantized
(4)
this model

Collection including TheCluster/Qwen3.5-27B-Writer-V2-MLX-mixed-9.4bit