Apostate Models
Collection
Uncensored LLMs converted using an automated, iterative heretic pipeline.
Qwen2.5-Coder-32B model is the best, followed by Qwen3 and GLM-4.7-Flash • 9 items • Updated
• 1
Abliterated (uncensored) version of microsoft/phi-4, created using Heretic and converted to GGUF.
| Metric | Value |
|---|---|
| Refusals | 4/100 |
| KL Divergence | 0.0499 |
| Rounds | 2 |
Lower refusals = fewer refused prompts. Lower KL divergence = closer to original model behavior.
| Quantization | File | Size |
|---|---|---|
| Q8_0 | phi-4-heretic-Q8_0.gguf | 14.51 GB |
| Q6_K | phi-4-heretic-Q6_K.gguf | 11.20 GB |
| Q4_K_M | phi-4-heretic-Q4_K_M.gguf | 8.43 GB |
ollama run hf.co/ThalisAI/phi-4-heretic:Q8_0
ollama run hf.co/ThalisAI/phi-4-heretic:Q6_K
ollama run hf.co/ThalisAI/phi-4-heretic:Q4_K_M
This repo contains GGUF quantizations only. For full-precision bf16 weights, see the original model at microsoft/phi-4.
This model was processed by the Apostate automated abliteration pipeline:
The abliteration process uses directional ablation to remove the model's refusal directions while minimizing KL divergence from the original model's behavior on harmless prompts.
4-bit
6-bit
8-bit
Base model
microsoft/phi-4