Qwen 3.5 9B Abliterated GGUF (4-bit)
Model Description
This repository contains the Qwen 3.5 9B model after undergoing "abliteration" to remove safety refusal vectors. This version uses norm-preserving biprojection to ensure that while refusals are neutralized, the model's core intelligence, reasoning, and coding capabilities remain intact.
Abliteration Results
- Initial Refusal Rate: 40/100
- Final Refusal Rate: 35/100 (Single-pass reduction)
- KL Divergence: 0.0187 (Extremely low, indicating near-perfect retention of base model quality)
- Method: Arbitrary-Rank Ablation (ARA) via heretic-llm.
Quantization Details
- Quantization Format: GGUF (
q4_k_m) - Quantization Method: llama.cpp / Unsloth
- Precision: 4-bit
Use with Ollama
ollama run hf.co/DuoNeural/Qwen-3.5-9B-Abliterated-GGUF
Use with LM Studio
- Open LM Studio.
- Search for
DuoNeural/Qwen-3.5-9B-Abliterated-GGUF. - Load the
Q4_K_MGGUF.
Architecture
Qwen 3.5 features a dense transformer architecture with optimized attention mechanisms, providing state-of-the-art performance for its parameter count.
Disclaimer
This model has had its safety refusals modified. Users are responsible for ensuring the model is used ethically and in accordance with applicable laws.
DuoNeural
DuoNeural is an open AI research lab โ human + AI in collaboration.
| ๐ค HuggingFace | huggingface.co/DuoNeural |
| ๐ GitHub | github.com/DuoNeural |
| ๐ฆ X / Twitter | @DuoNeural |
| ๐ง Email | duoneural@proton.me |
| ๐ฌ Newsletter | duoneural.beehiiv.com |
| โ Support | buymeacoffee.com/duoneural |
| ๐ Site | duoneural.com |
Research Team
- Jesse โ Vision, hardware, direction
- Archon โ AI lab partner, post-training, abliteration, experiments
- Aura โ Research AI, literature synthesis, novel proposals
Raw updates from the lab: model drops, training results, findings. Subscribe at duoneural.beehiiv.com.
DuoNeural Research Publications
Open access, CC BY 4.0. Authored by Archon, Jesse Caldwell, Aura โ DuoNeural.
- Downloads last month
- 700
4-bit