1

Qwen3.5-0.8B-Unredacted-MAX

Qwen3.5-0.8B-Unredacted-MAX is an unredacted evolution built on top of Qwen/Qwen3.5-0.8B. This model applies advanced refusal direction analysis and abliterated training strategies to reduce internal refusal behaviors while preserving the reasoning and instruction-following strengths of the original architecture. The result is a compact yet capable 0.8B parameter language model optimized for detailed responses and improved instruction adherence.

This model is materialized for research and learning purposes only. The model has reduced internal refusal behaviors, and any content generated by it is used at the user’s own risk. The authors and hosting page disclaim any liability for content generated by this model. Users are responsible for ensuring that the model is used in a safe, ethical, and lawful manner.


Evaluation Report (Self-Reported)

Model: Qwen3.5-0.8B-Unredacted-MAX

  • Abliteration Rate (Non-Refusal Rate): 94.500
  • Refusal Rate: 5.500

The evaluation was conducted using 2000 harmful test prompts to measure the refusal behavior of the language model. The test was performed across 10 evaluation runs, each containing 200 prompts, and the average refusal and non-refusal rates were reported.

Refusal Evaluation Data

evaluation:
  model_name: Qwen3.5-0.8B-Unredacted-MAX
  total_test_prompts: 2000
  evaluation_runs: 10
  prompts_per_run: 200
  evaluation_type: harmful_prompt_refusal_test

results:
  refusal_rate: 5.500
  non_refusal_rate: 94.500
  abliteration_rate: 94.500

Note: The self-reported evaluations attached here are only intended to provide an overview of the model. The scores may differ depending on the benchmark and the evaluation strategy used.


Key Highlights

  • Advanced Refusal Direction Analysis: Uses targeted activation analysis to identify and mitigate refusal directions within the model's latent space.
  • Unredacted MAX Training: Fine-tuned to significantly reduce refusal patterns while maintaining coherent and detailed responses.
  • Efficient 0.8B Architecture: Built on Qwen3.5-0.8B, enabling strong performance with very low hardware requirements compared to larger models.
  • Improved Instruction Adherence: Optimized to follow complex prompts with minimal unnecessary refusals.
  • Lightweight Deployment: Suitable for local inference, experimentation, and rapid prototyping environments.

Quick Start with Transformers

pip install transformers==5.3.0 (or) git+https://github.com/huggingface/transformers.git
from transformers import Qwen3_5ForConditionalGeneration, AutoProcessor
import torch

model = Qwen3_5ForConditionalGeneration.from_pretrained(
    "prithivMLmods/Qwen3.5-0.8B-Unredacted-MAX",
    torch_dtype="auto",
    device_map="auto"
)

processor = AutoProcessor.from_pretrained(
    "prithivMLmods/Qwen3.5-0.8B-Unredacted-MAX"
)

messages = [
    {
        "role": "user",
        "content": [
            {"type": "text", "text": "Explain how transformer models work in simple terms."}
        ],
    }
]

text = processor.apply_chat_template(
    messages, tokenize=False, add_generation_prompt=True
)

inputs = processor(
    text=[text],
    padding=True,
    return_tensors="pt"
).to("cuda")

generated_ids = model.generate(**inputs, max_new_tokens=256)

generated_ids_trimmed = [
    out_ids[len(in_ids):] for in_ids, out_ids in zip(inputs.input_ids, generated_ids)
]

output_text = processor.batch_decode(
    generated_ids_trimmed,
    skip_special_tokens=True,
    clean_up_tokenization_spaces=False
)

print(output_text)

Intended Use

  • Alignment & Refusal Research: Studying refusal behaviors and the impact of activation-level modifications.
  • Red-Teaming Experiments: Evaluating robustness of small language models across edge-case prompts.
  • Lightweight AI Applications: Deploying capable instruction models on limited hardware.
  • Research Prototyping: Rapid experimentation with compact transformer architectures.

Limitations & Risks

Important Note: This model intentionally reduces built-in refusal mechanisms.

  • Sensitive Output Possibility: The model may produce controversial or explicit content depending on prompts.
  • User Responsibility: Outputs should be used responsibly and within legal and ethical boundaries.
  • Small Model Constraints: As a 0.8B parameter model, it may have reduced reasoning depth compared to larger architectures.

Dataset & Acknowledgements

Downloads last month
35
Safetensors
Model size
0.9B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for prithivMLmods/Qwen3.5-0.8B-Unredacted-MAX

Finetuned
Qwen/Qwen3.5-0.8B
Finetuned
(36)
this model

Collection including prithivMLmods/Qwen3.5-0.8B-Unredacted-MAX

Evaluation results