Qwopus3.5-9B-v4 / README.md
MeridianVector's picture
Upload README.md with huggingface_hub
954b5eb verified
---
base_model: Jackrong/Qwopus3.5-9B-v3
tags: [qwen3.5, reasoning, grpo, dapo, dpo, self-correction, verifiable-rewards]
license: apache-2.0
language: [en]
pipeline_tag: text-generation
---
# Qwopus3.5-9B-v4
## Training Pipeline
```
Jackrong/Qwopus3.5-9B-v3 (clean base, 87.8% HumanEval)
-> Phase 1: DAPO-GRPO 150 steps (native TRL)
loss=dapo, clip=[0.2, 0.28], beta=0, mask_truncated=True
Multiplicative reward: tags as entry fee, filler penalized
-> Phase 2: SAI-DPO Round 1 (self-generated pairs)
-> Phase 3: SAI-DPO Round 2 (hard example mining)
-> This Model
```
## Key Innovation: Multiplicative Rewards
- No tags + correct = score x 0.1 (format bad)
- Filler thinking = score x 0.15 - 0.5 (canned phrase penalty)
- Full tags + correct = score x 1.0 + bonuses (full credit)
## DAPO Features (Native TRL)
- Token-level loss (no length bias)
- Asymmetric clipping (epsilon_high=0.28)
- Overlong filtering (mask_truncated=True)
- No KL penalty (beta=0)
## Files
- `merged-model/` — Full merged safetensors
## Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained(
'MeridianVector/Qwopus3.5-9B-v4', subfolder='merged-model',
torch_dtype='auto', device_map='auto', trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained(
'MeridianVector/Qwopus3.5-9B-v4', subfolder='merged-model', trust_remote_code=True)
```
## Acknowledgements
- [Jackrong](https://huggingface.co/Jackrong) for Qwopus3.5-9B-v3
- [Qwen Team](https://huggingface.co/Qwen) for Qwen3.5-9B
Trained on Google Colab G4 (RTX PRO 6000 Blackwell, 96GB).
Benchmarks pending.