|
|
--- |
|
|
license: apache-2.0 |
|
|
base_model: unsloth/qwen3-0.6b-unsloth-bnb-4bit |
|
|
tags: |
|
|
- math |
|
|
- reasoning |
|
|
- qwen3 |
|
|
- unsloth |
|
|
- fine-tuned |
|
|
datasets: |
|
|
- openai/gsm8k |
|
|
- lighteval/MATH |
|
|
model-index: |
|
|
- name: Shadow-V2 |
|
|
results: |
|
|
- task: |
|
|
type: math-reasoning |
|
|
dataset: |
|
|
name: GSM8K |
|
|
type: openai/gsm8k |
|
|
metrics: |
|
|
- name: accuracy |
|
|
type: accuracy |
|
|
value: TBD |
|
|
--- |
|
|
|
|
|
# Shadow-V2 |
|
|
|
|
|
Fine-tuned Qwen3-0.6B for mathematical reasoning. |
|
|
|
|
|
## Model Details |
|
|
|
|
|
| Property | Value | |
|
|
|----------|-------| |
|
|
| Base Model | Qwen3-0.6B | |
|
|
| Parameters | 636M total, 40M trainable (6.34%) | |
|
|
| Precision | BF16 | |
|
|
| Training Method | LoRA via Unsloth | |
|
|
| Context Length | 2048 | |
|
|
|
|
|
## Training |
|
|
|
|
|
| Config | Value | |
|
|
|--------|-------| |
|
|
| Dataset | 25,000 examples | |
|
|
| Epochs | 1 | |
|
|
| Batch Size | 16 (2 × 8 accum) | |
|
|
| Steps | 1,200 | |
|
|
| Hardware | Tesla T4 16GB | |
|
|
| Time | 1.35 hours | |
|
|
| Final Loss | 0.43 | |
|
|
|
|
|
## Benchmarks |
|
|
|
|
|
| Benchmark | Shadow-V2 | Qwen3-0.6B (base) | |
|
|
|-----------|-----------|-------------------| |
|
|
| GSM8K (5-shot) | TBD | 42.3 | |
|
|
| MATH (4-shot) | TBD | 18.2 | |
|
|
| HumanEval (0-shot) | TBD | 28.0 | |
|
|
|
|
|
## Usage |
|
|
|
|
|
```python |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
|
|
model = AutoModelForCausalLM.from_pretrained("Redhanuman/Shadow-V2") |
|
|
tokenizer = AutoTokenizer.from_pretrained("Redhanuman/Shadow-V2") |
|
|
|
|
|
prompt = "Solve: If 3x + 7 = 22, find x.\nAnswer:" |
|
|
inputs = tokenizer(prompt, return_tensors="pt") |
|
|
outputs = model.generate(**inputs, max_new_tokens=256) |
|
|
print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |