File size: 2,189 Bytes
1c40cb7
 
 
 
 
 
 
20931d0
1c40cb7
20931d0
 
1c40cb7
 
 
 
a614fea
 
 
 
 
 
 
1c40cb7
 
 
 
 
 
20931d0
 
 
 
 
1c40cb7
 
a614fea
 
 
 
 
 
 
 
 
 
 
 
1c40cb7
a614fea
1c40cb7
a614fea
1c40cb7
a614fea
1c40cb7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20931d0
 
 
 
 
 
 
 
 
1c40cb7
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
---
base_model: google/gemma-3-270m-it
library_name: transformers
model_name: Router
tags:
- trl
- sft
- gemma3
licence: license
datasets:
- d-s-b/synthetic-reasoning-dataset
---

# Model Card for Router

This model is fine-tuned to serve as a router for reasoning tasks, classifying input queries into one of three categories:

  no_reasoning – Direct factual lookup or simple recall (e.g., "What is the capital of France?")
  
  low_reasoning – Requires light reasoning such as simple arithmetic, comparisons, or single logical steps (e.g., "If John has 5 apples and eats 2, how many are left?")
  
  high_reasoning – Requires multi-step reasoning, deep logical chains, or complex problem-solving (e.g., "Prove that the sum of two even numbers is always even").

## Quick start

```python
from transformers import pipeline

pipe = pipeline("text-generation", model="d-s-b/Router")
messages = [
    {"role": "user", "content": "what is capital of india"}
]
pipe(messages)
```

## Training Details

Method: Supervised fine-tuning with SFTTrainer

Objective: Multi-class classification with labels (no_reasoning, low_reasoning, high_reasoning)

Dataset: Custom dataset of queries annotated with reasoning levels.




## Limitations & Bias

May misclassify borderline queries (e.g., between low_reasoning and high_reasoning).

Performance depends on the diversity of training data.

Inherits any biases from the base Gemma 3 270M model.

### Framework versions

- TRL: 0.21.0
- Transformers: 4.55.1
- Pytorch: 2.6.0+cu124
- Datasets: 4.0.0
- Tokenizers: 0.21.4

## Citations



    
```bibtex
@misc{vonwerra2022trl,
	title        = {{TRL: Transformer Reinforcement Learning}},
	author       = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
	year         = 2020,
	journal      = {GitHub repository},
	publisher    = {GitHub},
	howpublished = {\url{https://github.com/huggingface/trl}}
}
@article{gemma_2025,
    title={Gemma 3},
    url={https://arxiv.org/abs/2503.19786},
    publisher={Google DeepMind},
    author={Gemma Team},
    year={2025}
}


```