File size: 3,526 Bytes
531035b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
---
license: cc-by-4.0
language:
  - en
base_model: Qwen/Qwen2.5-1.5B
pipeline_tag: text-generation
library_name: transformers
tags:
  - DeepMiddleGo
  - math-reasoning
  - fine-tuned
  - qwen
model-index:
  - name: Mobile-ReasoningLLM-v0-1.5B
    results:
      - task:
          type: text-generation
          name: Math Reasoning
        dataset:
          name: AIME 2024
          type: aime-2024
        metrics:
          - name: Pass@1 (avg16)
            type: pass@1
            value: 90.0
      - task:
          type: text-generation
          name: Math Reasoning
        dataset:
          name: AIME 2025
          type: aime-2025
        metrics:
          - name: Pass@1 (avg16)
            type: pass@1
            value: 76.7
---
# Mobile-Flash-v1-1.5B

## Model Description
Mobile-Flash-v1-1.5B is a fine-tuned derivative of [Qwen2.5-1.5B](https://huggingface.co/Qwen/Qwen2.5-1.5B), optimized for reasoning tasks in mathematics generation. It supports up to 40K output tokens for math problems. This model is designed for both commercial and non-commercial research use.
This repository contains the evluation code of Mobile-Flash-v1-1.5B which **starts to explore Self-RL learning** besides of **sparse reward** learning in the reinforcement learning. In this work, I start to explore the self-rl training algorithm after pre-training, r1-reinforcement learning,
r1-curriculumn reinforcement learning to reduce the difficulty of sparse reward and inefficiency in the RL-Post training stage.

- **Architecture**: Dense decoder-only Transformer
- **Base Model**: Qwen2.5-1.5B
- **Parameters**: 1.5 billion
- **Version**: v1 (released Feb 12, 2026)

## Intended Use
- **Primary Use**: Solving complex math problems.
- **Applications**: Research, education, software development, and math reasoning tasks.
- **Limitations**: May not handle ambiguous or poorly formatted inputs well. Ethical use is encouraged to avoid harmful applications.

## Benchmarks
The model was post-trained on a hybrid dataset (automated, human, synthetic) including:
- Math datasets: AIME 2024, AIME 2025

## Evaluation
The model was evaluated on the following benchmarks, achieving strong performance pass1@avg16:

| Model(1.5B)              | AIME24 | AIME25 |
|--------------------------|--------|--------|
| Mobile-ReasoningLLM-v0-1.5B | 60.0 | 45.0 |
| Mobile-Flash-ReasoningLLM-v0-1.5B | 70.0 | 60.0 |
| Viber-Thinker-1.5B | 78.0 | 70.0 |
| **Mobile-Flash-v1-1.5B** | **90.0** | **76.7** |
| Model(>235B)             | AIME24 | AIME25 |
| GPT-5.2      |97.0+   |97.0+ |
| Grok-4       |97.0+   |97.0+ |
| Gemini-3-Pro |97.0+   |97.0+ |
| GPT-OSS-120B | 96.6 | 97.9 |
| GPT-OSS-20B | 96.0 | 98.7 |
| Grok 3 Mini | 95.8 | 93.3 |
| o4-mini | 93.4 | 92.7 |
| o3 | 91.6 | 86.5 |
| DeepSeek-R1-0528(671B) | 91.4 | 87.5 |
| Qwen-3(235B) | 85.7 | 81.5 |

## How to Use
### Requirements
- **Library**: `transformers`, `torch`, `vLLM` or `TensorRT-LLM`
- **Hardware**: Trained and Tested on NVIDIA 8xA100-80GB GPUs
- **Environment**: Python 3.10+ (e.g., Conda `hug` environment)

### Inference Example
```python
import transformers
import torch
model_id = "deepgo/Mobile-Flash-v1-1.5B"
pipeline = transformers.pipeline(
    "text-generation",
    model=model_id,
    model_kwargs={"torch_dtype": torch.bfloat16},
    device_map="auto",
)
# Math problem prompt
prompt = """Solve the following math problem. Make sure to put the answer (and only answer) inside \\boxed{}."""
max-length=40,000 is recommend.(reduced from 48,000 to 40,000)