File size: 2,907 Bytes
db2a9b9
 
a52da4d
db2a9b9
 
 
 
 
d539efa
 
 
 
db2a9b9
 
 
d539efa
 
 
 
a52da4d
 
 
 
db2a9b9
 
a52da4d
db2a9b9
a52da4d
db2a9b9
d539efa
a52da4d
d539efa
a52da4d
db2a9b9
 
d539efa
a52da4d
 
d539efa
 
 
a52da4d
 
d539efa
a52da4d
 
d539efa
a52da4d
 
 
 
 
d539efa
a52da4d
d539efa
 
 
 
 
a52da4d
 
d539efa
 
 
 
a52da4d
d539efa
 
a52da4d
 
d539efa
 
 
 
 
 
a52da4d
d539efa
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
---
base_model: unsloth/gpt-oss-120b-unsloth-bnb-4bit
repo_name: Azmainadeeb/MathGPT
tags:
- text-generation-inference
- transformers
- unsloth
- gpt_oss
- mathematics
- olympiad-math
- reasoning
- chain-of-thought
license: apache-2.0
language:
- en
datasets:
- HuggingFaceH4/Multilingual-Thinking
- Goedel-LM/MathOlympiadBench
- hf-imo-colab/olympiads-ref-base-math-word
- alejopaullier/aimo-external-dataset
- imbishal7/math-olympiad-problems-and-solutions-aops
- baidalinadilzhan/problems-and-solutions-interantional-phos
- kishanvavdara/aimo-olympiadbench-math-dataset
---

# MathGPT (GPT-OSS-120B-Olympiad)

**MathGPT** is a high-performance reasoning model fine-tuned from **GPT-OSS 120B**. It is engineered specifically for solving complex mathematical theorems, competition-level problems (AIME/IMO), and advanced scientific reasoning.

- **Developed by:** Azmainadeeb
- **Model Type:** Causal Language Model (Fine-tuned for Mathematical Reasoning)
- **Base Model:** unsloth/gpt-oss-120b-unsloth-bnb-4bit
- **Training Framework:** Unsloth + TRL

[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)

## 🧩 Model Architecture
MathGPT leverages the **Mixture-of-Experts (MoE)** architecture of the GPT-OSS family, utilizing 117B total parameters with 5.1B active parameters per token. This allows the model to maintain state-of-the-art reasoning depth while remaining computationally efficient during inference.



## 📚 Training Data
The model was trained on a massive synthesis of reasoning-dense datasets to ensure "Chain of Thought" consistency:

### Primary Thinking Dataset
* **[Multilingual-Thinking](https://huggingface.co/datasets/HuggingFaceH4/Multilingual-Thinking):** Instills the core "Thinking" trace and multi-step internal monologue.

### Olympiad & Competition Sets
* **OlympiadBench & MathOlympiadBench:** High-difficulty benchmark problems.
* **IMO Math Boxed:** Problems curated from the International Mathematical Olympiad.
* **AoPS (Art of Problem Solving):** Diverse competition-style math problems.
* **AIMO External Data:** Specific sets designed for the AI Mathematical Olympiad.

## 🚀 Quickstart Usage

```python
from unsloth import FastLanguageModel

model, tokenizer = FastLanguageModel.from_pretrained(
    model_name = "Azmainadeeb/MathGPT",
    max_seq_length = 4096,
    load_in_4bit = True,
)

messages = [
    {"role": "user", "content": "Find all real numbers x such that 8^x + 2^x = 130."}
]

# Apply the template with reasoning_effort to trigger the "Thinking" mode
inputs = tokenizer.apply_chat_template(
    messages,
    add_generation_prompt = True,
    reasoning_effort = "medium", # Options: low, medium, high
    return_tensors = "pt"
).to("cuda")

outputs = model.generate(inputs, max_new_tokens = 1024)
print(tokenizer.decode(outputs[0]))