File size: 3,999 Bytes
e598549
 
 
82a2a73
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e598549
 
82a2a73
e598549
82a2a73
e598549
82a2a73
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e598549
82a2a73
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
---
license: apache-2.0
language:
  - en
tags:
  - text-generation-inference
  - transformers
  - unsloth
  - llama
  - gguf
library_name: transformers
pipeline_tag: text-generation
datasets:
  - Rimyy/problemMath-Llama3.5K
base_model: unsloth/llama-3.2-3b-instruct-bnb-4bit
model_name: llama-3.2-3b-instruct-bnb-4bit-math-gguf
---

# ๐Ÿงฎ LLaMA 3.2 3B Instruct (Unsloth 4-bit) โ€” Finetuned on Rimyy/problemMath-Llama3.5K (GGUF)

This model is a **4-bit GGUF** variant of [`unsloth/llama-3.2-3b-instruct-bnb-4bit`](https://huggingface.co/unsloth/llama-3.2-3b-instruct-bnb-4bit), fine-tuned on [`Rimyy/problemMath-Llama3.5K`](https://huggingface.co/datasets/Rimyy/problemMath-Llama3.5K), a high-quality dataset of math reasoning and problem-solving questions. The model is tailored for **math instruction**, **step-by-step reasoning**, and educational applications.

> ๐Ÿšจ Designed to reason, not just regurgitate. Small model, big brain.

---

## ๐Ÿง  Model Details

| Feature           | Value                                                                 |
|-------------------|-----------------------------------------------------------------------|
| Base              | [`unsloth/llama-3.2-3b-instruct-bnb-4bit`](https://huggingface.co/unsloth/llama-3.2-3b-instruct-bnb-4bit) |
| Finetuned Dataset | [`Rimyy/problemMath-Llama3.5K`](https://huggingface.co/datasets/Rimyy/problemMath-Llama3.5K)              |
| Quantization      | 4-bit GGUF (compatible with llama.cpp/text-generation-webui)         |
| Format            | GGUF                                                                 |
| Language          | English                                                              |
| Instruction Tuned | โœ… Yes                                                               |

---

## ๐Ÿ“š Dataset: `Rimyy/problemMath-Llama3.5K`

- ~3.5K math word problems and reasoning tasks  
- Emphasizes chain-of-thought (CoT) explanations  
- Covers arithmetic, algebra, and word problems  
- Aligns with OpenAI-style "question โ†’ step-by-step answer" format  

---

## ๐Ÿ”ง Quick Usage Example (llama.cpp)

```bash
./main -m llama-3.2-3b-math.gguf   --prompt "### Question: What is the value of x if x + 3 = 7?
### Answer:"
```

Expected output:
```
To solve for x, subtract 3 from both sides of the equation:
x + 3 = 7
x = 7 - 3
x = 4

Answer: 4
```

---

## ๐Ÿงช Usage in Python

```python
from llama_cpp import Llama

llm = Llama(
    model_path="llama-3.2-3b-instruct-math.q4_K.gguf",
    n_ctx=2048,
    n_gpu_layers=32,  # adjust based on your GPU
)

prompt = (
    "### Question: If a rectangle has length 10 and width 5, what is its area?
"
    "### Answer:"
)
response = llm(prompt)
print(response["choices"][0]["text"])
```

---

## ๐Ÿ“ฆ Applications

- ๐Ÿค– Math tutoring agents  
- ๐Ÿ“š AI-driven educational platforms  
- ๐Ÿงฉ RAG pipelines for mathematical queries  
- ๐Ÿ“ Automated solution generators  

---

## โš ๏ธ Limitations

- Occasional step hallucinations  
- Not optimized for LaTeX-heavy symbolic math  
- May struggle on very long multi-step problems  

---

## ๐Ÿ“Š Qualitative Benchmark

| Task Type         | Performance        |
|-------------------|--------------------|
| Simple Arithmetic | โœ… Excellent        |
| One-Step Algebra  | โœ… Strong           |
| Multi-Step CoT    | โš ๏ธ Good (some drift)|
| Logic Puzzles     | โš ๏ธ Mixed            |

> ๐Ÿ“Œ Quantitative benchmarks forthcoming.

---

## ๐Ÿ”— Citation

If you use this model, please cite:

```bibtex
@misc{rimyy2025math,
  author = {Rimyy},
  title = {ProblemMath-Llama3.5K: A Dataset for Math Problem Solving},
  year = {2025},
  url = {https://huggingface.co/datasets/Rimyy/problemMath-Llama3.5K}
}
```

---

## ๐Ÿ™Œ Acknowledgements

- **Meta** for LLaMA 3.  
- **Unsloth** for the 4-bit instruct base.  
- **Rimyy** for an excellent math dataset.  
- **llama.cpp & GGUF** community for stellar tooling.

---

๐Ÿ”ข *Small enough to run on your laptop, smart enough to teach algebra.*