File size: 3,999 Bytes
e598549 82a2a73 e598549 82a2a73 e598549 82a2a73 e598549 82a2a73 e598549 82a2a73 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 | ---
license: apache-2.0
language:
- en
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- gguf
library_name: transformers
pipeline_tag: text-generation
datasets:
- Rimyy/problemMath-Llama3.5K
base_model: unsloth/llama-3.2-3b-instruct-bnb-4bit
model_name: llama-3.2-3b-instruct-bnb-4bit-math-gguf
---
# ๐งฎ LLaMA 3.2 3B Instruct (Unsloth 4-bit) โ Finetuned on Rimyy/problemMath-Llama3.5K (GGUF)
This model is a **4-bit GGUF** variant of [`unsloth/llama-3.2-3b-instruct-bnb-4bit`](https://huggingface.co/unsloth/llama-3.2-3b-instruct-bnb-4bit), fine-tuned on [`Rimyy/problemMath-Llama3.5K`](https://huggingface.co/datasets/Rimyy/problemMath-Llama3.5K), a high-quality dataset of math reasoning and problem-solving questions. The model is tailored for **math instruction**, **step-by-step reasoning**, and educational applications.
> ๐จ Designed to reason, not just regurgitate. Small model, big brain.
---
## ๐ง Model Details
| Feature | Value |
|-------------------|-----------------------------------------------------------------------|
| Base | [`unsloth/llama-3.2-3b-instruct-bnb-4bit`](https://huggingface.co/unsloth/llama-3.2-3b-instruct-bnb-4bit) |
| Finetuned Dataset | [`Rimyy/problemMath-Llama3.5K`](https://huggingface.co/datasets/Rimyy/problemMath-Llama3.5K) |
| Quantization | 4-bit GGUF (compatible with llama.cpp/text-generation-webui) |
| Format | GGUF |
| Language | English |
| Instruction Tuned | โ
Yes |
---
## ๐ Dataset: `Rimyy/problemMath-Llama3.5K`
- ~3.5K math word problems and reasoning tasks
- Emphasizes chain-of-thought (CoT) explanations
- Covers arithmetic, algebra, and word problems
- Aligns with OpenAI-style "question โ step-by-step answer" format
---
## ๐ง Quick Usage Example (llama.cpp)
```bash
./main -m llama-3.2-3b-math.gguf --prompt "### Question: What is the value of x if x + 3 = 7?
### Answer:"
```
Expected output:
```
To solve for x, subtract 3 from both sides of the equation:
x + 3 = 7
x = 7 - 3
x = 4
Answer: 4
```
---
## ๐งช Usage in Python
```python
from llama_cpp import Llama
llm = Llama(
model_path="llama-3.2-3b-instruct-math.q4_K.gguf",
n_ctx=2048,
n_gpu_layers=32, # adjust based on your GPU
)
prompt = (
"### Question: If a rectangle has length 10 and width 5, what is its area?
"
"### Answer:"
)
response = llm(prompt)
print(response["choices"][0]["text"])
```
---
## ๐ฆ Applications
- ๐ค Math tutoring agents
- ๐ AI-driven educational platforms
- ๐งฉ RAG pipelines for mathematical queries
- ๐ Automated solution generators
---
## โ ๏ธ Limitations
- Occasional step hallucinations
- Not optimized for LaTeX-heavy symbolic math
- May struggle on very long multi-step problems
---
## ๐ Qualitative Benchmark
| Task Type | Performance |
|-------------------|--------------------|
| Simple Arithmetic | โ
Excellent |
| One-Step Algebra | โ
Strong |
| Multi-Step CoT | โ ๏ธ Good (some drift)|
| Logic Puzzles | โ ๏ธ Mixed |
> ๐ Quantitative benchmarks forthcoming.
---
## ๐ Citation
If you use this model, please cite:
```bibtex
@misc{rimyy2025math,
author = {Rimyy},
title = {ProblemMath-Llama3.5K: A Dataset for Math Problem Solving},
year = {2025},
url = {https://huggingface.co/datasets/Rimyy/problemMath-Llama3.5K}
}
```
---
## ๐ Acknowledgements
- **Meta** for LLaMA 3.
- **Unsloth** for the 4-bit instruct base.
- **Rimyy** for an excellent math dataset.
- **llama.cpp & GGUF** community for stellar tooling.
---
๐ข *Small enough to run on your laptop, smart enough to teach algebra.*
|