File size: 1,284 Bytes
40e0f0c c9f8a72 3b7f247 c9f8a72 3b7f247 c9f8a72 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 | ---
language:
- en
license: apache-2.0
tags:
- math
- reasoning
- mathematics
- causal-lm
- text-generation
library_name: transformers
pipeline_tag: text-generation
model_name: Math-A1-2B
---
# 🐟 Math-A1-2B
**Math-A1-2B** is a 2B-parameter language model by **Kitefish**, focused on mathematical reasoning, symbolic understanding, and structured problem solving.
This is an early release and part of our ongoing effort to build strong, efficient models for reasoning-heavy tasks.
---
## ✨ What this model is good at
- Basic to intermediate **math problem solving**
- **Step-by-step reasoning** for equations and word problems
- Understanding **mathematical symbols and structure**
- Educational and experimentation use cases
---
## 🚀 Quick start
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("kitefish/math-a1-2B")
model = AutoModelForCausalLM.from_pretrained(
"kitefish/math-a1-2B",
torch_dtype="auto",
device_map="auto"
)
prompt = "Solve: 2x + 5 = 13"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|