Update README.md
Browse files# ALLaM-Thinking: Arabic Large Language Model with Enhanced Reasoning Capabilities
[](https://opensource.org/licenses/Apache-2.0)
[](https://huggingface.co/almaghrabima/ALLaM-Thinking)
[](https://github.com/unslothai/unsloth)
## Overview
ALLaM-Thinking is an advanced Arabic Large Language Model specifically optimized for reasoning and mathematical problem-solving tasks. This model builds on state-of-the-art language model architecture and has been fine-tuned using the Unsloth library for improved performance and efficiency.
## Key Features
- **Arabic-First Design**: Built from the ground up to excel at understanding and generating high-quality Arabic text
- **Enhanced Reasoning**: Specialized in step-by-step problem solving, particularly for mathematical questions
- **Optimized Performance**: Accelerated using Unsloth for faster inference and reduced computational requirements
- **GRPO Implementation**: Utilizes Generalized Reinforced Preference Optimization for improved alignment
## Usage Example
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
from vllm import SamplingParams
import torch
# Load the model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("almaghrabima/ALLaM-Thinking")
model = AutoModelForCausalLM.from_pretrained("almaghrabima/ALLaM-Thinking")
# For faster inference with vLLM
from vllm import LLM, SamplingParams
model = LLM(model="almaghrabima/ALLaM-Thinking")
# Format the prompt using chat template
text = tokenizer.apply_chat_template([
{"role": "user", "content": "ูู ูุฑูู ู
ููู ู
ู 15 ูุงุนุจุงูุ 40% ู
ููู
ูุณุฌููู ุงูุฃูุฏุงู. ุฅุฐุง ุณุฌู ูู ูุงุนุจ ู
ู ุงููุงุนุจูู ุงูุฐูู ูุณุฌููู ุงูุฃูุฏุงู ูู ุงูู
ุชูุณุท 5 ุฃูุฏุงู ุฎูุงู ุงูู
ูุณู
ุ ููู
ุนุฏุฏ ุงูุฃูุฏุงู ุงูููู ุงูุชู ุณุฌููุง ุงููุงุนุจูู ุงูุฐูู ูุณุฌููู ุงูุฃูุฏุงูุ"}
], tokenize=False, add_generation_prompt=True)
# Configure sampling parameters
sampling_params = SamplingParams(
temperature=0.8,
top_p=0.95,
max_tokens=1024,
)
# Generate response
output = model.fast_generate(
[text],
sampling_params=sampling_params,
lora_request=None,
)[0].outputs[0].text
print(output)
```
## Unsloth Optimization
This model has been optimized using [Unsloth](https://github.com/unslothai/unsloth), which provides significant speedups for training and inference.
## Training Details
ALLaM-Thinking was trained using a combination of techniques:
- Base architecture fine-tuned on diverse Arabic datasets
- GRPO (Generalized Reinforced Preference Optimization) for better alignment
- Specialized training on mathematical reasoning and step-by-step problem solving
## Performance
ALLaM-Thinking demonstrates strong capabilities in:
- Mathematical problem solving with step-by-step reasoning
- Logical analysis and deduction
- Maintaining coherence in long-form responses
- Domain-specific reasoning in technical fields
## Limitations
- Model outputs should always be verified by human experts, especially for critical applications
- May occasionally produce incorrect mathematical reasoning despite step-by-step approach
- Limited context window compared to some larger models
- Performance may vary based on query complexity and domain specificity
## Citation
If you use ALLaM-Thinking in your research or applications, please cite:
```bibtex
@misc
{almaghrabima2025allam,
author = {Al-Maghrabima Research},
title = {ALLaM-Thinking: Arabic Large Language Model with Enhanced Reasoning Capabilities},
year = {2025},
publisher = {Hugging Face},
howpublished = {\url{https://huggingface.co/almaghrabima/ALLaM-Thinking}}
}
```
## License
This model is released under the [Apache 2.0 License](https://opensource.org/licenses/Apache-2.0).