bigcodemax / README.md
1kz's picture
Update README.md
a4e0ec8 verified
|
raw
history blame
1.71 kB
---
language:
- en
license: apache-2.0
tags:
- reasoning
- instruct
- 8b
- 1kz
- lfm-inspiration
library_name: transformers
pipeline_tag: text-generation
inference: true
---
# 1kz-Reasoning-8B
**A compact 8B reasoning model trained by 1kz**
Strong at logical deduction, math, coding, multi-step problem solving, and long-context reasoning while staying efficient enough to run on a single consumer GPU.
## Model Details
- **Developer**: [1kz](https://huggingface.co/1kz)
- **Parameters**: 8.0B (dense)
- **Context length**: 128K (RoPE scaled)
- **Architecture**: Llama-3.1 style (same tokenizer & chat template as Meta-Llama-3.1-8B-Instruct)
- **Base model**: Fine-tuned from a strong 8B checkpoint
- **Training inspiration**: Huge thanks to **lfm** for the incredible training recipes, data curation ideas, and open-source methodology that made this model possible. Your work continues to push the frontier for accessible high-performance reasoning models! ❤️
## Intended Use
- Chain-of-thought reasoning
- Complex math & science problems
- Code generation + debugging
- Agentic workflows
- Research & education
## Quick Start
```python
from transformers import pipeline
pipe = pipeline(
"text-generation",
model="1kz/1kz-Reasoning-8B",
device_map="auto",
torch_dtype="auto"
)
messages = [
{"role": "system", "content": "You are a world-class reasoning assistant."},
{"role": "user", "content": "Solve this step-by-step: A bat and a ball cost $1.10 in total. The bat costs $1.00 more than the ball. How much does the ball cost?"}
]
output = pipe(messages, max_new_tokens=2048, temperature=0.7, do_sample=True)
print(output[0]["generated_text"][-1]["content"])