| | --- |
| | language: |
| | - en |
| | license: apache-2.0 |
| | tags: |
| | - reasoning |
| | - instruct |
| | - 8b |
| | - 1kz |
| | - lfm-inspiration |
| | library_name: transformers |
| | pipeline_tag: text-generation |
| | inference: true |
| | --- |
| | |
| | # 1kz-Reasoning-8B |
| |
|
| | **A compact 8B reasoning model trained by 1kz** |
| | Strong at logical deduction, math, coding, multi-step problem solving, and long-context reasoning while staying efficient enough to run on a single consumer GPU. |
| |
|
| | ## Model Details |
| |
|
| | - **Developer**: [1kz](https://huggingface.co/1kz) |
| | - **Parameters**: 8.0B (dense) |
| | - **Context length**: 128K (RoPE scaled) |
| | - **Architecture**: Llama-3.1 style (same tokenizer & chat template as Meta-Llama-3.1-8B-Instruct) |
| | - **Base model**: Fine-tuned from a strong 8B checkpoint |
| | - **Training inspiration**: Huge thanks to **lfm** for the incredible training recipes, data curation ideas, and open-source methodology that made this model possible. Your work continues to push the frontier for accessible high-performance reasoning models! ❤️ |
| |
|
| | ## Intended Use |
| |
|
| | - Chain-of-thought reasoning |
| | - Complex math & science problems |
| | - Code generation + debugging |
| | - Agentic workflows |
| | - Research & education |
| |
|
| | ## Quick Start |
| |
|
| | ```python |
| | from transformers import pipeline |
| | |
| | pipe = pipeline( |
| | "text-generation", |
| | model="1kz/1kz-Reasoning-8B", |
| | device_map="auto", |
| | torch_dtype="auto" |
| | ) |
| | |
| | messages = [ |
| | {"role": "system", "content": "You are a world-class reasoning assistant."}, |
| | {"role": "user", "content": "Solve this step-by-step: A bat and a ball cost $1.10 in total. The bat costs $1.00 more than the ball. How much does the ball cost?"} |
| | ] |
| | |
| | output = pipe(messages, max_new_tokens=2048, temperature=0.7, do_sample=True) |
| | print(output[0]["generated_text"][-1]["content"]) |