| | --- |
| | language: |
| | - en |
| | license: apache-2.0 |
| | tags: |
| | - code |
| | - reasoning |
| | - coding |
| | - instruct |
| | - 8b |
| | - 1kz |
| | - lfm-inspiration |
| | library_name: transformers |
| | pipeline_tag: text-generation |
| | inference: true |
| | --- |
| | |
| | # bigcodemax |
| |
|
| | **Maximum coding + reasoning power in 8B parameters** |
| | Created by **[1kz](https://huggingface.co/1kz)** |
| |
|
| | An 8B model that punches way above its weight in code generation, software engineering, advanced reasoning, math, and long-context understanding. |
| |
|
| | ## Model Details |
| |
|
| | - **Developer**: [1kz](https://huggingface.co/1kz) |
| | - **Parameters**: 8.0B (dense) |
| | - **Context length**: 128K (RoPE scaled) |
| | - **Architecture**: Llama-3.1 style (same tokenizer & chat template as Meta-Llama-3.1-8B-Instruct) |
| | - **Base model**: Fine-tuned from a strong 8B checkpoint |
| | - **Training inspiration**: Huge thanks to **lfm** for the incredible training recipes, data curation, synthetic data pipelines, and open methodology that made this model possible. Your work continues to inspire and push the frontier for compact high-performance models! ❤️ |
| |
|
| | ## Strengths |
| |
|
| | - Best-in-class code generation, editing, and debugging |
| | - Strong mathematical & logical reasoning (CoT & ToT) |
| | - Excellent at understanding and refactoring large codebases |
| | - Agentic coding, tool use, and multi-step problem solving |
| | - Fast inference on consumer hardware (single 4090 / 24GB VRAM) |
| |
|
| | ## Quick Start |
| |
|
| | ```python |
| | from transformers import pipeline |
| | |
| | pipe = pipeline( |
| | "text-generation", |
| | model="1kz/bigcodemax", |
| | device_map="auto", |
| | torch_dtype="auto" |
| | ) |
| | |
| | messages = [ |
| | {"role": "system", "content": "You are bigcodemax, an expert coding and reasoning assistant."}, |
| | {"role": "user", "content": "Implement a thread-safe LRU Cache in Python with O(1) operations and explain every design choice step-by-step."} |
| | ] |
| | |
| | output = pipe(messages, max_new_tokens=2048, temperature=0.6, top_p=0.95, do_sample=True) |
| | print(output[0]["generated_text"][-1]["content"]) |
| | Benchmarks (internal eval) |
| | |
| | |
| | |
| | Massive thank you to lfm — without your public training logs, data mixing strategies, and relentless open-source experimentation, a model this capable at only 8B would not exist. You're building the future of accessible frontier intelligence. 🚀 |