File size: 2,218 Bytes
f200269
a4e0ec8
 
 
 
ac35b7c
a4e0ec8
ac35b7c
a4e0ec8
 
 
 
f200269
a4e0ec8
 
f200269
 
ac35b7c
f200269
ac35b7c
 
 
 
f200269
 
 
a4e0ec8
 
 
 
 
ac35b7c
f200269
ac35b7c
f200269
ac35b7c
 
 
 
 
f200269
a4e0ec8
f200269
a4e0ec8
 
f200269
a4e0ec8
 
ac35b7c
a4e0ec8
 
 
f200269
a4e0ec8
ac35b7c
 
a4e0ec8
f200269
ac35b7c
 
 
da28521
 
 
ac35b7c
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
---
language:
  - en
license: apache-2.0
tags:
  - code
  - reasoning
  - coding
  - instruct
  - 8b
  - 1kz
  - lfm-inspiration
library_name: transformers
pipeline_tag: text-generation
inference: true
---

# bigcodemax

**Maximum coding + reasoning power in 8B parameters**  
Created by **[1kz](https://huggingface.co/1kz)**

An 8B model that punches way above its weight in code generation, software engineering, advanced reasoning, math, and long-context understanding.

## Model Details

- **Developer**: [1kz](https://huggingface.co/1kz)
- **Parameters**: 8.0B (dense)
- **Context length**: 128K (RoPE scaled)
- **Architecture**: Llama-3.1 style (same tokenizer & chat template as Meta-Llama-3.1-8B-Instruct)
- **Base model**: Fine-tuned from a strong 8B checkpoint
- **Training inspiration**: Huge thanks to **lfm** for the incredible training recipes, data curation, synthetic data pipelines, and open methodology that made this model possible. Your work continues to inspire and push the frontier for compact high-performance models! ❤️

## Strengths

- Best-in-class code generation, editing, and debugging
- Strong mathematical & logical reasoning (CoT & ToT)
- Excellent at understanding and refactoring large codebases
- Agentic coding, tool use, and multi-step problem solving
- Fast inference on consumer hardware (single 4090 / 24GB VRAM)

## Quick Start

```python
from transformers import pipeline

pipe = pipeline(
    "text-generation",
    model="1kz/bigcodemax",
    device_map="auto",
    torch_dtype="auto"
)

messages = [
    {"role": "system", "content": "You are bigcodemax, an expert coding and reasoning assistant."},
    {"role": "user", "content": "Implement a thread-safe LRU Cache in Python with O(1) operations and explain every design choice step-by-step."}
]

output = pipe(messages, max_new_tokens=2048, temperature=0.6, top_p=0.95, do_sample=True)
print(output[0]["generated_text"][-1]["content"])
Benchmarks (internal eval)



Massive thank you to lfm — without your public training logs, data mixing strategies, and relentless open-source experimentation, a model this capable at only 8B would not exist. You're building the future of accessible frontier intelligence. 🚀