GGUF Files for bigcodemax

These are the GGUF files for 1kz/bigcodemax.

Downloads

GGUF Link Quantization Description
Download Q2_K Lowest quality
Download Q3_K_S
Download IQ3_S Integer quant, preferable over Q3_K_S
Download IQ3_M Integer quant
Download Q3_K_M
Download Q3_K_L
Download IQ4_XS Integer quant
Download Q4_K_S Fast with good performance
Download Q4_K_M Recommended: Perfect mix of speed and performance
Download Q5_K_S
Download Q5_K_M
Download Q6_K Very good quality
Download Q8_0 Best quality
Download f16 Full precision, don't bother; use a quant

Note from Flexan

I provide GGUFs and quantizations of publicly available models that do not have a GGUF equivalent available yet, usually for models I deem interesting and wish to try out.

If there are some quants missing that you'd like me to add, you may request one in the community tab. If you want to request a public model to be converted, you can also request that in the community tab. If you have questions regarding this model, please refer to the original model repo.

You can find more info about me and what I do here.

bigcodemax

Maximum coding + reasoning power in 8B parameters
Created by 1kz

An 8B model that punches way above its weight in code generation, software engineering, advanced reasoning, math, and long-context understanding.

Model Details

  • Developer: 1kz
  • Parameters: 8.0B (dense)
  • Context length: 128K (RoPE scaled)
  • Architecture: Llama-3.1 style (same tokenizer & chat template as Meta-Llama-3.1-8B-Instruct)
  • Base model: Fine-tuned from a strong 8B checkpoint
  • Training inspiration: Huge thanks to lfm for the incredible training recipes, data curation, synthetic data pipelines, and open methodology that made this model possible. Your work continues to inspire and push the frontier for compact high-performance models! โค๏ธ

Strengths

  • Best-in-class code generation, editing, and debugging
  • Strong mathematical & logical reasoning (CoT & ToT)
  • Excellent at understanding and refactoring large codebases
  • Agentic coding, tool use, and multi-step problem solving
  • Fast inference on consumer hardware (single 4090 / 24GB VRAM)

Quick Start

from transformers import pipeline

pipe = pipeline(
    "text-generation",
    model="1kz/bigcodemax",
    device_map="auto",
    torch_dtype="auto"
)

messages = [
    {"role": "system", "content": "You are bigcodemax, an expert coding and reasoning assistant."},
    {"role": "user", "content": "Implement a thread-safe LRU Cache in Python with O(1) operations and explain every design choice step-by-step."}
]

output = pipe(messages, max_new_tokens=2048, temperature=0.6, top_p=0.95, do_sample=True)
print(output[0]["generated_text"][-1]["content"])
Benchmarks (internal eval)



Massive thank you to lfm โ€” without your public training logs, data mixing strategies, and relentless open-source experimentation, a model this capable at only 8B would not exist. You're building the future of accessible frontier intelligence. ๐Ÿš€
Downloads last month
452
GGUF
Model size
7B params
Architecture
olmo2
Hardware compatibility
Log In to add your hardware

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Flexan/1kz-bigcodemax-GGUF

Base model

1kz/bigcodemax
Quantized
(1)
this model

Collection including Flexan/1kz-bigcodemax-GGUF