1kz commited on
Commit
ac35b7c
·
verified ·
1 Parent(s): a4e0ec8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -15
README.md CHANGED
@@ -3,7 +3,9 @@ language:
3
  - en
4
  license: apache-2.0
5
  tags:
 
6
  - reasoning
 
7
  - instruct
8
  - 8b
9
  - 1kz
@@ -13,10 +15,12 @@ pipeline_tag: text-generation
13
  inference: true
14
  ---
15
 
16
- # 1kz-Reasoning-8B
17
 
18
- **A compact 8B reasoning model trained by 1kz**
19
- Strong at logical deduction, math, coding, multi-step problem solving, and long-context reasoning while staying efficient enough to run on a single consumer GPU.
 
 
20
 
21
  ## Model Details
22
 
@@ -25,15 +29,15 @@ Strong at logical deduction, math, coding, multi-step problem solving, and long-
25
  - **Context length**: 128K (RoPE scaled)
26
  - **Architecture**: Llama-3.1 style (same tokenizer & chat template as Meta-Llama-3.1-8B-Instruct)
27
  - **Base model**: Fine-tuned from a strong 8B checkpoint
28
- - **Training inspiration**: Huge thanks to **lfm** for the incredible training recipes, data curation ideas, and open-source methodology that made this model possible. Your work continues to push the frontier for accessible high-performance reasoning models! ❤️
29
 
30
- ## Intended Use
31
 
32
- - Chain-of-thought reasoning
33
- - Complex math & science problems
34
- - Code generation + debugging
35
- - Agentic workflows
36
- - Research & education
37
 
38
  ## Quick Start
39
 
@@ -42,15 +46,17 @@ from transformers import pipeline
42
 
43
  pipe = pipeline(
44
  "text-generation",
45
- model="1kz/1kz-Reasoning-8B",
46
  device_map="auto",
47
  torch_dtype="auto"
48
  )
49
 
50
  messages = [
51
- {"role": "system", "content": "You are a world-class reasoning assistant."},
52
- {"role": "user", "content": "Solve this step-by-step: A bat and a ball cost $1.10 in total. The bat costs $1.00 more than the ball. How much does the ball cost?"}
53
  ]
54
 
55
- output = pipe(messages, max_new_tokens=2048, temperature=0.7, do_sample=True)
56
- print(output[0]["generated_text"][-1]["content"])
 
 
 
3
  - en
4
  license: apache-2.0
5
  tags:
6
+ - code
7
  - reasoning
8
+ - coding
9
  - instruct
10
  - 8b
11
  - 1kz
 
15
  inference: true
16
  ---
17
 
18
+ # bigcodemax
19
 
20
+ **Maximum coding + reasoning power in 8B parameters**
21
+ Created by **[1kz](https://huggingface.co/1kz)**
22
+
23
+ An 8B model that punches way above its weight in code generation, software engineering, advanced reasoning, math, and long-context understanding.
24
 
25
  ## Model Details
26
 
 
29
  - **Context length**: 128K (RoPE scaled)
30
  - **Architecture**: Llama-3.1 style (same tokenizer & chat template as Meta-Llama-3.1-8B-Instruct)
31
  - **Base model**: Fine-tuned from a strong 8B checkpoint
32
+ - **Training inspiration**: Huge thanks to **lfm** for the incredible training recipes, data curation, synthetic data pipelines, and open methodology that made this model possible. Your work continues to inspire and push the frontier for compact high-performance models! ❤️
33
 
34
+ ## Strengths
35
 
36
+ - Best-in-class code generation, editing, and debugging
37
+ - Strong mathematical & logical reasoning (CoT & ToT)
38
+ - Excellent at understanding and refactoring large codebases
39
+ - Agentic coding, tool use, and multi-step problem solving
40
+ - Fast inference on consumer hardware (single 4090 / 24GB VRAM)
41
 
42
  ## Quick Start
43
 
 
46
 
47
  pipe = pipeline(
48
  "text-generation",
49
+ model="1kz/bigcodemax",
50
  device_map="auto",
51
  torch_dtype="auto"
52
  )
53
 
54
  messages = [
55
+ {"role": "system", "content": "You are bigcodemax, an expert coding and reasoning assistant."},
56
+ {"role": "user", "content": "Implement a thread-safe LRU Cache in Python with O(1) operations and explain every design choice step-by-step."}
57
  ]
58
 
59
+ output = pipe(messages, max_new_tokens=2048, temperature=0.6, top_p=0.95, do_sample=True)
60
+ print(output[0]["generated_text"][-1]["content"])
61
+ Benchmarks (internal eval)
62
+ Massive thank you to lfm — without your public training logs, data mixing strategies, and relentless open-source experimentation, a model this capable at only 8B would not exist. You're building the future of accessible frontier intelligence. 🚀