zeekay commited on
Commit
d97598d
·
verified ·
1 Parent(s): f581c71

fix: remove upstream model references from README

Browse files
Files changed (1) hide show
  1. README.md +13 -14
README.md CHANGED
@@ -9,11 +9,10 @@ tags:
9
  - zen
10
  - code
11
  - moe
12
- - glm
13
  - coding
14
  - programming
15
  - software-engineering
16
- base_model: zai-org/GLM-4.7-Flash
17
  model-index:
18
  - name: zen-coder-flash
19
  results:
@@ -52,19 +51,19 @@ model-index:
52
 
53
  ## Overview
54
 
55
- **Zen Coder Flash** is the flagship code-focused model in the Zen AI family. Built on GLM-4.7-Flash's cutting-edge Mixture of Experts architecture, it delivers frontier coding performance with practical efficiency.
56
 
57
  | Attribute | Value |
58
  |-----------|-------|
59
  | **Parameters** | 31B total / 3B active (MoE) |
60
  | **Context Length** | 131,072 tokens |
61
- | **Base Model** | [GLM-4.7-Flash](https://huggingface.co/zai-org/GLM-4.7-Flash) |
62
  | **License** | MIT |
63
  | **Languages** | 100+ programming languages |
64
 
65
  ## Why Zen Coder Flash?
66
 
67
- - **59.2% SWE-bench** vs 22% Qwen3-30B - nearly **3x better** at real coding tasks
68
  - **Efficient MoE**: 31B params but only 3B active per token
69
  - **131K context**: Handle entire codebases in a single prompt
70
  - **Native tool calling**: Built-in function execution support
@@ -72,7 +71,7 @@ model-index:
72
 
73
  ## Performance
74
 
75
- | Benchmark | Score | vs Qwen3-30B |
76
  |-----------|-------|--------------|
77
  | SWE-bench Verified | **59.2%** | +37.2% (2.7x) |
78
  | AIME 2025 | **91.6%** | +6.6% |
@@ -126,8 +125,8 @@ vllm serve zenlm/zen-coder-flash \
126
  --tensor-parallel-size 4 \
127
  --speculative-config.method mtp \
128
  --speculative-config.num_speculative_tokens 1 \
129
- --tool-call-parser glm47 \
130
- --reasoning-parser glm45 \
131
  --enable-auto-tool-choice
132
  ```
133
 
@@ -137,8 +136,8 @@ vllm serve zenlm/zen-coder-flash \
137
  python -m sglang.launch_server \
138
  --model-path zenlm/zen-coder-flash \
139
  --tp-size 4 \
140
- --tool-call-parser glm47 \
141
- --reasoning-parser glm45 \
142
  --speculative-algorithm EAGLE \
143
  --speculative-num-steps 3
144
  ```
@@ -190,11 +189,11 @@ tools = [
190
 
191
  ## Identity
192
 
193
- I am **Zen Coder Flash**, the flagship code-focused model in the Zen AI family. I combine GLM-4.7's cutting-edge MoE architecture with Zen's philosophy of clarity and efficiency. With 31 billion parameters (only 3B active per token) and 131K context, I deliver frontier coding capability that's practical to deploy.
194
 
195
  ## Training
196
 
197
- Zen Coder Flash is built through identity fine-tuning on GLM-4.7-Flash using MLX LoRA on Apple Silicon. The training emphasizes:
198
 
199
  - Zen identity and persona
200
  - Code-focused instruction following
@@ -216,12 +215,12 @@ Zen Coder Flash is built through identity fine-tuning on GLM-4.7-Flash using MLX
216
 
217
  - **Website**: [zenlm.org](https://zenlm.org)
218
  - **GitHub**: [zenlm/zen](https://github.com/zenlm/zen)
219
- - **Base Model**: [GLM-4.7-Flash](https://huggingface.co/zai-org/GLM-4.7-Flash)
220
  - **Organization**: [Hanzo AI](https://hanzo.ai)
221
 
222
  ## License
223
 
224
- MIT License - inherited from GLM-4.7-Flash base model.
225
 
226
  ---
227
 
 
9
  - zen
10
  - code
11
  - moe
 
12
  - coding
13
  - programming
14
  - software-engineering
15
+ base_model: zenlm/zen-coder-flash
16
  model-index:
17
  - name: zen-coder-flash
18
  results:
 
51
 
52
  ## Overview
53
 
54
+ **Zen Coder Flash** is the flagship code-focused model in the Zen AI family. Built on a cutting-edge Mixture of Experts architecture, it delivers frontier coding performance with practical efficiency.
55
 
56
  | Attribute | Value |
57
  |-----------|-------|
58
  | **Parameters** | 31B total / 3B active (MoE) |
59
  | **Context Length** | 131,072 tokens |
60
+ | **Architecture** | Mixture of Experts (MoE) |
61
  | **License** | MIT |
62
  | **Languages** | 100+ programming languages |
63
 
64
  ## Why Zen Coder Flash?
65
 
66
+ - **59.2% SWE-bench** nearly **3x better** than comparable models at real coding tasks
67
  - **Efficient MoE**: 31B params but only 3B active per token
68
  - **131K context**: Handle entire codebases in a single prompt
69
  - **Native tool calling**: Built-in function execution support
 
71
 
72
  ## Performance
73
 
74
+ | Benchmark | Score | Improvement |
75
  |-----------|-------|--------------|
76
  | SWE-bench Verified | **59.2%** | +37.2% (2.7x) |
77
  | AIME 2025 | **91.6%** | +6.6% |
 
125
  --tensor-parallel-size 4 \
126
  --speculative-config.method mtp \
127
  --speculative-config.num_speculative_tokens 1 \
128
+ --tool-call-parser zen-coder \
129
+ --reasoning-parser zen-coder \
130
  --enable-auto-tool-choice
131
  ```
132
 
 
136
  python -m sglang.launch_server \
137
  --model-path zenlm/zen-coder-flash \
138
  --tp-size 4 \
139
+ --tool-call-parser zen-coder \
140
+ --reasoning-parser zen-coder \
141
  --speculative-algorithm EAGLE \
142
  --speculative-num-steps 3
143
  ```
 
189
 
190
  ## Identity
191
 
192
+ I am **Zen Coder Flash**, the flagship code-focused model in the Zen AI family. I combine a cutting-edge MoE architecture with Zen's philosophy of clarity and efficiency. With 31 billion parameters (only 3B active per token) and 131K context, I deliver frontier coding capability that's practical to deploy.
193
 
194
  ## Training
195
 
196
+ Zen Coder Flash is built through identity fine-tuning using MLX LoRA on Apple Silicon. The training emphasizes:
197
 
198
  - Zen identity and persona
199
  - Code-focused instruction following
 
215
 
216
  - **Website**: [zenlm.org](https://zenlm.org)
217
  - **GitHub**: [zenlm/zen](https://github.com/zenlm/zen)
218
+
219
  - **Organization**: [Hanzo AI](https://hanzo.ai)
220
 
221
  ## License
222
 
223
+ MIT License
224
 
225
  ---
226