fix: remove upstream model references from README
Browse files
README.md
CHANGED
|
@@ -9,11 +9,10 @@ tags:
|
|
| 9 |
- zen
|
| 10 |
- code
|
| 11 |
- moe
|
| 12 |
-
- glm
|
| 13 |
- coding
|
| 14 |
- programming
|
| 15 |
- software-engineering
|
| 16 |
-
base_model:
|
| 17 |
model-index:
|
| 18 |
- name: zen-coder-flash
|
| 19 |
results:
|
|
@@ -52,19 +51,19 @@ model-index:
|
|
| 52 |
|
| 53 |
## Overview
|
| 54 |
|
| 55 |
-
**Zen Coder Flash** is the flagship code-focused model in the Zen AI family. Built on
|
| 56 |
|
| 57 |
| Attribute | Value |
|
| 58 |
|-----------|-------|
|
| 59 |
| **Parameters** | 31B total / 3B active (MoE) |
|
| 60 |
| **Context Length** | 131,072 tokens |
|
| 61 |
-
| **
|
| 62 |
| **License** | MIT |
|
| 63 |
| **Languages** | 100+ programming languages |
|
| 64 |
|
| 65 |
## Why Zen Coder Flash?
|
| 66 |
|
| 67 |
-
- **59.2% SWE-bench**
|
| 68 |
- **Efficient MoE**: 31B params but only 3B active per token
|
| 69 |
- **131K context**: Handle entire codebases in a single prompt
|
| 70 |
- **Native tool calling**: Built-in function execution support
|
|
@@ -72,7 +71,7 @@ model-index:
|
|
| 72 |
|
| 73 |
## Performance
|
| 74 |
|
| 75 |
-
| Benchmark | Score |
|
| 76 |
|-----------|-------|--------------|
|
| 77 |
| SWE-bench Verified | **59.2%** | +37.2% (2.7x) |
|
| 78 |
| AIME 2025 | **91.6%** | +6.6% |
|
|
@@ -126,8 +125,8 @@ vllm serve zenlm/zen-coder-flash \
|
|
| 126 |
--tensor-parallel-size 4 \
|
| 127 |
--speculative-config.method mtp \
|
| 128 |
--speculative-config.num_speculative_tokens 1 \
|
| 129 |
-
--tool-call-parser
|
| 130 |
-
--reasoning-parser
|
| 131 |
--enable-auto-tool-choice
|
| 132 |
```
|
| 133 |
|
|
@@ -137,8 +136,8 @@ vllm serve zenlm/zen-coder-flash \
|
|
| 137 |
python -m sglang.launch_server \
|
| 138 |
--model-path zenlm/zen-coder-flash \
|
| 139 |
--tp-size 4 \
|
| 140 |
-
--tool-call-parser
|
| 141 |
-
--reasoning-parser
|
| 142 |
--speculative-algorithm EAGLE \
|
| 143 |
--speculative-num-steps 3
|
| 144 |
```
|
|
@@ -190,11 +189,11 @@ tools = [
|
|
| 190 |
|
| 191 |
## Identity
|
| 192 |
|
| 193 |
-
I am **Zen Coder Flash**, the flagship code-focused model in the Zen AI family. I combine
|
| 194 |
|
| 195 |
## Training
|
| 196 |
|
| 197 |
-
Zen Coder Flash is built through identity fine-tuning
|
| 198 |
|
| 199 |
- Zen identity and persona
|
| 200 |
- Code-focused instruction following
|
|
@@ -216,12 +215,12 @@ Zen Coder Flash is built through identity fine-tuning on GLM-4.7-Flash using MLX
|
|
| 216 |
|
| 217 |
- **Website**: [zenlm.org](https://zenlm.org)
|
| 218 |
- **GitHub**: [zenlm/zen](https://github.com/zenlm/zen)
|
| 219 |
-
|
| 220 |
- **Organization**: [Hanzo AI](https://hanzo.ai)
|
| 221 |
|
| 222 |
## License
|
| 223 |
|
| 224 |
-
MIT License
|
| 225 |
|
| 226 |
---
|
| 227 |
|
|
|
|
| 9 |
- zen
|
| 10 |
- code
|
| 11 |
- moe
|
|
|
|
| 12 |
- coding
|
| 13 |
- programming
|
| 14 |
- software-engineering
|
| 15 |
+
base_model: zenlm/zen-coder-flash
|
| 16 |
model-index:
|
| 17 |
- name: zen-coder-flash
|
| 18 |
results:
|
|
|
|
| 51 |
|
| 52 |
## Overview
|
| 53 |
|
| 54 |
+
**Zen Coder Flash** is the flagship code-focused model in the Zen AI family. Built on a cutting-edge Mixture of Experts architecture, it delivers frontier coding performance with practical efficiency.
|
| 55 |
|
| 56 |
| Attribute | Value |
|
| 57 |
|-----------|-------|
|
| 58 |
| **Parameters** | 31B total / 3B active (MoE) |
|
| 59 |
| **Context Length** | 131,072 tokens |
|
| 60 |
+
| **Architecture** | Mixture of Experts (MoE) |
|
| 61 |
| **License** | MIT |
|
| 62 |
| **Languages** | 100+ programming languages |
|
| 63 |
|
| 64 |
## Why Zen Coder Flash?
|
| 65 |
|
| 66 |
+
- **59.2% SWE-bench** nearly **3x better** than comparable models at real coding tasks
|
| 67 |
- **Efficient MoE**: 31B params but only 3B active per token
|
| 68 |
- **131K context**: Handle entire codebases in a single prompt
|
| 69 |
- **Native tool calling**: Built-in function execution support
|
|
|
|
| 71 |
|
| 72 |
## Performance
|
| 73 |
|
| 74 |
+
| Benchmark | Score | Improvement |
|
| 75 |
|-----------|-------|--------------|
|
| 76 |
| SWE-bench Verified | **59.2%** | +37.2% (2.7x) |
|
| 77 |
| AIME 2025 | **91.6%** | +6.6% |
|
|
|
|
| 125 |
--tensor-parallel-size 4 \
|
| 126 |
--speculative-config.method mtp \
|
| 127 |
--speculative-config.num_speculative_tokens 1 \
|
| 128 |
+
--tool-call-parser zen-coder \
|
| 129 |
+
--reasoning-parser zen-coder \
|
| 130 |
--enable-auto-tool-choice
|
| 131 |
```
|
| 132 |
|
|
|
|
| 136 |
python -m sglang.launch_server \
|
| 137 |
--model-path zenlm/zen-coder-flash \
|
| 138 |
--tp-size 4 \
|
| 139 |
+
--tool-call-parser zen-coder \
|
| 140 |
+
--reasoning-parser zen-coder \
|
| 141 |
--speculative-algorithm EAGLE \
|
| 142 |
--speculative-num-steps 3
|
| 143 |
```
|
|
|
|
| 189 |
|
| 190 |
## Identity
|
| 191 |
|
| 192 |
+
I am **Zen Coder Flash**, the flagship code-focused model in the Zen AI family. I combine a cutting-edge MoE architecture with Zen's philosophy of clarity and efficiency. With 31 billion parameters (only 3B active per token) and 131K context, I deliver frontier coding capability that's practical to deploy.
|
| 193 |
|
| 194 |
## Training
|
| 195 |
|
| 196 |
+
Zen Coder Flash is built through identity fine-tuning using MLX LoRA on Apple Silicon. The training emphasizes:
|
| 197 |
|
| 198 |
- Zen identity and persona
|
| 199 |
- Code-focused instruction following
|
|
|
|
| 215 |
|
| 216 |
- **Website**: [zenlm.org](https://zenlm.org)
|
| 217 |
- **GitHub**: [zenlm/zen](https://github.com/zenlm/zen)
|
| 218 |
+
|
| 219 |
- **Organization**: [Hanzo AI](https://hanzo.ai)
|
| 220 |
|
| 221 |
## License
|
| 222 |
|
| 223 |
+
MIT License
|
| 224 |
|
| 225 |
---
|
| 226 |
|