smirki commited on
Commit
3c02ff8
·
verified ·
1 Parent(s): a4b862f

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +8 -26
README.md CHANGED
@@ -53,30 +53,21 @@ model-index:
53
 
54
  # OmniCoder-9B
55
 
 
 
56
  [![License](https://img.shields.io/badge/License-Apache_2.0-blue.svg)](https://opensource.org/licenses/Apache-2.0)
57
  [![Base Model](https://img.shields.io/badge/Base-Qwen3.5--9B-purple)](https://huggingface.co/Qwen/Qwen3.5-9B)
58
  [![GGUF](https://img.shields.io/badge/GGUF-Available-green)](https://huggingface.co/Tesslate/OmniCoder-9B-GGUF)
59
- [![Tesslate](https://img.shields.io/badge/Tesslate-Website-orange)](https://tesslate.com)
60
 
61
- [Get Started](#quickstart) | [Benchmarks](#benchmarks) | [GGUF Downloads](https://huggingface.co/Tesslate/OmniCoder-9B-GGUF) | [Website](https://tesslate.com)
62
 
63
  ---
64
 
65
  </div>
66
 
67
- ## Why OmniCoder?
68
-
69
- Most open coding models are trained on synthetic instruction data. OmniCoder is different. It was trained on **425,000+ real agentic coding trajectories** from the best frontier models in the world: Claude Opus 4.6, GPT-5.4, GPT-5.3-Codex, and Gemini 3.1 Pro. It learned how top-tier agents actually write code, recover from errors, use tools, and solve problems end-to-end.
70
-
71
- The result: a 9B model that scores **83.8 on GPQA Diamond** (above GPT-OSS-120B's 80.1 and Claude Haiku 4.5's 73), hits **90 on AIME 2025**, and solves **+8 more Terminal-Bench tasks** than its base model (25/89 vs 18/89).
72
-
73
- You can run it locally. Right now. On a single GPU. [Jump to Quickstart.](#quickstart)
74
-
75
- ---
76
-
77
  ## Overview
78
 
79
- **OmniCoder-9B** is built by [Tesslate](https://tesslate.com), fine-tuned on top of [Qwen3.5-9B](https://huggingface.co/Qwen/Qwen3.5-9B)'s hybrid architecture (Gated Delta Networks interleaved with standard attention).
80
 
81
  The training data was specifically built from **Claude Opus 4.6 agentic and coding reasoning traces**, targeting scaffolding patterns from Claude Code, OpenCode, Codex, and Droid. The dataset includes successful trajectories from models like Claude Opus 4.6, GPT-5.4, GPT-5.3-Codex, and Gemini 3.1 Pro.
82
 
@@ -106,12 +97,9 @@ The model shows strong agentic behavior: it recovers from errors (read-before-wr
106
 
107
  </div>
108
 
109
- **Highlights:**
110
- - **GPQA Diamond pass@1: 83.8** (166/198 correct). Beats GPT-OSS-120B (80.1), Qwen3.5-9B (81.7), Qwen3-Next-80B (77.2), GPT-OSS-20B (71.5), and Claude Haiku 4.5 (73). At pass@3 it reaches **86.4** (171/198).
111
- - **AIME 2025 pass@5: 90** (27/30 correct). Competitive with GPT-OSS-20B (91.7) and GLM-4.7-Flash (91.6).
112
- - **Terminal-Bench 2.0: 28.1** (25/89 tasks solved). +8.1 points over the Qwen3.5-9B base model (20) and above Claude Haiku 4.5 (27).
113
-
114
- > [Try it yourself.](#quickstart) | [Run it locally with GGUF.](https://huggingface.co/Tesslate/OmniCoder-9B-GGUF)
115
 
116
  ---
117
 
@@ -159,15 +147,11 @@ print(response.choices[0].message.content)
159
 
160
  ### llama.cpp (GGUF)
161
 
162
- Run it locally on your laptop:
163
-
164
  ```bash
165
  llama-cli --hf-repo Tesslate/OmniCoder-9B-GGUF --hf-file omnicoder-9b-q4_k_m.gguf -p "Your prompt" -c 8192
166
  ```
167
 
168
- The Q4_K_M quantization (5.7 GB) fits comfortably on most consumer GPUs and Apple Silicon Macs.
169
-
170
- **[Browse all quantizations here.](https://huggingface.co/Tesslate/OmniCoder-9B-GGUF)**
171
 
172
  ---
173
 
@@ -238,6 +222,4 @@ Special thanks to the [Axolotl](https://github.com/axolotl-ai-cloud/axolotl) tea
238
 
239
  **Built by [Tesslate](https://tesslate.com)**
240
 
241
- [Get the model](https://huggingface.co/Tesslate/OmniCoder-9B) | [GGUF quantizations](https://huggingface.co/Tesslate/OmniCoder-9B-GGUF) | [Website](https://tesslate.com)
242
-
243
  </div>
 
53
 
54
  # OmniCoder-9B
55
 
56
+ ### A 9B coding agent fine-tuned on 425K agentic trajectories.
57
+
58
  [![License](https://img.shields.io/badge/License-Apache_2.0-blue.svg)](https://opensource.org/licenses/Apache-2.0)
59
  [![Base Model](https://img.shields.io/badge/Base-Qwen3.5--9B-purple)](https://huggingface.co/Qwen/Qwen3.5-9B)
60
  [![GGUF](https://img.shields.io/badge/GGUF-Available-green)](https://huggingface.co/Tesslate/OmniCoder-9B-GGUF)
 
61
 
62
+ [Get Started](#quickstart) | [Benchmarks](#benchmarks) | [GGUF Downloads](https://huggingface.co/Tesslate/OmniCoder-9B-GGUF)
63
 
64
  ---
65
 
66
  </div>
67
 
 
 
 
 
 
 
 
 
 
 
68
  ## Overview
69
 
70
+ **OmniCoder-9B** is a 9-billion parameter coding agent model built by [Tesslate](https://tesslate.com), fine-tuned on top of [Qwen3.5-9B](https://huggingface.co/Qwen/Qwen3.5-9B)'s hybrid architecture (Gated Delta Networks interleaved with standard attention). It was trained on **425,000+ curated agentic coding trajectories** spanning real-world software engineering tasks, tool use, terminal operations, and multi-step reasoning.
71
 
72
  The training data was specifically built from **Claude Opus 4.6 agentic and coding reasoning traces**, targeting scaffolding patterns from Claude Code, OpenCode, Codex, and Droid. The dataset includes successful trajectories from models like Claude Opus 4.6, GPT-5.4, GPT-5.3-Codex, and Gemini 3.1 Pro.
73
 
 
97
 
98
  </div>
99
 
100
+ - **GPQA Diamond pass@1: 83.8** (166/198). +2.1 points over the Qwen3.5-9B base model (81.7). At pass@3: **86.4** (171/198).
101
+ - **AIME 2025 pass@5: 90** (27/30).
102
+ - **Terminal-Bench 2.0: 28.1** (25/89). +8.1 points over the Qwen3.5-9B base model (20).
 
 
 
103
 
104
  ---
105
 
 
147
 
148
  ### llama.cpp (GGUF)
149
 
 
 
150
  ```bash
151
  llama-cli --hf-repo Tesslate/OmniCoder-9B-GGUF --hf-file omnicoder-9b-q4_k_m.gguf -p "Your prompt" -c 8192
152
  ```
153
 
154
+ All quantizations: [Tesslate/OmniCoder-9B-GGUF](https://huggingface.co/Tesslate/OmniCoder-9B-GGUF)
 
 
155
 
156
  ---
157
 
 
222
 
223
  **Built by [Tesslate](https://tesslate.com)**
224
 
 
 
225
  </div>