Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
|
@@ -53,30 +53,21 @@ model-index:
|
|
| 53 |
|
| 54 |
# OmniCoder-9B
|
| 55 |
|
|
|
|
|
|
|
| 56 |
[](https://opensource.org/licenses/Apache-2.0)
|
| 57 |
[](https://huggingface.co/Qwen/Qwen3.5-9B)
|
| 58 |
[](https://huggingface.co/Tesslate/OmniCoder-9B-GGUF)
|
| 59 |
-
[](https://tesslate.com)
|
| 60 |
|
| 61 |
-
[Get Started](#quickstart) | [Benchmarks](#benchmarks) | [GGUF Downloads](https://huggingface.co/Tesslate/OmniCoder-9B-GGUF)
|
| 62 |
|
| 63 |
---
|
| 64 |
|
| 65 |
</div>
|
| 66 |
|
| 67 |
-
## Why OmniCoder?
|
| 68 |
-
|
| 69 |
-
Most open coding models are trained on synthetic instruction data. OmniCoder is different. It was trained on **425,000+ real agentic coding trajectories** from the best frontier models in the world: Claude Opus 4.6, GPT-5.4, GPT-5.3-Codex, and Gemini 3.1 Pro. It learned how top-tier agents actually write code, recover from errors, use tools, and solve problems end-to-end.
|
| 70 |
-
|
| 71 |
-
The result: a 9B model that scores **83.8 on GPQA Diamond** (above GPT-OSS-120B's 80.1 and Claude Haiku 4.5's 73), hits **90 on AIME 2025**, and solves **+8 more Terminal-Bench tasks** than its base model (25/89 vs 18/89).
|
| 72 |
-
|
| 73 |
-
You can run it locally. Right now. On a single GPU. [Jump to Quickstart.](#quickstart)
|
| 74 |
-
|
| 75 |
-
---
|
| 76 |
-
|
| 77 |
## Overview
|
| 78 |
|
| 79 |
-
**OmniCoder-9B** is built by [Tesslate](https://tesslate.com), fine-tuned on top of [Qwen3.5-9B](https://huggingface.co/Qwen/Qwen3.5-9B)'s hybrid architecture (Gated Delta Networks interleaved with standard attention).
|
| 80 |
|
| 81 |
The training data was specifically built from **Claude Opus 4.6 agentic and coding reasoning traces**, targeting scaffolding patterns from Claude Code, OpenCode, Codex, and Droid. The dataset includes successful trajectories from models like Claude Opus 4.6, GPT-5.4, GPT-5.3-Codex, and Gemini 3.1 Pro.
|
| 82 |
|
|
@@ -106,12 +97,9 @@ The model shows strong agentic behavior: it recovers from errors (read-before-wr
|
|
| 106 |
|
| 107 |
</div>
|
| 108 |
|
| 109 |
-
**
|
| 110 |
-
- **
|
| 111 |
-
- **
|
| 112 |
-
- **Terminal-Bench 2.0: 28.1** (25/89 tasks solved). +8.1 points over the Qwen3.5-9B base model (20) and above Claude Haiku 4.5 (27).
|
| 113 |
-
|
| 114 |
-
> [Try it yourself.](#quickstart) | [Run it locally with GGUF.](https://huggingface.co/Tesslate/OmniCoder-9B-GGUF)
|
| 115 |
|
| 116 |
---
|
| 117 |
|
|
@@ -159,15 +147,11 @@ print(response.choices[0].message.content)
|
|
| 159 |
|
| 160 |
### llama.cpp (GGUF)
|
| 161 |
|
| 162 |
-
Run it locally on your laptop:
|
| 163 |
-
|
| 164 |
```bash
|
| 165 |
llama-cli --hf-repo Tesslate/OmniCoder-9B-GGUF --hf-file omnicoder-9b-q4_k_m.gguf -p "Your prompt" -c 8192
|
| 166 |
```
|
| 167 |
|
| 168 |
-
|
| 169 |
-
|
| 170 |
-
**[Browse all quantizations here.](https://huggingface.co/Tesslate/OmniCoder-9B-GGUF)**
|
| 171 |
|
| 172 |
---
|
| 173 |
|
|
@@ -238,6 +222,4 @@ Special thanks to the [Axolotl](https://github.com/axolotl-ai-cloud/axolotl) tea
|
|
| 238 |
|
| 239 |
**Built by [Tesslate](https://tesslate.com)**
|
| 240 |
|
| 241 |
-
[Get the model](https://huggingface.co/Tesslate/OmniCoder-9B) | [GGUF quantizations](https://huggingface.co/Tesslate/OmniCoder-9B-GGUF) | [Website](https://tesslate.com)
|
| 242 |
-
|
| 243 |
</div>
|
|
|
|
| 53 |
|
| 54 |
# OmniCoder-9B
|
| 55 |
|
| 56 |
+
### A 9B coding agent fine-tuned on 425K agentic trajectories.
|
| 57 |
+
|
| 58 |
[](https://opensource.org/licenses/Apache-2.0)
|
| 59 |
[](https://huggingface.co/Qwen/Qwen3.5-9B)
|
| 60 |
[](https://huggingface.co/Tesslate/OmniCoder-9B-GGUF)
|
|
|
|
| 61 |
|
| 62 |
+
[Get Started](#quickstart) | [Benchmarks](#benchmarks) | [GGUF Downloads](https://huggingface.co/Tesslate/OmniCoder-9B-GGUF)
|
| 63 |
|
| 64 |
---
|
| 65 |
|
| 66 |
</div>
|
| 67 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 68 |
## Overview
|
| 69 |
|
| 70 |
+
**OmniCoder-9B** is a 9-billion parameter coding agent model built by [Tesslate](https://tesslate.com), fine-tuned on top of [Qwen3.5-9B](https://huggingface.co/Qwen/Qwen3.5-9B)'s hybrid architecture (Gated Delta Networks interleaved with standard attention). It was trained on **425,000+ curated agentic coding trajectories** spanning real-world software engineering tasks, tool use, terminal operations, and multi-step reasoning.
|
| 71 |
|
| 72 |
The training data was specifically built from **Claude Opus 4.6 agentic and coding reasoning traces**, targeting scaffolding patterns from Claude Code, OpenCode, Codex, and Droid. The dataset includes successful trajectories from models like Claude Opus 4.6, GPT-5.4, GPT-5.3-Codex, and Gemini 3.1 Pro.
|
| 73 |
|
|
|
|
| 97 |
|
| 98 |
</div>
|
| 99 |
|
| 100 |
+
- **GPQA Diamond pass@1: 83.8** (166/198). +2.1 points over the Qwen3.5-9B base model (81.7). At pass@3: **86.4** (171/198).
|
| 101 |
+
- **AIME 2025 pass@5: 90** (27/30).
|
| 102 |
+
- **Terminal-Bench 2.0: 28.1** (25/89). +8.1 points over the Qwen3.5-9B base model (20).
|
|
|
|
|
|
|
|
|
|
| 103 |
|
| 104 |
---
|
| 105 |
|
|
|
|
| 147 |
|
| 148 |
### llama.cpp (GGUF)
|
| 149 |
|
|
|
|
|
|
|
| 150 |
```bash
|
| 151 |
llama-cli --hf-repo Tesslate/OmniCoder-9B-GGUF --hf-file omnicoder-9b-q4_k_m.gguf -p "Your prompt" -c 8192
|
| 152 |
```
|
| 153 |
|
| 154 |
+
All quantizations: [Tesslate/OmniCoder-9B-GGUF](https://huggingface.co/Tesslate/OmniCoder-9B-GGUF)
|
|
|
|
|
|
|
| 155 |
|
| 156 |
---
|
| 157 |
|
|
|
|
| 222 |
|
| 223 |
**Built by [Tesslate](https://tesslate.com)**
|
| 224 |
|
|
|
|
|
|
|
| 225 |
</div>
|