File size: 6,186 Bytes
c38b2b2 5740268 c38b2b2 02ce519 c38b2b2 fd2d219 0305080 fd2d219 9bb7f29 02ce519 9bb7f29 e058a39 02ce519 fd2d219 ae93770 6adcc70 ae93770 a69a8f3 ae93770 fd2d219 c38b2b2 13911cf c38b2b2 13911cf c38b2b2 13911cf c38b2b2 13911cf c38b2b2 13911cf 2b6ecaa 13911cf c38b2b2 7353728 81f1ccc 7353728 7b80db9 7353728 2c17f69 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 | ---
license: mit
base_model:
- ByteDance-Seed/Seed-Coder-8B-Base
---
# Seed-Coder-8B-Reasoning
## Introduction
We are thrilled to introduce Seed-Coder, a powerful, transparent, and parameter-efficient family of open-source code models at the 8B scale, featuring base, instruct, and reasoning variants. Seed-Coder contributes to promote the evolution of open code models through the following highlights.
- Model-centric: Seed-Coder predominantly leverages LLMs instead of hand-crafted rules for code data filtering, minimizing manual effort in pretraining data construction.
- Transparent: We openly share detailed insights into our model-centric data pipeline, including methods for curating GitHub data, commits data, and code-related web data.
- Powerful: Seed-Coder achieves state-of-the-art performance among open-source models of comparable size across a diverse range of coding tasks.
<p align="center">
<img width="100%" src="imgs/seed-coder_intro_performance.jpg">
</p>
This repo contains Seed-Coder-8B-Reasoning model, which has the following features:
- Type: Causal Language Models
- Data source: Public Dataset
- Training Stage: Pretraining & Post-training
- Context Length: 32,768
## Highlight
**Seed-Coder-8B-Reasoning** is an 8-billion-parameter model further optimized for **code reasoning**, **problem-solving**, and **algorithmic thinking** tasks.
- **Reinforcement Learning** to specifically strengthen **multi-step reasoning**, **algorithm design**, and **competitive programming** capabilities.
- Achieves superior performance across various challenging tasks, rivaling or surpassing larger open-source code models.
## Model Downloads
| Model Name | Length | Download | Notes |
|---------------------------------------------------------|-----------|------------------------------------|-----------------------|
| Seed-Coder-8B-Base | 32K | 🤗 [Model](https://huggingface.co/ByteDance-Seed/Seed-Coder-8B-Base) | Pretrained on our model-centric code data. |
| Seed-Coder-8B-Instruct | 32K | 🤗 [Model](https://huggingface.co/ByteDance-Seed/Seed-Coder-8B-Instruct) | Instruction-tuned for alignment with user intent. |
| 👉 **Seed-Coder-8B-Reasoning** | 32K | 🤗 [Model](https://huggingface.co/ByteDance-Seed/Seed-Coder-8B-Reasoning) | RL trained to boost reasoning capabilities. |
## Requirements
You will need to install the latest versions of `transformers` and `accelerate`:
```bash
pip install -U transformers accelerate
```
## Quickstart
Here is a simple example demonstrating how to load the model and perform code generation using the Hugging Face `pipeline` API:
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
model_id = "ByteDance-Seed/Seed-Coder-8B-Reasoning"
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16, device_map="auto", trust_remote_code=True)
messages = [
{"role": "user", "content": "Write a quick sort algorithm."},
]
input_ids = tokenizer.apply_chat_template(
messages,
tokenize=True,
return_tensors="pt",
add_generation_prompt=True,
).to(model.device)
outputs = model.generate(input_ids, max_new_tokens=16384)
response = tokenizer.decode(outputs[0][input_ids.shape[-1]:], skip_special_tokens=True)
print(response)
```
## Evaluation
Seed-Coder-8B-Reasoning has been evaluated extensively on reasoning-intensive code benchmarks, showing:
- Significant improvements on **competitive programming** datasets and coding challenges.
- Enhanced ability to **break down complex problems**, **design correct algorithms**, and **produce efficient implementations**.
- Strong generalization to unseen problems across multiple domains (math, strings, arrays, graphs, DP, etc.).
<table>
<tr>
<th rowspan="2">Model</th>
<th colspan="3">LiveCodeBench-Hard</th>
<th colspan="3">LiveCodeBench-Medium</th>
<th colspan="3">LiveCodeBench-Easy</th>
<th rowspan="2">Overall</th>
</tr>
<tr>
<th>4mon</th><th>3mon</th><th>2mon</th>
<th>4mon</th><th>3mon</th><th>2mon</th>
<th>4mon</th><th>3mon</th><th>2mon</th>
</tr>
<!-- ~8B Models -->
<tr><td colspan="11"><b>~8B Models</b></td></tr>
<tr>
<td>DeepSeek-R1-Distill-Qwen-7B</td>
<td>11.3</td><td>10.7</td><td>9.6</td>
<td>39.6</td><td>37.2</td><td>37.1</td>
<td>76.2</td><td>77.1</td><td>67.1</td>
<td>36.5</td>
</tr>
<tr>
<td>DeepSeek-R1-Distill-Seed-Coder-8B</td>
<td>13.6</td><td>13.9</td><td>13.4</td>
<td>39.6</td><td>38.7</td><td>39.3</td>
<td>79.8</td><td>80.2</td><td>73.2</td>
<td>39.0</td>
</tr>
<tr>
<td>OlympicCoder-7B</td>
<td>12.7</td><td>11.8</td><td>12.5</td>
<td>40.8</td><td>39.0</td><td>38.7</td>
<td>78.0</td><td>77.1</td><td>67.8</td>
<td>37.9</td>
</tr>
<tr>
<td>Qwen3-8B-thinking</td>
<td>27.5</td><td>23.5</td><td>19.7</td>
<td>65.7</td><td>59.7</td><td>58.5</td>
<td>98.0</td><td>98.1</td><td>97.3</td>
<td>57.4</td>
</tr>
<tr>
<td>Seed-Coder-8B-Reasoning</td>
<td>27.6</td><td>28.0</td><td>31.0</td>
<td>65.8</td><td>59.2</td><td>57.5</td>
<td>87.8</td><td>88.0</td><td>80.1</td>
<td>53.6</td>
</tr>
<!-- 13B+ Models -->
<tr><td colspan="11"><b>13B+ Models</b></td></tr>
<tr>
<td>DeepSeek-R1-Distill-Qwen-14B</td>
<td>21.3</td><td>20.5</td><td>16.1</td>
<td>58.1</td><td>53.4</td><td>51.4</td>
<td>93.3</td><td>94.2</td><td>93.7</td>
<td>51.9</td>
</tr>
<tr>
<td>Claude-3.7-Sonnet-thinking</td>
<td>27.3</td><td>30.8</td><td>31.0</td>
<td>54.5</td><td>55.1</td><td>51.4</td>
<td>96.2</td><td>100.0</td><td>100.0</td>
<td>53.3</td>
</tr>
<tr>
<td>o3-mini-low</td>
<td>30.3</td><td>32.3</td><td>28.6</td>
<td>69.6</td><td>61.2</td><td>54.1</td>
<td>98.7</td><td>100.0</td><td>100.0</td>
<td>59.4</td>
</tr>
</table>
For detailed benchmark performance, please refer to our [📑 technical report](https://github.com/ByteDance-Seed/Seed-Coder/blob/master/Seed-Coder.pdf).
|