yuyuzhang commited on
Commit
02ce519
·
verified ·
1 Parent(s): 2d926d0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -5
README.md CHANGED
@@ -7,16 +7,31 @@ base_model:
7
  # Seed-Coder-8B-Reasoning
8
 
9
  ## Introduction
10
- **Seed-Coder-8B-Reasoning** is an 8-billion-parameter model further optimized for **code reasoning**, **problem-solving**, and **algorithmic thinking** tasks.
11
- Built upon the strong base of Seed-Coder, it undergoes additional training in sandbox environments to significantly enhance its ability to tackle complex coding problems and competitions. It features:
12
- - Trained on a **massively curated corpus**, filtered using an **LLM-based method** to ensure high-quality real-world code, text-code alignment, and synthetic datasets.
13
- - **Sandbox fine-tuning** to specifically strengthen **multi-step reasoning**, **algorithm design**, and **competitive programming** capabilities.
14
- - Maintains **long-context handling** up to 32K tokens, enabling it to reason over extended problem descriptions and large input-output examples.
15
 
16
  <p align="center">
17
  <img width="100%" src="imgs/seed-coder_intro_performance.jpg">
18
  </p>
19
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
  ## Model Downloads
21
  | Model Name | Length | Download | Notes |
22
  |---------------------------------------------------------|-----------|------------------------------------|-----------------------|
 
7
  # Seed-Coder-8B-Reasoning
8
 
9
  ## Introduction
10
+ We are thrilled to introduce Seed-Coder, a powerful, transparent, and parameter-efficient family of open-source code models at the 8B scale, featuring base, instruct, and reasoning variants. Seed-Coder contributes to promote the evolution of open code models through the following highlights.
11
+
12
+ - Model-centric: Seed-Coder predominantly leverages LLMs instead of hand-crafted rules for code data filtering, minimizing manual effort in pretraining data construction.
13
+ - Transparent: We openly share detailed insights into our model-centric data pipeline, including methods for curating GitHub data, commits data, and code-related web data.
14
+ - Powerful: Seed-Coder achieves state-of-the-art performance among open-source models of comparable size across a diverse range of coding tasks.
15
 
16
  <p align="center">
17
  <img width="100%" src="imgs/seed-coder_intro_performance.jpg">
18
  </p>
19
 
20
+ This repo contains Seed-Coder-8B-Base model, which has the following features:
21
+ - Type: Causal Language Models
22
+ - Data source: Public Dataset
23
+ - Training Stage: Pretraining & Post-training
24
+ - Context Length: 32,768
25
+
26
+
27
+ ## Highlight
28
+
29
+ **Seed-Coder-8B-Reasoning** is an 8-billion-parameter model further optimized for **code reasoning**, **problem-solving**, and **algorithmic thinking** tasks.
30
+ - Trained on a **massively curated corpus**, filtered using an **LLM-based method** to ensure high-quality real-world code, text-code alignment, and synthetic datasets.
31
+ - **Sandbox fine-tuning** to specifically strengthen **multi-step reasoning**, **algorithm design**, and **competitive programming** capabilities.
32
+
33
+
34
+
35
  ## Model Downloads
36
  | Model Name | Length | Download | Notes |
37
  |---------------------------------------------------------|-----------|------------------------------------|-----------------------|