nielsr HF Staff commited on
Commit
5775766
·
verified ·
1 Parent(s): d4a5491

Add metadata and link to paper/code

Browse files

Hi, I'm Niels from the Hugging Face community team.

This PR improves the model card for X-Coder-RL-Qwen3-8B by:
- Adding `library_name: transformers` to the metadata.
- Adding `pipeline_tag: text-generation` to categorize the model.
- Adding the Arxiv ID (`2601.06953`) to link the model to the paper [X-Coder: Advancing Competitive Programming with Fully Synthetic Tasks, Solutions, and Tests](https://huggingface.co/papers/2601.06953).
- Ensuring direct links to the GitHub repository are present.

These additions will enable the inference snippet button and improve the model's discoverability on the Hub.

Best regards,
Niels

Files changed (1) hide show
  1. README.md +18 -11
README.md CHANGED
@@ -1,23 +1,30 @@
1
  ---
2
- license: apache-2.0
3
  base_model:
4
- - IIGroup/X-Coder-SFT-Qwen3-8B
5
  datasets:
6
- - IIGroup/X-Coder-RL-40k
7
  language:
8
- - en
 
 
 
9
  tags:
10
- - code
11
- - rl
12
- - competitive-programming
 
13
  ---
14
 
15
  # X-Coder-RL-Qwen3-8B
16
 
17
- X-Coder-RL-Qwen3-8B is a code reasoning foundation model trained with RLVR on fully synthetic rl data, achieving strong performance on competitive programming.
 
 
18
 
19
  ## Model Description
20
 
 
 
21
  - **Base Model**: [IIGroup/X-Coder-SFT-Qwen3-8B](https://huggingface.co/IIGroup/X-Coder-SFT-Qwen3-8B)
22
  - **Training Method**: GRPO
23
  - **Training Data**: [IIGroup/X-Coder-RL-40k](https://huggingface.co/datasets/IIGroup/X-Coder-RL-40k)
@@ -25,11 +32,11 @@ X-Coder-RL-Qwen3-8B is a code reasoning foundation model trained with RLVR on fu
25
 
26
  ## Training
27
 
28
- This model was trained using the X-Coder RLVR recipe. For training details and code, please refer to the [X-Coder GitHub repository](https://github.com/JieWu02/X-Coder).
29
 
30
  ## Performance
31
 
32
- **Average Performance on LiveCodeBench v5 & v6.**
33
 
34
  ![Results](results.png)
35
 
@@ -80,4 +87,4 @@ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
80
 
81
  ## License
82
 
83
- This project is licensed under the Apache License 2.0.
 
1
  ---
 
2
  base_model:
3
+ - IIGroup/X-Coder-SFT-Qwen3-8B
4
  datasets:
5
+ - IIGroup/X-Coder-RL-40k
6
  language:
7
+ - en
8
+ license: apache-2.0
9
+ library_name: transformers
10
+ pipeline_tag: text-generation
11
  tags:
12
+ - code
13
+ - rl
14
+ - competitive-programming
15
+ - arxiv:2601.06953
16
  ---
17
 
18
  # X-Coder-RL-Qwen3-8B
19
 
20
+ X-Coder-RL-Qwen3-8B is a code reasoning foundation model trained with RLVR on fully synthetic reinforcement learning data, achieving strong performance on competitive programming. It was introduced in the paper [X-Coder: Advancing Competitive Programming with Fully Synthetic Tasks, Solutions, and Tests](https://huggingface.co/papers/2601.06953).
21
+
22
+ The model leverages the **SynthSmith** pipeline, which generates diverse and challenging tasks, verified solutions, and tests to advance code reasoning without relying on real-world data.
23
 
24
  ## Model Description
25
 
26
+ - **Paper**: [X-Coder: Advancing Competitive Programming with Fully Synthetic Tasks, Solutions, and Tests](https://arxiv.org/abs/2601.06953)
27
+ - **Code**: [X-Coder GitHub repository](https://github.com/JieWu02/X-Coder)
28
  - **Base Model**: [IIGroup/X-Coder-SFT-Qwen3-8B](https://huggingface.co/IIGroup/X-Coder-SFT-Qwen3-8B)
29
  - **Training Method**: GRPO
30
  - **Training Data**: [IIGroup/X-Coder-RL-40k](https://huggingface.co/datasets/IIGroup/X-Coder-RL-40k)
 
32
 
33
  ## Training
34
 
35
+ This model was trained using the X-Coder RLVR recipe. For training details and environment setup, please refer to the [official GitHub repository](https://github.com/JieWu02/X-Coder).
36
 
37
  ## Performance
38
 
39
+ X-Coder-RL-Qwen3-8B achieves significant performance gains on competitive programming using fully synthetic data, as shown in results for LiveCodeBench v5 and v6.
40
 
41
  ![Results](results.png)
42
 
 
87
 
88
  ## License
89
 
90
+ This project is licensed under the Apache License 2.0.