21223wj commited on
Commit
8b0b3df
·
verified ·
1 Parent(s): faac808

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. README.md +84 -10
  3. results.png +3 -0
.gitattributes CHANGED
@@ -34,3 +34,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ results.png filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,13 +1,87 @@
1
  ---
2
- license: mit
3
- language:
4
- - en
5
- metrics:
6
- - code_eval
7
  base_model:
8
- - Qwen/Qwen3-8B-Base
9
- pipeline_tag: text-generation
 
 
 
10
  tags:
11
- - code
12
- - competitive programming
13
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: apache-2.0
 
 
 
 
3
  base_model:
4
+ - Qwen/Qwen3-8B-Base
5
+ datasets:
6
+ - IIGroup/X-Coder-SFT-376k
7
+ language:
8
+ - en
9
  tags:
10
+ - code
11
+ - sft
12
+ - competitive-programming
13
+ ---
14
+
15
+ # X-Coder-SFT-Qwen3-8B
16
+
17
+ X-Coder-SFT-Qwen3-8B is a code generation model fine-tuned on fully synthetic instruction data, designed for competitive programming tasks. It serves as the foundation for subsequent RLVR training.
18
+
19
+ ## Model Description
20
+
21
+ - **Base Model**: [Qwen/Qwen3-8B-Base](https://huggingface.co/Qwen/Qwen3-8B-Base)
22
+ - **Training Method**: Supervised Fine-Tuning (SFT)
23
+ - **Training Data**: [IIGroup/X-Coder-SFT-376k](https://huggingface.co/datasets/IIGroup/X-Coder-SFT-376k) (376k fully synthetic samples)
24
+ - **Parameters**: 8B
25
+
26
+ ## Training
27
+
28
+ This model was trained using [ms-swift](https://github.com/modelscope/ms-swift). For training details and code, please refer to the [X-Coder GitHub repository](https://github.com/JieWu02/X-Coder).
29
+
30
+ ## Performance
31
+
32
+ ![Results](results.png)
33
+
34
+ **Performance on LiveCodeBench v5.** X-Coder-SFT demonstrates strong coding capabilities trained entirely on synthetic data.
35
+
36
+ ## Recommended Inference Parameters
37
+
38
+ | Parameter | Value |
39
+ |-----------|-------|
40
+ | temperature | 0.6 |
41
+ | top_p | 0.95 |
42
+ | top_k | 20 (or -1 to disable) |
43
+ | max_new_tokens | 32768 |
44
+
45
+ ## Usage
46
+
47
+ ```python
48
+ from transformers import AutoModelForCausalLM, AutoTokenizer
49
+
50
+ model_name = "IIGroup/X-Coder-SFT-Qwen3-8B"
51
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
52
+ model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto")
53
+
54
+ prompt = "Write a Python function to solve the two sum problem."
55
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
56
+ outputs = model.generate(
57
+ **inputs,
58
+ max_new_tokens=32768,
59
+ temperature=0.6,
60
+ top_p=0.95,
61
+ top_k=20,
62
+ do_sample=True
63
+ )
64
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
65
+ ```
66
+
67
+ ## Related Models
68
+
69
+ - **RL Model**: [IIGroup/X-Coder-RL-Qwen3-8B](https://huggingface.co/IIGroup/X-Coder-RL-Qwen3-8B) - RLVR trained version achieving 64.0 on LiveCodeBench
70
+
71
+ ## Citation
72
+
73
+ ```bibtex
74
+ @inproceedings{
75
+ anonymous2025xcoder,
76
+ title={X-Coder: Advancing Competitive Programming with Fully Synthetic Tasks, Solutions, and Tests},
77
+ author={Anonymous},
78
+ booktitle={Submitted to The Fourteenth International Conference on Learning Representations},
79
+ year={2025},
80
+ url={https://openreview.net/forum?id=jp4dzBilqH},
81
+ note={under review}
82
+ }
83
+ ```
84
+
85
+ ## License
86
+
87
+ This project is licensed under the Apache License 2.0.
results.png ADDED

Git LFS Details

  • SHA256: 1a65e6a94d2e445ded612e3b2f401e9cbd64766893462f545dcb4baefd18c60c
  • Pointer size: 131 Bytes
  • Size of remote file: 445 kB