File size: 1,444 Bytes
0287376
 
 
 
 
1d972c9
 
0287376
 
 
 
 
4a78e4f
0deb257
 
 
 
693f5a2
905958a
 
c3610aa
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
---
license: apache-2.0
datasets:
- livecodebench/code_generation_lite
- agentica-org/DeepCoder-Preview-Dataset
- NousResearch/lcb_test
- NousResearch/RLVR_Coding_Problems
base_model:
- Qwen/Qwen3-14B
pipeline_tag: text-generation
---

# NousCoder-14B

[![](https://img.shields.io/badge/X-NousResearch-000000?logo=x&logoColor=white)](https://x.com/NousResearch)
[![apache 2.0](https://img.shields.io/badge/License-Apache%202.0-orange?logoColor=white&logoUrl=https://pbs.twimg.com/profile_images/1816254738234761216/TX7TW-Mp_400x400.jpg)](https://www.apache.org/licenses/LICENSE-2.0)

We introduce *NousCoder-14B*, a competitive programming model post-trained on [Qwen3-14B](https://huggingface.co/Qwen/Qwen3-14B) via reinforcement learning.
On LiveCodeBench v6 (08/01/2024 - 05/01/2025), we achieve a Pass@1 accuracy of 67.87\%, up 7.08\% from the baseline Pass@1 accuracy of 60.79\%
of Qwen3-14B. We trained on 24k verifiable coding problems using 48 B200s over the course of four days.

![test](./lcb_score_vs_step.png)
![test](./performance_params_ratio.png)

# Acknowledgements
I would like to thank my mentor, Roger Jin, Dakota Mahan, Teknium, and others at the Nous Research team for their invaluable support throughout this project. I would also like to thank Together AI and Agentica for their immensely helpful blog posts on DeepCoder-14B. Finally, thank you to Modal and Lambda for their generous support by providing me with credits.