NousCoder-14B / README.md
jli505's picture
Update README.md
905958a verified
---
license: apache-2.0
datasets:
- livecodebench/code_generation_lite
- agentica-org/DeepCoder-Preview-Dataset
- NousResearch/lcb_test
- NousResearch/RLVR_Coding_Problems
base_model:
- Qwen/Qwen3-14B
pipeline_tag: text-generation
---
# NousCoder-14B
[![](https://img.shields.io/badge/X-NousResearch-000000?logo=x&logoColor=white)](https://x.com/NousResearch)
[![apache 2.0](https://img.shields.io/badge/License-Apache%202.0-orange?logoColor=white&logoUrl=https://pbs.twimg.com/profile_images/1816254738234761216/TX7TW-Mp_400x400.jpg)](https://www.apache.org/licenses/LICENSE-2.0)
We introduce *NousCoder-14B*, a competitive programming model post-trained on [Qwen3-14B](https://huggingface.co/Qwen/Qwen3-14B) via reinforcement learning.
On LiveCodeBench v6 (08/01/2024 - 05/01/2025), we achieve a Pass@1 accuracy of 67.87\%, up 7.08\% from the baseline Pass@1 accuracy of 60.79\%
of Qwen3-14B. We trained on 24k verifiable coding problems using 48 B200s over the course of four days.
![test](./lcb_score_vs_step.png)
![test](./performance_params_ratio.png)
# Acknowledgements
I would like to thank my mentor, Roger Jin, Dakota Mahan, Teknium, and others at the Nous Research team for their invaluable support throughout this project. I would also like to thank Together AI and Agentica for their immensely helpful blog posts on DeepCoder-14B. Finally, thank you to Modal and Lambda for their generous support by providing me with credits.