|
|
--- |
|
|
license: apache-2.0 |
|
|
datasets: |
|
|
- livecodebench/code_generation_lite |
|
|
- agentica-org/DeepCoder-Preview-Dataset |
|
|
- NousResearch/lcb_test |
|
|
- NousResearch/RLVR_Coding_Problems |
|
|
base_model: |
|
|
- Qwen/Qwen3-14B |
|
|
pipeline_tag: text-generation |
|
|
--- |
|
|
|
|
|
# NousCoder-14B |
|
|
|
|
|
[](https://x.com/NousResearch) |
|
|
[](https://www.apache.org/licenses/LICENSE-2.0) |
|
|
|
|
|
We introduce *NousCoder-14B*, a competitive programming model post-trained on [Qwen3-14B](https://huggingface.co/Qwen/Qwen3-14B) via reinforcement learning. |
|
|
On LiveCodeBench v6 (08/01/2024 - 05/01/2025), we achieve a Pass@1 accuracy of 67.87\%, up 7.08\% from the baseline Pass@1 accuracy of 60.79\% |
|
|
of Qwen3-14B. We trained on 24k verifiable coding problems using 48 B200s over the course of four days. |
|
|
|
|
|
 |
|
|
 |
|
|
|
|
|
# Acknowledgements |
|
|
I would like to thank my mentor, Roger Jin, Dakota Mahan, Teknium, and others at the Nous Research team for their invaluable support throughout this project. I would also like to thank Together AI and Agentica for their immensely helpful blog posts on DeepCoder-14B. Finally, thank you to Modal and Lambda for their generous support by providing me with credits. |