| license: apache-2.0 | |
| datasets: | |
| - livecodebench/code_generation_lite | |
| - agentica-org/DeepCoder-Preview-Dataset | |
| - NousResearch/lcb_test | |
| - NousResearch/RLVR_Coding_Problems | |
| base_model: | |
| - Qwen/Qwen3-14B | |
| pipeline_tag: text-generation | |
| library_name: transformers | |
| # NousCoder-14B | |
| [](https://x.com/NousResearch) | |
| [](https://www.apache.org/licenses/LICENSE-2.0) | |
| We introduce *NousCoder-14B*, a competitive programming model post-trained on [Qwen3-14B](https://huggingface.co/Qwen/Qwen3-14B) via reinforcement learning. | |
| On LiveCodeBench v6 (08/01/2024 - 05/01/2025), we achieve a Pass@1 accuracy of 67.87\%, up 7.08\% from the baseline Pass@1 accuracy of 60.79\% | |
| of Qwen3-14B. We trained on 24k verifiable coding problems using 48 B200s over the course of four days. | |
|  | |
|  | |
| # Acknowledgements | |
| I would like to thank my mentor, Roger Jin, Dakota Mahan, Teknium, and others at the Nous Research team for their invaluable support throughout this project. I would also like to thank Together AI and Agentica for their immensely helpful blog posts on DeepCoder-14B. Finally, thank you to Modal and Lambda for their generous support by providing me with credits. |