| license: llama3.2 | |
| datasets: | |
| - LLM360/MegaMath | |
| language: | |
| - en | |
| pipeline_tag: text-generation | |
| library_name: transformers | |
| tags: | |
| - math | |
| - code | |
| - cot | |
| - pal | |
| # MegaMath-Llama-3.2-3B | |
| [Arxiv](https://arxiv.org/abs/2504.02807) | [Datasets](https://huggingface.co/datasets/LLM360/MegaMath) | |
| A proof-of-concept model train on [MegaMath](https://huggingface.co/datasets/LLM360/MegaMath) dataset, capable of both Chain-of-Thought and Program-Aided-Language problem solving. | |
|  | |
| ## Performance | |
|  | |
| ## Citation | |
| If you find our work useful, please cite | |
| ```bibtex | |
| @article{zhou2025megamath, | |
| title = {MegaMath: Pushing the Limits of Open Math Corpora}, | |
| author = {Zhou, Fan and Wang, Zengzhi and Ranjan, Nikhil and Cheng, Zhoujun and Tang, Liping and He, Guowei and Liu, Zhengzhong and Xing, Eric P.}, | |
| journal = {arXiv preprint arXiv:2504.02807}, | |
| year = {2025}, | |
| note = {Preprint} | |
| } | |
| ``` |