Add model card for DeepSeekMath-7B-Caco with metadata and links

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +33 -0
README.md ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: transformers
4
+ pipeline_tag: text-generation
5
+ ---
6
+
7
+ # DeepSeekMath-7B-Caco
8
+
9
+ This repository hosts the **DeepSeekMath-7B-Caco** model, which is fine-tuned using the Caco framework. Caco (Code-Assisted Chain-of-ThOught) is a novel framework that automates the synthesis of high-quality, verifiable, and diverse instruction-Chain-of-Thought (CoT) reasoning data through code-driven augmentation. It leverages executable code steps to generate comprehensive and logically correct reasoning paths, enhancing the reasoning capabilities of Large Language Models (LLMs).
10
+
11
+ The **DeepSeekMath-7B-Caco** model is built upon the `deepseek-math-7b-base` and has been trained on the Caco-1.3M dataset, achieving strong competitive performance on mathematical reasoning benchmarks, as detailed in the research paper.
12
+
13
+ For more details on the Caco framework, its methodology, and experimental results, please refer to the official paper:
14
+ [**Scaling Code-Assisted Chain-of-Thoughts and Instructions for Model Reasoning**](https://huggingface.co/papers/2510.04081) (accepted by NeurIPS 2025).
15
+
16
+ ## Code and Further Information
17
+ The official implementation, detailed instructions for data generation, training, and evaluation, can be found on the GitHub repository:
18
+ [https://github.com/LHL3341/Caco](https://github.com/LHL3341/Caco)
19
+
20
+ You can also find the Caco dataset and other Caco-trained models within the Hugging Face collection:
21
+ [https://huggingface.co/collections/LHL3341/caco-68e0cb7b8a5f0071fac1f611](https://huggingface.co/collections/LHL3341/caco-68e0cb7b8a5f0071fac1f611)
22
+
23
+ ## Citation
24
+ If you find this work useful, please cite the paper:
25
+
26
+ ```bibtex
27
+ @article{caco,
28
+ title={Scaling Code-Assisted Chain-of-Thoughts and Instructions for Model Reasoning},
29
+ author={Honglin Lin and Qizhi Pei and Xin Gao and Zhuoshi Pan and Yu Li and Juntao Li and Conghui He and Lijun Wu},
30
+ journal={arXiv preprint arXiv:2510.04081},
31
+ year={2025}
32
+ }
33
+ ```