| --- |
| license: apache-2.0 |
| base_model: Qwen/Qwen2.5-Math-7B |
| datasets: |
| - kfdong/STP |
| - internlm/Lean-Workbook |
| - AI-MO/NuminaMath-LEAN |
| language: |
| - en |
| tags: |
| - theorem-proving |
| - automated-theorem-proving |
| - lean4 |
| - mathematical-reasoning |
| - llm |
| - qwen2.5-math |
| pipeline_tag: text-generation |
| --- |
| |
|
|
| # SEG-ATP: Segment-Level Learning for LLM-Based Theorem Proving |
|
|
| This repository contains the policy model used in the paper **"Rethinking Supervision Granularity: Segment-Level Learning for LLM-Based Theorem Proving"**. |
|
|
| ## Model |
|
|
| The model is fully fine-tuned from [Qwen2.5-Math-7B](https://huggingface.co/Qwen/Qwen2.5-Math-7B) using segment-level supervision constructed from Lean 4 proof trajectories. |
|
|
| ## Training Data |
|
|
| The model is trained using proof trajectories derived from STP, LeanWorkbook, and NuminaMath-LEAN. |
|
|
| ## Usage |
|
|
| Please refer to the GitHub repository for preprocessing, inference, and evaluation details: |
|
|
| https://github.com/NJUDeepEngine/SEG-ATP |
|
|
| ## Citation |
|
|
| ```bibtex |
| @article{xu2026rethinking, |
| title={Rethinking Supervision Granularity: Segment-Level Learning for LLM-Based Theorem Proving}, |
| author={Xu, Shuo and Zhang, Jiakun and Lai, Junyu and Cao, Chun and Xu, Jingwei}, |
| journal={arXiv preprint}, |
| year={2026} |
| } |