Datasets:

Modalities:
Text
Formats:
parquet
Size:
< 1K
ArXiv:
Libraries:
Datasets
pandas
License:
zzzzzhy commited on
Commit
45d199b
·
verified ·
1 Parent(s): a6addab

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -1
README.md CHANGED
@@ -27,4 +27,38 @@ configs:
27
  - "real.parquet"
28
  ---
29
 
30
- The proof of the seed problems and the evaluation scripts can be found at https://github.com/haoyuzhao123/LeanIneqComp
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27
  - "real.parquet"
28
  ---
29
 
30
+ <div align="center">
31
+ <h1> <a href="https://arxiv.org">Ineq-Comp: Benchmarking Human-Intuitive Compositional Reasoning in Automated Theorem Proving on Inequalities</a></h1>
32
+ </div>
33
+
34
+ <div align="center">
35
+
36
+ [![GitHub](https://img.shields.io/badge/GitHub-181717?style=flat&logo=github&logoColor=white)](https://github.com/haoyuzhao123/LeanIneqComp)
37
+ [![arXiv](https://img.shields.io/badge/arXiv-2505.12680-b31b1b.svg?style=flat)](https://arxiv.org/abs/2505.12680)
38
+ [![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](https://opensource.org/licenses/Apache-2.0)
39
+ [![Email me](https://img.shields.io/badge/Contact-6fcf97?logo=gmail)](mailto:thomaszhao1998@gmail.com)
40
+
41
+ </div>
42
+
43
+ ## Introduction
44
+
45
+ We introduce Ineq-Comp, a benchmark built from elementary inequalities through systematic transformations, including variable duplication, algebraic rewriting, and multi-step composition. Although these problems remain easy for humans, we find that most provers&mdash;including Goedel, STP, and Kimina-7B&mdash;struggle significantly. DeepSeek-Prover-V2 shows relative robustness&mdash;possibly because it is trained to decompose the problems into sub-problems&mdash;but still suffers a 20\% performance drop (pass@32). Strikingly, performance remains poor for all models even when formal proofs of the constituent parts are provided in context, revealing that the source of weakness is indeed in compositional reasoning. Our results expose a persisting gap between the generalization behavior of current AI provers and human mathematical intuition.
46
+
47
+ ## Quick Start
48
+
49
+ The proof of the seed problems and the evaluation scripts can be found at https://github.com/haoyuzhao123/LeanIneqComp
50
+
51
+ ## Ineq-Comp Benchmark
52
+
53
+ We provide 5 splits: seed, type1, type2, mix, and real. For seed, type1, and type2 splits, each contains 75 problems. mix split contains 100 problems generated by Ineq-Mix, and real split contains 50 real-world inequality problems. Please refer to the github repo for more fine-grained splits.
54
+
55
+ ## Citation
56
+
57
+ ```{bibtex}
58
+ @article{zhao2025ineq,
59
+ title={Ineq-Comp: Benchmarking Human-Intuitive Compositional Reasoning in Automated Theorem Proving on Inequalities},
60
+ author={Zhao, Haoyu and Geng, Yihan and Tang, Shange and Lin, Yong and Lyu, Bohan and Lin, Hongzhou and Jin, Chi and Arora, Sanjeev},
61
+ journal={arXiv preprint arXiv:2505.12680},
62
+ year={2025}
63
+ }
64
+ ```