|
|
--- |
|
|
license: apache-2.0 |
|
|
task_categories: |
|
|
- text-generation |
|
|
tags: |
|
|
- code |
|
|
size_categories: |
|
|
- n<1K |
|
|
language: |
|
|
- en |
|
|
configs: |
|
|
- config_name: default |
|
|
data_files: |
|
|
- split: original |
|
|
path: "dataset.jsonl" |
|
|
- split: verified |
|
|
path: "dataset_verified.jsonl" |
|
|
--- |
|
|
|
|
|
# CP-Bench: A dataset for evaluating LLM-driven constraint modelling |
|
|
|
|
|
[](https://huggingface.co/spaces/kostis-init/CP-Bench-Leaderboard) |
|
|
|
|
|
This dataset is designed to facilitate the evaluation of LLM-based methods for translating natural language problem descriptions into accurate constraint specifications. It contains diverse combinatorial problems, and is sourced from various well-established sources from the Constraint Programming community. |
|
|
|
|
|
--- |
|
|
|
|
|
## Dataset Versions |
|
|
|
|
|
[//]: # (> **tl;dr:** Use `v1_verified`, it's the latest!) |
|
|
|
|
|
You will notice that the dataset contains various splits (which are not exactly splits, but rather different versions of the dataset): |
|
|
- `original`: The original dataset, which contains all problems initially designed, including those with known issues. More details about the issues can be found in the [changelog](cp_bench_changes.md). |
|
|
- `verified`: A stripped-down version of the original dataset that has been verified to contain complete problem specifications with matching ground-truth models. This version is recommended for use in evaluations. |
|
|
|
|
|
--- |
|
|
|
|
|
## 📊 Leaderboard |
|
|
|
|
|
You can easily submit your results and view the global leaderboard here: |
|
|
👉 **[CP-Bench Leaderboard](https://huggingface.co/spaces/kostis-init/CP-Bench-Leaderboard)** |
|
|
|
|
|
--- |
|
|
|
|
|
## Dataset Breakdown |
|
|
|
|
|
The dataset contains problems from the following sources: |
|
|
|
|
|
- `aplai_course`: Problems from the APLAI course of KU Leuven, 2023-2024. As modelled [here](https://github.com/kostis-init/CP-LLMs-ICL/tree/main/data/APLAI_course). |
|
|
- `cpmpy_examples`: Problems from the [CPMpy repository](https://github.com/CPMpy/cpmpy/tree/master/examples) |
|
|
- All included, except for the ones that require enumeration of all solutions (e.g. `solveAll`). |
|
|
- [`csplib`](https://www.csplib.org/Problems/) |
|
|
- For now, only the ones modelled in the [CPMpy repository](https://github.com/CPMpy/cpmpy/tree/master/examples/csplib) are included, and the ones modelled by [Hakan Kjellerstrand](http://www.hakank.org/cpmpy/). |
|
|
- `hakan_examples`: Models created by [Hakan Kjellerstrand](http://www.hakank.org/cpmpy/) |
|
|
- In progress with alphabetical order, excluding the following: |
|
|
- Those already modelled from other sources (e.g. aplai_course, cpmpy_examples, csplib) |
|
|
- Those that contain `solveAll` (counting solutions). |
|
|
- Global constraints tests, e.g. http://www.hakank.org/cpmpy/atmost_test.py |
|
|
|
|
|
--- |
|
|
|
|
|
## Diversity |
|
|
|
|
|
We attempted to include unique problems from different sources, in order to provide a diverse set of problems. |
|
|
However, as this was a manual process, there might be duplicates or similar problems. If you notice any issues, please let us know. |
|
|
|
|
|
--- |
|
|
|
|
|
## Citation |
|
|
|
|
|
If you found our work useful, please consider citing it: |
|
|
|
|
|
```bib |
|
|
@misc{michailidis2025cpbench, |
|
|
title={CP-Bench: Evaluating Large Language Models for Constraint Modelling}, |
|
|
author={Kostis Michailidis and Dimos Tsouros and Tias Guns}, |
|
|
year={2025}, |
|
|
eprint={2506.06052}, |
|
|
archivePrefix={arXiv}, |
|
|
primaryClass={cs.AI}, |
|
|
url={https://arxiv.org/abs/2506.06052}, |
|
|
} |
|
|
``` |