The viewer is disabled because this dataset repo requires arbitrary Python code execution. Please consider
removing the
loading script
and relying on
automated data support
(you can use
convert_to_parquet
from the datasets library). If this is not possible, please
open a discussion
for direct help.
HeuriGen
HeuriGen is a benchmark and agentic evaluation framework designed to rigorously assess Large Language Models (LLMs) on combinatorial optimization (CO) problems — a domain where success requires more than pattern recognition: it demands creative algorithm design, multi-step planning, tool use, and adaptive reasoning.
🧠 Motivation
While LLMs have shown impressive capabilities in coding and open-ended reasoning, existing benchmarks fall short:
- Objective benchmarks (e.g., HumanEval, AIME) are prone to saturation and fail to test creativity or multi-step reasoning.
- Subjective evaluations (e.g., Chatbot Arena) allow diverse outputs but often rely on noisy or superficial feedback.
To bridge this gap, HeuriGen introduces real-world CO tasks that:
- Feature well-defined objectives with expansive solution spaces.
- Require heuristic design, not just memorized answers.
- Enable quantitative and automated evaluation through code execution.
Problem Set
| Problem | Domain |
|---|---|
| Operator Scheduling | Electronic Design Automation |
| E-Graph Extraction | Compilers |
| Pickup and Delivery w/ Time Windoes | Logistics |
| Technology Mapping | Electronic Design Automation |
| Global Routing | Electronic Design Automation |
| Protein Sequence Design | Computational Biology |
| Airline Crew Pairing | Logistics |
| Pedigree | Computational Biology |
| Intra-Op Parallelism | Compilers |
- Downloads last month
- 104