---
task_categories:
- image-text-to-text
language:
- en
tags:
- visual-reasoning
- symbolic-reasoning
- math
- matchstick-puzzles
- benchmark
---
# MathSticks: A Benchmark for Visual Symbolic Compositional Reasoning with Matchstick Puzzles
**Paper:** [MathSticks: A Benchmark for Visual Symbolic Compositional Reasoning with Matchstick Puzzles](https://huggingface.co/papers/2510.00483)
**Code:** [https://github.com/Yuheng2000/MathSticks](https://github.com/Yuheng2000/MathSticks)
## Overview
MathSticks is a benchmark for Visual Symbolic Compositional Reasoning (VSCR) that unifies visual perception, symbolic manipulation, and arithmetic consistency. Each task presents an incorrect matchstick equation in a seven-segment style. The goal is to move exactly one or two sticks—under strict stick-conservation and digit-legibility constraints—to make the equation mathematically correct.
- Two evaluation regimes:
- Text-guided: the equation string is provided to the model.
- Pure-visual: only the rendered puzzle image is provided.
- Systematic coverage: digit scale (Levels 1–4), move complexity (1/2 sticks), solution multiplicity, and operator variation.
- Scale: 1.4M generated instances; a curated test set of 400 items is released.
Example: input puzzle, reasoning trace, and move-format predictions.
Evaluations across 14 VLMs reveal substantial limitations: closed-source models succeed only on simple cases, open-source models fail in the pure-visual regime, while humans exceed 90% accuracy. These results establish MathSticks as a rigorous, diagnostic testbed for advancing compositional reasoning across vision and symbols.
## Task Definition
- Input: a rendered image of an incorrect equation composed of matchsticks (seven-segment digits). Optionally, a text equation string (text-guided regime).
- Constraints: move one or two sticks only; no addition/removal; preserve digit legibility.
- Objective: reach a valid arithmetic equation (addition or subtraction). Operator flips may be required by the minimal-move constraint in some cases.
- Output format: a boxed sequence of Move operations, e.g. `\boxed{Move(A0, C3)}` or `\boxed{Move(A0, C3), Move(E1, F4)}`.
## Data Format (Benchmark JSONL)
Each line is one sample with the following fields:
- `id` (string): unique sample identifier, e.g., `"00075585"`.
- `level` (int): difficulty level (1–4) indicating digit scale.
- `image` (string): image path relative to repo root, e.g., `level1/00075585_8-9=3.png`.
- `problem` (string): the displayed (incorrect) equation string, e.g., `8-9=3`.
- `solution_num` (list[int, int]): counts of solvable solutions by move budget `[one_move_count, two_move_count]`.
- `mode_1_solution` (list): list of one-move solutions. Empty when `solution_num[0] == 0`.
- `mode_2_solution` (list): list of two-move solutions. Each item has:
- `solution` (string): corrected equation (e.g., `"8 - 6 = 2"`).
- `moves` (list[string]): standardized move format strings, e.g., `["Move(B2, B5)", "Move(C3, C5)"]`.
- `option_answer` (object): order-invariant representation of moves, for robust parsing:
- `mode_1` (list): each one-move answer as `{ "pick": [from_label], "place": [to_label] }`.
- `mode_2` (list): each two-move answer as `{ "pick": [from_label_1, from_label_2], "place": [to_label_1, to_label_2] }`.
Example:
```json
{
"id": "00075585",
"level": 1,
"problem": "8-9=3",
"image": "level1/00075585_8-9=3.png",
"solution_num": [0, 4],
"mode_1_solution": [],
"mode_2_solution": [
{"solution": "8 - 6 = 2", "moves": ["Move(B2, B5)", "Move(C3, C5)"]},
{"solution": "9 - 9 = 0", "moves": ["Move(A5, C5)", "Move(C0, C6)"]},
{"solution": "6 + 3 = 9", "moves": ["Move(A2, G0)", "Move(B6, C6)"]},
{"solution": "9 - 0 = 9", "moves": ["Move(A5, B5)", "Move(B0, C6)"]}
],
"option_answer": {
"mode_1": [],
"mode_2": [
{"pick": ["B2", "C3"], "place": ["B5", "C5"]},
{"pick": ["A5", "C0"], "place": ["C5", "C6"]},
{"pick": ["A2", "B6"], "place": ["G0", "C6"]},
{"pick": ["A5", "B0"], "place": ["B5", "C6"]}
]
}
}
```
Notes:
- Pure-visual regime uses only `image` for model input; text-guided may also use `problem`.
- The string move format is strict for parsing; `option_answer` provides an order-invariant equivalent when needed.
## Sample Usage
### 1) Run evaluation
```bash
python eval.py \
--input MathSticks_bench_400.jsonl \
--image-dir ./image \
--output predictions.jsonl
```
### 2) Score predictions
```bash
python cal_score.py \
--pred predictions.jsonl \
--label MathSticks_bench_400.jsonl \
--output score.json
```
### 3) Generate data (research use only)
```bash
python match_gen_flt.py
```
This enumerates positional encodings and writes the discovered solvable cases to `match_gen.jsonl`. It is computationally intensive and may take a long time.
## Citation
If you find MathSticks useful, please cite the paper:
```bibtex
@article{mathsticks2025,
title = {MathSticks: A Benchmark for Visual Symbolic Compositional Reasoning with Matchstick Puzzles},
author = {Anonymous},
journal = {arXiv preprint arXiv:TODO},
year = {2025},
url = {https://huggingface.co/papers/2510.00483}
}
```