|
|
--- |
|
|
license: apache-2.0 |
|
|
language: |
|
|
- zh |
|
|
--- |
|
|
# QZhou-Flowchart-QA-Benchmark: Real-World Flowchart Understanding Benchmark |
|
|
|
|
|
## Overview |
|
|
|
|
|
While the open-source community has various chart and document benchmarks, there is no specialized evaluation set for flowchart understanding. **QZhou-Flowchart-QA-Benchmark** fills this gap by providing a dedicated benchmark to effectively assess multimodal models' flowchart comprehension abilities. |
|
|
|
|
|
## Dataset Composition |
|
|
|
|
|
### Part 1: Web-Collected Real-World Flowcharts (Public) |
|
|
|
|
|
Manually curated flowcharts from image search engines, covering actual deployment scenarios including: |
|
|
- Government services and administrative processes |
|
|
- Banking and financial operations |
|
|
- Campus management systems |
|
|
- Daily office workflows |
|
|
- Financial processing procedures |
|
|
|
|
|
**Quality Diversity:** We deliberately control the distribution of image resolution and clarity, introducing varying degrees of blur and size differences to better reflect real-world application environments. |
|
|
|
|
|
**Annotation:** All questions and answers are carefully labeled and verified by human annotators. |
|
|
|
|
|
### Part 2: Enterprise Office Flowcharts (Coming Soon) |
|
|
|
|
|
Real flowcharts from production office environments, including: |
|
|
- HR management processes |
|
|
- Financial reimbursement workflows |
|
|
- Internal approval procedures |
|
|
|
|
|
*Note: This portion is currently undergoing anonymization and will be released in a future update.* |
|
|
|
|
|
## Question Diversity |
|
|
|
|
|
FlowchartBench ensures comprehensive query coverage, considering various questioning angles and possibilities: |
|
|
- Upstream and downstream node queries |
|
|
- Conditional branch reasoning |
|
|
- Path analysis and node relationships |
|
|
- Structural understanding |
|
|
- Spatial reasoning with X/Y axes |
|
|
|
|
|
## Performance Leaderboard |
|
|
|
|
|
State-of-the-art results on QZhou-Flowchart-QA-Benchmark: |
|
|
|
|
|
| Model | QZhou-Flowchart-QA-Benchmark (%) | |
|
|
|-------|-------------------| |
|
|
| **QZhou-Flowchart-VL-32B (Ours)** | **87.83** | |
|
|
| Qwen3-VL-Plus-Thinking (235B) | 86.09 | |
|
|
| Gemini-2.5-Pro | 84.42 | |
|
|
| doubao-seed-1-6 | 83.83 | |
|
|
| GPT-5 | 79.29 | |
|
|
| GLM-4.5V | 75.97 | |
|
|
| Qwen2.5-VL-32B | 73.90 | |
|
|
|
|
|
### Comparison with Base Model |
|
|
|
|
|
| Model | MMMU | CMMU | MathVista | DocVQA | QZhou-Flowchart-QA-Benchmark | |
|
|
|-------|------|------|-----------|--------|----------------| |
|
|
| Qwen2.5-VL-32B | 66.67 | 76.38 | 74.20 | 93.96 | 73.90 | |
|
|
| **QZhou-Flowchart-VL-32B** | **67.78** | **76.46** | **76.50** | 93.87 | **87.83** | |
|
|
|
|
|
## Usage |
|
|
|
|
|
```python |
|
|
from datasets import load_dataset |
|
|
|
|
|
# Load benchmark |
|
|
benchmark = load_dataset("Kingsoft-LLM/QZhou-Flowchart-QA-Benchmark", split="test") |
|
|
|
|
|
# Evaluate your model |
|
|
for sample in benchmark: |
|
|
prediction = model.predict(sample['image'], sample['question']) |
|
|
accuracy = evaluate(prediction, sample['answer']) |
|
|
``` |
|
|
|
|
|
## Evaluation Protocol |
|
|
|
|
|
- **Answer Matching:** Two evaluation methods based on question type: |
|
|
- **Exact Match:** For multiple-choice questions, direct comparison with ground truth |
|
|
- **Normalized Edit Distance:** For open-ended questions, score calculated as `1 - (edit_distance / max_length)` |
|
|
- **Metrics:** Overall accuracy, breakdown by question type, domain, and complexity level |
|
|
- **Submission:** Open a GitHub issue with your model predictions to be added to the leaderboard |
|
|
|
|
|
## Key Features |
|
|
|
|
|
β
**Real-world scenarios** - Flowcharts from actual deployments |
|
|
β
**Manual annotation** - Human-verified questions and answers |
|
|
β
**Quality diversity** - Various resolutions, clarity levels, and sizes |
|
|
β
**Comprehensive coverage** - 20+ question types across multiple domains |
|
|
β
**Rigorous evaluation** - Standardized protocol for fair comparison |
|
|
|