nezhazheng's picture
Update README.md
b5efedf verified
---
license: apache-2.0
language:
- zh
---
# QZhou-Flowchart-QA-Benchmark: Real-World Flowchart Understanding Benchmark
## Overview
While the open-source community has various chart and document benchmarks, there is no specialized evaluation set for flowchart understanding. **QZhou-Flowchart-QA-Benchmark** fills this gap by providing a dedicated benchmark to effectively assess multimodal models' flowchart comprehension abilities.
## Dataset Composition
### Part 1: Web-Collected Real-World Flowcharts (Public)
Manually curated flowcharts from image search engines, covering actual deployment scenarios including:
- Government services and administrative processes
- Banking and financial operations
- Campus management systems
- Daily office workflows
- Financial processing procedures
**Quality Diversity:** We deliberately control the distribution of image resolution and clarity, introducing varying degrees of blur and size differences to better reflect real-world application environments.
**Annotation:** All questions and answers are carefully labeled and verified by human annotators.
### Part 2: Enterprise Office Flowcharts (Coming Soon)
Real flowcharts from production office environments, including:
- HR management processes
- Financial reimbursement workflows
- Internal approval procedures
*Note: This portion is currently undergoing anonymization and will be released in a future update.*
## Question Diversity
FlowchartBench ensures comprehensive query coverage, considering various questioning angles and possibilities:
- Upstream and downstream node queries
- Conditional branch reasoning
- Path analysis and node relationships
- Structural understanding
- Spatial reasoning with X/Y axes
## Performance Leaderboard
State-of-the-art results on QZhou-Flowchart-QA-Benchmark:
| Model | QZhou-Flowchart-QA-Benchmark (%) |
|-------|-------------------|
| **QZhou-Flowchart-VL-32B (Ours)** | **87.83** |
| Qwen3-VL-Plus-Thinking (235B) | 86.09 |
| Gemini-2.5-Pro | 84.42 |
| doubao-seed-1-6 | 83.83 |
| GPT-5 | 79.29 |
| GLM-4.5V | 75.97 |
| Qwen2.5-VL-32B | 73.90 |
### Comparison with Base Model
| Model | MMMU | CMMU | MathVista | DocVQA | QZhou-Flowchart-QA-Benchmark |
|-------|------|------|-----------|--------|----------------|
| Qwen2.5-VL-32B | 66.67 | 76.38 | 74.20 | 93.96 | 73.90 |
| **QZhou-Flowchart-VL-32B** | **67.78** | **76.46** | **76.50** | 93.87 | **87.83** |
## Usage
```python
from datasets import load_dataset
# Load benchmark
benchmark = load_dataset("Kingsoft-LLM/QZhou-Flowchart-QA-Benchmark", split="test")
# Evaluate your model
for sample in benchmark:
prediction = model.predict(sample['image'], sample['question'])
accuracy = evaluate(prediction, sample['answer'])
```
## Evaluation Protocol
- **Answer Matching:** Two evaluation methods based on question type:
- **Exact Match:** For multiple-choice questions, direct comparison with ground truth
- **Normalized Edit Distance:** For open-ended questions, score calculated as `1 - (edit_distance / max_length)`
- **Metrics:** Overall accuracy, breakdown by question type, domain, and complexity level
- **Submission:** Open a GitHub issue with your model predictions to be added to the leaderboard
## Key Features
βœ… **Real-world scenarios** - Flowcharts from actual deployments
βœ… **Manual annotation** - Human-verified questions and answers
βœ… **Quality diversity** - Various resolutions, clarity levels, and sizes
βœ… **Comprehensive coverage** - 20+ question types across multiple domains
βœ… **Rigorous evaluation** - Standardized protocol for fair comparison