RuChartQA / README.md
romath's picture
Upload README.md with huggingface_hub
82a31ad verified
---
language:
- ru
- en
license: other
license_name: mixed-research-use
size_categories:
- 1K<n<10K
task_categories:
- visual-question-answering
- question-answering
task_ids:
- visual-question-answering
pretty_name: RuChartQA
tags:
- chart-understanding
- visual-reasoning
- russian
- benchmark
- vlm-evaluation
configs:
- config_name: chartbasic
data_files: synthetic/chartbasic.jsonl
- config_name: chartreasoning
data_files: synthetic/chartreasoning.jsonl
- config_name: chartperception
data_files: synthetic/chartperception.jsonl
- config_name: chartreal
data_files: chartreal/data.jsonl
---
# RuChartQA
A Russian-language chart question answering benchmark for evaluating Vision-Language Models, with both **synthetic** and **real-world** evaluation sets.
## Dataset summary
| Split | Examples | Charts | Source |
|---|---|---|---|
| Synthetic ChartBasic | 360 | 90 (×4 variants) | Generated |
| Synthetic ChartReasoning | 480 | 120 (×4 variants) | Generated |
| Synthetic ChartPerception | 360 | 90 (×4 variants) | Generated |
| ChartReal | 242 QA | 96 charts | Rosstat, Bank of Russia (PDF) |
| **Total** | **1442 QA** | **396 unique charts** | |
The synthetic split has 4 variants per chart: `ru_image`, `en_image`, `ru_text` (text description instead of image), `en_text` — enabling controlled language and modality ablations. ChartReal is `ru_image` only.
## Why this benchmark
Most chart-QA benchmarks (ChartQA, PlotQA, FigureQA) are English-only. Existing Russian-language chart evaluation has been limited to translated subsets. This benchmark addresses two gaps:
1. **Language coverage.** Native Russian questions, Russian-language axis labels, captions, and currencies (₽).
2. **Real-world distribution shift.** Synthetic-only benchmarks systematically overestimate VLM performance on real-world graphs from government statistics and central bank publications. Our analysis (see `results/leaderboard.csv` and the [accompanying paper]) shows performance gaps of **+11 to +41 percentage points** between synthetic and real-world splits across three modern VLMs.
## Loading
```python
from datasets import load_dataset
# Real-world split (the one with the bigger story)
chartreal = load_dataset("romath/RuChartQA", "chartreal")
# Synthetic splits
chartbasic = load_dataset("romath/RuChartQA", "chartbasic")
chartreasoning = load_dataset("romath/RuChartQA", "chartreasoning")
chartperception = load_dataset("romath/RuChartQA", "chartperception")
```
## Schema
Each row has:
| Field | Type | Description |
|---|---|---|
| `example_id` | string | Unique identifier (e.g. `chartreal_007_q2_ru_image`) |
| `subdataset` | string | `ChartBasic`, `ChartReasoning`, `ChartPerception`, or `ChartReal` |
| `variant` | string | `ru_image`, `en_image`, `ru_text`, `en_text` |
| `language` | string | `ru` or `en` |
| `modality` | string | `image` or `text` |
| `chart_type` | string | `bar`, `line`, `mixed`, `pie` |
| `chart_id` | string | Chart identifier (multiple QA may share one chart) |
| `question_type` | string | `lookup`, `comparison`, `min`, `max`, `difference`, `conditional` |
| `question` | string | Natural-language question |
| `answer` | string | Gold answer |
| `answer_numeric` | float \| null | Numeric form if applicable (for tolerance scoring) |
| `answer_type` | string | `numeric` or `categorical` |
| `image_path` | string \| null | Relative path to PNG (for image variants) |
| `text_description` | string \| null | Text description of the chart (for text variants) |
## Evaluation
We provide a normalizer (`eval/normalize.py`) that handles:
- Numeric tolerance (5%, the ChartQA standard) with a year-as-numeric exception requiring exact match (1900–2100)
- Bidirectional substring matching for categorical answers (`gold ⊆ pred` or `pred ⊆ gold`), disabled when gold contains compound markers (`и`, `or`, `,`)
- Lower/strip/punctuation normalization
Minimal example:
```bash
python3 eval/eval_example.py predictions.jsonl chartreal/data.jsonl
```
A prediction file is JSONL with `{"example_id": ..., "prediction_raw": "..."}` per line.
## Baselines
Predictions on ChartReal from four systems are included in `baselines/`:
| System | ChartReal Accuracy | Synthetic ru_image |
|---|---|---|
| Qwen3-VL 32B Instruct | **75.2%** | 86.3% |
| Gemini 2.5 Flash | 71.1% | 92.7% |
| Nemotron Nano 12B v2 VL | 45.9% | 86.7% |
| OCR + Llama 3.3 70B (text-only baseline) | 34.7% | n/a |
All gaps between systems on ChartReal are statistically significant (95% bootstrap CI) **except Qwen vs Gemini** (Δ=+4.1pp, CI [−1.2, +9.5], p=0.16). See `results/leaderboard.csv`.
## Construction
### Synthetic
Generated from category templates (cities, products, demographics, etc.) with controlled distributions over chart types (bar) and question types. Each chart was rendered in Russian and English; for each language, both an image and a text-description variant exist. This 4-way structure allows clean ablations of language and modality effects.
### ChartReal
Charts were extracted from public PDF reports of:
- **Rosstat** (Russian Federal State Statistics Service) — annual and monthly statistical bulletins
- **Bank of Russia (CBR)** — financial stability reports, monetary policy commentary
Each chart received 1–4 questions covering different reasoning types. Charts span four types (bar, line, mixed, pie) with realistic noise: small fonts, dense legends, multi-axis scales, and stylistic conventions specific to Russian government publications.
## Licenses
This dataset uses **mixed licensing**:
- **Code** (`eval/normalize.py`, `eval/eval_example.py`): Apache 2.0
- **Synthetic QA + images** (`synthetic/`): CC-BY 4.0 — author's original work
- **ChartReal QA annotations** (`chartreal/data.jsonl`): CC-BY 4.0 — author's original annotations
- **ChartReal images** (`chartreal/images/`): **research use only, original copyright preserved**. These are derivative works (PNG renderings of pages from public-domain government PDFs). Original publishers (Rosstat, Bank of Russia) retain copyright on the visual material. Re-use beyond academic research may require permission from the original publishers.
By using the `chartreal/images/` portion, you agree to:
1. Use it only for academic / non-commercial research
2. Cite both this dataset and the original publisher
3. Not redistribute the images independently of the QA annotations
## Citation
```bibtex
@dataset{ruchartqa_2026,
title = {RuChartQA: A Russian-Language Chart Question Answering Benchmark with Synthetic and Real-World Splits},
author = {Roman <last name>},
year = {2026},
url = {https://huggingface.co/datasets/romath/RuChartQA},
note = {HSE Bachelor's thesis}
}
```
## Limitations
- **ChartReal is image-only.** A `text_description` variant for real-world charts is not provided — automatic transcription of complex line/mixed charts to faithful text without losing information turned out to be infeasible in practice.
- **Bar-bias in synthetic.** All synthetic charts are bar-type. Comparison fairness across chart types should use the bar-only subset of ChartReal (n=67) — see `results/leaderboard.csv`.
- **Answer normalizer judgement calls.** A small number of answers (≤2pp of total) are influenced by language-drift conventions: yes/no in English vs Russian, Roman vs Cyrillic month numerals. We chose conservative scoring (mismatch counted as wrong); reasonable alternatives exist.
## Contact
Questions, errata, or contributions: [your email or GitHub username].