Datasets:
File size: 7,590 Bytes
58bd51a 82a31ad 58bd51a 82a31ad 58bd51a 82a31ad 58bd51a | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 | ---
language:
- ru
- en
license: other
license_name: mixed-research-use
size_categories:
- 1K<n<10K
task_categories:
- visual-question-answering
- question-answering
task_ids:
- visual-question-answering
pretty_name: RuChartQA
tags:
- chart-understanding
- visual-reasoning
- russian
- benchmark
- vlm-evaluation
configs:
- config_name: chartbasic
data_files: synthetic/chartbasic.jsonl
- config_name: chartreasoning
data_files: synthetic/chartreasoning.jsonl
- config_name: chartperception
data_files: synthetic/chartperception.jsonl
- config_name: chartreal
data_files: chartreal/data.jsonl
---
# RuChartQA
A Russian-language chart question answering benchmark for evaluating Vision-Language Models, with both **synthetic** and **real-world** evaluation sets.
## Dataset summary
| Split | Examples | Charts | Source |
|---|---|---|---|
| Synthetic ChartBasic | 360 | 90 (×4 variants) | Generated |
| Synthetic ChartReasoning | 480 | 120 (×4 variants) | Generated |
| Synthetic ChartPerception | 360 | 90 (×4 variants) | Generated |
| ChartReal | 242 QA | 96 charts | Rosstat, Bank of Russia (PDF) |
| **Total** | **1442 QA** | **396 unique charts** | |
The synthetic split has 4 variants per chart: `ru_image`, `en_image`, `ru_text` (text description instead of image), `en_text` — enabling controlled language and modality ablations. ChartReal is `ru_image` only.
## Why this benchmark
Most chart-QA benchmarks (ChartQA, PlotQA, FigureQA) are English-only. Existing Russian-language chart evaluation has been limited to translated subsets. This benchmark addresses two gaps:
1. **Language coverage.** Native Russian questions, Russian-language axis labels, captions, and currencies (₽).
2. **Real-world distribution shift.** Synthetic-only benchmarks systematically overestimate VLM performance on real-world graphs from government statistics and central bank publications. Our analysis (see `results/leaderboard.csv` and the [accompanying paper]) shows performance gaps of **+11 to +41 percentage points** between synthetic and real-world splits across three modern VLMs.
## Loading
```python
from datasets import load_dataset
# Real-world split (the one with the bigger story)
chartreal = load_dataset("romath/RuChartQA", "chartreal")
# Synthetic splits
chartbasic = load_dataset("romath/RuChartQA", "chartbasic")
chartreasoning = load_dataset("romath/RuChartQA", "chartreasoning")
chartperception = load_dataset("romath/RuChartQA", "chartperception")
```
## Schema
Each row has:
| Field | Type | Description |
|---|---|---|
| `example_id` | string | Unique identifier (e.g. `chartreal_007_q2_ru_image`) |
| `subdataset` | string | `ChartBasic`, `ChartReasoning`, `ChartPerception`, or `ChartReal` |
| `variant` | string | `ru_image`, `en_image`, `ru_text`, `en_text` |
| `language` | string | `ru` or `en` |
| `modality` | string | `image` or `text` |
| `chart_type` | string | `bar`, `line`, `mixed`, `pie` |
| `chart_id` | string | Chart identifier (multiple QA may share one chart) |
| `question_type` | string | `lookup`, `comparison`, `min`, `max`, `difference`, `conditional` |
| `question` | string | Natural-language question |
| `answer` | string | Gold answer |
| `answer_numeric` | float \| null | Numeric form if applicable (for tolerance scoring) |
| `answer_type` | string | `numeric` or `categorical` |
| `image_path` | string \| null | Relative path to PNG (for image variants) |
| `text_description` | string \| null | Text description of the chart (for text variants) |
## Evaluation
We provide a normalizer (`eval/normalize.py`) that handles:
- Numeric tolerance (5%, the ChartQA standard) with a year-as-numeric exception requiring exact match (1900–2100)
- Bidirectional substring matching for categorical answers (`gold ⊆ pred` or `pred ⊆ gold`), disabled when gold contains compound markers (`и`, `or`, `,`)
- Lower/strip/punctuation normalization
Minimal example:
```bash
python3 eval/eval_example.py predictions.jsonl chartreal/data.jsonl
```
A prediction file is JSONL with `{"example_id": ..., "prediction_raw": "..."}` per line.
## Baselines
Predictions on ChartReal from four systems are included in `baselines/`:
| System | ChartReal Accuracy | Synthetic ru_image |
|---|---|---|
| Qwen3-VL 32B Instruct | **75.2%** | 86.3% |
| Gemini 2.5 Flash | 71.1% | 92.7% |
| Nemotron Nano 12B v2 VL | 45.9% | 86.7% |
| OCR + Llama 3.3 70B (text-only baseline) | 34.7% | n/a |
All gaps between systems on ChartReal are statistically significant (95% bootstrap CI) **except Qwen vs Gemini** (Δ=+4.1pp, CI [−1.2, +9.5], p=0.16). See `results/leaderboard.csv`.
## Construction
### Synthetic
Generated from category templates (cities, products, demographics, etc.) with controlled distributions over chart types (bar) and question types. Each chart was rendered in Russian and English; for each language, both an image and a text-description variant exist. This 4-way structure allows clean ablations of language and modality effects.
### ChartReal
Charts were extracted from public PDF reports of:
- **Rosstat** (Russian Federal State Statistics Service) — annual and monthly statistical bulletins
- **Bank of Russia (CBR)** — financial stability reports, monetary policy commentary
Each chart received 1–4 questions covering different reasoning types. Charts span four types (bar, line, mixed, pie) with realistic noise: small fonts, dense legends, multi-axis scales, and stylistic conventions specific to Russian government publications.
## Licenses
This dataset uses **mixed licensing**:
- **Code** (`eval/normalize.py`, `eval/eval_example.py`): Apache 2.0
- **Synthetic QA + images** (`synthetic/`): CC-BY 4.0 — author's original work
- **ChartReal QA annotations** (`chartreal/data.jsonl`): CC-BY 4.0 — author's original annotations
- **ChartReal images** (`chartreal/images/`): **research use only, original copyright preserved**. These are derivative works (PNG renderings of pages from public-domain government PDFs). Original publishers (Rosstat, Bank of Russia) retain copyright on the visual material. Re-use beyond academic research may require permission from the original publishers.
By using the `chartreal/images/` portion, you agree to:
1. Use it only for academic / non-commercial research
2. Cite both this dataset and the original publisher
3. Not redistribute the images independently of the QA annotations
## Citation
```bibtex
@dataset{ruchartqa_2026,
title = {RuChartQA: A Russian-Language Chart Question Answering Benchmark with Synthetic and Real-World Splits},
author = {Roman <last name>},
year = {2026},
url = {https://huggingface.co/datasets/romath/RuChartQA},
note = {HSE Bachelor's thesis}
}
```
## Limitations
- **ChartReal is image-only.** A `text_description` variant for real-world charts is not provided — automatic transcription of complex line/mixed charts to faithful text without losing information turned out to be infeasible in practice.
- **Bar-bias in synthetic.** All synthetic charts are bar-type. Comparison fairness across chart types should use the bar-only subset of ChartReal (n=67) — see `results/leaderboard.csv`.
- **Answer normalizer judgement calls.** A small number of answers (≤2pp of total) are influenced by language-drift conventions: yes/no in English vs Russian, Roman vs Cyrillic month numerals. We chose conservative scoring (mismatch counted as wrong); reasonable alternatives exist.
## Contact
Questions, errata, or contributions: [your email or GitHub username].
|