IndicVisionBench / README.md
AtharvImmverse's picture
Duplicate from krutrim-ai-labs/IndicVisionBench
1804ff6
metadata
dataset_info:
  - config_name: mmt
    features:
      - name: id
        dtype: string
      - name: image
        dtype: image
      - name: topic
        dtype: string
      - name: State/UT
        dtype: string
      - name: English
        dtype: string
      - name: Hindi
        dtype: string
      - name: Bengali
        dtype: string
      - name: Gujarati
        dtype: string
      - name: Kannada
        dtype: string
      - name: Malayalam
        dtype: string
      - name: Marathi
        dtype: string
      - name: Odia
        dtype: string
      - name: Punjabi
        dtype: string
      - name: Tamil
        dtype: string
      - name: Telugu
        dtype: string
      - name: source_url
        dtype: string
    splits:
      - name: test
        num_bytes: 14424797
        num_examples: 106
    download_size: 13255747
    dataset_size: 14424797
  - config_name: ocr
    features:
      - name: id
        dtype: string
      - name: image
        dtype: image
      - name: text
        dtype: string
      - name: language
        dtype: string
      - name: page_url
        dtype: string
    splits:
      - name: test
        num_bytes: 614014454
        num_examples: 876
    download_size: 612223184
    dataset_size: 614014454
  - config_name: vqa_en
    features:
      - name: id
        dtype: string
      - name: image
        dtype: image
      - name: topic
        dtype: string
      - name: State/UT
        dtype: string
      - name: language
        dtype: string
      - name: short_q1
        dtype: string
      - name: short_a1
        dtype: string
      - name: short_q2
        dtype: string
      - name: short_a2
        dtype: string
      - name: mcq
        dtype: string
      - name: mcq_a
        dtype: string
      - name: mcq_opt1
        dtype: string
      - name: mcq_opt2
        dtype: string
      - name: mcq_opt3
        dtype: string
      - name: mcq_opt4
        dtype: string
      - name: true_false_q
        dtype: string
      - name: true_false_a
        dtype: string
      - name: long_q
        dtype: string
      - name: long_a
        dtype: string
      - name: adversarial_question
        dtype: string
      - name: adversarial_answer
        dtype: string
      - name: source_url
        dtype: string
    splits:
      - name: test
        num_bytes: 1131332865
        num_examples: 4117
    download_size: 1127187152
    dataset_size: 1131332865
  - config_name: vqa_indic
    features:
      - name: id
        dtype: string
      - name: image
        dtype: image
      - name: topic
        dtype: string
      - name: State/UT
        dtype: string
      - name: language
        dtype: string
      - name: short_q1
        dtype: string
      - name: short_a1
        dtype: string
      - name: short_q2
        dtype: string
      - name: short_a2
        dtype: string
      - name: mcq
        dtype: string
      - name: mcq_a
        dtype: string
      - name: mcq_opt1
        dtype: string
      - name: mcq_opt2
        dtype: string
      - name: mcq_opt3
        dtype: string
      - name: mcq_opt4
        dtype: string
      - name: true_false_q
        dtype: string
      - name: true_false_a
        dtype: string
      - name: long_q
        dtype: string
      - name: long_a
        dtype: string
      - name: adversarial_question
        dtype: string
      - name: adversarial_answer
        dtype: string
      - name: source_url
        dtype: string
    splits:
      - name: test
        num_bytes: 276711951
        num_examples: 1007
    download_size: 273419974
    dataset_size: 276711951
  - config_name: vqa_parallel
    features:
      - name: id
        dtype: string
      - name: image
        dtype: image
      - name: topic
        dtype: string
      - name: State/UT
        dtype: string
      - name: language
        dtype: string
      - name: short_q1
        dtype: string
      - name: short_a1
        dtype: string
      - name: short_q2
        dtype: string
      - name: short_a2
        dtype: string
      - name: mcq
        dtype: string
      - name: mcq_a
        dtype: string
      - name: mcq_opt1
        dtype: string
      - name: mcq_opt2
        dtype: string
      - name: mcq_opt3
        dtype: string
      - name: mcq_opt4
        dtype: string
      - name: true_false_q
        dtype: string
      - name: true_false_a
        dtype: string
      - name: long_q
        dtype: string
      - name: long_a
        dtype: string
      - name: adversarial_question
        dtype: string
      - name: adversarial_answer
        dtype: string
      - name: source_url
        dtype: string
    splits:
      - name: test
        num_bytes: 324650384
        num_examples: 1166
    download_size: 321701661
    dataset_size: 324650384
configs:
  - config_name: mmt
    data_files:
      - split: test
        path: mmt/test-*
  - config_name: ocr
    data_files:
      - split: test
        path: ocr/test-*
  - config_name: vqa_en
    data_files:
      - split: test
        path: vqa_en/test-*
  - config_name: vqa_indic
    data_files:
      - split: test
        path: vqa_indic/test-*
  - config_name: vqa_parallel
    data_files:
      - split: test
        path: vqa_parallel/test-*
task_categories:
  - visual-question-answering
language:
  - en
  - hi
  - ta
  - te
  - ml
  - mr
  - gu
  - pa
  - or
  - kn
  - bn
tags:
  - vision
  - ocr
  - vqa
  - indic
  - benchmark
  - cultural
  - mmt
  - multimodal
size_categories:
  - 10K<n<100K

IndicVisionBench

ICLR 2026 arXiv IndicVisionBench-Github

This repository contains the dataset for IndicVisionBench, introduced in

“IndicVisionBench: Benchmarking Cultural and Multilingual Understanding in VLMs”
📄 arXiv:2511.04727
🏛️ Accepted at ICLR 2026
🔗 OpenReview: https://openreview.net/forum?id=LmJoLn04iL

IndicVisionBench is a culturally grounded, multilingual vision-language benchmark designed to evaluate Vision–Language Models (VLMs) on visual understanding tasks in the Indian context. The benchmark focuses on:

  • Multilingual Visual Question Answering (VQA)
  • Culturally-aware reasoning
  • Adversarial robustness
  • Parallel cross-lingual consistency
  • Optical Character Recognition (OCR) in Indic scripts
  • Multimodal Machine Translation (MMT)

Unlike generic VQA datasets, IndicVisionBench emphasizes Indian cultural context, regional diversity, and Indic language coverage, enabling systematic evaluation of multilingual and culturally-aware VLMs.


Languages Covered

  • English
  • Hindi
  • Tamil
  • Telugu
  • Malayalam
  • Marathi
  • Gujarati
  • Punjabi
  • Odia
  • Kannada
  • Bengali

Benchmark Overview

IndicVisionBench consists of five main configurations:

Config Task #Images Description
mmt Multimodal Machine Translation 106 Image-grounded translations across Indic languages
ocr Optical Character Recognition 876 OCR in multiple Indic scripts
vqa_en Visual Question Answering 4,117 Culturally grounded VQA in English
vqa_indic Visual Question Answering 1,007 Culturally grounded VQA in Indic languages
vqa_parallel Visual Question Answering 1,166 Same QA pairs across multiple languages for cross-lingual consistency
  • Total images across all configs: 4993
  • Total questions across VQA En, Indic and Parallel: (4117 + 1007 + 1166)*6 = 37,740

Subset Descriptions

1️⃣ Multimodal Machine Translation (mmt)

Image-grounded translation benchmark with aligned captions across multiple Indic languages.

Features:

  • image
  • topic
  • State/UT
  • Parallel captions in 11 languages
  • source_url

This subset evaluates:

  • Cultural terminology consistency
  • Visual grounding in translation

2️⃣ Optical Character Recognition (ocr)

OCR dataset consisting of scanned pages in Indic scripts from Wikisource.

Features:

  • image
  • text
  • language
  • page_url

This subset evaluates OCR capabitilies on Indic scripts/languages.

3️⃣ English VQA (vqa_en)

Culturally grounded VQA in English.

Each example includes:

  • 2 short-answer questions
  • 1 multiple-choice question (4 options)
  • 1 true/false question
  • 1 long-form reasoning question
  • 1 adversarial question
  • Metadata: topic, language, State/UT, 'source_url'

This subset evaluates:

  • Object & scene understanding
  • Cultural knowledge
  • Fine-grained attribute recognition
  • Robustness to false assumptions in the adversarial questions

4️⃣ Indic VQA (vqa_indic)

Same VQA format as in vqa_en, but in Indic languages.

This subset evaluates:

  • Multilingual reasoning
  • Cultural alignment in local languages

5️⃣ Parallel VQA (vqa_parallel)

Same VQA format as in vqa_en. Parallel multilingual QA pairs for the same image.

This subset enables the study of

  • cross-lingual performance of VLMs across 11 languages (English and 10 Indic languages)
  • region-specific strengths or biases

Usage

All configurations can be loaded using datasets:

from datasets import load_dataset

# Example: load English VQA split
ds = load_dataset("krutrim-ai-labs/IndicVisionBench", "vqa_en")["test"]

print(ds[0])

The following five configurations/splits are present in the dataset:

  • mmt
  • ocr
  • vqa_en
  • vqa_indic
  • vqa_parallel

Images are stored directly within the dataset and loaded automatically by 🤗 Datasets.

Evaluation Dimensions

IndicVisionBench is designed to measure:

  • Scene & contextual understanding
  • Attribute detection
  • Cultural understanding
  • Bias & adversarial robustness
  • Cross-lingual consistency
  • OCR performance
  • Image-grounded translation capability

Code & Evaluation

The official inference and evaluation codebase for IndicVisionBench is available on GitHub.

GitHub Repository: https://github.com/ola-krutrim/IndicVisionBench

The repository provides the complete pipeline for running inference and reproducing benchmark results across all evaluation tracks.

The codebase includes:

  • End-to-end inference pipelines for Vision-Language Models (VLMs) and OCR systems
  • Modular wrappers enabling easy integration of API-based models and open-source models
  • Evaluation pipelines for all benchmark tasks:
    • OCR evaluation
    • Visual Question Answering (VQA)
      • Structured questions (MCQ, True/False)
      • Open-ended questions (short answer, long answer, adversarial)
    • Multimodal Machine Translation (MMT)
  • LLM-as-a-judge evaluation for open-ended VQA responses
  • Data generation scripts for constructing a similar multimodal benchmark.

Citation

If you use this dataset, please cite:

@inproceedings{faraz2026indicvisionbench,
  title={IndicVisionBench: Benchmarking Cultural and Multilingual Understanding in VLMs},
  author={Ali Faraz and Akash and Shaharukh Khan and Raja Kolla and Akshat Patidar and Suranjan Goswami and Abhinav Ravi and Chandra Khatri and Shubham Agarwal},
  booktitle={International Conference on Learning Representations (ICLR)},
  year={2026},
  url={https://openreview.net/forum?id=LmJoLn04iL}
}