Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
Size:
< 1K
ArXiv:
FRIEDA / README.md
YOO-uN-ee's picture
Update README.md
8b25237 verified
metadata
task_categories:
  - visual-question-answering
language:
  - en
size_categories:
  - n<1K
configs:
  - config_name: default
    data_files:
      - split: data
        path: data/data-*
dataset_info:
  features:
    - name: question_ref
      dtype: string
    - name: images
      list: string
    - name: question_text
      dtype: string
    - name: expected_answer
      dtype: string
    - name: map_count
      dtype: string
    - name: spatial_relationship
      dtype: string
    - name: answer_type
      dtype: string
    - name: domain
      dtype: string
    - name: map_elements
      list: string
    - name: context_images
      list: string
  splits:
    - name: data
      num_bytes: 576010
      num_examples: 500
  download_size: 120231
  dataset_size: 576010
pretty_name: FRIEDA

arXiv Website Code

FRIEDA is a multimodal benchmark for open-ended cartographic reasoning over real-world map images.
Each example pairs reference maps (and optional contextual maps) with a natural-language question and a reference answer. The benchmark targets common GIS relation types (i.e., topological, metric, directional) and includes questions that require multi-step reasoning and cross-map grounding.

Dataset Summary

  • Modality: image + text
  • # Examples: 500
  • Input: map image(s) + question text
  • Output: expected answer (textual)
  • Metadata: map_count, domain, relationship type, map elements

Languages

The dataset questions and answers are in English.


How to use it

from datasets import load_dataset

# Full dataset (split name = "data")
ds = load_dataset("knowledge-computing/FRIEDA", split="data")
print(ds[0].keys())
print(ds[0]["question_text"])    # Actual question being asked
print(ds[0]["images"])           # List of string paths to images (e.g., "images/...png")
print(ds[0]["context_images"])   # List of string paths to contextual images