jaagli commited on
Commit
453e94f
Β·
verified Β·
1 Parent(s): c1b5421

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +118 -0
README.md ADDED
@@ -0,0 +1,118 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ task_categories:
3
+ - image-to-text
4
+ viewer: false
5
+ ---
6
+ # RAVENEA
7
+
8
+ [πŸ“ƒ PAPER](https://arxiv.org/abs/2505.14462) | [πŸ’» GITHUB](https://github.com/yfyuan01/RAVENEA)
9
+
10
+ **RAVENEA** is a multimodal benchmark designed to comprehensively evaluate the capabilities of VLMs in **cultural understanding through RAG**, introduced in [RAVENEA: A Benchmark for Multimodal Retrieval-Augmented Visual Culture Understanding](https://arxiv.org/abs/2505.14462).
11
+
12
+ It provides:
13
+
14
+ - **A large-scale cultural retrieval-generation corpus** featuring 1,868 culturally grounded images paired with over 10,000 **human-ranked** Wikipedia documents.
15
+
16
+ - **Two downstream tasks** for assessing culture-centric visual understanding (cVQA) and culture-informed image captioning (cIC).
17
+
18
+ - **Broad cross-cultural coverage spanning 8 countries and 11 categories**, including China, India, Indonesia, Korea, Mexico, Nigeria, Russia, and Spain. The benchmark encompasses a diverse taxonomic spectrum: Architecture, Cuisine, History, Art, Daily Life, Companies, Sports & Recreation, Transportation, Religion, Nature, and Tools.
19
+
20
+
21
+ ## Dataset Structure
22
+
23
+ The dataset is organized as follows:
24
+
25
+ ```
26
+ ravenea/
27
+ β”œβ”€β”€ images/ # Directory containing all images
28
+ β”œβ”€β”€ metadata_train.jsonl # Training split metadata
29
+ β”œβ”€β”€ metadata_val.jsonl # Validation split metadata
30
+ β”œβ”€β”€ metadata_test.jsonl # Test split metadata
31
+ β”œβ”€β”€ metadata.jsonl # Full metadata
32
+ β”œβ”€β”€ cic_downstream.jsonl # culture-informed image captioning task
33
+ β”œβ”€β”€ cvqa_downstream.jsonl # culture-centric visual question answering task
34
+ └── wiki_documents.jsonl # Corpus of Wikipedia articles for retrieval
35
+ ```
36
+
37
+ ## Schema
38
+
39
+ ### Metadata (`metadata_*.jsonl`)
40
+
41
+ Each line is a JSON object representing a data sample:
42
+
43
+ - `file_name`: Path to the image file (e.g., `./ravenea/images/ccub_101_China_38.jpg`).
44
+ - `country`: Country of origin for the cultural content.
45
+ - `task_type`: Task category (e.g., `cIC` for image captioning/QA).
46
+ - `category`: Broad cultural category (e.g., `Daily Life`).
47
+ - `human_captions`: Human-written caption describing the image.
48
+ - `questions`: List of questions associated with the image.
49
+ - `options`: Multiple-choice options for the questions.
50
+ - `answers`: Correct answers for the questions.
51
+ - `enwiki_ids`: List of relevant Wikipedia article IDs.
52
+ - `culture_relevance`: Score or indicator of cultural relevance.
53
+
54
+ ### Wikipedia Corpus (`wiki_documents.jsonl`)
55
+
56
+ Contains the knowledge base for retrieval:
57
+
58
+ - `id`: Unique identifier for the article (e.g., `enwiki/65457597`).
59
+ - `text`: Full text content of the Wikipedia article.
60
+ - `date_modified`: Last modification date of the article.
61
+
62
+ ## Usage
63
+
64
+ ### Download the Dataset
65
+
66
+ Please download the dataset then unzip it to the current directory.
67
+ ```python
68
+ from huggingface_hub import hf_hub_download
69
+
70
+ local_path = hf_hub_download(
71
+ repo_id="jaagli/ravenea",
72
+ filename="./ravenea.zip",
73
+ repo_type="dataset",
74
+ local_dir="./",
75
+ )
76
+ print(f"File downloaded to: {local_path}")
77
+ ```
78
+
79
+ ### Loading the Data
80
+ You can load the dataset using standard Python libraries.:
81
+
82
+ ```python
83
+ import json
84
+ from pathlib import Path
85
+
86
+ def load_jsonl(file_path):
87
+ data = []
88
+ with open(file_path, 'r', encoding='utf-8') as f:
89
+ for line in f:
90
+ data.append(json.loads(line))
91
+ return data
92
+
93
+ # Load metadata
94
+ train_data = load_jsonl("./ravenea/metadata_train.jsonl")
95
+ test_data = load_jsonl("./ravenea/metadata_test.jsonl")
96
+
97
+ # Load Wikipedia corpus
98
+ wiki_docs = load_jsonl("./ravenea/wiki_documents.jsonl")
99
+ doc_id_to_text = {doc['id']: doc['text'] for doc in wiki_docs}
100
+
101
+ # Example: Accessing a sample
102
+ sample = train_data[0]
103
+ print(f"Image: {sample['file_name']}")
104
+ print(f"Caption: {sample['human_captions']}")
105
+ print(f"Docs: {sample['enwiki_ids']}")
106
+ ```
107
+
108
+ ## BibTeX Citation
109
+ ```bibtex
110
+ @inproceedings{
111
+ li2026ravenea,
112
+ title={{RAVENEA}: A Benchmark for Multimodal Retrieval-Augmented Visual Culture Understanding},
113
+ author={Jiaang Li and Yifei Yuan and Wenyan Li and Mohammad Aliannejadi and Daniel Hershcovich and Anders S{\o}gaard and Ivan Vuli{\'c} and Wenxuan Zhang and Paul Pu Liang and Yang Deng and Serge Belongie},
114
+ booktitle={The Fourteenth International Conference on Learning Representations},
115
+ year={2026},
116
+ url={https://openreview.net/forum?id=4zAbkxQ23i}
117
+ }
118
+ ```