Datasets:

Modalities:
Image
Text
Formats:
webdataset
Languages:
English
ArXiv:
Libraries:
Datasets
WebDataset
License:
saehyungl commited on
Commit
fa9fbb6
·
verified ·
1 Parent(s): 0c4e274

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +58 -0
README.md CHANGED
@@ -2,4 +2,62 @@
2
  license: other
3
  license_name: adobe-research-license
4
  license_link: LICENSE
 
 
5
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  license: other
3
  license_name: adobe-research-license
4
  license_link: LICENSE
5
+ language:
6
+ - en
7
  ---
8
+ # [ICML 2025] Toward Robust Hyper-Detailed Image Captioning: A Multiagent Approach and Dual Evaluation Metrics for Factuality and Coverage
9
+
10
+ This dataset is associated with the evaluation in our ICML 2025 paper, [Toward Robust Hyper-Detailed Image Captioning: A Multiagent Approach and Dual Evaluation Metrics for Factuality and Coverage](https://arxiv.org/abs/2412.15484).
11
+
12
+ ## Prerequisites
13
+ ### Packages
14
+ * openai>=1.14.1
15
+ * python-dotenv==1.0.1
16
+
17
+ ### Dataset download
18
+ ```dataset download
19
+ from huggingface_hub import hf_hub_download
20
+
21
+ local_path = hf_hub_download(
22
+ repo_id="saehyungl/CapMAS",
23
+ filename="images_capmas.tar.gz",
24
+ repo_type="dataset"
25
+ )
26
+ print("Downloaded to:", local_path)
27
+ ```
28
+ Or you can download it using [this URL](https://huggingface.co/datasets/saehyungl/CapMAS/resolve/main/images_capmas.tar.gz?download=true).
29
+ Our evaluation uses a subset of the [DOCCI](https://google.github.io/docci/) images.
30
+
31
+ ## Captioning
32
+ Please generate captions for the 1,000 downloaded images for captioning evaluation.
33
+ Summarize the generated captions into a dictionary where the key is the corresponding image file name, and save it as a .json file.
34
+ ```captions file
35
+ {
36
+ "aar_test_04600.jpg": <caption_aar_test_04600>,
37
+ "aar_test_04601.jpg": <caption_aar_test_04601>,
38
+ ...
39
+ "test_00599.json": <caption_test_00599>,
40
+ }
41
+ ```
42
+ You may refer to the [sample captions](https://github.com/david-yoon/CapMAS_private/blob/main/sample_captions/llava1.6-vicuna_llama3_th1.0/captions_final.json) for guidance.
43
+
44
+ ## Evaluation
45
+ Please visit our [GitHub repository](https://github.com/david-yoon/CapMAS_private).
46
+ We provide the evaluation codes for the three metrics used in our paper: **Factuality**, **Coverage**, and **CLAIR** (Chan et al., EMNLP 2023). These evaluations rely on GPT-4o, so please fill in your OpenAI API key **OR** Azure OpenAI credentials in the `conf/gpt4o` file.
47
+ ### Factuality (ours)
48
+ ```factuality
49
+ python eval_factuality.py --image-dir <the image directory path> --captions-file <the caption .json file path>
50
+ ```
51
+ ### Coverage (ours)
52
+ ```coverage
53
+ python eval_coverage.py --vqa-dir data/COVERAGE_TEST_VQA --captions-file <the caption .json file path>
54
+ ```
55
+ ### CLAIR
56
+ ```clair
57
+ python eval_clair.py --captions-file <the caption .json file path>
58
+ ```
59
+
60
+ ## References
61
+ 1. [DOCCI (Onoe et al., ECCV 2024)](https://google.github.io/docci/#downloads)
62
+ 2. [ImageInWords (Garg et al., EMNLP 2024)](https://github.com/google/imageinwords)
63
+ 3. [CLAIR (Chan et al., EMNLP 2023)](https://github.com/davidmchan/clair)