Datasets:

Modalities:
Image
Text
Formats:
webdataset
Languages:
English
ArXiv:
Libraries:
Datasets
WebDataset
License:

Improve dataset card: Add task category, tags, and code block formatting

#2
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +18 -10
README.md CHANGED
@@ -1,10 +1,18 @@
1
  ---
 
 
2
  license: other
3
  license_name: adobe-research-license
4
  license_link: LICENSE
5
- language:
6
- - en
 
 
 
 
 
7
  ---
 
8
  # [ICML 2025] Toward Robust Hyper-Detailed Image Captioning: A Multiagent Approach and Dual Evaluation Metrics for Factuality and Coverage
9
 
10
  This dataset is associated with the evaluation in our ICML 2025 paper, [Toward Robust Hyper-Detailed Image Captioning: A Multiagent Approach and Dual Evaluation Metrics for Factuality and Coverage](https://arxiv.org/abs/2412.15484).
@@ -15,7 +23,7 @@ This dataset is associated with the evaluation in our ICML 2025 paper, [Toward R
15
  * python-dotenv==1.0.1
16
 
17
  ### Dataset download
18
- ```dataset download
19
  from huggingface_hub import hf_hub_download
20
 
21
  local_path = hf_hub_download(
@@ -31,12 +39,12 @@ Our evaluation uses a subset of the [DOCCI](https://google.github.io/docci/) ima
31
  ## Captioning
32
  Please generate captions for the 1,000 downloaded images for captioning evaluation.
33
  Summarize the generated captions into a dictionary where the key is the corresponding image file name, and save it as a .json file.
34
- ```captions file
35
  {
36
- "aar_test_04600.jpg": <caption_aar_test_04600>,
37
- "aar_test_04601.jpg": <caption_aar_test_04601>,
38
  ...
39
- "test_00599.json": <caption_test_00599>,
40
  }
41
  ```
42
  You may refer to the [sample captions](https://github.com/adobe-research/CapMAS/blob/master/sample_captions/llava1.6-vicuna_llama3_th1.0/captions_final.json) for guidance.
@@ -45,15 +53,15 @@ You may refer to the [sample captions](https://github.com/adobe-research/CapMAS/
45
  Please visit our [GitHub repository](https://github.com/adobe-research/CapMAS).
46
  We provide the evaluation codes for the three metrics used in our paper: **Factuality**, **Coverage**, and **CLAIR** (Chan et al., EMNLP 2023). These evaluations rely on GPT-4o, so please fill in your OpenAI API key **OR** Azure OpenAI credentials in the `conf/gpt4o` file.
47
  ### Factuality (ours)
48
- ```factuality
49
  python eval_factuality.py --image-dir <the image directory path> --captions-file <the caption .json file path>
50
  ```
51
  ### Coverage (ours)
52
- ```coverage
53
  python eval_coverage.py --vqa-dir data/COVERAGE_TEST_VQA --captions-file <the caption .json file path>
54
  ```
55
  ### CLAIR
56
- ```clair
57
  python eval_clair.py --captions-file <the caption .json file path>
58
  ```
59
 
 
1
  ---
2
+ language:
3
+ - en
4
  license: other
5
  license_name: adobe-research-license
6
  license_link: LICENSE
7
+ task_categories:
8
+ - image-to-text
9
+ tags:
10
+ - image-captioning
11
+ - hallucination-detection
12
+ - evaluation
13
+ - multimodal
14
  ---
15
+
16
  # [ICML 2025] Toward Robust Hyper-Detailed Image Captioning: A Multiagent Approach and Dual Evaluation Metrics for Factuality and Coverage
17
 
18
  This dataset is associated with the evaluation in our ICML 2025 paper, [Toward Robust Hyper-Detailed Image Captioning: A Multiagent Approach and Dual Evaluation Metrics for Factuality and Coverage](https://arxiv.org/abs/2412.15484).
 
23
  * python-dotenv==1.0.1
24
 
25
  ### Dataset download
26
+ ```python
27
  from huggingface_hub import hf_hub_download
28
 
29
  local_path = hf_hub_download(
 
39
  ## Captioning
40
  Please generate captions for the 1,000 downloaded images for captioning evaluation.
41
  Summarize the generated captions into a dictionary where the key is the corresponding image file name, and save it as a .json file.
42
+ ```json
43
  {
44
+ "aar_test_04600.jpg": "<caption_aar_test_04600>",
45
+ "aar_test_04601.jpg": "<caption_aar_test_04601>",
46
  ...
47
+ "test_00599.json": "<caption_test_00599>",
48
  }
49
  ```
50
  You may refer to the [sample captions](https://github.com/adobe-research/CapMAS/blob/master/sample_captions/llava1.6-vicuna_llama3_th1.0/captions_final.json) for guidance.
 
53
  Please visit our [GitHub repository](https://github.com/adobe-research/CapMAS).
54
  We provide the evaluation codes for the three metrics used in our paper: **Factuality**, **Coverage**, and **CLAIR** (Chan et al., EMNLP 2023). These evaluations rely on GPT-4o, so please fill in your OpenAI API key **OR** Azure OpenAI credentials in the `conf/gpt4o` file.
55
  ### Factuality (ours)
56
+ ```bash
57
  python eval_factuality.py --image-dir <the image directory path> --captions-file <the caption .json file path>
58
  ```
59
  ### Coverage (ours)
60
+ ```bash
61
  python eval_coverage.py --vqa-dir data/COVERAGE_TEST_VQA --captions-file <the caption .json file path>
62
  ```
63
  ### CLAIR
64
+ ```bash
65
  python eval_clair.py --captions-file <the caption .json file path>
66
  ```
67