File size: 5,072 Bytes
7f3465a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
151f10d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
63ae3a0
151f10d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
---
license: cc0-1.0
dataset_info:
  features:
  - name: uuid
    dtype: string
  - name: image
    dtype: image
  - name: reference
    dtype: string
  splits:
  - name: train
    num_bytes: 3513634757
    num_examples: 1000
  - name: val
    num_bytes: 866045160
    num_examples: 250
  - name: test
    num_bytes: 1722946225
    num_examples: 500
  download_size: 6101723890
  dataset_size: 6102626142
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
  - split: val
    path: data/val-*
  - split: test
    path: data/test-*
task_categories:
- image-to-text
language:
- en
tags:
- art
size_categories:
- 1K<n<10K
---
# Dataset Card for docent

This dataset contains works of art with expert-written detailed descriptions from the U.S. National Gallery of Art, published as part of DOCENT.  It was introduced in "PoSh: Using Scene Graphs To Guide LLMs-as-a-Judge For Detailed Image Descriptions".  You can find a full description of its collection methodology in the paper: https://arxiv.org/abs/2510.19060.

## Dataset Details

- **Language:** English
- **License:** CC-0

### Dataset Sources 

- **Images:** images in the public domain, from the U.S. National Gallery of Art
- **Reference Descriptions:** expert-written references from the U.S. National Gallery of Art, published as part of their Open Data Initiative (https://github.com/NationalGalleryOfArt/opendata) 
- **Repository:** https://github.com/amith-ananthram/posh
- **Paper:** https://arxiv.org/abs/2510.19060

## Uses

The intended use of this dataset is as a benchmark for evaluating detailed image description, in particular for artwork.  It contains three splits: a training set of 1,000 images, a validation set of 250 images and a test set of 500 images.  When evaluating model generations, we recommend reporting PoSh scores (https://github.com/amith-ananthram/posh) or using a replicable metric that produces stronger correlations with the judgments in https://huggingface.co/datasets/amitha/docent-eval-coarse. 

## Dataset Structure

Each row in the dataset corresponds to a work of art. 
- uuid: a unique identifier for work of art
- image: an image of the work of art (useful for multimodal metrics)
- reference: an expert-written reference description of this artwork from the U.S. National Gallery of Art

## Dataset Creation

### Curation Rationale

This dataset was collected to evaluate detailed image description, especially for artwork.

### Source Data

The images/artwork are all in the public domain and provided by the U.S. National Gallery of Art.

The expert written references were published by the U.S. National Gallery of Art as part of their Open Data Initiative (https://github.com/NationalGalleryOfArt/opendata).

### Annotations 

#### Annotation process

The expert written reference descriptions were composed according to the U.S. National Gallery of Art's Accessibility Guidelines: https://www.nga.gov/visit/accessibility/collection-image-descriptions.

#### Who are the annotators?

An expert in art history from the U.S. National Gallery of Art.

## Bias, Risks, and Limitations

While this work aims to benefit accessibility applications for blind and low-vision users (as reference descriptions were written according to the U.S. National Gallery of Art's Accessibility Guidelines: https://www.nga.gov/visit/accessibility/collection-image-descriptions), we acknowledge that it assumes a one-size-fits-all approach to assistive text.  Ideally, such a benchmark would include different styles of accessibility text more representative of diverse user needs.  However, it is our hope that by including reference descriptions that are extremely detailed, models that perform well in this more challenging setting will be able to adapt to a wide number of description needs.

Additionally, as with other computer vision systems, this work could theoretically be applied to surveillance contexts, but our focus on detailed description does not introduce novel privacy risks beyond those inherent to existing image analysis technologies. The primary intended application—-improving accessibility—-aligns with beneficial societal outcomes.

## Citation 

**BibTeX:**

@misc{ananthram2025poshusingscenegraphs,
      title={PoSh: Using Scene Graphs To Guide LLMs-as-a-Judge For Detailed Image Descriptions}, 
      author={Amith Ananthram and Elias Stengel-Eskin and Lorena A. Bradford and Julia Demarest and Adam Purvis and Keith Krut and Robert Stein and Rina Elster Pantalony and Mohit Bansal and Kathleen McKeown},
      year={2025},
      eprint={2510.19060},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2510.19060}, 
}

**APA:**

Ananthram, A., Stengel-Eskin, E., Bradford, L.A., Demarest, J., Purvis, A., Krut, K., Stein, R., Pantalony, R.E., Bansal, M., McKeown, K. (2025). PoSh: Using Scene Graphs To Guide LLMs-as-a-Judge For Detailed Image Descriptions. arXiv preprint arXiv:2510.19060.

## Dataset Card Authors

Amith Ananthram

## Dataset Card Contact

amith@cs.columbia.edu