maelic commited on
Commit
9efb951
·
verified ·
1 Parent(s): 3c987f7

Add dataset card

Browse files
Files changed (1) hide show
  1. README.md +167 -52
README.md CHANGED
@@ -1,54 +1,169 @@
1
  ---
2
- dataset_info:
3
- features:
4
- - name: image
5
- dtype: image
6
- - name: image_id
7
- dtype: int64
8
- - name: width
9
- dtype: int64
10
- - name: height
11
- dtype: int64
12
- - name: file_name
13
- dtype: string
14
- - name: objects
15
- list:
16
- - name: area
17
- dtype: float64
18
- - name: bbox
19
- list: float64
20
- - name: category_id
21
- dtype: int64
22
- - name: id
23
- dtype: int64
24
- - name: iscrowd
25
- dtype: int64
26
- - name: segmentation
27
- list: 'null'
28
- - name: relations
29
- list:
30
- - name: id
31
- dtype: int64
32
- - name: object_id
33
- dtype: int64
34
- - name: predicate_id
35
- dtype: int64
36
- - name: subject_id
37
- dtype: int64
38
- splits:
39
- - name: train
40
- num_bytes: 8148164061
41
- num_examples: 57623
42
- - name: val
43
- num_bytes: 1190973439
44
- num_examples: 8209
45
- download_size: 9391851086
46
- dataset_size: 9339137500
47
- configs:
48
- - config_name: default
49
- data_files:
50
- - split: train
51
- path: data/train-*
52
- - split: val
53
- path: data/val-*
54
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: mit
3
+ task_categories:
4
+ - object-detection
5
+ tags:
6
+ - scene-graph-generation
7
+ - visual-relationship-detection
8
+ - gqa
9
+ - coco-format
10
+ language:
11
+ - en
12
+ pretty_name: GQA — General Question Answering (COCO format)
13
+ size_categories:
14
+ - 100K<n<1M
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  ---
16
+
17
+ # GQA — General Question Answering (COCO format)
18
+
19
+ This dataset is the **GQA200** split of
20
+ [the GQA dataset](https://cs.stanford.edu/people/dorarad/gqa/about.html)
21
+ (Hudson et al., 2019), reformatted in standard COCO-JSON format.
22
+ GQA200 contains the top 200 object categories
23
+ and 100 relations from the original GQA dataset, selected by frequency in the
24
+ [Stacked hybrid-attention and group collaborative learning for unbiased scene graph generation
25
+ paper](https://arxiv.org/abs/2203.09811). This dataset has no official test split since it was used
26
+ for question answering rather than scene graph generation (for test there is no scene graph annotations).
27
+
28
+ This version in COCO format was produced as part of the
29
+ [SGG-Benchmark](https://github.com/Maelic/SGG-Benchmark) framework and used to train
30
+ the models described in the **REACT++** paper
31
+ ([Neau et al., 2026](https://arxiv.org/abs/2603.06386)).
32
+
33
+ ---
34
+
35
+ ## Annotation overview
36
+
37
+ Each image comes with:
38
+ - **Object bounding boxes** — 200 GQA object categories.
39
+ - **Scene-graph relations** — 100 predicate categories connecting pairs of objects as
40
+ directed `(subject, predicate, object)` triplets.
41
+
42
+ ![Annotation example — val split](gqa200_samples_val.png)
43
+
44
+ *Four random validation images with bounding boxes (coloured by category) and
45
+ relation arrows (yellow, labelled with the predicate name).*
46
+
47
+ ---
48
+
49
+ ## Dataset statistics
50
+
51
+ | Split | Images | Object annotations | Relations |
52
+ |-------|--------:|-------------------:|-----------:|
53
+ | train | 57 623 | 775 744 | 238 720 |
54
+ | val | 8 209 | 110 030 | 33 487 |
55
+
56
+ ---
57
+
58
+ ## Object categories (200)
59
+
60
+ Top-200 GQA object vocabulary used by the standard SGG split. Full list
61
+ embedded in `dataset_info.description`.
62
+
63
+ ## Predicate categories (100)
64
+
65
+ Top 100 GQA predicate vocabulary used by the standard SGG split. Full list
66
+ embedded in `dataset_info.description`.
67
+
68
+ ---
69
+
70
+ ## Dataset structure
71
+
72
+ ```python
73
+ DatasetDict({
74
+ train: Dataset({
75
+ features: ['image', 'image_id', 'width', 'height', 'file_name',
76
+ 'objects', 'relations'],
77
+ num_rows: 57623
78
+ }),
79
+ val: Dataset({
80
+ features: ['image', 'image_id', 'width', 'height', 'file_name',
81
+ 'objects', 'relations'],
82
+ num_rows: 8209
83
+ }),
84
+ })
85
+ ```
86
+
87
+ Each row contains:
88
+
89
+ | Field | Type | Description |
90
+ |-------|------|-------------|
91
+ | `image` | `Image` | PIL image |
92
+ | `image_id` | `int` | Original GQA200 image id |
93
+ | `width` / `height` | `int` | Image dimensions |
94
+ | `file_name` | `str` | Original filename |
95
+ | `objects` | `List[dict]` | `{id, category_id, bbox (xywh), area, iscrowd, segmentation}` |
96
+ | `relations` | `List[dict]` | `{id, subject_id, object_id, predicate_id}` — ids refer to `objects[*].id` |
97
+
98
+ ---
99
+
100
+ ## Usage
101
+
102
+ ```python
103
+ from datasets import load_dataset
104
+ import json
105
+
106
+ ds = load_dataset("maelic/GQA200-coco-format")
107
+
108
+ # Recover label maps from the embedded metadata
109
+ meta = json.loads(ds["train"].info.description)
110
+ cat_id2name = {c["id"]: c["name"] for c in meta["categories"]}
111
+ pred_id2name = {c["id"]: c["name"] for c in meta["rel_categories"]}
112
+
113
+ sample = ds["train"][0]
114
+ image = sample["image"] # PIL Image
115
+ for obj in sample["objects"]:
116
+ print(cat_id2name[obj["category_id"]], obj["bbox"])
117
+ for rel in sample["relations"]:
118
+ print(rel["subject_id"], "--", pred_id2name[rel["predicate_id"]], "->", rel["object_id"])
119
+ ```
120
+
121
+ ---
122
+
123
+ ## Citation
124
+
125
+ If you use this dataset, please cite GQA:
126
+
127
+ ```bibtex
128
+ @inproceedings{hudson2019gqa,
129
+ title={Gqa: A new dataset for real-world visual reasoning and compositional question answering},
130
+ author={Hudson, Drew A and Manning, Christopher D},
131
+ booktitle={Proceedings of the IEEE/CVF conference on computer vision and pattern recognition},
132
+ pages={6700--6709},
133
+ year={2019}
134
+ }
135
+ ```
136
+
137
+ And also the paper that established the GQA-200 split:
138
+
139
+ ```bibtex
140
+ @inproceedings{dong2022stacked,
141
+ title={Stacked hybrid-attention and group collaborative learning for unbiased scene graph generation},
142
+ author={Dong, Xingning and Gan, Tian and Song, Xuemeng and Wu, Jianlong and Cheng, Yuan and Nie, Liqiang},
143
+ booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
144
+ pages={19427--19436},
145
+ year={2022}
146
+ }
147
+ ```
148
+
149
+ And the REACT paper if you use the SGG-Benchmark models:
150
+
151
+ ```bibtex
152
+ @inproceedings{Neau_2025_BMVC,
153
+ author = {Ma\"elic Neau and Paulo Eduardo Santos and Anne-Gwenn Bosser
154
+ and Akihiro Sugimoto and Cedric Buche},
155
+ title = {REACT: Real-time Efficiency and Accuracy Compromise for Tradeoffs
156
+ in Scene Graph Generation},
157
+ booktitle = {36th British Machine Vision Conference 2025, {BMVC} 2025,
158
+ Sheffield, UK, November 24-27, 2025},
159
+ publisher = {BMVA},
160
+ year = {2025},
161
+ url = {https://bmva-archive.org.uk/bmvc/2025/assets/papers/Paper_239/paper.pdf},
162
+ }
163
+ ```
164
+
165
+ ---
166
+
167
+ ## License
168
+
169
+ The GQA images and annotations are released under the [Creative Commons Attribution 4.0 International (CC BY 4.0)](https://creativecommons.org/licenses/by/4.0/) license.