nihalnayak commited on
Commit
44a49dc
·
verified ·
1 Parent(s): ee4dd73

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +59 -0
README.md ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ task_categories:
3
+ - image-classification
4
+ tags:
5
+ - composition
6
+ ---
7
+
8
+ # Dataset Card for CGQA dataset
9
+ This is the CGQA dataset from the [Learning Graph Embeddings for Compositional Zero-shot Learning](https://arxiv.org/abs/2102.01987) paper.
10
+
11
+
12
+ ## Citation
13
+ If you use this dataset, please cite the following papers:
14
+
15
+ ```
16
+ @inproceedings{naeem2021learning,
17
+ title={Learning graph embeddings for compositional zero-shot learning},
18
+ author={Naeem, Muhammad Ferjad and Xian, Yongqin and Tombari, Federico and Akata, Zeynep},
19
+ booktitle={Proceedings of the IEEE/CVF conference on computer vision and pattern recognition},
20
+ pages={953--962},
21
+ year={2021}
22
+ }
23
+ ```
24
+
25
+ CGQA is derived from the GQA datset
26
+ ```
27
+ @inproceedings{hudson2019gqa,
28
+ title={Gqa: A new dataset for real-world visual reasoning and compositional question answering},
29
+ author={Hudson, Drew A and Manning, Christopher D},
30
+ booktitle={Proceedings of the IEEE/CVF conference on computer vision and pattern recognition},
31
+ pages={6700--6709},
32
+ year={2019}
33
+ }
34
+ ```
35
+
36
+ The GQA dataset is derived from Visual Genome
37
+ ```
38
+ @article{krishna2017visual,
39
+ title={Visual genome: Connecting language and vision using crowdsourced dense image annotations},
40
+ author={Krishna, Ranjay and Zhu, Yuke and Groth, Oliver and Johnson, Justin and Hata, Kenji and Kravitz, Joshua and Chen, Stephanie and Kalantidis, Yannis and Li, Li-Jia and Shamma, David A and others},
41
+ journal={International journal of computer vision},
42
+ volume={123},
43
+ number={1},
44
+ pages={32--73},
45
+ year={2017},
46
+ publisher={Springer}
47
+ }
48
+ ```
49
+
50
+ If you use this dataset with [compositional soft prompting](https://arxiv.org/abs/2204.03574), then cite this paper:
51
+ ```
52
+ @inproceedings{
53
+ csp2023,
54
+ title={Learning to Compose Soft Prompts for Compositional Zero-Shot Learning},
55
+ author={Nihal V. Nayak and Peilin Yu and Stephen H. Bach},
56
+ booktitle={International Conference on Learning Representations},
57
+ year={2023}
58
+ }
59
+ ```