peinan commited on
Commit
99a6999
·
verified ·
1 Parent(s): b80d2a1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +161 -0
README.md CHANGED
@@ -1,3 +1,164 @@
1
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  license: cc-by-nc-sa-4.0
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ dataset_info:
3
+ - config_name: default
4
+ features:
5
+ - name: LPimage
6
+ dtype: image
7
+ - name: image1
8
+ dtype: image
9
+ - name: image2
10
+ dtype: image
11
+ - name: image3
12
+ dtype: image
13
+ - name: image4
14
+ dtype: image
15
+ - name: image5
16
+ dtype: image
17
+ - name: annotator1_ranking
18
+ sequence: int32
19
+ length: 5
20
+ - name: annotator1_best
21
+ dtype: int32
22
+ - name: annotator1_worst
23
+ dtype: int32
24
+ - name: annotator2_ranking
25
+ sequence: int32
26
+ length: 5
27
+ - name: annotator2_best
28
+ dtype: int32
29
+ - name: annotator2_worst
30
+ dtype: int32
31
+ - name: annotator3_ranking
32
+ sequence: int32
33
+ length: 5
34
+ - name: annotator3_best
35
+ dtype: int32
36
+ - name: annotator3_worst
37
+ dtype: int32
38
+ - name: annotator4_ranking
39
+ sequence: int32
40
+ length: 5
41
+ - name: annotator4_best
42
+ dtype: int32
43
+ - name: annotator4_worst
44
+ dtype: int32
45
+ - name: annotator5_ranking
46
+ sequence: int32
47
+ length: 5
48
+ - name: annotator5_best
49
+ dtype: int32
50
+ - name: annotator5_worst
51
+ dtype: int32
52
+ - name: best_annotator
53
+ dtype: string
54
+ - name: average_rank_correlation
55
+ dtype: float32
56
  license: cc-by-nc-sa-4.0
57
+ task_categories:
58
+ - visual-question-answering
59
+ language:
60
+ - ja
61
+ size_categories:
62
+ - n<1K
63
  ---
64
+
65
+ # BannerBench: Benchmarking Vision Language Models for Multi-Ad Selection with Human Preferences
66
+
67
+ ### Dataset Summary
68
+ > The BannerBench is designed to evaluate the ability of VLMs to identify the banner that best matches human preferences from a set of candidates.
69
+
70
+ ## Dataset Structure
71
+ The structure of the raw dataset is as follows:
72
+
73
+ ```JSON
74
+ {
75
+ "train": Dataset({
76
+ "features": [
77
+ 'LPimage', 'image1', 'image2', 'image3', 'image4', 'image5',
78
+ 'annotator1_ranking', 'annotator1_best', 'annotator1_worst',
79
+ 'annotator2_ranking', 'annotator2_best', 'annotator2_worst',
80
+ 'annotator3_ranking', 'annotator3_best', 'annotator3_worst',
81
+ 'annotator4_ranking', 'annotator4_best', 'annotator4_worst',
82
+ 'annotator5_ranking', 'annotator5_best', 'annotator5_worst',
83
+ 'best_annotator', 'average_rank_correlation'
84
+ ],
85
+ })
86
+ }
87
+ ```
88
+
89
+ ### Example
90
+ ```Python
91
+ from datasets import load_dataset
92
+
93
+ dataset = load_dataset("cyberagent/BannerBench")
94
+
95
+ print(dataset)
96
+ # DatasetDict({
97
+ # train: Dataset({
98
+ # features: ['LPimage', 'image1', 'image2', 'image3', 'image4', 'image5', 'annotator1_ranking', 'annotator1_best', 'annotator1_worst', 'annotator2_ranking', 'annotator2_best', 'annotator2_worst', 'annotator3_ranking', 'annotator3_best', 'annotator3_worst', 'annotator4_ranking', 'annotator4_best', 'annotator4_worst', 'annotator5_ranking', 'annotator5_best', 'annotator5_worst', 'best_annotator', 'average_rank_correlation'],
99
+ # num_rows: 900
100
+ # })
101
+ # })
102
+ ```
103
+
104
+ An example of the dataset is as follows:
105
+
106
+ ```JSON
107
+ {
108
+ "LPimage": <PIL.PngImagePlugin.PngImageFile image mode=RGB size=1280x5352 at 0x7F09A24675D0>,
109
+ "image1": <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=1080x1080 at 0x7F09A1C9B250>,
110
+ "image2": <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=1080x1080 at 0x7F09A1CB52D0>,
111
+ "image3": <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=1080x1080 at 0x7F09A1CB5810>,
112
+ "image4": <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=1080x1080 at 0x7F09A1CB5E50>,
113
+ "image5": <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=1080x1080 at 0x7F09A1CB6490>,
114
+ "annotator1_ranking": [2, 4, 1, 3, 5],
115
+ "annotator1_best": 3,
116
+ "annotator1_worst": 5,
117
+ "annotator2_ranking": [4, 5, 1, 2, 3],
118
+ "annotator2_best": 3,
119
+ "annotator2_worst": 2,
120
+ "annotator3_ranking": [3, 2, 1, 4, 5],
121
+ "annotator3_best": 3,
122
+ "annotator3_worst": 5,
123
+ "annotator4_ranking": [3, 4, 5, 2, 1],
124
+ "annotator4_best": 5,
125
+ "annotator4_worst": 3,
126
+ "annotator5_ranking": [1, 4, 2, 3, 5],
127
+ "annotator5_best": 1,
128
+ "annotator5_worst": 5,
129
+ "best_annotator": "annotator1",
130
+ "average_rank_correlation": 0.6534000039100647
131
+ }
132
+ ```
133
+
134
+ ### Data Fields
135
+
136
+ - LPimage: The Landing-Page image related image[1-5].
137
+ - image[1-5]: The Banners derived from a "LPimage".
138
+ - annotator[1-5]_ranking: Ranking of the advertisemental images in most prefered order by annotators 1 to 5.
139
+ - annotator[1-5]_best: The advertisement image is the most preferred one by annotators 1 to 5 in the Best-Choice task.
140
+ - annotator[1-5]_worst: The advertisement image is the least preferred one by annotators 1 to 5 in the Best-Choice task.
141
+ - best_annotator: The annotator whose average rank correlation with the other four annotators is the highest
142
+ - average_rank_correlation: The average of the top half of all possible annotator pairs, selected based on their rank correlation.
143
+
144
+ ## Dataset Creation
145
+
146
+ > BannerBench construction process consists of the following 3 steps;
147
+ > (1)we collected sets of five banners derived from a single LP (Banner Sets; BSs), (2)we annotated human preference to the BSs, (3)we propose two subtasks: Ranking and Best-Choice.
148
+
149
+ ## Considerations for Using the Data
150
+ > Since BannerBench is intended solely for evaluation purposes, it is not designed for training use; the benchmark focuses on assessing the inductive capabilities of VLMs.
151
+
152
+ ## License
153
+ AdTEC dataset is released under the [CreativeCommons Attribution-NonCommercial-ShareAlike 4.0 International license](./LICENSE).
154
+
155
+ ### Citation Information
156
+ To cite this work, please use the following format:
157
+ ```
158
+ {
159
+ author = {Hiroto Otake and Peinan Zhang and Yusuke Sakai and Masato Mita and Hiroki Ouchi and Taro Watanabe},
160
+ title = {BannerBench: Benchmarking Vision Language Models for Multi-Ad Selection with Human Preferences},
161
+ year = {2025},
162
+ url = {}
163
+ }
164
+ ```