PeterAM4 commited on
Commit
0554d0c
·
verified ·
1 Parent(s): 9bb1ebc

Add SQuID dataset card with statistics

Browse files
Files changed (1) hide show
  1. README.md +230 -31
README.md CHANGED
@@ -1,33 +1,232 @@
1
  ---
2
- dataset_info:
3
- features:
4
- - name: id
5
- dtype: string
6
- - name: image
7
- dtype: image
8
- - name: question
9
- dtype: string
10
- - name: answer
11
- dtype: string
12
- - name: type
13
- dtype: string
14
- - name: tier
15
- dtype: int32
16
- - name: gsd
17
- dtype: float32
18
- - name: acceptable_range_lower
19
- dtype: float64
20
- - name: acceptable_range_upper
21
- dtype: float64
22
- splits:
23
- - name: train
24
- num_bytes: 3939200871
25
- num_examples: 2000
26
- download_size: 3958457415
27
- dataset_size: 3939200871
28
- configs:
29
- - config_name: default
30
- data_files:
31
- - split: train
32
- path: data/train-*
33
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: cc-by-4.0
3
+ task_categories:
4
+ - visual-question-answering
5
+ language:
6
+ - en
7
+ tags:
8
+ - satellite-imagery
9
+ - spatial-reasoning
10
+ - benchmark
11
+ - quantitative-reasoning
12
+ size_categories:
13
+ - 1K<n<10K
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  ---
15
+
16
+ # SQuID: Satellite Quantitative Intelligence Dataset
17
+
18
+ A comprehensive benchmark for evaluating quantitative spatial reasoning in Vision-Language Models using satellite imagery.
19
+
20
+ ## Dataset Overview
21
+
22
+ - **2000 questions** testing spatial reasoning on satellite imagery
23
+ - **587 unique images** across four datasets
24
+ - **1950 auto-labeled** questions from segmentation masks (DeepGlobe, EarthVQA, Solar Panels)
25
+ - **50 human-annotated** questions from NAIP imagery with consensus answers
26
+ - **1577 questions** include human-agreement ranges for numeric answers
27
+ - **3 difficulty tiers**: Basic (710), Spatial (616), Complex (674)
28
+ - **3 resolution levels**: 0.3m, 0.5m, 1.0m GSD
29
+
30
+ ## Human Annotation & Agreement Methodology
31
+
32
+ ### Human Annotation Process
33
+ - **50 questions** on NAIP 1.0m GSD imagery were annotated by humans
34
+ - **10 annotators per question** resulting in 500 total annotations
35
+ - **Answer aggregation**:
36
+ - **Numeric questions**: Used MEDIAN of all responses for robustness
37
+ - **Categorical questions** (connected/fragmented): Used MAJORITY voting
38
+ - **Binary questions**: Converted yes/no to 1/0 and used majority
39
+
40
+ ### Human Agreement Quantification
41
+ From the 500 human annotations, we computed the Mean Median Absolute Deviation (MAD) for each question type:
42
+ - **Percentage questions**: MAD = ±1.74 percentage points
43
+ - **Proximity questions**: MAD = ±2.25 percentage points
44
+ - **Count questions**: Normalized MADc = 0.19 (proportional to count magnitude)
45
+
46
+ For count questions, we use a normalized MAD (MADc) that makes the acceptable range proportional to the count value:
47
+ ```
48
+ MADc = median(|Xi - median(X)|) / median(X) = 0.19
49
+ ```
50
+
51
+ ### Acceptable Range Calculation
52
+ These MAD values were applied to ALL numeric questions in the benchmark to define acceptable ranges:
53
+
54
+ ```python
55
+ import math
56
+
57
+ # For percentage questions (absolute deviation)
58
+ if question_type == 'percentage':
59
+ lower = max(0.0, answer - 1.74)
60
+ upper = min(100.0, answer + 1.74)
61
+
62
+ # For count questions (proportional deviation)
63
+ # range(C) = [C - max(1, C × MADc), C + max(1, C × MADc)]
64
+ elif question_type in ['count', 'building_proximity', 'building_flood_risk',
65
+ 'building_fire_risk', 'connectivity']:
66
+ MADc = 0.19
67
+ dr = max(1, answer * MADc) # At least ±1 deviation
68
+ lower = max(0, math.floor(answer - dr))
69
+ upper = math.ceil(answer + dr)
70
+
71
+ # For proximity percentage questions (absolute deviation)
72
+ elif 'within' in question and 'm of' in question:
73
+ lower = max(0.0, answer - 2.25)
74
+ upper = min(100.0, answer + 2.25)
75
+ ```
76
+
77
+ **Example count ranges with MADc = 0.19:**
78
+ - C=5 → range [4, 7]
79
+ - C=10 → range [8, 12]
80
+ - C=50 → range [40, 60]
81
+ - C=100 → range [81, 120]
82
+
83
+ Special cases:
84
+ - Zero values have no range (exact match required)
85
+ - Binary/fragmentation questions have no range (exact match)
86
+ - Ranges are capped at valid bounds (0-100 for percentages, ≥0 for counts)
87
+
88
+ ## Question Types
89
+
90
+ The benchmark includes 24 distinct question types organized into three tiers:
91
+
92
+ ### Tier 1: Basic Questions (710 questions)
93
+ - **percentage**: Coverage percentage of a land use class
94
+ - **count**: Number of separate regions or objects
95
+ - **size**: Area measurements of regions
96
+ - **total_area**: Total area covered by a class
97
+ - **binary_comparison**: Comparing quantities between two classes
98
+ - **binary_presence**: Checking if a class exists
99
+ - **binary_threshold**: Testing if values exceed thresholds
100
+ - **binary_multiple**: Checking for multiple instances
101
+
102
+ ### Tier 2: Spatial Questions (616 questions)
103
+ - **proximity_percentage**: Percentage of one class near another
104
+ - **proximity_area**: Area of one class near another
105
+ - **binary_proximity**: Presence of one class near another
106
+ - **building_proximity**: Number of buildings near other features
107
+ - **building_flood_risk**: Buildings at flood risk (near water)
108
+ - **building_fire_risk**: Buildings at fire risk (near forest)
109
+ - **connectivity**: Counting isolated patches by size
110
+ - **fragmentation**: Assessing if regions are connected or fragmented
111
+ - **power_calculation**: Calculating solar panel power output
112
+
113
+ ### Tier 3: Complex Questions (674 questions)
114
+ - **complex_multi_condition**: Areas meeting multiple spatial criteria
115
+ - **complex_urban_flood_risk**: Urban areas at flood risk (near water)
116
+ - **complex_urban_fire_risk**: Urban areas at fire risk (near forest)
117
+ - **complex_agriculture_water_access**: Agricultural land with irrigation potential
118
+ - **complex_size_filter**: Filtering by size thresholds
119
+ - **complex_average**: Average sizes of regions
120
+
121
+ ## Loading the Dataset
122
+
123
+ ```python
124
+ from datasets import load_dataset
125
+
126
+ # Load dataset
127
+ dataset = load_dataset("PeterAM4/SQuID")
128
+
129
+ # Access a sample
130
+ sample = dataset['train'][0]
131
+ image = sample['image'] # PIL Image
132
+ question = sample['question']
133
+ answer = sample['answer'] # String or numeric
134
+ type = sample['type']
135
+
136
+ # Convert answer based on type
137
+ if type in ['percentage', 'count', 'proximity_percentage', 'proximity_area',
138
+ 'building_proximity', 'building_flood_risk', 'building_fire_risk',
139
+ 'connectivity', 'size', 'total_area', 'power_calculation'] or 'complex' in type:
140
+ answer_value = float(answer)
141
+ elif 'binary' in type:
142
+ answer_value = int(answer) # 0 or 1
143
+ elif type == 'fragmentation':
144
+ answer_value = answer # "connected" or "fragmented"
145
+ ```
146
+
147
+ ## Fields
148
+
149
+ - **id**: Question identifier (e.g., "SQuID_0001")
150
+ - **image**: Satellite image path
151
+ - **question**: Question text with GSD notation
152
+ - **answer**: Ground truth answer
153
+ - **type**: One of 24 question types
154
+ - **tier**: Difficulty level (1=Basic, 2=Spatial, 3=Complex)
155
+ - **gsd**: Ground sampling distance in meters
156
+ - **acceptable_range**: [lower, upper] bounds for numeric questions (when applicable)
157
+
158
+ ## Evaluation
159
+
160
+ For numeric questions, check if predictions fall within the acceptable range:
161
+
162
+ ```python
163
+ import math
164
+
165
+ def evaluate(prediction, sample):
166
+ if 'acceptable_range' in sample:
167
+ # Numeric question - check if within human agreement range
168
+ lower, upper = sample['acceptable_range']
169
+ return lower <= float(prediction) <= upper
170
+ else:
171
+ # Non-numeric question - exact match required
172
+ return str(prediction).lower() == str(sample['answer']).lower()
173
+ ```
174
+
175
+ The acceptable ranges represent the natural variation in human perception for spatial measurements.
176
+
177
+ ## Dataset Distribution
178
+
179
+ ### By Tier
180
+ - **Tier 1 (Basic)**: 710 questions (35.5%)
181
+ - **Tier 2 (Spatial)**: 616 questions (30.8%)
182
+ - **Tier 3 (Complex)**: 674 questions (33.7%)
183
+
184
+ ### Top Question Types
185
+ - **complex_multi_condition**: 490 questions (24.5%)
186
+ - **count**: 178 questions (8.9%)
187
+ - **binary_comparison**: 172 questions (8.6%)
188
+ - **size**: 166 questions (8.3%)
189
+ - **percentage**: 157 questions (7.8%)
190
+ - **proximity_percentage**: 123 questions (6.2%)
191
+ - **binary_proximity**: 122 questions (6.1%)
192
+ - **proximity_area**: 107 questions (5.3%)
193
+ - **connectivity**: 104 questions (5.2%)
194
+ - **fragmentation**: 98 questions (4.9%)
195
+
196
+ ### By Source
197
+ - **DeepGlobe (0.5m GSD)**: 612 questions, 174 images - Land use classification masks
198
+ - **EarthVQA (0.3m GSD)**: 1241 questions, 364 images - Building detection and land cover masks
199
+ - **Solar Panels (0.3m GSD)**: 97 questions, 35 images - Solar panel segmentation masks
200
+ - **NAIP (1.0m GSD)**: 50 questions, 14 images - Human-annotated diverse scenes
201
+
202
+
203
+ ## Statistics Summary
204
+
205
+ - **Zero-valued answers**: 102 (5.1%)
206
+ - **Questions with ranges**: 1577 (78.8%)
207
+ - **Average questions per image**: 3.4
208
+
209
+ ## Notes
210
+
211
+ - Questions explicitly state minimum area thresholds (e.g., "ignore patches smaller than 0.125 hectares")
212
+ - Zero-valued answers indicate absence of features (intentionally included for robustness testing)
213
+ - The benchmark tests both presence and absence of spatial features to avoid positive-only bias
214
+ - Human agreement ranges allow for natural variation in spatial perception and counting
215
+ - All measurements use metric units based on the specified GSD (Ground Sampling Distance)
216
+ - Count ranges use proportional MADc (0.19) so larger counts have wider acceptable ranges
217
+
218
+ ## Citation
219
+
220
+ If you use this dataset, please cite:
221
+
222
+ ```bibtex
223
+ @article{massih2025squid,
224
+ title={Preserving Pixel-Level Precision: SQuID Dataset and QVLM Architecture for Quantitative Geospatial Reasoning},
225
+ author={Peter A. Massih, Eric Cosatto},
226
+ journal={arXiv preprint arXiv:XXXX.XXXXX},
227
+ year=2025
228
+ }
229
+ ```
230
+
231
+ ---
232
+ *Generated on 2026-01-18 17:13:25*