EvalData commited on
Commit
afaeb05
·
verified ·
1 Parent(s): 3fb68c1

croissant: add rai:hasSyntheticData, absolute HF contentUrls, JSONL column extracts (all 4 NeurIPS checker checks pass)

Browse files
Files changed (1) hide show
  1. croissant.json +12 -11
croissant.json CHANGED
@@ -77,6 +77,7 @@
77
  "rai:dataLimitations": "Not intended as a deployment decision tool for safety-critical domains without additional validation. Template-generated questions do not fully capture naturalistic queries. Retrievability scores and failure buckets are heuristic and lack human validation; they should be treated as exploratory. API-based model results may vary across calls due to provider-side updates.",
78
  "rai:personalSensitiveInformation": "None. All data are synthetic or from public datasets with no personal or sensitive information. SciTabAlign contains scientific paper tables; ETT contains electricity transformer temperature measurements; NetworkX graphs are classic social-network datasets without personal identifiers.",
79
  "rai:dataUseCases": "Primary: benchmarking MLLM visualization-format dependence; secondary: representation-invariance research, evaluation-protocol research, diagnostic analysis of structured-data reasoning.",
 
80
  "keywords": [
81
  "multimodal",
82
  "benchmark",
@@ -91,7 +92,7 @@
91
  "@id": "base-items-file",
92
  "name": "base_items.jsonl",
93
  "description": "Synthetic-only single-modality items: 1,500 questions (500 tabular + 500 time-series + 500 graph) from programmatically generated data.",
94
- "contentUrl": "benchmark/base_items.jsonl",
95
  "encodingFormat": "application/jsonlines",
96
  "sha256": "5d77094f209cb6e4d2fc2cbcd87da1f5dc55679b606f28798ddd26e0d9bc7e5b"
97
  },
@@ -100,7 +101,7 @@
100
  "@id": "mixed-items-file",
101
  "name": "mixed_items.jsonl",
102
  "description": "Cross-modal composite evaluation items (600 items) requiring reasoning across two data modalities simultaneously (tabular+timeseries, tabular+graph, timeseries+graph).",
103
- "contentUrl": "benchmark/mixed_items.jsonl",
104
  "encodingFormat": "application/jsonlines",
105
  "sha256": "06f022e16dda6f35e8dbb01e1c8dd73ad6b981b399a4c7606b1943da78313d91"
106
  },
@@ -109,7 +110,7 @@
109
  "@id": "realworld-test-file",
110
  "name": "realworld_test.jsonl",
111
  "description": "Full benchmark (3,795 items): superset containing all 1,500 synthetic items plus 2,295 real-world items (1,265 SciTabAlign tabular + 870 ETT time-series + 160 NetworkX graph). This is the primary evaluation file. Use the source field to filter by provenance.",
112
- "contentUrl": "benchmark/realworld_test.jsonl",
113
  "encodingFormat": "application/jsonlines",
114
  "sha256": "6ef2baf69daefac216200fbcc87df7087bb88e6d987390f64ce24364a2138779"
115
  }
@@ -132,7 +133,7 @@
132
  "@id": "base-items-file"
133
  },
134
  "extract": {
135
- "jsonPath": "$.question_id"
136
  }
137
  }
138
  },
@@ -147,7 +148,7 @@
147
  "@id": "base-items-file"
148
  },
149
  "extract": {
150
- "jsonPath": "$.question"
151
  }
152
  }
153
  },
@@ -162,7 +163,7 @@
162
  "@id": "base-items-file"
163
  },
164
  "extract": {
165
- "jsonPath": "$.answer"
166
  }
167
  }
168
  },
@@ -177,7 +178,7 @@
177
  "@id": "base-items-file"
178
  },
179
  "extract": {
180
- "jsonPath": "$.modality"
181
  }
182
  }
183
  },
@@ -192,7 +193,7 @@
192
  "@id": "base-items-file"
193
  },
194
  "extract": {
195
- "jsonPath": "$.source"
196
  }
197
  }
198
  },
@@ -207,7 +208,7 @@
207
  "@id": "base-items-file"
208
  },
209
  "extract": {
210
- "jsonPath": "$.viz_methods"
211
  }
212
  }
213
  },
@@ -222,7 +223,7 @@
222
  "@id": "base-items-file"
223
  },
224
  "extract": {
225
- "jsonPath": "$.difficulty"
226
  }
227
  }
228
  },
@@ -237,7 +238,7 @@
237
  "@id": "base-items-file"
238
  },
239
  "extract": {
240
- "jsonPath": "$.task"
241
  }
242
  }
243
  }
 
77
  "rai:dataLimitations": "Not intended as a deployment decision tool for safety-critical domains without additional validation. Template-generated questions do not fully capture naturalistic queries. Retrievability scores and failure buckets are heuristic and lack human validation; they should be treated as exploratory. API-based model results may vary across calls due to provider-side updates.",
78
  "rai:personalSensitiveInformation": "None. All data are synthetic or from public datasets with no personal or sensitive information. SciTabAlign contains scientific paper tables; ETT contains electricity transformer temperature measurements; NetworkX graphs are classic social-network datasets without personal identifiers.",
79
  "rai:dataUseCases": "Primary: benchmarking MLLM visualization-format dependence; secondary: representation-invariance research, evaluation-protocol research, diagnostic analysis of structured-data reasoning.",
80
+ "rai:hasSyntheticData": true,
81
  "keywords": [
82
  "multimodal",
83
  "benchmark",
 
92
  "@id": "base-items-file",
93
  "name": "base_items.jsonl",
94
  "description": "Synthetic-only single-modality items: 1,500 questions (500 tabular + 500 time-series + 500 graph) from programmatically generated data.",
95
+ "contentUrl": "https://huggingface.co/datasets/EvalData/StructViz-Bench/resolve/main/benchmark/base_items.jsonl",
96
  "encodingFormat": "application/jsonlines",
97
  "sha256": "5d77094f209cb6e4d2fc2cbcd87da1f5dc55679b606f28798ddd26e0d9bc7e5b"
98
  },
 
101
  "@id": "mixed-items-file",
102
  "name": "mixed_items.jsonl",
103
  "description": "Cross-modal composite evaluation items (600 items) requiring reasoning across two data modalities simultaneously (tabular+timeseries, tabular+graph, timeseries+graph).",
104
+ "contentUrl": "https://huggingface.co/datasets/EvalData/StructViz-Bench/resolve/main/benchmark/mixed_items.jsonl",
105
  "encodingFormat": "application/jsonlines",
106
  "sha256": "06f022e16dda6f35e8dbb01e1c8dd73ad6b981b399a4c7606b1943da78313d91"
107
  },
 
110
  "@id": "realworld-test-file",
111
  "name": "realworld_test.jsonl",
112
  "description": "Full benchmark (3,795 items): superset containing all 1,500 synthetic items plus 2,295 real-world items (1,265 SciTabAlign tabular + 870 ETT time-series + 160 NetworkX graph). This is the primary evaluation file. Use the source field to filter by provenance.",
113
+ "contentUrl": "https://huggingface.co/datasets/EvalData/StructViz-Bench/resolve/main/benchmark/realworld_test.jsonl",
114
  "encodingFormat": "application/jsonlines",
115
  "sha256": "6ef2baf69daefac216200fbcc87df7087bb88e6d987390f64ce24364a2138779"
116
  }
 
133
  "@id": "base-items-file"
134
  },
135
  "extract": {
136
+ "column": "question_id"
137
  }
138
  }
139
  },
 
148
  "@id": "base-items-file"
149
  },
150
  "extract": {
151
+ "column": "question"
152
  }
153
  }
154
  },
 
163
  "@id": "base-items-file"
164
  },
165
  "extract": {
166
+ "column": "answer"
167
  }
168
  }
169
  },
 
178
  "@id": "base-items-file"
179
  },
180
  "extract": {
181
+ "column": "modality"
182
  }
183
  }
184
  },
 
193
  "@id": "base-items-file"
194
  },
195
  "extract": {
196
+ "column": "source"
197
  }
198
  }
199
  },
 
208
  "@id": "base-items-file"
209
  },
210
  "extract": {
211
+ "column": "viz_methods"
212
  }
213
  }
214
  },
 
223
  "@id": "base-items-file"
224
  },
225
  "extract": {
226
+ "column": "difficulty"
227
  }
228
  }
229
  },
 
238
  "@id": "base-items-file"
239
  },
240
  "extract": {
241
+ "column": "task"
242
  }
243
  }
244
  }