|
|
--- |
|
|
annotations_creators: |
|
|
- expert-generated |
|
|
language_creators: |
|
|
- other |
|
|
language: en |
|
|
license: cc-by-4.0 |
|
|
multilinguality: |
|
|
- monolingual |
|
|
size_categories: |
|
|
- 1K<n<10K |
|
|
source_datasets: |
|
|
- combination |
|
|
task_categories: |
|
|
- other |
|
|
task_ids: |
|
|
- multi-label-classification |
|
|
pretty_name: CUEBench |
|
|
configs: |
|
|
- config_name: clue |
|
|
default: true |
|
|
data_files: |
|
|
- split: train |
|
|
path: data/clue/train.jsonl |
|
|
- config_name: mep |
|
|
data_files: |
|
|
- split: train |
|
|
path: data/mep/train.jsonl |
|
|
dataset_info: |
|
|
- config_name: clue |
|
|
features: |
|
|
- name: id |
|
|
dtype: int64 |
|
|
- name: seq_name |
|
|
dtype: string |
|
|
- name: frame_count |
|
|
dtype: int64 |
|
|
- name: aligned_id |
|
|
dtype: string |
|
|
- name: image_id |
|
|
dtype: string |
|
|
- name: observed_classes |
|
|
sequence: string |
|
|
- name: target_classes |
|
|
sequence: string |
|
|
- name: detected_classes |
|
|
sequence: string |
|
|
- name: image_path |
|
|
dtype: string |
|
|
- name: image |
|
|
dtype: image |
|
|
splits: |
|
|
- name: train |
|
|
num_bytes: 1101143 |
|
|
num_examples: 1648 |
|
|
download_size: 1101143 |
|
|
dataset_size: 1101143 |
|
|
- config_name: mep |
|
|
features: |
|
|
- name: id |
|
|
dtype: int64 |
|
|
- name: seq_name |
|
|
dtype: string |
|
|
- name: frame_count |
|
|
dtype: int64 |
|
|
- name: aligned_id |
|
|
dtype: string |
|
|
- name: image_id |
|
|
dtype: string |
|
|
- name: observed_classes |
|
|
- name: image |
|
|
dtype: image |
|
|
sequence: string |
|
|
- name: target_classes |
|
|
sequence: string |
|
|
- name: detected_classes |
|
|
sequence: string |
|
|
- name: image_path |
|
|
dtype: string |
|
|
splits: |
|
|
- name: train |
|
|
num_bytes: 845579 |
|
|
num_examples: 1216 |
|
|
download_size: 845579 |
|
|
dataset_size: 845579 |
|
|
--- |
|
|
|
|
|
# CUEBench: Contextual Unobserved Entity Benchmark |
|
|
|
|
|
CUEBench is a neurosymbolic benchmark that emphasizes **contextual entity prediction** in autonomous driving scenes. Unlike traditional detection tasks, CUEBench focuses on reasoning over **unobserved entities** — objects that may be occluded, out-of-frame, or affected by sensor failures. |
|
|
|
|
|
## Dataset Summary |
|
|
|
|
|
- **Modalities**: RGB dashcam imagery + symbolic annotations (provided as metadata) |
|
|
- **Primary task**: Predict unobserved `target_classes` given the set of `observed_classes` in a scene |
|
|
- **Geography / Scenario**: Urban autonomous driving across diverse traffic densities |
|
|
- **License**: CC-BY-4.0 (you may adapt if different licensing is desired) |
|
|
|
|
|
### Configurations |
|
|
|
|
|
| Config | File | Description | |
|
|
| --- | --- | --- | |
|
|
| `clue` *(default)* | `data/clue/train.jsonl` | Contextual Unobserved Entity (CLUE) frames with heavy occlusions and single-target predictions. | |
|
|
| `mep` | `data/mep/train.jsonl` | Multi-Entity Prediction (MEP) split that introduces complementary metadata and more diverse target sets. | |
|
|
|
|
|
When this dataset is viewed on Hugging Face, the dataset viewer automatically exposes a **config dropdown** so you can switch between `clue` and `mep` without leaving the UI. |
|
|
|
|
|
## Dataset Structure |
|
|
|
|
|
### Data Fields |
|
|
| Field | Type | Description | |
|
|
| --- | --- | --- | |
|
|
| `image_id` | `string` | Unique identifier for each frame (`aligned_id` in the raw metadata). |
|
|
| `image_path` | `string` | Relative path to the rendered frame image. |
|
|
| `observed_classes` | `list[string]` | Entity classes detected in-frame (cars, cones, pedestrians, etc.). |
|
|
| `target_classes` | `list[string]` | Entities inferred to exist but unobserved (occluded, off-frame, sensor failure). |
|
|
|
|
|
### Splits |
|
|
Each configuration exposes a single **train** split sourced from either `clue_metadata.jsonl` or `mep_metadata.jsonl`. Feel free to carve out validation/test subsets before upload if you need them. |
|
|
|
|
|
### Label Taxonomy |
|
|
Representative classes include: `Car`, `Bus`, `Pedestrian`, `PickupTruck`, `MediumSizedTruck`, `Animal`, `Standing`, `VehicleWithRider`, `ConstructionSign`, `TrafficCone`, and more (~40 classes). Extend this section with the final taxonomy before publication if you want exhaustive documentation. |
|
|
|
|
|
## Example Record |
|
|
```json |
|
|
{ |
|
|
"image_id": "00003.00019", |
|
|
"observed_classes": ["Car", "Bus", "Pedestrian"], |
|
|
"target_classes": ["PickupTruck"], |
|
|
"image_path": "images/00003.00019.png" |
|
|
} |
|
|
``` |
|
|
|
|
|
## Usage |
|
|
|
|
|
### Loading with `datasets` |
|
|
```python |
|
|
from datasets import load_dataset |
|
|
|
|
|
dataset = load_dataset( |
|
|
path="ishwarbb23/cuebench", |
|
|
split="train", |
|
|
config_name="clue", # or "mep" |
|
|
) |
|
|
``` |
|
|
|
|
|
### Working From Source |
|
|
```python |
|
|
from datasets import load_dataset |
|
|
|
|
|
dataset = load_dataset( |
|
|
path="json", |
|
|
data_files={"train": "data/clue/train.jsonl"}, # swap with data/mep/train.jsonl |
|
|
split="train", |
|
|
) |
|
|
``` |
|
|
|
|
|
> **Tip:** From source, you can still switch configurations by pointing `data_files` to `data/mep/train.jsonl`. |
|
|
|
|
|
### Regenerating viewer files |
|
|
|
|
|
The repository keeps the original metadata dumps under `raw/`. To refresh the |
|
|
viewer-friendly JSONL files (e.g. after updating the raw annotations), run: |
|
|
|
|
|
```bash |
|
|
/.venv/bin/python scripts/build_viewer_files.py |
|
|
``` |
|
|
|
|
|
This script adds the derived columns (`image_id`, `observed_classes`, etc.) and |
|
|
drops the converted files into `data/clue/train.jsonl` and |
|
|
`data/mep/train.jsonl`. It also updates `data/stats.json`, which is referenced by |
|
|
the dataset card to keep `dataset_info` counters accurate. |
|
|
|
|
|
## Metrics |
|
|
|
|
|
`metric.py` defines **Mean Reciprocal Rank**, **Hits@K (1/3/5/10)**, and **Coverage@K (1/3/5/10)** over the predicted class rankings. When publishing to the Hugging Face Metrics Hub, expose the `compute(predictions, references)` signature so leaderboard integrations can consume it. |
|
|
|
|
|
## Licensing |
|
|
|
|
|
The dataset is currently tagged as **CC-BY-4.0**. Update this section if you select a different license. |
|
|
|
|
|
## Citation |
|
|
|
|
|
``` |
|
|
@misc{cuebench2025, |
|
|
title = {CUEBench: Contextual Unobserved Entity Benchmark}, |
|
|
author = {CUEBench Authors}, |
|
|
year = {2025} |
|
|
} |
|
|
``` |
|
|
|
|
|
## Hugging Face Upload Checklist |
|
|
|
|
|
1. Install tools: `pip install datasets huggingface_hub` and run `huggingface-cli login`. |
|
|
2. Create the dataset repo: `huggingface-cli repo create cuebench --type dataset` (or via UI). |
|
|
3. Ensure directory layout: |
|
|
``` |
|
|
cuebench/ |
|
|
README.md |
|
|
data/ |
|
|
clue/train.jsonl |
|
|
mep/train.jsonl |
|
|
raw/ |
|
|
clue_metadata.jsonl |
|
|
mep_metadata.jsonl |
|
|
metric.py # optional metric script |
|
|
scripts/build_viewer_files.py |
|
|
scripts/push_to_hub.py |
|
|
images/... # optional or host separately |
|
|
``` |
|
|
4. Initialize Git + LFS: |
|
|
```bash |
|
|
cd cuebench |
|
|
git init |
|
|
git lfs install |
|
|
git lfs track "*.jsonl" "images/*" |
|
|
git remote add origin https://huggingface.co/datasets/ishwarbb23/cuebench |
|
|
git add . |
|
|
git commit -m "Initial CUEBench dataset" |
|
|
git push origin main |
|
|
``` |
|
|
5. Regenerate viewer files anytime the raw metadata changes: `/.venv/bin/python scripts/build_viewer_files.py` |
|
|
6. Push the prepared splits to the Hub (per config) using `/.venv/bin/python scripts/push_to_hub.py --repo ishwarbb23/cuebench` |
|
|
7. On the Hub page, trigger the dataset preview to ensure the loader runs. |
|
|
8. (Optional) Publish the metric under `metrics/cuebench-metric` following the Metrics Hub template and link it from the dataset card. |
|
|
|
|
|
Update these steps with any organization-specific tooling you use. |
|
|
|