The dataset is currently empty. Upload or create new data files. Then, you will be able to explore them in the Dataset Viewer.

Mistake Attribution: Fine-Grained Mistake Understanding in Egocentric Videos

CVPR 2026

Yayuan Li1, Aadit Jain1, Filippos Bellos1, Jason J. Corso1,2

1University of Michigan, 2Voxel51

[Paper] [Code] [Project Page]


Dataset coming soon. We are preparing the data for public release. Stay tuned!

MATT-Bench Overview

MATT-Bench provides two large-scale benchmarks for Mistake Attribution (MATT) — a task that goes beyond binary mistake detection to attribute what semantic role was violated, when the mistake became irreversible (Point-of-No-Return), and where the mistake occurred in the frame.

The benchmarks are constructed by MisEngine, a data engine that automatically creates mistake samples with attribution-rich annotations from existing egocentric action datasets:

Dataset Samples Instruction Texts Semantic Temporal Spatial
Ego4D-M 257,584 16,099
EPIC-KITCHENS-M 221,094 12,283

These are at least two orders of magnitude larger than any existing mistake dataset.

Annotations

Each sample consists of an instruction text and an attempt video, annotated with:

  • Semantic Attribution: Which semantic role (predicate, object) in the instruction is violated in the attempt video
  • Temporal Attribution: The Point-of-No-Return (PNR) frame where the mistake becomes irreversible (Ego4D-M)
  • Spatial Attribution: Bounding box localizing the mistake region in the PNR frame (Ego4D-M)

Citation

@inproceedings{li2026mistakeattribution,
  title     = {Mistake Attribution: Fine-Grained Mistake Understanding in Egocentric Videos},
  author    = {Li, Yayuan and Jain, Aadit and Bellos, Filippos and Corso, Jason J.},
  booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
  year      = {2026},
}
Downloads last month
19

Paper for yayuanli/MATT-Bench