--- license: cc-by-4.0 task_categories: - video-classification - video-text-to-text - object-detection tags: - egocentric-video - mistake-detection - temporal-localization - video-language-grounding - hand-object-interaction - action-recognition - procedural-activities - semantic-role-labeling - ego4d - epic-kitchens - point-of-no-return - cvpr2026 pretty_name: MATT-Bench size_categories: - 100K1, [Aadit Jain](https://www.linkedin.com/in/jain-aadit/)1, [Filippos Bellos](https://www.linkedin.com/in/filippos-bellos-168595156/)1, [Jason J. Corso](https://www.linkedin.com/in/jason-corso/)1,2 1University of Michigan, 2Voxel51 [[Paper](https://arxiv.org/abs/2511.20525)] [[Code](https://github.com/yayuanli/MATT)] [[Project Page](https://yayuanli.github.io/MATT/)] --- > **Dataset coming soon.** We are preparing the data for public release. Stay tuned! ## MATT-Bench Overview MATT-Bench provides two large-scale benchmarks for **Mistake Attribution (MATT)** — a task that goes beyond binary mistake detection to attribute *what* semantic role was violated, *when* the mistake became irreversible (Point-of-No-Return), and *where* the mistake occurred in the frame. The benchmarks are constructed by **MisEngine**, a data engine that automatically creates mistake samples with attribution-rich annotations from existing egocentric action datasets: | Dataset | Samples | Instruction Texts | Semantic | Temporal | Spatial | |---|---|---|---|---|---| | **Ego4D-M** | 257,584 | 16,099 | ✓ | ✓ | ✓ | | **EPIC-KITCHENS-M** | 221,094 | 12,283 | ✓ | — | — | These are at least **two orders of magnitude larger** than any existing mistake dataset. ## Annotations Each sample consists of an instruction text and an attempt video, annotated with: - **Semantic Attribution**: Which semantic role (predicate, object) in the instruction is violated in the attempt video - **Temporal Attribution**: The Point-of-No-Return (PNR) frame where the mistake becomes irreversible (Ego4D-M) - **Spatial Attribution**: Bounding box localizing the mistake region in the PNR frame (Ego4D-M) ## Citation ```bibtex @inproceedings{li2026mistakeattribution, title = {Mistake Attribution: Fine-Grained Mistake Understanding in Egocentric Videos}, author = {Li, Yayuan and Jain, Aadit and Bellos, Filippos and Corso, Jason J.}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)}, year = {2026}, } ```