Ctrl-Crash: Controllable Diffusion for Realistic Car Crashes
Paper • 2506.00227 • Published • 12
import torch
from diffusers import DiffusionPipeline
# switch to "mps" for apple devices
pipe = DiffusionPipeline.from_pretrained("AnthonyGosselin/Ctrl-Crash", dtype=torch.bfloat16, device_map="cuda")
prompt = "Astronaut in a jungle, cold color palette, muted colors, detailed, 8k"
image = pipe(prompt).images[0]Generate car crash videos from an initial frame, using bounding-box and crash type control signals.
![]() |
![]() |
![]() |
This model uses the Stability AI Image-to-Video model (SVD 1.1) as a base model: https://huggingface.co/stabilityai/stable-video-diffusion-img2vid-xt-1-1
Ctrl-Crash supports different task settings, each enabled by varying the available control signals, namely:
Despite its strong performance, our approach has several limitations, which motivates future work in this direction.
BibTeX:
@misc{gosselin2025ctrlcrashcontrollablediffusionrealistic,
title={Ctrl-Crash: Controllable Diffusion for Realistic Car Crashes},
author={Anthony Gosselin and Ge Ya Luo and Luis Lara and Florian Golemo and Derek Nowrouzezahrai and Liam Paull and Alexia Jolicoeur-Martineau and Christopher Pal},
year={2025},
eprint={2506.00227},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2506.00227},
}