--- license: apache-2.0 pipeline_tag: image-to-3d --- # Baking Gaussian Splatting into Diffusion Denoiser for Fast and Scalable Single-stage Image-to-3D Generation and Reconstruction This repository contains **DiffusionGS**, a novel single-stage 3D diffusion model for object generation and scene reconstruction from a single view. As presented in the paper, DiffusionGS directly outputs 3D Gaussian point clouds at each timestep to enforce view consistency and allows the model to generate robustly given prompt views of any directions, beyond object-centric inputs. It also features a scene-object mixed training strategy to improve capability and generality. Our method enjoys over 5× faster speed (~6s on an A100 GPU) compared to state-of-the-art methods. * **Paper**: [Baking Gaussian Splatting into Diffusion Denoiser for Fast and Scalable Single-stage Image-to-3D Generation and Reconstruction](https://huggingface.co/papers/2411.14384) * **Project Page**: [https://caiyuanhao1998.github.io/project/DiffusionGS/](https://caiyuanhao1998.github.io/project/DiffusionGS/) * **Code**: [https://github.com/caiyuanhao1998/Open-DiffusionGS](https://github.com/caiyuanhao1998/Open-DiffusionGS)