DiffusionGS / README.md
nielsr's picture
nielsr HF Staff
Enhance model card: Add pipeline tag, links, visuals, and usage
ee2228b verified
|
raw
history blame
2.64 kB
metadata
license: apache-2.0
pipeline_tag: image-to-3d

Baking Gaussian Splatting into Diffusion Denoiser for Fast and Scalable Single-stage Image-to-3D Generation and Reconstruction

This repository contains DiffusionGS, a novel single-stage 3D diffusion model for object generation and scene reconstruction from a single view. As presented in the paper, DiffusionGS directly outputs 3D Gaussian point clouds at each timestep to enforce view consistency and allows the model to generate robustly given prompt views of any directions, beyond object-centric inputs. It also features a scene-object mixed training strategy to improve capability and generality. Our method enjoys over 5× faster speed (~6s on an A100 GPU) compared to state-of-the-art methods.

abo gso real_img wild
DiffusionGS Pipeline

Quick Demo

For object-centric image-to-3D generation, a single-line script is provided to use the code:

python run.py

This code will automatically download the model checkpoints and config files from Hugging Face.

Citation

If you find our work useful, please consider citing our paper:

@inproceedings{diffusiongs,
  title={Baking Gaussian Splatting into Diffusion Denoiser for Fast and Scalable Single-stage Image-to-3D Generation and Reconstruction},
  author={Yuanhao Cai and He Zhang and Kai Zhang and Yixun Liang and Mengwei Ren and Fujun Luan and Qing Liu and Soo Ye Kim and Jianming Zhang and Zhifei Zhang and Yuqian Zhou and Yulun Zhang and Xiaokang Yang and Zhe Lin and Alan Yuille},
  booktitle={ICCV},
  year={2025}
}