Add model card with metadata and usage instructions

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +55 -3
README.md CHANGED
@@ -1,3 +1,55 @@
1
- ---
2
- license: cc-by-nc-nd-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-nd-4.0
3
+ pipeline_tag: image-to-image
4
+ ---
5
+
6
+ # Splatent: Splatting Diffusion Latents for Novel View Synthesis
7
+
8
+ [**Project Website**](https://orhir.github.io/Splatent/) | [**Paper (arXiv)**](https://arxiv.org/abs/2512.09923) | [**GitHub**](https://github.com/orhir/Splatent)
9
+
10
+ Splatent is a diffusion-based enhancement framework designed to operate on top of 3D Gaussian Splatting (3DGS) in the latent space of VAEs. It enhances novel view synthesis by leveraging Stable Diffusion Turbo and multi-view attention mechanisms to recover fine-grained details while maintaining multi-view consistency.
11
+
12
+ ![Splatent Architecture](https://github.com/orhir/Splatent/raw/main/assets/splatent_architecture.png)
13
+
14
+ ## Installation
15
+
16
+ To set up the environment, run:
17
+
18
+ ```bash
19
+ conda env create -f environment.yml
20
+ conda activate splatent
21
+ pip install --no-build-isolation submodules/feature-3dgs/submodules/custom_rasterizer
22
+ pip install --no-build-isolation submodules/feature-3dgs/submodules/simple-knn
23
+ ```
24
+
25
+ ## Custom Inference
26
+
27
+ For inference on custom images using the pre-trained weights, you can use the following command from the source repository:
28
+
29
+ ```bash
30
+ python src/inference_custom.py \
31
+ --input_image /path/to/input \
32
+ --ref_image /path/to/reference \
33
+ --prompt "A photo of a scene" \
34
+ --model_path /path/to/checkpoint.pkl \
35
+ --output_dir ./results
36
+ ```
37
+
38
+ ## Citation
39
+
40
+ If you find Splatent useful for your research, please cite the following paper:
41
+
42
+ ```bibtex
43
+ @article{splatent2025,
44
+ title={Splatent: Splatting Diffusion Latents for Novel View Synthesis},
45
+ author={Or Hirschorn and Omer Sela and Inbar Huberman-Spiegelglas and Netalee Efrat and Eli Alshan and Ianir Ideses and Frederic Devernay and Yochai Zvik and Lior Fritz},
46
+ year={2025},
47
+ eprint={2512.09923},
48
+ archivePrefix={arXiv},
49
+ primaryClass={cs.CV},
50
+ url={https://arxiv.org/abs/2512.09923},
51
+ }
52
+ ```
53
+
54
+ ## Acknowledgments
55
+ We thank the excellent repositories of [Stable Diffusion](https://github.com/Stability-AI/stablediffusion), [Diffusers](https://github.com/huggingface/diffusers), [Diffix3d+](https://github.com/nv-tlabs/Difix3D), [Feature 3DGS](https://github.com/ShijieZhou-UCLA/feature-3dgs), [3D Gaussian Splatting](https://github.com/graphdeco-inria/gaussian-splatting), and [LPIPS](https://github.com/richzhang/PerceptualSimilarity).