File size: 1,615 Bytes
670f393 e7ce604 817b35b 78096ea 5e103a3 78096ea e7ce604 5e103a3 11c9442 5e103a3 78096ea e7ce604 11c9442 e7ce604 11c9442 e7ce604 78096ea e7ce604 11c9442 e7ce604 78096ea 11c9442 e7ce604 78096ea e7ce604 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 |
---
license: mit
---
# Unique3d-MVImage-Diffuser Model Card
[🌟GitHub](https://github.com/TingtingLiao/unique3d_diffuser) | [🦸 Project Page](https://wukailu.github.io/Unique3D/) | [🔋Normal Diffuser](https://huggingface.co/Luffuly/unique3d-normal-diffuser)</a>
## Example
Note the input image is required to be **white background**.

```bash
import torch
import numpy as np
from PIL import Image
from pipeline import StableDiffusionImage2MVCustomPipeline
pipe = Unique3dDiffusionPipeline.from_pretrained(
"Luffuly/unique3d-mvimage-diffuser",
torch_dtype=torch.float16,
trust_remote_code=True,
class_labels=torch.tensor(range(4)),
).to("cuda")
seed = -1
generator = torch.Generator(device='cuda').manual_seed(-1)
image = Image.open('data/boy.png')
forward_args = dict(
width=256,
height=256,
num_images_per_prompt=4,
num_inference_steps=50,
width_cond=256,
height_cond=256,
generator=generator,
guidance_scale=1.5,
)
out = pipe(image, **forward_args).images
rgb_np = np.hstack([np.array(img) for img in out])
Image.fromarray(rgb_np).save(f"mv-boy.png")
```
## Citation
```bash
@misc{wu2024unique3d,
title={Unique3D: High-Quality and Efficient 3D Mesh Generation from a Single Image},
author={Kailu Wu and Fangfu Liu and Zhihan Cai and Runjie Yan and Hanyang Wang and Yating Hu and Yueqi Duan and Kaisheng Ma},
year={2024},
eprint={2405.20343},
archivePrefix={arXiv},
primaryClass={cs.CV}
}
```
|