Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,39 @@
|
|
| 1 |
-
|
| 2 |
-
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
<div align="center">
|
| 2 |
+
|
| 3 |
+
# Assembler: Scalable 3D Part Assembly via Anchor Point Diffusion
|
| 4 |
+
|
| 5 |
+
<a href="https://assembler3d.github.io/"><img src="https://img.shields.io/static/v1?label=Project%20Page&message=Github&color=blue&logo=github-pages"></a> <a href="https://arxiv.org/abs/2506.17074"><img src="https://img.shields.io/badge/ArXiv-2506.17074-brightgreen"></a> <br>
|
| 6 |
+
|
| 7 |
+
**[Wang Zhao<sup>1</sup>](https://thuzhaowang.github.io), [Yan-Pei Cao<sup>2</sup>](https://yanpei.me/), [Jiale Xu<sup>1</sup>](https://bluestyle97.github.io/), [Yuejiang Dong<sup>1,3</sup>](https://scholar.google.com.hk/citations?user=0i7bPj8AAAAJ&hl=zh-CN), [Ying Shan<sup>1</sup>](https://scholar.google.com/citations?user=4oXBp9UAAAAJ&hl=en)**
|
| 8 |
+
|
| 9 |
+
<sup>1</sup>ARC Lab, Tencent PCG   <sup>2</sup>VAST   <sup>3</sup>Tsinghua University
|
| 10 |
+
|
| 11 |
+
|
| 12 |
+
**SIGGRAPH ASIA 2025**
|
| 13 |
+
|
| 14 |
+
</div>
|
| 15 |
+
|
| 16 |
+
---
|
| 17 |
+
|
| 18 |
+
## 🚩 Overview
|
| 19 |
+
This repository contains code release for our SIGGRAPH ASIA 2025 paper "Assembler: Scalable 3D Part Assembly via Anchor Point Diffusion".
|
| 20 |
+
|
| 21 |
+
|
| 22 |
+
## ⚙️ Installation
|
| 23 |
+
We recommend using anaconda to install the dependencies:
|
| 24 |
+
```
|
| 25 |
+
conda create -n assembler python=3.10.16
|
| 26 |
+
conda activate assembler
|
| 27 |
+
conda install pytorch==2.6.0 torchvision==0.21.0 torchaudio==2.6.0 pytorch-cuda=12.4 -c pytorch -c nvidia
|
| 28 |
+
pip install -r requirements.txt
|
| 29 |
+
```
|
| 30 |
+
|
| 31 |
+
## 🚀 Usage
|
| 32 |
+
### Inference
|
| 33 |
+
To run the inference demo, simply use:
|
| 34 |
+
```
|
| 35 |
+
python ./scripts/demo.py --config ./configs/demo/demo.yaml --input_dir ./examples/4ef447cbb4a72f0a0e5941c9073c4baa0babd3f93ec55d62b040915f8bf3f49c --output_dir ./outputs/4ef447
|
| 36 |
+
```
|
| 37 |
+
This script runs for example data inside `./examples` from Toys4k dataset. You could use your own data to assemble. Put all the part meshes (in GLB format) and reference image (in PNG format) into a single folder, and set the `input_dir` argument to that folder.
|
| 38 |
+
|
| 39 |
+
```
|