Improve model card title and integrate paper information
Browse filesHi, I'm Niels from the Hugging Face community team. This PR improves the model card for AniGen by:
1. Updating the title to the full paper title for better discoverability.
2. Adding the training dataset (`VAST-AI/AniGen-Sample-Dataset`) to the YAML metadata.
3. Linking the model checkpoints to the research paper and authors for better attribution.
4. Refined the layout to ensure easy navigation for users.
README.md
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
---
|
| 2 |
-
license: mit
|
| 3 |
-
language:
|
| 4 |
-
- en
|
| 5 |
base_model:
|
| 6 |
- microsoft/TRELLIS-image-large
|
|
|
|
|
|
|
|
|
|
| 7 |
pipeline_tag: image-to-3d
|
| 8 |
tags:
|
| 9 |
- animatable
|
|
@@ -11,10 +11,15 @@ tags:
|
|
| 11 |
- 3D
|
| 12 |
- Tripo
|
| 13 |
- VAST
|
|
|
|
|
|
|
| 14 |
---
|
| 15 |
-
# AniGen_Weights
|
| 16 |
|
| 17 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 18 |
|
| 19 |
<p align="center">
|
| 20 |
<a href="https://arxiv.org/pdf/2604.08746"><img src="https://img.shields.io/badge/arXiv-Paper-red?logo=arxiv&logoColor=white" alt="arXiv"></a>
|
|
@@ -24,8 +29,9 @@ Pretrained checkpoints for [AniGen](https://github.com/VAST-AI-Research/AniGen),
|
|
| 24 |
<a href="https://github.com/VAST-AI-Research/AniGen"><img src="https://img.shields.io/badge/GitHub-Repository-black?logo=github&logoColor=white" alt="GitHub"></a>
|
| 25 |
</p>
|
| 26 |
|
| 27 |
-
|
| 28 |
|
|
|
|
| 29 |
- AniGen stage checkpoints
|
| 30 |
- DINOv2 vision encoder weights
|
| 31 |
- DSINE normal estimation weights
|
|
@@ -33,7 +39,7 @@ This repository stores the contents of the `ckpts/` directory used by the AniGen
|
|
| 33 |
|
| 34 |
## What Is Included
|
| 35 |
|
| 36 |
-
The repository is organized exactly like the `ckpts/` folder expected by the main AniGen repo:
|
| 37 |
|
| 38 |
```text
|
| 39 |
ckpts/
|
|
@@ -56,11 +62,8 @@ Approximate total size: about 23 GB.
|
|
| 56 |
## Recommended Checkpoints
|
| 57 |
|
| 58 |
For most users, we recommend:
|
| 59 |
-
|
| 60 |
-
- `
|
| 61 |
-
- `slat_flow_auto` for structured latent generation
|
| 62 |
-
|
| 63 |
-
This combination is also the default setup used by the AniGen inference example.
|
| 64 |
|
| 65 |
## Checkpoint Overview
|
| 66 |
|
|
@@ -87,37 +90,24 @@ This combination is also the default setup used by the AniGen inference example.
|
|
| 87 |
|
| 88 |
## How To Use
|
| 89 |
|
| 90 |
-
|
| 91 |
|
| 92 |
```bash
|
| 93 |
git clone --recurse-submodules https://github.com/VAST-AI-Research/AniGen.git
|
| 94 |
cd AniGen
|
| 95 |
```
|
| 96 |
|
| 97 |
-
|
| 98 |
|
| 99 |
-
|
| 100 |
|
| 101 |
```bash
|
| 102 |
python -c "from huggingface_hub import snapshot_download; snapshot_download(repo_id='VAST-AI/AniGen_Weights', repo_type='model', local_dir='.', local_dir_use_symlinks=False)"
|
| 103 |
```
|
| 104 |
|
| 105 |
-
|
| 106 |
-
|
| 107 |
-
```text
|
| 108 |
-
ckpts/anigen/ss_flow_duet/ckpts/denoiser.pt
|
| 109 |
-
ckpts/anigen/slat_flow_auto/ckpts/denoiser.pt
|
| 110 |
-
ckpts/dsine/dsine.pt
|
| 111 |
-
ckpts/vgg/vgg16-397923af.pth
|
| 112 |
-
```
|
| 113 |
-
|
| 114 |
-
### Option 2: Download from the web UI
|
| 115 |
-
|
| 116 |
-
You can also download this repository from the Hugging Face file browser and place the entire `ckpts/` folder at the root of the AniGen project.
|
| 117 |
|
| 118 |
-
|
| 119 |
-
|
| 120 |
-
Once the `ckpts/` folder is in place, you can run:
|
| 121 |
|
| 122 |
```bash
|
| 123 |
python example.py --image_path assets/cond_images/trex.png
|
|
@@ -129,20 +119,6 @@ Or launch the Gradio demo:
|
|
| 129 |
python app.py
|
| 130 |
```
|
| 131 |
|
| 132 |
-
## Notes
|
| 133 |
-
|
| 134 |
-
- Keep the directory names unchanged. The AniGen code expects the exact `ckpts/...` layout shown above.
|
| 135 |
-
- The code repository may automatically fetch missing files in some setups, but this weights repository is the recommended way to pre-download and manage checkpoints explicitly.
|
| 136 |
-
- `slat_flow_control` supports joint density control, while `slat_flow_auto` is the best default for general use.
|
| 137 |
-
|
| 138 |
-
## Related Links
|
| 139 |
-
|
| 140 |
-
- Best AI 3D studio -- Tripo: https://www.tripo3d.ai
|
| 141 |
-
- Main code repository: https://github.com/VAST-AI-Research/AniGen
|
| 142 |
-
- Project page: https://yihua7.github.io/AniGen-web/
|
| 143 |
-
- Demo: https://huggingface.co/spaces/VAST-AI/AniGen
|
| 144 |
-
- Paper: https://arxiv.org/pdf/2604.08746
|
| 145 |
-
|
| 146 |
## Citation
|
| 147 |
|
| 148 |
```bibtex
|
|
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
| 2 |
base_model:
|
| 3 |
- microsoft/TRELLIS-image-large
|
| 4 |
+
language:
|
| 5 |
+
- en
|
| 6 |
+
license: mit
|
| 7 |
pipeline_tag: image-to-3d
|
| 8 |
tags:
|
| 9 |
- animatable
|
|
|
|
| 11 |
- 3D
|
| 12 |
- Tripo
|
| 13 |
- VAST
|
| 14 |
+
datasets:
|
| 15 |
+
- VAST-AI/AniGen-Sample-Dataset
|
| 16 |
---
|
|
|
|
| 17 |
|
| 18 |
+
# AniGen: Unified $S^3$ Fields for Animatable 3D Asset Generation
|
| 19 |
+
|
| 20 |
+
Pretrained checkpoints for **AniGen**, a unified framework for generating animatable 3D assets from a single image.
|
| 21 |
+
|
| 22 |
+
**Authors**: Yi-Hua Huang, Zi-Xin Zou, Yuting He, Chirui Chang, Cheng-Feng Pu, Ziyi Yang, Yuan-Chen Guo, Yan-Pei Cao, Xiaojuan Qi.
|
| 23 |
|
| 24 |
<p align="center">
|
| 25 |
<a href="https://arxiv.org/pdf/2604.08746"><img src="https://img.shields.io/badge/arXiv-Paper-red?logo=arxiv&logoColor=white" alt="arXiv"></a>
|
|
|
|
| 29 |
<a href="https://github.com/VAST-AI-Research/AniGen"><img src="https://img.shields.io/badge/GitHub-Repository-black?logo=github&logoColor=white" alt="GitHub"></a>
|
| 30 |
</p>
|
| 31 |
|
| 32 |
+
AniGen represents shape, skeleton, and skinning as mutually consistent $S^3$ Fields (Shape, Skeleton, Skin) defined over a shared spatial domain. Built upon a two-stage flow-matching pipeline, it first synthesizes a sparse structural scaffold and then generates dense geometry and articulation in a structured latent space.
|
| 33 |
|
| 34 |
+
This repository stores the contents of the `ckpts/` directory used by the AniGen codebase, including:
|
| 35 |
- AniGen stage checkpoints
|
| 36 |
- DINOv2 vision encoder weights
|
| 37 |
- DSINE normal estimation weights
|
|
|
|
| 39 |
|
| 40 |
## What Is Included
|
| 41 |
|
| 42 |
+
The repository is organized exactly like the `ckpts/` folder expected by the [main AniGen repo](https://github.com/VAST-AI-Research/AniGen):
|
| 43 |
|
| 44 |
```text
|
| 45 |
ckpts/
|
|
|
|
| 62 |
## Recommended Checkpoints
|
| 63 |
|
| 64 |
For most users, we recommend:
|
| 65 |
+
- `ss_flow_duet` for sparse structure generation (stronger skeleton detail)
|
| 66 |
+
- `slat_flow_auto` for structured latent generation (automatic joint-count prediction)
|
|
|
|
|
|
|
|
|
|
| 67 |
|
| 68 |
## Checkpoint Overview
|
| 69 |
|
|
|
|
| 90 |
|
| 91 |
## How To Use
|
| 92 |
|
| 93 |
+
First, clone the main AniGen repository:
|
| 94 |
|
| 95 |
```bash
|
| 96 |
git clone --recurse-submodules https://github.com/VAST-AI-Research/AniGen.git
|
| 97 |
cd AniGen
|
| 98 |
```
|
| 99 |
|
| 100 |
+
### Download with `huggingface_hub`
|
| 101 |
|
| 102 |
+
Download this weights repository so that the folder structure is preserved under the project root:
|
| 103 |
|
| 104 |
```bash
|
| 105 |
python -c "from huggingface_hub import snapshot_download; snapshot_download(repo_id='VAST-AI/AniGen_Weights', repo_type='model', local_dir='.', local_dir_use_symlinks=False)"
|
| 106 |
```
|
| 107 |
|
| 108 |
+
### Run Inference
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 109 |
|
| 110 |
+
Once the `ckpts/` folder is in place, you can run the minimal example:
|
|
|
|
|
|
|
| 111 |
|
| 112 |
```bash
|
| 113 |
python example.py --image_path assets/cond_images/trex.png
|
|
|
|
| 119 |
python app.py
|
| 120 |
```
|
| 121 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 122 |
## Citation
|
| 123 |
|
| 124 |
```bibtex
|