nielsr HF Staff commited on
Commit
af602e4
·
verified ·
1 Parent(s): 77be88b

Improve model card: add metadata, library name, and paper/code links

Browse files

Hi! I'm Niels from the community science team at Hugging Face.

This PR improves the model card for CharacterShot. Key changes include:
- Adding the `pipeline_tag: image-to-video` and `library_name: diffusers` metadata to ensure the model is discoverable and provides the correct code snippets.
- Adding links to the paper and the official GitHub repository.
- Providing a concise summary of the framework and a BibTeX citation.

Please let me know if you have any questions!

Files changed (1) hide show
  1. README.md +34 -1
README.md CHANGED
@@ -1,4 +1,37 @@
1
  ---
2
  license: mit
 
 
 
 
 
 
3
  ---
4
- arxiv.org/abs/2508.07409
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
+ library_name: diffusers
4
+ pipeline_tag: image-to-video
5
+ tags:
6
+ - character-animation
7
+ - 4d
8
+ - image-to-video
9
  ---
10
+ ---
11
+
12
+ # CharacterShot: Controllable and Consistent 4D Character Animation
13
+
14
+ [**CharacterShot**](https://arxiv.org/abs/2508.07409) is a controllable and consistent 4D character animation framework that enables the creation of dynamic 3D characters (i.e., 4D character animation) from a single reference character image and a 2D pose sequence.
15
+
16
+ - **Paper:** [CharacterShot: Controllable and Consistent 4D Character Animation](https://arxiv.org/abs/2508.07409)
17
+ - **Repository:** [https://github.com/Jeoyal/CharacterShot](https://github.com/Jeoyal/CharacterShot)
18
+
19
+ ## Introduction
20
+
21
+ CharacterShot utilizes a powerful 2D character animation model based on a DiT image-to-video architecture. It lifts these animations to 3D using dual-attention modules and camera priors to ensure spatial-temporal and spatial-view consistency. The final representation is optimized using neighbor-constrained 4D Gaussian Splatting, resulting in stable and continuous character representations.
22
+
23
+ The model was trained on **Character4D**, a large-scale dataset containing 13,115 unique characters with diverse appearances and motions.
24
+
25
+ ## Citation
26
+
27
+ ```bibtex
28
+ @article{gao2025charactershot,
29
+ title={CharacterShot: Controllable and Consistent 4D Character Animation},
30
+ author={Gao, Junyao and Li, Jiaxing and Liu, Wenran and Zeng, Yanhong and Shen, Fei and Chen, Kai and Sun, Yanan and Zhao, Cairong},
31
+ journal={arXiv preprint arXiv:2508.07409},
32
+ year={2025}
33
+ }
34
+ ```
35
+
36
+ ## Acknowledgements
37
+ The code is built upon [CogVideo](https://github.com/THUDM/CogVideo).