Improve model card: add metadata, library name, and paper/code links

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +34 -1
README.md CHANGED
@@ -1,4 +1,37 @@
1
  ---
2
  license: mit
 
 
 
 
 
 
3
  ---
4
- arxiv.org/abs/2508.07409
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
+ library_name: diffusers
4
+ pipeline_tag: image-to-video
5
+ tags:
6
+ - character-animation
7
+ - 4d
8
+ - image-to-video
9
  ---
10
+ ---
11
+
12
+ # CharacterShot: Controllable and Consistent 4D Character Animation
13
+
14
+ [**CharacterShot**](https://arxiv.org/abs/2508.07409) is a controllable and consistent 4D character animation framework that enables the creation of dynamic 3D characters (i.e., 4D character animation) from a single reference character image and a 2D pose sequence.
15
+
16
+ - **Paper:** [CharacterShot: Controllable and Consistent 4D Character Animation](https://arxiv.org/abs/2508.07409)
17
+ - **Repository:** [https://github.com/Jeoyal/CharacterShot](https://github.com/Jeoyal/CharacterShot)
18
+
19
+ ## Introduction
20
+
21
+ CharacterShot utilizes a powerful 2D character animation model based on a DiT image-to-video architecture. It lifts these animations to 3D using dual-attention modules and camera priors to ensure spatial-temporal and spatial-view consistency. The final representation is optimized using neighbor-constrained 4D Gaussian Splatting, resulting in stable and continuous character representations.
22
+
23
+ The model was trained on **Character4D**, a large-scale dataset containing 13,115 unique characters with diverse appearances and motions.
24
+
25
+ ## Citation
26
+
27
+ ```bibtex
28
+ @article{gao2025charactershot,
29
+ title={CharacterShot: Controllable and Consistent 4D Character Animation},
30
+ author={Gao, Junyao and Li, Jiaxing and Liu, Wenran and Zeng, Yanhong and Shen, Fei and Chen, Kai and Sun, Yanan and Zhao, Cairong},
31
+ journal={arXiv preprint arXiv:2508.07409},
32
+ year={2025}
33
+ }
34
+ ```
35
+
36
+ ## Acknowledgements
37
+ The code is built upon [CogVideo](https://github.com/THUDM/CogVideo).