SetuG commited on
Commit
22700ae
·
verified ·
1 Parent(s): b75f6e9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +88 -88
README.md CHANGED
@@ -1,89 +1,89 @@
1
- ---
2
- title: SadTalker
3
- emoji: 😭
4
- colorFrom: blue
5
- colorTo: red
6
- sdk: gradio
7
- sdk_version: 3.0.0
8
- app_file: app.py
9
- pinned: false
10
- license: mit
11
- ---
12
-
13
- <!-- Alternative deployment options:
14
-
15
- For Streamlit:
16
- sdk: streamlit
17
- app_file: app_streamlit.py
18
-
19
- For FastAPI:
20
- sdk: docker
21
- app_port: 7860
22
-
23
- For Docker:
24
- sdk: docker
25
- app_port: 7860
26
- -->
27
-
28
- # SadTalker: Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation (CVPR 2023)
29
-
30
- This is a Gradio app for SadTalker, which can generate talking face videos from a single image and audio.
31
-
32
- ## Features
33
- - Generate talking face videos from single image + audio
34
- - Multiple preprocessing options
35
- - Face enhancement with GFPGAN
36
- - Multiple pose styles
37
- - Still mode for fewer head movements
38
-
39
- ## Setup
40
-
41
- After cloning this repository to your Hugging Face Space, you'll need to:
42
-
43
- 1. **Upload model files**: Download the following model files and upload them to your repository:
44
-
45
- ### Required Model Files:
46
-
47
- #### SadTalker Models (upload to `checkpoints/` folder):
48
- - `SadTalker_V0.0.2_256.safetensors`
49
- - `SadTalker_V0.0.2_512.safetensors`
50
- - `mapping_00109-model.pth.tar`
51
- - `mapping_00229-model.pth.tar`
52
-
53
- #### GFPGAN Models (upload to `gfpgan/weights/` folder):
54
- - `alignment_WFLW_4HG.pth`
55
- - `detection_Resnet50_Final.pth`
56
- - `GFPGANv1.4.pth`
57
- - `parsing_parsenet.pth`
58
-
59
- ### Where to get the models:
60
- 1. Download from the original SadTalker repository: https://github.com/OpenTalker/SadTalker
61
- 2. Or from the model links provided in their documentation
62
-
63
- ### Upload Instructions:
64
- 1. Go to your Hugging Face Space repository
65
- 2. Click "Upload files"
66
- 3. Create the folder structure and upload the model files
67
- 4. Make sure the files are in the correct paths as listed above
68
-
69
- ## Usage
70
- 1. Upload a source image (preferably a portrait with clear face)
71
- 2. Upload an audio file
72
- 3. Adjust settings as needed
73
- 4. Click Generate to create your talking face video
74
-
75
- ## Citation
76
- ```
77
- @InProceedings{zhang2023sadtalker,
78
- author={Zhang, Wenxuan and Cun, Xiaodong and Wang, Xuan and Zhang, Yong and Shen, Xi and Guo, Yu and Shan, Ying and Wang, Fei},
79
- title={SadTalker: Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation},
80
- booktitle={The IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
81
- month={June},
82
- year={2023}
83
- }
84
- ```
85
-
86
- ## Links
87
- - [Paper](https://arxiv.org/abs/2211.12194)
88
- - [Project Page](https://sadtalker.github.io)
89
  - [Original Repository](https://github.com/OpenTalker/SadTalker)
 
1
+ ---
2
+ title: SadTalker
3
+ emoji: 😭
4
+ colorFrom: blue
5
+ colorTo: red
6
+ sdk: gradio
7
+ sdk_version: 6.5.1
8
+ app_file: app.py
9
+ pinned: false
10
+ license: mit
11
+ ---
12
+
13
+ <!-- Alternative deployment options:
14
+
15
+ For Streamlit:
16
+ sdk: streamlit
17
+ app_file: app_streamlit.py
18
+
19
+ For FastAPI:
20
+ sdk: docker
21
+ app_port: 7860
22
+
23
+ For Docker:
24
+ sdk: docker
25
+ app_port: 7860
26
+ -->
27
+
28
+ # SadTalker: Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation (CVPR 2023)
29
+
30
+ This is a Gradio app for SadTalker, which can generate talking face videos from a single image and audio.
31
+
32
+ ## Features
33
+ - Generate talking face videos from single image + audio
34
+ - Multiple preprocessing options
35
+ - Face enhancement with GFPGAN
36
+ - Multiple pose styles
37
+ - Still mode for fewer head movements
38
+
39
+ ## Setup
40
+
41
+ After cloning this repository to your Hugging Face Space, you'll need to:
42
+
43
+ 1. **Upload model files**: Download the following model files and upload them to your repository:
44
+
45
+ ### Required Model Files:
46
+
47
+ #### SadTalker Models (upload to `checkpoints/` folder):
48
+ - `SadTalker_V0.0.2_256.safetensors`
49
+ - `SadTalker_V0.0.2_512.safetensors`
50
+ - `mapping_00109-model.pth.tar`
51
+ - `mapping_00229-model.pth.tar`
52
+
53
+ #### GFPGAN Models (upload to `gfpgan/weights/` folder):
54
+ - `alignment_WFLW_4HG.pth`
55
+ - `detection_Resnet50_Final.pth`
56
+ - `GFPGANv1.4.pth`
57
+ - `parsing_parsenet.pth`
58
+
59
+ ### Where to get the models:
60
+ 1. Download from the original SadTalker repository: https://github.com/OpenTalker/SadTalker
61
+ 2. Or from the model links provided in their documentation
62
+
63
+ ### Upload Instructions:
64
+ 1. Go to your Hugging Face Space repository
65
+ 2. Click "Upload files"
66
+ 3. Create the folder structure and upload the model files
67
+ 4. Make sure the files are in the correct paths as listed above
68
+
69
+ ## Usage
70
+ 1. Upload a source image (preferably a portrait with clear face)
71
+ 2. Upload an audio file
72
+ 3. Adjust settings as needed
73
+ 4. Click Generate to create your talking face video
74
+
75
+ ## Citation
76
+ ```
77
+ @InProceedings{zhang2023sadtalker,
78
+ author={Zhang, Wenxuan and Cun, Xiaodong and Wang, Xuan and Zhang, Yong and Shen, Xi and Guo, Yu and Shan, Ying and Wang, Fei},
79
+ title={SadTalker: Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation},
80
+ booktitle={The IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
81
+ month={June},
82
+ year={2023}
83
+ }
84
+ ```
85
+
86
+ ## Links
87
+ - [Paper](https://arxiv.org/abs/2211.12194)
88
+ - [Project Page](https://sadtalker.github.io)
89
  - [Original Repository](https://github.com/OpenTalker/SadTalker)