Update app.py
Browse files
app.py
CHANGED
|
@@ -317,29 +317,9 @@ def generate(image_input, audio_input, pose_input, width, height, length, steps,
|
|
| 317 |
|
| 318 |
with gr.Blocks() as demo:
|
| 319 |
gr.Markdown("""
|
| 320 |
-
#
|
| 321 |
-
|
| 322 |
-
⚠️ This demonstration is for academic research and experiential use only.
|
| 323 |
""")
|
| 324 |
-
|
| 325 |
-
<div style="display:flex;column-gap:4px;">
|
| 326 |
-
<a href="https://github.com/antgroup/echomimic_v2">
|
| 327 |
-
<img src='https://img.shields.io/badge/GitHub-Repo-blue'>
|
| 328 |
-
</a>
|
| 329 |
-
<a href="https://antgroup.github.io/ai/echomimic_v2/">
|
| 330 |
-
<img src='https://img.shields.io/badge/Project-Page-green'>
|
| 331 |
-
</a>
|
| 332 |
-
<a href="https://arxiv.org/abs/2411.10061">
|
| 333 |
-
<img src='https://img.shields.io/badge/ArXiv-Paper-red'>
|
| 334 |
-
</a>
|
| 335 |
-
<a href="https://huggingface.co/spaces/fffiloni/echomimic-v2?duplicate=true">
|
| 336 |
-
<img src="https://huggingface.co/datasets/huggingface/badges/resolve/main/duplicate-this-space-sm.svg" alt="Duplicate this Space">
|
| 337 |
-
</a>
|
| 338 |
-
<a href="https://huggingface.co/fffiloni">
|
| 339 |
-
<img src="https://huggingface.co/datasets/huggingface/badges/resolve/main/follow-me-on-HF-sm-dark.svg" alt="Follow me on HF">
|
| 340 |
-
</a>
|
| 341 |
-
</div>
|
| 342 |
-
""")
|
| 343 |
with gr.Column():
|
| 344 |
with gr.Row():
|
| 345 |
with gr.Column():
|
|
@@ -369,13 +349,7 @@ with gr.Blocks() as demo:
|
|
| 369 |
seed_text = gr.Textbox(label="Seed", interactive=False, visible=False)
|
| 370 |
gr.Examples(
|
| 371 |
examples=[
|
| 372 |
-
["EMTD_dataset/ref_imgs_by_FLUX/man/0001.png", "assets/halfbody_demo/audio/chinese/echomimicv2_man.wav"]
|
| 373 |
-
["EMTD_dataset/ref_imgs_by_FLUX/woman/0077.png", "assets/halfbody_demo/audio/chinese/echomimicv2_woman.wav"],
|
| 374 |
-
["EMTD_dataset/ref_imgs_by_FLUX/man/0003.png", "assets/halfbody_demo/audio/chinese/fighting.wav"],
|
| 375 |
-
["EMTD_dataset/ref_imgs_by_FLUX/woman/0033.png", "assets/halfbody_demo/audio/chinese/good.wav"],
|
| 376 |
-
["EMTD_dataset/ref_imgs_by_FLUX/man/0010.png", "assets/halfbody_demo/audio/chinese/news.wav"],
|
| 377 |
-
["EMTD_dataset/ref_imgs_by_FLUX/man/1168.png", "assets/halfbody_demo/audio/chinese/no_smoking.wav"],
|
| 378 |
-
["EMTD_dataset/ref_imgs_by_FLUX/woman/0057.png", "assets/halfbody_demo/audio/chinese/ultraman.wav"]
|
| 379 |
],
|
| 380 |
inputs=[image_input, audio_input],
|
| 381 |
label="Preset Characters and Audio",
|
|
|
|
| 317 |
|
| 318 |
with gr.Blocks() as demo:
|
| 319 |
gr.Markdown("""
|
| 320 |
+
# GiniGen Face2Video
|
|
|
|
|
|
|
| 321 |
""")
|
| 322 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 323 |
with gr.Column():
|
| 324 |
with gr.Row():
|
| 325 |
with gr.Column():
|
|
|
|
| 349 |
seed_text = gr.Textbox(label="Seed", interactive=False, visible=False)
|
| 350 |
gr.Examples(
|
| 351 |
examples=[
|
| 352 |
+
["EMTD_dataset/ref_imgs_by_FLUX/man/0001.png", "assets/halfbody_demo/audio/chinese/echomimicv2_man.wav"]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 353 |
],
|
| 354 |
inputs=[image_input, audio_input],
|
| 355 |
label="Preset Characters and Audio",
|