|
|
import gradio as gr |
|
|
import os |
|
|
import shutil |
|
|
import subprocess |
|
|
import cv2 |
|
|
import numpy as np |
|
|
import math |
|
|
|
|
|
from huggingface_hub import snapshot_download |
|
|
|
|
|
os.environ['CUDA_LAUNCH_BLOCKING'] = '1' |
|
|
|
|
|
|
|
|
model_ids = [ |
|
|
'runwayml/stable-diffusion-v1-5', |
|
|
'lllyasviel/sd-controlnet-depth', |
|
|
'lllyasviel/sd-controlnet-canny', |
|
|
'lllyasviel/sd-controlnet-openpose', |
|
|
"lllyasviel/control_v11p_sd15_softedge", |
|
|
"lllyasviel/control_v11p_sd15_scribble", |
|
|
"lllyasviel/control_v11p_sd15_lineart_anime", |
|
|
"lllyasviel/control_v11p_sd15_lineart", |
|
|
"lllyasviel/control_v11f1p_sd15_depth", |
|
|
"lllyasviel/control_v11p_sd15_canny", |
|
|
"lllyasviel/control_v11p_sd15_openpose", |
|
|
"lllyasviel/control_v11p_sd15_normalbae" |
|
|
] |
|
|
|
|
|
|
|
|
for model_id in model_ids: |
|
|
model_name = model_id.split('/')[-1] |
|
|
snapshot_download(model_id, cache_dir=f'checkpoints/{model_name}') |
|
|
|
|
|
def load_model(model_id): |
|
|
local_dir = f'checkpoints/stable-diffusion-v1-5' |
|
|
|
|
|
if os.path.exists(local_dir): |
|
|
|
|
|
shutil.rmtree(local_dir) |
|
|
|
|
|
model_name = model_id.split('/')[-1] |
|
|
snapshot_download(model_id, local_dir=f'checkpoints/{model_name}') |
|
|
os.rename(f'checkpoints/{model_name}', f'checkpoints/stable-diffusion-v1-5') |
|
|
return "model loaded" |
|
|
|
|
|
def get_frame_count(filepath): |
|
|
if filepath is not None: |
|
|
video = cv2.VideoCapture(filepath) |
|
|
frame_count = int(video.get(cv2.CAP_PROP_FRAME_COUNT)) |
|
|
|
|
|
video.release() |
|
|
|
|
|
|
|
|
if frame_count > 100 : |
|
|
frame_count = 100 |
|
|
|
|
|
return gr.update(maximum=frame_count) |
|
|
|
|
|
else: |
|
|
return gr.update(value=1, maximum=100 ) |
|
|
|
|
|
def get_video_dimension(filepath): |
|
|
video = cv2.VideoCapture(filepath) |
|
|
width = int(video.get(cv2.CAP_PROP_FRAME_WIDTH)) |
|
|
height = int(video.get(cv2.CAP_PROP_FRAME_HEIGHT)) |
|
|
fps = int(video.get(cv2.CAP_PROP_FPS)) |
|
|
frame_count = int(video.get(cv2.CAP_PROP_FRAME_COUNT)) |
|
|
video.release() |
|
|
return width, height, fps, frame_count |
|
|
|
|
|
def resize_video(input_vid, output_vid, width, height, fps): |
|
|
print(f"RESIZING ...") |
|
|
|
|
|
video = cv2.VideoCapture(input_vid) |
|
|
|
|
|
|
|
|
fourcc = cv2.VideoWriter_fourcc(*'mp4v') |
|
|
output_video = cv2.VideoWriter(output_vid, fourcc, fps, (width, height)) |
|
|
|
|
|
while True: |
|
|
|
|
|
ret, frame = video.read() |
|
|
if not ret: |
|
|
break |
|
|
|
|
|
|
|
|
resized_frame = cv2.resize(frame, (width, height)) |
|
|
|
|
|
|
|
|
output_video.write(resized_frame) |
|
|
|
|
|
|
|
|
video.release() |
|
|
output_video.release() |
|
|
print(f"RESIZE VIDEO DONE!") |
|
|
return output_vid |
|
|
|
|
|
def make_nearest_multiple_of_32(number): |
|
|
remainder = number % 32 |
|
|
if remainder <= 16: |
|
|
number -= remainder |
|
|
else: |
|
|
number += 32 - remainder |
|
|
return number |
|
|
|
|
|
def change_video_fps(input_path): |
|
|
print(f"CHANGING FIANL OUTPUT FPS") |
|
|
cap = cv2.VideoCapture(input_path) |
|
|
|
|
|
if os.path.exists('output_video.mp4'): |
|
|
|
|
|
os.remove('output_video.mp4') |
|
|
output_path = 'output_video.mp4' |
|
|
fourcc = cv2.VideoWriter_fourcc(*'mp4v') |
|
|
output_fps = 12 |
|
|
output_size = (int(cap.get(cv2.CAP_PROP_FRAME_WIDTH)), int(cap.get(cv2.CAP_PROP_FRAME_HEIGHT))) |
|
|
out = cv2.VideoWriter(output_path, fourcc, output_fps, output_size) |
|
|
|
|
|
frame_count = 0 |
|
|
while cap.isOpened(): |
|
|
ret, frame = cap.read() |
|
|
if not ret: |
|
|
break |
|
|
|
|
|
|
|
|
for _ in range(output_fps // 8): |
|
|
out.write(frame) |
|
|
|
|
|
frame_count += 1 |
|
|
print(f'Processed frame {frame_count}') |
|
|
|
|
|
cap.release() |
|
|
out.release() |
|
|
cv2.destroyAllWindows() |
|
|
|
|
|
return 'output_video.mp4' |
|
|
|
|
|
def run_inference(prompt, video_path, version_condition, video_length, seed): |
|
|
|
|
|
seed = math.floor(seed) |
|
|
o_width = get_video_dimension(video_path)[0] |
|
|
o_height = get_video_dimension(video_path)[1] |
|
|
version, condition = version_condition.split("+") |
|
|
|
|
|
|
|
|
if o_width > 512 : |
|
|
|
|
|
n_height = int(o_height / o_width * 512) |
|
|
n_width = 512 |
|
|
else: |
|
|
n_height = o_height |
|
|
n_width = o_width |
|
|
|
|
|
|
|
|
r_width = make_nearest_multiple_of_32(n_width) |
|
|
r_height = make_nearest_multiple_of_32(n_height) |
|
|
print(f"multiple of 32 sizes : {r_width}x{r_height}") |
|
|
|
|
|
|
|
|
original_fps = get_video_dimension(video_path)[2] |
|
|
if original_fps > 12 : |
|
|
print(f"FPS is too high: {original_fps}") |
|
|
target_fps = 12 |
|
|
else : |
|
|
target_fps = original_fps |
|
|
print(f"NEW INPUT FPS: {target_fps}, NEW LENGTH: {video_length}") |
|
|
|
|
|
|
|
|
if os.path.exists('resized.mp4'): |
|
|
|
|
|
os.remove('resized.mp4') |
|
|
|
|
|
resized = resize_video(video_path, 'resized.mp4', r_width, r_height, target_fps) |
|
|
resized_video_fcount = get_video_dimension(resized)[3] |
|
|
print(f"RESIZED VIDEO FRAME COUNT: {resized_video_fcount}") |
|
|
|
|
|
|
|
|
if video_length > resized_video_fcount : |
|
|
video_length = resized_video_fcount |
|
|
|
|
|
|
|
|
output_path = 'output/' |
|
|
os.makedirs(output_path, exist_ok=True) |
|
|
|
|
|
|
|
|
if os.path.exists(os.path.join(output_path, f"result.mp4")): |
|
|
|
|
|
os.remove(os.path.join(output_path, f"result.mp4")) |
|
|
|
|
|
print(f"RUNNING INFERENCE ...") |
|
|
if video_length > 16: |
|
|
command = f"python inference.py --prompt '{prompt}' --condition '{condition}' --video_path '{resized}' --output_path '{output_path}' --temp_video_name 'result' --width {r_width} --height {r_height} --seed {seed} --video_length {video_length} --smoother_steps 19 20 --version {version} --is_long_video" |
|
|
else: |
|
|
command = f"python inference.py --prompt '{prompt}' --condition '{condition}' --video_path '{resized}' --output_path '{output_path}' --temp_video_name 'result' --width {r_width} --height {r_height} --seed {seed} --video_length {video_length} --smoother_steps 19 20 --version {version} " |
|
|
|
|
|
try: |
|
|
subprocess.run(command, shell=True) |
|
|
except cuda.Error as e: |
|
|
return f"CUDA Error: {e}", None |
|
|
except RuntimeError as e: |
|
|
return f"Runtime Error: {e}", None |
|
|
|
|
|
|
|
|
video_path_output = os.path.join(output_path, f"result.mp4") |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
gen_fps = get_video_dimension(video_path_output)[2] |
|
|
print(f"GEN VIDEO FPS: {gen_fps}") |
|
|
final = change_video_fps(video_path_output) |
|
|
print(f"FINISHED !") |
|
|
|
|
|
return final |
|
|
|
|
|
|
|
|
|
|
|
css=""" |
|
|
#col-container {max-width: 810px; margin-left: auto; margin-right: auto;} |
|
|
.animate-spin { |
|
|
animation: spin 1s linear infinite; |
|
|
} |
|
|
@keyframes spin { |
|
|
from { |
|
|
transform: rotate(0deg); |
|
|
} |
|
|
to { |
|
|
transform: rotate(360deg); |
|
|
} |
|
|
} |
|
|
#share-btn-container { |
|
|
display: flex; |
|
|
padding-left: 0.5rem !important; |
|
|
padding-right: 0.5rem !important; |
|
|
background-color: #000000; |
|
|
justify-content: center; |
|
|
align-items: center; |
|
|
border-radius: 9999px !important; |
|
|
max-width: 13rem; |
|
|
} |
|
|
#share-btn-container:hover { |
|
|
background-color: #060606; |
|
|
} |
|
|
#share-btn { |
|
|
all: initial; |
|
|
color: #ffffff; |
|
|
font-weight: 600; |
|
|
cursor:pointer; |
|
|
font-family: 'IBM Plex Sans', sans-serif; |
|
|
margin-left: 0.5rem !important; |
|
|
padding-top: 0.5rem !important; |
|
|
padding-bottom: 0.5rem !important; |
|
|
right:0; |
|
|
} |
|
|
#share-btn * { |
|
|
all: unset; |
|
|
} |
|
|
#share-btn-container div:nth-child(-n+2){ |
|
|
width: auto !important; |
|
|
min-height: 0px !important; |
|
|
} |
|
|
#share-btn-container .wrap { |
|
|
display: none !important; |
|
|
} |
|
|
#share-btn-container.hidden { |
|
|
display: none!important; |
|
|
} |
|
|
img[src*='#center'] { |
|
|
display: block; |
|
|
margin: auto; |
|
|
} |
|
|
""" |
|
|
with gr.Blocks(css=css) as demo: |
|
|
with gr.Column(elem_id="col-container"): |
|
|
gr.Markdown(""" |
|
|
<h1 style="text-align: center;">ControlVideo: Training-free Controllable Text-to-Video Generation</h1> |
|
|
<p style="text-align: center;"> |
|
|
[<a href="https://arxiv.org/abs/2305.13077" style="color:blue;">arXiv</a>] |
|
|
[<a href="https://github.com/YBYBZhang/ControlVideo" style="color:blue;">GitHub</a>] |
|
|
</p> |
|
|
<p style="text-align: center;"> ControlVideo adapts ControlNet to the video counterpart without any finetuning, aiming to directly inherit its high-quality and consistent generation. </p> |
|
|
""") |
|
|
|
|
|
with gr.Column(): |
|
|
with gr.Row(): |
|
|
video_path = gr.Video(label="Input video", source="upload", type="filepath", visible=True, elem_id="video-in") |
|
|
video_res = gr.Video(label="result", elem_id="video-out") |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
with gr.Row(): |
|
|
chosen_model = gr.Dropdown(label="Diffusion model (*1.5)", choices=['runwayml/stable-diffusion-v1-5','nitrosocke/Ghibli-Diffusion'], value="runwayml/stable-diffusion-v1-5", allow_custom_value=True) |
|
|
model_status = gr.Textbox(label="status") |
|
|
load_model_btn = gr.Button("load model (optional)") |
|
|
prompt = gr.Textbox(label="prompt", info="If you loaded a custom model, do not forget to include Prompt trigger", elem_id="prompt-in") |
|
|
with gr.Column(): |
|
|
video_length = gr.Slider(label="Video length", info="How many frames do you want to process ? For demo purpose, max is set to 24", minimum=1, maximum=12, step=1, value=2) |
|
|
with gr.Row(): |
|
|
|
|
|
version_condition = gr.Dropdown(label="ControlNet version + Condition", |
|
|
choices=["v10+depth_midas", "v10+canny", "v10+openpose", "v11+softedge_pidinet", "v11+softedge_pidsafe", |
|
|
"v11+softedge_hed", "v11+softedge_hedsafe", "v11+scribble_hed", "v11+scribble_pidinet", "v11+lineart_anime", |
|
|
"v11+lineart_coarse", "v11+lineart_realistic", "v11+depth_midas", "v11+depth_leres", "v11+depth_leres++", |
|
|
"v11+depth_zoe", "v11+canny", "v11+openpose", "v11+openpose_face", "v11+openpose_faceonly", "v11+openpose_full", |
|
|
"v11+openpose_hand", "v11+normal_bae"], value="v10+depth_midas") |
|
|
seed = gr.Number(label="seed", value=42) |
|
|
submit_btn = gr.Button("Submit") |
|
|
|
|
|
|
|
|
gr.Examples( |
|
|
examples=[["James bond moonwalks on the beach.", "./data/moonwalk.mp4", 'v10+openpose', 15, 42], |
|
|
["A striking mallard floats effortlessly on the sparkling pond.", "./data/mallard-water.mp4", "v11+depth_midas", 15, 42]], |
|
|
fn=run_inference, |
|
|
inputs=[prompt, |
|
|
video_path, |
|
|
version_condition, |
|
|
video_length, |
|
|
seed, |
|
|
], |
|
|
|
|
|
outputs=video_res, |
|
|
cache_examples=False |
|
|
) |
|
|
|
|
|
|
|
|
load_model_btn.click(fn=load_model, inputs=[chosen_model], outputs=[model_status], queue=False) |
|
|
video_path.change(fn=get_frame_count, |
|
|
inputs=[video_path], |
|
|
outputs=[video_length], |
|
|
queue=False |
|
|
) |
|
|
submit_btn.click(fn=run_inference, |
|
|
inputs=[prompt, |
|
|
video_path, |
|
|
version_condition, |
|
|
video_length, |
|
|
seed, |
|
|
], |
|
|
outputs=video_res) |
|
|
|
|
|
demo.queue(max_size=12).launch() |