|
|
|
|
|
|
|
|
import os |
|
|
import numpy as np |
|
|
import argparse |
|
|
import imageio |
|
|
import torch |
|
|
|
|
|
from einops import rearrange |
|
|
from diffusers import DDIMScheduler, AutoencoderKL |
|
|
from transformers import CLIPTextModel, CLIPTokenizer |
|
|
import controlnet_aux |
|
|
from controlnet_aux import OpenposeDetector, CannyDetector, MidasDetector |
|
|
|
|
|
from models.pipeline_controlvideo import ControlVideoPipeline |
|
|
from models.util import save_videos_grid, read_video, get_annotation |
|
|
from models.unet import UNet3DConditionModel |
|
|
from models.controlnet import ControlNetModel3D |
|
|
from models.RIFE.IFNet_HDv3 import IFNet |
|
|
from cog import BasePredictor, Input, Path |
|
|
|
|
|
|
|
|
sd_path = "checkpoints/stable-diffusion-v1-5" |
|
|
inter_path = "checkpoints/flownet.pkl" |
|
|
controlnet_dict = { |
|
|
"pose": "checkpoints/sd-controlnet-openpose", |
|
|
"depth": "checkpoints/sd-controlnet-depth", |
|
|
"canny": "checkpoints/sd-controlnet-canny", |
|
|
} |
|
|
|
|
|
controlnet_parser_dict = { |
|
|
"pose": OpenposeDetector, |
|
|
"depth": MidasDetector, |
|
|
"canny": CannyDetector, |
|
|
} |
|
|
|
|
|
POS_PROMPT = " ,best quality, extremely detailed, HD, ultra-realistic, 8K, HQ, masterpiece, trending on artstation, art, smooth" |
|
|
NEG_PROMPT = "longbody, lowres, bad anatomy, bad hands, missing fingers, extra digit, fewer digits, cropped, worst quality, low quality, deformed body, bloated, ugly, unrealistic" |
|
|
|
|
|
|
|
|
class Predictor(BasePredictor): |
|
|
def setup(self): |
|
|
"""Load the model into memory to make running multiple predictions efficient""" |
|
|
|
|
|
self.tokenizer = CLIPTokenizer.from_pretrained(sd_path, subfolder="tokenizer") |
|
|
self.text_encoder = CLIPTextModel.from_pretrained( |
|
|
sd_path, subfolder="text_encoder" |
|
|
).to(dtype=torch.float16) |
|
|
self.vae = AutoencoderKL.from_pretrained(sd_path, subfolder="vae").to( |
|
|
dtype=torch.float16 |
|
|
) |
|
|
self.unet = UNet3DConditionModel.from_pretrained_2d( |
|
|
sd_path, subfolder="unet" |
|
|
).to(dtype=torch.float16) |
|
|
self.interpolater = IFNet(ckpt_path=inter_path).to(dtype=torch.float16) |
|
|
self.scheduler = DDIMScheduler.from_pretrained(sd_path, subfolder="scheduler") |
|
|
self.controlnet = { |
|
|
k: ControlNetModel3D.from_pretrained_2d(controlnet_dict[k]).to( |
|
|
dtype=torch.float16 |
|
|
) |
|
|
for k in ["depth", "canny", "pose"] |
|
|
} |
|
|
self.annotator = {k: controlnet_parser_dict[k]() for k in ["depth", "canny"]} |
|
|
self.annotator["pose"] = controlnet_parser_dict["pose"].from_pretrained( |
|
|
"lllyasviel/ControlNet", cache_dir="checkpoints" |
|
|
) |
|
|
|
|
|
def predict( |
|
|
self, |
|
|
prompt: str = Input( |
|
|
description="Text description of target video", |
|
|
default="A striking mallard floats effortlessly on the sparkling pond.", |
|
|
), |
|
|
video_path: Path = Input(description="source video"), |
|
|
condition: str = Input( |
|
|
default="depth", |
|
|
choices=["depth", "canny", "pose"], |
|
|
description="Condition of structure sequence", |
|
|
), |
|
|
video_length: int = Input( |
|
|
default=15, description="Length of synthesized video" |
|
|
), |
|
|
smoother_steps: str = Input( |
|
|
default="19, 20", |
|
|
description="Timesteps at which using interleaved-frame smoother, separate with comma", |
|
|
), |
|
|
is_long_video: bool = Input( |
|
|
default=False, |
|
|
description="Whether to use hierarchical sampler to produce long video", |
|
|
), |
|
|
num_inference_steps: int = Input( |
|
|
description="Number of denoising steps", default=50 |
|
|
), |
|
|
guidance_scale: float = Input( |
|
|
description="Scale for classifier-free guidance", ge=1, le=20, default=12.5 |
|
|
), |
|
|
seed: str = Input( |
|
|
default=None, description="Random seed. Leave blank to randomize the seed" |
|
|
), |
|
|
) -> Path: |
|
|
"""Run a single prediction on the model""" |
|
|
if seed is None: |
|
|
seed = int.from_bytes(os.urandom(2), "big") |
|
|
else: |
|
|
seed = int(seed) |
|
|
print(f"Using seed: {seed}") |
|
|
|
|
|
generator = torch.Generator(device="cuda") |
|
|
generator.manual_seed(seed) |
|
|
|
|
|
pipe = ControlVideoPipeline( |
|
|
vae=self.vae, |
|
|
text_encoder=self.text_encoder, |
|
|
tokenizer=self.tokenizer, |
|
|
unet=self.unet, |
|
|
controlnet=self.controlnet[condition], |
|
|
interpolater=self.interpolater, |
|
|
scheduler=self.scheduler, |
|
|
) |
|
|
|
|
|
pipe.enable_vae_slicing() |
|
|
pipe.enable_xformers_memory_efficient_attention() |
|
|
pipe.to("cuda") |
|
|
|
|
|
|
|
|
video = read_video(video_path=str(video_path), video_length=video_length) |
|
|
|
|
|
|
|
|
pil_annotation = get_annotation(video, self.annotator[condition]) |
|
|
|
|
|
|
|
|
smoother_steps = [int(s) for s in smoother_steps.split(",")] |
|
|
|
|
|
if is_long_video: |
|
|
window_size = int(np.sqrt(video_length)) |
|
|
sample = pipe.generate_long_video( |
|
|
prompt + POS_PROMPT, |
|
|
video_length=video_length, |
|
|
frames=pil_annotation, |
|
|
num_inference_steps=num_inference_steps, |
|
|
smooth_steps=smoother_steps, |
|
|
window_size=window_size, |
|
|
generator=generator, |
|
|
guidance_scale=guidance_scale, |
|
|
negative_prompt=NEG_PROMPT, |
|
|
).videos |
|
|
else: |
|
|
sample = pipe( |
|
|
prompt + POS_PROMPT, |
|
|
video_length=video_length, |
|
|
frames=pil_annotation, |
|
|
num_inference_steps=num_inference_steps, |
|
|
smooth_steps=smoother_steps, |
|
|
generator=generator, |
|
|
guidance_scale=guidance_scale, |
|
|
negative_prompt=NEG_PROMPT, |
|
|
).videos |
|
|
|
|
|
out_path = "/tmp/out.mp4" |
|
|
save_videos_grid(sample, out_path) |
|
|
del pipe |
|
|
torch.cuda.empty_cache() |
|
|
|
|
|
return Path(out_path) |
|
|
|