Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -4,9 +4,8 @@ os.environ['HF_HOME'] = os.path.join(os.path.dirname(__file__), 'hf_download')
|
|
| 4 |
result_dir = os.path.join('./', 'results')
|
| 5 |
os.makedirs(result_dir, exist_ok=True)
|
| 6 |
|
| 7 |
-
|
| 8 |
import functools
|
| 9 |
-
import os
|
| 10 |
import random
|
| 11 |
import gradio as gr
|
| 12 |
import numpy as np
|
|
@@ -23,7 +22,6 @@ from diffusers.models.attention_processor import AttnProcessor2_0
|
|
| 23 |
from transformers import CLIPTextModel, CLIPTokenizer
|
| 24 |
from diffusers_vdm.pipeline import LatentVideoDiffusionPipeline
|
| 25 |
from diffusers_vdm.utils import resize_and_center_crop, save_bcthw_as_mp4
|
| 26 |
-
import spaces
|
| 27 |
|
| 28 |
# Disable gradients globally
|
| 29 |
torch.set_grad_enabled(False)
|
|
|
|
| 4 |
result_dir = os.path.join('./', 'results')
|
| 5 |
os.makedirs(result_dir, exist_ok=True)
|
| 6 |
|
| 7 |
+
import spaces
|
| 8 |
import functools
|
|
|
|
| 9 |
import random
|
| 10 |
import gradio as gr
|
| 11 |
import numpy as np
|
|
|
|
| 22 |
from transformers import CLIPTextModel, CLIPTokenizer
|
| 23 |
from diffusers_vdm.pipeline import LatentVideoDiffusionPipeline
|
| 24 |
from diffusers_vdm.utils import resize_and_center_crop, save_bcthw_as_mp4
|
|
|
|
| 25 |
|
| 26 |
# Disable gradients globally
|
| 27 |
torch.set_grad_enabled(False)
|