Update app.py--no-half-vae
Browse files- app.py--no-half-vae +1 -2
app.py--no-half-vae
CHANGED
|
@@ -1,4 +1,4 @@
|
|
| 1 |
-
from diffusers import
|
| 2 |
import gradio as gr
|
| 3 |
import torch
|
| 4 |
from PIL import Image
|
|
@@ -27,7 +27,6 @@ def error_str(error, title="Error"):
|
|
| 27 |
{error}""" if error else ""
|
| 28 |
|
| 29 |
def inference(prompt, guidance, steps, width=512, height=512, seed=0, img=None, strength=0.5, neg_prompt="", auto_prefix=False):
|
| 30 |
-
os.system("python3 -m --no-half-vae -o out")
|
| 31 |
generator = torch.Generator('cuda').manual_seed(seed) if seed != 0 else None
|
| 32 |
prompt = f"{prefix} {prompt}" if auto_prefix else prompt
|
| 33 |
|
|
|
|
| 1 |
+
from diffusers import StableDiffusionPipeline, StableDiffusionImg2ImgPipeline, DPMSolverMultistepScheduler
|
| 2 |
import gradio as gr
|
| 3 |
import torch
|
| 4 |
from PIL import Image
|
|
|
|
| 27 |
{error}""" if error else ""
|
| 28 |
|
| 29 |
def inference(prompt, guidance, steps, width=512, height=512, seed=0, img=None, strength=0.5, neg_prompt="", auto_prefix=False):
|
|
|
|
| 30 |
generator = torch.Generator('cuda').manual_seed(seed) if seed != 0 else None
|
| 31 |
prompt = f"{prefix} {prompt}" if auto_prefix else prompt
|
| 32 |
|