Update app.py
Browse files
app.py
CHANGED
|
@@ -61,9 +61,6 @@ if torch.cuda.is_available():
|
|
| 61 |
|
| 62 |
print("Loaded on Device!")
|
| 63 |
|
| 64 |
-
if USE_TORCH_COMPILE:
|
| 65 |
-
pipe.unet = torch.compile(pipe.unet, mode="reduce-overhead", fullgraph=True)
|
| 66 |
-
print("Model Compiled!")
|
| 67 |
|
| 68 |
|
| 69 |
def save_image(img):
|
|
@@ -91,13 +88,13 @@ def generate(
|
|
| 91 |
randomize_seed: bool = False,
|
| 92 |
progress=gr.Progress(track_tqdm=True),
|
| 93 |
):
|
|
|
|
| 94 |
|
| 95 |
-
pipe.to(device)
|
| 96 |
seed = int(randomize_seed_fn(seed, randomize_seed))
|
| 97 |
|
| 98 |
if not use_negative_prompt:
|
| 99 |
negative_prompt = "" # type: ignore
|
| 100 |
-
if model == "Fluently
|
| 101 |
images = pipe_3_5(
|
| 102 |
prompt=prompt,
|
| 103 |
negative_prompt=negative_prompt,
|
|
|
|
| 61 |
|
| 62 |
print("Loaded on Device!")
|
| 63 |
|
|
|
|
|
|
|
|
|
|
| 64 |
|
| 65 |
|
| 66 |
def save_image(img):
|
|
|
|
| 88 |
randomize_seed: bool = False,
|
| 89 |
progress=gr.Progress(track_tqdm=True),
|
| 90 |
):
|
| 91 |
+
|
| 92 |
|
|
|
|
| 93 |
seed = int(randomize_seed_fn(seed, randomize_seed))
|
| 94 |
|
| 95 |
if not use_negative_prompt:
|
| 96 |
negative_prompt = "" # type: ignore
|
| 97 |
+
if model == "Fluently v3.5":
|
| 98 |
images = pipe_3_5(
|
| 99 |
prompt=prompt,
|
| 100 |
negative_prompt=negative_prompt,
|