Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -7,11 +7,9 @@ device = "cuda" if torch.cuda.is_available() else "cpu"
|
|
| 7 |
|
| 8 |
pipe = DiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-xl-base-1.0", use_safetensors=True)
|
| 9 |
pipe = pipe.to(device)
|
| 10 |
-
pipe.unet = torch.compile(pipe.unet, mode="reduce-overhead", fullgraph=True)
|
| 11 |
|
| 12 |
refiner = DiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-xl-refiner-1.0", use_safetensors=True)
|
| 13 |
refiner = refiner.to(device)
|
| 14 |
-
refiner.unet = torch.compile(refiner.unet, mode="reduce-overhead", fullgraph=True)
|
| 15 |
|
| 16 |
def genie (prompt, negative_prompt, scale, steps, seed):
|
| 17 |
generator = torch.Generator(device=device).manual_seed(seed)
|
|
|
|
| 7 |
|
| 8 |
pipe = DiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-xl-base-1.0", use_safetensors=True)
|
| 9 |
pipe = pipe.to(device)
|
|
|
|
| 10 |
|
| 11 |
refiner = DiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-xl-refiner-1.0", use_safetensors=True)
|
| 12 |
refiner = refiner.to(device)
|
|
|
|
| 13 |
|
| 14 |
def genie (prompt, negative_prompt, scale, steps, seed):
|
| 15 |
generator = torch.Generator(device=device).manual_seed(seed)
|