Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -5,6 +5,7 @@ import random
|
|
| 5 |
# import spaces #[uncomment to use ZeroGPU]
|
| 6 |
from diffusers import DiffusionPipeline, StableDiffusionPipeline
|
| 7 |
import torch
|
|
|
|
| 8 |
|
| 9 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 10 |
|
|
@@ -50,8 +51,19 @@ def get_lora_pipeline(base_model_id, lora_model_id, lora_scale=0.8):
|
|
| 50 |
requires_safety_checker=False
|
| 51 |
)
|
| 52 |
|
| 53 |
-
|
| 54 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 55 |
|
| 56 |
pipe = pipe.to(device)
|
| 57 |
lora_pipelines[cache_key] = pipe
|
|
|
|
| 5 |
# import spaces #[uncomment to use ZeroGPU]
|
| 6 |
from diffusers import DiffusionPipeline, StableDiffusionPipeline
|
| 7 |
import torch
|
| 8 |
+
from peft import PeftModel
|
| 9 |
|
| 10 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 11 |
|
|
|
|
| 51 |
requires_safety_checker=False
|
| 52 |
)
|
| 53 |
|
| 54 |
+
pipe.unet = PeftModel.from_pretrained(
|
| 55 |
+
pipe.unet,
|
| 56 |
+
f"{lora_model_id}/unet",
|
| 57 |
+
adapter_name="default"
|
| 58 |
+
)
|
| 59 |
+
|
| 60 |
+
pipe.text_encoder = PeftModel.from_pretrained(
|
| 61 |
+
pipe.text_encoder,
|
| 62 |
+
f"{lora_model_id}/text_encoder",
|
| 63 |
+
adapter_name="default"
|
| 64 |
+
)
|
| 65 |
+
|
| 66 |
+
pipe = pipe.to(device)
|
| 67 |
|
| 68 |
pipe = pipe.to(device)
|
| 69 |
lora_pipelines[cache_key] = pipe
|