Spaces:
Running
on
Zero
Running
on
Zero
small change
Browse files
app.py
CHANGED
|
@@ -193,15 +193,9 @@ PIPELINE = FluxKontextSliderPipeline.from_pretrained(
|
|
| 193 |
_log("[worker] pipeline assembled.")
|
| 194 |
|
| 195 |
_log(f"[worker] loading LoRA from: {trained_models_path}")
|
| 196 |
-
|
| 197 |
_log("[worker] LoRA loaded.")
|
| 198 |
|
| 199 |
-
# if cuda_ok:
|
| 200 |
-
# free, total = torch.cuda.mem_get_info()
|
| 201 |
-
# _log(f"[worker] VRAM free/total: {free/1e9:.2f}/{total/1e9:.2f} GB")
|
| 202 |
-
# _log("[worker] PIPELINE ready.")
|
| 203 |
-
|
| 204 |
-
|
| 205 |
# moving the pipeline to GPU
|
| 206 |
PIPELINE.to('cuda')
|
| 207 |
|
|
|
|
| 193 |
_log("[worker] pipeline assembled.")
|
| 194 |
|
| 195 |
_log(f"[worker] loading LoRA from: {trained_models_path}")
|
| 196 |
+
PIPELINE.load_lora_weights(trained_models_path)
|
| 197 |
_log("[worker] LoRA loaded.")
|
| 198 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 199 |
# moving the pipeline to GPU
|
| 200 |
PIPELINE.to('cuda')
|
| 201 |
|