Spaces:
Running
on
Zero
Running
on
Zero
Upload dc.py
Browse files
dc.py
CHANGED
|
@@ -308,7 +308,6 @@ class GuiSD:
|
|
| 308 |
self.model = None
|
| 309 |
|
| 310 |
print("Loading model...")
|
| 311 |
-
self.model.device = torch.device("cpu")
|
| 312 |
self.model = Model_Diffusers(
|
| 313 |
base_model_id="cagliostrolab/animagine-xl-3.1",
|
| 314 |
task_name="txt2img",
|
|
@@ -317,6 +316,7 @@ class GuiSD:
|
|
| 317 |
retain_task_model_in_cache=False,
|
| 318 |
#device="cpu",
|
| 319 |
)
|
|
|
|
| 320 |
|
| 321 |
def infer_short(self, model, pipe_params, progress=gr.Progress(track_tqdm=True)):
|
| 322 |
progress(0, desc="Start inference...")
|
|
|
|
| 308 |
self.model = None
|
| 309 |
|
| 310 |
print("Loading model...")
|
|
|
|
| 311 |
self.model = Model_Diffusers(
|
| 312 |
base_model_id="cagliostrolab/animagine-xl-3.1",
|
| 313 |
task_name="txt2img",
|
|
|
|
| 316 |
retain_task_model_in_cache=False,
|
| 317 |
#device="cpu",
|
| 318 |
)
|
| 319 |
+
self.model.device = torch.device("cpu")
|
| 320 |
|
| 321 |
def infer_short(self, model, pipe_params, progress=gr.Progress(track_tqdm=True)):
|
| 322 |
progress(0, desc="Start inference...")
|