Spaces:
Runtime error
Runtime error
Load float32 model
Browse files
app.py
CHANGED
|
@@ -135,7 +135,7 @@ def inference(
|
|
| 135 |
raise gr.Error("cuda is not available")
|
| 136 |
|
| 137 |
device = torch.device("cuda")
|
| 138 |
-
model.to(device)
|
| 139 |
|
| 140 |
seed_everything(seed)
|
| 141 |
start_code = torch.randn([len(prompts), 4, 128, 128], device=device)
|
|
@@ -162,7 +162,7 @@ def inference(
|
|
| 162 |
register_attention_editor_diffusers(model, editor)
|
| 163 |
images = model(prompts, latents=start_code, guidance_scale=classifier_free_guidance_scale).images
|
| 164 |
unregister_attention_editor_diffusers(model)
|
| 165 |
-
model.to(torch.device("cpu"))
|
| 166 |
return images
|
| 167 |
|
| 168 |
|
|
@@ -256,7 +256,7 @@ def main():
|
|
| 256 |
nltk.download("averaged_perceptron_tagger")
|
| 257 |
|
| 258 |
scheduler = DDIMScheduler(beta_start=0.00085, beta_end=0.012, beta_schedule="scaled_linear", clip_sample=False, set_alpha_to_one=False)
|
| 259 |
-
model = StableDiffusionXLPipeline.from_pretrained(MODEL_PATH, scheduler=scheduler
|
| 260 |
model.unet.set_attn_processor(AttnProcessor2_0())
|
| 261 |
model.enable_sequential_cpu_offload()
|
| 262 |
|
|
|
|
| 135 |
raise gr.Error("cuda is not available")
|
| 136 |
|
| 137 |
device = torch.device("cuda")
|
| 138 |
+
model.to(device).half()
|
| 139 |
|
| 140 |
seed_everything(seed)
|
| 141 |
start_code = torch.randn([len(prompts), 4, 128, 128], device=device)
|
|
|
|
| 162 |
register_attention_editor_diffusers(model, editor)
|
| 163 |
images = model(prompts, latents=start_code, guidance_scale=classifier_free_guidance_scale).images
|
| 164 |
unregister_attention_editor_diffusers(model)
|
| 165 |
+
model.double().to(torch.device("cpu"))
|
| 166 |
return images
|
| 167 |
|
| 168 |
|
|
|
|
| 256 |
nltk.download("averaged_perceptron_tagger")
|
| 257 |
|
| 258 |
scheduler = DDIMScheduler(beta_start=0.00085, beta_end=0.012, beta_schedule="scaled_linear", clip_sample=False, set_alpha_to_one=False)
|
| 259 |
+
model = StableDiffusionXLPipeline.from_pretrained(MODEL_PATH, scheduler=scheduler)
|
| 260 |
model.unet.set_attn_processor(AttnProcessor2_0())
|
| 261 |
model.enable_sequential_cpu_offload()
|
| 262 |
|