Update app.py
Browse files
app.py
CHANGED
|
@@ -29,7 +29,7 @@ torch.backends.cudnn.allow_tf32 = False
|
|
| 29 |
torch.backends.cudnn.deterministic = False
|
| 30 |
#torch.backends.cudnn.benchmark = False
|
| 31 |
torch.backends.cuda.preferred_blas_library="cublas"
|
| 32 |
-
torch.backends.cuda.preferred_linalg_library="cusolver"
|
| 33 |
|
| 34 |
torch.set_float32_matmul_precision("highest")
|
| 35 |
|
|
@@ -53,7 +53,7 @@ torch_dtype = torch.bfloat16
|
|
| 53 |
|
| 54 |
checkpoint = "microsoft/Phi-3.5-mini-instruct"
|
| 55 |
#vae = AutoencoderKL.from_pretrained("madebyollin/sdxl-vae-fp16-fix", torch_dtype=torch.float16)
|
| 56 |
-
vae = AutoencoderKL.from_pretrained("ford442/sdxl-vae-bf16", torch_dtype=torch.bfloat16
|
| 57 |
#vae = AutoencoderKL.from_pretrained("ford442/sdxl-vae-bf16")
|
| 58 |
|
| 59 |
pipe = StableDiffusion3Pipeline.from_pretrained("ford442/stable-diffusion-3.5-medium-bf16", torch_dtype=torch.bfloat16).to(torch.device("cuda:0"))
|
|
|
|
| 29 |
torch.backends.cudnn.deterministic = False
|
| 30 |
#torch.backends.cudnn.benchmark = False
|
| 31 |
torch.backends.cuda.preferred_blas_library="cublas"
|
| 32 |
+
#torch.backends.cuda.preferred_linalg_library="cusolver"
|
| 33 |
|
| 34 |
torch.set_float32_matmul_precision("highest")
|
| 35 |
|
|
|
|
| 53 |
|
| 54 |
checkpoint = "microsoft/Phi-3.5-mini-instruct"
|
| 55 |
#vae = AutoencoderKL.from_pretrained("madebyollin/sdxl-vae-fp16-fix", torch_dtype=torch.float16)
|
| 56 |
+
vae = AutoencoderKL.from_pretrained("ford442/sdxl-vae-bf16", torch_dtype=torch.bfloat16)
|
| 57 |
#vae = AutoencoderKL.from_pretrained("ford442/sdxl-vae-bf16")
|
| 58 |
|
| 59 |
pipe = StableDiffusion3Pipeline.from_pretrained("ford442/stable-diffusion-3.5-medium-bf16", torch_dtype=torch.bfloat16).to(torch.device("cuda:0"))
|