Spaces:
Running
on
Zero
Running
on
Zero
Update hf_demo.py
Browse files- hf_demo.py +2 -2
hf_demo.py
CHANGED
|
@@ -10,7 +10,7 @@ from PIL import Image
|
|
| 10 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 11 |
dtype=torch.bfloat16 if torch.cuda.is_available() else torch.float16
|
| 12 |
print(f"Using {device} device, dtype={dtype}")
|
| 13 |
-
pipe = DiffusionPipeline.from_pretrained("rhfeiyang/
|
| 14 |
torch_dtype=dtype).to(device)
|
| 15 |
|
| 16 |
from inference import get_lora_network, inference, get_validation_dataloader
|
|
@@ -118,7 +118,7 @@ block = gr.Blocks()
|
|
| 118 |
# Direct infer
|
| 119 |
with block:
|
| 120 |
with gr.Group():
|
| 121 |
-
gr.Markdown(" #
|
| 122 |
with gr.Row():
|
| 123 |
text = gr.Textbox(
|
| 124 |
label="Prompt:",
|
|
|
|
| 10 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 11 |
dtype=torch.bfloat16 if torch.cuda.is_available() else torch.float16
|
| 12 |
print(f"Using {device} device, dtype={dtype}")
|
| 13 |
+
pipe = DiffusionPipeline.from_pretrained("rhfeiyang/Blank-Canvas-Diffusion-v1",
|
| 14 |
torch_dtype=dtype).to(device)
|
| 15 |
|
| 16 |
from inference import get_lora_network, inference, get_validation_dataloader
|
|
|
|
| 118 |
# Direct infer
|
| 119 |
with block:
|
| 120 |
with gr.Group():
|
| 121 |
+
gr.Markdown(" # Blank Canvas Diffusion Demo")
|
| 122 |
with gr.Row():
|
| 123 |
text = gr.Textbox(
|
| 124 |
label="Prompt:",
|