Spaces:
Running on A10G
Running on A10G
work with f16
Browse files
app.py
CHANGED
|
@@ -32,18 +32,24 @@ device = "cuda" if torch.cuda.is_available() else "cpu"
|
|
| 32 |
|
| 33 |
if torch.cuda.is_available():
|
| 34 |
power_device = "GPU"
|
|
|
|
| 35 |
else:
|
| 36 |
power_device = "CPU"
|
| 37 |
|
| 38 |
scheduler_class = MyEulerAncestralDiscreteScheduler
|
| 39 |
|
| 40 |
-
pipe_inversion = SDXLDDIMPipeline.from_pretrained("stabilityai/sdxl-turbo", use_safetensors=True) # .to('cpu')
|
| 41 |
-
pipe_inference = AutoPipelineForImage2Image.from_pretrained("stabilityai/sdxl-turbo",
|
| 42 |
use_safetensors=True) # .to('cpu')
|
| 43 |
pipe_inference.scheduler = scheduler_class.from_config(pipe_inference.scheduler.config)
|
| 44 |
pipe_inversion.scheduler = scheduler_class.from_config(pipe_inversion.scheduler.config)
|
| 45 |
pipe_inversion.scheduler_inference = scheduler_class.from_config(pipe_inference.scheduler.config)
|
| 46 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 47 |
# with gr.Blocks(css=css) as demo:
|
| 48 |
with gr.Blocks(css="style.css") as demo:
|
| 49 |
gr.Markdown(f""" # Real Time Editing with RNRI Inversion 🍎⚡️
|
|
|
|
| 32 |
|
| 33 |
if torch.cuda.is_available():
|
| 34 |
power_device = "GPU"
|
| 35 |
+
torch.cuda.max_memory_allocated(device=device)
|
| 36 |
else:
|
| 37 |
power_device = "CPU"
|
| 38 |
|
| 39 |
scheduler_class = MyEulerAncestralDiscreteScheduler
|
| 40 |
|
| 41 |
+
pipe_inversion = SDXLDDIMPipeline.from_pretrained("stabilityai/sdxl-turbo", torch_dtype=torch.float16, variant="fp16", use_safetensors=True) # .to('cpu')
|
| 42 |
+
pipe_inference = AutoPipelineForImage2Image.from_pretrained("stabilityai/sdxl-turbo", torch_dtype=torch.float16, variant="fp16",
|
| 43 |
use_safetensors=True) # .to('cpu')
|
| 44 |
pipe_inference.scheduler = scheduler_class.from_config(pipe_inference.scheduler.config)
|
| 45 |
pipe_inversion.scheduler = scheduler_class.from_config(pipe_inversion.scheduler.config)
|
| 46 |
pipe_inversion.scheduler_inference = scheduler_class.from_config(pipe_inference.scheduler.config)
|
| 47 |
|
| 48 |
+
pipe_inference.enable_xformers_memory_efficient_attention()
|
| 49 |
+
pipe_inversion.enable_xformers_memory_efficient_attention()
|
| 50 |
+
|
| 51 |
+
|
| 52 |
+
|
| 53 |
# with gr.Blocks(css=css) as demo:
|
| 54 |
with gr.Blocks(css="style.css") as demo:
|
| 55 |
gr.Markdown(f""" # Real Time Editing with RNRI Inversion 🍎⚡️
|