Spaces:
Running
on
Zero
Running
on
Zero
Update raw.py
Browse files
raw.py
CHANGED
|
@@ -7,11 +7,11 @@ from diffusers import BitsAndBytesConfig as DiffusersBitsAndBytesConfig
|
|
| 7 |
from transformers import T5EncoderModel
|
| 8 |
from transformers import BitsAndBytesConfig as TransformersBitsAndBytesConfig
|
| 9 |
from peft import PeftModel, PeftConfig
|
| 10 |
-
from attention_map_diffusers import (
|
| 11 |
-
|
| 12 |
-
|
| 13 |
-
|
| 14 |
-
)
|
| 15 |
import gradio as gr
|
| 16 |
huggingface_token = os.getenv("HUGGINFACE_TOKEN")
|
| 17 |
|
|
@@ -50,7 +50,6 @@ pipe.load_lora_weights(adapter_id3, weight_name="lora.safetensors", adapter_name
|
|
| 50 |
pipe.set_adapters(["turbo", "real", "enhance"], adapter_weights=[0.9, 0.66, 0.6])
|
| 51 |
pipe.fuse_lora(adapter_names=["turbo", "real", "enhance"], lora_scale=1.0)
|
| 52 |
pipe.unload_lora_weights()
|
| 53 |
-
pipe = init_pipeline(pipe)
|
| 54 |
# pipe.enable_xformers_memory_efficient_attention()
|
| 55 |
# save to the Hub
|
| 56 |
# pipe.push_to_hub("fused-t-r")
|
|
|
|
| 7 |
from transformers import T5EncoderModel
|
| 8 |
from transformers import BitsAndBytesConfig as TransformersBitsAndBytesConfig
|
| 9 |
from peft import PeftModel, PeftConfig
|
| 10 |
+
# from attention_map_diffusers import (
|
| 11 |
+
# attn_maps,
|
| 12 |
+
# init_pipeline,
|
| 13 |
+
# save_attention_maps
|
| 14 |
+
# )
|
| 15 |
import gradio as gr
|
| 16 |
huggingface_token = os.getenv("HUGGINFACE_TOKEN")
|
| 17 |
|
|
|
|
| 50 |
pipe.set_adapters(["turbo", "real", "enhance"], adapter_weights=[0.9, 0.66, 0.6])
|
| 51 |
pipe.fuse_lora(adapter_names=["turbo", "real", "enhance"], lora_scale=1.0)
|
| 52 |
pipe.unload_lora_weights()
|
|
|
|
| 53 |
# pipe.enable_xformers_memory_efficient_attention()
|
| 54 |
# save to the Hub
|
| 55 |
# pipe.push_to_hub("fused-t-r")
|