Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -73,11 +73,10 @@ pipe.load_lora_weights(
|
|
| 73 |
kwargs_lora_h = {}
|
| 74 |
kwargs_lora_h["load_into_transformer"] = True
|
| 75 |
pipe.load_lora_weights(
|
| 76 |
-
"
|
| 77 |
-
weight_name="
|
| 78 |
adapter_name="lora_h", **kwargs_lora_h
|
| 79 |
)
|
| 80 |
-
|
| 81 |
kwargs_lora = {}
|
| 82 |
kwargs_lora["load_into_transformer_2"] = True
|
| 83 |
pipe.load_lora_weights(
|
|
@@ -86,11 +85,20 @@ pipe.load_lora_weights(
|
|
| 86 |
adapter_name="lightx2v_2", **kwargs_lora
|
| 87 |
)
|
| 88 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 89 |
|
| 90 |
-
|
|
|
|
| 91 |
pipe.fuse_lora(adapter_names=["lightx2v"], lora_scale=3., components=["transformer"])
|
| 92 |
-
pipe.fuse_lora(adapter_names=["lora_h"], lora_scale=0.
|
| 93 |
pipe.fuse_lora(adapter_names=["lightx2v_2"], lora_scale=1., components=["transformer_2"])
|
|
|
|
| 94 |
pipe.unload_lora_weights()
|
| 95 |
|
| 96 |
quantize_(pipe.text_encoder, Int8WeightOnlyConfig())
|
|
@@ -370,4 +378,5 @@ def check_ffmpeg():
|
|
| 370 |
return False
|
| 371 |
|
| 372 |
if __name__ == "__main__":
|
| 373 |
-
demo.queue().launch(mcp_server=True)
|
|
|
|
|
|
| 73 |
kwargs_lora_h = {}
|
| 74 |
kwargs_lora_h["load_into_transformer"] = True
|
| 75 |
pipe.load_lora_weights(
|
| 76 |
+
"kingofpersia/wan-22-nsfw-loras",
|
| 77 |
+
weight_name="iGOON_Blink_Blowjob_I2V_HIGH.safetensors",
|
| 78 |
adapter_name="lora_h", **kwargs_lora_h
|
| 79 |
)
|
|
|
|
| 80 |
kwargs_lora = {}
|
| 81 |
kwargs_lora["load_into_transformer_2"] = True
|
| 82 |
pipe.load_lora_weights(
|
|
|
|
| 85 |
adapter_name="lightx2v_2", **kwargs_lora
|
| 86 |
)
|
| 87 |
|
| 88 |
+
kwargs_lora_l = {}
|
| 89 |
+
kwargs_lora_l["load_into_transformer_2"] = True
|
| 90 |
+
pipe.load_lora_weights(
|
| 91 |
+
"kingofpersia/wan-22-nsfw-loras",
|
| 92 |
+
weight_name="iGOON_Blink_Blowjob_I2V_LOW.safetensors",
|
| 93 |
+
adapter_name="lora_l", **kwargs_lora_l
|
| 94 |
+
)
|
| 95 |
|
| 96 |
+
|
| 97 |
+
pipe.set_adapters(["lightx2v", "lora_h", "lightx2v_2", "lora_l"], adapter_weights=[1., 1., 1., 1.])
|
| 98 |
pipe.fuse_lora(adapter_names=["lightx2v"], lora_scale=3., components=["transformer"])
|
| 99 |
+
pipe.fuse_lora(adapter_names=["lora_h"], lora_scale=0.6, components=["transformer"])
|
| 100 |
pipe.fuse_lora(adapter_names=["lightx2v_2"], lora_scale=1., components=["transformer_2"])
|
| 101 |
+
pipe.fuse_lora(adapter_names=["lora_l"], lora_scale=1., components=["transformer_2"])
|
| 102 |
pipe.unload_lora_weights()
|
| 103 |
|
| 104 |
quantize_(pipe.text_encoder, Int8WeightOnlyConfig())
|
|
|
|
| 378 |
return False
|
| 379 |
|
| 380 |
if __name__ == "__main__":
|
| 381 |
+
demo.queue().launch(mcp_server=True)
|
| 382 |
+
|