Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -83,6 +83,12 @@ with torch.inference_mode():
|
|
| 83 |
type="flux"
|
| 84 |
)
|
| 85 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 86 |
# Style Model
|
| 87 |
stylemodelloader = NODE_CLASS_MAPPINGS["StyleModelLoader"]()
|
| 88 |
stylemodelloader_441 = stylemodelloader.load_style_model(
|
|
@@ -127,6 +133,7 @@ def generate_image(prompt, input_image, lora_weight, guidance, downsampling_fact
|
|
| 127 |
weight=weight,
|
| 128 |
conditioning=flux_guidance[0],
|
| 129 |
style_model=stylemodelloader_441[0],
|
|
|
|
| 130 |
image=loaded_image[0]
|
| 131 |
)
|
| 132 |
|
|
|
|
| 83 |
type="flux"
|
| 84 |
)
|
| 85 |
|
| 86 |
+
# CLIP Vision
|
| 87 |
+
clipvisionloader = NODE_CLASS_MAPPINGS["CLIPVisionLoader"]()
|
| 88 |
+
clip_vision = clipvisionloader.load_clip(
|
| 89 |
+
clip_name="model.safetensors"
|
| 90 |
+
)
|
| 91 |
+
|
| 92 |
# Style Model
|
| 93 |
stylemodelloader = NODE_CLASS_MAPPINGS["StyleModelLoader"]()
|
| 94 |
stylemodelloader_441 = stylemodelloader.load_style_model(
|
|
|
|
| 133 |
weight=weight,
|
| 134 |
conditioning=flux_guidance[0],
|
| 135 |
style_model=stylemodelloader_441[0],
|
| 136 |
+
clip_vision=clip_vision[0],
|
| 137 |
image=loaded_image[0]
|
| 138 |
)
|
| 139 |
|