Spaces:
Running
on
Zero
Running
on
Zero
Upload dc.py
Browse files
dc.py
CHANGED
|
@@ -823,7 +823,7 @@ def infer(prompt, negative_prompt, seed, randomize_seed, width, height, guidance
|
|
| 823 |
import numpy as np
|
| 824 |
MAX_SEED = np.iinfo(np.int32).max
|
| 825 |
|
| 826 |
-
load_lora_cpu =
|
| 827 |
verbose_info = False
|
| 828 |
gpu_duration = 59
|
| 829 |
|
|
@@ -842,8 +842,8 @@ def infer(prompt, negative_prompt, seed, randomize_seed, width, height, guidance
|
|
| 842 |
|
| 843 |
prompt, negative_prompt = insert_model_recom_prompt(prompt, negative_prompt, model_name)
|
| 844 |
progress(0.5, desc="Preparing...")
|
| 845 |
-
|
| 846 |
-
|
| 847 |
lora1 = get_valid_lora_path(lora1)
|
| 848 |
lora2 = get_valid_lora_path(lora2)
|
| 849 |
lora3 = get_valid_lora_path(lora3)
|
|
|
|
| 823 |
import numpy as np
|
| 824 |
MAX_SEED = np.iinfo(np.int32).max
|
| 825 |
|
| 826 |
+
load_lora_cpu = False
|
| 827 |
verbose_info = False
|
| 828 |
gpu_duration = 59
|
| 829 |
|
|
|
|
| 842 |
|
| 843 |
prompt, negative_prompt = insert_model_recom_prompt(prompt, negative_prompt, model_name)
|
| 844 |
progress(0.5, desc="Preparing...")
|
| 845 |
+
lora1, lora1_wt, lora2, lora2_wt, lora3, lora3_wt, lora4, lora4_wt, lora5, lora5_wt = \
|
| 846 |
+
set_prompt_loras(prompt, model_name, lora1, lora1_wt, lora2, lora2_wt, lora3, lora3_wt, lora4, lora4_wt, lora5, lora5_wt)
|
| 847 |
lora1 = get_valid_lora_path(lora1)
|
| 848 |
lora2 = get_valid_lora_path(lora2)
|
| 849 |
lora3 = get_valid_lora_path(lora3)
|