Spaces:
Runtime error
Runtime error
Commit
·
3c2cdca
1
Parent(s):
fcc8012
update
Browse files- app.py +16 -11
- requirements.txt +6 -6
app.py
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
from email.policy import default
|
| 2 |
from json import encoder
|
| 3 |
import gradio as gr
|
| 4 |
-
import spaces
|
| 5 |
import numpy as np
|
| 6 |
import torch
|
| 7 |
import requests
|
|
@@ -311,7 +311,8 @@ def set_attention_processor(unet,id_length,is_ipadapter = False):
|
|
| 311 |
hidden_size = unet.config.block_out_channels[block_id]
|
| 312 |
if cross_attention_dim is None:
|
| 313 |
if name.startswith("up_blocks") :
|
| 314 |
-
|
|
|
|
| 315 |
total_count +=1
|
| 316 |
else:
|
| 317 |
attn_procs[name] = AttnProcessor()
|
|
@@ -418,7 +419,8 @@ cur_step = 0
|
|
| 418 |
id_length = 4
|
| 419 |
total_length = 5
|
| 420 |
cur_model_type = ""
|
| 421 |
-
device="cuda"
|
|
|
|
| 422 |
global attn_procs,unet
|
| 423 |
attn_procs = {}
|
| 424 |
###
|
|
@@ -441,27 +443,27 @@ use_safetensors= False
|
|
| 441 |
###
|
| 442 |
pipe2 = PhotoMakerStableDiffusionXLPipeline.from_pretrained(
|
| 443 |
models_dict["Unstable"], torch_dtype=torch.float16, use_safetensors=use_safetensors)
|
| 444 |
-
pipe2 = pipe2.to(
|
| 445 |
pipe2.load_photomaker_adapter(
|
| 446 |
os.path.dirname(photomaker_path),
|
| 447 |
subfolder="",
|
| 448 |
weight_name=os.path.basename(photomaker_path),
|
| 449 |
trigger_word="img" # define the trigger word
|
| 450 |
)
|
| 451 |
-
pipe2 = pipe2.to(
|
| 452 |
pipe2.enable_freeu(s1=0.6, s2=0.4, b1=1.1, b2=1.2)
|
| 453 |
pipe2.fuse_lora()
|
| 454 |
|
| 455 |
pipe4 = PhotoMakerStableDiffusionXLPipeline.from_pretrained(
|
| 456 |
models_dict["RealVision"], torch_dtype=torch.float16, use_safetensors=True)
|
| 457 |
-
pipe4 = pipe4.to(
|
| 458 |
pipe4.load_photomaker_adapter(
|
| 459 |
os.path.dirname(photomaker_path),
|
| 460 |
subfolder="",
|
| 461 |
weight_name=os.path.basename(photomaker_path),
|
| 462 |
trigger_word="img" # define the trigger word
|
| 463 |
)
|
| 464 |
-
pipe4 = pipe4.to(
|
| 465 |
pipe4.enable_freeu(s1=0.6, s2=0.4, b1=1.1, b2=1.2)
|
| 466 |
pipe4.fuse_lora()
|
| 467 |
|
|
@@ -502,7 +504,7 @@ def change_visiale_by_model_type(_model_type):
|
|
| 502 |
|
| 503 |
|
| 504 |
######### Image Generation ##############
|
| 505 |
-
|
| 506 |
def process_generation(_sd_type,_model_type,_upload_images, _num_steps,style_name, _Ip_Adapter_Strength ,_style_strength_ratio, guidance_scale, seed_, sa32_, sa64_, id_length_, general_prompt, negative_prompt,prompt_array,G_height,G_width,_comic_type):
|
| 507 |
_model_type = "Photomaker" if _model_type == "Using Ref Images" else "original"
|
| 508 |
if _model_type == "Photomaker" and "img" not in general_prompt:
|
|
@@ -557,7 +559,8 @@ def process_generation(_sd_type,_model_type,_upload_images, _num_steps,style_nam
|
|
| 557 |
if start_merge_step > 30:
|
| 558 |
start_merge_step = 30
|
| 559 |
print(f"start_merge_step:{start_merge_step}")
|
| 560 |
-
|
|
|
|
| 561 |
sa32, sa64 = sa32_, sa64_
|
| 562 |
id_length = id_length_
|
| 563 |
clipped_prompts = prompts[:]
|
|
@@ -566,7 +569,8 @@ def process_generation(_sd_type,_model_type,_upload_images, _num_steps,style_nam
|
|
| 566 |
print(prompts)
|
| 567 |
id_prompts = prompts[:id_length]
|
| 568 |
real_prompts = prompts[id_length:]
|
| 569 |
-
|
|
|
|
| 570 |
write = True
|
| 571 |
cur_step = 0
|
| 572 |
|
|
@@ -804,4 +808,5 @@ with gr.Blocks(css=css) as demo:
|
|
| 804 |
|
| 805 |
# demo.load(None, None, None, _js=load_js)
|
| 806 |
|
| 807 |
-
demo.launch(server_name="0.0.0.0", share = True if use_va else False)
|
|
|
|
|
|
| 1 |
from email.policy import default
|
| 2 |
from json import encoder
|
| 3 |
import gradio as gr
|
| 4 |
+
#import spaces
|
| 5 |
import numpy as np
|
| 6 |
import torch
|
| 7 |
import requests
|
|
|
|
| 311 |
hidden_size = unet.config.block_out_channels[block_id]
|
| 312 |
if cross_attention_dim is None:
|
| 313 |
if name.startswith("up_blocks") :
|
| 314 |
+
device = devictorch.get(torch)
|
| 315 |
+
attn_procs[name] = SpatialAttnProcessor2_0(id_length = id_length, device=device)
|
| 316 |
total_count +=1
|
| 317 |
else:
|
| 318 |
attn_procs[name] = AttnProcessor()
|
|
|
|
| 419 |
id_length = 4
|
| 420 |
total_length = 5
|
| 421 |
cur_model_type = ""
|
| 422 |
+
#device="cuda"
|
| 423 |
+
device = devicetorch.get(torch)
|
| 424 |
global attn_procs,unet
|
| 425 |
attn_procs = {}
|
| 426 |
###
|
|
|
|
| 443 |
###
|
| 444 |
pipe2 = PhotoMakerStableDiffusionXLPipeline.from_pretrained(
|
| 445 |
models_dict["Unstable"], torch_dtype=torch.float16, use_safetensors=use_safetensors)
|
| 446 |
+
pipe2 = pipe2.to(device)
|
| 447 |
pipe2.load_photomaker_adapter(
|
| 448 |
os.path.dirname(photomaker_path),
|
| 449 |
subfolder="",
|
| 450 |
weight_name=os.path.basename(photomaker_path),
|
| 451 |
trigger_word="img" # define the trigger word
|
| 452 |
)
|
| 453 |
+
pipe2 = pipe2.to(device)
|
| 454 |
pipe2.enable_freeu(s1=0.6, s2=0.4, b1=1.1, b2=1.2)
|
| 455 |
pipe2.fuse_lora()
|
| 456 |
|
| 457 |
pipe4 = PhotoMakerStableDiffusionXLPipeline.from_pretrained(
|
| 458 |
models_dict["RealVision"], torch_dtype=torch.float16, use_safetensors=True)
|
| 459 |
+
pipe4 = pipe4.to(device)
|
| 460 |
pipe4.load_photomaker_adapter(
|
| 461 |
os.path.dirname(photomaker_path),
|
| 462 |
subfolder="",
|
| 463 |
weight_name=os.path.basename(photomaker_path),
|
| 464 |
trigger_word="img" # define the trigger word
|
| 465 |
)
|
| 466 |
+
pipe4 = pipe4.to(device)
|
| 467 |
pipe4.enable_freeu(s1=0.6, s2=0.4, b1=1.1, b2=1.2)
|
| 468 |
pipe4.fuse_lora()
|
| 469 |
|
|
|
|
| 504 |
|
| 505 |
|
| 506 |
######### Image Generation ##############
|
| 507 |
+
#@spaces.GPU(duration=120)
|
| 508 |
def process_generation(_sd_type,_model_type,_upload_images, _num_steps,style_name, _Ip_Adapter_Strength ,_style_strength_ratio, guidance_scale, seed_, sa32_, sa64_, id_length_, general_prompt, negative_prompt,prompt_array,G_height,G_width,_comic_type):
|
| 509 |
_model_type = "Photomaker" if _model_type == "Using Ref Images" else "original"
|
| 510 |
if _model_type == "Photomaker" and "img" not in general_prompt:
|
|
|
|
| 559 |
if start_merge_step > 30:
|
| 560 |
start_merge_step = 30
|
| 561 |
print(f"start_merge_step:{start_merge_step}")
|
| 562 |
+
device = devictorch.get(torch)
|
| 563 |
+
generator = torch.Generator(device=device).manual_seed(seed_)
|
| 564 |
sa32, sa64 = sa32_, sa64_
|
| 565 |
id_length = id_length_
|
| 566 |
clipped_prompts = prompts[:]
|
|
|
|
| 569 |
print(prompts)
|
| 570 |
id_prompts = prompts[:id_length]
|
| 571 |
real_prompts = prompts[id_length:]
|
| 572 |
+
devicetorch.empty_cache(torch)
|
| 573 |
+
#torch.cuda.empty_cache()
|
| 574 |
write = True
|
| 575 |
cur_step = 0
|
| 576 |
|
|
|
|
| 808 |
|
| 809 |
# demo.load(None, None, None, _js=load_js)
|
| 810 |
|
| 811 |
+
#demo.launch(server_name="0.0.0.0", share = True if use_va else False)
|
| 812 |
+
demo.launch()
|
requirements.txt
CHANGED
|
@@ -1,14 +1,14 @@
|
|
| 1 |
-
xformers==0.0.20
|
| 2 |
-
torch==2.0.1
|
| 3 |
-
torchvision==0.15.2
|
| 4 |
diffusers==0.25.0
|
| 5 |
transformers==4.36.2
|
| 6 |
huggingface-hub==0.20.2
|
| 7 |
-
spaces==0.19.4
|
| 8 |
numpy
|
| 9 |
accelerate
|
| 10 |
safetensors
|
| 11 |
omegaconf
|
| 12 |
peft
|
| 13 |
-
spaces==0.19.4
|
| 14 |
-
Pillow==9.5.0
|
|
|
|
| 1 |
+
#xformers==0.0.20
|
| 2 |
+
#torch==2.0.1
|
| 3 |
+
#torchvision==0.15.2
|
| 4 |
diffusers==0.25.0
|
| 5 |
transformers==4.36.2
|
| 6 |
huggingface-hub==0.20.2
|
| 7 |
+
#spaces==0.19.4
|
| 8 |
numpy
|
| 9 |
accelerate
|
| 10 |
safetensors
|
| 11 |
omegaconf
|
| 12 |
peft
|
| 13 |
+
#spaces==0.19.4
|
| 14 |
+
Pillow==9.5.0
|