Spaces:
Runtime error
Runtime error
Commit ·
c563c9c
1
Parent(s): b3aaea2
app.py
CHANGED
|
@@ -39,17 +39,17 @@ start_time = time.time()
|
|
| 39 |
model_id = "stabilityai/stable-diffusion-xl-base-1.0"
|
| 40 |
sdxl_lightening = "ByteDance/SDXL-Lightning"
|
| 41 |
ckpt = "sdxl_lightning_2step_unet.safetensors"
|
| 42 |
-
unet = UNet2DConditionModel.from_config(model_id, subfolder="unet", low_cpu_mem_usage=True).to(torch.float16)
|
| 43 |
unet.load_state_dict(load_file(hf_hub_download(sdxl_lightening, ckpt)))
|
| 44 |
|
| 45 |
-
image_encoder = CLIPVisionModelWithProjection.from_pretrained("h94/IP-Adapter", subfolder="models/image_encoder", torch_dtype=torch.float16, low_cpu_mem_usage=True)
|
| 46 |
-
pipe = AutoPipelineForText2Image.from_pretrained(model_id, unet=unet, torch_dtype=torch.float16, variant="fp16", image_encoder=image_encoder, low_cpu_mem_usage=True)
|
| 47 |
pipe.unet._load_ip_adapter_weights(torch.load(hf_hub_download('h94/IP-Adapter', 'sdxl_models/ip-adapter_sdxl_vit-h.bin')))
|
| 48 |
pipe.load_ip_adapter("h94/IP-Adapter", subfolder="sdxl_models", weight_name="ip-adapter_sdxl_vit-h.bin")
|
| 49 |
pipe.register_modules(image_encoder = image_encoder)
|
| 50 |
pipe.set_ip_adapter_scale(0.8)
|
| 51 |
|
| 52 |
-
pipe.vae = AutoencoderTiny.from_pretrained("madebyollin/taesdxl", torch_dtype=torch.float16, low_cpu_mem_usage=True)
|
| 53 |
pipe.scheduler = EulerDiscreteScheduler.from_config(pipe.scheduler.config, timestep_spacing="trailing")
|
| 54 |
|
| 55 |
pipe.to(device=DEVICE)
|
|
@@ -75,7 +75,7 @@ class BottleneckT5Autoencoder:
|
|
| 75 |
def __init__(self, model_path: str, device='cuda'):
|
| 76 |
self.device = device
|
| 77 |
self.tokenizer = AutoTokenizer.from_pretrained(model_path, model_max_length=512, torch_dtype=torch.bfloat16)
|
| 78 |
-
self.model = AutoModelForCausalLM.from_pretrained(model_path, trust_remote_code=True
|
| 79 |
self.model.eval()
|
| 80 |
|
| 81 |
|
|
|
|
| 39 |
model_id = "stabilityai/stable-diffusion-xl-base-1.0"
|
| 40 |
sdxl_lightening = "ByteDance/SDXL-Lightning"
|
| 41 |
ckpt = "sdxl_lightning_2step_unet.safetensors"
|
| 42 |
+
unet = UNet2DConditionModel.from_config(model_id, subfolder="unet", low_cpu_mem_usage=True, device=DEVICE).to(torch.float16)
|
| 43 |
unet.load_state_dict(load_file(hf_hub_download(sdxl_lightening, ckpt)))
|
| 44 |
|
| 45 |
+
image_encoder = CLIPVisionModelWithProjection.from_pretrained("h94/IP-Adapter", subfolder="models/image_encoder", torch_dtype=torch.float16, low_cpu_mem_usage=True, device=DEVICE)
|
| 46 |
+
pipe = AutoPipelineForText2Image.from_pretrained(model_id, unet=unet, torch_dtype=torch.float16, variant="fp16", image_encoder=image_encoder, low_cpu_mem_usage=True, device=DEVICE)
|
| 47 |
pipe.unet._load_ip_adapter_weights(torch.load(hf_hub_download('h94/IP-Adapter', 'sdxl_models/ip-adapter_sdxl_vit-h.bin')))
|
| 48 |
pipe.load_ip_adapter("h94/IP-Adapter", subfolder="sdxl_models", weight_name="ip-adapter_sdxl_vit-h.bin")
|
| 49 |
pipe.register_modules(image_encoder = image_encoder)
|
| 50 |
pipe.set_ip_adapter_scale(0.8)
|
| 51 |
|
| 52 |
+
pipe.vae = AutoencoderTiny.from_pretrained("madebyollin/taesdxl", torch_dtype=torch.float16, low_cpu_mem_usage=True, device=DEVICE)
|
| 53 |
pipe.scheduler = EulerDiscreteScheduler.from_config(pipe.scheduler.config, timestep_spacing="trailing")
|
| 54 |
|
| 55 |
pipe.to(device=DEVICE)
|
|
|
|
| 75 |
def __init__(self, model_path: str, device='cuda'):
|
| 76 |
self.device = device
|
| 77 |
self.tokenizer = AutoTokenizer.from_pretrained(model_path, model_max_length=512, torch_dtype=torch.bfloat16)
|
| 78 |
+
self.model = AutoModelForCausalLM.from_pretrained(model_path, trust_remote_code=True, device=DEVICE, low_cpu_mem_usage=True)
|
| 79 |
self.model.eval()
|
| 80 |
|
| 81 |
|