| | from lcm_pipeline import LatentConsistencyModelPipeline |
| | from lcm_scheduler import LCMScheduler |
| |
|
| | from diffusers import AutoencoderKL, UNet2DConditionModel |
| | from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker |
| | from transformers import CLIPTokenizer, CLIPTextModel, CLIPImageProcessor |
| |
|
| | import os |
| | import torch |
| | from tqdm import tqdm |
| | from safetensors.torch import load_file |
| |
|
| | |
| | prompt = "Self-portrait oil painting, a beautiful cyborg with golden hair" |
| |
|
| | |
| | save_path = "./lcm_images" |
| | os.makedirs(save_path, exist_ok=True) |
| |
|
| |
|
| | |
| | model_id = "digiplay/DreamShaper_7" |
| |
|
| |
|
| | |
| | vae = AutoencoderKL.from_pretrained(model_id, subfolder="vae") |
| | text_encoder = CLIPTextModel.from_pretrained(model_id, subfolder="text_encoder") |
| | tokenizer = CLIPTokenizer.from_pretrained(model_id, subfolder="tokenizer") |
| | unet = UNet2DConditionModel.from_pretrained(model_id, subfolder="unet", device_map=None, low_cpu_mem_usage=False, local_files_only=True) |
| | safety_checker = StableDiffusionSafetyChecker.from_pretrained(model_id, subfolder="safety_checker") |
| | feature_extractor = CLIPImageProcessor.from_pretrained(model_id, subfolder="feature_extractor") |
| |
|
| |
|
| | |
| | scheduler = LCMScheduler(beta_start=0.00085, beta_end=0.0120, beta_schedule="scaled_linear", prediction_type="epsilon") |
| |
|
| |
|
| | |
| | lcm_unet_ckpt = "./LCM_Dreamshaper_v7_4k.safetensors" |
| | ckpt = load_file(lcm_unet_ckpt) |
| | m, u = unet.load_state_dict(ckpt, strict=False) |
| | if len(m) > 0: |
| | print("missing keys:") |
| | print(m) |
| | if len(u) > 0: |
| | print("unexpected keys:") |
| | print(u) |
| |
|
| |
|
| | |
| | pipe = LatentConsistencyModelPipeline(vae=vae, text_encoder=text_encoder, tokenizer=tokenizer, unet=unet, scheduler=scheduler, safety_checker=safety_checker, feature_extractor=feature_extractor) |
| | pipe = pipe.to("cuda") |
| |
|
| |
|
| | |
| | images = pipe(prompt=prompt, num_images_per_prompt=4, num_inference_steps=4, guidance_scale=8.0, lcm_origin_steps=50).images |
| |
|
| | |
| | for i in tqdm(range(len(images))): |
| | output_path = os.path.join(save_path, "{}.png".format(i)) |
| | image = images[i] |
| | image.save(output_path) |
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|