| | import spaces |
| |
|
| | import sys |
| | sys.path.append('./') |
| |
|
| | from diffusers import ( |
| | StableDiffusionPipeline, |
| | UNet2DConditionModel, |
| | DPMSolverMultistepScheduler, |
| | LCMScheduler |
| | ) |
| |
|
| | from arc2face import CLIPTextModelWrapper, project_face_embs |
| |
|
| | import torch |
| | from insightface.app import FaceAnalysis |
| | from PIL import Image |
| | import numpy as np |
| | import random |
| | |
| |
|
| | import gradio as gr |
| |
|
| | |
| | MAX_SEED = np.iinfo(np.int32).max |
| | if torch.cuda.is_available(): |
| | device = "cuda" |
| | dtype = torch.float16 |
| | else: |
| | device = "cpu" |
| | dtype = torch.float32 |
| |
|
| |
|
| | |
| | from huggingface_hub import hf_hub_download |
| | |
| | |
| |
|
| | hf_hub_download(repo_id="FoivosPar/Arc2Face", filename="arc2face/config.json", local_dir="./models") |
| | hf_hub_download(repo_id="FoivosPar/Arc2Face", filename="arc2face/diffusion_pytorch_model.safetensors", local_dir="./models") |
| | hf_hub_download(repo_id="FoivosPar/Arc2Face", filename="encoder/config.json", local_dir="./models") |
| | hf_hub_download(repo_id="FoivosPar/Arc2Face", filename="encoder/pytorch_model.bin", local_dir="./models") |
| | hf_hub_download(repo_id="FoivosPar/Arc2Face", filename="arcface.onnx", local_dir="./models/antelopev2") |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | |
| | app = FaceAnalysis(name='antelopev2', root='./', providers=['CPUExecutionProvider']) |
| | app.prepare(ctx_id=0, det_size=(640, 640)) |
| |
|
| | |
| | |
| | base_model = 'stable-diffusion-v1-5/stable-diffusion-v1-5' |
| | encoder = CLIPTextModelWrapper.from_pretrained( |
| | 'models', subfolder="encoder", torch_dtype=dtype |
| | ) |
| | unet = UNet2DConditionModel.from_pretrained( |
| | 'models', subfolder="arc2face", torch_dtype=dtype |
| | ) |
| | pipeline = StableDiffusionPipeline.from_pretrained( |
| | base_model, |
| | text_encoder=encoder, |
| | unet=unet, |
| | torch_dtype=dtype, |
| | safety_checker=None, |
| | ) |
| | pipeline.scheduler = DPMSolverMultistepScheduler.from_config(pipeline.scheduler.config) |
| | pipeline = pipeline.to(device) |
| |
|
| | |
| | pipeline.load_lora_weights("latent-consistency/lcm-lora-sdv1-5") |
| | pipeline.disable_lora() |
| |
|
| | def toggle_lcm_ui(value): |
| | if value: |
| | return ( |
| | gr.update(minimum=1, maximum=20, step=1, value=3), |
| | gr.update(minimum=0.1, maximum=10.0, step=0.1, value=1.0), |
| | ) |
| | else: |
| | return ( |
| | gr.update(minimum=1, maximum=100, step=1, value=25), |
| | gr.update(minimum=0.1, maximum=10.0, step=0.1, value=3.0), |
| | ) |
| |
|
| | def randomize_seed_fn(seed: int, randomize_seed: bool) -> int: |
| | if randomize_seed: |
| | seed = random.randint(0, MAX_SEED) |
| | return seed |
| |
|
| | def get_example(): |
| | case = [ |
| | [ |
| | './assets/examples/freeman.jpg', |
| | ], |
| | [ |
| | './assets/examples/lily.png', |
| | ], |
| | [ |
| | './assets/examples/joacquin.png', |
| | ], |
| | [ |
| | './assets/examples/jackie.png', |
| | ], |
| | [ |
| | './assets/examples/freddie.png', |
| | ], |
| | [ |
| | './assets/examples/hepburn.png', |
| | ], |
| | ] |
| | return case |
| |
|
| | def run_example(img_file): |
| | return generate_image(img_file, 25, 3, 23, 2, False) |
| |
|
| | @spaces.GPU |
| | def generate_image(image_path, num_steps, guidance_scale, seed, num_images, use_lcm, progress=gr.Progress(track_tqdm=True)): |
| |
|
| | if use_lcm: |
| | pipeline.scheduler = LCMScheduler.from_config(pipeline.scheduler.config) |
| | pipeline.enable_lora() |
| | pipeline.to(device) |
| | else: |
| | pipeline.disable_lora() |
| | pipeline.scheduler = DPMSolverMultistepScheduler.from_config(pipeline.scheduler.config) |
| | pipeline.to(device) |
| |
|
| | if image_path is None: |
| | raise gr.Error(f"Cannot find any input face image! Please upload a face image.") |
| | |
| | img = np.array(Image.open(image_path))[:,:,::-1] |
| |
|
| | |
| | faces = app.get(img) |
| | |
| | if len(faces) == 0: |
| | raise gr.Error(f"Face detection failed! Please try with another image.") |
| | |
| | faces = sorted(faces, key=lambda x:(x['bbox'][2]-x['bbox'][0])*(x['bbox'][3]-x['bbox'][1]))[-1] |
| | id_emb = torch.tensor(faces['embedding'], dtype=dtype)[None].to(device) |
| | id_emb = id_emb/torch.norm(id_emb, dim=1, keepdim=True) |
| | id_emb = project_face_embs(pipeline, id_emb) |
| | |
| | generator = torch.Generator(device=device).manual_seed(seed) |
| | |
| | print("Start inference...") |
| | images = pipeline( |
| | prompt_embeds=id_emb, |
| | num_inference_steps=num_steps, |
| | guidance_scale=guidance_scale, |
| | num_images_per_prompt=num_images, |
| | generator=generator |
| | ).images |
| |
|
| | return images |
| |
|
| | |
| | title = r""" |
| | <h1>Arc2Face: A Foundation Model for ID-Consistent Human Faces</h1> |
| | """ |
| |
|
| | description = r""" |
| | <b>Official 🤗 Gradio demo</b> for <a href='https://arc2face.github.io/' target='_blank'><b>Arc2Face: A Foundation Model for ID-Consistent Human Faces</b></a>.<br> |
| | |
| | Steps:<br> |
| | 1. Upload an image with a face. If multiple faces are detected, we use the largest one. For images with already tightly cropped faces, detection may fail, try images with a larger margin. |
| | 2. Click <b>Submit</b> to generate new images of the subject. |
| | """ |
| |
|
| | Footer = r""" |
| | --- |
| | 📝 **Citation** |
| | <br> |
| | If you find Arc2Face helpful for your research, please consider citing our paper: |
| | ```bibtex |
| | @inproceedings{paraperas2024arc2face, |
| | title={Arc2Face: A Foundation Model for ID-Consistent Human Faces}, |
| | author={Paraperas Papantoniou, Foivos and Lattas, Alexandros and Moschoglou, Stylianos and Deng, Jiankang and Kainz, Bernhard and Zafeiriou, Stefanos}, |
| | booktitle={Proceedings of the European Conference on Computer Vision (ECCV)}, |
| | year={2024} |
| | } |
| | ``` |
| | """ |
| |
|
| | css = ''' |
| | .gradio-container {width: 85% !important} |
| | ''' |
| | with gr.Blocks(css=css) as demo: |
| |
|
| | |
| | gr.Markdown(title) |
| | gr.Markdown(description) |
| |
|
| | with gr.Row(): |
| | with gr.Column(): |
| | |
| | |
| | img_file = gr.Image(label="Upload a photo with a face", type="filepath") |
| | |
| | submit = gr.Button("Submit", variant="primary") |
| |
|
| | use_lcm = gr.Checkbox( |
| | label="Use LCM-LoRA to accelerate sampling", value=False, |
| | info="Reduces sampling steps significantly, but may decrease quality.", |
| | ) |
| | |
| | with gr.Accordion(open=False, label="Advanced Options"): |
| | num_steps = gr.Slider( |
| | label="Number of sample steps", |
| | minimum=1, |
| | maximum=100, |
| | step=1, |
| | value=25, |
| | ) |
| | guidance_scale = gr.Slider( |
| | label="Guidance scale", |
| | minimum=0.1, |
| | maximum=10.0, |
| | step=0.1, |
| | value=3.0, |
| | ) |
| | num_images = gr.Slider( |
| | label="Number of output images", |
| | minimum=1, |
| | maximum=4, |
| | step=1, |
| | value=2, |
| | ) |
| | seed = gr.Slider( |
| | label="Seed", |
| | minimum=0, |
| | maximum=MAX_SEED, |
| | step=1, |
| | value=0, |
| | ) |
| | randomize_seed = gr.Checkbox(label="Randomize seed", value=True) |
| |
|
| | with gr.Column(): |
| | gallery = gr.Gallery(label="Generated Images") |
| |
|
| | submit.click( |
| | fn=randomize_seed_fn, |
| | inputs=[seed, randomize_seed], |
| | outputs=seed, |
| | queue=False, |
| | api_name=False, |
| | ).then( |
| | fn=generate_image, |
| | inputs=[img_file, num_steps, guidance_scale, seed, num_images, use_lcm], |
| | outputs=[gallery] |
| | ) |
| |
|
| | use_lcm.input( |
| | fn=toggle_lcm_ui, |
| | inputs=[use_lcm], |
| | outputs=[num_steps, guidance_scale], |
| | queue=False, |
| | ) |
| | |
| | gr.Examples( |
| | examples=get_example(), |
| | inputs=[img_file], |
| | run_on_click=True, |
| | fn=run_example, |
| | outputs=[gallery], |
| | ) |
| | |
| | gr.Markdown(Footer) |
| |
|
| | demo.launch() |