| | import os |
| | import gradio as gr |
| | import numpy as np |
| | import random |
| | import spaces |
| | import torch |
| | from diffusers.pipelines.glm_image import GlmImagePipeline |
| | from PIL import Image |
| |
|
| | dtype = torch.bfloat16 |
| | device = "cuda" if torch.cuda.is_available() else "cpu" |
| |
|
| | MAX_SEED = np.iinfo(np.int32).max |
| | MAX_IMAGE_SIZE = 2048 |
| |
|
| | |
| | pipe = GlmImagePipeline.from_pretrained( |
| | "zai-org/GLM-Image", |
| | torch_dtype=torch.bfloat16, |
| | ).to("cuda") |
| |
|
| |
|
| | @spaces.GPU(duration=120) |
| | def infer(prompt, input_images=None, seed=42, randomize_seed=False, width=1024, height=1024, |
| | num_inference_steps=50, guidance_scale=1.5, progress=gr.Progress(track_tqdm=True)): |
| | """Main inference function""" |
| | print("Randomizing seed") |
| | if randomize_seed: |
| | seed = random.randint(0, MAX_SEED) |
| | |
| | |
| | width = (width // 32) * 32 |
| | height = (height // 32) * 32 |
| | |
| | generator = torch.Generator(device="cuda").manual_seed(seed) |
| |
|
| | print("preparing iages") |
| | |
| | image_list = None |
| | if input_images is not None and len(input_images) > 0: |
| | image_list = [] |
| | for item in input_images: |
| | img = item[0] if isinstance(item, tuple) else item |
| | if isinstance(img, str): |
| | img = Image.open(img).convert("RGB") |
| | elif isinstance(img, Image.Image): |
| | img = img.convert("RGB") |
| | image_list.append(img) |
| | print("handling kwargs") |
| | pipe_kwargs = { |
| | "prompt": prompt, |
| | "height": height, |
| | "width": width, |
| | "num_inference_steps": num_inference_steps, |
| | "guidance_scale": guidance_scale, |
| | "generator": generator, |
| | } |
| | print("adding images") |
| | |
| | if image_list is not None: |
| | pipe_kwargs["image"] = image_list |
| | print("running kwargs") |
| | image = pipe(**pipe_kwargs).images[0] |
| | |
| | return image, seed |
| |
|
| |
|
| | def update_dimensions_from_image(image_list): |
| | """Update width/height sliders based on uploaded image aspect ratio. |
| | Keeps dimensions proportional with both sides as multiples of 32.""" |
| | if image_list is None or len(image_list) == 0: |
| | return 1024, 1024 |
| | |
| | |
| | item = image_list[0] |
| | img = item[0] if isinstance(item, tuple) else item |
| | |
| | if isinstance(img, str): |
| | img = Image.open(img) |
| | |
| | img_width, img_height = img.size |
| | aspect_ratio = img_width / img_height |
| | |
| | if aspect_ratio >= 1: |
| | new_width = 1024 |
| | new_height = int(1024 / aspect_ratio) |
| | else: |
| | new_height = 1024 |
| | new_width = int(1024 * aspect_ratio) |
| | |
| | |
| | new_width = round(new_width / 32) * 32 |
| | new_height = round(new_height / 32) * 32 |
| | |
| | |
| | new_width = max(256, min(MAX_IMAGE_SIZE, new_width)) |
| | new_height = max(256, min(MAX_IMAGE_SIZE, new_height)) |
| | |
| | return new_width, new_height |
| |
|
| | css = """ |
| | #col-container { |
| | margin: 0 auto; |
| | max-width: 1200px; |
| | } |
| | .gallery-container img { |
| | object-fit: contain; |
| | } |
| | """ |
| |
|
| | with gr.Blocks() as demo: |
| | |
| | with gr.Column(elem_id="col-container"): |
| | gr.Markdown("""# GLM-Image |
| | GLM-Image is a hybrid auto-regressive + diffusion 9B parameters model by z.ai |
| | [[Model](https://huggingface.co/zai-org/GLM-Image)] |
| | """) |
| | |
| | with gr.Row(): |
| | with gr.Column(): |
| | prompt = gr.Text( |
| | label="Prompt", |
| | show_label=False, |
| | max_lines=4, |
| | placeholder="Enter your prompt (for text-to-image) or editing instructions (for image-to-image)", |
| | container=False, |
| | scale=3 |
| | ) |
| | |
| | run_button = gr.Button("π¨ Generate", variant="primary", scale=1) |
| | |
| | with gr.Accordion("π· Input Image(s) (optional - for image-to-image mode)", open=True): |
| | input_images = gr.Gallery( |
| | label="Input Image(s)", |
| | type="pil", |
| | columns=3, |
| | rows=1, |
| | elem_classes="gallery-container" |
| | ) |
| | gr.Markdown("*Upload one or more images for image-to-image generation. Leave empty for text-to-image mode.*") |
| | |
| | with gr.Accordion("βοΈ Advanced Settings", open=False): |
| | seed = gr.Slider( |
| | label="Seed", |
| | minimum=0, |
| | maximum=MAX_SEED, |
| | step=1, |
| | value=42, |
| | ) |
| | |
| | randomize_seed = gr.Checkbox(label="Randomize seed", value=True) |
| | |
| | with gr.Row(): |
| | width = gr.Slider( |
| | label="Width", |
| | minimum=256, |
| | maximum=MAX_IMAGE_SIZE, |
| | step=32, |
| | value=1024, |
| | info="Must be a multiple of 32" |
| | ) |
| | |
| | height = gr.Slider( |
| | label="Height", |
| | minimum=256, |
| | maximum=MAX_IMAGE_SIZE, |
| | step=32, |
| | value=1024, |
| | info="Must be a multiple of 32" |
| | ) |
| | |
| | with gr.Row(): |
| | num_inference_steps = gr.Slider( |
| | label="Number of inference steps", |
| | minimum=1, |
| | maximum=100, |
| | step=1, |
| | value=50, |
| | ) |
| | |
| | guidance_scale = gr.Slider( |
| | label="Guidance scale", |
| | minimum=0.0, |
| | maximum=10.0, |
| | step=0.1, |
| | value=1.5, |
| | ) |
| | |
| | with gr.Column(): |
| | result = gr.Image(label="Result", show_label=False) |
| |
|
| | |
| | input_images.upload( |
| | fn=update_dimensions_from_image, |
| | inputs=[input_images], |
| | outputs=[width, height] |
| | ) |
| |
|
| | gr.on( |
| | triggers=[run_button.click, prompt.submit], |
| | fn=infer, |
| | inputs=[prompt, input_images, seed, randomize_seed, width, height, num_inference_steps, guidance_scale], |
| | outputs=[result, seed] |
| | ) |
| |
|
| | demo.launch(theme=gr.themes.Citrus(), css=css) |