Spaces:
Paused
Paused
| import gradio as gr | |
| import numpy as np | |
| from PIL import Image, ImageDraw | |
| from gradio_client import Client, handle_file | |
| import random | |
| import tempfile | |
| import os | |
| import logging | |
| import torch | |
| from diffusers import AutoencoderKL, TCDScheduler | |
| from diffusers.models.model_loading_utils import load_state_dict | |
| from huggingface_hub import hf_hub_download | |
| # Spaces GPU | |
| try: | |
| import spaces | |
| except: | |
| # GPU λ°μ½λ μ΄ν°κ° μμ λλ₯Ό μν λλ―Έ λ°μ½λ μ΄ν° | |
| class spaces: | |
| def GPU(duration=None): | |
| def decorator(func): | |
| return func | |
| return decorator | |
| # νκ²½ λ³μ μ€μ | |
| os.environ["TRANSFORMERS_ALLOW_UNSAFE_DESERIALIZATION"] = "1" | |
| # MMAudio κ΄λ ¨ μν¬νΈ | |
| try: | |
| import mmaudio | |
| from mmaudio.eval_utils import (ModelConfig, all_model_cfg, generate, load_video, make_video, | |
| setup_eval_logging) | |
| from mmaudio.model.flow_matching import FlowMatching | |
| from mmaudio.model.networks import MMAudio, get_my_mmaudio | |
| from mmaudio.model.sequence_config import SequenceConfig | |
| from mmaudio.model.utils.features_utils import FeaturesUtils | |
| MMAUDIO_AVAILABLE = True | |
| except ImportError: | |
| MMAUDIO_AVAILABLE = False | |
| logging.warning("MMAudio not available. Sound generation will be disabled.") | |
| # ControlNet λͺ¨λΈ λ‘λ | |
| try: | |
| from controlnet_union import ControlNetModel_Union | |
| from pipeline_fill_sd_xl import StableDiffusionXLFillPipeline | |
| # ControlNet μ€μ λ° λ‘λ | |
| config_file = hf_hub_download( | |
| "xinsir/controlnet-union-sdxl-1.0", | |
| filename="config_promax.json", | |
| ) | |
| config = ControlNetModel_Union.load_config(config_file) | |
| controlnet_model = ControlNetModel_Union.from_config(config) | |
| model_file = hf_hub_download( | |
| "xinsir/controlnet-union-sdxl-1.0", | |
| filename="diffusion_pytorch_model_promax.safetensors", | |
| ) | |
| state_dict = load_state_dict(model_file) | |
| loaded_keys = list(state_dict.keys()) | |
| result = ControlNetModel_Union._load_pretrained_model( | |
| controlnet_model, state_dict, model_file, "xinsir/controlnet-union-sdxl-1.0", loaded_keys | |
| ) | |
| model = result[0] | |
| model = model.to(device="cuda", dtype=torch.float16) | |
| # VAE λ‘λ | |
| vae = AutoencoderKL.from_pretrained( | |
| "madebyollin/sdxl-vae-fp16-fix", torch_dtype=torch.float16 | |
| ).to("cuda") | |
| # νμ΄νλΌμΈ λ‘λ | |
| pipe = StableDiffusionXLFillPipeline.from_pretrained( | |
| "SG161222/RealVisXL_V5.0_Lightning", | |
| torch_dtype=torch.float16, | |
| vae=vae, | |
| controlnet=model, | |
| variant="fp16", | |
| ).to("cuda") | |
| pipe.scheduler = TCDScheduler.from_config(pipe.scheduler.config) | |
| OUTPAINT_MODEL_LOADED = True | |
| except Exception as e: | |
| logging.error(f"Failed to load outpainting models: {str(e)}") | |
| OUTPAINT_MODEL_LOADED = False | |
| # MMAudio λͺ¨λΈ μ€μ λ° λ‘λ | |
| if MMAUDIO_AVAILABLE: | |
| try: | |
| # CUDA μ€μ | |
| if torch.cuda.is_available(): | |
| device = torch.device("cuda") | |
| torch.backends.cuda.matmul.allow_tf32 = True | |
| torch.backends.cudnn.allow_tf32 = True | |
| torch.backends.cudnn.benchmark = True | |
| else: | |
| device = torch.device("cpu") | |
| dtype = torch.bfloat16 | |
| # λͺ¨λΈ μ€μ | |
| model_cfg: ModelConfig = all_model_cfg['large_44k_v2'] | |
| model_cfg.download_if_needed() | |
| setup_eval_logging() | |
| # λͺ¨λΈ λ‘λ | |
| def get_mmaudio_model(): | |
| with torch.cuda.device(device): | |
| seq_cfg = model_cfg.seq_cfg | |
| net: MMAudio = get_my_mmaudio(model_cfg.model_name).to(device, dtype).eval() | |
| net.load_weights(torch.load(model_cfg.model_path, map_location=device, weights_only=True)) | |
| logging.info(f'Loaded MMAudio weights from {model_cfg.model_path}') | |
| feature_utils = FeaturesUtils( | |
| tod_vae_ckpt=model_cfg.vae_path, | |
| synchformer_ckpt=model_cfg.synchformer_ckpt, | |
| enable_conditions=True, | |
| mode=model_cfg.mode, | |
| bigvgan_vocoder_ckpt=model_cfg.bigvgan_16k_path, | |
| need_vae_encoder=False | |
| ).to(device, dtype).eval() | |
| return net, feature_utils, seq_cfg | |
| mmaudio_net, mmaudio_feature_utils, mmaudio_seq_cfg = get_mmaudio_model() | |
| MMAUDIO_LOADED = True | |
| except Exception as e: | |
| logging.error(f"Failed to load MMAudio models: {str(e)}") | |
| MMAUDIO_LOADED = False | |
| else: | |
| MMAUDIO_LOADED = False | |
| # API URLs | |
| TEXT2IMG_API_URL = "http://211.233.58.201:7896" | |
| VIDEO_API_URL = "http://211.233.58.201:7875" | |
| # λ‘κΉ μ€μ | |
| logging.basicConfig(level=logging.INFO) | |
| # Image size presets | |
| IMAGE_PRESETS = { | |
| "컀μ€ν ": {"width": 1024, "height": 1024}, | |
| "1:1 μ μ¬κ°ν": {"width": 1024, "height": 1024}, | |
| "4:3 νμ€": {"width": 1024, "height": 768}, | |
| "16:9 μμ΄λμ€ν¬λ¦°": {"width": 1024, "height": 576}, | |
| "9:16 μΈλ‘ν": {"width": 576, "height": 1024}, | |
| "6:19 νΉμ μΈλ‘ν": {"width": 324, "height": 1024}, | |
| "Instagram μ μ¬κ°ν": {"width": 1080, "height": 1080}, | |
| "Instagram μ€ν 리": {"width": 1080, "height": 1920}, | |
| "Instagram κ°λ‘ν": {"width": 1080, "height": 566}, | |
| "Facebook 컀λ²": {"width": 820, "height": 312}, | |
| "Twitter ν€λ": {"width": 1500, "height": 500}, | |
| "YouTube μΈλ€μΌ": {"width": 1280, "height": 720}, | |
| "LinkedIn λ°°λ": {"width": 1584, "height": 396}, | |
| } | |
| def update_dimensions(preset): | |
| if preset in IMAGE_PRESETS: | |
| return IMAGE_PRESETS[preset]["width"], IMAGE_PRESETS[preset]["height"] | |
| return 1024, 1024 | |
| def generate_text_to_image(prompt, width, height, guidance, inference_steps, seed): | |
| if not prompt: | |
| return None, "ν둬ννΈλ₯Ό μ λ ₯ν΄μ£ΌμΈμ" | |
| try: | |
| client = Client(TEXT2IMG_API_URL) | |
| if seed == -1: | |
| seed = random.randint(0, 9999999) | |
| result = client.predict( | |
| prompt=prompt, | |
| width=int(width), | |
| height=int(height), | |
| guidance=float(guidance), | |
| inference_steps=int(inference_steps), | |
| seed=int(seed), | |
| do_img2img=False, | |
| init_image=None, | |
| image2image_strength=0.8, | |
| resize_img=True, | |
| api_name="/generate_image" | |
| ) | |
| return result[0], f"μ¬μ©λ μλ: {result[1]}" | |
| except Exception as e: | |
| logging.error(f"Image generation error: {str(e)}") | |
| return None, f"μ€λ₯: {str(e)}" | |
| def video_to_audio(video_path, prompt, negative_prompt="music", seed=0, num_steps=25, cfg_strength=4.5, duration=8.0): | |
| """λΉλμ€μ μ¬μ΄λλ₯Ό μΆκ°νλ ν¨μ""" | |
| if not MMAUDIO_LOADED: | |
| logging.error("MMAudio model not loaded") | |
| return video_path | |
| try: | |
| rng = torch.Generator(device=device) | |
| rng.manual_seed(seed) | |
| fm = FlowMatching(min_sigma=0, inference_mode='euler', num_steps=num_steps) | |
| # λΉλμ€ λ‘λ | |
| clip_frames, sync_frames, actual_duration = load_video(video_path, duration) | |
| clip_frames = clip_frames.unsqueeze(0) | |
| sync_frames = sync_frames.unsqueeze(0) | |
| mmaudio_seq_cfg.duration = actual_duration | |
| mmaudio_net.update_seq_lengths(mmaudio_seq_cfg.latent_seq_len, mmaudio_seq_cfg.clip_seq_len, mmaudio_seq_cfg.sync_seq_len) | |
| # μ€λμ€ μμ± | |
| audios = generate(clip_frames, | |
| sync_frames, [prompt], | |
| negative_text=[negative_prompt], | |
| feature_utils=mmaudio_feature_utils, | |
| net=mmaudio_net, | |
| fm=fm, | |
| rng=rng, | |
| cfg_strength=cfg_strength) | |
| audio = audios.float().cpu()[0] | |
| # λΉλμ€μ μ€λμ€ κ²°ν© | |
| video_save_path = tempfile.NamedTemporaryFile(delete=False, suffix='.mp4').name | |
| make_video(video_path, | |
| video_save_path, | |
| audio, | |
| sampling_rate=mmaudio_seq_cfg.sampling_rate, | |
| duration_sec=mmaudio_seq_cfg.duration) | |
| return video_save_path | |
| except Exception as e: | |
| logging.error(f"Video to audio error: {str(e)}") | |
| return video_path | |
| def generate_video_from_image(image, prompt="", length=4.0, sound_generation="μ¬μ΄λ μμ", sound_prompt="", sound_negative_prompt="music"): | |
| if image is None: | |
| return None | |
| try: | |
| # μ΄λ―Έμ§ μ μ₯ | |
| with tempfile.NamedTemporaryFile(suffix=".png", delete=False) as fp: | |
| temp_path = fp.name | |
| Image.fromarray(image).save(temp_path) | |
| # λΉλμ€ μμ± API νΈμΆ | |
| client = Client(VIDEO_API_URL) | |
| result = client.predict( | |
| input_image=handle_file(temp_path), | |
| prompt=prompt if prompt else "Generate natural motion", | |
| n_prompt="", | |
| seed=random.randint(0, 9999999), | |
| use_teacache=True, | |
| video_length=float(length), | |
| api_name="/process" | |
| ) | |
| os.unlink(temp_path) | |
| if result and len(result) > 0: | |
| video_dict = result[0] | |
| video_path = video_dict.get("video") if isinstance(video_dict, dict) else None | |
| # μ¬μ΄λ μμ± μ΅μ μ΄ μ νλ κ²½μ° | |
| if video_path and sound_generation == "μ¬μ΄λ μμ±" and MMAUDIO_LOADED: | |
| # μ¬μ΄λ ν둬ννΈκ° λΉμ΄μμΌλ©΄ κΈ°λ³Έκ° μ¬μ© | |
| if not sound_prompt: | |
| sound_prompt = prompt if prompt else "ambient sound" | |
| # λΉλμ€μ μ¬μ΄λ μΆκ° | |
| video_with_sound = video_to_audio( | |
| video_path, | |
| sound_prompt, | |
| sound_negative_prompt, | |
| duration=length | |
| ) | |
| return video_with_sound | |
| return video_path | |
| except Exception as e: | |
| logging.error(f"Video generation error: {str(e)}") | |
| return None | |
| def prepare_image_and_mask(image, width, height, overlap_percentage, alignment): | |
| """μ΄λ―Έμ§μ λ§μ€ν¬λ₯Ό μ€λΉνλ ν¨μ""" | |
| if image is None: | |
| return None, None | |
| # PIL μ΄λ―Έμ§λ‘ λ³ν | |
| if isinstance(image, np.ndarray): | |
| image = Image.fromarray(image).convert('RGB') | |
| target_size = (width, height) | |
| # μ΄λ―Έμ§λ₯Ό νκ² ν¬κΈ°μ λ§κ² μ‘°μ | |
| scale_factor = min(target_size[0] / image.width, target_size[1] / image.height) | |
| new_width = int(image.width * scale_factor) | |
| new_height = int(image.height * scale_factor) | |
| # μ΄λ―Έμ§ 리μ¬μ΄μ¦ | |
| source = image.resize((new_width, new_height), Image.LANCZOS) | |
| # μ€λ²λ© κ³μ° | |
| overlap_x = int(new_width * (overlap_percentage / 100)) | |
| overlap_y = int(new_height * (overlap_percentage / 100)) | |
| overlap_x = max(overlap_x, 1) | |
| overlap_y = max(overlap_y, 1) | |
| # μ λ ¬μ λ°λ₯Έ λ§μ§ κ³μ° | |
| if alignment == "κ°μ΄λ°": | |
| margin_x = (target_size[0] - new_width) // 2 | |
| margin_y = (target_size[1] - new_height) // 2 | |
| elif alignment == "μΌμͺ½": | |
| margin_x = 0 | |
| margin_y = (target_size[1] - new_height) // 2 | |
| elif alignment == "μ€λ₯Έμͺ½": | |
| margin_x = target_size[0] - new_width | |
| margin_y = (target_size[1] - new_height) // 2 | |
| elif alignment == "μ": | |
| margin_x = (target_size[0] - new_width) // 2 | |
| margin_y = 0 | |
| elif alignment == "μλ": | |
| margin_x = (target_size[0] - new_width) // 2 | |
| margin_y = target_size[1] - new_height | |
| # λ°°κ²½ μ΄λ―Έμ§ μμ± | |
| background = Image.new('RGB', target_size, (255, 255, 255)) | |
| background.paste(source, (margin_x, margin_y)) | |
| # λ§μ€ν¬ μμ± | |
| mask = Image.new('L', target_size, 255) | |
| mask_draw = ImageDraw.Draw(mask) | |
| # λ§μ€ν¬ μμ 그리기 | |
| white_gaps_patch = 2 | |
| left_overlap = margin_x + overlap_x if alignment != "μΌμͺ½" else margin_x | |
| right_overlap = margin_x + new_width - overlap_x if alignment != "μ€λ₯Έμͺ½" else margin_x + new_width | |
| top_overlap = margin_y + overlap_y if alignment != "μ" else margin_y | |
| bottom_overlap = margin_y + new_height - overlap_y if alignment != "μλ" else margin_y + new_height | |
| mask_draw.rectangle([ | |
| (left_overlap, top_overlap), | |
| (right_overlap, bottom_overlap) | |
| ], fill=0) | |
| return background, mask | |
| def outpaint_image(image, prompt, width, height, overlap_percentage, alignment, num_steps=8): | |
| """μ΄λ―Έμ§ μμνμΈν μ€ν""" | |
| if image is None: | |
| return None | |
| if not OUTPAINT_MODEL_LOADED: | |
| return Image.new('RGB', (width, height), (200, 200, 200)) | |
| try: | |
| # μ΄λ―Έμ§μ λ§μ€ν¬ μ€λΉ | |
| background, mask = prepare_image_and_mask(image, width, height, overlap_percentage, alignment) | |
| if background is None: | |
| return None | |
| # cnet_image μμ± (λ§μ€ν¬ μμμ κ²μμμΌλ‘) | |
| cnet_image = background.copy() | |
| cnet_image.paste(0, (0, 0), mask) | |
| # ν둬ννΈ μ€λΉ | |
| final_prompt = f"{prompt}, high quality, 4k" if prompt else "high quality, 4k" | |
| # GPUμμ μ€ν | |
| with torch.autocast(device_type="cuda", dtype=torch.float16): | |
| ( | |
| prompt_embeds, | |
| negative_prompt_embeds, | |
| pooled_prompt_embeds, | |
| negative_pooled_prompt_embeds, | |
| ) = pipe.encode_prompt(final_prompt, "cuda", True) | |
| # μμ± νλ‘μΈμ€ | |
| for generated_image in pipe( | |
| prompt_embeds=prompt_embeds, | |
| negative_prompt_embeds=negative_prompt_embeds, | |
| pooled_prompt_embeds=pooled_prompt_embeds, | |
| negative_pooled_prompt_embeds=negative_pooled_prompt_embeds, | |
| image=cnet_image, | |
| num_inference_steps=num_steps | |
| ): | |
| # μ€κ° κ²°κ³Ό (νμμ μ¬μ©) | |
| pass | |
| # μ΅μ’ μ΄λ―Έμ§ | |
| final_image = generated_image | |
| # RGBAλ‘ λ³ννκ³ λ§μ€ν¬ μ μ© | |
| final_image = final_image.convert("RGBA") | |
| cnet_image.paste(final_image, (0, 0), mask) | |
| return cnet_image | |
| except Exception as e: | |
| logging.error(f"Outpainting error: {str(e)}") | |
| return background if 'background' in locals() else None | |
| # CSS | |
| css = """ | |
| :root { | |
| --primary-color: #f8c3cd; | |
| --secondary-color: #b3e5fc; | |
| --background-color: #f5f5f7; | |
| --card-background: #ffffff; | |
| --text-color: #424242; | |
| --accent-color: #ffb6c1; | |
| --success-color: #c8e6c9; | |
| --warning-color: #fff9c4; | |
| --shadow-color: rgba(0, 0, 0, 0.1); | |
| --border-radius: 12px; | |
| } | |
| .gradio-container { | |
| max-width: 1200px !important; | |
| margin: 0 auto !important; | |
| } | |
| .panel-box { | |
| border-radius: var(--border-radius) !important; | |
| box-shadow: 0 8px 16px var(--shadow-color) !important; | |
| background-color: var(--card-background) !important; | |
| padding: 20px !important; | |
| margin-bottom: 20px !important; | |
| } | |
| #generate-btn, #video-btn, #outpaint-btn { | |
| background: linear-gradient(135deg, #ff9a9e, #fad0c4) !important; | |
| font-size: 1.1rem !important; | |
| padding: 12px 24px !important; | |
| margin-top: 10px !important; | |
| width: 100% !important; | |
| } | |
| .tabitem { | |
| min-height: 700px !important; | |
| } | |
| """ | |
| # Gradio Interface | |
| demo = gr.Blocks(css=css, title="AI μ΄λ―Έμ§ & λΉλμ€ μμ±κΈ°") | |
| with demo: | |
| gr.Markdown("# π¨ Ginigen μ€νλμ€") | |
| with gr.Tabs() as tabs: | |
| # 첫 λ²μ§Έ ν: ν μ€νΈ to μ΄λ―Έμ§ | |
| with gr.Tab("ν μ€νΈβμ΄λ―Έμ§βλΉλμ€", elem_classes="tabitem"): | |
| with gr.Row(equal_height=True): | |
| # μ λ ₯ μ»¬λΌ | |
| with gr.Column(scale=1): | |
| with gr.Group(elem_classes="panel-box"): | |
| gr.Markdown("### π μ΄λ―Έμ§ μμ± μ€μ ") | |
| prompt = gr.Textbox( | |
| label="ν둬ννΈ(νκΈ/μμ΄ κ°λ₯)", | |
| placeholder="μμ±νκ³ μΆμ μ΄λ―Έμ§λ₯Ό μ€λͺ νμΈμ...", | |
| lines=3 | |
| ) | |
| size_preset = gr.Dropdown( | |
| choices=list(IMAGE_PRESETS.keys()), | |
| value="1:1 μ μ¬κ°ν", | |
| label="ν¬κΈ° ν리μ " | |
| ) | |
| with gr.Row(): | |
| width = gr.Slider(256, 2048, 1024, step=64, label="λλΉ") | |
| height = gr.Slider(256, 2048, 1024, step=64, label="λμ΄") | |
| with gr.Row(): | |
| guidance = gr.Slider(1.0, 20.0, 3.5, step=0.1, label="κ°μ΄λμ€") | |
| steps = gr.Slider(1, 50, 30, step=1, label="μ€ν ") | |
| seed = gr.Number(label="μλ (-1=λλ€)", value=-1) | |
| generate_btn = gr.Button("π¨ μ΄λ―Έμ§ μμ±", variant="primary", elem_id="generate-btn") | |
| with gr.Group(elem_classes="panel-box"): | |
| gr.Markdown("### π¬ λΉλμ€ μμ± μ€μ ") | |
| video_prompt = gr.Textbox( | |
| label="(μ ν) λΉλμ€ ν둬ννΈ(μμ΄λ‘ μ λ ₯)", | |
| placeholder="λΉλμ€μ μμ§μμ μ€λͺ νμΈμ... (λΉμλλ©΄ κΈ°λ³Έ μμ§μ μ μ©)", | |
| lines=2 | |
| ) | |
| video_length = gr.Slider( | |
| minimum=1, | |
| maximum=60, | |
| value=4, | |
| step=0.5, | |
| label="λΉλμ€ κΈΈμ΄ (μ΄)", | |
| info="1μ΄μμ 60μ΄κΉμ§ μ ν κ°λ₯ν©λλ€" | |
| ) | |
| # μ¬μ΄λ μμ± μ΅μ μΆκ° | |
| sound_generation = gr.Radio( | |
| choices=["μ¬μ΄λ μμ", "μ¬μ΄λ μμ±"], | |
| value="μ¬μ΄λ μμ", | |
| label="μ¬μ΄λ μ΅μ ", | |
| info="λΉλμ€μ μ¬μ΄λλ₯Ό μΆκ°ν μ§ μ ννμΈμ" | |
| ) | |
| # μ¬μ΄λ κ΄λ ¨ μ λ ₯ νλ (μ‘°κ±΄λΆ νμ) | |
| with gr.Column(visible=False) as sound_options: | |
| sound_prompt = gr.Textbox( | |
| label="μ¬μ΄λ ν둬ννΈ (μ ν)", | |
| placeholder="μμ±ν μ¬μ΄λλ₯Ό μ€λͺ νμΈμ... (λΉμλλ©΄ λΉλμ€ ν둬ννΈ μ¬μ©)", | |
| lines=2 | |
| ) | |
| sound_negative_prompt = gr.Textbox( | |
| label="μ¬μ΄λ λ€κ±°ν°λΈ ν둬ννΈ", | |
| value="music", | |
| lines=1 | |
| ) | |
| video_btn = gr.Button("π¬ λΉλμ€λ‘ λ³ν", variant="secondary", elem_id="video-btn") | |
| # μΆλ ₯ μ»¬λΌ | |
| with gr.Column(scale=1): | |
| with gr.Group(elem_classes="panel-box"): | |
| gr.Markdown("### πΌοΈ μμ± κ²°κ³Ό") | |
| output_image = gr.Image(label="μμ±λ μ΄λ―Έμ§", type="numpy") | |
| output_seed = gr.Textbox(label="μλ μ 보") | |
| output_video = gr.Video(label="μμ±λ λΉλμ€") | |
| # λ λ²μ§Έ ν: μ΄λ―Έμ§ μμνμΈν | |
| with gr.Tab("μ΄λ―Έμ§ λΉμ¨ λ³κ²½/μμ±", elem_classes="tabitem"): | |
| with gr.Row(equal_height=True): | |
| # μ λ ₯ μ»¬λΌ | |
| with gr.Column(scale=1): | |
| with gr.Group(elem_classes="panel-box"): | |
| gr.Markdown("### πΌοΈ μ΄λ―Έμ§ μ λ‘λ") | |
| input_image = gr.Image( | |
| label="μλ³Έ μ΄λ―Έμ§", | |
| type="numpy" | |
| ) | |
| outpaint_prompt = gr.Textbox( | |
| label="ν둬ννΈ (μ ν)", | |
| placeholder="νμ₯ν μμμ λν μ€λͺ ...", | |
| lines=2 | |
| ) | |
| with gr.Group(elem_classes="panel-box"): | |
| gr.Markdown("### βοΈ μμνμΈν μ€μ ") | |
| outpaint_size_preset = gr.Dropdown( | |
| choices=list(IMAGE_PRESETS.keys()), | |
| value="16:9 μμ΄λμ€ν¬λ¦°", | |
| label="λͺ©ν ν¬κΈ° ν리μ " | |
| ) | |
| with gr.Row(): | |
| outpaint_width = gr.Slider(256, 2048, 1280, step=64, label="λͺ©ν λλΉ") | |
| outpaint_height = gr.Slider(256, 2048, 720, step=64, label="λͺ©ν λμ΄") | |
| alignment = gr.Dropdown( | |
| choices=["κ°μ΄λ°", "μΌμͺ½", "μ€λ₯Έμͺ½", "μ", "μλ"], | |
| value="κ°μ΄λ°", | |
| label="μ λ ¬" | |
| ) | |
| overlap_percentage = gr.Slider( | |
| minimum=1, | |
| maximum=50, | |
| value=10, | |
| step=1, | |
| label="λ§μ€ν¬ μ€λ²λ© (%)" | |
| ) | |
| outpaint_steps = gr.Slider( | |
| minimum=4, | |
| maximum=12, | |
| value=8, | |
| step=1, | |
| label="μΆλ‘ μ€ν " | |
| ) | |
| outpaint_btn = gr.Button("π¨ μμνμΈν μ€ν", variant="primary", elem_id="outpaint-btn") | |
| # μΆλ ₯ μ»¬λΌ | |
| with gr.Column(scale=1): | |
| with gr.Group(elem_classes="panel-box"): | |
| gr.Markdown("### πΌοΈ κ²°κ³Ό") | |
| outpaint_result = gr.Image(label="μμνμΈν κ²°κ³Ό") | |
| # μ΄λ²€νΈ μ°κ²° - 첫 λ²μ§Έ ν | |
| size_preset.change(update_dimensions, [size_preset], [width, height]) | |
| generate_btn.click( | |
| generate_text_to_image, | |
| [prompt, width, height, guidance, steps, seed], | |
| [output_image, output_seed] | |
| ) | |
| # μ¬μ΄λ μ΅μ νμ/μ¨κΉ | |
| def toggle_sound_options(choice): | |
| return gr.update(visible=(choice == "μ¬μ΄λ μμ±")) | |
| sound_generation.change( | |
| toggle_sound_options, | |
| [sound_generation], | |
| [sound_options] | |
| ) | |
| video_btn.click( | |
| generate_video_from_image, | |
| [output_image, video_prompt, video_length, sound_generation, sound_prompt, sound_negative_prompt], | |
| [output_video] | |
| ) | |
| # μ΄λ²€νΈ μ°κ²° - λ λ²μ§Έ ν | |
| outpaint_size_preset.change(update_dimensions, [outpaint_size_preset], [outpaint_width, outpaint_height]) | |
| outpaint_btn.click( | |
| outpaint_image, | |
| [input_image, outpaint_prompt, outpaint_width, outpaint_height, overlap_percentage, alignment, outpaint_steps], | |
| [outpaint_result] | |
| ) | |
| demo.launch() |