""" Sentinel Tiny Image Space — CIFAR-10 diffusion with Sentinel noise schedule """ import gradio as gr import torch from diffusers import DDPMScheduler, UNet2DModel, DDPMPipeline from PIL import Image import numpy as np import json # ─── Load Model ────────────────────────────────────────────────────────────── MODEL_ID = "5dimension/sentinel-tiny-image" model_status = "⏳ Loading model..." pipeline = None try: pipeline = DDPMPipeline.from_pretrained(MODEL_ID) pipeline.unet.eval() model_status = f"✅ Model loaded — {sum(p.numel() for p in pipeline.unet.parameters()):,} params" except Exception as e: # Fallback: create from config try: unet = UNet2DModel( sample_size=32, in_channels=3, out_channels=3, layers_per_block=1, block_out_channels=(32, 64, 64, 128), down_block_types=("DownBlock2D",)*4, up_block_types=("UpBlock2D",)*4, time_embedding_type="positional", ) scheduler = DDPMScheduler(num_train_timesteps=1000) pipeline = DDPMPipeline(unet=unet, scheduler=scheduler) pipeline.unet.eval() model_status = f"⚠️ Using fresh UNet ({sum(p.numel() for p in unet.parameters()):,} params)" except: model_status = f"❌ Failed: {str(e)[:200]}" # ─── Generate Images ───────────────────────────────────────────────────────── def generate_images(batch_size=4, num_steps=50, seed=42): if pipeline is None: return [Image.new('RGB', (256, 256), color='gray') for _ in range(batch_size)] generator = torch.manual_seed(seed) if seed > 0 else None with torch.no_grad(): results = pipeline( batch_size=batch_size, num_inference_steps=num_steps, generator=generator, ).images # Upscale from 32x32 to 256x256 for display upscaled = [] for img in results: upscaled.append(img.resize((256, 256), Image.NEAREST)) return upscaled # ─── UI ──────────────────────────────────────────────────────────────────────── with gr.Blocks(title="🎨 Sentinel Tiny Image", css=""" .gradio-container { max-width: 1000px; margin: 0 auto; } .title { text-align: center; font-size: 2em; font-weight: bold; color: #6b4c9a; } .subtitle { text-align: center; color: #888; margin-bottom: 1em; } """) as demo: gr.Markdown("""