File size: 13,085 Bytes
17dd751
a533ab1
e6c83b5
02623e7
dbc4275
02623e7
dbc4275
e6c83b5
8de6538
e6c83b5
f753191
e6c83b5
 
02623e7
e6c83b5
 
02623e7
 
dbc4275
 
 
 
 
02623e7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
169cdb3
 
 
f753191
169cdb3
 
 
 
 
 
02623e7
169cdb3
17ddc19
02623e7
17ddc19
8de6538
 
17ddc19
8de6538
17ddc19
 
 
 
 
 
02623e7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1c883f5
02623e7
1c883f5
02623e7
 
dbc4275
02623e7
 
 
dbc4275
 
 
02623e7
e6c83b5
 
02623e7
dbc4275
17ddc19
8de6538
17ddc19
a533ab1
17ddc19
02623e7
17ddc19
a533ab1
02623e7
 
dbc4275
02623e7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
dbc4275
1c883f5
169cdb3
02623e7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8de6538
1c883f5
a533ab1
8de6538
a533ab1
 
8de6538
a533ab1
 
 
 
 
8de6538
 
 
 
 
 
 
 
 
1c883f5
02623e7
 
8de6538
 
 
 
 
 
 
 
 
 
 
 
1c883f5
02623e7
a533ab1
169cdb3
8de6538
 
 
 
e6c83b5
02623e7
169cdb3
e6c83b5
02623e7
 
e6c83b5
169cdb3
02623e7
169cdb3
e6c83b5
02623e7
 
 
 
 
 
 
 
 
 
 
169cdb3
1c883f5
dbc4275
1c883f5
8de6538
1c883f5
 
 
169cdb3
e6c83b5
 
1c883f5
169cdb3
e6c83b5
 
8de6538
169cdb3
02623e7
e6c83b5
02623e7
 
 
1c883f5
8de6538
1c883f5
02623e7
169cdb3
 
a533ab1
02623e7
 
a533ab1
8de6538
 
1c883f5
8de6538
 
02623e7
8de6538
 
 
 
 
 
 
1c883f5
e6c83b5
1c883f5
e6c83b5
 
 
 
 
8de6538
169cdb3
e6c83b5
 
1c883f5
e6c83b5
 
169cdb3
 
1c883f5
169cdb3
 
e6c83b5
17dd751
e6c83b5
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
import streamlit as st
from PIL import Image
import torch
import torch.nn.functional as F
from torchvision import transforms
from transformers import AutoModelForImageSegmentation, AutoImageProcessor, Swin2SRForImageSuperResolution, VitMatteForImageMatting
import io
import numpy as np
import gc

# Page Configuration
st.set_page_config(layout="wide", page_title="AI Image Lab")

# --- 1. MODEL LOADING (Cached) ---

@st.cache_resource
def load_rmbg_model():
    """Option 1: The Lightweight Specialist"""
    model = AutoModelForImageSegmentation.from_pretrained("briaai/RMBG-1.4", trust_remote_code=True)
    device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
    model.to(device)
    return model, device

@st.cache_resource
def load_birefnet_model():
    """Option 2: The Heavyweight Generalist"""
    # This requires 'timm' installed
    model = AutoModelForImageSegmentation.from_pretrained("ZhengPeng7/BiRefNet", trust_remote_code=True)
    device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
    model.to(device)
    return model, device

@st.cache_resource
def load_vitmatte_model():
    """Option 3: The Refiner (Matting)"""
    # VitMatte requires a rough mask first (we use RMBG for that)
    processor = AutoImageProcessor.from_pretrained("hustvl/vitmatte-small-composition-1k")
    model = VitMatteForImageMatting.from_pretrained("hustvl/vitmatte-small-composition-1k")
    device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
    model.to(device)
    return processor, model, device

@st.cache_resource
def load_upscaler(scale=2):
    if scale == 4:
        model_id = "caidas/swin2SR-realworld-sr-x4-64-bsrgan-psnr"
    else:
        model_id = "caidas/swin2SR-classical-sr-x2-64"
    processor = AutoImageProcessor.from_pretrained(model_id)
    model = Swin2SRForImageSuperResolution.from_pretrained(model_id)
    return processor, model

# --- 2. HELPER FUNCTIONS ---

def find_mask_tensor(output):
    """Recursively finds the mask tensor in complex model outputs."""
    if isinstance(output, torch.Tensor):
        if output.dim() == 4 and output.shape[1] == 1: return output
        elif output.dim() == 3 and output.shape[0] == 1: return output
        return None
    if hasattr(output, "logits"): return find_mask_tensor(output.logits)
    elif isinstance(output, (list, tuple)):
        for item in output:
            found = find_mask_tensor(item)
            if found is not None: return found
    return None

def generate_trimap(mask_tensor, erode_kernel_size=10, dilate_kernel_size=10):
    """
    Generates a trimap (Foreground, Background, Unknown) from a binary mask 
    using Pure PyTorch (No OpenCV required).
    Values: 1=FG, 0=BG, 0.5=Unknown (Edge)
    """
    # Ensure mask is Bx1xHxW
    if mask_tensor.dim() == 3: mask_tensor = mask_tensor.unsqueeze(0)
    
    # Create kernels
    erode_k = erode_kernel_size
    dilate_k = dilate_kernel_size
    
    # Dilation (Max Pooling) - Expands the white area
    # We pad to keep size same
    dilated = F.max_pool2d(mask_tensor, kernel_size=dilate_k, stride=1, padding=dilate_k//2)
    
    # Erosion (Negative Max Pooling) - Shrinks the white area
    eroded = -F.max_pool2d(-mask_tensor, kernel_size=erode_k, stride=1, padding=erode_k//2)
    
    # Trimap construction
    # Pixels that are 1 in eroded are definitely FG (1.0)
    # Pixels that are 0 in dilated are definitely BG (0.0)
    # Everything else is the "Unknown" zone (0.5)
    
    # Start with Unknown (0.5)
    trimap = torch.full_like(mask_tensor, 0.5)
    
    # Set definites
    trimap[eroded > 0.5] = 1.0
    trimap[dilated < 0.5] = 0.0
    
    return trimap

# --- 3. INFERENCE LOGIC ---

def inference_segmentation(model, image, device, resolution=1024):
    """Generic inference for RMBG and BiRefNet."""
    w, h = image.size
    
    transform = transforms.Compose([
        transforms.Resize((resolution, resolution)),
        transforms.ToTensor(),
        transforms.Normalize([0.485, 0.456, 0.406], [0.229, 0.224, 0.225])
    ])
    input_tensor = transform(image).unsqueeze(0).to(device)

    with torch.no_grad():
        outputs = model(input_tensor)
    
    result_tensor = find_mask_tensor(outputs)
    if result_tensor is None: result_tensor = outputs[0] if isinstance(outputs, (list, tuple)) else outputs
    if not isinstance(result_tensor, torch.Tensor):
         if isinstance(result_tensor, (list, tuple)): result_tensor = result_tensor[0]

    # Get binary-ish mask (logits or sigmoid)
    pred = result_tensor.squeeze().cpu()
    if pred.max() > 1 or pred.min() < 0: pred = pred.sigmoid()
    
    # Resize back to original
    pred_pil = transforms.ToPILImage()(pred)
    mask = pred_pil.resize((w, h), resample=Image.LANCZOS)
    return mask

def inference_vitmatte(image, device):
    """
    Runs pipeline: RMBG (Rough Mask) -> Trimap -> VitMatte (Refined Mask)
    """
    # 1. Get Rough Mask using RMBG (Fast)
    rmbg_model, _ = load_rmbg_model() # Re-use loaded model
    rough_mask_pil = inference_segmentation(rmbg_model, image, device, resolution=1024)
    
    # 2. Create Trimap
    # Convert PIL mask to Tensor
    mask_tensor = transforms.ToTensor()(rough_mask_pil).to(device)
    # Generate trimap (1=FG, 0=BG, 0.5=Unknown)
    trimap_tensor = generate_trimap(mask_tensor, erode_kernel_size=25, dilate_kernel_size=25)
    
    # 3. VitMatte Inference
    processor, model, _ = load_vitmatte_model()
    
    # VitMatte expects inputs: pixel_values (image) and mask_labels (trimap)
    inputs = processor(images=image, trimaps=trimap_tensor, return_tensors="pt").to(device)
    
    with torch.no_grad():
        outputs = model(**inputs)
    
    # Output is the refined alphas
    alphas = outputs.alphas
    
    # 4. Post-process
    # Extract alpha, resize to original
    alpha_np = alphas.squeeze().cpu().numpy()
    alpha_pil = Image.fromarray((alpha_np * 255).astype("uint8"), mode="L")
    alpha_pil = alpha_pil.resize(image.size, resample=Image.LANCZOS)
    
    return alpha_pil


@st.cache_data(show_spinner=False)
def process_background_removal(image_bytes, method="RMBG-1.4"):
    image = Image.open(io.BytesIO(image_bytes)).convert("RGB")
    
    if method == "RMBG-1.4":
        model, device = load_rmbg_model()
        mask = inference_segmentation(model, image, device)
        
    elif method == "BiRefNet (Heavy)":
        model, device = load_birefnet_model()
        mask = inference_segmentation(model, image, device, resolution=1024)
        
    elif method == "VitMatte (Refiner)":
        # VitMatte needs GPU ideally, works on CPU but slow
        device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
        mask = inference_vitmatte(image, device)
    
    else:
        # Fallback
        return image

    # Apply mask
    image.putalpha(mask)
    return image

# --- Upscaling Logic (Same as before) ---
def run_swin_inference(image, processor, model):
    inputs = processor(image, return_tensors="pt")
    with torch.no_grad():
        outputs = model(**inputs)
    output = outputs.reconstruction.data.squeeze().float().cpu().clamp_(0, 1).numpy()
    output = np.moveaxis(output, 0, -1)
    output = (output * 255.0).round().astype(np.uint8)
    return Image.fromarray(output)

def upscale_chunk_logic(image, processor, model):
    if image.mode == 'RGBA':
        r, g, b, a = image.split()
        rgb_image = Image.merge('RGB', (r, g, b))
        upscaled_rgb = run_swin_inference(rgb_image, processor, model)
        upscaled_a = a.resize(upscaled_rgb.size, Image.Resampling.LANCZOS)
        return Image.merge('RGBA', (*upscaled_rgb.split(), upscaled_a))
    else:
        return run_swin_inference(image, processor, model)

def process_tiled_upscale(image, scale_factor, grid_n, progress_bar):
    processor, model = load_upscaler(scale_factor)
    w, h = image.size
    rows = cols = grid_n
    tile_w = w // cols
    tile_h = h // rows
    overlap = 32 
    full_image = Image.new(image.mode, (w * scale_factor, h * scale_factor))
    total_tiles = rows * cols
    count = 0
    for y in range(rows):
        for x in range(cols):
            target_left = x * tile_w
            target_upper = y * tile_h
            target_right = w if x == cols - 1 else (x + 1) * tile_w
            target_lower = h if y == rows - 1 else (y + 1) * tile_h
            source_left = max(0, target_left - overlap)
            source_upper = max(0, target_upper - overlap)
            source_right = min(w, target_right + overlap)
            source_lower = min(h, target_lower + overlap)
            tile = image.crop((source_left, source_upper, source_right, source_lower))
            upscaled_tile = upscale_chunk_logic(tile, processor, model)
            target_w = target_right - target_left
            target_h = target_lower - target_upper
            extra_left = target_left - source_left
            extra_upper = target_upper - source_upper
            crop_x = extra_left * scale_factor
            crop_y = extra_upper * scale_factor
            crop_w = target_w * scale_factor
            crop_h = target_h * scale_factor
            clean_tile = upscaled_tile.crop((crop_x, crop_y, crop_x + crop_w, crop_y + crop_h))
            paste_x = target_left * scale_factor
            paste_y = target_upper * scale_factor
            full_image.paste(clean_tile, (paste_x, paste_y))
            del tile, upscaled_tile, clean_tile
            gc.collect()
            count += 1
            progress_bar.progress(count / total_tiles, text=f"Upscaling Tile {count}/{total_tiles}...")
    return full_image

def convert_image_to_bytes(img):
    buf = io.BytesIO()
    img.save(buf, format="PNG")
    return buf.getvalue()

# --- 4. MAIN APP ---

def main():
    st.title("✨ AI Image Lab: Ultimate Edition")
    st.markdown("Features: **Multi-Model Background** | **Swin2SR** | **Progress Bar**")

    # --- Sidebar ---
    st.sidebar.header("1. Background Removal")
    remove_bg = st.sidebar.checkbox("Remove Background", value=False)
    
    # NEW: Model Selector
    if remove_bg:
        bg_model = st.sidebar.selectbox(
            "Select AI Model",
            ["RMBG-1.4", "BiRefNet (Heavy)", "VitMatte (Refiner)"],
            index=0,
            help="RMBG: Fast, Standard Quality.\nBiRefNet: Slower, Better Edges.\nVitMatte: Slowest, Best for Hair/Transparency."
        )
    else:
        bg_model = "None"

    st.sidebar.header("2. AI Upscaling")
    upscale_mode = st.sidebar.radio("Magnification", ["None", "2x", "4x"])
    
    if upscale_mode != "None":
        grid_n = st.sidebar.slider("Grid Split", 2, 8, 4, help="Higher = Safer RAM usage")
    else:
        grid_n = 2

    st.sidebar.header("3. Geometry")
    rotate_angle = st.sidebar.slider("Rotate", -180, 180, 0, 1)

    # --- Main Logic ---
    uploaded_file = st.file_uploader("Upload Image", type=["png", "jpg", "jpeg", "webp"])

    if uploaded_file is not None:
        file_bytes = uploaded_file.getvalue()
        
        # 1. Background
        if remove_bg:
            # We add the model name to the spinner text so user knows what's happening
            with st.spinner(f"Removing background using {bg_model}..."):
                processed_image = process_background_removal(file_bytes, bg_model)
        else:
            processed_image = Image.open(io.BytesIO(file_bytes)).convert("RGB")

        # 2. Upscaling
        if upscale_mode != "None":
            scale = 4 if "4x" in upscale_mode else 2
            
            # Cache Key includes model name now
            cache_key = f"{uploaded_file.name}_{bg_model}_{scale}_{grid_n}_v5"
            
            if "upscale_cache" not in st.session_state:
                st.session_state.upscale_cache = {}
            
            if cache_key in st.session_state.upscale_cache:
                processed_image = st.session_state.upscale_cache[cache_key]
                st.info("✅ Loaded upscaled image from cache")
            else:
                progress_bar = st.progress(0, text="Initializing AI models...")
                processed_image = process_tiled_upscale(processed_image, scale, grid_n, progress_bar)
                progress_bar.empty()
                st.session_state.upscale_cache[cache_key] = processed_image
        
        # 3. Geometry
        final_image = processed_image.copy()
        if rotate_angle != 0:
            final_image = final_image.rotate(rotate_angle, expand=True)

        # --- Display ---
        col1, col2 = st.columns(2)
        with col1:
            st.subheader("Original")
            st.image(Image.open(io.BytesIO(file_bytes)), use_container_width=True)
        
        with col2:
            st.subheader("Result")
            st.image(final_image, use_container_width=True)

        st.markdown("---")
        st.download_button(
            label="💾 Download Result (PNG)",
            data=convert_image_to_bytes(final_image),
            file_name="processed_image.png",
            mime="image/png"
        )

if __name__ == "__main__":
    main()