Spaces:
Build error
Build error
Update app.py
Browse files
app.py
CHANGED
|
@@ -65,7 +65,9 @@ from PIL import Image
|
|
| 65 |
import numpy as np
|
| 66 |
|
| 67 |
# Simple Generative Inference function
|
| 68 |
-
def simple_generative_inference(
|
|
|
|
|
|
|
| 69 |
"""
|
| 70 |
Perform Generative Perceptual Inference on the input image.
|
| 71 |
:param image: Input image as a PIL image.
|
|
@@ -73,6 +75,8 @@ def simple_generative_inference(image, mode, model, n_iterations=10, step_size=0
|
|
| 73 |
:param model: Pretrained PyTorch model.
|
| 74 |
:param n_iterations: Number of inference iterations.
|
| 75 |
:param step_size: Step size for gradient-based updates.
|
|
|
|
|
|
|
| 76 |
:return: Processed image and gradient visualization.
|
| 77 |
"""
|
| 78 |
# Preprocess image
|
|
@@ -81,51 +85,83 @@ def simple_generative_inference(image, mode, model, n_iterations=10, step_size=0
|
|
| 81 |
transforms.ToTensor(),
|
| 82 |
transforms.Normalize(mean=[0.5, 0.5, 0.5], std=[0.5, 0.5, 0.5]) # Adjust normalization as needed
|
| 83 |
])
|
| 84 |
-
image_tensor = transform(image).unsqueeze(0)
|
|
|
|
| 85 |
|
| 86 |
optimizer = torch.optim.SGD([image_tensor], lr=step_size)
|
| 87 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 88 |
for _ in range(n_iterations):
|
| 89 |
optimizer.zero_grad()
|
| 90 |
output = model(image_tensor)
|
| 91 |
|
| 92 |
# Define inference loss based on mode
|
| 93 |
if mode == "increase confidence":
|
| 94 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 95 |
elif mode == "ReverseDiffuse":
|
| 96 |
-
noisy_image = image_tensor + torch.randn_like(image_tensor) * 0.1
|
| 97 |
loss = torch.nn.functional.mse_loss(image_tensor, noisy_image)
|
| 98 |
else:
|
| 99 |
raise ValueError("Invalid mode selected. Choose 'increase confidence' or 'ReverseDiffuse'.")
|
| 100 |
|
|
|
|
| 101 |
loss.backward()
|
| 102 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 103 |
|
| 104 |
# Generate gradient visualization
|
| 105 |
-
|
| 106 |
-
grad_image = (
|
| 107 |
grad_image = Image.fromarray((grad_image * 255).astype(np.uint8))
|
| 108 |
|
| 109 |
# Convert final processed image back to PIL format
|
| 110 |
processed_image = image_tensor.detach().squeeze().permute(1, 2, 0).cpu().numpy()
|
| 111 |
-
processed_image = (processed_image - processed_image.min()) / (processed_image.max() - processed_image.min())
|
| 112 |
processed_image = Image.fromarray((processed_image * 255).astype(np.uint8))
|
| 113 |
|
| 114 |
return processed_image, grad_image
|
| 115 |
|
| 116 |
# Gradio Interface
|
| 117 |
iface = gr.Interface(
|
| 118 |
-
fn=lambda image, mode
|
|
|
|
|
|
|
| 119 |
inputs=[
|
| 120 |
gr.Image(type="pil", label="Input Image"), # Input image
|
| 121 |
-
gr.Radio(["increase confidence", "ReverseDiffuse"], label="Inference Mode") # Mode selection
|
|
|
|
|
|
|
|
|
|
|
|
|
| 122 |
],
|
| 123 |
outputs=[
|
| 124 |
gr.Image(label="Processed Image"), # Processed image
|
| 125 |
gr.Image(label="Gradient Visualization") # Gradient visualization
|
| 126 |
],
|
| 127 |
-
title="Generative Inference"
|
|
|
|
| 128 |
)
|
| 129 |
|
| 130 |
|
|
|
|
| 131 |
iface.launch()
|
|
|
|
| 65 |
import numpy as np
|
| 66 |
|
| 67 |
# Simple Generative Inference function
|
| 68 |
+
def simple_generative_inference(
|
| 69 |
+
image, mode, model, n_iterations=10, step_size=0.01, noise_ratio=0.1, eps=0.1
|
| 70 |
+
):
|
| 71 |
"""
|
| 72 |
Perform Generative Perceptual Inference on the input image.
|
| 73 |
:param image: Input image as a PIL image.
|
|
|
|
| 75 |
:param model: Pretrained PyTorch model.
|
| 76 |
:param n_iterations: Number of inference iterations.
|
| 77 |
:param step_size: Step size for gradient-based updates.
|
| 78 |
+
:param noise_ratio: Ratio of noise to be added in ReverseDiffuse mode.
|
| 79 |
+
:param eps: Constraint on perturbation magnitude.
|
| 80 |
:return: Processed image and gradient visualization.
|
| 81 |
"""
|
| 82 |
# Preprocess image
|
|
|
|
| 85 |
transforms.ToTensor(),
|
| 86 |
transforms.Normalize(mean=[0.5, 0.5, 0.5], std=[0.5, 0.5, 0.5]) # Adjust normalization as needed
|
| 87 |
])
|
| 88 |
+
image_tensor = transform(image).unsqueeze(0)
|
| 89 |
+
image_tensor.requires_grad_(True) # Enable gradient computation for the image tensor
|
| 90 |
|
| 91 |
optimizer = torch.optim.SGD([image_tensor], lr=step_size)
|
| 92 |
|
| 93 |
+
# Define least likely classes for "increase confidence" mode
|
| 94 |
+
if mode == "increase confidence":
|
| 95 |
+
with torch.no_grad():
|
| 96 |
+
output = model(image_tensor)
|
| 97 |
+
probs = torch.nn.functional.softmax(output, dim=1)
|
| 98 |
+
_, least_likely_classes = torch.topk(probs, k=5, largest=False, dim=1)
|
| 99 |
+
|
| 100 |
+
# Create noisy image (only for ReverseDiffuse mode)
|
| 101 |
+
if mode == "ReverseDiffuse":
|
| 102 |
+
noisy_image = image_tensor + torch.randn_like(image_tensor) * noise_ratio
|
| 103 |
+
|
| 104 |
for _ in range(n_iterations):
|
| 105 |
optimizer.zero_grad()
|
| 106 |
output = model(image_tensor)
|
| 107 |
|
| 108 |
# Define inference loss based on mode
|
| 109 |
if mode == "increase confidence":
|
| 110 |
+
losses = []
|
| 111 |
+
for idx in least_likely_classes[0]: # Iterate over least likely classes
|
| 112 |
+
target = torch.full((1,), idx, dtype=torch.long, device=output.device)
|
| 113 |
+
loss = torch.nn.functional.cross_entropy(output, target)
|
| 114 |
+
losses.append(loss)
|
| 115 |
+
loss = torch.stack(losses).mean() # Average loss over least likely classes
|
| 116 |
elif mode == "ReverseDiffuse":
|
|
|
|
| 117 |
loss = torch.nn.functional.mse_loss(image_tensor, noisy_image)
|
| 118 |
else:
|
| 119 |
raise ValueError("Invalid mode selected. Choose 'increase confidence' or 'ReverseDiffuse'.")
|
| 120 |
|
| 121 |
+
# Compute gradients and update the image
|
| 122 |
loss.backward()
|
| 123 |
+
grad = image_tensor.grad.data
|
| 124 |
+
grad_norm = grad.view(grad.shape[0], -1).norm(dim=1, keepdim=True).view_as(image_tensor)
|
| 125 |
+
scaled_grad = grad / (grad_norm + 1e-10)
|
| 126 |
+
image_tensor = torch.clamp(
|
| 127 |
+
image_tensor + step_size * scaled_grad,
|
| 128 |
+
min=image_tensor - eps,
|
| 129 |
+
max=image_tensor + eps
|
| 130 |
+
)
|
| 131 |
|
| 132 |
# Generate gradient visualization
|
| 133 |
+
grad_visualization = image_tensor.grad.abs().mean(dim=1).squeeze().cpu().numpy()
|
| 134 |
+
grad_image = (grad_visualization - grad_visualization.min()) / (grad_visualization.max() - grad_visualization.min())
|
| 135 |
grad_image = Image.fromarray((grad_image * 255).astype(np.uint8))
|
| 136 |
|
| 137 |
# Convert final processed image back to PIL format
|
| 138 |
processed_image = image_tensor.detach().squeeze().permute(1, 2, 0).cpu().numpy()
|
| 139 |
+
processed_image = (processed_image - processed_image.min()) / (processed_image.max() - processed_image.min())
|
| 140 |
processed_image = Image.fromarray((processed_image * 255).astype(np.uint8))
|
| 141 |
|
| 142 |
return processed_image, grad_image
|
| 143 |
|
| 144 |
# Gradio Interface
|
| 145 |
iface = gr.Interface(
|
| 146 |
+
fn=lambda image, mode, step_size, eps, noise_ratio, n_iterations: simple_generative_inference(
|
| 147 |
+
image, mode, model, step_size=step_size, eps=eps, noise_ratio=noise_ratio, n_iterations=n_iterations
|
| 148 |
+
),
|
| 149 |
inputs=[
|
| 150 |
gr.Image(type="pil", label="Input Image"), # Input image
|
| 151 |
+
gr.Radio(["increase confidence", "ReverseDiffuse"], label="Inference Mode"), # Mode selection
|
| 152 |
+
gr.Slider(0.001, 1.0, value=0.01, step=0.001, label="Step Size"), # Step size
|
| 153 |
+
gr.Slider(0.001, 0.5, value=0.1, step=0.001, label="Epsilon (eps)"), # Epsilon constraint
|
| 154 |
+
gr.Slider(0.0, 0.5, value=0.1, step=0.01, label="Noise Ratio"), # Noise ratio
|
| 155 |
+
gr.Slider(1, 100, value=10, step=1, label="Number of Iterations"), # Number of iterations
|
| 156 |
],
|
| 157 |
outputs=[
|
| 158 |
gr.Image(label="Processed Image"), # Processed image
|
| 159 |
gr.Image(label="Gradient Visualization") # Gradient visualization
|
| 160 |
],
|
| 161 |
+
title="Generative Perceptual Inference (GPI)",
|
| 162 |
+
description="Perform GPI on input images using adjustable parameters such as step size, epsilon, noise ratio, and number of iterations."
|
| 163 |
)
|
| 164 |
|
| 165 |
|
| 166 |
+
|
| 167 |
iface.launch()
|