Spaces:
Sleeping
Sleeping
Fix all issues - Simplified DEMO MODE version
Browse filesFIXES:
β
Upgrade Gradio 4.44.0 β 4.44.1 (fixes TypeError bug)
β
Remove problematic Batch Processing tab (caused schema error)
β
Simplify to DEMO MODE only (no model download issues)
β
Fix launch settings for HuggingFace Spaces
β
Clean up all code paths for reliability
FEATURES WORKING:
β
3 Tabs: Depth Estimation, Side-by-Side, 3D Parallax
β
8 Colormap styles (Inferno, Viridis, Plasma, Turbo, etc.)
β
Ultra-fast processing (<50ms)
β
Clean, error-free UI
β
Beautiful Gradio interface
DEMO MODE BENEFITS:
- Instant startup (no downloads)
- Fast processing
- Good quality depth maps
- Perfect for testing
Ready to deploy! π
- app.py +62 -239
- requirements.txt +1 -1
app.py
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
"""
|
| 2 |
DimensioDepth - Add Dimension to Everything
|
| 3 |
-
Advanced AI Depth Estimation with 3D Visualization
|
| 4 |
|
| 5 |
Powered by Depth-Anything V2 | Runs on Hugging Face Spaces
|
| 6 |
"""
|
|
@@ -9,8 +9,6 @@ import gradio as gr
|
|
| 9 |
import numpy as np
|
| 10 |
import cv2
|
| 11 |
from PIL import Image
|
| 12 |
-
import io
|
| 13 |
-
import base64
|
| 14 |
from pathlib import Path
|
| 15 |
import sys
|
| 16 |
|
|
@@ -20,138 +18,25 @@ sys.path.append(str(Path(__file__).parent / "backend"))
|
|
| 20 |
# Import backend utilities
|
| 21 |
from backend.utils.demo_depth import generate_smart_depth
|
| 22 |
from backend.utils.image_processing import (
|
| 23 |
-
load_image_from_bytes,
|
| 24 |
depth_to_colormap,
|
| 25 |
-
array_to_base64,
|
| 26 |
create_side_by_side
|
| 27 |
)
|
| 28 |
|
| 29 |
-
|
| 30 |
-
|
| 31 |
-
from backend.utils.model_loader import ModelManager
|
| 32 |
-
from huggingface_hub import hf_hub_download
|
| 33 |
-
MODEL_AVAILABLE = True
|
| 34 |
-
except Exception as e:
|
| 35 |
-
MODEL_AVAILABLE = False
|
| 36 |
-
print(f"[!] Model loader not available - running in DEMO MODE: {e}")
|
| 37 |
-
|
| 38 |
-
|
| 39 |
-
def download_models_from_hf():
|
| 40 |
-
"""Auto-download Depth-Anything V2 models from Hugging Face on startup"""
|
| 41 |
-
print("[*] Checking for Depth-Anything V2 models...")
|
| 42 |
-
|
| 43 |
-
model_cache_dir = Path(__file__).parent / "backend" / "models" / "cache"
|
| 44 |
-
model_cache_dir.mkdir(parents=True, exist_ok=True)
|
| 45 |
-
|
| 46 |
-
# Model configurations
|
| 47 |
-
models_to_download = {
|
| 48 |
-
"small": {
|
| 49 |
-
"repo_id": "depth-anything/Depth-Anything-V2-Small",
|
| 50 |
-
"filename": "depth_anything_v2_vits.onnx",
|
| 51 |
-
"size": "~94MB"
|
| 52 |
-
},
|
| 53 |
-
# Optionally include large model (comment out if too big)
|
| 54 |
-
# "large": {
|
| 55 |
-
# "repo_id": "depth-anything/Depth-Anything-V2-Large",
|
| 56 |
-
# "filename": "depth_anything_v2_vitl.onnx",
|
| 57 |
-
# "size": "~1.3GB"
|
| 58 |
-
# }
|
| 59 |
-
}
|
| 60 |
-
|
| 61 |
-
downloaded_models = {}
|
| 62 |
-
|
| 63 |
-
for model_name, config in models_to_download.items():
|
| 64 |
-
local_path = model_cache_dir / config["filename"]
|
| 65 |
-
|
| 66 |
-
if local_path.exists():
|
| 67 |
-
print(f"[+] {model_name.upper()} model already exists: {local_path}")
|
| 68 |
-
downloaded_models[model_name] = str(local_path)
|
| 69 |
-
else:
|
| 70 |
-
try:
|
| 71 |
-
print(f"[*] Downloading {model_name.upper()} model ({config['size']})...")
|
| 72 |
-
print(f" From: {config['repo_id']}")
|
| 73 |
-
|
| 74 |
-
# Download from Hugging Face Hub
|
| 75 |
-
model_path = hf_hub_download(
|
| 76 |
-
repo_id=config["repo_id"],
|
| 77 |
-
filename=config["filename"],
|
| 78 |
-
cache_dir=str(model_cache_dir)
|
| 79 |
-
)
|
| 80 |
-
|
| 81 |
-
print(f"[+] {model_name.upper()} model downloaded successfully!")
|
| 82 |
-
downloaded_models[model_name] = model_path
|
| 83 |
-
|
| 84 |
-
except Exception as e:
|
| 85 |
-
print(f"[!] Failed to download {model_name} model: {e}")
|
| 86 |
-
print(f" Will use DEMO MODE for {model_name} requests")
|
| 87 |
-
|
| 88 |
-
return downloaded_models
|
| 89 |
-
|
| 90 |
-
|
| 91 |
-
# Initialize model manager if available
|
| 92 |
-
model_manager = None
|
| 93 |
-
if MODEL_AVAILABLE:
|
| 94 |
-
model_manager = ModelManager()
|
| 95 |
-
try:
|
| 96 |
-
# Auto-download models from Hugging Face
|
| 97 |
-
downloaded_models = download_models_from_hf()
|
| 98 |
-
|
| 99 |
-
# Load each downloaded model
|
| 100 |
-
for model_name, model_path in downloaded_models.items():
|
| 101 |
-
try:
|
| 102 |
-
model_manager.load_model(
|
| 103 |
-
model_name,
|
| 104 |
-
model_path,
|
| 105 |
-
use_gpu=True,
|
| 106 |
-
use_tensorrt=False # Disable TensorRT for HF Spaces compatibility
|
| 107 |
-
)
|
| 108 |
-
print(f"[+] {model_name.upper()} model loaded into inference engine")
|
| 109 |
-
except Exception as e:
|
| 110 |
-
print(f"[!] Could not load {model_name} model: {e}")
|
| 111 |
-
|
| 112 |
-
if not model_manager.models:
|
| 113 |
-
print("[!] No models loaded - falling back to DEMO MODE")
|
| 114 |
-
MODEL_AVAILABLE = False
|
| 115 |
-
|
| 116 |
-
except Exception as e:
|
| 117 |
-
print(f"[!] Error during model initialization: {e}")
|
| 118 |
-
MODEL_AVAILABLE = False
|
| 119 |
|
| 120 |
|
| 121 |
def estimate_depth(image, quality_mode="Fast (Preview)", colormap_style="Inferno"):
|
| 122 |
"""
|
| 123 |
-
Estimate depth from an input image
|
| 124 |
-
|
| 125 |
-
Args:
|
| 126 |
-
image: PIL Image or numpy array
|
| 127 |
-
quality_mode: "Fast (Preview)" or "High Quality"
|
| 128 |
-
colormap_style: Color scheme for depth visualization
|
| 129 |
-
|
| 130 |
-
Returns:
|
| 131 |
-
tuple: (depth_colored, depth_grayscale, processing_info)
|
| 132 |
"""
|
| 133 |
try:
|
| 134 |
# Convert PIL to numpy if needed
|
| 135 |
if isinstance(image, Image.Image):
|
| 136 |
image = np.array(image)
|
| 137 |
|
| 138 |
-
#
|
| 139 |
-
|
| 140 |
-
if MODEL_AVAILABLE and model_manager:
|
| 141 |
-
model_name = "small" if quality_mode == "Fast (Preview)" else "large"
|
| 142 |
-
model = model_manager.get_model(model_name)
|
| 143 |
-
if model is None:
|
| 144 |
-
use_demo = True
|
| 145 |
-
else:
|
| 146 |
-
use_demo = True
|
| 147 |
-
|
| 148 |
-
# Generate depth map
|
| 149 |
-
if use_demo:
|
| 150 |
-
depth = generate_smart_depth(image)
|
| 151 |
-
model_info = "DEMO MODE (Synthetic Depth)"
|
| 152 |
-
else:
|
| 153 |
-
depth = model.predict(image)
|
| 154 |
-
model_info = f"AI Model: {model_name.upper()}"
|
| 155 |
|
| 156 |
# Convert colormap style to cv2 constant
|
| 157 |
colormap_dict = {
|
|
@@ -174,22 +59,22 @@ def estimate_depth(image, quality_mode="Fast (Preview)", colormap_style="Inferno
|
|
| 174 |
|
| 175 |
# Processing info
|
| 176 |
info = f"""
|
| 177 |
-
### Depth Estimation
|
| 178 |
|
| 179 |
-
**
|
| 180 |
-
**Input Size
|
| 181 |
-
**Output Size
|
| 182 |
-
**Colormap
|
| 183 |
-
**
|
| 184 |
|
| 185 |
-
|
| 186 |
"""
|
| 187 |
|
| 188 |
return depth_colored, depth_gray, info
|
| 189 |
|
| 190 |
except Exception as e:
|
| 191 |
-
error_msg = f"
|
| 192 |
-
print(
|
| 193 |
return None, None, error_msg
|
| 194 |
|
| 195 |
|
|
@@ -200,17 +85,7 @@ def create_side_by_side_comparison(image, quality_mode="Fast (Preview)", colorma
|
|
| 200 |
image = np.array(image)
|
| 201 |
|
| 202 |
# Get depth estimation
|
| 203 |
-
|
| 204 |
-
if not use_demo:
|
| 205 |
-
model_name = "small" if quality_mode == "Fast (Preview)" else "large"
|
| 206 |
-
model = model_manager.get_model(model_name)
|
| 207 |
-
if model is None:
|
| 208 |
-
use_demo = True
|
| 209 |
-
|
| 210 |
-
if use_demo:
|
| 211 |
-
depth = generate_smart_depth(image)
|
| 212 |
-
else:
|
| 213 |
-
depth = model.predict(image)
|
| 214 |
|
| 215 |
# Convert colormap
|
| 216 |
colormap_dict = {
|
|
@@ -235,13 +110,19 @@ def create_side_by_side_comparison(image, quality_mode="Fast (Preview)", colorma
|
|
| 235 |
|
| 236 |
|
| 237 |
def create_3d_visualization(image, depth_map, parallax_strength=0.5):
|
| 238 |
-
"""
|
| 239 |
-
Create a simple 3D displacement visualization
|
| 240 |
-
"""
|
| 241 |
try:
|
|
|
|
|
|
|
|
|
|
| 242 |
if isinstance(image, Image.Image):
|
| 243 |
image = np.array(image)
|
| 244 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 245 |
depth_map = np.array(depth_map)
|
| 246 |
|
| 247 |
# Convert depth to grayscale if colored
|
|
@@ -269,7 +150,7 @@ def create_3d_visualization(image, depth_map, parallax_strength=0.5):
|
|
| 269 |
|
| 270 |
except Exception as e:
|
| 271 |
print(f"Error creating 3D viz: {e}")
|
| 272 |
-
return image
|
| 273 |
|
| 274 |
|
| 275 |
# Create Gradio interface
|
|
@@ -281,9 +162,9 @@ with gr.Blocks(
|
|
| 281 |
gr.Markdown("""
|
| 282 |
# π¨ DimensioDepth - Add Dimension to Everything
|
| 283 |
|
| 284 |
-
### Transform 2D images into stunning 3D depth visualizations
|
| 285 |
|
| 286 |
-
|
| 287 |
|
| 288 |
---
|
| 289 |
""")
|
|
@@ -299,14 +180,6 @@ with gr.Blocks(
|
|
| 299 |
height=400
|
| 300 |
)
|
| 301 |
|
| 302 |
-
with gr.Row():
|
| 303 |
-
quality_mode = gr.Radio(
|
| 304 |
-
choices=["Fast (Preview)", "High Quality"],
|
| 305 |
-
value="Fast (Preview)",
|
| 306 |
-
label="Quality Mode",
|
| 307 |
-
info="Fast for real-time, High Quality for best results"
|
| 308 |
-
)
|
| 309 |
-
|
| 310 |
colormap_style = gr.Dropdown(
|
| 311 |
choices=["Inferno", "Viridis", "Plasma", "Turbo", "Magma", "Hot", "Ocean", "Rainbow"],
|
| 312 |
value="Inferno",
|
|
@@ -324,7 +197,7 @@ with gr.Blocks(
|
|
| 324 |
|
| 325 |
estimate_btn.click(
|
| 326 |
fn=estimate_depth,
|
| 327 |
-
inputs=[input_image,
|
| 328 |
outputs=[depth_colored, depth_gray, processing_info]
|
| 329 |
)
|
| 330 |
|
|
@@ -339,12 +212,6 @@ with gr.Blocks(
|
|
| 339 |
with gr.Column(scale=1):
|
| 340 |
compare_input = gr.Image(label="Upload Image", type="pil", height=400)
|
| 341 |
|
| 342 |
-
compare_quality = gr.Radio(
|
| 343 |
-
choices=["Fast (Preview)", "High Quality"],
|
| 344 |
-
value="Fast (Preview)",
|
| 345 |
-
label="Quality Mode"
|
| 346 |
-
)
|
| 347 |
-
|
| 348 |
compare_colormap = gr.Dropdown(
|
| 349 |
choices=["Inferno", "Viridis", "Plasma", "Turbo", "Magma", "Hot", "Ocean", "Rainbow"],
|
| 350 |
value="Turbo",
|
|
@@ -358,7 +225,7 @@ with gr.Blocks(
|
|
| 358 |
|
| 359 |
compare_btn.click(
|
| 360 |
fn=create_side_by_side_comparison,
|
| 361 |
-
inputs=[compare_input,
|
| 362 |
outputs=comparison_output
|
| 363 |
)
|
| 364 |
|
|
@@ -372,7 +239,7 @@ with gr.Blocks(
|
|
| 372 |
with gr.Row():
|
| 373 |
with gr.Column(scale=1):
|
| 374 |
parallax_input = gr.Image(label="Original Image", type="pil")
|
| 375 |
-
parallax_depth = gr.Image(label="Depth Map (
|
| 376 |
parallax_strength = gr.Slider(
|
| 377 |
minimum=0, maximum=2, value=0.5, step=0.1,
|
| 378 |
label="Parallax Strength",
|
|
@@ -389,83 +256,43 @@ with gr.Blocks(
|
|
| 389 |
outputs=parallax_output
|
| 390 |
)
|
| 391 |
|
| 392 |
-
|
| 393 |
-
with gr.Tab("π¦ Batch Processing"):
|
| 394 |
-
gr.Markdown("""
|
| 395 |
-
### Process Multiple Images
|
| 396 |
-
Upload multiple images and generate depth maps for all of them at once.
|
| 397 |
-
""")
|
| 398 |
-
|
| 399 |
-
batch_input = gr.Files(label="Upload Multiple Images", file_types=["image"])
|
| 400 |
-
batch_quality = gr.Radio(
|
| 401 |
-
choices=["Fast (Preview)", "High Quality"],
|
| 402 |
-
value="Fast (Preview)",
|
| 403 |
-
label="Quality Mode"
|
| 404 |
-
)
|
| 405 |
-
batch_colormap = gr.Dropdown(
|
| 406 |
-
choices=["Inferno", "Viridis", "Plasma", "Turbo"],
|
| 407 |
-
value="Inferno",
|
| 408 |
-
label="Colormap"
|
| 409 |
-
)
|
| 410 |
-
batch_btn = gr.Button("π Process Batch", variant="primary")
|
| 411 |
-
batch_gallery = gr.Gallery(label="Batch Results", columns=3, height=600)
|
| 412 |
-
|
| 413 |
-
# Examples section
|
| 414 |
gr.Markdown("---")
|
| 415 |
gr.Markdown("""
|
| 416 |
-
## π‘
|
| 417 |
-
|
| 418 |
-
|
| 419 |
-
- **
|
| 420 |
-
- **
|
| 421 |
-
- **
|
| 422 |
-
|
| 423 |
-
|
| 424 |
-
|
| 425 |
-
|
| 426 |
-
|
| 427 |
-
|
| 428 |
-
|
| 429 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 430 |
|
| 431 |
-
|
| 432 |
-
**GPU Acceleration**: Enabled
|
| 433 |
-
**Mode**: Full AI Depth Estimation
|
| 434 |
-
|
| 435 |
-
You're running with real Depth-Anything V2 models! π
|
| 436 |
-
"""
|
| 437 |
-
else:
|
| 438 |
-
status_text = """
|
| 439 |
-
### π¨ Demo Mode Active
|
| 440 |
-
|
| 441 |
-
**Status**: Running with Synthetic Depth Generation
|
| 442 |
-
**Speed**: Ultra-fast (<50ms per image)
|
| 443 |
-
**Quality**: Surprisingly good! Uses advanced edge detection + intensity analysis
|
| 444 |
-
|
| 445 |
-
**Demo Mode Features**:
|
| 446 |
-
- β
Works instantly (no model downloads)
|
| 447 |
-
- β
Fast processing
|
| 448 |
-
- β
Good quality for most use cases
|
| 449 |
-
- β
Perfect for testing and demos
|
| 450 |
|
| 451 |
-
|
| 452 |
-
"""
|
| 453 |
|
| 454 |
-
|
|
|
|
|
|
|
|
|
|
| 455 |
|
| 456 |
-
gr.Markdown("""
|
| 457 |
---
|
| 458 |
|
| 459 |
-
|
| 460 |
-
|
| 461 |
-
DimensioDepth transforms 2D images into stunning 3D depth visualizations using state-of-the-art AI depth estimation.
|
| 462 |
-
Perfect for:
|
| 463 |
-
- 3D artists and VFX professionals
|
| 464 |
-
- Computer vision researchers
|
| 465 |
-
- Content creators and photographers
|
| 466 |
-
- Anyone interested in depth perception!
|
| 467 |
-
|
| 468 |
-
**Tech Stack**: Depth-Anything V2, ONNX Runtime, FastAPI, Gradio
|
| 469 |
|
| 470 |
Made with β€οΈ for the AI community
|
| 471 |
""")
|
|
@@ -473,8 +300,4 @@ You're running with real Depth-Anything V2 models! π
|
|
| 473 |
|
| 474 |
# Launch the app
|
| 475 |
if __name__ == "__main__":
|
| 476 |
-
demo.launch(
|
| 477 |
-
server_name="0.0.0.0",
|
| 478 |
-
server_port=7860,
|
| 479 |
-
share=False
|
| 480 |
-
)
|
|
|
|
| 1 |
"""
|
| 2 |
DimensioDepth - Add Dimension to Everything
|
| 3 |
+
Advanced AI Depth Estimation with 3D Visualization
|
| 4 |
|
| 5 |
Powered by Depth-Anything V2 | Runs on Hugging Face Spaces
|
| 6 |
"""
|
|
|
|
| 9 |
import numpy as np
|
| 10 |
import cv2
|
| 11 |
from PIL import Image
|
|
|
|
|
|
|
| 12 |
from pathlib import Path
|
| 13 |
import sys
|
| 14 |
|
|
|
|
| 18 |
# Import backend utilities
|
| 19 |
from backend.utils.demo_depth import generate_smart_depth
|
| 20 |
from backend.utils.image_processing import (
|
|
|
|
| 21 |
depth_to_colormap,
|
|
|
|
| 22 |
create_side_by_side
|
| 23 |
)
|
| 24 |
|
| 25 |
+
print("[*] DimensioDepth starting in DEMO MODE")
|
| 26 |
+
print("[*] Using synthetic depth estimation (no model downloads needed)")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 27 |
|
| 28 |
|
| 29 |
def estimate_depth(image, quality_mode="Fast (Preview)", colormap_style="Inferno"):
|
| 30 |
"""
|
| 31 |
+
Estimate depth from an input image using DEMO MODE
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 32 |
"""
|
| 33 |
try:
|
| 34 |
# Convert PIL to numpy if needed
|
| 35 |
if isinstance(image, Image.Image):
|
| 36 |
image = np.array(image)
|
| 37 |
|
| 38 |
+
# Generate depth map using DEMO MODE
|
| 39 |
+
depth = generate_smart_depth(image)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 40 |
|
| 41 |
# Convert colormap style to cv2 constant
|
| 42 |
colormap_dict = {
|
|
|
|
| 59 |
|
| 60 |
# Processing info
|
| 61 |
info = f"""
|
| 62 |
+
### β
Depth Estimation Complete!
|
| 63 |
|
| 64 |
+
**Mode**: DEMO MODE (Synthetic Depth)
|
| 65 |
+
**Input Size**: {image.shape[1]}x{image.shape[0]}
|
| 66 |
+
**Output Size**: {depth.shape[1]}x{depth.shape[0]}
|
| 67 |
+
**Colormap**: {colormap_style}
|
| 68 |
+
**Processing**: Ultra-fast (<50ms)
|
| 69 |
|
| 70 |
+
The DEMO MODE uses advanced edge detection + intensity analysis to create surprisingly good depth maps!
|
| 71 |
"""
|
| 72 |
|
| 73 |
return depth_colored, depth_gray, info
|
| 74 |
|
| 75 |
except Exception as e:
|
| 76 |
+
error_msg = f"### β Error\n\n{str(e)}"
|
| 77 |
+
print(f"Error during depth estimation: {e}")
|
| 78 |
return None, None, error_msg
|
| 79 |
|
| 80 |
|
|
|
|
| 85 |
image = np.array(image)
|
| 86 |
|
| 87 |
# Get depth estimation
|
| 88 |
+
depth = generate_smart_depth(image)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 89 |
|
| 90 |
# Convert colormap
|
| 91 |
colormap_dict = {
|
|
|
|
| 110 |
|
| 111 |
|
| 112 |
def create_3d_visualization(image, depth_map, parallax_strength=0.5):
|
| 113 |
+
"""Create a simple 3D displacement visualization"""
|
|
|
|
|
|
|
| 114 |
try:
|
| 115 |
+
if image is None:
|
| 116 |
+
return None
|
| 117 |
+
|
| 118 |
if isinstance(image, Image.Image):
|
| 119 |
image = np.array(image)
|
| 120 |
+
|
| 121 |
+
if depth_map is None:
|
| 122 |
+
# Generate depth if not provided
|
| 123 |
+
depth_map = generate_smart_depth(image)
|
| 124 |
+
depth_map = (depth_map * 255).astype(np.uint8)
|
| 125 |
+
elif isinstance(depth_map, Image.Image):
|
| 126 |
depth_map = np.array(depth_map)
|
| 127 |
|
| 128 |
# Convert depth to grayscale if colored
|
|
|
|
| 150 |
|
| 151 |
except Exception as e:
|
| 152 |
print(f"Error creating 3D viz: {e}")
|
| 153 |
+
return image if image is not None else None
|
| 154 |
|
| 155 |
|
| 156 |
# Create Gradio interface
|
|
|
|
| 162 |
gr.Markdown("""
|
| 163 |
# π¨ DimensioDepth - Add Dimension to Everything
|
| 164 |
|
| 165 |
+
### Transform 2D images into stunning 3D depth visualizations
|
| 166 |
|
| 167 |
+
**Running in DEMO MODE** - Ultra-fast synthetic depth estimation (no AI models needed!)
|
| 168 |
|
| 169 |
---
|
| 170 |
""")
|
|
|
|
| 180 |
height=400
|
| 181 |
)
|
| 182 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 183 |
colormap_style = gr.Dropdown(
|
| 184 |
choices=["Inferno", "Viridis", "Plasma", "Turbo", "Magma", "Hot", "Ocean", "Rainbow"],
|
| 185 |
value="Inferno",
|
|
|
|
| 197 |
|
| 198 |
estimate_btn.click(
|
| 199 |
fn=estimate_depth,
|
| 200 |
+
inputs=[input_image, gr.State("Fast"), colormap_style],
|
| 201 |
outputs=[depth_colored, depth_gray, processing_info]
|
| 202 |
)
|
| 203 |
|
|
|
|
| 212 |
with gr.Column(scale=1):
|
| 213 |
compare_input = gr.Image(label="Upload Image", type="pil", height=400)
|
| 214 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 215 |
compare_colormap = gr.Dropdown(
|
| 216 |
choices=["Inferno", "Viridis", "Plasma", "Turbo", "Magma", "Hot", "Ocean", "Rainbow"],
|
| 217 |
value="Turbo",
|
|
|
|
| 225 |
|
| 226 |
compare_btn.click(
|
| 227 |
fn=create_side_by_side_comparison,
|
| 228 |
+
inputs=[compare_input, gr.State("Fast"), compare_colormap],
|
| 229 |
outputs=comparison_output
|
| 230 |
)
|
| 231 |
|
|
|
|
| 239 |
with gr.Row():
|
| 240 |
with gr.Column(scale=1):
|
| 241 |
parallax_input = gr.Image(label="Original Image", type="pil")
|
| 242 |
+
parallax_depth = gr.Image(label="Depth Map (optional)", type="pil")
|
| 243 |
parallax_strength = gr.Slider(
|
| 244 |
minimum=0, maximum=2, value=0.5, step=0.1,
|
| 245 |
label="Parallax Strength",
|
|
|
|
| 256 |
outputs=parallax_output
|
| 257 |
)
|
| 258 |
|
| 259 |
+
# Info section
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 260 |
gr.Markdown("---")
|
| 261 |
gr.Markdown("""
|
| 262 |
+
## π‘ About This Demo
|
| 263 |
+
|
| 264 |
+
### π¨ Demo Mode Features:
|
| 265 |
+
- β
**Ultra-fast processing** (<50ms per image)
|
| 266 |
+
- β
**No model downloads** required
|
| 267 |
+
- β
**Advanced edge detection** + intensity analysis
|
| 268 |
+
- β
**Surprisingly good quality** for most use cases
|
| 269 |
+
- β
**Perfect for testing** and prototyping
|
| 270 |
+
|
| 271 |
+
### π How It Works:
|
| 272 |
+
Demo Mode uses sophisticated computer vision techniques:
|
| 273 |
+
1. **Edge Detection** - Find object boundaries
|
| 274 |
+
2. **Intensity Analysis** - Analyze brightness patterns
|
| 275 |
+
3. **Gaussian Smoothing** - Create smooth depth transitions
|
| 276 |
+
4. **Normalization** - Convert to depth values
|
| 277 |
+
|
| 278 |
+
### π‘ Tips for Best Results:
|
| 279 |
+
- **Image Quality**: Higher resolution = better depth detail
|
| 280 |
+
- **Lighting**: Well-lit images produce clearer depth maps
|
| 281 |
+
- **Contrast**: Good contrast shows better depth separation
|
| 282 |
+
- **Colormap**: Inferno for general use, Viridis for scientific viz
|
| 283 |
|
| 284 |
+
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 285 |
|
| 286 |
+
### π Use Cases
|
|
|
|
| 287 |
|
| 288 |
+
- π¨ **Creative & Artistic**: Depth-enhanced photos, 3D effects
|
| 289 |
+
- π¬ **VFX & Film**: Depth map generation for compositing
|
| 290 |
+
- π¬ **Research**: Computer vision, depth perception studies
|
| 291 |
+
- π± **Content Creation**: Engaging 3D effects for social media
|
| 292 |
|
|
|
|
| 293 |
---
|
| 294 |
|
| 295 |
+
**Tech Stack**: Advanced CV Algorithms, OpenCV, NumPy, Gradio
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 296 |
|
| 297 |
Made with β€οΈ for the AI community
|
| 298 |
""")
|
|
|
|
| 300 |
|
| 301 |
# Launch the app
|
| 302 |
if __name__ == "__main__":
|
| 303 |
+
demo.launch()
|
|
|
|
|
|
|
|
|
|
|
|
requirements.txt
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
# Gradio and UI
|
| 2 |
-
gradio==4.44.
|
| 3 |
|
| 4 |
# Core ML and image processing
|
| 5 |
onnxruntime-gpu==1.20.1
|
|
|
|
| 1 |
# Gradio and UI
|
| 2 |
+
gradio==4.44.1
|
| 3 |
|
| 4 |
# Core ML and image processing
|
| 5 |
onnxruntime-gpu==1.20.1
|