rooftop_segmentation / model_and_lightning_module.py
hmaissoro's picture
add RoofSegmentationDataset
896f322 verified
"""
Modules for roof segmentation.
"""
from pathlib import Path
from typing import Any, Callable, Dict, Optional, Tuple
import albumentations as A
import cv2
import numpy as np
import pytorch_lightning as pl
import torch
import torch.nn as nn
import torch.nn.functional as F
from albumentations.pytorch import ToTensorV2
from torch.utils.data import Dataset
class DoubleConv(nn.Module):
"""Double convolution block: (conv => BN => ReLU) * 2"""
def __init__(self, in_channels, out_channels):
super().__init__()
self.double_conv = nn.Sequential(
nn.Conv2d(in_channels, out_channels, kernel_size=3, padding=1),
nn.BatchNorm2d(out_channels),
nn.ReLU(inplace=True),
nn.Conv2d(out_channels, out_channels, kernel_size=3, padding=1),
nn.BatchNorm2d(out_channels),
nn.ReLU(inplace=True)
)
def forward(self, x):
return self.double_conv(x)
class Down(nn.Module):
"""Downscaling with maxpool then double conv"""
def __init__(self, in_channels, out_channels):
super().__init__()
self.maxpool_conv = nn.Sequential(
nn.MaxPool2d(2),
DoubleConv(in_channels, out_channels)
)
def forward(self, x):
return self.maxpool_conv(x)
class Up(nn.Module):
"""Upscaling then double conv"""
def __init__(self, in_channels, out_channels, bilinear=True):
super().__init__()
# Use bilinear upsampling or transpose convolution
if bilinear:
self.up = nn.Upsample(scale_factor=2, mode='bilinear', align_corners=True)
self.conv = DoubleConv(in_channels, out_channels)
else:
self.up = nn.ConvTranspose2d(in_channels, in_channels // 2, kernel_size=2, stride=2)
self.conv = DoubleConv(in_channels, out_channels)
def forward(self, x1, x2):
x1 = self.up(x1)
# Input is CHW
diffY = x2.size()[2] - x1.size()[2]
diffX = x2.size()[3] - x1.size()[3]
x1 = F.pad(x1, [diffX // 2, diffX - diffX // 2,
diffY // 2, diffY - diffY // 2])
# Concatenate along channel dimension
x = torch.cat([x2, x1], dim=1)
return self.conv(x)
class OutConv(nn.Module):
"""Output convolution"""
def __init__(self, in_channels, out_channels):
super().__init__()
self.conv = nn.Conv2d(in_channels, out_channels, kernel_size=1)
def forward(self, x):
return self.conv(x)
class UNet(nn.Module):
"""Simple U-Net implementation with configurable base channels"""
def __init__(self, n_channels=3, n_classes=1, base_channels=32, bilinear=True):
super().__init__()
self.n_channels = n_channels
self.n_classes = n_classes
self.bilinear = bilinear
# Use configurable base channels (default 32 instead of 64)
c = base_channels
# Encoder (downsampling path)
self.inc = DoubleConv(n_channels, c)
self.down1 = Down(c, c*2)
self.down2 = Down(c*2, c*4)
self.down3 = Down(c*4, c*8)
factor = 2 if bilinear else 1
self.down4 = Down(c*8, c*16 // factor)
# Decoder (upsampling path)
self.up1 = Up(c*16, c*8 // factor, bilinear)
self.up2 = Up(c*8, c*4 // factor, bilinear)
self.up3 = Up(c*4, c*2 // factor, bilinear)
self.up4 = Up(c*2, c, bilinear)
self.outc = OutConv(c, n_classes)
def forward(self, x):
# Encoder
x1 = self.inc(x)
x2 = self.down1(x1)
x3 = self.down2(x2)
x4 = self.down3(x3)
x5 = self.down4(x4)
# Decoder with skip connections
x = self.up1(x5, x4)
x = self.up2(x, x3)
x = self.up3(x, x2)
x = self.up4(x, x1)
logits = self.outc(x)
return logits
def get_unet_model(n_channels=3, n_classes=1, base_channels=32, bilinear=True):
"""
Create a U-Net model.
Args:
n_channels: Number of input channels
n_classes: Number of output classes
base_channels: Base number of channels (32 = lighter, 64 = standard)
bilinear: Use bilinear upsampling
"""
return UNet(n_channels=n_channels, n_classes=n_classes, base_channels=base_channels, bilinear=bilinear)
class SegmentationLightningModule(pl.LightningModule):
def __init__(self, config: Dict[str, Any]):
super().__init__()
self.model = get_unet_model(
n_channels=config["in_channels"],
n_classes=config["classes"],
base_channels=config.get("base_channels", 32),
bilinear=config.get("bilinear", True)
)
def forward(self, x):
return self.model(x)
class RoofSegmentationDataset(Dataset):
"""Dataset for roof segmentation with images and masks."""
def __init__(
self,
images_dir: Path,
masks_dir: Path,
transform: Optional[Callable] = None,
image_size: Tuple[int, int] = (512, 512)
):
"""
Args:
images_dir: Directory containing input images
masks_dir: Directory containing segmentation masks
transform: Albumentations transforms to apply
image_size: Target size for images (height, width)
"""
self.images_dir = Path(images_dir)
self.masks_dir = Path(masks_dir)
self.image_size = image_size
self.transform = transform
# Get all image files
self.image_files = []
for ext in ['.jpg', '.jpeg', '.png', '.tiff', '.tif']:
self.image_files.extend(self.images_dir.glob(f'*{ext}'))
self.image_files.extend(self.images_dir.glob(f'*{ext.upper()}'))
self.image_files = sorted(self.image_files)
# Verify that corresponding masks exist
self.valid_pairs = []
for image_path in self.image_files:
mask_candidates = []
for ext in ['.jpg', '.jpeg', '.png', '.tiff', '.tif']:
mask_path = self.masks_dir / f"{image_path.stem}{ext}"
if mask_path.exists():
mask_candidates.append(mask_path)
if mask_candidates:
self.valid_pairs.append((image_path, mask_candidates[0]))
print(f"Dataset initialized with {len(self.valid_pairs)} image-mask pairs")
def __len__(self) -> int:
return len(self.valid_pairs)
def __getitem__(self, idx: int) -> dict:
image_path, mask_path = self.valid_pairs[idx]
# Load image
image = cv2.imread(str(image_path))
image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)
# Load mask
mask = cv2.imread(str(mask_path), cv2.IMREAD_GRAYSCALE)
# Resize to target size
image = cv2.resize(image, self.image_size)
mask = cv2.resize(mask, self.image_size, interpolation=cv2.INTER_NEAREST)
# Normalize mask to 0-1 (assuming binary segmentation)
mask = (mask > 127).astype(np.uint8)
# Apply transforms
if self.transform:
transformed = self.transform(image=image, mask=mask)
image = transformed['image']
mask = transformed['mask']
# Ensure mask is float for loss calculations
if isinstance(mask, torch.Tensor):
mask = mask.float()
else:
# Convert to tensors manually if no transforms
image = torch.from_numpy(image.transpose(2, 0, 1)).float()
mask = torch.from_numpy(mask).float()
return {
'image': image,
'mask': mask,
'image_path': str(image_path),
'mask_path': str(mask_path)
}
def get_training_transforms(image_size: Tuple[int, int] = (512, 512)) -> A.Compose:
"""Get augmentation transforms for training."""
return A.Compose([
A.HorizontalFlip(p=0.5),
A.VerticalFlip(p=0.5),
A.RandomRotate90(p=0.5),
A.ShiftScaleRotate(
shift_limit=0.1,
scale_limit=0.2,
rotate_limit=45,
border_mode=cv2.BORDER_CONSTANT,
value=0,
p=0.5
),
A.OneOf([
A.RandomBrightnessContrast(brightness_limit=0.2, contrast_limit=0.2, p=1.0),
A.HueSaturationValue(hue_shift_limit=20, sat_shift_limit=30, val_shift_limit=20, p=1.0),
], p=0.5),
A.OneOf([
A.GaussianBlur(blur_limit=(3, 7), p=1.0),
A.MedianBlur(blur_limit=5, p=1.0),
], p=0.3),
A.Resize(image_size[0], image_size[1]),
A.Normalize(
mean=[0.485, 0.456, 0.406],
std=[0.229, 0.224, 0.225]
),
ToTensorV2()
])
def get_validation_transforms(image_size: Tuple[int, int] = (512, 512)) -> A.Compose:
"""Get transforms for validation (no augmentation)."""
return A.Compose([
A.Resize(image_size[0], image_size[1]),
A.Normalize(
mean=[0.485, 0.456, 0.406],
std=[0.229, 0.224, 0.225]
),
ToTensorV2()
])
def create_dataloaders(
train_images_dir: Path,
train_masks_dir: Path,
val_images_dir: Path,
val_masks_dir: Path,
batch_size: int = 8,
num_workers: int = 4,
image_size: Tuple[int, int] = (512, 512)
) -> Tuple[torch.utils.data.DataLoader, torch.utils.data.DataLoader]:
"""Create training and validation dataloaders."""
# Create datasets
train_dataset = RoofSegmentationDataset(
images_dir=train_images_dir,
masks_dir=train_masks_dir,
transform=get_training_transforms(image_size),
image_size=image_size
)
val_dataset = RoofSegmentationDataset(
images_dir=val_images_dir,
masks_dir=val_masks_dir,
transform=get_validation_transforms(image_size),
image_size=image_size
)
# Create dataloaders
train_loader = torch.utils.data.DataLoader(
train_dataset,
batch_size=batch_size,
shuffle=True,
num_workers=num_workers,
pin_memory=True,
drop_last=True
)
val_loader = torch.utils.data.DataLoader(
val_dataset,
batch_size=batch_size,
shuffle=False,
num_workers=num_workers,
pin_memory=True
)
return train_loader, val_loader