|
|
"""
|
|
|
LaborView AI - MedSigLIP Training Script
|
|
|
Fine-tune MedSigLIP vision encoder for ultrasound segmentation
|
|
|
Self-contained script for HuggingFace Jobs
|
|
|
"""
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import os
|
|
|
import sys
|
|
|
import json
|
|
|
import zipfile
|
|
|
import urllib.request
|
|
|
from pathlib import Path
|
|
|
from dataclasses import dataclass
|
|
|
from typing import Dict, List, Optional, Tuple
|
|
|
from datetime import datetime
|
|
|
|
|
|
import torch
|
|
|
import torch.nn as nn
|
|
|
import torch.nn.functional as F
|
|
|
from torch.amp import GradScaler, autocast
|
|
|
from torch.optim import AdamW
|
|
|
from torch.optim.lr_scheduler import OneCycleLR
|
|
|
from torch.utils.data import Dataset, DataLoader
|
|
|
from PIL import Image
|
|
|
import numpy as np
|
|
|
from tqdm import tqdm
|
|
|
|
|
|
try:
|
|
|
import albumentations as A
|
|
|
from albumentations.pytorch import ToTensorV2
|
|
|
ALBU_AVAILABLE = True
|
|
|
except ImportError:
|
|
|
ALBU_AVAILABLE = False
|
|
|
import torchvision.transforms as T
|
|
|
|
|
|
|
|
|
@dataclass
|
|
|
class Config:
|
|
|
|
|
|
data_url: str = "https://zenodo.org/records/17655183/files/DatasetV3.zip?download=1"
|
|
|
data_dir: Path = Path("./data")
|
|
|
image_size: int = 448
|
|
|
|
|
|
|
|
|
encoder_name: str = "medsiglip"
|
|
|
encoder_pretrained: str = "google/medsiglip-448"
|
|
|
encoder_hidden_dim: int = 1152
|
|
|
projection_dim: int = 256
|
|
|
|
|
|
|
|
|
num_plane_classes: int = 2
|
|
|
num_seg_classes: int = 3
|
|
|
|
|
|
|
|
|
batch_size: int = 4
|
|
|
num_epochs: int = 30
|
|
|
learning_rate: float = 5e-5
|
|
|
weight_decay: float = 0.01
|
|
|
warmup_epochs: int = 2
|
|
|
gradient_accumulation: int = 8
|
|
|
freeze_encoder_epochs: int = 3
|
|
|
use_gradient_checkpointing: bool = True
|
|
|
|
|
|
|
|
|
output_dir: Path = Path("./outputs")
|
|
|
hub_model_id: str = "samwell/laborview-medsiglip"
|
|
|
push_to_hub: bool = True
|
|
|
seed: int = 42
|
|
|
|
|
|
|
|
|
class UltrasoundDataset(Dataset):
|
|
|
"""Dataset for ultrasound segmentation"""
|
|
|
|
|
|
def __init__(self, data_dir: Path, split: str = "train", image_size: int = 448, augment: bool = True):
|
|
|
self.data_dir = Path(data_dir)
|
|
|
self.split = split
|
|
|
self.image_size = image_size
|
|
|
self.samples = self._find_samples()
|
|
|
print(f"Found {len(self.samples)} samples for {split}")
|
|
|
self.transform = self._get_transform(augment and split == "train")
|
|
|
|
|
|
def _find_samples(self) -> List[Dict]:
|
|
|
samples = []
|
|
|
seg_dir = self.data_dir / self.split / "seg"
|
|
|
|
|
|
if not seg_dir.exists():
|
|
|
print(f"Warning: {seg_dir} not found")
|
|
|
return samples
|
|
|
|
|
|
for video_dir in seg_dir.iterdir():
|
|
|
if not video_dir.is_dir():
|
|
|
continue
|
|
|
|
|
|
|
|
|
image_dir = video_dir / "image"
|
|
|
mask_dir = video_dir / "mask"
|
|
|
|
|
|
if mask_dir.exists():
|
|
|
for mask_path in mask_dir.glob("*.png"):
|
|
|
|
|
|
image_path = None
|
|
|
if image_dir.exists():
|
|
|
potential_image = image_dir / mask_path.name
|
|
|
if potential_image.exists():
|
|
|
image_path = str(potential_image)
|
|
|
|
|
|
samples.append({
|
|
|
"mask_path": str(mask_path),
|
|
|
"image_path": image_path,
|
|
|
"video_id": video_dir.name,
|
|
|
})
|
|
|
|
|
|
return samples
|
|
|
|
|
|
def _get_transform(self, augment: bool):
|
|
|
if ALBU_AVAILABLE:
|
|
|
if augment:
|
|
|
return A.Compose([
|
|
|
A.Resize(self.image_size, self.image_size),
|
|
|
A.HorizontalFlip(p=0.5),
|
|
|
A.RandomBrightnessContrast(p=0.3),
|
|
|
A.GaussNoise(var_limit=(10, 50), p=0.2),
|
|
|
A.ShiftScaleRotate(shift_limit=0.1, scale_limit=0.1, rotate_limit=10, p=0.3),
|
|
|
A.Normalize(mean=[0.5, 0.5, 0.5], std=[0.5, 0.5, 0.5]),
|
|
|
ToTensorV2()
|
|
|
])
|
|
|
else:
|
|
|
return A.Compose([
|
|
|
A.Resize(self.image_size, self.image_size),
|
|
|
A.Normalize(mean=[0.5, 0.5, 0.5], std=[0.5, 0.5, 0.5]),
|
|
|
ToTensorV2()
|
|
|
])
|
|
|
else:
|
|
|
return T.Compose([
|
|
|
T.Resize((self.image_size, self.image_size)),
|
|
|
T.ToTensor(),
|
|
|
T.Normalize(mean=[0.5, 0.5, 0.5], std=[0.5, 0.5, 0.5])
|
|
|
])
|
|
|
|
|
|
def __len__(self):
|
|
|
return len(self.samples)
|
|
|
|
|
|
def __getitem__(self, idx):
|
|
|
sample = self.samples[idx]
|
|
|
|
|
|
|
|
|
mask = Image.open(sample["mask_path"]).convert("L")
|
|
|
mask = np.array(mask)
|
|
|
|
|
|
|
|
|
if sample["image_path"] and os.path.exists(sample["image_path"]):
|
|
|
image = Image.open(sample["image_path"]).convert("RGB")
|
|
|
image = np.array(image)
|
|
|
else:
|
|
|
|
|
|
image = np.stack([mask, mask, mask], axis=-1)
|
|
|
|
|
|
|
|
|
|
|
|
mask_classes = np.zeros_like(mask, dtype=np.int64)
|
|
|
mask_classes[mask > 0] = 1
|
|
|
mask_classes[mask > 127] = 2
|
|
|
|
|
|
if ALBU_AVAILABLE:
|
|
|
transformed = self.transform(image=image, mask=mask_classes)
|
|
|
image, mask = transformed["image"], transformed["mask"]
|
|
|
else:
|
|
|
image = self.transform(Image.fromarray(image))
|
|
|
mask = torch.from_numpy(
|
|
|
np.array(Image.fromarray(mask_classes.astype(np.uint8)).resize(
|
|
|
(self.image_size, self.image_size), Image.NEAREST
|
|
|
))
|
|
|
).long()
|
|
|
|
|
|
return {
|
|
|
"pixel_values": image,
|
|
|
"seg_labels": mask,
|
|
|
"plane_labels": torch.tensor(1, dtype=torch.long)
|
|
|
}
|
|
|
|
|
|
|
|
|
class SegmentationDecoder(nn.Module):
|
|
|
"""Decoder for upsampling vision features to segmentation mask"""
|
|
|
|
|
|
def __init__(self, input_dim: int, num_classes: int, decoder_channels=[512, 256, 128, 64]):
|
|
|
super().__init__()
|
|
|
|
|
|
self.input_proj = nn.Conv2d(input_dim, decoder_channels[0], 1)
|
|
|
|
|
|
self.up_blocks = nn.ModuleList()
|
|
|
in_ch = decoder_channels[0]
|
|
|
for out_ch in decoder_channels[1:]:
|
|
|
self.up_blocks.append(nn.Sequential(
|
|
|
nn.ConvTranspose2d(in_ch, out_ch, 4, stride=2, padding=1),
|
|
|
nn.BatchNorm2d(out_ch),
|
|
|
nn.GELU()
|
|
|
))
|
|
|
in_ch = out_ch
|
|
|
|
|
|
|
|
|
self.final_up = nn.Sequential(
|
|
|
nn.ConvTranspose2d(decoder_channels[-1], 32, 4, stride=2, padding=1),
|
|
|
nn.BatchNorm2d(32),
|
|
|
nn.GELU(),
|
|
|
nn.ConvTranspose2d(32, 32, 4, stride=2, padding=1),
|
|
|
nn.BatchNorm2d(32),
|
|
|
nn.GELU(),
|
|
|
)
|
|
|
|
|
|
self.classifier = nn.Conv2d(32, num_classes, 1)
|
|
|
|
|
|
def forward(self, x, target_size=None):
|
|
|
B = x.shape[0]
|
|
|
|
|
|
|
|
|
if x.dim() == 3:
|
|
|
|
|
|
num_patches = x.shape[1]
|
|
|
H = W = int(num_patches ** 0.5)
|
|
|
x = x.transpose(1, 2).reshape(B, -1, H, W)
|
|
|
|
|
|
x = self.input_proj(x)
|
|
|
|
|
|
for block in self.up_blocks:
|
|
|
x = block(x)
|
|
|
|
|
|
x = self.final_up(x)
|
|
|
x = self.classifier(x)
|
|
|
|
|
|
if target_size:
|
|
|
x = F.interpolate(x, size=target_size, mode='bilinear', align_corners=False)
|
|
|
|
|
|
return x
|
|
|
|
|
|
|
|
|
class LaborViewMedSigLIP(nn.Module):
|
|
|
"""LaborView model with MedSigLIP vision encoder"""
|
|
|
|
|
|
def __init__(self, config: Config):
|
|
|
super().__init__()
|
|
|
self.config = config
|
|
|
|
|
|
|
|
|
print(f"Loading MedSigLIP from {config.encoder_pretrained}...")
|
|
|
from transformers import AutoModel
|
|
|
|
|
|
self.encoder = AutoModel.from_pretrained(
|
|
|
config.encoder_pretrained,
|
|
|
trust_remote_code=True
|
|
|
)
|
|
|
|
|
|
|
|
|
if hasattr(self.encoder, 'vision_model'):
|
|
|
self.vision_encoder = self.encoder.vision_model
|
|
|
else:
|
|
|
self.vision_encoder = self.encoder
|
|
|
|
|
|
|
|
|
if hasattr(self.vision_encoder.config, 'hidden_size'):
|
|
|
hidden_dim = self.vision_encoder.config.hidden_size
|
|
|
else:
|
|
|
hidden_dim = config.encoder_hidden_dim
|
|
|
|
|
|
print(f"Vision encoder hidden dim: {hidden_dim}")
|
|
|
|
|
|
|
|
|
self.projector = nn.Sequential(
|
|
|
nn.Linear(hidden_dim, config.projection_dim),
|
|
|
nn.LayerNorm(config.projection_dim),
|
|
|
nn.GELU(),
|
|
|
nn.Linear(config.projection_dim, config.projection_dim)
|
|
|
)
|
|
|
|
|
|
|
|
|
self.cls_head = nn.Linear(config.projection_dim, config.num_plane_classes)
|
|
|
|
|
|
|
|
|
self.seg_decoder = SegmentationDecoder(hidden_dim, config.num_seg_classes)
|
|
|
|
|
|
def forward(self, pixel_values):
|
|
|
|
|
|
if hasattr(self, 'vision_encoder'):
|
|
|
outputs = self.vision_encoder(pixel_values)
|
|
|
else:
|
|
|
outputs = self.encoder.get_image_features(pixel_values, return_dict=True)
|
|
|
|
|
|
|
|
|
if hasattr(outputs, 'last_hidden_state'):
|
|
|
hidden = outputs.last_hidden_state
|
|
|
elif hasattr(outputs, 'pooler_output'):
|
|
|
hidden = outputs.pooler_output
|
|
|
else:
|
|
|
hidden = outputs
|
|
|
|
|
|
|
|
|
if hidden.dim() == 2:
|
|
|
|
|
|
pooled = hidden
|
|
|
|
|
|
B, D = hidden.shape
|
|
|
seq = hidden.unsqueeze(1).expand(B, 32*32, D)
|
|
|
elif hidden.dim() == 3:
|
|
|
|
|
|
pooled = hidden.mean(dim=1)
|
|
|
seq = hidden
|
|
|
else:
|
|
|
|
|
|
B, D, H, W = hidden.shape
|
|
|
pooled = hidden.mean(dim=[2, 3])
|
|
|
seq = hidden.flatten(2).transpose(1, 2)
|
|
|
|
|
|
|
|
|
projected = self.projector(pooled)
|
|
|
plane_logits = self.cls_head(projected)
|
|
|
|
|
|
|
|
|
seg_masks = self.seg_decoder(seq, target_size=pixel_values.shape[-2:])
|
|
|
|
|
|
return plane_logits, seg_masks
|
|
|
|
|
|
def compute_loss(self, plane_logits, seg_masks, plane_labels, seg_labels):
|
|
|
losses = {}
|
|
|
|
|
|
|
|
|
if plane_labels is not None:
|
|
|
losses["cls"] = F.cross_entropy(plane_logits, plane_labels)
|
|
|
|
|
|
|
|
|
if seg_labels is not None:
|
|
|
|
|
|
ce_loss = F.cross_entropy(seg_masks, seg_labels.long())
|
|
|
|
|
|
|
|
|
seg_probs = F.softmax(seg_masks, dim=1)
|
|
|
target_oh = F.one_hot(seg_labels.long(), self.config.num_seg_classes).permute(0, 3, 1, 2).float()
|
|
|
|
|
|
intersection = (seg_probs * target_oh).sum(dim=(2, 3))
|
|
|
union = seg_probs.sum(dim=(2, 3)) + target_oh.sum(dim=(2, 3))
|
|
|
dice_loss = 1 - ((2 * intersection + 1e-6) / (union + 1e-6)).mean()
|
|
|
|
|
|
losses["seg"] = dice_loss + ce_loss
|
|
|
|
|
|
return sum(losses.values()), losses
|
|
|
|
|
|
def freeze_encoder(self):
|
|
|
"""Freeze the vision encoder"""
|
|
|
for param in self.vision_encoder.parameters():
|
|
|
param.requires_grad = False
|
|
|
print("Encoder frozen")
|
|
|
|
|
|
def unfreeze_encoder(self, use_gradient_checkpointing=True):
|
|
|
"""Unfreeze the vision encoder with optional gradient checkpointing"""
|
|
|
for param in self.vision_encoder.parameters():
|
|
|
param.requires_grad = True
|
|
|
|
|
|
|
|
|
if use_gradient_checkpointing:
|
|
|
if hasattr(self.vision_encoder, 'gradient_checkpointing_enable'):
|
|
|
self.vision_encoder.gradient_checkpointing_enable()
|
|
|
print("Encoder unfrozen with gradient checkpointing")
|
|
|
else:
|
|
|
print("Encoder unfrozen (gradient checkpointing not available)")
|
|
|
else:
|
|
|
print("Encoder unfrozen")
|
|
|
|
|
|
|
|
|
def train_epoch(model, loader, optimizer, scheduler, scaler, device, config, epoch):
|
|
|
model.train()
|
|
|
total_loss, num_batches = 0, 0
|
|
|
|
|
|
pbar = tqdm(loader, desc=f"Epoch {epoch+1} Training")
|
|
|
for batch_idx, batch in enumerate(pbar):
|
|
|
pixel_values = batch["pixel_values"].to(device)
|
|
|
seg_labels = batch["seg_labels"].to(device)
|
|
|
plane_labels = batch["plane_labels"].to(device)
|
|
|
|
|
|
with autocast("cuda", enabled=True):
|
|
|
plane_logits, seg_masks = model(pixel_values)
|
|
|
loss, _ = model.compute_loss(plane_logits, seg_masks, plane_labels, seg_labels)
|
|
|
|
|
|
loss = loss / config.gradient_accumulation
|
|
|
scaler.scale(loss).backward()
|
|
|
|
|
|
if (batch_idx + 1) % config.gradient_accumulation == 0:
|
|
|
scaler.step(optimizer)
|
|
|
scaler.update()
|
|
|
optimizer.zero_grad()
|
|
|
scheduler.step()
|
|
|
|
|
|
total_loss += loss.item() * config.gradient_accumulation
|
|
|
num_batches += 1
|
|
|
pbar.set_postfix({"loss": f"{loss.item() * config.gradient_accumulation:.4f}"})
|
|
|
|
|
|
return total_loss / num_batches
|
|
|
|
|
|
|
|
|
@torch.no_grad()
|
|
|
def validate(model, loader, device):
|
|
|
model.eval()
|
|
|
total_loss, total_iou, num_batches = 0, 0, 0
|
|
|
|
|
|
for batch in tqdm(loader, desc="Validating"):
|
|
|
pixel_values = batch["pixel_values"].to(device)
|
|
|
seg_labels = batch["seg_labels"].to(device)
|
|
|
plane_labels = batch["plane_labels"].to(device)
|
|
|
|
|
|
plane_logits, seg_masks = model(pixel_values)
|
|
|
loss, _ = model.compute_loss(plane_logits, seg_masks, plane_labels, seg_labels)
|
|
|
|
|
|
|
|
|
seg_preds = seg_masks.argmax(dim=1)
|
|
|
intersection = ((seg_preds == 1) & (seg_labels == 1)).sum().item()
|
|
|
union = ((seg_preds == 1) | (seg_labels == 1)).sum().item()
|
|
|
|
|
|
total_loss += loss.item()
|
|
|
total_iou += intersection / (union + 1e-6)
|
|
|
num_batches += 1
|
|
|
|
|
|
return total_loss / num_batches, total_iou / num_batches
|
|
|
|
|
|
|
|
|
def download_with_retry(url, dest_path, max_retries=3):
|
|
|
"""Download file with retry logic using subprocess for robustness"""
|
|
|
import subprocess
|
|
|
import shutil
|
|
|
|
|
|
|
|
|
if shutil.which("wget"):
|
|
|
for attempt in range(max_retries):
|
|
|
try:
|
|
|
print(f"Download attempt {attempt + 1}/{max_retries} with wget...")
|
|
|
result = subprocess.run(
|
|
|
["wget", "-c", "-O", str(dest_path), url],
|
|
|
check=True, capture_output=True, text=True
|
|
|
)
|
|
|
if dest_path.exists() and dest_path.stat().st_size > 0:
|
|
|
return True
|
|
|
except subprocess.CalledProcessError as e:
|
|
|
print(f"wget failed: {e}")
|
|
|
if attempt < max_retries - 1:
|
|
|
import time
|
|
|
time.sleep(5)
|
|
|
|
|
|
|
|
|
if shutil.which("curl"):
|
|
|
for attempt in range(max_retries):
|
|
|
try:
|
|
|
print(f"Download attempt {attempt + 1}/{max_retries} with curl...")
|
|
|
result = subprocess.run(
|
|
|
["curl", "-L", "-C", "-", "-o", str(dest_path), url],
|
|
|
check=True, capture_output=True, text=True
|
|
|
)
|
|
|
if dest_path.exists() and dest_path.stat().st_size > 0:
|
|
|
return True
|
|
|
except subprocess.CalledProcessError as e:
|
|
|
print(f"curl failed: {e}")
|
|
|
if attempt < max_retries - 1:
|
|
|
import time
|
|
|
time.sleep(5)
|
|
|
|
|
|
|
|
|
print("Falling back to urllib chunked download...")
|
|
|
import urllib.request
|
|
|
for attempt in range(max_retries):
|
|
|
try:
|
|
|
with urllib.request.urlopen(url, timeout=300) as response:
|
|
|
total_size = int(response.headers.get('content-length', 0))
|
|
|
downloaded = 0
|
|
|
chunk_size = 8192 * 16
|
|
|
with open(dest_path, 'wb') as f:
|
|
|
while True:
|
|
|
chunk = response.read(chunk_size)
|
|
|
if not chunk:
|
|
|
break
|
|
|
f.write(chunk)
|
|
|
downloaded += len(chunk)
|
|
|
if total_size > 0:
|
|
|
pct = (downloaded / total_size) * 100
|
|
|
print(f"\rDownloaded {downloaded / 1e6:.1f}/{total_size / 1e6:.1f} MB ({pct:.1f}%)", end="", flush=True)
|
|
|
print()
|
|
|
return True
|
|
|
except Exception as e:
|
|
|
print(f"urllib attempt {attempt + 1} failed: {e}")
|
|
|
if attempt < max_retries - 1:
|
|
|
import time
|
|
|
time.sleep(5)
|
|
|
|
|
|
raise Exception(f"Failed to download {url} after {max_retries} attempts")
|
|
|
|
|
|
|
|
|
def download_dataset(config):
|
|
|
"""Download and extract dataset"""
|
|
|
config.data_dir.mkdir(parents=True, exist_ok=True)
|
|
|
zip_path = config.data_dir / "dataset.zip"
|
|
|
|
|
|
if not (config.data_dir / "train").exists():
|
|
|
print(f"Downloading dataset from {config.data_url}...")
|
|
|
download_with_retry(config.data_url, zip_path)
|
|
|
|
|
|
print("Extracting...")
|
|
|
with zipfile.ZipFile(zip_path, 'r') as z:
|
|
|
z.extractall(config.data_dir)
|
|
|
|
|
|
|
|
|
inner_zip = config.data_dir / "DatasetV3.zip"
|
|
|
if inner_zip.exists():
|
|
|
with zipfile.ZipFile(inner_zip, 'r') as z:
|
|
|
z.extractall(config.data_dir)
|
|
|
|
|
|
|
|
|
dataset_dir = config.data_dir / "DatasetV3"
|
|
|
if dataset_dir.exists():
|
|
|
for split in ["train", "val", "test"]:
|
|
|
for sz in dataset_dir.glob(f"{split}*.zip"):
|
|
|
print(f"Extracting {sz.name}...")
|
|
|
with zipfile.ZipFile(sz, 'r') as z:
|
|
|
z.extractall(dataset_dir)
|
|
|
|
|
|
|
|
|
zip_path.unlink(missing_ok=True)
|
|
|
inner_zip.unlink(missing_ok=True)
|
|
|
|
|
|
|
|
|
dataset_v3 = config.data_dir / "DatasetV3"
|
|
|
if dataset_v3.exists():
|
|
|
return dataset_v3
|
|
|
return config.data_dir
|
|
|
|
|
|
|
|
|
def main():
|
|
|
config = Config()
|
|
|
|
|
|
|
|
|
torch.manual_seed(config.seed)
|
|
|
np.random.seed(config.seed)
|
|
|
|
|
|
|
|
|
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
|
|
print(f"Device: {device}")
|
|
|
if device.type == "cuda":
|
|
|
print(f"GPU: {torch.cuda.get_device_name(0)}")
|
|
|
print(f"Memory: {torch.cuda.get_device_properties(0).total_memory / 1e9:.1f} GB")
|
|
|
|
|
|
|
|
|
data_dir = download_dataset(config)
|
|
|
print(f"Data directory: {data_dir}")
|
|
|
|
|
|
|
|
|
train_dataset = UltrasoundDataset(data_dir, "train", config.image_size, augment=True)
|
|
|
val_dataset = UltrasoundDataset(data_dir, "val", config.image_size, augment=False)
|
|
|
|
|
|
if len(val_dataset) == 0:
|
|
|
print("No validation data, using 10% of train")
|
|
|
train_size = int(0.9 * len(train_dataset))
|
|
|
train_dataset, val_dataset = torch.utils.data.random_split(
|
|
|
train_dataset, [train_size, len(train_dataset) - train_size]
|
|
|
)
|
|
|
|
|
|
train_loader = DataLoader(
|
|
|
train_dataset, batch_size=config.batch_size, shuffle=True,
|
|
|
num_workers=4, pin_memory=True, drop_last=True
|
|
|
)
|
|
|
val_loader = DataLoader(
|
|
|
val_dataset, batch_size=config.batch_size, shuffle=False,
|
|
|
num_workers=4, pin_memory=True
|
|
|
)
|
|
|
|
|
|
print(f"Train: {len(train_dataset)}, Val: {len(val_dataset)}")
|
|
|
|
|
|
|
|
|
print(f"Creating model with {config.encoder_name} encoder...")
|
|
|
model = LaborViewMedSigLIP(config).to(device)
|
|
|
|
|
|
|
|
|
total_params = sum(p.numel() for p in model.parameters())
|
|
|
trainable_params = sum(p.numel() for p in model.parameters() if p.requires_grad)
|
|
|
print(f"Total parameters: {total_params:,}")
|
|
|
print(f"Trainable parameters: {trainable_params:,}")
|
|
|
|
|
|
|
|
|
model.freeze_encoder()
|
|
|
|
|
|
|
|
|
optimizer = AdamW(
|
|
|
filter(lambda p: p.requires_grad, model.parameters()),
|
|
|
lr=config.learning_rate,
|
|
|
weight_decay=config.weight_decay
|
|
|
)
|
|
|
|
|
|
|
|
|
total_steps = len(train_loader) * config.num_epochs
|
|
|
scheduler = OneCycleLR(
|
|
|
optimizer,
|
|
|
max_lr=config.learning_rate,
|
|
|
total_steps=total_steps,
|
|
|
pct_start=config.warmup_epochs / config.num_epochs
|
|
|
)
|
|
|
|
|
|
|
|
|
scaler = GradScaler("cuda")
|
|
|
|
|
|
|
|
|
config.output_dir.mkdir(parents=True, exist_ok=True)
|
|
|
|
|
|
|
|
|
best_val_loss = float("inf")
|
|
|
print("Starting training")
|
|
|
|
|
|
for epoch in range(config.num_epochs):
|
|
|
|
|
|
if epoch == config.freeze_encoder_epochs:
|
|
|
|
|
|
torch.cuda.empty_cache()
|
|
|
model.unfreeze_encoder(use_gradient_checkpointing=config.use_gradient_checkpointing)
|
|
|
|
|
|
optimizer = AdamW(
|
|
|
model.parameters(),
|
|
|
lr=config.learning_rate * 0.1,
|
|
|
weight_decay=config.weight_decay
|
|
|
)
|
|
|
scheduler = OneCycleLR(
|
|
|
optimizer,
|
|
|
max_lr=config.learning_rate * 0.1,
|
|
|
total_steps=len(train_loader) * (config.num_epochs - epoch),
|
|
|
pct_start=0.1
|
|
|
)
|
|
|
|
|
|
train_loss = train_epoch(model, train_loader, optimizer, scheduler, scaler, device, config, epoch)
|
|
|
val_loss, val_iou = validate(model, val_loader, device)
|
|
|
|
|
|
print(f"Epoch {epoch+1}/{config.num_epochs}")
|
|
|
print(f" Train: {train_loss:.4f}, Val: {val_loss:.4f}, IoU: {val_iou:.4f}")
|
|
|
|
|
|
if val_loss < best_val_loss:
|
|
|
best_val_loss = val_loss
|
|
|
torch.save({
|
|
|
"epoch": epoch,
|
|
|
"model_state_dict": model.state_dict(),
|
|
|
"val_loss": val_loss,
|
|
|
"val_iou": val_iou,
|
|
|
"config": vars(config)
|
|
|
}, config.output_dir / "best.pt")
|
|
|
print(" >>> New best!")
|
|
|
|
|
|
|
|
|
torch.save({
|
|
|
"model_state_dict": model.state_dict(),
|
|
|
"config": vars(config)
|
|
|
}, config.output_dir / "final.pt")
|
|
|
|
|
|
|
|
|
if config.push_to_hub:
|
|
|
try:
|
|
|
from huggingface_hub import HfApi, create_repo
|
|
|
print(f"Pushing to Hub: {config.hub_model_id}")
|
|
|
create_repo(config.hub_model_id, exist_ok=True)
|
|
|
HfApi().upload_folder(
|
|
|
folder_path=str(config.output_dir),
|
|
|
repo_id=config.hub_model_id,
|
|
|
commit_message=f"LaborView MedSigLIP v1 - IoU: {val_iou:.4f}"
|
|
|
)
|
|
|
print(f"Uploaded to https://huggingface.co/{config.hub_model_id}")
|
|
|
except Exception as e:
|
|
|
print(f"Hub upload failed: {e}")
|
|
|
|
|
|
print(f"Training complete! Best Val Loss: {best_val_loss:.4f}")
|
|
|
|
|
|
|
|
|
if __name__ == "__main__":
|
|
|
main()
|
|
|
|