#!/usr/bin/env python3 """ Training script for FSD-Level5-CoT model on SADC real driving dataset. Dataset: jHaselberger/SADC-Situation-Awareness-for-Driver-Centric-Driving-Style-Adaptation - 100K+ real driving frames with camera images - Speed, acceleration, steering, yaw rate, lane position - Multiple road types: rural, federal, highway Maps SADC columns → FSD model inputs: frame → camera_images (replicated across 6 virtual cameras) v_kmph → ego_state[0] (converted to m/s) ax_mpss → ego_state[1] steering_rack_pos_m → ego_state[2], gt_steering yaw_rate_radps → ego_state[3] d_lanecenter_m → used for waypoint GT lane_curvature_radpm → used for trajectory generation road_type → nav_command mapping """ import os import sys import time import json import math import torch import torch.nn as nn import torch.nn.functional as F from torch.utils.data import Dataset, DataLoader import numpy as np # ── Config ── DATASET_NAME = "jHaselberger/SADC-Situation-Awareness-for-Driver-Centric-Driving-Style-Adaptation" SPLIT = "pretrain_train" VAL_SPLIT = "pretrain_val" HUB_MODEL_ID = "Reality123b/FSD-Level5-CoT" # Model config (smaller for training feasibility) BEV_SIZE = 100 BEV_FEATURE_DIM = 128 PLANNING_D_MODEL = 128 IMG_H, IMG_W = 120, 160 NUM_WAYPOINTS = 20 COT_ACTOR_QUERIES = 32 COT_ROAD_QUERIES = 16 # Training config BATCH_SIZE = 8 LEARNING_RATE = 3e-4 WEIGHT_DECAY = 1e-4 NUM_EPOCHS = 5 GRAD_ACCUM = 4 MAX_GRAD_NORM = 5.0 WARMUP_STEPS = 200 LOG_EVERY = 10 EVAL_EVERY = 500 MAX_TRAIN_SAMPLES = 50000 # cap for reasonable training time MAX_VAL_SAMPLES = 2000 NUM_WORKERS = 4 # Derived EFFECTIVE_BATCH = BATCH_SIZE * GRAD_ACCUM MAX_SPEED_MS = 20.0 * 0.44704 # 20 mph in m/s # ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ # Dataset # ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ ROAD_TYPE_MAP = { "misc": 0, "rural": 1, "federal": 2, "highway": 3, "city": 4, "parking": 5, "intersection": 6, } class SADCDrivingDataset(Dataset): """Wraps the SADC dataset for FSD model training.""" def __init__(self, hf_dataset, max_samples=None, img_size=(IMG_H, IMG_W)): self.ds = hf_dataset self.img_h, self.img_w = img_size if max_samples and len(self.ds) > max_samples: self.ds = self.ds.select(range(max_samples)) def __len__(self): return len(self.ds) def __getitem__(self, idx): row = self.ds[idx] # ── Image processing ── img = row["frame"] if img is None: # Fallback: black image img_tensor = torch.zeros(3, self.img_h, self.img_w) else: from torchvision import transforms transform = transforms.Compose([ transforms.Resize((self.img_h, self.img_w)), transforms.ToTensor(), transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]), ]) try: if hasattr(img, 'convert'): img = img.convert('RGB') img_tensor = transform(img) except Exception: img_tensor = torch.zeros(3, self.img_h, self.img_w) # Replicate single camera to 6 virtual cameras (with augmented transforms) camera_images = img_tensor.unsqueeze(0).expand(6, -1, -1, -1).clone() # Add slight noise to simulate different camera views for i in range(1, 6): camera_images[i] = camera_images[i] + torch.randn_like(camera_images[i]) * 0.01 # ── Ego state ── speed_ms = float(row.get("v_kmph", 0.0)) / 3.6 ax = float(row.get("ax_mpss", 0.0)) steering = float(row.get("steering_rack_pos_m", 0.0)) yaw_rate = float(row.get("yaw_rate_radps", 0.0)) lane_center = float(row.get("d_lanecenter_m", 0.0)) curvature = float(row.get("lane_curvature_radpm", 0.0)) ego_state = torch.tensor([ speed_ms, # speed m/s ax, # longitudinal acceleration steering, # steering position yaw_rate, # yaw rate 0.0, # x position (relative) lane_center, # y position (lane offset) ], dtype=torch.float32) # ── Navigation command ── road_type = str(row.get("road_type", "misc")) nav_cmd = ROAD_TYPE_MAP.get(road_type, 0) # ── Camera intrinsics/extrinsics (synthetic) ── K = torch.zeros(6, 3, 3) K[:, 0, 0] = 200.0 # fx K[:, 1, 1] = 200.0 # fy K[:, 0, 2] = self.img_w / 2 K[:, 1, 2] = self.img_h / 2 K[:, 2, 2] = 1.0 E = torch.eye(4).unsqueeze(0).expand(6, -1, -1).clone() # Offset each camera slightly yaw_offsets = [-45, 45, -135, 135, -90, 90] for i, yaw_deg in enumerate(yaw_offsets): yaw_r = math.radians(yaw_deg) E[i, 0, 0] = math.cos(yaw_r) E[i, 0, 1] = -math.sin(yaw_r) E[i, 1, 0] = math.sin(yaw_r) E[i, 1, 1] = math.cos(yaw_r) # ── Ultrasonic data (simulated from scene context) ── # Use lane_center as proxy for side distance base_dist = max(0.5, abs(lane_center)) us_distances = torch.ones(20, 1) * base_dist us_distances[:7] = torch.clamp(torch.randn(7, 1) * 0.5 + 3.0, 0.3, 5.0) us_distances[7:14] = torch.clamp(torch.randn(7, 1) * 0.5 + 3.5, 0.3, 5.0) us_distances[14:17] = torch.clamp(torch.tensor([[base_dist]] * 3) + torch.randn(3, 1) * 0.2, 0.3, 5.0) us_distances[17:20] = torch.clamp(torch.tensor([[base_dist]] * 3) + torch.randn(3, 1) * 0.2, 0.3, 5.0) us_placements = torch.zeros(20, 6) # Simplified: positions along vehicle perimeter for i in range(7): us_placements[i] = torch.tensor([2.25, (i-3)*0.3, 0.4, (i-3)*10, 0, 0]) for i in range(7): us_placements[7+i] = torch.tensor([-2.25, (i-3)*0.3, 0.4, 180+(i-3)*10, 0, 0]) for i in range(3): us_placements[14+i] = torch.tensor([(1-i)*1.0, 0.9, 0.6, -90, 0, 0]) us_placements[17+i] = torch.tensor([(1-i)*1.0, -0.9, 0.6, 90, 0, 0]) # ── Ground truth targets ── # Steering target (normalize to degrees, roughly) gt_steering = torch.tensor(steering * 20.0) # rack pos → approximate degrees # Throttle/brake from acceleration gt_throttle = torch.tensor(max(0.0, ax / 3.0)).clamp(0, 1) gt_brake = torch.tensor(max(0.0, -ax / 8.0)).clamp(0, 1) # Waypoints: simulate straight-ahead driving with lane-following gt_waypoints = torch.zeros(NUM_WAYPOINTS, 4) for t in range(NUM_WAYPOINTS): dt = (t + 1) * 0.5 # 0.5s intervals # Forward motion gt_waypoints[t, 0] = speed_ms * dt # Lateral: correct toward lane center gt_waypoints[t, 1] = -lane_center * min(1.0, dt / 3.0) # Heading: based on curvature gt_waypoints[t, 2] = curvature * speed_ms * dt # Speed: maintain current (clamped to max) gt_waypoints[t, 3] = min(speed_ms, MAX_SPEED_MS) # Behavior label if abs(steering) > 0.3: if steering > 0: gt_behavior = 1 # turn_left else: gt_behavior = 2 # turn_right elif abs(ax) < 0.1 and speed_ms < 0.5: gt_behavior = 5 # stop else: gt_behavior = 0 # keep_lane # Segmentation (simplified: center = drivable, edges = not) bev = BEV_SIZE gt_seg = torch.zeros(bev, bev, dtype=torch.long) gt_seg[bev//4:3*bev//4, :] = 1 # drivable area # Heatmap (empty — no object annotations in SADC) gt_heatmap = torch.zeros(10, bev, bev) # Occupancy (edges occupied) gt_occ = torch.zeros(1, bev, bev) gt_occ[:, :bev//4, :] = 1.0 gt_occ[:, 3*bev//4:, :] = 1.0 inputs = { "camera_images": camera_images, "camera_intrinsics": K, "camera_extrinsics": E, "ultrasonic_distances": us_distances, "ultrasonic_placements": us_placements, "ego_state": ego_state, "nav_command": torch.tensor(nav_cmd, dtype=torch.long), } targets = { "gt_steering": gt_steering, "gt_throttle": gt_throttle, "gt_brake": gt_brake, "gt_waypoints": gt_waypoints, "gt_behavior": torch.tensor(gt_behavior, dtype=torch.long), "gt_segmentation": gt_seg, "gt_heatmap": gt_heatmap, "gt_occupancy": gt_occ, } return inputs, targets def collate_fn(batch): """Custom collate for dict-based batches.""" inputs_list, targets_list = zip(*batch) inputs = {} for k in inputs_list[0]: inputs[k] = torch.stack([d[k] for d in inputs_list]) targets = {} for k in targets_list[0]: targets[k] = torch.stack([d[k] for d in targets_list]) return inputs, targets # ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ # Training Loop # ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ def main(): print("=" * 70) print(" FSD-Level5-CoT Training on SADC Real Driving Data") print("=" * 70) device = torch.device("cuda" if torch.cuda.is_available() else "cpu") print(f"Device: {device}") if device.type == "cuda": print(f"GPU: {torch.cuda.get_device_name()}") print(f"VRAM: {torch.cuda.get_device_properties(0).total_mem / 1e9:.1f} GB") # ── Init tracking ── try: import trackio trackio.init(project="fsd-level5-cot", name="sadc-training") HAS_TRACKIO = True print("Trackio initialized") except Exception as e: print(f"Trackio not available: {e}") HAS_TRACKIO = False # ── Load dataset ── print(f"\nLoading dataset: {DATASET_NAME}") print(f" Train split: {SPLIT} (max {MAX_TRAIN_SAMPLES} samples)") print(f" Val split: {VAL_SPLIT} (max {MAX_VAL_SAMPLES} samples)") from datasets import load_dataset ds = load_dataset(DATASET_NAME, split=SPLIT, streaming=False) val_ds = load_dataset(DATASET_NAME, split=VAL_SPLIT, streaming=False) print(f" Loaded train: {len(ds)} rows") print(f" Loaded val: {len(val_ds)} rows") train_dataset = SADCDrivingDataset(ds, max_samples=MAX_TRAIN_SAMPLES) val_dataset = SADCDrivingDataset(val_ds, max_samples=MAX_VAL_SAMPLES) train_loader = DataLoader( train_dataset, batch_size=BATCH_SIZE, shuffle=True, num_workers=NUM_WORKERS, collate_fn=collate_fn, pin_memory=True, drop_last=True, ) val_loader = DataLoader( val_dataset, batch_size=BATCH_SIZE, shuffle=False, num_workers=NUM_WORKERS, collate_fn=collate_fn, pin_memory=True, drop_last=True, ) print(f" Train batches/epoch: {len(train_loader)}") print(f" Val batches: {len(val_loader)}") # ── Build model ── print("\nBuilding model...") sys.path.insert(0, os.path.dirname(os.path.abspath(__file__))) from fsd_model.config import VehicleConfig from fsd_model.model import FullSelfDrivingModel, FSDLoss config = VehicleConfig() model = FullSelfDrivingModel( vehicle_config=config, bev_size=BEV_SIZE, bev_resolution=0.5, bev_feature_dim=BEV_FEATURE_DIM, num_object_classes=10, num_seg_classes=7, num_waypoints=NUM_WAYPOINTS, planning_d_model=PLANNING_D_MODEL, future_steps=6, num_forecast_modes=6, forecast_steps=12, num_behaviors=10, enable_cot=True, cot_num_actor_queries=COT_ACTOR_QUERIES, cot_num_road_queries=COT_ROAD_QUERIES, ).to(device) param_counts = model.count_parameters() total_params = param_counts["total"] print(f" Total parameters: {total_params:,}") for k, v in param_counts.items(): if k not in ["total", "total_trainable"]: print(f" {k}: {v:,}") # ── Loss + Optimizer ── loss_fn = FSDLoss( learnable_weights=True, w_detection=0.5, w_segmentation=1.0, w_occupancy=1.0, w_motion=0.5, w_behavior=1.0, w_trajectory=3.0, # prioritize trajectory w_control=3.0, # prioritize control w_safety=2.0, # safety matters most ).to(device) all_params = list(model.parameters()) + list(loss_fn.parameters()) optimizer = torch.optim.AdamW(all_params, lr=LEARNING_RATE, weight_decay=WEIGHT_DECAY) total_steps = len(train_loader) * NUM_EPOCHS // GRAD_ACCUM scheduler = torch.optim.lr_scheduler.OneCycleLR( optimizer, max_lr=LEARNING_RATE, total_steps=total_steps + 10, pct_start=0.1, anneal_strategy='cos', ) # Enable gradient checkpointing if possible if hasattr(model, 'gradient_checkpointing_enable'): model.gradient_checkpointing_enable() # ── Training loop ── print(f"\nStarting training: {NUM_EPOCHS} epochs, effective batch={EFFECTIVE_BATCH}") print(f"Total steps: ~{total_steps}") global_step = 0 best_val_loss = float('inf') t0 = time.time() for epoch in range(NUM_EPOCHS): model.train() epoch_losses = [] optimizer.zero_grad() for batch_idx, (inputs, targets) in enumerate(train_loader): # Move to device inputs = {k: v.to(device, non_blocking=True) for k, v in inputs.items()} targets = {k: v.to(device, non_blocking=True) for k, v in targets.items()} # Forward try: output = model(**inputs) losses = loss_fn(output, targets) loss = losses["total"] / GRAD_ACCUM except RuntimeError as e: if "out of memory" in str(e): torch.cuda.empty_cache() print(f" OOM at batch {batch_idx}, skipping") continue raise # Backward loss.backward() if (batch_idx + 1) % GRAD_ACCUM == 0: torch.nn.utils.clip_grad_norm_(all_params, MAX_GRAD_NORM) optimizer.step() scheduler.step() optimizer.zero_grad() global_step += 1 total_loss_val = losses["total"].item() epoch_losses.append(total_loss_val) # Log if (batch_idx + 1) % LOG_EVERY == 0: elapsed = time.time() - t0 lr = scheduler.get_last_lr()[0] avg_loss = np.mean(epoch_losses[-LOG_EVERY:]) ctrl_loss = losses.get("control", torch.tensor(0.0)).item() traj_loss = losses.get("trajectory", torch.tensor(0.0)).item() seg_loss = losses.get("segmentation", torch.tensor(0.0)).item() safety_loss = losses.get("safety", torch.tensor(0.0)).item() print(f" [E{epoch+1}/{NUM_EPOCHS}][{batch_idx+1}/{len(train_loader)}] " f"loss={avg_loss:.4f} ctrl={ctrl_loss:.4f} traj={traj_loss:.4f} " f"seg={seg_loss:.4f} safety={safety_loss:.4f} " f"lr={lr:.2e} t={elapsed:.0f}s") if HAS_TRACKIO: trackio.log({ "train/loss": avg_loss, "train/control_loss": ctrl_loss, "train/trajectory_loss": traj_loss, "train/segmentation_loss": seg_loss, "train/safety_loss": safety_loss, "train/lr": lr, "train/epoch": epoch + batch_idx / len(train_loader), }) # Eval if global_step > 0 and global_step % EVAL_EVERY == 0: val_loss = evaluate(model, loss_fn, val_loader, device) print(f" ── EVAL step {global_step}: val_loss={val_loss:.4f} " f"(best={best_val_loss:.4f})") if HAS_TRACKIO: trackio.log({"val/loss": val_loss, "val/step": global_step}) if val_loss < best_val_loss: best_val_loss = val_loss save_dir = "/app/best_model" model.save_pretrained(save_dir) print(f" ── Saved best model (val_loss={val_loss:.4f})") model.train() # End of epoch eval val_loss = evaluate(model, loss_fn, val_loader, device) avg_epoch_loss = np.mean(epoch_losses) print(f"\n Epoch {epoch+1}/{NUM_EPOCHS} complete: " f"train_loss={avg_epoch_loss:.4f} val_loss={val_loss:.4f}") if val_loss < best_val_loss: best_val_loss = val_loss model.save_pretrained("/app/best_model") print(f" ── Saved best model (val_loss={val_loss:.4f})") # ── Final save + push to Hub ── total_time = time.time() - t0 print(f"\nTraining complete in {total_time/60:.1f} min") print(f"Best val loss: {best_val_loss:.4f}") model.save_pretrained("/app/final_model") print("\nPushing model to Hub...") try: from huggingface_hub import HfApi api = HfApi() api.upload_folder( folder_path="/app/best_model", repo_id=HUB_MODEL_ID, path_in_repo="trained_model", commit_message=f"Upload trained model (best val_loss={best_val_loss:.4f})", ) print(f" ✓ Pushed to {HUB_MODEL_ID}/trained_model") except Exception as e: print(f" Push failed: {e}") # Save training metadata meta = { "dataset": DATASET_NAME, "split": SPLIT, "num_epochs": NUM_EPOCHS, "best_val_loss": best_val_loss, "total_params": total_params, "training_time_min": total_time / 60, "device": str(device), "batch_size": BATCH_SIZE, "grad_accum": GRAD_ACCUM, "learning_rate": LEARNING_RATE, } with open("/app/training_meta.json", "w") as f: json.dump(meta, f, indent=2) try: api.upload_file( path_or_fileobj="/app/training_meta.json", path_in_repo="trained_model/training_meta.json", repo_id=HUB_MODEL_ID, ) except: pass print("\nDone!") @torch.no_grad() def evaluate(model, loss_fn, val_loader, device, max_batches=50): """Quick validation pass.""" model.eval() val_losses = [] for i, (inputs, targets) in enumerate(val_loader): if i >= max_batches: break inputs = {k: v.to(device, non_blocking=True) for k, v in inputs.items()} targets = {k: v.to(device, non_blocking=True) for k, v in targets.items()} try: output = model(**inputs) losses = loss_fn(output, targets) val_losses.append(losses["total"].item()) except RuntimeError: continue return np.mean(val_losses) if val_losses else float('inf') if __name__ == "__main__": main()