File size: 41,277 Bytes
fe39cc9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
"""
Flexible Training Pipeline for Multi-Task Comorbidity Detection
Uses the flexible biomarker configuration system for any task structure
"""

import os
import random
import numpy as np
import torch
import torch.optim as optim
import torch.nn as nn
import torch.nn.functional as F
from torch.utils.data import DataLoader, WeightedRandomSampler
from torch.utils.tensorboard import SummaryWriter
from torchvision import transforms
from sklearn.utils.class_weight import compute_class_weight
from tqdm import tqdm
import json
import time
import logging
import sys
from argparse import ArgumentParser
from typing import Dict, List, Tuple, Any

# Import our custom modules
from dataset import ClassifierDataset
from model.model_factory import ModelFactory
from model.flexible_multitask_head import FlexibleMultiTaskLoss, FlexibleMetricsCalculator
from model.gradnorm_loss import GradNormLoss, GradNormTrainer
from config.biomarker_config import FlexibleBiomarkerConfig
from config.experiment_config import (
    ExperimentConfig, get_model_defaults, DEFAULT_AUGMENTATIONS,
    parse_augmentation_string, create_optimizer, create_scheduler
)
from utils.checkpoints import save_checkpoint, load_checkpoint

# Set device
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
print(f"Using device: {device}")


def set_global_seed(seed: int, deterministic: bool = False) -> None:
    """Set global random seeds for reproducible training runs."""
    random.seed(seed)
    np.random.seed(seed)
    torch.manual_seed(seed)
    if torch.cuda.is_available():
        torch.cuda.manual_seed_all(seed)

    if deterministic:
        torch.backends.cudnn.deterministic = True
        torch.backends.cudnn.benchmark = False
    else:
        torch.backends.cudnn.deterministic = False
        torch.backends.cudnn.benchmark = True


def normalize_fine_tuning_strategy(strategy: str) -> str:
    """Normalize fine-tuning strategy values to canonical internal names."""
    normalized = strategy.strip().lower().replace("-", "_").replace(" ", "_")
    if normalized in {"linear_probe", "linearprobe"}:
        return "linear_probe"
    return "full"


def compute_class_weights_for_dataset(dataset, biomarker_config: FlexibleBiomarkerConfig):
    """Compute class weights for balanced training"""
    class_weights = {}
    
    # Get all target tensors
    all_targets = dataset.targets
    
    # Compute weights for binary tasks
    for biomarker in biomarker_config.binary_biomarkers:
        layout = biomarker_config.get_tensor_layout()[biomarker.name]
        labels = all_targets[:, layout.start_idx]
        
        unique_classes = np.unique(labels)
        if len(unique_classes) > 1:
            weights = compute_class_weight('balanced', classes=unique_classes, y=labels)
            # Use positive class weight for BCE
            pos_weight = weights[1] / weights[0] if len(weights) > 1 else 1.0
            class_weights[biomarker.name] = pos_weight
        else:
            class_weights[biomarker.name] = 1.0
    
    return class_weights


def create_data_transforms(config: ExperimentConfig, is_training=True):
    """Create data transforms optimized for different pre-trained models on medical images"""
    aug_params = parse_augmentation_string(config.image_augmentations)
    
    if is_training:
        transform_list = []
        
        # Different augmentation strategies based on pre-training source
        if config.pretrained_weights == "RadImageNet":
            # RadImageNet-specific augmentations (more conservative for medical domain)
            if aug_params['horizontal_flip']:
                # Very conservative for medical images
                transform_list.append(transforms.RandomHorizontalFlip(p=0.2))
            
            # Use RandomApply to prevent over-augmentation
            geometric_augs = []
            
            if aug_params['rotation'] > 0:
                # Very conservative rotation for medical images
                geometric_augs.append(transforms.RandomRotation(degrees=aug_params['rotation']//3))  # 1/3 the rotation
            
            if aug_params['random_crop']:
                # Minimal cropping to preserve anatomical features
                geometric_augs.append(transforms.RandomResizedCrop(
                    256, 
                    scale=(0.95, 1.0),     # Very conservative cropping
                    ratio=(0.9, 1.1)       # Minimal aspect ratio change
                ))
            
            # Apply geometric augmentations with low probability
            if geometric_augs:
                transform_list.append(transforms.RandomApply(geometric_augs, p=0.4))
            
            # Minimal color augmentations for medical images
            if aug_params['color_jitter']:
                transform_list.append(transforms.RandomApply([
                    transforms.ColorJitter(
                        brightness=aug_params['brightness'] * 0.3,  # Very reduced intensity
                        contrast=aug_params['contrast'] * 0.3       # Very reduced intensity
                    )
                ], p=0.3))  # Very low probability
            
        else:
            # ImageNet or non-pretrained augmentations (more aggressive)
            if aug_params['horizontal_flip']:
                # Reduced probability for medical images (anatomical consistency)
                transform_list.append(transforms.RandomHorizontalFlip(p=0.3))
            
            # Use RandomApply to prevent over-augmentation
            geometric_augs = []
            
            if aug_params['rotation'] > 0:
                # More conservative rotation for medical images
                geometric_augs.append(transforms.RandomRotation(degrees=aug_params['rotation']//2))  # Half the rotation
            
            if aug_params['random_crop']:
                # Less aggressive cropping to preserve anatomical features
                geometric_augs.append(transforms.RandomResizedCrop(
                    256, 
                    scale=(0.9, 1.0),      # Less aggressive cropping
                    ratio=(0.8, 1.2)       # Slightly wider aspect ratio
                ))
            
            # Apply geometric augmentations with moderate probability
            if geometric_augs:
                transform_list.append(transforms.RandomApply(geometric_augs, p=0.6))
            
            # Color augmentations (less critical for grayscale, but helps with domain adaptation)
            if aug_params['color_jitter']:
                transform_list.append(transforms.RandomApply([
                    transforms.ColorJitter(
                        brightness=aug_params['brightness'] * 0.5,  # Reduced intensity
                        contrast=aug_params['contrast'] * 0.5       # Reduced intensity
                    )
                ], p=0.4))  # Lower probability
        
        # Convert to tensor
        transform_list.append(transforms.ToTensor())
        
        # CRITICAL: Convert grayscale to 3-channel for pre-trained models
        transform_list.append(transforms.Lambda(lambda x: x.repeat(3, 1, 1)))
        
        # Add normalization - use appropriate stats based on pre-training
        if aug_params['imagenet_norm']:
            if config.pretrained_weights == "ImageNet":
                # Use ImageNet normalization for ImageNet pre-trained models
                transform_list.append(transforms.Normalize(
                    mean=[0.485, 0.456, 0.406], 
                    std=[0.229, 0.224, 0.225]
                ))
            elif config.pretrained_weights == "RadImageNet":
                # Use RadImageNet normalization (medical imaging specific)
                # Note: These are estimated values - you may need to adjust based on actual RadImageNet stats
                transform_list.append(transforms.Normalize(
                    mean=[0.485, 0.456, 0.406],  # Using ImageNet stats as fallback
                    std=[0.229, 0.224, 0.225]   # RadImageNet likely uses similar normalization
                ))
            else:
                # Use CT-specific normalization for non-pretrained models
                transform_list.append(transforms.Normalize(
                    mean=[0.55001191, 0.55001191, 0.55001191], 
                    std=[0.18854326, 0.18854326, 0.18854326]
                ))
        
        return transforms.Compose(transform_list)
    
    else:
        # Validation/test transforms (no augmentation)
        transform_list = [transforms.ToTensor()]
        
        # CRITICAL: Convert grayscale to 3-channel for pre-trained models
        transform_list.append(transforms.Lambda(lambda x: x.repeat(3, 1, 1)))
        
        if aug_params['imagenet_norm']:
            if config.pretrained_weights == "ImageNet":
                # Use ImageNet normalization for ImageNet pre-trained models
                transform_list.append(transforms.Normalize(
                    mean=[0.485, 0.456, 0.406], 
                    std=[0.229, 0.224, 0.225]
                ))
            elif config.pretrained_weights == "RadImageNet":
                # Use RadImageNet normalization (medical imaging specific)
                # Note: These are estimated values - you may need to adjust based on actual RadImageNet stats
                transform_list.append(transforms.Normalize(
                    mean=[0.485, 0.456, 0.406],  # Using ImageNet stats as fallback
                    std=[0.229, 0.224, 0.225]   # RadImageNet likely uses similar normalization
                ))
            else:
                # Use CT-specific normalization for non-pretrained models
                transform_list.append(transforms.Normalize(
                    mean=[0.55001191, 0.55001191, 0.55001191], 
                    std=[0.18854326, 0.18854326, 0.18854326]
                ))
        
        return transforms.Compose(transform_list)


def setup_logging(output_dir: str, experiment_name: str):
    """Set up comprehensive logging for the experiment"""
    # Create logs directory
    logs_dir = os.path.join(output_dir, 'logs')
    os.makedirs(logs_dir, exist_ok=True)
    
    # Set up main logger
    logger = logging.getLogger('experiment')
    logger.setLevel(logging.INFO)
    
    # Clear existing handlers
    logger.handlers.clear()
    
    # Create formatters
    detailed_formatter = logging.Formatter(
        '%(asctime)s - %(name)s - %(levelname)s - %(message)s'
    )
    simple_formatter = logging.Formatter('%(asctime)s - %(message)s')
    
    # File handler for detailed logs
    detailed_log_file = os.path.join(logs_dir, 'experiment_detailed.log')
    file_handler = logging.FileHandler(detailed_log_file)
    file_handler.setLevel(logging.INFO)
    file_handler.setFormatter(detailed_formatter)
    logger.addHandler(file_handler)
    
    # Console handler
    console_handler = logging.StreamHandler(sys.stdout)
    console_handler.setLevel(logging.INFO)
    console_handler.setFormatter(simple_formatter)
    logger.addHandler(console_handler)
    
    # Create separate logger for training progress
    training_logger = logging.getLogger('training')
    training_logger.setLevel(logging.INFO)
    training_logger.handlers.clear()
    
    training_log_file = os.path.join(logs_dir, 'training_progress.log')
    training_handler = logging.FileHandler(training_log_file)
    training_handler.setLevel(logging.INFO)
    training_handler.setFormatter(simple_formatter)
    training_logger.addHandler(training_handler)
    training_logger.addHandler(console_handler)
    
    # Log experiment start
    logger.info(f"Starting experiment: {experiment_name}")
    logger.info(f"Output directory: {output_dir}")
    logger.info(f"Logs directory: {logs_dir}")
    
    return logger, training_logger


def create_balanced_sampler(dataset, biomarker_config: FlexibleBiomarkerConfig):
    """Create balanced sampler for training"""
    # Get all target tensors
    all_targets = dataset.targets
    
    # Create sample weights based on inverse frequency
    sample_weights = np.ones(len(dataset), dtype=np.float64)
    
    # Weight based on binary biomarkers
    for biomarker in biomarker_config.binary_biomarkers:
        layout = biomarker_config.get_tensor_layout()[biomarker.name]
        labels = all_targets[:, layout.start_idx]
        
        unique, counts = np.unique(labels, return_counts=True)
        class_weights = len(labels) / (len(unique) * counts)
        
        for j, label in enumerate(labels):
            sample_weights[j] *= class_weights[int(label)]
    
    # Convert to torch.DoubleTensor to prevent overflow and match WeightedRandomSampler expectations
    sample_weights_tensor = torch.from_numpy(sample_weights).double()
    
    return WeightedRandomSampler(sample_weights_tensor, len(sample_weights_tensor), replacement=True)


def train_epoch(model, dataloader, criterion, optimizer, device, metrics_calc, gradnorm_trainer=None):
    """Train for one epoch"""
    model.train()
    
    total_loss = 0
    all_predictions = []
    all_targets = []
    loss_components = {'total_loss': 0}
    
    for batch_idx, (images, targets) in enumerate(tqdm(dataloader, desc="Training")):
        images = images.to(device)
        targets = targets.to(device)
        
        # Convert single channel to 3-channel for models expecting RGB
        if images.shape[1] == 1:
            images = images.repeat(1, 3, 1, 1)
        
        # Forward pass
        predictions = model(images)
        
        # Calculate loss - use GradNorm if available
        if gradnorm_trainer is not None:
            loss, loss_dict = gradnorm_trainer.compute_loss(model, predictions, targets)
        else:
            loss, loss_dict = criterion(predictions, targets)
        
        # Backward pass
        optimizer.zero_grad()
        loss.backward()
        optimizer.step()

        # Accumulate metrics
        total_loss += loss.item()
        for key, value in loss_dict.items():
            if key not in loss_components:
                loss_components[key] = 0
            loss_components[key] += value
        
        # Store predictions and targets for metric calculation
        all_predictions.append(predictions.detach().cpu())
        all_targets.append(targets.detach().cpu())
    
    # Calculate metrics
    all_predictions = torch.cat(all_predictions, dim=0)
    all_targets = torch.cat(all_targets, dim=0)
    
    metrics = metrics_calc.calculate_all_metrics(all_predictions, all_targets)
    
    # Average losses
    avg_loss = total_loss / len(dataloader)
    for key in loss_components:
        loss_components[key] /= len(dataloader)
    
    return avg_loss, metrics, loss_components


def validate_epoch(model, dataloader, criterion, device, metrics_calc):
    """Validate for one epoch"""
    model.eval()
    
    total_loss = 0
    all_predictions = []
    all_targets = []
    loss_components = {'total_loss': 0}
    
    with torch.no_grad():
        for batch_idx, (images, targets) in enumerate(tqdm(dataloader, desc="Validation")):
            images = images.to(device)
            targets = targets.to(device)
            
            # Convert single channel to 3-channel for models expecting RGB
            if images.shape[1] == 1:
                images = images.repeat(1, 3, 1, 1)
            
            # Forward pass
            predictions = model(images)
            
            # Calculate loss
            loss, loss_dict = criterion(predictions, targets)
            
            # Accumulate metrics
            total_loss += loss.item()
            for key, value in loss_dict.items():
                if key not in loss_components:
                    loss_components[key] = 0
                loss_components[key] += value
            
            # Store predictions and targets for metric calculation
            all_predictions.append(predictions.detach().cpu())
            all_targets.append(targets.detach().cpu())
    
    # Calculate metrics with threshold optimization
    all_predictions = torch.cat(all_predictions, dim=0)
    all_targets = torch.cat(all_targets, dim=0)
    
    # Update optimal thresholds based on validation data
    metrics_calc.update_optimal_thresholds(all_predictions, all_targets)
    
    # Calculate metrics using optimal thresholds
    metrics = metrics_calc.calculate_all_metrics(all_predictions, all_targets)
    
    # Average losses
    avg_loss = total_loss / len(dataloader)
    for key in loss_components:
        loss_components[key] /= len(dataloader)
    
    return avg_loss, metrics, loss_components


def train_model(
    config: ExperimentConfig,
    data_dir: str,
    output_dir: str,
    biomarker_config: FlexibleBiomarkerConfig,
    epochs: int = 100,
    seed: int = 42,
):
    """Main training function"""
    
    # Safety check: if directory exists and has important files, create a new one
    if os.path.exists(output_dir):
        important_files = ['best_checkpoint.pth', 'config.json', 'experiment_results.csv']
        has_important_files = any(os.path.exists(os.path.join(output_dir, f)) for f in important_files)
        
        if has_important_files:
            import datetime
            timestamp = datetime.datetime.now().strftime("%Y%m%d_%H%M%S")
            original_output_dir = output_dir
            output_dir = f"{output_dir}_{timestamp}"
            print(f"Warning: Output directory {original_output_dir} exists with important files.")
            print(f"Using new directory: {output_dir}")
    
    # Create output directory
    os.makedirs(output_dir, exist_ok=True)
    
    # Setup comprehensive logging
    logger, training_logger = setup_logging(output_dir, config.experiment_name)
    
    # Setup tensorboard logging
    writer = SummaryWriter(log_dir=os.path.join(output_dir, 'tensorboard'))
    
    # Save configuration
    config_file = os.path.join(output_dir, 'config.json')
    with open(config_file, 'w') as f:
        json.dump(config.to_dict(), f, indent=2)
    logger.info(f"Configuration saved to: {config_file}")
    
    # Save biomarker configuration
    biomarker_config_file = os.path.join(output_dir, 'biomarker_config.json')
    biomarker_config.save_to_file(biomarker_config_file)
    logger.info(f"Biomarker configuration saved to: {biomarker_config_file}")
    
    logger.info(f"Model: {config.model}")
    logger.info(f"Single-target strategy: {config.single_target_strategy}")
    logger.info(f"Multi-target strategy: {config.multi_target_strategy}")
    logger.info(f"Expected GPU memory: {config.expected_gpu_memory}")
    logger.info(f"Training epochs: {epochs}")
    logger.info(f"Data directory: {data_dir}")
    logger.info(f"Biomarker configuration: {biomarker_config.experiment_name}")
    logger.info(f"Total output size: {biomarker_config.total_output_size}")
    logger.info(f"Seed: {seed}")
    
    # Create data transforms
    train_transform = create_data_transforms(config, is_training=True)
    val_transform = create_data_transforms(config, is_training=False)
    
    # Load datasets
    logger.info("Loading datasets...")
    train_dataset = ClassifierDataset(
        data_dir, biomarker_config, transforms=train_transform, 
        size=256, train=True
    )
    
    val_dataset = ClassifierDataset(
        data_dir, biomarker_config, transforms=val_transform,
        size=256, train=False
    )
    
    logger.info(f"Train dataset size: {len(train_dataset)}")
    logger.info(f"Validation dataset size: {len(val_dataset)}")
    
    # Compute class weights if specified
    class_weights = None
    if config.class_weighting == 'inverse_frequency':
        logger.info("Computing class weights...")
        class_weights = compute_class_weights_for_dataset(train_dataset, biomarker_config)
        logger.info(f"Class weights computed for {len(class_weights)} binary biomarkers")
    
    # Create data loaders
    data_loader_generator = torch.Generator()
    data_loader_generator.manual_seed(seed)

    def _seed_worker(worker_id: int) -> None:
        worker_seed = seed + worker_id
        np.random.seed(worker_seed)
        random.seed(worker_seed)
        torch.manual_seed(worker_seed)

    if config.sampling_strategy == 'balanced_batch':
        train_sampler = create_balanced_sampler(train_dataset, biomarker_config)
        train_loader = DataLoader(
            train_dataset, batch_size=config.batch_size, 
            sampler=train_sampler, num_workers=8, pin_memory=True,
            worker_init_fn=_seed_worker, generator=data_loader_generator
        )
    else:
        train_loader = DataLoader(
            train_dataset, batch_size=config.batch_size, 
            shuffle=True, num_workers=8, pin_memory=True,
            worker_init_fn=_seed_worker, generator=data_loader_generator
        )
    
    val_loader = DataLoader(
        val_dataset, batch_size=config.batch_size, 
        shuffle=False, num_workers=8, pin_memory=True,
        worker_init_fn=_seed_worker, generator=data_loader_generator
    )
    
    # Create model
    logger.info("Creating model...")
    model = ModelFactory.create_model(
        architecture=config.model,
        num_classes=biomarker_config.total_output_size,
        pretrained_weights=config.pretrained_weights,
        fine_tuning_strategy=config.fine_tuning_strategy,
        dropout=config.dropout,
        biomarker_config=biomarker_config,
        single_target_strategy=config.single_target_strategy
    )

    model = model.to(device)

    # Log model info
    total_params = sum(p.numel() for p in model.parameters())
    trainable_params = sum(p.numel() for p in model.parameters() if p.requires_grad)
    logger.info(f"Total parameters: {total_params:,}")
    logger.info(f"Trainable parameters: {trainable_params:,}")
    
    # Create loss function - check if GradNorm is enabled
    use_gradnorm = getattr(config, 'use_gradnorm', False)
    gradnorm_trainer = None
    
    if use_gradnorm:
        logger.info("Using GradNorm for loss balancing")
        gradnorm_alpha = getattr(config, 'gradnorm_alpha', 0.16)
        gradnorm_update_freq = getattr(config, 'gradnorm_update_freq', 10)
        
        gradnorm_loss = GradNormLoss(
            biomarker_config=biomarker_config,
            class_weights=class_weights,
            alpha=gradnorm_alpha,
            update_weights_every=gradnorm_update_freq,
            initial_task_loss_average_window=20,
            normalize_losses=True,
            restoring_force_factor=0.1
        )
        gradnorm_loss = gradnorm_loss.to(device)  # Move to correct device
        gradnorm_trainer = GradNormTrainer(gradnorm_loss)
        criterion = gradnorm_loss  # For validation
    else:
        criterion = FlexibleMultiTaskLoss(biomarker_config, class_weights=class_weights)
    
    # Create optimizer and scheduler
    # Note: GradNorm task weights are updated manually, not through optimizer
    optimizer = create_optimizer(model.parameters(), config)
    
    scheduler = create_scheduler(optimizer, config, epochs)
    
    # Create metrics calculator
    metrics_calc = FlexibleMetricsCalculator(biomarker_config)
    
    # Training loop
    best_median_auroc = 0.0
    best_mae = float('inf')  # Initialize MAE tracking for continuous-only scenarios
    best_epoch = 0
    patience = 10
    patience_counter = 0
    
    # Determine task types for logging and selection
    has_classification_tasks = (len(biomarker_config.binary_biomarkers) > 0 or 
                               len(biomarker_config.multiclass_biomarkers) > 0)
    has_continuous_tasks = len(biomarker_config.continuous_biomarkers) > 0
    
    logger.info(f"Starting training for {epochs} epochs with early stopping (patience: {patience})")
    if has_classification_tasks and has_continuous_tasks:
        logger.info("Multi-task training: Classification + Regression")
    elif has_classification_tasks:
        logger.info("Classification training: Using median AUROC for best model selection")
    elif has_continuous_tasks:
        logger.info("Regression training: Using MAE for best model selection")
    
    for epoch in range(epochs):
        epoch_start_time = time.time()
        
        training_logger.info(f"Starting Epoch {epoch+1}/{epochs}")
        
        # Training phase
        train_loss, train_metrics, train_loss_components = train_epoch(
            model, train_loader, criterion, optimizer, device, metrics_calc, gradnorm_trainer
        )
        
        # Validation phase
        val_loss, val_metrics, val_loss_components = validate_epoch(
            model, val_loader, criterion, device, metrics_calc
        )
        
        # Update scheduler
        if isinstance(scheduler, torch.optim.lr_scheduler.ReduceLROnPlateau):
            scheduler.step(val_loss)
        else:
            scheduler.step()
        
        epoch_time = time.time() - epoch_start_time
        
        # Log to tensorboard
        writer.add_scalar('Loss/Train', train_loss, epoch)
        writer.add_scalar('Loss/Validation', val_loss, epoch)
        writer.add_scalar('Metrics/Average_AUROC_Train', train_metrics['average_auroc'], epoch)
        writer.add_scalar('Metrics/Average_AUROC_Val', val_metrics['average_auroc'], epoch)
        writer.add_scalar('Metrics/Median_AUROC_Train', train_metrics['median_auroc'], epoch)
        writer.add_scalar('Metrics/Median_AUROC_Val', val_metrics['median_auroc'], epoch)
        writer.add_scalar('Learning_Rate', optimizer.param_groups[0]['lr'], epoch)
        
        # Log GradNorm weights if using GradNorm
        if gradnorm_trainer is not None:
            gradnorm_stats = gradnorm_trainer.get_training_stats()
            task_weights = gradnorm_stats['task_weights']
            
            for task_name, weight in task_weights.items():
                writer.add_scalar(f'GradNorm_Weights/{task_name}', weight, epoch)
            
            # Log GradNorm info
            if gradnorm_stats['initial_losses_computed']:
                training_logger.info(f"GradNorm task weights: {task_weights}")
        
        # Log individual biomarker metrics
        for biomarker in biomarker_config.binary_biomarkers:
            if biomarker.name in train_metrics and biomarker.name in val_metrics:
                train_biomarker_metrics = train_metrics[biomarker.name]
                val_biomarker_metrics = val_metrics[biomarker.name]
                
                if 'auroc' in train_biomarker_metrics and 'auroc' in val_biomarker_metrics:
                    writer.add_scalar(f'AUROC_Train/{biomarker.name}', train_biomarker_metrics['auroc'], epoch)
                    writer.add_scalar(f'AUROC_Val/{biomarker.name}', val_biomarker_metrics['auroc'], epoch)
        
        # Log epoch results
        training_logger.info(f"Epoch {epoch+1} completed in {epoch_time:.2f}s")
        training_logger.info(f"Train Loss: {train_loss:.4f}, Val Loss: {val_loss:.4f}")
        training_logger.info(f"Train Avg AUROC: {train_metrics['average_auroc']:.4f}, Val Avg AUROC: {val_metrics['average_auroc']:.4f}")
        training_logger.info(f"Train Median AUROC: {train_metrics['median_auroc']:.4f}, Val Median AUROC: {val_metrics['median_auroc']:.4f}")
        
        # Log individual biomarker validation metrics
        training_logger.info("Validation metrics per biomarker:")
        for biomarker in biomarker_config.binary_biomarkers:
            if biomarker.name in val_metrics:
                biomarker_metrics = val_metrics[biomarker.name]
                if 'auroc' in biomarker_metrics and 'accuracy' in biomarker_metrics:
                    training_logger.info(f"  {biomarker.name}: AUROC={biomarker_metrics['auroc']:.4f}, "
                                       f"Acc={biomarker_metrics['accuracy']:.4f}, "
                                       f"F1={biomarker_metrics.get('f1', 0.0):.4f}")
        
        # Log multiclass biomarker metrics if any exist
        for biomarker in biomarker_config.multiclass_biomarkers:
            if biomarker.name in val_metrics:
                biomarker_metrics = val_metrics[biomarker.name]
                if 'accuracy' in biomarker_metrics:
                    training_logger.info(f"  {biomarker.name}: Acc={biomarker_metrics['accuracy']:.4f}, "
                                       f"F1={biomarker_metrics.get('f1_weighted', 0.0):.4f}")
        
        # Log continuous biomarker metrics if any exist
        for biomarker in biomarker_config.continuous_biomarkers:
            if biomarker.name in val_metrics:
                biomarker_metrics = val_metrics[biomarker.name]
                if 'mse' in biomarker_metrics:
                    training_logger.info(f"  {biomarker.name}: MSE={biomarker_metrics['mse']:.4f}, "
                                       f"MAE={biomarker_metrics.get('mae', 0.0):.4f}")
        
        # Save checkpoint if best model
        # Use MAE for continuous-only scenarios, AUROC for classification scenarios
        
        if has_classification_tasks:
            # Use median AUROC for classification scenarios
            # Check if we actually have AUROC values (not just 0.0)
            if val_metrics['median_auroc'] > 0.0:
                is_best = val_metrics['median_auroc'] > best_median_auroc
                if is_best:
                    best_median_auroc = val_metrics['median_auroc']
                    best_epoch = epoch + 1
                    patience_counter = 0  # Reset patience counter
                    training_logger.info(f"New best model! Median AUROC: {best_median_auroc:.4f} (Avg: {val_metrics['average_auroc']:.4f})")
                else:
                    patience_counter += 1
                    training_logger.info(f"No improvement. Patience: {patience_counter}/{patience}")
                    
                    # Early stopping check
                    if patience_counter >= patience:
                        training_logger.info("Early stopping triggered!")
                        break
            else:
                # No actual AUROC values available - fall back to MAE if continuous tasks exist
                if has_continuous_tasks:
                    training_logger.warning("No AUROC values available for classification tasks, falling back to MAE for model selection")
                    # Use MAE logic (will be handled in the elif block below)
                    pass
                else:
                    # No meaningful metrics available
                    is_best = False
                    patience_counter += 1
                    training_logger.warning("No AUROC values available and no continuous tasks - cannot determine best model")
        elif has_continuous_tasks:
            # Use MAE for continuous-only scenarios (lower MAE is better)
            mae_values = []
            for biomarker in biomarker_config.continuous_biomarkers:
                if biomarker.name in val_metrics:
                    mae = val_metrics[biomarker.name].get('mae', float('inf'))
                    mae_values.append(mae)
                    training_logger.info(f"  {biomarker.name}: MAE={mae:.4f}")
            
            if mae_values:
                current_mae = np.mean(mae_values)
                is_best = current_mae < best_mae
                if is_best:
                    best_mae = current_mae
                    best_epoch = epoch + 1
                    patience_counter = 0  # Reset patience counter
                    training_logger.info(f"New best model! Average MAE: {best_mae:.4f}")
                else:
                    patience_counter += 1
                    training_logger.info(f"No improvement. Current MAE: {current_mae:.4f}, Best MAE: {best_mae:.4f}. Patience: {patience_counter}/{patience}")
                    
                    # Early stopping check
                    if patience_counter >= patience:
                        training_logger.info("Early stopping triggered!")
                        break
            else:
                # Fallback if no MAE values available
                is_best = False
                patience_counter += 1
                training_logger.warning("No MAE values available for continuous biomarkers")
        else:
            # No biomarkers configured - should not happen
            is_best = False
            training_logger.error("No biomarkers configured!")
            break
        
        # Save checkpoint
        checkpoint = {
            'epoch': epoch + 1,
            'model_state_dict': model.state_dict(),
            'optimizer_state_dict': optimizer.state_dict(),
            'scheduler_state_dict': scheduler.state_dict(),
            'train_loss': train_loss,
            'val_loss': val_loss,
            'train_metrics': train_metrics,
            'val_metrics': val_metrics,
            'config': config.to_dict(),
            'biomarker_config': biomarker_config.experiment_name,
            'best_median_auroc': best_median_auroc,
            'best_mae': best_mae,
            'best_epoch': best_epoch,
            'optimal_thresholds': metrics_calc.optimal_thresholds  # Save optimal thresholds
        }
        
        # Save latest and best checkpoints
        torch.save(checkpoint, os.path.join(output_dir, 'latest_checkpoint.pth'))
        if is_best:
            torch.save(checkpoint, os.path.join(output_dir, 'best_checkpoint.pth'))
    
    # Close tensorboard writer
    writer.close()
    
    # Final logging based on task type and actual metrics used
    if has_classification_tasks and best_median_auroc > 0.0:
        # Classification tasks with actual AUROC values
        logger.info(f"Training completed! Best model at epoch {best_epoch} with Median AUROC: {best_median_auroc:.4f}")
        return model, best_median_auroc
    elif has_continuous_tasks:
        # Continuous tasks (either only continuous, or classification tasks fell back to MAE)
        logger.info(f"Training completed! Best model at epoch {best_epoch} with MAE: {best_mae:.4f}")
        return model, best_mae
    else:
        # No meaningful metrics available
        logger.info(f"Training completed! Best model at epoch {best_epoch} (no meaningful metrics available)")
        return model, 0.0


def main():
    parser = ArgumentParser(description='Flexible Multi-Task Training')

    # --- Required arguments ---
    parser.add_argument('--model', required=True,
                        help='Model architecture to train (e.g. "ResNet-18", "ViT-Small (DINOv2)"). '
                             'See README for the full list of supported models.')
    parser.add_argument('--data_dir', required=True,
                        help='Path to dataset directory. Must contain train.csv, val.csv, and a data/ subfolder with PNG images.')
    parser.add_argument('--biomarker_config', required=True,
                        help='Path to biomarker configuration file (YAML or JSON). '
                             'See config/biomarker_config_multitask_example.yaml for the full multi-task config used in the paper.')

    # --- Output ---
    parser.add_argument('--output_dir', default='./outputs',
                        help='Directory to save checkpoints, logs, and TensorBoard (default: ./outputs)')
    parser.add_argument('--experiment_name',
                        help='Custom name for this run. Auto-generated from model/lr/batch if not provided.')

    # --- Training hyperparameters (defaults match the published experiments) ---
    parser.add_argument('--epochs', type=int, default=100,
                        help='Number of training epochs (default: 100)')
    parser.add_argument('--learning_rate', type=float, default=1e-4,
                        help='Learning rate (default: 1e-4)')
    parser.add_argument('--batch_size', type=int, default=16,
                        help='Batch size (default: 16)')
    parser.add_argument('--weight_decay', type=float, default=1e-4,
                        help='Weight decay (default: 1e-4)')
    parser.add_argument('--optimizer', default='AdamW', choices=['AdamW', 'Adam', 'SGD'],
                        help='Optimizer (default: AdamW)')
    parser.add_argument('--scheduler', default='CosineAnnealing',
                        choices=['CosineAnnealing', 'CosineAnnealingWarmRestarts',
                                 'ReduceLROnPlateau', 'StepLR', 'ExponentialLR'],
                        help='LR scheduler (default: CosineAnnealing)')
    parser.add_argument('--dropout', type=float, default=0.2,
                        help='Dropout rate (default: 0.2)')
    parser.add_argument('--class_weighting', default='inverse_frequency',
                        choices=['inverse_frequency', 'none'],
                        help='Class weighting strategy (default: inverse_frequency)')
    parser.add_argument('--sampling_strategy', default='balanced_batch',
                        choices=['balanced_batch', 'random'],
                        help='Sampling strategy for training DataLoader (default: balanced_batch)')
    parser.add_argument('--fine_tuning_strategy', default='Full fine-tuning',
                        choices=['Full fine-tuning', 'full', 'linear_probe'],
                        help='Fine-tuning strategy (default: Full fine-tuning)')
    parser.add_argument('--seed', type=int, default=42,
                        help='Global random seed for reproducibility (default: 42)')
    parser.add_argument('--deterministic', action='store_true',
                        help='Enable deterministic backend behavior (can reduce throughput)')

    # --- Per-model defaults (auto-detected from model name if not specified) ---
    parser.add_argument('--pretrained_weights',
                        help='Pretrained weights to use. Auto-detected from --model if not provided.')
    parser.add_argument('--single_target_strategy',
                        help='Single-target strategy. Auto-detected from --model if not provided.')

    # --- GradNorm ---
    parser.add_argument('--use_gradnorm', action='store_true',
                        help='Enable GradNorm adaptive loss balancing')
    parser.add_argument('--gradnorm_alpha', type=float, default=0.16,
                        help='GradNorm restoring force strength (default: 0.16)')
    parser.add_argument('--gradnorm_update_freq', type=int, default=10,
                        help='Update GradNorm weights every N iterations (default: 10)')

    args = parser.parse_args()
    set_global_seed(args.seed, deterministic=args.deterministic)

    # Load biomarker configuration
    print(f"Loading biomarker configuration from: {args.biomarker_config}")
    biomarker_config = FlexibleBiomarkerConfig(args.biomarker_config)

    print("Biomarker configuration loaded:")
    biomarker_config.print_summary()

    # Resolve per-model defaults for pretrained_weights and single_target_strategy
    model_defaults = get_model_defaults(args.model)
    pretrained_weights = args.pretrained_weights or model_defaults['pretrained_weights']
    single_target_strategy = args.single_target_strategy or model_defaults['single_target_strategy']
    fine_tuning_strategy = normalize_fine_tuning_strategy(args.fine_tuning_strategy)

    # Build ExperimentConfig directly from CLI args
    config = ExperimentConfig(
        model=args.model,
        loss_function='CE',
        must_include=True,
        learning_rate=[args.learning_rate],
        batch_size=args.batch_size,
        weight_decay=args.weight_decay,
        optimizer=args.optimizer,
        scheduler=args.scheduler,
        image_augmentations=DEFAULT_AUGMENTATIONS.copy(),
        dropout=args.dropout,
        loss_specific_params='class_weights=inverse_frequency',
        multi_target_strategy='Shared backbone + task-specific heads',
        single_target_strategy=single_target_strategy,
        pretrained_weights=pretrained_weights,
        fine_tuning_strategy=fine_tuning_strategy,
        expected_gpu_memory='',
        architectural_family='',
        class_weighting=args.class_weighting,
        sampling_strategy=args.sampling_strategy,
        threshold_selection='F1_optimal',
        experiment_name=args.experiment_name or '',
        use_gradnorm=args.use_gradnorm,
        gradnorm_alpha=args.gradnorm_alpha,
        gradnorm_update_freq=args.gradnorm_update_freq,
    )

    output_dir = os.path.join(args.output_dir, config.experiment_name)

    print(f"\n{'='*50}")
    print(f"Training: {config.model}")
    print(f"  Pretrained weights:      {pretrained_weights}")
    print(f"  Single-target strategy:  {single_target_strategy}")
    print(f"  Learning rate:           {args.learning_rate}")
    print(f"  Batch size:              {args.batch_size}")
    print(f"  Epochs:                  {args.epochs}")
    print(f"  Seed:                    {args.seed}")
    print(f"  Deterministic:           {args.deterministic}")
    print(f"  Output dir:              {output_dir}")
    print(f"{'='*50}\n")

    train_model(
        config=config,
        data_dir=args.data_dir,
        output_dir=output_dir,
        biomarker_config=biomarker_config,
        epochs=args.epochs,
        seed=args.seed,
    )


if __name__ == "__main__":
    main()