bruniss commited on
Commit
5d8b7d7
·
verified ·
1 Parent(s): 95a9dc0

Upload 12 files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ fold_0/progress.png filter=lfs diff=lfs merge=lfs -text
dataset.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "channel_names": {
3
+ "0": "CT"
4
+ },
5
+ "labels": {
6
+ "background": 0,
7
+ "fiber": 1,
8
+ "ignore": 2
9
+ },
10
+ "numTraining": 1854,
11
+ "file_ending": ".tif",
12
+ "overwrite_image_reader_writer": "Tiff3DIO"
13
+ }
dataset_fingerprint.json ADDED
The diff for this file is too large to render. See raw diff
 
fold_0/checkpoint_best.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0857f609010abebffe51b1de5362a077475384c324f77d37fee695bcdcd5ea70
3
+ size 819301071
fold_0/checkpoint_final.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84425d0a69db5ba49a55427eb2b3229c3ca4d28216a9fba053117631ac9ff690
3
+ size 819310459
fold_0/debug.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_best_ema": "None",
3
+ "batch_size": "2",
4
+ "configuration_manager": "{'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [192, 192, 192], 'median_image_size_in_voxels': [236.0, 236.0, 236.0], 'spacing': [1.0, 1.0, 1.0], 'normalization_schemes': ['CTNormalization'], 'use_mask_for_norm': [False], 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'architecture': {'network_class_name': 'dynamic_network_architectures.architectures.unet.ResidualEncoderUNet', 'arch_kwargs': {'n_stages': 6, 'features_per_stage': [32, 64, 128, 256, 320, 320], 'conv_op': 'torch.nn.modules.conv.Conv3d', 'kernel_sizes': [[3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'strides': [[1, 1, 1], [2, 2, 2], [2, 2, 2], [2, 2, 2], [2, 2, 2], [2, 2, 2]], 'n_blocks_per_stage': [1, 3, 4, 6, 6, 6], 'n_conv_per_stage_decoder': [1, 1, 1, 1, 1], 'conv_bias': True, 'norm_op': 'torch.nn.modules.instancenorm.InstanceNorm3d', 'norm_op_kwargs': {'eps': 1e-05, 'affine': True}, 'dropout_op': None, 'dropout_op_kwargs': None, 'nonlin': 'torch.nn.LeakyReLU', 'nonlin_kwargs': {'inplace': True}}, '_kw_requires_import': ['conv_op', 'norm_op', 'dropout_op', 'nonlin']}, 'batch_dice': False}",
5
+ "configuration_name": "3d_fullres",
6
+ "cudnn_version": 91002,
7
+ "current_epoch": "0",
8
+ "dataloader_train": "<batchgenerators.dataloading.nondet_multi_threaded_augmenter.NonDetMultiThreadedAugmenter object at 0x79652e4b8da0>",
9
+ "dataloader_train.generator": "<nnunetv2.training.dataloading.data_loader_3d_skel.nnUNetDataLoader3DSkel object at 0x79652e5452e0>",
10
+ "dataloader_train.num_processes": "16",
11
+ "dataloader_train.transform": "None",
12
+ "dataloader_val": "<batchgenerators.dataloading.nondet_multi_threaded_augmenter.NonDetMultiThreadedAugmenter object at 0x79652eaeeea0>",
13
+ "dataloader_val.generator": "<nnunetv2.training.dataloading.data_loader_3d_skel.nnUNetDataLoader3DSkel object at 0x79652e545340>",
14
+ "dataloader_val.num_processes": "8",
15
+ "dataloader_val.transform": "None",
16
+ "dataset_json": "{'channel_names': {'0': 'CT'}, 'labels': {'background': 0, 'fiber': 1, 'ignore': 2}, 'numTraining': 1854, 'file_ending': '.tif', 'overwrite_image_reader_writer': 'Tiff3DIO'}",
17
+ "device": "cuda:0",
18
+ "disable_checkpointing": "False",
19
+ "enable_deep_supervision": "True",
20
+ "fold": "0",
21
+ "folder_with_segs_from_previous_stage": "None",
22
+ "gpu_name": "NVIDIA GeForce RTX 5090",
23
+ "grad_scaler": "<torch.amp.grad_scaler.GradScaler object at 0x79652ecd11f0>",
24
+ "hostname": "seanpc",
25
+ "inference_allowed_mirroring_axes": "(0, 1, 2)",
26
+ "initial_lr": "0.01",
27
+ "is_cascaded": "False",
28
+ "is_ddp": "False",
29
+ "label_manager": "<nnunetv2.utilities.label_handling.label_handling.LabelManager object at 0x79652e898d40>",
30
+ "local_rank": "0",
31
+ "log_file": "/mnt/raid_nvme/datasets/results/Dataset104_059plus0500p2/nnUNetTrainerMedialSurfaceRecall__nnUNetResEncUNetLPlans__3d_fullres/fold_0/training_log_2025_8_23_02_35_03.txt",
32
+ "logger": "<nnunetv2.training.logging.nnunet_logger.nnUNetLogger object at 0x79652ea6ddf0>",
33
+ "loss": "DeepSupervisionWrapper(\n (loss): DC_SkelREC_and_CE_loss(\n (ce): RobustCrossEntropyLoss()\n (dc): MemoryEfficientSoftDiceLoss()\n (srec): SoftSkeletonRecallLoss()\n )\n)",
34
+ "lr_scheduler": "<nnunetv2.training.lr_scheduler.polylr.PolyLRScheduler object at 0x79652c52bd70>",
35
+ "my_init_kwargs": "{'plans': {'dataset_name': 'Dataset104_059plus0500p2', 'plans_name': 'nnUNetResEncUNetLPlans', 'original_median_spacing_after_transp': [1.0, 1.0, 1.0], 'original_median_shape_after_transp': [236, 236, 236], 'image_reader_writer': 'Tiff3DIO', 'transpose_forward': [0, 1, 2], 'transpose_backward': [0, 1, 2], 'configurations': {'2d': {'data_identifier': 'nnUNetPlans_2d', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 141, 'patch_size': [256, 256], 'median_image_size_in_voxels': [236.0, 236.0], 'spacing': [1.0, 1.0], 'normalization_schemes': ['CTNormalization'], 'use_mask_for_norm': [False], 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'architecture': {'network_class_name': 'dynamic_network_architectures.architectures.unet.ResidualEncoderUNet', 'arch_kwargs': {'n_stages': 7, 'features_per_stage': [32, 64, 128, 256, 512, 512, 512], 'conv_op': 'torch.nn.modules.conv.Conv2d', 'kernel_sizes': [[3, 3], [3, 3], [3, 3], [3, 3], [3, 3], [3, 3], [3, 3]], 'strides': [[1, 1], [2, 2], [2, 2], [2, 2], [2, 2], [2, 2], [2, 2]], 'n_blocks_per_stage': [1, 3, 4, 6, 6, 6, 6], 'n_conv_per_stage_decoder': [1, 1, 1, 1, 1, 1], 'conv_bias': True, 'norm_op': 'torch.nn.modules.instancenorm.InstanceNorm2d', 'norm_op_kwargs': {'eps': 1e-05, 'affine': True}, 'dropout_op': None, 'dropout_op_kwargs': None, 'nonlin': 'torch.nn.LeakyReLU', 'nonlin_kwargs': {'inplace': True}}, '_kw_requires_import': ['conv_op', 'norm_op', 'dropout_op', 'nonlin']}, 'batch_dice': True}, '3d_fullres': {'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [192, 192, 192], 'median_image_size_in_voxels': [236.0, 236.0, 236.0], 'spacing': [1.0, 1.0, 1.0], 'normalization_schemes': ['CTNormalization'], 'use_mask_for_norm': [False], 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'architecture': {'network_class_name': 'dynamic_network_architectures.architectures.unet.ResidualEncoderUNet', 'arch_kwargs': {'n_stages': 6, 'features_per_stage': [32, 64, 128, 256, 320, 320], 'conv_op': 'torch.nn.modules.conv.Conv3d', 'kernel_sizes': [[3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'strides': [[1, 1, 1], [2, 2, 2], [2, 2, 2], [2, 2, 2], [2, 2, 2], [2, 2, 2]], 'n_blocks_per_stage': [1, 3, 4, 6, 6, 6], 'n_conv_per_stage_decoder': [1, 1, 1, 1, 1], 'conv_bias': True, 'norm_op': 'torch.nn.modules.instancenorm.InstanceNorm3d', 'norm_op_kwargs': {'eps': 1e-05, 'affine': True}, 'dropout_op': None, 'dropout_op_kwargs': None, 'nonlin': 'torch.nn.LeakyReLU', 'nonlin_kwargs': {'inplace': True}}, '_kw_requires_import': ['conv_op', 'norm_op', 'dropout_op', 'nonlin']}, 'batch_dice': False}}, 'experiment_planner_used': 'nnUNetPlannerResEncL', 'label_manager': 'LabelManager', 'foreground_intensity_properties_per_channel': {'0': {'max': 255.0, 'mean': 126.20806121826172, 'median': 126.0, 'min': 0.0, 'percentile_00_5': 0.0, 'percentile_99_5': 235.0, 'std': 43.85218048095703}}}, 'configuration': '3d_fullres', 'fold': 0, 'dataset_json': {'channel_names': {'0': 'CT'}, 'labels': {'background': 0, 'fiber': 1, 'ignore': 2}, 'numTraining': 1854, 'file_ending': '.tif', 'overwrite_image_reader_writer': 'Tiff3DIO'}, 'unpack_dataset': True, 'device': device(type='cuda'), 'yaml_config_path': None}",
36
+ "network": "OptimizedModule",
37
+ "num_epochs": "1000",
38
+ "num_input_channels": "1",
39
+ "num_iterations_per_epoch": "250",
40
+ "num_val_iterations_per_epoch": "50",
41
+ "optimizer": "SGD (\nParameter Group 0\n dampening: 0\n differentiable: False\n foreach: None\n fused: None\n initial_lr: 0.01\n lr: 0.01\n maximize: False\n momentum: 0.99\n nesterov: True\n weight_decay: 3e-05\n)",
42
+ "output_folder": "/mnt/raid_nvme/datasets/results/Dataset104_059plus0500p2/nnUNetTrainerMedialSurfaceRecall__nnUNetResEncUNetLPlans__3d_fullres/fold_0",
43
+ "output_folder_base": "/mnt/raid_nvme/datasets/results/Dataset104_059plus0500p2/nnUNetTrainerMedialSurfaceRecall__nnUNetResEncUNetLPlans__3d_fullres",
44
+ "oversample_foreground_percent": "0.33",
45
+ "plans_manager": "{'dataset_name': 'Dataset104_059plus0500p2', 'plans_name': 'nnUNetResEncUNetLPlans', 'original_median_spacing_after_transp': [1.0, 1.0, 1.0], 'original_median_shape_after_transp': [236, 236, 236], 'image_reader_writer': 'Tiff3DIO', 'transpose_forward': [0, 1, 2], 'transpose_backward': [0, 1, 2], 'configurations': {'2d': {'data_identifier': 'nnUNetPlans_2d', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 141, 'patch_size': [256, 256], 'median_image_size_in_voxels': [236.0, 236.0], 'spacing': [1.0, 1.0], 'normalization_schemes': ['CTNormalization'], 'use_mask_for_norm': [False], 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'architecture': {'network_class_name': 'dynamic_network_architectures.architectures.unet.ResidualEncoderUNet', 'arch_kwargs': {'n_stages': 7, 'features_per_stage': [32, 64, 128, 256, 512, 512, 512], 'conv_op': 'torch.nn.modules.conv.Conv2d', 'kernel_sizes': [[3, 3], [3, 3], [3, 3], [3, 3], [3, 3], [3, 3], [3, 3]], 'strides': [[1, 1], [2, 2], [2, 2], [2, 2], [2, 2], [2, 2], [2, 2]], 'n_blocks_per_stage': [1, 3, 4, 6, 6, 6, 6], 'n_conv_per_stage_decoder': [1, 1, 1, 1, 1, 1], 'conv_bias': True, 'norm_op': 'torch.nn.modules.instancenorm.InstanceNorm2d', 'norm_op_kwargs': {'eps': 1e-05, 'affine': True}, 'dropout_op': None, 'dropout_op_kwargs': None, 'nonlin': 'torch.nn.LeakyReLU', 'nonlin_kwargs': {'inplace': True}}, '_kw_requires_import': ['conv_op', 'norm_op', 'dropout_op', 'nonlin']}, 'batch_dice': True}, '3d_fullres': {'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [192, 192, 192], 'median_image_size_in_voxels': [236.0, 236.0, 236.0], 'spacing': [1.0, 1.0, 1.0], 'normalization_schemes': ['CTNormalization'], 'use_mask_for_norm': [False], 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'architecture': {'network_class_name': 'dynamic_network_architectures.architectures.unet.ResidualEncoderUNet', 'arch_kwargs': {'n_stages': 6, 'features_per_stage': [32, 64, 128, 256, 320, 320], 'conv_op': 'torch.nn.modules.conv.Conv3d', 'kernel_sizes': [[3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'strides': [[1, 1, 1], [2, 2, 2], [2, 2, 2], [2, 2, 2], [2, 2, 2], [2, 2, 2]], 'n_blocks_per_stage': [1, 3, 4, 6, 6, 6], 'n_conv_per_stage_decoder': [1, 1, 1, 1, 1], 'conv_bias': True, 'norm_op': 'torch.nn.modules.instancenorm.InstanceNorm3d', 'norm_op_kwargs': {'eps': 1e-05, 'affine': True}, 'dropout_op': None, 'dropout_op_kwargs': None, 'nonlin': 'torch.nn.LeakyReLU', 'nonlin_kwargs': {'inplace': True}}, '_kw_requires_import': ['conv_op', 'norm_op', 'dropout_op', 'nonlin']}, 'batch_dice': False}}, 'experiment_planner_used': 'nnUNetPlannerResEncL', 'label_manager': 'LabelManager', 'foreground_intensity_properties_per_channel': {'0': {'max': 255.0, 'mean': 126.20806121826172, 'median': 126.0, 'min': 0.0, 'percentile_00_5': 0.0, 'percentile_99_5': 235.0, 'std': 43.85218048095703}}}",
46
+ "preprocessed_dataset_folder": "/mnt/raid_nvme/datasets/preprocessed/Dataset104_059plus0500p2/nnUNetPlans_3d_fullres",
47
+ "preprocessed_dataset_folder_base": "/mnt/raid_nvme/datasets/preprocessed/Dataset104_059plus0500p2",
48
+ "save_every": "50",
49
+ "torch_version": "2.9.0.dev20250822+cu128",
50
+ "unpack_dataset": "True",
51
+ "wandb": "<nnunetv2.training.nnUNetTrainer.variants.WandbWrapper.WandbWrapper object at 0x79652eaef3b0>",
52
+ "was_initialized": "True",
53
+ "weight_decay": "3e-05",
54
+ "weight_srec": "1"
55
+ }
fold_0/progress.png ADDED

Git LFS Details

  • SHA256: 95225a7bbf92a042629a66b073271315f3f2a0225d6a96edb410fe7c0ac23aaf
  • Pointer size: 132 Bytes
  • Size of remote file: 1.02 MB
fold_0/training_log_2025_8_23_02_20_27.txt ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ #######################################################################
3
+ Please cite the following paper when using nnU-Net:
4
+ Isensee, F., Jaeger, P. F., Kohl, S. A., Petersen, J., & Maier-Hein, K. H. (2021). nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nature methods, 18(2), 203-211.
5
+ #######################################################################
6
+
7
+ 2025-08-23 02:20:28.719060: do_dummy_2d_data_aug: False
8
+ 2025-08-23 02:20:28.763258: Creating new 5-fold cross-validation split...
9
+ 2025-08-23 02:20:28.768607: Desired fold for training: 0
10
+ 2025-08-23 02:20:28.768805: This split has 1481 training and 371 validation cases.
fold_0/training_log_2025_8_23_02_22_39.txt ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ #######################################################################
3
+ Please cite the following paper when using nnU-Net:
4
+ Isensee, F., Jaeger, P. F., Kohl, S. A., Petersen, J., & Maier-Hein, K. H. (2021). nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nature methods, 18(2), 203-211.
5
+ #######################################################################
6
+
7
+ 2025-08-23 02:22:39.902989: do_dummy_2d_data_aug: False
8
+ 2025-08-23 02:22:39.934233: Using splits from existing split file: /mnt/raid_nvme/datasets/preprocessed/Dataset104_059plus0500p2/splits_final.json
9
+ 2025-08-23 02:22:39.935047: The split file contains 5 splits.
10
+ 2025-08-23 02:22:39.935127: Desired fold for training: 0
11
+ 2025-08-23 02:22:39.935181: This split has 1805 training and 47 validation cases.
12
+ 2025-08-23 02:22:45.526000: Using torch.compile...
13
+
14
+ This is the configuration used by this training:
15
+ Configuration name: 3d_fullres
16
+ {'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [192, 192, 192], 'median_image_size_in_voxels': [236.0, 236.0, 236.0], 'spacing': [1.0, 1.0, 1.0], 'normalization_schemes': ['CTNormalization'], 'use_mask_for_norm': [False], 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'architecture': {'network_class_name': 'dynamic_network_architectures.architectures.unet.ResidualEncoderUNet', 'arch_kwargs': {'n_stages': 6, 'features_per_stage': [32, 64, 128, 256, 320, 320], 'conv_op': 'torch.nn.modules.conv.Conv3d', 'kernel_sizes': [[3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'strides': [[1, 1, 1], [2, 2, 2], [2, 2, 2], [2, 2, 2], [2, 2, 2], [2, 2, 2]], 'n_blocks_per_stage': [1, 3, 4, 6, 6, 6], 'n_conv_per_stage_decoder': [1, 1, 1, 1, 1], 'conv_bias': True, 'norm_op': 'torch.nn.modules.instancenorm.InstanceNorm3d', 'norm_op_kwargs': {'eps': 1e-05, 'affine': True}, 'dropout_op': None, 'dropout_op_kwargs': None, 'nonlin': 'torch.nn.LeakyReLU', 'nonlin_kwargs': {'inplace': True}}, '_kw_requires_import': ['conv_op', 'norm_op', 'dropout_op', 'nonlin']}, 'batch_dice': False}
17
+
18
+ These are the global plan.json settings:
19
+ {'dataset_name': 'Dataset104_059plus0500p2', 'plans_name': 'nnUNetResEncUNetLPlans', 'original_median_spacing_after_transp': [1.0, 1.0, 1.0], 'original_median_shape_after_transp': [236, 236, 236], 'image_reader_writer': 'Tiff3DIO', 'transpose_forward': [0, 1, 2], 'transpose_backward': [0, 1, 2], 'experiment_planner_used': 'nnUNetPlannerResEncL', 'label_manager': 'LabelManager', 'foreground_intensity_properties_per_channel': {'0': {'max': 255.0, 'mean': 126.20806121826172, 'median': 126.0, 'min': 0.0, 'percentile_00_5': 0.0, 'percentile_99_5': 235.0, 'std': 43.85218048095703}}}
20
+
21
+ 2025-08-23 02:22:47.060563: unpacking dataset...
22
+ 2025-08-23 02:23:48.445971: unpacking done...
23
+ 2025-08-23 02:23:48.469908: Unable to plot network architecture: nnUNet_compile is enabled!
24
+ 2025-08-23 02:23:48.538867:
25
+ 2025-08-23 02:23:48.539452: Epoch 0
26
+ 2025-08-23 02:23:48.543053: Current learning rate: 0.01
fold_0/training_log_2025_8_23_02_27_10.txt ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ #######################################################################
3
+ Please cite the following paper when using nnU-Net:
4
+ Isensee, F., Jaeger, P. F., Kohl, S. A., Petersen, J., & Maier-Hein, K. H. (2021). nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nature methods, 18(2), 203-211.
5
+ #######################################################################
6
+
7
+ 2025-08-23 02:27:11.308838: do_dummy_2d_data_aug: False
8
+ 2025-08-23 02:27:11.368202: Using splits from existing split file: /mnt/raid_nvme/datasets/preprocessed/Dataset104_059plus0500p2/splits_final.json
9
+ 2025-08-23 02:27:11.380205: The split file contains 5 splits.
10
+ 2025-08-23 02:27:11.380355: Desired fold for training: 0
11
+ 2025-08-23 02:27:11.380435: This split has 1805 training and 47 validation cases.
12
+ 2025-08-23 02:27:18.279987: Using torch.compile...
13
+
14
+ This is the configuration used by this training:
15
+ Configuration name: 3d_fullres
16
+ {'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [192, 192, 192], 'median_image_size_in_voxels': [236.0, 236.0, 236.0], 'spacing': [1.0, 1.0, 1.0], 'normalization_schemes': ['CTNormalization'], 'use_mask_for_norm': [False], 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'architecture': {'network_class_name': 'dynamic_network_architectures.architectures.unet.ResidualEncoderUNet', 'arch_kwargs': {'n_stages': 6, 'features_per_stage': [32, 64, 128, 256, 320, 320], 'conv_op': 'torch.nn.modules.conv.Conv3d', 'kernel_sizes': [[3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'strides': [[1, 1, 1], [2, 2, 2], [2, 2, 2], [2, 2, 2], [2, 2, 2], [2, 2, 2]], 'n_blocks_per_stage': [1, 3, 4, 6, 6, 6], 'n_conv_per_stage_decoder': [1, 1, 1, 1, 1], 'conv_bias': True, 'norm_op': 'torch.nn.modules.instancenorm.InstanceNorm3d', 'norm_op_kwargs': {'eps': 1e-05, 'affine': True}, 'dropout_op': None, 'dropout_op_kwargs': None, 'nonlin': 'torch.nn.LeakyReLU', 'nonlin_kwargs': {'inplace': True}}, '_kw_requires_import': ['conv_op', 'norm_op', 'dropout_op', 'nonlin']}, 'batch_dice': False}
17
+
18
+ These are the global plan.json settings:
19
+ {'dataset_name': 'Dataset104_059plus0500p2', 'plans_name': 'nnUNetResEncUNetLPlans', 'original_median_spacing_after_transp': [1.0, 1.0, 1.0], 'original_median_shape_after_transp': [236, 236, 236], 'image_reader_writer': 'Tiff3DIO', 'transpose_forward': [0, 1, 2], 'transpose_backward': [0, 1, 2], 'experiment_planner_used': 'nnUNetPlannerResEncL', 'label_manager': 'LabelManager', 'foreground_intensity_properties_per_channel': {'0': {'max': 255.0, 'mean': 126.20806121826172, 'median': 126.0, 'min': 0.0, 'percentile_00_5': 0.0, 'percentile_99_5': 235.0, 'std': 43.85218048095703}}}
20
+
21
+ 2025-08-23 02:27:20.052681: unpacking dataset...
22
+ 2025-08-23 02:27:25.110170: unpacking done...
23
+ 2025-08-23 02:27:25.111650: Unable to plot network architecture: nnUNet_compile is enabled!
24
+ 2025-08-23 02:27:25.117785:
25
+ 2025-08-23 02:27:25.118071: Epoch 0
26
+ 2025-08-23 02:27:25.118300: Current learning rate: 0.01
fold_0/training_log_2025_8_23_02_34_17.txt ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ #######################################################################
3
+ Please cite the following paper when using nnU-Net:
4
+ Isensee, F., Jaeger, P. F., Kohl, S. A., Petersen, J., & Maier-Hein, K. H. (2021). nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nature methods, 18(2), 203-211.
5
+ #######################################################################
6
+
7
+ 2025-08-23 02:34:18.589556: do_dummy_2d_data_aug: False
8
+ 2025-08-23 02:34:18.623172: Using splits from existing split file: /mnt/raid_nvme/datasets/preprocessed/Dataset104_059plus0500p2/splits_final.json
9
+ 2025-08-23 02:34:18.624004: The split file contains 5 splits.
10
+ 2025-08-23 02:34:18.624081: Desired fold for training: 0
11
+ 2025-08-23 02:34:18.624135: This split has 1805 training and 47 validation cases.
12
+ 2025-08-23 02:34:28.195994: Using torch.compile...
13
+
14
+ This is the configuration used by this training:
15
+ Configuration name: 3d_fullres
16
+ {'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [192, 192, 192], 'median_image_size_in_voxels': [236.0, 236.0, 236.0], 'spacing': [1.0, 1.0, 1.0], 'normalization_schemes': ['CTNormalization'], 'use_mask_for_norm': [False], 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'architecture': {'network_class_name': 'dynamic_network_architectures.architectures.unet.ResidualEncoderUNet', 'arch_kwargs': {'n_stages': 6, 'features_per_stage': [32, 64, 128, 256, 320, 320], 'conv_op': 'torch.nn.modules.conv.Conv3d', 'kernel_sizes': [[3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'strides': [[1, 1, 1], [2, 2, 2], [2, 2, 2], [2, 2, 2], [2, 2, 2], [2, 2, 2]], 'n_blocks_per_stage': [1, 3, 4, 6, 6, 6], 'n_conv_per_stage_decoder': [1, 1, 1, 1, 1], 'conv_bias': True, 'norm_op': 'torch.nn.modules.instancenorm.InstanceNorm3d', 'norm_op_kwargs': {'eps': 1e-05, 'affine': True}, 'dropout_op': None, 'dropout_op_kwargs': None, 'nonlin': 'torch.nn.LeakyReLU', 'nonlin_kwargs': {'inplace': True}}, '_kw_requires_import': ['conv_op', 'norm_op', 'dropout_op', 'nonlin']}, 'batch_dice': False}
17
+
18
+ These are the global plan.json settings:
19
+ {'dataset_name': 'Dataset104_059plus0500p2', 'plans_name': 'nnUNetResEncUNetLPlans', 'original_median_spacing_after_transp': [1.0, 1.0, 1.0], 'original_median_shape_after_transp': [236, 236, 236], 'image_reader_writer': 'Tiff3DIO', 'transpose_forward': [0, 1, 2], 'transpose_backward': [0, 1, 2], 'experiment_planner_used': 'nnUNetPlannerResEncL', 'label_manager': 'LabelManager', 'foreground_intensity_properties_per_channel': {'0': {'max': 255.0, 'mean': 126.20806121826172, 'median': 126.0, 'min': 0.0, 'percentile_00_5': 0.0, 'percentile_99_5': 235.0, 'std': 43.85218048095703}}}
20
+
21
+ 2025-08-23 02:34:30.899173: unpacking dataset...
fold_0/training_log_2025_8_23_02_35_03.txt ADDED
The diff for this file is too large to render. See raw diff
 
plans.json ADDED
@@ -0,0 +1,356 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_name": "Dataset104_059plus0500p2",
3
+ "plans_name": "nnUNetResEncUNetLPlans",
4
+ "original_median_spacing_after_transp": [
5
+ 1.0,
6
+ 1.0,
7
+ 1.0
8
+ ],
9
+ "original_median_shape_after_transp": [
10
+ 236,
11
+ 236,
12
+ 236
13
+ ],
14
+ "image_reader_writer": "Tiff3DIO",
15
+ "transpose_forward": [
16
+ 0,
17
+ 1,
18
+ 2
19
+ ],
20
+ "transpose_backward": [
21
+ 0,
22
+ 1,
23
+ 2
24
+ ],
25
+ "configurations": {
26
+ "2d": {
27
+ "data_identifier": "nnUNetPlans_2d",
28
+ "preprocessor_name": "DefaultPreprocessor",
29
+ "batch_size": 141,
30
+ "patch_size": [
31
+ 256,
32
+ 256
33
+ ],
34
+ "median_image_size_in_voxels": [
35
+ 236.0,
36
+ 236.0
37
+ ],
38
+ "spacing": [
39
+ 1.0,
40
+ 1.0
41
+ ],
42
+ "normalization_schemes": [
43
+ "CTNormalization"
44
+ ],
45
+ "use_mask_for_norm": [
46
+ false
47
+ ],
48
+ "resampling_fn_data": "resample_data_or_seg_to_shape",
49
+ "resampling_fn_seg": "resample_data_or_seg_to_shape",
50
+ "resampling_fn_data_kwargs": {
51
+ "is_seg": false,
52
+ "order": 3,
53
+ "order_z": 0,
54
+ "force_separate_z": null
55
+ },
56
+ "resampling_fn_seg_kwargs": {
57
+ "is_seg": true,
58
+ "order": 1,
59
+ "order_z": 0,
60
+ "force_separate_z": null
61
+ },
62
+ "resampling_fn_probabilities": "resample_data_or_seg_to_shape",
63
+ "resampling_fn_probabilities_kwargs": {
64
+ "is_seg": false,
65
+ "order": 1,
66
+ "order_z": 0,
67
+ "force_separate_z": null
68
+ },
69
+ "architecture": {
70
+ "network_class_name": "dynamic_network_architectures.architectures.unet.ResidualEncoderUNet",
71
+ "arch_kwargs": {
72
+ "n_stages": 7,
73
+ "features_per_stage": [
74
+ 32,
75
+ 64,
76
+ 128,
77
+ 256,
78
+ 512,
79
+ 512,
80
+ 512
81
+ ],
82
+ "conv_op": "torch.nn.modules.conv.Conv2d",
83
+ "kernel_sizes": [
84
+ [
85
+ 3,
86
+ 3
87
+ ],
88
+ [
89
+ 3,
90
+ 3
91
+ ],
92
+ [
93
+ 3,
94
+ 3
95
+ ],
96
+ [
97
+ 3,
98
+ 3
99
+ ],
100
+ [
101
+ 3,
102
+ 3
103
+ ],
104
+ [
105
+ 3,
106
+ 3
107
+ ],
108
+ [
109
+ 3,
110
+ 3
111
+ ]
112
+ ],
113
+ "strides": [
114
+ [
115
+ 1,
116
+ 1
117
+ ],
118
+ [
119
+ 2,
120
+ 2
121
+ ],
122
+ [
123
+ 2,
124
+ 2
125
+ ],
126
+ [
127
+ 2,
128
+ 2
129
+ ],
130
+ [
131
+ 2,
132
+ 2
133
+ ],
134
+ [
135
+ 2,
136
+ 2
137
+ ],
138
+ [
139
+ 2,
140
+ 2
141
+ ]
142
+ ],
143
+ "n_blocks_per_stage": [
144
+ 1,
145
+ 3,
146
+ 4,
147
+ 6,
148
+ 6,
149
+ 6,
150
+ 6
151
+ ],
152
+ "n_conv_per_stage_decoder": [
153
+ 1,
154
+ 1,
155
+ 1,
156
+ 1,
157
+ 1,
158
+ 1
159
+ ],
160
+ "conv_bias": true,
161
+ "norm_op": "torch.nn.modules.instancenorm.InstanceNorm2d",
162
+ "norm_op_kwargs": {
163
+ "eps": 1e-05,
164
+ "affine": true
165
+ },
166
+ "dropout_op": null,
167
+ "dropout_op_kwargs": null,
168
+ "nonlin": "torch.nn.LeakyReLU",
169
+ "nonlin_kwargs": {
170
+ "inplace": true
171
+ }
172
+ },
173
+ "_kw_requires_import": [
174
+ "conv_op",
175
+ "norm_op",
176
+ "dropout_op",
177
+ "nonlin"
178
+ ]
179
+ },
180
+ "batch_dice": true
181
+ },
182
+ "3d_fullres": {
183
+ "data_identifier": "nnUNetPlans_3d_fullres",
184
+ "preprocessor_name": "DefaultPreprocessor",
185
+ "batch_size": 2,
186
+ "patch_size": [
187
+ 192,
188
+ 192,
189
+ 192
190
+ ],
191
+ "median_image_size_in_voxels": [
192
+ 236.0,
193
+ 236.0,
194
+ 236.0
195
+ ],
196
+ "spacing": [
197
+ 1.0,
198
+ 1.0,
199
+ 1.0
200
+ ],
201
+ "normalization_schemes": [
202
+ "CTNormalization"
203
+ ],
204
+ "use_mask_for_norm": [
205
+ false
206
+ ],
207
+ "resampling_fn_data": "resample_data_or_seg_to_shape",
208
+ "resampling_fn_seg": "resample_data_or_seg_to_shape",
209
+ "resampling_fn_data_kwargs": {
210
+ "is_seg": false,
211
+ "order": 3,
212
+ "order_z": 0,
213
+ "force_separate_z": null
214
+ },
215
+ "resampling_fn_seg_kwargs": {
216
+ "is_seg": true,
217
+ "order": 1,
218
+ "order_z": 0,
219
+ "force_separate_z": null
220
+ },
221
+ "resampling_fn_probabilities": "resample_data_or_seg_to_shape",
222
+ "resampling_fn_probabilities_kwargs": {
223
+ "is_seg": false,
224
+ "order": 1,
225
+ "order_z": 0,
226
+ "force_separate_z": null
227
+ },
228
+ "architecture": {
229
+ "network_class_name": "dynamic_network_architectures.architectures.unet.ResidualEncoderUNet",
230
+ "arch_kwargs": {
231
+ "n_stages": 6,
232
+ "features_per_stage": [
233
+ 32,
234
+ 64,
235
+ 128,
236
+ 256,
237
+ 320,
238
+ 320
239
+ ],
240
+ "conv_op": "torch.nn.modules.conv.Conv3d",
241
+ "kernel_sizes": [
242
+ [
243
+ 3,
244
+ 3,
245
+ 3
246
+ ],
247
+ [
248
+ 3,
249
+ 3,
250
+ 3
251
+ ],
252
+ [
253
+ 3,
254
+ 3,
255
+ 3
256
+ ],
257
+ [
258
+ 3,
259
+ 3,
260
+ 3
261
+ ],
262
+ [
263
+ 3,
264
+ 3,
265
+ 3
266
+ ],
267
+ [
268
+ 3,
269
+ 3,
270
+ 3
271
+ ]
272
+ ],
273
+ "strides": [
274
+ [
275
+ 1,
276
+ 1,
277
+ 1
278
+ ],
279
+ [
280
+ 2,
281
+ 2,
282
+ 2
283
+ ],
284
+ [
285
+ 2,
286
+ 2,
287
+ 2
288
+ ],
289
+ [
290
+ 2,
291
+ 2,
292
+ 2
293
+ ],
294
+ [
295
+ 2,
296
+ 2,
297
+ 2
298
+ ],
299
+ [
300
+ 2,
301
+ 2,
302
+ 2
303
+ ]
304
+ ],
305
+ "n_blocks_per_stage": [
306
+ 1,
307
+ 3,
308
+ 4,
309
+ 6,
310
+ 6,
311
+ 6
312
+ ],
313
+ "n_conv_per_stage_decoder": [
314
+ 1,
315
+ 1,
316
+ 1,
317
+ 1,
318
+ 1
319
+ ],
320
+ "conv_bias": true,
321
+ "norm_op": "torch.nn.modules.instancenorm.InstanceNorm3d",
322
+ "norm_op_kwargs": {
323
+ "eps": 1e-05,
324
+ "affine": true
325
+ },
326
+ "dropout_op": null,
327
+ "dropout_op_kwargs": null,
328
+ "nonlin": "torch.nn.LeakyReLU",
329
+ "nonlin_kwargs": {
330
+ "inplace": true
331
+ }
332
+ },
333
+ "_kw_requires_import": [
334
+ "conv_op",
335
+ "norm_op",
336
+ "dropout_op",
337
+ "nonlin"
338
+ ]
339
+ },
340
+ "batch_dice": false
341
+ }
342
+ },
343
+ "experiment_planner_used": "nnUNetPlannerResEncL",
344
+ "label_manager": "LabelManager",
345
+ "foreground_intensity_properties_per_channel": {
346
+ "0": {
347
+ "max": 255.0,
348
+ "mean": 126.20806121826172,
349
+ "median": 126.0,
350
+ "min": 0.0,
351
+ "percentile_00_5": 0.0,
352
+ "percentile_99_5": 235.0,
353
+ "std": 43.85218048095703
354
+ }
355
+ }
356
+ }