FelixzeroSun commited on
Commit
b6319c5
·
verified ·
1 Parent(s): 6db5668

Upload folder using huggingface_hub

Browse files
Files changed (22) hide show
  1. .gitattributes +3 -0
  2. Dataset540/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/dataset.json +14 -0
  3. Dataset540/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/dataset_fingerprint.json +3108 -0
  4. Dataset540/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/checkpoint_final.pth +3 -0
  5. Dataset540/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/debug.json +0 -0
  6. Dataset540/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/progress.png +3 -0
  7. Dataset540/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/validation/summary.json +52 -0
  8. Dataset540/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/plans.json +521 -0
  9. Dataset542/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/dataset.json +14 -0
  10. Dataset542/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/dataset_fingerprint.json +3268 -0
  11. Dataset542/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/checkpoint_final.pth +3 -0
  12. Dataset542/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/debug.json +58 -0
  13. Dataset542/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/progress.png +3 -0
  14. Dataset542/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/validation/summary.json +52 -0
  15. Dataset542/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/plans.json +356 -0
  16. Dataset544/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/dataset.json +14 -0
  17. Dataset544/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/dataset_fingerprint.json +3228 -0
  18. Dataset544/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/checkpoint_final.pth +3 -0
  19. Dataset544/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/debug.json +0 -0
  20. Dataset544/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/progress.png +3 -0
  21. Dataset544/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/validation/summary.json +52 -0
  22. Dataset544/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/plans.json +521 -0
.gitattributes CHANGED
@@ -33,3 +33,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Dataset540/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/progress.png filter=lfs diff=lfs merge=lfs -text
37
+ Dataset542/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/progress.png filter=lfs diff=lfs merge=lfs -text
38
+ Dataset544/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/progress.png filter=lfs diff=lfs merge=lfs -text
Dataset540/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/dataset.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "labels": {
3
+ "label_001": "1",
4
+ "background": 0
5
+ },
6
+ "channel_names": {
7
+ "0": "CT_zscore_synthrad"
8
+ },
9
+ "numTraining": 309,
10
+ "file_ending": ".nii.gz",
11
+ "test_set": [
12
+ "case_0000"
13
+ ]
14
+ }
Dataset540/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/dataset_fingerprint.json ADDED
@@ -0,0 +1,3108 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "foreground_intensity_properties_per_channel": {
3
+ "0": {
4
+ "max": 3071.0,
5
+ "mean": -239.3903045654297,
6
+ "median": -185.0,
7
+ "min": -1024.0,
8
+ "percentile_00_5": -1024.0,
9
+ "percentile_99_5": 479.0,
10
+ "std": 328.2787170410156
11
+ }
12
+ },
13
+ "median_relative_size_after_cropping": 1.0,
14
+ "shapes_after_crop": [
15
+ [
16
+ 88,
17
+ 309,
18
+ 310
19
+ ],
20
+ [
21
+ 88,
22
+ 309,
23
+ 289
24
+ ],
25
+ [
26
+ 87,
27
+ 310,
28
+ 312
29
+ ],
30
+ [
31
+ 86,
32
+ 308,
33
+ 315
34
+ ],
35
+ [
36
+ 87,
37
+ 313,
38
+ 311
39
+ ],
40
+ [
41
+ 82,
42
+ 309,
43
+ 308
44
+ ],
45
+ [
46
+ 87,
47
+ 310,
48
+ 311
49
+ ],
50
+ [
51
+ 88,
52
+ 308,
53
+ 309
54
+ ],
55
+ [
56
+ 82,
57
+ 310,
58
+ 308
59
+ ],
60
+ [
61
+ 81,
62
+ 448,
63
+ 449
64
+ ],
65
+ [
66
+ 85,
67
+ 317,
68
+ 309
69
+ ],
70
+ [
71
+ 86,
72
+ 311,
73
+ 314
74
+ ],
75
+ [
76
+ 76,
77
+ 459,
78
+ 455
79
+ ],
80
+ [
81
+ 88,
82
+ 308,
83
+ 310
84
+ ],
85
+ [
86
+ 50,
87
+ 510,
88
+ 510
89
+ ],
90
+ [
91
+ 86,
92
+ 310,
93
+ 311
94
+ ],
95
+ [
96
+ 78,
97
+ 311,
98
+ 309
99
+ ],
100
+ [
101
+ 87,
102
+ 290,
103
+ 291
104
+ ],
105
+ [
106
+ 87,
107
+ 309,
108
+ 308
109
+ ],
110
+ [
111
+ 94,
112
+ 393,
113
+ 393
114
+ ],
115
+ [
116
+ 87,
117
+ 311,
118
+ 308
119
+ ],
120
+ [
121
+ 87,
122
+ 314,
123
+ 309
124
+ ],
125
+ [
126
+ 87,
127
+ 309,
128
+ 309
129
+ ],
130
+ [
131
+ 100,
132
+ 368,
133
+ 370
134
+ ],
135
+ [
136
+ 88,
137
+ 308,
138
+ 307
139
+ ],
140
+ [
141
+ 100,
142
+ 498,
143
+ 499
144
+ ],
145
+ [
146
+ 87,
147
+ 308,
148
+ 313
149
+ ],
150
+ [
151
+ 87,
152
+ 313,
153
+ 310
154
+ ],
155
+ [
156
+ 86,
157
+ 314,
158
+ 311
159
+ ],
160
+ [
161
+ 88,
162
+ 309,
163
+ 309
164
+ ],
165
+ [
166
+ 87,
167
+ 312,
168
+ 309
169
+ ],
170
+ [
171
+ 80,
172
+ 308,
173
+ 313
174
+ ],
175
+ [
176
+ 87,
177
+ 309,
178
+ 310
179
+ ],
180
+ [
181
+ 86,
182
+ 309,
183
+ 308
184
+ ],
185
+ [
186
+ 88,
187
+ 308,
188
+ 309
189
+ ],
190
+ [
191
+ 88,
192
+ 307,
193
+ 308
194
+ ],
195
+ [
196
+ 79,
197
+ 311,
198
+ 310
199
+ ],
200
+ [
201
+ 115,
202
+ 560,
203
+ 560
204
+ ],
205
+ [
206
+ 79,
207
+ 450,
208
+ 451
209
+ ],
210
+ [
211
+ 85,
212
+ 308,
213
+ 309
214
+ ],
215
+ [
216
+ 87,
217
+ 308,
218
+ 307
219
+ ],
220
+ [
221
+ 88,
222
+ 310,
223
+ 305
224
+ ],
225
+ [
226
+ 86,
227
+ 313,
228
+ 311
229
+ ],
230
+ [
231
+ 87,
232
+ 310,
233
+ 308
234
+ ],
235
+ [
236
+ 85,
237
+ 316,
238
+ 308
239
+ ],
240
+ [
241
+ 73,
242
+ 314,
243
+ 313
244
+ ],
245
+ [
246
+ 86,
247
+ 312,
248
+ 311
249
+ ],
250
+ [
251
+ 87,
252
+ 309,
253
+ 310
254
+ ],
255
+ [
256
+ 86,
257
+ 311,
258
+ 308
259
+ ],
260
+ [
261
+ 87,
262
+ 306,
263
+ 307
264
+ ],
265
+ [
266
+ 82,
267
+ 308,
268
+ 307
269
+ ],
270
+ [
271
+ 87,
272
+ 310,
273
+ 308
274
+ ],
275
+ [
276
+ 80,
277
+ 288,
278
+ 303
279
+ ],
280
+ [
281
+ 81,
282
+ 450,
283
+ 428
284
+ ],
285
+ [
286
+ 86,
287
+ 308,
288
+ 311
289
+ ],
290
+ [
291
+ 87,
292
+ 311,
293
+ 310
294
+ ],
295
+ [
296
+ 86,
297
+ 310,
298
+ 317
299
+ ],
300
+ [
301
+ 87,
302
+ 311,
303
+ 310
304
+ ],
305
+ [
306
+ 87,
307
+ 309,
308
+ 307
309
+ ],
310
+ [
311
+ 77,
312
+ 453,
313
+ 456
314
+ ],
315
+ [
316
+ 100,
317
+ 374,
318
+ 405
319
+ ],
320
+ [
321
+ 80,
322
+ 310,
323
+ 312
324
+ ],
325
+ [
326
+ 87,
327
+ 308,
328
+ 308
329
+ ],
330
+ [
331
+ 81,
332
+ 430,
333
+ 449
334
+ ],
335
+ [
336
+ 78,
337
+ 452,
338
+ 433
339
+ ],
340
+ [
341
+ 80,
342
+ 454,
343
+ 450
344
+ ],
345
+ [
346
+ 82,
347
+ 450,
348
+ 451
349
+ ],
350
+ [
351
+ 80,
352
+ 432,
353
+ 450
354
+ ],
355
+ [
356
+ 80,
357
+ 448,
358
+ 452
359
+ ],
360
+ [
361
+ 81,
362
+ 451,
363
+ 450
364
+ ],
365
+ [
366
+ 80,
367
+ 451,
368
+ 449
369
+ ],
370
+ [
371
+ 80,
372
+ 449,
373
+ 453
374
+ ],
375
+ [
376
+ 81,
377
+ 448,
378
+ 412
379
+ ],
380
+ [
381
+ 80,
382
+ 449,
383
+ 433
384
+ ],
385
+ [
386
+ 81,
387
+ 449,
388
+ 449
389
+ ],
390
+ [
391
+ 81,
392
+ 452,
393
+ 421
394
+ ],
395
+ [
396
+ 77,
397
+ 458,
398
+ 450
399
+ ],
400
+ [
401
+ 81,
402
+ 450,
403
+ 451
404
+ ],
405
+ [
406
+ 73,
407
+ 453,
408
+ 412
409
+ ],
410
+ [
411
+ 77,
412
+ 452,
413
+ 417
414
+ ],
415
+ [
416
+ 80,
417
+ 449,
418
+ 431
419
+ ],
420
+ [
421
+ 81,
422
+ 451,
423
+ 442
424
+ ],
425
+ [
426
+ 81,
427
+ 450,
428
+ 413
429
+ ],
430
+ [
431
+ 81,
432
+ 450,
433
+ 450
434
+ ],
435
+ [
436
+ 79,
437
+ 451,
438
+ 454
439
+ ],
440
+ [
441
+ 80,
442
+ 449,
443
+ 453
444
+ ],
445
+ [
446
+ 80,
447
+ 452,
448
+ 431
449
+ ],
450
+ [
451
+ 80,
452
+ 455,
453
+ 451
454
+ ],
455
+ [
456
+ 56,
457
+ 451,
458
+ 450
459
+ ],
460
+ [
461
+ 81,
462
+ 452,
463
+ 448
464
+ ],
465
+ [
466
+ 80,
467
+ 452,
468
+ 450
469
+ ],
470
+ [
471
+ 74,
472
+ 452,
473
+ 418
474
+ ],
475
+ [
476
+ 68,
477
+ 454,
478
+ 449
479
+ ],
480
+ [
481
+ 51,
482
+ 449,
483
+ 427
484
+ ],
485
+ [
486
+ 70,
487
+ 456,
488
+ 393
489
+ ],
490
+ [
491
+ 80,
492
+ 454,
493
+ 450
494
+ ],
495
+ [
496
+ 78,
497
+ 439,
498
+ 450
499
+ ],
500
+ [
501
+ 82,
502
+ 449,
503
+ 431
504
+ ],
505
+ [
506
+ 79,
507
+ 451,
508
+ 445
509
+ ],
510
+ [
511
+ 80,
512
+ 430,
513
+ 455
514
+ ],
515
+ [
516
+ 82,
517
+ 449,
518
+ 447
519
+ ],
520
+ [
521
+ 82,
522
+ 400,
523
+ 449
524
+ ],
525
+ [
526
+ 81,
527
+ 402,
528
+ 449
529
+ ],
530
+ [
531
+ 75,
532
+ 451,
533
+ 449
534
+ ],
535
+ [
536
+ 81,
537
+ 452,
538
+ 448
539
+ ],
540
+ [
541
+ 80,
542
+ 452,
543
+ 452
544
+ ],
545
+ [
546
+ 56,
547
+ 451,
548
+ 450
549
+ ],
550
+ [
551
+ 70,
552
+ 455,
553
+ 393
554
+ ],
555
+ [
556
+ 80,
557
+ 451,
558
+ 452
559
+ ],
560
+ [
561
+ 71,
562
+ 451,
563
+ 449
564
+ ],
565
+ [
566
+ 80,
567
+ 421,
568
+ 449
569
+ ],
570
+ [
571
+ 80,
572
+ 448,
573
+ 434
574
+ ],
575
+ [
576
+ 73,
577
+ 454,
578
+ 393
579
+ ],
580
+ [
581
+ 81,
582
+ 449,
583
+ 452
584
+ ],
585
+ [
586
+ 75,
587
+ 452,
588
+ 434
589
+ ],
590
+ [
591
+ 74,
592
+ 449,
593
+ 449
594
+ ],
595
+ [
596
+ 82,
597
+ 427,
598
+ 396
599
+ ],
600
+ [
601
+ 75,
602
+ 398,
603
+ 398
604
+ ],
605
+ [
606
+ 80,
607
+ 454,
608
+ 430
609
+ ],
610
+ [
611
+ 78,
612
+ 451,
613
+ 456
614
+ ],
615
+ [
616
+ 81,
617
+ 450,
618
+ 451
619
+ ],
620
+ [
621
+ 80,
622
+ 449,
623
+ 431
624
+ ],
625
+ [
626
+ 79,
627
+ 448,
628
+ 454
629
+ ],
630
+ [
631
+ 80,
632
+ 451,
633
+ 452
634
+ ],
635
+ [
636
+ 79,
637
+ 452,
638
+ 454
639
+ ],
640
+ [
641
+ 79,
642
+ 450,
643
+ 455
644
+ ],
645
+ [
646
+ 82,
647
+ 449,
648
+ 447
649
+ ],
650
+ [
651
+ 81,
652
+ 450,
653
+ 448
654
+ ],
655
+ [
656
+ 81,
657
+ 451,
658
+ 451
659
+ ],
660
+ [
661
+ 81,
662
+ 449,
663
+ 449
664
+ ],
665
+ [
666
+ 79,
667
+ 408,
668
+ 457
669
+ ],
670
+ [
671
+ 81,
672
+ 449,
673
+ 452
674
+ ],
675
+ [
676
+ 80,
677
+ 451,
678
+ 453
679
+ ],
680
+ [
681
+ 81,
682
+ 448,
683
+ 450
684
+ ],
685
+ [
686
+ 77,
687
+ 449,
688
+ 457
689
+ ],
690
+ [
691
+ 81,
692
+ 450,
693
+ 451
694
+ ],
695
+ [
696
+ 81,
697
+ 450,
698
+ 450
699
+ ],
700
+ [
701
+ 87,
702
+ 310,
703
+ 310
704
+ ],
705
+ [
706
+ 81,
707
+ 449,
708
+ 451
709
+ ],
710
+ [
711
+ 80,
712
+ 453,
713
+ 450
714
+ ],
715
+ [
716
+ 80,
717
+ 448,
718
+ 452
719
+ ],
720
+ [
721
+ 78,
722
+ 451,
723
+ 457
724
+ ],
725
+ [
726
+ 81,
727
+ 430,
728
+ 391
729
+ ],
730
+ [
731
+ 78,
732
+ 449,
733
+ 455
734
+ ],
735
+ [
736
+ 81,
737
+ 452,
738
+ 412
739
+ ],
740
+ [
741
+ 81,
742
+ 449,
743
+ 449
744
+ ],
745
+ [
746
+ 80,
747
+ 444,
748
+ 432
749
+ ],
750
+ [
751
+ 81,
752
+ 449,
753
+ 451
754
+ ],
755
+ [
756
+ 81,
757
+ 432,
758
+ 450
759
+ ],
760
+ [
761
+ 81,
762
+ 448,
763
+ 453
764
+ ],
765
+ [
766
+ 79,
767
+ 455,
768
+ 449
769
+ ],
770
+ [
771
+ 81,
772
+ 449,
773
+ 450
774
+ ],
775
+ [
776
+ 81,
777
+ 450,
778
+ 385
779
+ ],
780
+ [
781
+ 81,
782
+ 434,
783
+ 436
784
+ ],
785
+ [
786
+ 82,
787
+ 416,
788
+ 449
789
+ ],
790
+ [
791
+ 82,
792
+ 442,
793
+ 450
794
+ ],
795
+ [
796
+ 81,
797
+ 439,
798
+ 451
799
+ ],
800
+ [
801
+ 81,
802
+ 447,
803
+ 450
804
+ ],
805
+ [
806
+ 81,
807
+ 415,
808
+ 450
809
+ ],
810
+ [
811
+ 79,
812
+ 450,
813
+ 453
814
+ ],
815
+ [
816
+ 76,
817
+ 430,
818
+ 450
819
+ ],
820
+ [
821
+ 81,
822
+ 450,
823
+ 450
824
+ ],
825
+ [
826
+ 81,
827
+ 448,
828
+ 450
829
+ ],
830
+ [
831
+ 81,
832
+ 448,
833
+ 422
834
+ ],
835
+ [
836
+ 82,
837
+ 448,
838
+ 448
839
+ ],
840
+ [
841
+ 79,
842
+ 454,
843
+ 429
844
+ ],
845
+ [
846
+ 81,
847
+ 451,
848
+ 420
849
+ ],
850
+ [
851
+ 75,
852
+ 449,
853
+ 462
854
+ ],
855
+ [
856
+ 82,
857
+ 450,
858
+ 450
859
+ ],
860
+ [
861
+ 81,
862
+ 450,
863
+ 451
864
+ ],
865
+ [
866
+ 82,
867
+ 448,
868
+ 451
869
+ ],
870
+ [
871
+ 82,
872
+ 448,
873
+ 451
874
+ ],
875
+ [
876
+ 75,
877
+ 449,
878
+ 448
879
+ ],
880
+ [
881
+ 80,
882
+ 451,
883
+ 454
884
+ ],
885
+ [
886
+ 80,
887
+ 449,
888
+ 453
889
+ ],
890
+ [
891
+ 81,
892
+ 449,
893
+ 452
894
+ ],
895
+ [
896
+ 82,
897
+ 450,
898
+ 450
899
+ ],
900
+ [
901
+ 78,
902
+ 450,
903
+ 428
904
+ ],
905
+ [
906
+ 80,
907
+ 442,
908
+ 453
909
+ ],
910
+ [
911
+ 81,
912
+ 440,
913
+ 449
914
+ ],
915
+ [
916
+ 82,
917
+ 441,
918
+ 449
919
+ ],
920
+ [
921
+ 82,
922
+ 450,
923
+ 411
924
+ ],
925
+ [
926
+ 81,
927
+ 430,
928
+ 449
929
+ ],
930
+ [
931
+ 82,
932
+ 413,
933
+ 450
934
+ ],
935
+ [
936
+ 81,
937
+ 448,
938
+ 404
939
+ ],
940
+ [
941
+ 81,
942
+ 450,
943
+ 449
944
+ ],
945
+ [
946
+ 81,
947
+ 449,
948
+ 451
949
+ ],
950
+ [
951
+ 81,
952
+ 450,
953
+ 377
954
+ ],
955
+ [
956
+ 80,
957
+ 455,
958
+ 453
959
+ ],
960
+ [
961
+ 80,
962
+ 454,
963
+ 448
964
+ ],
965
+ [
966
+ 82,
967
+ 447,
968
+ 449
969
+ ],
970
+ [
971
+ 99,
972
+ 452,
973
+ 435
974
+ ],
975
+ [
976
+ 68,
977
+ 450,
978
+ 430
979
+ ],
980
+ [
981
+ 99,
982
+ 314,
983
+ 310
984
+ ],
985
+ [
986
+ 99,
987
+ 453,
988
+ 452
989
+ ],
990
+ [
991
+ 99,
992
+ 312,
993
+ 311
994
+ ],
995
+ [
996
+ 75,
997
+ 459,
998
+ 414
999
+ ],
1000
+ [
1001
+ 69,
1002
+ 432,
1003
+ 449
1004
+ ],
1005
+ [
1006
+ 99,
1007
+ 452,
1008
+ 434
1009
+ ],
1010
+ [
1011
+ 99,
1012
+ 456,
1013
+ 404
1014
+ ],
1015
+ [
1016
+ 71,
1017
+ 453,
1018
+ 400
1019
+ ],
1020
+ [
1021
+ 98,
1022
+ 310,
1023
+ 311
1024
+ ],
1025
+ [
1026
+ 99,
1027
+ 311,
1028
+ 309
1029
+ ],
1030
+ [
1031
+ 99,
1032
+ 309,
1033
+ 310
1034
+ ],
1035
+ [
1036
+ 99,
1037
+ 315,
1038
+ 308
1039
+ ],
1040
+ [
1041
+ 98,
1042
+ 308,
1043
+ 308
1044
+ ],
1045
+ [
1046
+ 99,
1047
+ 450,
1048
+ 453
1049
+ ],
1050
+ [
1051
+ 99,
1052
+ 383,
1053
+ 349
1054
+ ],
1055
+ [
1056
+ 99,
1057
+ 314,
1058
+ 310
1059
+ ],
1060
+ [
1061
+ 70,
1062
+ 452,
1063
+ 451
1064
+ ],
1065
+ [
1066
+ 69,
1067
+ 452,
1068
+ 447
1069
+ ],
1070
+ [
1071
+ 99,
1072
+ 452,
1073
+ 455
1074
+ ],
1075
+ [
1076
+ 68,
1077
+ 433,
1078
+ 447
1079
+ ],
1080
+ [
1081
+ 100,
1082
+ 308,
1083
+ 318
1084
+ ],
1085
+ [
1086
+ 98,
1087
+ 319,
1088
+ 309
1089
+ ],
1090
+ [
1091
+ 69,
1092
+ 452,
1093
+ 448
1094
+ ],
1095
+ [
1096
+ 99,
1097
+ 435,
1098
+ 454
1099
+ ],
1100
+ [
1101
+ 91,
1102
+ 315,
1103
+ 310
1104
+ ],
1105
+ [
1106
+ 99,
1107
+ 450,
1108
+ 424
1109
+ ],
1110
+ [
1111
+ 99,
1112
+ 452,
1113
+ 452
1114
+ ],
1115
+ [
1116
+ 98,
1117
+ 450,
1118
+ 448
1119
+ ],
1120
+ [
1121
+ 100,
1122
+ 455,
1123
+ 451
1124
+ ],
1125
+ [
1126
+ 91,
1127
+ 435,
1128
+ 452
1129
+ ],
1130
+ [
1131
+ 99,
1132
+ 313,
1133
+ 315
1134
+ ],
1135
+ [
1136
+ 99,
1137
+ 308,
1138
+ 310
1139
+ ],
1140
+ [
1141
+ 70,
1142
+ 450,
1143
+ 450
1144
+ ],
1145
+ [
1146
+ 99,
1147
+ 448,
1148
+ 448
1149
+ ],
1150
+ [
1151
+ 99,
1152
+ 309,
1153
+ 308
1154
+ ],
1155
+ [
1156
+ 79,
1157
+ 311,
1158
+ 307
1159
+ ],
1160
+ [
1161
+ 94,
1162
+ 312,
1163
+ 312
1164
+ ],
1165
+ [
1166
+ 99,
1167
+ 326,
1168
+ 315
1169
+ ],
1170
+ [
1171
+ 74,
1172
+ 467,
1173
+ 497
1174
+ ],
1175
+ [
1176
+ 71,
1177
+ 465,
1178
+ 499
1179
+ ],
1180
+ [
1181
+ 74,
1182
+ 467,
1183
+ 499
1184
+ ],
1185
+ [
1186
+ 73,
1187
+ 472,
1188
+ 499
1189
+ ],
1190
+ [
1191
+ 98,
1192
+ 448,
1193
+ 448
1194
+ ],
1195
+ [
1196
+ 100,
1197
+ 314,
1198
+ 316
1199
+ ],
1200
+ [
1201
+ 98,
1202
+ 431,
1203
+ 417
1204
+ ],
1205
+ [
1206
+ 100,
1207
+ 315,
1208
+ 309
1209
+ ],
1210
+ [
1211
+ 98,
1212
+ 307,
1213
+ 308
1214
+ ],
1215
+ [
1216
+ 98,
1217
+ 312,
1218
+ 319
1219
+ ],
1220
+ [
1221
+ 98,
1222
+ 308,
1223
+ 308
1224
+ ],
1225
+ [
1226
+ 98,
1227
+ 310,
1228
+ 265
1229
+ ],
1230
+ [
1231
+ 128,
1232
+ 390,
1233
+ 391
1234
+ ],
1235
+ [
1236
+ 68,
1237
+ 492,
1238
+ 473
1239
+ ],
1240
+ [
1241
+ 66,
1242
+ 497,
1243
+ 497
1244
+ ],
1245
+ [
1246
+ 67,
1247
+ 497,
1248
+ 498
1249
+ ],
1250
+ [
1251
+ 67,
1252
+ 474,
1253
+ 468
1254
+ ],
1255
+ [
1256
+ 66,
1257
+ 460,
1258
+ 493
1259
+ ],
1260
+ [
1261
+ 74,
1262
+ 445,
1263
+ 506
1264
+ ],
1265
+ [
1266
+ 70,
1267
+ 485,
1268
+ 468
1269
+ ],
1270
+ [
1271
+ 68,
1272
+ 501,
1273
+ 497
1274
+ ],
1275
+ [
1276
+ 67,
1277
+ 497,
1278
+ 498
1279
+ ],
1280
+ [
1281
+ 65,
1282
+ 498,
1283
+ 477
1284
+ ],
1285
+ [
1286
+ 72,
1287
+ 470,
1288
+ 503
1289
+ ],
1290
+ [
1291
+ 66,
1292
+ 498,
1293
+ 497
1294
+ ],
1295
+ [
1296
+ 67,
1297
+ 479,
1298
+ 498
1299
+ ],
1300
+ [
1301
+ 68,
1302
+ 499,
1303
+ 497
1304
+ ],
1305
+ [
1306
+ 69,
1307
+ 499,
1308
+ 501
1309
+ ],
1310
+ [
1311
+ 67,
1312
+ 497,
1313
+ 497
1314
+ ],
1315
+ [
1316
+ 69,
1317
+ 498,
1318
+ 499
1319
+ ],
1320
+ [
1321
+ 67,
1322
+ 497,
1323
+ 498
1324
+ ],
1325
+ [
1326
+ 68,
1327
+ 478,
1328
+ 500
1329
+ ],
1330
+ [
1331
+ 68,
1332
+ 501,
1333
+ 487
1334
+ ],
1335
+ [
1336
+ 70,
1337
+ 484,
1338
+ 500
1339
+ ],
1340
+ [
1341
+ 66,
1342
+ 473,
1343
+ 497
1344
+ ],
1345
+ [
1346
+ 66,
1347
+ 497,
1348
+ 497
1349
+ ],
1350
+ [
1351
+ 67,
1352
+ 499,
1353
+ 498
1354
+ ],
1355
+ [
1356
+ 67,
1357
+ 498,
1358
+ 497
1359
+ ],
1360
+ [
1361
+ 66,
1362
+ 498,
1363
+ 496
1364
+ ],
1365
+ [
1366
+ 67,
1367
+ 490,
1368
+ 497
1369
+ ],
1370
+ [
1371
+ 67,
1372
+ 498,
1373
+ 498
1374
+ ],
1375
+ [
1376
+ 66,
1377
+ 497,
1378
+ 497
1379
+ ],
1380
+ [
1381
+ 66,
1382
+ 479,
1383
+ 498
1384
+ ],
1385
+ [
1386
+ 67,
1387
+ 482,
1388
+ 499
1389
+ ],
1390
+ [
1391
+ 70,
1392
+ 478,
1393
+ 499
1394
+ ],
1395
+ [
1396
+ 68,
1397
+ 499,
1398
+ 500
1399
+ ],
1400
+ [
1401
+ 66,
1402
+ 497,
1403
+ 472
1404
+ ],
1405
+ [
1406
+ 70,
1407
+ 499,
1408
+ 482
1409
+ ],
1410
+ [
1411
+ 71,
1412
+ 479,
1413
+ 481
1414
+ ],
1415
+ [
1416
+ 79,
1417
+ 454,
1418
+ 499
1419
+ ],
1420
+ [
1421
+ 66,
1422
+ 497,
1423
+ 496
1424
+ ],
1425
+ [
1426
+ 67,
1427
+ 485,
1428
+ 500
1429
+ ],
1430
+ [
1431
+ 70,
1432
+ 499,
1433
+ 496
1434
+ ],
1435
+ [
1436
+ 59,
1437
+ 501,
1438
+ 498
1439
+ ],
1440
+ [
1441
+ 68,
1442
+ 471,
1443
+ 501
1444
+ ],
1445
+ [
1446
+ 64,
1447
+ 499,
1448
+ 498
1449
+ ],
1450
+ [
1451
+ 66,
1452
+ 498,
1453
+ 497
1454
+ ],
1455
+ [
1456
+ 67,
1457
+ 452,
1458
+ 469
1459
+ ],
1460
+ [
1461
+ 75,
1462
+ 435,
1463
+ 484
1464
+ ],
1465
+ [
1466
+ 66,
1467
+ 486,
1468
+ 492
1469
+ ],
1470
+ [
1471
+ 66,
1472
+ 467,
1473
+ 497
1474
+ ],
1475
+ [
1476
+ 75,
1477
+ 486,
1478
+ 498
1479
+ ],
1480
+ [
1481
+ 67,
1482
+ 498,
1483
+ 497
1484
+ ],
1485
+ [
1486
+ 67,
1487
+ 497,
1488
+ 497
1489
+ ],
1490
+ [
1491
+ 67,
1492
+ 498,
1493
+ 499
1494
+ ],
1495
+ [
1496
+ 66,
1497
+ 497,
1498
+ 497
1499
+ ],
1500
+ [
1501
+ 67,
1502
+ 498,
1503
+ 496
1504
+ ],
1505
+ [
1506
+ 66,
1507
+ 466,
1508
+ 498
1509
+ ],
1510
+ [
1511
+ 70,
1512
+ 482,
1513
+ 501
1514
+ ],
1515
+ [
1516
+ 69,
1517
+ 478,
1518
+ 479
1519
+ ],
1520
+ [
1521
+ 67,
1522
+ 500,
1523
+ 497
1524
+ ],
1525
+ [
1526
+ 71,
1527
+ 469,
1528
+ 502
1529
+ ],
1530
+ [
1531
+ 72,
1532
+ 466,
1533
+ 500
1534
+ ],
1535
+ [
1536
+ 66,
1537
+ 486,
1538
+ 497
1539
+ ],
1540
+ [
1541
+ 68,
1542
+ 497,
1543
+ 500
1544
+ ],
1545
+ [
1546
+ 67,
1547
+ 498,
1548
+ 498
1549
+ ],
1550
+ [
1551
+ 71,
1552
+ 481,
1553
+ 501
1554
+ ],
1555
+ [
1556
+ 70,
1557
+ 501,
1558
+ 499
1559
+ ]
1560
+ ],
1561
+ "spacings": [
1562
+ [
1563
+ 3.0,
1564
+ 1.0,
1565
+ 1.0
1566
+ ],
1567
+ [
1568
+ 3.0,
1569
+ 1.0,
1570
+ 1.0
1571
+ ],
1572
+ [
1573
+ 3.0,
1574
+ 1.0,
1575
+ 1.0
1576
+ ],
1577
+ [
1578
+ 3.0,
1579
+ 1.0,
1580
+ 1.0
1581
+ ],
1582
+ [
1583
+ 3.0,
1584
+ 1.0,
1585
+ 1.0
1586
+ ],
1587
+ [
1588
+ 3.0,
1589
+ 1.0,
1590
+ 1.0
1591
+ ],
1592
+ [
1593
+ 3.0,
1594
+ 1.0,
1595
+ 1.0
1596
+ ],
1597
+ [
1598
+ 3.0,
1599
+ 1.0,
1600
+ 1.0
1601
+ ],
1602
+ [
1603
+ 3.0,
1604
+ 1.0,
1605
+ 1.0
1606
+ ],
1607
+ [
1608
+ 3.0,
1609
+ 1.0,
1610
+ 1.0
1611
+ ],
1612
+ [
1613
+ 3.0,
1614
+ 1.0,
1615
+ 1.0
1616
+ ],
1617
+ [
1618
+ 3.0,
1619
+ 1.0,
1620
+ 1.0
1621
+ ],
1622
+ [
1623
+ 3.0,
1624
+ 1.0,
1625
+ 1.0
1626
+ ],
1627
+ [
1628
+ 3.0,
1629
+ 1.0,
1630
+ 1.0
1631
+ ],
1632
+ [
1633
+ 3.0,
1634
+ 1.0,
1635
+ 1.0
1636
+ ],
1637
+ [
1638
+ 3.0,
1639
+ 1.0,
1640
+ 1.0
1641
+ ],
1642
+ [
1643
+ 3.0,
1644
+ 1.0,
1645
+ 1.0
1646
+ ],
1647
+ [
1648
+ 3.0,
1649
+ 1.0,
1650
+ 1.0
1651
+ ],
1652
+ [
1653
+ 3.0,
1654
+ 1.0,
1655
+ 1.0
1656
+ ],
1657
+ [
1658
+ 3.0,
1659
+ 1.0,
1660
+ 1.0
1661
+ ],
1662
+ [
1663
+ 3.0,
1664
+ 1.0,
1665
+ 1.0
1666
+ ],
1667
+ [
1668
+ 3.0,
1669
+ 1.0,
1670
+ 1.0
1671
+ ],
1672
+ [
1673
+ 3.0,
1674
+ 1.0,
1675
+ 1.0
1676
+ ],
1677
+ [
1678
+ 3.0,
1679
+ 1.0,
1680
+ 1.0
1681
+ ],
1682
+ [
1683
+ 3.0,
1684
+ 1.0,
1685
+ 1.0
1686
+ ],
1687
+ [
1688
+ 3.0,
1689
+ 1.0,
1690
+ 1.0
1691
+ ],
1692
+ [
1693
+ 3.0,
1694
+ 1.0,
1695
+ 1.0
1696
+ ],
1697
+ [
1698
+ 3.0,
1699
+ 1.0,
1700
+ 1.0
1701
+ ],
1702
+ [
1703
+ 3.0,
1704
+ 1.0,
1705
+ 1.0
1706
+ ],
1707
+ [
1708
+ 3.0,
1709
+ 1.0,
1710
+ 1.0
1711
+ ],
1712
+ [
1713
+ 3.0,
1714
+ 1.0,
1715
+ 1.0
1716
+ ],
1717
+ [
1718
+ 3.0,
1719
+ 1.0,
1720
+ 1.0
1721
+ ],
1722
+ [
1723
+ 3.0,
1724
+ 1.0,
1725
+ 1.0
1726
+ ],
1727
+ [
1728
+ 3.0,
1729
+ 1.0,
1730
+ 1.0
1731
+ ],
1732
+ [
1733
+ 3.0,
1734
+ 1.0,
1735
+ 1.0
1736
+ ],
1737
+ [
1738
+ 3.0,
1739
+ 1.0,
1740
+ 1.0
1741
+ ],
1742
+ [
1743
+ 3.0,
1744
+ 1.0,
1745
+ 1.0
1746
+ ],
1747
+ [
1748
+ 3.0,
1749
+ 1.0,
1750
+ 1.0
1751
+ ],
1752
+ [
1753
+ 3.0,
1754
+ 1.0,
1755
+ 1.0
1756
+ ],
1757
+ [
1758
+ 3.0,
1759
+ 1.0,
1760
+ 1.0
1761
+ ],
1762
+ [
1763
+ 3.0,
1764
+ 1.0,
1765
+ 1.0
1766
+ ],
1767
+ [
1768
+ 3.0,
1769
+ 1.0,
1770
+ 1.0
1771
+ ],
1772
+ [
1773
+ 3.0,
1774
+ 1.0,
1775
+ 1.0
1776
+ ],
1777
+ [
1778
+ 3.0,
1779
+ 1.0,
1780
+ 1.0
1781
+ ],
1782
+ [
1783
+ 3.0,
1784
+ 1.0,
1785
+ 1.0
1786
+ ],
1787
+ [
1788
+ 3.0,
1789
+ 1.0,
1790
+ 1.0
1791
+ ],
1792
+ [
1793
+ 3.0,
1794
+ 1.0,
1795
+ 1.0
1796
+ ],
1797
+ [
1798
+ 3.0,
1799
+ 1.0,
1800
+ 1.0
1801
+ ],
1802
+ [
1803
+ 3.0,
1804
+ 1.0,
1805
+ 1.0
1806
+ ],
1807
+ [
1808
+ 3.0,
1809
+ 1.0,
1810
+ 1.0
1811
+ ],
1812
+ [
1813
+ 3.0,
1814
+ 1.0,
1815
+ 1.0
1816
+ ],
1817
+ [
1818
+ 3.0,
1819
+ 1.0,
1820
+ 1.0
1821
+ ],
1822
+ [
1823
+ 3.0,
1824
+ 1.0,
1825
+ 1.0
1826
+ ],
1827
+ [
1828
+ 3.0,
1829
+ 1.0,
1830
+ 1.0
1831
+ ],
1832
+ [
1833
+ 3.0,
1834
+ 1.0,
1835
+ 1.0
1836
+ ],
1837
+ [
1838
+ 3.0,
1839
+ 1.0,
1840
+ 1.0
1841
+ ],
1842
+ [
1843
+ 3.0,
1844
+ 1.0,
1845
+ 1.0
1846
+ ],
1847
+ [
1848
+ 3.0,
1849
+ 1.0,
1850
+ 1.0
1851
+ ],
1852
+ [
1853
+ 3.0,
1854
+ 1.0,
1855
+ 1.0
1856
+ ],
1857
+ [
1858
+ 3.0,
1859
+ 1.0,
1860
+ 1.0
1861
+ ],
1862
+ [
1863
+ 3.0,
1864
+ 1.0,
1865
+ 1.0
1866
+ ],
1867
+ [
1868
+ 3.0,
1869
+ 1.0,
1870
+ 1.0
1871
+ ],
1872
+ [
1873
+ 3.0,
1874
+ 1.0,
1875
+ 1.0
1876
+ ],
1877
+ [
1878
+ 3.0,
1879
+ 1.0,
1880
+ 1.0
1881
+ ],
1882
+ [
1883
+ 3.0,
1884
+ 1.0,
1885
+ 1.0
1886
+ ],
1887
+ [
1888
+ 3.0,
1889
+ 1.0,
1890
+ 1.0
1891
+ ],
1892
+ [
1893
+ 3.0,
1894
+ 1.0,
1895
+ 1.0
1896
+ ],
1897
+ [
1898
+ 3.0,
1899
+ 1.0,
1900
+ 1.0
1901
+ ],
1902
+ [
1903
+ 3.0,
1904
+ 1.0,
1905
+ 1.0
1906
+ ],
1907
+ [
1908
+ 3.0,
1909
+ 1.0,
1910
+ 1.0
1911
+ ],
1912
+ [
1913
+ 3.0,
1914
+ 1.0,
1915
+ 1.0
1916
+ ],
1917
+ [
1918
+ 3.0,
1919
+ 1.0,
1920
+ 1.0
1921
+ ],
1922
+ [
1923
+ 3.0,
1924
+ 1.0,
1925
+ 1.0
1926
+ ],
1927
+ [
1928
+ 3.0,
1929
+ 1.0,
1930
+ 1.0
1931
+ ],
1932
+ [
1933
+ 3.0,
1934
+ 1.0,
1935
+ 1.0
1936
+ ],
1937
+ [
1938
+ 3.0,
1939
+ 1.0,
1940
+ 1.0
1941
+ ],
1942
+ [
1943
+ 3.0,
1944
+ 1.0,
1945
+ 1.0
1946
+ ],
1947
+ [
1948
+ 3.0,
1949
+ 1.0,
1950
+ 1.0
1951
+ ],
1952
+ [
1953
+ 3.0,
1954
+ 1.0,
1955
+ 1.0
1956
+ ],
1957
+ [
1958
+ 3.0,
1959
+ 1.0,
1960
+ 1.0
1961
+ ],
1962
+ [
1963
+ 3.0,
1964
+ 1.0,
1965
+ 1.0
1966
+ ],
1967
+ [
1968
+ 3.0,
1969
+ 1.0,
1970
+ 1.0
1971
+ ],
1972
+ [
1973
+ 3.0,
1974
+ 1.0,
1975
+ 1.0
1976
+ ],
1977
+ [
1978
+ 3.0,
1979
+ 1.0,
1980
+ 1.0
1981
+ ],
1982
+ [
1983
+ 3.0,
1984
+ 1.0,
1985
+ 1.0
1986
+ ],
1987
+ [
1988
+ 3.0,
1989
+ 1.0,
1990
+ 1.0
1991
+ ],
1992
+ [
1993
+ 3.0,
1994
+ 1.0,
1995
+ 1.0
1996
+ ],
1997
+ [
1998
+ 3.0,
1999
+ 1.0,
2000
+ 1.0
2001
+ ],
2002
+ [
2003
+ 3.0,
2004
+ 1.0,
2005
+ 1.0
2006
+ ],
2007
+ [
2008
+ 3.0,
2009
+ 1.0,
2010
+ 1.0
2011
+ ],
2012
+ [
2013
+ 3.0,
2014
+ 1.0,
2015
+ 1.0
2016
+ ],
2017
+ [
2018
+ 3.0,
2019
+ 1.0,
2020
+ 1.0
2021
+ ],
2022
+ [
2023
+ 3.0,
2024
+ 1.0,
2025
+ 1.0
2026
+ ],
2027
+ [
2028
+ 3.0,
2029
+ 1.0,
2030
+ 1.0
2031
+ ],
2032
+ [
2033
+ 3.0,
2034
+ 1.0,
2035
+ 1.0
2036
+ ],
2037
+ [
2038
+ 3.0,
2039
+ 1.0,
2040
+ 1.0
2041
+ ],
2042
+ [
2043
+ 3.0,
2044
+ 1.0,
2045
+ 1.0
2046
+ ],
2047
+ [
2048
+ 3.0,
2049
+ 1.0,
2050
+ 1.0
2051
+ ],
2052
+ [
2053
+ 3.0,
2054
+ 1.0,
2055
+ 1.0
2056
+ ],
2057
+ [
2058
+ 3.0,
2059
+ 1.0,
2060
+ 1.0
2061
+ ],
2062
+ [
2063
+ 3.0,
2064
+ 1.0,
2065
+ 1.0
2066
+ ],
2067
+ [
2068
+ 3.0,
2069
+ 1.0,
2070
+ 1.0
2071
+ ],
2072
+ [
2073
+ 3.0,
2074
+ 1.0,
2075
+ 1.0
2076
+ ],
2077
+ [
2078
+ 3.0,
2079
+ 1.0,
2080
+ 1.0
2081
+ ],
2082
+ [
2083
+ 3.0,
2084
+ 1.0,
2085
+ 1.0
2086
+ ],
2087
+ [
2088
+ 3.0,
2089
+ 1.0,
2090
+ 1.0
2091
+ ],
2092
+ [
2093
+ 3.0,
2094
+ 1.0,
2095
+ 1.0
2096
+ ],
2097
+ [
2098
+ 3.0,
2099
+ 1.0,
2100
+ 1.0
2101
+ ],
2102
+ [
2103
+ 3.0,
2104
+ 1.0,
2105
+ 1.0
2106
+ ],
2107
+ [
2108
+ 3.0,
2109
+ 1.0,
2110
+ 1.0
2111
+ ],
2112
+ [
2113
+ 3.0,
2114
+ 1.0,
2115
+ 1.0
2116
+ ],
2117
+ [
2118
+ 3.0,
2119
+ 1.0,
2120
+ 1.0
2121
+ ],
2122
+ [
2123
+ 3.0,
2124
+ 1.0,
2125
+ 1.0
2126
+ ],
2127
+ [
2128
+ 3.0,
2129
+ 1.0,
2130
+ 1.0
2131
+ ],
2132
+ [
2133
+ 3.0,
2134
+ 1.0,
2135
+ 1.0
2136
+ ],
2137
+ [
2138
+ 3.0,
2139
+ 1.0,
2140
+ 1.0
2141
+ ],
2142
+ [
2143
+ 3.0,
2144
+ 1.0,
2145
+ 1.0
2146
+ ],
2147
+ [
2148
+ 3.0,
2149
+ 1.0,
2150
+ 1.0
2151
+ ],
2152
+ [
2153
+ 3.0,
2154
+ 1.0,
2155
+ 1.0
2156
+ ],
2157
+ [
2158
+ 3.0,
2159
+ 1.0,
2160
+ 1.0
2161
+ ],
2162
+ [
2163
+ 3.0,
2164
+ 1.0,
2165
+ 1.0
2166
+ ],
2167
+ [
2168
+ 3.0,
2169
+ 1.0,
2170
+ 1.0
2171
+ ],
2172
+ [
2173
+ 3.0,
2174
+ 1.0,
2175
+ 1.0
2176
+ ],
2177
+ [
2178
+ 3.0,
2179
+ 1.0,
2180
+ 1.0
2181
+ ],
2182
+ [
2183
+ 3.0,
2184
+ 1.0,
2185
+ 1.0
2186
+ ],
2187
+ [
2188
+ 3.0,
2189
+ 1.0,
2190
+ 1.0
2191
+ ],
2192
+ [
2193
+ 3.0,
2194
+ 1.0,
2195
+ 1.0
2196
+ ],
2197
+ [
2198
+ 3.0,
2199
+ 1.0,
2200
+ 1.0
2201
+ ],
2202
+ [
2203
+ 3.0,
2204
+ 1.0,
2205
+ 1.0
2206
+ ],
2207
+ [
2208
+ 3.0,
2209
+ 1.0,
2210
+ 1.0
2211
+ ],
2212
+ [
2213
+ 3.0,
2214
+ 1.0,
2215
+ 1.0
2216
+ ],
2217
+ [
2218
+ 3.0,
2219
+ 1.0,
2220
+ 1.0
2221
+ ],
2222
+ [
2223
+ 3.0,
2224
+ 1.0,
2225
+ 1.0
2226
+ ],
2227
+ [
2228
+ 3.0,
2229
+ 1.0,
2230
+ 1.0
2231
+ ],
2232
+ [
2233
+ 3.0,
2234
+ 1.0,
2235
+ 1.0
2236
+ ],
2237
+ [
2238
+ 3.0,
2239
+ 1.0,
2240
+ 1.0
2241
+ ],
2242
+ [
2243
+ 3.0,
2244
+ 1.0,
2245
+ 1.0
2246
+ ],
2247
+ [
2248
+ 3.0,
2249
+ 1.0,
2250
+ 1.0
2251
+ ],
2252
+ [
2253
+ 3.0,
2254
+ 1.0,
2255
+ 1.0
2256
+ ],
2257
+ [
2258
+ 3.0,
2259
+ 1.0,
2260
+ 1.0
2261
+ ],
2262
+ [
2263
+ 3.0,
2264
+ 1.0,
2265
+ 1.0
2266
+ ],
2267
+ [
2268
+ 3.0,
2269
+ 1.0,
2270
+ 1.0
2271
+ ],
2272
+ [
2273
+ 3.0,
2274
+ 1.0,
2275
+ 1.0
2276
+ ],
2277
+ [
2278
+ 3.0,
2279
+ 1.0,
2280
+ 1.0
2281
+ ],
2282
+ [
2283
+ 3.0,
2284
+ 1.0,
2285
+ 1.0
2286
+ ],
2287
+ [
2288
+ 3.0,
2289
+ 1.0,
2290
+ 1.0
2291
+ ],
2292
+ [
2293
+ 3.0,
2294
+ 1.0,
2295
+ 1.0
2296
+ ],
2297
+ [
2298
+ 3.0,
2299
+ 1.0,
2300
+ 1.0
2301
+ ],
2302
+ [
2303
+ 3.0,
2304
+ 1.0,
2305
+ 1.0
2306
+ ],
2307
+ [
2308
+ 3.0,
2309
+ 1.0,
2310
+ 1.0
2311
+ ],
2312
+ [
2313
+ 3.0,
2314
+ 1.0,
2315
+ 1.0
2316
+ ],
2317
+ [
2318
+ 3.0,
2319
+ 1.0,
2320
+ 1.0
2321
+ ],
2322
+ [
2323
+ 3.0,
2324
+ 1.0,
2325
+ 1.0
2326
+ ],
2327
+ [
2328
+ 3.0,
2329
+ 1.0,
2330
+ 1.0
2331
+ ],
2332
+ [
2333
+ 3.0,
2334
+ 1.0,
2335
+ 1.0
2336
+ ],
2337
+ [
2338
+ 3.0,
2339
+ 1.0,
2340
+ 1.0
2341
+ ],
2342
+ [
2343
+ 3.0,
2344
+ 1.0,
2345
+ 1.0
2346
+ ],
2347
+ [
2348
+ 3.0,
2349
+ 1.0,
2350
+ 1.0
2351
+ ],
2352
+ [
2353
+ 3.0,
2354
+ 1.0,
2355
+ 1.0
2356
+ ],
2357
+ [
2358
+ 3.0,
2359
+ 1.0,
2360
+ 1.0
2361
+ ],
2362
+ [
2363
+ 3.0,
2364
+ 1.0,
2365
+ 1.0
2366
+ ],
2367
+ [
2368
+ 3.0,
2369
+ 1.0,
2370
+ 1.0
2371
+ ],
2372
+ [
2373
+ 3.0,
2374
+ 1.0,
2375
+ 1.0
2376
+ ],
2377
+ [
2378
+ 3.0,
2379
+ 1.0,
2380
+ 1.0
2381
+ ],
2382
+ [
2383
+ 3.0,
2384
+ 1.0,
2385
+ 1.0
2386
+ ],
2387
+ [
2388
+ 3.0,
2389
+ 1.0,
2390
+ 1.0
2391
+ ],
2392
+ [
2393
+ 3.0,
2394
+ 1.0,
2395
+ 1.0
2396
+ ],
2397
+ [
2398
+ 3.0,
2399
+ 1.0,
2400
+ 1.0
2401
+ ],
2402
+ [
2403
+ 3.0,
2404
+ 1.0,
2405
+ 1.0
2406
+ ],
2407
+ [
2408
+ 3.0,
2409
+ 1.0,
2410
+ 1.0
2411
+ ],
2412
+ [
2413
+ 3.0,
2414
+ 1.0,
2415
+ 1.0
2416
+ ],
2417
+ [
2418
+ 3.0,
2419
+ 1.0,
2420
+ 1.0
2421
+ ],
2422
+ [
2423
+ 3.0,
2424
+ 1.0,
2425
+ 1.0
2426
+ ],
2427
+ [
2428
+ 3.0,
2429
+ 1.0,
2430
+ 1.0
2431
+ ],
2432
+ [
2433
+ 3.0,
2434
+ 1.0,
2435
+ 1.0
2436
+ ],
2437
+ [
2438
+ 3.0,
2439
+ 1.0,
2440
+ 1.0
2441
+ ],
2442
+ [
2443
+ 3.0,
2444
+ 1.0,
2445
+ 1.0
2446
+ ],
2447
+ [
2448
+ 3.0,
2449
+ 1.0,
2450
+ 1.0
2451
+ ],
2452
+ [
2453
+ 3.0,
2454
+ 1.0,
2455
+ 1.0
2456
+ ],
2457
+ [
2458
+ 3.0,
2459
+ 1.0,
2460
+ 1.0
2461
+ ],
2462
+ [
2463
+ 3.0,
2464
+ 1.0,
2465
+ 1.0
2466
+ ],
2467
+ [
2468
+ 3.0,
2469
+ 1.0,
2470
+ 1.0
2471
+ ],
2472
+ [
2473
+ 3.0,
2474
+ 1.0,
2475
+ 1.0
2476
+ ],
2477
+ [
2478
+ 3.0,
2479
+ 1.0,
2480
+ 1.0
2481
+ ],
2482
+ [
2483
+ 3.0,
2484
+ 1.0,
2485
+ 1.0
2486
+ ],
2487
+ [
2488
+ 3.0,
2489
+ 1.0,
2490
+ 1.0
2491
+ ],
2492
+ [
2493
+ 3.0,
2494
+ 1.0,
2495
+ 1.0
2496
+ ],
2497
+ [
2498
+ 3.0,
2499
+ 1.0,
2500
+ 1.0
2501
+ ],
2502
+ [
2503
+ 3.0,
2504
+ 1.0,
2505
+ 1.0
2506
+ ],
2507
+ [
2508
+ 3.0,
2509
+ 1.0,
2510
+ 1.0
2511
+ ],
2512
+ [
2513
+ 3.0,
2514
+ 1.0,
2515
+ 1.0
2516
+ ],
2517
+ [
2518
+ 3.0,
2519
+ 1.0,
2520
+ 1.0
2521
+ ],
2522
+ [
2523
+ 3.0,
2524
+ 1.0,
2525
+ 1.0
2526
+ ],
2527
+ [
2528
+ 3.0,
2529
+ 1.0,
2530
+ 1.0
2531
+ ],
2532
+ [
2533
+ 3.0,
2534
+ 1.0,
2535
+ 1.0
2536
+ ],
2537
+ [
2538
+ 3.0,
2539
+ 1.0,
2540
+ 1.0
2541
+ ],
2542
+ [
2543
+ 3.0,
2544
+ 1.0,
2545
+ 1.0
2546
+ ],
2547
+ [
2548
+ 3.0,
2549
+ 1.0,
2550
+ 1.0
2551
+ ],
2552
+ [
2553
+ 3.0,
2554
+ 1.0,
2555
+ 1.0
2556
+ ],
2557
+ [
2558
+ 3.0,
2559
+ 1.0,
2560
+ 1.0
2561
+ ],
2562
+ [
2563
+ 3.0,
2564
+ 1.0,
2565
+ 1.0
2566
+ ],
2567
+ [
2568
+ 3.0,
2569
+ 1.0,
2570
+ 1.0
2571
+ ],
2572
+ [
2573
+ 3.0,
2574
+ 1.0,
2575
+ 1.0
2576
+ ],
2577
+ [
2578
+ 3.0,
2579
+ 1.0,
2580
+ 1.0
2581
+ ],
2582
+ [
2583
+ 3.0,
2584
+ 1.0,
2585
+ 1.0
2586
+ ],
2587
+ [
2588
+ 3.0,
2589
+ 1.0,
2590
+ 1.0
2591
+ ],
2592
+ [
2593
+ 3.0,
2594
+ 1.0,
2595
+ 1.0
2596
+ ],
2597
+ [
2598
+ 3.0,
2599
+ 1.0,
2600
+ 1.0
2601
+ ],
2602
+ [
2603
+ 3.0,
2604
+ 1.0,
2605
+ 1.0
2606
+ ],
2607
+ [
2608
+ 3.0,
2609
+ 1.0,
2610
+ 1.0
2611
+ ],
2612
+ [
2613
+ 3.0,
2614
+ 1.0,
2615
+ 1.0
2616
+ ],
2617
+ [
2618
+ 3.0,
2619
+ 1.0,
2620
+ 1.0
2621
+ ],
2622
+ [
2623
+ 3.0,
2624
+ 1.0,
2625
+ 1.0
2626
+ ],
2627
+ [
2628
+ 3.0,
2629
+ 1.0,
2630
+ 1.0
2631
+ ],
2632
+ [
2633
+ 3.0,
2634
+ 1.0,
2635
+ 1.0
2636
+ ],
2637
+ [
2638
+ 3.0,
2639
+ 1.0,
2640
+ 1.0
2641
+ ],
2642
+ [
2643
+ 3.0,
2644
+ 1.0,
2645
+ 1.0
2646
+ ],
2647
+ [
2648
+ 3.0,
2649
+ 1.0,
2650
+ 1.0
2651
+ ],
2652
+ [
2653
+ 3.0,
2654
+ 1.0,
2655
+ 1.0
2656
+ ],
2657
+ [
2658
+ 3.0,
2659
+ 1.0,
2660
+ 1.0
2661
+ ],
2662
+ [
2663
+ 3.0,
2664
+ 1.0,
2665
+ 1.0
2666
+ ],
2667
+ [
2668
+ 3.0,
2669
+ 1.0,
2670
+ 1.0
2671
+ ],
2672
+ [
2673
+ 3.0,
2674
+ 1.0,
2675
+ 1.0
2676
+ ],
2677
+ [
2678
+ 3.0,
2679
+ 1.0,
2680
+ 1.0
2681
+ ],
2682
+ [
2683
+ 3.0,
2684
+ 1.0,
2685
+ 1.0
2686
+ ],
2687
+ [
2688
+ 3.0,
2689
+ 1.0,
2690
+ 1.0
2691
+ ],
2692
+ [
2693
+ 3.0,
2694
+ 1.0,
2695
+ 1.0
2696
+ ],
2697
+ [
2698
+ 3.0,
2699
+ 1.0,
2700
+ 1.0
2701
+ ],
2702
+ [
2703
+ 3.0,
2704
+ 1.0,
2705
+ 1.0
2706
+ ],
2707
+ [
2708
+ 3.0,
2709
+ 1.0,
2710
+ 1.0
2711
+ ],
2712
+ [
2713
+ 3.0,
2714
+ 1.0,
2715
+ 1.0
2716
+ ],
2717
+ [
2718
+ 3.0,
2719
+ 1.0,
2720
+ 1.0
2721
+ ],
2722
+ [
2723
+ 3.0,
2724
+ 1.0,
2725
+ 1.0
2726
+ ],
2727
+ [
2728
+ 3.0,
2729
+ 1.0,
2730
+ 1.0
2731
+ ],
2732
+ [
2733
+ 3.0,
2734
+ 1.0,
2735
+ 1.0
2736
+ ],
2737
+ [
2738
+ 3.0,
2739
+ 1.0,
2740
+ 1.0
2741
+ ],
2742
+ [
2743
+ 3.0,
2744
+ 1.0,
2745
+ 1.0
2746
+ ],
2747
+ [
2748
+ 3.0,
2749
+ 1.0,
2750
+ 1.0
2751
+ ],
2752
+ [
2753
+ 3.0,
2754
+ 1.0,
2755
+ 1.0
2756
+ ],
2757
+ [
2758
+ 3.0,
2759
+ 1.0,
2760
+ 1.0
2761
+ ],
2762
+ [
2763
+ 3.0,
2764
+ 1.0,
2765
+ 1.0
2766
+ ],
2767
+ [
2768
+ 3.0,
2769
+ 1.0,
2770
+ 1.0
2771
+ ],
2772
+ [
2773
+ 3.0,
2774
+ 1.0,
2775
+ 1.0
2776
+ ],
2777
+ [
2778
+ 3.0,
2779
+ 1.0,
2780
+ 1.0
2781
+ ],
2782
+ [
2783
+ 3.0,
2784
+ 1.0,
2785
+ 1.0
2786
+ ],
2787
+ [
2788
+ 3.0,
2789
+ 1.0,
2790
+ 1.0
2791
+ ],
2792
+ [
2793
+ 3.0,
2794
+ 1.0,
2795
+ 1.0
2796
+ ],
2797
+ [
2798
+ 3.0,
2799
+ 1.0,
2800
+ 1.0
2801
+ ],
2802
+ [
2803
+ 3.0,
2804
+ 1.0,
2805
+ 1.0
2806
+ ],
2807
+ [
2808
+ 3.0,
2809
+ 1.0,
2810
+ 1.0
2811
+ ],
2812
+ [
2813
+ 3.0,
2814
+ 1.0,
2815
+ 1.0
2816
+ ],
2817
+ [
2818
+ 3.0,
2819
+ 1.0,
2820
+ 1.0
2821
+ ],
2822
+ [
2823
+ 3.0,
2824
+ 1.0,
2825
+ 1.0
2826
+ ],
2827
+ [
2828
+ 3.0,
2829
+ 1.0,
2830
+ 1.0
2831
+ ],
2832
+ [
2833
+ 3.0,
2834
+ 1.0,
2835
+ 1.0
2836
+ ],
2837
+ [
2838
+ 3.0,
2839
+ 1.0,
2840
+ 1.0
2841
+ ],
2842
+ [
2843
+ 3.0,
2844
+ 1.0,
2845
+ 1.0
2846
+ ],
2847
+ [
2848
+ 3.0,
2849
+ 1.0,
2850
+ 1.0
2851
+ ],
2852
+ [
2853
+ 3.0,
2854
+ 1.0,
2855
+ 1.0
2856
+ ],
2857
+ [
2858
+ 3.0,
2859
+ 1.0,
2860
+ 1.0
2861
+ ],
2862
+ [
2863
+ 3.0,
2864
+ 1.0,
2865
+ 1.0
2866
+ ],
2867
+ [
2868
+ 3.0,
2869
+ 1.0,
2870
+ 1.0
2871
+ ],
2872
+ [
2873
+ 3.0,
2874
+ 1.0,
2875
+ 1.0
2876
+ ],
2877
+ [
2878
+ 3.0,
2879
+ 1.0,
2880
+ 1.0
2881
+ ],
2882
+ [
2883
+ 3.0,
2884
+ 1.0,
2885
+ 1.0
2886
+ ],
2887
+ [
2888
+ 3.0,
2889
+ 1.0,
2890
+ 1.0
2891
+ ],
2892
+ [
2893
+ 3.0,
2894
+ 1.0,
2895
+ 1.0
2896
+ ],
2897
+ [
2898
+ 3.0,
2899
+ 1.0,
2900
+ 1.0
2901
+ ],
2902
+ [
2903
+ 3.0,
2904
+ 1.0,
2905
+ 1.0
2906
+ ],
2907
+ [
2908
+ 3.0,
2909
+ 1.0,
2910
+ 1.0
2911
+ ],
2912
+ [
2913
+ 3.0,
2914
+ 1.0,
2915
+ 1.0
2916
+ ],
2917
+ [
2918
+ 3.0,
2919
+ 1.0,
2920
+ 1.0
2921
+ ],
2922
+ [
2923
+ 3.0,
2924
+ 1.0,
2925
+ 1.0
2926
+ ],
2927
+ [
2928
+ 3.0,
2929
+ 1.0,
2930
+ 1.0
2931
+ ],
2932
+ [
2933
+ 3.0,
2934
+ 1.0,
2935
+ 1.0
2936
+ ],
2937
+ [
2938
+ 3.0,
2939
+ 1.0,
2940
+ 1.0
2941
+ ],
2942
+ [
2943
+ 3.0,
2944
+ 1.0,
2945
+ 1.0
2946
+ ],
2947
+ [
2948
+ 3.0,
2949
+ 1.0,
2950
+ 1.0
2951
+ ],
2952
+ [
2953
+ 3.0,
2954
+ 1.0,
2955
+ 1.0
2956
+ ],
2957
+ [
2958
+ 3.0,
2959
+ 1.0,
2960
+ 1.0
2961
+ ],
2962
+ [
2963
+ 3.0,
2964
+ 1.0,
2965
+ 1.0
2966
+ ],
2967
+ [
2968
+ 3.0,
2969
+ 1.0,
2970
+ 1.0
2971
+ ],
2972
+ [
2973
+ 3.0,
2974
+ 1.0,
2975
+ 1.0
2976
+ ],
2977
+ [
2978
+ 3.0,
2979
+ 1.0,
2980
+ 1.0
2981
+ ],
2982
+ [
2983
+ 3.0,
2984
+ 1.0,
2985
+ 1.0
2986
+ ],
2987
+ [
2988
+ 3.0,
2989
+ 1.0,
2990
+ 1.0
2991
+ ],
2992
+ [
2993
+ 3.0,
2994
+ 1.0,
2995
+ 1.0
2996
+ ],
2997
+ [
2998
+ 3.0,
2999
+ 1.0,
3000
+ 1.0
3001
+ ],
3002
+ [
3003
+ 3.0,
3004
+ 1.0,
3005
+ 1.0
3006
+ ],
3007
+ [
3008
+ 3.0,
3009
+ 1.0,
3010
+ 1.0
3011
+ ],
3012
+ [
3013
+ 3.0,
3014
+ 1.0,
3015
+ 1.0
3016
+ ],
3017
+ [
3018
+ 3.0,
3019
+ 1.0,
3020
+ 1.0
3021
+ ],
3022
+ [
3023
+ 3.0,
3024
+ 1.0,
3025
+ 1.0
3026
+ ],
3027
+ [
3028
+ 3.0,
3029
+ 1.0,
3030
+ 1.0
3031
+ ],
3032
+ [
3033
+ 3.0,
3034
+ 1.0,
3035
+ 1.0
3036
+ ],
3037
+ [
3038
+ 3.0,
3039
+ 1.0,
3040
+ 1.0
3041
+ ],
3042
+ [
3043
+ 3.0,
3044
+ 1.0,
3045
+ 1.0
3046
+ ],
3047
+ [
3048
+ 3.0,
3049
+ 1.0,
3050
+ 1.0
3051
+ ],
3052
+ [
3053
+ 3.0,
3054
+ 1.0,
3055
+ 1.0
3056
+ ],
3057
+ [
3058
+ 3.0,
3059
+ 1.0,
3060
+ 1.0
3061
+ ],
3062
+ [
3063
+ 3.0,
3064
+ 1.0,
3065
+ 1.0
3066
+ ],
3067
+ [
3068
+ 3.0,
3069
+ 1.0,
3070
+ 1.0
3071
+ ],
3072
+ [
3073
+ 3.0,
3074
+ 1.0,
3075
+ 1.0
3076
+ ],
3077
+ [
3078
+ 3.0,
3079
+ 1.0,
3080
+ 1.0
3081
+ ],
3082
+ [
3083
+ 3.0,
3084
+ 1.0,
3085
+ 1.0
3086
+ ],
3087
+ [
3088
+ 3.0,
3089
+ 1.0,
3090
+ 1.0
3091
+ ],
3092
+ [
3093
+ 3.0,
3094
+ 1.0,
3095
+ 1.0
3096
+ ],
3097
+ [
3098
+ 3.0,
3099
+ 1.0,
3100
+ 1.0
3101
+ ],
3102
+ [
3103
+ 3.0,
3104
+ 1.0,
3105
+ 1.0
3106
+ ]
3107
+ ]
3108
+ }
Dataset540/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/checkpoint_final.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aaef86084b37113ab519ff31157a2fdb0ddedf6df7290e480c15b724f6b9f1da
3
+ size 815581485
Dataset540/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/debug.json ADDED
The diff for this file is too large to render. See raw diff
 
Dataset540/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/progress.png ADDED

Git LFS Details

  • SHA256: 3bd925c152d4d2bc2e3b25a334f6e30e896d0dc1c9edfc5522ef549fcb7f75ed
  • Pointer size: 131 Bytes
  • Size of remote file: 524 kB
Dataset540/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/validation/summary.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "mae": {
3
+ "mean": 55.60064848848964,
4
+ "std": 14.338210868800594,
5
+ "max": 105.47389638122534,
6
+ "min": 35.20833348049252,
7
+ "25pc": 44.134254115405234,
8
+ "50pc": 53.92539230685208,
9
+ "75pc": 62.27850375865349,
10
+ "count": 62
11
+ },
12
+ "psnr": {
13
+ "mean": 30.48458344031393,
14
+ "std": 2.38001206819551,
15
+ "max": 35.77302372613302,
16
+ "min": 25.190641684003367,
17
+ "25pc": 28.694927361710747,
18
+ "50pc": 29.964147419029842,
19
+ "75pc": 32.48056910462868,
20
+ "count": 62
21
+ },
22
+ "ms_ssim": {
23
+ "mean": 0.9189507095920884,
24
+ "std": 0.05408709159150134,
25
+ "max": 0.9836779174167589,
26
+ "min": 0.715160057895457,
27
+ "25pc": 0.8933213351088732,
28
+ "50pc": 0.9349496548077856,
29
+ "75pc": 0.9545689319733478,
30
+ "count": 62
31
+ },
32
+ "DICE": {
33
+ "mean": 0.7588220357162053,
34
+ "std": 0.15028017475619626,
35
+ "max": 0.9240115880966187,
36
+ "min": 0.12273936718702316,
37
+ "25pc": 0.7277467846870422,
38
+ "50pc": 0.8152719140052795,
39
+ "75pc": 0.8395445346832275,
40
+ "count": 62
41
+ },
42
+ "HD95": {
43
+ "mean": 5.62839973168295,
44
+ "std": 4.9812358838641115,
45
+ "max": 27.221315383911133,
46
+ "min": 1.3090169429779053,
47
+ "25pc": 2.8109169006347656,
48
+ "50pc": 4.001410961151123,
49
+ "75pc": 6.309501647949219,
50
+ "count": 62
51
+ }
52
+ }
Dataset540/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/plans.json ADDED
@@ -0,0 +1,521 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_name": "Dataset540_synthrad2025_task2_CBCT_AB_pre_v2r_stitched_masked_both",
3
+ "plans_name": "nnUNetResEncUNetLPlans",
4
+ "original_median_spacing_after_transp": [
5
+ 3.0,
6
+ 1.0,
7
+ 1.0
8
+ ],
9
+ "original_median_shape_after_transp": [
10
+ 81,
11
+ 449,
12
+ 449
13
+ ],
14
+ "image_reader_writer": "SimpleITKIO",
15
+ "transpose_forward": [
16
+ 0,
17
+ 1,
18
+ 2
19
+ ],
20
+ "transpose_backward": [
21
+ 0,
22
+ 1,
23
+ 2
24
+ ],
25
+ "configurations": {
26
+ "2d": {
27
+ "data_identifier": "nnUNetPlans_2d",
28
+ "preprocessor_name": "DefaultPreprocessor",
29
+ "batch_size": 13,
30
+ "patch_size": [
31
+ 512,
32
+ 448
33
+ ],
34
+ "median_image_size_in_voxels": [
35
+ 449.0,
36
+ 449.0
37
+ ],
38
+ "spacing": [
39
+ 1.0,
40
+ 1.0
41
+ ],
42
+ "normalization_schemes": [
43
+ "CTNormalizationClippingSynthrad2025"
44
+ ],
45
+ "use_mask_for_norm": [
46
+ false
47
+ ],
48
+ "resampling_fn_data": "resample_data_or_seg_to_shape",
49
+ "resampling_fn_seg": "resample_data_or_seg_to_shape",
50
+ "resampling_fn_data_kwargs": {
51
+ "is_seg": false,
52
+ "order": 3,
53
+ "order_z": 0,
54
+ "force_separate_z": null
55
+ },
56
+ "resampling_fn_seg_kwargs": {
57
+ "is_seg": true,
58
+ "order": 1,
59
+ "order_z": 0,
60
+ "force_separate_z": null
61
+ },
62
+ "resampling_fn_probabilities": "resample_data_or_seg_to_shape",
63
+ "resampling_fn_probabilities_kwargs": {
64
+ "is_seg": false,
65
+ "order": 1,
66
+ "order_z": 0,
67
+ "force_separate_z": null
68
+ },
69
+ "architecture": {
70
+ "network_class_name": "dynamic_network_architectures.architectures.unet.ResidualEncoderUNet",
71
+ "arch_kwargs": {
72
+ "n_stages": 7,
73
+ "features_per_stage": [
74
+ 32,
75
+ 64,
76
+ 128,
77
+ 256,
78
+ 512,
79
+ 512,
80
+ 512
81
+ ],
82
+ "conv_op": "torch.nn.modules.conv.Conv2d",
83
+ "kernel_sizes": [
84
+ [
85
+ 3,
86
+ 3
87
+ ],
88
+ [
89
+ 3,
90
+ 3
91
+ ],
92
+ [
93
+ 3,
94
+ 3
95
+ ],
96
+ [
97
+ 3,
98
+ 3
99
+ ],
100
+ [
101
+ 3,
102
+ 3
103
+ ],
104
+ [
105
+ 3,
106
+ 3
107
+ ],
108
+ [
109
+ 3,
110
+ 3
111
+ ]
112
+ ],
113
+ "strides": [
114
+ [
115
+ 1,
116
+ 1
117
+ ],
118
+ [
119
+ 2,
120
+ 2
121
+ ],
122
+ [
123
+ 2,
124
+ 2
125
+ ],
126
+ [
127
+ 2,
128
+ 2
129
+ ],
130
+ [
131
+ 2,
132
+ 2
133
+ ],
134
+ [
135
+ 2,
136
+ 2
137
+ ],
138
+ [
139
+ 2,
140
+ 2
141
+ ]
142
+ ],
143
+ "n_blocks_per_stage": [
144
+ 1,
145
+ 3,
146
+ 4,
147
+ 6,
148
+ 6,
149
+ 6,
150
+ 6
151
+ ],
152
+ "n_conv_per_stage_decoder": [
153
+ 1,
154
+ 1,
155
+ 1,
156
+ 1,
157
+ 1,
158
+ 1
159
+ ],
160
+ "conv_bias": true,
161
+ "norm_op": "torch.nn.modules.instancenorm.InstanceNorm2d",
162
+ "norm_op_kwargs": {
163
+ "eps": 1e-05,
164
+ "affine": true
165
+ },
166
+ "dropout_op": null,
167
+ "dropout_op_kwargs": null,
168
+ "nonlin": "torch.nn.LeakyReLU",
169
+ "nonlin_kwargs": {
170
+ "inplace": true
171
+ }
172
+ },
173
+ "_kw_requires_import": [
174
+ "conv_op",
175
+ "norm_op",
176
+ "dropout_op",
177
+ "nonlin"
178
+ ]
179
+ },
180
+ "batch_dice": true
181
+ },
182
+ "3d_lowres": {
183
+ "data_identifier": "nnUNetResEncUNetLPlans_3d_lowres",
184
+ "preprocessor_name": "DefaultPreprocessor",
185
+ "batch_size": 2,
186
+ "patch_size": [
187
+ 56,
188
+ 224,
189
+ 192
190
+ ],
191
+ "median_image_size_in_voxels": [
192
+ 81,
193
+ 297,
194
+ 297
195
+ ],
196
+ "spacing": [
197
+ 3.0,
198
+ 1.512589724855112,
199
+ 1.512589724855112
200
+ ],
201
+ "normalization_schemes": [
202
+ "CTNormalizationClippingSynthrad2025"
203
+ ],
204
+ "use_mask_for_norm": [
205
+ false
206
+ ],
207
+ "resampling_fn_data": "resample_data_or_seg_to_shape",
208
+ "resampling_fn_seg": "resample_data_or_seg_to_shape",
209
+ "resampling_fn_data_kwargs": {
210
+ "is_seg": false,
211
+ "order": 3,
212
+ "order_z": 0,
213
+ "force_separate_z": null
214
+ },
215
+ "resampling_fn_seg_kwargs": {
216
+ "is_seg": true,
217
+ "order": 1,
218
+ "order_z": 0,
219
+ "force_separate_z": null
220
+ },
221
+ "resampling_fn_probabilities": "resample_data_or_seg_to_shape",
222
+ "resampling_fn_probabilities_kwargs": {
223
+ "is_seg": false,
224
+ "order": 1,
225
+ "order_z": 0,
226
+ "force_separate_z": null
227
+ },
228
+ "architecture": {
229
+ "network_class_name": "dynamic_network_architectures.architectures.unet.ResidualEncoderUNet",
230
+ "arch_kwargs": {
231
+ "n_stages": 6,
232
+ "features_per_stage": [
233
+ 32,
234
+ 64,
235
+ 128,
236
+ 256,
237
+ 320,
238
+ 320
239
+ ],
240
+ "conv_op": "torch.nn.modules.conv.Conv3d",
241
+ "kernel_sizes": [
242
+ [
243
+ 3,
244
+ 3,
245
+ 3
246
+ ],
247
+ [
248
+ 3,
249
+ 3,
250
+ 3
251
+ ],
252
+ [
253
+ 3,
254
+ 3,
255
+ 3
256
+ ],
257
+ [
258
+ 3,
259
+ 3,
260
+ 3
261
+ ],
262
+ [
263
+ 3,
264
+ 3,
265
+ 3
266
+ ],
267
+ [
268
+ 3,
269
+ 3,
270
+ 3
271
+ ]
272
+ ],
273
+ "strides": [
274
+ [
275
+ 1,
276
+ 1,
277
+ 1
278
+ ],
279
+ [
280
+ 2,
281
+ 2,
282
+ 2
283
+ ],
284
+ [
285
+ 2,
286
+ 2,
287
+ 2
288
+ ],
289
+ [
290
+ 2,
291
+ 2,
292
+ 2
293
+ ],
294
+ [
295
+ 1,
296
+ 2,
297
+ 2
298
+ ],
299
+ [
300
+ 1,
301
+ 2,
302
+ 2
303
+ ]
304
+ ],
305
+ "n_blocks_per_stage": [
306
+ 1,
307
+ 3,
308
+ 4,
309
+ 6,
310
+ 6,
311
+ 6
312
+ ],
313
+ "n_conv_per_stage_decoder": [
314
+ 1,
315
+ 1,
316
+ 1,
317
+ 1,
318
+ 1
319
+ ],
320
+ "conv_bias": true,
321
+ "norm_op": "torch.nn.modules.instancenorm.InstanceNorm3d",
322
+ "norm_op_kwargs": {
323
+ "eps": 1e-05,
324
+ "affine": true
325
+ },
326
+ "dropout_op": null,
327
+ "dropout_op_kwargs": null,
328
+ "nonlin": "torch.nn.LeakyReLU",
329
+ "nonlin_kwargs": {
330
+ "inplace": true
331
+ }
332
+ },
333
+ "_kw_requires_import": [
334
+ "conv_op",
335
+ "norm_op",
336
+ "dropout_op",
337
+ "nonlin"
338
+ ]
339
+ },
340
+ "batch_dice": false,
341
+ "next_stage": "3d_cascade_fullres"
342
+ },
343
+ "3d_fullres": {
344
+ "data_identifier": "nnUNetPlans_3d_fullres",
345
+ "preprocessor_name": "DefaultPreprocessor",
346
+ "batch_size": 2,
347
+ "patch_size": [
348
+ 40,
349
+ 224,
350
+ 192
351
+ ],
352
+ "median_image_size_in_voxels": [
353
+ 81.0,
354
+ 449.0,
355
+ 449.0
356
+ ],
357
+ "spacing": [
358
+ 3.0,
359
+ 1.0,
360
+ 1.0
361
+ ],
362
+ "normalization_schemes": [
363
+ "CTNormalizationClippingSynthrad2025"
364
+ ],
365
+ "use_mask_for_norm": [
366
+ false
367
+ ],
368
+ "resampling_fn_data": "resample_data_or_seg_to_shape",
369
+ "resampling_fn_seg": "resample_data_or_seg_to_shape",
370
+ "resampling_fn_data_kwargs": {
371
+ "is_seg": false,
372
+ "order": 3,
373
+ "order_z": 0,
374
+ "force_separate_z": null
375
+ },
376
+ "resampling_fn_seg_kwargs": {
377
+ "is_seg": true,
378
+ "order": 1,
379
+ "order_z": 0,
380
+ "force_separate_z": null
381
+ },
382
+ "resampling_fn_probabilities": "resample_data_or_seg_to_shape",
383
+ "resampling_fn_probabilities_kwargs": {
384
+ "is_seg": false,
385
+ "order": 1,
386
+ "order_z": 0,
387
+ "force_separate_z": null
388
+ },
389
+ "architecture": {
390
+ "network_class_name": "dynamic_network_architectures.architectures.unet.ResidualEncoderUNet",
391
+ "arch_kwargs": {
392
+ "n_stages": 6,
393
+ "features_per_stage": [
394
+ 32,
395
+ 64,
396
+ 128,
397
+ 256,
398
+ 320,
399
+ 320
400
+ ],
401
+ "conv_op": "torch.nn.modules.conv.Conv3d",
402
+ "kernel_sizes": [
403
+ [
404
+ 1,
405
+ 3,
406
+ 3
407
+ ],
408
+ [
409
+ 3,
410
+ 3,
411
+ 3
412
+ ],
413
+ [
414
+ 3,
415
+ 3,
416
+ 3
417
+ ],
418
+ [
419
+ 3,
420
+ 3,
421
+ 3
422
+ ],
423
+ [
424
+ 3,
425
+ 3,
426
+ 3
427
+ ],
428
+ [
429
+ 3,
430
+ 3,
431
+ 3
432
+ ]
433
+ ],
434
+ "strides": [
435
+ [
436
+ 1,
437
+ 1,
438
+ 1
439
+ ],
440
+ [
441
+ 1,
442
+ 2,
443
+ 2
444
+ ],
445
+ [
446
+ 2,
447
+ 2,
448
+ 2
449
+ ],
450
+ [
451
+ 2,
452
+ 2,
453
+ 2
454
+ ],
455
+ [
456
+ 2,
457
+ 2,
458
+ 2
459
+ ],
460
+ [
461
+ 1,
462
+ 2,
463
+ 2
464
+ ]
465
+ ],
466
+ "n_blocks_per_stage": [
467
+ 1,
468
+ 3,
469
+ 4,
470
+ 6,
471
+ 6,
472
+ 6
473
+ ],
474
+ "n_conv_per_stage_decoder": [
475
+ 1,
476
+ 1,
477
+ 1,
478
+ 1,
479
+ 1
480
+ ],
481
+ "conv_bias": true,
482
+ "norm_op": "torch.nn.modules.instancenorm.InstanceNorm3d",
483
+ "norm_op_kwargs": {
484
+ "eps": 1e-05,
485
+ "affine": true
486
+ },
487
+ "dropout_op": null,
488
+ "dropout_op_kwargs": null,
489
+ "nonlin": "torch.nn.LeakyReLU",
490
+ "nonlin_kwargs": {
491
+ "inplace": true
492
+ }
493
+ },
494
+ "_kw_requires_import": [
495
+ "conv_op",
496
+ "norm_op",
497
+ "dropout_op",
498
+ "nonlin"
499
+ ]
500
+ },
501
+ "batch_dice": true
502
+ },
503
+ "3d_cascade_fullres": {
504
+ "inherits_from": "3d_fullres",
505
+ "previous_stage": "3d_lowres"
506
+ }
507
+ },
508
+ "experiment_planner_used": "nnUNetPlannerResEncL",
509
+ "label_manager": "LabelManager",
510
+ "foreground_intensity_properties_per_channel": {
511
+ "0": {
512
+ "max": 3071.0,
513
+ "mean": -239.3903045654297,
514
+ "median": -185.0,
515
+ "min": -1024.0,
516
+ "percentile_00_5": -1024.0,
517
+ "percentile_99_5": 479.0,
518
+ "std": 328.2787170410156
519
+ }
520
+ }
521
+ }
Dataset542/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/dataset.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "labels": {
3
+ "label_001": "1",
4
+ "background": 0
5
+ },
6
+ "channel_names": {
7
+ "0": "CT_zscore_synthrad"
8
+ },
9
+ "numTraining": 325,
10
+ "file_ending": ".nii.gz",
11
+ "test_set": [
12
+ "case_0000"
13
+ ]
14
+ }
Dataset542/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/dataset_fingerprint.json ADDED
@@ -0,0 +1,3268 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "foreground_intensity_properties_per_channel": {
3
+ "0": {
4
+ "max": 3071.0,
5
+ "mean": -215.1082763671875,
6
+ "median": -133.0,
7
+ "min": -1024.0,
8
+ "percentile_00_5": -1024.0,
9
+ "percentile_99_5": 1345.0,
10
+ "std": 459.63665771484375
11
+ }
12
+ },
13
+ "median_relative_size_after_cropping": 1.0,
14
+ "shapes_after_crop": [
15
+ [
16
+ 88,
17
+ 307,
18
+ 309
19
+ ],
20
+ [
21
+ 83,
22
+ 307,
23
+ 311
24
+ ],
25
+ [
26
+ 88,
27
+ 309,
28
+ 308
29
+ ],
30
+ [
31
+ 79,
32
+ 313,
33
+ 314
34
+ ],
35
+ [
36
+ 87,
37
+ 309,
38
+ 310
39
+ ],
40
+ [
41
+ 88,
42
+ 308,
43
+ 308
44
+ ],
45
+ [
46
+ 86,
47
+ 309,
48
+ 312
49
+ ],
50
+ [
51
+ 88,
52
+ 307,
53
+ 308
54
+ ],
55
+ [
56
+ 87,
57
+ 309,
58
+ 309
59
+ ],
60
+ [
61
+ 87,
62
+ 308,
63
+ 309
64
+ ],
65
+ [
66
+ 88,
67
+ 309,
68
+ 308
69
+ ],
70
+ [
71
+ 88,
72
+ 308,
73
+ 307
74
+ ],
75
+ [
76
+ 86,
77
+ 313,
78
+ 313
79
+ ],
80
+ [
81
+ 82,
82
+ 309,
83
+ 309
84
+ ],
85
+ [
86
+ 86,
87
+ 308,
88
+ 313
89
+ ],
90
+ [
91
+ 86,
92
+ 312,
93
+ 307
94
+ ],
95
+ [
96
+ 87,
97
+ 307,
98
+ 309
99
+ ],
100
+ [
101
+ 87,
102
+ 309,
103
+ 309
104
+ ],
105
+ [
106
+ 87,
107
+ 307,
108
+ 308
109
+ ],
110
+ [
111
+ 87,
112
+ 308,
113
+ 309
114
+ ],
115
+ [
116
+ 86,
117
+ 309,
118
+ 314
119
+ ],
120
+ [
121
+ 88,
122
+ 308,
123
+ 310
124
+ ],
125
+ [
126
+ 88,
127
+ 307,
128
+ 308
129
+ ],
130
+ [
131
+ 87,
132
+ 308,
133
+ 307
134
+ ],
135
+ [
136
+ 87,
137
+ 311,
138
+ 314
139
+ ],
140
+ [
141
+ 87,
142
+ 309,
143
+ 309
144
+ ],
145
+ [
146
+ 87,
147
+ 307,
148
+ 306
149
+ ],
150
+ [
151
+ 86,
152
+ 308,
153
+ 310
154
+ ],
155
+ [
156
+ 87,
157
+ 307,
158
+ 312
159
+ ],
160
+ [
161
+ 87,
162
+ 308,
163
+ 311
164
+ ],
165
+ [
166
+ 87,
167
+ 307,
168
+ 307
169
+ ],
170
+ [
171
+ 51,
172
+ 282,
173
+ 311
174
+ ],
175
+ [
176
+ 87,
177
+ 308,
178
+ 310
179
+ ],
180
+ [
181
+ 88,
182
+ 310,
183
+ 309
184
+ ],
185
+ [
186
+ 87,
187
+ 310,
188
+ 308
189
+ ],
190
+ [
191
+ 88,
192
+ 308,
193
+ 308
194
+ ],
195
+ [
196
+ 88,
197
+ 308,
198
+ 308
199
+ ],
200
+ [
201
+ 87,
202
+ 308,
203
+ 309
204
+ ],
205
+ [
206
+ 87,
207
+ 308,
208
+ 309
209
+ ],
210
+ [
211
+ 87,
212
+ 309,
213
+ 310
214
+ ],
215
+ [
216
+ 87,
217
+ 308,
218
+ 312
219
+ ],
220
+ [
221
+ 82,
222
+ 309,
223
+ 310
224
+ ],
225
+ [
226
+ 88,
227
+ 310,
228
+ 308
229
+ ],
230
+ [
231
+ 88,
232
+ 311,
233
+ 308
234
+ ],
235
+ [
236
+ 88,
237
+ 309,
238
+ 309
239
+ ],
240
+ [
241
+ 87,
242
+ 308,
243
+ 312
244
+ ],
245
+ [
246
+ 88,
247
+ 309,
248
+ 310
249
+ ],
250
+ [
251
+ 86,
252
+ 310,
253
+ 312
254
+ ],
255
+ [
256
+ 87,
257
+ 310,
258
+ 309
259
+ ],
260
+ [
261
+ 87,
262
+ 309,
263
+ 310
264
+ ],
265
+ [
266
+ 87,
267
+ 307,
268
+ 308
269
+ ],
270
+ [
271
+ 87,
272
+ 308,
273
+ 310
274
+ ],
275
+ [
276
+ 82,
277
+ 288,
278
+ 273
279
+ ],
280
+ [
281
+ 88,
282
+ 309,
283
+ 309
284
+ ],
285
+ [
286
+ 88,
287
+ 307,
288
+ 307
289
+ ],
290
+ [
291
+ 87,
292
+ 310,
293
+ 307
294
+ ],
295
+ [
296
+ 86,
297
+ 311,
298
+ 309
299
+ ],
300
+ [
301
+ 87,
302
+ 308,
303
+ 311
304
+ ],
305
+ [
306
+ 87,
307
+ 309,
308
+ 309
309
+ ],
310
+ [
311
+ 87,
312
+ 310,
313
+ 308
314
+ ],
315
+ [
316
+ 88,
317
+ 225,
318
+ 308
319
+ ],
320
+ [
321
+ 87,
322
+ 308,
323
+ 307
324
+ ],
325
+ [
326
+ 81,
327
+ 311,
328
+ 308
329
+ ],
330
+ [
331
+ 87,
332
+ 310,
333
+ 310
334
+ ],
335
+ [
336
+ 87,
337
+ 309,
338
+ 307
339
+ ],
340
+ [
341
+ 88,
342
+ 310,
343
+ 308
344
+ ],
345
+ [
346
+ 87,
347
+ 313,
348
+ 311
349
+ ],
350
+ [
351
+ 88,
352
+ 309,
353
+ 309
354
+ ],
355
+ [
356
+ 88,
357
+ 309,
358
+ 307
359
+ ],
360
+ [
361
+ 88,
362
+ 309,
363
+ 308
364
+ ],
365
+ [
366
+ 88,
367
+ 309,
368
+ 309
369
+ ],
370
+ [
371
+ 87,
372
+ 308,
373
+ 312
374
+ ],
375
+ [
376
+ 87,
377
+ 308,
378
+ 310
379
+ ],
380
+ [
381
+ 88,
382
+ 311,
383
+ 309
384
+ ],
385
+ [
386
+ 88,
387
+ 308,
388
+ 308
389
+ ],
390
+ [
391
+ 88,
392
+ 308,
393
+ 308
394
+ ],
395
+ [
396
+ 49,
397
+ 307,
398
+ 307
399
+ ],
400
+ [
401
+ 88,
402
+ 308,
403
+ 309
404
+ ],
405
+ [
406
+ 88,
407
+ 310,
408
+ 307
409
+ ],
410
+ [
411
+ 86,
412
+ 315,
413
+ 315
414
+ ],
415
+ [
416
+ 88,
417
+ 306,
418
+ 307
419
+ ],
420
+ [
421
+ 87,
422
+ 310,
423
+ 311
424
+ ],
425
+ [
426
+ 87,
427
+ 311,
428
+ 308
429
+ ],
430
+ [
431
+ 88,
432
+ 309,
433
+ 308
434
+ ],
435
+ [
436
+ 87,
437
+ 310,
438
+ 309
439
+ ],
440
+ [
441
+ 88,
442
+ 309,
443
+ 309
444
+ ],
445
+ [
446
+ 87,
447
+ 312,
448
+ 309
449
+ ],
450
+ [
451
+ 87,
452
+ 310,
453
+ 311
454
+ ],
455
+ [
456
+ 88,
457
+ 310,
458
+ 311
459
+ ],
460
+ [
461
+ 88,
462
+ 308,
463
+ 309
464
+ ],
465
+ [
466
+ 87,
467
+ 309,
468
+ 310
469
+ ],
470
+ [
471
+ 87,
472
+ 310,
473
+ 311
474
+ ],
475
+ [
476
+ 87,
477
+ 312,
478
+ 311
479
+ ],
480
+ [
481
+ 88,
482
+ 309,
483
+ 311
484
+ ],
485
+ [
486
+ 88,
487
+ 308,
488
+ 310
489
+ ],
490
+ [
491
+ 88,
492
+ 310,
493
+ 309
494
+ ],
495
+ [
496
+ 87,
497
+ 308,
498
+ 310
499
+ ],
500
+ [
501
+ 87,
502
+ 311,
503
+ 308
504
+ ],
505
+ [
506
+ 88,
507
+ 309,
508
+ 308
509
+ ],
510
+ [
511
+ 87,
512
+ 311,
513
+ 311
514
+ ],
515
+ [
516
+ 87,
517
+ 307,
518
+ 309
519
+ ],
520
+ [
521
+ 87,
522
+ 310,
523
+ 311
524
+ ],
525
+ [
526
+ 88,
527
+ 309,
528
+ 309
529
+ ],
530
+ [
531
+ 82,
532
+ 309,
533
+ 308
534
+ ],
535
+ [
536
+ 88,
537
+ 309,
538
+ 309
539
+ ],
540
+ [
541
+ 87,
542
+ 311,
543
+ 308
544
+ ],
545
+ [
546
+ 87,
547
+ 310,
548
+ 313
549
+ ],
550
+ [
551
+ 88,
552
+ 310,
553
+ 309
554
+ ],
555
+ [
556
+ 87,
557
+ 309,
558
+ 311
559
+ ],
560
+ [
561
+ 88,
562
+ 310,
563
+ 309
564
+ ],
565
+ [
566
+ 88,
567
+ 308,
568
+ 308
569
+ ],
570
+ [
571
+ 87,
572
+ 310,
573
+ 309
574
+ ],
575
+ [
576
+ 87,
577
+ 309,
578
+ 310
579
+ ],
580
+ [
581
+ 87,
582
+ 312,
583
+ 312
584
+ ],
585
+ [
586
+ 87,
587
+ 309,
588
+ 311
589
+ ],
590
+ [
591
+ 87,
592
+ 312,
593
+ 311
594
+ ],
595
+ [
596
+ 87,
597
+ 312,
598
+ 310
599
+ ],
600
+ [
601
+ 87,
602
+ 307,
603
+ 309
604
+ ],
605
+ [
606
+ 87,
607
+ 312,
608
+ 308
609
+ ],
610
+ [
611
+ 88,
612
+ 307,
613
+ 309
614
+ ],
615
+ [
616
+ 87,
617
+ 308,
618
+ 312
619
+ ],
620
+ [
621
+ 87,
622
+ 311,
623
+ 307
624
+ ],
625
+ [
626
+ 87,
627
+ 312,
628
+ 311
629
+ ],
630
+ [
631
+ 81,
632
+ 310,
633
+ 311
634
+ ],
635
+ [
636
+ 87,
637
+ 308,
638
+ 312
639
+ ],
640
+ [
641
+ 88,
642
+ 310,
643
+ 308
644
+ ],
645
+ [
646
+ 87,
647
+ 311,
648
+ 309
649
+ ],
650
+ [
651
+ 87,
652
+ 309,
653
+ 310
654
+ ],
655
+ [
656
+ 86,
657
+ 311,
658
+ 317
659
+ ],
660
+ [
661
+ 88,
662
+ 308,
663
+ 308
664
+ ],
665
+ [
666
+ 88,
667
+ 311,
668
+ 309
669
+ ],
670
+ [
671
+ 51,
672
+ 311,
673
+ 311
674
+ ],
675
+ [
676
+ 87,
677
+ 311,
678
+ 309
679
+ ],
680
+ [
681
+ 86,
682
+ 310,
683
+ 307
684
+ ],
685
+ [
686
+ 88,
687
+ 311,
688
+ 308
689
+ ],
690
+ [
691
+ 88,
692
+ 309,
693
+ 309
694
+ ],
695
+ [
696
+ 87,
697
+ 311,
698
+ 312
699
+ ],
700
+ [
701
+ 87,
702
+ 310,
703
+ 307
704
+ ],
705
+ [
706
+ 86,
707
+ 309,
708
+ 311
709
+ ],
710
+ [
711
+ 88,
712
+ 310,
713
+ 311
714
+ ],
715
+ [
716
+ 87,
717
+ 312,
718
+ 308
719
+ ],
720
+ [
721
+ 86,
722
+ 308,
723
+ 311
724
+ ],
725
+ [
726
+ 87,
727
+ 311,
728
+ 312
729
+ ],
730
+ [
731
+ 85,
732
+ 313,
733
+ 314
734
+ ],
735
+ [
736
+ 86,
737
+ 311,
738
+ 311
739
+ ],
740
+ [
741
+ 88,
742
+ 308,
743
+ 309
744
+ ],
745
+ [
746
+ 86,
747
+ 314,
748
+ 313
749
+ ],
750
+ [
751
+ 87,
752
+ 312,
753
+ 312
754
+ ],
755
+ [
756
+ 87,
757
+ 314,
758
+ 311
759
+ ],
760
+ [
761
+ 86,
762
+ 308,
763
+ 312
764
+ ],
765
+ [
766
+ 87,
767
+ 308,
768
+ 312
769
+ ],
770
+ [
771
+ 88,
772
+ 313,
773
+ 308
774
+ ],
775
+ [
776
+ 87,
777
+ 307,
778
+ 309
779
+ ],
780
+ [
781
+ 87,
782
+ 311,
783
+ 308
784
+ ],
785
+ [
786
+ 85,
787
+ 317,
788
+ 311
789
+ ],
790
+ [
791
+ 87,
792
+ 307,
793
+ 311
794
+ ],
795
+ [
796
+ 88,
797
+ 308,
798
+ 312
799
+ ],
800
+ [
801
+ 87,
802
+ 310,
803
+ 310
804
+ ],
805
+ [
806
+ 87,
807
+ 311,
808
+ 311
809
+ ],
810
+ [
811
+ 52,
812
+ 308,
813
+ 310
814
+ ],
815
+ [
816
+ 87,
817
+ 309,
818
+ 307
819
+ ],
820
+ [
821
+ 87,
822
+ 308,
823
+ 311
824
+ ],
825
+ [
826
+ 87,
827
+ 308,
828
+ 310
829
+ ],
830
+ [
831
+ 87,
832
+ 310,
833
+ 309
834
+ ],
835
+ [
836
+ 88,
837
+ 309,
838
+ 309
839
+ ],
840
+ [
841
+ 87,
842
+ 309,
843
+ 309
844
+ ],
845
+ [
846
+ 86,
847
+ 310,
848
+ 314
849
+ ],
850
+ [
851
+ 88,
852
+ 309,
853
+ 308
854
+ ],
855
+ [
856
+ 86,
857
+ 314,
858
+ 312
859
+ ],
860
+ [
861
+ 88,
862
+ 308,
863
+ 309
864
+ ],
865
+ [
866
+ 87,
867
+ 309,
868
+ 310
869
+ ],
870
+ [
871
+ 87,
872
+ 309,
873
+ 310
874
+ ],
875
+ [
876
+ 87,
877
+ 312,
878
+ 309
879
+ ],
880
+ [
881
+ 88,
882
+ 310,
883
+ 310
884
+ ],
885
+ [
886
+ 86,
887
+ 313,
888
+ 311
889
+ ],
890
+ [
891
+ 87,
892
+ 311,
893
+ 310
894
+ ],
895
+ [
896
+ 87,
897
+ 311,
898
+ 309
899
+ ],
900
+ [
901
+ 86,
902
+ 311,
903
+ 312
904
+ ],
905
+ [
906
+ 87,
907
+ 311,
908
+ 310
909
+ ],
910
+ [
911
+ 87,
912
+ 309,
913
+ 307
914
+ ],
915
+ [
916
+ 85,
917
+ 310,
918
+ 311
919
+ ],
920
+ [
921
+ 88,
922
+ 312,
923
+ 308
924
+ ],
925
+ [
926
+ 85,
927
+ 293,
928
+ 319
929
+ ],
930
+ [
931
+ 86,
932
+ 308,
933
+ 308
934
+ ],
935
+ [
936
+ 87,
937
+ 310,
938
+ 309
939
+ ],
940
+ [
941
+ 87,
942
+ 309,
943
+ 310
944
+ ],
945
+ [
946
+ 53,
947
+ 307,
948
+ 310
949
+ ],
950
+ [
951
+ 87,
952
+ 308,
953
+ 312
954
+ ],
955
+ [
956
+ 54,
957
+ 308,
958
+ 307
959
+ ],
960
+ [
961
+ 88,
962
+ 309,
963
+ 307
964
+ ],
965
+ [
966
+ 88,
967
+ 308,
968
+ 307
969
+ ],
970
+ [
971
+ 87,
972
+ 307,
973
+ 310
974
+ ],
975
+ [
976
+ 88,
977
+ 307,
978
+ 308
979
+ ],
980
+ [
981
+ 87,
982
+ 310,
983
+ 307
984
+ ],
985
+ [
986
+ 77,
987
+ 311,
988
+ 311
989
+ ],
990
+ [
991
+ 99,
992
+ 315,
993
+ 316
994
+ ],
995
+ [
996
+ 128,
997
+ 305,
998
+ 307
999
+ ],
1000
+ [
1001
+ 129,
1002
+ 300,
1003
+ 305
1004
+ ],
1005
+ [
1006
+ 128,
1007
+ 303,
1008
+ 305
1009
+ ],
1010
+ [
1011
+ 121,
1012
+ 299,
1013
+ 300
1014
+ ],
1015
+ [
1016
+ 128,
1017
+ 304,
1018
+ 301
1019
+ ],
1020
+ [
1021
+ 70,
1022
+ 300,
1023
+ 298
1024
+ ],
1025
+ [
1026
+ 70,
1027
+ 300,
1028
+ 303
1029
+ ],
1030
+ [
1031
+ 71,
1032
+ 304,
1033
+ 305
1034
+ ],
1035
+ [
1036
+ 71,
1037
+ 302,
1038
+ 303
1039
+ ],
1040
+ [
1041
+ 72,
1042
+ 302,
1043
+ 299
1044
+ ],
1045
+ [
1046
+ 70,
1047
+ 300,
1048
+ 301
1049
+ ],
1050
+ [
1051
+ 99,
1052
+ 308,
1053
+ 312
1054
+ ],
1055
+ [
1056
+ 70,
1057
+ 303,
1058
+ 299
1059
+ ],
1060
+ [
1061
+ 71,
1062
+ 300,
1063
+ 300
1064
+ ],
1065
+ [
1066
+ 70,
1067
+ 301,
1068
+ 299
1069
+ ],
1070
+ [
1071
+ 72,
1072
+ 307,
1073
+ 306
1074
+ ],
1075
+ [
1076
+ 70,
1077
+ 304,
1078
+ 304
1079
+ ],
1080
+ [
1081
+ 121,
1082
+ 303,
1083
+ 306
1084
+ ],
1085
+ [
1086
+ 119,
1087
+ 303,
1088
+ 303
1089
+ ],
1090
+ [
1091
+ 122,
1092
+ 308,
1093
+ 301
1094
+ ],
1095
+ [
1096
+ 130,
1097
+ 310,
1098
+ 301
1099
+ ],
1100
+ [
1101
+ 71,
1102
+ 300,
1103
+ 300
1104
+ ],
1105
+ [
1106
+ 128,
1107
+ 301,
1108
+ 301
1109
+ ],
1110
+ [
1111
+ 130,
1112
+ 392,
1113
+ 400
1114
+ ],
1115
+ [
1116
+ 73,
1117
+ 302,
1118
+ 302
1119
+ ],
1120
+ [
1121
+ 70,
1122
+ 301,
1123
+ 302
1124
+ ],
1125
+ [
1126
+ 70,
1127
+ 301,
1128
+ 301
1129
+ ],
1130
+ [
1131
+ 71,
1132
+ 299,
1133
+ 301
1134
+ ],
1135
+ [
1136
+ 71,
1137
+ 304,
1138
+ 303
1139
+ ],
1140
+ [
1141
+ 71,
1142
+ 302,
1143
+ 298
1144
+ ],
1145
+ [
1146
+ 70,
1147
+ 299,
1148
+ 300
1149
+ ],
1150
+ [
1151
+ 130,
1152
+ 371,
1153
+ 394
1154
+ ],
1155
+ [
1156
+ 123,
1157
+ 285,
1158
+ 306
1159
+ ],
1160
+ [
1161
+ 70,
1162
+ 300,
1163
+ 299
1164
+ ],
1165
+ [
1166
+ 70,
1167
+ 301,
1168
+ 299
1169
+ ],
1170
+ [
1171
+ 116,
1172
+ 302,
1173
+ 302
1174
+ ],
1175
+ [
1176
+ 129,
1177
+ 299,
1178
+ 302
1179
+ ],
1180
+ [
1181
+ 127,
1182
+ 304,
1183
+ 303
1184
+ ],
1185
+ [
1186
+ 115,
1187
+ 311,
1188
+ 311
1189
+ ],
1190
+ [
1191
+ 71,
1192
+ 299,
1193
+ 302
1194
+ ],
1195
+ [
1196
+ 70,
1197
+ 300,
1198
+ 301
1199
+ ],
1200
+ [
1201
+ 72,
1202
+ 304,
1203
+ 300
1204
+ ],
1205
+ [
1206
+ 72,
1207
+ 305,
1208
+ 300
1209
+ ],
1210
+ [
1211
+ 99,
1212
+ 290,
1213
+ 313
1214
+ ],
1215
+ [
1216
+ 70,
1217
+ 300,
1218
+ 299
1219
+ ],
1220
+ [
1221
+ 111,
1222
+ 302,
1223
+ 303
1224
+ ],
1225
+ [
1226
+ 128,
1227
+ 300,
1228
+ 298
1229
+ ],
1230
+ [
1231
+ 122,
1232
+ 305,
1233
+ 302
1234
+ ],
1235
+ [
1236
+ 70,
1237
+ 300,
1238
+ 299
1239
+ ],
1240
+ [
1241
+ 129,
1242
+ 303,
1243
+ 301
1244
+ ],
1245
+ [
1246
+ 121,
1247
+ 310,
1248
+ 305
1249
+ ],
1250
+ [
1251
+ 122,
1252
+ 301,
1253
+ 302
1254
+ ],
1255
+ [
1256
+ 110,
1257
+ 306,
1258
+ 298
1259
+ ],
1260
+ [
1261
+ 119,
1262
+ 310,
1263
+ 308
1264
+ ],
1265
+ [
1266
+ 121,
1267
+ 302,
1268
+ 304
1269
+ ],
1270
+ [
1271
+ 116,
1272
+ 304,
1273
+ 303
1274
+ ],
1275
+ [
1276
+ 124,
1277
+ 308,
1278
+ 313
1279
+ ],
1280
+ [
1281
+ 72,
1282
+ 304,
1283
+ 303
1284
+ ],
1285
+ [
1286
+ 69,
1287
+ 298,
1288
+ 299
1289
+ ],
1290
+ [
1291
+ 71,
1292
+ 299,
1293
+ 301
1294
+ ],
1295
+ [
1296
+ 122,
1297
+ 303,
1298
+ 299
1299
+ ],
1300
+ [
1301
+ 70,
1302
+ 301,
1303
+ 300
1304
+ ],
1305
+ [
1306
+ 131,
1307
+ 313,
1308
+ 310
1309
+ ],
1310
+ [
1311
+ 70,
1312
+ 302,
1313
+ 301
1314
+ ],
1315
+ [
1316
+ 72,
1317
+ 299,
1318
+ 298
1319
+ ],
1320
+ [
1321
+ 66,
1322
+ 496,
1323
+ 498
1324
+ ],
1325
+ [
1326
+ 71,
1327
+ 297,
1328
+ 298
1329
+ ],
1330
+ [
1331
+ 74,
1332
+ 305,
1333
+ 304
1334
+ ],
1335
+ [
1336
+ 72,
1337
+ 298,
1338
+ 297
1339
+ ],
1340
+ [
1341
+ 71,
1342
+ 297,
1343
+ 297
1344
+ ],
1345
+ [
1346
+ 71,
1347
+ 299,
1348
+ 297
1349
+ ],
1350
+ [
1351
+ 73,
1352
+ 298,
1353
+ 299
1354
+ ],
1355
+ [
1356
+ 71,
1357
+ 298,
1358
+ 297
1359
+ ],
1360
+ [
1361
+ 68,
1362
+ 457,
1363
+ 498
1364
+ ],
1365
+ [
1366
+ 71,
1367
+ 298,
1368
+ 297
1369
+ ],
1370
+ [
1371
+ 71,
1372
+ 298,
1373
+ 296
1374
+ ],
1375
+ [
1376
+ 72,
1377
+ 297,
1378
+ 298
1379
+ ],
1380
+ [
1381
+ 72,
1382
+ 297,
1383
+ 301
1384
+ ],
1385
+ [
1386
+ 66,
1387
+ 482,
1388
+ 498
1389
+ ],
1390
+ [
1391
+ 67,
1392
+ 496,
1393
+ 498
1394
+ ],
1395
+ [
1396
+ 70,
1397
+ 500,
1398
+ 498
1399
+ ],
1400
+ [
1401
+ 71,
1402
+ 298,
1403
+ 296
1404
+ ],
1405
+ [
1406
+ 71,
1407
+ 296,
1408
+ 295
1409
+ ],
1410
+ [
1411
+ 66,
1412
+ 497,
1413
+ 497
1414
+ ],
1415
+ [
1416
+ 73,
1417
+ 300,
1418
+ 299
1419
+ ],
1420
+ [
1421
+ 71,
1422
+ 297,
1423
+ 297
1424
+ ],
1425
+ [
1426
+ 71,
1427
+ 296,
1428
+ 297
1429
+ ],
1430
+ [
1431
+ 69,
1432
+ 480,
1433
+ 499
1434
+ ],
1435
+ [
1436
+ 71,
1437
+ 298,
1438
+ 298
1439
+ ],
1440
+ [
1441
+ 71,
1442
+ 297,
1443
+ 297
1444
+ ],
1445
+ [
1446
+ 73,
1447
+ 297,
1448
+ 303
1449
+ ],
1450
+ [
1451
+ 73,
1452
+ 299,
1453
+ 299
1454
+ ],
1455
+ [
1456
+ 72,
1457
+ 299,
1458
+ 298
1459
+ ],
1460
+ [
1461
+ 71,
1462
+ 297,
1463
+ 295
1464
+ ],
1465
+ [
1466
+ 71,
1467
+ 299,
1468
+ 299
1469
+ ],
1470
+ [
1471
+ 71,
1472
+ 298,
1473
+ 296
1474
+ ],
1475
+ [
1476
+ 70,
1477
+ 493,
1478
+ 499
1479
+ ],
1480
+ [
1481
+ 71,
1482
+ 296,
1483
+ 297
1484
+ ],
1485
+ [
1486
+ 68,
1487
+ 497,
1488
+ 500
1489
+ ],
1490
+ [
1491
+ 66,
1492
+ 496,
1493
+ 496
1494
+ ],
1495
+ [
1496
+ 71,
1497
+ 484,
1498
+ 501
1499
+ ],
1500
+ [
1501
+ 74,
1502
+ 297,
1503
+ 301
1504
+ ],
1505
+ [
1506
+ 73,
1507
+ 297,
1508
+ 297
1509
+ ],
1510
+ [
1511
+ 73,
1512
+ 300,
1513
+ 298
1514
+ ],
1515
+ [
1516
+ 72,
1517
+ 298,
1518
+ 298
1519
+ ],
1520
+ [
1521
+ 71,
1522
+ 499,
1523
+ 470
1524
+ ],
1525
+ [
1526
+ 71,
1527
+ 297,
1528
+ 298
1529
+ ],
1530
+ [
1531
+ 72,
1532
+ 498,
1533
+ 503
1534
+ ],
1535
+ [
1536
+ 71,
1537
+ 297,
1538
+ 296
1539
+ ],
1540
+ [
1541
+ 71,
1542
+ 300,
1543
+ 298
1544
+ ],
1545
+ [
1546
+ 71,
1547
+ 298,
1548
+ 296
1549
+ ],
1550
+ [
1551
+ 67,
1552
+ 497,
1553
+ 497
1554
+ ],
1555
+ [
1556
+ 71,
1557
+ 501,
1558
+ 499
1559
+ ],
1560
+ [
1561
+ 69,
1562
+ 498,
1563
+ 501
1564
+ ],
1565
+ [
1566
+ 70,
1567
+ 297,
1568
+ 297
1569
+ ],
1570
+ [
1571
+ 66,
1572
+ 498,
1573
+ 477
1574
+ ],
1575
+ [
1576
+ 71,
1577
+ 500,
1578
+ 500
1579
+ ],
1580
+ [
1581
+ 68,
1582
+ 483,
1583
+ 499
1584
+ ],
1585
+ [
1586
+ 66,
1587
+ 497,
1588
+ 498
1589
+ ],
1590
+ [
1591
+ 76,
1592
+ 306,
1593
+ 301
1594
+ ],
1595
+ [
1596
+ 73,
1597
+ 299,
1598
+ 298
1599
+ ],
1600
+ [
1601
+ 71,
1602
+ 296,
1603
+ 298
1604
+ ],
1605
+ [
1606
+ 70,
1607
+ 499,
1608
+ 498
1609
+ ],
1610
+ [
1611
+ 70,
1612
+ 499,
1613
+ 499
1614
+ ],
1615
+ [
1616
+ 72,
1617
+ 498,
1618
+ 504
1619
+ ],
1620
+ [
1621
+ 71,
1622
+ 490,
1623
+ 502
1624
+ ],
1625
+ [
1626
+ 71,
1627
+ 501,
1628
+ 500
1629
+ ],
1630
+ [
1631
+ 71,
1632
+ 501,
1633
+ 499
1634
+ ],
1635
+ [
1636
+ 71,
1637
+ 499,
1638
+ 501
1639
+ ]
1640
+ ],
1641
+ "spacings": [
1642
+ [
1643
+ 3.0,
1644
+ 1.0,
1645
+ 1.0
1646
+ ],
1647
+ [
1648
+ 3.0,
1649
+ 1.0,
1650
+ 1.0
1651
+ ],
1652
+ [
1653
+ 3.0,
1654
+ 1.0,
1655
+ 1.0
1656
+ ],
1657
+ [
1658
+ 3.0,
1659
+ 1.0,
1660
+ 1.0
1661
+ ],
1662
+ [
1663
+ 3.0,
1664
+ 1.0,
1665
+ 1.0
1666
+ ],
1667
+ [
1668
+ 3.0,
1669
+ 1.0,
1670
+ 1.0
1671
+ ],
1672
+ [
1673
+ 3.0,
1674
+ 1.0,
1675
+ 1.0
1676
+ ],
1677
+ [
1678
+ 3.0,
1679
+ 1.0,
1680
+ 1.0
1681
+ ],
1682
+ [
1683
+ 3.0,
1684
+ 1.0,
1685
+ 1.0
1686
+ ],
1687
+ [
1688
+ 3.0,
1689
+ 1.0,
1690
+ 1.0
1691
+ ],
1692
+ [
1693
+ 3.0,
1694
+ 1.0,
1695
+ 1.0
1696
+ ],
1697
+ [
1698
+ 3.0,
1699
+ 1.0,
1700
+ 1.0
1701
+ ],
1702
+ [
1703
+ 3.0,
1704
+ 1.0,
1705
+ 1.0
1706
+ ],
1707
+ [
1708
+ 3.0,
1709
+ 1.0,
1710
+ 1.0
1711
+ ],
1712
+ [
1713
+ 3.0,
1714
+ 1.0,
1715
+ 1.0
1716
+ ],
1717
+ [
1718
+ 3.0,
1719
+ 1.0,
1720
+ 1.0
1721
+ ],
1722
+ [
1723
+ 3.0,
1724
+ 1.0,
1725
+ 1.0
1726
+ ],
1727
+ [
1728
+ 3.0,
1729
+ 1.0,
1730
+ 1.0
1731
+ ],
1732
+ [
1733
+ 3.0,
1734
+ 1.0,
1735
+ 1.0
1736
+ ],
1737
+ [
1738
+ 3.0,
1739
+ 1.0,
1740
+ 1.0
1741
+ ],
1742
+ [
1743
+ 3.0,
1744
+ 1.0,
1745
+ 1.0
1746
+ ],
1747
+ [
1748
+ 3.0,
1749
+ 1.0,
1750
+ 1.0
1751
+ ],
1752
+ [
1753
+ 3.0,
1754
+ 1.0,
1755
+ 1.0
1756
+ ],
1757
+ [
1758
+ 3.0,
1759
+ 1.0,
1760
+ 1.0
1761
+ ],
1762
+ [
1763
+ 3.0,
1764
+ 1.0,
1765
+ 1.0
1766
+ ],
1767
+ [
1768
+ 3.0,
1769
+ 1.0,
1770
+ 1.0
1771
+ ],
1772
+ [
1773
+ 3.0,
1774
+ 1.0,
1775
+ 1.0
1776
+ ],
1777
+ [
1778
+ 3.0,
1779
+ 1.0,
1780
+ 1.0
1781
+ ],
1782
+ [
1783
+ 3.0,
1784
+ 1.0,
1785
+ 1.0
1786
+ ],
1787
+ [
1788
+ 3.0,
1789
+ 1.0,
1790
+ 1.0
1791
+ ],
1792
+ [
1793
+ 3.0,
1794
+ 1.0,
1795
+ 1.0
1796
+ ],
1797
+ [
1798
+ 3.0,
1799
+ 1.0,
1800
+ 1.0
1801
+ ],
1802
+ [
1803
+ 3.0,
1804
+ 1.0,
1805
+ 1.0
1806
+ ],
1807
+ [
1808
+ 3.0,
1809
+ 1.0,
1810
+ 1.0
1811
+ ],
1812
+ [
1813
+ 3.0,
1814
+ 1.0,
1815
+ 1.0
1816
+ ],
1817
+ [
1818
+ 3.0,
1819
+ 1.0,
1820
+ 1.0
1821
+ ],
1822
+ [
1823
+ 3.0,
1824
+ 1.0,
1825
+ 1.0
1826
+ ],
1827
+ [
1828
+ 3.0,
1829
+ 1.0,
1830
+ 1.0
1831
+ ],
1832
+ [
1833
+ 3.0,
1834
+ 1.0,
1835
+ 1.0
1836
+ ],
1837
+ [
1838
+ 3.0,
1839
+ 1.0,
1840
+ 1.0
1841
+ ],
1842
+ [
1843
+ 3.0,
1844
+ 1.0,
1845
+ 1.0
1846
+ ],
1847
+ [
1848
+ 3.0,
1849
+ 1.0,
1850
+ 1.0
1851
+ ],
1852
+ [
1853
+ 3.0,
1854
+ 1.0,
1855
+ 1.0
1856
+ ],
1857
+ [
1858
+ 3.0,
1859
+ 1.0,
1860
+ 1.0
1861
+ ],
1862
+ [
1863
+ 3.0,
1864
+ 1.0,
1865
+ 1.0
1866
+ ],
1867
+ [
1868
+ 3.0,
1869
+ 1.0,
1870
+ 1.0
1871
+ ],
1872
+ [
1873
+ 3.0,
1874
+ 1.0,
1875
+ 1.0
1876
+ ],
1877
+ [
1878
+ 3.0,
1879
+ 1.0,
1880
+ 1.0
1881
+ ],
1882
+ [
1883
+ 3.0,
1884
+ 1.0,
1885
+ 1.0
1886
+ ],
1887
+ [
1888
+ 3.0,
1889
+ 1.0,
1890
+ 1.0
1891
+ ],
1892
+ [
1893
+ 3.0,
1894
+ 1.0,
1895
+ 1.0
1896
+ ],
1897
+ [
1898
+ 3.0,
1899
+ 1.0,
1900
+ 1.0
1901
+ ],
1902
+ [
1903
+ 3.0,
1904
+ 1.0,
1905
+ 1.0
1906
+ ],
1907
+ [
1908
+ 3.0,
1909
+ 1.0,
1910
+ 1.0
1911
+ ],
1912
+ [
1913
+ 3.0,
1914
+ 1.0,
1915
+ 1.0
1916
+ ],
1917
+ [
1918
+ 3.0,
1919
+ 1.0,
1920
+ 1.0
1921
+ ],
1922
+ [
1923
+ 3.0,
1924
+ 1.0,
1925
+ 1.0
1926
+ ],
1927
+ [
1928
+ 3.0,
1929
+ 1.0,
1930
+ 1.0
1931
+ ],
1932
+ [
1933
+ 3.0,
1934
+ 1.0,
1935
+ 1.0
1936
+ ],
1937
+ [
1938
+ 3.0,
1939
+ 1.0,
1940
+ 1.0
1941
+ ],
1942
+ [
1943
+ 3.0,
1944
+ 1.0,
1945
+ 1.0
1946
+ ],
1947
+ [
1948
+ 3.0,
1949
+ 1.0,
1950
+ 1.0
1951
+ ],
1952
+ [
1953
+ 3.0,
1954
+ 1.0,
1955
+ 1.0
1956
+ ],
1957
+ [
1958
+ 3.0,
1959
+ 1.0,
1960
+ 1.0
1961
+ ],
1962
+ [
1963
+ 3.0,
1964
+ 1.0,
1965
+ 1.0
1966
+ ],
1967
+ [
1968
+ 3.0,
1969
+ 1.0,
1970
+ 1.0
1971
+ ],
1972
+ [
1973
+ 3.0,
1974
+ 1.0,
1975
+ 1.0
1976
+ ],
1977
+ [
1978
+ 3.0,
1979
+ 1.0,
1980
+ 1.0
1981
+ ],
1982
+ [
1983
+ 3.0,
1984
+ 1.0,
1985
+ 1.0
1986
+ ],
1987
+ [
1988
+ 3.0,
1989
+ 1.0,
1990
+ 1.0
1991
+ ],
1992
+ [
1993
+ 3.0,
1994
+ 1.0,
1995
+ 1.0
1996
+ ],
1997
+ [
1998
+ 3.0,
1999
+ 1.0,
2000
+ 1.0
2001
+ ],
2002
+ [
2003
+ 3.0,
2004
+ 1.0,
2005
+ 1.0
2006
+ ],
2007
+ [
2008
+ 3.0,
2009
+ 1.0,
2010
+ 1.0
2011
+ ],
2012
+ [
2013
+ 3.0,
2014
+ 1.0,
2015
+ 1.0
2016
+ ],
2017
+ [
2018
+ 3.0,
2019
+ 1.0,
2020
+ 1.0
2021
+ ],
2022
+ [
2023
+ 3.0,
2024
+ 1.0,
2025
+ 1.0
2026
+ ],
2027
+ [
2028
+ 3.0,
2029
+ 1.0,
2030
+ 1.0
2031
+ ],
2032
+ [
2033
+ 3.0,
2034
+ 1.0,
2035
+ 1.0
2036
+ ],
2037
+ [
2038
+ 3.0,
2039
+ 1.0,
2040
+ 1.0
2041
+ ],
2042
+ [
2043
+ 3.0,
2044
+ 1.0,
2045
+ 1.0
2046
+ ],
2047
+ [
2048
+ 3.0,
2049
+ 1.0,
2050
+ 1.0
2051
+ ],
2052
+ [
2053
+ 3.0,
2054
+ 1.0,
2055
+ 1.0
2056
+ ],
2057
+ [
2058
+ 3.0,
2059
+ 1.0,
2060
+ 1.0
2061
+ ],
2062
+ [
2063
+ 3.0,
2064
+ 1.0,
2065
+ 1.0
2066
+ ],
2067
+ [
2068
+ 3.0,
2069
+ 1.0,
2070
+ 1.0
2071
+ ],
2072
+ [
2073
+ 3.0,
2074
+ 1.0,
2075
+ 1.0
2076
+ ],
2077
+ [
2078
+ 3.0,
2079
+ 1.0,
2080
+ 1.0
2081
+ ],
2082
+ [
2083
+ 3.0,
2084
+ 1.0,
2085
+ 1.0
2086
+ ],
2087
+ [
2088
+ 3.0,
2089
+ 1.0,
2090
+ 1.0
2091
+ ],
2092
+ [
2093
+ 3.0,
2094
+ 1.0,
2095
+ 1.0
2096
+ ],
2097
+ [
2098
+ 3.0,
2099
+ 1.0,
2100
+ 1.0
2101
+ ],
2102
+ [
2103
+ 3.0,
2104
+ 1.0,
2105
+ 1.0
2106
+ ],
2107
+ [
2108
+ 3.0,
2109
+ 1.0,
2110
+ 1.0
2111
+ ],
2112
+ [
2113
+ 3.0,
2114
+ 1.0,
2115
+ 1.0
2116
+ ],
2117
+ [
2118
+ 3.0,
2119
+ 1.0,
2120
+ 1.0
2121
+ ],
2122
+ [
2123
+ 3.0,
2124
+ 1.0,
2125
+ 1.0
2126
+ ],
2127
+ [
2128
+ 3.0,
2129
+ 1.0,
2130
+ 1.0
2131
+ ],
2132
+ [
2133
+ 3.0,
2134
+ 1.0,
2135
+ 1.0
2136
+ ],
2137
+ [
2138
+ 3.0,
2139
+ 1.0,
2140
+ 1.0
2141
+ ],
2142
+ [
2143
+ 3.0,
2144
+ 1.0,
2145
+ 1.0
2146
+ ],
2147
+ [
2148
+ 3.0,
2149
+ 1.0,
2150
+ 1.0
2151
+ ],
2152
+ [
2153
+ 3.0,
2154
+ 1.0,
2155
+ 1.0
2156
+ ],
2157
+ [
2158
+ 3.0,
2159
+ 1.0,
2160
+ 1.0
2161
+ ],
2162
+ [
2163
+ 3.0,
2164
+ 1.0,
2165
+ 1.0
2166
+ ],
2167
+ [
2168
+ 3.0,
2169
+ 1.0,
2170
+ 1.0
2171
+ ],
2172
+ [
2173
+ 3.0,
2174
+ 1.0,
2175
+ 1.0
2176
+ ],
2177
+ [
2178
+ 3.0,
2179
+ 1.0,
2180
+ 1.0
2181
+ ],
2182
+ [
2183
+ 3.0,
2184
+ 1.0,
2185
+ 1.0
2186
+ ],
2187
+ [
2188
+ 3.0,
2189
+ 1.0,
2190
+ 1.0
2191
+ ],
2192
+ [
2193
+ 3.0,
2194
+ 1.0,
2195
+ 1.0
2196
+ ],
2197
+ [
2198
+ 3.0,
2199
+ 1.0,
2200
+ 1.0
2201
+ ],
2202
+ [
2203
+ 3.0,
2204
+ 1.0,
2205
+ 1.0
2206
+ ],
2207
+ [
2208
+ 3.0,
2209
+ 1.0,
2210
+ 1.0
2211
+ ],
2212
+ [
2213
+ 3.0,
2214
+ 1.0,
2215
+ 1.0
2216
+ ],
2217
+ [
2218
+ 3.0,
2219
+ 1.0,
2220
+ 1.0
2221
+ ],
2222
+ [
2223
+ 3.0,
2224
+ 1.0,
2225
+ 1.0
2226
+ ],
2227
+ [
2228
+ 3.0,
2229
+ 1.0,
2230
+ 1.0
2231
+ ],
2232
+ [
2233
+ 3.0,
2234
+ 1.0,
2235
+ 1.0
2236
+ ],
2237
+ [
2238
+ 3.0,
2239
+ 1.0,
2240
+ 1.0
2241
+ ],
2242
+ [
2243
+ 3.0,
2244
+ 1.0,
2245
+ 1.0
2246
+ ],
2247
+ [
2248
+ 3.0,
2249
+ 1.0,
2250
+ 1.0
2251
+ ],
2252
+ [
2253
+ 3.0,
2254
+ 1.0,
2255
+ 1.0
2256
+ ],
2257
+ [
2258
+ 3.0,
2259
+ 1.0,
2260
+ 1.0
2261
+ ],
2262
+ [
2263
+ 3.0,
2264
+ 1.0,
2265
+ 1.0
2266
+ ],
2267
+ [
2268
+ 3.0,
2269
+ 1.0,
2270
+ 1.0
2271
+ ],
2272
+ [
2273
+ 3.0,
2274
+ 1.0,
2275
+ 1.0
2276
+ ],
2277
+ [
2278
+ 3.0,
2279
+ 1.0,
2280
+ 1.0
2281
+ ],
2282
+ [
2283
+ 3.0,
2284
+ 1.0,
2285
+ 1.0
2286
+ ],
2287
+ [
2288
+ 3.0,
2289
+ 1.0,
2290
+ 1.0
2291
+ ],
2292
+ [
2293
+ 3.0,
2294
+ 1.0,
2295
+ 1.0
2296
+ ],
2297
+ [
2298
+ 3.0,
2299
+ 1.0,
2300
+ 1.0
2301
+ ],
2302
+ [
2303
+ 3.0,
2304
+ 1.0,
2305
+ 1.0
2306
+ ],
2307
+ [
2308
+ 3.0,
2309
+ 1.0,
2310
+ 1.0
2311
+ ],
2312
+ [
2313
+ 3.0,
2314
+ 1.0,
2315
+ 1.0
2316
+ ],
2317
+ [
2318
+ 3.0,
2319
+ 1.0,
2320
+ 1.0
2321
+ ],
2322
+ [
2323
+ 3.0,
2324
+ 1.0,
2325
+ 1.0
2326
+ ],
2327
+ [
2328
+ 3.0,
2329
+ 1.0,
2330
+ 1.0
2331
+ ],
2332
+ [
2333
+ 3.0,
2334
+ 1.0,
2335
+ 1.0
2336
+ ],
2337
+ [
2338
+ 3.0,
2339
+ 1.0,
2340
+ 1.0
2341
+ ],
2342
+ [
2343
+ 3.0,
2344
+ 1.0,
2345
+ 1.0
2346
+ ],
2347
+ [
2348
+ 3.0,
2349
+ 1.0,
2350
+ 1.0
2351
+ ],
2352
+ [
2353
+ 3.0,
2354
+ 1.0,
2355
+ 1.0
2356
+ ],
2357
+ [
2358
+ 3.0,
2359
+ 1.0,
2360
+ 1.0
2361
+ ],
2362
+ [
2363
+ 3.0,
2364
+ 1.0,
2365
+ 1.0
2366
+ ],
2367
+ [
2368
+ 3.0,
2369
+ 1.0,
2370
+ 1.0
2371
+ ],
2372
+ [
2373
+ 3.0,
2374
+ 1.0,
2375
+ 1.0
2376
+ ],
2377
+ [
2378
+ 3.0,
2379
+ 1.0,
2380
+ 1.0
2381
+ ],
2382
+ [
2383
+ 3.0,
2384
+ 1.0,
2385
+ 1.0
2386
+ ],
2387
+ [
2388
+ 3.0,
2389
+ 1.0,
2390
+ 1.0
2391
+ ],
2392
+ [
2393
+ 3.0,
2394
+ 1.0,
2395
+ 1.0
2396
+ ],
2397
+ [
2398
+ 3.0,
2399
+ 1.0,
2400
+ 1.0
2401
+ ],
2402
+ [
2403
+ 3.0,
2404
+ 1.0,
2405
+ 1.0
2406
+ ],
2407
+ [
2408
+ 3.0,
2409
+ 1.0,
2410
+ 1.0
2411
+ ],
2412
+ [
2413
+ 3.0,
2414
+ 1.0,
2415
+ 1.0
2416
+ ],
2417
+ [
2418
+ 3.0,
2419
+ 1.0,
2420
+ 1.0
2421
+ ],
2422
+ [
2423
+ 3.0,
2424
+ 1.0,
2425
+ 1.0
2426
+ ],
2427
+ [
2428
+ 3.0,
2429
+ 1.0,
2430
+ 1.0
2431
+ ],
2432
+ [
2433
+ 3.0,
2434
+ 1.0,
2435
+ 1.0
2436
+ ],
2437
+ [
2438
+ 3.0,
2439
+ 1.0,
2440
+ 1.0
2441
+ ],
2442
+ [
2443
+ 3.0,
2444
+ 1.0,
2445
+ 1.0
2446
+ ],
2447
+ [
2448
+ 3.0,
2449
+ 1.0,
2450
+ 1.0
2451
+ ],
2452
+ [
2453
+ 3.0,
2454
+ 1.0,
2455
+ 1.0
2456
+ ],
2457
+ [
2458
+ 3.0,
2459
+ 1.0,
2460
+ 1.0
2461
+ ],
2462
+ [
2463
+ 3.0,
2464
+ 1.0,
2465
+ 1.0
2466
+ ],
2467
+ [
2468
+ 3.0,
2469
+ 1.0,
2470
+ 1.0
2471
+ ],
2472
+ [
2473
+ 3.0,
2474
+ 1.0,
2475
+ 1.0
2476
+ ],
2477
+ [
2478
+ 3.0,
2479
+ 1.0,
2480
+ 1.0
2481
+ ],
2482
+ [
2483
+ 3.0,
2484
+ 1.0,
2485
+ 1.0
2486
+ ],
2487
+ [
2488
+ 3.0,
2489
+ 1.0,
2490
+ 1.0
2491
+ ],
2492
+ [
2493
+ 3.0,
2494
+ 1.0,
2495
+ 1.0
2496
+ ],
2497
+ [
2498
+ 3.0,
2499
+ 1.0,
2500
+ 1.0
2501
+ ],
2502
+ [
2503
+ 3.0,
2504
+ 1.0,
2505
+ 1.0
2506
+ ],
2507
+ [
2508
+ 3.0,
2509
+ 1.0,
2510
+ 1.0
2511
+ ],
2512
+ [
2513
+ 3.0,
2514
+ 1.0,
2515
+ 1.0
2516
+ ],
2517
+ [
2518
+ 3.0,
2519
+ 1.0,
2520
+ 1.0
2521
+ ],
2522
+ [
2523
+ 3.0,
2524
+ 1.0,
2525
+ 1.0
2526
+ ],
2527
+ [
2528
+ 3.0,
2529
+ 1.0,
2530
+ 1.0
2531
+ ],
2532
+ [
2533
+ 3.0,
2534
+ 1.0,
2535
+ 1.0
2536
+ ],
2537
+ [
2538
+ 3.0,
2539
+ 1.0,
2540
+ 1.0
2541
+ ],
2542
+ [
2543
+ 3.0,
2544
+ 1.0,
2545
+ 1.0
2546
+ ],
2547
+ [
2548
+ 3.0,
2549
+ 1.0,
2550
+ 1.0
2551
+ ],
2552
+ [
2553
+ 3.0,
2554
+ 1.0,
2555
+ 1.0
2556
+ ],
2557
+ [
2558
+ 3.0,
2559
+ 1.0,
2560
+ 1.0
2561
+ ],
2562
+ [
2563
+ 3.0,
2564
+ 1.0,
2565
+ 1.0
2566
+ ],
2567
+ [
2568
+ 3.0,
2569
+ 1.0,
2570
+ 1.0
2571
+ ],
2572
+ [
2573
+ 3.0,
2574
+ 1.0,
2575
+ 1.0
2576
+ ],
2577
+ [
2578
+ 3.0,
2579
+ 1.0,
2580
+ 1.0
2581
+ ],
2582
+ [
2583
+ 3.0,
2584
+ 1.0,
2585
+ 1.0
2586
+ ],
2587
+ [
2588
+ 3.0,
2589
+ 1.0,
2590
+ 1.0
2591
+ ],
2592
+ [
2593
+ 3.0,
2594
+ 1.0,
2595
+ 1.0
2596
+ ],
2597
+ [
2598
+ 3.0,
2599
+ 1.0,
2600
+ 1.0
2601
+ ],
2602
+ [
2603
+ 3.0,
2604
+ 1.0,
2605
+ 1.0
2606
+ ],
2607
+ [
2608
+ 3.0,
2609
+ 1.0,
2610
+ 1.0
2611
+ ],
2612
+ [
2613
+ 3.0,
2614
+ 1.0,
2615
+ 1.0
2616
+ ],
2617
+ [
2618
+ 3.0,
2619
+ 1.0,
2620
+ 1.0
2621
+ ],
2622
+ [
2623
+ 3.0,
2624
+ 1.0,
2625
+ 1.0
2626
+ ],
2627
+ [
2628
+ 3.0,
2629
+ 1.0,
2630
+ 1.0
2631
+ ],
2632
+ [
2633
+ 3.0,
2634
+ 1.0,
2635
+ 1.0
2636
+ ],
2637
+ [
2638
+ 3.0,
2639
+ 1.0,
2640
+ 1.0
2641
+ ],
2642
+ [
2643
+ 3.0,
2644
+ 1.0,
2645
+ 1.0
2646
+ ],
2647
+ [
2648
+ 3.0,
2649
+ 1.0,
2650
+ 1.0
2651
+ ],
2652
+ [
2653
+ 3.0,
2654
+ 1.0,
2655
+ 1.0
2656
+ ],
2657
+ [
2658
+ 3.0,
2659
+ 1.0,
2660
+ 1.0
2661
+ ],
2662
+ [
2663
+ 3.0,
2664
+ 1.0,
2665
+ 1.0
2666
+ ],
2667
+ [
2668
+ 3.0,
2669
+ 1.0,
2670
+ 1.0
2671
+ ],
2672
+ [
2673
+ 3.0,
2674
+ 1.0,
2675
+ 1.0
2676
+ ],
2677
+ [
2678
+ 3.0,
2679
+ 1.0,
2680
+ 1.0
2681
+ ],
2682
+ [
2683
+ 3.0,
2684
+ 1.0,
2685
+ 1.0
2686
+ ],
2687
+ [
2688
+ 3.0,
2689
+ 1.0,
2690
+ 1.0
2691
+ ],
2692
+ [
2693
+ 3.0,
2694
+ 1.0,
2695
+ 1.0
2696
+ ],
2697
+ [
2698
+ 3.0,
2699
+ 1.0,
2700
+ 1.0
2701
+ ],
2702
+ [
2703
+ 3.0,
2704
+ 1.0,
2705
+ 1.0
2706
+ ],
2707
+ [
2708
+ 3.0,
2709
+ 1.0,
2710
+ 1.0
2711
+ ],
2712
+ [
2713
+ 3.0,
2714
+ 1.0,
2715
+ 1.0
2716
+ ],
2717
+ [
2718
+ 3.0,
2719
+ 1.0,
2720
+ 1.0
2721
+ ],
2722
+ [
2723
+ 3.0,
2724
+ 1.0,
2725
+ 1.0
2726
+ ],
2727
+ [
2728
+ 3.0,
2729
+ 1.0,
2730
+ 1.0
2731
+ ],
2732
+ [
2733
+ 3.0,
2734
+ 1.0,
2735
+ 1.0
2736
+ ],
2737
+ [
2738
+ 3.0,
2739
+ 1.0,
2740
+ 1.0
2741
+ ],
2742
+ [
2743
+ 3.0,
2744
+ 1.0,
2745
+ 1.0
2746
+ ],
2747
+ [
2748
+ 3.0,
2749
+ 1.0,
2750
+ 1.0
2751
+ ],
2752
+ [
2753
+ 3.0,
2754
+ 1.0,
2755
+ 1.0
2756
+ ],
2757
+ [
2758
+ 3.0,
2759
+ 1.0,
2760
+ 1.0
2761
+ ],
2762
+ [
2763
+ 3.0,
2764
+ 1.0,
2765
+ 1.0
2766
+ ],
2767
+ [
2768
+ 3.0,
2769
+ 1.0,
2770
+ 1.0
2771
+ ],
2772
+ [
2773
+ 3.0,
2774
+ 1.0,
2775
+ 1.0
2776
+ ],
2777
+ [
2778
+ 3.0,
2779
+ 1.0,
2780
+ 1.0
2781
+ ],
2782
+ [
2783
+ 3.0,
2784
+ 1.0,
2785
+ 1.0
2786
+ ],
2787
+ [
2788
+ 3.0,
2789
+ 1.0,
2790
+ 1.0
2791
+ ],
2792
+ [
2793
+ 3.0,
2794
+ 1.0,
2795
+ 1.0
2796
+ ],
2797
+ [
2798
+ 3.0,
2799
+ 1.0,
2800
+ 1.0
2801
+ ],
2802
+ [
2803
+ 3.0,
2804
+ 1.0,
2805
+ 1.0
2806
+ ],
2807
+ [
2808
+ 3.0,
2809
+ 1.0,
2810
+ 1.0
2811
+ ],
2812
+ [
2813
+ 3.0,
2814
+ 1.0,
2815
+ 1.0
2816
+ ],
2817
+ [
2818
+ 3.0,
2819
+ 1.0,
2820
+ 1.0
2821
+ ],
2822
+ [
2823
+ 3.0,
2824
+ 1.0,
2825
+ 1.0
2826
+ ],
2827
+ [
2828
+ 3.0,
2829
+ 1.0,
2830
+ 1.0
2831
+ ],
2832
+ [
2833
+ 3.0,
2834
+ 1.0,
2835
+ 1.0
2836
+ ],
2837
+ [
2838
+ 3.0,
2839
+ 1.0,
2840
+ 1.0
2841
+ ],
2842
+ [
2843
+ 3.0,
2844
+ 1.0,
2845
+ 1.0
2846
+ ],
2847
+ [
2848
+ 3.0,
2849
+ 1.0,
2850
+ 1.0
2851
+ ],
2852
+ [
2853
+ 3.0,
2854
+ 1.0,
2855
+ 1.0
2856
+ ],
2857
+ [
2858
+ 3.0,
2859
+ 1.0,
2860
+ 1.0
2861
+ ],
2862
+ [
2863
+ 3.0,
2864
+ 1.0,
2865
+ 1.0
2866
+ ],
2867
+ [
2868
+ 3.0,
2869
+ 1.0,
2870
+ 1.0
2871
+ ],
2872
+ [
2873
+ 3.0,
2874
+ 1.0,
2875
+ 1.0
2876
+ ],
2877
+ [
2878
+ 3.0,
2879
+ 1.0,
2880
+ 1.0
2881
+ ],
2882
+ [
2883
+ 3.0,
2884
+ 1.0,
2885
+ 1.0
2886
+ ],
2887
+ [
2888
+ 3.0,
2889
+ 1.0,
2890
+ 1.0
2891
+ ],
2892
+ [
2893
+ 3.0,
2894
+ 1.0,
2895
+ 1.0
2896
+ ],
2897
+ [
2898
+ 3.0,
2899
+ 1.0,
2900
+ 1.0
2901
+ ],
2902
+ [
2903
+ 3.0,
2904
+ 1.0,
2905
+ 1.0
2906
+ ],
2907
+ [
2908
+ 3.0,
2909
+ 1.0,
2910
+ 1.0
2911
+ ],
2912
+ [
2913
+ 3.0,
2914
+ 1.0,
2915
+ 1.0
2916
+ ],
2917
+ [
2918
+ 3.0,
2919
+ 1.0,
2920
+ 1.0
2921
+ ],
2922
+ [
2923
+ 3.0,
2924
+ 1.0,
2925
+ 1.0
2926
+ ],
2927
+ [
2928
+ 3.0,
2929
+ 1.0,
2930
+ 1.0
2931
+ ],
2932
+ [
2933
+ 3.0,
2934
+ 1.0,
2935
+ 1.0
2936
+ ],
2937
+ [
2938
+ 3.0,
2939
+ 1.0,
2940
+ 1.0
2941
+ ],
2942
+ [
2943
+ 3.0,
2944
+ 1.0,
2945
+ 1.0
2946
+ ],
2947
+ [
2948
+ 3.0,
2949
+ 1.0,
2950
+ 1.0
2951
+ ],
2952
+ [
2953
+ 3.0,
2954
+ 1.0,
2955
+ 1.0
2956
+ ],
2957
+ [
2958
+ 3.0,
2959
+ 1.0,
2960
+ 1.0
2961
+ ],
2962
+ [
2963
+ 3.0,
2964
+ 1.0,
2965
+ 1.0
2966
+ ],
2967
+ [
2968
+ 3.0,
2969
+ 1.0,
2970
+ 1.0
2971
+ ],
2972
+ [
2973
+ 3.0,
2974
+ 1.0,
2975
+ 1.0
2976
+ ],
2977
+ [
2978
+ 3.0,
2979
+ 1.0,
2980
+ 1.0
2981
+ ],
2982
+ [
2983
+ 3.0,
2984
+ 1.0,
2985
+ 1.0
2986
+ ],
2987
+ [
2988
+ 3.0,
2989
+ 1.0,
2990
+ 1.0
2991
+ ],
2992
+ [
2993
+ 3.0,
2994
+ 1.0,
2995
+ 1.0
2996
+ ],
2997
+ [
2998
+ 3.0,
2999
+ 1.0,
3000
+ 1.0
3001
+ ],
3002
+ [
3003
+ 3.0,
3004
+ 1.0,
3005
+ 1.0
3006
+ ],
3007
+ [
3008
+ 3.0,
3009
+ 1.0,
3010
+ 1.0
3011
+ ],
3012
+ [
3013
+ 3.0,
3014
+ 1.0,
3015
+ 1.0
3016
+ ],
3017
+ [
3018
+ 3.0,
3019
+ 1.0,
3020
+ 1.0
3021
+ ],
3022
+ [
3023
+ 3.0,
3024
+ 1.0,
3025
+ 1.0
3026
+ ],
3027
+ [
3028
+ 3.0,
3029
+ 1.0,
3030
+ 1.0
3031
+ ],
3032
+ [
3033
+ 3.0,
3034
+ 1.0,
3035
+ 1.0
3036
+ ],
3037
+ [
3038
+ 3.0,
3039
+ 1.0,
3040
+ 1.0
3041
+ ],
3042
+ [
3043
+ 3.0,
3044
+ 1.0,
3045
+ 1.0
3046
+ ],
3047
+ [
3048
+ 3.0,
3049
+ 1.0,
3050
+ 1.0
3051
+ ],
3052
+ [
3053
+ 3.0,
3054
+ 1.0,
3055
+ 1.0
3056
+ ],
3057
+ [
3058
+ 3.0,
3059
+ 1.0,
3060
+ 1.0
3061
+ ],
3062
+ [
3063
+ 3.0,
3064
+ 1.0,
3065
+ 1.0
3066
+ ],
3067
+ [
3068
+ 3.0,
3069
+ 1.0,
3070
+ 1.0
3071
+ ],
3072
+ [
3073
+ 3.0,
3074
+ 1.0,
3075
+ 1.0
3076
+ ],
3077
+ [
3078
+ 3.0,
3079
+ 1.0,
3080
+ 1.0
3081
+ ],
3082
+ [
3083
+ 3.0,
3084
+ 1.0,
3085
+ 1.0
3086
+ ],
3087
+ [
3088
+ 3.0,
3089
+ 1.0,
3090
+ 1.0
3091
+ ],
3092
+ [
3093
+ 3.0,
3094
+ 1.0,
3095
+ 1.0
3096
+ ],
3097
+ [
3098
+ 3.0,
3099
+ 1.0,
3100
+ 1.0
3101
+ ],
3102
+ [
3103
+ 3.0,
3104
+ 1.0,
3105
+ 1.0
3106
+ ],
3107
+ [
3108
+ 3.0,
3109
+ 1.0,
3110
+ 1.0
3111
+ ],
3112
+ [
3113
+ 3.0,
3114
+ 1.0,
3115
+ 1.0
3116
+ ],
3117
+ [
3118
+ 3.0,
3119
+ 1.0,
3120
+ 1.0
3121
+ ],
3122
+ [
3123
+ 3.0,
3124
+ 1.0,
3125
+ 1.0
3126
+ ],
3127
+ [
3128
+ 3.0,
3129
+ 1.0,
3130
+ 1.0
3131
+ ],
3132
+ [
3133
+ 3.0,
3134
+ 1.0,
3135
+ 1.0
3136
+ ],
3137
+ [
3138
+ 3.0,
3139
+ 1.0,
3140
+ 1.0
3141
+ ],
3142
+ [
3143
+ 3.0,
3144
+ 1.0,
3145
+ 1.0
3146
+ ],
3147
+ [
3148
+ 3.0,
3149
+ 1.0,
3150
+ 1.0
3151
+ ],
3152
+ [
3153
+ 3.0,
3154
+ 1.0,
3155
+ 1.0
3156
+ ],
3157
+ [
3158
+ 3.0,
3159
+ 1.0,
3160
+ 1.0
3161
+ ],
3162
+ [
3163
+ 3.0,
3164
+ 1.0,
3165
+ 1.0
3166
+ ],
3167
+ [
3168
+ 3.0,
3169
+ 1.0,
3170
+ 1.0
3171
+ ],
3172
+ [
3173
+ 3.0,
3174
+ 1.0,
3175
+ 1.0
3176
+ ],
3177
+ [
3178
+ 3.0,
3179
+ 1.0,
3180
+ 1.0
3181
+ ],
3182
+ [
3183
+ 3.0,
3184
+ 1.0,
3185
+ 1.0
3186
+ ],
3187
+ [
3188
+ 3.0,
3189
+ 1.0,
3190
+ 1.0
3191
+ ],
3192
+ [
3193
+ 3.0,
3194
+ 1.0,
3195
+ 1.0
3196
+ ],
3197
+ [
3198
+ 3.0,
3199
+ 1.0,
3200
+ 1.0
3201
+ ],
3202
+ [
3203
+ 3.0,
3204
+ 1.0,
3205
+ 1.0
3206
+ ],
3207
+ [
3208
+ 3.0,
3209
+ 1.0,
3210
+ 1.0
3211
+ ],
3212
+ [
3213
+ 3.0,
3214
+ 1.0,
3215
+ 1.0
3216
+ ],
3217
+ [
3218
+ 3.0,
3219
+ 1.0,
3220
+ 1.0
3221
+ ],
3222
+ [
3223
+ 3.0,
3224
+ 1.0,
3225
+ 1.0
3226
+ ],
3227
+ [
3228
+ 3.0,
3229
+ 1.0,
3230
+ 1.0
3231
+ ],
3232
+ [
3233
+ 3.0,
3234
+ 1.0,
3235
+ 1.0
3236
+ ],
3237
+ [
3238
+ 3.0,
3239
+ 1.0,
3240
+ 1.0
3241
+ ],
3242
+ [
3243
+ 3.0,
3244
+ 1.0,
3245
+ 1.0
3246
+ ],
3247
+ [
3248
+ 3.0,
3249
+ 1.0,
3250
+ 1.0
3251
+ ],
3252
+ [
3253
+ 3.0,
3254
+ 1.0,
3255
+ 1.0
3256
+ ],
3257
+ [
3258
+ 3.0,
3259
+ 1.0,
3260
+ 1.0
3261
+ ],
3262
+ [
3263
+ 3.0,
3264
+ 1.0,
3265
+ 1.0
3266
+ ]
3267
+ ]
3268
+ }
Dataset542/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/checkpoint_final.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ced0929a6ae2327f388ca4bb74dfca15acc111a68340b10442a2840895b5b5d8
3
+ size 815577709
Dataset542/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/debug.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_best_ema": "0.047194228",
3
+ "aim_run": "<Run#-1202935887677095771 name=7ec9a6e8af2847c58549736c repo=<Repo#-3891852473720853603 path=/datasets/work/hb-synthrad2023/work/synthrad2025/bw_workplace/data/nnunet_struct/results/runs_aim/.aim read_only=None>>",
4
+ "batch_size": "2",
5
+ "configuration_manager": "{'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [48, 192, 192], 'median_image_size_in_voxels': [87.0, 308.0, 309.0], 'spacing': [3.0, 1.0, 1.0], 'normalization_schemes': ['CTNormalizationClippingSynthrad2025'], 'use_mask_for_norm': [False], 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'architecture': {'network_class_name': 'dynamic_network_architectures.architectures.unet.ResidualEncoderUNet', 'arch_kwargs': {'n_stages': 6, 'features_per_stage': [32, 64, 128, 256, 320, 320], 'conv_op': 'torch.nn.modules.conv.Conv3d', 'kernel_sizes': [[1, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'strides': [[1, 1, 1], [1, 2, 2], [2, 2, 2], [2, 2, 2], [2, 2, 2], [1, 2, 2]], 'n_blocks_per_stage': [1, 3, 4, 6, 6, 6], 'n_conv_per_stage_decoder': [1, 1, 1, 1, 1], 'conv_bias': True, 'norm_op': 'torch.nn.modules.instancenorm.InstanceNorm3d', 'norm_op_kwargs': {'eps': 1e-05, 'affine': True}, 'dropout_op': None, 'dropout_op_kwargs': None, 'nonlin': 'torch.nn.LeakyReLU', 'nonlin_kwargs': {'inplace': True}, 'deep_supervision': False}, '_kw_requires_import': ['conv_op', 'norm_op', 'dropout_op', 'nonlin']}, 'batch_dice': False}",
6
+ "configuration_name": "3d_fullres",
7
+ "cudnn_version": 90501,
8
+ "current_epoch": "27",
9
+ "dataloader_train": "<nnunetv2.training.data_augmentation.custom_transforms.limited_length_multithreaded_augmenter.LimitedLenWrapper object at 0x15543c17dc50>",
10
+ "dataloader_train.generator": "<nnunetv2.training.dataloading.data_loader_3d.nnUNetDataLoader3D_MRCT_mask object at 0x1551f9793150>",
11
+ "dataloader_train.num_processes": "12",
12
+ "dataloader_train.transform": "Compose ( [RemoveLabelTransform( output_key = 'seg', input_key = 'seg', replace_with = 0, remove_label = -1 ), RenameTransform( delete_old = True, out_key = 'target', in_key = 'seg' ), NumpyToTensor( keys = ['data', 'target', 'mask'], cast_to = 'float' )] )",
13
+ "dataloader_val": "<nnunetv2.training.data_augmentation.custom_transforms.limited_length_multithreaded_augmenter.LimitedLenWrapper object at 0x15543c17f1d0>",
14
+ "dataloader_val.generator": "<nnunetv2.training.dataloading.data_loader_3d.nnUNetDataLoader3D_MRCT_mask object at 0x1551f97937d0>",
15
+ "dataloader_val.num_processes": "6",
16
+ "dataloader_val.transform": "Compose ( [RemoveLabelTransform( output_key = 'seg', input_key = 'seg', replace_with = 0, remove_label = -1 ), RenameTransform( delete_old = True, out_key = 'target', in_key = 'seg' ), NumpyToTensor( keys = ['data', 'target', 'mask'], cast_to = 'float' )] )",
17
+ "dataset_json": "{'labels': {'label_001': '1', 'background': 0}, 'channel_names': {'0': 'CT_zscore_synthrad'}, 'numTraining': 325, 'file_ending': '.mha'}",
18
+ "debug_custom": "{'run_hash': 'de0b335742114ac0a04510ad'}",
19
+ "decoder_type": "standard",
20
+ "device": "cuda:0",
21
+ "disable_checkpointing": "False",
22
+ "enable_deep_supervision": "False",
23
+ "fold": "0",
24
+ "folder_with_segs_from_previous_stage": "None",
25
+ "gpu_name": "NVIDIA H100",
26
+ "grad_scaler": "<torch.amp.grad_scaler.GradScaler object at 0x15543db00690>",
27
+ "hostname": "g030",
28
+ "image_loss_weight": "0.5",
29
+ "inference_allowed_mirroring_axes": "None",
30
+ "initial_lr": "0.01",
31
+ "is_cascaded": "False",
32
+ "is_ddp": "False",
33
+ "label_manager": "<nnunetv2.utilities.label_handling.label_handling.LabelManager object at 0x15543db00790>",
34
+ "local_rank": "0",
35
+ "log_file": "/datasets/work/hb-synthrad2023/work/synthrad2025/bw_workplace/data/nnunet_struct/results/Dataset542_synthrad2025_task2_CBCT_HN_pre_v2r_stitched_masked_both/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/training_log_2025_8_14_18_25_01.txt",
36
+ "logger": "<nnunetv2.training.logging.nnunet_logger.nnUNetLogger object at 0x15543e6adb50>",
37
+ "loss": "SynPerceptionLoss(\n (seg_model): ResidualEncoderUNet(\n (encoder): ResidualEncoder(\n (stem): StackedConvBlocks(\n (convs): Sequential(\n (0): ConvDropoutNormReLU(\n (conv): Conv3d(1, 32, kernel_size=(1, 3, 3), stride=(1, 1, 1), padding=(0, 1, 1))\n (norm): InstanceNorm3d(32, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(1, 32, kernel_size=(1, 3, 3), stride=(1, 1, 1), padding=(0, 1, 1))\n (1): InstanceNorm3d(32, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n )\n (stages): Sequential(\n (0): StackedResidualBlocks(\n (blocks): Sequential(\n (0): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(32, 32, kernel_size=(1, 3, 3), stride=(1, 1, 1), padding=(0, 1, 1))\n (norm): InstanceNorm3d(32, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(32, 32, kernel_size=(1, 3, 3), stride=(1, 1, 1), padding=(0, 1, 1))\n (1): InstanceNorm3d(32, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(32, 32, kernel_size=(1, 3, 3), stride=(1, 1, 1), padding=(0, 1, 1))\n (norm): InstanceNorm3d(32, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(32, 32, kernel_size=(1, 3, 3), stride=(1, 1, 1), padding=(0, 1, 1))\n (1): InstanceNorm3d(32, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n (1): StackedResidualBlocks(\n (blocks): Sequential(\n (0): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(32, 64, kernel_size=(3, 3, 3), stride=(1, 2, 2), padding=(1, 1, 1))\n (norm): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(32, 64, kernel_size=(3, 3, 3), stride=(1, 2, 2), padding=(1, 1, 1))\n (1): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n (skip): Sequential(\n (0): AvgPool3d(kernel_size=[1, 2, 2], stride=[1, 2, 2], padding=0)\n (1): ConvDropoutNormReLU(\n (conv): Conv3d(32, 64, kernel_size=(1, 1, 1), stride=(1, 1, 1), bias=False)\n (norm): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(32, 64, kernel_size=(1, 1, 1), stride=(1, 1, 1), bias=False)\n (1): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n )\n )\n (1): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (2): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n (2): StackedResidualBlocks(\n (blocks): Sequential(\n (0): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(64, 128, kernel_size=(3, 3, 3), stride=(2, 2, 2), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(64, 128, kernel_size=(3, 3, 3), stride=(2, 2, 2), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n (skip): Sequential(\n (0): AvgPool3d(kernel_size=[2, 2, 2], stride=[2, 2, 2], padding=0)\n (1): ConvDropoutNormReLU(\n (conv): Conv3d(64, 128, kernel_size=(1, 1, 1), stride=(1, 1, 1), bias=False)\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(64, 128, kernel_size=(1, 1, 1), stride=(1, 1, 1), bias=False)\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n )\n )\n (1): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (2): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (3): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n (3): StackedResidualBlocks(\n (blocks): Sequential(\n (0): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(128, 256, kernel_size=(3, 3, 3), stride=(2, 2, 2), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(128, 256, kernel_size=(3, 3, 3), stride=(2, 2, 2), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n (skip): Sequential(\n (0): AvgPool3d(kernel_size=[2, 2, 2], stride=[2, 2, 2], padding=0)\n (1): ConvDropoutNormReLU(\n (conv): Conv3d(128, 256, kernel_size=(1, 1, 1), stride=(1, 1, 1), bias=False)\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(128, 256, kernel_size=(1, 1, 1), stride=(1, 1, 1), bias=False)\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n )\n )\n (1): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (2): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (3): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (4): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (5): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n (4): StackedResidualBlocks(\n (blocks): Sequential(\n (0): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(256, 320, kernel_size=(3, 3, 3), stride=(2, 2, 2), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(256, 320, kernel_size=(3, 3, 3), stride=(2, 2, 2), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n (skip): Sequential(\n (0): AvgPool3d(kernel_size=[2, 2, 2], stride=[2, 2, 2], padding=0)\n (1): ConvDropoutNormReLU(\n (conv): Conv3d(256, 320, kernel_size=(1, 1, 1), stride=(1, 1, 1), bias=False)\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(256, 320, kernel_size=(1, 1, 1), stride=(1, 1, 1), bias=False)\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n )\n )\n (1): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (2): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (3): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (4): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (5): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n (5): StackedResidualBlocks(\n (blocks): Sequential(\n (0): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 2, 2), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 2, 2), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n (skip): Sequential(\n (0): AvgPool3d(kernel_size=[1, 2, 2], stride=[1, 2, 2], padding=0)\n )\n )\n (1): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (2): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (3): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (4): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (5): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n )\n )\n (decoder): UNetDecoder(\n (encoder): ResidualEncoder(\n (stem): StackedConvBlocks(\n (convs): Sequential(\n (0): ConvDropoutNormReLU(\n (conv): Conv3d(1, 32, kernel_size=(1, 3, 3), stride=(1, 1, 1), padding=(0, 1, 1))\n (norm): InstanceNorm3d(32, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(1, 32, kernel_size=(1, 3, 3), stride=(1, 1, 1), padding=(0, 1, 1))\n (1): InstanceNorm3d(32, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n )\n (stages): Sequential(\n (0): StackedResidualBlocks(\n (blocks): Sequential(\n (0): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(32, 32, kernel_size=(1, 3, 3), stride=(1, 1, 1), padding=(0, 1, 1))\n (norm): InstanceNorm3d(32, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(32, 32, kernel_size=(1, 3, 3), stride=(1, 1, 1), padding=(0, 1, 1))\n (1): InstanceNorm3d(32, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(32, 32, kernel_size=(1, 3, 3), stride=(1, 1, 1), padding=(0, 1, 1))\n (norm): InstanceNorm3d(32, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(32, 32, kernel_size=(1, 3, 3), stride=(1, 1, 1), padding=(0, 1, 1))\n (1): InstanceNorm3d(32, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n (1): StackedResidualBlocks(\n (blocks): Sequential(\n (0): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(32, 64, kernel_size=(3, 3, 3), stride=(1, 2, 2), padding=(1, 1, 1))\n (norm): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(32, 64, kernel_size=(3, 3, 3), stride=(1, 2, 2), padding=(1, 1, 1))\n (1): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n (skip): Sequential(\n (0): AvgPool3d(kernel_size=[1, 2, 2], stride=[1, 2, 2], padding=0)\n (1): ConvDropoutNormReLU(\n (conv): Conv3d(32, 64, kernel_size=(1, 1, 1), stride=(1, 1, 1), bias=False)\n (norm): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(32, 64, kernel_size=(1, 1, 1), stride=(1, 1, 1), bias=False)\n (1): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n )\n )\n (1): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (2): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(64, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n (2): StackedResidualBlocks(\n (blocks): Sequential(\n (0): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(64, 128, kernel_size=(3, 3, 3), stride=(2, 2, 2), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(64, 128, kernel_size=(3, 3, 3), stride=(2, 2, 2), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n (skip): Sequential(\n (0): AvgPool3d(kernel_size=[2, 2, 2], stride=[2, 2, 2], padding=0)\n (1): ConvDropoutNormReLU(\n (conv): Conv3d(64, 128, kernel_size=(1, 1, 1), stride=(1, 1, 1), bias=False)\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(64, 128, kernel_size=(1, 1, 1), stride=(1, 1, 1), bias=False)\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n )\n )\n (1): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (2): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (3): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(128, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n (3): StackedResidualBlocks(\n (blocks): Sequential(\n (0): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(128, 256, kernel_size=(3, 3, 3), stride=(2, 2, 2), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(128, 256, kernel_size=(3, 3, 3), stride=(2, 2, 2), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n (skip): Sequential(\n (0): AvgPool3d(kernel_size=[2, 2, 2], stride=[2, 2, 2], padding=0)\n (1): ConvDropoutNormReLU(\n (conv): Conv3d(128, 256, kernel_size=(1, 1, 1), stride=(1, 1, 1), bias=False)\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(128, 256, kernel_size=(1, 1, 1), stride=(1, 1, 1), bias=False)\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n )\n )\n (1): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (2): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (3): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (4): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (5): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(256, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n (4): StackedResidualBlocks(\n (blocks): Sequential(\n (0): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(256, 320, kernel_size=(3, 3, 3), stride=(2, 2, 2), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(256, 320, kernel_size=(3, 3, 3), stride=(2, 2, 2), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n (skip): Sequential(\n (0): AvgPool3d(kernel_size=[2, 2, 2], stride=[2, 2, 2], padding=0)\n (1): ConvDropoutNormReLU(\n (conv): Conv3d(256, 320, kernel_size=(1, 1, 1), stride=(1, 1, 1), bias=False)\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(256, 320, kernel_size=(1, 1, 1), stride=(1, 1, 1), bias=False)\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n )\n )\n (1): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (2): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (3): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (4): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (5): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n (5): StackedResidualBlocks(\n (blocks): Sequential(\n (0): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 2, 2), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 2, 2), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n (skip): Sequential(\n (0): AvgPool3d(kernel_size=[1, 2, 2], stride=[1, 2, 2], padding=0)\n )\n )\n (1): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (2): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (3): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (4): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n (5): BasicBlockD(\n (conv1): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n (conv2): ConvDropoutNormReLU(\n (conv): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (all_modules): Sequential(\n (0): Conv3d(320, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n )\n )\n (nonlin2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n )\n )\n (stages): ModuleList(\n (0): StackedConvBlocks(\n (convs): Sequential(\n (0): ConvDropoutNormReLU(\n (conv): Conv3d(640, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(640, 320, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(320, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n )\n (1): StackedConvBlocks(\n (convs): Sequential(\n (0): ConvDropoutNormReLU(\n (conv): Conv3d(512, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(512, 256, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n )\n (2): StackedConvBlocks(\n (convs): Sequential(\n (0): ConvDropoutNormReLU(\n (conv): Conv3d(256, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(256, 128, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n )\n (3): StackedConvBlocks(\n (convs): Sequential(\n (0): ConvDropoutNormReLU(\n (conv): Conv3d(128, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (norm): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(128, 64, kernel_size=(3, 3, 3), stride=(1, 1, 1), padding=(1, 1, 1))\n (1): InstanceNorm3d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n )\n (4): StackedConvBlocks(\n (convs): Sequential(\n (0): ConvDropoutNormReLU(\n (conv): Conv3d(64, 32, kernel_size=(1, 3, 3), stride=(1, 1, 1), padding=(0, 1, 1))\n (norm): InstanceNorm3d(32, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (nonlin): LeakyReLU(negative_slope=0.01, inplace=True)\n (all_modules): Sequential(\n (0): Conv3d(64, 32, kernel_size=(1, 3, 3), stride=(1, 1, 1), padding=(0, 1, 1))\n (1): InstanceNorm3d(32, eps=1e-05, momentum=0.1, affine=True, track_running_stats=False)\n (2): LeakyReLU(negative_slope=0.01, inplace=True)\n )\n )\n )\n )\n )\n (transpconvs): ModuleList(\n (0): ConvTranspose3d(320, 320, kernel_size=(1, 2, 2), stride=(1, 2, 2))\n (1): ConvTranspose3d(320, 256, kernel_size=(2, 2, 2), stride=(2, 2, 2))\n (2): ConvTranspose3d(256, 128, kernel_size=(2, 2, 2), stride=(2, 2, 2))\n (3): ConvTranspose3d(128, 64, kernel_size=(2, 2, 2), stride=(2, 2, 2))\n (4): ConvTranspose3d(64, 32, kernel_size=(1, 2, 2), stride=(1, 2, 2))\n )\n (seg_layers): ModuleList(\n (0): Conv3d(320, 32, kernel_size=(1, 1, 1), stride=(1, 1, 1))\n (1): Conv3d(256, 32, kernel_size=(1, 1, 1), stride=(1, 1, 1))\n (2): Conv3d(128, 32, kernel_size=(1, 1, 1), stride=(1, 1, 1))\n (3): Conv3d(64, 32, kernel_size=(1, 1, 1), stride=(1, 1, 1))\n (4): Conv3d(32, 32, kernel_size=(1, 1, 1), stride=(1, 1, 1))\n )\n )\n )\n (L1): L1Loss()\n (image_loss): myMaskedMSE(\n (mse): myMSE()\n )\n)",
38
+ "lr_scheduler": "<nnunetv2.training.lr_scheduler.polylr.PolyLRScheduler object at 0x15543c17d590>",
39
+ "my_init_kwargs": "{'plans': {'dataset_name': 'Dataset542_synthrad2025_task2_CBCT_HN_pre_v2r_stitched_masked_both', 'plans_name': 'nnUNetResEncUNetLPlans', 'original_median_spacing_after_transp': [3.0, 1.0, 1.0], 'original_median_shape_after_transp': [87, 308, 309], 'image_reader_writer': 'SimpleITKIO', 'transpose_forward': [0, 1, 2], 'transpose_backward': [0, 1, 2], 'configurations': {'2d': {'data_identifier': 'nnUNetPlans_2d', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 30, 'patch_size': [320, 320], 'median_image_size_in_voxels': [308.0, 309.0], 'spacing': [1.0, 1.0], 'normalization_schemes': ['CTNormalizationClippingSynthrad2025'], 'use_mask_for_norm': [False], 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'architecture': {'network_class_name': 'dynamic_network_architectures.architectures.unet.ResidualEncoderUNet', 'arch_kwargs': {'n_stages': 7, 'features_per_stage': [32, 64, 128, 256, 512, 512, 512], 'conv_op': 'torch.nn.modules.conv.Conv2d', 'kernel_sizes': [[3, 3], [3, 3], [3, 3], [3, 3], [3, 3], [3, 3], [3, 3]], 'strides': [[1, 1], [2, 2], [2, 2], [2, 2], [2, 2], [2, 2], [2, 2]], 'n_blocks_per_stage': [1, 3, 4, 6, 6, 6, 6], 'n_conv_per_stage_decoder': [1, 1, 1, 1, 1, 1], 'conv_bias': True, 'norm_op': 'torch.nn.modules.instancenorm.InstanceNorm2d', 'norm_op_kwargs': {'eps': 1e-05, 'affine': True}, 'dropout_op': None, 'dropout_op_kwargs': None, 'nonlin': 'torch.nn.LeakyReLU', 'nonlin_kwargs': {'inplace': True}}, '_kw_requires_import': ['conv_op', 'norm_op', 'dropout_op', 'nonlin']}, 'batch_dice': True}, '3d_fullres': {'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [48, 192, 192], 'median_image_size_in_voxels': [87.0, 308.0, 309.0], 'spacing': [3.0, 1.0, 1.0], 'normalization_schemes': ['CTNormalizationClippingSynthrad2025'], 'use_mask_for_norm': [False], 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'architecture': {'network_class_name': 'dynamic_network_architectures.architectures.unet.ResidualEncoderUNet', 'arch_kwargs': {'n_stages': 6, 'features_per_stage': [32, 64, 128, 256, 320, 320], 'conv_op': 'torch.nn.modules.conv.Conv3d', 'kernel_sizes': [[1, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'strides': [[1, 1, 1], [1, 2, 2], [2, 2, 2], [2, 2, 2], [2, 2, 2], [1, 2, 2]], 'n_blocks_per_stage': [1, 3, 4, 6, 6, 6], 'n_conv_per_stage_decoder': [1, 1, 1, 1, 1], 'conv_bias': True, 'norm_op': 'torch.nn.modules.instancenorm.InstanceNorm3d', 'norm_op_kwargs': {'eps': 1e-05, 'affine': True}, 'dropout_op': None, 'dropout_op_kwargs': None, 'nonlin': 'torch.nn.LeakyReLU', 'nonlin_kwargs': {'inplace': True}}, '_kw_requires_import': ['conv_op', 'norm_op', 'dropout_op', 'nonlin']}, 'batch_dice': False}}, 'experiment_planner_used': 'nnUNetPlannerResEncL', 'label_manager': 'LabelManager', 'foreground_intensity_properties_per_channel': {'0': {'max': 3071.0, 'mean': -215.1082763671875, 'median': -133.0, 'min': -1024.0, 'percentile_00_5': -1024.0, 'percentile_99_5': 1345.0, 'std': 459.63665771484375}}}, 'configuration': '3d_fullres', 'fold': 0, 'dataset_json': {'labels': {'label_001': '1', 'background': 0}, 'channel_names': {'0': 'CT_zscore_synthrad'}, 'numTraining': 325, 'file_ending': '.mha'}, 'unpack_dataset': True, 'device': device(type='cuda')}",
40
+ "network": "ResidualEncoderUNet",
41
+ "num_epochs": "1000",
42
+ "num_input_channels": "1",
43
+ "num_iterations_per_epoch": "250",
44
+ "num_val_iterations_per_epoch": "50",
45
+ "optimizer": "SGD (\nParameter Group 0\n dampening: 0\n differentiable: False\n foreach: None\n fused: None\n initial_lr: 0.01\n lr: 0.009765692859724779\n maximize: False\n momentum: 0.99\n nesterov: True\n weight_decay: 3e-05\n)",
46
+ "output_folder": "/datasets/work/hb-synthrad2023/work/synthrad2025/bw_workplace/data/nnunet_struct/results/Dataset542_synthrad2025_task2_CBCT_HN_pre_v2r_stitched_masked_both/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0",
47
+ "output_folder_base": "/datasets/work/hb-synthrad2023/work/synthrad2025/bw_workplace/data/nnunet_struct/results/Dataset542_synthrad2025_task2_CBCT_HN_pre_v2r_stitched_masked_both/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres",
48
+ "oversample_foreground_percent": "0.33",
49
+ "perception_masked": "True",
50
+ "plans_manager": "{'dataset_name': 'Dataset542_synthrad2025_task2_CBCT_HN_pre_v2r_stitched_masked_both', 'plans_name': 'nnUNetResEncUNetLPlans', 'original_median_spacing_after_transp': [3.0, 1.0, 1.0], 'original_median_shape_after_transp': [87, 308, 309], 'image_reader_writer': 'SimpleITKIO', 'transpose_forward': [0, 1, 2], 'transpose_backward': [0, 1, 2], 'configurations': {'2d': {'data_identifier': 'nnUNetPlans_2d', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 30, 'patch_size': [320, 320], 'median_image_size_in_voxels': [308.0, 309.0], 'spacing': [1.0, 1.0], 'normalization_schemes': ['CTNormalizationClippingSynthrad2025'], 'use_mask_for_norm': [False], 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'architecture': {'network_class_name': 'dynamic_network_architectures.architectures.unet.ResidualEncoderUNet', 'arch_kwargs': {'n_stages': 7, 'features_per_stage': [32, 64, 128, 256, 512, 512, 512], 'conv_op': 'torch.nn.modules.conv.Conv2d', 'kernel_sizes': [[3, 3], [3, 3], [3, 3], [3, 3], [3, 3], [3, 3], [3, 3]], 'strides': [[1, 1], [2, 2], [2, 2], [2, 2], [2, 2], [2, 2], [2, 2]], 'n_blocks_per_stage': [1, 3, 4, 6, 6, 6, 6], 'n_conv_per_stage_decoder': [1, 1, 1, 1, 1, 1], 'conv_bias': True, 'norm_op': 'torch.nn.modules.instancenorm.InstanceNorm2d', 'norm_op_kwargs': {'eps': 1e-05, 'affine': True}, 'dropout_op': None, 'dropout_op_kwargs': None, 'nonlin': 'torch.nn.LeakyReLU', 'nonlin_kwargs': {'inplace': True}}, '_kw_requires_import': ['conv_op', 'norm_op', 'dropout_op', 'nonlin']}, 'batch_dice': True}, '3d_fullres': {'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [48, 192, 192], 'median_image_size_in_voxels': [87.0, 308.0, 309.0], 'spacing': [3.0, 1.0, 1.0], 'normalization_schemes': ['CTNormalizationClippingSynthrad2025'], 'use_mask_for_norm': [False], 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'architecture': {'network_class_name': 'dynamic_network_architectures.architectures.unet.ResidualEncoderUNet', 'arch_kwargs': {'n_stages': 6, 'features_per_stage': [32, 64, 128, 256, 320, 320], 'conv_op': 'torch.nn.modules.conv.Conv3d', 'kernel_sizes': [[1, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'strides': [[1, 1, 1], [1, 2, 2], [2, 2, 2], [2, 2, 2], [2, 2, 2], [1, 2, 2]], 'n_blocks_per_stage': [1, 3, 4, 6, 6, 6], 'n_conv_per_stage_decoder': [1, 1, 1, 1, 1], 'conv_bias': True, 'norm_op': 'torch.nn.modules.instancenorm.InstanceNorm3d', 'norm_op_kwargs': {'eps': 1e-05, 'affine': True}, 'dropout_op': None, 'dropout_op_kwargs': None, 'nonlin': 'torch.nn.LeakyReLU', 'nonlin_kwargs': {'inplace': True}}, '_kw_requires_import': ['conv_op', 'norm_op', 'dropout_op', 'nonlin']}, 'batch_dice': False}}, 'experiment_planner_used': 'nnUNetPlannerResEncL', 'label_manager': 'LabelManager', 'foreground_intensity_properties_per_channel': {'0': {'max': 3071.0, 'mean': -215.1082763671875, 'median': -133.0, 'min': -1024.0, 'percentile_00_5': -1024.0, 'percentile_99_5': 1345.0, 'std': 459.63665771484375}}}",
51
+ "preprocessed_dataset_folder": "/datasets/work/hb-synthrad2023/work/synthrad2025/bw_workplace/data/nnunet_struct/preprocessed/Dataset542_synthrad2025_task2_CBCT_HN_pre_v2r_stitched_masked_both/nnUNetPlans_3d_fullres",
52
+ "preprocessed_dataset_folder_base": "/datasets/work/hb-synthrad2023/work/synthrad2025/bw_workplace/data/nnunet_struct/preprocessed/Dataset542_synthrad2025_task2_CBCT_HN_pre_v2r_stitched_masked_both",
53
+ "save_every": "50",
54
+ "torch_version": "2.7.1+cu126",
55
+ "unpack_dataset": "True",
56
+ "was_initialized": "True",
57
+ "weight_decay": "3e-05"
58
+ }
Dataset542/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/progress.png ADDED

Git LFS Details

  • SHA256: 5d87d6b783b54c578c4fb7db56c814d908815889146a7c6aad34673d1790dadc
  • Pointer size: 131 Bytes
  • Size of remote file: 491 kB
Dataset542/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/validation/summary.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "mae": {
3
+ "mean": 59.6831403703124,
4
+ "std": 13.282268575250406,
5
+ "max": 135.26334083952912,
6
+ "min": 43.70561993673689,
7
+ "25pc": 50.90544320106635,
8
+ "50pc": 57.59766416973342,
9
+ "75pc": 63.61140017268492,
10
+ "count": 65
11
+ },
12
+ "psnr": {
13
+ "mean": 30.41879186477032,
14
+ "std": 1.7822310331910258,
15
+ "max": 33.303479514662435,
16
+ "min": 21.938598217631913,
17
+ "25pc": 29.489967207335592,
18
+ "50pc": 30.549653942101994,
19
+ "75pc": 31.678537891293097,
20
+ "count": 65
21
+ },
22
+ "ms_ssim": {
23
+ "mean": 0.972537866604284,
24
+ "std": 0.023400539772748585,
25
+ "max": 0.9896922348656185,
26
+ "min": 0.8256950304841117,
27
+ "25pc": 0.972346153083238,
28
+ "50pc": 0.9793511650351168,
29
+ "75pc": 0.9834134511089297,
30
+ "count": 65
31
+ },
32
+ "DICE": {
33
+ "mean": 0.8757181415191063,
34
+ "std": 0.04152858997642459,
35
+ "max": 0.9331912994384766,
36
+ "min": 0.7493325471878052,
37
+ "25pc": 0.8536831736564636,
38
+ "50pc": 0.885571300983429,
39
+ "75pc": 0.907253623008728,
40
+ "count": 65
41
+ },
42
+ "HD95": {
43
+ "mean": 1.9038109889397254,
44
+ "std": 0.8689346520839407,
45
+ "max": 4.7832865715026855,
46
+ "min": 1.1818181276321411,
47
+ "25pc": 1.3928370475769043,
48
+ "50pc": 1.54265558719635,
49
+ "75pc": 2.0603861808776855,
50
+ "count": 65
51
+ }
52
+ }
Dataset542/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/plans.json ADDED
@@ -0,0 +1,356 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_name": "Dataset542_synthrad2025_task2_CBCT_HN_pre_v2r_stitched_masked_both",
3
+ "plans_name": "nnUNetResEncUNetLPlans",
4
+ "original_median_spacing_after_transp": [
5
+ 3.0,
6
+ 1.0,
7
+ 1.0
8
+ ],
9
+ "original_median_shape_after_transp": [
10
+ 87,
11
+ 308,
12
+ 309
13
+ ],
14
+ "image_reader_writer": "SimpleITKIO",
15
+ "transpose_forward": [
16
+ 0,
17
+ 1,
18
+ 2
19
+ ],
20
+ "transpose_backward": [
21
+ 0,
22
+ 1,
23
+ 2
24
+ ],
25
+ "configurations": {
26
+ "2d": {
27
+ "data_identifier": "nnUNetPlans_2d",
28
+ "preprocessor_name": "DefaultPreprocessor",
29
+ "batch_size": 30,
30
+ "patch_size": [
31
+ 320,
32
+ 320
33
+ ],
34
+ "median_image_size_in_voxels": [
35
+ 308.0,
36
+ 309.0
37
+ ],
38
+ "spacing": [
39
+ 1.0,
40
+ 1.0
41
+ ],
42
+ "normalization_schemes": [
43
+ "CTNormalizationClippingSynthrad2025"
44
+ ],
45
+ "use_mask_for_norm": [
46
+ false
47
+ ],
48
+ "resampling_fn_data": "resample_data_or_seg_to_shape",
49
+ "resampling_fn_seg": "resample_data_or_seg_to_shape",
50
+ "resampling_fn_data_kwargs": {
51
+ "is_seg": false,
52
+ "order": 3,
53
+ "order_z": 0,
54
+ "force_separate_z": null
55
+ },
56
+ "resampling_fn_seg_kwargs": {
57
+ "is_seg": true,
58
+ "order": 1,
59
+ "order_z": 0,
60
+ "force_separate_z": null
61
+ },
62
+ "resampling_fn_probabilities": "resample_data_or_seg_to_shape",
63
+ "resampling_fn_probabilities_kwargs": {
64
+ "is_seg": false,
65
+ "order": 1,
66
+ "order_z": 0,
67
+ "force_separate_z": null
68
+ },
69
+ "architecture": {
70
+ "network_class_name": "dynamic_network_architectures.architectures.unet.ResidualEncoderUNet",
71
+ "arch_kwargs": {
72
+ "n_stages": 7,
73
+ "features_per_stage": [
74
+ 32,
75
+ 64,
76
+ 128,
77
+ 256,
78
+ 512,
79
+ 512,
80
+ 512
81
+ ],
82
+ "conv_op": "torch.nn.modules.conv.Conv2d",
83
+ "kernel_sizes": [
84
+ [
85
+ 3,
86
+ 3
87
+ ],
88
+ [
89
+ 3,
90
+ 3
91
+ ],
92
+ [
93
+ 3,
94
+ 3
95
+ ],
96
+ [
97
+ 3,
98
+ 3
99
+ ],
100
+ [
101
+ 3,
102
+ 3
103
+ ],
104
+ [
105
+ 3,
106
+ 3
107
+ ],
108
+ [
109
+ 3,
110
+ 3
111
+ ]
112
+ ],
113
+ "strides": [
114
+ [
115
+ 1,
116
+ 1
117
+ ],
118
+ [
119
+ 2,
120
+ 2
121
+ ],
122
+ [
123
+ 2,
124
+ 2
125
+ ],
126
+ [
127
+ 2,
128
+ 2
129
+ ],
130
+ [
131
+ 2,
132
+ 2
133
+ ],
134
+ [
135
+ 2,
136
+ 2
137
+ ],
138
+ [
139
+ 2,
140
+ 2
141
+ ]
142
+ ],
143
+ "n_blocks_per_stage": [
144
+ 1,
145
+ 3,
146
+ 4,
147
+ 6,
148
+ 6,
149
+ 6,
150
+ 6
151
+ ],
152
+ "n_conv_per_stage_decoder": [
153
+ 1,
154
+ 1,
155
+ 1,
156
+ 1,
157
+ 1,
158
+ 1
159
+ ],
160
+ "conv_bias": true,
161
+ "norm_op": "torch.nn.modules.instancenorm.InstanceNorm2d",
162
+ "norm_op_kwargs": {
163
+ "eps": 1e-05,
164
+ "affine": true
165
+ },
166
+ "dropout_op": null,
167
+ "dropout_op_kwargs": null,
168
+ "nonlin": "torch.nn.LeakyReLU",
169
+ "nonlin_kwargs": {
170
+ "inplace": true
171
+ }
172
+ },
173
+ "_kw_requires_import": [
174
+ "conv_op",
175
+ "norm_op",
176
+ "dropout_op",
177
+ "nonlin"
178
+ ]
179
+ },
180
+ "batch_dice": true
181
+ },
182
+ "3d_fullres": {
183
+ "data_identifier": "nnUNetPlans_3d_fullres",
184
+ "preprocessor_name": "DefaultPreprocessor",
185
+ "batch_size": 2,
186
+ "patch_size": [
187
+ 48,
188
+ 192,
189
+ 192
190
+ ],
191
+ "median_image_size_in_voxels": [
192
+ 87.0,
193
+ 308.0,
194
+ 309.0
195
+ ],
196
+ "spacing": [
197
+ 3.0,
198
+ 1.0,
199
+ 1.0
200
+ ],
201
+ "normalization_schemes": [
202
+ "CTNormalizationClippingSynthrad2025"
203
+ ],
204
+ "use_mask_for_norm": [
205
+ false
206
+ ],
207
+ "resampling_fn_data": "resample_data_or_seg_to_shape",
208
+ "resampling_fn_seg": "resample_data_or_seg_to_shape",
209
+ "resampling_fn_data_kwargs": {
210
+ "is_seg": false,
211
+ "order": 3,
212
+ "order_z": 0,
213
+ "force_separate_z": null
214
+ },
215
+ "resampling_fn_seg_kwargs": {
216
+ "is_seg": true,
217
+ "order": 1,
218
+ "order_z": 0,
219
+ "force_separate_z": null
220
+ },
221
+ "resampling_fn_probabilities": "resample_data_or_seg_to_shape",
222
+ "resampling_fn_probabilities_kwargs": {
223
+ "is_seg": false,
224
+ "order": 1,
225
+ "order_z": 0,
226
+ "force_separate_z": null
227
+ },
228
+ "architecture": {
229
+ "network_class_name": "dynamic_network_architectures.architectures.unet.ResidualEncoderUNet",
230
+ "arch_kwargs": {
231
+ "n_stages": 6,
232
+ "features_per_stage": [
233
+ 32,
234
+ 64,
235
+ 128,
236
+ 256,
237
+ 320,
238
+ 320
239
+ ],
240
+ "conv_op": "torch.nn.modules.conv.Conv3d",
241
+ "kernel_sizes": [
242
+ [
243
+ 1,
244
+ 3,
245
+ 3
246
+ ],
247
+ [
248
+ 3,
249
+ 3,
250
+ 3
251
+ ],
252
+ [
253
+ 3,
254
+ 3,
255
+ 3
256
+ ],
257
+ [
258
+ 3,
259
+ 3,
260
+ 3
261
+ ],
262
+ [
263
+ 3,
264
+ 3,
265
+ 3
266
+ ],
267
+ [
268
+ 3,
269
+ 3,
270
+ 3
271
+ ]
272
+ ],
273
+ "strides": [
274
+ [
275
+ 1,
276
+ 1,
277
+ 1
278
+ ],
279
+ [
280
+ 1,
281
+ 2,
282
+ 2
283
+ ],
284
+ [
285
+ 2,
286
+ 2,
287
+ 2
288
+ ],
289
+ [
290
+ 2,
291
+ 2,
292
+ 2
293
+ ],
294
+ [
295
+ 2,
296
+ 2,
297
+ 2
298
+ ],
299
+ [
300
+ 1,
301
+ 2,
302
+ 2
303
+ ]
304
+ ],
305
+ "n_blocks_per_stage": [
306
+ 1,
307
+ 3,
308
+ 4,
309
+ 6,
310
+ 6,
311
+ 6
312
+ ],
313
+ "n_conv_per_stage_decoder": [
314
+ 1,
315
+ 1,
316
+ 1,
317
+ 1,
318
+ 1
319
+ ],
320
+ "conv_bias": true,
321
+ "norm_op": "torch.nn.modules.instancenorm.InstanceNorm3d",
322
+ "norm_op_kwargs": {
323
+ "eps": 1e-05,
324
+ "affine": true
325
+ },
326
+ "dropout_op": null,
327
+ "dropout_op_kwargs": null,
328
+ "nonlin": "torch.nn.LeakyReLU",
329
+ "nonlin_kwargs": {
330
+ "inplace": true
331
+ }
332
+ },
333
+ "_kw_requires_import": [
334
+ "conv_op",
335
+ "norm_op",
336
+ "dropout_op",
337
+ "nonlin"
338
+ ]
339
+ },
340
+ "batch_dice": false
341
+ }
342
+ },
343
+ "experiment_planner_used": "nnUNetPlannerResEncL",
344
+ "label_manager": "LabelManager",
345
+ "foreground_intensity_properties_per_channel": {
346
+ "0": {
347
+ "max": 3071.0,
348
+ "mean": -215.1082763671875,
349
+ "median": -133.0,
350
+ "min": -1024.0,
351
+ "percentile_00_5": -1024.0,
352
+ "percentile_99_5": 1345.0,
353
+ "std": 459.63665771484375
354
+ }
355
+ }
356
+ }
Dataset544/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/dataset.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "labels": {
3
+ "label_001": "1",
4
+ "background": 0
5
+ },
6
+ "channel_names": {
7
+ "0": "CT_zscore_synthrad"
8
+ },
9
+ "numTraining": 321,
10
+ "file_ending": ".nii.gz",
11
+ "test_set": [
12
+ "case_0000"
13
+ ]
14
+ }
Dataset544/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/dataset_fingerprint.json ADDED
@@ -0,0 +1,3228 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "foreground_intensity_properties_per_channel": {
3
+ "0": {
4
+ "max": 3071.0,
5
+ "mean": -330.8976135253906,
6
+ "median": -243.0,
7
+ "min": -1024.0,
8
+ "percentile_00_5": -1024.0,
9
+ "percentile_99_5": 563.0,
10
+ "std": 371.2642822265625
11
+ }
12
+ },
13
+ "median_relative_size_after_cropping": 1.0,
14
+ "shapes_after_crop": [
15
+ [
16
+ 81,
17
+ 311,
18
+ 310
19
+ ],
20
+ [
21
+ 78,
22
+ 307,
23
+ 308
24
+ ],
25
+ [
26
+ 80,
27
+ 359,
28
+ 451
29
+ ],
30
+ [
31
+ 81,
32
+ 450,
33
+ 450
34
+ ],
35
+ [
36
+ 88,
37
+ 310,
38
+ 311
39
+ ],
40
+ [
41
+ 81,
42
+ 377,
43
+ 451
44
+ ],
45
+ [
46
+ 87,
47
+ 309,
48
+ 309
49
+ ],
50
+ [
51
+ 80,
52
+ 370,
53
+ 451
54
+ ],
55
+ [
56
+ 81,
57
+ 370,
58
+ 451
59
+ ],
60
+ [
61
+ 80,
62
+ 449,
63
+ 453
64
+ ],
65
+ [
66
+ 81,
67
+ 380,
68
+ 448
69
+ ],
70
+ [
71
+ 87,
72
+ 311,
73
+ 311
74
+ ],
75
+ [
76
+ 80,
77
+ 429,
78
+ 448
79
+ ],
80
+ [
81
+ 87,
82
+ 311,
83
+ 307
84
+ ],
85
+ [
86
+ 86,
87
+ 316,
88
+ 311
89
+ ],
90
+ [
91
+ 87,
92
+ 308,
93
+ 311
94
+ ],
95
+ [
96
+ 87,
97
+ 312,
98
+ 308
99
+ ],
100
+ [
101
+ 88,
102
+ 307,
103
+ 308
104
+ ],
105
+ [
106
+ 86,
107
+ 314,
108
+ 310
109
+ ],
110
+ [
111
+ 79,
112
+ 312,
113
+ 310
114
+ ],
115
+ [
116
+ 87,
117
+ 308,
118
+ 311
119
+ ],
120
+ [
121
+ 87,
122
+ 310,
123
+ 312
124
+ ],
125
+ [
126
+ 85,
127
+ 310,
128
+ 319
129
+ ],
130
+ [
131
+ 87,
132
+ 310,
133
+ 309
134
+ ],
135
+ [
136
+ 87,
137
+ 310,
138
+ 309
139
+ ],
140
+ [
141
+ 81,
142
+ 308,
143
+ 309
144
+ ],
145
+ [
146
+ 81,
147
+ 290,
148
+ 285
149
+ ],
150
+ [
151
+ 88,
152
+ 288,
153
+ 291
154
+ ],
155
+ [
156
+ 88,
157
+ 310,
158
+ 310
159
+ ],
160
+ [
161
+ 88,
162
+ 308,
163
+ 297
164
+ ],
165
+ [
166
+ 87,
167
+ 310,
168
+ 309
169
+ ],
170
+ [
171
+ 86,
172
+ 316,
173
+ 309
174
+ ],
175
+ [
176
+ 84,
177
+ 308,
178
+ 309
179
+ ],
180
+ [
181
+ 87,
182
+ 310,
183
+ 311
184
+ ],
185
+ [
186
+ 87,
187
+ 311,
188
+ 308
189
+ ],
190
+ [
191
+ 86,
192
+ 314,
193
+ 309
194
+ ],
195
+ [
196
+ 88,
197
+ 310,
198
+ 299
199
+ ],
200
+ [
201
+ 87,
202
+ 312,
203
+ 309
204
+ ],
205
+ [
206
+ 88,
207
+ 309,
208
+ 309
209
+ ],
210
+ [
211
+ 87,
212
+ 310,
213
+ 289
214
+ ],
215
+ [
216
+ 88,
217
+ 308,
218
+ 309
219
+ ],
220
+ [
221
+ 74,
222
+ 308,
223
+ 313
224
+ ],
225
+ [
226
+ 88,
227
+ 308,
228
+ 309
229
+ ],
230
+ [
231
+ 82,
232
+ 310,
233
+ 308
234
+ ],
235
+ [
236
+ 88,
237
+ 309,
238
+ 307
239
+ ],
240
+ [
241
+ 88,
242
+ 311,
243
+ 309
244
+ ],
245
+ [
246
+ 67,
247
+ 309,
248
+ 312
249
+ ],
250
+ [
251
+ 86,
252
+ 314,
253
+ 309
254
+ ],
255
+ [
256
+ 87,
257
+ 309,
258
+ 313
259
+ ],
260
+ [
261
+ 80,
262
+ 296,
263
+ 286
264
+ ],
265
+ [
266
+ 87,
267
+ 308,
268
+ 309
269
+ ],
270
+ [
271
+ 81,
272
+ 309,
273
+ 311
274
+ ],
275
+ [
276
+ 88,
277
+ 308,
278
+ 307
279
+ ],
280
+ [
281
+ 87,
282
+ 307,
283
+ 309
284
+ ],
285
+ [
286
+ 86,
287
+ 308,
288
+ 313
289
+ ],
290
+ [
291
+ 86,
292
+ 314,
293
+ 309
294
+ ],
295
+ [
296
+ 87,
297
+ 312,
298
+ 308
299
+ ],
300
+ [
301
+ 87,
302
+ 311,
303
+ 311
304
+ ],
305
+ [
306
+ 87,
307
+ 311,
308
+ 309
309
+ ],
310
+ [
311
+ 86,
312
+ 313,
313
+ 312
314
+ ],
315
+ [
316
+ 86,
317
+ 317,
318
+ 311
319
+ ],
320
+ [
321
+ 87,
322
+ 310,
323
+ 312
324
+ ],
325
+ [
326
+ 87,
327
+ 313,
328
+ 310
329
+ ],
330
+ [
331
+ 87,
332
+ 312,
333
+ 309
334
+ ],
335
+ [
336
+ 76,
337
+ 307,
338
+ 293
339
+ ],
340
+ [
341
+ 78,
342
+ 450,
343
+ 434
344
+ ],
345
+ [
346
+ 79,
347
+ 450,
348
+ 456
349
+ ],
350
+ [
351
+ 77,
352
+ 425,
353
+ 446
354
+ ],
355
+ [
356
+ 76,
357
+ 452,
358
+ 432
359
+ ],
360
+ [
361
+ 82,
362
+ 449,
363
+ 449
364
+ ],
365
+ [
366
+ 81,
367
+ 451,
368
+ 448
369
+ ],
370
+ [
371
+ 80,
372
+ 449,
373
+ 425
374
+ ],
375
+ [
376
+ 81,
377
+ 452,
378
+ 451
379
+ ],
380
+ [
381
+ 79,
382
+ 449,
383
+ 455
384
+ ],
385
+ [
386
+ 73,
387
+ 440,
388
+ 456
389
+ ],
390
+ [
391
+ 77,
392
+ 452,
393
+ 442
394
+ ],
395
+ [
396
+ 80,
397
+ 449,
398
+ 453
399
+ ],
400
+ [
401
+ 78,
402
+ 450,
403
+ 436
404
+ ],
405
+ [
406
+ 80,
407
+ 451,
408
+ 454
409
+ ],
410
+ [
411
+ 81,
412
+ 451,
413
+ 450
414
+ ],
415
+ [
416
+ 80,
417
+ 452,
418
+ 453
419
+ ],
420
+ [
421
+ 80,
422
+ 452,
423
+ 449
424
+ ],
425
+ [
426
+ 74,
427
+ 454,
428
+ 433
429
+ ],
430
+ [
431
+ 79,
432
+ 449,
433
+ 455
434
+ ],
435
+ [
436
+ 81,
437
+ 448,
438
+ 452
439
+ ],
440
+ [
441
+ 80,
442
+ 453,
443
+ 450
444
+ ],
445
+ [
446
+ 80,
447
+ 449,
448
+ 449
449
+ ],
450
+ [
451
+ 81,
452
+ 451,
453
+ 450
454
+ ],
455
+ [
456
+ 78,
457
+ 437,
458
+ 438
459
+ ],
460
+ [
461
+ 79,
462
+ 454,
463
+ 429
464
+ ],
465
+ [
466
+ 77,
467
+ 454,
468
+ 457
469
+ ],
470
+ [
471
+ 81,
472
+ 451,
473
+ 433
474
+ ],
475
+ [
476
+ 79,
477
+ 455,
478
+ 450
479
+ ],
480
+ [
481
+ 79,
482
+ 457,
483
+ 452
484
+ ],
485
+ [
486
+ 82,
487
+ 451,
488
+ 449
489
+ ],
490
+ [
491
+ 80,
492
+ 449,
493
+ 455
494
+ ],
495
+ [
496
+ 81,
497
+ 450,
498
+ 449
499
+ ],
500
+ [
501
+ 79,
502
+ 447,
503
+ 448
504
+ ],
505
+ [
506
+ 80,
507
+ 450,
508
+ 452
509
+ ],
510
+ [
511
+ 81,
512
+ 449,
513
+ 451
514
+ ],
515
+ [
516
+ 81,
517
+ 450,
518
+ 446
519
+ ],
520
+ [
521
+ 81,
522
+ 451,
523
+ 450
524
+ ],
525
+ [
526
+ 62,
527
+ 451,
528
+ 449
529
+ ],
530
+ [
531
+ 80,
532
+ 449,
533
+ 454
534
+ ],
535
+ [
536
+ 81,
537
+ 450,
538
+ 450
539
+ ],
540
+ [
541
+ 80,
542
+ 451,
543
+ 453
544
+ ],
545
+ [
546
+ 82,
547
+ 450,
548
+ 447
549
+ ],
550
+ [
551
+ 82,
552
+ 449,
553
+ 449
554
+ ],
555
+ [
556
+ 80,
557
+ 450,
558
+ 425
559
+ ],
560
+ [
561
+ 78,
562
+ 452,
563
+ 457
564
+ ],
565
+ [
566
+ 80,
567
+ 430,
568
+ 423
569
+ ],
570
+ [
571
+ 77,
572
+ 448,
573
+ 429
574
+ ],
575
+ [
576
+ 78,
577
+ 449,
578
+ 458
579
+ ],
580
+ [
581
+ 80,
582
+ 449,
583
+ 451
584
+ ],
585
+ [
586
+ 81,
587
+ 451,
588
+ 449
589
+ ],
590
+ [
591
+ 81,
592
+ 452,
593
+ 442
594
+ ],
595
+ [
596
+ 79,
597
+ 454,
598
+ 456
599
+ ],
600
+ [
601
+ 79,
602
+ 450,
603
+ 402
604
+ ],
605
+ [
606
+ 79,
607
+ 450,
608
+ 456
609
+ ],
610
+ [
611
+ 81,
612
+ 452,
613
+ 449
614
+ ],
615
+ [
616
+ 80,
617
+ 453,
618
+ 449
619
+ ],
620
+ [
621
+ 77,
622
+ 457,
623
+ 453
624
+ ],
625
+ [
626
+ 82,
627
+ 450,
628
+ 449
629
+ ],
630
+ [
631
+ 79,
632
+ 450,
633
+ 442
634
+ ],
635
+ [
636
+ 81,
637
+ 450,
638
+ 450
639
+ ],
640
+ [
641
+ 80,
642
+ 449,
643
+ 432
644
+ ],
645
+ [
646
+ 81,
647
+ 448,
648
+ 454
649
+ ],
650
+ [
651
+ 70,
652
+ 451,
653
+ 457
654
+ ],
655
+ [
656
+ 80,
657
+ 451,
658
+ 454
659
+ ],
660
+ [
661
+ 80,
662
+ 425,
663
+ 449
664
+ ],
665
+ [
666
+ 87,
667
+ 310,
668
+ 298
669
+ ],
670
+ [
671
+ 87,
672
+ 289,
673
+ 309
674
+ ],
675
+ [
676
+ 87,
677
+ 293,
678
+ 309
679
+ ],
680
+ [
681
+ 86,
682
+ 307,
683
+ 314
684
+ ],
685
+ [
686
+ 88,
687
+ 308,
688
+ 309
689
+ ],
690
+ [
691
+ 88,
692
+ 309,
693
+ 308
694
+ ],
695
+ [
696
+ 75,
697
+ 449,
698
+ 462
699
+ ],
700
+ [
701
+ 88,
702
+ 309,
703
+ 309
704
+ ],
705
+ [
706
+ 87,
707
+ 311,
708
+ 308
709
+ ],
710
+ [
711
+ 87,
712
+ 278,
713
+ 315
714
+ ],
715
+ [
716
+ 87,
717
+ 308,
718
+ 312
719
+ ],
720
+ [
721
+ 81,
722
+ 450,
723
+ 451
724
+ ],
725
+ [
726
+ 88,
727
+ 289,
728
+ 308
729
+ ],
730
+ [
731
+ 85,
732
+ 317,
733
+ 314
734
+ ],
735
+ [
736
+ 86,
737
+ 311,
738
+ 315
739
+ ],
740
+ [
741
+ 82,
742
+ 421,
743
+ 450
744
+ ],
745
+ [
746
+ 87,
747
+ 309,
748
+ 284
749
+ ],
750
+ [
751
+ 81,
752
+ 450,
753
+ 447
754
+ ],
755
+ [
756
+ 88,
757
+ 311,
758
+ 311
759
+ ],
760
+ [
761
+ 88,
762
+ 311,
763
+ 312
764
+ ],
765
+ [
766
+ 81,
767
+ 422,
768
+ 394
769
+ ],
770
+ [
771
+ 87,
772
+ 288,
773
+ 312
774
+ ],
775
+ [
776
+ 87,
777
+ 312,
778
+ 300
779
+ ],
780
+ [
781
+ 81,
782
+ 452,
783
+ 451
784
+ ],
785
+ [
786
+ 87,
787
+ 311,
788
+ 311
789
+ ],
790
+ [
791
+ 84,
792
+ 316,
793
+ 318
794
+ ],
795
+ [
796
+ 80,
797
+ 451,
798
+ 452
799
+ ],
800
+ [
801
+ 87,
802
+ 289,
803
+ 311
804
+ ],
805
+ [
806
+ 81,
807
+ 450,
808
+ 451
809
+ ],
810
+ [
811
+ 86,
812
+ 316,
813
+ 309
814
+ ],
815
+ [
816
+ 81,
817
+ 448,
818
+ 449
819
+ ],
820
+ [
821
+ 86,
822
+ 310,
823
+ 302
824
+ ],
825
+ [
826
+ 87,
827
+ 314,
828
+ 310
829
+ ],
830
+ [
831
+ 81,
832
+ 450,
833
+ 448
834
+ ],
835
+ [
836
+ 88,
837
+ 309,
838
+ 309
839
+ ],
840
+ [
841
+ 86,
842
+ 313,
843
+ 314
844
+ ],
845
+ [
846
+ 80,
847
+ 448,
848
+ 453
849
+ ],
850
+ [
851
+ 87,
852
+ 308,
853
+ 314
854
+ ],
855
+ [
856
+ 84,
857
+ 320,
858
+ 310
859
+ ],
860
+ [
861
+ 81,
862
+ 451,
863
+ 451
864
+ ],
865
+ [
866
+ 88,
867
+ 310,
868
+ 306
869
+ ],
870
+ [
871
+ 82,
872
+ 308,
873
+ 309
874
+ ],
875
+ [
876
+ 87,
877
+ 290,
878
+ 312
879
+ ],
880
+ [
881
+ 88,
882
+ 289,
883
+ 280
884
+ ],
885
+ [
886
+ 88,
887
+ 309,
888
+ 308
889
+ ],
890
+ [
891
+ 75,
892
+ 427,
893
+ 452
894
+ ],
895
+ [
896
+ 80,
897
+ 435,
898
+ 433
899
+ ],
900
+ [
901
+ 86,
902
+ 294,
903
+ 310
904
+ ],
905
+ [
906
+ 88,
907
+ 310,
908
+ 304
909
+ ],
910
+ [
911
+ 80,
912
+ 453,
913
+ 451
914
+ ],
915
+ [
916
+ 86,
917
+ 297,
918
+ 315
919
+ ],
920
+ [
921
+ 85,
922
+ 313,
923
+ 314
924
+ ],
925
+ [
926
+ 87,
927
+ 311,
928
+ 308
929
+ ],
930
+ [
931
+ 88,
932
+ 307,
933
+ 309
934
+ ],
935
+ [
936
+ 87,
937
+ 308,
938
+ 311
939
+ ],
940
+ [
941
+ 84,
942
+ 321,
943
+ 315
944
+ ],
945
+ [
946
+ 87,
947
+ 309,
948
+ 310
949
+ ],
950
+ [
951
+ 85,
952
+ 309,
953
+ 299
954
+ ],
955
+ [
956
+ 81,
957
+ 309,
958
+ 314
959
+ ],
960
+ [
961
+ 74,
962
+ 309,
963
+ 308
964
+ ],
965
+ [
966
+ 86,
967
+ 297,
968
+ 313
969
+ ],
970
+ [
971
+ 87,
972
+ 310,
973
+ 298
974
+ ],
975
+ [
976
+ 88,
977
+ 310,
978
+ 310
979
+ ],
980
+ [
981
+ 99,
982
+ 308,
983
+ 308
984
+ ],
985
+ [
986
+ 128,
987
+ 384,
988
+ 392
989
+ ],
990
+ [
991
+ 70,
992
+ 392,
993
+ 393
994
+ ],
995
+ [
996
+ 128,
997
+ 392,
998
+ 393
999
+ ],
1000
+ [
1001
+ 71,
1002
+ 472,
1003
+ 499
1004
+ ],
1005
+ [
1006
+ 71,
1007
+ 463,
1008
+ 499
1009
+ ],
1010
+ [
1011
+ 75,
1012
+ 468,
1013
+ 494
1014
+ ],
1015
+ [
1016
+ 130,
1017
+ 471,
1018
+ 499
1019
+ ],
1020
+ [
1021
+ 133,
1022
+ 479,
1023
+ 499
1024
+ ],
1025
+ [
1026
+ 114,
1027
+ 394,
1028
+ 406
1029
+ ],
1030
+ [
1031
+ 137,
1032
+ 469,
1033
+ 499
1034
+ ],
1035
+ [
1036
+ 128,
1037
+ 467,
1038
+ 499
1039
+ ],
1040
+ [
1041
+ 70,
1042
+ 466,
1043
+ 498
1044
+ ],
1045
+ [
1046
+ 99,
1047
+ 310,
1048
+ 311
1049
+ ],
1050
+ [
1051
+ 131,
1052
+ 477,
1053
+ 499
1054
+ ],
1055
+ [
1056
+ 71,
1057
+ 467,
1058
+ 499
1059
+ ],
1060
+ [
1061
+ 74,
1062
+ 499,
1063
+ 496
1064
+ ],
1065
+ [
1066
+ 132,
1067
+ 469,
1068
+ 499
1069
+ ],
1070
+ [
1071
+ 99,
1072
+ 308,
1073
+ 307
1074
+ ],
1075
+ [
1076
+ 130,
1077
+ 472,
1078
+ 499
1079
+ ],
1080
+ [
1081
+ 93,
1082
+ 310,
1083
+ 309
1084
+ ],
1085
+ [
1086
+ 128,
1087
+ 465,
1088
+ 499
1089
+ ],
1090
+ [
1091
+ 98,
1092
+ 308,
1093
+ 315
1094
+ ],
1095
+ [
1096
+ 71,
1097
+ 467,
1098
+ 499
1099
+ ],
1100
+ [
1101
+ 91,
1102
+ 312,
1103
+ 307
1104
+ ],
1105
+ [
1106
+ 71,
1107
+ 468,
1108
+ 499
1109
+ ],
1110
+ [
1111
+ 75,
1112
+ 464,
1113
+ 499
1114
+ ],
1115
+ [
1116
+ 72,
1117
+ 469,
1118
+ 499
1119
+ ],
1120
+ [
1121
+ 123,
1122
+ 499,
1123
+ 499
1124
+ ],
1125
+ [
1126
+ 72,
1127
+ 469,
1128
+ 499
1129
+ ],
1130
+ [
1131
+ 131,
1132
+ 499,
1133
+ 499
1134
+ ],
1135
+ [
1136
+ 131,
1137
+ 467,
1138
+ 499
1139
+ ],
1140
+ [
1141
+ 71,
1142
+ 465,
1143
+ 499
1144
+ ],
1145
+ [
1146
+ 139,
1147
+ 467,
1148
+ 499
1149
+ ],
1150
+ [
1151
+ 137,
1152
+ 466,
1153
+ 499
1154
+ ],
1155
+ [
1156
+ 71,
1157
+ 465,
1158
+ 495
1159
+ ],
1160
+ [
1161
+ 129,
1162
+ 499,
1163
+ 499
1164
+ ],
1165
+ [
1166
+ 70,
1167
+ 466,
1168
+ 494
1169
+ ],
1170
+ [
1171
+ 133,
1172
+ 473,
1173
+ 499
1174
+ ],
1175
+ [
1176
+ 131,
1177
+ 468,
1178
+ 499
1179
+ ],
1180
+ [
1181
+ 131,
1182
+ 471,
1183
+ 499
1184
+ ],
1185
+ [
1186
+ 128,
1187
+ 467,
1188
+ 499
1189
+ ],
1190
+ [
1191
+ 131,
1192
+ 473,
1193
+ 499
1194
+ ],
1195
+ [
1196
+ 122,
1197
+ 499,
1198
+ 498
1199
+ ],
1200
+ [
1201
+ 132,
1202
+ 446,
1203
+ 499
1204
+ ],
1205
+ [
1206
+ 73,
1207
+ 468,
1208
+ 499
1209
+ ],
1210
+ [
1211
+ 99,
1212
+ 316,
1213
+ 308
1214
+ ],
1215
+ [
1216
+ 133,
1217
+ 474,
1218
+ 499
1219
+ ],
1220
+ [
1221
+ 131,
1222
+ 499,
1223
+ 487
1224
+ ],
1225
+ [
1226
+ 130,
1227
+ 470,
1228
+ 496
1229
+ ],
1230
+ [
1231
+ 79,
1232
+ 474,
1233
+ 499
1234
+ ],
1235
+ [
1236
+ 70,
1237
+ 467,
1238
+ 499
1239
+ ],
1240
+ [
1241
+ 128,
1242
+ 472,
1243
+ 499
1244
+ ],
1245
+ [
1246
+ 71,
1247
+ 471,
1248
+ 499
1249
+ ],
1250
+ [
1251
+ 72,
1252
+ 540,
1253
+ 539
1254
+ ],
1255
+ [
1256
+ 79,
1257
+ 467,
1258
+ 499
1259
+ ],
1260
+ [
1261
+ 75,
1262
+ 468,
1263
+ 492
1264
+ ],
1265
+ [
1266
+ 74,
1267
+ 470,
1268
+ 492
1269
+ ],
1270
+ [
1271
+ 72,
1272
+ 468,
1273
+ 499
1274
+ ],
1275
+ [
1276
+ 71,
1277
+ 466,
1278
+ 499
1279
+ ],
1280
+ [
1281
+ 132,
1282
+ 470,
1283
+ 499
1284
+ ],
1285
+ [
1286
+ 131,
1287
+ 472,
1288
+ 499
1289
+ ],
1290
+ [
1291
+ 130,
1292
+ 469,
1293
+ 499
1294
+ ],
1295
+ [
1296
+ 67,
1297
+ 497,
1298
+ 499
1299
+ ],
1300
+ [
1301
+ 66,
1302
+ 477,
1303
+ 497
1304
+ ],
1305
+ [
1306
+ 70,
1307
+ 478,
1308
+ 502
1309
+ ],
1310
+ [
1311
+ 69,
1312
+ 499,
1313
+ 500
1314
+ ],
1315
+ [
1316
+ 69,
1317
+ 497,
1318
+ 486
1319
+ ],
1320
+ [
1321
+ 66,
1322
+ 496,
1323
+ 498
1324
+ ],
1325
+ [
1326
+ 67,
1327
+ 449,
1328
+ 498
1329
+ ],
1330
+ [
1331
+ 71,
1332
+ 478,
1333
+ 503
1334
+ ],
1335
+ [
1336
+ 66,
1337
+ 496,
1338
+ 497
1339
+ ],
1340
+ [
1341
+ 68,
1342
+ 499,
1343
+ 499
1344
+ ],
1345
+ [
1346
+ 55,
1347
+ 480,
1348
+ 501
1349
+ ],
1350
+ [
1351
+ 66,
1352
+ 497,
1353
+ 498
1354
+ ],
1355
+ [
1356
+ 60,
1357
+ 477,
1358
+ 500
1359
+ ],
1360
+ [
1361
+ 69,
1362
+ 482,
1363
+ 479
1364
+ ],
1365
+ [
1366
+ 70,
1367
+ 497,
1368
+ 503
1369
+ ],
1370
+ [
1371
+ 68,
1372
+ 500,
1373
+ 495
1374
+ ],
1375
+ [
1376
+ 65,
1377
+ 500,
1378
+ 499
1379
+ ],
1380
+ [
1381
+ 70,
1382
+ 477,
1383
+ 502
1384
+ ],
1385
+ [
1386
+ 68,
1387
+ 499,
1388
+ 499
1389
+ ],
1390
+ [
1391
+ 67,
1392
+ 490,
1393
+ 497
1394
+ ],
1395
+ [
1396
+ 67,
1397
+ 496,
1398
+ 499
1399
+ ],
1400
+ [
1401
+ 65,
1402
+ 477,
1403
+ 497
1404
+ ],
1405
+ [
1406
+ 68,
1407
+ 492,
1408
+ 498
1409
+ ],
1410
+ [
1411
+ 69,
1412
+ 499,
1413
+ 498
1414
+ ],
1415
+ [
1416
+ 72,
1417
+ 440,
1418
+ 494
1419
+ ],
1420
+ [
1421
+ 73,
1422
+ 503,
1423
+ 501
1424
+ ],
1425
+ [
1426
+ 74,
1427
+ 505,
1428
+ 502
1429
+ ],
1430
+ [
1431
+ 67,
1432
+ 497,
1433
+ 478
1434
+ ],
1435
+ [
1436
+ 67,
1437
+ 499,
1438
+ 482
1439
+ ],
1440
+ [
1441
+ 70,
1442
+ 440,
1443
+ 482
1444
+ ],
1445
+ [
1446
+ 66,
1447
+ 498,
1448
+ 496
1449
+ ],
1450
+ [
1451
+ 67,
1452
+ 464,
1453
+ 499
1454
+ ],
1455
+ [
1456
+ 70,
1457
+ 481,
1458
+ 501
1459
+ ],
1460
+ [
1461
+ 66,
1462
+ 490,
1463
+ 478
1464
+ ],
1465
+ [
1466
+ 67,
1467
+ 498,
1468
+ 497
1469
+ ],
1470
+ [
1471
+ 65,
1472
+ 497,
1473
+ 496
1474
+ ],
1475
+ [
1476
+ 68,
1477
+ 499,
1478
+ 497
1479
+ ],
1480
+ [
1481
+ 67,
1482
+ 478,
1483
+ 498
1484
+ ],
1485
+ [
1486
+ 72,
1487
+ 501,
1488
+ 504
1489
+ ],
1490
+ [
1491
+ 70,
1492
+ 494,
1493
+ 507
1494
+ ],
1495
+ [
1496
+ 66,
1497
+ 498,
1498
+ 496
1499
+ ],
1500
+ [
1501
+ 71,
1502
+ 297,
1503
+ 296
1504
+ ],
1505
+ [
1506
+ 69,
1507
+ 500,
1508
+ 498
1509
+ ],
1510
+ [
1511
+ 69,
1512
+ 500,
1513
+ 501
1514
+ ],
1515
+ [
1516
+ 68,
1517
+ 498,
1518
+ 500
1519
+ ],
1520
+ [
1521
+ 69,
1522
+ 500,
1523
+ 499
1524
+ ],
1525
+ [
1526
+ 66,
1527
+ 497,
1528
+ 498
1529
+ ],
1530
+ [
1531
+ 68,
1532
+ 499,
1533
+ 498
1534
+ ],
1535
+ [
1536
+ 68,
1537
+ 477,
1538
+ 500
1539
+ ],
1540
+ [
1541
+ 75,
1542
+ 481,
1543
+ 500
1544
+ ],
1545
+ [
1546
+ 66,
1547
+ 488,
1548
+ 497
1549
+ ],
1550
+ [
1551
+ 73,
1552
+ 505,
1553
+ 500
1554
+ ],
1555
+ [
1556
+ 78,
1557
+ 512,
1558
+ 503
1559
+ ],
1560
+ [
1561
+ 67,
1562
+ 498,
1563
+ 499
1564
+ ],
1565
+ [
1566
+ 71,
1567
+ 500,
1568
+ 503
1569
+ ],
1570
+ [
1571
+ 72,
1572
+ 501,
1573
+ 502
1574
+ ],
1575
+ [
1576
+ 72,
1577
+ 503,
1578
+ 499
1579
+ ],
1580
+ [
1581
+ 72,
1582
+ 299,
1583
+ 297
1584
+ ],
1585
+ [
1586
+ 68,
1587
+ 500,
1588
+ 500
1589
+ ],
1590
+ [
1591
+ 68,
1592
+ 499,
1593
+ 499
1594
+ ],
1595
+ [
1596
+ 69,
1597
+ 493,
1598
+ 501
1599
+ ],
1600
+ [
1601
+ 70,
1602
+ 500,
1603
+ 501
1604
+ ],
1605
+ [
1606
+ 67,
1607
+ 499,
1608
+ 498
1609
+ ],
1610
+ [
1611
+ 67,
1612
+ 498,
1613
+ 497
1614
+ ],
1615
+ [
1616
+ 68,
1617
+ 498,
1618
+ 498
1619
+ ]
1620
+ ],
1621
+ "spacings": [
1622
+ [
1623
+ 3.0,
1624
+ 1.0,
1625
+ 1.0
1626
+ ],
1627
+ [
1628
+ 3.0,
1629
+ 1.0,
1630
+ 1.0
1631
+ ],
1632
+ [
1633
+ 3.0,
1634
+ 1.0,
1635
+ 1.0
1636
+ ],
1637
+ [
1638
+ 3.0,
1639
+ 1.0,
1640
+ 1.0
1641
+ ],
1642
+ [
1643
+ 3.0,
1644
+ 1.0,
1645
+ 1.0
1646
+ ],
1647
+ [
1648
+ 3.0,
1649
+ 1.0,
1650
+ 1.0
1651
+ ],
1652
+ [
1653
+ 3.0,
1654
+ 1.0,
1655
+ 1.0
1656
+ ],
1657
+ [
1658
+ 3.0,
1659
+ 1.0,
1660
+ 1.0
1661
+ ],
1662
+ [
1663
+ 3.0,
1664
+ 1.0,
1665
+ 1.0
1666
+ ],
1667
+ [
1668
+ 3.0,
1669
+ 1.0,
1670
+ 1.0
1671
+ ],
1672
+ [
1673
+ 3.0,
1674
+ 1.0,
1675
+ 1.0
1676
+ ],
1677
+ [
1678
+ 3.0,
1679
+ 1.0,
1680
+ 1.0
1681
+ ],
1682
+ [
1683
+ 3.0,
1684
+ 1.0,
1685
+ 1.0
1686
+ ],
1687
+ [
1688
+ 3.0,
1689
+ 1.0,
1690
+ 1.0
1691
+ ],
1692
+ [
1693
+ 3.0,
1694
+ 1.0,
1695
+ 1.0
1696
+ ],
1697
+ [
1698
+ 3.0,
1699
+ 1.0,
1700
+ 1.0
1701
+ ],
1702
+ [
1703
+ 3.0,
1704
+ 1.0,
1705
+ 1.0
1706
+ ],
1707
+ [
1708
+ 3.0,
1709
+ 1.0,
1710
+ 1.0
1711
+ ],
1712
+ [
1713
+ 3.0,
1714
+ 1.0,
1715
+ 1.0
1716
+ ],
1717
+ [
1718
+ 3.0,
1719
+ 1.0,
1720
+ 1.0
1721
+ ],
1722
+ [
1723
+ 3.0,
1724
+ 1.0,
1725
+ 1.0
1726
+ ],
1727
+ [
1728
+ 3.0,
1729
+ 1.0,
1730
+ 1.0
1731
+ ],
1732
+ [
1733
+ 3.0,
1734
+ 1.0,
1735
+ 1.0
1736
+ ],
1737
+ [
1738
+ 3.0,
1739
+ 1.0,
1740
+ 1.0
1741
+ ],
1742
+ [
1743
+ 3.0,
1744
+ 1.0,
1745
+ 1.0
1746
+ ],
1747
+ [
1748
+ 3.0,
1749
+ 1.0,
1750
+ 1.0
1751
+ ],
1752
+ [
1753
+ 3.0,
1754
+ 1.0,
1755
+ 1.0
1756
+ ],
1757
+ [
1758
+ 3.0,
1759
+ 1.0,
1760
+ 1.0
1761
+ ],
1762
+ [
1763
+ 3.0,
1764
+ 1.0,
1765
+ 1.0
1766
+ ],
1767
+ [
1768
+ 3.0,
1769
+ 1.0,
1770
+ 1.0
1771
+ ],
1772
+ [
1773
+ 3.0,
1774
+ 1.0,
1775
+ 1.0
1776
+ ],
1777
+ [
1778
+ 3.0,
1779
+ 1.0,
1780
+ 1.0
1781
+ ],
1782
+ [
1783
+ 3.0,
1784
+ 1.0,
1785
+ 1.0
1786
+ ],
1787
+ [
1788
+ 3.0,
1789
+ 1.0,
1790
+ 1.0
1791
+ ],
1792
+ [
1793
+ 3.0,
1794
+ 1.0,
1795
+ 1.0
1796
+ ],
1797
+ [
1798
+ 3.0,
1799
+ 1.0,
1800
+ 1.0
1801
+ ],
1802
+ [
1803
+ 3.0,
1804
+ 1.0,
1805
+ 1.0
1806
+ ],
1807
+ [
1808
+ 3.0,
1809
+ 1.0,
1810
+ 1.0
1811
+ ],
1812
+ [
1813
+ 3.0,
1814
+ 1.0,
1815
+ 1.0
1816
+ ],
1817
+ [
1818
+ 3.0,
1819
+ 1.0,
1820
+ 1.0
1821
+ ],
1822
+ [
1823
+ 3.0,
1824
+ 1.0,
1825
+ 1.0
1826
+ ],
1827
+ [
1828
+ 3.0,
1829
+ 1.0,
1830
+ 1.0
1831
+ ],
1832
+ [
1833
+ 3.0,
1834
+ 1.0,
1835
+ 1.0
1836
+ ],
1837
+ [
1838
+ 3.0,
1839
+ 1.0,
1840
+ 1.0
1841
+ ],
1842
+ [
1843
+ 3.0,
1844
+ 1.0,
1845
+ 1.0
1846
+ ],
1847
+ [
1848
+ 3.0,
1849
+ 1.0,
1850
+ 1.0
1851
+ ],
1852
+ [
1853
+ 3.0,
1854
+ 1.0,
1855
+ 1.0
1856
+ ],
1857
+ [
1858
+ 3.0,
1859
+ 1.0,
1860
+ 1.0
1861
+ ],
1862
+ [
1863
+ 3.0,
1864
+ 1.0,
1865
+ 1.0
1866
+ ],
1867
+ [
1868
+ 3.0,
1869
+ 1.0,
1870
+ 1.0
1871
+ ],
1872
+ [
1873
+ 3.0,
1874
+ 1.0,
1875
+ 1.0
1876
+ ],
1877
+ [
1878
+ 3.0,
1879
+ 1.0,
1880
+ 1.0
1881
+ ],
1882
+ [
1883
+ 3.0,
1884
+ 1.0,
1885
+ 1.0
1886
+ ],
1887
+ [
1888
+ 3.0,
1889
+ 1.0,
1890
+ 1.0
1891
+ ],
1892
+ [
1893
+ 3.0,
1894
+ 1.0,
1895
+ 1.0
1896
+ ],
1897
+ [
1898
+ 3.0,
1899
+ 1.0,
1900
+ 1.0
1901
+ ],
1902
+ [
1903
+ 3.0,
1904
+ 1.0,
1905
+ 1.0
1906
+ ],
1907
+ [
1908
+ 3.0,
1909
+ 1.0,
1910
+ 1.0
1911
+ ],
1912
+ [
1913
+ 3.0,
1914
+ 1.0,
1915
+ 1.0
1916
+ ],
1917
+ [
1918
+ 3.0,
1919
+ 1.0,
1920
+ 1.0
1921
+ ],
1922
+ [
1923
+ 3.0,
1924
+ 1.0,
1925
+ 1.0
1926
+ ],
1927
+ [
1928
+ 3.0,
1929
+ 1.0,
1930
+ 1.0
1931
+ ],
1932
+ [
1933
+ 3.0,
1934
+ 1.0,
1935
+ 1.0
1936
+ ],
1937
+ [
1938
+ 3.0,
1939
+ 1.0,
1940
+ 1.0
1941
+ ],
1942
+ [
1943
+ 3.0,
1944
+ 1.0,
1945
+ 1.0
1946
+ ],
1947
+ [
1948
+ 3.0,
1949
+ 1.0,
1950
+ 1.0
1951
+ ],
1952
+ [
1953
+ 3.0,
1954
+ 1.0,
1955
+ 1.0
1956
+ ],
1957
+ [
1958
+ 3.0,
1959
+ 1.0,
1960
+ 1.0
1961
+ ],
1962
+ [
1963
+ 3.0,
1964
+ 1.0,
1965
+ 1.0
1966
+ ],
1967
+ [
1968
+ 3.0,
1969
+ 1.0,
1970
+ 1.0
1971
+ ],
1972
+ [
1973
+ 3.0,
1974
+ 1.0,
1975
+ 1.0
1976
+ ],
1977
+ [
1978
+ 3.0,
1979
+ 1.0,
1980
+ 1.0
1981
+ ],
1982
+ [
1983
+ 3.0,
1984
+ 1.0,
1985
+ 1.0
1986
+ ],
1987
+ [
1988
+ 3.0,
1989
+ 1.0,
1990
+ 1.0
1991
+ ],
1992
+ [
1993
+ 3.0,
1994
+ 1.0,
1995
+ 1.0
1996
+ ],
1997
+ [
1998
+ 3.0,
1999
+ 1.0,
2000
+ 1.0
2001
+ ],
2002
+ [
2003
+ 3.0,
2004
+ 1.0,
2005
+ 1.0
2006
+ ],
2007
+ [
2008
+ 3.0,
2009
+ 1.0,
2010
+ 1.0
2011
+ ],
2012
+ [
2013
+ 3.0,
2014
+ 1.0,
2015
+ 1.0
2016
+ ],
2017
+ [
2018
+ 3.0,
2019
+ 1.0,
2020
+ 1.0
2021
+ ],
2022
+ [
2023
+ 3.0,
2024
+ 1.0,
2025
+ 1.0
2026
+ ],
2027
+ [
2028
+ 3.0,
2029
+ 1.0,
2030
+ 1.0
2031
+ ],
2032
+ [
2033
+ 3.0,
2034
+ 1.0,
2035
+ 1.0
2036
+ ],
2037
+ [
2038
+ 3.0,
2039
+ 1.0,
2040
+ 1.0
2041
+ ],
2042
+ [
2043
+ 3.0,
2044
+ 1.0,
2045
+ 1.0
2046
+ ],
2047
+ [
2048
+ 3.0,
2049
+ 1.0,
2050
+ 1.0
2051
+ ],
2052
+ [
2053
+ 3.0,
2054
+ 1.0,
2055
+ 1.0
2056
+ ],
2057
+ [
2058
+ 3.0,
2059
+ 1.0,
2060
+ 1.0
2061
+ ],
2062
+ [
2063
+ 3.0,
2064
+ 1.0,
2065
+ 1.0
2066
+ ],
2067
+ [
2068
+ 3.0,
2069
+ 1.0,
2070
+ 1.0
2071
+ ],
2072
+ [
2073
+ 3.0,
2074
+ 1.0,
2075
+ 1.0
2076
+ ],
2077
+ [
2078
+ 3.0,
2079
+ 1.0,
2080
+ 1.0
2081
+ ],
2082
+ [
2083
+ 3.0,
2084
+ 1.0,
2085
+ 1.0
2086
+ ],
2087
+ [
2088
+ 3.0,
2089
+ 1.0,
2090
+ 1.0
2091
+ ],
2092
+ [
2093
+ 3.0,
2094
+ 1.0,
2095
+ 1.0
2096
+ ],
2097
+ [
2098
+ 3.0,
2099
+ 1.0,
2100
+ 1.0
2101
+ ],
2102
+ [
2103
+ 3.0,
2104
+ 1.0,
2105
+ 1.0
2106
+ ],
2107
+ [
2108
+ 3.0,
2109
+ 1.0,
2110
+ 1.0
2111
+ ],
2112
+ [
2113
+ 3.0,
2114
+ 1.0,
2115
+ 1.0
2116
+ ],
2117
+ [
2118
+ 3.0,
2119
+ 1.0,
2120
+ 1.0
2121
+ ],
2122
+ [
2123
+ 3.0,
2124
+ 1.0,
2125
+ 1.0
2126
+ ],
2127
+ [
2128
+ 3.0,
2129
+ 1.0,
2130
+ 1.0
2131
+ ],
2132
+ [
2133
+ 3.0,
2134
+ 1.0,
2135
+ 1.0
2136
+ ],
2137
+ [
2138
+ 3.0,
2139
+ 1.0,
2140
+ 1.0
2141
+ ],
2142
+ [
2143
+ 3.0,
2144
+ 1.0,
2145
+ 1.0
2146
+ ],
2147
+ [
2148
+ 3.0,
2149
+ 1.0,
2150
+ 1.0
2151
+ ],
2152
+ [
2153
+ 3.0,
2154
+ 1.0,
2155
+ 1.0
2156
+ ],
2157
+ [
2158
+ 3.0,
2159
+ 1.0,
2160
+ 1.0
2161
+ ],
2162
+ [
2163
+ 3.0,
2164
+ 1.0,
2165
+ 1.0
2166
+ ],
2167
+ [
2168
+ 3.0,
2169
+ 1.0,
2170
+ 1.0
2171
+ ],
2172
+ [
2173
+ 3.0,
2174
+ 1.0,
2175
+ 1.0
2176
+ ],
2177
+ [
2178
+ 3.0,
2179
+ 1.0,
2180
+ 1.0
2181
+ ],
2182
+ [
2183
+ 3.0,
2184
+ 1.0,
2185
+ 1.0
2186
+ ],
2187
+ [
2188
+ 3.0,
2189
+ 1.0,
2190
+ 1.0
2191
+ ],
2192
+ [
2193
+ 3.0,
2194
+ 1.0,
2195
+ 1.0
2196
+ ],
2197
+ [
2198
+ 3.0,
2199
+ 1.0,
2200
+ 1.0
2201
+ ],
2202
+ [
2203
+ 3.0,
2204
+ 1.0,
2205
+ 1.0
2206
+ ],
2207
+ [
2208
+ 3.0,
2209
+ 1.0,
2210
+ 1.0
2211
+ ],
2212
+ [
2213
+ 3.0,
2214
+ 1.0,
2215
+ 1.0
2216
+ ],
2217
+ [
2218
+ 3.0,
2219
+ 1.0,
2220
+ 1.0
2221
+ ],
2222
+ [
2223
+ 3.0,
2224
+ 1.0,
2225
+ 1.0
2226
+ ],
2227
+ [
2228
+ 3.0,
2229
+ 1.0,
2230
+ 1.0
2231
+ ],
2232
+ [
2233
+ 3.0,
2234
+ 1.0,
2235
+ 1.0
2236
+ ],
2237
+ [
2238
+ 3.0,
2239
+ 1.0,
2240
+ 1.0
2241
+ ],
2242
+ [
2243
+ 3.0,
2244
+ 1.0,
2245
+ 1.0
2246
+ ],
2247
+ [
2248
+ 3.0,
2249
+ 1.0,
2250
+ 1.0
2251
+ ],
2252
+ [
2253
+ 3.0,
2254
+ 1.0,
2255
+ 1.0
2256
+ ],
2257
+ [
2258
+ 3.0,
2259
+ 1.0,
2260
+ 1.0
2261
+ ],
2262
+ [
2263
+ 3.0,
2264
+ 1.0,
2265
+ 1.0
2266
+ ],
2267
+ [
2268
+ 3.0,
2269
+ 1.0,
2270
+ 1.0
2271
+ ],
2272
+ [
2273
+ 3.0,
2274
+ 1.0,
2275
+ 1.0
2276
+ ],
2277
+ [
2278
+ 3.0,
2279
+ 1.0,
2280
+ 1.0
2281
+ ],
2282
+ [
2283
+ 3.0,
2284
+ 1.0,
2285
+ 1.0
2286
+ ],
2287
+ [
2288
+ 3.0,
2289
+ 1.0,
2290
+ 1.0
2291
+ ],
2292
+ [
2293
+ 3.0,
2294
+ 1.0,
2295
+ 1.0
2296
+ ],
2297
+ [
2298
+ 3.0,
2299
+ 1.0,
2300
+ 1.0
2301
+ ],
2302
+ [
2303
+ 3.0,
2304
+ 1.0,
2305
+ 1.0
2306
+ ],
2307
+ [
2308
+ 3.0,
2309
+ 1.0,
2310
+ 1.0
2311
+ ],
2312
+ [
2313
+ 3.0,
2314
+ 1.0,
2315
+ 1.0
2316
+ ],
2317
+ [
2318
+ 3.0,
2319
+ 1.0,
2320
+ 1.0
2321
+ ],
2322
+ [
2323
+ 3.0,
2324
+ 1.0,
2325
+ 1.0
2326
+ ],
2327
+ [
2328
+ 3.0,
2329
+ 1.0,
2330
+ 1.0
2331
+ ],
2332
+ [
2333
+ 3.0,
2334
+ 1.0,
2335
+ 1.0
2336
+ ],
2337
+ [
2338
+ 3.0,
2339
+ 1.0,
2340
+ 1.0
2341
+ ],
2342
+ [
2343
+ 3.0,
2344
+ 1.0,
2345
+ 1.0
2346
+ ],
2347
+ [
2348
+ 3.0,
2349
+ 1.0,
2350
+ 1.0
2351
+ ],
2352
+ [
2353
+ 3.0,
2354
+ 1.0,
2355
+ 1.0
2356
+ ],
2357
+ [
2358
+ 3.0,
2359
+ 1.0,
2360
+ 1.0
2361
+ ],
2362
+ [
2363
+ 3.0,
2364
+ 1.0,
2365
+ 1.0
2366
+ ],
2367
+ [
2368
+ 3.0,
2369
+ 1.0,
2370
+ 1.0
2371
+ ],
2372
+ [
2373
+ 3.0,
2374
+ 1.0,
2375
+ 1.0
2376
+ ],
2377
+ [
2378
+ 3.0,
2379
+ 1.0,
2380
+ 1.0
2381
+ ],
2382
+ [
2383
+ 3.0,
2384
+ 1.0,
2385
+ 1.0
2386
+ ],
2387
+ [
2388
+ 3.0,
2389
+ 1.0,
2390
+ 1.0
2391
+ ],
2392
+ [
2393
+ 3.0,
2394
+ 1.0,
2395
+ 1.0
2396
+ ],
2397
+ [
2398
+ 3.0,
2399
+ 1.0,
2400
+ 1.0
2401
+ ],
2402
+ [
2403
+ 3.0,
2404
+ 1.0,
2405
+ 1.0
2406
+ ],
2407
+ [
2408
+ 3.0,
2409
+ 1.0,
2410
+ 1.0
2411
+ ],
2412
+ [
2413
+ 3.0,
2414
+ 1.0,
2415
+ 1.0
2416
+ ],
2417
+ [
2418
+ 3.0,
2419
+ 1.0,
2420
+ 1.0
2421
+ ],
2422
+ [
2423
+ 3.0,
2424
+ 1.0,
2425
+ 1.0
2426
+ ],
2427
+ [
2428
+ 3.0,
2429
+ 1.0,
2430
+ 1.0
2431
+ ],
2432
+ [
2433
+ 3.0,
2434
+ 1.0,
2435
+ 1.0
2436
+ ],
2437
+ [
2438
+ 3.0,
2439
+ 1.0,
2440
+ 1.0
2441
+ ],
2442
+ [
2443
+ 3.0,
2444
+ 1.0,
2445
+ 1.0
2446
+ ],
2447
+ [
2448
+ 3.0,
2449
+ 1.0,
2450
+ 1.0
2451
+ ],
2452
+ [
2453
+ 3.0,
2454
+ 1.0,
2455
+ 1.0
2456
+ ],
2457
+ [
2458
+ 3.0,
2459
+ 1.0,
2460
+ 1.0
2461
+ ],
2462
+ [
2463
+ 3.0,
2464
+ 1.0,
2465
+ 1.0
2466
+ ],
2467
+ [
2468
+ 3.0,
2469
+ 1.0,
2470
+ 1.0
2471
+ ],
2472
+ [
2473
+ 3.0,
2474
+ 1.0,
2475
+ 1.0
2476
+ ],
2477
+ [
2478
+ 3.0,
2479
+ 1.0,
2480
+ 1.0
2481
+ ],
2482
+ [
2483
+ 3.0,
2484
+ 1.0,
2485
+ 1.0
2486
+ ],
2487
+ [
2488
+ 3.0,
2489
+ 1.0,
2490
+ 1.0
2491
+ ],
2492
+ [
2493
+ 3.0,
2494
+ 1.0,
2495
+ 1.0
2496
+ ],
2497
+ [
2498
+ 3.0,
2499
+ 1.0,
2500
+ 1.0
2501
+ ],
2502
+ [
2503
+ 3.0,
2504
+ 1.0,
2505
+ 1.0
2506
+ ],
2507
+ [
2508
+ 3.0,
2509
+ 1.0,
2510
+ 1.0
2511
+ ],
2512
+ [
2513
+ 3.0,
2514
+ 1.0,
2515
+ 1.0
2516
+ ],
2517
+ [
2518
+ 3.0,
2519
+ 1.0,
2520
+ 1.0
2521
+ ],
2522
+ [
2523
+ 3.0,
2524
+ 1.0,
2525
+ 1.0
2526
+ ],
2527
+ [
2528
+ 3.0,
2529
+ 1.0,
2530
+ 1.0
2531
+ ],
2532
+ [
2533
+ 3.0,
2534
+ 1.0,
2535
+ 1.0
2536
+ ],
2537
+ [
2538
+ 3.0,
2539
+ 1.0,
2540
+ 1.0
2541
+ ],
2542
+ [
2543
+ 3.0,
2544
+ 1.0,
2545
+ 1.0
2546
+ ],
2547
+ [
2548
+ 3.0,
2549
+ 1.0,
2550
+ 1.0
2551
+ ],
2552
+ [
2553
+ 3.0,
2554
+ 1.0,
2555
+ 1.0
2556
+ ],
2557
+ [
2558
+ 3.0,
2559
+ 1.0,
2560
+ 1.0
2561
+ ],
2562
+ [
2563
+ 3.0,
2564
+ 1.0,
2565
+ 1.0
2566
+ ],
2567
+ [
2568
+ 3.0,
2569
+ 1.0,
2570
+ 1.0
2571
+ ],
2572
+ [
2573
+ 3.0,
2574
+ 1.0,
2575
+ 1.0
2576
+ ],
2577
+ [
2578
+ 3.0,
2579
+ 1.0,
2580
+ 1.0
2581
+ ],
2582
+ [
2583
+ 3.0,
2584
+ 1.0,
2585
+ 1.0
2586
+ ],
2587
+ [
2588
+ 3.0,
2589
+ 1.0,
2590
+ 1.0
2591
+ ],
2592
+ [
2593
+ 3.0,
2594
+ 1.0,
2595
+ 1.0
2596
+ ],
2597
+ [
2598
+ 3.0,
2599
+ 1.0,
2600
+ 1.0
2601
+ ],
2602
+ [
2603
+ 3.0,
2604
+ 1.0,
2605
+ 1.0
2606
+ ],
2607
+ [
2608
+ 3.0,
2609
+ 1.0,
2610
+ 1.0
2611
+ ],
2612
+ [
2613
+ 3.0,
2614
+ 1.0,
2615
+ 1.0
2616
+ ],
2617
+ [
2618
+ 3.0,
2619
+ 1.0,
2620
+ 1.0
2621
+ ],
2622
+ [
2623
+ 3.0,
2624
+ 1.0,
2625
+ 1.0
2626
+ ],
2627
+ [
2628
+ 3.0,
2629
+ 1.0,
2630
+ 1.0
2631
+ ],
2632
+ [
2633
+ 3.0,
2634
+ 1.0,
2635
+ 1.0
2636
+ ],
2637
+ [
2638
+ 3.0,
2639
+ 1.0,
2640
+ 1.0
2641
+ ],
2642
+ [
2643
+ 3.0,
2644
+ 1.0,
2645
+ 1.0
2646
+ ],
2647
+ [
2648
+ 3.0,
2649
+ 1.0,
2650
+ 1.0
2651
+ ],
2652
+ [
2653
+ 3.0,
2654
+ 1.0,
2655
+ 1.0
2656
+ ],
2657
+ [
2658
+ 3.0,
2659
+ 1.0,
2660
+ 1.0
2661
+ ],
2662
+ [
2663
+ 3.0,
2664
+ 1.0,
2665
+ 1.0
2666
+ ],
2667
+ [
2668
+ 3.0,
2669
+ 1.0,
2670
+ 1.0
2671
+ ],
2672
+ [
2673
+ 3.0,
2674
+ 1.0,
2675
+ 1.0
2676
+ ],
2677
+ [
2678
+ 3.0,
2679
+ 1.0,
2680
+ 1.0
2681
+ ],
2682
+ [
2683
+ 3.0,
2684
+ 1.0,
2685
+ 1.0
2686
+ ],
2687
+ [
2688
+ 3.0,
2689
+ 1.0,
2690
+ 1.0
2691
+ ],
2692
+ [
2693
+ 3.0,
2694
+ 1.0,
2695
+ 1.0
2696
+ ],
2697
+ [
2698
+ 3.0,
2699
+ 1.0,
2700
+ 1.0
2701
+ ],
2702
+ [
2703
+ 3.0,
2704
+ 1.0,
2705
+ 1.0
2706
+ ],
2707
+ [
2708
+ 3.0,
2709
+ 1.0,
2710
+ 1.0
2711
+ ],
2712
+ [
2713
+ 3.0,
2714
+ 1.0,
2715
+ 1.0
2716
+ ],
2717
+ [
2718
+ 3.0,
2719
+ 1.0,
2720
+ 1.0
2721
+ ],
2722
+ [
2723
+ 3.0,
2724
+ 1.0,
2725
+ 1.0
2726
+ ],
2727
+ [
2728
+ 3.0,
2729
+ 1.0,
2730
+ 1.0
2731
+ ],
2732
+ [
2733
+ 3.0,
2734
+ 1.0,
2735
+ 1.0
2736
+ ],
2737
+ [
2738
+ 3.0,
2739
+ 1.0,
2740
+ 1.0
2741
+ ],
2742
+ [
2743
+ 3.0,
2744
+ 1.0,
2745
+ 1.0
2746
+ ],
2747
+ [
2748
+ 3.0,
2749
+ 1.0,
2750
+ 1.0
2751
+ ],
2752
+ [
2753
+ 3.0,
2754
+ 1.0,
2755
+ 1.0
2756
+ ],
2757
+ [
2758
+ 3.0,
2759
+ 1.0,
2760
+ 1.0
2761
+ ],
2762
+ [
2763
+ 3.0,
2764
+ 1.0,
2765
+ 1.0
2766
+ ],
2767
+ [
2768
+ 3.0,
2769
+ 1.0,
2770
+ 1.0
2771
+ ],
2772
+ [
2773
+ 3.0,
2774
+ 1.0,
2775
+ 1.0
2776
+ ],
2777
+ [
2778
+ 3.0,
2779
+ 1.0,
2780
+ 1.0
2781
+ ],
2782
+ [
2783
+ 3.0,
2784
+ 1.0,
2785
+ 1.0
2786
+ ],
2787
+ [
2788
+ 3.0,
2789
+ 1.0,
2790
+ 1.0
2791
+ ],
2792
+ [
2793
+ 3.0,
2794
+ 1.0,
2795
+ 1.0
2796
+ ],
2797
+ [
2798
+ 3.0,
2799
+ 1.0,
2800
+ 1.0
2801
+ ],
2802
+ [
2803
+ 3.0,
2804
+ 1.0,
2805
+ 1.0
2806
+ ],
2807
+ [
2808
+ 3.0,
2809
+ 1.0,
2810
+ 1.0
2811
+ ],
2812
+ [
2813
+ 3.0,
2814
+ 1.0,
2815
+ 1.0
2816
+ ],
2817
+ [
2818
+ 3.0,
2819
+ 1.0,
2820
+ 1.0
2821
+ ],
2822
+ [
2823
+ 3.0,
2824
+ 1.0,
2825
+ 1.0
2826
+ ],
2827
+ [
2828
+ 3.0,
2829
+ 1.0,
2830
+ 1.0
2831
+ ],
2832
+ [
2833
+ 3.0,
2834
+ 1.0,
2835
+ 1.0
2836
+ ],
2837
+ [
2838
+ 3.0,
2839
+ 1.0,
2840
+ 1.0
2841
+ ],
2842
+ [
2843
+ 3.0,
2844
+ 1.0,
2845
+ 1.0
2846
+ ],
2847
+ [
2848
+ 3.0,
2849
+ 1.0,
2850
+ 1.0
2851
+ ],
2852
+ [
2853
+ 3.0,
2854
+ 1.0,
2855
+ 1.0
2856
+ ],
2857
+ [
2858
+ 3.0,
2859
+ 1.0,
2860
+ 1.0
2861
+ ],
2862
+ [
2863
+ 3.0,
2864
+ 1.0,
2865
+ 1.0
2866
+ ],
2867
+ [
2868
+ 3.0,
2869
+ 1.0,
2870
+ 1.0
2871
+ ],
2872
+ [
2873
+ 3.0,
2874
+ 1.0,
2875
+ 1.0
2876
+ ],
2877
+ [
2878
+ 3.0,
2879
+ 1.0,
2880
+ 1.0
2881
+ ],
2882
+ [
2883
+ 3.0,
2884
+ 1.0,
2885
+ 1.0
2886
+ ],
2887
+ [
2888
+ 3.0,
2889
+ 1.0,
2890
+ 1.0
2891
+ ],
2892
+ [
2893
+ 3.0,
2894
+ 1.0,
2895
+ 1.0
2896
+ ],
2897
+ [
2898
+ 3.0,
2899
+ 1.0,
2900
+ 1.0
2901
+ ],
2902
+ [
2903
+ 3.0,
2904
+ 1.0,
2905
+ 1.0
2906
+ ],
2907
+ [
2908
+ 3.0,
2909
+ 1.0,
2910
+ 1.0
2911
+ ],
2912
+ [
2913
+ 3.0,
2914
+ 1.0,
2915
+ 1.0
2916
+ ],
2917
+ [
2918
+ 3.0,
2919
+ 1.0,
2920
+ 1.0
2921
+ ],
2922
+ [
2923
+ 3.0,
2924
+ 1.0,
2925
+ 1.0
2926
+ ],
2927
+ [
2928
+ 3.0,
2929
+ 1.0,
2930
+ 1.0
2931
+ ],
2932
+ [
2933
+ 3.0,
2934
+ 1.0,
2935
+ 1.0
2936
+ ],
2937
+ [
2938
+ 3.0,
2939
+ 1.0,
2940
+ 1.0
2941
+ ],
2942
+ [
2943
+ 3.0,
2944
+ 1.0,
2945
+ 1.0
2946
+ ],
2947
+ [
2948
+ 3.0,
2949
+ 1.0,
2950
+ 1.0
2951
+ ],
2952
+ [
2953
+ 3.0,
2954
+ 1.0,
2955
+ 1.0
2956
+ ],
2957
+ [
2958
+ 3.0,
2959
+ 1.0,
2960
+ 1.0
2961
+ ],
2962
+ [
2963
+ 3.0,
2964
+ 1.0,
2965
+ 1.0
2966
+ ],
2967
+ [
2968
+ 3.0,
2969
+ 1.0,
2970
+ 1.0
2971
+ ],
2972
+ [
2973
+ 3.0,
2974
+ 1.0,
2975
+ 1.0
2976
+ ],
2977
+ [
2978
+ 3.0,
2979
+ 1.0,
2980
+ 1.0
2981
+ ],
2982
+ [
2983
+ 3.0,
2984
+ 1.0,
2985
+ 1.0
2986
+ ],
2987
+ [
2988
+ 3.0,
2989
+ 1.0,
2990
+ 1.0
2991
+ ],
2992
+ [
2993
+ 3.0,
2994
+ 1.0,
2995
+ 1.0
2996
+ ],
2997
+ [
2998
+ 3.0,
2999
+ 1.0,
3000
+ 1.0
3001
+ ],
3002
+ [
3003
+ 3.0,
3004
+ 1.0,
3005
+ 1.0
3006
+ ],
3007
+ [
3008
+ 3.0,
3009
+ 1.0,
3010
+ 1.0
3011
+ ],
3012
+ [
3013
+ 3.0,
3014
+ 1.0,
3015
+ 1.0
3016
+ ],
3017
+ [
3018
+ 3.0,
3019
+ 1.0,
3020
+ 1.0
3021
+ ],
3022
+ [
3023
+ 3.0,
3024
+ 1.0,
3025
+ 1.0
3026
+ ],
3027
+ [
3028
+ 3.0,
3029
+ 1.0,
3030
+ 1.0
3031
+ ],
3032
+ [
3033
+ 3.0,
3034
+ 1.0,
3035
+ 1.0
3036
+ ],
3037
+ [
3038
+ 3.0,
3039
+ 1.0,
3040
+ 1.0
3041
+ ],
3042
+ [
3043
+ 3.0,
3044
+ 1.0,
3045
+ 1.0
3046
+ ],
3047
+ [
3048
+ 3.0,
3049
+ 1.0,
3050
+ 1.0
3051
+ ],
3052
+ [
3053
+ 3.0,
3054
+ 1.0,
3055
+ 1.0
3056
+ ],
3057
+ [
3058
+ 3.0,
3059
+ 1.0,
3060
+ 1.0
3061
+ ],
3062
+ [
3063
+ 3.0,
3064
+ 1.0,
3065
+ 1.0
3066
+ ],
3067
+ [
3068
+ 3.0,
3069
+ 1.0,
3070
+ 1.0
3071
+ ],
3072
+ [
3073
+ 3.0,
3074
+ 1.0,
3075
+ 1.0
3076
+ ],
3077
+ [
3078
+ 3.0,
3079
+ 1.0,
3080
+ 1.0
3081
+ ],
3082
+ [
3083
+ 3.0,
3084
+ 1.0,
3085
+ 1.0
3086
+ ],
3087
+ [
3088
+ 3.0,
3089
+ 1.0,
3090
+ 1.0
3091
+ ],
3092
+ [
3093
+ 3.0,
3094
+ 1.0,
3095
+ 1.0
3096
+ ],
3097
+ [
3098
+ 3.0,
3099
+ 1.0,
3100
+ 1.0
3101
+ ],
3102
+ [
3103
+ 3.0,
3104
+ 1.0,
3105
+ 1.0
3106
+ ],
3107
+ [
3108
+ 3.0,
3109
+ 1.0,
3110
+ 1.0
3111
+ ],
3112
+ [
3113
+ 3.0,
3114
+ 1.0,
3115
+ 1.0
3116
+ ],
3117
+ [
3118
+ 3.0,
3119
+ 1.0,
3120
+ 1.0
3121
+ ],
3122
+ [
3123
+ 3.0,
3124
+ 1.0,
3125
+ 1.0
3126
+ ],
3127
+ [
3128
+ 3.0,
3129
+ 1.0,
3130
+ 1.0
3131
+ ],
3132
+ [
3133
+ 3.0,
3134
+ 1.0,
3135
+ 1.0
3136
+ ],
3137
+ [
3138
+ 3.0,
3139
+ 1.0,
3140
+ 1.0
3141
+ ],
3142
+ [
3143
+ 3.0,
3144
+ 1.0,
3145
+ 1.0
3146
+ ],
3147
+ [
3148
+ 3.0,
3149
+ 1.0,
3150
+ 1.0
3151
+ ],
3152
+ [
3153
+ 3.0,
3154
+ 1.0,
3155
+ 1.0
3156
+ ],
3157
+ [
3158
+ 3.0,
3159
+ 1.0,
3160
+ 1.0
3161
+ ],
3162
+ [
3163
+ 3.0,
3164
+ 1.0,
3165
+ 1.0
3166
+ ],
3167
+ [
3168
+ 3.0,
3169
+ 1.0,
3170
+ 1.0
3171
+ ],
3172
+ [
3173
+ 3.0,
3174
+ 1.0,
3175
+ 1.0
3176
+ ],
3177
+ [
3178
+ 3.0,
3179
+ 1.0,
3180
+ 1.0
3181
+ ],
3182
+ [
3183
+ 3.0,
3184
+ 1.0,
3185
+ 1.0
3186
+ ],
3187
+ [
3188
+ 3.0,
3189
+ 1.0,
3190
+ 1.0
3191
+ ],
3192
+ [
3193
+ 3.0,
3194
+ 1.0,
3195
+ 1.0
3196
+ ],
3197
+ [
3198
+ 3.0,
3199
+ 1.0,
3200
+ 1.0
3201
+ ],
3202
+ [
3203
+ 3.0,
3204
+ 1.0,
3205
+ 1.0
3206
+ ],
3207
+ [
3208
+ 3.0,
3209
+ 1.0,
3210
+ 1.0
3211
+ ],
3212
+ [
3213
+ 3.0,
3214
+ 1.0,
3215
+ 1.0
3216
+ ],
3217
+ [
3218
+ 3.0,
3219
+ 1.0,
3220
+ 1.0
3221
+ ],
3222
+ [
3223
+ 3.0,
3224
+ 1.0,
3225
+ 1.0
3226
+ ]
3227
+ ]
3228
+ }
Dataset544/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/checkpoint_final.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1695e6f921488dbc372d8c7600f748c8d56fcf28cf71ea3c14bcacc78cf2aabf
3
+ size 815580525
Dataset544/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/debug.json ADDED
The diff for this file is too large to render. See raw diff
 
Dataset544/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/progress.png ADDED

Git LFS Details

  • SHA256: ddf72b0f141c0785923ac73f5821c75e7f6d5daa970742590d9792210414ca78
  • Pointer size: 131 Bytes
  • Size of remote file: 618 kB
Dataset544/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/fold_0/validation/summary.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "mae": {
3
+ "mean": 52.80768787193642,
4
+ "std": 39.25411217734592,
5
+ "max": 349.4295323444315,
6
+ "min": 30.53782153506361,
7
+ "25pc": 39.67259547669727,
8
+ "50pc": 44.417437064864046,
9
+ "75pc": 56.57117028196328,
10
+ "count": 65
11
+ },
12
+ "psnr": {
13
+ "mean": 32.25481819350785,
14
+ "std": 2.9858789867732187,
15
+ "max": 36.561440579128075,
16
+ "min": 17.068229372936194,
17
+ "25pc": 30.947542147774367,
18
+ "50pc": 32.57369524597034,
19
+ "75pc": 34.17962467969445,
20
+ "count": 65
21
+ },
22
+ "ms_ssim": {
23
+ "mean": 0.9536834487302785,
24
+ "std": 0.08019347284318636,
25
+ "max": 0.9915810123537975,
26
+ "min": 0.3802523438307981,
27
+ "25pc": 0.9586479045957355,
28
+ "50pc": 0.9741513563984106,
29
+ "75pc": 0.9824337046998175,
30
+ "count": 65
31
+ },
32
+ "DICE": {
33
+ "mean": 0.8256282861416157,
34
+ "std": 0.06491444113684663,
35
+ "max": 0.9275574088096619,
36
+ "min": 0.5898815393447876,
37
+ "25pc": 0.7742491364479065,
38
+ "50pc": 0.8340623378753662,
39
+ "75pc": 0.8758383393287659,
40
+ "count": 65
41
+ },
42
+ "HD95": {
43
+ "mean": 4.208385368493887,
44
+ "std": 3.814029574946056,
45
+ "max": 30.388620376586914,
46
+ "min": 1.1522142887115479,
47
+ "25pc": 2.4057328701019287,
48
+ "50pc": 3.2774264812469482,
49
+ "75pc": 4.871885299682617,
50
+ "count": 65
51
+ }
52
+ }
Dataset544/nnUNetTrainerMRCT_loss_masked_perception_masked__nnUNetResEncUNetLPlans__3d_fullres/plans.json ADDED
@@ -0,0 +1,521 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_name": "Dataset544_synthrad2025_task2_CBCT_TH_pre_v2r_stitched_masked_both",
3
+ "plans_name": "nnUNetResEncUNetLPlans",
4
+ "original_median_spacing_after_transp": [
5
+ 3.0,
6
+ 1.0,
7
+ 1.0
8
+ ],
9
+ "original_median_shape_after_transp": [
10
+ 81,
11
+ 450,
12
+ 450
13
+ ],
14
+ "image_reader_writer": "SimpleITKIO",
15
+ "transpose_forward": [
16
+ 0,
17
+ 1,
18
+ 2
19
+ ],
20
+ "transpose_backward": [
21
+ 0,
22
+ 1,
23
+ 2
24
+ ],
25
+ "configurations": {
26
+ "2d": {
27
+ "data_identifier": "nnUNetPlans_2d",
28
+ "preprocessor_name": "DefaultPreprocessor",
29
+ "batch_size": 13,
30
+ "patch_size": [
31
+ 512,
32
+ 448
33
+ ],
34
+ "median_image_size_in_voxels": [
35
+ 450.0,
36
+ 450.0
37
+ ],
38
+ "spacing": [
39
+ 1.0,
40
+ 1.0
41
+ ],
42
+ "normalization_schemes": [
43
+ "CTNormalizationClippingSynthrad2025"
44
+ ],
45
+ "use_mask_for_norm": [
46
+ false
47
+ ],
48
+ "resampling_fn_data": "resample_data_or_seg_to_shape",
49
+ "resampling_fn_seg": "resample_data_or_seg_to_shape",
50
+ "resampling_fn_data_kwargs": {
51
+ "is_seg": false,
52
+ "order": 3,
53
+ "order_z": 0,
54
+ "force_separate_z": null
55
+ },
56
+ "resampling_fn_seg_kwargs": {
57
+ "is_seg": true,
58
+ "order": 1,
59
+ "order_z": 0,
60
+ "force_separate_z": null
61
+ },
62
+ "resampling_fn_probabilities": "resample_data_or_seg_to_shape",
63
+ "resampling_fn_probabilities_kwargs": {
64
+ "is_seg": false,
65
+ "order": 1,
66
+ "order_z": 0,
67
+ "force_separate_z": null
68
+ },
69
+ "architecture": {
70
+ "network_class_name": "dynamic_network_architectures.architectures.unet.ResidualEncoderUNet",
71
+ "arch_kwargs": {
72
+ "n_stages": 7,
73
+ "features_per_stage": [
74
+ 32,
75
+ 64,
76
+ 128,
77
+ 256,
78
+ 512,
79
+ 512,
80
+ 512
81
+ ],
82
+ "conv_op": "torch.nn.modules.conv.Conv2d",
83
+ "kernel_sizes": [
84
+ [
85
+ 3,
86
+ 3
87
+ ],
88
+ [
89
+ 3,
90
+ 3
91
+ ],
92
+ [
93
+ 3,
94
+ 3
95
+ ],
96
+ [
97
+ 3,
98
+ 3
99
+ ],
100
+ [
101
+ 3,
102
+ 3
103
+ ],
104
+ [
105
+ 3,
106
+ 3
107
+ ],
108
+ [
109
+ 3,
110
+ 3
111
+ ]
112
+ ],
113
+ "strides": [
114
+ [
115
+ 1,
116
+ 1
117
+ ],
118
+ [
119
+ 2,
120
+ 2
121
+ ],
122
+ [
123
+ 2,
124
+ 2
125
+ ],
126
+ [
127
+ 2,
128
+ 2
129
+ ],
130
+ [
131
+ 2,
132
+ 2
133
+ ],
134
+ [
135
+ 2,
136
+ 2
137
+ ],
138
+ [
139
+ 2,
140
+ 2
141
+ ]
142
+ ],
143
+ "n_blocks_per_stage": [
144
+ 1,
145
+ 3,
146
+ 4,
147
+ 6,
148
+ 6,
149
+ 6,
150
+ 6
151
+ ],
152
+ "n_conv_per_stage_decoder": [
153
+ 1,
154
+ 1,
155
+ 1,
156
+ 1,
157
+ 1,
158
+ 1
159
+ ],
160
+ "conv_bias": true,
161
+ "norm_op": "torch.nn.modules.instancenorm.InstanceNorm2d",
162
+ "norm_op_kwargs": {
163
+ "eps": 1e-05,
164
+ "affine": true
165
+ },
166
+ "dropout_op": null,
167
+ "dropout_op_kwargs": null,
168
+ "nonlin": "torch.nn.LeakyReLU",
169
+ "nonlin_kwargs": {
170
+ "inplace": true
171
+ }
172
+ },
173
+ "_kw_requires_import": [
174
+ "conv_op",
175
+ "norm_op",
176
+ "dropout_op",
177
+ "nonlin"
178
+ ]
179
+ },
180
+ "batch_dice": true
181
+ },
182
+ "3d_lowres": {
183
+ "data_identifier": "nnUNetResEncUNetLPlans_3d_lowres",
184
+ "preprocessor_name": "DefaultPreprocessor",
185
+ "batch_size": 2,
186
+ "patch_size": [
187
+ 56,
188
+ 224,
189
+ 192
190
+ ],
191
+ "median_image_size_in_voxels": [
192
+ 81,
193
+ 298,
194
+ 298
195
+ ],
196
+ "spacing": [
197
+ 3.0,
198
+ 1.512589724855112,
199
+ 1.512589724855112
200
+ ],
201
+ "normalization_schemes": [
202
+ "CTNormalizationClippingSynthrad2025"
203
+ ],
204
+ "use_mask_for_norm": [
205
+ false
206
+ ],
207
+ "resampling_fn_data": "resample_data_or_seg_to_shape",
208
+ "resampling_fn_seg": "resample_data_or_seg_to_shape",
209
+ "resampling_fn_data_kwargs": {
210
+ "is_seg": false,
211
+ "order": 3,
212
+ "order_z": 0,
213
+ "force_separate_z": null
214
+ },
215
+ "resampling_fn_seg_kwargs": {
216
+ "is_seg": true,
217
+ "order": 1,
218
+ "order_z": 0,
219
+ "force_separate_z": null
220
+ },
221
+ "resampling_fn_probabilities": "resample_data_or_seg_to_shape",
222
+ "resampling_fn_probabilities_kwargs": {
223
+ "is_seg": false,
224
+ "order": 1,
225
+ "order_z": 0,
226
+ "force_separate_z": null
227
+ },
228
+ "architecture": {
229
+ "network_class_name": "dynamic_network_architectures.architectures.unet.ResidualEncoderUNet",
230
+ "arch_kwargs": {
231
+ "n_stages": 6,
232
+ "features_per_stage": [
233
+ 32,
234
+ 64,
235
+ 128,
236
+ 256,
237
+ 320,
238
+ 320
239
+ ],
240
+ "conv_op": "torch.nn.modules.conv.Conv3d",
241
+ "kernel_sizes": [
242
+ [
243
+ 3,
244
+ 3,
245
+ 3
246
+ ],
247
+ [
248
+ 3,
249
+ 3,
250
+ 3
251
+ ],
252
+ [
253
+ 3,
254
+ 3,
255
+ 3
256
+ ],
257
+ [
258
+ 3,
259
+ 3,
260
+ 3
261
+ ],
262
+ [
263
+ 3,
264
+ 3,
265
+ 3
266
+ ],
267
+ [
268
+ 3,
269
+ 3,
270
+ 3
271
+ ]
272
+ ],
273
+ "strides": [
274
+ [
275
+ 1,
276
+ 1,
277
+ 1
278
+ ],
279
+ [
280
+ 2,
281
+ 2,
282
+ 2
283
+ ],
284
+ [
285
+ 2,
286
+ 2,
287
+ 2
288
+ ],
289
+ [
290
+ 2,
291
+ 2,
292
+ 2
293
+ ],
294
+ [
295
+ 1,
296
+ 2,
297
+ 2
298
+ ],
299
+ [
300
+ 1,
301
+ 2,
302
+ 2
303
+ ]
304
+ ],
305
+ "n_blocks_per_stage": [
306
+ 1,
307
+ 3,
308
+ 4,
309
+ 6,
310
+ 6,
311
+ 6
312
+ ],
313
+ "n_conv_per_stage_decoder": [
314
+ 1,
315
+ 1,
316
+ 1,
317
+ 1,
318
+ 1
319
+ ],
320
+ "conv_bias": true,
321
+ "norm_op": "torch.nn.modules.instancenorm.InstanceNorm3d",
322
+ "norm_op_kwargs": {
323
+ "eps": 1e-05,
324
+ "affine": true
325
+ },
326
+ "dropout_op": null,
327
+ "dropout_op_kwargs": null,
328
+ "nonlin": "torch.nn.LeakyReLU",
329
+ "nonlin_kwargs": {
330
+ "inplace": true
331
+ }
332
+ },
333
+ "_kw_requires_import": [
334
+ "conv_op",
335
+ "norm_op",
336
+ "dropout_op",
337
+ "nonlin"
338
+ ]
339
+ },
340
+ "batch_dice": false,
341
+ "next_stage": "3d_cascade_fullres"
342
+ },
343
+ "3d_fullres": {
344
+ "data_identifier": "nnUNetPlans_3d_fullres",
345
+ "preprocessor_name": "DefaultPreprocessor",
346
+ "batch_size": 2,
347
+ "patch_size": [
348
+ 40,
349
+ 224,
350
+ 192
351
+ ],
352
+ "median_image_size_in_voxels": [
353
+ 81.0,
354
+ 450.0,
355
+ 450.0
356
+ ],
357
+ "spacing": [
358
+ 3.0,
359
+ 1.0,
360
+ 1.0
361
+ ],
362
+ "normalization_schemes": [
363
+ "CTNormalizationClippingSynthrad2025"
364
+ ],
365
+ "use_mask_for_norm": [
366
+ false
367
+ ],
368
+ "resampling_fn_data": "resample_data_or_seg_to_shape",
369
+ "resampling_fn_seg": "resample_data_or_seg_to_shape",
370
+ "resampling_fn_data_kwargs": {
371
+ "is_seg": false,
372
+ "order": 3,
373
+ "order_z": 0,
374
+ "force_separate_z": null
375
+ },
376
+ "resampling_fn_seg_kwargs": {
377
+ "is_seg": true,
378
+ "order": 1,
379
+ "order_z": 0,
380
+ "force_separate_z": null
381
+ },
382
+ "resampling_fn_probabilities": "resample_data_or_seg_to_shape",
383
+ "resampling_fn_probabilities_kwargs": {
384
+ "is_seg": false,
385
+ "order": 1,
386
+ "order_z": 0,
387
+ "force_separate_z": null
388
+ },
389
+ "architecture": {
390
+ "network_class_name": "dynamic_network_architectures.architectures.unet.ResidualEncoderUNet",
391
+ "arch_kwargs": {
392
+ "n_stages": 6,
393
+ "features_per_stage": [
394
+ 32,
395
+ 64,
396
+ 128,
397
+ 256,
398
+ 320,
399
+ 320
400
+ ],
401
+ "conv_op": "torch.nn.modules.conv.Conv3d",
402
+ "kernel_sizes": [
403
+ [
404
+ 1,
405
+ 3,
406
+ 3
407
+ ],
408
+ [
409
+ 3,
410
+ 3,
411
+ 3
412
+ ],
413
+ [
414
+ 3,
415
+ 3,
416
+ 3
417
+ ],
418
+ [
419
+ 3,
420
+ 3,
421
+ 3
422
+ ],
423
+ [
424
+ 3,
425
+ 3,
426
+ 3
427
+ ],
428
+ [
429
+ 3,
430
+ 3,
431
+ 3
432
+ ]
433
+ ],
434
+ "strides": [
435
+ [
436
+ 1,
437
+ 1,
438
+ 1
439
+ ],
440
+ [
441
+ 1,
442
+ 2,
443
+ 2
444
+ ],
445
+ [
446
+ 2,
447
+ 2,
448
+ 2
449
+ ],
450
+ [
451
+ 2,
452
+ 2,
453
+ 2
454
+ ],
455
+ [
456
+ 2,
457
+ 2,
458
+ 2
459
+ ],
460
+ [
461
+ 1,
462
+ 2,
463
+ 2
464
+ ]
465
+ ],
466
+ "n_blocks_per_stage": [
467
+ 1,
468
+ 3,
469
+ 4,
470
+ 6,
471
+ 6,
472
+ 6
473
+ ],
474
+ "n_conv_per_stage_decoder": [
475
+ 1,
476
+ 1,
477
+ 1,
478
+ 1,
479
+ 1
480
+ ],
481
+ "conv_bias": true,
482
+ "norm_op": "torch.nn.modules.instancenorm.InstanceNorm3d",
483
+ "norm_op_kwargs": {
484
+ "eps": 1e-05,
485
+ "affine": true
486
+ },
487
+ "dropout_op": null,
488
+ "dropout_op_kwargs": null,
489
+ "nonlin": "torch.nn.LeakyReLU",
490
+ "nonlin_kwargs": {
491
+ "inplace": true
492
+ }
493
+ },
494
+ "_kw_requires_import": [
495
+ "conv_op",
496
+ "norm_op",
497
+ "dropout_op",
498
+ "nonlin"
499
+ ]
500
+ },
501
+ "batch_dice": true
502
+ },
503
+ "3d_cascade_fullres": {
504
+ "inherits_from": "3d_fullres",
505
+ "previous_stage": "3d_lowres"
506
+ }
507
+ },
508
+ "experiment_planner_used": "nnUNetPlannerResEncL",
509
+ "label_manager": "LabelManager",
510
+ "foreground_intensity_properties_per_channel": {
511
+ "0": {
512
+ "max": 3071.0,
513
+ "mean": -330.8976135253906,
514
+ "median": -243.0,
515
+ "min": -1024.0,
516
+ "percentile_00_5": -1024.0,
517
+ "percentile_99_5": 563.0,
518
+ "std": 371.2642822265625
519
+ }
520
+ }
521
+ }