Add files using upload-large-folder tool
Browse files- .DS_Store +0 -0
- .gitattributes +1 -0
- BUSI_with_classes/.DS_Store +0 -0
- BUSI_with_classes/pipeline_check.json +0 -0
- changed_mlr_all_strategies (5).py +0 -0
- dataset_splits_busi_with_classes_stratified.json +0 -0
- hard_coded_param/strat1_pct10.json +8 -0
- hard_coded_param/strat3_pct10.json +8 -0
- requirements_runpod.txt +13 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/norm_stats_BUSI_with_classes_stratified_80_10_10_50pct.json +266 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/best_params.json +8 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0030.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0040.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0050.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0060.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0070.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0080.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0090.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0100.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/evaluation.json +619 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/evaluation_results.xlsx +0 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/evaluation_summary.json +4 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/history.json +2242 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/low_iou_samples.csv +4 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/run_config.json +40 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/summary.json +11 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64/pct_50/norm_stats_BUSI_with_classes_stratified_80_10_10_50pct.json +266 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64/pct_50/strategy_3/final/evaluation_summary.json +4 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64/pct_50/strategy_3/final/low_iou_samples.csv +5 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/norm_stats_BUSI_with_classes_stratified_80_10_10_50pct.json +266 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/best_params.json +8 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/best.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0010.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0020.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0030.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0040.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0050.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0060.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0070.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0080.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0090.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0100.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/latest.pt.meta.json +58 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/evaluation.json +619 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/evaluation_results.xlsx +0 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/evaluation_summary.json +4 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/history.json +2242 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/low_iou_samples.csv +3 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/run_config.json +40 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/summary.json +11 -0
.DS_Store
ADDED
|
Binary file (6.15 kB). View file
|
|
|
.gitattributes
CHANGED
|
@@ -58,3 +58,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 58 |
# Video files - compressed
|
| 59 |
*.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 60 |
*.webm filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 58 |
# Video files - compressed
|
| 59 |
*.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 60 |
*.webm filter=lfs diff=lfs merge=lfs -text
|
| 61 |
+
__pycache__/Changed_MLR_All_Strategies.cpython-312.pyc filter=lfs diff=lfs merge=lfs -text
|
BUSI_with_classes/.DS_Store
ADDED
|
Binary file (6.15 kB). View file
|
|
|
BUSI_with_classes/pipeline_check.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
changed_mlr_all_strategies (5).py
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
dataset_splits_busi_with_classes_stratified.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
hard_coded_param/strat1_pct10.json
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"head_lr": 1e-3,
|
| 3 |
+
"encoder_lr": 1e-3,
|
| 4 |
+
"weight_decay": 1e-4,
|
| 5 |
+
"dropout_p": 0.2,
|
| 6 |
+
"tmax": 6,
|
| 7 |
+
"entropy_lr": 2e-05
|
| 8 |
+
}
|
hard_coded_param/strat3_pct10.json
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"head_lr": 2e-3,
|
| 3 |
+
"encoder_lr": 2e-4,
|
| 4 |
+
"weight_decay": 2e-3,
|
| 5 |
+
"dropout_p": 0.3,
|
| 6 |
+
"tmax": 6,
|
| 7 |
+
"entropy_lr": 2e-04
|
| 8 |
+
}
|
requirements_runpod.txt
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Use this with a RunPod PyTorch image that already includes torch/torchvision + CUDA.
|
| 2 |
+
# This avoids wasting setup time reinstalling large GPU wheels.
|
| 3 |
+
|
| 4 |
+
numpy>=1.24.0
|
| 5 |
+
scipy>=1.10.0
|
| 6 |
+
pillow>=10.0.0
|
| 7 |
+
tqdm>=4.66.0
|
| 8 |
+
matplotlib>=3.8.0
|
| 9 |
+
pandas>=2.2.0
|
| 10 |
+
openpyxl>=3.1.0
|
| 11 |
+
optuna>=3.6.0
|
| 12 |
+
segmentation-models-pytorch>=0.5.0
|
| 13 |
+
timm>=1.0.0
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/norm_stats_BUSI_with_classes_stratified_80_10_10_50pct.json
ADDED
|
@@ -0,0 +1,266 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"global_mean": 85.12605579417583,
|
| 3 |
+
"global_std": 56.68613974260357,
|
| 4 |
+
"total_pixels": 244630659,
|
| 5 |
+
"num_images": 258,
|
| 6 |
+
"filenames": [
|
| 7 |
+
"BUSI_0034_B.png",
|
| 8 |
+
"BUSI_0155_M.png",
|
| 9 |
+
"BUSI_0005_M.png",
|
| 10 |
+
"BUSI_0260_B.png",
|
| 11 |
+
"BUSI_0386_B.png",
|
| 12 |
+
"BUSI_0280_B.png",
|
| 13 |
+
"BUSI_0047_B.png",
|
| 14 |
+
"BUSI_0222_B.png",
|
| 15 |
+
"BUSI_0205_B.png",
|
| 16 |
+
"BUSI_0403_B.png",
|
| 17 |
+
"BUSI_0133_B.png",
|
| 18 |
+
"BUSI_0016_B.png",
|
| 19 |
+
"BUSI_0105_M.png",
|
| 20 |
+
"BUSI_0220_B.png",
|
| 21 |
+
"BUSI_0066_B.png",
|
| 22 |
+
"BUSI_0060_B.png",
|
| 23 |
+
"BUSI_0351_B.png",
|
| 24 |
+
"BUSI_0029_M.png",
|
| 25 |
+
"BUSI_0287_B.png",
|
| 26 |
+
"BUSI_0046_M.png",
|
| 27 |
+
"BUSI_0123_B.png",
|
| 28 |
+
"BUSI_0359_B.png",
|
| 29 |
+
"BUSI_0215_B.png",
|
| 30 |
+
"BUSI_0377_B.png",
|
| 31 |
+
"BUSI_0029_B.png",
|
| 32 |
+
"BUSI_0158_M.png",
|
| 33 |
+
"BUSI_0375_B.png",
|
| 34 |
+
"BUSI_0221_B.png",
|
| 35 |
+
"BUSI_0120_B.png",
|
| 36 |
+
"BUSI_0404_B.png",
|
| 37 |
+
"BUSI_0093_M.png",
|
| 38 |
+
"BUSI_0106_M.png",
|
| 39 |
+
"BUSI_0030_B.png",
|
| 40 |
+
"BUSI_0407_B.png",
|
| 41 |
+
"BUSI_0131_B.png",
|
| 42 |
+
"BUSI_0092_B.png",
|
| 43 |
+
"BUSI_0112_M.png",
|
| 44 |
+
"BUSI_0177_M.png",
|
| 45 |
+
"BUSI_0333_B.png",
|
| 46 |
+
"BUSI_0206_B.png",
|
| 47 |
+
"BUSI_0239_B.png",
|
| 48 |
+
"BUSI_0165_B.png",
|
| 49 |
+
"BUSI_0396_B.png",
|
| 50 |
+
"BUSI_0207_B.png",
|
| 51 |
+
"BUSI_0158_B.png",
|
| 52 |
+
"BUSI_0160_B.png",
|
| 53 |
+
"BUSI_0091_B.png",
|
| 54 |
+
"BUSI_0090_M.png",
|
| 55 |
+
"BUSI_0035_B.png",
|
| 56 |
+
"BUSI_0281_B.png",
|
| 57 |
+
"BUSI_0410_B.png",
|
| 58 |
+
"BUSI_0322_B.png",
|
| 59 |
+
"BUSI_0236_B.png",
|
| 60 |
+
"BUSI_0385_B.png",
|
| 61 |
+
"BUSI_0193_M.png",
|
| 62 |
+
"BUSI_0388_B.png",
|
| 63 |
+
"BUSI_0144_M.png",
|
| 64 |
+
"BUSI_0293_B.png",
|
| 65 |
+
"BUSI_0088_M.png",
|
| 66 |
+
"BUSI_0052_B.png",
|
| 67 |
+
"BUSI_0346_B.png",
|
| 68 |
+
"BUSI_0262_B.png",
|
| 69 |
+
"BUSI_0408_B.png",
|
| 70 |
+
"BUSI_0146_M.png",
|
| 71 |
+
"BUSI_0309_B.png",
|
| 72 |
+
"BUSI_0083_B.png",
|
| 73 |
+
"BUSI_0190_B.png",
|
| 74 |
+
"BUSI_0426_B.png",
|
| 75 |
+
"BUSI_0019_B.png",
|
| 76 |
+
"BUSI_0274_B.png",
|
| 77 |
+
"BUSI_0162_M.png",
|
| 78 |
+
"BUSI_0179_M.png",
|
| 79 |
+
"BUSI_0078_B.png",
|
| 80 |
+
"BUSI_0328_B.png",
|
| 81 |
+
"BUSI_0001_M.png",
|
| 82 |
+
"BUSI_0301_B.png",
|
| 83 |
+
"BUSI_0109_M.png",
|
| 84 |
+
"BUSI_0011_B.png",
|
| 85 |
+
"BUSI_0063_B.png",
|
| 86 |
+
"BUSI_0360_B.png",
|
| 87 |
+
"BUSI_0114_B.png",
|
| 88 |
+
"BUSI_0430_B.png",
|
| 89 |
+
"BUSI_0391_B.png",
|
| 90 |
+
"BUSI_0133_M.png",
|
| 91 |
+
"BUSI_0188_M.png",
|
| 92 |
+
"BUSI_0424_B.png",
|
| 93 |
+
"BUSI_0179_B.png",
|
| 94 |
+
"BUSI_0390_B.png",
|
| 95 |
+
"BUSI_0030_M.png",
|
| 96 |
+
"BUSI_0235_B.png",
|
| 97 |
+
"BUSI_0065_M.png",
|
| 98 |
+
"BUSI_0101_M.png",
|
| 99 |
+
"BUSI_0355_B.png",
|
| 100 |
+
"BUSI_0053_M.png",
|
| 101 |
+
"BUSI_0203_B.png",
|
| 102 |
+
"BUSI_0007_B.png",
|
| 103 |
+
"BUSI_0344_B.png",
|
| 104 |
+
"BUSI_0129_B.png",
|
| 105 |
+
"BUSI_0271_B.png",
|
| 106 |
+
"BUSI_0252_B.png",
|
| 107 |
+
"BUSI_0296_B.png",
|
| 108 |
+
"BUSI_0183_M.png",
|
| 109 |
+
"BUSI_0310_B.png",
|
| 110 |
+
"BUSI_0409_B.png",
|
| 111 |
+
"BUSI_0041_M.png",
|
| 112 |
+
"BUSI_0092_M.png",
|
| 113 |
+
"BUSI_0273_B.png",
|
| 114 |
+
"BUSI_0206_M.png",
|
| 115 |
+
"BUSI_0024_M.png",
|
| 116 |
+
"BUSI_0015_B.png",
|
| 117 |
+
"BUSI_0094_M.png",
|
| 118 |
+
"BUSI_0055_B.png",
|
| 119 |
+
"BUSI_0094_B.png",
|
| 120 |
+
"BUSI_0077_B.png",
|
| 121 |
+
"BUSI_0347_B.png",
|
| 122 |
+
"BUSI_0126_M.png",
|
| 123 |
+
"BUSI_0186_M.png",
|
| 124 |
+
"BUSI_0145_M.png",
|
| 125 |
+
"BUSI_0054_M.png",
|
| 126 |
+
"BUSI_0074_M.png",
|
| 127 |
+
"BUSI_0111_B.png",
|
| 128 |
+
"BUSI_0032_M.png",
|
| 129 |
+
"BUSI_0176_M.png",
|
| 130 |
+
"BUSI_0002_M.png",
|
| 131 |
+
"BUSI_0069_B.png",
|
| 132 |
+
"BUSI_0027_B.png",
|
| 133 |
+
"BUSI_0109_B.png",
|
| 134 |
+
"BUSI_0389_B.png",
|
| 135 |
+
"BUSI_0088_B.png",
|
| 136 |
+
"BUSI_0169_B.png",
|
| 137 |
+
"BUSI_0157_B.png",
|
| 138 |
+
"BUSI_0058_B.png",
|
| 139 |
+
"BUSI_0231_B.png",
|
| 140 |
+
"BUSI_0166_B.png",
|
| 141 |
+
"BUSI_0098_M.png",
|
| 142 |
+
"BUSI_0020_M.png",
|
| 143 |
+
"BUSI_0182_B.png",
|
| 144 |
+
"BUSI_0397_B.png",
|
| 145 |
+
"BUSI_0045_M.png",
|
| 146 |
+
"BUSI_0138_M.png",
|
| 147 |
+
"BUSI_0036_M.png",
|
| 148 |
+
"BUSI_0380_B.png",
|
| 149 |
+
"BUSI_0137_B.png",
|
| 150 |
+
"BUSI_0205_M.png",
|
| 151 |
+
"BUSI_0107_M.png",
|
| 152 |
+
"BUSI_0044_M.png",
|
| 153 |
+
"BUSI_0213_B.png",
|
| 154 |
+
"BUSI_0070_B.png",
|
| 155 |
+
"BUSI_0295_B.png",
|
| 156 |
+
"BUSI_0173_B.png",
|
| 157 |
+
"BUSI_0164_M.png",
|
| 158 |
+
"BUSI_0039_M.png",
|
| 159 |
+
"BUSI_0421_B.png",
|
| 160 |
+
"BUSI_0174_B.png",
|
| 161 |
+
"BUSI_0416_B.png",
|
| 162 |
+
"BUSI_0124_M.png",
|
| 163 |
+
"BUSI_0428_B.png",
|
| 164 |
+
"BUSI_0181_B.png",
|
| 165 |
+
"BUSI_0076_B.png",
|
| 166 |
+
"BUSI_0073_B.png",
|
| 167 |
+
"BUSI_0045_B.png",
|
| 168 |
+
"BUSI_0104_M.png",
|
| 169 |
+
"BUSI_0272_B.png",
|
| 170 |
+
"BUSI_0405_B.png",
|
| 171 |
+
"BUSI_0425_B.png",
|
| 172 |
+
"BUSI_0212_B.png",
|
| 173 |
+
"BUSI_0156_M.png",
|
| 174 |
+
"BUSI_0339_B.png",
|
| 175 |
+
"BUSI_0160_M.png",
|
| 176 |
+
"BUSI_0204_M.png",
|
| 177 |
+
"BUSI_0082_B.png",
|
| 178 |
+
"BUSI_0070_M.png",
|
| 179 |
+
"BUSI_0061_M.png",
|
| 180 |
+
"BUSI_0136_M.png",
|
| 181 |
+
"BUSI_0024_B.png",
|
| 182 |
+
"BUSI_0241_B.png",
|
| 183 |
+
"BUSI_0420_B.png",
|
| 184 |
+
"BUSI_0196_B.png",
|
| 185 |
+
"BUSI_0406_B.png",
|
| 186 |
+
"BUSI_0374_B.png",
|
| 187 |
+
"BUSI_0113_M.png",
|
| 188 |
+
"BUSI_0009_B.png",
|
| 189 |
+
"BUSI_0292_B.png",
|
| 190 |
+
"BUSI_0079_M.png",
|
| 191 |
+
"BUSI_0118_B.png",
|
| 192 |
+
"BUSI_0175_M.png",
|
| 193 |
+
"BUSI_0011_M.png",
|
| 194 |
+
"BUSI_0187_B.png",
|
| 195 |
+
"BUSI_0034_M.png",
|
| 196 |
+
"BUSI_0321_B.png",
|
| 197 |
+
"BUSI_0078_M.png",
|
| 198 |
+
"BUSI_0180_B.png",
|
| 199 |
+
"BUSI_0398_B.png",
|
| 200 |
+
"BUSI_0195_B.png",
|
| 201 |
+
"BUSI_0203_M.png",
|
| 202 |
+
"BUSI_0167_B.png",
|
| 203 |
+
"BUSI_0263_B.png",
|
| 204 |
+
"BUSI_0012_B.png",
|
| 205 |
+
"BUSI_0177_B.png",
|
| 206 |
+
"BUSI_0147_M.png",
|
| 207 |
+
"BUSI_0005_B.png",
|
| 208 |
+
"BUSI_0012_M.png",
|
| 209 |
+
"BUSI_0170_B.png",
|
| 210 |
+
"BUSI_0025_B.png",
|
| 211 |
+
"BUSI_0136_B.png",
|
| 212 |
+
"BUSI_0089_B.png",
|
| 213 |
+
"BUSI_0018_B.png",
|
| 214 |
+
"BUSI_0202_M.png",
|
| 215 |
+
"BUSI_0071_B.png",
|
| 216 |
+
"BUSI_0185_M.png",
|
| 217 |
+
"BUSI_0096_B.png",
|
| 218 |
+
"BUSI_0184_M.png",
|
| 219 |
+
"BUSI_0248_B.png",
|
| 220 |
+
"BUSI_0113_B.png",
|
| 221 |
+
"BUSI_0246_B.png",
|
| 222 |
+
"BUSI_0014_M.png",
|
| 223 |
+
"BUSI_0245_B.png",
|
| 224 |
+
"BUSI_0135_B.png",
|
| 225 |
+
"BUSI_0031_B.png",
|
| 226 |
+
"BUSI_0365_B.png",
|
| 227 |
+
"BUSI_0155_B.png",
|
| 228 |
+
"BUSI_0103_M.png",
|
| 229 |
+
"BUSI_0023_B.png",
|
| 230 |
+
"BUSI_0242_B.png",
|
| 231 |
+
"BUSI_0370_B.png",
|
| 232 |
+
"BUSI_0367_B.png",
|
| 233 |
+
"BUSI_0035_M.png",
|
| 234 |
+
"BUSI_0240_B.png",
|
| 235 |
+
"BUSI_0148_M.png",
|
| 236 |
+
"BUSI_0064_B.png",
|
| 237 |
+
"BUSI_0435_B.png",
|
| 238 |
+
"BUSI_0373_B.png",
|
| 239 |
+
"BUSI_0065_B.png",
|
| 240 |
+
"BUSI_0312_B.png",
|
| 241 |
+
"BUSI_0164_B.png",
|
| 242 |
+
"BUSI_0049_B.png",
|
| 243 |
+
"BUSI_0181_M.png",
|
| 244 |
+
"BUSI_0353_B.png",
|
| 245 |
+
"BUSI_0038_M.png",
|
| 246 |
+
"BUSI_0356_B.png",
|
| 247 |
+
"BUSI_0159_B.png",
|
| 248 |
+
"BUSI_0049_M.png",
|
| 249 |
+
"BUSI_0051_B.png",
|
| 250 |
+
"BUSI_0308_B.png",
|
| 251 |
+
"BUSI_0368_B.png",
|
| 252 |
+
"BUSI_0084_B.png",
|
| 253 |
+
"BUSI_0072_M.png",
|
| 254 |
+
"BUSI_0197_M.png",
|
| 255 |
+
"BUSI_0329_B.png",
|
| 256 |
+
"BUSI_0174_M.png",
|
| 257 |
+
"BUSI_0196_M.png",
|
| 258 |
+
"BUSI_0169_M.png",
|
| 259 |
+
"BUSI_0101_B.png",
|
| 260 |
+
"BUSI_0086_M.png",
|
| 261 |
+
"BUSI_0317_B.png",
|
| 262 |
+
"BUSI_0077_M.png",
|
| 263 |
+
"BUSI_0193_B.png",
|
| 264 |
+
"BUSI_0031_M.png"
|
| 265 |
+
]
|
| 266 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/best_params.json
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"head_lr": 0.002,
|
| 3 |
+
"encoder_lr": 0.0002,
|
| 4 |
+
"weight_decay": 0.002,
|
| 5 |
+
"dropout_p": 0.3,
|
| 6 |
+
"tmax": 6,
|
| 7 |
+
"entropy_lr": 0.0002
|
| 8 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0030.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0030.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 30,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0040.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0040.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 40,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0050.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0050.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 50,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0060.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0060.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 60,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0070.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0070.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 70,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0080.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0080.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 80,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0090.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0090.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 90,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0100.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/epoch_0100.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 100,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/evaluation.json
ADDED
|
@@ -0,0 +1,619 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"strategy": 3,
|
| 3 |
+
"best_metric_name": "val_iou",
|
| 4 |
+
"metrics": {
|
| 5 |
+
"dice": {
|
| 6 |
+
"mean": 0.7611992359161377,
|
| 7 |
+
"std": 0.2595278322696686
|
| 8 |
+
},
|
| 9 |
+
"ppv": {
|
| 10 |
+
"mean": 0.7929834127426147,
|
| 11 |
+
"std": 0.25871941447257996
|
| 12 |
+
},
|
| 13 |
+
"sen": {
|
| 14 |
+
"mean": 0.8039218187332153,
|
| 15 |
+
"std": 0.26019707322120667
|
| 16 |
+
},
|
| 17 |
+
"iou": {
|
| 18 |
+
"mean": 0.6688458919525146,
|
| 19 |
+
"std": 0.2634357810020447
|
| 20 |
+
},
|
| 21 |
+
"biou": {
|
| 22 |
+
"mean": 0.10918061435222626,
|
| 23 |
+
"std": 0.10209918022155762
|
| 24 |
+
},
|
| 25 |
+
"hd95": {
|
| 26 |
+
"mean": 13.542085647583008,
|
| 27 |
+
"std": 16.97413444519043
|
| 28 |
+
}
|
| 29 |
+
},
|
| 30 |
+
"per_sample": [
|
| 31 |
+
{
|
| 32 |
+
"sample_id": "BUSI_0097_B",
|
| 33 |
+
"dice": 0.6917765287345359,
|
| 34 |
+
"ppv": 0.851184353698025,
|
| 35 |
+
"sen": 0.5826577517568645,
|
| 36 |
+
"iou": 0.5287908020220472,
|
| 37 |
+
"biou": 0.011267884149891789,
|
| 38 |
+
"hd95": 21.947614669799805
|
| 39 |
+
},
|
| 40 |
+
{
|
| 41 |
+
"sample_id": "BUSI_0415_B",
|
| 42 |
+
"dice": 0.912280702337316,
|
| 43 |
+
"ppv": 0.9188755030940361,
|
| 44 |
+
"sen": 0.9057798903961481,
|
| 45 |
+
"iou": 0.8387096793901554,
|
| 46 |
+
"biou": 0.023006284814986993,
|
| 47 |
+
"hd95": 16.970561981201172
|
| 48 |
+
},
|
| 49 |
+
{
|
| 50 |
+
"sample_id": "BUSI_0290_B",
|
| 51 |
+
"dice": 0.9323008864535202,
|
| 52 |
+
"ppv": 0.8935538637169693,
|
| 53 |
+
"sen": 0.9745605932210641,
|
| 54 |
+
"iou": 0.8731869095239574,
|
| 55 |
+
"biou": 0.039039327615817535,
|
| 56 |
+
"hd95": 3.0
|
| 57 |
+
},
|
| 58 |
+
{
|
| 59 |
+
"sample_id": "BUSI_0354_B",
|
| 60 |
+
"dice": 0.8286445061482095,
|
| 61 |
+
"ppv": 0.707423594981873,
|
| 62 |
+
"sen": 1.0,
|
| 63 |
+
"iou": 0.707423594981873,
|
| 64 |
+
"biou": 0.02649038857934153,
|
| 65 |
+
"hd95": 10.833560943603516
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"sample_id": "BUSI_0067_B",
|
| 69 |
+
"dice": 0.9320945957416453,
|
| 70 |
+
"ppv": 0.9490884090440864,
|
| 71 |
+
"sen": 0.9156986420279243,
|
| 72 |
+
"iou": 0.8728250593854774,
|
| 73 |
+
"biou": 0.15160374588812076,
|
| 74 |
+
"hd95": 7.0
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"sample_id": "BUSI_0165_M",
|
| 78 |
+
"dice": 0.8645071950103317,
|
| 79 |
+
"ppv": 0.921285668722767,
|
| 80 |
+
"sen": 0.8143209187661847,
|
| 81 |
+
"iou": 0.7613497796266462,
|
| 82 |
+
"biou": 0.03463869960260548,
|
| 83 |
+
"hd95": 19.979145050048828
|
| 84 |
+
},
|
| 85 |
+
{
|
| 86 |
+
"sample_id": "BUSI_0198_B",
|
| 87 |
+
"dice": 0.9495772017044409,
|
| 88 |
+
"ppv": 0.9238269394377245,
|
| 89 |
+
"sen": 0.9768041252059197,
|
| 90 |
+
"iou": 0.9039952353014171,
|
| 91 |
+
"biou": 0.09090943526157756,
|
| 92 |
+
"hd95": 2.0
|
| 93 |
+
},
|
| 94 |
+
{
|
| 95 |
+
"sample_id": "BUSI_0361_B",
|
| 96 |
+
"dice": 0.9217585722146583,
|
| 97 |
+
"ppv": 0.8861031600212637,
|
| 98 |
+
"sen": 0.9604037297823191,
|
| 99 |
+
"iou": 0.8548721593039281,
|
| 100 |
+
"biou": 0.14693912369831688,
|
| 101 |
+
"hd95": 3.0
|
| 102 |
+
},
|
| 103 |
+
{
|
| 104 |
+
"sample_id": "BUSI_0095_B",
|
| 105 |
+
"dice": 0.9304556521689084,
|
| 106 |
+
"ppv": 0.965174146679529,
|
| 107 |
+
"sen": 0.8981481953017615,
|
| 108 |
+
"iou": 0.8699552152667196,
|
| 109 |
+
"biou": 0.32857238775373176,
|
| 110 |
+
"hd95": 1.4142135381698608
|
| 111 |
+
},
|
| 112 |
+
{
|
| 113 |
+
"sample_id": "BUSI_0057_B",
|
| 114 |
+
"dice": 0.9166666859567857,
|
| 115 |
+
"ppv": 0.9428571700680143,
|
| 116 |
+
"sen": 0.891891940589216,
|
| 117 |
+
"iou": 0.8461539119000376,
|
| 118 |
+
"biou": 0.20253265502195567,
|
| 119 |
+
"hd95": 1.4142135381698608
|
| 120 |
+
},
|
| 121 |
+
{
|
| 122 |
+
"sample_id": "BUSI_0009_M",
|
| 123 |
+
"dice": 0.5284553228898111,
|
| 124 |
+
"ppv": 0.8904109889284961,
|
| 125 |
+
"sen": 0.37572261552339703,
|
| 126 |
+
"iou": 0.35911609291534885,
|
| 127 |
+
"biou": 0.07784486356594995,
|
| 128 |
+
"hd95": 15.265974044799805
|
| 129 |
+
},
|
| 130 |
+
{
|
| 131 |
+
"sample_id": "BUSI_0399_B",
|
| 132 |
+
"dice": 0.9445490700031582,
|
| 133 |
+
"ppv": 0.9053349416231472,
|
| 134 |
+
"sen": 0.9873140862942219,
|
| 135 |
+
"iou": 0.8949246671322495,
|
| 136 |
+
"biou": 0.12420410057194252,
|
| 137 |
+
"hd95": 4.123105525970459
|
| 138 |
+
},
|
| 139 |
+
{
|
| 140 |
+
"sample_id": "BUSI_0099_M",
|
| 141 |
+
"dice": 0.7458823678892724,
|
| 142 |
+
"ppv": 0.8475936032628872,
|
| 143 |
+
"sen": 0.6659664216421826,
|
| 144 |
+
"iou": 0.5947467547141881,
|
| 145 |
+
"biou": 0.05882399077255354,
|
| 146 |
+
"hd95": 11.176175117492676
|
| 147 |
+
},
|
| 148 |
+
{
|
| 149 |
+
"sample_id": "BUSI_0307_B",
|
| 150 |
+
"dice": 0.9220183575667021,
|
| 151 |
+
"ppv": 0.9095022829180356,
|
| 152 |
+
"sen": 0.9348837360735498,
|
| 153 |
+
"iou": 0.8553191797193235,
|
| 154 |
+
"biou": 0.2096780567112446,
|
| 155 |
+
"hd95": 2.0
|
| 156 |
+
},
|
| 157 |
+
{
|
| 158 |
+
"sample_id": "BUSI_0039_B",
|
| 159 |
+
"dice": 0.8976572259362237,
|
| 160 |
+
"ppv": 0.8272727665289167,
|
| 161 |
+
"sen": 0.9811320805573907,
|
| 162 |
+
"iou": 0.8143177149177372,
|
| 163 |
+
"biou": 0.15929277938692088,
|
| 164 |
+
"hd95": 2.2360680103302
|
| 165 |
+
},
|
| 166 |
+
{
|
| 167 |
+
"sample_id": "BUSI_0189_M",
|
| 168 |
+
"dice": 0.16069297608443917,
|
| 169 |
+
"ppv": 0.08739444160511885,
|
| 170 |
+
"sen": 0.996296297668038,
|
| 171 |
+
"iou": 0.08736605757174219,
|
| 172 |
+
"biou": 0.02181853752053181,
|
| 173 |
+
"hd95": 46.010868072509766
|
| 174 |
+
},
|
| 175 |
+
{
|
| 176 |
+
"sample_id": "BUSI_0021_B",
|
| 177 |
+
"dice": 0.8686131866375232,
|
| 178 |
+
"ppv": 0.7880796105433043,
|
| 179 |
+
"sen": 0.9674797012360152,
|
| 180 |
+
"iou": 0.7677420853276868,
|
| 181 |
+
"biou": 0.1666679292910162,
|
| 182 |
+
"hd95": 2.0
|
| 183 |
+
},
|
| 184 |
+
{
|
| 185 |
+
"sample_id": "BUSI_0311_B",
|
| 186 |
+
"dice": 0.5763503773328357,
|
| 187 |
+
"ppv": 0.9030192220476466,
|
| 188 |
+
"sen": 0.4232418772194735,
|
| 189 |
+
"iou": 0.40484005722559235,
|
| 190 |
+
"biou": 0.02676422706466495,
|
| 191 |
+
"hd95": 11.661903381347656
|
| 192 |
+
},
|
| 193 |
+
{
|
| 194 |
+
"sample_id": "BUSI_0080_M",
|
| 195 |
+
"dice": 0.6462027084889466,
|
| 196 |
+
"ppv": 0.5611213436984676,
|
| 197 |
+
"sen": 0.7616968333314514,
|
| 198 |
+
"iou": 0.4773260563985122,
|
| 199 |
+
"biou": 0.04054079985383788,
|
| 200 |
+
"hd95": 22.19007110595703
|
| 201 |
+
},
|
| 202 |
+
{
|
| 203 |
+
"sample_id": "BUSI_0130_B",
|
| 204 |
+
"dice": 0.9573361104403688,
|
| 205 |
+
"ppv": 0.924623123153455,
|
| 206 |
+
"sen": 0.9924487602536397,
|
| 207 |
+
"iou": 0.918163680821988,
|
| 208 |
+
"biou": 0.21893537341102168,
|
| 209 |
+
"hd95": 1.4142135381698608
|
| 210 |
+
},
|
| 211 |
+
{
|
| 212 |
+
"sample_id": "BUSI_0401_B",
|
| 213 |
+
"dice": 0.8443240034831676,
|
| 214 |
+
"ppv": 0.9127604223463267,
|
| 215 |
+
"sen": 0.7854341856899616,
|
| 216 |
+
"iou": 0.7305888623976622,
|
| 217 |
+
"biou": 0.05118147591280476,
|
| 218 |
+
"hd95": 10.770329475402832
|
| 219 |
+
},
|
| 220 |
+
{
|
| 221 |
+
"sample_id": "BUSI_0326_B",
|
| 222 |
+
"dice": 0.9327995443866302,
|
| 223 |
+
"ppv": 0.8830536067648291,
|
| 224 |
+
"sen": 0.9884848491827364,
|
| 225 |
+
"iou": 0.8740621718080294,
|
| 226 |
+
"biou": 0.050584027010106225,
|
| 227 |
+
"hd95": 2.8284270763397217
|
| 228 |
+
},
|
| 229 |
+
{
|
| 230 |
+
"sample_id": "BUSI_0119_B",
|
| 231 |
+
"dice": 0.16860990275291912,
|
| 232 |
+
"ppv": 0.09206664583415054,
|
| 233 |
+
"sen": 1.0,
|
| 234 |
+
"iou": 0.09206664583415054,
|
| 235 |
+
"biou": 4.0322564386062747e-07,
|
| 236 |
+
"hd95": 33.48348617553711
|
| 237 |
+
},
|
| 238 |
+
{
|
| 239 |
+
"sample_id": "BUSI_0372_B",
|
| 240 |
+
"dice": 0.9119403017932723,
|
| 241 |
+
"ppv": 0.8527564652647268,
|
| 242 |
+
"sen": 0.9799518861305625,
|
| 243 |
+
"iou": 0.8381344418289135,
|
| 244 |
+
"biou": 0.09012914586731936,
|
| 245 |
+
"hd95": 4.4721360206604
|
| 246 |
+
},
|
| 247 |
+
{
|
| 248 |
+
"sample_id": "BUSI_0003_M",
|
| 249 |
+
"dice": 0.7971864128261188,
|
| 250 |
+
"ppv": 0.9550561860876143,
|
| 251 |
+
"sen": 0.6841046595468149,
|
| 252 |
+
"iou": 0.6627680640576935,
|
| 253 |
+
"biou": 0.11734738910847495,
|
| 254 |
+
"hd95": 11.058856964111328
|
| 255 |
+
},
|
| 256 |
+
{
|
| 257 |
+
"sample_id": "BUSI_0007_M",
|
| 258 |
+
"dice": 0.1505618931952929,
|
| 259 |
+
"ppv": 1.0,
|
| 260 |
+
"sen": 0.08140958913613741,
|
| 261 |
+
"iou": 0.08140958913613741,
|
| 262 |
+
"biou": 0.027778452931629907,
|
| 263 |
+
"hd95": 31.72361946105957
|
| 264 |
+
},
|
| 265 |
+
{
|
| 266 |
+
"sample_id": "BUSI_0411_B",
|
| 267 |
+
"dice": 0.9211723112386986,
|
| 268 |
+
"ppv": 0.873921386676827,
|
| 269 |
+
"sen": 0.9738247877230348,
|
| 270 |
+
"iou": 0.8538641754630362,
|
| 271 |
+
"biou": 0.05678263192976911,
|
| 272 |
+
"hd95": 4.0
|
| 273 |
+
},
|
| 274 |
+
{
|
| 275 |
+
"sample_id": "BUSI_0211_B",
|
| 276 |
+
"dice": 0.46938789046227286,
|
| 277 |
+
"ppv": 0.8518519890259361,
|
| 278 |
+
"sen": 0.32394390001975354,
|
| 279 |
+
"iou": 0.3066668977777008,
|
| 280 |
+
"biou": 0.06976852352497263,
|
| 281 |
+
"hd95": 12.80700969696045
|
| 282 |
+
},
|
| 283 |
+
{
|
| 284 |
+
"sample_id": "BUSI_0202_B",
|
| 285 |
+
"dice": 0.8864865478451093,
|
| 286 |
+
"ppv": 0.9111112098764335,
|
| 287 |
+
"sen": 0.8631580387810118,
|
| 288 |
+
"iou": 0.7961167027993177,
|
| 289 |
+
"biou": 0.2448995002051016,
|
| 290 |
+
"hd95": 1.0
|
| 291 |
+
},
|
| 292 |
+
{
|
| 293 |
+
"sample_id": "BUSI_0194_M",
|
| 294 |
+
"dice": 0.7320636646996175,
|
| 295 |
+
"ppv": 0.8199883214501273,
|
| 296 |
+
"sen": 0.6611687247328593,
|
| 297 |
+
"iou": 0.5773662725363673,
|
| 298 |
+
"biou": 0.02752323317332319,
|
| 299 |
+
"hd95": 12.041594505310059
|
| 300 |
+
},
|
| 301 |
+
{
|
| 302 |
+
"sample_id": "BUSI_0141_B",
|
| 303 |
+
"dice": 0.5810111008548204,
|
| 304 |
+
"ppv": 0.9393939410049458,
|
| 305 |
+
"sen": 0.42056409115120685,
|
| 306 |
+
"iou": 0.40945429950811846,
|
| 307 |
+
"biou": 0.01883578444592732,
|
| 308 |
+
"hd95": 53.375057220458984
|
| 309 |
+
},
|
| 310 |
+
{
|
| 311 |
+
"sample_id": "BUSI_0251_B",
|
| 312 |
+
"dice": 0.8521462874171244,
|
| 313 |
+
"ppv": 0.9305555796682015,
|
| 314 |
+
"sen": 0.7859238164446286,
|
| 315 |
+
"iou": 0.742382342830376,
|
| 316 |
+
"biou": 0.18446681119727068,
|
| 317 |
+
"hd95": 4.454683303833008
|
| 318 |
+
},
|
| 319 |
+
{
|
| 320 |
+
"sample_id": "BUSI_0161_M",
|
| 321 |
+
"dice": 0.8134579462497452,
|
| 322 |
+
"ppv": 0.797264294155733,
|
| 323 |
+
"sen": 0.830323077325793,
|
| 324 |
+
"iou": 0.6855702649533656,
|
| 325 |
+
"biou": 0.026859705194275788,
|
| 326 |
+
"hd95": 15.053184509277344
|
| 327 |
+
},
|
| 328 |
+
{
|
| 329 |
+
"sample_id": "BUSI_0055_M",
|
| 330 |
+
"dice": 0.8746321439298326,
|
| 331 |
+
"ppv": 0.8084875072374856,
|
| 332 |
+
"sen": 0.9525641086456271,
|
| 333 |
+
"iou": 0.7771966760254523,
|
| 334 |
+
"biou": 0.13471547424068692,
|
| 335 |
+
"hd95": 6.008275508880615
|
| 336 |
+
},
|
| 337 |
+
{
|
| 338 |
+
"sample_id": "BUSI_0059_M",
|
| 339 |
+
"dice": 0.8924312497878895,
|
| 340 |
+
"ppv": 0.8362599809290768,
|
| 341 |
+
"sen": 0.9566918873808534,
|
| 342 |
+
"iou": 0.8057569807568231,
|
| 343 |
+
"biou": 0.047984827642067636,
|
| 344 |
+
"hd95": 7.262857913970947
|
| 345 |
+
},
|
| 346 |
+
{
|
| 347 |
+
"sample_id": "BUSI_0285_B",
|
| 348 |
+
"dice": 0.9231541977320845,
|
| 349 |
+
"ppv": 0.8948393481169087,
|
| 350 |
+
"sen": 0.9533195069170636,
|
| 351 |
+
"iou": 0.8572761327167787,
|
| 352 |
+
"biou": 0.121053094182582,
|
| 353 |
+
"hd95": 3.605551242828369
|
| 354 |
+
},
|
| 355 |
+
{
|
| 356 |
+
"sample_id": "BUSI_0163_B",
|
| 357 |
+
"dice": 0.2118538194383582,
|
| 358 |
+
"ppv": 0.11847685211892073,
|
| 359 |
+
"sen": 1.0,
|
| 360 |
+
"iou": 0.11847685211892073,
|
| 361 |
+
"biou": 0.015625769042367937,
|
| 362 |
+
"hd95": 24.102724075317383
|
| 363 |
+
},
|
| 364 |
+
{
|
| 365 |
+
"sample_id": "BUSI_0224_B",
|
| 366 |
+
"dice": 2.2784232791427824e-08,
|
| 367 |
+
"ppv": 4.2844899621041144e-08,
|
| 368 |
+
"sen": 4.866179811864729e-08,
|
| 369 |
+
"iou": 2.2784232791427824e-08,
|
| 370 |
+
"biou": 2.932550459662622e-07,
|
| 371 |
+
"hd95": 59.36328887939453
|
| 372 |
+
},
|
| 373 |
+
{
|
| 374 |
+
"sample_id": "BUSI_0316_B",
|
| 375 |
+
"dice": 0.8419295330644976,
|
| 376 |
+
"ppv": 0.8553829699092489,
|
| 377 |
+
"sen": 0.8288927365243471,
|
| 378 |
+
"iou": 0.7270106262972591,
|
| 379 |
+
"biou": 0.029275958509097304,
|
| 380 |
+
"hd95": 16.0
|
| 381 |
+
},
|
| 382 |
+
{
|
| 383 |
+
"sample_id": "BUSI_0116_M",
|
| 384 |
+
"dice": 6.013228742439643e-08,
|
| 385 |
+
"ppv": 1.2820511176857542e-07,
|
| 386 |
+
"sen": 1.1325027030008265e-07,
|
| 387 |
+
"iou": 6.013228742439643e-08,
|
| 388 |
+
"biou": 4.7169789070854215e-07,
|
| 389 |
+
"hd95": 91.25376892089844
|
| 390 |
+
},
|
| 391 |
+
{
|
| 392 |
+
"sample_id": "BUSI_0214_B",
|
| 393 |
+
"dice": 0.9151094546124157,
|
| 394 |
+
"ppv": 0.8862461337904722,
|
| 395 |
+
"sen": 0.94591612075981,
|
| 396 |
+
"iou": 0.8435039524110283,
|
| 397 |
+
"biou": 0.14673959416326404,
|
| 398 |
+
"hd95": 3.3839144706726074
|
| 399 |
+
},
|
| 400 |
+
{
|
| 401 |
+
"sample_id": "BUSI_0207_M",
|
| 402 |
+
"dice": 5.274261325197188e-08,
|
| 403 |
+
"ppv": 1.2345663770785468e-06,
|
| 404 |
+
"sen": 5.5096415697167185e-08,
|
| 405 |
+
"iou": 5.274261325197188e-08,
|
| 406 |
+
"biou": 4.5248848303688553e-07,
|
| 407 |
+
"hd95": 61.188232421875
|
| 408 |
+
},
|
| 409 |
+
{
|
| 410 |
+
"sample_id": "BUSI_0016_M",
|
| 411 |
+
"dice": 0.6726003547539205,
|
| 412 |
+
"ppv": 0.7985909739498145,
|
| 413 |
+
"sen": 0.5809466511622957,
|
| 414 |
+
"iou": 0.5067052456822181,
|
| 415 |
+
"biou": 0.03125020182287462,
|
| 416 |
+
"hd95": 20.10287857055664
|
| 417 |
+
},
|
| 418 |
+
{
|
| 419 |
+
"sample_id": "BUSI_0067_M",
|
| 420 |
+
"dice": 0.4770642379522855,
|
| 421 |
+
"ppv": 0.31325304269285836,
|
| 422 |
+
"sen": 1.0,
|
| 423 |
+
"iou": 0.31325304269285836,
|
| 424 |
+
"biou": 2.8011196635519154e-07,
|
| 425 |
+
"hd95": 24.0
|
| 426 |
+
},
|
| 427 |
+
{
|
| 428 |
+
"sample_id": "BUSI_0184_B",
|
| 429 |
+
"dice": 0.9675491040232899,
|
| 430 |
+
"ppv": 0.970865468257692,
|
| 431 |
+
"sen": 0.9642553206699863,
|
| 432 |
+
"iou": 0.9371381332862642,
|
| 433 |
+
"biou": 0.28445768784232034,
|
| 434 |
+
"hd95": 2.0
|
| 435 |
+
},
|
| 436 |
+
{
|
| 437 |
+
"sample_id": "BUSI_0099_B",
|
| 438 |
+
"dice": 0.97163120768573,
|
| 439 |
+
"ppv": 0.9675141288821851,
|
| 440 |
+
"sen": 0.9757834792331226,
|
| 441 |
+
"iou": 0.9448275938168836,
|
| 442 |
+
"biou": 0.4354843262223176,
|
| 443 |
+
"hd95": 1.4142135381698608
|
| 444 |
+
},
|
| 445 |
+
{
|
| 446 |
+
"sample_id": "BUSI_0412_B",
|
| 447 |
+
"dice": 0.8218966861826721,
|
| 448 |
+
"ppv": 0.9404038435521693,
|
| 449 |
+
"sen": 0.7299147941278942,
|
| 450 |
+
"iou": 0.6976439834548577,
|
| 451 |
+
"biou": 0.07446824920775724,
|
| 452 |
+
"hd95": 30.236526489257812
|
| 453 |
+
},
|
| 454 |
+
{
|
| 455 |
+
"sample_id": "BUSI_0100_M",
|
| 456 |
+
"dice": 0.7011027451478383,
|
| 457 |
+
"ppv": 0.819538694770869,
|
| 458 |
+
"sen": 0.6125761042012065,
|
| 459 |
+
"iou": 0.5397676908161134,
|
| 460 |
+
"biou": 0.041841405087278204,
|
| 461 |
+
"hd95": 12.60145092010498
|
| 462 |
+
},
|
| 463 |
+
{
|
| 464 |
+
"sample_id": "BUSI_0073_M",
|
| 465 |
+
"dice": 0.8393113458169766,
|
| 466 |
+
"ppv": 0.7509628047544538,
|
| 467 |
+
"sen": 0.9512195201269072,
|
| 468 |
+
"iou": 0.7231149909623003,
|
| 469 |
+
"biou": 0.11578993905792681,
|
| 470 |
+
"hd95": 5.0
|
| 471 |
+
},
|
| 472 |
+
{
|
| 473 |
+
"sample_id": "BUSI_0357_B",
|
| 474 |
+
"dice": 0.8186596702588915,
|
| 475 |
+
"ppv": 0.6945373807650634,
|
| 476 |
+
"sen": 0.9968000005119999,
|
| 477 |
+
"iou": 0.6929922477205509,
|
| 478 |
+
"biou": 0.14450916502360403,
|
| 479 |
+
"hd95": 10.029705047607422
|
| 480 |
+
},
|
| 481 |
+
{
|
| 482 |
+
"sample_id": "BUSI_0229_B",
|
| 483 |
+
"dice": 0.9578059093984225,
|
| 484 |
+
"ppv": 0.9487983334588993,
|
| 485 |
+
"sen": 0.9669861590004091,
|
| 486 |
+
"iou": 0.9190283482764829,
|
| 487 |
+
"biou": 0.18750046164746498,
|
| 488 |
+
"hd95": 1.2485243082046509
|
| 489 |
+
},
|
| 490 |
+
{
|
| 491 |
+
"sample_id": "BUSI_0129_M",
|
| 492 |
+
"dice": 0.851133807762387,
|
| 493 |
+
"ppv": 0.9746283453988435,
|
| 494 |
+
"sen": 0.7554155821875007,
|
| 495 |
+
"iou": 0.7408467720227999,
|
| 496 |
+
"biou": 0.021775708245275003,
|
| 497 |
+
"hd95": 13.0
|
| 498 |
+
},
|
| 499 |
+
{
|
| 500 |
+
"sample_id": "BUSI_0156_B",
|
| 501 |
+
"dice": 0.9533333411111098,
|
| 502 |
+
"ppv": 0.9597315571370614,
|
| 503 |
+
"sen": 0.9470198850927533,
|
| 504 |
+
"iou": 0.910828053876416,
|
| 505 |
+
"biou": 0.3625007968740039,
|
| 506 |
+
"hd95": 1.0
|
| 507 |
+
},
|
| 508 |
+
{
|
| 509 |
+
"sample_id": "BUSI_0320_B",
|
| 510 |
+
"dice": 0.9270946760506337,
|
| 511 |
+
"ppv": 0.8729508457067939,
|
| 512 |
+
"sen": 0.9883990746173841,
|
| 513 |
+
"iou": 0.8640973906496165,
|
| 514 |
+
"biou": 0.22222288698898546,
|
| 515 |
+
"hd95": 2.0
|
| 516 |
+
},
|
| 517 |
+
{
|
| 518 |
+
"sample_id": "BUSI_0104_B",
|
| 519 |
+
"dice": 0.9039146249414147,
|
| 520 |
+
"ppv": 0.9270073525493777,
|
| 521 |
+
"sen": 0.8819445264274122,
|
| 522 |
+
"iou": 0.8246754385224425,
|
| 523 |
+
"biou": 0.19354968782308415,
|
| 524 |
+
"hd95": 1.0
|
| 525 |
+
},
|
| 526 |
+
{
|
| 527 |
+
"sample_id": "BUSI_0340_B",
|
| 528 |
+
"dice": 0.8899498116054619,
|
| 529 |
+
"ppv": 0.9242308689491261,
|
| 530 |
+
"sen": 0.8581208720934639,
|
| 531 |
+
"iou": 0.8017203396179898,
|
| 532 |
+
"biou": 0.07674440075711611,
|
| 533 |
+
"hd95": 13.0
|
| 534 |
+
},
|
| 535 |
+
{
|
| 536 |
+
"sample_id": "BUSI_0121_B",
|
| 537 |
+
"dice": 0.9532836863044221,
|
| 538 |
+
"ppv": 0.9399199012122962,
|
| 539 |
+
"sen": 0.967032971561405,
|
| 540 |
+
"iou": 0.9107373983522123,
|
| 541 |
+
"biou": 0.3787883494027656,
|
| 542 |
+
"hd95": 2.2360680103302
|
| 543 |
+
},
|
| 544 |
+
{
|
| 545 |
+
"sample_id": "BUSI_0358_B",
|
| 546 |
+
"dice": 0.860164520554376,
|
| 547 |
+
"ppv": 0.7546392005526598,
|
| 548 |
+
"sen": 1.0,
|
| 549 |
+
"iou": 0.7546392005526598,
|
| 550 |
+
"biou": 0.08421100830999563,
|
| 551 |
+
"hd95": 5.0
|
| 552 |
+
},
|
| 553 |
+
{
|
| 554 |
+
"sample_id": "BUSI_0270_B",
|
| 555 |
+
"dice": 0.8934010287562146,
|
| 556 |
+
"ppv": 0.8380952766439818,
|
| 557 |
+
"sen": 0.9565217509451763,
|
| 558 |
+
"iou": 0.8073394937294739,
|
| 559 |
+
"biou": 0.17857216198914108,
|
| 560 |
+
"hd95": 3.0
|
| 561 |
+
},
|
| 562 |
+
{
|
| 563 |
+
"sample_id": "BUSI_0023_M",
|
| 564 |
+
"dice": 0.744337981877853,
|
| 565 |
+
"ppv": 0.8973483880979682,
|
| 566 |
+
"sen": 0.6359069835180097,
|
| 567 |
+
"iou": 0.592785300159811,
|
| 568 |
+
"biou": 0.03243260676889968,
|
| 569 |
+
"hd95": 31.15386962890625
|
| 570 |
+
},
|
| 571 |
+
{
|
| 572 |
+
"sample_id": "BUSI_0020_B",
|
| 573 |
+
"dice": 0.9017857205493801,
|
| 574 |
+
"ppv": 0.8240093445210555,
|
| 575 |
+
"sen": 0.9957746484824439,
|
| 576 |
+
"iou": 0.8211382321558383,
|
| 577 |
+
"biou": 0.07070754004669694,
|
| 578 |
+
"hd95": 2.0
|
| 579 |
+
},
|
| 580 |
+
{
|
| 581 |
+
"sample_id": "BUSI_0153_M",
|
| 582 |
+
"dice": 0.8708170476611969,
|
| 583 |
+
"ppv": 0.9686108172086784,
|
| 584 |
+
"sen": 0.7909594771222813,
|
| 585 |
+
"iou": 0.771192168672878,
|
| 586 |
+
"biou": 0.01818199669418242,
|
| 587 |
+
"hd95": 11.0
|
| 588 |
+
},
|
| 589 |
+
{
|
| 590 |
+
"sample_id": "BUSI_0154_B",
|
| 591 |
+
"dice": 0.9130435097668443,
|
| 592 |
+
"ppv": 0.8750000868054952,
|
| 593 |
+
"sen": 0.9545454889806901,
|
| 594 |
+
"iou": 0.8400001066665955,
|
| 595 |
+
"biou": 0.28070301631049766,
|
| 596 |
+
"hd95": 1.165686011314392
|
| 597 |
+
},
|
| 598 |
+
{
|
| 599 |
+
"sample_id": "BUSI_0062_M",
|
| 600 |
+
"dice": 0.810285224496862,
|
| 601 |
+
"ppv": 0.7965165761887605,
|
| 602 |
+
"sen": 0.8245382662911931,
|
| 603 |
+
"iou": 0.681075202285681,
|
| 604 |
+
"biou": 0.03381665849839167,
|
| 605 |
+
"hd95": 10.93582820892334
|
| 606 |
+
},
|
| 607 |
+
{
|
| 608 |
+
"sample_id": "BUSI_0075_B",
|
| 609 |
+
"dice": 0.9297012364058668,
|
| 610 |
+
"ppv": 0.9635701341402306,
|
| 611 |
+
"sen": 0.8981324451388039,
|
| 612 |
+
"iou": 0.8686371315866779,
|
| 613 |
+
"biou": 0.2061074762538349,
|
| 614 |
+
"hd95": 2.2360680103302
|
| 615 |
+
}
|
| 616 |
+
],
|
| 617 |
+
"checkpoint_mode": "latest",
|
| 618 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/checkpoints/latest.pt"
|
| 619 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/evaluation_results.xlsx
ADDED
|
Binary file (35.1 kB). View file
|
|
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/evaluation_summary.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"mean_iou": 0.6688458919525146,
|
| 3 |
+
"mean_dice": 0.7611992359161377
|
| 4 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/history.json
ADDED
|
@@ -0,0 +1,2242 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"epoch": 1,
|
| 4 |
+
"train_loss": 1.0584740731269555,
|
| 5 |
+
"train_actor_loss": -0.3708620821691338,
|
| 6 |
+
"train_critic_loss": 0.1442930435921373,
|
| 7 |
+
"train_mean_reward": 0.00036421462253728194,
|
| 8 |
+
"train_entropy": 0.6123965491061832,
|
| 9 |
+
"train_ce_loss": 1.1726361124725766,
|
| 10 |
+
"train_dice_loss": 1.5417431245632565,
|
| 11 |
+
"train_dice": 0.32734915183324587,
|
| 12 |
+
"train_iou": 0.22729327462042734,
|
| 13 |
+
"grad_norm": 12.769804340924404,
|
| 14 |
+
"lr": 0.001999509027563903,
|
| 15 |
+
"encoder_lr": 0.0001999531232347445,
|
| 16 |
+
"alpha": 0.49042823910713196,
|
| 17 |
+
"validated_this_epoch": false,
|
| 18 |
+
"val_loss": null,
|
| 19 |
+
"val_dice": null,
|
| 20 |
+
"val_iou": null,
|
| 21 |
+
"val_reward": null,
|
| 22 |
+
"val_entropy": null
|
| 23 |
+
},
|
| 24 |
+
{
|
| 25 |
+
"epoch": 2,
|
| 26 |
+
"train_loss": 0.6699016441973721,
|
| 27 |
+
"train_actor_loss": -0.37100063641471376,
|
| 28 |
+
"train_critic_loss": 0.0006068380439207434,
|
| 29 |
+
"train_mean_reward": -0.0060538716521755225,
|
| 30 |
+
"train_entropy": 0.6410567822844483,
|
| 31 |
+
"train_ce_loss": 0.694799918147136,
|
| 32 |
+
"train_dice_loss": 1.3863977829923309,
|
| 33 |
+
"train_dice": 0.44094900537170323,
|
| 34 |
+
"train_iou": 0.32866166023422005,
|
| 35 |
+
"grad_norm": 7.105209102002225,
|
| 36 |
+
"lr": 0.0019980365947861304,
|
| 37 |
+
"encoder_lr": 0.00019981253920068581,
|
| 38 |
+
"alpha": 0.4801827669143677,
|
| 39 |
+
"validated_this_epoch": false,
|
| 40 |
+
"val_loss": null,
|
| 41 |
+
"val_dice": null,
|
| 42 |
+
"val_iou": null,
|
| 43 |
+
"val_reward": null,
|
| 44 |
+
"val_entropy": null
|
| 45 |
+
},
|
| 46 |
+
{
|
| 47 |
+
"epoch": 3,
|
| 48 |
+
"train_loss": 0.6044829953245299,
|
| 49 |
+
"train_actor_loss": -0.3912744443257188,
|
| 50 |
+
"train_critic_loss": 0.0005253357889423028,
|
| 51 |
+
"train_mean_reward": -0.006731109448294372,
|
| 52 |
+
"train_entropy": 0.6847654770703586,
|
| 53 |
+
"train_ce_loss": 0.6383650734929189,
|
| 54 |
+
"train_dice_loss": 1.3526244525453532,
|
| 55 |
+
"train_dice": 0.49477523933083556,
|
| 56 |
+
"train_iou": 0.3718987799493475,
|
| 57 |
+
"grad_norm": 5.6952283890672435,
|
| 58 |
+
"lr": 0.001995584154780065,
|
| 59 |
+
"encoder_lr": 0.0001995783866372926,
|
| 60 |
+
"alpha": 0.46894803643226624,
|
| 61 |
+
"validated_this_epoch": false,
|
| 62 |
+
"val_loss": null,
|
| 63 |
+
"val_dice": null,
|
| 64 |
+
"val_iou": null,
|
| 65 |
+
"val_reward": null,
|
| 66 |
+
"val_entropy": null
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"epoch": 4,
|
| 70 |
+
"train_loss": 0.5724230632402523,
|
| 71 |
+
"train_actor_loss": -0.39071922276640736,
|
| 72 |
+
"train_critic_loss": 0.00047149525420940743,
|
| 73 |
+
"train_mean_reward": -0.0067716842457545,
|
| 74 |
+
"train_entropy": 0.7006937515096454,
|
| 75 |
+
"train_ce_loss": 0.6100725016246239,
|
| 76 |
+
"train_dice_loss": 1.3157405616085045,
|
| 77 |
+
"train_dice": 0.527922768882871,
|
| 78 |
+
"train_iou": 0.4040958187929779,
|
| 79 |
+
"grad_norm": 4.745243258254472,
|
| 80 |
+
"lr": 0.001992154127807906,
|
| 81 |
+
"encoder_lr": 0.0001992508966248754,
|
| 82 |
+
"alpha": 0.4576987624168396,
|
| 83 |
+
"validated_this_epoch": false,
|
| 84 |
+
"val_loss": null,
|
| 85 |
+
"val_dice": null,
|
| 86 |
+
"val_iou": null,
|
| 87 |
+
"val_reward": null,
|
| 88 |
+
"val_entropy": null
|
| 89 |
+
},
|
| 90 |
+
{
|
| 91 |
+
"epoch": 5,
|
| 92 |
+
"train_loss": 0.5469325371844651,
|
| 93 |
+
"train_actor_loss": -0.3785795701104541,
|
| 94 |
+
"train_critic_loss": 0.0004715523620961669,
|
| 95 |
+
"train_mean_reward": -0.00690545157892046,
|
| 96 |
+
"train_entropy": 0.6709512266154745,
|
| 97 |
+
"train_ce_loss": 0.5703006941170579,
|
| 98 |
+
"train_dice_loss": 1.280251953490945,
|
| 99 |
+
"train_dice": 0.5307233496164627,
|
| 100 |
+
"train_iou": 0.4092315009579271,
|
| 101 |
+
"grad_norm": 4.396956750126773,
|
| 102 |
+
"lr": 0.0019877498988921626,
|
| 103 |
+
"encoder_lr": 0.0001988303923565381,
|
| 104 |
+
"alpha": 0.44726115465164185,
|
| 105 |
+
"validated_this_epoch": false,
|
| 106 |
+
"val_loss": null,
|
| 107 |
+
"val_dice": null,
|
| 108 |
+
"val_iou": null,
|
| 109 |
+
"val_reward": null,
|
| 110 |
+
"val_entropy": null
|
| 111 |
+
},
|
| 112 |
+
{
|
| 113 |
+
"epoch": 6,
|
| 114 |
+
"train_loss": 0.5355125919865199,
|
| 115 |
+
"train_actor_loss": -0.34525162971624257,
|
| 116 |
+
"train_critic_loss": 0.00042556139174209566,
|
| 117 |
+
"train_mean_reward": -0.006230154039269947,
|
| 118 |
+
"train_entropy": 0.597773207693291,
|
| 119 |
+
"train_ce_loss": 0.5069862835973173,
|
| 120 |
+
"train_dice_loss": 1.2541165843182447,
|
| 121 |
+
"train_dice": 0.5814218008870045,
|
| 122 |
+
"train_iou": 0.458288984026859,
|
| 123 |
+
"grad_norm": 4.42461252674576,
|
| 124 |
+
"lr": 0.0019823758144750458,
|
| 125 |
+
"encoder_lr": 0.00019831728881922545,
|
| 126 |
+
"alpha": 0.43896469473838806,
|
| 127 |
+
"validated_this_epoch": false,
|
| 128 |
+
"val_loss": null,
|
| 129 |
+
"val_dice": null,
|
| 130 |
+
"val_iou": null,
|
| 131 |
+
"val_reward": null,
|
| 132 |
+
"val_entropy": null
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"epoch": 7,
|
| 136 |
+
"train_loss": 0.5283558962888323,
|
| 137 |
+
"train_actor_loss": -0.35748377442359924,
|
| 138 |
+
"train_critic_loss": 0.00041388250464260355,
|
| 139 |
+
"train_mean_reward": -0.006353063866781409,
|
| 140 |
+
"train_entropy": 0.6386820398444353,
|
| 141 |
+
"train_ce_loss": 0.5225535242900117,
|
| 142 |
+
"train_dice_loss": 1.2487119171354508,
|
| 143 |
+
"train_dice": 0.585187339823131,
|
| 144 |
+
"train_iou": 0.4587046930903811,
|
| 145 |
+
"grad_norm": 3.9337960190551224,
|
| 146 |
+
"lr": 0.001976037178129054,
|
| 147 |
+
"encoder_lr": 0.00019771209238418101,
|
| 148 |
+
"alpha": 0.42962440848350525,
|
| 149 |
+
"validated_this_epoch": false,
|
| 150 |
+
"val_loss": null,
|
| 151 |
+
"val_dice": null,
|
| 152 |
+
"val_iou": null,
|
| 153 |
+
"val_reward": null,
|
| 154 |
+
"val_entropy": null
|
| 155 |
+
},
|
| 156 |
+
{
|
| 157 |
+
"epoch": 8,
|
| 158 |
+
"train_loss": 0.511005380270324,
|
| 159 |
+
"train_actor_loss": -0.3669324039989355,
|
| 160 |
+
"train_critic_loss": 0.0003826832748234416,
|
| 161 |
+
"train_mean_reward": -0.006282203196285013,
|
| 162 |
+
"train_entropy": 0.6669621354772756,
|
| 163 |
+
"train_ce_loss": 0.5081055811209272,
|
| 164 |
+
"train_dice_loss": 1.2473872911560444,
|
| 165 |
+
"train_dice": 0.5991991453610351,
|
| 166 |
+
"train_iou": 0.474563320711657,
|
| 167 |
+
"grad_norm": 4.075359844884207,
|
| 168 |
+
"lr": 0.001968740245322988,
|
| 169 |
+
"encoder_lr": 0.00019701540030721995,
|
| 170 |
+
"alpha": 0.4195512533187866,
|
| 171 |
+
"validated_this_epoch": false,
|
| 172 |
+
"val_loss": null,
|
| 173 |
+
"val_dice": null,
|
| 174 |
+
"val_iou": null,
|
| 175 |
+
"val_reward": null,
|
| 176 |
+
"val_entropy": null
|
| 177 |
+
},
|
| 178 |
+
{
|
| 179 |
+
"epoch": 9,
|
| 180 |
+
"train_loss": 0.5260217207693314,
|
| 181 |
+
"train_actor_loss": -0.33067883112171825,
|
| 182 |
+
"train_critic_loss": 0.00036695755093193907,
|
| 183 |
+
"train_mean_reward": -0.005957611770614577,
|
| 184 |
+
"train_entropy": 0.6033429998991101,
|
| 185 |
+
"train_ce_loss": 0.48840847373783625,
|
| 186 |
+
"train_dice_loss": 1.2246256587733286,
|
| 187 |
+
"train_dice": 0.6230224750580919,
|
| 188 |
+
"train_iou": 0.498122586893859,
|
| 189 |
+
"grad_norm": 4.094718871190566,
|
| 190 |
+
"lr": 0.0019604922172485584,
|
| 191 |
+
"encoder_lr": 0.0001962279001393096,
|
| 192 |
+
"alpha": 0.41127151250839233,
|
| 193 |
+
"validated_this_epoch": false,
|
| 194 |
+
"val_loss": null,
|
| 195 |
+
"val_dice": null,
|
| 196 |
+
"val_iou": null,
|
| 197 |
+
"val_reward": null,
|
| 198 |
+
"val_entropy": null
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"epoch": 10,
|
| 202 |
+
"train_loss": 0.5189683107774734,
|
| 203 |
+
"train_actor_loss": -0.313856204302659,
|
| 204 |
+
"train_critic_loss": 0.0003426296938677665,
|
| 205 |
+
"train_mean_reward": -0.00604343147621643,
|
| 206 |
+
"train_entropy": 0.5891070610022977,
|
| 207 |
+
"train_ce_loss": 0.4610433480449656,
|
| 208 |
+
"train_dice_loss": 1.2042630359491944,
|
| 209 |
+
"train_dice": 0.6337237518959358,
|
| 210 |
+
"train_iou": 0.5118181521282327,
|
| 211 |
+
"grad_norm": 4.0946759909622426,
|
| 212 |
+
"lr": 0.0019513012337136779,
|
| 213 |
+
"encoder_lr": 0.00019535036904803958,
|
| 214 |
+
"alpha": 0.4030348062515259,
|
| 215 |
+
"validated_this_epoch": true,
|
| 216 |
+
"val_loss": 0.8658681987629583,
|
| 217 |
+
"val_dice": 0.6483120921833402,
|
| 218 |
+
"val_iou": 0.5412297563748109,
|
| 219 |
+
"val_reward": 0.0010978090273134383,
|
| 220 |
+
"val_entropy": 0.5185394353336759,
|
| 221 |
+
"val_actor_loss": -0.013089635326512946,
|
| 222 |
+
"val_critic_loss": 0.00012033747017737526,
|
| 223 |
+
"val_ce_loss": 0.44879946107458735,
|
| 224 |
+
"val_dice_loss": 1.3089958450408898
|
| 225 |
+
},
|
| 226 |
+
{
|
| 227 |
+
"epoch": 11,
|
| 228 |
+
"train_loss": 0.5169719918378333,
|
| 229 |
+
"train_actor_loss": -0.30092242963669835,
|
| 230 |
+
"train_critic_loss": 0.00032993409273490866,
|
| 231 |
+
"train_mean_reward": -0.005009823866250081,
|
| 232 |
+
"train_entropy": 0.5672867706656765,
|
| 233 |
+
"train_ce_loss": 0.4728587061607157,
|
| 234 |
+
"train_dice_loss": 1.162600189447403,
|
| 235 |
+
"train_dice": 0.6484574124733716,
|
| 236 |
+
"train_iou": 0.529883171955006,
|
| 237 |
+
"grad_norm": 4.911387224530065,
|
| 238 |
+
"lr": 0.0019411763651094546,
|
| 239 |
+
"encoder_lr": 0.0001943836730506514,
|
| 240 |
+
"alpha": 0.39621788263320923,
|
| 241 |
+
"validated_this_epoch": false,
|
| 242 |
+
"val_loss": null,
|
| 243 |
+
"val_dice": null,
|
| 244 |
+
"val_iou": null,
|
| 245 |
+
"val_reward": null,
|
| 246 |
+
"val_entropy": null
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"epoch": 12,
|
| 250 |
+
"train_loss": 0.5156925811518936,
|
| 251 |
+
"train_actor_loss": -0.21270607940332834,
|
| 252 |
+
"train_critic_loss": 0.00021040216463655433,
|
| 253 |
+
"train_mean_reward": -0.0015707019343298191,
|
| 254 |
+
"train_entropy": 0.47500704094112994,
|
| 255 |
+
"train_ce_loss": 0.549091537220726,
|
| 256 |
+
"train_dice_loss": 0.9074953658174175,
|
| 257 |
+
"train_dice": 0.6966743995530615,
|
| 258 |
+
"train_iou": 0.5857623692100483,
|
| 259 |
+
"grad_norm": 8.718096688274265,
|
| 260 |
+
"lr": 0.0019301276034588103,
|
| 261 |
+
"encoder_lr": 0.00019332876615938386,
|
| 262 |
+
"alpha": 0.39151766896247864,
|
| 263 |
+
"validated_this_epoch": false,
|
| 264 |
+
"val_loss": null,
|
| 265 |
+
"val_dice": null,
|
| 266 |
+
"val_iou": null,
|
| 267 |
+
"val_reward": null,
|
| 268 |
+
"val_entropy": null
|
| 269 |
+
},
|
| 270 |
+
{
|
| 271 |
+
"epoch": 13,
|
| 272 |
+
"train_loss": 0.5113261588555833,
|
| 273 |
+
"train_actor_loss": -0.18855736370966109,
|
| 274 |
+
"train_critic_loss": 0.00020809433632396787,
|
| 275 |
+
"train_mean_reward": -0.0005119910320178603,
|
| 276 |
+
"train_entropy": 0.460748731647197,
|
| 277 |
+
"train_ce_loss": 0.5882928621827671,
|
| 278 |
+
"train_dice_loss": 0.811266074808993,
|
| 279 |
+
"train_dice": 0.6974849103066867,
|
| 280 |
+
"train_iou": 0.5952277725331597,
|
| 281 |
+
"grad_norm": 10.336135561152021,
|
| 282 |
+
"lr": 0.0019181658525555612,
|
| 283 |
+
"encoder_lr": 0.00019218668943997818,
|
| 284 |
+
"alpha": 0.38730087876319885,
|
| 285 |
+
"validated_this_epoch": false,
|
| 286 |
+
"val_loss": null,
|
| 287 |
+
"val_dice": null,
|
| 288 |
+
"val_iou": null,
|
| 289 |
+
"val_reward": null,
|
| 290 |
+
"val_entropy": null
|
| 291 |
+
},
|
| 292 |
+
{
|
| 293 |
+
"epoch": 14,
|
| 294 |
+
"train_loss": 0.473906206902001,
|
| 295 |
+
"train_actor_loss": -0.18250627628014993,
|
| 296 |
+
"train_critic_loss": 0.0001614137087024348,
|
| 297 |
+
"train_mean_reward": -0.0002394148408716472,
|
| 298 |
+
"train_entropy": 0.47662893222591973,
|
| 299 |
+
"train_ce_loss": 0.6125215898563926,
|
| 300 |
+
"train_dice_loss": 0.7001419497090716,
|
| 301 |
+
"train_dice": 0.7135963475255609,
|
| 302 |
+
"train_iou": 0.6107922743086078,
|
| 303 |
+
"grad_norm": 16.26865666504054,
|
| 304 |
+
"lr": 0.0019053029172036895,
|
| 305 |
+
"encoder_lr": 0.00019095856998427185,
|
| 306 |
+
"alpha": 0.38257333636283875,
|
| 307 |
+
"validated_this_epoch": false,
|
| 308 |
+
"val_loss": null,
|
| 309 |
+
"val_dice": null,
|
| 310 |
+
"val_iou": null,
|
| 311 |
+
"val_reward": null,
|
| 312 |
+
"val_entropy": null
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"epoch": 15,
|
| 316 |
+
"train_loss": 0.4633492227795339,
|
| 317 |
+
"train_actor_loss": -0.1753022367235764,
|
| 318 |
+
"train_critic_loss": 0.00017698028348848977,
|
| 319 |
+
"train_mean_reward": 0.00013446423788295893,
|
| 320 |
+
"train_entropy": 0.4722871264052946,
|
| 321 |
+
"train_ce_loss": 0.5853927693418792,
|
| 322 |
+
"train_dice_loss": 0.6917331533684595,
|
| 323 |
+
"train_dice": 0.7295051299633056,
|
| 324 |
+
"train_iou": 0.6284848608078466,
|
| 325 |
+
"grad_norm": 21.209616538166074,
|
| 326 |
+
"lr": 0.001891551491567426,
|
| 327 |
+
"encoder_lr": 0.00018964561979789495,
|
| 328 |
+
"alpha": 0.3774704337120056,
|
| 329 |
+
"validated_this_epoch": false,
|
| 330 |
+
"val_loss": null,
|
| 331 |
+
"val_dice": null,
|
| 332 |
+
"val_iou": null,
|
| 333 |
+
"val_reward": null,
|
| 334 |
+
"val_entropy": null
|
| 335 |
+
},
|
| 336 |
+
{
|
| 337 |
+
"epoch": 16,
|
| 338 |
+
"train_loss": 0.45069999386404835,
|
| 339 |
+
"train_actor_loss": -0.1719807679415981,
|
| 340 |
+
"train_critic_loss": 0.00017143654359273857,
|
| 341 |
+
"train_mean_reward": 0.00012184302997848697,
|
| 342 |
+
"train_entropy": 0.4701696360527083,
|
| 343 |
+
"train_ce_loss": 0.5797492340844825,
|
| 344 |
+
"train_dice_loss": 0.6654408376192245,
|
| 345 |
+
"train_dice": 0.7421282550434533,
|
| 346 |
+
"train_iou": 0.6424375749275147,
|
| 347 |
+
"grad_norm": 25.159671474796856,
|
| 348 |
+
"lr": 0.0018769251466436443,
|
| 349 |
+
"encoder_lr": 0.00018824913460416704,
|
| 350 |
+
"alpha": 0.37229958176612854,
|
| 351 |
+
"validated_this_epoch": false,
|
| 352 |
+
"val_loss": null,
|
| 353 |
+
"val_dice": null,
|
| 354 |
+
"val_iou": null,
|
| 355 |
+
"val_reward": null,
|
| 356 |
+
"val_entropy": null
|
| 357 |
+
},
|
| 358 |
+
{
|
| 359 |
+
"epoch": 17,
|
| 360 |
+
"train_loss": 0.42364747036436157,
|
| 361 |
+
"train_actor_loss": -0.172354863739984,
|
| 362 |
+
"train_critic_loss": 0.00015902851578351873,
|
| 363 |
+
"train_mean_reward": 3.5833351123276044e-05,
|
| 364 |
+
"train_entropy": 0.47168740861175595,
|
| 365 |
+
"train_ce_loss": 0.5588817212668076,
|
| 366 |
+
"train_dice_loss": 0.6329639064249142,
|
| 367 |
+
"train_dice": 0.7427122128899768,
|
| 368 |
+
"train_iou": 0.6444893287770052,
|
| 369 |
+
"grad_norm": 20.865994857263196,
|
| 370 |
+
"lr": 0.001861438316868924,
|
| 371 |
+
"encoder_lr": 0.00018677049256537467,
|
| 372 |
+
"alpha": 0.3669697642326355,
|
| 373 |
+
"validated_this_epoch": false,
|
| 374 |
+
"val_loss": null,
|
| 375 |
+
"val_dice": null,
|
| 376 |
+
"val_iou": null,
|
| 377 |
+
"val_reward": null,
|
| 378 |
+
"val_entropy": null
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"epoch": 18,
|
| 382 |
+
"train_loss": 0.4278344268325803,
|
| 383 |
+
"train_actor_loss": -0.16659095633091547,
|
| 384 |
+
"train_critic_loss": 0.00016343044311142125,
|
| 385 |
+
"train_mean_reward": -0.0001825981043069522,
|
| 386 |
+
"train_entropy": 0.4700656973345335,
|
| 387 |
+
"train_ce_loss": 0.5608913859867465,
|
| 388 |
+
"train_dice_loss": 0.6277959380655017,
|
| 389 |
+
"train_dice": 0.739135205211056,
|
| 390 |
+
"train_iou": 0.6385081814756682,
|
| 391 |
+
"grad_norm": 22.34932634488557,
|
| 392 |
+
"lr": 0.0018451062858745053,
|
| 393 |
+
"encoder_lr": 0.00018521115292269145,
|
| 394 |
+
"alpha": 0.36155298352241516,
|
| 395 |
+
"validated_this_epoch": false,
|
| 396 |
+
"val_loss": null,
|
| 397 |
+
"val_dice": null,
|
| 398 |
+
"val_iou": null,
|
| 399 |
+
"val_reward": null,
|
| 400 |
+
"val_entropy": null
|
| 401 |
+
},
|
| 402 |
+
{
|
| 403 |
+
"epoch": 19,
|
| 404 |
+
"train_loss": 0.41978721616269893,
|
| 405 |
+
"train_actor_loss": -0.1580277794408937,
|
| 406 |
+
"train_critic_loss": 0.0001490891467237205,
|
| 407 |
+
"train_mean_reward": 0.00015340800874190315,
|
| 408 |
+
"train_entropy": 0.45316800831285253,
|
| 409 |
+
"train_ce_loss": 0.5566257887257585,
|
| 410 |
+
"train_dice_loss": 0.5988550999367884,
|
| 411 |
+
"train_dice": 0.7589593601497644,
|
| 412 |
+
"train_iou": 0.6631517153801573,
|
| 413 |
+
"grad_norm": 23.58096486884494,
|
| 414 |
+
"lr": 0.0018279451714031893,
|
| 415 |
+
"encoder_lr": 0.0001835726545560834,
|
| 416 |
+
"alpha": 0.3565765619277954,
|
| 417 |
+
"validated_this_epoch": false,
|
| 418 |
+
"val_loss": null,
|
| 419 |
+
"val_dice": null,
|
| 420 |
+
"val_iou": null,
|
| 421 |
+
"val_reward": null,
|
| 422 |
+
"val_entropy": null
|
| 423 |
+
},
|
| 424 |
+
{
|
| 425 |
+
"epoch": 20,
|
| 426 |
+
"train_loss": 0.4144181594209269,
|
| 427 |
+
"train_actor_loss": -0.15806331774292065,
|
| 428 |
+
"train_critic_loss": 0.00014961237979296314,
|
| 429 |
+
"train_mean_reward": 0.0001406732154024936,
|
| 430 |
+
"train_entropy": 0.4587987050106051,
|
| 431 |
+
"train_ce_loss": 0.5281236314332685,
|
| 432 |
+
"train_dice_loss": 0.6166896969465014,
|
| 433 |
+
"train_dice": 0.747118808771198,
|
| 434 |
+
"train_iou": 0.6504663208258449,
|
| 435 |
+
"grad_norm": 29.991744668908822,
|
| 436 |
+
"lr": 0.001809971909403073,
|
| 437 |
+
"encoder_lr": 0.00018185661446562003,
|
| 438 |
+
"alpha": 0.35118064284324646,
|
| 439 |
+
"validated_this_epoch": true,
|
| 440 |
+
"val_loss": 0.9268459208436649,
|
| 441 |
+
"val_dice": 0.6792677168026484,
|
| 442 |
+
"val_iou": 0.5805517275109805,
|
| 443 |
+
"val_reward": 0.00037374342522253096,
|
| 444 |
+
"val_entropy": 0.758493410517471,
|
| 445 |
+
"val_actor_loss": -0.004558008116542955,
|
| 446 |
+
"val_critic_loss": 3.888744741485917e-05,
|
| 447 |
+
"val_ce_loss": 0.9429843647715946,
|
| 448 |
+
"val_dice_loss": 0.9197845705831893
|
| 449 |
+
},
|
| 450 |
+
{
|
| 451 |
+
"epoch": 21,
|
| 452 |
+
"train_loss": 0.4077782355963664,
|
| 453 |
+
"train_actor_loss": -0.1490337310638038,
|
| 454 |
+
"train_critic_loss": 0.00013294521661247,
|
| 455 |
+
"train_mean_reward": 0.00012037717441288634,
|
| 456 |
+
"train_entropy": 0.4389037002619231,
|
| 457 |
+
"train_ce_loss": 0.5254090449674109,
|
| 458 |
+
"train_dice_loss": 0.5880819268004839,
|
| 459 |
+
"train_dice": 0.7646179686779582,
|
| 460 |
+
"train_iou": 0.6694972253854804,
|
| 461 |
+
"grad_norm": 34.367439948773196,
|
| 462 |
+
"lr": 0.001791204237313812,
|
| 463 |
+
"encoder_lr": 0.0001800647261756906,
|
| 464 |
+
"alpha": 0.34665411710739136,
|
| 465 |
+
"validated_this_epoch": false,
|
| 466 |
+
"val_loss": null,
|
| 467 |
+
"val_dice": null,
|
| 468 |
+
"val_iou": null,
|
| 469 |
+
"val_reward": null,
|
| 470 |
+
"val_entropy": null
|
| 471 |
+
},
|
| 472 |
+
{
|
| 473 |
+
"epoch": 22,
|
| 474 |
+
"train_loss": 0.433384067544988,
|
| 475 |
+
"train_actor_loss": -0.14995086329774956,
|
| 476 |
+
"train_critic_loss": 0.00015422982842116327,
|
| 477 |
+
"train_mean_reward": 0.0001497393264947087,
|
| 478 |
+
"train_entropy": 0.44661826728927384,
|
| 479 |
+
"train_ce_loss": 0.5397826065164955,
|
| 480 |
+
"train_dice_loss": 0.6267330112845398,
|
| 481 |
+
"train_dice": 0.7525465459856076,
|
| 482 |
+
"train_iou": 0.657785617767998,
|
| 483 |
+
"grad_norm": 30.673339942166972,
|
| 484 |
+
"lr": 0.0017716606765619107,
|
| 485 |
+
"encoder_lr": 0.0001781987580637,
|
| 486 |
+
"alpha": 0.3416270911693573,
|
| 487 |
+
"validated_this_epoch": false,
|
| 488 |
+
"val_loss": null,
|
| 489 |
+
"val_dice": null,
|
| 490 |
+
"val_iou": null,
|
| 491 |
+
"val_reward": null,
|
| 492 |
+
"val_entropy": null
|
| 493 |
+
},
|
| 494 |
+
{
|
| 495 |
+
"epoch": 23,
|
| 496 |
+
"train_loss": 0.40507646549363124,
|
| 497 |
+
"train_actor_loss": -0.14455326414578026,
|
| 498 |
+
"train_critic_loss": 0.00014203655907649753,
|
| 499 |
+
"train_mean_reward": 0.00032320420844938807,
|
| 500 |
+
"train_entropy": 0.44618050338223925,
|
| 501 |
+
"train_ce_loss": 0.5112710448610055,
|
| 502 |
+
"train_dice_loss": 0.5878463627756104,
|
| 503 |
+
"train_dice": 0.7673078520605532,
|
| 504 |
+
"train_iou": 0.6719448287667493,
|
| 505 |
+
"grad_norm": 26.01113082965215,
|
| 506 |
+
"lr": 0.0017513605142823077,
|
| 507 |
+
"encoder_lr": 0.00017626055161489368,
|
| 508 |
+
"alpha": 0.3365047574043274,
|
| 509 |
+
"validated_this_epoch": false,
|
| 510 |
+
"val_loss": null,
|
| 511 |
+
"val_dice": null,
|
| 512 |
+
"val_iou": null,
|
| 513 |
+
"val_reward": null,
|
| 514 |
+
"val_entropy": null
|
| 515 |
+
},
|
| 516 |
+
{
|
| 517 |
+
"epoch": 24,
|
| 518 |
+
"train_loss": 0.3780487252565375,
|
| 519 |
+
"train_actor_loss": -0.14499196673452236,
|
| 520 |
+
"train_critic_loss": 0.00014369825797255246,
|
| 521 |
+
"train_mean_reward": 0.0001885495209655149,
|
| 522 |
+
"train_entropy": 0.4499398101053804,
|
| 523 |
+
"train_ce_loss": 0.4997276534863029,
|
| 524 |
+
"train_dice_loss": 0.5462100193174002,
|
| 525 |
+
"train_dice": 0.7845861859088455,
|
| 526 |
+
"train_iou": 0.6885019467525786,
|
| 527 |
+
"grad_norm": 26.614973334841025,
|
| 528 |
+
"lr": 0.001730323784284305,
|
| 529 |
+
"encoder_lr": 0.0001742520196050341,
|
| 530 |
+
"alpha": 0.33124810457229614,
|
| 531 |
+
"validated_this_epoch": false,
|
| 532 |
+
"val_loss": null,
|
| 533 |
+
"val_dice": null,
|
| 534 |
+
"val_iou": null,
|
| 535 |
+
"val_reward": null,
|
| 536 |
+
"val_entropy": null
|
| 537 |
+
},
|
| 538 |
+
{
|
| 539 |
+
"epoch": 25,
|
| 540 |
+
"train_loss": 0.38908639487759894,
|
| 541 |
+
"train_actor_loss": -0.13329775343126876,
|
| 542 |
+
"train_critic_loss": 0.00016965053916843335,
|
| 543 |
+
"train_mean_reward": 2.6888248793481033e-05,
|
| 544 |
+
"train_entropy": 0.42154980593334485,
|
| 545 |
+
"train_ce_loss": 0.48580549833381625,
|
| 546 |
+
"train_dice_loss": 0.5587931333124175,
|
| 547 |
+
"train_dice": 0.7790614682217857,
|
| 548 |
+
"train_iou": 0.6818256805813048,
|
| 549 |
+
"grad_norm": 26.845777614172114,
|
| 550 |
+
"lr": 0.0017085712472806151,
|
| 551 |
+
"encoder_lr": 0.000172175144212722,
|
| 552 |
+
"alpha": 0.3269284963607788,
|
| 553 |
+
"validated_this_epoch": false,
|
| 554 |
+
"val_loss": null,
|
| 555 |
+
"val_dice": null,
|
| 556 |
+
"val_iou": null,
|
| 557 |
+
"val_reward": null,
|
| 558 |
+
"val_entropy": null
|
| 559 |
+
},
|
| 560 |
+
{
|
| 561 |
+
"epoch": 26,
|
| 562 |
+
"train_loss": 0.39532285311454257,
|
| 563 |
+
"train_actor_loss": -0.13579296406348854,
|
| 564 |
+
"train_critic_loss": 0.00015299524621985102,
|
| 565 |
+
"train_mean_reward": 0.00015512171313312182,
|
| 566 |
+
"train_entropy": 0.44275329762574,
|
| 567 |
+
"train_ce_loss": 0.5049350640744944,
|
| 568 |
+
"train_dice_loss": 0.5571435626009023,
|
| 569 |
+
"train_dice": 0.7806902893939869,
|
| 570 |
+
"train_iou": 0.6857736482088257,
|
| 571 |
+
"grad_norm": 39.3334833642309,
|
| 572 |
+
"lr": 0.0016861243703990456,
|
| 573 |
+
"encoder_lr": 0.00017003197506322543,
|
| 574 |
+
"alpha": 0.32178056240081787,
|
| 575 |
+
"validated_this_epoch": false,
|
| 576 |
+
"val_loss": null,
|
| 577 |
+
"val_dice": null,
|
| 578 |
+
"val_iou": null,
|
| 579 |
+
"val_reward": null,
|
| 580 |
+
"val_entropy": null
|
| 581 |
+
},
|
| 582 |
+
{
|
| 583 |
+
"epoch": 27,
|
| 584 |
+
"train_loss": 0.3828199585618148,
|
| 585 |
+
"train_actor_loss": -0.13163483681912877,
|
| 586 |
+
"train_critic_loss": 0.00013637513745628805,
|
| 587 |
+
"train_mean_reward": 0.00015548940206688296,
|
| 588 |
+
"train_entropy": 0.43242431505166895,
|
| 589 |
+
"train_ce_loss": 0.4881270981011118,
|
| 590 |
+
"train_dice_loss": 0.5406461055426635,
|
| 591 |
+
"train_dice": 0.7886346166343351,
|
| 592 |
+
"train_iou": 0.6938208513665279,
|
| 593 |
+
"grad_norm": 37.907179278920786,
|
| 594 |
+
"lr": 0.001663005305997034,
|
| 595 |
+
"encoder_lr": 0.00016782462720574693,
|
| 596 |
+
"alpha": 0.3169545531272888,
|
| 597 |
+
"validated_this_epoch": false,
|
| 598 |
+
"val_loss": null,
|
| 599 |
+
"val_dice": null,
|
| 600 |
+
"val_iou": null,
|
| 601 |
+
"val_reward": null,
|
| 602 |
+
"val_entropy": null
|
| 603 |
+
},
|
| 604 |
+
{
|
| 605 |
+
"epoch": 28,
|
| 606 |
+
"train_loss": 0.3429493787427714,
|
| 607 |
+
"train_actor_loss": -0.13037152397771215,
|
| 608 |
+
"train_critic_loss": 0.0001438095249130796,
|
| 609 |
+
"train_mean_reward": 0.0003648013825353306,
|
| 610 |
+
"train_entropy": 0.43418115483312947,
|
| 611 |
+
"train_ce_loss": 0.46587762550886436,
|
| 612 |
+
"train_dice_loss": 0.4806203594195442,
|
| 613 |
+
"train_dice": 0.8047064984877329,
|
| 614 |
+
"train_iou": 0.7090507375866408,
|
| 615 |
+
"grad_norm": 18.304460791192312,
|
| 616 |
+
"lr": 0.0016392368697999468,
|
| 617 |
+
"encoder_lr": 0.00016555527902612554,
|
| 618 |
+
"alpha": 0.3119323253631592,
|
| 619 |
+
"validated_this_epoch": false,
|
| 620 |
+
"val_loss": null,
|
| 621 |
+
"val_dice": null,
|
| 622 |
+
"val_iou": null,
|
| 623 |
+
"val_reward": null,
|
| 624 |
+
"val_entropy": null
|
| 625 |
+
},
|
| 626 |
+
{
|
| 627 |
+
"epoch": 29,
|
| 628 |
+
"train_loss": 0.3824699216652228,
|
| 629 |
+
"train_actor_loss": -0.1247279809436518,
|
| 630 |
+
"train_critic_loss": 0.0001429658757382222,
|
| 631 |
+
"train_mean_reward": 0.00011097060011991435,
|
| 632 |
+
"train_entropy": 0.42478789997501265,
|
| 633 |
+
"train_ce_loss": 0.4865120022957158,
|
| 634 |
+
"train_dice_loss": 0.5277408230520344,
|
| 635 |
+
"train_dice": 0.7835890101750125,
|
| 636 |
+
"train_iou": 0.68982450170887,
|
| 637 |
+
"grad_norm": 25.796629567016925,
|
| 638 |
+
"lr": 0.0016148425183847121,
|
| 639 |
+
"encoder_lr": 0.0001632261700970328,
|
| 640 |
+
"alpha": 0.30725884437561035,
|
| 641 |
+
"validated_this_epoch": false,
|
| 642 |
+
"val_loss": null,
|
| 643 |
+
"val_dice": null,
|
| 644 |
+
"val_iou": null,
|
| 645 |
+
"val_reward": null,
|
| 646 |
+
"val_entropy": null
|
| 647 |
+
},
|
| 648 |
+
{
|
| 649 |
+
"epoch": 30,
|
| 650 |
+
"train_loss": 0.3841695500783218,
|
| 651 |
+
"train_actor_loss": -0.12016126437916466,
|
| 652 |
+
"train_critic_loss": 0.000157365113309589,
|
| 653 |
+
"train_mean_reward": 0.00015016873281504297,
|
| 654 |
+
"train_entropy": 0.4186583744886617,
|
| 655 |
+
"train_ce_loss": 0.4737400686300964,
|
| 656 |
+
"train_dice_loss": 0.534764184816247,
|
| 657 |
+
"train_dice": 0.7947439405220528,
|
| 658 |
+
"train_iou": 0.7026476943567499,
|
| 659 |
+
"grad_norm": 25.832807304323183,
|
| 660 |
+
"lr": 0.0015898463260310114,
|
| 661 |
+
"encoder_lr": 0.00016083959896778498,
|
| 662 |
+
"alpha": 0.3028067350387573,
|
| 663 |
+
"validated_this_epoch": true,
|
| 664 |
+
"val_loss": 0.8005378840872172,
|
| 665 |
+
"val_dice": 0.719336415518507,
|
| 666 |
+
"val_iou": 0.6182636885216048,
|
| 667 |
+
"val_reward": 0.00017815031171009807,
|
| 668 |
+
"val_entropy": 0.7006657554043665,
|
| 669 |
+
"val_actor_loss": 0.005542401223786815,
|
| 670 |
+
"val_critic_loss": 0.00015388492244616916,
|
| 671 |
+
"val_ce_loss": 0.7663840318917099,
|
| 672 |
+
"val_dice_loss": 0.8234530148482082
|
| 673 |
+
},
|
| 674 |
+
{
|
| 675 |
+
"epoch": 31,
|
| 676 |
+
"train_loss": 0.3483792745811475,
|
| 677 |
+
"train_actor_loss": -0.11970010543728213,
|
| 678 |
+
"train_critic_loss": 0.00014264401010486298,
|
| 679 |
+
"train_mean_reward": 0.0003100846606079926,
|
| 680 |
+
"train_entropy": 0.4194148445853276,
|
| 681 |
+
"train_ce_loss": 0.45510414917411846,
|
| 682 |
+
"train_dice_loss": 0.48091195489085,
|
| 683 |
+
"train_dice": 0.8001322847189095,
|
| 684 |
+
"train_iou": 0.7070290593958339,
|
| 685 |
+
"grad_norm": 24.737091465051783,
|
| 686 |
+
"lr": 0.0015642729609628705,
|
| 687 |
+
"encoder_lr": 0.00015839792089595244,
|
| 688 |
+
"alpha": 0.2984693646430969,
|
| 689 |
+
"validated_this_epoch": false,
|
| 690 |
+
"val_loss": null,
|
| 691 |
+
"val_dice": null,
|
| 692 |
+
"val_iou": null,
|
| 693 |
+
"val_reward": null,
|
| 694 |
+
"val_entropy": null
|
| 695 |
+
},
|
| 696 |
+
{
|
| 697 |
+
"epoch": 32,
|
| 698 |
+
"train_loss": 0.36958907622584075,
|
| 699 |
+
"train_actor_loss": -0.1152340413121752,
|
| 700 |
+
"train_critic_loss": 0.0001391026548515804,
|
| 701 |
+
"train_mean_reward": 0.00017435385433828423,
|
| 702 |
+
"train_entropy": 0.4164304467211706,
|
| 703 |
+
"train_ce_loss": 0.47014906941564116,
|
| 704 |
+
"train_dice_loss": 0.4993580527391852,
|
| 705 |
+
"train_dice": 0.7966122601177305,
|
| 706 |
+
"train_iou": 0.7061758950702463,
|
| 707 |
+
"grad_norm": 24.56620332067327,
|
| 708 |
+
"lr": 0.001538147661004102,
|
| 709 |
+
"encoder_lr": 0.0001559035455230047,
|
| 710 |
+
"alpha": 0.2939406633377075,
|
| 711 |
+
"validated_this_epoch": false,
|
| 712 |
+
"val_loss": null,
|
| 713 |
+
"val_dice": null,
|
| 714 |
+
"val_iou": null,
|
| 715 |
+
"val_reward": null,
|
| 716 |
+
"val_entropy": null
|
| 717 |
+
},
|
| 718 |
+
{
|
| 719 |
+
"epoch": 33,
|
| 720 |
+
"train_loss": 0.36723682661937884,
|
| 721 |
+
"train_actor_loss": -0.1084495876465734,
|
| 722 |
+
"train_critic_loss": 0.0001362829561384028,
|
| 723 |
+
"train_mean_reward": 0.0002597813012711018,
|
| 724 |
+
"train_entropy": 0.39607080638177633,
|
| 725 |
+
"train_ce_loss": 0.44253188522186043,
|
| 726 |
+
"train_dice_loss": 0.5087046493855558,
|
| 727 |
+
"train_dice": 0.8035252065650282,
|
| 728 |
+
"train_iou": 0.7096314218783073,
|
| 729 |
+
"grad_norm": 20.297288823035338,
|
| 730 |
+
"lr": 0.0015114962086716199,
|
| 731 |
+
"encoder_lr": 0.0001533589344962853,
|
| 732 |
+
"alpha": 0.29033252596855164,
|
| 733 |
+
"validated_this_epoch": false,
|
| 734 |
+
"val_loss": null,
|
| 735 |
+
"val_dice": null,
|
| 736 |
+
"val_iou": null,
|
| 737 |
+
"val_reward": null,
|
| 738 |
+
"val_entropy": null
|
| 739 |
+
},
|
| 740 |
+
{
|
| 741 |
+
"epoch": 34,
|
| 742 |
+
"train_loss": 0.37566004671488507,
|
| 743 |
+
"train_actor_loss": -0.11033118959487409,
|
| 744 |
+
"train_critic_loss": 0.0001425033112439472,
|
| 745 |
+
"train_mean_reward": 0.00014549040100343967,
|
| 746 |
+
"train_entropy": 0.4101937913771439,
|
| 747 |
+
"train_ce_loss": 0.44928009827195325,
|
| 748 |
+
"train_dice_loss": 0.5225598561825371,
|
| 749 |
+
"train_dice": 0.7964956943647349,
|
| 750 |
+
"train_iou": 0.7042059690209154,
|
| 751 |
+
"grad_norm": 28.742233728253563,
|
| 752 |
+
"lr": 0.0014843449057312071,
|
| 753 |
+
"encoder_lr": 0.00015076659903966297,
|
| 754 |
+
"alpha": 0.2861836552619934,
|
| 755 |
+
"validated_this_epoch": false,
|
| 756 |
+
"val_loss": null,
|
| 757 |
+
"val_dice": null,
|
| 758 |
+
"val_iou": null,
|
| 759 |
+
"val_reward": null,
|
| 760 |
+
"val_entropy": null
|
| 761 |
+
},
|
| 762 |
+
{
|
| 763 |
+
"epoch": 35,
|
| 764 |
+
"train_loss": 0.37246958079004244,
|
| 765 |
+
"train_actor_loss": -0.10782961958408047,
|
| 766 |
+
"train_critic_loss": 0.00013862756627713917,
|
| 767 |
+
"train_mean_reward": 0.00018686077199525183,
|
| 768 |
+
"train_entropy": 0.4078874550663532,
|
| 769 |
+
"train_ce_loss": 0.4632008154094412,
|
| 770 |
+
"train_dice_loss": 0.4972589435657481,
|
| 771 |
+
"train_dice": 0.7994193666937086,
|
| 772 |
+
"train_iou": 0.7088677735598863,
|
| 773 |
+
"grad_norm": 27.656757638898007,
|
| 774 |
+
"lr": 0.0014567205472408495,
|
| 775 |
+
"encoder_lr": 0.00014812909747525697,
|
| 776 |
+
"alpha": 0.28203633427619934,
|
| 777 |
+
"validated_this_epoch": false,
|
| 778 |
+
"val_loss": null,
|
| 779 |
+
"val_dice": null,
|
| 780 |
+
"val_iou": null,
|
| 781 |
+
"val_reward": null,
|
| 782 |
+
"val_entropy": null
|
| 783 |
+
},
|
| 784 |
+
{
|
| 785 |
+
"epoch": 36,
|
| 786 |
+
"train_loss": 0.36569868010115025,
|
| 787 |
+
"train_actor_loss": -0.10511455387454624,
|
| 788 |
+
"train_critic_loss": 0.00013444818822619134,
|
| 789 |
+
"train_mean_reward": 0.00020051913024731185,
|
| 790 |
+
"train_entropy": 0.40169064635146184,
|
| 791 |
+
"train_ce_loss": 0.4443169142256758,
|
| 792 |
+
"train_dice_loss": 0.4971750939231203,
|
| 793 |
+
"train_dice": 0.8128801082313178,
|
| 794 |
+
"train_iou": 0.7203668184998436,
|
| 795 |
+
"grad_norm": 26.154768455398177,
|
| 796 |
+
"lr": 0.0014286503951072478,
|
| 797 |
+
"encoder_lr": 0.00014544903269868192,
|
| 798 |
+
"alpha": 0.27821218967437744,
|
| 799 |
+
"validated_this_epoch": false,
|
| 800 |
+
"val_loss": null,
|
| 801 |
+
"val_dice": null,
|
| 802 |
+
"val_iou": null,
|
| 803 |
+
"val_reward": null,
|
| 804 |
+
"val_entropy": null
|
| 805 |
+
},
|
| 806 |
+
{
|
| 807 |
+
"epoch": 37,
|
| 808 |
+
"train_loss": 0.3712652839301515,
|
| 809 |
+
"train_actor_loss": -0.10037048568694014,
|
| 810 |
+
"train_critic_loss": 0.00013630793239048856,
|
| 811 |
+
"train_mean_reward": 4.5652922321195115e-05,
|
| 812 |
+
"train_entropy": 0.39490996079774476,
|
| 813 |
+
"train_ce_loss": 0.45512808208625755,
|
| 814 |
+
"train_dice_loss": 0.48800713532346773,
|
| 815 |
+
"train_dice": 0.8035842073731768,
|
| 816 |
+
"train_iou": 0.7133901172020373,
|
| 817 |
+
"grad_norm": 23.78031743358272,
|
| 818 |
+
"lr": 0.001400162151181607,
|
| 819 |
+
"encoder_lr": 0.00014272904961030417,
|
| 820 |
+
"alpha": 0.2747458517551422,
|
| 821 |
+
"validated_this_epoch": false,
|
| 822 |
+
"val_loss": null,
|
| 823 |
+
"val_dice": null,
|
| 824 |
+
"val_iou": null,
|
| 825 |
+
"val_reward": null,
|
| 826 |
+
"val_entropy": null
|
| 827 |
+
},
|
| 828 |
+
{
|
| 829 |
+
"epoch": 38,
|
| 830 |
+
"train_loss": 0.34659122102785594,
|
| 831 |
+
"train_actor_loss": -0.0976431970448457,
|
| 832 |
+
"train_critic_loss": 0.0001405083436048001,
|
| 833 |
+
"train_mean_reward": 0.00015749140965153177,
|
| 834 |
+
"train_entropy": 0.39074589623961337,
|
| 835 |
+
"train_ce_loss": 0.43406275981069764,
|
| 836 |
+
"train_dice_loss": 0.45426555613214653,
|
| 837 |
+
"train_dice": 0.8197832531580147,
|
| 838 |
+
"train_iou": 0.72957933267766,
|
| 839 |
+
"grad_norm": 23.137535401554995,
|
| 840 |
+
"lr": 0.0013712839299212553,
|
| 841 |
+
"encoder_lr": 0.00013997183250504444,
|
| 842 |
+
"alpha": 0.2713741958141327,
|
| 843 |
+
"validated_this_epoch": false,
|
| 844 |
+
"val_loss": null,
|
| 845 |
+
"val_dice": null,
|
| 846 |
+
"val_iou": null,
|
| 847 |
+
"val_reward": null,
|
| 848 |
+
"val_entropy": null
|
| 849 |
+
},
|
| 850 |
+
{
|
| 851 |
+
"epoch": 39,
|
| 852 |
+
"train_loss": 0.34516546673812837,
|
| 853 |
+
"train_actor_loss": -0.0988163874323362,
|
| 854 |
+
"train_critic_loss": 0.00013380287337614585,
|
| 855 |
+
"train_mean_reward": 0.00015119109996959793,
|
| 856 |
+
"train_entropy": 0.38998730222692174,
|
| 857 |
+
"train_ce_loss": 0.4231410857136012,
|
| 858 |
+
"train_dice_loss": 0.4646888094971038,
|
| 859 |
+
"train_dice": 0.8202555758321584,
|
| 860 |
+
"train_iou": 0.731533634370036,
|
| 861 |
+
"grad_norm": 23.102389688870705,
|
| 862 |
+
"lr": 0.0013420442306440656,
|
| 863 |
+
"encoder_lr": 0.0001371801024233027,
|
| 864 |
+
"alpha": 0.2682475745677948,
|
| 865 |
+
"validated_this_epoch": false,
|
| 866 |
+
"val_loss": null,
|
| 867 |
+
"val_dice": null,
|
| 868 |
+
"val_iou": null,
|
| 869 |
+
"val_reward": null,
|
| 870 |
+
"val_entropy": null
|
| 871 |
+
},
|
| 872 |
+
{
|
| 873 |
+
"epoch": 40,
|
| 874 |
+
"train_loss": 0.34337753655980957,
|
| 875 |
+
"train_actor_loss": -0.09437107133561172,
|
| 876 |
+
"train_critic_loss": 0.00012730427457387187,
|
| 877 |
+
"train_mean_reward": 0.00012083658739991437,
|
| 878 |
+
"train_entropy": 0.37829540974101966,
|
| 879 |
+
"train_ce_loss": 0.4199352459356343,
|
| 880 |
+
"train_dice_loss": 0.4554346518769129,
|
| 881 |
+
"train_dice": 0.8226035641539043,
|
| 882 |
+
"train_iou": 0.7337379314716742,
|
| 883 |
+
"grad_norm": 16.85659814620203,
|
| 884 |
+
"lr": 0.0013124719094030732,
|
| 885 |
+
"encoder_lr": 0.00013435661446562,
|
| 886 |
+
"alpha": 0.2651420831680298,
|
| 887 |
+
"validated_this_epoch": true,
|
| 888 |
+
"val_loss": 0.6661472226471894,
|
| 889 |
+
"val_dice": 0.7398004646205255,
|
| 890 |
+
"val_iou": 0.6408889753679513,
|
| 891 |
+
"val_reward": 4.633968136190449e-06,
|
| 892 |
+
"val_entropy": 0.5061837929968882,
|
| 893 |
+
"val_actor_loss": -0.0028216223123088257,
|
| 894 |
+
"val_critic_loss": 8.342342312751111e-05,
|
| 895 |
+
"val_ce_loss": 0.5672927435576878,
|
| 896 |
+
"val_dice_loss": 0.7705614946105265
|
| 897 |
+
},
|
| 898 |
+
{
|
| 899 |
+
"epoch": 41,
|
| 900 |
+
"train_loss": 0.3639304505646691,
|
| 901 |
+
"train_actor_loss": -0.09513970169428862,
|
| 902 |
+
"train_critic_loss": 0.00014166066102093834,
|
| 903 |
+
"train_mean_reward": 0.0001632288085496585,
|
| 904 |
+
"train_entropy": 0.38950260008567367,
|
| 905 |
+
"train_ce_loss": 0.42250009601002014,
|
| 906 |
+
"train_dice_loss": 0.4954985340128265,
|
| 907 |
+
"train_dice": 0.8072809255234856,
|
| 908 |
+
"train_iou": 0.7181567705886672,
|
| 909 |
+
"grad_norm": 19.02943361752717,
|
| 910 |
+
"lr": 0.0012825961505090336,
|
| 911 |
+
"encoder_lr": 0.00013150415507372682,
|
| 912 |
+
"alpha": 0.26216208934783936,
|
| 913 |
+
"validated_this_epoch": false,
|
| 914 |
+
"val_loss": null,
|
| 915 |
+
"val_dice": null,
|
| 916 |
+
"val_iou": null,
|
| 917 |
+
"val_reward": null,
|
| 918 |
+
"val_entropy": null
|
| 919 |
+
},
|
| 920 |
+
{
|
| 921 |
+
"epoch": 42,
|
| 922 |
+
"train_loss": 0.38788169173364956,
|
| 923 |
+
"train_actor_loss": -0.09157998811511338,
|
| 924 |
+
"train_critic_loss": 0.0001348569546887992,
|
| 925 |
+
"train_mean_reward": 0.00013418730785526504,
|
| 926 |
+
"train_entropy": 0.37884149700403214,
|
| 927 |
+
"train_ce_loss": 0.4306592201207613,
|
| 928 |
+
"train_dice_loss": 0.5281292702954562,
|
| 929 |
+
"train_dice": 0.8050203055498506,
|
| 930 |
+
"train_iou": 0.7146521429808815,
|
| 931 |
+
"grad_norm": 24.381595764973365,
|
| 932 |
+
"lr": 0.0012524464377290307,
|
| 933 |
+
"encoder_lr": 0.00012862553928066122,
|
| 934 |
+
"alpha": 0.2591218054294586,
|
| 935 |
+
"validated_this_epoch": false,
|
| 936 |
+
"val_loss": null,
|
| 937 |
+
"val_dice": null,
|
| 938 |
+
"val_iou": null,
|
| 939 |
+
"val_reward": null,
|
| 940 |
+
"val_entropy": null
|
| 941 |
+
},
|
| 942 |
+
{
|
| 943 |
+
"epoch": 43,
|
| 944 |
+
"train_loss": 0.34964677242403347,
|
| 945 |
+
"train_actor_loss": -0.08890084440047427,
|
| 946 |
+
"train_critic_loss": 0.00012647496734699166,
|
| 947 |
+
"train_mean_reward": 0.0002560999898674328,
|
| 948 |
+
"train_entropy": 0.3753304744105622,
|
| 949 |
+
"train_ce_loss": 0.41325800430549436,
|
| 950 |
+
"train_dice_loss": 0.4637107400161043,
|
| 951 |
+
"train_dice": 0.8212461406535518,
|
| 952 |
+
"train_iou": 0.7353043384630651,
|
| 953 |
+
"grad_norm": 18.612451695656592,
|
| 954 |
+
"lr": 0.0012220525251895604,
|
| 955 |
+
"encoder_lr": 0.0001257236079326716,
|
| 956 |
+
"alpha": 0.25658485293388367,
|
| 957 |
+
"validated_this_epoch": false,
|
| 958 |
+
"val_loss": null,
|
| 959 |
+
"val_dice": null,
|
| 960 |
+
"val_iou": null,
|
| 961 |
+
"val_reward": null,
|
| 962 |
+
"val_entropy": null
|
| 963 |
+
},
|
| 964 |
+
{
|
| 965 |
+
"epoch": 44,
|
| 966 |
+
"train_loss": 0.3581098036031968,
|
| 967 |
+
"train_actor_loss": -0.08932060533195965,
|
| 968 |
+
"train_critic_loss": 0.00014833953186643595,
|
| 969 |
+
"train_mean_reward": 0.00019072714067713307,
|
| 970 |
+
"train_entropy": 0.3795520902565591,
|
| 971 |
+
"train_ce_loss": 0.420504329547416,
|
| 972 |
+
"train_dice_loss": 0.47420813516744964,
|
| 973 |
+
"train_dice": 0.8151747006899632,
|
| 974 |
+
"train_iou": 0.7265285797567105,
|
| 975 |
+
"grad_norm": 23.536205614260002,
|
| 976 |
+
"lr": 0.0011914444080127966,
|
| 977 |
+
"encoder_lr": 0.0001228012248856439,
|
| 978 |
+
"alpha": 0.2537332773208618,
|
| 979 |
+
"validated_this_epoch": false,
|
| 980 |
+
"val_loss": null,
|
| 981 |
+
"val_dice": null,
|
| 982 |
+
"val_iou": null,
|
| 983 |
+
"val_reward": null,
|
| 984 |
+
"val_entropy": null
|
| 985 |
+
},
|
| 986 |
+
{
|
| 987 |
+
"epoch": 45,
|
| 988 |
+
"train_loss": 0.3296818918685696,
|
| 989 |
+
"train_actor_loss": -0.08440704489829504,
|
| 990 |
+
"train_critic_loss": 0.00012505974848908496,
|
| 991 |
+
"train_mean_reward": 0.000270947692055514,
|
| 992 |
+
"train_entropy": 0.3576581258451908,
|
| 993 |
+
"train_ce_loss": 0.39342997775555083,
|
| 994 |
+
"train_dice_loss": 0.4346228246540986,
|
| 995 |
+
"train_dice": 0.8301809164656195,
|
| 996 |
+
"train_iou": 0.7426396492908831,
|
| 997 |
+
"grad_norm": 15.553295236687328,
|
| 998 |
+
"lr": 0.0011606522927150302,
|
| 999 |
+
"encoder_lr": 0.00011986127417882199,
|
| 1000 |
+
"alpha": 0.25193050503730774,
|
| 1001 |
+
"validated_this_epoch": false,
|
| 1002 |
+
"val_loss": null,
|
| 1003 |
+
"val_dice": null,
|
| 1004 |
+
"val_iou": null,
|
| 1005 |
+
"val_reward": null,
|
| 1006 |
+
"val_entropy": null
|
| 1007 |
+
},
|
| 1008 |
+
{
|
| 1009 |
+
"epoch": 46,
|
| 1010 |
+
"train_loss": 0.33681043100998154,
|
| 1011 |
+
"train_actor_loss": -0.08502839836328066,
|
| 1012 |
+
"train_critic_loss": 0.0001308202621098656,
|
| 1013 |
+
"train_mean_reward": 0.00014770419428166798,
|
| 1014 |
+
"train_entropy": 0.36429003649210745,
|
| 1015 |
+
"train_ce_loss": 0.3938134777328379,
|
| 1016 |
+
"train_dice_loss": 0.44973334862588293,
|
| 1017 |
+
"train_dice": 0.823815928785608,
|
| 1018 |
+
"train_iou": 0.7347555344395293,
|
| 1019 |
+
"grad_norm": 17.340920525465826,
|
| 1020 |
+
"lr": 0.0011297065673964834,
|
| 1021 |
+
"encoder_lr": 0.00011690665718860897,
|
| 1022 |
+
"alpha": 0.24997247755527496,
|
| 1023 |
+
"validated_this_epoch": false,
|
| 1024 |
+
"val_loss": null,
|
| 1025 |
+
"val_dice": null,
|
| 1026 |
+
"val_iou": null,
|
| 1027 |
+
"val_reward": null,
|
| 1028 |
+
"val_entropy": null
|
| 1029 |
+
},
|
| 1030 |
+
{
|
| 1031 |
+
"epoch": 47,
|
| 1032 |
+
"train_loss": 0.3300588635020247,
|
| 1033 |
+
"train_actor_loss": -0.08265878778018383,
|
| 1034 |
+
"train_critic_loss": 0.00012799435771457794,
|
| 1035 |
+
"train_mean_reward": 0.0002089108672292591,
|
| 1036 |
+
"train_entropy": 0.3630187226704849,
|
| 1037 |
+
"train_ce_loss": 0.4005389582433133,
|
| 1038 |
+
"train_dice_loss": 0.42476833852378587,
|
| 1039 |
+
"train_dice": 0.8402610316690685,
|
| 1040 |
+
"train_iou": 0.7523163984788156,
|
| 1041 |
+
"grad_norm": 12.577170362306196,
|
| 1042 |
+
"lr": 0.001098637771751922,
|
| 1043 |
+
"encoder_lr": 0.0001139402897652589,
|
| 1044 |
+
"alpha": 0.2479257732629776,
|
| 1045 |
+
"validated_this_epoch": false,
|
| 1046 |
+
"val_loss": null,
|
| 1047 |
+
"val_dice": null,
|
| 1048 |
+
"val_iou": null,
|
| 1049 |
+
"val_reward": null,
|
| 1050 |
+
"val_entropy": null
|
| 1051 |
+
},
|
| 1052 |
+
{
|
| 1053 |
+
"epoch": 48,
|
| 1054 |
+
"train_loss": 0.3352510346417395,
|
| 1055 |
+
"train_actor_loss": -0.08028301221946645,
|
| 1056 |
+
"train_critic_loss": 0.00012717796010901342,
|
| 1057 |
+
"train_mean_reward": 0.0002520704256042193,
|
| 1058 |
+
"train_entropy": 0.34976153693688933,
|
| 1059 |
+
"train_ce_loss": 0.39535445724770385,
|
| 1060 |
+
"train_dice_loss": 0.43558644547634967,
|
| 1061 |
+
"train_dice": 0.827534154009597,
|
| 1062 |
+
"train_iou": 0.7404953736419453,
|
| 1063 |
+
"grad_norm": 20.311514901560408,
|
| 1064 |
+
"lr": 0.0010674765669316673,
|
| 1065 |
+
"encoder_lr": 0.00011096509935528483,
|
| 1066 |
+
"alpha": 0.2465425580739975,
|
| 1067 |
+
"validated_this_epoch": false,
|
| 1068 |
+
"val_loss": null,
|
| 1069 |
+
"val_dice": null,
|
| 1070 |
+
"val_iou": null,
|
| 1071 |
+
"val_reward": null,
|
| 1072 |
+
"val_entropy": null
|
| 1073 |
+
},
|
| 1074 |
+
{
|
| 1075 |
+
"epoch": 49,
|
| 1076 |
+
"train_loss": 0.3290987055127819,
|
| 1077 |
+
"train_actor_loss": -0.08222802513993757,
|
| 1078 |
+
"train_critic_loss": 0.00012018642991154636,
|
| 1079 |
+
"train_mean_reward": 0.00013203583620063322,
|
| 1080 |
+
"train_entropy": 0.3606551154600866,
|
| 1081 |
+
"train_ce_loss": 0.3977279713141518,
|
| 1082 |
+
"train_dice_loss": 0.4248052921714093,
|
| 1083 |
+
"train_dice": 0.8269430755728544,
|
| 1084 |
+
"train_iou": 0.7412282039856082,
|
| 1085 |
+
"grad_norm": 11.538063617177713,
|
| 1086 |
+
"lr": 0.001036253705282738,
|
| 1087 |
+
"encoder_lr": 0.00010798402211242223,
|
| 1088 |
+
"alpha": 0.24484525620937347,
|
| 1089 |
+
"validated_this_epoch": false,
|
| 1090 |
+
"val_loss": null,
|
| 1091 |
+
"val_dice": null,
|
| 1092 |
+
"val_iou": null,
|
| 1093 |
+
"val_reward": null,
|
| 1094 |
+
"val_entropy": null
|
| 1095 |
+
},
|
| 1096 |
+
{
|
| 1097 |
+
"epoch": 50,
|
| 1098 |
+
"train_loss": 0.32694140568288954,
|
| 1099 |
+
"train_actor_loss": -0.07954849014708469,
|
| 1100 |
+
"train_critic_loss": 0.00011943667615773508,
|
| 1101 |
+
"train_mean_reward": 0.00014478434278746924,
|
| 1102 |
+
"train_entropy": 0.3522500717670727,
|
| 1103 |
+
"train_ce_loss": 0.3900441239920062,
|
| 1104 |
+
"train_dice_loss": 0.42281621871684566,
|
| 1105 |
+
"train_dice": 0.8420062959898956,
|
| 1106 |
+
"train_iou": 0.7574317933854343,
|
| 1107 |
+
"grad_norm": 17.142169914504354,
|
| 1108 |
+
"lr": 0.0010050000000000003,
|
| 1109 |
+
"encoder_lr": 0.00010500000000000005,
|
| 1110 |
+
"alpha": 0.24330276250839233,
|
| 1111 |
+
"validated_this_epoch": true,
|
| 1112 |
+
"val_loss": 0.6609939909469562,
|
| 1113 |
+
"val_dice": 0.7257673794089163,
|
| 1114 |
+
"val_iou": 0.6307605070392235,
|
| 1115 |
+
"val_reward": -0.00047722498947214175,
|
| 1116 |
+
"val_entropy": 0.5358675189993597,
|
| 1117 |
+
"val_actor_loss": -0.009008371088926615,
|
| 1118 |
+
"val_critic_loss": 0.00011903740661834648,
|
| 1119 |
+
"val_ce_loss": 0.6363068507429249,
|
| 1120 |
+
"val_dice_loss": 0.7035788138105411
|
| 1121 |
+
},
|
| 1122 |
+
{
|
| 1123 |
+
"epoch": 51,
|
| 1124 |
+
"train_loss": 0.32810530010177646,
|
| 1125 |
+
"train_actor_loss": -0.08019897433657208,
|
| 1126 |
+
"train_critic_loss": 0.00012296190519864986,
|
| 1127 |
+
"train_mean_reward": 0.00020249696290816025,
|
| 1128 |
+
"train_entropy": 0.35550544172217374,
|
| 1129 |
+
"train_ce_loss": 0.39523161621042346,
|
| 1130 |
+
"train_dice_loss": 0.42125395972291324,
|
| 1131 |
+
"train_dice": 0.83986629286193,
|
| 1132 |
+
"train_iou": 0.7533175089949959,
|
| 1133 |
+
"grad_norm": 12.47872456792713,
|
| 1134 |
+
"lr": 0.0009737462947172626,
|
| 1135 |
+
"encoder_lr": 0.00010201597788757786,
|
| 1136 |
+
"alpha": 0.2417379468679428,
|
| 1137 |
+
"validated_this_epoch": false,
|
| 1138 |
+
"val_loss": null,
|
| 1139 |
+
"val_dice": null,
|
| 1140 |
+
"val_iou": null,
|
| 1141 |
+
"val_reward": null,
|
| 1142 |
+
"val_entropy": null
|
| 1143 |
+
},
|
| 1144 |
+
{
|
| 1145 |
+
"epoch": 52,
|
| 1146 |
+
"train_loss": 0.3195515123390874,
|
| 1147 |
+
"train_actor_loss": -0.08083896847918283,
|
| 1148 |
+
"train_critic_loss": 0.00012635513610789072,
|
| 1149 |
+
"train_mean_reward": 0.00022060030848657677,
|
| 1150 |
+
"train_entropy": 0.35755756607120354,
|
| 1151 |
+
"train_ce_loss": 0.3861291893139404,
|
| 1152 |
+
"train_dice_loss": 0.41452540722928305,
|
| 1153 |
+
"train_dice": 0.8412893763029696,
|
| 1154 |
+
"train_iou": 0.7540153808170909,
|
| 1155 |
+
"grad_norm": 11.124208023381788,
|
| 1156 |
+
"lr": 0.0009425234330683334,
|
| 1157 |
+
"encoder_lr": 9.903490064471527e-05,
|
| 1158 |
+
"alpha": 0.24000048637390137,
|
| 1159 |
+
"validated_this_epoch": false,
|
| 1160 |
+
"val_loss": null,
|
| 1161 |
+
"val_dice": null,
|
| 1162 |
+
"val_iou": null,
|
| 1163 |
+
"val_reward": null,
|
| 1164 |
+
"val_entropy": null
|
| 1165 |
+
},
|
| 1166 |
+
{
|
| 1167 |
+
"epoch": 53,
|
| 1168 |
+
"train_loss": 0.31989620574713906,
|
| 1169 |
+
"train_actor_loss": -0.07736056789669186,
|
| 1170 |
+
"train_critic_loss": 0.00013042327430456767,
|
| 1171 |
+
"train_mean_reward": 0.0002896128579406405,
|
| 1172 |
+
"train_entropy": 0.3445096579633018,
|
| 1173 |
+
"train_ce_loss": 0.36765422839526984,
|
| 1174 |
+
"train_dice_loss": 0.4267288845330862,
|
| 1175 |
+
"train_dice": 0.8393492094799264,
|
| 1176 |
+
"train_iou": 0.7518908600332397,
|
| 1177 |
+
"grad_norm": 11.346680256516434,
|
| 1178 |
+
"lr": 0.0009113622282480784,
|
| 1179 |
+
"encoder_lr": 9.605971023474117e-05,
|
| 1180 |
+
"alpha": 0.23904910683631897,
|
| 1181 |
+
"validated_this_epoch": false,
|
| 1182 |
+
"val_loss": null,
|
| 1183 |
+
"val_dice": null,
|
| 1184 |
+
"val_iou": null,
|
| 1185 |
+
"val_reward": null,
|
| 1186 |
+
"val_entropy": null
|
| 1187 |
+
},
|
| 1188 |
+
{
|
| 1189 |
+
"epoch": 54,
|
| 1190 |
+
"train_loss": 0.3126838584755396,
|
| 1191 |
+
"train_actor_loss": -0.07909460926731658,
|
| 1192 |
+
"train_critic_loss": 0.00011665948026940137,
|
| 1193 |
+
"train_mean_reward": 0.0002334243813540233,
|
| 1194 |
+
"train_entropy": 0.35079435155142186,
|
| 1195 |
+
"train_ce_loss": 0.3847371113501027,
|
| 1196 |
+
"train_dice_loss": 0.39870315459039485,
|
| 1197 |
+
"train_dice": 0.8424953708421018,
|
| 1198 |
+
"train_iou": 0.7568916018308116,
|
| 1199 |
+
"grad_norm": 9.734733078137848,
|
| 1200 |
+
"lr": 0.0008802934326035174,
|
| 1201 |
+
"encoder_lr": 9.309334281139113e-05,
|
| 1202 |
+
"alpha": 0.23761622607707977,
|
| 1203 |
+
"validated_this_epoch": false,
|
| 1204 |
+
"val_loss": null,
|
| 1205 |
+
"val_dice": null,
|
| 1206 |
+
"val_iou": null,
|
| 1207 |
+
"val_reward": null,
|
| 1208 |
+
"val_entropy": null
|
| 1209 |
+
},
|
| 1210 |
+
{
|
| 1211 |
+
"epoch": 55,
|
| 1212 |
+
"train_loss": 0.3203974377739337,
|
| 1213 |
+
"train_actor_loss": -0.07711033656163194,
|
| 1214 |
+
"train_critic_loss": 0.00011569674984104861,
|
| 1215 |
+
"train_mean_reward": 0.0002693216590126962,
|
| 1216 |
+
"train_entropy": 0.3434054397566374,
|
| 1217 |
+
"train_ce_loss": 0.37575268809435747,
|
| 1218 |
+
"train_dice_loss": 0.41914715331038144,
|
| 1219 |
+
"train_dice": 0.8392136805264592,
|
| 1220 |
+
"train_iou": 0.7538897563710298,
|
| 1221 |
+
"grad_norm": 8.833165427742077,
|
| 1222 |
+
"lr": 0.0008493477072849704,
|
| 1223 |
+
"encoder_lr": 9.013872582117809e-05,
|
| 1224 |
+
"alpha": 0.23675352334976196,
|
| 1225 |
+
"validated_this_epoch": false,
|
| 1226 |
+
"val_loss": null,
|
| 1227 |
+
"val_dice": null,
|
| 1228 |
+
"val_iou": null,
|
| 1229 |
+
"val_reward": null,
|
| 1230 |
+
"val_entropy": null
|
| 1231 |
+
},
|
| 1232 |
+
{
|
| 1233 |
+
"epoch": 56,
|
| 1234 |
+
"train_loss": 0.3169864391612048,
|
| 1235 |
+
"train_actor_loss": -0.0783995640649429,
|
| 1236 |
+
"train_critic_loss": 0.00013419919499038164,
|
| 1237 |
+
"train_mean_reward": 0.0002763906886294002,
|
| 1238 |
+
"train_entropy": 0.35053005814552307,
|
| 1239 |
+
"train_ce_loss": 0.3813023854465794,
|
| 1240 |
+
"train_dice_loss": 0.40933541041011956,
|
| 1241 |
+
"train_dice": 0.8370911792398404,
|
| 1242 |
+
"train_iou": 0.7506825233081726,
|
| 1243 |
+
"grad_norm": 12.303973652133646,
|
| 1244 |
+
"lr": 0.0008185555919872042,
|
| 1245 |
+
"encoder_lr": 8.719877511435618e-05,
|
| 1246 |
+
"alpha": 0.2352609634399414,
|
| 1247 |
+
"validated_this_epoch": false,
|
| 1248 |
+
"val_loss": null,
|
| 1249 |
+
"val_dice": null,
|
| 1250 |
+
"val_iou": null,
|
| 1251 |
+
"val_reward": null,
|
| 1252 |
+
"val_entropy": null
|
| 1253 |
+
},
|
| 1254 |
+
{
|
| 1255 |
+
"epoch": 57,
|
| 1256 |
+
"train_loss": 0.32172405992656256,
|
| 1257 |
+
"train_actor_loss": -0.07612577078320105,
|
| 1258 |
+
"train_critic_loss": 0.00011965676430354364,
|
| 1259 |
+
"train_mean_reward": 0.0003141100375593618,
|
| 1260 |
+
"train_entropy": 0.34466766704653584,
|
| 1261 |
+
"train_ce_loss": 0.3667684007707974,
|
| 1262 |
+
"train_dice_loss": 0.4288115936502314,
|
| 1263 |
+
"train_dice": 0.8384048968619499,
|
| 1264 |
+
"train_iou": 0.7540147290688785,
|
| 1265 |
+
"grad_norm": 10.873141574305157,
|
| 1266 |
+
"lr": 0.0007879474748104406,
|
| 1267 |
+
"encoder_lr": 8.42763920673285e-05,
|
| 1268 |
+
"alpha": 0.23414088785648346,
|
| 1269 |
+
"validated_this_epoch": false,
|
| 1270 |
+
"val_loss": null,
|
| 1271 |
+
"val_dice": null,
|
| 1272 |
+
"val_iou": null,
|
| 1273 |
+
"val_reward": null,
|
| 1274 |
+
"val_entropy": null
|
| 1275 |
+
},
|
| 1276 |
+
{
|
| 1277 |
+
"epoch": 58,
|
| 1278 |
+
"train_loss": 0.30342912203957173,
|
| 1279 |
+
"train_actor_loss": -0.07662009767398663,
|
| 1280 |
+
"train_critic_loss": 0.00012699133806220504,
|
| 1281 |
+
"train_mean_reward": 0.00039101014844227183,
|
| 1282 |
+
"train_entropy": 0.3492025148691441,
|
| 1283 |
+
"train_ce_loss": 0.3667694697137296,
|
| 1284 |
+
"train_dice_loss": 0.39320196877462305,
|
| 1285 |
+
"train_dice": 0.8477996347051621,
|
| 1286 |
+
"train_iou": 0.7629780981745732,
|
| 1287 |
+
"grad_norm": 7.260664137065873,
|
| 1288 |
+
"lr": 0.0007575535622709696,
|
| 1289 |
+
"encoder_lr": 8.13744607193388e-05,
|
| 1290 |
+
"alpha": 0.23308241367340088,
|
| 1291 |
+
"validated_this_epoch": false,
|
| 1292 |
+
"val_loss": null,
|
| 1293 |
+
"val_dice": null,
|
| 1294 |
+
"val_iou": null,
|
| 1295 |
+
"val_reward": null,
|
| 1296 |
+
"val_entropy": null
|
| 1297 |
+
},
|
| 1298 |
+
{
|
| 1299 |
+
"epoch": 59,
|
| 1300 |
+
"train_loss": 0.33081871811156127,
|
| 1301 |
+
"train_actor_loss": -0.07658997324153358,
|
| 1302 |
+
"train_critic_loss": 0.0001299005210325571,
|
| 1303 |
+
"train_mean_reward": 0.0002709054761926993,
|
| 1304 |
+
"train_entropy": 0.3497036518360601,
|
| 1305 |
+
"train_ce_loss": 0.37487363598540485,
|
| 1306 |
+
"train_dice_loss": 0.43981383529128343,
|
| 1307 |
+
"train_dice": 0.8387148987247811,
|
| 1308 |
+
"train_iou": 0.7541987773036851,
|
| 1309 |
+
"grad_norm": 13.621012659497964,
|
| 1310 |
+
"lr": 0.0007274038494909672,
|
| 1311 |
+
"encoder_lr": 7.849584492627325e-05,
|
| 1312 |
+
"alpha": 0.23157989978790283,
|
| 1313 |
+
"validated_this_epoch": false,
|
| 1314 |
+
"val_loss": null,
|
| 1315 |
+
"val_dice": null,
|
| 1316 |
+
"val_iou": null,
|
| 1317 |
+
"val_reward": null,
|
| 1318 |
+
"val_entropy": null
|
| 1319 |
+
},
|
| 1320 |
+
{
|
| 1321 |
+
"epoch": 60,
|
| 1322 |
+
"train_loss": 0.30376051975659624,
|
| 1323 |
+
"train_actor_loss": -0.07472321491698115,
|
| 1324 |
+
"train_critic_loss": 0.00012578927674343558,
|
| 1325 |
+
"train_mean_reward": 0.0003138011309188386,
|
| 1326 |
+
"train_entropy": 0.3418057337210776,
|
| 1327 |
+
"train_ce_loss": 0.3591565573293215,
|
| 1328 |
+
"train_dice_loss": 0.3976851114160947,
|
| 1329 |
+
"train_dice": 0.8463827972728685,
|
| 1330 |
+
"train_iou": 0.7581544463489015,
|
| 1331 |
+
"grad_norm": 7.946312002433363,
|
| 1332 |
+
"lr": 0.0006975280905969278,
|
| 1333 |
+
"encoder_lr": 7.564338553438004e-05,
|
| 1334 |
+
"alpha": 0.23056729137897491,
|
| 1335 |
+
"validated_this_epoch": true,
|
| 1336 |
+
"val_loss": 0.7106220282854118,
|
| 1337 |
+
"val_dice": 0.7019209265288548,
|
| 1338 |
+
"val_iou": 0.6110423824257872,
|
| 1339 |
+
"val_reward": -0.00017641598912972528,
|
| 1340 |
+
"val_entropy": 0.4816596681120419,
|
| 1341 |
+
"val_actor_loss": -0.005822640944817903,
|
| 1342 |
+
"val_critic_loss": 6.893903112844174e-05,
|
| 1343 |
+
"val_ce_loss": 0.6329004717473355,
|
| 1344 |
+
"val_dice_loss": 0.7999199014721494
|
| 1345 |
+
},
|
| 1346 |
+
{
|
| 1347 |
+
"epoch": 61,
|
| 1348 |
+
"train_loss": 0.34061276014937564,
|
| 1349 |
+
"train_actor_loss": -0.07629987845336897,
|
| 1350 |
+
"train_critic_loss": 0.00011843174980494763,
|
| 1351 |
+
"train_mean_reward": 0.0003106296192095978,
|
| 1352 |
+
"train_entropy": 0.34895620497006163,
|
| 1353 |
+
"train_ce_loss": 0.3895268828730828,
|
| 1354 |
+
"train_dice_loss": 0.44417995045043396,
|
| 1355 |
+
"train_dice": 0.8333412182458753,
|
| 1356 |
+
"train_iou": 0.7483183806291988,
|
| 1357 |
+
"grad_norm": 9.332444371179093,
|
| 1358 |
+
"lr": 0.0006679557693559353,
|
| 1359 |
+
"encoder_lr": 7.281989757669733e-05,
|
| 1360 |
+
"alpha": 0.22940221428871155,
|
| 1361 |
+
"validated_this_epoch": false,
|
| 1362 |
+
"val_loss": null,
|
| 1363 |
+
"val_dice": null,
|
| 1364 |
+
"val_iou": null,
|
| 1365 |
+
"val_reward": null,
|
| 1366 |
+
"val_entropy": null
|
| 1367 |
+
},
|
| 1368 |
+
{
|
| 1369 |
+
"epoch": 62,
|
| 1370 |
+
"train_loss": 0.3187312413252486,
|
| 1371 |
+
"train_actor_loss": -0.07293276306366889,
|
| 1372 |
+
"train_critic_loss": 0.000127791242215777,
|
| 1373 |
+
"train_mean_reward": 0.0002824390816801828,
|
| 1374 |
+
"train_entropy": 0.33108046603718777,
|
| 1375 |
+
"train_ce_loss": 0.36541023498367353,
|
| 1376 |
+
"train_dice_loss": 0.41778997156663145,
|
| 1377 |
+
"train_dice": 0.831938644931682,
|
| 1378 |
+
"train_iou": 0.747280623564462,
|
| 1379 |
+
"grad_norm": 9.38257000243017,
|
| 1380 |
+
"lr": 0.000638716070078746,
|
| 1381 |
+
"encoder_lr": 7.002816749495564e-05,
|
| 1382 |
+
"alpha": 0.2293829321861267,
|
| 1383 |
+
"validated_this_epoch": false,
|
| 1384 |
+
"val_loss": null,
|
| 1385 |
+
"val_dice": null,
|
| 1386 |
+
"val_iou": null,
|
| 1387 |
+
"val_reward": null,
|
| 1388 |
+
"val_entropy": null
|
| 1389 |
+
},
|
| 1390 |
+
{
|
| 1391 |
+
"epoch": 63,
|
| 1392 |
+
"train_loss": 0.305091810849488,
|
| 1393 |
+
"train_actor_loss": -0.07417594880173681,
|
| 1394 |
+
"train_critic_loss": 0.00012104062561986718,
|
| 1395 |
+
"train_mean_reward": 0.0003602059929455598,
|
| 1396 |
+
"train_entropy": 0.33797220598896655,
|
| 1397 |
+
"train_ce_loss": 0.35662980153928064,
|
| 1398 |
+
"train_dice_loss": 0.4017846661328654,
|
| 1399 |
+
"train_dice": 0.845640668067181,
|
| 1400 |
+
"train_iou": 0.7627700658709685,
|
| 1401 |
+
"grad_norm": 7.5072468442972315,
|
| 1402 |
+
"lr": 0.0006098378488183934,
|
| 1403 |
+
"encoder_lr": 6.727095038969585e-05,
|
| 1404 |
+
"alpha": 0.2287590354681015,
|
| 1405 |
+
"validated_this_epoch": false,
|
| 1406 |
+
"val_loss": null,
|
| 1407 |
+
"val_dice": null,
|
| 1408 |
+
"val_iou": null,
|
| 1409 |
+
"val_reward": null,
|
| 1410 |
+
"val_entropy": null
|
| 1411 |
+
},
|
| 1412 |
+
{
|
| 1413 |
+
"epoch": 64,
|
| 1414 |
+
"train_loss": 0.30916663881954415,
|
| 1415 |
+
"train_actor_loss": -0.07265147338898068,
|
| 1416 |
+
"train_critic_loss": 0.00011829692118828006,
|
| 1417 |
+
"train_mean_reward": 0.00031331498586730145,
|
| 1418 |
+
"train_entropy": 0.33687925708386324,
|
| 1419 |
+
"train_ce_loss": 0.35709881396405885,
|
| 1420 |
+
"train_dice_loss": 0.4064191006874853,
|
| 1421 |
+
"train_dice": 0.84684892193631,
|
| 1422 |
+
"train_iou": 0.7594184435233344,
|
| 1423 |
+
"grad_norm": 7.820028717203658,
|
| 1424 |
+
"lr": 0.0005813496048927529,
|
| 1425 |
+
"encoder_lr": 6.455096730131812e-05,
|
| 1426 |
+
"alpha": 0.22815224528312683,
|
| 1427 |
+
"validated_this_epoch": false,
|
| 1428 |
+
"val_loss": null,
|
| 1429 |
+
"val_dice": null,
|
| 1430 |
+
"val_iou": null,
|
| 1431 |
+
"val_reward": null,
|
| 1432 |
+
"val_entropy": null
|
| 1433 |
+
},
|
| 1434 |
+
{
|
| 1435 |
+
"epoch": 65,
|
| 1436 |
+
"train_loss": 0.30473408729646556,
|
| 1437 |
+
"train_actor_loss": -0.07506646286667318,
|
| 1438 |
+
"train_critic_loss": 0.00012803353832213448,
|
| 1439 |
+
"train_mean_reward": 0.000337804655438103,
|
| 1440 |
+
"train_entropy": 0.3389589909743277,
|
| 1441 |
+
"train_ce_loss": 0.3583998876366312,
|
| 1442 |
+
"train_dice_loss": 0.4010731673363876,
|
| 1443 |
+
"train_dice": 0.8379877805545787,
|
| 1444 |
+
"train_iou": 0.7529568941959792,
|
| 1445 |
+
"grad_norm": 7.492030368294826,
|
| 1446 |
+
"lr": 0.0005532794527591512,
|
| 1447 |
+
"encoder_lr": 6.187090252474308e-05,
|
| 1448 |
+
"alpha": 0.22725585103034973,
|
| 1449 |
+
"validated_this_epoch": false,
|
| 1450 |
+
"val_loss": null,
|
| 1451 |
+
"val_dice": null,
|
| 1452 |
+
"val_iou": null,
|
| 1453 |
+
"val_reward": null,
|
| 1454 |
+
"val_entropy": null
|
| 1455 |
+
},
|
| 1456 |
+
{
|
| 1457 |
+
"epoch": 66,
|
| 1458 |
+
"train_loss": 0.29755060085954593,
|
| 1459 |
+
"train_actor_loss": -0.07396081913484005,
|
| 1460 |
+
"train_critic_loss": 0.00011773357540455517,
|
| 1461 |
+
"train_mean_reward": 0.00037845082483692056,
|
| 1462 |
+
"train_entropy": 0.33654534046656087,
|
| 1463 |
+
"train_ce_loss": 0.3550966944803004,
|
| 1464 |
+
"train_dice_loss": 0.3878084018402937,
|
| 1465 |
+
"train_dice": 0.8486660593986286,
|
| 1466 |
+
"train_iou": 0.7640734506027482,
|
| 1467 |
+
"grad_norm": 5.849660679813503,
|
| 1468 |
+
"lr": 0.0005256550942687934,
|
| 1469 |
+
"encoder_lr": 5.9233400960337056e-05,
|
| 1470 |
+
"alpha": 0.22692431509494781,
|
| 1471 |
+
"validated_this_epoch": false,
|
| 1472 |
+
"val_loss": null,
|
| 1473 |
+
"val_dice": null,
|
| 1474 |
+
"val_iou": null,
|
| 1475 |
+
"val_reward": null,
|
| 1476 |
+
"val_entropy": null
|
| 1477 |
+
},
|
| 1478 |
+
{
|
| 1479 |
+
"epoch": 67,
|
| 1480 |
+
"train_loss": 0.32860205865414566,
|
| 1481 |
+
"train_actor_loss": -0.07382364720539299,
|
| 1482 |
+
"train_critic_loss": 0.00013191550129123929,
|
| 1483 |
+
"train_mean_reward": 0.000363645237766315,
|
| 1484 |
+
"train_entropy": 0.3390335294896934,
|
| 1485 |
+
"train_ce_loss": 0.3690149215704615,
|
| 1486 |
+
"train_dice_loss": 0.4357045620140795,
|
| 1487 |
+
"train_dice": 0.8404538364642208,
|
| 1488 |
+
"train_iou": 0.7565638090073977,
|
| 1489 |
+
"grad_norm": 7.07320559071016,
|
| 1490 |
+
"lr": 0.0004985037913283807,
|
| 1491 |
+
"encoder_lr": 5.6641065503714745e-05,
|
| 1492 |
+
"alpha": 0.22625866532325745,
|
| 1493 |
+
"validated_this_epoch": false,
|
| 1494 |
+
"val_loss": null,
|
| 1495 |
+
"val_dice": null,
|
| 1496 |
+
"val_iou": null,
|
| 1497 |
+
"val_reward": null,
|
| 1498 |
+
"val_entropy": null
|
| 1499 |
+
},
|
| 1500 |
+
{
|
| 1501 |
+
"epoch": 68,
|
| 1502 |
+
"train_loss": 0.2919036108033024,
|
| 1503 |
+
"train_actor_loss": -0.07404205512053283,
|
| 1504 |
+
"train_critic_loss": 0.0001149739954109948,
|
| 1505 |
+
"train_mean_reward": 0.00033341147210154366,
|
| 1506 |
+
"train_entropy": 0.33607516686121625,
|
| 1507 |
+
"train_ce_loss": 0.3559206936385685,
|
| 1508 |
+
"train_dice_loss": 0.3758556544010645,
|
| 1509 |
+
"train_dice": 0.8515192324740475,
|
| 1510 |
+
"train_iou": 0.7692301099886145,
|
| 1511 |
+
"grad_norm": 5.008078430281129,
|
| 1512 |
+
"lr": 0.0004718523389958983,
|
| 1513 |
+
"encoder_lr": 5.4096454476995315e-05,
|
| 1514 |
+
"alpha": 0.22566978633403778,
|
| 1515 |
+
"validated_this_epoch": false,
|
| 1516 |
+
"val_loss": null,
|
| 1517 |
+
"val_dice": null,
|
| 1518 |
+
"val_iou": null,
|
| 1519 |
+
"val_reward": null,
|
| 1520 |
+
"val_entropy": null
|
| 1521 |
+
},
|
| 1522 |
+
{
|
| 1523 |
+
"epoch": 69,
|
| 1524 |
+
"train_loss": 0.30045714176642574,
|
| 1525 |
+
"train_actor_loss": -0.07472459624403516,
|
| 1526 |
+
"train_critic_loss": 0.00012446951021550352,
|
| 1527 |
+
"train_mean_reward": 0.00032924411503977276,
|
| 1528 |
+
"train_entropy": 0.341039406621795,
|
| 1529 |
+
"train_ce_loss": 0.35282586111617337,
|
| 1530 |
+
"train_dice_loss": 0.39741313549899315,
|
| 1531 |
+
"train_dice": 0.8511249706053692,
|
| 1532 |
+
"train_iou": 0.7681291154930762,
|
| 1533 |
+
"grad_norm": 5.12398703985436,
|
| 1534 |
+
"lr": 0.0004457270390371302,
|
| 1535 |
+
"encoder_lr": 5.1602079104047604e-05,
|
| 1536 |
+
"alpha": 0.22514598071575165,
|
| 1537 |
+
"validated_this_epoch": false,
|
| 1538 |
+
"val_loss": null,
|
| 1539 |
+
"val_dice": null,
|
| 1540 |
+
"val_iou": null,
|
| 1541 |
+
"val_reward": null,
|
| 1542 |
+
"val_entropy": null
|
| 1543 |
+
},
|
| 1544 |
+
{
|
| 1545 |
+
"epoch": 70,
|
| 1546 |
+
"train_loss": 0.3023134499405648,
|
| 1547 |
+
"train_actor_loss": -0.07259652735208202,
|
| 1548 |
+
"train_critic_loss": 0.00012518320061372158,
|
| 1549 |
+
"train_mean_reward": 0.0003855950714753439,
|
| 1550 |
+
"train_entropy": 0.3310071336784104,
|
| 1551 |
+
"train_ce_loss": 0.3490266974459496,
|
| 1552 |
+
"train_dice_loss": 0.4006680628279999,
|
| 1553 |
+
"train_dice": 0.8481722826609699,
|
| 1554 |
+
"train_iou": 0.7625283523995638,
|
| 1555 |
+
"grad_norm": 6.663285031568172,
|
| 1556 |
+
"lr": 0.00042015367396898955,
|
| 1557 |
+
"encoder_lr": 4.916040103221508e-05,
|
| 1558 |
+
"alpha": 0.22464090585708618,
|
| 1559 |
+
"validated_this_epoch": true,
|
| 1560 |
+
"val_loss": 0.7145497286268934,
|
| 1561 |
+
"val_dice": 0.6912404706453031,
|
| 1562 |
+
"val_iou": 0.6037589098436458,
|
| 1563 |
+
"val_reward": -0.00047079030568828313,
|
| 1564 |
+
"val_entropy": 0.49360059367285825,
|
| 1565 |
+
"val_actor_loss": -0.0104366628567199,
|
| 1566 |
+
"val_critic_loss": 0.00010624083522933953,
|
| 1567 |
+
"val_ce_loss": 0.6408483187451657,
|
| 1568 |
+
"val_dice_loss": 0.8090181994919824
|
| 1569 |
+
},
|
| 1570 |
+
{
|
| 1571 |
+
"epoch": 71,
|
| 1572 |
+
"train_loss": 0.30499391168962386,
|
| 1573 |
+
"train_actor_loss": -0.07412265131408854,
|
| 1574 |
+
"train_critic_loss": 0.00012848690242558416,
|
| 1575 |
+
"train_mean_reward": 0.0004997858340990789,
|
| 1576 |
+
"train_entropy": 0.3364787173941154,
|
| 1577 |
+
"train_ce_loss": 0.3572229222618332,
|
| 1578 |
+
"train_dice_loss": 0.4008817070522357,
|
| 1579 |
+
"train_dice": 0.8490629131403404,
|
| 1580 |
+
"train_iou": 0.7657471504419211,
|
| 1581 |
+
"grad_norm": 6.832022731156313,
|
| 1582 |
+
"lr": 0.0003951574816152885,
|
| 1583 |
+
"encoder_lr": 4.6773829902967244e-05,
|
| 1584 |
+
"alpha": 0.22418074309825897,
|
| 1585 |
+
"validated_this_epoch": false,
|
| 1586 |
+
"val_loss": null,
|
| 1587 |
+
"val_dice": null,
|
| 1588 |
+
"val_iou": null,
|
| 1589 |
+
"val_reward": null,
|
| 1590 |
+
"val_entropy": null
|
| 1591 |
+
},
|
| 1592 |
+
{
|
| 1593 |
+
"epoch": 72,
|
| 1594 |
+
"train_loss": 0.3118327355820888,
|
| 1595 |
+
"train_actor_loss": -0.07266073473741225,
|
| 1596 |
+
"train_critic_loss": 0.0001306005791715605,
|
| 1597 |
+
"train_mean_reward": 0.00031827906143996963,
|
| 1598 |
+
"train_entropy": 0.33669408222116554,
|
| 1599 |
+
"train_ce_loss": 0.36209461886317834,
|
| 1600 |
+
"train_dice_loss": 0.40676170964881736,
|
| 1601 |
+
"train_dice": 0.8424260610287432,
|
| 1602 |
+
"train_iou": 0.7600064773104147,
|
| 1603 |
+
"grad_norm": 9.566452346568884,
|
| 1604 |
+
"lr": 0.0003707631302000539,
|
| 1605 |
+
"encoder_lr": 4.444472097387449e-05,
|
| 1606 |
+
"alpha": 0.22372505068778992,
|
| 1607 |
+
"validated_this_epoch": false,
|
| 1608 |
+
"val_loss": null,
|
| 1609 |
+
"val_dice": null,
|
| 1610 |
+
"val_iou": null,
|
| 1611 |
+
"val_reward": null,
|
| 1612 |
+
"val_entropy": null
|
| 1613 |
+
},
|
| 1614 |
+
{
|
| 1615 |
+
"epoch": 73,
|
| 1616 |
+
"train_loss": 0.294036972062128,
|
| 1617 |
+
"train_actor_loss": -0.07193701197838445,
|
| 1618 |
+
"train_critic_loss": 0.00012495978486886113,
|
| 1619 |
+
"train_mean_reward": 0.0003658884601727009,
|
| 1620 |
+
"train_entropy": 0.33350245410080714,
|
| 1621 |
+
"train_ce_loss": 0.34289108710700417,
|
| 1622 |
+
"train_dice_loss": 0.3889319105049745,
|
| 1623 |
+
"train_dice": 0.8528327441438501,
|
| 1624 |
+
"train_iou": 0.7687465509533236,
|
| 1625 |
+
"grad_norm": 5.324581496937331,
|
| 1626 |
+
"lr": 0.00034699469400296674,
|
| 1627 |
+
"encoder_lr": 4.2175372794253104e-05,
|
| 1628 |
+
"alpha": 0.22345323860645294,
|
| 1629 |
+
"validated_this_epoch": false,
|
| 1630 |
+
"val_loss": null,
|
| 1631 |
+
"val_dice": null,
|
| 1632 |
+
"val_iou": null,
|
| 1633 |
+
"val_reward": null,
|
| 1634 |
+
"val_entropy": null
|
| 1635 |
+
},
|
| 1636 |
+
{
|
| 1637 |
+
"epoch": 74,
|
| 1638 |
+
"train_loss": 0.3021333998126924,
|
| 1639 |
+
"train_actor_loss": -0.07250877590585278,
|
| 1640 |
+
"train_critic_loss": 0.00012220305825629474,
|
| 1641 |
+
"train_mean_reward": 0.0004024200848777067,
|
| 1642 |
+
"train_entropy": 0.3336720071367207,
|
| 1643 |
+
"train_ce_loss": 0.34372401130397734,
|
| 1644 |
+
"train_dice_loss": 0.4054381247484715,
|
| 1645 |
+
"train_dice": 0.8509219215626117,
|
| 1646 |
+
"train_iou": 0.7671072159163831,
|
| 1647 |
+
"grad_norm": 6.59460845196894,
|
| 1648 |
+
"lr": 0.0003238756296009549,
|
| 1649 |
+
"encoder_lr": 3.996802493677459e-05,
|
| 1650 |
+
"alpha": 0.22300010919570923,
|
| 1651 |
+
"validated_this_epoch": false,
|
| 1652 |
+
"val_loss": null,
|
| 1653 |
+
"val_dice": null,
|
| 1654 |
+
"val_iou": null,
|
| 1655 |
+
"val_reward": null,
|
| 1656 |
+
"val_entropy": null
|
| 1657 |
+
},
|
| 1658 |
+
{
|
| 1659 |
+
"epoch": 75,
|
| 1660 |
+
"train_loss": 0.31672749468739064,
|
| 1661 |
+
"train_actor_loss": -0.0733969451387916,
|
| 1662 |
+
"train_critic_loss": 0.0001242369075876941,
|
| 1663 |
+
"train_mean_reward": 0.0004109462970107236,
|
| 1664 |
+
"train_entropy": 0.3357755857536651,
|
| 1665 |
+
"train_ce_loss": 0.34997719376124153,
|
| 1666 |
+
"train_dice_loss": 0.4301474389333749,
|
| 1667 |
+
"train_dice": 0.844357328736083,
|
| 1668 |
+
"train_iou": 0.7597672407808863,
|
| 1669 |
+
"grad_norm": 5.663388241862142,
|
| 1670 |
+
"lr": 0.0003014287527193855,
|
| 1671 |
+
"encoder_lr": 3.782485578727801e-05,
|
| 1672 |
+
"alpha": 0.2226247638463974,
|
| 1673 |
+
"validated_this_epoch": false,
|
| 1674 |
+
"val_loss": null,
|
| 1675 |
+
"val_dice": null,
|
| 1676 |
+
"val_iou": null,
|
| 1677 |
+
"val_reward": null,
|
| 1678 |
+
"val_entropy": null
|
| 1679 |
+
},
|
| 1680 |
+
{
|
| 1681 |
+
"epoch": 76,
|
| 1682 |
+
"train_loss": 0.2779201053646068,
|
| 1683 |
+
"train_actor_loss": -0.07282891482817494,
|
| 1684 |
+
"train_critic_loss": 0.00012327545328498813,
|
| 1685 |
+
"train_mean_reward": 0.00043906504535489176,
|
| 1686 |
+
"train_entropy": 0.3337099741018096,
|
| 1687 |
+
"train_ce_loss": 0.33426290947958365,
|
| 1688 |
+
"train_dice_loss": 0.36711184453286555,
|
| 1689 |
+
"train_dice": 0.854956048777491,
|
| 1690 |
+
"train_iou": 0.7703340781443433,
|
| 1691 |
+
"grad_norm": 5.9125312254872435,
|
| 1692 |
+
"lr": 0.0002796762157156959,
|
| 1693 |
+
"encoder_lr": 3.574798039496594e-05,
|
| 1694 |
+
"alpha": 0.2222006767988205,
|
| 1695 |
+
"validated_this_epoch": false,
|
| 1696 |
+
"val_loss": null,
|
| 1697 |
+
"val_dice": null,
|
| 1698 |
+
"val_iou": null,
|
| 1699 |
+
"val_reward": null,
|
| 1700 |
+
"val_entropy": null
|
| 1701 |
+
},
|
| 1702 |
+
{
|
| 1703 |
+
"epoch": 77,
|
| 1704 |
+
"train_loss": 0.3142224269009434,
|
| 1705 |
+
"train_actor_loss": -0.0722625620797931,
|
| 1706 |
+
"train_critic_loss": 0.00012830420399649711,
|
| 1707 |
+
"train_mean_reward": 0.0003481147887649091,
|
| 1708 |
+
"train_entropy": 0.33544559382113986,
|
| 1709 |
+
"train_ce_loss": 0.35628506827507483,
|
| 1710 |
+
"train_dice_loss": 0.4165565952928183,
|
| 1711 |
+
"train_dice": 0.8436234909006203,
|
| 1712 |
+
"train_iou": 0.759317984863462,
|
| 1713 |
+
"grad_norm": 6.832556002592856,
|
| 1714 |
+
"lr": 0.00025863948571769286,
|
| 1715 |
+
"encoder_lr": 3.373944838510636e-05,
|
| 1716 |
+
"alpha": 0.22175610065460205,
|
| 1717 |
+
"validated_this_epoch": false,
|
| 1718 |
+
"val_loss": null,
|
| 1719 |
+
"val_dice": null,
|
| 1720 |
+
"val_iou": null,
|
| 1721 |
+
"val_reward": null,
|
| 1722 |
+
"val_entropy": null
|
| 1723 |
+
},
|
| 1724 |
+
{
|
| 1725 |
+
"epoch": 78,
|
| 1726 |
+
"train_loss": 0.2987207806337076,
|
| 1727 |
+
"train_actor_loss": -0.07171638469026456,
|
| 1728 |
+
"train_critic_loss": 0.000122379408974041,
|
| 1729 |
+
"train_mean_reward": 0.0003365349849206208,
|
| 1730 |
+
"train_entropy": 0.33023583592524514,
|
| 1731 |
+
"train_ce_loss": 0.3456724862055492,
|
| 1732 |
+
"train_dice_loss": 0.3950794541712571,
|
| 1733 |
+
"train_dice": 0.8446218460940892,
|
| 1734 |
+
"train_iou": 0.7632167434505799,
|
| 1735 |
+
"grad_norm": 7.457615356112635,
|
| 1736 |
+
"lr": 0.00023833932343808993,
|
| 1737 |
+
"encoder_lr": 3.180124193630005e-05,
|
| 1738 |
+
"alpha": 0.22160348296165466,
|
| 1739 |
+
"validated_this_epoch": false,
|
| 1740 |
+
"val_loss": null,
|
| 1741 |
+
"val_dice": null,
|
| 1742 |
+
"val_iou": null,
|
| 1743 |
+
"val_reward": null,
|
| 1744 |
+
"val_entropy": null
|
| 1745 |
+
},
|
| 1746 |
+
{
|
| 1747 |
+
"epoch": 79,
|
| 1748 |
+
"train_loss": 0.3011552453033857,
|
| 1749 |
+
"train_actor_loss": -0.07330376887814327,
|
| 1750 |
+
"train_critic_loss": 0.00011982823424969942,
|
| 1751 |
+
"train_mean_reward": 0.00047051392455153785,
|
| 1752 |
+
"train_entropy": 0.3302778948377577,
|
| 1753 |
+
"train_ce_loss": 0.3443866588695105,
|
| 1754 |
+
"train_dice_loss": 0.40441152961679205,
|
| 1755 |
+
"train_dice": 0.8504072967197733,
|
| 1756 |
+
"train_iou": 0.7653278481360787,
|
| 1757 |
+
"grad_norm": 4.6133805575758915,
|
| 1758 |
+
"lr": 0.00021879576268618817,
|
| 1759 |
+
"encoder_lr": 2.993527382430943e-05,
|
| 1760 |
+
"alpha": 0.2214362621307373,
|
| 1761 |
+
"validated_this_epoch": false,
|
| 1762 |
+
"val_loss": null,
|
| 1763 |
+
"val_dice": null,
|
| 1764 |
+
"val_iou": null,
|
| 1765 |
+
"val_reward": null,
|
| 1766 |
+
"val_entropy": null
|
| 1767 |
+
},
|
| 1768 |
+
{
|
| 1769 |
+
"epoch": 80,
|
| 1770 |
+
"train_loss": 0.2905335736410447,
|
| 1771 |
+
"train_actor_loss": -0.07300525020509574,
|
| 1772 |
+
"train_critic_loss": 0.00011988496681831194,
|
| 1773 |
+
"train_mean_reward": 0.0004030720341646418,
|
| 1774 |
+
"train_entropy": 0.33324251801391597,
|
| 1775 |
+
"train_ce_loss": 0.3442871042518836,
|
| 1776 |
+
"train_dice_loss": 0.3826706478762072,
|
| 1777 |
+
"train_dice": 0.8532768000601139,
|
| 1778 |
+
"train_iou": 0.7697927085745738,
|
| 1779 |
+
"grad_norm": 5.163090821615485,
|
| 1780 |
+
"lr": 0.0002000280905969275,
|
| 1781 |
+
"encoder_lr": 2.814338553438002e-05,
|
| 1782 |
+
"alpha": 0.22123309969902039,
|
| 1783 |
+
"validated_this_epoch": true,
|
| 1784 |
+
"val_loss": 0.6838661219614247,
|
| 1785 |
+
"val_dice": 0.7104630094377155,
|
| 1786 |
+
"val_iou": 0.6215935392087729,
|
| 1787 |
+
"val_reward": -0.000355479123313809,
|
| 1788 |
+
"val_entropy": 0.493805588947402,
|
| 1789 |
+
"val_actor_loss": -0.008708937118897647,
|
| 1790 |
+
"val_critic_loss": 9.855451011242972e-05,
|
| 1791 |
+
"val_ce_loss": 0.6103711776868112,
|
| 1792 |
+
"val_dice_loss": 0.7746803600980777
|
| 1793 |
+
},
|
| 1794 |
+
{
|
| 1795 |
+
"epoch": 81,
|
| 1796 |
+
"train_loss": 0.30171160252273893,
|
| 1797 |
+
"train_actor_loss": -0.07213261174832204,
|
| 1798 |
+
"train_critic_loss": 0.00011506211908935908,
|
| 1799 |
+
"train_mean_reward": 0.0003728851701516549,
|
| 1800 |
+
"train_entropy": 0.330792365552381,
|
| 1801 |
+
"train_ce_loss": 0.3559774713631652,
|
| 1802 |
+
"train_dice_loss": 0.39159588458002076,
|
| 1803 |
+
"train_dice": 0.85154616867088,
|
| 1804 |
+
"train_iou": 0.7701923551888791,
|
| 1805 |
+
"grad_norm": 3.7083017994729124,
|
| 1806 |
+
"lr": 0.00018205482859681116,
|
| 1807 |
+
"encoder_lr": 2.6427345443916652e-05,
|
| 1808 |
+
"alpha": 0.22108666598796844,
|
| 1809 |
+
"validated_this_epoch": false,
|
| 1810 |
+
"val_loss": null,
|
| 1811 |
+
"val_dice": null,
|
| 1812 |
+
"val_iou": null,
|
| 1813 |
+
"val_reward": null,
|
| 1814 |
+
"val_entropy": null
|
| 1815 |
+
},
|
| 1816 |
+
{
|
| 1817 |
+
"epoch": 82,
|
| 1818 |
+
"train_loss": 0.30057397876801184,
|
| 1819 |
+
"train_actor_loss": -0.07186304960274727,
|
| 1820 |
+
"train_critic_loss": 0.00011579467679300747,
|
| 1821 |
+
"train_mean_reward": 0.00038280587102425945,
|
| 1822 |
+
"train_entropy": 0.33068286279067205,
|
| 1823 |
+
"train_ce_loss": 0.35046436714317686,
|
| 1824 |
+
"train_dice_loss": 0.3942938827421006,
|
| 1825 |
+
"train_dice": 0.8485180186160597,
|
| 1826 |
+
"train_iou": 0.7650724621037117,
|
| 1827 |
+
"grad_norm": 3.922348376161368,
|
| 1828 |
+
"lr": 0.00016489371412549528,
|
| 1829 |
+
"encoder_lr": 2.4788847077308606e-05,
|
| 1830 |
+
"alpha": 0.22114522755146027,
|
| 1831 |
+
"validated_this_epoch": false,
|
| 1832 |
+
"val_loss": null,
|
| 1833 |
+
"val_dice": null,
|
| 1834 |
+
"val_iou": null,
|
| 1835 |
+
"val_reward": null,
|
| 1836 |
+
"val_entropy": null
|
| 1837 |
+
},
|
| 1838 |
+
{
|
| 1839 |
+
"epoch": 83,
|
| 1840 |
+
"train_loss": 0.29423536059930344,
|
| 1841 |
+
"train_actor_loss": -0.07249550441477343,
|
| 1842 |
+
"train_critic_loss": 0.00011824716955083116,
|
| 1843 |
+
"train_mean_reward": 0.0004294131553773371,
|
| 1844 |
+
"train_entropy": 0.32967492506793605,
|
| 1845 |
+
"train_ce_loss": 0.34995244776478723,
|
| 1846 |
+
"train_dice_loss": 0.3833910249924475,
|
| 1847 |
+
"train_dice": 0.8522250481106907,
|
| 1848 |
+
"train_iou": 0.7695109080997942,
|
| 1849 |
+
"grad_norm": 4.646700745405153,
|
| 1850 |
+
"lr": 0.00014856168313107603,
|
| 1851 |
+
"encoder_lr": 2.322950743462536e-05,
|
| 1852 |
+
"alpha": 0.22111012041568756,
|
| 1853 |
+
"validated_this_epoch": false,
|
| 1854 |
+
"val_loss": null,
|
| 1855 |
+
"val_dice": null,
|
| 1856 |
+
"val_iou": null,
|
| 1857 |
+
"val_reward": null,
|
| 1858 |
+
"val_entropy": null
|
| 1859 |
+
},
|
| 1860 |
+
{
|
| 1861 |
+
"epoch": 84,
|
| 1862 |
+
"train_loss": 0.3042767142061173,
|
| 1863 |
+
"train_actor_loss": -0.07132348046669355,
|
| 1864 |
+
"train_critic_loss": 0.0001261125137878748,
|
| 1865 |
+
"train_mean_reward": 0.0003745551838882789,
|
| 1866 |
+
"train_entropy": 0.33019461825451496,
|
| 1867 |
+
"train_ce_loss": 0.35154136319497953,
|
| 1868 |
+
"train_dice_loss": 0.399532902687403,
|
| 1869 |
+
"train_dice": 0.8446554858571187,
|
| 1870 |
+
"train_iou": 0.7612405642045641,
|
| 1871 |
+
"grad_norm": 6.631763223760812,
|
| 1872 |
+
"lr": 0.0001330748533563558,
|
| 1873 |
+
"encoder_lr": 2.1750865395832973e-05,
|
| 1874 |
+
"alpha": 0.22095707058906555,
|
| 1875 |
+
"validated_this_epoch": false,
|
| 1876 |
+
"val_loss": null,
|
| 1877 |
+
"val_dice": null,
|
| 1878 |
+
"val_iou": null,
|
| 1879 |
+
"val_reward": null,
|
| 1880 |
+
"val_entropy": null
|
| 1881 |
+
},
|
| 1882 |
+
{
|
| 1883 |
+
"epoch": 85,
|
| 1884 |
+
"train_loss": 0.28466739658944945,
|
| 1885 |
+
"train_actor_loss": -0.0730642176663999,
|
| 1886 |
+
"train_critic_loss": 0.00011755592657935676,
|
| 1887 |
+
"train_mean_reward": 0.0003950850994181727,
|
| 1888 |
+
"train_entropy": 0.33199412156214086,
|
| 1889 |
+
"train_ce_loss": 0.3458855807112208,
|
| 1890 |
+
"train_dice_loss": 0.369460081792309,
|
| 1891 |
+
"train_dice": 0.8575569248690216,
|
| 1892 |
+
"train_iou": 0.7752594128349084,
|
| 1893 |
+
"grad_norm": 3.146349541900694,
|
| 1894 |
+
"lr": 0.0001184485084325741,
|
| 1895 |
+
"encoder_lr": 2.035438020210507e-05,
|
| 1896 |
+
"alpha": 0.22062376141548157,
|
| 1897 |
+
"validated_this_epoch": false,
|
| 1898 |
+
"val_loss": null,
|
| 1899 |
+
"val_dice": null,
|
| 1900 |
+
"val_iou": null,
|
| 1901 |
+
"val_reward": null,
|
| 1902 |
+
"val_entropy": null
|
| 1903 |
+
},
|
| 1904 |
+
{
|
| 1905 |
+
"epoch": 86,
|
| 1906 |
+
"train_loss": 0.30655952215613436,
|
| 1907 |
+
"train_actor_loss": -0.0723763510815857,
|
| 1908 |
+
"train_critic_loss": 0.00012025530802868904,
|
| 1909 |
+
"train_mean_reward": 0.0004013156939382105,
|
| 1910 |
+
"train_entropy": 0.3311550373100803,
|
| 1911 |
+
"train_ce_loss": 0.3596116368892273,
|
| 1912 |
+
"train_dice_loss": 0.39813984426109067,
|
| 1913 |
+
"train_dice": 0.8516724470429881,
|
| 1914 |
+
"train_iou": 0.7716768695522656,
|
| 1915 |
+
"grad_norm": 4.173769120098084,
|
| 1916 |
+
"lr": 0.00010469708279631079,
|
| 1917 |
+
"encoder_lr": 1.904143001572817e-05,
|
| 1918 |
+
"alpha": 0.22068528831005096,
|
| 1919 |
+
"validated_this_epoch": false,
|
| 1920 |
+
"val_loss": null,
|
| 1921 |
+
"val_dice": null,
|
| 1922 |
+
"val_iou": null,
|
| 1923 |
+
"val_reward": null,
|
| 1924 |
+
"val_entropy": null
|
| 1925 |
+
},
|
| 1926 |
+
{
|
| 1927 |
+
"epoch": 87,
|
| 1928 |
+
"train_loss": 0.2830173387444874,
|
| 1929 |
+
"train_actor_loss": -0.0724798035916201,
|
| 1930 |
+
"train_critic_loss": 0.00011784090247660606,
|
| 1931 |
+
"train_mean_reward": 0.0004453247080088943,
|
| 1932 |
+
"train_entropy": 0.32887790019043966,
|
| 1933 |
+
"train_ce_loss": 0.3388439235133083,
|
| 1934 |
+
"train_dice_loss": 0.37203251129589027,
|
| 1935 |
+
"train_dice": 0.8542728868752464,
|
| 1936 |
+
"train_iou": 0.7733409025539238,
|
| 1937 |
+
"grad_norm": 3.44946292273758,
|
| 1938 |
+
"lr": 9.183414744443903e-05,
|
| 1939 |
+
"encoder_lr": 1.781331056002182e-05,
|
| 1940 |
+
"alpha": 0.22063101828098297,
|
| 1941 |
+
"validated_this_epoch": false,
|
| 1942 |
+
"val_loss": null,
|
| 1943 |
+
"val_dice": null,
|
| 1944 |
+
"val_iou": null,
|
| 1945 |
+
"val_reward": null,
|
| 1946 |
+
"val_entropy": null
|
| 1947 |
+
},
|
| 1948 |
+
{
|
| 1949 |
+
"epoch": 88,
|
| 1950 |
+
"train_loss": 0.28595402206532483,
|
| 1951 |
+
"train_actor_loss": -0.07255098667929494,
|
| 1952 |
+
"train_critic_loss": 0.00012356264467784178,
|
| 1953 |
+
"train_mean_reward": 0.0003911677741452689,
|
| 1954 |
+
"train_entropy": 0.3309807311887889,
|
| 1955 |
+
"train_ce_loss": 0.3456270301513595,
|
| 1956 |
+
"train_dice_loss": 0.37125941317826894,
|
| 1957 |
+
"train_dice": 0.8545613734738329,
|
| 1958 |
+
"train_iou": 0.7723519333668014,
|
| 1959 |
+
"grad_norm": 3.209707435249358,
|
| 1960 |
+
"lr": 7.987239654118994e-05,
|
| 1961 |
+
"encoder_lr": 1.6671233840616128e-05,
|
| 1962 |
+
"alpha": 0.2207200974225998,
|
| 1963 |
+
"validated_this_epoch": false,
|
| 1964 |
+
"val_loss": null,
|
| 1965 |
+
"val_dice": null,
|
| 1966 |
+
"val_iou": null,
|
| 1967 |
+
"val_reward": null,
|
| 1968 |
+
"val_entropy": null
|
| 1969 |
+
},
|
| 1970 |
+
{
|
| 1971 |
+
"epoch": 89,
|
| 1972 |
+
"train_loss": 0.30518444239833326,
|
| 1973 |
+
"train_actor_loss": -0.07393080627098102,
|
| 1974 |
+
"train_critic_loss": 0.00012344244997663291,
|
| 1975 |
+
"train_mean_reward": 0.0004168651600596226,
|
| 1976 |
+
"train_entropy": 0.33114878032404627,
|
| 1977 |
+
"train_ce_loss": 0.35808215933690657,
|
| 1978 |
+
"train_dice_loss": 0.4000248840458942,
|
| 1979 |
+
"train_dice": 0.846365768460868,
|
| 1980 |
+
"train_iou": 0.7643701411667106,
|
| 1981 |
+
"grad_norm": 4.160948542661445,
|
| 1982 |
+
"lr": 6.88236348905457e-05,
|
| 1983 |
+
"encoder_lr": 1.5616326949348587e-05,
|
| 1984 |
+
"alpha": 0.22024644911289215,
|
| 1985 |
+
"validated_this_epoch": false,
|
| 1986 |
+
"val_loss": null,
|
| 1987 |
+
"val_dice": null,
|
| 1988 |
+
"val_iou": null,
|
| 1989 |
+
"val_reward": null,
|
| 1990 |
+
"val_entropy": null
|
| 1991 |
+
},
|
| 1992 |
+
{
|
| 1993 |
+
"epoch": 90,
|
| 1994 |
+
"train_loss": 0.29456700422880494,
|
| 1995 |
+
"train_actor_loss": -0.07267641179379104,
|
| 1996 |
+
"train_critic_loss": 0.00012243569635380996,
|
| 1997 |
+
"train_mean_reward": 0.00043230136513724087,
|
| 1998 |
+
"train_entropy": 0.3302888144433344,
|
| 1999 |
+
"train_ce_loss": 0.3388812345120507,
|
| 2000 |
+
"train_dice_loss": 0.39548314980758253,
|
| 2001 |
+
"train_dice": 0.8561749730699215,
|
| 2002 |
+
"train_iou": 0.7730351031584265,
|
| 2003 |
+
"grad_norm": 3.4153452765109926,
|
| 2004 |
+
"lr": 5.869876628632226e-05,
|
| 2005 |
+
"encoder_lr": 1.4649630951960419e-05,
|
| 2006 |
+
"alpha": 0.2204536348581314,
|
| 2007 |
+
"validated_this_epoch": true,
|
| 2008 |
+
"val_loss": 0.6775146138247554,
|
| 2009 |
+
"val_dice": 0.706986393743568,
|
| 2010 |
+
"val_iou": 0.6163869310813893,
|
| 2011 |
+
"val_reward": -0.0004311679855953741,
|
| 2012 |
+
"val_entropy": 0.4883407793863855,
|
| 2013 |
+
"val_actor_loss": -0.009531207823058156,
|
| 2014 |
+
"val_critic_loss": 0.00011411243016105671,
|
| 2015 |
+
"val_ce_loss": 0.5846136145039715,
|
| 2016 |
+
"val_dice_loss": 0.7893638902842397
|
| 2017 |
+
},
|
| 2018 |
+
{
|
| 2019 |
+
"epoch": 91,
|
| 2020 |
+
"train_loss": 0.29661607545215724,
|
| 2021 |
+
"train_actor_loss": -0.07188104666788757,
|
| 2022 |
+
"train_critic_loss": 0.00012089872763072662,
|
| 2023 |
+
"train_mean_reward": 0.0003427809425711084,
|
| 2024 |
+
"train_entropy": 0.33183658467629773,
|
| 2025 |
+
"train_ce_loss": 0.3505657425815678,
|
| 2026 |
+
"train_dice_loss": 0.3863075930018758,
|
| 2027 |
+
"train_dice": 0.8520865877334147,
|
| 2028 |
+
"train_iou": 0.7703478072208753,
|
| 2029 |
+
"grad_norm": 4.115103510923164,
|
| 2030 |
+
"lr": 4.9507782751441774e-05,
|
| 2031 |
+
"encoder_lr": 1.3772099860690422e-05,
|
| 2032 |
+
"alpha": 0.22023090720176697,
|
| 2033 |
+
"validated_this_epoch": false,
|
| 2034 |
+
"val_loss": null,
|
| 2035 |
+
"val_dice": null,
|
| 2036 |
+
"val_iou": null,
|
| 2037 |
+
"val_reward": null,
|
| 2038 |
+
"val_entropy": null
|
| 2039 |
+
},
|
| 2040 |
+
{
|
| 2041 |
+
"epoch": 92,
|
| 2042 |
+
"train_loss": 0.2919624967233444,
|
| 2043 |
+
"train_actor_loss": -0.07301800076678895,
|
| 2044 |
+
"train_critic_loss": 0.00012191792358230404,
|
| 2045 |
+
"train_mean_reward": 0.0003810117018380553,
|
| 2046 |
+
"train_entropy": 0.32967362144735746,
|
| 2047 |
+
"train_ce_loss": 0.34240928628946377,
|
| 2048 |
+
"train_dice_loss": 0.38742978027624675,
|
| 2049 |
+
"train_dice": 0.8549640766549117,
|
| 2050 |
+
"train_iou": 0.7713086833325667,
|
| 2051 |
+
"grad_norm": 3.6402773584506307,
|
| 2052 |
+
"lr": 4.12597546770122e-05,
|
| 2053 |
+
"encoder_lr": 1.298459969278006e-05,
|
| 2054 |
+
"alpha": 0.22014546394348145,
|
| 2055 |
+
"validated_this_epoch": false,
|
| 2056 |
+
"val_loss": null,
|
| 2057 |
+
"val_dice": null,
|
| 2058 |
+
"val_iou": null,
|
| 2059 |
+
"val_reward": null,
|
| 2060 |
+
"val_entropy": null
|
| 2061 |
+
},
|
| 2062 |
+
{
|
| 2063 |
+
"epoch": 93,
|
| 2064 |
+
"train_loss": 0.29143034370774906,
|
| 2065 |
+
"train_actor_loss": -0.07214952321862682,
|
| 2066 |
+
"train_critic_loss": 0.00012014422068663293,
|
| 2067 |
+
"train_mean_reward": 0.00043519376021432845,
|
| 2068 |
+
"train_entropy": 0.32858847749710696,
|
| 2069 |
+
"train_ce_loss": 0.3387124215521024,
|
| 2070 |
+
"train_dice_loss": 0.38832715631147374,
|
| 2071 |
+
"train_dice": 0.8596259128347951,
|
| 2072 |
+
"train_iou": 0.7769101457383847,
|
| 2073 |
+
"grad_norm": 3.2182745358278586,
|
| 2074 |
+
"lr": 3.396282187094642e-05,
|
| 2075 |
+
"encoder_lr": 1.2287907615819006e-05,
|
| 2076 |
+
"alpha": 0.2203262895345688,
|
| 2077 |
+
"validated_this_epoch": false,
|
| 2078 |
+
"val_loss": null,
|
| 2079 |
+
"val_dice": null,
|
| 2080 |
+
"val_iou": null,
|
| 2081 |
+
"val_reward": null,
|
| 2082 |
+
"val_entropy": null
|
| 2083 |
+
},
|
| 2084 |
+
{
|
| 2085 |
+
"epoch": 94,
|
| 2086 |
+
"train_loss": 0.2755323467530714,
|
| 2087 |
+
"train_actor_loss": -0.07318119826974485,
|
| 2088 |
+
"train_critic_loss": 0.00012057591710170064,
|
| 2089 |
+
"train_mean_reward": 0.00036238677654835873,
|
| 2090 |
+
"train_entropy": 0.3320165549522839,
|
| 2091 |
+
"train_ce_loss": 0.3335853564800834,
|
| 2092 |
+
"train_dice_loss": 0.3637211469870822,
|
| 2093 |
+
"train_dice": 0.8579416991629952,
|
| 2094 |
+
"train_iou": 0.7767182181978003,
|
| 2095 |
+
"grad_norm": 3.4153079307356546,
|
| 2096 |
+
"lr": 2.762418552495473e-05,
|
| 2097 |
+
"encoder_lr": 1.1682711180774574e-05,
|
| 2098 |
+
"alpha": 0.22009976208209991,
|
| 2099 |
+
"validated_this_epoch": false,
|
| 2100 |
+
"val_loss": null,
|
| 2101 |
+
"val_dice": null,
|
| 2102 |
+
"val_iou": null,
|
| 2103 |
+
"val_reward": null,
|
| 2104 |
+
"val_entropy": null
|
| 2105 |
+
},
|
| 2106 |
+
{
|
| 2107 |
+
"epoch": 95,
|
| 2108 |
+
"train_loss": 0.2836655908329196,
|
| 2109 |
+
"train_actor_loss": -0.07266705442768659,
|
| 2110 |
+
"train_critic_loss": 0.00013612801661449424,
|
| 2111 |
+
"train_mean_reward": 0.0004221114861508186,
|
| 2112 |
+
"train_entropy": 0.3299900178414907,
|
| 2113 |
+
"train_ce_loss": 0.333192195198246,
|
| 2114 |
+
"train_dice_loss": 0.3793369562434904,
|
| 2115 |
+
"train_dice": 0.8606499562006772,
|
| 2116 |
+
"train_iou": 0.7791500019013545,
|
| 2117 |
+
"grad_norm": 3.5880267021729964,
|
| 2118 |
+
"lr": 2.2250101107838036e-05,
|
| 2119 |
+
"encoder_lr": 1.1169607643461924e-05,
|
| 2120 |
+
"alpha": 0.22008110582828522,
|
| 2121 |
+
"validated_this_epoch": false,
|
| 2122 |
+
"val_loss": null,
|
| 2123 |
+
"val_dice": null,
|
| 2124 |
+
"val_iou": null,
|
| 2125 |
+
"val_reward": null,
|
| 2126 |
+
"val_entropy": null
|
| 2127 |
+
},
|
| 2128 |
+
{
|
| 2129 |
+
"epoch": 96,
|
| 2130 |
+
"train_loss": 0.2804659482402106,
|
| 2131 |
+
"train_actor_loss": -0.07287914884933668,
|
| 2132 |
+
"train_critic_loss": 0.00011778175776471141,
|
| 2133 |
+
"train_mean_reward": 0.0003677834477725422,
|
| 2134 |
+
"train_entropy": 0.33113240997381604,
|
| 2135 |
+
"train_ce_loss": 0.3359914323855288,
|
| 2136 |
+
"train_dice_loss": 0.3705809710869802,
|
| 2137 |
+
"train_dice": 0.8590769734068892,
|
| 2138 |
+
"train_iou": 0.7782296418034993,
|
| 2139 |
+
"grad_norm": 2.525416620479998,
|
| 2140 |
+
"lr": 1.784587219209463e-05,
|
| 2141 |
+
"encoder_lr": 1.0749103375124614e-05,
|
| 2142 |
+
"alpha": 0.21981731057167053,
|
| 2143 |
+
"validated_this_epoch": false,
|
| 2144 |
+
"val_loss": null,
|
| 2145 |
+
"val_dice": null,
|
| 2146 |
+
"val_iou": null,
|
| 2147 |
+
"val_reward": null,
|
| 2148 |
+
"val_entropy": null
|
| 2149 |
+
},
|
| 2150 |
+
{
|
| 2151 |
+
"epoch": 97,
|
| 2152 |
+
"train_loss": 0.28411368406630416,
|
| 2153 |
+
"train_actor_loss": -0.07260279489310521,
|
| 2154 |
+
"train_critic_loss": 0.00012272602945334135,
|
| 2155 |
+
"train_mean_reward": 0.00037436771377585056,
|
| 2156 |
+
"train_entropy": 0.32934633091823695,
|
| 2157 |
+
"train_ce_loss": 0.3374680800300852,
|
| 2158 |
+
"train_dice_loss": 0.37584214085756346,
|
| 2159 |
+
"train_dice": 0.861275109495895,
|
| 2160 |
+
"train_iou": 0.7795144575080115,
|
| 2161 |
+
"grad_norm": 3.490209317946619,
|
| 2162 |
+
"lr": 1.4415845219935406e-05,
|
| 2163 |
+
"encoder_lr": 1.0421613362707402e-05,
|
| 2164 |
+
"alpha": 0.2197108268737793,
|
| 2165 |
+
"validated_this_epoch": false,
|
| 2166 |
+
"val_loss": null,
|
| 2167 |
+
"val_dice": null,
|
| 2168 |
+
"val_iou": null,
|
| 2169 |
+
"val_reward": null,
|
| 2170 |
+
"val_entropy": null
|
| 2171 |
+
},
|
| 2172 |
+
{
|
| 2173 |
+
"epoch": 98,
|
| 2174 |
+
"train_loss": 0.2874124547639905,
|
| 2175 |
+
"train_actor_loss": -0.07238130806644448,
|
| 2176 |
+
"train_critic_loss": 0.00011691069013592119,
|
| 2177 |
+
"train_mean_reward": 0.0004247713193020411,
|
| 2178 |
+
"train_entropy": 0.32734987311045943,
|
| 2179 |
+
"train_ce_loss": 0.34002001251717334,
|
| 2180 |
+
"train_dice_loss": 0.37945059157465166,
|
| 2181 |
+
"train_dice": 0.8557656035842773,
|
| 2182 |
+
"train_iou": 0.7772688591842246,
|
| 2183 |
+
"grad_norm": 2.909817350465198,
|
| 2184 |
+
"lr": 1.19634052138698e-05,
|
| 2185 |
+
"encoder_lr": 1.0187460799314204e-05,
|
| 2186 |
+
"alpha": 0.21998971700668335,
|
| 2187 |
+
"validated_this_epoch": false,
|
| 2188 |
+
"val_loss": null,
|
| 2189 |
+
"val_dice": null,
|
| 2190 |
+
"val_iou": null,
|
| 2191 |
+
"val_reward": null,
|
| 2192 |
+
"val_entropy": null
|
| 2193 |
+
},
|
| 2194 |
+
{
|
| 2195 |
+
"epoch": 99,
|
| 2196 |
+
"train_loss": 0.2879619414362219,
|
| 2197 |
+
"train_actor_loss": -0.07192628221822185,
|
| 2198 |
+
"train_critic_loss": 0.00011812157636932381,
|
| 2199 |
+
"train_mean_reward": 0.0003382137320959819,
|
| 2200 |
+
"train_entropy": 0.3282493023678314,
|
| 2201 |
+
"train_ce_loss": 0.3379050381974984,
|
| 2202 |
+
"train_dice_loss": 0.38175327733197567,
|
| 2203 |
+
"train_dice": 0.8539977130043642,
|
| 2204 |
+
"train_iou": 0.7728905618656741,
|
| 2205 |
+
"grad_norm": 3.7788753990055053,
|
| 2206 |
+
"lr": 1.049097243609706e-05,
|
| 2207 |
+
"encoder_lr": 1.0046876765255499e-05,
|
| 2208 |
+
"alpha": 0.22014373540878296,
|
| 2209 |
+
"validated_this_epoch": false,
|
| 2210 |
+
"val_loss": null,
|
| 2211 |
+
"val_dice": null,
|
| 2212 |
+
"val_iou": null,
|
| 2213 |
+
"val_reward": null,
|
| 2214 |
+
"val_entropy": null
|
| 2215 |
+
},
|
| 2216 |
+
{
|
| 2217 |
+
"epoch": 100,
|
| 2218 |
+
"train_loss": 0.293190891092802,
|
| 2219 |
+
"train_actor_loss": -0.07237436432266267,
|
| 2220 |
+
"train_critic_loss": 0.00012451195190267997,
|
| 2221 |
+
"train_mean_reward": 0.00042688932931613557,
|
| 2222 |
+
"train_entropy": 0.3290513631639863,
|
| 2223 |
+
"train_ce_loss": 0.34630919320241715,
|
| 2224 |
+
"train_dice_loss": 0.3846967955738383,
|
| 2225 |
+
"train_dice": 0.8555878325024208,
|
| 2226 |
+
"train_iou": 0.7732638274599075,
|
| 2227 |
+
"grad_norm": 3.228855545437613,
|
| 2228 |
+
"lr": 1e-05,
|
| 2229 |
+
"encoder_lr": 1e-05,
|
| 2230 |
+
"alpha": 0.22020697593688965,
|
| 2231 |
+
"validated_this_epoch": true,
|
| 2232 |
+
"val_loss": 0.674074559274948,
|
| 2233 |
+
"val_dice": 0.7063104816199213,
|
| 2234 |
+
"val_iou": 0.6163285774561079,
|
| 2235 |
+
"val_reward": -0.0004757653101900948,
|
| 2236 |
+
"val_entropy": 0.4805109150180913,
|
| 2237 |
+
"val_actor_loss": -0.008498934451537678,
|
| 2238 |
+
"val_critic_loss": 0.00010437710152289563,
|
| 2239 |
+
"val_ce_loss": 0.593108140985773,
|
| 2240 |
+
"val_dice_loss": 0.7719344433509943
|
| 2241 |
+
}
|
| 2242 |
+
]
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/low_iou_samples.csv
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
sample_id
|
| 2 |
+
BUSI_0224_B
|
| 3 |
+
BUSI_0116_M
|
| 4 |
+
BUSI_0207_M
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/run_config.json
ADDED
|
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"project_dir": "/workspace",
|
| 3 |
+
"data_root": "/workspace/BUSI_with_classes",
|
| 4 |
+
"run_type": "final",
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"dataset_name": "BUSI_with_classes",
|
| 8 |
+
"dataset_split_policy": "stratified",
|
| 9 |
+
"dataset_splits_path": "/workspace/dataset_splits_busi_with_classes_stratified.json",
|
| 10 |
+
"split_type": "80_10_10",
|
| 11 |
+
"train_subset_key": "50",
|
| 12 |
+
"normalization_cache_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/norm_stats_BUSI_with_classes_stratified_80_10_10_50pct.json",
|
| 13 |
+
"head_lr": 0.002,
|
| 14 |
+
"encoder_lr": 0.0002,
|
| 15 |
+
"weight_decay": 0.002,
|
| 16 |
+
"dropout_p": 0.3,
|
| 17 |
+
"tmax": 6,
|
| 18 |
+
"entropy_lr": 0.0002,
|
| 19 |
+
"max_epochs": 100,
|
| 20 |
+
"gamma": 0.95,
|
| 21 |
+
"grad_clip_norm": 6.0,
|
| 22 |
+
"scheduler_factor": 0.5,
|
| 23 |
+
"scheduler_patience": 5,
|
| 24 |
+
"scheduler_threshold": 0.001,
|
| 25 |
+
"scheduler_min_lr": 1e-05,
|
| 26 |
+
"execution_mode": "train_eval",
|
| 27 |
+
"evaluation_checkpoint_mode": "latest",
|
| 28 |
+
"strategy2_checkpoint_mode": "specific",
|
| 29 |
+
"train_resume_mode": "off",
|
| 30 |
+
"train_resume_specific_checkpoint": "",
|
| 31 |
+
"backbone_family": "smp",
|
| 32 |
+
"smp_encoder_name": "efficientnet-b0",
|
| 33 |
+
"smp_encoder_weights": "imagenet",
|
| 34 |
+
"smp_encoder_depth": 5,
|
| 35 |
+
"smp_encoder_proj_dim": 32,
|
| 36 |
+
"smp_decoder_type": "Unet",
|
| 37 |
+
"vgg_feature_scales": 4,
|
| 38 |
+
"vgg_feature_dilation": 1,
|
| 39 |
+
"strategy2_checkpoint_path": "/workspace/50%_Strategt2_latest.pt"
|
| 40 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=32/pct_50/strategy_3/final/summary.json
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_val_iou": 0.6408889753679513,
|
| 3 |
+
"best_val_dice": 0.7398004646205255,
|
| 4 |
+
"final_epoch": 100,
|
| 5 |
+
"elapsed_seconds": 1503.7406146526337,
|
| 6 |
+
"seconds_per_epoch": 15.037406146526337,
|
| 7 |
+
"device_used": "cuda",
|
| 8 |
+
"strategy": 3,
|
| 9 |
+
"run_type": "final",
|
| 10 |
+
"resumed": false
|
| 11 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64/pct_50/norm_stats_BUSI_with_classes_stratified_80_10_10_50pct.json
ADDED
|
@@ -0,0 +1,266 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"global_mean": 85.12605579417583,
|
| 3 |
+
"global_std": 56.68613974260357,
|
| 4 |
+
"total_pixels": 244630659,
|
| 5 |
+
"num_images": 258,
|
| 6 |
+
"filenames": [
|
| 7 |
+
"BUSI_0034_B.png",
|
| 8 |
+
"BUSI_0155_M.png",
|
| 9 |
+
"BUSI_0005_M.png",
|
| 10 |
+
"BUSI_0260_B.png",
|
| 11 |
+
"BUSI_0386_B.png",
|
| 12 |
+
"BUSI_0280_B.png",
|
| 13 |
+
"BUSI_0047_B.png",
|
| 14 |
+
"BUSI_0222_B.png",
|
| 15 |
+
"BUSI_0205_B.png",
|
| 16 |
+
"BUSI_0403_B.png",
|
| 17 |
+
"BUSI_0133_B.png",
|
| 18 |
+
"BUSI_0016_B.png",
|
| 19 |
+
"BUSI_0105_M.png",
|
| 20 |
+
"BUSI_0220_B.png",
|
| 21 |
+
"BUSI_0066_B.png",
|
| 22 |
+
"BUSI_0060_B.png",
|
| 23 |
+
"BUSI_0351_B.png",
|
| 24 |
+
"BUSI_0029_M.png",
|
| 25 |
+
"BUSI_0287_B.png",
|
| 26 |
+
"BUSI_0046_M.png",
|
| 27 |
+
"BUSI_0123_B.png",
|
| 28 |
+
"BUSI_0359_B.png",
|
| 29 |
+
"BUSI_0215_B.png",
|
| 30 |
+
"BUSI_0377_B.png",
|
| 31 |
+
"BUSI_0029_B.png",
|
| 32 |
+
"BUSI_0158_M.png",
|
| 33 |
+
"BUSI_0375_B.png",
|
| 34 |
+
"BUSI_0221_B.png",
|
| 35 |
+
"BUSI_0120_B.png",
|
| 36 |
+
"BUSI_0404_B.png",
|
| 37 |
+
"BUSI_0093_M.png",
|
| 38 |
+
"BUSI_0106_M.png",
|
| 39 |
+
"BUSI_0030_B.png",
|
| 40 |
+
"BUSI_0407_B.png",
|
| 41 |
+
"BUSI_0131_B.png",
|
| 42 |
+
"BUSI_0092_B.png",
|
| 43 |
+
"BUSI_0112_M.png",
|
| 44 |
+
"BUSI_0177_M.png",
|
| 45 |
+
"BUSI_0333_B.png",
|
| 46 |
+
"BUSI_0206_B.png",
|
| 47 |
+
"BUSI_0239_B.png",
|
| 48 |
+
"BUSI_0165_B.png",
|
| 49 |
+
"BUSI_0396_B.png",
|
| 50 |
+
"BUSI_0207_B.png",
|
| 51 |
+
"BUSI_0158_B.png",
|
| 52 |
+
"BUSI_0160_B.png",
|
| 53 |
+
"BUSI_0091_B.png",
|
| 54 |
+
"BUSI_0090_M.png",
|
| 55 |
+
"BUSI_0035_B.png",
|
| 56 |
+
"BUSI_0281_B.png",
|
| 57 |
+
"BUSI_0410_B.png",
|
| 58 |
+
"BUSI_0322_B.png",
|
| 59 |
+
"BUSI_0236_B.png",
|
| 60 |
+
"BUSI_0385_B.png",
|
| 61 |
+
"BUSI_0193_M.png",
|
| 62 |
+
"BUSI_0388_B.png",
|
| 63 |
+
"BUSI_0144_M.png",
|
| 64 |
+
"BUSI_0293_B.png",
|
| 65 |
+
"BUSI_0088_M.png",
|
| 66 |
+
"BUSI_0052_B.png",
|
| 67 |
+
"BUSI_0346_B.png",
|
| 68 |
+
"BUSI_0262_B.png",
|
| 69 |
+
"BUSI_0408_B.png",
|
| 70 |
+
"BUSI_0146_M.png",
|
| 71 |
+
"BUSI_0309_B.png",
|
| 72 |
+
"BUSI_0083_B.png",
|
| 73 |
+
"BUSI_0190_B.png",
|
| 74 |
+
"BUSI_0426_B.png",
|
| 75 |
+
"BUSI_0019_B.png",
|
| 76 |
+
"BUSI_0274_B.png",
|
| 77 |
+
"BUSI_0162_M.png",
|
| 78 |
+
"BUSI_0179_M.png",
|
| 79 |
+
"BUSI_0078_B.png",
|
| 80 |
+
"BUSI_0328_B.png",
|
| 81 |
+
"BUSI_0001_M.png",
|
| 82 |
+
"BUSI_0301_B.png",
|
| 83 |
+
"BUSI_0109_M.png",
|
| 84 |
+
"BUSI_0011_B.png",
|
| 85 |
+
"BUSI_0063_B.png",
|
| 86 |
+
"BUSI_0360_B.png",
|
| 87 |
+
"BUSI_0114_B.png",
|
| 88 |
+
"BUSI_0430_B.png",
|
| 89 |
+
"BUSI_0391_B.png",
|
| 90 |
+
"BUSI_0133_M.png",
|
| 91 |
+
"BUSI_0188_M.png",
|
| 92 |
+
"BUSI_0424_B.png",
|
| 93 |
+
"BUSI_0179_B.png",
|
| 94 |
+
"BUSI_0390_B.png",
|
| 95 |
+
"BUSI_0030_M.png",
|
| 96 |
+
"BUSI_0235_B.png",
|
| 97 |
+
"BUSI_0065_M.png",
|
| 98 |
+
"BUSI_0101_M.png",
|
| 99 |
+
"BUSI_0355_B.png",
|
| 100 |
+
"BUSI_0053_M.png",
|
| 101 |
+
"BUSI_0203_B.png",
|
| 102 |
+
"BUSI_0007_B.png",
|
| 103 |
+
"BUSI_0344_B.png",
|
| 104 |
+
"BUSI_0129_B.png",
|
| 105 |
+
"BUSI_0271_B.png",
|
| 106 |
+
"BUSI_0252_B.png",
|
| 107 |
+
"BUSI_0296_B.png",
|
| 108 |
+
"BUSI_0183_M.png",
|
| 109 |
+
"BUSI_0310_B.png",
|
| 110 |
+
"BUSI_0409_B.png",
|
| 111 |
+
"BUSI_0041_M.png",
|
| 112 |
+
"BUSI_0092_M.png",
|
| 113 |
+
"BUSI_0273_B.png",
|
| 114 |
+
"BUSI_0206_M.png",
|
| 115 |
+
"BUSI_0024_M.png",
|
| 116 |
+
"BUSI_0015_B.png",
|
| 117 |
+
"BUSI_0094_M.png",
|
| 118 |
+
"BUSI_0055_B.png",
|
| 119 |
+
"BUSI_0094_B.png",
|
| 120 |
+
"BUSI_0077_B.png",
|
| 121 |
+
"BUSI_0347_B.png",
|
| 122 |
+
"BUSI_0126_M.png",
|
| 123 |
+
"BUSI_0186_M.png",
|
| 124 |
+
"BUSI_0145_M.png",
|
| 125 |
+
"BUSI_0054_M.png",
|
| 126 |
+
"BUSI_0074_M.png",
|
| 127 |
+
"BUSI_0111_B.png",
|
| 128 |
+
"BUSI_0032_M.png",
|
| 129 |
+
"BUSI_0176_M.png",
|
| 130 |
+
"BUSI_0002_M.png",
|
| 131 |
+
"BUSI_0069_B.png",
|
| 132 |
+
"BUSI_0027_B.png",
|
| 133 |
+
"BUSI_0109_B.png",
|
| 134 |
+
"BUSI_0389_B.png",
|
| 135 |
+
"BUSI_0088_B.png",
|
| 136 |
+
"BUSI_0169_B.png",
|
| 137 |
+
"BUSI_0157_B.png",
|
| 138 |
+
"BUSI_0058_B.png",
|
| 139 |
+
"BUSI_0231_B.png",
|
| 140 |
+
"BUSI_0166_B.png",
|
| 141 |
+
"BUSI_0098_M.png",
|
| 142 |
+
"BUSI_0020_M.png",
|
| 143 |
+
"BUSI_0182_B.png",
|
| 144 |
+
"BUSI_0397_B.png",
|
| 145 |
+
"BUSI_0045_M.png",
|
| 146 |
+
"BUSI_0138_M.png",
|
| 147 |
+
"BUSI_0036_M.png",
|
| 148 |
+
"BUSI_0380_B.png",
|
| 149 |
+
"BUSI_0137_B.png",
|
| 150 |
+
"BUSI_0205_M.png",
|
| 151 |
+
"BUSI_0107_M.png",
|
| 152 |
+
"BUSI_0044_M.png",
|
| 153 |
+
"BUSI_0213_B.png",
|
| 154 |
+
"BUSI_0070_B.png",
|
| 155 |
+
"BUSI_0295_B.png",
|
| 156 |
+
"BUSI_0173_B.png",
|
| 157 |
+
"BUSI_0164_M.png",
|
| 158 |
+
"BUSI_0039_M.png",
|
| 159 |
+
"BUSI_0421_B.png",
|
| 160 |
+
"BUSI_0174_B.png",
|
| 161 |
+
"BUSI_0416_B.png",
|
| 162 |
+
"BUSI_0124_M.png",
|
| 163 |
+
"BUSI_0428_B.png",
|
| 164 |
+
"BUSI_0181_B.png",
|
| 165 |
+
"BUSI_0076_B.png",
|
| 166 |
+
"BUSI_0073_B.png",
|
| 167 |
+
"BUSI_0045_B.png",
|
| 168 |
+
"BUSI_0104_M.png",
|
| 169 |
+
"BUSI_0272_B.png",
|
| 170 |
+
"BUSI_0405_B.png",
|
| 171 |
+
"BUSI_0425_B.png",
|
| 172 |
+
"BUSI_0212_B.png",
|
| 173 |
+
"BUSI_0156_M.png",
|
| 174 |
+
"BUSI_0339_B.png",
|
| 175 |
+
"BUSI_0160_M.png",
|
| 176 |
+
"BUSI_0204_M.png",
|
| 177 |
+
"BUSI_0082_B.png",
|
| 178 |
+
"BUSI_0070_M.png",
|
| 179 |
+
"BUSI_0061_M.png",
|
| 180 |
+
"BUSI_0136_M.png",
|
| 181 |
+
"BUSI_0024_B.png",
|
| 182 |
+
"BUSI_0241_B.png",
|
| 183 |
+
"BUSI_0420_B.png",
|
| 184 |
+
"BUSI_0196_B.png",
|
| 185 |
+
"BUSI_0406_B.png",
|
| 186 |
+
"BUSI_0374_B.png",
|
| 187 |
+
"BUSI_0113_M.png",
|
| 188 |
+
"BUSI_0009_B.png",
|
| 189 |
+
"BUSI_0292_B.png",
|
| 190 |
+
"BUSI_0079_M.png",
|
| 191 |
+
"BUSI_0118_B.png",
|
| 192 |
+
"BUSI_0175_M.png",
|
| 193 |
+
"BUSI_0011_M.png",
|
| 194 |
+
"BUSI_0187_B.png",
|
| 195 |
+
"BUSI_0034_M.png",
|
| 196 |
+
"BUSI_0321_B.png",
|
| 197 |
+
"BUSI_0078_M.png",
|
| 198 |
+
"BUSI_0180_B.png",
|
| 199 |
+
"BUSI_0398_B.png",
|
| 200 |
+
"BUSI_0195_B.png",
|
| 201 |
+
"BUSI_0203_M.png",
|
| 202 |
+
"BUSI_0167_B.png",
|
| 203 |
+
"BUSI_0263_B.png",
|
| 204 |
+
"BUSI_0012_B.png",
|
| 205 |
+
"BUSI_0177_B.png",
|
| 206 |
+
"BUSI_0147_M.png",
|
| 207 |
+
"BUSI_0005_B.png",
|
| 208 |
+
"BUSI_0012_M.png",
|
| 209 |
+
"BUSI_0170_B.png",
|
| 210 |
+
"BUSI_0025_B.png",
|
| 211 |
+
"BUSI_0136_B.png",
|
| 212 |
+
"BUSI_0089_B.png",
|
| 213 |
+
"BUSI_0018_B.png",
|
| 214 |
+
"BUSI_0202_M.png",
|
| 215 |
+
"BUSI_0071_B.png",
|
| 216 |
+
"BUSI_0185_M.png",
|
| 217 |
+
"BUSI_0096_B.png",
|
| 218 |
+
"BUSI_0184_M.png",
|
| 219 |
+
"BUSI_0248_B.png",
|
| 220 |
+
"BUSI_0113_B.png",
|
| 221 |
+
"BUSI_0246_B.png",
|
| 222 |
+
"BUSI_0014_M.png",
|
| 223 |
+
"BUSI_0245_B.png",
|
| 224 |
+
"BUSI_0135_B.png",
|
| 225 |
+
"BUSI_0031_B.png",
|
| 226 |
+
"BUSI_0365_B.png",
|
| 227 |
+
"BUSI_0155_B.png",
|
| 228 |
+
"BUSI_0103_M.png",
|
| 229 |
+
"BUSI_0023_B.png",
|
| 230 |
+
"BUSI_0242_B.png",
|
| 231 |
+
"BUSI_0370_B.png",
|
| 232 |
+
"BUSI_0367_B.png",
|
| 233 |
+
"BUSI_0035_M.png",
|
| 234 |
+
"BUSI_0240_B.png",
|
| 235 |
+
"BUSI_0148_M.png",
|
| 236 |
+
"BUSI_0064_B.png",
|
| 237 |
+
"BUSI_0435_B.png",
|
| 238 |
+
"BUSI_0373_B.png",
|
| 239 |
+
"BUSI_0065_B.png",
|
| 240 |
+
"BUSI_0312_B.png",
|
| 241 |
+
"BUSI_0164_B.png",
|
| 242 |
+
"BUSI_0049_B.png",
|
| 243 |
+
"BUSI_0181_M.png",
|
| 244 |
+
"BUSI_0353_B.png",
|
| 245 |
+
"BUSI_0038_M.png",
|
| 246 |
+
"BUSI_0356_B.png",
|
| 247 |
+
"BUSI_0159_B.png",
|
| 248 |
+
"BUSI_0049_M.png",
|
| 249 |
+
"BUSI_0051_B.png",
|
| 250 |
+
"BUSI_0308_B.png",
|
| 251 |
+
"BUSI_0368_B.png",
|
| 252 |
+
"BUSI_0084_B.png",
|
| 253 |
+
"BUSI_0072_M.png",
|
| 254 |
+
"BUSI_0197_M.png",
|
| 255 |
+
"BUSI_0329_B.png",
|
| 256 |
+
"BUSI_0174_M.png",
|
| 257 |
+
"BUSI_0196_M.png",
|
| 258 |
+
"BUSI_0169_M.png",
|
| 259 |
+
"BUSI_0101_B.png",
|
| 260 |
+
"BUSI_0086_M.png",
|
| 261 |
+
"BUSI_0317_B.png",
|
| 262 |
+
"BUSI_0077_M.png",
|
| 263 |
+
"BUSI_0193_B.png",
|
| 264 |
+
"BUSI_0031_M.png"
|
| 265 |
+
]
|
| 266 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64/pct_50/strategy_3/final/evaluation_summary.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"mean_iou": 0.6765084862709045,
|
| 3 |
+
"mean_dice": 0.7636451721191406
|
| 4 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64/pct_50/strategy_3/final/low_iou_samples.csv
ADDED
|
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
sample_id
|
| 2 |
+
BUSI_0224_B
|
| 3 |
+
BUSI_0116_M
|
| 4 |
+
BUSI_0207_M
|
| 5 |
+
BUSI_0067_M
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/norm_stats_BUSI_with_classes_stratified_80_10_10_50pct.json
ADDED
|
@@ -0,0 +1,266 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"global_mean": 85.12605579417583,
|
| 3 |
+
"global_std": 56.68613974260357,
|
| 4 |
+
"total_pixels": 244630659,
|
| 5 |
+
"num_images": 258,
|
| 6 |
+
"filenames": [
|
| 7 |
+
"BUSI_0034_B.png",
|
| 8 |
+
"BUSI_0155_M.png",
|
| 9 |
+
"BUSI_0005_M.png",
|
| 10 |
+
"BUSI_0260_B.png",
|
| 11 |
+
"BUSI_0386_B.png",
|
| 12 |
+
"BUSI_0280_B.png",
|
| 13 |
+
"BUSI_0047_B.png",
|
| 14 |
+
"BUSI_0222_B.png",
|
| 15 |
+
"BUSI_0205_B.png",
|
| 16 |
+
"BUSI_0403_B.png",
|
| 17 |
+
"BUSI_0133_B.png",
|
| 18 |
+
"BUSI_0016_B.png",
|
| 19 |
+
"BUSI_0105_M.png",
|
| 20 |
+
"BUSI_0220_B.png",
|
| 21 |
+
"BUSI_0066_B.png",
|
| 22 |
+
"BUSI_0060_B.png",
|
| 23 |
+
"BUSI_0351_B.png",
|
| 24 |
+
"BUSI_0029_M.png",
|
| 25 |
+
"BUSI_0287_B.png",
|
| 26 |
+
"BUSI_0046_M.png",
|
| 27 |
+
"BUSI_0123_B.png",
|
| 28 |
+
"BUSI_0359_B.png",
|
| 29 |
+
"BUSI_0215_B.png",
|
| 30 |
+
"BUSI_0377_B.png",
|
| 31 |
+
"BUSI_0029_B.png",
|
| 32 |
+
"BUSI_0158_M.png",
|
| 33 |
+
"BUSI_0375_B.png",
|
| 34 |
+
"BUSI_0221_B.png",
|
| 35 |
+
"BUSI_0120_B.png",
|
| 36 |
+
"BUSI_0404_B.png",
|
| 37 |
+
"BUSI_0093_M.png",
|
| 38 |
+
"BUSI_0106_M.png",
|
| 39 |
+
"BUSI_0030_B.png",
|
| 40 |
+
"BUSI_0407_B.png",
|
| 41 |
+
"BUSI_0131_B.png",
|
| 42 |
+
"BUSI_0092_B.png",
|
| 43 |
+
"BUSI_0112_M.png",
|
| 44 |
+
"BUSI_0177_M.png",
|
| 45 |
+
"BUSI_0333_B.png",
|
| 46 |
+
"BUSI_0206_B.png",
|
| 47 |
+
"BUSI_0239_B.png",
|
| 48 |
+
"BUSI_0165_B.png",
|
| 49 |
+
"BUSI_0396_B.png",
|
| 50 |
+
"BUSI_0207_B.png",
|
| 51 |
+
"BUSI_0158_B.png",
|
| 52 |
+
"BUSI_0160_B.png",
|
| 53 |
+
"BUSI_0091_B.png",
|
| 54 |
+
"BUSI_0090_M.png",
|
| 55 |
+
"BUSI_0035_B.png",
|
| 56 |
+
"BUSI_0281_B.png",
|
| 57 |
+
"BUSI_0410_B.png",
|
| 58 |
+
"BUSI_0322_B.png",
|
| 59 |
+
"BUSI_0236_B.png",
|
| 60 |
+
"BUSI_0385_B.png",
|
| 61 |
+
"BUSI_0193_M.png",
|
| 62 |
+
"BUSI_0388_B.png",
|
| 63 |
+
"BUSI_0144_M.png",
|
| 64 |
+
"BUSI_0293_B.png",
|
| 65 |
+
"BUSI_0088_M.png",
|
| 66 |
+
"BUSI_0052_B.png",
|
| 67 |
+
"BUSI_0346_B.png",
|
| 68 |
+
"BUSI_0262_B.png",
|
| 69 |
+
"BUSI_0408_B.png",
|
| 70 |
+
"BUSI_0146_M.png",
|
| 71 |
+
"BUSI_0309_B.png",
|
| 72 |
+
"BUSI_0083_B.png",
|
| 73 |
+
"BUSI_0190_B.png",
|
| 74 |
+
"BUSI_0426_B.png",
|
| 75 |
+
"BUSI_0019_B.png",
|
| 76 |
+
"BUSI_0274_B.png",
|
| 77 |
+
"BUSI_0162_M.png",
|
| 78 |
+
"BUSI_0179_M.png",
|
| 79 |
+
"BUSI_0078_B.png",
|
| 80 |
+
"BUSI_0328_B.png",
|
| 81 |
+
"BUSI_0001_M.png",
|
| 82 |
+
"BUSI_0301_B.png",
|
| 83 |
+
"BUSI_0109_M.png",
|
| 84 |
+
"BUSI_0011_B.png",
|
| 85 |
+
"BUSI_0063_B.png",
|
| 86 |
+
"BUSI_0360_B.png",
|
| 87 |
+
"BUSI_0114_B.png",
|
| 88 |
+
"BUSI_0430_B.png",
|
| 89 |
+
"BUSI_0391_B.png",
|
| 90 |
+
"BUSI_0133_M.png",
|
| 91 |
+
"BUSI_0188_M.png",
|
| 92 |
+
"BUSI_0424_B.png",
|
| 93 |
+
"BUSI_0179_B.png",
|
| 94 |
+
"BUSI_0390_B.png",
|
| 95 |
+
"BUSI_0030_M.png",
|
| 96 |
+
"BUSI_0235_B.png",
|
| 97 |
+
"BUSI_0065_M.png",
|
| 98 |
+
"BUSI_0101_M.png",
|
| 99 |
+
"BUSI_0355_B.png",
|
| 100 |
+
"BUSI_0053_M.png",
|
| 101 |
+
"BUSI_0203_B.png",
|
| 102 |
+
"BUSI_0007_B.png",
|
| 103 |
+
"BUSI_0344_B.png",
|
| 104 |
+
"BUSI_0129_B.png",
|
| 105 |
+
"BUSI_0271_B.png",
|
| 106 |
+
"BUSI_0252_B.png",
|
| 107 |
+
"BUSI_0296_B.png",
|
| 108 |
+
"BUSI_0183_M.png",
|
| 109 |
+
"BUSI_0310_B.png",
|
| 110 |
+
"BUSI_0409_B.png",
|
| 111 |
+
"BUSI_0041_M.png",
|
| 112 |
+
"BUSI_0092_M.png",
|
| 113 |
+
"BUSI_0273_B.png",
|
| 114 |
+
"BUSI_0206_M.png",
|
| 115 |
+
"BUSI_0024_M.png",
|
| 116 |
+
"BUSI_0015_B.png",
|
| 117 |
+
"BUSI_0094_M.png",
|
| 118 |
+
"BUSI_0055_B.png",
|
| 119 |
+
"BUSI_0094_B.png",
|
| 120 |
+
"BUSI_0077_B.png",
|
| 121 |
+
"BUSI_0347_B.png",
|
| 122 |
+
"BUSI_0126_M.png",
|
| 123 |
+
"BUSI_0186_M.png",
|
| 124 |
+
"BUSI_0145_M.png",
|
| 125 |
+
"BUSI_0054_M.png",
|
| 126 |
+
"BUSI_0074_M.png",
|
| 127 |
+
"BUSI_0111_B.png",
|
| 128 |
+
"BUSI_0032_M.png",
|
| 129 |
+
"BUSI_0176_M.png",
|
| 130 |
+
"BUSI_0002_M.png",
|
| 131 |
+
"BUSI_0069_B.png",
|
| 132 |
+
"BUSI_0027_B.png",
|
| 133 |
+
"BUSI_0109_B.png",
|
| 134 |
+
"BUSI_0389_B.png",
|
| 135 |
+
"BUSI_0088_B.png",
|
| 136 |
+
"BUSI_0169_B.png",
|
| 137 |
+
"BUSI_0157_B.png",
|
| 138 |
+
"BUSI_0058_B.png",
|
| 139 |
+
"BUSI_0231_B.png",
|
| 140 |
+
"BUSI_0166_B.png",
|
| 141 |
+
"BUSI_0098_M.png",
|
| 142 |
+
"BUSI_0020_M.png",
|
| 143 |
+
"BUSI_0182_B.png",
|
| 144 |
+
"BUSI_0397_B.png",
|
| 145 |
+
"BUSI_0045_M.png",
|
| 146 |
+
"BUSI_0138_M.png",
|
| 147 |
+
"BUSI_0036_M.png",
|
| 148 |
+
"BUSI_0380_B.png",
|
| 149 |
+
"BUSI_0137_B.png",
|
| 150 |
+
"BUSI_0205_M.png",
|
| 151 |
+
"BUSI_0107_M.png",
|
| 152 |
+
"BUSI_0044_M.png",
|
| 153 |
+
"BUSI_0213_B.png",
|
| 154 |
+
"BUSI_0070_B.png",
|
| 155 |
+
"BUSI_0295_B.png",
|
| 156 |
+
"BUSI_0173_B.png",
|
| 157 |
+
"BUSI_0164_M.png",
|
| 158 |
+
"BUSI_0039_M.png",
|
| 159 |
+
"BUSI_0421_B.png",
|
| 160 |
+
"BUSI_0174_B.png",
|
| 161 |
+
"BUSI_0416_B.png",
|
| 162 |
+
"BUSI_0124_M.png",
|
| 163 |
+
"BUSI_0428_B.png",
|
| 164 |
+
"BUSI_0181_B.png",
|
| 165 |
+
"BUSI_0076_B.png",
|
| 166 |
+
"BUSI_0073_B.png",
|
| 167 |
+
"BUSI_0045_B.png",
|
| 168 |
+
"BUSI_0104_M.png",
|
| 169 |
+
"BUSI_0272_B.png",
|
| 170 |
+
"BUSI_0405_B.png",
|
| 171 |
+
"BUSI_0425_B.png",
|
| 172 |
+
"BUSI_0212_B.png",
|
| 173 |
+
"BUSI_0156_M.png",
|
| 174 |
+
"BUSI_0339_B.png",
|
| 175 |
+
"BUSI_0160_M.png",
|
| 176 |
+
"BUSI_0204_M.png",
|
| 177 |
+
"BUSI_0082_B.png",
|
| 178 |
+
"BUSI_0070_M.png",
|
| 179 |
+
"BUSI_0061_M.png",
|
| 180 |
+
"BUSI_0136_M.png",
|
| 181 |
+
"BUSI_0024_B.png",
|
| 182 |
+
"BUSI_0241_B.png",
|
| 183 |
+
"BUSI_0420_B.png",
|
| 184 |
+
"BUSI_0196_B.png",
|
| 185 |
+
"BUSI_0406_B.png",
|
| 186 |
+
"BUSI_0374_B.png",
|
| 187 |
+
"BUSI_0113_M.png",
|
| 188 |
+
"BUSI_0009_B.png",
|
| 189 |
+
"BUSI_0292_B.png",
|
| 190 |
+
"BUSI_0079_M.png",
|
| 191 |
+
"BUSI_0118_B.png",
|
| 192 |
+
"BUSI_0175_M.png",
|
| 193 |
+
"BUSI_0011_M.png",
|
| 194 |
+
"BUSI_0187_B.png",
|
| 195 |
+
"BUSI_0034_M.png",
|
| 196 |
+
"BUSI_0321_B.png",
|
| 197 |
+
"BUSI_0078_M.png",
|
| 198 |
+
"BUSI_0180_B.png",
|
| 199 |
+
"BUSI_0398_B.png",
|
| 200 |
+
"BUSI_0195_B.png",
|
| 201 |
+
"BUSI_0203_M.png",
|
| 202 |
+
"BUSI_0167_B.png",
|
| 203 |
+
"BUSI_0263_B.png",
|
| 204 |
+
"BUSI_0012_B.png",
|
| 205 |
+
"BUSI_0177_B.png",
|
| 206 |
+
"BUSI_0147_M.png",
|
| 207 |
+
"BUSI_0005_B.png",
|
| 208 |
+
"BUSI_0012_M.png",
|
| 209 |
+
"BUSI_0170_B.png",
|
| 210 |
+
"BUSI_0025_B.png",
|
| 211 |
+
"BUSI_0136_B.png",
|
| 212 |
+
"BUSI_0089_B.png",
|
| 213 |
+
"BUSI_0018_B.png",
|
| 214 |
+
"BUSI_0202_M.png",
|
| 215 |
+
"BUSI_0071_B.png",
|
| 216 |
+
"BUSI_0185_M.png",
|
| 217 |
+
"BUSI_0096_B.png",
|
| 218 |
+
"BUSI_0184_M.png",
|
| 219 |
+
"BUSI_0248_B.png",
|
| 220 |
+
"BUSI_0113_B.png",
|
| 221 |
+
"BUSI_0246_B.png",
|
| 222 |
+
"BUSI_0014_M.png",
|
| 223 |
+
"BUSI_0245_B.png",
|
| 224 |
+
"BUSI_0135_B.png",
|
| 225 |
+
"BUSI_0031_B.png",
|
| 226 |
+
"BUSI_0365_B.png",
|
| 227 |
+
"BUSI_0155_B.png",
|
| 228 |
+
"BUSI_0103_M.png",
|
| 229 |
+
"BUSI_0023_B.png",
|
| 230 |
+
"BUSI_0242_B.png",
|
| 231 |
+
"BUSI_0370_B.png",
|
| 232 |
+
"BUSI_0367_B.png",
|
| 233 |
+
"BUSI_0035_M.png",
|
| 234 |
+
"BUSI_0240_B.png",
|
| 235 |
+
"BUSI_0148_M.png",
|
| 236 |
+
"BUSI_0064_B.png",
|
| 237 |
+
"BUSI_0435_B.png",
|
| 238 |
+
"BUSI_0373_B.png",
|
| 239 |
+
"BUSI_0065_B.png",
|
| 240 |
+
"BUSI_0312_B.png",
|
| 241 |
+
"BUSI_0164_B.png",
|
| 242 |
+
"BUSI_0049_B.png",
|
| 243 |
+
"BUSI_0181_M.png",
|
| 244 |
+
"BUSI_0353_B.png",
|
| 245 |
+
"BUSI_0038_M.png",
|
| 246 |
+
"BUSI_0356_B.png",
|
| 247 |
+
"BUSI_0159_B.png",
|
| 248 |
+
"BUSI_0049_M.png",
|
| 249 |
+
"BUSI_0051_B.png",
|
| 250 |
+
"BUSI_0308_B.png",
|
| 251 |
+
"BUSI_0368_B.png",
|
| 252 |
+
"BUSI_0084_B.png",
|
| 253 |
+
"BUSI_0072_M.png",
|
| 254 |
+
"BUSI_0197_M.png",
|
| 255 |
+
"BUSI_0329_B.png",
|
| 256 |
+
"BUSI_0174_M.png",
|
| 257 |
+
"BUSI_0196_M.png",
|
| 258 |
+
"BUSI_0169_M.png",
|
| 259 |
+
"BUSI_0101_B.png",
|
| 260 |
+
"BUSI_0086_M.png",
|
| 261 |
+
"BUSI_0317_B.png",
|
| 262 |
+
"BUSI_0077_M.png",
|
| 263 |
+
"BUSI_0193_B.png",
|
| 264 |
+
"BUSI_0031_M.png"
|
| 265 |
+
]
|
| 266 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/best_params.json
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"head_lr": 0.002,
|
| 3 |
+
"encoder_lr": 0.0002,
|
| 4 |
+
"weight_decay": 0.002,
|
| 5 |
+
"dropout_p": 0.3,
|
| 6 |
+
"tmax": 6,
|
| 7 |
+
"entropy_lr": 0.0002
|
| 8 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/best.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/best.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 80,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0010.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0010.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 10,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0020.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0020.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 20,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0030.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0030.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 30,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0040.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0040.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 40,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0050.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0050.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 50,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0060.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0060.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 60,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0070.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0070.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 70,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0080.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0080.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 80,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0090.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0090.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 90,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0100.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/epoch_0100.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 100,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/latest.pt.meta.json
ADDED
|
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/latest.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 100,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
},
|
| 39 |
+
"last_restore": {
|
| 40 |
+
"restored_keys": {
|
| 41 |
+
"model_state_dict": true,
|
| 42 |
+
"optimizer_state_dict": true,
|
| 43 |
+
"scheduler_state_dict": true,
|
| 44 |
+
"scaler_state_dict": false,
|
| 45 |
+
"log_alpha": true,
|
| 46 |
+
"alpha_optimizer_state_dict": true,
|
| 47 |
+
"best_metric_name": true,
|
| 48 |
+
"best_metric_value": true,
|
| 49 |
+
"patience_counter": true,
|
| 50 |
+
"elapsed_seconds": true,
|
| 51 |
+
"run_config": true,
|
| 52 |
+
"epoch_metrics": true,
|
| 53 |
+
"resume_source": false
|
| 54 |
+
},
|
| 55 |
+
"restored_at_epoch": 100,
|
| 56 |
+
"expected_run_type": null
|
| 57 |
+
}
|
| 58 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/evaluation.json
ADDED
|
@@ -0,0 +1,619 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"strategy": 3,
|
| 3 |
+
"best_metric_name": "val_iou",
|
| 4 |
+
"metrics": {
|
| 5 |
+
"dice": {
|
| 6 |
+
"mean": 0.774315595626831,
|
| 7 |
+
"std": 0.24029327929019928
|
| 8 |
+
},
|
| 9 |
+
"ppv": {
|
| 10 |
+
"mean": 0.807403028011322,
|
| 11 |
+
"std": 0.23015764355659485
|
| 12 |
+
},
|
| 13 |
+
"sen": {
|
| 14 |
+
"mean": 0.8042696118354797,
|
| 15 |
+
"std": 0.2399255484342575
|
| 16 |
+
},
|
| 17 |
+
"iou": {
|
| 18 |
+
"mean": 0.6796422600746155,
|
| 19 |
+
"std": 0.2480257898569107
|
| 20 |
+
},
|
| 21 |
+
"biou": {
|
| 22 |
+
"mean": 0.1120559424161911,
|
| 23 |
+
"std": 0.09220939874649048
|
| 24 |
+
},
|
| 25 |
+
"hd95": {
|
| 26 |
+
"mean": 14.11050033569336,
|
| 27 |
+
"std": 17.762704849243164
|
| 28 |
+
}
|
| 29 |
+
},
|
| 30 |
+
"per_sample": [
|
| 31 |
+
{
|
| 32 |
+
"sample_id": "BUSI_0097_B",
|
| 33 |
+
"dice": 0.807166608659649,
|
| 34 |
+
"ppv": 0.8478851436598702,
|
| 35 |
+
"sen": 0.7701797754607058,
|
| 36 |
+
"iou": 0.6766800967271571,
|
| 37 |
+
"biou": 0.024000260266597264,
|
| 38 |
+
"hd95": 11.704699516296387
|
| 39 |
+
},
|
| 40 |
+
{
|
| 41 |
+
"sample_id": "BUSI_0415_B",
|
| 42 |
+
"dice": 0.8128742528976478,
|
| 43 |
+
"ppv": 0.9391214123292734,
|
| 44 |
+
"sen": 0.7165479055615215,
|
| 45 |
+
"iou": 0.6847414919956937,
|
| 46 |
+
"biou": 0.06624620406211293,
|
| 47 |
+
"hd95": 30.0
|
| 48 |
+
},
|
| 49 |
+
{
|
| 50 |
+
"sample_id": "BUSI_0290_B",
|
| 51 |
+
"dice": 0.9208633110777564,
|
| 52 |
+
"ppv": 0.8709278403741096,
|
| 53 |
+
"sen": 0.9768732665646037,
|
| 54 |
+
"iou": 0.853333339259259,
|
| 55 |
+
"biou": 0.09748456053944637,
|
| 56 |
+
"hd95": 5.830951690673828
|
| 57 |
+
},
|
| 58 |
+
{
|
| 59 |
+
"sample_id": "BUSI_0354_B",
|
| 60 |
+
"dice": 0.8605200986843943,
|
| 61 |
+
"ppv": 0.7559709368654758,
|
| 62 |
+
"sen": 0.9986282579816009,
|
| 63 |
+
"iou": 0.7551867346894847,
|
| 64 |
+
"biou": 0.05033588914903049,
|
| 65 |
+
"hd95": 10.889991760253906
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"sample_id": "BUSI_0067_B",
|
| 69 |
+
"dice": 0.921733174829026,
|
| 70 |
+
"ppv": 0.9522293010534572,
|
| 71 |
+
"sen": 0.8931297745393371,
|
| 72 |
+
"iou": 0.8548284671274312,
|
| 73 |
+
"biou": 0.13235319636670698,
|
| 74 |
+
"hd95": 8.209418296813965
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"sample_id": "BUSI_0165_M",
|
| 78 |
+
"dice": 0.8612094294660032,
|
| 79 |
+
"ppv": 0.8962382461404663,
|
| 80 |
+
"sen": 0.828815772882798,
|
| 81 |
+
"iou": 0.7562491766135541,
|
| 82 |
+
"biou": 0.04186810919998242,
|
| 83 |
+
"hd95": 18.815441131591797
|
| 84 |
+
},
|
| 85 |
+
{
|
| 86 |
+
"sample_id": "BUSI_0198_B",
|
| 87 |
+
"dice": 0.9607097604337845,
|
| 88 |
+
"ppv": 0.9451371605276084,
|
| 89 |
+
"sen": 0.9768041252059197,
|
| 90 |
+
"iou": 0.9243902485127897,
|
| 91 |
+
"biou": 0.20920535179692396,
|
| 92 |
+
"hd95": 2.0
|
| 93 |
+
},
|
| 94 |
+
{
|
| 95 |
+
"sample_id": "BUSI_0361_B",
|
| 96 |
+
"dice": 0.9222681215821209,
|
| 97 |
+
"ppv": 0.8930909168661151,
|
| 98 |
+
"sen": 0.9534161526850813,
|
| 99 |
+
"iou": 0.8557491389721854,
|
| 100 |
+
"biou": 0.06513445780288973,
|
| 101 |
+
"hd95": 3.0
|
| 102 |
+
},
|
| 103 |
+
{
|
| 104 |
+
"sample_id": "BUSI_0095_B",
|
| 105 |
+
"dice": 0.9199084851467997,
|
| 106 |
+
"ppv": 0.909502303392623,
|
| 107 |
+
"sen": 0.9305555877057464,
|
| 108 |
+
"iou": 0.8516949780953482,
|
| 109 |
+
"biou": 0.21250098437376952,
|
| 110 |
+
"hd95": 1.0
|
| 111 |
+
},
|
| 112 |
+
{
|
| 113 |
+
"sample_id": "BUSI_0057_B",
|
| 114 |
+
"dice": 0.9046563404309667,
|
| 115 |
+
"ppv": 0.8908297419957458,
|
| 116 |
+
"sen": 0.9189189554419119,
|
| 117 |
+
"iou": 0.8259110016554649,
|
| 118 |
+
"biou": 0.1034493063801076,
|
| 119 |
+
"hd95": 1.0
|
| 120 |
+
},
|
| 121 |
+
{
|
| 122 |
+
"sample_id": "BUSI_0009_M",
|
| 123 |
+
"dice": 0.7105624341178028,
|
| 124 |
+
"ppv": 0.8735244732673738,
|
| 125 |
+
"sen": 0.598843977012257,
|
| 126 |
+
"iou": 0.551063877546396,
|
| 127 |
+
"biou": 0.08888939506144719,
|
| 128 |
+
"hd95": 12.154613494873047
|
| 129 |
+
},
|
| 130 |
+
{
|
| 131 |
+
"sample_id": "BUSI_0399_B",
|
| 132 |
+
"dice": 0.9481765845984946,
|
| 133 |
+
"ppv": 0.9250936360760035,
|
| 134 |
+
"sen": 0.9724409460874477,
|
| 135 |
+
"iou": 0.9014598580105492,
|
| 136 |
+
"biou": 0.1684590794053479,
|
| 137 |
+
"hd95": 4.4721360206604
|
| 138 |
+
},
|
| 139 |
+
{
|
| 140 |
+
"sample_id": "BUSI_0099_M",
|
| 141 |
+
"dice": 0.7714285856255537,
|
| 142 |
+
"ppv": 0.9437690055062302,
|
| 143 |
+
"sen": 0.6523109608917058,
|
| 144 |
+
"iou": 0.6279070143673393,
|
| 145 |
+
"biou": 0.0964471591638786,
|
| 146 |
+
"hd95": 11.746427536010742
|
| 147 |
+
},
|
| 148 |
+
{
|
| 149 |
+
"sample_id": "BUSI_0307_B",
|
| 150 |
+
"dice": 0.925754068938043,
|
| 151 |
+
"ppv": 0.9236111287937202,
|
| 152 |
+
"sen": 0.9279069935100015,
|
| 153 |
+
"iou": 0.8617710881703913,
|
| 154 |
+
"biou": 0.1692314082835321,
|
| 155 |
+
"hd95": 1.4142135381698608
|
| 156 |
+
},
|
| 157 |
+
{
|
| 158 |
+
"sample_id": "BUSI_0039_B",
|
| 159 |
+
"dice": 0.8948004963965542,
|
| 160 |
+
"ppv": 0.8114035501308005,
|
| 161 |
+
"sen": 0.99730458293677,
|
| 162 |
+
"iou": 0.8096280504096169,
|
| 163 |
+
"biou": 0.1696435985325013,
|
| 164 |
+
"hd95": 2.8284270763397217
|
| 165 |
+
},
|
| 166 |
+
{
|
| 167 |
+
"sample_id": "BUSI_0189_M",
|
| 168 |
+
"dice": 0.19722428059809052,
|
| 169 |
+
"ppv": 0.10940036023499349,
|
| 170 |
+
"sen": 1.0,
|
| 171 |
+
"iou": 0.10940036023499349,
|
| 172 |
+
"biou": 0.02745136178377969,
|
| 173 |
+
"hd95": 39.5467643737793
|
| 174 |
+
},
|
| 175 |
+
{
|
| 176 |
+
"sample_id": "BUSI_0021_B",
|
| 177 |
+
"dice": 0.8491228599568913,
|
| 178 |
+
"ppv": 0.7469137364730022,
|
| 179 |
+
"sen": 0.9837398506180076,
|
| 180 |
+
"iou": 0.7378050379237574,
|
| 181 |
+
"biou": 0.09722347607850544,
|
| 182 |
+
"hd95": 2.0
|
| 183 |
+
},
|
| 184 |
+
{
|
| 185 |
+
"sample_id": "BUSI_0311_B",
|
| 186 |
+
"dice": 0.3339047194410449,
|
| 187 |
+
"ppv": 0.8324786611147588,
|
| 188 |
+
"sen": 0.2088336531374934,
|
| 189 |
+
"iou": 0.20041155553861909,
|
| 190 |
+
"biou": 0.006968987118819821,
|
| 191 |
+
"hd95": 64.56619262695312
|
| 192 |
+
},
|
| 193 |
+
{
|
| 194 |
+
"sample_id": "BUSI_0080_M",
|
| 195 |
+
"dice": 0.6093575193786094,
|
| 196 |
+
"ppv": 0.475209515530394,
|
| 197 |
+
"sen": 0.8490330724246368,
|
| 198 |
+
"iou": 0.4381841777790026,
|
| 199 |
+
"biou": 0.05066691982215472,
|
| 200 |
+
"hd95": 34.46880340576172
|
| 201 |
+
},
|
| 202 |
+
{
|
| 203 |
+
"sample_id": "BUSI_0130_B",
|
| 204 |
+
"dice": 0.9532133700147367,
|
| 205 |
+
"ppv": 0.9106090461091311,
|
| 206 |
+
"sen": 1.0,
|
| 207 |
+
"iou": 0.9106090461091311,
|
| 208 |
+
"biou": 0.18644113760387707,
|
| 209 |
+
"hd95": 1.4142135381698608
|
| 210 |
+
},
|
| 211 |
+
{
|
| 212 |
+
"sample_id": "BUSI_0401_B",
|
| 213 |
+
"dice": 0.8358672095604341,
|
| 214 |
+
"ppv": 0.9367176678221372,
|
| 215 |
+
"sen": 0.7546218624861701,
|
| 216 |
+
"iou": 0.7180170726003693,
|
| 217 |
+
"biou": 0.07200037119985152,
|
| 218 |
+
"hd95": 10.800439834594727
|
| 219 |
+
},
|
| 220 |
+
{
|
| 221 |
+
"sample_id": "BUSI_0326_B",
|
| 222 |
+
"dice": 0.9385187320984063,
|
| 223 |
+
"ppv": 0.8884677916368278,
|
| 224 |
+
"sen": 0.9945454548760331,
|
| 225 |
+
"iou": 0.8841594890000275,
|
| 226 |
+
"biou": 0.09387792086615475,
|
| 227 |
+
"hd95": 2.8284270763397217
|
| 228 |
+
},
|
| 229 |
+
{
|
| 230 |
+
"sample_id": "BUSI_0119_B",
|
| 231 |
+
"dice": 0.14594934162341736,
|
| 232 |
+
"ppv": 0.08257529163223992,
|
| 233 |
+
"sen": 0.6276597725213976,
|
| 234 |
+
"iou": 0.0787192075570909,
|
| 235 |
+
"biou": 0.006734341166888497,
|
| 236 |
+
"hd95": 55.453575134277344
|
| 237 |
+
},
|
| 238 |
+
{
|
| 239 |
+
"sample_id": "BUSI_0372_B",
|
| 240 |
+
"dice": 0.9234909679549801,
|
| 241 |
+
"ppv": 0.8870014854504072,
|
| 242 |
+
"sen": 0.9631114704802349,
|
| 243 |
+
"iou": 0.8578571530102034,
|
| 244 |
+
"biou": 0.11206934824597058,
|
| 245 |
+
"hd95": 4.4721360206604
|
| 246 |
+
},
|
| 247 |
+
{
|
| 248 |
+
"sample_id": "BUSI_0003_M",
|
| 249 |
+
"dice": 0.8090017421233847,
|
| 250 |
+
"ppv": 0.9485791679865808,
|
| 251 |
+
"sen": 0.7052314179847667,
|
| 252 |
+
"iou": 0.6792635969705817,
|
| 253 |
+
"biou": 0.08457756986190554,
|
| 254 |
+
"hd95": 10.401281356811523
|
| 255 |
+
},
|
| 256 |
+
{
|
| 257 |
+
"sample_id": "BUSI_0007_M",
|
| 258 |
+
"dice": 0.5889262089827005,
|
| 259 |
+
"ppv": 0.9512195254147627,
|
| 260 |
+
"sen": 0.42648852655060426,
|
| 261 |
+
"iou": 0.41736035465394117,
|
| 262 |
+
"biou": 0.12359599797977641,
|
| 263 |
+
"hd95": 12.04366683959961
|
| 264 |
+
},
|
| 265 |
+
{
|
| 266 |
+
"sample_id": "BUSI_0411_B",
|
| 267 |
+
"dice": 0.929952955307032,
|
| 268 |
+
"ppv": 0.9104401274083865,
|
| 269 |
+
"sen": 0.9503205154743314,
|
| 270 |
+
"iou": 0.8690767040021151,
|
| 271 |
+
"biou": 0.12413823305578173,
|
| 272 |
+
"hd95": 4.8680338859558105
|
| 273 |
+
},
|
| 274 |
+
{
|
| 275 |
+
"sample_id": "BUSI_0211_B",
|
| 276 |
+
"dice": 0.6939655831970726,
|
| 277 |
+
"ppv": 0.8944445030863871,
|
| 278 |
+
"sen": 0.5669015609501546,
|
| 279 |
+
"iou": 0.5313532899824126,
|
| 280 |
+
"biou": 0.11538546597551348,
|
| 281 |
+
"hd95": 5.0962910652160645
|
| 282 |
+
},
|
| 283 |
+
{
|
| 284 |
+
"sample_id": "BUSI_0202_B",
|
| 285 |
+
"dice": 0.8877551593085922,
|
| 286 |
+
"ppv": 0.8613862758551725,
|
| 287 |
+
"sen": 0.9157895623267766,
|
| 288 |
+
"iou": 0.7981653227841076,
|
| 289 |
+
"biou": 0.21153997780773498,
|
| 290 |
+
"hd95": 1.0
|
| 291 |
+
},
|
| 292 |
+
{
|
| 293 |
+
"sample_id": "BUSI_0194_M",
|
| 294 |
+
"dice": 0.6504398929489768,
|
| 295 |
+
"ppv": 0.8610248555104926,
|
| 296 |
+
"sen": 0.5226201921479645,
|
| 297 |
+
"iou": 0.4819643858337946,
|
| 298 |
+
"biou": 0.026756178342415272,
|
| 299 |
+
"hd95": 17.77068519592285
|
| 300 |
+
},
|
| 301 |
+
{
|
| 302 |
+
"sample_id": "BUSI_0141_B",
|
| 303 |
+
"dice": 0.5470874639324538,
|
| 304 |
+
"ppv": 0.9282255788030311,
|
| 305 |
+
"sen": 0.38783768430515675,
|
| 306 |
+
"iou": 0.3765453567123587,
|
| 307 |
+
"biou": 0.03723421325634517,
|
| 308 |
+
"hd95": 52.57560729980469
|
| 309 |
+
},
|
| 310 |
+
{
|
| 311 |
+
"sample_id": "BUSI_0251_B",
|
| 312 |
+
"dice": 0.8440945127410217,
|
| 313 |
+
"ppv": 0.9115646559303892,
|
| 314 |
+
"sen": 0.7859238164446286,
|
| 315 |
+
"iou": 0.7302453051102711,
|
| 316 |
+
"biou": 0.08256964901867062,
|
| 317 |
+
"hd95": 5.057775497436523
|
| 318 |
+
},
|
| 319 |
+
{
|
| 320 |
+
"sample_id": "BUSI_0161_M",
|
| 321 |
+
"dice": 0.8028890373881938,
|
| 322 |
+
"ppv": 0.8298045648804407,
|
| 323 |
+
"sen": 0.7776647220131081,
|
| 324 |
+
"iou": 0.6706889058646577,
|
| 325 |
+
"biou": 0.03132012722144805,
|
| 326 |
+
"hd95": 16.552946090698242
|
| 327 |
+
},
|
| 328 |
+
{
|
| 329 |
+
"sample_id": "BUSI_0055_M",
|
| 330 |
+
"dice": 0.8983566708243049,
|
| 331 |
+
"ppv": 0.8551564478381868,
|
| 332 |
+
"sen": 0.9461538530571983,
|
| 333 |
+
"iou": 0.8154696336497642,
|
| 334 |
+
"biou": 0.16577584717869134,
|
| 335 |
+
"hd95": 5.477386474609375
|
| 336 |
+
},
|
| 337 |
+
{
|
| 338 |
+
"sample_id": "BUSI_0059_M",
|
| 339 |
+
"dice": 0.8764978066615519,
|
| 340 |
+
"ppv": 0.8037423889525155,
|
| 341 |
+
"sen": 0.9637359779875821,
|
| 342 |
+
"iou": 0.7801478399124004,
|
| 343 |
+
"biou": 0.04882831077572056,
|
| 344 |
+
"hd95": 9.919593811035156
|
| 345 |
+
},
|
| 346 |
+
{
|
| 347 |
+
"sample_id": "BUSI_0285_B",
|
| 348 |
+
"dice": 0.9318854920096006,
|
| 349 |
+
"ppv": 0.8888888993513278,
|
| 350 |
+
"sen": 0.9792531141853617,
|
| 351 |
+
"iou": 0.872458422138778,
|
| 352 |
+
"biou": 0.14438548428583728,
|
| 353 |
+
"hd95": 3.0
|
| 354 |
+
},
|
| 355 |
+
{
|
| 356 |
+
"sample_id": "BUSI_0163_B",
|
| 357 |
+
"dice": 0.20802014937090862,
|
| 358 |
+
"ppv": 0.11624662237442264,
|
| 359 |
+
"sen": 0.9880952522675568,
|
| 360 |
+
"iou": 0.11608403970852592,
|
| 361 |
+
"biou": 0.030769976330787437,
|
| 362 |
+
"hd95": 27.93022346496582
|
| 363 |
+
},
|
| 364 |
+
{
|
| 365 |
+
"sample_id": "BUSI_0224_B",
|
| 366 |
+
"dice": 2.2614201207277222e-08,
|
| 367 |
+
"ppv": 4.224756897982387e-08,
|
| 368 |
+
"sen": 4.866179811864729e-08,
|
| 369 |
+
"iou": 2.2614201207277222e-08,
|
| 370 |
+
"biou": 2.9069758991349133e-07,
|
| 371 |
+
"hd95": 58.11451721191406
|
| 372 |
+
},
|
| 373 |
+
{
|
| 374 |
+
"sample_id": "BUSI_0316_B",
|
| 375 |
+
"dice": 0.8549763046921677,
|
| 376 |
+
"ppv": 0.945492663616506,
|
| 377 |
+
"sen": 0.7802768204104357,
|
| 378 |
+
"iou": 0.7466887459157493,
|
| 379 |
+
"biou": 0.06859770295766723,
|
| 380 |
+
"hd95": 15.0
|
| 381 |
+
},
|
| 382 |
+
{
|
| 383 |
+
"sample_id": "BUSI_0116_M",
|
| 384 |
+
"dice": 0.17761037795423154,
|
| 385 |
+
"ppv": 0.1692308544378611,
|
| 386 |
+
"sen": 0.18686305924540667,
|
| 387 |
+
"iou": 0.09746018325692953,
|
| 388 |
+
"biou": 3.9525676076807874e-07,
|
| 389 |
+
"hd95": 91.3189468383789
|
| 390 |
+
},
|
| 391 |
+
{
|
| 392 |
+
"sample_id": "BUSI_0214_B",
|
| 393 |
+
"dice": 0.9120762758434175,
|
| 394 |
+
"ppv": 0.876782089940724,
|
| 395 |
+
"sen": 0.9503311313100297,
|
| 396 |
+
"iou": 0.83836418321673,
|
| 397 |
+
"biou": 0.1555560246910974,
|
| 398 |
+
"hd95": 4.391811370849609
|
| 399 |
+
},
|
| 400 |
+
{
|
| 401 |
+
"sample_id": "BUSI_0207_M",
|
| 402 |
+
"dice": 0.005476505724178219,
|
| 403 |
+
"ppv": 0.454550413178062,
|
| 404 |
+
"sen": 0.002754875881274056,
|
| 405 |
+
"iou": 0.002745798860746905,
|
| 406 |
+
"biou": 5.747123133837279e-07,
|
| 407 |
+
"hd95": 39.1464958190918
|
| 408 |
+
},
|
| 409 |
+
{
|
| 410 |
+
"sample_id": "BUSI_0016_M",
|
| 411 |
+
"dice": 0.6339020839409851,
|
| 412 |
+
"ppv": 0.8236144663318329,
|
| 413 |
+
"sen": 0.5152246151575334,
|
| 414 |
+
"iou": 0.4640239081177326,
|
| 415 |
+
"biou": 0.019693868775958693,
|
| 416 |
+
"hd95": 25.89861297607422
|
| 417 |
+
},
|
| 418 |
+
{
|
| 419 |
+
"sample_id": "BUSI_0067_M",
|
| 420 |
+
"dice": 0.5841969181219613,
|
| 421 |
+
"ppv": 0.5356295088325999,
|
| 422 |
+
"sen": 0.6424501933831633,
|
| 423 |
+
"iou": 0.4126258542885769,
|
| 424 |
+
"biou": 0.03964800015506601,
|
| 425 |
+
"hd95": 13.0
|
| 426 |
+
},
|
| 427 |
+
{
|
| 428 |
+
"sample_id": "BUSI_0184_B",
|
| 429 |
+
"dice": 0.9733939006404088,
|
| 430 |
+
"ppv": 0.9898812146994627,
|
| 431 |
+
"sen": 0.9574468103214123,
|
| 432 |
+
"iou": 0.9481668795547038,
|
| 433 |
+
"biou": 0.3589745644312294,
|
| 434 |
+
"hd95": 2.0
|
| 435 |
+
},
|
| 436 |
+
{
|
| 437 |
+
"sample_id": "BUSI_0099_B",
|
| 438 |
+
"dice": 0.965468642320744,
|
| 439 |
+
"ppv": 0.9553696017615618,
|
| 440 |
+
"sen": 0.9757834792331226,
|
| 441 |
+
"iou": 0.9332425159070141,
|
| 442 |
+
"biou": 0.37984544198027753,
|
| 443 |
+
"hd95": 2.0
|
| 444 |
+
},
|
| 445 |
+
{
|
| 446 |
+
"sample_id": "BUSI_0412_B",
|
| 447 |
+
"dice": 0.794729295493642,
|
| 448 |
+
"ppv": 0.9670593375423355,
|
| 449 |
+
"sen": 0.674528306839192,
|
| 450 |
+
"iou": 0.6593782588222779,
|
| 451 |
+
"biou": 0.07939525909352381,
|
| 452 |
+
"hd95": 28.62643051147461
|
| 453 |
+
},
|
| 454 |
+
{
|
| 455 |
+
"sample_id": "BUSI_0100_M",
|
| 456 |
+
"dice": 0.7312676207736551,
|
| 457 |
+
"ppv": 0.822560225277538,
|
| 458 |
+
"sen": 0.6582150448057764,
|
| 459 |
+
"iou": 0.5763765917960398,
|
| 460 |
+
"biou": 0.056277464814950294,
|
| 461 |
+
"hd95": 13.821033477783203
|
| 462 |
+
},
|
| 463 |
+
{
|
| 464 |
+
"sample_id": "BUSI_0073_M",
|
| 465 |
+
"dice": 0.8752735320734112,
|
| 466 |
+
"ppv": 0.7936508209456586,
|
| 467 |
+
"sen": 0.9756097600634537,
|
| 468 |
+
"iou": 0.7782101454980356,
|
| 469 |
+
"biou": 0.13966528510319268,
|
| 470 |
+
"hd95": 4.4721360206604
|
| 471 |
+
},
|
| 472 |
+
{
|
| 473 |
+
"sample_id": "BUSI_0357_B",
|
| 474 |
+
"dice": 0.8212624703480086,
|
| 475 |
+
"ppv": 0.7022727611053681,
|
| 476 |
+
"sen": 0.9888000017919997,
|
| 477 |
+
"iou": 0.6967305866143645,
|
| 478 |
+
"biou": 0.0847462798043617,
|
| 479 |
+
"hd95": 10.283563613891602
|
| 480 |
+
},
|
| 481 |
+
{
|
| 482 |
+
"sample_id": "BUSI_0229_B",
|
| 483 |
+
"dice": 0.9443277340164006,
|
| 484 |
+
"ppv": 0.9316062247040182,
|
| 485 |
+
"sen": 0.9574014954843988,
|
| 486 |
+
"iou": 0.8945273736788683,
|
| 487 |
+
"biou": 0.18079142328168177,
|
| 488 |
+
"hd95": 2.0
|
| 489 |
+
},
|
| 490 |
+
{
|
| 491 |
+
"sample_id": "BUSI_0129_M",
|
| 492 |
+
"dice": 0.8341248915204494,
|
| 493 |
+
"ppv": 0.9755194409479646,
|
| 494 |
+
"sen": 0.7285297322395187,
|
| 495 |
+
"iou": 0.71544961201796,
|
| 496 |
+
"biou": 0.026217410820709584,
|
| 497 |
+
"hd95": 15.264337539672852
|
| 498 |
+
},
|
| 499 |
+
{
|
| 500 |
+
"sample_id": "BUSI_0156_B",
|
| 501 |
+
"dice": 0.9492635107588362,
|
| 502 |
+
"ppv": 0.938511346760082,
|
| 503 |
+
"sen": 0.9602649138195649,
|
| 504 |
+
"iou": 0.903426821362361,
|
| 505 |
+
"biou": 0.3095246315182958,
|
| 506 |
+
"hd95": 1.0
|
| 507 |
+
},
|
| 508 |
+
{
|
| 509 |
+
"sample_id": "BUSI_0320_B",
|
| 510 |
+
"dice": 0.9186295590332378,
|
| 511 |
+
"ppv": 0.8528827330253016,
|
| 512 |
+
"sen": 0.9953596298469536,
|
| 513 |
+
"iou": 0.8495049802960435,
|
| 514 |
+
"biou": 0.16935550858426726,
|
| 515 |
+
"hd95": 2.0
|
| 516 |
+
},
|
| 517 |
+
{
|
| 518 |
+
"sample_id": "BUSI_0104_B",
|
| 519 |
+
"dice": 0.9183673747049745,
|
| 520 |
+
"ppv": 0.9000000666666222,
|
| 521 |
+
"sen": 0.9375000434027476,
|
| 522 |
+
"iou": 0.8490566987064788,
|
| 523 |
+
"biou": 0.26666788888685183,
|
| 524 |
+
"hd95": 1.0
|
| 525 |
+
},
|
| 526 |
+
{
|
| 527 |
+
"sample_id": "BUSI_0340_B",
|
| 528 |
+
"dice": 0.8854581687567459,
|
| 529 |
+
"ppv": 0.9367755548794637,
|
| 530 |
+
"sen": 0.839471203034202,
|
| 531 |
+
"iou": 0.7944593432873248,
|
| 532 |
+
"biou": 0.04796185900195228,
|
| 533 |
+
"hd95": 15.072942733764648
|
| 534 |
+
},
|
| 535 |
+
{
|
| 536 |
+
"sample_id": "BUSI_0121_B",
|
| 537 |
+
"dice": 0.9489728330700574,
|
| 538 |
+
"ppv": 0.9167733781340105,
|
| 539 |
+
"sen": 0.9835164857807025,
|
| 540 |
+
"iou": 0.9029003905548058,
|
| 541 |
+
"biou": 0.30496503193969365,
|
| 542 |
+
"hd95": 2.2360680103302
|
| 543 |
+
},
|
| 544 |
+
{
|
| 545 |
+
"sample_id": "BUSI_0358_B",
|
| 546 |
+
"dice": 0.8566413190964705,
|
| 547 |
+
"ppv": 0.7492323695770349,
|
| 548 |
+
"sen": 1.0,
|
| 549 |
+
"iou": 0.7492323695770349,
|
| 550 |
+
"biou": 0.04787284687614528,
|
| 551 |
+
"hd95": 5.0
|
| 552 |
+
},
|
| 553 |
+
{
|
| 554 |
+
"sample_id": "BUSI_0270_B",
|
| 555 |
+
"dice": 0.9042821279241653,
|
| 556 |
+
"ppv": 0.8427230416143094,
|
| 557 |
+
"sen": 0.9755434849066618,
|
| 558 |
+
"iou": 0.825287396485656,
|
| 559 |
+
"biou": 0.17857216198914108,
|
| 560 |
+
"hd95": 3.1622776985168457
|
| 561 |
+
},
|
| 562 |
+
{
|
| 563 |
+
"sample_id": "BUSI_0023_M",
|
| 564 |
+
"dice": 0.8133824451260238,
|
| 565 |
+
"ppv": 0.9036144600224038,
|
| 566 |
+
"sen": 0.7395348885667927,
|
| 567 |
+
"iou": 0.6854630162879287,
|
| 568 |
+
"biou": 0.07079662463776557,
|
| 569 |
+
"hd95": 23.15347671508789
|
| 570 |
+
},
|
| 571 |
+
{
|
| 572 |
+
"sample_id": "BUSI_0020_B",
|
| 573 |
+
"dice": 0.8886107704248579,
|
| 574 |
+
"ppv": 0.799549572122796,
|
| 575 |
+
"sen": 1.0,
|
| 576 |
+
"iou": 0.799549572122796,
|
| 577 |
+
"biou": 0.0191392252922367,
|
| 578 |
+
"hd95": 2.8284270763397217
|
| 579 |
+
},
|
| 580 |
+
{
|
| 581 |
+
"sample_id": "BUSI_0153_M",
|
| 582 |
+
"dice": 0.8590025858236566,
|
| 583 |
+
"ppv": 0.9601968765205033,
|
| 584 |
+
"sen": 0.7771042643383397,
|
| 585 |
+
"iou": 0.7528523531400612,
|
| 586 |
+
"biou": 0.0277779578188967,
|
| 587 |
+
"hd95": 12.01247787475586
|
| 588 |
+
},
|
| 589 |
+
{
|
| 590 |
+
"sample_id": "BUSI_0154_B",
|
| 591 |
+
"dice": 0.9090909408772934,
|
| 592 |
+
"ppv": 0.8441559453532822,
|
| 593 |
+
"sen": 0.9848484963268967,
|
| 594 |
+
"iou": 0.8333334401708717,
|
| 595 |
+
"biou": 0.29310466706091887,
|
| 596 |
+
"hd95": 1.1242622137069702
|
| 597 |
+
},
|
| 598 |
+
{
|
| 599 |
+
"sample_id": "BUSI_0062_M",
|
| 600 |
+
"dice": 0.8233000286790146,
|
| 601 |
+
"ppv": 0.9234554443162688,
|
| 602 |
+
"sen": 0.7427440746374636,
|
| 603 |
+
"iou": 0.6996686122755338,
|
| 604 |
+
"biou": 0.06887778855158455,
|
| 605 |
+
"hd95": 8.9442720413208
|
| 606 |
+
},
|
| 607 |
+
{
|
| 608 |
+
"sample_id": "BUSI_0075_B",
|
| 609 |
+
"dice": 0.949269135918389,
|
| 610 |
+
"ppv": 0.961672480544864,
|
| 611 |
+
"sen": 0.9371816745022624,
|
| 612 |
+
"iou": 0.9034370043474624,
|
| 613 |
+
"biou": 0.31405015367755895,
|
| 614 |
+
"hd95": 2.0
|
| 615 |
+
}
|
| 616 |
+
],
|
| 617 |
+
"checkpoint_mode": "latest",
|
| 618 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/checkpoints/latest.pt"
|
| 619 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/evaluation_results.xlsx
ADDED
|
Binary file (35.2 kB). View file
|
|
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/evaluation_summary.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"mean_iou": 0.6796422600746155,
|
| 3 |
+
"mean_dice": 0.774315595626831
|
| 4 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/history.json
ADDED
|
@@ -0,0 +1,2242 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"epoch": 1,
|
| 4 |
+
"train_loss": 1.1891985079643583,
|
| 5 |
+
"train_actor_loss": -0.14385482020846396,
|
| 6 |
+
"train_critic_loss": 0.013641612457290096,
|
| 7 |
+
"train_mean_reward": -0.005757260079732565,
|
| 8 |
+
"train_entropy": 0.28889599936453714,
|
| 9 |
+
"train_ce_loss": 1.2596830325826849,
|
| 10 |
+
"train_dice_loss": 1.3927819658003422,
|
| 11 |
+
"train_dice": 0.16055311882263193,
|
| 12 |
+
"train_iou": 0.10825497923894342,
|
| 13 |
+
"grad_norm": 16.694549272226734,
|
| 14 |
+
"lr": 0.001999509027563903,
|
| 15 |
+
"encoder_lr": 0.0001999531232347445,
|
| 16 |
+
"alpha": 0.49568524956703186,
|
| 17 |
+
"validated_this_epoch": false,
|
| 18 |
+
"val_loss": null,
|
| 19 |
+
"val_dice": null,
|
| 20 |
+
"val_iou": null,
|
| 21 |
+
"val_reward": null,
|
| 22 |
+
"val_entropy": null
|
| 23 |
+
},
|
| 24 |
+
{
|
| 25 |
+
"epoch": 2,
|
| 26 |
+
"train_loss": 0.818282509260332,
|
| 27 |
+
"train_actor_loss": -0.14629023636508376,
|
| 28 |
+
"train_critic_loss": 0.001870527719642532,
|
| 29 |
+
"train_mean_reward": -0.00531817088800893,
|
| 30 |
+
"train_entropy": 0.19809083641598182,
|
| 31 |
+
"train_ce_loss": 0.8355272378993899,
|
| 32 |
+
"train_dice_loss": 1.0917477011680603,
|
| 33 |
+
"train_dice": 0.4397409178477557,
|
| 34 |
+
"train_iou": 0.3295219844213137,
|
| 35 |
+
"grad_norm": 11.557924665452898,
|
| 36 |
+
"lr": 0.0019980365947861304,
|
| 37 |
+
"encoder_lr": 0.00019981253920068581,
|
| 38 |
+
"alpha": 0.49613648653030396,
|
| 39 |
+
"validated_this_epoch": false,
|
| 40 |
+
"val_loss": null,
|
| 41 |
+
"val_dice": null,
|
| 42 |
+
"val_iou": null,
|
| 43 |
+
"val_reward": null,
|
| 44 |
+
"val_entropy": null
|
| 45 |
+
},
|
| 46 |
+
{
|
| 47 |
+
"epoch": 3,
|
| 48 |
+
"train_loss": 0.635328826890829,
|
| 49 |
+
"train_actor_loss": -0.15330567914277834,
|
| 50 |
+
"train_critic_loss": 0.0020154548127549214,
|
| 51 |
+
"train_mean_reward": -0.0057115005348240214,
|
| 52 |
+
"train_entropy": 0.1984728891035489,
|
| 53 |
+
"train_ce_loss": 0.5709510658836374,
|
| 54 |
+
"train_dice_loss": 1.0043024703359726,
|
| 55 |
+
"train_dice": 0.4930809456533083,
|
| 56 |
+
"train_iou": 0.3717657259660212,
|
| 57 |
+
"grad_norm": 8.382133818874063,
|
| 58 |
+
"lr": 0.001995584154780065,
|
| 59 |
+
"encoder_lr": 0.0001995783866372926,
|
| 60 |
+
"alpha": 0.4964904487133026,
|
| 61 |
+
"validated_this_epoch": false,
|
| 62 |
+
"val_loss": null,
|
| 63 |
+
"val_dice": null,
|
| 64 |
+
"val_iou": null,
|
| 65 |
+
"val_reward": null,
|
| 66 |
+
"val_entropy": null
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"epoch": 4,
|
| 70 |
+
"train_loss": 0.5540803857150549,
|
| 71 |
+
"train_actor_loss": -0.15356863105010232,
|
| 72 |
+
"train_critic_loss": 0.0016151386606829627,
|
| 73 |
+
"train_mean_reward": -0.004284753409573867,
|
| 74 |
+
"train_entropy": 0.17806554686769036,
|
| 75 |
+
"train_ce_loss": 0.4961163886059894,
|
| 76 |
+
"train_dice_loss": 0.9175664900347243,
|
| 77 |
+
"train_dice": 0.5539812596228515,
|
| 78 |
+
"train_iou": 0.4316342645845245,
|
| 79 |
+
"grad_norm": 7.555573263602663,
|
| 80 |
+
"lr": 0.001992154127807906,
|
| 81 |
+
"encoder_lr": 0.0001992508966248754,
|
| 82 |
+
"alpha": 0.4994765818119049,
|
| 83 |
+
"validated_this_epoch": false,
|
| 84 |
+
"val_loss": null,
|
| 85 |
+
"val_dice": null,
|
| 86 |
+
"val_iou": null,
|
| 87 |
+
"val_reward": null,
|
| 88 |
+
"val_entropy": null
|
| 89 |
+
},
|
| 90 |
+
{
|
| 91 |
+
"epoch": 5,
|
| 92 |
+
"train_loss": 0.5524988963583798,
|
| 93 |
+
"train_actor_loss": -0.14277176459883015,
|
| 94 |
+
"train_critic_loss": 0.0018230974611758784,
|
| 95 |
+
"train_mean_reward": -0.003748167719393828,
|
| 96 |
+
"train_entropy": 0.15688260864930886,
|
| 97 |
+
"train_ce_loss": 0.5368485971358762,
|
| 98 |
+
"train_dice_loss": 0.8518696125471623,
|
| 99 |
+
"train_dice": 0.5879836920314611,
|
| 100 |
+
"train_iou": 0.4633833715527618,
|
| 101 |
+
"grad_norm": 7.4680814107721165,
|
| 102 |
+
"lr": 0.0019877498988921626,
|
| 103 |
+
"encoder_lr": 0.0001988303923565381,
|
| 104 |
+
"alpha": 0.5024347901344299,
|
| 105 |
+
"validated_this_epoch": false,
|
| 106 |
+
"val_loss": null,
|
| 107 |
+
"val_dice": null,
|
| 108 |
+
"val_iou": null,
|
| 109 |
+
"val_reward": null,
|
| 110 |
+
"val_entropy": null
|
| 111 |
+
},
|
| 112 |
+
{
|
| 113 |
+
"epoch": 6,
|
| 114 |
+
"train_loss": 0.4704458591334888,
|
| 115 |
+
"train_actor_loss": -0.12353817232266594,
|
| 116 |
+
"train_critic_loss": 0.0018596988828320877,
|
| 117 |
+
"train_mean_reward": -0.0036054538273540596,
|
| 118 |
+
"train_entropy": 0.11523698880373276,
|
| 119 |
+
"train_ce_loss": 0.42850772723752895,
|
| 120 |
+
"train_dice_loss": 0.7576006261261243,
|
| 121 |
+
"train_dice": 0.6269239983034779,
|
| 122 |
+
"train_iou": 0.5068712139931751,
|
| 123 |
+
"grad_norm": 7.174212845952012,
|
| 124 |
+
"lr": 0.0019823758144750458,
|
| 125 |
+
"encoder_lr": 0.00019831728881922545,
|
| 126 |
+
"alpha": 0.5088079571723938,
|
| 127 |
+
"validated_this_epoch": false,
|
| 128 |
+
"val_loss": null,
|
| 129 |
+
"val_dice": null,
|
| 130 |
+
"val_iou": null,
|
| 131 |
+
"val_reward": null,
|
| 132 |
+
"val_entropy": null
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"epoch": 7,
|
| 136 |
+
"train_loss": 0.4845495873960599,
|
| 137 |
+
"train_actor_loss": -0.1325493694092953,
|
| 138 |
+
"train_critic_loss": 0.0013554611838958465,
|
| 139 |
+
"train_mean_reward": -0.0023723470695081903,
|
| 140 |
+
"train_entropy": 0.14502612589040254,
|
| 141 |
+
"train_ce_loss": 0.440391869780557,
|
| 142 |
+
"train_dice_loss": 0.7924505673175635,
|
| 143 |
+
"train_dice": 0.6351170796091885,
|
| 144 |
+
"train_iou": 0.5188012910991857,
|
| 145 |
+
"grad_norm": 6.873728731343912,
|
| 146 |
+
"lr": 0.001976037178129054,
|
| 147 |
+
"encoder_lr": 0.00019771209238418101,
|
| 148 |
+
"alpha": 0.5143405795097351,
|
| 149 |
+
"validated_this_epoch": false,
|
| 150 |
+
"val_loss": null,
|
| 151 |
+
"val_dice": null,
|
| 152 |
+
"val_iou": null,
|
| 153 |
+
"val_reward": null,
|
| 154 |
+
"val_entropy": null
|
| 155 |
+
},
|
| 156 |
+
{
|
| 157 |
+
"epoch": 8,
|
| 158 |
+
"train_loss": 0.45745840563910134,
|
| 159 |
+
"train_actor_loss": -0.09643839902631958,
|
| 160 |
+
"train_critic_loss": 0.0010265517476341008,
|
| 161 |
+
"train_mean_reward": -0.0017459925424630639,
|
| 162 |
+
"train_entropy": 0.1051378345254746,
|
| 163 |
+
"train_ce_loss": 0.38999933367101747,
|
| 164 |
+
"train_dice_loss": 0.7167677097696359,
|
| 165 |
+
"train_dice": 0.6688116205029115,
|
| 166 |
+
"train_iou": 0.5602378654574829,
|
| 167 |
+
"grad_norm": 6.862808584704879,
|
| 168 |
+
"lr": 0.001968740245322988,
|
| 169 |
+
"encoder_lr": 0.00019701540030721995,
|
| 170 |
+
"alpha": 0.5218983292579651,
|
| 171 |
+
"validated_this_epoch": false,
|
| 172 |
+
"val_loss": null,
|
| 173 |
+
"val_dice": null,
|
| 174 |
+
"val_iou": null,
|
| 175 |
+
"val_reward": null,
|
| 176 |
+
"val_entropy": null
|
| 177 |
+
},
|
| 178 |
+
{
|
| 179 |
+
"epoch": 9,
|
| 180 |
+
"train_loss": 0.42765265260078555,
|
| 181 |
+
"train_actor_loss": -0.09352477382807162,
|
| 182 |
+
"train_critic_loss": 0.0010350641177891767,
|
| 183 |
+
"train_mean_reward": -0.0017285095820116629,
|
| 184 |
+
"train_entropy": 0.08897589071225424,
|
| 185 |
+
"train_ce_loss": 0.3772868796285406,
|
| 186 |
+
"train_dice_loss": 0.6640328974378816,
|
| 187 |
+
"train_dice": 0.6855635435724505,
|
| 188 |
+
"train_iou": 0.5720665179864213,
|
| 189 |
+
"grad_norm": 7.253844627345255,
|
| 190 |
+
"lr": 0.0019604922172485584,
|
| 191 |
+
"encoder_lr": 0.0001962279001393096,
|
| 192 |
+
"alpha": 0.5313951373100281,
|
| 193 |
+
"validated_this_epoch": false,
|
| 194 |
+
"val_loss": null,
|
| 195 |
+
"val_dice": null,
|
| 196 |
+
"val_iou": null,
|
| 197 |
+
"val_reward": null,
|
| 198 |
+
"val_entropy": null
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"epoch": 10,
|
| 202 |
+
"train_loss": 0.41356122419269675,
|
| 203 |
+
"train_actor_loss": -0.11736918808092203,
|
| 204 |
+
"train_critic_loss": 0.000826721060900376,
|
| 205 |
+
"train_mean_reward": -0.0016534036600847164,
|
| 206 |
+
"train_entropy": 0.1261659822281142,
|
| 207 |
+
"train_ce_loss": 0.3702138133632207,
|
| 208 |
+
"train_dice_loss": 0.6908202818197797,
|
| 209 |
+
"train_dice": 0.6805355253188674,
|
| 210 |
+
"train_iou": 0.571171953733401,
|
| 211 |
+
"grad_norm": 6.580118854378545,
|
| 212 |
+
"lr": 0.0019513012337136779,
|
| 213 |
+
"encoder_lr": 0.00019535036904803958,
|
| 214 |
+
"alpha": 0.5382658839225769,
|
| 215 |
+
"validated_this_epoch": true,
|
| 216 |
+
"val_loss": 0.6691178178386481,
|
| 217 |
+
"val_dice": 0.6582080869381314,
|
| 218 |
+
"val_iou": 0.5582093128452004,
|
| 219 |
+
"val_reward": 0.0010208340474601948,
|
| 220 |
+
"val_entropy": 0.011444261270537597,
|
| 221 |
+
"val_actor_loss": 0.0102237103194718,
|
| 222 |
+
"val_critic_loss": 0.00033531971266487556,
|
| 223 |
+
"val_ce_loss": 0.5222147997195898,
|
| 224 |
+
"val_dice_loss": 0.7952380680074594
|
| 225 |
+
},
|
| 226 |
+
{
|
| 227 |
+
"epoch": 11,
|
| 228 |
+
"train_loss": 0.4245665888602937,
|
| 229 |
+
"train_actor_loss": -0.11189932351133566,
|
| 230 |
+
"train_critic_loss": 0.0008305284401357375,
|
| 231 |
+
"train_mean_reward": -0.0011164111023688897,
|
| 232 |
+
"train_entropy": 0.12529789316931195,
|
| 233 |
+
"train_ce_loss": 0.36888449644737276,
|
| 234 |
+
"train_dice_loss": 0.7032167929241514,
|
| 235 |
+
"train_dice": 0.6986143531762166,
|
| 236 |
+
"train_iou": 0.5936553059035486,
|
| 237 |
+
"grad_norm": 6.461455593737521,
|
| 238 |
+
"lr": 0.0019411763651094546,
|
| 239 |
+
"encoder_lr": 0.0001943836730506514,
|
| 240 |
+
"alpha": 0.5459710955619812,
|
| 241 |
+
"validated_this_epoch": false,
|
| 242 |
+
"val_loss": null,
|
| 243 |
+
"val_dice": null,
|
| 244 |
+
"val_iou": null,
|
| 245 |
+
"val_reward": null,
|
| 246 |
+
"val_entropy": null
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"epoch": 12,
|
| 250 |
+
"train_loss": 0.40071953776266106,
|
| 251 |
+
"train_actor_loss": -0.11561252579311972,
|
| 252 |
+
"train_critic_loss": 0.0007962394205719355,
|
| 253 |
+
"train_mean_reward": -0.001026278586407919,
|
| 254 |
+
"train_entropy": 0.12654679661432586,
|
| 255 |
+
"train_ce_loss": 0.3548377711489527,
|
| 256 |
+
"train_dice_loss": 0.6770301041523,
|
| 257 |
+
"train_dice": 0.7120413399868962,
|
| 258 |
+
"train_iou": 0.6065762666666911,
|
| 259 |
+
"grad_norm": 6.42586472076039,
|
| 260 |
+
"lr": 0.0019301276034588103,
|
| 261 |
+
"encoder_lr": 0.00019332876615938386,
|
| 262 |
+
"alpha": 0.5526116490364075,
|
| 263 |
+
"validated_this_epoch": false,
|
| 264 |
+
"val_loss": null,
|
| 265 |
+
"val_dice": null,
|
| 266 |
+
"val_iou": null,
|
| 267 |
+
"val_reward": null,
|
| 268 |
+
"val_entropy": null
|
| 269 |
+
},
|
| 270 |
+
{
|
| 271 |
+
"epoch": 13,
|
| 272 |
+
"train_loss": 0.3900934287680578,
|
| 273 |
+
"train_actor_loss": -0.12012952791457124,
|
| 274 |
+
"train_critic_loss": 0.0009460315189077426,
|
| 275 |
+
"train_mean_reward": -0.0014665966582305485,
|
| 276 |
+
"train_entropy": 0.12971411958921614,
|
| 277 |
+
"train_ce_loss": 0.34212315788803,
|
| 278 |
+
"train_dice_loss": 0.6773767142486818,
|
| 279 |
+
"train_dice": 0.6945299318404636,
|
| 280 |
+
"train_iou": 0.5903638599273804,
|
| 281 |
+
"grad_norm": 5.599468872528668,
|
| 282 |
+
"lr": 0.0019181658525555612,
|
| 283 |
+
"encoder_lr": 0.00019218668943997818,
|
| 284 |
+
"alpha": 0.5603895783424377,
|
| 285 |
+
"validated_this_epoch": false,
|
| 286 |
+
"val_loss": null,
|
| 287 |
+
"val_dice": null,
|
| 288 |
+
"val_iou": null,
|
| 289 |
+
"val_reward": null,
|
| 290 |
+
"val_entropy": null
|
| 291 |
+
},
|
| 292 |
+
{
|
| 293 |
+
"epoch": 14,
|
| 294 |
+
"train_loss": 0.38763296501362443,
|
| 295 |
+
"train_actor_loss": -0.11083099663164947,
|
| 296 |
+
"train_critic_loss": 0.0006457861516940423,
|
| 297 |
+
"train_mean_reward": -0.001140778198519718,
|
| 298 |
+
"train_entropy": 0.11926065123031152,
|
| 299 |
+
"train_ce_loss": 0.3462269149909581,
|
| 300 |
+
"train_dice_loss": 0.6500552131070031,
|
| 301 |
+
"train_dice": 0.7082473524010235,
|
| 302 |
+
"train_iou": 0.602831241122943,
|
| 303 |
+
"grad_norm": 5.907783779525017,
|
| 304 |
+
"lr": 0.0019053029172036895,
|
| 305 |
+
"encoder_lr": 0.00019095856998427185,
|
| 306 |
+
"alpha": 0.5690038204193115,
|
| 307 |
+
"validated_this_epoch": false,
|
| 308 |
+
"val_loss": null,
|
| 309 |
+
"val_dice": null,
|
| 310 |
+
"val_iou": null,
|
| 311 |
+
"val_reward": null,
|
| 312 |
+
"val_entropy": null
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"epoch": 15,
|
| 316 |
+
"train_loss": 0.37242659079878676,
|
| 317 |
+
"train_actor_loss": -0.11219951566245566,
|
| 318 |
+
"train_critic_loss": 0.0006818748898066099,
|
| 319 |
+
"train_mean_reward": -0.0010600401826714968,
|
| 320 |
+
"train_entropy": 0.11697332309414596,
|
| 321 |
+
"train_ce_loss": 0.3335571032732205,
|
| 322 |
+
"train_dice_loss": 0.6350132263138005,
|
| 323 |
+
"train_dice": 0.7213671361705907,
|
| 324 |
+
"train_iou": 0.6161926555335705,
|
| 325 |
+
"grad_norm": 5.924156122198401,
|
| 326 |
+
"lr": 0.001891551491567426,
|
| 327 |
+
"encoder_lr": 0.00018964561979789495,
|
| 328 |
+
"alpha": 0.5773556232452393,
|
| 329 |
+
"validated_this_epoch": false,
|
| 330 |
+
"val_loss": null,
|
| 331 |
+
"val_dice": null,
|
| 332 |
+
"val_iou": null,
|
| 333 |
+
"val_reward": null,
|
| 334 |
+
"val_entropy": null
|
| 335 |
+
},
|
| 336 |
+
{
|
| 337 |
+
"epoch": 16,
|
| 338 |
+
"train_loss": 0.37308159086015347,
|
| 339 |
+
"train_actor_loss": -0.11912248727958168,
|
| 340 |
+
"train_critic_loss": 0.0006558383563034931,
|
| 341 |
+
"train_mean_reward": -0.0009862760316536878,
|
| 342 |
+
"train_entropy": 0.11742194459198442,
|
| 343 |
+
"train_ce_loss": 0.3612682700015181,
|
| 344 |
+
"train_dice_loss": 0.622484038889562,
|
| 345 |
+
"train_dice": 0.7270059274570869,
|
| 346 |
+
"train_iou": 0.6204479040213374,
|
| 347 |
+
"grad_norm": 5.802861483291138,
|
| 348 |
+
"lr": 0.0018769251466436443,
|
| 349 |
+
"encoder_lr": 0.00018824913460416704,
|
| 350 |
+
"alpha": 0.5872713923454285,
|
| 351 |
+
"validated_this_epoch": false,
|
| 352 |
+
"val_loss": null,
|
| 353 |
+
"val_dice": null,
|
| 354 |
+
"val_iou": null,
|
| 355 |
+
"val_reward": null,
|
| 356 |
+
"val_entropy": null
|
| 357 |
+
},
|
| 358 |
+
{
|
| 359 |
+
"epoch": 17,
|
| 360 |
+
"train_loss": 0.3403013715889206,
|
| 361 |
+
"train_actor_loss": -0.12918154034313556,
|
| 362 |
+
"train_critic_loss": 0.000671508310180524,
|
| 363 |
+
"train_mean_reward": -0.0010539740271084035,
|
| 364 |
+
"train_entropy": 0.13631959537342578,
|
| 365 |
+
"train_ce_loss": 0.3131163409713762,
|
| 366 |
+
"train_dice_loss": 0.6251779683800631,
|
| 367 |
+
"train_dice": 0.7330847738183486,
|
| 368 |
+
"train_iou": 0.6301571954434319,
|
| 369 |
+
"grad_norm": 4.818354519762734,
|
| 370 |
+
"lr": 0.001861438316868924,
|
| 371 |
+
"encoder_lr": 0.00018677049256537467,
|
| 372 |
+
"alpha": 0.5947402119636536,
|
| 373 |
+
"validated_this_epoch": false,
|
| 374 |
+
"val_loss": null,
|
| 375 |
+
"val_dice": null,
|
| 376 |
+
"val_iou": null,
|
| 377 |
+
"val_reward": null,
|
| 378 |
+
"val_entropy": null
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"epoch": 18,
|
| 382 |
+
"train_loss": 0.3465674067000261,
|
| 383 |
+
"train_actor_loss": -0.11124682430543718,
|
| 384 |
+
"train_critic_loss": 0.0004608817075294879,
|
| 385 |
+
"train_mean_reward": -0.0006208453414975627,
|
| 386 |
+
"train_entropy": 0.11393350369636286,
|
| 387 |
+
"train_ce_loss": 0.310555939055445,
|
| 388 |
+
"train_dice_loss": 0.604611631665735,
|
| 389 |
+
"train_dice": 0.7408692987317794,
|
| 390 |
+
"train_iou": 0.638657558827756,
|
| 391 |
+
"grad_norm": 6.679722870274108,
|
| 392 |
+
"lr": 0.0018451062858745053,
|
| 393 |
+
"encoder_lr": 0.00018521115292269145,
|
| 394 |
+
"alpha": 0.6039740443229675,
|
| 395 |
+
"validated_this_epoch": false,
|
| 396 |
+
"val_loss": null,
|
| 397 |
+
"val_dice": null,
|
| 398 |
+
"val_iou": null,
|
| 399 |
+
"val_reward": null,
|
| 400 |
+
"val_entropy": null
|
| 401 |
+
},
|
| 402 |
+
{
|
| 403 |
+
"epoch": 19,
|
| 404 |
+
"train_loss": 0.3196958774559641,
|
| 405 |
+
"train_actor_loss": -0.10207490796988555,
|
| 406 |
+
"train_critic_loss": 0.0005179192286761589,
|
| 407 |
+
"train_mean_reward": -0.0007859211989767727,
|
| 408 |
+
"train_entropy": 0.09792058131999265,
|
| 409 |
+
"train_ce_loss": 0.2854182712387207,
|
| 410 |
+
"train_dice_loss": 0.5576053729199008,
|
| 411 |
+
"train_dice": 0.7518243967488492,
|
| 412 |
+
"train_iou": 0.6510606179708539,
|
| 413 |
+
"grad_norm": 5.146859123725299,
|
| 414 |
+
"lr": 0.0018279451714031893,
|
| 415 |
+
"encoder_lr": 0.0001835726545560834,
|
| 416 |
+
"alpha": 0.6161861419677734,
|
| 417 |
+
"validated_this_epoch": false,
|
| 418 |
+
"val_loss": null,
|
| 419 |
+
"val_dice": null,
|
| 420 |
+
"val_iou": null,
|
| 421 |
+
"val_reward": null,
|
| 422 |
+
"val_entropy": null
|
| 423 |
+
},
|
| 424 |
+
{
|
| 425 |
+
"epoch": 20,
|
| 426 |
+
"train_loss": 0.33993886335874,
|
| 427 |
+
"train_actor_loss": -0.11230295782250616,
|
| 428 |
+
"train_critic_loss": 0.0005854193985872692,
|
| 429 |
+
"train_mean_reward": -0.0005493046321432849,
|
| 430 |
+
"train_entropy": 0.11678799156740895,
|
| 431 |
+
"train_ce_loss": 0.3032244033551984,
|
| 432 |
+
"train_dice_loss": 0.6006738093010214,
|
| 433 |
+
"train_dice": 0.7504697387494383,
|
| 434 |
+
"train_iou": 0.6525684242515419,
|
| 435 |
+
"grad_norm": 6.595465942639713,
|
| 436 |
+
"lr": 0.001809971909403073,
|
| 437 |
+
"encoder_lr": 0.00018185661446562003,
|
| 438 |
+
"alpha": 0.6258824467658997,
|
| 439 |
+
"validated_this_epoch": true,
|
| 440 |
+
"val_loss": 0.6058488682434527,
|
| 441 |
+
"val_dice": 0.6758976004974464,
|
| 442 |
+
"val_iou": 0.5798828121374697,
|
| 443 |
+
"val_reward": 0.0009105781605647031,
|
| 444 |
+
"val_entropy": 0.08111995256933938,
|
| 445 |
+
"val_actor_loss": 0.01578864925705611,
|
| 446 |
+
"val_critic_loss": 0.00012280528117022035,
|
| 447 |
+
"val_ce_loss": 0.3955631865249627,
|
| 448 |
+
"val_dice_loss": 0.7844344214959579
|
| 449 |
+
},
|
| 450 |
+
{
|
| 451 |
+
"epoch": 21,
|
| 452 |
+
"train_loss": 0.33783783532309697,
|
| 453 |
+
"train_actor_loss": -0.12112886593174742,
|
| 454 |
+
"train_critic_loss": 0.0004999653698821792,
|
| 455 |
+
"train_mean_reward": -0.0006548256951709341,
|
| 456 |
+
"train_entropy": 0.12140276391529634,
|
| 457 |
+
"train_ce_loss": 0.3139488244674696,
|
| 458 |
+
"train_dice_loss": 0.6034846033236777,
|
| 459 |
+
"train_dice": 0.7565286667517149,
|
| 460 |
+
"train_iou": 0.6552796659072054,
|
| 461 |
+
"grad_norm": 7.98160846566045,
|
| 462 |
+
"lr": 0.001791204237313812,
|
| 463 |
+
"encoder_lr": 0.0001800647261756906,
|
| 464 |
+
"alpha": 0.6371986269950867,
|
| 465 |
+
"validated_this_epoch": false,
|
| 466 |
+
"val_loss": null,
|
| 467 |
+
"val_dice": null,
|
| 468 |
+
"val_iou": null,
|
| 469 |
+
"val_reward": null,
|
| 470 |
+
"val_entropy": null
|
| 471 |
+
},
|
| 472 |
+
{
|
| 473 |
+
"epoch": 22,
|
| 474 |
+
"train_loss": 0.32119292539385613,
|
| 475 |
+
"train_actor_loss": -0.1414668701672734,
|
| 476 |
+
"train_critic_loss": 0.0005037010120514665,
|
| 477 |
+
"train_mean_reward": -0.0005201104737794401,
|
| 478 |
+
"train_entropy": 0.14940759286679117,
|
| 479 |
+
"train_ce_loss": 0.30971103774212805,
|
| 480 |
+
"train_dice_loss": 0.615104848023225,
|
| 481 |
+
"train_dice": 0.7505479799613791,
|
| 482 |
+
"train_iou": 0.650304530271963,
|
| 483 |
+
"grad_norm": 5.128550249014714,
|
| 484 |
+
"lr": 0.0017716606765619107,
|
| 485 |
+
"encoder_lr": 0.0001781987580637,
|
| 486 |
+
"alpha": 0.6430078744888306,
|
| 487 |
+
"validated_this_epoch": false,
|
| 488 |
+
"val_loss": null,
|
| 489 |
+
"val_dice": null,
|
| 490 |
+
"val_iou": null,
|
| 491 |
+
"val_reward": null,
|
| 492 |
+
"val_entropy": null
|
| 493 |
+
},
|
| 494 |
+
{
|
| 495 |
+
"epoch": 23,
|
| 496 |
+
"train_loss": 0.32069559561438465,
|
| 497 |
+
"train_actor_loss": -0.11642256766817598,
|
| 498 |
+
"train_critic_loss": 0.00039907622603068045,
|
| 499 |
+
"train_mean_reward": -0.0003467290809431484,
|
| 500 |
+
"train_entropy": 0.11859938290658426,
|
| 501 |
+
"train_ce_loss": 0.29202824156221185,
|
| 502 |
+
"train_dice_loss": 0.581809002761693,
|
| 503 |
+
"train_dice": 0.756445268682098,
|
| 504 |
+
"train_iou": 0.6590281282195251,
|
| 505 |
+
"grad_norm": 5.741100971088853,
|
| 506 |
+
"lr": 0.0017513605142823077,
|
| 507 |
+
"encoder_lr": 0.00017626055161489368,
|
| 508 |
+
"alpha": 0.6538273096084595,
|
| 509 |
+
"validated_this_epoch": false,
|
| 510 |
+
"val_loss": null,
|
| 511 |
+
"val_dice": null,
|
| 512 |
+
"val_iou": null,
|
| 513 |
+
"val_reward": null,
|
| 514 |
+
"val_entropy": null
|
| 515 |
+
},
|
| 516 |
+
{
|
| 517 |
+
"epoch": 24,
|
| 518 |
+
"train_loss": 0.29897785745569583,
|
| 519 |
+
"train_actor_loss": -0.12957355188138456,
|
| 520 |
+
"train_critic_loss": 0.00041756645586441216,
|
| 521 |
+
"train_mean_reward": -0.0005477840249582645,
|
| 522 |
+
"train_entropy": 0.13350534411637566,
|
| 523 |
+
"train_ce_loss": 0.29774519978440067,
|
| 524 |
+
"train_dice_loss": 0.5589400465174239,
|
| 525 |
+
"train_dice": 0.7655308687615067,
|
| 526 |
+
"train_iou": 0.668239688240459,
|
| 527 |
+
"grad_norm": 4.87719123400459,
|
| 528 |
+
"lr": 0.001730323784284305,
|
| 529 |
+
"encoder_lr": 0.0001742520196050341,
|
| 530 |
+
"alpha": 0.6636579632759094,
|
| 531 |
+
"validated_this_epoch": false,
|
| 532 |
+
"val_loss": null,
|
| 533 |
+
"val_dice": null,
|
| 534 |
+
"val_iou": null,
|
| 535 |
+
"val_reward": null,
|
| 536 |
+
"val_entropy": null
|
| 537 |
+
},
|
| 538 |
+
{
|
| 539 |
+
"epoch": 25,
|
| 540 |
+
"train_loss": 0.2942871344795806,
|
| 541 |
+
"train_actor_loss": -0.12450371159827793,
|
| 542 |
+
"train_critic_loss": 0.0005274401361702471,
|
| 543 |
+
"train_mean_reward": -0.000649438754344585,
|
| 544 |
+
"train_entropy": 0.113507084427245,
|
| 545 |
+
"train_ce_loss": 0.28789521227001147,
|
| 546 |
+
"train_dice_loss": 0.5491590321218013,
|
| 547 |
+
"train_dice": 0.7636851966996399,
|
| 548 |
+
"train_iou": 0.6662332365956998,
|
| 549 |
+
"grad_norm": 5.003853906494703,
|
| 550 |
+
"lr": 0.0017085712472806151,
|
| 551 |
+
"encoder_lr": 0.000172175144212722,
|
| 552 |
+
"alpha": 0.6747679710388184,
|
| 553 |
+
"validated_this_epoch": false,
|
| 554 |
+
"val_loss": null,
|
| 555 |
+
"val_dice": null,
|
| 556 |
+
"val_iou": null,
|
| 557 |
+
"val_reward": null,
|
| 558 |
+
"val_entropy": null
|
| 559 |
+
},
|
| 560 |
+
{
|
| 561 |
+
"epoch": 26,
|
| 562 |
+
"train_loss": 0.27197905851273896,
|
| 563 |
+
"train_actor_loss": -0.12968303817333365,
|
| 564 |
+
"train_critic_loss": 0.00043207868618891745,
|
| 565 |
+
"train_mean_reward": -0.00033076796249924435,
|
| 566 |
+
"train_entropy": 0.12936882866087468,
|
| 567 |
+
"train_ce_loss": 0.27031334647271077,
|
| 568 |
+
"train_dice_loss": 0.5325787626495657,
|
| 569 |
+
"train_dice": 0.7862254934461822,
|
| 570 |
+
"train_iou": 0.6910171577350163,
|
| 571 |
+
"grad_norm": 5.357030190931734,
|
| 572 |
+
"lr": 0.0016861243703990456,
|
| 573 |
+
"encoder_lr": 0.00017003197506322543,
|
| 574 |
+
"alpha": 0.684677004814148,
|
| 575 |
+
"validated_this_epoch": false,
|
| 576 |
+
"val_loss": null,
|
| 577 |
+
"val_dice": null,
|
| 578 |
+
"val_iou": null,
|
| 579 |
+
"val_reward": null,
|
| 580 |
+
"val_entropy": null
|
| 581 |
+
},
|
| 582 |
+
{
|
| 583 |
+
"epoch": 27,
|
| 584 |
+
"train_loss": 0.2883123110995391,
|
| 585 |
+
"train_actor_loss": -0.12412422681552206,
|
| 586 |
+
"train_critic_loss": 0.000414339517297773,
|
| 587 |
+
"train_mean_reward": -0.0004206856636515915,
|
| 588 |
+
"train_entropy": 0.11815022285146283,
|
| 589 |
+
"train_ce_loss": 0.2753041366239148,
|
| 590 |
+
"train_dice_loss": 0.5491545912677311,
|
| 591 |
+
"train_dice": 0.7789285341871623,
|
| 592 |
+
"train_iou": 0.6811039755281605,
|
| 593 |
+
"grad_norm": 6.211623058993687,
|
| 594 |
+
"lr": 0.001663005305997034,
|
| 595 |
+
"encoder_lr": 0.00016782462720574693,
|
| 596 |
+
"alpha": 0.6961227059364319,
|
| 597 |
+
"validated_this_epoch": false,
|
| 598 |
+
"val_loss": null,
|
| 599 |
+
"val_dice": null,
|
| 600 |
+
"val_iou": null,
|
| 601 |
+
"val_reward": null,
|
| 602 |
+
"val_entropy": null
|
| 603 |
+
},
|
| 604 |
+
{
|
| 605 |
+
"epoch": 28,
|
| 606 |
+
"train_loss": 0.24405958539090386,
|
| 607 |
+
"train_actor_loss": -0.12702207463806695,
|
| 608 |
+
"train_critic_loss": 0.00037900887518953376,
|
| 609 |
+
"train_mean_reward": -0.00025710482833823836,
|
| 610 |
+
"train_entropy": 0.12001601498675094,
|
| 611 |
+
"train_ce_loss": 0.25536012724138196,
|
| 612 |
+
"train_dice_loss": 0.4864241788091586,
|
| 613 |
+
"train_dice": 0.7958033615568744,
|
| 614 |
+
"train_iou": 0.6996788281554621,
|
| 615 |
+
"grad_norm": 4.225626177565996,
|
| 616 |
+
"lr": 0.0016392368697999468,
|
| 617 |
+
"encoder_lr": 0.00016555527902612554,
|
| 618 |
+
"alpha": 0.7079911828041077,
|
| 619 |
+
"validated_this_epoch": false,
|
| 620 |
+
"val_loss": null,
|
| 621 |
+
"val_dice": null,
|
| 622 |
+
"val_iou": null,
|
| 623 |
+
"val_reward": null,
|
| 624 |
+
"val_entropy": null
|
| 625 |
+
},
|
| 626 |
+
{
|
| 627 |
+
"epoch": 29,
|
| 628 |
+
"train_loss": 0.2620487071750783,
|
| 629 |
+
"train_actor_loss": -0.1383464525658902,
|
| 630 |
+
"train_critic_loss": 0.0004014990134032403,
|
| 631 |
+
"train_mean_reward": -0.0002986126402375874,
|
| 632 |
+
"train_entropy": 0.13120045638061129,
|
| 633 |
+
"train_ce_loss": 0.2789849668545095,
|
| 634 |
+
"train_dice_loss": 0.5214038475077282,
|
| 635 |
+
"train_dice": 0.7894119900967371,
|
| 636 |
+
"train_iou": 0.6956790585935629,
|
| 637 |
+
"grad_norm": 3.81498890329701,
|
| 638 |
+
"lr": 0.0016148425183847121,
|
| 639 |
+
"encoder_lr": 0.0001632261700970328,
|
| 640 |
+
"alpha": 0.7190180420875549,
|
| 641 |
+
"validated_this_epoch": false,
|
| 642 |
+
"val_loss": null,
|
| 643 |
+
"val_dice": null,
|
| 644 |
+
"val_iou": null,
|
| 645 |
+
"val_reward": null,
|
| 646 |
+
"val_entropy": null
|
| 647 |
+
},
|
| 648 |
+
{
|
| 649 |
+
"epoch": 30,
|
| 650 |
+
"train_loss": 0.2681546088477389,
|
| 651 |
+
"train_actor_loss": -0.1450507549078081,
|
| 652 |
+
"train_critic_loss": 0.00036110722065043347,
|
| 653 |
+
"train_mean_reward": -0.00020131954692439237,
|
| 654 |
+
"train_entropy": 0.1428435419289665,
|
| 655 |
+
"train_ce_loss": 0.27884036997147865,
|
| 656 |
+
"train_dice_loss": 0.5472092441348142,
|
| 657 |
+
"train_dice": 0.7922666869068672,
|
| 658 |
+
"train_iou": 0.6981795122681079,
|
| 659 |
+
"grad_norm": 5.026507161846457,
|
| 660 |
+
"lr": 0.0015898463260310114,
|
| 661 |
+
"encoder_lr": 0.00016083959896778498,
|
| 662 |
+
"alpha": 0.7274738550186157,
|
| 663 |
+
"validated_this_epoch": true,
|
| 664 |
+
"val_loss": 0.7430638820638485,
|
| 665 |
+
"val_dice": 0.6872590205363176,
|
| 666 |
+
"val_iou": 0.5869224157201949,
|
| 667 |
+
"val_reward": 0.0006182405235437812,
|
| 668 |
+
"val_entropy": 0.3462533910918717,
|
| 669 |
+
"val_actor_loss": -0.009860609485203344,
|
| 670 |
+
"val_critic_loss": 2.61106471334137e-05,
|
| 671 |
+
"val_ce_loss": 0.5418442606530858,
|
| 672 |
+
"val_dice_loss": 0.9639785976120921
|
| 673 |
+
},
|
| 674 |
+
{
|
| 675 |
+
"epoch": 31,
|
| 676 |
+
"train_loss": 0.2419653381414458,
|
| 677 |
+
"train_actor_loss": -0.13977525247912492,
|
| 678 |
+
"train_critic_loss": 0.00040694229643068624,
|
| 679 |
+
"train_mean_reward": -0.00024306075673912128,
|
| 680 |
+
"train_entropy": 0.13699384201955753,
|
| 681 |
+
"train_ce_loss": 0.25766247116602015,
|
| 682 |
+
"train_dice_loss": 0.5054117649870633,
|
| 683 |
+
"train_dice": 0.7930538086762543,
|
| 684 |
+
"train_iou": 0.7004955207772444,
|
| 685 |
+
"grad_norm": 5.029161268657492,
|
| 686 |
+
"lr": 0.0015642729609628705,
|
| 687 |
+
"encoder_lr": 0.00015839792089595244,
|
| 688 |
+
"alpha": 0.7372567057609558,
|
| 689 |
+
"validated_this_epoch": false,
|
| 690 |
+
"val_loss": null,
|
| 691 |
+
"val_dice": null,
|
| 692 |
+
"val_iou": null,
|
| 693 |
+
"val_reward": null,
|
| 694 |
+
"val_entropy": null
|
| 695 |
+
},
|
| 696 |
+
{
|
| 697 |
+
"epoch": 32,
|
| 698 |
+
"train_loss": 0.25898992450897784,
|
| 699 |
+
"train_actor_loss": -0.16107231655786203,
|
| 700 |
+
"train_critic_loss": 0.00037360776567725313,
|
| 701 |
+
"train_mean_reward": -0.0004209323508555565,
|
| 702 |
+
"train_entropy": 0.16398638646925728,
|
| 703 |
+
"train_ce_loss": 0.2916806083911361,
|
| 704 |
+
"train_dice_loss": 0.5480702604896338,
|
| 705 |
+
"train_dice": 0.7811459824714374,
|
| 706 |
+
"train_iou": 0.6868544911376306,
|
| 707 |
+
"grad_norm": 4.4459205255951995,
|
| 708 |
+
"lr": 0.001538147661004102,
|
| 709 |
+
"encoder_lr": 0.0001559035455230047,
|
| 710 |
+
"alpha": 0.7429198622703552,
|
| 711 |
+
"validated_this_epoch": false,
|
| 712 |
+
"val_loss": null,
|
| 713 |
+
"val_dice": null,
|
| 714 |
+
"val_iou": null,
|
| 715 |
+
"val_reward": null,
|
| 716 |
+
"val_entropy": null
|
| 717 |
+
},
|
| 718 |
+
{
|
| 719 |
+
"epoch": 33,
|
| 720 |
+
"train_loss": 0.2335554674774435,
|
| 721 |
+
"train_actor_loss": -0.14542618503003532,
|
| 722 |
+
"train_critic_loss": 0.00034765358440057216,
|
| 723 |
+
"train_mean_reward": -0.0001484340938312084,
|
| 724 |
+
"train_entropy": 0.13918635872862295,
|
| 725 |
+
"train_ce_loss": 0.24909859949130883,
|
| 726 |
+
"train_dice_loss": 0.5085170472315116,
|
| 727 |
+
"train_dice": 0.8066853675066915,
|
| 728 |
+
"train_iou": 0.7147664998030524,
|
| 729 |
+
"grad_norm": 3.9813086838685265,
|
| 730 |
+
"lr": 0.0015114962086716199,
|
| 731 |
+
"encoder_lr": 0.0001533589344962853,
|
| 732 |
+
"alpha": 0.7527349591255188,
|
| 733 |
+
"validated_this_epoch": false,
|
| 734 |
+
"val_loss": null,
|
| 735 |
+
"val_dice": null,
|
| 736 |
+
"val_iou": null,
|
| 737 |
+
"val_reward": null,
|
| 738 |
+
"val_entropy": null
|
| 739 |
+
},
|
| 740 |
+
{
|
| 741 |
+
"epoch": 34,
|
| 742 |
+
"train_loss": 0.24259487278564396,
|
| 743 |
+
"train_actor_loss": -0.166359303764369,
|
| 744 |
+
"train_critic_loss": 0.00034587143462580934,
|
| 745 |
+
"train_mean_reward": -0.0001772796735751675,
|
| 746 |
+
"train_entropy": 0.16630749488748775,
|
| 747 |
+
"train_ce_loss": 0.2649123764666669,
|
| 748 |
+
"train_dice_loss": 0.5526501008860517,
|
| 749 |
+
"train_dice": 0.7976520299655028,
|
| 750 |
+
"train_iou": 0.7011901887642519,
|
| 751 |
+
"grad_norm": 4.824721911156824,
|
| 752 |
+
"lr": 0.0014843449057312071,
|
| 753 |
+
"encoder_lr": 0.00015076659903966297,
|
| 754 |
+
"alpha": 0.759739339351654,
|
| 755 |
+
"validated_this_epoch": false,
|
| 756 |
+
"val_loss": null,
|
| 757 |
+
"val_dice": null,
|
| 758 |
+
"val_iou": null,
|
| 759 |
+
"val_reward": null,
|
| 760 |
+
"val_entropy": null
|
| 761 |
+
},
|
| 762 |
+
{
|
| 763 |
+
"epoch": 35,
|
| 764 |
+
"train_loss": 0.2452491777983441,
|
| 765 |
+
"train_actor_loss": -0.165889281065995,
|
| 766 |
+
"train_critic_loss": 0.00030275803681001183,
|
| 767 |
+
"train_mean_reward": -8.916573227371239e-05,
|
| 768 |
+
"train_entropy": 0.16797401592284633,
|
| 769 |
+
"train_ce_loss": 0.27974583001239395,
|
| 770 |
+
"train_dice_loss": 0.5422283230055827,
|
| 771 |
+
"train_dice": 0.7967665333893651,
|
| 772 |
+
"train_iou": 0.7015099449278955,
|
| 773 |
+
"grad_norm": 5.058081422665323,
|
| 774 |
+
"lr": 0.0014567205472408495,
|
| 775 |
+
"encoder_lr": 0.00014812909747525697,
|
| 776 |
+
"alpha": 0.7649304270744324,
|
| 777 |
+
"validated_this_epoch": false,
|
| 778 |
+
"val_loss": null,
|
| 779 |
+
"val_dice": null,
|
| 780 |
+
"val_iou": null,
|
| 781 |
+
"val_reward": null,
|
| 782 |
+
"val_entropy": null
|
| 783 |
+
},
|
| 784 |
+
{
|
| 785 |
+
"epoch": 36,
|
| 786 |
+
"train_loss": 0.2417166044870336,
|
| 787 |
+
"train_actor_loss": -0.20451328885056877,
|
| 788 |
+
"train_critic_loss": 0.0003731843245630188,
|
| 789 |
+
"train_mean_reward": -0.0002134889531042002,
|
| 790 |
+
"train_entropy": 0.21378324490304657,
|
| 791 |
+
"train_ce_loss": 0.3008104628030191,
|
| 792 |
+
"train_dice_loss": 0.591276133784336,
|
| 793 |
+
"train_dice": 0.8011073714781418,
|
| 794 |
+
"train_iou": 0.7042845662591657,
|
| 795 |
+
"grad_norm": 4.392330631036167,
|
| 796 |
+
"lr": 0.0014286503951072478,
|
| 797 |
+
"encoder_lr": 0.00014544903269868192,
|
| 798 |
+
"alpha": 0.766568660736084,
|
| 799 |
+
"validated_this_epoch": false,
|
| 800 |
+
"val_loss": null,
|
| 801 |
+
"val_dice": null,
|
| 802 |
+
"val_iou": null,
|
| 803 |
+
"val_reward": null,
|
| 804 |
+
"val_entropy": null
|
| 805 |
+
},
|
| 806 |
+
{
|
| 807 |
+
"epoch": 37,
|
| 808 |
+
"train_loss": 0.2268339839012605,
|
| 809 |
+
"train_actor_loss": -0.2725019927277381,
|
| 810 |
+
"train_critic_loss": 0.00041547275399529953,
|
| 811 |
+
"train_mean_reward": -0.0005525801379682526,
|
| 812 |
+
"train_entropy": 0.297990503814643,
|
| 813 |
+
"train_ce_loss": 0.333135462184954,
|
| 814 |
+
"train_dice_loss": 0.6651210148821198,
|
| 815 |
+
"train_dice": 0.8049763572205624,
|
| 816 |
+
"train_iou": 0.710061384210109,
|
| 817 |
+
"grad_norm": 3.7980856692144114,
|
| 818 |
+
"lr": 0.001400162151181607,
|
| 819 |
+
"encoder_lr": 0.00014272904961030417,
|
| 820 |
+
"alpha": 0.75355064868927,
|
| 821 |
+
"validated_this_epoch": false,
|
| 822 |
+
"val_loss": null,
|
| 823 |
+
"val_dice": null,
|
| 824 |
+
"val_iou": null,
|
| 825 |
+
"val_reward": null,
|
| 826 |
+
"val_entropy": null
|
| 827 |
+
},
|
| 828 |
+
{
|
| 829 |
+
"epoch": 38,
|
| 830 |
+
"train_loss": 0.23498452874121914,
|
| 831 |
+
"train_actor_loss": -0.30966703845452737,
|
| 832 |
+
"train_critic_loss": 0.0002865542463869027,
|
| 833 |
+
"train_mean_reward": -0.00016223431195838258,
|
| 834 |
+
"train_entropy": 0.3629140889209524,
|
| 835 |
+
"train_ce_loss": 0.3516756117606213,
|
| 836 |
+
"train_dice_loss": 0.7373409640111357,
|
| 837 |
+
"train_dice": 0.8047426790602573,
|
| 838 |
+
"train_iou": 0.7123635327741342,
|
| 839 |
+
"grad_norm": 5.456004706002021,
|
| 840 |
+
"lr": 0.0013712839299212553,
|
| 841 |
+
"encoder_lr": 0.00013997183250504444,
|
| 842 |
+
"alpha": 0.7326399683952332,
|
| 843 |
+
"validated_this_epoch": false,
|
| 844 |
+
"val_loss": null,
|
| 845 |
+
"val_dice": null,
|
| 846 |
+
"val_iou": null,
|
| 847 |
+
"val_reward": null,
|
| 848 |
+
"val_entropy": null
|
| 849 |
+
},
|
| 850 |
+
{
|
| 851 |
+
"epoch": 39,
|
| 852 |
+
"train_loss": 0.24540536295498613,
|
| 853 |
+
"train_actor_loss": -0.16366779255663894,
|
| 854 |
+
"train_critic_loss": 0.00027790911189908083,
|
| 855 |
+
"train_mean_reward": -0.00017703560533329003,
|
| 856 |
+
"train_entropy": 0.17281000839681085,
|
| 857 |
+
"train_ce_loss": 0.2660327444252502,
|
| 858 |
+
"train_dice_loss": 0.5518356541638535,
|
| 859 |
+
"train_dice": 0.8095257314740014,
|
| 860 |
+
"train_iou": 0.7186081626685688,
|
| 861 |
+
"grad_norm": 4.3213203387667045,
|
| 862 |
+
"lr": 0.0013420442306440656,
|
| 863 |
+
"encoder_lr": 0.0001371801024233027,
|
| 864 |
+
"alpha": 0.7366441488265991,
|
| 865 |
+
"validated_this_epoch": false,
|
| 866 |
+
"val_loss": null,
|
| 867 |
+
"val_dice": null,
|
| 868 |
+
"val_iou": null,
|
| 869 |
+
"val_reward": null,
|
| 870 |
+
"val_entropy": null
|
| 871 |
+
},
|
| 872 |
+
{
|
| 873 |
+
"epoch": 40,
|
| 874 |
+
"train_loss": 0.22665752784842888,
|
| 875 |
+
"train_actor_loss": -0.14277830604750794,
|
| 876 |
+
"train_critic_loss": 0.00027942328337015145,
|
| 877 |
+
"train_mean_reward": -0.00014903082476743626,
|
| 878 |
+
"train_entropy": 0.13653502300338344,
|
| 879 |
+
"train_ce_loss": 0.25464586519758947,
|
| 880 |
+
"train_dice_loss": 0.48394637506753585,
|
| 881 |
+
"train_dice": 0.8137895945860739,
|
| 882 |
+
"train_iou": 0.7245843358363597,
|
| 883 |
+
"grad_norm": 3.7977688589761422,
|
| 884 |
+
"lr": 0.0013124719094030732,
|
| 885 |
+
"encoder_lr": 0.00013435661446562,
|
| 886 |
+
"alpha": 0.7459734082221985,
|
| 887 |
+
"validated_this_epoch": true,
|
| 888 |
+
"val_loss": 0.7060043448935065,
|
| 889 |
+
"val_dice": 0.6944229048344814,
|
| 890 |
+
"val_iou": 0.6006167833361351,
|
| 891 |
+
"val_reward": 0.00025978301050018484,
|
| 892 |
+
"val_entropy": 0.25707643581178,
|
| 893 |
+
"val_actor_loss": 0.0002591476512301451,
|
| 894 |
+
"val_critic_loss": 1.1533491465974699e-05,
|
| 895 |
+
"val_ce_loss": 0.5313696203885054,
|
| 896 |
+
"val_dice_loss": 0.8801092210442129
|
| 897 |
+
},
|
| 898 |
+
{
|
| 899 |
+
"epoch": 41,
|
| 900 |
+
"train_loss": 0.22827605276792165,
|
| 901 |
+
"train_actor_loss": -0.13935510288516073,
|
| 902 |
+
"train_critic_loss": 0.0003057238575523488,
|
| 903 |
+
"train_mean_reward": -5.3540082862050194e-05,
|
| 904 |
+
"train_entropy": 0.1290512340594846,
|
| 905 |
+
"train_ce_loss": 0.24137248971032638,
|
| 906 |
+
"train_dice_loss": 0.4935840934283974,
|
| 907 |
+
"train_dice": 0.8110878260994983,
|
| 908 |
+
"train_iou": 0.7185375429496027,
|
| 909 |
+
"grad_norm": 3.6863142249658125,
|
| 910 |
+
"lr": 0.0012825961505090336,
|
| 911 |
+
"encoder_lr": 0.00013150415507372682,
|
| 912 |
+
"alpha": 0.7564293146133423,
|
| 913 |
+
"validated_this_epoch": false,
|
| 914 |
+
"val_loss": null,
|
| 915 |
+
"val_dice": null,
|
| 916 |
+
"val_iou": null,
|
| 917 |
+
"val_reward": null,
|
| 918 |
+
"val_entropy": null
|
| 919 |
+
},
|
| 920 |
+
{
|
| 921 |
+
"epoch": 42,
|
| 922 |
+
"train_loss": 0.2505705078750512,
|
| 923 |
+
"train_actor_loss": -0.21421807959257055,
|
| 924 |
+
"train_critic_loss": 0.00032212619224719613,
|
| 925 |
+
"train_mean_reward": -0.0001298960555058134,
|
| 926 |
+
"train_entropy": 0.22823248900116952,
|
| 927 |
+
"train_ce_loss": 0.296694058667416,
|
| 928 |
+
"train_dice_loss": 0.6325609861110225,
|
| 929 |
+
"train_dice": 0.8057127760604219,
|
| 930 |
+
"train_iou": 0.7118997799954666,
|
| 931 |
+
"grad_norm": 4.6142573592274685,
|
| 932 |
+
"lr": 0.0012524464377290307,
|
| 933 |
+
"encoder_lr": 0.00012862553928066122,
|
| 934 |
+
"alpha": 0.7529605627059937,
|
| 935 |
+
"validated_this_epoch": false,
|
| 936 |
+
"val_loss": null,
|
| 937 |
+
"val_dice": null,
|
| 938 |
+
"val_iou": null,
|
| 939 |
+
"val_reward": null,
|
| 940 |
+
"val_entropy": null
|
| 941 |
+
},
|
| 942 |
+
{
|
| 943 |
+
"epoch": 43,
|
| 944 |
+
"train_loss": 0.22293822620056308,
|
| 945 |
+
"train_actor_loss": -0.16250154380602355,
|
| 946 |
+
"train_critic_loss": 0.0002661989683860831,
|
| 947 |
+
"train_mean_reward": 1.0269669334095748e-06,
|
| 948 |
+
"train_entropy": 0.1652221239790999,
|
| 949 |
+
"train_ce_loss": 0.2388527302874026,
|
| 950 |
+
"train_dice_loss": 0.5317606046526314,
|
| 951 |
+
"train_dice": 0.8200685175565087,
|
| 952 |
+
"train_iou": 0.7313523545669718,
|
| 953 |
+
"grad_norm": 4.8361687565496725,
|
| 954 |
+
"lr": 0.0012220525251895604,
|
| 955 |
+
"encoder_lr": 0.0001257236079326716,
|
| 956 |
+
"alpha": 0.7592724561691284,
|
| 957 |
+
"validated_this_epoch": false,
|
| 958 |
+
"val_loss": null,
|
| 959 |
+
"val_dice": null,
|
| 960 |
+
"val_iou": null,
|
| 961 |
+
"val_reward": null,
|
| 962 |
+
"val_entropy": null
|
| 963 |
+
},
|
| 964 |
+
{
|
| 965 |
+
"epoch": 44,
|
| 966 |
+
"train_loss": 0.21762242550260474,
|
| 967 |
+
"train_actor_loss": -0.16106510639686414,
|
| 968 |
+
"train_critic_loss": 0.0002910420823219079,
|
| 969 |
+
"train_mean_reward": -0.00020571770305803476,
|
| 970 |
+
"train_entropy": 0.1526970639737631,
|
| 971 |
+
"train_ce_loss": 0.24921473733977315,
|
| 972 |
+
"train_dice_loss": 0.5078692797388525,
|
| 973 |
+
"train_dice": 0.816686018552457,
|
| 974 |
+
"train_iou": 0.7243800738289251,
|
| 975 |
+
"grad_norm": 3.2378107840700667,
|
| 976 |
+
"lr": 0.0011914444080127966,
|
| 977 |
+
"encoder_lr": 0.0001228012248856439,
|
| 978 |
+
"alpha": 0.7673797011375427,
|
| 979 |
+
"validated_this_epoch": false,
|
| 980 |
+
"val_loss": null,
|
| 981 |
+
"val_dice": null,
|
| 982 |
+
"val_iou": null,
|
| 983 |
+
"val_reward": null,
|
| 984 |
+
"val_entropy": null
|
| 985 |
+
},
|
| 986 |
+
{
|
| 987 |
+
"epoch": 45,
|
| 988 |
+
"train_loss": 0.19360105940993594,
|
| 989 |
+
"train_actor_loss": -0.1493051653416245,
|
| 990 |
+
"train_critic_loss": 0.00030364165918202694,
|
| 991 |
+
"train_mean_reward": -0.00015405603163739316,
|
| 992 |
+
"train_entropy": 0.1412152343148965,
|
| 993 |
+
"train_ce_loss": 0.22449993756804273,
|
| 994 |
+
"train_dice_loss": 0.46100886709006256,
|
| 995 |
+
"train_dice": 0.8279386565064007,
|
| 996 |
+
"train_iou": 0.7362816179990674,
|
| 997 |
+
"grad_norm": 3.42305705459543,
|
| 998 |
+
"lr": 0.0011606522927150302,
|
| 999 |
+
"encoder_lr": 0.00011986127417882199,
|
| 1000 |
+
"alpha": 0.7759975790977478,
|
| 1001 |
+
"validated_this_epoch": false,
|
| 1002 |
+
"val_loss": null,
|
| 1003 |
+
"val_dice": null,
|
| 1004 |
+
"val_iou": null,
|
| 1005 |
+
"val_reward": null,
|
| 1006 |
+
"val_entropy": null
|
| 1007 |
+
},
|
| 1008 |
+
{
|
| 1009 |
+
"epoch": 46,
|
| 1010 |
+
"train_loss": 0.19809391227143125,
|
| 1011 |
+
"train_actor_loss": -0.15470215332150178,
|
| 1012 |
+
"train_critic_loss": 0.00027187143541509723,
|
| 1013 |
+
"train_mean_reward": -1.955718293546774e-05,
|
| 1014 |
+
"train_entropy": 0.149146339762986,
|
| 1015 |
+
"train_ce_loss": 0.22930508960779636,
|
| 1016 |
+
"train_dice_loss": 0.4760151647936158,
|
| 1017 |
+
"train_dice": 0.8210904568455468,
|
| 1018 |
+
"train_iou": 0.7321297108166892,
|
| 1019 |
+
"grad_norm": 3.2063551235568615,
|
| 1020 |
+
"lr": 0.0011297065673964834,
|
| 1021 |
+
"encoder_lr": 0.00011690665718860897,
|
| 1022 |
+
"alpha": 0.7846786975860596,
|
| 1023 |
+
"validated_this_epoch": false,
|
| 1024 |
+
"val_loss": null,
|
| 1025 |
+
"val_dice": null,
|
| 1026 |
+
"val_iou": null,
|
| 1027 |
+
"val_reward": null,
|
| 1028 |
+
"val_entropy": null
|
| 1029 |
+
},
|
| 1030 |
+
{
|
| 1031 |
+
"epoch": 47,
|
| 1032 |
+
"train_loss": 0.20463162070501673,
|
| 1033 |
+
"train_actor_loss": -0.15526240608337255,
|
| 1034 |
+
"train_critic_loss": 0.00030984192344802176,
|
| 1035 |
+
"train_mean_reward": -0.0001875786568467278,
|
| 1036 |
+
"train_entropy": 0.14836972930531186,
|
| 1037 |
+
"train_ce_loss": 0.247041476594443,
|
| 1038 |
+
"train_dice_loss": 0.47243673194594465,
|
| 1039 |
+
"train_dice": 0.8252368338421961,
|
| 1040 |
+
"train_iou": 0.7349258905130092,
|
| 1041 |
+
"grad_norm": 3.315960697656454,
|
| 1042 |
+
"lr": 0.001098637771751922,
|
| 1043 |
+
"encoder_lr": 0.0001139402897652589,
|
| 1044 |
+
"alpha": 0.7939295768737793,
|
| 1045 |
+
"validated_this_epoch": false,
|
| 1046 |
+
"val_loss": null,
|
| 1047 |
+
"val_dice": null,
|
| 1048 |
+
"val_iou": null,
|
| 1049 |
+
"val_reward": null,
|
| 1050 |
+
"val_entropy": null
|
| 1051 |
+
},
|
| 1052 |
+
{
|
| 1053 |
+
"epoch": 48,
|
| 1054 |
+
"train_loss": 0.18540352444162897,
|
| 1055 |
+
"train_actor_loss": -0.3196820626853549,
|
| 1056 |
+
"train_critic_loss": 0.0002807973353104751,
|
| 1057 |
+
"train_mean_reward": -0.0001917376270877105,
|
| 1058 |
+
"train_entropy": 0.3533610503913628,
|
| 1059 |
+
"train_ce_loss": 0.3237076615613888,
|
| 1060 |
+
"train_dice_loss": 0.686182710782502,
|
| 1061 |
+
"train_dice": 0.8303664617548598,
|
| 1062 |
+
"train_iou": 0.740450200784089,
|
| 1063 |
+
"grad_norm": 2.572279711333356,
|
| 1064 |
+
"lr": 0.0010674765669316673,
|
| 1065 |
+
"encoder_lr": 0.00011096509935528483,
|
| 1066 |
+
"alpha": 0.7756125926971436,
|
| 1067 |
+
"validated_this_epoch": false,
|
| 1068 |
+
"val_loss": null,
|
| 1069 |
+
"val_dice": null,
|
| 1070 |
+
"val_iou": null,
|
| 1071 |
+
"val_reward": null,
|
| 1072 |
+
"val_entropy": null
|
| 1073 |
+
},
|
| 1074 |
+
{
|
| 1075 |
+
"epoch": 49,
|
| 1076 |
+
"train_loss": 0.1961687517927936,
|
| 1077 |
+
"train_actor_loss": -0.35685843416039636,
|
| 1078 |
+
"train_critic_loss": 0.00023025654816407357,
|
| 1079 |
+
"train_mean_reward": -0.0001445000375056683,
|
| 1080 |
+
"train_entropy": 0.4168141348147885,
|
| 1081 |
+
"train_ce_loss": 0.35553827329998405,
|
| 1082 |
+
"train_dice_loss": 0.7502858367693209,
|
| 1083 |
+
"train_dice": 0.8250803735152323,
|
| 1084 |
+
"train_iou": 0.7364160488065451,
|
| 1085 |
+
"grad_norm": 2.7564951251874596,
|
| 1086 |
+
"lr": 0.001036253705282738,
|
| 1087 |
+
"encoder_lr": 0.00010798402211242223,
|
| 1088 |
+
"alpha": 0.7476118206977844,
|
| 1089 |
+
"validated_this_epoch": false,
|
| 1090 |
+
"val_loss": null,
|
| 1091 |
+
"val_dice": null,
|
| 1092 |
+
"val_iou": null,
|
| 1093 |
+
"val_reward": null,
|
| 1094 |
+
"val_entropy": null
|
| 1095 |
+
},
|
| 1096 |
+
{
|
| 1097 |
+
"epoch": 50,
|
| 1098 |
+
"train_loss": 0.20967349763571314,
|
| 1099 |
+
"train_actor_loss": -0.3224664660079226,
|
| 1100 |
+
"train_critic_loss": 0.00020030396571885746,
|
| 1101 |
+
"train_mean_reward": 5.219104939154434e-06,
|
| 1102 |
+
"train_entropy": 0.39053801337415855,
|
| 1103 |
+
"train_ce_loss": 0.3344985051309204,
|
| 1104 |
+
"train_dice_loss": 0.7295811151657301,
|
| 1105 |
+
"train_dice": 0.8311792915010486,
|
| 1106 |
+
"train_iou": 0.7435864765531297,
|
| 1107 |
+
"grad_norm": 4.13983164092367,
|
| 1108 |
+
"lr": 0.0010050000000000003,
|
| 1109 |
+
"encoder_lr": 0.00010500000000000005,
|
| 1110 |
+
"alpha": 0.7262772917747498,
|
| 1111 |
+
"validated_this_epoch": true,
|
| 1112 |
+
"val_loss": 0.8986567843416141,
|
| 1113 |
+
"val_dice": 0.6819966748084384,
|
| 1114 |
+
"val_iou": 0.5892341475374794,
|
| 1115 |
+
"val_reward": 0.00013612646343826097,
|
| 1116 |
+
"val_entropy": 0.6055493098918838,
|
| 1117 |
+
"val_actor_loss": -0.06335509269535919,
|
| 1118 |
+
"val_critic_loss": 7.196651630253122e-06,
|
| 1119 |
+
"val_ce_loss": 0.7981973983220417,
|
| 1120 |
+
"val_dice_loss": 1.1258191210452957
|
| 1121 |
+
},
|
| 1122 |
+
{
|
| 1123 |
+
"epoch": 51,
|
| 1124 |
+
"train_loss": 0.2152879249062833,
|
| 1125 |
+
"train_actor_loss": -0.22433636481187835,
|
| 1126 |
+
"train_critic_loss": 0.0002474339772373571,
|
| 1127 |
+
"train_mean_reward": -0.00011234690190663614,
|
| 1128 |
+
"train_entropy": 0.258694730780884,
|
| 1129 |
+
"train_ce_loss": 0.2758197087183286,
|
| 1130 |
+
"train_dice_loss": 0.6031814314598261,
|
| 1131 |
+
"train_dice": 0.8281835404755985,
|
| 1132 |
+
"train_iou": 0.7396459541424982,
|
| 1133 |
+
"grad_norm": 3.171783290399137,
|
| 1134 |
+
"lr": 0.0009737462947172626,
|
| 1135 |
+
"encoder_lr": 0.00010201597788757786,
|
| 1136 |
+
"alpha": 0.7190115451812744,
|
| 1137 |
+
"validated_this_epoch": false,
|
| 1138 |
+
"val_loss": null,
|
| 1139 |
+
"val_dice": null,
|
| 1140 |
+
"val_iou": null,
|
| 1141 |
+
"val_reward": null,
|
| 1142 |
+
"val_entropy": null
|
| 1143 |
+
},
|
| 1144 |
+
{
|
| 1145 |
+
"epoch": 52,
|
| 1146 |
+
"train_loss": 0.21913143885931347,
|
| 1147 |
+
"train_actor_loss": -0.12770263460502077,
|
| 1148 |
+
"train_critic_loss": 0.00027524231112191044,
|
| 1149 |
+
"train_mean_reward": -0.0001909137164777809,
|
| 1150 |
+
"train_entropy": 0.12714165257237664,
|
| 1151 |
+
"train_ce_loss": 0.22825219654099105,
|
| 1152 |
+
"train_dice_loss": 0.46514070234249427,
|
| 1153 |
+
"train_dice": 0.8271630719840092,
|
| 1154 |
+
"train_iou": 0.7399610596129539,
|
| 1155 |
+
"grad_norm": 3.5046284820220266,
|
| 1156 |
+
"lr": 0.0009425234330683334,
|
| 1157 |
+
"encoder_lr": 9.903490064471527e-05,
|
| 1158 |
+
"alpha": 0.727668821811676,
|
| 1159 |
+
"validated_this_epoch": false,
|
| 1160 |
+
"val_loss": null,
|
| 1161 |
+
"val_dice": null,
|
| 1162 |
+
"val_iou": null,
|
| 1163 |
+
"val_reward": null,
|
| 1164 |
+
"val_entropy": null
|
| 1165 |
+
},
|
| 1166 |
+
{
|
| 1167 |
+
"epoch": 53,
|
| 1168 |
+
"train_loss": 0.1948301377060136,
|
| 1169 |
+
"train_actor_loss": -0.13600847273475422,
|
| 1170 |
+
"train_critic_loss": 0.0003649366933275839,
|
| 1171 |
+
"train_mean_reward": -0.00023356911426576315,
|
| 1172 |
+
"train_entropy": 0.13049386871173355,
|
| 1173 |
+
"train_ce_loss": 0.21625699299310627,
|
| 1174 |
+
"train_dice_loss": 0.4450552883998368,
|
| 1175 |
+
"train_dice": 0.8268062177808664,
|
| 1176 |
+
"train_iou": 0.7378000813584671,
|
| 1177 |
+
"grad_norm": 3.183339916920477,
|
| 1178 |
+
"lr": 0.0009113622282480784,
|
| 1179 |
+
"encoder_lr": 9.605971023474117e-05,
|
| 1180 |
+
"alpha": 0.7380543351173401,
|
| 1181 |
+
"validated_this_epoch": false,
|
| 1182 |
+
"val_loss": null,
|
| 1183 |
+
"val_dice": null,
|
| 1184 |
+
"val_iou": null,
|
| 1185 |
+
"val_reward": null,
|
| 1186 |
+
"val_entropy": null
|
| 1187 |
+
},
|
| 1188 |
+
{
|
| 1189 |
+
"epoch": 54,
|
| 1190 |
+
"train_loss": 0.1766825389237215,
|
| 1191 |
+
"train_actor_loss": -0.12638184072057843,
|
| 1192 |
+
"train_critic_loss": 0.00022454364877774998,
|
| 1193 |
+
"train_mean_reward": 4.699535979300677e-05,
|
| 1194 |
+
"train_entropy": 0.12382035084668601,
|
| 1195 |
+
"train_ce_loss": 0.20927528819810476,
|
| 1196 |
+
"train_dice_loss": 0.39662892381042164,
|
| 1197 |
+
"train_dice": 0.8438913834393494,
|
| 1198 |
+
"train_iou": 0.7551607306877318,
|
| 1199 |
+
"grad_norm": 2.4921159205972687,
|
| 1200 |
+
"lr": 0.0008802934326035174,
|
| 1201 |
+
"encoder_lr": 9.309334281139113e-05,
|
| 1202 |
+
"alpha": 0.7475517988204956,
|
| 1203 |
+
"validated_this_epoch": false,
|
| 1204 |
+
"val_loss": null,
|
| 1205 |
+
"val_dice": null,
|
| 1206 |
+
"val_iou": null,
|
| 1207 |
+
"val_reward": null,
|
| 1208 |
+
"val_entropy": null
|
| 1209 |
+
},
|
| 1210 |
+
{
|
| 1211 |
+
"epoch": 55,
|
| 1212 |
+
"train_loss": 0.18376757142313316,
|
| 1213 |
+
"train_actor_loss": -0.13605426495834028,
|
| 1214 |
+
"train_critic_loss": 0.0002440471370203707,
|
| 1215 |
+
"train_mean_reward": 2.7349599607616022e-05,
|
| 1216 |
+
"train_entropy": 0.13277763510300375,
|
| 1217 |
+
"train_ce_loss": 0.20885094217607122,
|
| 1218 |
+
"train_dice_loss": 0.43054867876592534,
|
| 1219 |
+
"train_dice": 0.8343481483428782,
|
| 1220 |
+
"train_iou": 0.7474028473200067,
|
| 1221 |
+
"grad_norm": 2.911448457444361,
|
| 1222 |
+
"lr": 0.0008493477072849704,
|
| 1223 |
+
"encoder_lr": 9.013872582117809e-05,
|
| 1224 |
+
"alpha": 0.7576013207435608,
|
| 1225 |
+
"validated_this_epoch": false,
|
| 1226 |
+
"val_loss": null,
|
| 1227 |
+
"val_dice": null,
|
| 1228 |
+
"val_iou": null,
|
| 1229 |
+
"val_reward": null,
|
| 1230 |
+
"val_entropy": null
|
| 1231 |
+
},
|
| 1232 |
+
{
|
| 1233 |
+
"epoch": 56,
|
| 1234 |
+
"train_loss": 0.19156841376940248,
|
| 1235 |
+
"train_actor_loss": -0.14204173592331057,
|
| 1236 |
+
"train_critic_loss": 0.0002921382484599291,
|
| 1237 |
+
"train_mean_reward": -2.301977274833514e-05,
|
| 1238 |
+
"train_entropy": 0.13757345951926347,
|
| 1239 |
+
"train_ce_loss": 0.22215836897230098,
|
| 1240 |
+
"train_dice_loss": 0.44476978661786065,
|
| 1241 |
+
"train_dice": 0.8318472642130662,
|
| 1242 |
+
"train_iou": 0.7431191569140458,
|
| 1243 |
+
"grad_norm": 3.7214020546092543,
|
| 1244 |
+
"lr": 0.0008185555919872042,
|
| 1245 |
+
"encoder_lr": 8.719877511435618e-05,
|
| 1246 |
+
"alpha": 0.766362190246582,
|
| 1247 |
+
"validated_this_epoch": false,
|
| 1248 |
+
"val_loss": null,
|
| 1249 |
+
"val_dice": null,
|
| 1250 |
+
"val_iou": null,
|
| 1251 |
+
"val_reward": null,
|
| 1252 |
+
"val_entropy": null
|
| 1253 |
+
},
|
| 1254 |
+
{
|
| 1255 |
+
"epoch": 57,
|
| 1256 |
+
"train_loss": 0.1910335316662627,
|
| 1257 |
+
"train_actor_loss": -0.1534685715527613,
|
| 1258 |
+
"train_critic_loss": 0.00025699430291346387,
|
| 1259 |
+
"train_mean_reward": 4.7306957105451796e-05,
|
| 1260 |
+
"train_entropy": 0.15236955194493987,
|
| 1261 |
+
"train_ce_loss": 0.22049266224948433,
|
| 1262 |
+
"train_dice_loss": 0.4682545470174893,
|
| 1263 |
+
"train_dice": 0.8356843691367334,
|
| 1264 |
+
"train_iou": 0.7508747995227614,
|
| 1265 |
+
"grad_norm": 3.7364069031652556,
|
| 1266 |
+
"lr": 0.0007879474748104406,
|
| 1267 |
+
"encoder_lr": 8.42763920673285e-05,
|
| 1268 |
+
"alpha": 0.7737754583358765,
|
| 1269 |
+
"validated_this_epoch": false,
|
| 1270 |
+
"val_loss": null,
|
| 1271 |
+
"val_dice": null,
|
| 1272 |
+
"val_iou": null,
|
| 1273 |
+
"val_reward": null,
|
| 1274 |
+
"val_entropy": null
|
| 1275 |
+
},
|
| 1276 |
+
{
|
| 1277 |
+
"epoch": 58,
|
| 1278 |
+
"train_loss": 0.1703913249243775,
|
| 1279 |
+
"train_actor_loss": -0.15045072542529345,
|
| 1280 |
+
"train_critic_loss": 0.00025059190623428096,
|
| 1281 |
+
"train_mean_reward": -1.7067346137620663e-05,
|
| 1282 |
+
"train_entropy": 0.14288238203566633,
|
| 1283 |
+
"train_ce_loss": 0.21381499299849321,
|
| 1284 |
+
"train_dice_loss": 0.42761851234953535,
|
| 1285 |
+
"train_dice": 0.842017969582979,
|
| 1286 |
+
"train_iou": 0.7527994345628402,
|
| 1287 |
+
"grad_norm": 2.900194777536762,
|
| 1288 |
+
"lr": 0.0007575535622709696,
|
| 1289 |
+
"encoder_lr": 8.13744607193388e-05,
|
| 1290 |
+
"alpha": 0.783578097820282,
|
| 1291 |
+
"validated_this_epoch": false,
|
| 1292 |
+
"val_loss": null,
|
| 1293 |
+
"val_dice": null,
|
| 1294 |
+
"val_iou": null,
|
| 1295 |
+
"val_reward": null,
|
| 1296 |
+
"val_entropy": null
|
| 1297 |
+
},
|
| 1298 |
+
{
|
| 1299 |
+
"epoch": 59,
|
| 1300 |
+
"train_loss": 0.19329094379417108,
|
| 1301 |
+
"train_actor_loss": -0.1571274292148416,
|
| 1302 |
+
"train_critic_loss": 0.0002410241053144829,
|
| 1303 |
+
"train_mean_reward": 1.0436220464403958e-05,
|
| 1304 |
+
"train_entropy": 0.15346740127683411,
|
| 1305 |
+
"train_ce_loss": 0.22447862494279672,
|
| 1306 |
+
"train_dice_loss": 0.4761170938957569,
|
| 1307 |
+
"train_dice": 0.8365017596261244,
|
| 1308 |
+
"train_iou": 0.7529517564205035,
|
| 1309 |
+
"grad_norm": 3.645170865132827,
|
| 1310 |
+
"lr": 0.0007274038494909672,
|
| 1311 |
+
"encoder_lr": 7.849584492627325e-05,
|
| 1312 |
+
"alpha": 0.7916308045387268,
|
| 1313 |
+
"validated_this_epoch": false,
|
| 1314 |
+
"val_loss": null,
|
| 1315 |
+
"val_dice": null,
|
| 1316 |
+
"val_iou": null,
|
| 1317 |
+
"val_reward": null,
|
| 1318 |
+
"val_entropy": null
|
| 1319 |
+
},
|
| 1320 |
+
{
|
| 1321 |
+
"epoch": 60,
|
| 1322 |
+
"train_loss": 0.16571185784266354,
|
| 1323 |
+
"train_actor_loss": -0.16073855055476244,
|
| 1324 |
+
"train_critic_loss": 0.00023862707734120393,
|
| 1325 |
+
"train_mean_reward": 8.963804767572067e-05,
|
| 1326 |
+
"train_entropy": 0.15593899034405279,
|
| 1327 |
+
"train_ce_loss": 0.21485867297748534,
|
| 1328 |
+
"train_dice_loss": 0.437803514706072,
|
| 1329 |
+
"train_dice": 0.8502178646112836,
|
| 1330 |
+
"train_iou": 0.7608532948904497,
|
| 1331 |
+
"grad_norm": 2.706126112808553,
|
| 1332 |
+
"lr": 0.0006975280905969278,
|
| 1333 |
+
"encoder_lr": 7.564338553438004e-05,
|
| 1334 |
+
"alpha": 0.7988956570625305,
|
| 1335 |
+
"validated_this_epoch": true,
|
| 1336 |
+
"val_loss": 0.6635697757627702,
|
| 1337 |
+
"val_dice": 0.6951244897142669,
|
| 1338 |
+
"val_iou": 0.6034134250515542,
|
| 1339 |
+
"val_reward": 6.882173567842903e-05,
|
| 1340 |
+
"val_entropy": 0.21648115364627704,
|
| 1341 |
+
"val_actor_loss": 0.003797827805027911,
|
| 1342 |
+
"val_critic_loss": 2.2509576867568827e-05,
|
| 1343 |
+
"val_ce_loss": 0.48172503826913954,
|
| 1344 |
+
"val_dice_loss": 0.8377963355695358
|
| 1345 |
+
},
|
| 1346 |
+
{
|
| 1347 |
+
"epoch": 61,
|
| 1348 |
+
"train_loss": 0.19566198594459772,
|
| 1349 |
+
"train_actor_loss": -0.16466724708930827,
|
| 1350 |
+
"train_critic_loss": 0.0002085545418287166,
|
| 1351 |
+
"train_mean_reward": -9.744621296732482e-05,
|
| 1352 |
+
"train_entropy": 0.15808398778627336,
|
| 1353 |
+
"train_ce_loss": 0.24173496568212116,
|
| 1354 |
+
"train_dice_loss": 0.4787149426240946,
|
| 1355 |
+
"train_dice": 0.834585681473012,
|
| 1356 |
+
"train_iou": 0.7474598492282799,
|
| 1357 |
+
"grad_norm": 3.3483349879582724,
|
| 1358 |
+
"lr": 0.0006679557693559353,
|
| 1359 |
+
"encoder_lr": 7.281989757669733e-05,
|
| 1360 |
+
"alpha": 0.8069456815719604,
|
| 1361 |
+
"validated_this_epoch": false,
|
| 1362 |
+
"val_loss": null,
|
| 1363 |
+
"val_dice": null,
|
| 1364 |
+
"val_iou": null,
|
| 1365 |
+
"val_reward": null,
|
| 1366 |
+
"val_entropy": null
|
| 1367 |
+
},
|
| 1368 |
+
{
|
| 1369 |
+
"epoch": 62,
|
| 1370 |
+
"train_loss": 0.1767019698157526,
|
| 1371 |
+
"train_actor_loss": -0.1660022865992896,
|
| 1372 |
+
"train_critic_loss": 0.00021094280346555213,
|
| 1373 |
+
"train_mean_reward": 6.161623775383761e-05,
|
| 1374 |
+
"train_entropy": 0.16003943242683757,
|
| 1375 |
+
"train_ce_loss": 0.22611310016616543,
|
| 1376 |
+
"train_dice_loss": 0.4590844673703807,
|
| 1377 |
+
"train_dice": 0.8376886267486875,
|
| 1378 |
+
"train_iou": 0.7503941471512698,
|
| 1379 |
+
"grad_norm": 3.0423298339511073,
|
| 1380 |
+
"lr": 0.000638716070078746,
|
| 1381 |
+
"encoder_lr": 7.002816749495564e-05,
|
| 1382 |
+
"alpha": 0.8135802149772644,
|
| 1383 |
+
"validated_this_epoch": false,
|
| 1384 |
+
"val_loss": null,
|
| 1385 |
+
"val_dice": null,
|
| 1386 |
+
"val_iou": null,
|
| 1387 |
+
"val_reward": null,
|
| 1388 |
+
"val_entropy": null
|
| 1389 |
+
},
|
| 1390 |
+
{
|
| 1391 |
+
"epoch": 63,
|
| 1392 |
+
"train_loss": 0.15656305300764461,
|
| 1393 |
+
"train_actor_loss": -0.24940603538112507,
|
| 1394 |
+
"train_critic_loss": 0.0002685534220951836,
|
| 1395 |
+
"train_mean_reward": -0.00016037248998646093,
|
| 1396 |
+
"train_entropy": 0.25489531159264067,
|
| 1397 |
+
"train_ce_loss": 0.2665173671495901,
|
| 1398 |
+
"train_dice_loss": 0.5451522535747951,
|
| 1399 |
+
"train_dice": 0.8508806595961862,
|
| 1400 |
+
"train_iou": 0.7618729682766019,
|
| 1401 |
+
"grad_norm": 2.50465609953385,
|
| 1402 |
+
"lr": 0.0006098378488183934,
|
| 1403 |
+
"encoder_lr": 6.727095038969585e-05,
|
| 1404 |
+
"alpha": 0.8100873827934265,
|
| 1405 |
+
"validated_this_epoch": false,
|
| 1406 |
+
"val_loss": null,
|
| 1407 |
+
"val_dice": null,
|
| 1408 |
+
"val_iou": null,
|
| 1409 |
+
"val_reward": null,
|
| 1410 |
+
"val_entropy": null
|
| 1411 |
+
},
|
| 1412 |
+
{
|
| 1413 |
+
"epoch": 64,
|
| 1414 |
+
"train_loss": 0.16336153899500241,
|
| 1415 |
+
"train_actor_loss": -0.3880823067069362,
|
| 1416 |
+
"train_critic_loss": 0.0002281419488079333,
|
| 1417 |
+
"train_mean_reward": -6.659597557025698e-05,
|
| 1418 |
+
"train_entropy": 0.44060070600928575,
|
| 1419 |
+
"train_ce_loss": 0.34770322979237756,
|
| 1420 |
+
"train_dice_loss": 0.7549563162702614,
|
| 1421 |
+
"train_dice": 0.8377770572230382,
|
| 1422 |
+
"train_iou": 0.7489864436237361,
|
| 1423 |
+
"grad_norm": 2.691737491031026,
|
| 1424 |
+
"lr": 0.0005813496048927529,
|
| 1425 |
+
"encoder_lr": 6.455096730131812e-05,
|
| 1426 |
+
"alpha": 0.7785568833351135,
|
| 1427 |
+
"validated_this_epoch": false,
|
| 1428 |
+
"val_loss": null,
|
| 1429 |
+
"val_dice": null,
|
| 1430 |
+
"val_iou": null,
|
| 1431 |
+
"val_reward": null,
|
| 1432 |
+
"val_entropy": null
|
| 1433 |
+
},
|
| 1434 |
+
{
|
| 1435 |
+
"epoch": 65,
|
| 1436 |
+
"train_loss": 0.17650862792669253,
|
| 1437 |
+
"train_actor_loss": -0.3694659742967103,
|
| 1438 |
+
"train_critic_loss": 0.00021740543091777702,
|
| 1439 |
+
"train_mean_reward": 2.112583727816866e-05,
|
| 1440 |
+
"train_entropy": 0.43385662330983527,
|
| 1441 |
+
"train_ce_loss": 0.34690162986011586,
|
| 1442 |
+
"train_dice_loss": 0.7448301661846249,
|
| 1443 |
+
"train_dice": 0.8411768975851796,
|
| 1444 |
+
"train_iou": 0.7545056907882479,
|
| 1445 |
+
"grad_norm": 2.24325423208318,
|
| 1446 |
+
"lr": 0.0005532794527591512,
|
| 1447 |
+
"encoder_lr": 6.187090252474308e-05,
|
| 1448 |
+
"alpha": 0.7512552738189697,
|
| 1449 |
+
"validated_this_epoch": false,
|
| 1450 |
+
"val_loss": null,
|
| 1451 |
+
"val_dice": null,
|
| 1452 |
+
"val_iou": null,
|
| 1453 |
+
"val_reward": null,
|
| 1454 |
+
"val_entropy": null
|
| 1455 |
+
},
|
| 1456 |
+
{
|
| 1457 |
+
"epoch": 66,
|
| 1458 |
+
"train_loss": 0.16945145106494816,
|
| 1459 |
+
"train_actor_loss": -0.33264579106054876,
|
| 1460 |
+
"train_critic_loss": 0.00019288668677982658,
|
| 1461 |
+
"train_mean_reward": 6.140212775753706e-05,
|
| 1462 |
+
"train_entropy": 0.3996078079176504,
|
| 1463 |
+
"train_ce_loss": 0.30971403589871654,
|
| 1464 |
+
"train_dice_loss": 0.6942875587817003,
|
| 1465 |
+
"train_dice": 0.8501023952673218,
|
| 1466 |
+
"train_iou": 0.7640104052839,
|
| 1467 |
+
"grad_norm": 1.9354052120863006,
|
| 1468 |
+
"lr": 0.0005256550942687934,
|
| 1469 |
+
"encoder_lr": 5.9233400960337056e-05,
|
| 1470 |
+
"alpha": 0.7295467257499695,
|
| 1471 |
+
"validated_this_epoch": false,
|
| 1472 |
+
"val_loss": null,
|
| 1473 |
+
"val_dice": null,
|
| 1474 |
+
"val_iou": null,
|
| 1475 |
+
"val_reward": null,
|
| 1476 |
+
"val_entropy": null
|
| 1477 |
+
},
|
| 1478 |
+
{
|
| 1479 |
+
"epoch": 67,
|
| 1480 |
+
"train_loss": 0.19790268460360672,
|
| 1481 |
+
"train_actor_loss": -0.3114694067077274,
|
| 1482 |
+
"train_critic_loss": 0.00030121911855226217,
|
| 1483 |
+
"train_mean_reward": -0.00017128807249150376,
|
| 1484 |
+
"train_entropy": 0.3746029567117839,
|
| 1485 |
+
"train_ce_loss": 0.3135900703458554,
|
| 1486 |
+
"train_dice_loss": 0.7048528903676557,
|
| 1487 |
+
"train_dice": 0.8392508010410212,
|
| 1488 |
+
"train_iou": 0.7507586108401063,
|
| 1489 |
+
"grad_norm": 2.4681112499431124,
|
| 1490 |
+
"lr": 0.0004985037913283807,
|
| 1491 |
+
"encoder_lr": 5.6641065503714745e-05,
|
| 1492 |
+
"alpha": 0.7107322216033936,
|
| 1493 |
+
"validated_this_epoch": false,
|
| 1494 |
+
"val_loss": null,
|
| 1495 |
+
"val_dice": null,
|
| 1496 |
+
"val_iou": null,
|
| 1497 |
+
"val_reward": null,
|
| 1498 |
+
"val_entropy": null
|
| 1499 |
+
},
|
| 1500 |
+
{
|
| 1501 |
+
"epoch": 68,
|
| 1502 |
+
"train_loss": 0.1744239007847675,
|
| 1503 |
+
"train_actor_loss": -0.10339116890197597,
|
| 1504 |
+
"train_critic_loss": 0.00017886468952056203,
|
| 1505 |
+
"train_mean_reward": 3.129754450303227e-05,
|
| 1506 |
+
"train_entropy": 0.10293221994945707,
|
| 1507 |
+
"train_ce_loss": 0.18557560108168875,
|
| 1508 |
+
"train_dice_loss": 0.3698756709425332,
|
| 1509 |
+
"train_dice": 0.8506117135623796,
|
| 1510 |
+
"train_iou": 0.7690501950903161,
|
| 1511 |
+
"grad_norm": 2.109886308742124,
|
| 1512 |
+
"lr": 0.0004718523389958983,
|
| 1513 |
+
"encoder_lr": 5.4096454476995315e-05,
|
| 1514 |
+
"alpha": 0.7208677530288696,
|
| 1515 |
+
"validated_this_epoch": false,
|
| 1516 |
+
"val_loss": null,
|
| 1517 |
+
"val_dice": null,
|
| 1518 |
+
"val_iou": null,
|
| 1519 |
+
"val_reward": null,
|
| 1520 |
+
"val_entropy": null
|
| 1521 |
+
},
|
| 1522 |
+
{
|
| 1523 |
+
"epoch": 69,
|
| 1524 |
+
"train_loss": 0.16875991365795504,
|
| 1525 |
+
"train_actor_loss": -0.1146977549394441,
|
| 1526 |
+
"train_critic_loss": 0.0002089566750811584,
|
| 1527 |
+
"train_mean_reward": 4.675462931127491e-05,
|
| 1528 |
+
"train_entropy": 0.1083037691476312,
|
| 1529 |
+
"train_ce_loss": 0.19058924686832676,
|
| 1530 |
+
"train_dice_loss": 0.3761171311987154,
|
| 1531 |
+
"train_dice": 0.8560199877292133,
|
| 1532 |
+
"train_iou": 0.7718941560926348,
|
| 1533 |
+
"grad_norm": 2.2549599707126617,
|
| 1534 |
+
"lr": 0.0004457270390371302,
|
| 1535 |
+
"encoder_lr": 5.1602079104047604e-05,
|
| 1536 |
+
"alpha": 0.7328778505325317,
|
| 1537 |
+
"validated_this_epoch": false,
|
| 1538 |
+
"val_loss": null,
|
| 1539 |
+
"val_dice": null,
|
| 1540 |
+
"val_iou": null,
|
| 1541 |
+
"val_reward": null,
|
| 1542 |
+
"val_entropy": null
|
| 1543 |
+
},
|
| 1544 |
+
{
|
| 1545 |
+
"epoch": 70,
|
| 1546 |
+
"train_loss": 0.17815924153040588,
|
| 1547 |
+
"train_actor_loss": -0.1275307896119261,
|
| 1548 |
+
"train_critic_loss": 0.00019872413947868224,
|
| 1549 |
+
"train_mean_reward": 0.00016344842689327625,
|
| 1550 |
+
"train_entropy": 0.12876040197368013,
|
| 1551 |
+
"train_ce_loss": 0.20345089225837293,
|
| 1552 |
+
"train_dice_loss": 0.4077304424549566,
|
| 1553 |
+
"train_dice": 0.8520825587866656,
|
| 1554 |
+
"train_iou": 0.7655679997176092,
|
| 1555 |
+
"grad_norm": 2.7472325283889623,
|
| 1556 |
+
"lr": 0.00042015367396898955,
|
| 1557 |
+
"encoder_lr": 4.916040103221508e-05,
|
| 1558 |
+
"alpha": 0.7412089109420776,
|
| 1559 |
+
"validated_this_epoch": true,
|
| 1560 |
+
"val_loss": 0.6863508209670809,
|
| 1561 |
+
"val_dice": 0.6821482281002661,
|
| 1562 |
+
"val_iou": 0.5939569225766711,
|
| 1563 |
+
"val_reward": 6.355834034394817e-06,
|
| 1564 |
+
"val_entropy": 0.22547387715541956,
|
| 1565 |
+
"val_actor_loss": -6.967572779912073e-05,
|
| 1566 |
+
"val_critic_loss": 3.0113368522519427e-05,
|
| 1567 |
+
"val_ce_loss": 0.5027745388979751,
|
| 1568 |
+
"val_dice_loss": 0.8700363217580199
|
| 1569 |
+
},
|
| 1570 |
+
{
|
| 1571 |
+
"epoch": 71,
|
| 1572 |
+
"train_loss": 0.17333358829252343,
|
| 1573 |
+
"train_actor_loss": -0.1321150205569319,
|
| 1574 |
+
"train_critic_loss": 0.00018025401000566264,
|
| 1575 |
+
"train_mean_reward": 0.00022727838300483485,
|
| 1576 |
+
"train_entropy": 0.12977316539743025,
|
| 1577 |
+
"train_ce_loss": 0.20141242139609367,
|
| 1578 |
+
"train_dice_loss": 0.4093045394395981,
|
| 1579 |
+
"train_dice": 0.8503221106168345,
|
| 1580 |
+
"train_iou": 0.7653329082419676,
|
| 1581 |
+
"grad_norm": 2.3786626878634904,
|
| 1582 |
+
"lr": 0.0003951574816152885,
|
| 1583 |
+
"encoder_lr": 4.6773829902967244e-05,
|
| 1584 |
+
"alpha": 0.7504040002822876,
|
| 1585 |
+
"validated_this_epoch": false,
|
| 1586 |
+
"val_loss": null,
|
| 1587 |
+
"val_dice": null,
|
| 1588 |
+
"val_iou": null,
|
| 1589 |
+
"val_reward": null,
|
| 1590 |
+
"val_entropy": null
|
| 1591 |
+
},
|
| 1592 |
+
{
|
| 1593 |
+
"epoch": 72,
|
| 1594 |
+
"train_loss": 0.16332370126747087,
|
| 1595 |
+
"train_actor_loss": -0.14536811089523988,
|
| 1596 |
+
"train_critic_loss": 0.0002427480646816593,
|
| 1597 |
+
"train_mean_reward": 6.861287548847826e-05,
|
| 1598 |
+
"train_entropy": 0.1454896568528254,
|
| 1599 |
+
"train_ce_loss": 0.20292035082249257,
|
| 1600 |
+
"train_dice_loss": 0.4142205246967247,
|
| 1601 |
+
"train_dice": 0.8465431448000059,
|
| 1602 |
+
"train_iou": 0.7625912184087085,
|
| 1603 |
+
"grad_norm": 2.3941543224708055,
|
| 1604 |
+
"lr": 0.0003707631302000539,
|
| 1605 |
+
"encoder_lr": 4.444472097387449e-05,
|
| 1606 |
+
"alpha": 0.7586029767990112,
|
| 1607 |
+
"validated_this_epoch": false,
|
| 1608 |
+
"val_loss": null,
|
| 1609 |
+
"val_dice": null,
|
| 1610 |
+
"val_iou": null,
|
| 1611 |
+
"val_reward": null,
|
| 1612 |
+
"val_entropy": null
|
| 1613 |
+
},
|
| 1614 |
+
{
|
| 1615 |
+
"epoch": 73,
|
| 1616 |
+
"train_loss": 0.16648753797386417,
|
| 1617 |
+
"train_actor_loss": -0.14332285741377485,
|
| 1618 |
+
"train_critic_loss": 0.00023685415124042995,
|
| 1619 |
+
"train_mean_reward": 6.147725720396456e-05,
|
| 1620 |
+
"train_entropy": 0.14026954176016596,
|
| 1621 |
+
"train_ce_loss": 0.20103773811154635,
|
| 1622 |
+
"train_dice_loss": 0.41834619625902303,
|
| 1623 |
+
"train_dice": 0.8502943282516033,
|
| 1624 |
+
"train_iou": 0.7664420242647398,
|
| 1625 |
+
"grad_norm": 2.3526900583921475,
|
| 1626 |
+
"lr": 0.00034699469400296674,
|
| 1627 |
+
"encoder_lr": 4.2175372794253104e-05,
|
| 1628 |
+
"alpha": 0.7670595049858093,
|
| 1629 |
+
"validated_this_epoch": false,
|
| 1630 |
+
"val_loss": null,
|
| 1631 |
+
"val_dice": null,
|
| 1632 |
+
"val_iou": null,
|
| 1633 |
+
"val_reward": null,
|
| 1634 |
+
"val_entropy": null
|
| 1635 |
+
},
|
| 1636 |
+
{
|
| 1637 |
+
"epoch": 74,
|
| 1638 |
+
"train_loss": 0.1864025143504689,
|
| 1639 |
+
"train_actor_loss": -0.1636330015456198,
|
| 1640 |
+
"train_critic_loss": 0.00021891480586267405,
|
| 1641 |
+
"train_mean_reward": 0.00018915478167875747,
|
| 1642 |
+
"train_entropy": 0.1676478852761262,
|
| 1643 |
+
"train_ce_loss": 0.22412041534527716,
|
| 1644 |
+
"train_dice_loss": 0.4757316981175149,
|
| 1645 |
+
"train_dice": 0.8465099269735887,
|
| 1646 |
+
"train_iou": 0.7634977530046909,
|
| 1647 |
+
"grad_norm": 3.1497551041056018,
|
| 1648 |
+
"lr": 0.0003238756296009549,
|
| 1649 |
+
"encoder_lr": 3.996802493677459e-05,
|
| 1650 |
+
"alpha": 0.77214115858078,
|
| 1651 |
+
"validated_this_epoch": false,
|
| 1652 |
+
"val_loss": null,
|
| 1653 |
+
"val_dice": null,
|
| 1654 |
+
"val_iou": null,
|
| 1655 |
+
"val_reward": null,
|
| 1656 |
+
"val_entropy": null
|
| 1657 |
+
},
|
| 1658 |
+
{
|
| 1659 |
+
"epoch": 75,
|
| 1660 |
+
"train_loss": 0.17251363464272926,
|
| 1661 |
+
"train_actor_loss": -0.16913796605864773,
|
| 1662 |
+
"train_critic_loss": 0.00023815348975063505,
|
| 1663 |
+
"train_mean_reward": 0.00020223084292180621,
|
| 1664 |
+
"train_entropy": 0.17127980025554645,
|
| 1665 |
+
"train_ce_loss": 0.21734169669099182,
|
| 1666 |
+
"train_dice_loss": 0.46572334844936697,
|
| 1667 |
+
"train_dice": 0.8486972486914746,
|
| 1668 |
+
"train_iou": 0.7635609623835686,
|
| 1669 |
+
"grad_norm": 2.2733089527649475,
|
| 1670 |
+
"lr": 0.0003014287527193855,
|
| 1671 |
+
"encoder_lr": 3.782485578727801e-05,
|
| 1672 |
+
"alpha": 0.7787152528762817,
|
| 1673 |
+
"validated_this_epoch": false,
|
| 1674 |
+
"val_loss": null,
|
| 1675 |
+
"val_dice": null,
|
| 1676 |
+
"val_iou": null,
|
| 1677 |
+
"val_reward": null,
|
| 1678 |
+
"val_entropy": null
|
| 1679 |
+
},
|
| 1680 |
+
{
|
| 1681 |
+
"epoch": 76,
|
| 1682 |
+
"train_loss": 0.16099657091410702,
|
| 1683 |
+
"train_actor_loss": -0.3585326444039973,
|
| 1684 |
+
"train_critic_loss": 0.00022158143960617415,
|
| 1685 |
+
"train_mean_reward": 0.00017724469076941412,
|
| 1686 |
+
"train_entropy": 0.42097398003404457,
|
| 1687 |
+
"train_ce_loss": 0.319164398167215,
|
| 1688 |
+
"train_dice_loss": 0.7196724472581878,
|
| 1689 |
+
"train_dice": 0.8552571074881123,
|
| 1690 |
+
"train_iou": 0.770026831687242,
|
| 1691 |
+
"grad_norm": 2.3694654321947763,
|
| 1692 |
+
"lr": 0.0002796762157156959,
|
| 1693 |
+
"encoder_lr": 3.574798039496594e-05,
|
| 1694 |
+
"alpha": 0.7524411678314209,
|
| 1695 |
+
"validated_this_epoch": false,
|
| 1696 |
+
"val_loss": null,
|
| 1697 |
+
"val_dice": null,
|
| 1698 |
+
"val_iou": null,
|
| 1699 |
+
"val_reward": null,
|
| 1700 |
+
"val_entropy": null
|
| 1701 |
+
},
|
| 1702 |
+
{
|
| 1703 |
+
"epoch": 77,
|
| 1704 |
+
"train_loss": 0.19603997080360056,
|
| 1705 |
+
"train_actor_loss": -0.34931063717495875,
|
| 1706 |
+
"train_critic_loss": 0.0002373255926114385,
|
| 1707 |
+
"train_mean_reward": 6.0014517314574754e-05,
|
| 1708 |
+
"train_entropy": 0.42060748951687676,
|
| 1709 |
+
"train_ce_loss": 0.3419758139523492,
|
| 1710 |
+
"train_dice_loss": 0.7484880747413143,
|
| 1711 |
+
"train_dice": 0.8399708753551846,
|
| 1712 |
+
"train_iou": 0.7537986866580183,
|
| 1713 |
+
"grad_norm": 2.2308089513649314,
|
| 1714 |
+
"lr": 0.00025863948571769286,
|
| 1715 |
+
"encoder_lr": 3.373944838510636e-05,
|
| 1716 |
+
"alpha": 0.727345883846283,
|
| 1717 |
+
"validated_this_epoch": false,
|
| 1718 |
+
"val_loss": null,
|
| 1719 |
+
"val_dice": null,
|
| 1720 |
+
"val_iou": null,
|
| 1721 |
+
"val_reward": null,
|
| 1722 |
+
"val_entropy": null
|
| 1723 |
+
},
|
| 1724 |
+
{
|
| 1725 |
+
"epoch": 78,
|
| 1726 |
+
"train_loss": 0.19064804471455396,
|
| 1727 |
+
"train_actor_loss": -0.3107402986180258,
|
| 1728 |
+
"train_critic_loss": 0.00019152150593753663,
|
| 1729 |
+
"train_mean_reward": 3.1469529604628736e-06,
|
| 1730 |
+
"train_entropy": 0.3855169350594205,
|
| 1731 |
+
"train_ce_loss": 0.29851927525708,
|
| 1732 |
+
"train_dice_loss": 0.7040658849616384,
|
| 1733 |
+
"train_dice": 0.8494665244007175,
|
| 1734 |
+
"train_iou": 0.7665395354958116,
|
| 1735 |
+
"grad_norm": 2.177723064903141,
|
| 1736 |
+
"lr": 0.00023833932343808993,
|
| 1737 |
+
"encoder_lr": 3.180124193630005e-05,
|
| 1738 |
+
"alpha": 0.7072717547416687,
|
| 1739 |
+
"validated_this_epoch": false,
|
| 1740 |
+
"val_loss": null,
|
| 1741 |
+
"val_dice": null,
|
| 1742 |
+
"val_iou": null,
|
| 1743 |
+
"val_reward": null,
|
| 1744 |
+
"val_entropy": null
|
| 1745 |
+
},
|
| 1746 |
+
{
|
| 1747 |
+
"epoch": 79,
|
| 1748 |
+
"train_loss": 0.19396903592593118,
|
| 1749 |
+
"train_actor_loss": -0.26468459025833957,
|
| 1750 |
+
"train_critic_loss": 0.0002043833121862514,
|
| 1751 |
+
"train_mean_reward": 8.722437308593956e-05,
|
| 1752 |
+
"train_entropy": 0.3299847322417476,
|
| 1753 |
+
"train_ce_loss": 0.26729249338997235,
|
| 1754 |
+
"train_dice_loss": 0.6498103708567854,
|
| 1755 |
+
"train_dice": 0.8513238843966868,
|
| 1756 |
+
"train_iou": 0.7663467231103386,
|
| 1757 |
+
"grad_norm": 2.300941375798957,
|
| 1758 |
+
"lr": 0.00021879576268618817,
|
| 1759 |
+
"encoder_lr": 2.993527382430943e-05,
|
| 1760 |
+
"alpha": 0.6939977407455444,
|
| 1761 |
+
"validated_this_epoch": false,
|
| 1762 |
+
"val_loss": null,
|
| 1763 |
+
"val_dice": null,
|
| 1764 |
+
"val_iou": null,
|
| 1765 |
+
"val_reward": null,
|
| 1766 |
+
"val_entropy": null
|
| 1767 |
+
},
|
| 1768 |
+
{
|
| 1769 |
+
"epoch": 80,
|
| 1770 |
+
"train_loss": 0.18400389912312676,
|
| 1771 |
+
"train_actor_loss": -0.13717832479557354,
|
| 1772 |
+
"train_critic_loss": 0.0002163584999777427,
|
| 1773 |
+
"train_mean_reward": 6.061659044671588e-05,
|
| 1774 |
+
"train_entropy": 0.14982667201134264,
|
| 1775 |
+
"train_ce_loss": 0.2004421017311985,
|
| 1776 |
+
"train_dice_loss": 0.4417059850477126,
|
| 1777 |
+
"train_dice": 0.8483958409801015,
|
| 1778 |
+
"train_iou": 0.7642456585214356,
|
| 1779 |
+
"grad_norm": 2.3113473721945934,
|
| 1780 |
+
"lr": 0.0002000280905969275,
|
| 1781 |
+
"encoder_lr": 2.814338553438002e-05,
|
| 1782 |
+
"alpha": 0.6992458701133728,
|
| 1783 |
+
"validated_this_epoch": true,
|
| 1784 |
+
"val_loss": 0.7117857244077392,
|
| 1785 |
+
"val_dice": 0.6950371827858323,
|
| 1786 |
+
"val_iou": 0.6056063691440434,
|
| 1787 |
+
"val_reward": 0.00018021898699599302,
|
| 1788 |
+
"val_entropy": 0.2857575181523582,
|
| 1789 |
+
"val_actor_loss": -0.0026461288992625973,
|
| 1790 |
+
"val_critic_loss": 1.7802196069139586e-05,
|
| 1791 |
+
"val_ce_loss": 0.5380321555451086,
|
| 1792 |
+
"val_dice_loss": 0.8908137161322314
|
| 1793 |
+
},
|
| 1794 |
+
{
|
| 1795 |
+
"epoch": 81,
|
| 1796 |
+
"train_loss": 0.18418789623150544,
|
| 1797 |
+
"train_actor_loss": -0.11318161402325427,
|
| 1798 |
+
"train_critic_loss": 0.00016862177067552728,
|
| 1799 |
+
"train_mean_reward": 9.918585761361641e-05,
|
| 1800 |
+
"train_entropy": 0.11332600283736087,
|
| 1801 |
+
"train_ce_loss": 0.19780750780544443,
|
| 1802 |
+
"train_dice_loss": 0.3967628900240866,
|
| 1803 |
+
"train_dice": 0.8479436340069079,
|
| 1804 |
+
"train_iou": 0.7650513197431223,
|
| 1805 |
+
"grad_norm": 2.563585849348889,
|
| 1806 |
+
"lr": 0.00018205482859681116,
|
| 1807 |
+
"encoder_lr": 2.6427345443916652e-05,
|
| 1808 |
+
"alpha": 0.7098958492279053,
|
| 1809 |
+
"validated_this_epoch": false,
|
| 1810 |
+
"val_loss": null,
|
| 1811 |
+
"val_dice": null,
|
| 1812 |
+
"val_iou": null,
|
| 1813 |
+
"val_reward": null,
|
| 1814 |
+
"val_entropy": null
|
| 1815 |
+
},
|
| 1816 |
+
{
|
| 1817 |
+
"epoch": 82,
|
| 1818 |
+
"train_loss": 0.1789410398722716,
|
| 1819 |
+
"train_actor_loss": -0.12827740686819877,
|
| 1820 |
+
"train_critic_loss": 0.0001697110170483639,
|
| 1821 |
+
"train_mean_reward": 0.0002111912606247661,
|
| 1822 |
+
"train_entropy": 0.1329572673639444,
|
| 1823 |
+
"train_ce_loss": 0.2004150915824753,
|
| 1824 |
+
"train_dice_loss": 0.41385208798009293,
|
| 1825 |
+
"train_dice": 0.8457427975943537,
|
| 1826 |
+
"train_iou": 0.7632460742681563,
|
| 1827 |
+
"grad_norm": 2.8753034236819244,
|
| 1828 |
+
"lr": 0.00016489371412549528,
|
| 1829 |
+
"encoder_lr": 2.4788847077308606e-05,
|
| 1830 |
+
"alpha": 0.7189814448356628,
|
| 1831 |
+
"validated_this_epoch": false,
|
| 1832 |
+
"val_loss": null,
|
| 1833 |
+
"val_dice": null,
|
| 1834 |
+
"val_iou": null,
|
| 1835 |
+
"val_reward": null,
|
| 1836 |
+
"val_entropy": null
|
| 1837 |
+
},
|
| 1838 |
+
{
|
| 1839 |
+
"epoch": 83,
|
| 1840 |
+
"train_loss": 0.17658287311983858,
|
| 1841 |
+
"train_actor_loss": -0.1292920145810791,
|
| 1842 |
+
"train_critic_loss": 0.00017113867515229774,
|
| 1843 |
+
"train_mean_reward": 0.00026606118202346385,
|
| 1844 |
+
"train_entropy": 0.13348660446294025,
|
| 1845 |
+
"train_ce_loss": 0.20380682547730744,
|
| 1846 |
+
"train_dice_loss": 0.40777180854048223,
|
| 1847 |
+
"train_dice": 0.8499716152564881,
|
| 1848 |
+
"train_iou": 0.7681526389108068,
|
| 1849 |
+
"grad_norm": 2.482911787984907,
|
| 1850 |
+
"lr": 0.00014856168313107603,
|
| 1851 |
+
"encoder_lr": 2.322950743462536e-05,
|
| 1852 |
+
"alpha": 0.7284263372421265,
|
| 1853 |
+
"validated_this_epoch": false,
|
| 1854 |
+
"val_loss": null,
|
| 1855 |
+
"val_dice": null,
|
| 1856 |
+
"val_iou": null,
|
| 1857 |
+
"val_reward": null,
|
| 1858 |
+
"val_entropy": null
|
| 1859 |
+
},
|
| 1860 |
+
{
|
| 1861 |
+
"epoch": 84,
|
| 1862 |
+
"train_loss": 0.17571877033423539,
|
| 1863 |
+
"train_actor_loss": -0.1395557042485336,
|
| 1864 |
+
"train_critic_loss": 0.00019218631963647306,
|
| 1865 |
+
"train_mean_reward": 0.00027871194264103013,
|
| 1866 |
+
"train_entropy": 0.14217421238491726,
|
| 1867 |
+
"train_ce_loss": 0.21189307623202322,
|
| 1868 |
+
"train_dice_loss": 0.4184636830975535,
|
| 1869 |
+
"train_dice": 0.8465016705687789,
|
| 1870 |
+
"train_iou": 0.7615497006222393,
|
| 1871 |
+
"grad_norm": 2.34209006100662,
|
| 1872 |
+
"lr": 0.0001330748533563558,
|
| 1873 |
+
"encoder_lr": 2.1750865395832973e-05,
|
| 1874 |
+
"alpha": 0.7367838025093079,
|
| 1875 |
+
"validated_this_epoch": false,
|
| 1876 |
+
"val_loss": null,
|
| 1877 |
+
"val_dice": null,
|
| 1878 |
+
"val_iou": null,
|
| 1879 |
+
"val_reward": null,
|
| 1880 |
+
"val_entropy": null
|
| 1881 |
+
},
|
| 1882 |
+
{
|
| 1883 |
+
"epoch": 85,
|
| 1884 |
+
"train_loss": 0.1528053823113278,
|
| 1885 |
+
"train_actor_loss": -0.14316263296117993,
|
| 1886 |
+
"train_critic_loss": 0.0002121986632837077,
|
| 1887 |
+
"train_mean_reward": 0.00016577295829344645,
|
| 1888 |
+
"train_entropy": 0.14427456186606924,
|
| 1889 |
+
"train_ce_loss": 0.20302122817898183,
|
| 1890 |
+
"train_dice_loss": 0.38870259942318425,
|
| 1891 |
+
"train_dice": 0.8613720082034959,
|
| 1892 |
+
"train_iou": 0.7789682926454943,
|
| 1893 |
+
"grad_norm": 2.125086767266887,
|
| 1894 |
+
"lr": 0.0001184485084325741,
|
| 1895 |
+
"encoder_lr": 2.035438020210507e-05,
|
| 1896 |
+
"alpha": 0.7445387840270996,
|
| 1897 |
+
"validated_this_epoch": false,
|
| 1898 |
+
"val_loss": null,
|
| 1899 |
+
"val_dice": null,
|
| 1900 |
+
"val_iou": null,
|
| 1901 |
+
"val_reward": null,
|
| 1902 |
+
"val_entropy": null
|
| 1903 |
+
},
|
| 1904 |
+
{
|
| 1905 |
+
"epoch": 86,
|
| 1906 |
+
"train_loss": 0.16931983796149686,
|
| 1907 |
+
"train_actor_loss": -0.14602539513876195,
|
| 1908 |
+
"train_critic_loss": 0.00018830968809474018,
|
| 1909 |
+
"train_mean_reward": 0.0001309802909387516,
|
| 1910 |
+
"train_entropy": 0.14672880917743936,
|
| 1911 |
+
"train_ce_loss": 0.211210356349034,
|
| 1912 |
+
"train_dice_loss": 0.4192917954552081,
|
| 1913 |
+
"train_dice": 0.8492452497283338,
|
| 1914 |
+
"train_iou": 0.7666334972420588,
|
| 1915 |
+
"grad_norm": 2.314589279104573,
|
| 1916 |
+
"lr": 0.00010469708279631079,
|
| 1917 |
+
"encoder_lr": 1.904143001572817e-05,
|
| 1918 |
+
"alpha": 0.7531952857971191,
|
| 1919 |
+
"validated_this_epoch": false,
|
| 1920 |
+
"val_loss": null,
|
| 1921 |
+
"val_dice": null,
|
| 1922 |
+
"val_iou": null,
|
| 1923 |
+
"val_reward": null,
|
| 1924 |
+
"val_entropy": null
|
| 1925 |
+
},
|
| 1926 |
+
{
|
| 1927 |
+
"epoch": 87,
|
| 1928 |
+
"train_loss": 0.15823414598029797,
|
| 1929 |
+
"train_actor_loss": -0.1351440753077833,
|
| 1930 |
+
"train_critic_loss": 0.0001876394967989208,
|
| 1931 |
+
"train_mean_reward": 0.00023312765219096594,
|
| 1932 |
+
"train_entropy": 0.13327255354245443,
|
| 1933 |
+
"train_ce_loss": 0.19285895224469315,
|
| 1934 |
+
"train_dice_loss": 0.39370984824744926,
|
| 1935 |
+
"train_dice": 0.853391345465348,
|
| 1936 |
+
"train_iou": 0.769790539176407,
|
| 1937 |
+
"grad_norm": 2.1255563461734344,
|
| 1938 |
+
"lr": 9.183414744443903e-05,
|
| 1939 |
+
"encoder_lr": 1.781331056002182e-05,
|
| 1940 |
+
"alpha": 0.7628701329231262,
|
| 1941 |
+
"validated_this_epoch": false,
|
| 1942 |
+
"val_loss": null,
|
| 1943 |
+
"val_dice": null,
|
| 1944 |
+
"val_iou": null,
|
| 1945 |
+
"val_reward": null,
|
| 1946 |
+
"val_entropy": null
|
| 1947 |
+
},
|
| 1948 |
+
{
|
| 1949 |
+
"epoch": 88,
|
| 1950 |
+
"train_loss": 0.15529527510151703,
|
| 1951 |
+
"train_actor_loss": -0.15254736937539176,
|
| 1952 |
+
"train_critic_loss": 0.00020523001152816198,
|
| 1953 |
+
"train_mean_reward": 0.0001299313019447225,
|
| 1954 |
+
"train_entropy": 0.15522287024689826,
|
| 1955 |
+
"train_ce_loss": 0.2077771317778102,
|
| 1956 |
+
"train_dice_loss": 0.40770292628643123,
|
| 1957 |
+
"train_dice": 0.8514592457776147,
|
| 1958 |
+
"train_iou": 0.7675236827726852,
|
| 1959 |
+
"grad_norm": 2.311247755390729,
|
| 1960 |
+
"lr": 7.987239654118994e-05,
|
| 1961 |
+
"encoder_lr": 1.6671233840616128e-05,
|
| 1962 |
+
"alpha": 0.7708621621131897,
|
| 1963 |
+
"validated_this_epoch": false,
|
| 1964 |
+
"val_loss": null,
|
| 1965 |
+
"val_dice": null,
|
| 1966 |
+
"val_iou": null,
|
| 1967 |
+
"val_reward": null,
|
| 1968 |
+
"val_entropy": null
|
| 1969 |
+
},
|
| 1970 |
+
{
|
| 1971 |
+
"epoch": 89,
|
| 1972 |
+
"train_loss": 0.16340521088767312,
|
| 1973 |
+
"train_actor_loss": -0.1669392089770977,
|
| 1974 |
+
"train_critic_loss": 0.00019239297025016794,
|
| 1975 |
+
"train_mean_reward": 0.00017165939379805504,
|
| 1976 |
+
"train_entropy": 0.16946179643622297,
|
| 1977 |
+
"train_ce_loss": 0.21881499802705762,
|
| 1978 |
+
"train_dice_loss": 0.4416814451839881,
|
| 1979 |
+
"train_dice": 0.8490378196353089,
|
| 1980 |
+
"train_iou": 0.765961689253112,
|
| 1981 |
+
"grad_norm": 2.187550806028898,
|
| 1982 |
+
"lr": 6.88236348905457e-05,
|
| 1983 |
+
"encoder_lr": 1.5616326949348587e-05,
|
| 1984 |
+
"alpha": 0.7758125066757202,
|
| 1985 |
+
"validated_this_epoch": false,
|
| 1986 |
+
"val_loss": null,
|
| 1987 |
+
"val_dice": null,
|
| 1988 |
+
"val_iou": null,
|
| 1989 |
+
"val_reward": null,
|
| 1990 |
+
"val_entropy": null
|
| 1991 |
+
},
|
| 1992 |
+
{
|
| 1993 |
+
"epoch": 90,
|
| 1994 |
+
"train_loss": 0.14845789145468918,
|
| 1995 |
+
"train_actor_loss": -0.17361601983294497,
|
| 1996 |
+
"train_critic_loss": 0.0002149038882360949,
|
| 1997 |
+
"train_mean_reward": 0.00020396311106725442,
|
| 1998 |
+
"train_entropy": 0.17577674381286917,
|
| 1999 |
+
"train_ce_loss": 0.20646615894102233,
|
| 2000 |
+
"train_dice_loss": 0.43746675878224134,
|
| 2001 |
+
"train_dice": 0.8566305560287784,
|
| 2002 |
+
"train_iou": 0.7724157795054248,
|
| 2003 |
+
"grad_norm": 2.419094535962556,
|
| 2004 |
+
"lr": 5.869876628632226e-05,
|
| 2005 |
+
"encoder_lr": 1.4649630951960419e-05,
|
| 2006 |
+
"alpha": 0.7809571623802185,
|
| 2007 |
+
"validated_this_epoch": true,
|
| 2008 |
+
"val_loss": 0.762551580852067,
|
| 2009 |
+
"val_dice": 0.6935831256254652,
|
| 2010 |
+
"val_iou": 0.6016614334988197,
|
| 2011 |
+
"val_reward": 0.0004320609756388022,
|
| 2012 |
+
"val_entropy": 0.37483241261368755,
|
| 2013 |
+
"val_actor_loss": -0.02379191335704592,
|
| 2014 |
+
"val_critic_loss": 1.4108355477642042e-05,
|
| 2015 |
+
"val_ce_loss": 0.6016430677187563,
|
| 2016 |
+
"val_dice_loss": 0.9710297894598257
|
| 2017 |
+
},
|
| 2018 |
+
{
|
| 2019 |
+
"epoch": 91,
|
| 2020 |
+
"train_loss": 0.15473357864058304,
|
| 2021 |
+
"train_actor_loss": -0.16770201316449274,
|
| 2022 |
+
"train_critic_loss": 0.0002083533466210298,
|
| 2023 |
+
"train_mean_reward": 0.0001703963217445205,
|
| 2024 |
+
"train_entropy": 0.16831023405820136,
|
| 2025 |
+
"train_ce_loss": 0.2106252686506816,
|
| 2026 |
+
"train_dice_loss": 0.4340375607943966,
|
| 2027 |
+
"train_dice": 0.8555939797907576,
|
| 2028 |
+
"train_iou": 0.772517290095208,
|
| 2029 |
+
"grad_norm": 2.134082082630128,
|
| 2030 |
+
"lr": 4.9507782751441774e-05,
|
| 2031 |
+
"encoder_lr": 1.3772099860690422e-05,
|
| 2032 |
+
"alpha": 0.7866286039352417,
|
| 2033 |
+
"validated_this_epoch": false,
|
| 2034 |
+
"val_loss": null,
|
| 2035 |
+
"val_dice": null,
|
| 2036 |
+
"val_iou": null,
|
| 2037 |
+
"val_reward": null,
|
| 2038 |
+
"val_entropy": null
|
| 2039 |
+
},
|
| 2040 |
+
{
|
| 2041 |
+
"epoch": 92,
|
| 2042 |
+
"train_loss": 0.16115680698795734,
|
| 2043 |
+
"train_actor_loss": -0.16973860407349214,
|
| 2044 |
+
"train_critic_loss": 0.00020795477385097757,
|
| 2045 |
+
"train_mean_reward": 0.00013245078033079913,
|
| 2046 |
+
"train_entropy": 0.17213721808289162,
|
| 2047 |
+
"train_ce_loss": 0.21866592448878583,
|
| 2048 |
+
"train_dice_loss": 0.4429169410420943,
|
| 2049 |
+
"train_dice": 0.852969726219291,
|
| 2050 |
+
"train_iou": 0.7701707081546183,
|
| 2051 |
+
"grad_norm": 2.2340754688710205,
|
| 2052 |
+
"lr": 4.12597546770122e-05,
|
| 2053 |
+
"encoder_lr": 1.298459969278006e-05,
|
| 2054 |
+
"alpha": 0.7919908165931702,
|
| 2055 |
+
"validated_this_epoch": false,
|
| 2056 |
+
"val_loss": null,
|
| 2057 |
+
"val_dice": null,
|
| 2058 |
+
"val_iou": null,
|
| 2059 |
+
"val_reward": null,
|
| 2060 |
+
"val_entropy": null
|
| 2061 |
+
},
|
| 2062 |
+
{
|
| 2063 |
+
"epoch": 93,
|
| 2064 |
+
"train_loss": 0.14667118563567622,
|
| 2065 |
+
"train_actor_loss": -0.18147251045557572,
|
| 2066 |
+
"train_critic_loss": 0.00021491104622725597,
|
| 2067 |
+
"train_mean_reward": 0.00017005381468133537,
|
| 2068 |
+
"train_entropy": 0.18147685209739514,
|
| 2069 |
+
"train_ce_loss": 0.2064232270333204,
|
| 2070 |
+
"train_dice_loss": 0.4496492517856972,
|
| 2071 |
+
"train_dice": 0.855426020196529,
|
| 2072 |
+
"train_iou": 0.7723631443538967,
|
| 2073 |
+
"grad_norm": 3.2533832443083903,
|
| 2074 |
+
"lr": 3.396282187094642e-05,
|
| 2075 |
+
"encoder_lr": 1.2287907615819006e-05,
|
| 2076 |
+
"alpha": 0.795804500579834,
|
| 2077 |
+
"validated_this_epoch": false,
|
| 2078 |
+
"val_loss": null,
|
| 2079 |
+
"val_dice": null,
|
| 2080 |
+
"val_iou": null,
|
| 2081 |
+
"val_reward": null,
|
| 2082 |
+
"val_entropy": null
|
| 2083 |
+
},
|
| 2084 |
+
{
|
| 2085 |
+
"epoch": 94,
|
| 2086 |
+
"train_loss": 0.14571193257006246,
|
| 2087 |
+
"train_actor_loss": -0.17757292256031176,
|
| 2088 |
+
"train_critic_loss": 0.0002029286401575219,
|
| 2089 |
+
"train_mean_reward": 0.00018972502286329973,
|
| 2090 |
+
"train_entropy": 0.17809738459210547,
|
| 2091 |
+
"train_ce_loss": 0.2113327354045936,
|
| 2092 |
+
"train_dice_loss": 0.43503404272003077,
|
| 2093 |
+
"train_dice": 0.8546282968637822,
|
| 2094 |
+
"train_iou": 0.7724071125191809,
|
| 2095 |
+
"grad_norm": 2.5069293365922083,
|
| 2096 |
+
"lr": 2.762418552495473e-05,
|
| 2097 |
+
"encoder_lr": 1.1682711180774574e-05,
|
| 2098 |
+
"alpha": 0.8005829453468323,
|
| 2099 |
+
"validated_this_epoch": false,
|
| 2100 |
+
"val_loss": null,
|
| 2101 |
+
"val_dice": null,
|
| 2102 |
+
"val_iou": null,
|
| 2103 |
+
"val_reward": null,
|
| 2104 |
+
"val_entropy": null
|
| 2105 |
+
},
|
| 2106 |
+
{
|
| 2107 |
+
"epoch": 95,
|
| 2108 |
+
"train_loss": 0.14423112103196578,
|
| 2109 |
+
"train_actor_loss": -0.17556476156324752,
|
| 2110 |
+
"train_critic_loss": 0.0002048066929312785,
|
| 2111 |
+
"train_mean_reward": 0.00025611642897417945,
|
| 2112 |
+
"train_entropy": 0.17595287655222708,
|
| 2113 |
+
"train_ce_loss": 0.20168037730081426,
|
| 2114 |
+
"train_dice_loss": 0.43770657683835773,
|
| 2115 |
+
"train_dice": 0.8554711375909557,
|
| 2116 |
+
"train_iou": 0.7739590366227652,
|
| 2117 |
+
"grad_norm": 2.684449559034303,
|
| 2118 |
+
"lr": 2.2250101107838036e-05,
|
| 2119 |
+
"encoder_lr": 1.1169607643461924e-05,
|
| 2120 |
+
"alpha": 0.8052234053611755,
|
| 2121 |
+
"validated_this_epoch": false,
|
| 2122 |
+
"val_loss": null,
|
| 2123 |
+
"val_dice": null,
|
| 2124 |
+
"val_iou": null,
|
| 2125 |
+
"val_reward": null,
|
| 2126 |
+
"val_entropy": null
|
| 2127 |
+
},
|
| 2128 |
+
{
|
| 2129 |
+
"epoch": 96,
|
| 2130 |
+
"train_loss": 0.14631287637499454,
|
| 2131 |
+
"train_actor_loss": -0.18440970697185116,
|
| 2132 |
+
"train_critic_loss": 0.00020343274860079783,
|
| 2133 |
+
"train_mean_reward": 0.00017452988981839213,
|
| 2134 |
+
"train_entropy": 0.18574097264995457,
|
| 2135 |
+
"train_ce_loss": 0.2126614818667141,
|
| 2136 |
+
"train_dice_loss": 0.44858024779524297,
|
| 2137 |
+
"train_dice": 0.8538192476914324,
|
| 2138 |
+
"train_iou": 0.7727546310077062,
|
| 2139 |
+
"grad_norm": 2.6177008787328884,
|
| 2140 |
+
"lr": 1.784587219209463e-05,
|
| 2141 |
+
"encoder_lr": 1.0749103375124614e-05,
|
| 2142 |
+
"alpha": 0.8083691596984863,
|
| 2143 |
+
"validated_this_epoch": false,
|
| 2144 |
+
"val_loss": null,
|
| 2145 |
+
"val_dice": null,
|
| 2146 |
+
"val_iou": null,
|
| 2147 |
+
"val_reward": null,
|
| 2148 |
+
"val_entropy": null
|
| 2149 |
+
},
|
| 2150 |
+
{
|
| 2151 |
+
"epoch": 97,
|
| 2152 |
+
"train_loss": 0.14009501250496728,
|
| 2153 |
+
"train_actor_loss": -0.19012411691979997,
|
| 2154 |
+
"train_critic_loss": 0.00019543305347324274,
|
| 2155 |
+
"train_mean_reward": 0.00015206075173017304,
|
| 2156 |
+
"train_entropy": 0.1901560930801569,
|
| 2157 |
+
"train_ce_loss": 0.21398438341212903,
|
| 2158 |
+
"train_dice_loss": 0.4462584397281479,
|
| 2159 |
+
"train_dice": 0.8586204484455152,
|
| 2160 |
+
"train_iou": 0.776544562536967,
|
| 2161 |
+
"grad_norm": 2.0971654593482496,
|
| 2162 |
+
"lr": 1.4415845219935406e-05,
|
| 2163 |
+
"encoder_lr": 1.0421613362707402e-05,
|
| 2164 |
+
"alpha": 0.8107907176017761,
|
| 2165 |
+
"validated_this_epoch": false,
|
| 2166 |
+
"val_loss": null,
|
| 2167 |
+
"val_dice": null,
|
| 2168 |
+
"val_iou": null,
|
| 2169 |
+
"val_reward": null,
|
| 2170 |
+
"val_entropy": null
|
| 2171 |
+
},
|
| 2172 |
+
{
|
| 2173 |
+
"epoch": 98,
|
| 2174 |
+
"train_loss": 0.14782394715748806,
|
| 2175 |
+
"train_actor_loss": -0.18745400663525424,
|
| 2176 |
+
"train_critic_loss": 0.0001820941067855296,
|
| 2177 |
+
"train_mean_reward": 0.00030522423972346476,
|
| 2178 |
+
"train_entropy": 0.189084270016052,
|
| 2179 |
+
"train_ce_loss": 0.21639102566157029,
|
| 2180 |
+
"train_dice_loss": 0.45398278538262815,
|
| 2181 |
+
"train_dice": 0.855614805566656,
|
| 2182 |
+
"train_iou": 0.7772229988712768,
|
| 2183 |
+
"grad_norm": 2.0551631196524744,
|
| 2184 |
+
"lr": 1.19634052138698e-05,
|
| 2185 |
+
"encoder_lr": 1.0187460799314204e-05,
|
| 2186 |
+
"alpha": 0.8139742016792297,
|
| 2187 |
+
"validated_this_epoch": false,
|
| 2188 |
+
"val_loss": null,
|
| 2189 |
+
"val_dice": null,
|
| 2190 |
+
"val_iou": null,
|
| 2191 |
+
"val_reward": null,
|
| 2192 |
+
"val_entropy": null
|
| 2193 |
+
},
|
| 2194 |
+
{
|
| 2195 |
+
"epoch": 99,
|
| 2196 |
+
"train_loss": 0.14752250479470125,
|
| 2197 |
+
"train_actor_loss": -0.1768703861944185,
|
| 2198 |
+
"train_critic_loss": 0.00018605505316639026,
|
| 2199 |
+
"train_mean_reward": 0.00025336058906783216,
|
| 2200 |
+
"train_entropy": 0.17714261470426618,
|
| 2201 |
+
"train_ce_loss": 0.21373744001696293,
|
| 2202 |
+
"train_dice_loss": 0.4348622848513206,
|
| 2203 |
+
"train_dice": 0.8571827775567559,
|
| 2204 |
+
"train_iou": 0.774586036474277,
|
| 2205 |
+
"grad_norm": 1.9410047665122867,
|
| 2206 |
+
"lr": 1.049097243609706e-05,
|
| 2207 |
+
"encoder_lr": 1.0046876765255499e-05,
|
| 2208 |
+
"alpha": 0.8186289072036743,
|
| 2209 |
+
"validated_this_epoch": false,
|
| 2210 |
+
"val_loss": null,
|
| 2211 |
+
"val_dice": null,
|
| 2212 |
+
"val_iou": null,
|
| 2213 |
+
"val_reward": null,
|
| 2214 |
+
"val_entropy": null
|
| 2215 |
+
},
|
| 2216 |
+
{
|
| 2217 |
+
"epoch": 100,
|
| 2218 |
+
"train_loss": 0.14635389277825106,
|
| 2219 |
+
"train_actor_loss": -0.19498363891237436,
|
| 2220 |
+
"train_critic_loss": 0.00019570429837603056,
|
| 2221 |
+
"train_mean_reward": 0.00038425886165249334,
|
| 2222 |
+
"train_entropy": 0.19594582247211492,
|
| 2223 |
+
"train_ce_loss": 0.2211552746311186,
|
| 2224 |
+
"train_dice_loss": 0.4613240829425881,
|
| 2225 |
+
"train_dice": 0.8532291285107462,
|
| 2226 |
+
"train_iou": 0.7701849428971537,
|
| 2227 |
+
"grad_norm": 2.5045857625876287,
|
| 2228 |
+
"lr": 1e-05,
|
| 2229 |
+
"encoder_lr": 1e-05,
|
| 2230 |
+
"alpha": 0.8207908272743225,
|
| 2231 |
+
"validated_this_epoch": true,
|
| 2232 |
+
"val_loss": 0.7522265741841473,
|
| 2233 |
+
"val_dice": 0.6935105822786541,
|
| 2234 |
+
"val_iou": 0.6035723638526949,
|
| 2235 |
+
"val_reward": 0.00024898100705191456,
|
| 2236 |
+
"val_entropy": 0.3568733804812184,
|
| 2237 |
+
"val_actor_loss": -0.02395725883120163,
|
| 2238 |
+
"val_critic_loss": 1.4051706136711166e-05,
|
| 2239 |
+
"val_ce_loss": 0.6048493145240678,
|
| 2240 |
+
"val_dice_loss": 0.9475042795894122
|
| 2241 |
+
}
|
| 2242 |
+
]
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/low_iou_samples.csv
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
sample_id
|
| 2 |
+
BUSI_0224_B
|
| 3 |
+
BUSI_0207_M
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/run_config.json
ADDED
|
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"project_dir": "/workspace",
|
| 3 |
+
"data_root": "/workspace/BUSI_with_classes",
|
| 4 |
+
"run_type": "final",
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"dataset_name": "BUSI_with_classes",
|
| 8 |
+
"dataset_split_policy": "stratified",
|
| 9 |
+
"dataset_splits_path": "/workspace/dataset_splits_busi_with_classes_stratified.json",
|
| 10 |
+
"split_type": "80_10_10",
|
| 11 |
+
"train_subset_key": "50",
|
| 12 |
+
"normalization_cache_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/norm_stats_BUSI_with_classes_stratified_80_10_10_50pct.json",
|
| 13 |
+
"head_lr": 0.002,
|
| 14 |
+
"encoder_lr": 0.0002,
|
| 15 |
+
"weight_decay": 0.002,
|
| 16 |
+
"dropout_p": 0.3,
|
| 17 |
+
"tmax": 6,
|
| 18 |
+
"entropy_lr": 0.0002,
|
| 19 |
+
"max_epochs": 100,
|
| 20 |
+
"gamma": 0.95,
|
| 21 |
+
"grad_clip_norm": 6.0,
|
| 22 |
+
"scheduler_factor": 0.5,
|
| 23 |
+
"scheduler_patience": 5,
|
| 24 |
+
"scheduler_threshold": 0.001,
|
| 25 |
+
"scheduler_min_lr": 1e-05,
|
| 26 |
+
"execution_mode": "train_eval",
|
| 27 |
+
"evaluation_checkpoint_mode": "latest",
|
| 28 |
+
"strategy2_checkpoint_mode": "specific",
|
| 29 |
+
"train_resume_mode": "off",
|
| 30 |
+
"train_resume_specific_checkpoint": "",
|
| 31 |
+
"backbone_family": "smp",
|
| 32 |
+
"smp_encoder_name": "efficientnet-b0",
|
| 33 |
+
"smp_encoder_weights": "imagenet",
|
| 34 |
+
"smp_encoder_depth": 5,
|
| 35 |
+
"smp_encoder_proj_dim": 64,
|
| 36 |
+
"smp_decoder_type": "Unet",
|
| 37 |
+
"vgg_feature_scales": 4,
|
| 38 |
+
"vgg_feature_dilation": 1,
|
| 39 |
+
"strategy2_checkpoint_path": "/workspace/50%_Strategt2_latest.pt"
|
| 40 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Proj=64_2actions/pct_50/strategy_3/final/summary.json
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_val_iou": 0.6056063691440434,
|
| 3 |
+
"best_val_dice": 0.6951244897142669,
|
| 4 |
+
"final_epoch": 100,
|
| 5 |
+
"elapsed_seconds": 1535.2277526855469,
|
| 6 |
+
"seconds_per_epoch": 15.352277526855469,
|
| 7 |
+
"device_used": "cuda",
|
| 8 |
+
"strategy": 3,
|
| 9 |
+
"run_type": "final",
|
| 10 |
+
"resumed": false
|
| 11 |
+
}
|