Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes. See raw diff
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/norm_stats_BUSI_with_classes_stratified_80_10_10_10pct.json +59 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/best_params.json +8 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/checkpoints/epoch_0100.pt.meta.json +40 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/evaluation.json +619 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/evaluation_results.xlsx +0 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/evaluation_summary.json +4 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/history.json +2602 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/low_iou_samples.csv +8 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/run_config.json +40 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/summary.json +11 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_100/norm_stats_BUSI_with_classes_stratified_80_10_10_100pct.json +525 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/norm_stats_BUSI_with_classes_stratified_80_10_10_50pct.json +266 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/best_params.json +8 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/best.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0010.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0020.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0030.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0040.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0050.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0060.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0070.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0080.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0090.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0100.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/latest.pt.meta.json +58 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/evaluation.json +619 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/evaluation_results.xlsx +0 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/evaluation_summary.json +4 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/history.json +2242 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/low_iou_samples.csv +4 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/run_config.json +40 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/summary.json +11 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/norm_stats_BUSI_with_classes_stratified_80_10_10_50pct.json +266 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/best_params.json +8 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/best.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0010.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0020.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0030.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0040.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0050.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0060.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0070.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0080.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0090.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0100.pt.meta.json +39 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/latest.pt.meta.json +58 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/evaluation.json +619 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/evaluation_results.xlsx +0 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/evaluation_summary.json +4 -0
- runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/history.json +2242 -0
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/norm_stats_BUSI_with_classes_stratified_80_10_10_10pct.json
ADDED
|
@@ -0,0 +1,59 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"global_mean": 90.04739151584342,
|
| 3 |
+
"global_std": 56.28519586200455,
|
| 4 |
+
"total_pixels": 45366348,
|
| 5 |
+
"num_images": 51,
|
| 6 |
+
"filenames": [
|
| 7 |
+
"BUSI_0016_B.png",
|
| 8 |
+
"BUSI_0002_M.png",
|
| 9 |
+
"BUSI_0346_B.png",
|
| 10 |
+
"BUSI_0287_B.png",
|
| 11 |
+
"BUSI_0242_B.png",
|
| 12 |
+
"BUSI_0074_M.png",
|
| 13 |
+
"BUSI_0124_M.png",
|
| 14 |
+
"BUSI_0049_B.png",
|
| 15 |
+
"BUSI_0389_B.png",
|
| 16 |
+
"BUSI_0190_B.png",
|
| 17 |
+
"BUSI_0273_B.png",
|
| 18 |
+
"BUSI_0236_B.png",
|
| 19 |
+
"BUSI_0034_B.png",
|
| 20 |
+
"BUSI_0281_B.png",
|
| 21 |
+
"BUSI_0156_M.png",
|
| 22 |
+
"BUSI_0066_B.png",
|
| 23 |
+
"BUSI_0113_B.png",
|
| 24 |
+
"BUSI_0086_M.png",
|
| 25 |
+
"BUSI_0333_B.png",
|
| 26 |
+
"BUSI_0138_M.png",
|
| 27 |
+
"BUSI_0220_B.png",
|
| 28 |
+
"BUSI_0158_M.png",
|
| 29 |
+
"BUSI_0011_B.png",
|
| 30 |
+
"BUSI_0183_M.png",
|
| 31 |
+
"BUSI_0420_B.png",
|
| 32 |
+
"BUSI_0011_M.png",
|
| 33 |
+
"BUSI_0397_B.png",
|
| 34 |
+
"BUSI_0144_M.png",
|
| 35 |
+
"BUSI_0160_B.png",
|
| 36 |
+
"BUSI_0029_B.png",
|
| 37 |
+
"BUSI_0239_B.png",
|
| 38 |
+
"BUSI_0390_B.png",
|
| 39 |
+
"BUSI_0031_M.png",
|
| 40 |
+
"BUSI_0321_B.png",
|
| 41 |
+
"BUSI_0030_M.png",
|
| 42 |
+
"BUSI_0136_B.png",
|
| 43 |
+
"BUSI_0248_B.png",
|
| 44 |
+
"BUSI_0187_B.png",
|
| 45 |
+
"BUSI_0098_M.png",
|
| 46 |
+
"BUSI_0344_B.png",
|
| 47 |
+
"BUSI_0053_M.png",
|
| 48 |
+
"BUSI_0065_B.png",
|
| 49 |
+
"BUSI_0407_B.png",
|
| 50 |
+
"BUSI_0107_M.png",
|
| 51 |
+
"BUSI_0049_M.png",
|
| 52 |
+
"BUSI_0263_B.png",
|
| 53 |
+
"BUSI_0212_B.png",
|
| 54 |
+
"BUSI_0205_B.png",
|
| 55 |
+
"BUSI_0136_M.png",
|
| 56 |
+
"BUSI_0007_B.png",
|
| 57 |
+
"BUSI_0370_B.png"
|
| 58 |
+
]
|
| 59 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/best_params.json
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"head_lr": 0.002,
|
| 3 |
+
"encoder_lr": 0.0002,
|
| 4 |
+
"weight_decay": 0.002,
|
| 5 |
+
"dropout_p": 0.3,
|
| 6 |
+
"tmax": 6,
|
| 7 |
+
"entropy_lr": 0.0002
|
| 8 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/checkpoints/epoch_0100.pt.meta.json
ADDED
|
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/checkpoints/epoch_0100.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 100,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.1,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scaler_state_dict",
|
| 23 |
+
"scheduler_state_dict"
|
| 24 |
+
],
|
| 25 |
+
"state_presence": {
|
| 26 |
+
"model_state_dict": true,
|
| 27 |
+
"optimizer_state_dict": true,
|
| 28 |
+
"scheduler_state_dict": true,
|
| 29 |
+
"scaler_state_dict": true,
|
| 30 |
+
"log_alpha": true,
|
| 31 |
+
"alpha_optimizer_state_dict": true,
|
| 32 |
+
"best_metric_name": true,
|
| 33 |
+
"best_metric_value": true,
|
| 34 |
+
"patience_counter": true,
|
| 35 |
+
"elapsed_seconds": true,
|
| 36 |
+
"run_config": true,
|
| 37 |
+
"epoch_metrics": true,
|
| 38 |
+
"resume_source": false
|
| 39 |
+
}
|
| 40 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/evaluation.json
ADDED
|
@@ -0,0 +1,619 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"strategy": 3,
|
| 3 |
+
"best_metric_name": "val_iou",
|
| 4 |
+
"metrics": {
|
| 5 |
+
"dice": {
|
| 6 |
+
"mean": 0.690712571144104,
|
| 7 |
+
"std": 0.32509860396385193
|
| 8 |
+
},
|
| 9 |
+
"ppv": {
|
| 10 |
+
"mean": 0.7483484745025635,
|
| 11 |
+
"std": 0.3237404227256775
|
| 12 |
+
},
|
| 13 |
+
"sen": {
|
| 14 |
+
"mean": 0.7158533930778503,
|
| 15 |
+
"std": 0.32736077904701233
|
| 16 |
+
},
|
| 17 |
+
"iou": {
|
| 18 |
+
"mean": 0.6050313115119934,
|
| 19 |
+
"std": 0.31567302346229553
|
| 20 |
+
},
|
| 21 |
+
"biou": {
|
| 22 |
+
"mean": 0.10852889716625214,
|
| 23 |
+
"std": 0.10821165144443512
|
| 24 |
+
},
|
| 25 |
+
"hd95": {
|
| 26 |
+
"mean": Infinity,
|
| 27 |
+
"std": NaN
|
| 28 |
+
}
|
| 29 |
+
},
|
| 30 |
+
"per_sample": [
|
| 31 |
+
{
|
| 32 |
+
"sample_id": "BUSI_0097_B",
|
| 33 |
+
"dice": 0.8881578969009695,
|
| 34 |
+
"ppv": 0.9845559852185335,
|
| 35 |
+
"sen": 0.8089531262265377,
|
| 36 |
+
"iou": 0.798816575049893,
|
| 37 |
+
"biou": 0.09810155123368633,
|
| 38 |
+
"hd95": 9.219544410705566
|
| 39 |
+
},
|
| 40 |
+
{
|
| 41 |
+
"sample_id": "BUSI_0415_B",
|
| 42 |
+
"dice": 0.9226056774718829,
|
| 43 |
+
"ppv": 0.9527625481124343,
|
| 44 |
+
"sen": 0.8942992888057629,
|
| 45 |
+
"iou": 0.8563305552649665,
|
| 46 |
+
"biou": 0.0757343005047449,
|
| 47 |
+
"hd95": 33.060386657714844
|
| 48 |
+
},
|
| 49 |
+
{
|
| 50 |
+
"sample_id": "BUSI_0290_B",
|
| 51 |
+
"dice": 0.9115675694796201,
|
| 52 |
+
"ppv": 0.8558668349221749,
|
| 53 |
+
"sen": 0.9750231278897721,
|
| 54 |
+
"iou": 0.8375049726855395,
|
| 55 |
+
"biou": 0.08125028710928528,
|
| 56 |
+
"hd95": 7.6157732009887695
|
| 57 |
+
},
|
| 58 |
+
{
|
| 59 |
+
"sample_id": "BUSI_0354_B",
|
| 60 |
+
"dice": 0.9435325034953216,
|
| 61 |
+
"ppv": 0.9049118447788511,
|
| 62 |
+
"sen": 0.98559670880681,
|
| 63 |
+
"iou": 0.8931013118022801,
|
| 64 |
+
"biou": 0.19831257455165632,
|
| 65 |
+
"hd95": 4.4721360206604
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"sample_id": "BUSI_0067_B",
|
| 69 |
+
"dice": 0.9035277195189395,
|
| 70 |
+
"ppv": 0.9870232014540621,
|
| 71 |
+
"sen": 0.83305675960648,
|
| 72 |
+
"iou": 0.8240315225203045,
|
| 73 |
+
"biou": 0.10714312287407059,
|
| 74 |
+
"hd95": 9.129254341125488
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"sample_id": "BUSI_0165_M",
|
| 78 |
+
"dice": 0.8620071694891938,
|
| 79 |
+
"ppv": 0.8889573403826068,
|
| 80 |
+
"sen": 0.836642994105769,
|
| 81 |
+
"iou": 0.757480318143303,
|
| 82 |
+
"biou": 0.026479902417460684,
|
| 83 |
+
"hd95": 13.03840446472168
|
| 84 |
+
},
|
| 85 |
+
{
|
| 86 |
+
"sample_id": "BUSI_0198_B",
|
| 87 |
+
"dice": 0.9594287582139319,
|
| 88 |
+
"ppv": 0.9666448681069413,
|
| 89 |
+
"sen": 0.9523195907010573,
|
| 90 |
+
"iou": 0.9220212150953703,
|
| 91 |
+
"biou": 0.202531982054016,
|
| 92 |
+
"hd95": 2.0
|
| 93 |
+
},
|
| 94 |
+
{
|
| 95 |
+
"sample_id": "BUSI_0361_B",
|
| 96 |
+
"dice": 0.9348159534196336,
|
| 97 |
+
"ppv": 0.9234848542814504,
|
| 98 |
+
"sen": 0.9464285755878435,
|
| 99 |
+
"iou": 0.8776098000280922,
|
| 100 |
+
"biou": 0.1219515764424486,
|
| 101 |
+
"hd95": 2.2360680103302
|
| 102 |
+
},
|
| 103 |
+
{
|
| 104 |
+
"sample_id": "BUSI_0095_B",
|
| 105 |
+
"dice": 0.9090909302818344,
|
| 106 |
+
"ppv": 0.915492997421128,
|
| 107 |
+
"sen": 0.902777822788045,
|
| 108 |
+
"iou": 0.8333334045583741,
|
| 109 |
+
"biou": 0.18750101562373045,
|
| 110 |
+
"hd95": 1.4142135381698608
|
| 111 |
+
},
|
| 112 |
+
{
|
| 113 |
+
"sample_id": "BUSI_0057_B",
|
| 114 |
+
"dice": 0.9414520043437928,
|
| 115 |
+
"ppv": 0.9804878143961882,
|
| 116 |
+
"sen": 0.905405448015564,
|
| 117 |
+
"iou": 0.8893805799200973,
|
| 118 |
+
"biou": 0.38806061483490323,
|
| 119 |
+
"hd95": 1.0
|
| 120 |
+
},
|
| 121 |
+
{
|
| 122 |
+
"sample_id": "BUSI_0009_M",
|
| 123 |
+
"dice": 0.811950801877516,
|
| 124 |
+
"ppv": 0.8230404010640854,
|
| 125 |
+
"sen": 0.8011560923518969,
|
| 126 |
+
"iou": 0.6834319838824473,
|
| 127 |
+
"biou": 0.056122930549525225,
|
| 128 |
+
"hd95": 5.94928503036499
|
| 129 |
+
},
|
| 130 |
+
{
|
| 131 |
+
"sample_id": "BUSI_0399_B",
|
| 132 |
+
"dice": 0.9486780726798918,
|
| 133 |
+
"ppv": 0.9638826201407396,
|
| 134 |
+
"sen": 0.9339457596699143,
|
| 135 |
+
"iou": 0.9023668680318314,
|
| 136 |
+
"biou": 0.11660808600420991,
|
| 137 |
+
"hd95": 3.2730960845947266
|
| 138 |
+
},
|
| 139 |
+
{
|
| 140 |
+
"sample_id": "BUSI_0099_M",
|
| 141 |
+
"dice": 0.8230547652418003,
|
| 142 |
+
"ppv": 0.9118774058904974,
|
| 143 |
+
"sen": 0.7500000262605014,
|
| 144 |
+
"iou": 0.6993144270994684,
|
| 145 |
+
"biou": 0.06190520861656732,
|
| 146 |
+
"hd95": 7.280109882354736
|
| 147 |
+
},
|
| 148 |
+
{
|
| 149 |
+
"sample_id": "BUSI_0307_B",
|
| 150 |
+
"dice": 0.9148936270811315,
|
| 151 |
+
"ppv": 0.9302884782960389,
|
| 152 |
+
"sen": 0.9000000232558085,
|
| 153 |
+
"iou": 0.8431372890768434,
|
| 154 |
+
"biou": 0.19841333459259158,
|
| 155 |
+
"hd95": 2.1180338859558105
|
| 156 |
+
},
|
| 157 |
+
{
|
| 158 |
+
"sample_id": "BUSI_0039_B",
|
| 159 |
+
"dice": 0.8970588361447505,
|
| 160 |
+
"ppv": 0.8224719500063034,
|
| 161 |
+
"sen": 0.9865229146838504,
|
| 162 |
+
"iou": 0.8133333748148056,
|
| 163 |
+
"biou": 0.14782682797667132,
|
| 164 |
+
"hd95": 2.2360680103302
|
| 165 |
+
},
|
| 166 |
+
{
|
| 167 |
+
"sample_id": "BUSI_0189_M",
|
| 168 |
+
"dice": 0.2916667100694418,
|
| 169 |
+
"ppv": 0.17474308555483953,
|
| 170 |
+
"sen": 0.8814815253772129,
|
| 171 |
+
"iou": 0.17073176680546867,
|
| 172 |
+
"biou": 0.08474615053129214,
|
| 173 |
+
"hd95": 53.49529266357422
|
| 174 |
+
},
|
| 175 |
+
{
|
| 176 |
+
"sample_id": "BUSI_0021_B",
|
| 177 |
+
"dice": 0.8727273190082477,
|
| 178 |
+
"ppv": 0.7894738227145903,
|
| 179 |
+
"sen": 0.9756097759270115,
|
| 180 |
+
"iou": 0.7741936940685845,
|
| 181 |
+
"biou": 0.203126245115242,
|
| 182 |
+
"hd95": 2.0
|
| 183 |
+
},
|
| 184 |
+
{
|
| 185 |
+
"sample_id": "BUSI_0311_B",
|
| 186 |
+
"dice": 0.11991437830426914,
|
| 187 |
+
"ppv": 0.35744694522405424,
|
| 188 |
+
"sen": 0.07204120617319013,
|
| 189 |
+
"iou": 0.06378135672811858,
|
| 190 |
+
"biou": 0.005682100658494132,
|
| 191 |
+
"hd95": 60.44804763793945
|
| 192 |
+
},
|
| 193 |
+
{
|
| 194 |
+
"sample_id": "BUSI_0080_M",
|
| 195 |
+
"dice": 0.4447108347404604,
|
| 196 |
+
"ppv": 0.427912374399517,
|
| 197 |
+
"sen": 0.46288212957691016,
|
| 198 |
+
"iou": 0.28593451691967176,
|
| 199 |
+
"biou": 0.02234664180261402,
|
| 200 |
+
"hd95": 28.160255432128906
|
| 201 |
+
},
|
| 202 |
+
{
|
| 203 |
+
"sample_id": "BUSI_0130_B",
|
| 204 |
+
"dice": 0.9400206856245767,
|
| 205 |
+
"ppv": 0.9026812410445639,
|
| 206 |
+
"sen": 0.980582526366502,
|
| 207 |
+
"iou": 0.8868292793337288,
|
| 208 |
+
"biou": 0.20114988439661818,
|
| 209 |
+
"hd95": 3.097365140914917
|
| 210 |
+
},
|
| 211 |
+
{
|
| 212 |
+
"sample_id": "BUSI_0401_B",
|
| 213 |
+
"dice": 0.8294098732857184,
|
| 214 |
+
"ppv": 0.9772036491486589,
|
| 215 |
+
"sen": 0.7204481949328743,
|
| 216 |
+
"iou": 0.7085399609619856,
|
| 217 |
+
"biou": 0.0850206133519784,
|
| 218 |
+
"hd95": 9.947099685668945
|
| 219 |
+
},
|
| 220 |
+
{
|
| 221 |
+
"sample_id": "BUSI_0326_B",
|
| 222 |
+
"dice": 0.9566229998071907,
|
| 223 |
+
"ppv": 0.9204481837283931,
|
| 224 |
+
"sen": 0.9957575760146924,
|
| 225 |
+
"iou": 0.9168526832113458,
|
| 226 |
+
"biou": 0.1186444412523554,
|
| 227 |
+
"hd95": 2.0
|
| 228 |
+
},
|
| 229 |
+
{
|
| 230 |
+
"sample_id": "BUSI_0119_B",
|
| 231 |
+
"dice": 0.03443711806499609,
|
| 232 |
+
"ppv": 0.01836161658327625,
|
| 233 |
+
"sen": 0.2765961294701439,
|
| 234 |
+
"iou": 0.017520248735840675,
|
| 235 |
+
"biou": 0.0072730882643315415,
|
| 236 |
+
"hd95": 63.06964111328125
|
| 237 |
+
},
|
| 238 |
+
{
|
| 239 |
+
"sample_id": "BUSI_0372_B",
|
| 240 |
+
"dice": 0.9237490101767669,
|
| 241 |
+
"ppv": 0.9150275440576283,
|
| 242 |
+
"sen": 0.9326383373986898,
|
| 243 |
+
"iou": 0.8583025934832035,
|
| 244 |
+
"biou": 0.10434821550077587,
|
| 245 |
+
"hd95": 4.123105525970459
|
| 246 |
+
},
|
| 247 |
+
{
|
| 248 |
+
"sample_id": "BUSI_0003_M",
|
| 249 |
+
"dice": 0.5980254162182358,
|
| 250 |
+
"ppv": 1.0,
|
| 251 |
+
"sen": 0.42655941382702073,
|
| 252 |
+
"iou": 0.42655941382702073,
|
| 253 |
+
"biou": 0.03723455466247092,
|
| 254 |
+
"hd95": 14.359317779541016
|
| 255 |
+
},
|
| 256 |
+
{
|
| 257 |
+
"sample_id": "BUSI_0007_M",
|
| 258 |
+
"dice": 0.8495788297485662,
|
| 259 |
+
"ppv": 0.8414779688345687,
|
| 260 |
+
"sen": 0.8578371983186879,
|
| 261 |
+
"iou": 0.7384937512035825,
|
| 262 |
+
"biou": 0.06532710285070208,
|
| 263 |
+
"hd95": 5.830951690673828
|
| 264 |
+
},
|
| 265 |
+
{
|
| 266 |
+
"sample_id": "BUSI_0411_B",
|
| 267 |
+
"dice": 0.9262720683729992,
|
| 268 |
+
"ppv": 0.9010101060095906,
|
| 269 |
+
"sen": 0.9529914555025932,
|
| 270 |
+
"iou": 0.8626692522887208,
|
| 271 |
+
"biou": 0.18181849173541983,
|
| 272 |
+
"hd95": 5.0
|
| 273 |
+
},
|
| 274 |
+
{
|
| 275 |
+
"sample_id": "BUSI_0211_B",
|
| 276 |
+
"dice": 0.4230770815722304,
|
| 277 |
+
"ppv": 0.9625000468749414,
|
| 278 |
+
"sen": 0.2711270172087968,
|
| 279 |
+
"iou": 0.2682929378770252,
|
| 280 |
+
"biou": 0.13158009002619733,
|
| 281 |
+
"hd95": 13.982097625732422
|
| 282 |
+
},
|
| 283 |
+
{
|
| 284 |
+
"sample_id": "BUSI_0202_B",
|
| 285 |
+
"dice": 3.1476234451487743e-08,
|
| 286 |
+
"ppv": 3.2446462282723484e-08,
|
| 287 |
+
"sen": 1.0526304709152937e-06,
|
| 288 |
+
"iou": 3.1476234451487743e-08,
|
| 289 |
+
"biou": 3.1446530991656923e-07,
|
| 290 |
+
"hd95": 100.16680908203125
|
| 291 |
+
},
|
| 292 |
+
{
|
| 293 |
+
"sample_id": "BUSI_0194_M",
|
| 294 |
+
"dice": 2.700513024560815e-08,
|
| 295 |
+
"ppv": 6.325110289366839e-08,
|
| 296 |
+
"sen": 4.712535121935197e-08,
|
| 297 |
+
"iou": 2.700513024560815e-08,
|
| 298 |
+
"biou": 3.278687449610673e-07,
|
| 299 |
+
"hd95": 54.42058563232422
|
| 300 |
+
},
|
| 301 |
+
{
|
| 302 |
+
"sample_id": "BUSI_0141_B",
|
| 303 |
+
"dice": 0.5530497219114153,
|
| 304 |
+
"ppv": 0.980018165909229,
|
| 305 |
+
"sen": 0.38521957175747273,
|
| 306 |
+
"iou": 0.3822175064090506,
|
| 307 |
+
"biou": 0.03421476917344181,
|
| 308 |
+
"hd95": 31.29216194152832
|
| 309 |
+
},
|
| 310 |
+
{
|
| 311 |
+
"sample_id": "BUSI_0251_B",
|
| 312 |
+
"dice": 2.932550459662622e-07,
|
| 313 |
+
"ppv": 1.0,
|
| 314 |
+
"sen": 2.932550459662622e-07,
|
| 315 |
+
"iou": 2.932550459662622e-07,
|
| 316 |
+
"biou": 1.6129006243538317e-06,
|
| 317 |
+
"hd95": Infinity
|
| 318 |
+
},
|
| 319 |
+
{
|
| 320 |
+
"sample_id": "BUSI_0161_M",
|
| 321 |
+
"dice": 0.6471834952907586,
|
| 322 |
+
"ppv": 0.5762812956638599,
|
| 323 |
+
"sen": 0.7379801643861571,
|
| 324 |
+
"iou": 0.4783971062269607,
|
| 325 |
+
"biou": 0.0036970418304913444,
|
| 326 |
+
"hd95": 24.734586715698242
|
| 327 |
+
},
|
| 328 |
+
{
|
| 329 |
+
"sample_id": "BUSI_0055_M",
|
| 330 |
+
"dice": 0.8985704629972455,
|
| 331 |
+
"ppv": 0.9579100206226385,
|
| 332 |
+
"sen": 0.8461538658777095,
|
| 333 |
+
"iou": 0.8158220252383158,
|
| 334 |
+
"biou": 0.08108157779374173,
|
| 335 |
+
"hd95": 2.8284270763397217
|
| 336 |
+
},
|
| 337 |
+
{
|
| 338 |
+
"sample_id": "BUSI_0059_M",
|
| 339 |
+
"dice": 0.7828631161665297,
|
| 340 |
+
"ppv": 0.6529545118885391,
|
| 341 |
+
"sen": 0.9773023747116522,
|
| 342 |
+
"iou": 0.6432005555769136,
|
| 343 |
+
"biou": 0.028674009198206236,
|
| 344 |
+
"hd95": 13.45362377166748
|
| 345 |
+
},
|
| 346 |
+
{
|
| 347 |
+
"sample_id": "BUSI_0285_B",
|
| 348 |
+
"dice": 0.9355657059581073,
|
| 349 |
+
"ppv": 0.9155908723345708,
|
| 350 |
+
"sen": 0.9564315397892593,
|
| 351 |
+
"iou": 0.8789323280331431,
|
| 352 |
+
"biou": 0.16759823039205005,
|
| 353 |
+
"hd95": 3.0
|
| 354 |
+
},
|
| 355 |
+
{
|
| 356 |
+
"sample_id": "BUSI_0163_B",
|
| 357 |
+
"dice": 4.4444424691366805e-07,
|
| 358 |
+
"ppv": 7.092193551635779e-07,
|
| 359 |
+
"sen": 1.1904747732443175e-06,
|
| 360 |
+
"iou": 4.4444424691366805e-07,
|
| 361 |
+
"biou": 1.0988998913188009e-06,
|
| 362 |
+
"hd95": 63.04317092895508
|
| 363 |
+
},
|
| 364 |
+
{
|
| 365 |
+
"sample_id": "BUSI_0224_B",
|
| 366 |
+
"dice": 0.704609398298866,
|
| 367 |
+
"ppv": 0.6344505208709851,
|
| 368 |
+
"sen": 0.7922141220333759,
|
| 369 |
+
"iou": 0.5439358655550997,
|
| 370 |
+
"biou": 0.017721767665375275,
|
| 371 |
+
"hd95": 33.078468322753906
|
| 372 |
+
},
|
| 373 |
+
{
|
| 374 |
+
"sample_id": "BUSI_0316_B",
|
| 375 |
+
"dice": 0.2154986227925498,
|
| 376 |
+
"ppv": 1.0,
|
| 377 |
+
"sen": 0.12076126088648337,
|
| 378 |
+
"iou": 0.12076126088648337,
|
| 379 |
+
"biou": 2.531644928697487e-07,
|
| 380 |
+
"hd95": 55.36330032348633
|
| 381 |
+
},
|
| 382 |
+
{
|
| 383 |
+
"sample_id": "BUSI_0116_M",
|
| 384 |
+
"dice": 6.242196863782967e-08,
|
| 385 |
+
"ppv": 1.3908203907064828e-07,
|
| 386 |
+
"sen": 1.1325027030008265e-07,
|
| 387 |
+
"iou": 6.242196863782967e-08,
|
| 388 |
+
"biou": 4.878046400952976e-07,
|
| 389 |
+
"hd95": 93.38094329833984
|
| 390 |
+
},
|
| 391 |
+
{
|
| 392 |
+
"sample_id": "BUSI_0214_B",
|
| 393 |
+
"dice": 0.9121004616381015,
|
| 394 |
+
"ppv": 0.9444444510112942,
|
| 395 |
+
"sen": 0.8818984677816261,
|
| 396 |
+
"iou": 0.8384050536825757,
|
| 397 |
+
"biou": 0.2562504648434595,
|
| 398 |
+
"hd95": 5.0
|
| 399 |
+
},
|
| 400 |
+
{
|
| 401 |
+
"sample_id": "BUSI_0207_M",
|
| 402 |
+
"dice": 0.7961015697332247,
|
| 403 |
+
"ppv": 0.7447217013089394,
|
| 404 |
+
"sen": 0.8550964267164504,
|
| 405 |
+
"iou": 0.6612697204401482,
|
| 406 |
+
"biou": 0.06497201554462838,
|
| 407 |
+
"hd95": 23.781862258911133
|
| 408 |
+
},
|
| 409 |
+
{
|
| 410 |
+
"sample_id": "BUSI_0016_M",
|
| 411 |
+
"dice": 0.5649971944288652,
|
| 412 |
+
"ppv": 0.7477656530404343,
|
| 413 |
+
"sen": 0.45402473759346584,
|
| 414 |
+
"iou": 0.39372550604639195,
|
| 415 |
+
"biou": 0.03991152108391994,
|
| 416 |
+
"hd95": 38.96393966674805
|
| 417 |
+
},
|
| 418 |
+
{
|
| 419 |
+
"sample_id": "BUSI_0067_M",
|
| 420 |
+
"dice": 0.27883851871840654,
|
| 421 |
+
"ppv": 0.16204347753020162,
|
| 422 |
+
"sen": 0.998575498778419,
|
| 423 |
+
"iou": 0.16200602814869358,
|
| 424 |
+
"biou": 0.03142884816318624,
|
| 425 |
+
"hd95": 35.22782897949219
|
| 426 |
+
},
|
| 427 |
+
{
|
| 428 |
+
"sample_id": "BUSI_0184_B",
|
| 429 |
+
"dice": 0.6870144370972687,
|
| 430 |
+
"ppv": 0.9872408303635701,
|
| 431 |
+
"sen": 0.526808530774105,
|
| 432 |
+
"iou": 0.5232460049346573,
|
| 433 |
+
"biou": 0.19594621758573733,
|
| 434 |
+
"hd95": 31.605533599853516
|
| 435 |
+
},
|
| 436 |
+
{
|
| 437 |
+
"sample_id": "BUSI_0099_B",
|
| 438 |
+
"dice": 0.9632829400084276,
|
| 439 |
+
"ppv": 0.9737991304513638,
|
| 440 |
+
"sen": 0.9529914596878263,
|
| 441 |
+
"iou": 0.9291666765046283,
|
| 442 |
+
"biou": 0.32575808654690414,
|
| 443 |
+
"hd95": 1.4142135381698608
|
| 444 |
+
},
|
| 445 |
+
{
|
| 446 |
+
"sample_id": "BUSI_0412_B",
|
| 447 |
+
"dice": 0.8176044385221894,
|
| 448 |
+
"ppv": 0.9919687433911711,
|
| 449 |
+
"sen": 0.6953743199121376,
|
| 450 |
+
"iou": 0.6914813180287287,
|
| 451 |
+
"biou": 0.09876561753793879,
|
| 452 |
+
"hd95": 29.89136505126953
|
| 453 |
+
},
|
| 454 |
+
{
|
| 455 |
+
"sample_id": "BUSI_0100_M",
|
| 456 |
+
"dice": 0.775696675390741,
|
| 457 |
+
"ppv": 0.7057356853087543,
|
| 458 |
+
"sen": 0.8610547808260871,
|
| 459 |
+
"iou": 0.633582116896857,
|
| 460 |
+
"biou": 0.0329222086739882,
|
| 461 |
+
"hd95": 8.485280990600586
|
| 462 |
+
},
|
| 463 |
+
{
|
| 464 |
+
"sample_id": "BUSI_0073_M",
|
| 465 |
+
"dice": 0.8361858324216925,
|
| 466 |
+
"ppv": 0.8382353205497842,
|
| 467 |
+
"sen": 0.8341463684314848,
|
| 468 |
+
"iou": 0.7184874343855134,
|
| 469 |
+
"biou": 0.1042950280398601,
|
| 470 |
+
"hd95": 5.0
|
| 471 |
+
},
|
| 472 |
+
{
|
| 473 |
+
"sample_id": "BUSI_0357_B",
|
| 474 |
+
"dice": 0.8555078787190359,
|
| 475 |
+
"ppv": 0.7736093436469155,
|
| 476 |
+
"sen": 0.9568000069119988,
|
| 477 |
+
"iou": 0.747500031562496,
|
| 478 |
+
"biou": 0.07100646686007878,
|
| 479 |
+
"hd95": 5.830951690673828
|
| 480 |
+
},
|
| 481 |
+
{
|
| 482 |
+
"sample_id": "BUSI_0229_B",
|
| 483 |
+
"dice": 0.9515608206910214,
|
| 484 |
+
"ppv": 0.9619151292801817,
|
| 485 |
+
"sen": 0.9414270562910483,
|
| 486 |
+
"iou": 0.9075975454211965,
|
| 487 |
+
"biou": 0.21893537341102168,
|
| 488 |
+
"hd95": 2.0
|
| 489 |
+
},
|
| 490 |
+
{
|
| 491 |
+
"sample_id": "BUSI_0129_M",
|
| 492 |
+
"dice": 0.7616263640568132,
|
| 493 |
+
"ppv": 0.8993723870424187,
|
| 494 |
+
"sen": 0.6604701235140555,
|
| 495 |
+
"iou": 0.6150214647350292,
|
| 496 |
+
"biou": 0.0038836885662740656,
|
| 497 |
+
"hd95": 25.38337516784668
|
| 498 |
+
},
|
| 499 |
+
{
|
| 500 |
+
"sample_id": "BUSI_0156_B",
|
| 501 |
+
"dice": 0.9493243328844032,
|
| 502 |
+
"ppv": 0.9689655279429213,
|
| 503 |
+
"sen": 0.9304635991842387,
|
| 504 |
+
"iou": 0.9035370085090004,
|
| 505 |
+
"biou": 0.3717956771850292,
|
| 506 |
+
"hd95": 1.4142135381698608
|
| 507 |
+
},
|
| 508 |
+
{
|
| 509 |
+
"sample_id": "BUSI_0320_B",
|
| 510 |
+
"dice": 0.949541290190219,
|
| 511 |
+
"ppv": 0.9387755240871828,
|
| 512 |
+
"sen": 0.9605568536991058,
|
| 513 |
+
"iou": 0.9039301519803161,
|
| 514 |
+
"biou": 0.33653909948163513,
|
| 515 |
+
"hd95": 2.0
|
| 516 |
+
},
|
| 517 |
+
{
|
| 518 |
+
"sample_id": "BUSI_0104_B",
|
| 519 |
+
"dice": 5.296609888950748e-08,
|
| 520 |
+
"ppv": 5.733944625347212e-08,
|
| 521 |
+
"sen": 6.94443962191693e-07,
|
| 522 |
+
"iou": 5.296609888950748e-08,
|
| 523 |
+
"biou": 5.076139555259109e-07,
|
| 524 |
+
"hd95": 70.69650268554688
|
| 525 |
+
},
|
| 526 |
+
{
|
| 527 |
+
"sample_id": "BUSI_0340_B",
|
| 528 |
+
"dice": 0.8188405821411624,
|
| 529 |
+
"ppv": 0.9486940314460811,
|
| 530 |
+
"sen": 0.7202549641110726,
|
| 531 |
+
"iou": 0.6932515407123031,
|
| 532 |
+
"biou": 0.03139035170619917,
|
| 533 |
+
"hd95": 20.83123016357422
|
| 534 |
+
},
|
| 535 |
+
{
|
| 536 |
+
"sample_id": "BUSI_0121_B",
|
| 537 |
+
"dice": 0.9299533068743624,
|
| 538 |
+
"ppv": 0.9040207647184482,
|
| 539 |
+
"sen": 0.9574175882668148,
|
| 540 |
+
"iou": 0.8690773230576904,
|
| 541 |
+
"biou": 0.2816906466967277,
|
| 542 |
+
"hd95": 3.1622776985168457
|
| 543 |
+
},
|
| 544 |
+
{
|
| 545 |
+
"sample_id": "BUSI_0358_B",
|
| 546 |
+
"dice": 0.2583377579259263,
|
| 547 |
+
"ppv": 0.14832828473498916,
|
| 548 |
+
"sen": 1.0,
|
| 549 |
+
"iou": 0.14832828473498916,
|
| 550 |
+
"biou": 0.012195423111151491,
|
| 551 |
+
"hd95": 45.79301071166992
|
| 552 |
+
},
|
| 553 |
+
{
|
| 554 |
+
"sample_id": "BUSI_0270_B",
|
| 555 |
+
"dice": 1.4662754448276476e-07,
|
| 556 |
+
"ppv": 3.1847123615565734e-07,
|
| 557 |
+
"sen": 2.717390565926477e-07,
|
| 558 |
+
"iou": 1.4662754448276476e-07,
|
| 559 |
+
"biou": 8.333326388894677e-07,
|
| 560 |
+
"hd95": 48.797027587890625
|
| 561 |
+
},
|
| 562 |
+
{
|
| 563 |
+
"sample_id": "BUSI_0023_M",
|
| 564 |
+
"dice": 0.846108837077152,
|
| 565 |
+
"ppv": 0.8891166245313256,
|
| 566 |
+
"sen": 0.8070697710312601,
|
| 567 |
+
"iou": 0.7332657245898289,
|
| 568 |
+
"biou": 0.0444841557857374,
|
| 569 |
+
"hd95": 17.48427391052246
|
| 570 |
+
},
|
| 571 |
+
{
|
| 572 |
+
"sample_id": "BUSI_0020_B",
|
| 573 |
+
"dice": 0.8599269269228424,
|
| 574 |
+
"ppv": 0.7575107556318932,
|
| 575 |
+
"sen": 0.9943661979765919,
|
| 576 |
+
"iou": 0.7542735305263322,
|
| 577 |
+
"biou": 0.12918701952774184,
|
| 578 |
+
"hd95": 13.265077590942383
|
| 579 |
+
},
|
| 580 |
+
{
|
| 581 |
+
"sample_id": "BUSI_0153_M",
|
| 582 |
+
"dice": 0.8575013456040388,
|
| 583 |
+
"ppv": 0.8885586944918519,
|
| 584 |
+
"sen": 0.8285417417987229,
|
| 585 |
+
"iou": 0.7505491096556463,
|
| 586 |
+
"biou": 0.010830503460197932,
|
| 587 |
+
"hd95": 14.040990829467773
|
| 588 |
+
},
|
| 589 |
+
{
|
| 590 |
+
"sample_id": "BUSI_0154_B",
|
| 591 |
+
"dice": 0.9318996659857828,
|
| 592 |
+
"ppv": 0.8843538201674693,
|
| 593 |
+
"sen": 0.9848484963268967,
|
| 594 |
+
"iou": 0.8724833070581832,
|
| 595 |
+
"biou": 0.46000107999783996,
|
| 596 |
+
"hd95": 1.4142135381698608
|
| 597 |
+
},
|
| 598 |
+
{
|
| 599 |
+
"sample_id": "BUSI_0062_M",
|
| 600 |
+
"dice": 0.8344308599616449,
|
| 601 |
+
"ppv": 0.8968655259420867,
|
| 602 |
+
"sen": 0.7801231407157809,
|
| 603 |
+
"iou": 0.7158999307546435,
|
| 604 |
+
"biou": 0.038860352626851655,
|
| 605 |
+
"hd95": 9.433980941772461
|
| 606 |
+
},
|
| 607 |
+
{
|
| 608 |
+
"sample_id": "BUSI_0075_B",
|
| 609 |
+
"dice": 0.9405594457552932,
|
| 610 |
+
"ppv": 0.9693693748884009,
|
| 611 |
+
"sen": 0.9134125783679833,
|
| 612 |
+
"iou": 0.8877887973945879,
|
| 613 |
+
"biou": 0.2440950833896981,
|
| 614 |
+
"hd95": 2.0
|
| 615 |
+
}
|
| 616 |
+
],
|
| 617 |
+
"checkpoint_mode": "latest",
|
| 618 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/checkpoints/latest.pt"
|
| 619 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/evaluation_results.xlsx
ADDED
|
Binary file (43.3 kB). View file
|
|
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/evaluation_summary.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"mean_iou": 0.6050313115119934,
|
| 3 |
+
"mean_dice": 0.690712571144104
|
| 4 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/history.json
ADDED
|
@@ -0,0 +1,2602 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"epoch": 1,
|
| 4 |
+
"train_loss": 4.124692803535324,
|
| 5 |
+
"train_actor_loss": -0.18059303024421797,
|
| 6 |
+
"train_critic_loss": 2.03618444023577,
|
| 7 |
+
"train_mean_reward": -0.006020427858927108,
|
| 8 |
+
"train_entropy": 0.34074558541536903,
|
| 9 |
+
"train_ce_loss": 5.015896934586076,
|
| 10 |
+
"train_dice_loss": 1.5584899806059325,
|
| 11 |
+
"train_dice": 0.2888267162087181,
|
| 12 |
+
"train_iou": 0.20086716557146847,
|
| 13 |
+
"grad_norm": NaN,
|
| 14 |
+
"lr": 0.001999509027563903,
|
| 15 |
+
"encoder_lr": 0.0001999531232347445,
|
| 16 |
+
"alpha": 0.4999520480632782,
|
| 17 |
+
"validated_this_epoch": true,
|
| 18 |
+
"val_loss": 1.2481977726248177,
|
| 19 |
+
"val_dice": 0.33219268283250747,
|
| 20 |
+
"val_iou": 0.24802803193665973,
|
| 21 |
+
"val_reward": -0.004744514358915492,
|
| 22 |
+
"val_entropy": 0.2566183015252604,
|
| 23 |
+
"val_actor_loss": -0.011879473347925332,
|
| 24 |
+
"val_critic_loss": 0.0006283396340755516,
|
| 25 |
+
"val_ce_loss": 0.9255689848402534,
|
| 26 |
+
"val_dice_loss": 1.5939571381819369
|
| 27 |
+
},
|
| 28 |
+
{
|
| 29 |
+
"epoch": 2,
|
| 30 |
+
"train_loss": 1.2309506776241155,
|
| 31 |
+
"train_actor_loss": -0.2894082183066087,
|
| 32 |
+
"train_critic_loss": 0.001025058574062617,
|
| 33 |
+
"train_mean_reward": -0.003943595153918394,
|
| 34 |
+
"train_entropy": 0.5190237962091581,
|
| 35 |
+
"train_ce_loss": 1.4911093896207137,
|
| 36 |
+
"train_dice_loss": 1.5485833019782336,
|
| 37 |
+
"train_dice": 0.2531890264649251,
|
| 38 |
+
"train_iou": 0.1766477307624061,
|
| 39 |
+
"grad_norm": 29.52235854130525,
|
| 40 |
+
"lr": 0.0019980365947861304,
|
| 41 |
+
"encoder_lr": 0.00019981253920068581,
|
| 42 |
+
"alpha": 0.49870994687080383,
|
| 43 |
+
"validated_this_epoch": true,
|
| 44 |
+
"val_loss": 1.6669950979677113,
|
| 45 |
+
"val_dice": 0.24753497359822357,
|
| 46 |
+
"val_iou": 0.1790499248930131,
|
| 47 |
+
"val_reward": 4.365709887190615e-05,
|
| 48 |
+
"val_entropy": 0.8971405366454462,
|
| 49 |
+
"val_actor_loss": -0.013084225648205087,
|
| 50 |
+
"val_critic_loss": 3.668231779510729e-05,
|
| 51 |
+
"val_ce_loss": 1.7032046027556813,
|
| 52 |
+
"val_dice_loss": 1.6569173068109186
|
| 53 |
+
},
|
| 54 |
+
{
|
| 55 |
+
"epoch": 3,
|
| 56 |
+
"train_loss": 0.9048539450248847,
|
| 57 |
+
"train_actor_loss": -0.3748849142247285,
|
| 58 |
+
"train_critic_loss": 0.0008304766358774849,
|
| 59 |
+
"train_mean_reward": -0.0036579365157358087,
|
| 60 |
+
"train_entropy": 0.6490705297925532,
|
| 61 |
+
"train_ce_loss": 1.049941763520623,
|
| 62 |
+
"train_dice_loss": 1.5087054417683528,
|
| 63 |
+
"train_dice": 0.31027342578140543,
|
| 64 |
+
"train_iou": 0.21426028343757578,
|
| 65 |
+
"grad_norm": 14.849804814045246,
|
| 66 |
+
"lr": 0.001995584154780065,
|
| 67 |
+
"encoder_lr": 0.0001995783866372926,
|
| 68 |
+
"alpha": 0.4967534840106964,
|
| 69 |
+
"validated_this_epoch": true,
|
| 70 |
+
"val_loss": 1.311537363876899,
|
| 71 |
+
"val_dice": 0.21748602584764165,
|
| 72 |
+
"val_iou": 0.16013215092125796,
|
| 73 |
+
"val_reward": -0.0020831877589425674,
|
| 74 |
+
"val_entropy": 0.8759821578107698,
|
| 75 |
+
"val_actor_loss": 0.0028199513673293165,
|
| 76 |
+
"val_critic_loss": 0.0006852614320322196,
|
| 77 |
+
"val_ce_loss": 1.0011459461817838,
|
| 78 |
+
"val_dice_loss": 1.6156035761038463
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"epoch": 4,
|
| 82 |
+
"train_loss": 0.8430438931458272,
|
| 83 |
+
"train_actor_loss": -0.3992089203152901,
|
| 84 |
+
"train_critic_loss": 0.0008437139081368701,
|
| 85 |
+
"train_mean_reward": -0.003583540964404491,
|
| 86 |
+
"train_entropy": 0.7006555653344363,
|
| 87 |
+
"train_ce_loss": 1.014484960394792,
|
| 88 |
+
"train_dice_loss": 1.4691769278202302,
|
| 89 |
+
"train_dice": 0.37631276985871226,
|
| 90 |
+
"train_iou": 0.26960600427108805,
|
| 91 |
+
"grad_norm": 11.449078403986418,
|
| 92 |
+
"lr": 0.001992154127807906,
|
| 93 |
+
"encoder_lr": 0.0001992508966248754,
|
| 94 |
+
"alpha": 0.49389657378196716,
|
| 95 |
+
"validated_this_epoch": true,
|
| 96 |
+
"val_loss": 1.1028843465628047,
|
| 97 |
+
"val_dice": 0.3450055919058855,
|
| 98 |
+
"val_iou": 0.2603861787226177,
|
| 99 |
+
"val_reward": -0.004785676271157255,
|
| 100 |
+
"val_entropy": 0.14012644680762532,
|
| 101 |
+
"val_actor_loss": 0.001968951509031349,
|
| 102 |
+
"val_critic_loss": 0.0007624242499525391,
|
| 103 |
+
"val_ce_loss": 0.6469488314004859,
|
| 104 |
+
"val_dice_loss": 1.5541195014510492
|
| 105 |
+
},
|
| 106 |
+
{
|
| 107 |
+
"epoch": 5,
|
| 108 |
+
"train_loss": 0.8307947535067797,
|
| 109 |
+
"train_actor_loss": -0.31379038835756295,
|
| 110 |
+
"train_critic_loss": 0.0008844932546288384,
|
| 111 |
+
"train_mean_reward": -0.006659194950211363,
|
| 112 |
+
"train_entropy": 0.5813854307127305,
|
| 113 |
+
"train_ce_loss": 0.8410559860177529,
|
| 114 |
+
"train_dice_loss": 1.4472297896177342,
|
| 115 |
+
"train_dice": 0.4250805232150118,
|
| 116 |
+
"train_iou": 0.31525707984510837,
|
| 117 |
+
"grad_norm": 8.589496713418226,
|
| 118 |
+
"lr": 0.0019877498988921626,
|
| 119 |
+
"encoder_lr": 0.0001988303923565381,
|
| 120 |
+
"alpha": 0.4920901954174042,
|
| 121 |
+
"validated_this_epoch": true,
|
| 122 |
+
"val_loss": 1.0295618523025152,
|
| 123 |
+
"val_dice": 0.3772401471932099,
|
| 124 |
+
"val_iou": 0.29026529405749285,
|
| 125 |
+
"val_reward": -0.008613752141815957,
|
| 126 |
+
"val_entropy": 0.6089879316512985,
|
| 127 |
+
"val_actor_loss": -0.013286725462724767,
|
| 128 |
+
"val_critic_loss": 0.0006117936610289781,
|
| 129 |
+
"val_ce_loss": 0.6055742956020614,
|
| 130 |
+
"val_dice_loss": 1.479511038221494
|
| 131 |
+
},
|
| 132 |
+
{
|
| 133 |
+
"epoch": 6,
|
| 134 |
+
"train_loss": 0.7223273553670599,
|
| 135 |
+
"train_actor_loss": -0.399001368584159,
|
| 136 |
+
"train_critic_loss": 0.0008885911050198886,
|
| 137 |
+
"train_mean_reward": -0.002872174513098658,
|
| 138 |
+
"train_entropy": 0.6880745768356018,
|
| 139 |
+
"train_ce_loss": 0.8429453945121705,
|
| 140 |
+
"train_dice_loss": 1.398823442367407,
|
| 141 |
+
"train_dice": 0.4658172537262241,
|
| 142 |
+
"train_iou": 0.3549031455928058,
|
| 143 |
+
"grad_norm": 6.742678358004643,
|
| 144 |
+
"lr": 0.0019823758144750458,
|
| 145 |
+
"encoder_lr": 0.00019831728881922545,
|
| 146 |
+
"alpha": 0.48958855867385864,
|
| 147 |
+
"validated_this_epoch": true,
|
| 148 |
+
"val_loss": 1.0685242997538862,
|
| 149 |
+
"val_dice": 0.4820899483330293,
|
| 150 |
+
"val_iou": 0.37349202623714367,
|
| 151 |
+
"val_reward": -0.004147537892673758,
|
| 152 |
+
"val_entropy": 0.8524880159382868,
|
| 153 |
+
"val_actor_loss": -0.009944345999621986,
|
| 154 |
+
"val_critic_loss": 0.00041824248356450965,
|
| 155 |
+
"val_ce_loss": 0.7303576633108385,
|
| 156 |
+
"val_dice_loss": 1.4261613542383365
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"epoch": 7,
|
| 160 |
+
"train_loss": 0.6978735817739596,
|
| 161 |
+
"train_actor_loss": -0.39513763699394,
|
| 162 |
+
"train_critic_loss": 0.0008355347454198635,
|
| 163 |
+
"train_mean_reward": -0.006269752712866578,
|
| 164 |
+
"train_entropy": 0.7092370855120513,
|
| 165 |
+
"train_ce_loss": 0.8028254902515657,
|
| 166 |
+
"train_dice_loss": 1.3823613952367735,
|
| 167 |
+
"train_dice": 0.43780810566728606,
|
| 168 |
+
"train_iou": 0.3209935353712781,
|
| 169 |
+
"grad_norm": 7.542747951470888,
|
| 170 |
+
"lr": 0.001976037178129054,
|
| 171 |
+
"encoder_lr": 0.00019771209238418101,
|
| 172 |
+
"alpha": 0.48685547709465027,
|
| 173 |
+
"validated_this_epoch": true,
|
| 174 |
+
"val_loss": 1.0627237464103736,
|
| 175 |
+
"val_dice": 0.42570455135171337,
|
| 176 |
+
"val_iou": 0.3345533035083298,
|
| 177 |
+
"val_reward": -0.00582087797058671,
|
| 178 |
+
"val_entropy": 0.44100411115872734,
|
| 179 |
+
"val_actor_loss": 0.0002196249387443366,
|
| 180 |
+
"val_critic_loss": 0.0006813791262377184,
|
| 181 |
+
"val_ce_loss": 0.6358245179583929,
|
| 182 |
+
"val_dice_loss": 1.4885023167037
|
| 183 |
+
},
|
| 184 |
+
{
|
| 185 |
+
"epoch": 8,
|
| 186 |
+
"train_loss": 0.6799760330468416,
|
| 187 |
+
"train_actor_loss": -0.39620169538718003,
|
| 188 |
+
"train_critic_loss": 0.0007662147344933308,
|
| 189 |
+
"train_mean_reward": -0.006801484580743516,
|
| 190 |
+
"train_entropy": 0.7274588676981436,
|
| 191 |
+
"train_ce_loss": 0.8132472054029887,
|
| 192 |
+
"train_dice_loss": 1.3383420155598564,
|
| 193 |
+
"train_dice": 0.4751703827550598,
|
| 194 |
+
"train_iou": 0.3532146776788959,
|
| 195 |
+
"grad_norm": 6.465040908409999,
|
| 196 |
+
"lr": 0.001968740245322988,
|
| 197 |
+
"encoder_lr": 0.00019701540030721995,
|
| 198 |
+
"alpha": 0.48420894145965576,
|
| 199 |
+
"validated_this_epoch": true,
|
| 200 |
+
"val_loss": 1.0429411728047964,
|
| 201 |
+
"val_dice": 0.5258612001348579,
|
| 202 |
+
"val_iou": 0.42542386446486347,
|
| 203 |
+
"val_reward": -0.0028045372007064278,
|
| 204 |
+
"val_entropy": 0.7958601117134094,
|
| 205 |
+
"val_actor_loss": -0.007244685599509851,
|
| 206 |
+
"val_critic_loss": 0.0005143134798290151,
|
| 207 |
+
"val_ce_loss": 0.678417847087287,
|
| 208 |
+
"val_dice_loss": 1.4214395290673383
|
| 209 |
+
},
|
| 210 |
+
{
|
| 211 |
+
"epoch": 9,
|
| 212 |
+
"train_loss": 0.6721410072910098,
|
| 213 |
+
"train_actor_loss": -0.37911497868406463,
|
| 214 |
+
"train_critic_loss": 0.0006894378362071271,
|
| 215 |
+
"train_mean_reward": -0.006053996880207425,
|
| 216 |
+
"train_entropy": 0.6732382929095854,
|
| 217 |
+
"train_ce_loss": 0.7564802700892473,
|
| 218 |
+
"train_dice_loss": 1.3453422433290727,
|
| 219 |
+
"train_dice": 0.5011906623703373,
|
| 220 |
+
"train_iou": 0.3809945909078142,
|
| 221 |
+
"grad_norm": 6.68727485033182,
|
| 222 |
+
"lr": 0.0019604922172485584,
|
| 223 |
+
"encoder_lr": 0.0001962279001393096,
|
| 224 |
+
"alpha": 0.48180973529815674,
|
| 225 |
+
"validated_this_epoch": true,
|
| 226 |
+
"val_loss": 1.046524871253606,
|
| 227 |
+
"val_dice": 0.5323191724065003,
|
| 228 |
+
"val_iou": 0.4267072540270193,
|
| 229 |
+
"val_reward": -0.0018958590408514318,
|
| 230 |
+
"val_entropy": 0.7997841467761031,
|
| 231 |
+
"val_actor_loss": -0.005723677820177994,
|
| 232 |
+
"val_critic_loss": 0.0003985267308099087,
|
| 233 |
+
"val_ce_loss": 0.6792289104425545,
|
| 234 |
+
"val_dice_loss": 1.4248696240511807
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"epoch": 10,
|
| 238 |
+
"train_loss": 0.7389564058528497,
|
| 239 |
+
"train_actor_loss": -0.3462359714441193,
|
| 240 |
+
"train_critic_loss": 0.0007087189578902698,
|
| 241 |
+
"train_mean_reward": -0.008154889420928577,
|
| 242 |
+
"train_entropy": 0.6210864984836333,
|
| 243 |
+
"train_ce_loss": 0.8403652332818662,
|
| 244 |
+
"train_dice_loss": 1.3293107858835127,
|
| 245 |
+
"train_dice": 0.48796748158101944,
|
| 246 |
+
"train_iou": 0.3614892897871779,
|
| 247 |
+
"grad_norm": 8.674190493730398,
|
| 248 |
+
"lr": 0.0019513012337136779,
|
| 249 |
+
"encoder_lr": 0.00019535036904803958,
|
| 250 |
+
"alpha": 0.47983747720718384,
|
| 251 |
+
"validated_this_epoch": true,
|
| 252 |
+
"val_loss": 0.9267651131658843,
|
| 253 |
+
"val_dice": 0.5515658567584873,
|
| 254 |
+
"val_iou": 0.44561434552902723,
|
| 255 |
+
"val_reward": -0.0006305682715049102,
|
| 256 |
+
"val_entropy": 0.4923267379553631,
|
| 257 |
+
"val_actor_loss": -0.01214622680775144,
|
| 258 |
+
"val_critic_loss": 0.00029055901833282513,
|
| 259 |
+
"val_ce_loss": 0.5030474558924184,
|
| 260 |
+
"val_dice_loss": 1.3744846344596209
|
| 261 |
+
},
|
| 262 |
+
{
|
| 263 |
+
"epoch": 11,
|
| 264 |
+
"train_loss": 0.6509717524481508,
|
| 265 |
+
"train_actor_loss": -0.3313698101884279,
|
| 266 |
+
"train_critic_loss": 0.0007098117410537385,
|
| 267 |
+
"train_mean_reward": -0.00622904040919354,
|
| 268 |
+
"train_entropy": 0.5706072659828724,
|
| 269 |
+
"train_ce_loss": 0.6570755381566974,
|
| 270 |
+
"train_dice_loss": 1.3068977582913182,
|
| 271 |
+
"train_dice": 0.5485902670924944,
|
| 272 |
+
"train_iou": 0.4249143655595633,
|
| 273 |
+
"grad_norm": 7.255294483441573,
|
| 274 |
+
"lr": 0.0019411763651094546,
|
| 275 |
+
"encoder_lr": 0.0001943836730506514,
|
| 276 |
+
"alpha": 0.47827038168907166,
|
| 277 |
+
"validated_this_epoch": true,
|
| 278 |
+
"val_loss": 1.0751198138477225,
|
| 279 |
+
"val_dice": 0.5475425656002981,
|
| 280 |
+
"val_iou": 0.4405522114454123,
|
| 281 |
+
"val_reward": 0.0010821664215680442,
|
| 282 |
+
"val_entropy": 0.8266470956079888,
|
| 283 |
+
"val_actor_loss": -0.000597621373023198,
|
| 284 |
+
"val_critic_loss": 0.00010793952133834028,
|
| 285 |
+
"val_ce_loss": 0.7394881703787379,
|
| 286 |
+
"val_dice_loss": 1.4118387298752564
|
| 287 |
+
},
|
| 288 |
+
{
|
| 289 |
+
"epoch": 12,
|
| 290 |
+
"train_loss": 0.6016950681805611,
|
| 291 |
+
"train_actor_loss": -0.396755995372167,
|
| 292 |
+
"train_critic_loss": 0.0005460568622182994,
|
| 293 |
+
"train_mean_reward": -0.005792910828276011,
|
| 294 |
+
"train_entropy": 0.7067173516903167,
|
| 295 |
+
"train_ce_loss": 0.7172150117082474,
|
| 296 |
+
"train_dice_loss": 1.2791410440053697,
|
| 297 |
+
"train_dice": 0.5604750568004877,
|
| 298 |
+
"train_iou": 0.4384531777857642,
|
| 299 |
+
"grad_norm": 4.683626351448206,
|
| 300 |
+
"lr": 0.0019301276034588103,
|
| 301 |
+
"encoder_lr": 0.00019332876615938386,
|
| 302 |
+
"alpha": 0.47619569301605225,
|
| 303 |
+
"validated_this_epoch": true,
|
| 304 |
+
"val_loss": 1.077100802100066,
|
| 305 |
+
"val_dice": 0.5708392699248853,
|
| 306 |
+
"val_iou": 0.47100899237778837,
|
| 307 |
+
"val_reward": 0.0009411644826130469,
|
| 308 |
+
"val_entropy": 0.8854880787507451,
|
| 309 |
+
"val_actor_loss": 0.0005156704650326357,
|
| 310 |
+
"val_critic_loss": 0.0002203960268271851,
|
| 311 |
+
"val_ce_loss": 0.7566127044082892,
|
| 312 |
+
"val_dice_loss": 1.3963371325622906
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"epoch": 13,
|
| 316 |
+
"train_loss": 0.6213419658418459,
|
| 317 |
+
"train_actor_loss": -0.40174111924492395,
|
| 318 |
+
"train_critic_loss": 0.0007154353506148953,
|
| 319 |
+
"train_mean_reward": -0.005412778684097785,
|
| 320 |
+
"train_entropy": 0.6890547539179142,
|
| 321 |
+
"train_ce_loss": 0.7297053460676509,
|
| 322 |
+
"train_dice_loss": 1.3157453621045136,
|
| 323 |
+
"train_dice": 0.5457709805409917,
|
| 324 |
+
"train_iou": 0.42751025366505574,
|
| 325 |
+
"grad_norm": 5.943756140195406,
|
| 326 |
+
"lr": 0.0019181658525555612,
|
| 327 |
+
"encoder_lr": 0.00019218668943997818,
|
| 328 |
+
"alpha": 0.4736851155757904,
|
| 329 |
+
"validated_this_epoch": true,
|
| 330 |
+
"val_loss": 0.9736593281568,
|
| 331 |
+
"val_dice": 0.5105214312041461,
|
| 332 |
+
"val_iou": 0.4070321482355667,
|
| 333 |
+
"val_reward": -0.008295658654815546,
|
| 334 |
+
"val_entropy": 0.5757494463463022,
|
| 335 |
+
"val_actor_loss": -0.013889056650276128,
|
| 336 |
+
"val_critic_loss": 0.0006183739253835731,
|
| 337 |
+
"val_ce_loss": 0.5348245901516592,
|
| 338 |
+
"val_dice_loss": 1.4396537765107975
|
| 339 |
+
},
|
| 340 |
+
{
|
| 341 |
+
"epoch": 14,
|
| 342 |
+
"train_loss": 0.578102036486738,
|
| 343 |
+
"train_actor_loss": -0.399635469302153,
|
| 344 |
+
"train_critic_loss": 0.0003892217091327975,
|
| 345 |
+
"train_mean_reward": -0.006363907533271301,
|
| 346 |
+
"train_entropy": 0.717309886446366,
|
| 347 |
+
"train_ce_loss": 0.6686721386340183,
|
| 348 |
+
"train_dice_loss": 1.2864136264110222,
|
| 349 |
+
"train_dice": 0.6148358661151838,
|
| 350 |
+
"train_iou": 0.49235872810651715,
|
| 351 |
+
"grad_norm": 4.8587432113977576,
|
| 352 |
+
"lr": 0.0019053029172036895,
|
| 353 |
+
"encoder_lr": 0.00019095856998427185,
|
| 354 |
+
"alpha": 0.4713596701622009,
|
| 355 |
+
"validated_this_epoch": true,
|
| 356 |
+
"val_loss": 0.9353481994546724,
|
| 357 |
+
"val_dice": 0.5808719194452687,
|
| 358 |
+
"val_iou": 0.47335888957189765,
|
| 359 |
+
"val_reward": -0.00330105487823444,
|
| 360 |
+
"val_entropy": 0.6282972597112559,
|
| 361 |
+
"val_actor_loss": -0.011903125377644718,
|
| 362 |
+
"val_critic_loss": 0.000270995189879099,
|
| 363 |
+
"val_ce_loss": 0.534408400117448,
|
| 364 |
+
"val_dice_loss": 1.3598232254235434
|
| 365 |
+
},
|
| 366 |
+
{
|
| 367 |
+
"epoch": 15,
|
| 368 |
+
"train_loss": 0.5584182431921363,
|
| 369 |
+
"train_actor_loss": -0.41171574057676846,
|
| 370 |
+
"train_critic_loss": 0.0006359130596441295,
|
| 371 |
+
"train_mean_reward": -0.006525103624041479,
|
| 372 |
+
"train_entropy": 0.7285724723568328,
|
| 373 |
+
"train_ce_loss": 0.6874248937297709,
|
| 374 |
+
"train_dice_loss": 1.2522071435665474,
|
| 375 |
+
"train_dice": 0.5810045037491649,
|
| 376 |
+
"train_iou": 0.4489766352207345,
|
| 377 |
+
"grad_norm": 6.053451258402604,
|
| 378 |
+
"lr": 0.001891551491567426,
|
| 379 |
+
"encoder_lr": 0.00018964561979789495,
|
| 380 |
+
"alpha": 0.4688485264778137,
|
| 381 |
+
"validated_this_epoch": true,
|
| 382 |
+
"val_loss": 0.9409299138368983,
|
| 383 |
+
"val_dice": 0.5661670531514399,
|
| 384 |
+
"val_iou": 0.46389403957973274,
|
| 385 |
+
"val_reward": -0.004888855943025908,
|
| 386 |
+
"val_entropy": 0.534103697297549,
|
| 387 |
+
"val_actor_loss": -0.0023174517137713427,
|
| 388 |
+
"val_critic_loss": 0.0005037098429359806,
|
| 389 |
+
"val_ce_loss": 0.49119276884529334,
|
| 390 |
+
"val_dice_loss": 1.3947982228163518
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"epoch": 16,
|
| 394 |
+
"train_loss": 0.5688677227411133,
|
| 395 |
+
"train_actor_loss": -0.36575121957904255,
|
| 396 |
+
"train_critic_loss": 0.0005062387153418221,
|
| 397 |
+
"train_mean_reward": -0.006006375938550963,
|
| 398 |
+
"train_entropy": 0.6541382865263865,
|
| 399 |
+
"train_ce_loss": 0.6072982923151591,
|
| 400 |
+
"train_dice_loss": 1.2614333339226553,
|
| 401 |
+
"train_dice": 0.6014604136985773,
|
| 402 |
+
"train_iou": 0.48738173702012694,
|
| 403 |
+
"grad_norm": 4.380005451349112,
|
| 404 |
+
"lr": 0.0018769251466436443,
|
| 405 |
+
"encoder_lr": 0.00018824913460416704,
|
| 406 |
+
"alpha": 0.4667417109012604,
|
| 407 |
+
"validated_this_epoch": true,
|
| 408 |
+
"val_loss": 1.0705410693631028,
|
| 409 |
+
"val_dice": 0.552851535049179,
|
| 410 |
+
"val_iou": 0.4420785912945701,
|
| 411 |
+
"val_reward": 0.0029147340451758782,
|
| 412 |
+
"val_entropy": 0.8061727288395468,
|
| 413 |
+
"val_actor_loss": 0.017850597762982506,
|
| 414 |
+
"val_critic_loss": 0.00038014610719175026,
|
| 415 |
+
"val_ce_loss": 0.6917726347843806,
|
| 416 |
+
"val_dice_loss": 1.4132281364214543
|
| 417 |
+
},
|
| 418 |
+
{
|
| 419 |
+
"epoch": 17,
|
| 420 |
+
"train_loss": 0.5323217533743725,
|
| 421 |
+
"train_actor_loss": -0.4315029327303935,
|
| 422 |
+
"train_critic_loss": 0.0005471801121016957,
|
| 423 |
+
"train_mean_reward": -0.005546206374091502,
|
| 424 |
+
"train_entropy": 0.7673012278018855,
|
| 425 |
+
"train_ce_loss": 0.6458314904369987,
|
| 426 |
+
"train_dice_loss": 1.281270696184574,
|
| 427 |
+
"train_dice": 0.6026678247669456,
|
| 428 |
+
"train_iou": 0.47407131383706835,
|
| 429 |
+
"grad_norm": 4.624961532079256,
|
| 430 |
+
"lr": 0.001861438316868924,
|
| 431 |
+
"encoder_lr": 0.00018677049256537467,
|
| 432 |
+
"alpha": 0.4641127586364746,
|
| 433 |
+
"validated_this_epoch": true,
|
| 434 |
+
"val_loss": 0.9380198700466391,
|
| 435 |
+
"val_dice": 0.6160756949092818,
|
| 436 |
+
"val_iou": 0.49941991857478524,
|
| 437 |
+
"val_reward": -6.568288466436866e-05,
|
| 438 |
+
"val_entropy": 0.6665921671824022,
|
| 439 |
+
"val_actor_loss": 0.0013212382657020682,
|
| 440 |
+
"val_critic_loss": 0.0003962774656548675,
|
| 441 |
+
"val_ce_loss": 0.5083370277037224,
|
| 442 |
+
"val_dice_loss": 1.3646639326606134
|
| 443 |
+
},
|
| 444 |
+
{
|
| 445 |
+
"epoch": 18,
|
| 446 |
+
"train_loss": 0.6178502566539325,
|
| 447 |
+
"train_actor_loss": -0.3371816512006215,
|
| 448 |
+
"train_critic_loss": 0.0005354955418834623,
|
| 449 |
+
"train_mean_reward": -0.006576303519698111,
|
| 450 |
+
"train_entropy": 0.6129026954563765,
|
| 451 |
+
"train_ce_loss": 0.6616315103064363,
|
| 452 |
+
"train_dice_loss": 1.2478967908865368,
|
| 453 |
+
"train_dice": 0.6383710791600992,
|
| 454 |
+
"train_iou": 0.5219767988501824,
|
| 455 |
+
"grad_norm": 5.774507054915795,
|
| 456 |
+
"lr": 0.0018451062858745053,
|
| 457 |
+
"encoder_lr": 0.00018521115292269145,
|
| 458 |
+
"alpha": 0.46186473965644836,
|
| 459 |
+
"validated_this_epoch": true,
|
| 460 |
+
"val_loss": 0.9476827522739768,
|
| 461 |
+
"val_dice": 0.5905338374490555,
|
| 462 |
+
"val_iou": 0.47460642974476186,
|
| 463 |
+
"val_reward": 0.0013835704784919424,
|
| 464 |
+
"val_entropy": 0.6066158997892129,
|
| 465 |
+
"val_actor_loss": -0.010646385336212455,
|
| 466 |
+
"val_critic_loss": 0.00014282232353647799,
|
| 467 |
+
"val_ce_loss": 0.5419457867327663,
|
| 468 |
+
"val_dice_loss": 1.3745696409182115
|
| 469 |
+
},
|
| 470 |
+
{
|
| 471 |
+
"epoch": 19,
|
| 472 |
+
"train_loss": 0.5352323791012168,
|
| 473 |
+
"train_actor_loss": -0.3727663930696554,
|
| 474 |
+
"train_critic_loss": 0.00043957579565847904,
|
| 475 |
+
"train_mean_reward": -0.007798636216205359,
|
| 476 |
+
"train_entropy": 0.6371233698267204,
|
| 477 |
+
"train_ce_loss": 0.587518162213457,
|
| 478 |
+
"train_dice_loss": 1.2280397904224885,
|
| 479 |
+
"train_dice": 0.638494775648795,
|
| 480 |
+
"train_iou": 0.5116780939712828,
|
| 481 |
+
"grad_norm": 5.645545601844788,
|
| 482 |
+
"lr": 0.0018279451714031893,
|
| 483 |
+
"encoder_lr": 0.0001835726545560834,
|
| 484 |
+
"alpha": 0.4601013660430908,
|
| 485 |
+
"validated_this_epoch": true,
|
| 486 |
+
"val_loss": 1.0493392618373036,
|
| 487 |
+
"val_dice": 0.46671734228199285,
|
| 488 |
+
"val_iou": 0.3684567259428228,
|
| 489 |
+
"val_reward": 0.00017481862898955654,
|
| 490 |
+
"val_entropy": 0.6736092802250024,
|
| 491 |
+
"val_actor_loss": -0.002973316388435172,
|
| 492 |
+
"val_critic_loss": 0.00011472162816172579,
|
| 493 |
+
"val_ce_loss": 0.6656508388355225,
|
| 494 |
+
"val_dice_loss": 1.4388595726152862
|
| 495 |
+
},
|
| 496 |
+
{
|
| 497 |
+
"epoch": 20,
|
| 498 |
+
"train_loss": 0.521827786659392,
|
| 499 |
+
"train_actor_loss": -0.37240563075129796,
|
| 500 |
+
"train_critic_loss": 0.0004011863822644302,
|
| 501 |
+
"train_mean_reward": -0.006314502775263136,
|
| 502 |
+
"train_entropy": 0.6671864271928102,
|
| 503 |
+
"train_ce_loss": 0.599286351687251,
|
| 504 |
+
"train_dice_loss": 1.1887792887595983,
|
| 505 |
+
"train_dice": 0.653482902061091,
|
| 506 |
+
"train_iou": 0.5338838891383032,
|
| 507 |
+
"grad_norm": 4.8063357541194325,
|
| 508 |
+
"lr": 0.001809971909403073,
|
| 509 |
+
"encoder_lr": 0.00018185661446562003,
|
| 510 |
+
"alpha": 0.4581984877586365,
|
| 511 |
+
"validated_this_epoch": true,
|
| 512 |
+
"val_loss": 1.2125592759782167,
|
| 513 |
+
"val_dice": 0.49222956679207874,
|
| 514 |
+
"val_iou": 0.39768468559667575,
|
| 515 |
+
"val_reward": 0.00026796316054782663,
|
| 516 |
+
"val_entropy": 0.9327594333826894,
|
| 517 |
+
"val_actor_loss": -0.0019848762789148496,
|
| 518 |
+
"val_critic_loss": 3.964200986772975e-05,
|
| 519 |
+
"val_ce_loss": 0.9674912371456321,
|
| 520 |
+
"val_dice_loss": 1.4615573582023083
|
| 521 |
+
},
|
| 522 |
+
{
|
| 523 |
+
"epoch": 21,
|
| 524 |
+
"train_loss": 0.5716410731323637,
|
| 525 |
+
"train_actor_loss": -0.3427719430854687,
|
| 526 |
+
"train_critic_loss": 0.0005440684444926237,
|
| 527 |
+
"train_mean_reward": -0.005628641693571505,
|
| 528 |
+
"train_entropy": 0.597291923295229,
|
| 529 |
+
"train_ce_loss": 0.5940375224305078,
|
| 530 |
+
"train_dice_loss": 1.2342444176857288,
|
| 531 |
+
"train_dice": 0.611721202932006,
|
| 532 |
+
"train_iou": 0.4905902642172341,
|
| 533 |
+
"grad_norm": 5.326637712808756,
|
| 534 |
+
"lr": 0.001791204237313812,
|
| 535 |
+
"encoder_lr": 0.0001800647261756906,
|
| 536 |
+
"alpha": 0.4561549425125122,
|
| 537 |
+
"validated_this_epoch": true,
|
| 538 |
+
"val_loss": 0.9314823713379375,
|
| 539 |
+
"val_dice": 0.569693804093768,
|
| 540 |
+
"val_iou": 0.46175878022001016,
|
| 541 |
+
"val_reward": -0.001326949094352195,
|
| 542 |
+
"val_entropy": 0.3209739920466837,
|
| 543 |
+
"val_actor_loss": 0.015528204020658378,
|
| 544 |
+
"val_critic_loss": 0.0004740314928340288,
|
| 545 |
+
"val_ce_loss": 0.443711575609867,
|
| 546 |
+
"val_dice_loss": 1.3877226999311734
|
| 547 |
+
},
|
| 548 |
+
{
|
| 549 |
+
"epoch": 22,
|
| 550 |
+
"train_loss": 0.5155629250220954,
|
| 551 |
+
"train_actor_loss": -0.3712368940886778,
|
| 552 |
+
"train_critic_loss": 0.0003984542297005716,
|
| 553 |
+
"train_mean_reward": -0.005965902420519762,
|
| 554 |
+
"train_entropy": 0.643367672100281,
|
| 555 |
+
"train_ce_loss": 0.5735999288467261,
|
| 556 |
+
"train_dice_loss": 1.1996012391188207,
|
| 557 |
+
"train_dice": 0.6514498393790906,
|
| 558 |
+
"train_iou": 0.53026083707079,
|
| 559 |
+
"grad_norm": 4.837438294520745,
|
| 560 |
+
"lr": 0.0017716606765619107,
|
| 561 |
+
"encoder_lr": 0.0001781987580637,
|
| 562 |
+
"alpha": 0.45449692010879517,
|
| 563 |
+
"validated_this_epoch": true,
|
| 564 |
+
"val_loss": 1.0097878981720319,
|
| 565 |
+
"val_dice": 0.5300866137243542,
|
| 566 |
+
"val_iou": 0.42341532862163717,
|
| 567 |
+
"val_reward": -0.0012309543248689313,
|
| 568 |
+
"val_entropy": 0.5555776898005996,
|
| 569 |
+
"val_actor_loss": 0.02268235039817301,
|
| 570 |
+
"val_critic_loss": 0.0004532229865773586,
|
| 571 |
+
"val_ce_loss": 0.5625683499707117,
|
| 572 |
+
"val_dice_loss": 1.4111894913996108
|
| 573 |
+
},
|
| 574 |
+
{
|
| 575 |
+
"epoch": 23,
|
| 576 |
+
"train_loss": 0.5090543657827836,
|
| 577 |
+
"train_actor_loss": -0.3717064187885859,
|
| 578 |
+
"train_critic_loss": 0.0004251376425926514,
|
| 579 |
+
"train_mean_reward": -0.006918917682457669,
|
| 580 |
+
"train_entropy": 0.6672209306405141,
|
| 581 |
+
"train_ce_loss": 0.5576843132670873,
|
| 582 |
+
"train_dice_loss": 1.2034120892102902,
|
| 583 |
+
"train_dice": 0.6986942014554317,
|
| 584 |
+
"train_iou": 0.5739240133721805,
|
| 585 |
+
"grad_norm": 4.3708025973576765,
|
| 586 |
+
"lr": 0.0017513605142823077,
|
| 587 |
+
"encoder_lr": 0.00017626055161489368,
|
| 588 |
+
"alpha": 0.45233091711997986,
|
| 589 |
+
"validated_this_epoch": true,
|
| 590 |
+
"val_loss": 0.9478603669752678,
|
| 591 |
+
"val_dice": 0.5910479322213954,
|
| 592 |
+
"val_iou": 0.4943642901967912,
|
| 593 |
+
"val_reward": -0.001200493672219425,
|
| 594 |
+
"val_entropy": 0.5320299665133158,
|
| 595 |
+
"val_actor_loss": 0.014567140895506423,
|
| 596 |
+
"val_critic_loss": 0.0003613642083376298,
|
| 597 |
+
"val_ce_loss": 0.49089928261107874,
|
| 598 |
+
"val_dice_loss": 1.3753257782170267
|
| 599 |
+
},
|
| 600 |
+
{
|
| 601 |
+
"epoch": 24,
|
| 602 |
+
"train_loss": 0.5092227396268684,
|
| 603 |
+
"train_actor_loss": -0.37688723932474094,
|
| 604 |
+
"train_critic_loss": 0.00048598178340725077,
|
| 605 |
+
"train_mean_reward": -0.005973524869403215,
|
| 606 |
+
"train_entropy": 0.6672153358276075,
|
| 607 |
+
"train_ce_loss": 0.5645445704173583,
|
| 608 |
+
"train_dice_loss": 1.2071893876179671,
|
| 609 |
+
"train_dice": 0.6591415298514163,
|
| 610 |
+
"train_iou": 0.5380809203814454,
|
| 611 |
+
"grad_norm": 4.776018261909485,
|
| 612 |
+
"lr": 0.001730323784284305,
|
| 613 |
+
"encoder_lr": 0.0001742520196050341,
|
| 614 |
+
"alpha": 0.45029836893081665,
|
| 615 |
+
"validated_this_epoch": true,
|
| 616 |
+
"val_loss": 0.9097729508454601,
|
| 617 |
+
"val_dice": 0.5204477280631221,
|
| 618 |
+
"val_iou": 0.4225808404764294,
|
| 619 |
+
"val_reward": -0.00903565783503129,
|
| 620 |
+
"val_entropy": 0.3755855101226556,
|
| 621 |
+
"val_actor_loss": -0.007876190943396508,
|
| 622 |
+
"val_critic_loss": 0.0005201082222636689,
|
| 623 |
+
"val_ce_loss": 0.45803871508123295,
|
| 624 |
+
"val_dice_loss": 1.3767394318123056
|
| 625 |
+
},
|
| 626 |
+
{
|
| 627 |
+
"epoch": 25,
|
| 628 |
+
"train_loss": 0.5169329488148483,
|
| 629 |
+
"train_actor_loss": -0.37725697954495746,
|
| 630 |
+
"train_critic_loss": 0.0004284516236601541,
|
| 631 |
+
"train_mean_reward": -0.007467282284125806,
|
| 632 |
+
"train_entropy": 0.6883107800132188,
|
| 633 |
+
"train_ce_loss": 0.5822334093734239,
|
| 634 |
+
"train_dice_loss": 1.2057179911778522,
|
| 635 |
+
"train_dice": 0.6639504256274771,
|
| 636 |
+
"train_iou": 0.5347485209760421,
|
| 637 |
+
"grad_norm": 4.419512734963344,
|
| 638 |
+
"lr": 0.0017085712472806151,
|
| 639 |
+
"encoder_lr": 0.000172175144212722,
|
| 640 |
+
"alpha": 0.4482811391353607,
|
| 641 |
+
"validated_this_epoch": true,
|
| 642 |
+
"val_loss": 0.9281574823808941,
|
| 643 |
+
"val_dice": 0.6043220068368382,
|
| 644 |
+
"val_iou": 0.5035993817374291,
|
| 645 |
+
"val_reward": -0.0019951673301590632,
|
| 646 |
+
"val_entropy": 0.5922944617993903,
|
| 647 |
+
"val_actor_loss": -0.00034574250751786466,
|
| 648 |
+
"val_critic_loss": 0.00026005728101484817,
|
| 649 |
+
"val_ce_loss": 0.4996511406556825,
|
| 650 |
+
"val_dice_loss": 1.3570952186680802
|
| 651 |
+
},
|
| 652 |
+
{
|
| 653 |
+
"epoch": 26,
|
| 654 |
+
"train_loss": 0.46593361287019575,
|
| 655 |
+
"train_actor_loss": -0.38484745568189865,
|
| 656 |
+
"train_critic_loss": 0.0004074945305686785,
|
| 657 |
+
"train_mean_reward": -0.007047869691375989,
|
| 658 |
+
"train_entropy": 0.6813409823255661,
|
| 659 |
+
"train_ce_loss": 0.547392190171358,
|
| 660 |
+
"train_dice_loss": 1.153762433773432,
|
| 661 |
+
"train_dice": 0.7085499278835806,
|
| 662 |
+
"train_iou": 0.5944846363272518,
|
| 663 |
+
"grad_norm": 4.200689107179642,
|
| 664 |
+
"lr": 0.0016861243703990456,
|
| 665 |
+
"encoder_lr": 0.00017003197506322543,
|
| 666 |
+
"alpha": 0.4461023807525635,
|
| 667 |
+
"validated_this_epoch": true,
|
| 668 |
+
"val_loss": 0.9061416189440272,
|
| 669 |
+
"val_dice": 0.5700205893244414,
|
| 670 |
+
"val_iou": 0.473531130399279,
|
| 671 |
+
"val_reward": -0.005436509036481959,
|
| 672 |
+
"val_entropy": 0.4344078625994499,
|
| 673 |
+
"val_actor_loss": -0.004749863597827804,
|
| 674 |
+
"val_critic_loss": 0.0003226743877284474,
|
| 675 |
+
"val_ce_loss": 0.4512566375085199,
|
| 676 |
+
"val_dice_loss": 1.370203620556629
|
| 677 |
+
},
|
| 678 |
+
{
|
| 679 |
+
"epoch": 27,
|
| 680 |
+
"train_loss": 0.5100976546080067,
|
| 681 |
+
"train_actor_loss": -0.3361852871588408,
|
| 682 |
+
"train_critic_loss": 0.00036971593883022125,
|
| 683 |
+
"train_mean_reward": -0.006196722914841629,
|
| 684 |
+
"train_entropy": 0.5878263140718143,
|
| 685 |
+
"train_ce_loss": 0.5171220913792077,
|
| 686 |
+
"train_dice_loss": 1.1750740626683602,
|
| 687 |
+
"train_dice": 0.706330878185291,
|
| 688 |
+
"train_iou": 0.5856191845537693,
|
| 689 |
+
"grad_norm": 5.500753943736736,
|
| 690 |
+
"lr": 0.001663005305997034,
|
| 691 |
+
"encoder_lr": 0.00016782462720574693,
|
| 692 |
+
"alpha": 0.44440510869026184,
|
| 693 |
+
"validated_this_epoch": true,
|
| 694 |
+
"val_loss": 1.0485106643119997,
|
| 695 |
+
"val_dice": 0.5565947839876758,
|
| 696 |
+
"val_iou": 0.45399743613296756,
|
| 697 |
+
"val_reward": 0.00025906190002746314,
|
| 698 |
+
"val_entropy": 0.7321080794238082,
|
| 699 |
+
"val_actor_loss": 0.006875541531527886,
|
| 700 |
+
"val_critic_loss": 0.0002066252955505707,
|
| 701 |
+
"val_ce_loss": 0.6854702495602947,
|
| 702 |
+
"val_dice_loss": 1.3975933377790932
|
| 703 |
+
},
|
| 704 |
+
{
|
| 705 |
+
"epoch": 28,
|
| 706 |
+
"train_loss": 0.44709886134100646,
|
| 707 |
+
"train_actor_loss": -0.38591816534216594,
|
| 708 |
+
"train_critic_loss": 0.00031500449795902777,
|
| 709 |
+
"train_mean_reward": -0.006214412949189771,
|
| 710 |
+
"train_entropy": 0.6848649518230021,
|
| 711 |
+
"train_ce_loss": 0.5030846432066307,
|
| 712 |
+
"train_dice_loss": 1.1626343891406672,
|
| 713 |
+
"train_dice": 0.7188990308452305,
|
| 714 |
+
"train_iou": 0.6035267862843952,
|
| 715 |
+
"grad_norm": 3.9835289945969214,
|
| 716 |
+
"lr": 0.0016392368697999468,
|
| 717 |
+
"encoder_lr": 0.00016555527902612554,
|
| 718 |
+
"alpha": 0.442297101020813,
|
| 719 |
+
"validated_this_epoch": true,
|
| 720 |
+
"val_loss": 0.9664471397515048,
|
| 721 |
+
"val_dice": 0.5888065701586435,
|
| 722 |
+
"val_iou": 0.48703203358395425,
|
| 723 |
+
"val_reward": 0.0002814172800175253,
|
| 724 |
+
"val_entropy": 0.5483520617990782,
|
| 725 |
+
"val_actor_loss": 0.0042537805657231744,
|
| 726 |
+
"val_critic_loss": 0.00022288397566525315,
|
| 727 |
+
"val_ce_loss": 0.5515529506773961,
|
| 728 |
+
"val_dice_loss": 1.37261084775732
|
| 729 |
+
},
|
| 730 |
+
{
|
| 731 |
+
"epoch": 29,
|
| 732 |
+
"train_loss": 0.4466173933245815,
|
| 733 |
+
"train_actor_loss": -0.3542197574980748,
|
| 734 |
+
"train_critic_loss": 0.00037170881548193464,
|
| 735 |
+
"train_mean_reward": -0.006261748330964623,
|
| 736 |
+
"train_entropy": 0.6198908950273807,
|
| 737 |
+
"train_ce_loss": 0.4632998152325551,
|
| 738 |
+
"train_dice_loss": 1.1380027681589127,
|
| 739 |
+
"train_dice": 0.7368941615408688,
|
| 740 |
+
"train_iou": 0.622616533427408,
|
| 741 |
+
"grad_norm": 3.8685836791992188,
|
| 742 |
+
"lr": 0.0016148425183847121,
|
| 743 |
+
"encoder_lr": 0.0001632261700970328,
|
| 744 |
+
"alpha": 0.4404059648513794,
|
| 745 |
+
"validated_this_epoch": true,
|
| 746 |
+
"val_loss": 1.0666786127751975,
|
| 747 |
+
"val_dice": 0.5952133384990907,
|
| 748 |
+
"val_iou": 0.494443452368194,
|
| 749 |
+
"val_reward": 0.0009663470733723002,
|
| 750 |
+
"val_entropy": 0.8454163053421059,
|
| 751 |
+
"val_actor_loss": 0.0010927279022371753,
|
| 752 |
+
"val_critic_loss": 8.34189270909804e-05,
|
| 753 |
+
"val_ce_loss": 0.7456756763047341,
|
| 754 |
+
"val_dice_loss": 1.385412640041775
|
| 755 |
+
},
|
| 756 |
+
{
|
| 757 |
+
"epoch": 30,
|
| 758 |
+
"train_loss": 0.4539631579357844,
|
| 759 |
+
"train_actor_loss": -0.3787951487570237,
|
| 760 |
+
"train_critic_loss": 0.0003260584723111839,
|
| 761 |
+
"train_mean_reward": -0.005491825730878657,
|
| 762 |
+
"train_entropy": 0.6765988700282879,
|
| 763 |
+
"train_ce_loss": 0.5002652784355749,
|
| 764 |
+
"train_dice_loss": 1.1649252634782057,
|
| 765 |
+
"train_dice": 0.7302734057565092,
|
| 766 |
+
"train_iou": 0.6188321534051177,
|
| 767 |
+
"grad_norm": 3.2881311854490867,
|
| 768 |
+
"lr": 0.0015898463260310114,
|
| 769 |
+
"encoder_lr": 0.00016083959896778498,
|
| 770 |
+
"alpha": 0.4383477568626404,
|
| 771 |
+
"validated_this_epoch": true,
|
| 772 |
+
"val_loss": 1.1206772290170193,
|
| 773 |
+
"val_dice": 0.5800200084556488,
|
| 774 |
+
"val_iou": 0.469510661858179,
|
| 775 |
+
"val_reward": 0.0010399065742154273,
|
| 776 |
+
"val_entropy": 0.8345904202774318,
|
| 777 |
+
"val_actor_loss": -0.0010422443391313082,
|
| 778 |
+
"val_critic_loss": 5.9686143327836007e-05,
|
| 779 |
+
"val_ce_loss": 0.8513938105016043,
|
| 780 |
+
"val_dice_loss": 1.3919854079834137
|
| 781 |
+
},
|
| 782 |
+
{
|
| 783 |
+
"epoch": 31,
|
| 784 |
+
"train_loss": 0.4675976040844734,
|
| 785 |
+
"train_actor_loss": -0.37801922227327645,
|
| 786 |
+
"train_critic_loss": 0.000433903732930328,
|
| 787 |
+
"train_mean_reward": -0.00689567717722056,
|
| 788 |
+
"train_entropy": 0.6895143133707536,
|
| 789 |
+
"train_ce_loss": 0.5382738136686385,
|
| 790 |
+
"train_dice_loss": 1.1525259124927032,
|
| 791 |
+
"train_dice": 0.706723137174313,
|
| 792 |
+
"train_iou": 0.5863540052326724,
|
| 793 |
+
"grad_norm": 4.119364953958071,
|
| 794 |
+
"lr": 0.0015642729609628705,
|
| 795 |
+
"encoder_lr": 0.00015839792089595244,
|
| 796 |
+
"alpha": 0.4362565279006958,
|
| 797 |
+
"validated_this_epoch": true,
|
| 798 |
+
"val_loss": 1.0093085204928436,
|
| 799 |
+
"val_dice": 0.5911643570386672,
|
| 800 |
+
"val_iou": 0.4900931004926501,
|
| 801 |
+
"val_reward": 0.00012592415092659956,
|
| 802 |
+
"val_entropy": 0.7652086419890626,
|
| 803 |
+
"val_actor_loss": -0.003916048056756456,
|
| 804 |
+
"val_critic_loss": 0.00016883788497341245,
|
| 805 |
+
"val_ce_loss": 0.6539990172975443,
|
| 806 |
+
"val_dice_loss": 1.3722812494244239
|
| 807 |
+
},
|
| 808 |
+
{
|
| 809 |
+
"epoch": 32,
|
| 810 |
+
"train_loss": 0.4460183047474577,
|
| 811 |
+
"train_actor_loss": -0.3839868192489331,
|
| 812 |
+
"train_critic_loss": 0.0003986711010675675,
|
| 813 |
+
"train_mean_reward": -0.005993056009081043,
|
| 814 |
+
"train_entropy": 0.6859309197618411,
|
| 815 |
+
"train_ce_loss": 0.4901390333588307,
|
| 816 |
+
"train_dice_loss": 1.1694725289558752,
|
| 817 |
+
"train_dice": 0.7514301035918441,
|
| 818 |
+
"train_iou": 0.6306527544762573,
|
| 819 |
+
"grad_norm": 4.120462527641883,
|
| 820 |
+
"lr": 0.001538147661004102,
|
| 821 |
+
"encoder_lr": 0.0001559035455230047,
|
| 822 |
+
"alpha": 0.43399524688720703,
|
| 823 |
+
"validated_this_epoch": true,
|
| 824 |
+
"val_loss": 0.9967452751924143,
|
| 825 |
+
"val_dice": 0.5612252810630465,
|
| 826 |
+
"val_iou": 0.45955895214672016,
|
| 827 |
+
"val_reward": 0.0010363999307343753,
|
| 828 |
+
"val_entropy": 0.5783539643191329,
|
| 829 |
+
"val_actor_loss": 0.0029968540184199814,
|
| 830 |
+
"val_critic_loss": 0.0001806213484160988,
|
| 831 |
+
"val_ce_loss": 0.5906040751836217,
|
| 832 |
+
"val_dice_loss": 1.3967121111022103
|
| 833 |
+
},
|
| 834 |
+
{
|
| 835 |
+
"epoch": 33,
|
| 836 |
+
"train_loss": 0.47304636605370504,
|
| 837 |
+
"train_actor_loss": -0.3684007794811175,
|
| 838 |
+
"train_critic_loss": 0.00036344878809848946,
|
| 839 |
+
"train_mean_reward": -0.005017452687845947,
|
| 840 |
+
"train_entropy": 0.6647009528600253,
|
| 841 |
+
"train_ce_loss": 0.5105359799539049,
|
| 842 |
+
"train_dice_loss": 1.171994848129077,
|
| 843 |
+
"train_dice": 0.7243718744298973,
|
| 844 |
+
"train_iou": 0.6102336073840977,
|
| 845 |
+
"grad_norm": 3.7008761259225698,
|
| 846 |
+
"lr": 0.0015114962086716199,
|
| 847 |
+
"encoder_lr": 0.0001533589344962853,
|
| 848 |
+
"alpha": 0.43197646737098694,
|
| 849 |
+
"validated_this_epoch": true,
|
| 850 |
+
"val_loss": 1.0182334237844881,
|
| 851 |
+
"val_dice": 0.5721745875358348,
|
| 852 |
+
"val_iou": 0.4709242833902396,
|
| 853 |
+
"val_reward": -0.0018724860011122538,
|
| 854 |
+
"val_entropy": 0.5548271387815475,
|
| 855 |
+
"val_actor_loss": 0.018485130254188645,
|
| 856 |
+
"val_critic_loss": 0.0003514501905106091,
|
| 857 |
+
"val_ce_loss": 0.6048358016826136,
|
| 858 |
+
"val_dice_loss": 1.3943093012679706
|
| 859 |
+
},
|
| 860 |
+
{
|
| 861 |
+
"epoch": 34,
|
| 862 |
+
"train_loss": 0.4521111491709374,
|
| 863 |
+
"train_actor_loss": -0.3617316567553923,
|
| 864 |
+
"train_critic_loss": 0.0003755203321746264,
|
| 865 |
+
"train_mean_reward": -0.007514555140425308,
|
| 866 |
+
"train_entropy": 0.6513997677427071,
|
| 867 |
+
"train_ce_loss": 0.4987090956658507,
|
| 868 |
+
"train_dice_loss": 1.1286009848117828,
|
| 869 |
+
"train_dice": 0.727860694714621,
|
| 870 |
+
"train_iou": 0.6064977575750912,
|
| 871 |
+
"grad_norm": 4.22557303080192,
|
| 872 |
+
"lr": 0.0014843449057312071,
|
| 873 |
+
"encoder_lr": 0.00015076659903966297,
|
| 874 |
+
"alpha": 0.4299250841140747,
|
| 875 |
+
"validated_this_epoch": true,
|
| 876 |
+
"val_loss": 0.9931692066743518,
|
| 877 |
+
"val_dice": 0.5792249596905733,
|
| 878 |
+
"val_iou": 0.48225509012248896,
|
| 879 |
+
"val_reward": -0.00014069161283480238,
|
| 880 |
+
"val_entropy": 0.47865820909389345,
|
| 881 |
+
"val_actor_loss": 0.007447211037159456,
|
| 882 |
+
"val_critic_loss": 0.00025460364298796225,
|
| 883 |
+
"val_ce_loss": 0.5852140955852738,
|
| 884 |
+
"val_dice_loss": 1.3859752612282532
|
| 885 |
+
},
|
| 886 |
+
{
|
| 887 |
+
"epoch": 35,
|
| 888 |
+
"train_loss": 0.45826140655061376,
|
| 889 |
+
"train_actor_loss": -0.3447150642482134,
|
| 890 |
+
"train_critic_loss": 0.0003537201548682881,
|
| 891 |
+
"train_mean_reward": -0.005238942036227783,
|
| 892 |
+
"train_entropy": 0.6311525259262477,
|
| 893 |
+
"train_ce_loss": 0.46436595005723535,
|
| 894 |
+
"train_dice_loss": 1.1412332554658253,
|
| 895 |
+
"train_dice": 0.7195140563371573,
|
| 896 |
+
"train_iou": 0.6091260020347202,
|
| 897 |
+
"grad_norm": 4.166423818239799,
|
| 898 |
+
"lr": 0.0014567205472408495,
|
| 899 |
+
"encoder_lr": 0.00014812909747525697,
|
| 900 |
+
"alpha": 0.42814216017723083,
|
| 901 |
+
"validated_this_epoch": true,
|
| 902 |
+
"val_loss": 0.9743582549871821,
|
| 903 |
+
"val_dice": 0.5993855916212806,
|
| 904 |
+
"val_iou": 0.5002673815217863,
|
| 905 |
+
"val_reward": -0.0007636598347009493,
|
| 906 |
+
"val_entropy": 0.6188808529063908,
|
| 907 |
+
"val_actor_loss": 6.852843775178735e-05,
|
| 908 |
+
"val_critic_loss": 0.00018786842244453234,
|
| 909 |
+
"val_ce_loss": 0.5814709786814873,
|
| 910 |
+
"val_dice_loss": 1.3669205765531522
|
| 911 |
+
},
|
| 912 |
+
{
|
| 913 |
+
"epoch": 36,
|
| 914 |
+
"train_loss": 0.4744221428409219,
|
| 915 |
+
"train_actor_loss": -0.3569041489122006,
|
| 916 |
+
"train_critic_loss": 0.0003805966694133717,
|
| 917 |
+
"train_mean_reward": -0.005978664583940722,
|
| 918 |
+
"train_entropy": 0.6339053958654404,
|
| 919 |
+
"train_ce_loss": 0.4796966288943226,
|
| 920 |
+
"train_dice_loss": 1.1825753381619089,
|
| 921 |
+
"train_dice": 0.7361298670955733,
|
| 922 |
+
"train_iou": 0.6190829128026962,
|
| 923 |
+
"grad_norm": 4.535672506460776,
|
| 924 |
+
"lr": 0.0014286503951072478,
|
| 925 |
+
"encoder_lr": 0.00014544903269868192,
|
| 926 |
+
"alpha": 0.4261617362499237,
|
| 927 |
+
"validated_this_epoch": true,
|
| 928 |
+
"val_loss": 0.9983611002451542,
|
| 929 |
+
"val_dice": 0.5851380908034461,
|
| 930 |
+
"val_iou": 0.4883595255594135,
|
| 931 |
+
"val_reward": 0.0006086710533477022,
|
| 932 |
+
"val_entropy": 0.6467966934045156,
|
| 933 |
+
"val_actor_loss": 0.0056402476920718045,
|
| 934 |
+
"val_critic_loss": 0.00019061703851823975,
|
| 935 |
+
"val_ce_loss": 0.600145113851988,
|
| 936 |
+
"val_dice_loss": 1.3851059437400164
|
| 937 |
+
},
|
| 938 |
+
{
|
| 939 |
+
"epoch": 37,
|
| 940 |
+
"train_loss": 0.48171053941433245,
|
| 941 |
+
"train_actor_loss": -0.3602291063811534,
|
| 942 |
+
"train_critic_loss": 0.00032436607057789865,
|
| 943 |
+
"train_mean_reward": -0.005994853362211964,
|
| 944 |
+
"train_entropy": 0.6731508744832797,
|
| 945 |
+
"train_ce_loss": 0.5125513766319133,
|
| 946 |
+
"train_dice_loss": 1.1710035315690897,
|
| 947 |
+
"train_dice": 0.7603298276662827,
|
| 948 |
+
"train_iou": 0.6483773518531748,
|
| 949 |
+
"grad_norm": 3.6725516319274902,
|
| 950 |
+
"lr": 0.001400162151181607,
|
| 951 |
+
"encoder_lr": 0.00014272904961030417,
|
| 952 |
+
"alpha": 0.42437827587127686,
|
| 953 |
+
"validated_this_epoch": true,
|
| 954 |
+
"val_loss": 0.9917462615183357,
|
| 955 |
+
"val_dice": 0.6000647880971535,
|
| 956 |
+
"val_iou": 0.5006580267792813,
|
| 957 |
+
"val_reward": 0.0006996945760179049,
|
| 958 |
+
"val_entropy": 0.7014988471161235,
|
| 959 |
+
"val_actor_loss": -0.008117274501751356,
|
| 960 |
+
"val_critic_loss": 8.761224529220585e-05,
|
| 961 |
+
"val_ce_loss": 0.6291443188345492,
|
| 962 |
+
"val_dice_loss": 1.3704951122553661
|
| 963 |
+
},
|
| 964 |
+
{
|
| 965 |
+
"epoch": 38,
|
| 966 |
+
"train_loss": 0.43163695702186,
|
| 967 |
+
"train_actor_loss": -0.35515828325580323,
|
| 968 |
+
"train_critic_loss": 0.000355380279847267,
|
| 969 |
+
"train_mean_reward": -0.005601540754772335,
|
| 970 |
+
"train_entropy": 0.6232071056579933,
|
| 971 |
+
"train_ce_loss": 0.4307928277979581,
|
| 972 |
+
"train_dice_loss": 1.1424422546839101,
|
| 973 |
+
"train_dice": 0.758382682314813,
|
| 974 |
+
"train_iou": 0.6436645330630154,
|
| 975 |
+
"grad_norm": 3.1323538422584534,
|
| 976 |
+
"lr": 0.0013712839299212553,
|
| 977 |
+
"encoder_lr": 0.00013997183250504444,
|
| 978 |
+
"alpha": 0.42254960536956787,
|
| 979 |
+
"validated_this_epoch": true,
|
| 980 |
+
"val_loss": 0.9447416806819312,
|
| 981 |
+
"val_dice": 0.5823208492173363,
|
| 982 |
+
"val_iou": 0.48725479184249215,
|
| 983 |
+
"val_reward": -0.0022871233585362344,
|
| 984 |
+
"val_entropy": 0.5277494472084623,
|
| 985 |
+
"val_actor_loss": 0.0004482340535638156,
|
| 986 |
+
"val_critic_loss": 0.00019819250567275278,
|
| 987 |
+
"val_ce_loss": 0.5252852672993235,
|
| 988 |
+
"val_dice_loss": 1.3631034002761646
|
| 989 |
+
},
|
| 990 |
+
{
|
| 991 |
+
"epoch": 39,
|
| 992 |
+
"train_loss": 0.4358260678127408,
|
| 993 |
+
"train_actor_loss": -0.3412730030906506,
|
| 994 |
+
"train_critic_loss": 0.0002938033546972572,
|
| 995 |
+
"train_mean_reward": -0.006202491661660576,
|
| 996 |
+
"train_entropy": 0.6009003785558236,
|
| 997 |
+
"train_ce_loss": 0.43094005995692736,
|
| 998 |
+
"train_dice_loss": 1.1229642595236118,
|
| 999 |
+
"train_dice": 0.7757681389019697,
|
| 1000 |
+
"train_iou": 0.664139337805739,
|
| 1001 |
+
"grad_norm": 2.808111289372811,
|
| 1002 |
+
"lr": 0.0013420442306440656,
|
| 1003 |
+
"encoder_lr": 0.0001371801024233027,
|
| 1004 |
+
"alpha": 0.4207606017589569,
|
| 1005 |
+
"validated_this_epoch": true,
|
| 1006 |
+
"val_loss": 0.9765053108108767,
|
| 1007 |
+
"val_dice": 0.5848608790244656,
|
| 1008 |
+
"val_iou": 0.4849027165132682,
|
| 1009 |
+
"val_reward": -1.0291009278840946e-05,
|
| 1010 |
+
"val_entropy": 0.6063887398652357,
|
| 1011 |
+
"val_actor_loss": -0.007628762292572193,
|
| 1012 |
+
"val_critic_loss": 0.00010877853919461786,
|
| 1013 |
+
"val_ce_loss": 0.5925832803953778,
|
| 1014 |
+
"val_dice_loss": 1.375576065947311
|
| 1015 |
+
},
|
| 1016 |
+
{
|
| 1017 |
+
"epoch": 40,
|
| 1018 |
+
"train_loss": 0.43519719733068574,
|
| 1019 |
+
"train_actor_loss": -0.3528290899136128,
|
| 1020 |
+
"train_critic_loss": 0.00037154424875879136,
|
| 1021 |
+
"train_mean_reward": -0.005653555018048231,
|
| 1022 |
+
"train_entropy": 0.6349980429961131,
|
| 1023 |
+
"train_ce_loss": 0.46302966537097323,
|
| 1024 |
+
"train_dice_loss": 1.1126513500244188,
|
| 1025 |
+
"train_dice": 0.7517919708637646,
|
| 1026 |
+
"train_iou": 0.6391540400766874,
|
| 1027 |
+
"grad_norm": 3.3850729511334348,
|
| 1028 |
+
"lr": 0.0013124719094030732,
|
| 1029 |
+
"encoder_lr": 0.00013435661446562,
|
| 1030 |
+
"alpha": 0.41894713044166565,
|
| 1031 |
+
"validated_this_epoch": true,
|
| 1032 |
+
"val_loss": 0.9671123823235658,
|
| 1033 |
+
"val_dice": 0.6021372623824597,
|
| 1034 |
+
"val_iou": 0.508104033643492,
|
| 1035 |
+
"val_reward": 0.00014083829138153382,
|
| 1036 |
+
"val_entropy": 0.5778696726668965,
|
| 1037 |
+
"val_actor_loss": -0.002354354345688897,
|
| 1038 |
+
"val_critic_loss": 0.00015545140487313003,
|
| 1039 |
+
"val_ce_loss": 0.5710232030468608,
|
| 1040 |
+
"val_dice_loss": 1.3677547881097503
|
| 1041 |
+
},
|
| 1042 |
+
{
|
| 1043 |
+
"epoch": 41,
|
| 1044 |
+
"train_loss": 0.41027157796690095,
|
| 1045 |
+
"train_actor_loss": -0.36369194462895393,
|
| 1046 |
+
"train_critic_loss": 0.00027834089047619503,
|
| 1047 |
+
"train_mean_reward": -0.005506115334029369,
|
| 1048 |
+
"train_entropy": 0.6481131227352679,
|
| 1049 |
+
"train_ce_loss": 0.45835875147858113,
|
| 1050 |
+
"train_dice_loss": 1.0892899410846906,
|
| 1051 |
+
"train_dice": 0.764532820225675,
|
| 1052 |
+
"train_iou": 0.6541578480197008,
|
| 1053 |
+
"grad_norm": 3.012432254277743,
|
| 1054 |
+
"lr": 0.0012825961505090336,
|
| 1055 |
+
"encoder_lr": 0.00013150415507372682,
|
| 1056 |
+
"alpha": 0.4171038866043091,
|
| 1057 |
+
"validated_this_epoch": true,
|
| 1058 |
+
"val_loss": 0.9524300348676854,
|
| 1059 |
+
"val_dice": 0.6049610217219984,
|
| 1060 |
+
"val_iou": 0.5103582773977223,
|
| 1061 |
+
"val_reward": -0.0007796270384469702,
|
| 1062 |
+
"val_entropy": 0.45061351780337516,
|
| 1063 |
+
"val_actor_loss": 0.007350031580750107,
|
| 1064 |
+
"val_critic_loss": 0.00019605499331035647,
|
| 1065 |
+
"val_ce_loss": 0.5327391567301344,
|
| 1066 |
+
"val_dice_loss": 1.357224767858332
|
| 1067 |
+
},
|
| 1068 |
+
{
|
| 1069 |
+
"epoch": 42,
|
| 1070 |
+
"train_loss": 0.44845502665982795,
|
| 1071 |
+
"train_actor_loss": -0.3152446856674476,
|
| 1072 |
+
"train_critic_loss": 0.0003483306755720734,
|
| 1073 |
+
"train_mean_reward": -0.006158825305763643,
|
| 1074 |
+
"train_entropy": 0.5772623611757388,
|
| 1075 |
+
"train_ce_loss": 0.4265357935681748,
|
| 1076 |
+
"train_dice_loss": 1.1005152845994022,
|
| 1077 |
+
"train_dice": 0.7632530857534969,
|
| 1078 |
+
"train_iou": 0.6498429150156239,
|
| 1079 |
+
"grad_norm": 3.7916071781745324,
|
| 1080 |
+
"lr": 0.0012524464377290307,
|
| 1081 |
+
"encoder_lr": 0.00012862553928066122,
|
| 1082 |
+
"alpha": 0.4155401587486267,
|
| 1083 |
+
"validated_this_epoch": true,
|
| 1084 |
+
"val_loss": 0.9896216293881562,
|
| 1085 |
+
"val_dice": 0.6112750386108796,
|
| 1086 |
+
"val_iou": 0.5096097420620453,
|
| 1087 |
+
"val_reward": 0.0006584465011904612,
|
| 1088 |
+
"val_entropy": 0.6842589498770356,
|
| 1089 |
+
"val_actor_loss": -0.01049841569520232,
|
| 1090 |
+
"val_critic_loss": 6.73761613951582e-05,
|
| 1091 |
+
"val_ce_loss": 0.6367772800208191,
|
| 1092 |
+
"val_dice_loss": 1.3633954040329868
|
| 1093 |
+
},
|
| 1094 |
+
{
|
| 1095 |
+
"epoch": 43,
|
| 1096 |
+
"train_loss": 0.419221063144505,
|
| 1097 |
+
"train_actor_loss": -0.3586601407673114,
|
| 1098 |
+
"train_critic_loss": 0.0002976230444358244,
|
| 1099 |
+
"train_mean_reward": -0.005665563095625169,
|
| 1100 |
+
"train_entropy": 0.6517653010594538,
|
| 1101 |
+
"train_ce_loss": 0.4376596640437269,
|
| 1102 |
+
"train_dice_loss": 1.117805098875975,
|
| 1103 |
+
"train_dice": 0.7767070718429179,
|
| 1104 |
+
"train_iou": 0.6672198976522458,
|
| 1105 |
+
"grad_norm": 2.7639757830363054,
|
| 1106 |
+
"lr": 0.0012220525251895604,
|
| 1107 |
+
"encoder_lr": 0.0001257236079326716,
|
| 1108 |
+
"alpha": 0.4136728048324585,
|
| 1109 |
+
"validated_this_epoch": true,
|
| 1110 |
+
"val_loss": 0.9794564454632839,
|
| 1111 |
+
"val_dice": 0.5870582070809978,
|
| 1112 |
+
"val_iou": 0.4947241870721633,
|
| 1113 |
+
"val_reward": -0.0013483005312063868,
|
| 1114 |
+
"val_entropy": 0.6078988542460432,
|
| 1115 |
+
"val_actor_loss": 0.0006784353730964915,
|
| 1116 |
+
"val_critic_loss": 0.00016044319634583913,
|
| 1117 |
+
"val_ce_loss": 0.5904078269424387,
|
| 1118 |
+
"val_dice_loss": 1.3669877278082296
|
| 1119 |
+
},
|
| 1120 |
+
{
|
| 1121 |
+
"epoch": 44,
|
| 1122 |
+
"train_loss": 0.40167351364373016,
|
| 1123 |
+
"train_actor_loss": -0.3597433136250728,
|
| 1124 |
+
"train_critic_loss": 0.00029776234481314296,
|
| 1125 |
+
"train_mean_reward": -0.006018124264161229,
|
| 1126 |
+
"train_entropy": 0.6609401433513716,
|
| 1127 |
+
"train_ce_loss": 0.43525927973529094,
|
| 1128 |
+
"train_dice_loss": 1.0872766004923062,
|
| 1129 |
+
"train_dice": 0.7956424787932751,
|
| 1130 |
+
"train_iou": 0.683820499947258,
|
| 1131 |
+
"grad_norm": 2.744549579345263,
|
| 1132 |
+
"lr": 0.0011914444080127966,
|
| 1133 |
+
"encoder_lr": 0.0001228012248856439,
|
| 1134 |
+
"alpha": 0.4116930365562439,
|
| 1135 |
+
"validated_this_epoch": true,
|
| 1136 |
+
"val_loss": 0.9630901383617047,
|
| 1137 |
+
"val_dice": 0.5959351835962927,
|
| 1138 |
+
"val_iou": 0.5012341325075262,
|
| 1139 |
+
"val_reward": -0.0003197197049780928,
|
| 1140 |
+
"val_entropy": 0.5540832329278039,
|
| 1141 |
+
"val_actor_loss": -0.008593487378381512,
|
| 1142 |
+
"val_critic_loss": 0.00014114359870254337,
|
| 1143 |
+
"val_ce_loss": 0.5749188384295187,
|
| 1144 |
+
"val_dice_loss": 1.3683072370712204
|
| 1145 |
+
},
|
| 1146 |
+
{
|
| 1147 |
+
"epoch": 45,
|
| 1148 |
+
"train_loss": 0.4038345699843306,
|
| 1149 |
+
"train_actor_loss": -0.34029908459155983,
|
| 1150 |
+
"train_critic_loss": 0.0002599890517441777,
|
| 1151 |
+
"train_mean_reward": -0.004970367209236242,
|
| 1152 |
+
"train_entropy": 0.6183586901961229,
|
| 1153 |
+
"train_ce_loss": 0.39680380688216055,
|
| 1154 |
+
"train_dice_loss": 1.0912035031196399,
|
| 1155 |
+
"train_dice": 0.7850258611835178,
|
| 1156 |
+
"train_iou": 0.683695480808039,
|
| 1157 |
+
"grad_norm": 2.30870613685021,
|
| 1158 |
+
"lr": 0.0011606522927150302,
|
| 1159 |
+
"encoder_lr": 0.00011986127417882199,
|
| 1160 |
+
"alpha": 0.40989047288894653,
|
| 1161 |
+
"validated_this_epoch": true,
|
| 1162 |
+
"val_loss": 0.9892087177458134,
|
| 1163 |
+
"val_dice": 0.6155704811744365,
|
| 1164 |
+
"val_iou": 0.5134376974588949,
|
| 1165 |
+
"val_reward": 1.616937306332127e-05,
|
| 1166 |
+
"val_entropy": 0.660823149211479,
|
| 1167 |
+
"val_actor_loss": -0.001626908473995034,
|
| 1168 |
+
"val_critic_loss": 0.00013798885369525503,
|
| 1169 |
+
"val_ce_loss": 0.6325338868555997,
|
| 1170 |
+
"val_dice_loss": 1.34899935156408
|
| 1171 |
+
},
|
| 1172 |
+
{
|
| 1173 |
+
"epoch": 46,
|
| 1174 |
+
"train_loss": 0.3916521086715735,
|
| 1175 |
+
"train_actor_loss": -0.3534373569373901,
|
| 1176 |
+
"train_critic_loss": 0.00027800446486463235,
|
| 1177 |
+
"train_mean_reward": -0.005609048159945255,
|
| 1178 |
+
"train_entropy": 0.6490224504317992,
|
| 1179 |
+
"train_ce_loss": 0.4172814865715993,
|
| 1180 |
+
"train_dice_loss": 1.0726194225060635,
|
| 1181 |
+
"train_dice": 0.7988420015284184,
|
| 1182 |
+
"train_iou": 0.6908461355523092,
|
| 1183 |
+
"grad_norm": 2.6501973512081,
|
| 1184 |
+
"lr": 0.0011297065673964834,
|
| 1185 |
+
"encoder_lr": 0.00011690665718860897,
|
| 1186 |
+
"alpha": 0.40802305936813354,
|
| 1187 |
+
"validated_this_epoch": true,
|
| 1188 |
+
"val_loss": 1.0283367871730165,
|
| 1189 |
+
"val_dice": 0.5932172502208353,
|
| 1190 |
+
"val_iou": 0.49888684023852187,
|
| 1191 |
+
"val_reward": 0.0006311665538305532,
|
| 1192 |
+
"val_entropy": 0.6628148733365415,
|
| 1193 |
+
"val_actor_loss": -0.007499737203511588,
|
| 1194 |
+
"val_critic_loss": 0.00010440275274610176,
|
| 1195 |
+
"val_ce_loss": 0.6956327190490985,
|
| 1196 |
+
"val_dice_loss": 1.3759358886516453
|
| 1197 |
+
},
|
| 1198 |
+
{
|
| 1199 |
+
"epoch": 47,
|
| 1200 |
+
"train_loss": 0.3928793285471889,
|
| 1201 |
+
"train_actor_loss": -0.33849631402737057,
|
| 1202 |
+
"train_critic_loss": 0.0002884241312830664,
|
| 1203 |
+
"train_mean_reward": -0.005427148230433559,
|
| 1204 |
+
"train_entropy": 0.6137915633809872,
|
| 1205 |
+
"train_ce_loss": 0.4009292839238277,
|
| 1206 |
+
"train_dice_loss": 1.0615335676150446,
|
| 1207 |
+
"train_dice": 0.7936766851766437,
|
| 1208 |
+
"train_iou": 0.6853872589109575,
|
| 1209 |
+
"grad_norm": 2.774618306985268,
|
| 1210 |
+
"lr": 0.001098637771751922,
|
| 1211 |
+
"encoder_lr": 0.0001139402897652589,
|
| 1212 |
+
"alpha": 0.406188428401947,
|
| 1213 |
+
"validated_this_epoch": true,
|
| 1214 |
+
"val_loss": 1.0211401376591036,
|
| 1215 |
+
"val_dice": 0.5918309723496331,
|
| 1216 |
+
"val_iou": 0.4934488360952767,
|
| 1217 |
+
"val_reward": 0.0005033920336089326,
|
| 1218 |
+
"val_entropy": 0.6002378346342029,
|
| 1219 |
+
"val_actor_loss": -0.006406879099556293,
|
| 1220 |
+
"val_critic_loss": 9.960826359056595e-05,
|
| 1221 |
+
"val_ce_loss": 0.6762496966567605,
|
| 1222 |
+
"val_dice_loss": 1.3787446973299742
|
| 1223 |
+
},
|
| 1224 |
+
{
|
| 1225 |
+
"epoch": 48,
|
| 1226 |
+
"train_loss": 0.40594040871096343,
|
| 1227 |
+
"train_actor_loss": -0.35219283354205966,
|
| 1228 |
+
"train_critic_loss": 0.00027113670815500274,
|
| 1229 |
+
"train_mean_reward": -0.005443790662236727,
|
| 1230 |
+
"train_entropy": 0.6499429960281421,
|
| 1231 |
+
"train_ce_loss": 0.42092169580670696,
|
| 1232 |
+
"train_dice_loss": 1.095073639200284,
|
| 1233 |
+
"train_dice": 0.78984078272433,
|
| 1234 |
+
"train_iou": 0.689760294329722,
|
| 1235 |
+
"grad_norm": 2.7991005663688364,
|
| 1236 |
+
"lr": 0.0010674765669316673,
|
| 1237 |
+
"encoder_lr": 0.00011096509935528483,
|
| 1238 |
+
"alpha": 0.4044100344181061,
|
| 1239 |
+
"validated_this_epoch": true,
|
| 1240 |
+
"val_loss": 0.9950734928222091,
|
| 1241 |
+
"val_dice": 0.6064484945134673,
|
| 1242 |
+
"val_iou": 0.5103638085602765,
|
| 1243 |
+
"val_reward": 0.00011931069545398441,
|
| 1244 |
+
"val_entropy": 0.6246733930375842,
|
| 1245 |
+
"val_actor_loss": -0.010457301404323627,
|
| 1246 |
+
"val_critic_loss": 9.503711840586743e-05,
|
| 1247 |
+
"val_ce_loss": 0.6561376606067173,
|
| 1248 |
+
"val_dice_loss": 1.3548288616267117
|
| 1249 |
+
},
|
| 1250 |
+
{
|
| 1251 |
+
"epoch": 49,
|
| 1252 |
+
"train_loss": 0.40061239446871555,
|
| 1253 |
+
"train_actor_loss": -0.32279307614916414,
|
| 1254 |
+
"train_critic_loss": 0.00025006945874431555,
|
| 1255 |
+
"train_mean_reward": -0.004406736936429157,
|
| 1256 |
+
"train_entropy": 0.5932777506800799,
|
| 1257 |
+
"train_ce_loss": 0.3798183309666525,
|
| 1258 |
+
"train_dice_loss": 1.0667425264150667,
|
| 1259 |
+
"train_dice": 0.8007408180365375,
|
| 1260 |
+
"train_iou": 0.6962133242803461,
|
| 1261 |
+
"grad_norm": 2.6737359853891225,
|
| 1262 |
+
"lr": 0.001036253705282738,
|
| 1263 |
+
"encoder_lr": 0.00010798402211242223,
|
| 1264 |
+
"alpha": 0.4027402400970459,
|
| 1265 |
+
"validated_this_epoch": true,
|
| 1266 |
+
"val_loss": 0.9955222581682558,
|
| 1267 |
+
"val_dice": 0.568637289993268,
|
| 1268 |
+
"val_iou": 0.4766859123750698,
|
| 1269 |
+
"val_reward": -0.001221412955426859,
|
| 1270 |
+
"val_entropy": 0.5546486856359425,
|
| 1271 |
+
"val_actor_loss": -0.003995603851202611,
|
| 1272 |
+
"val_critic_loss": 0.00011792266892058765,
|
| 1273 |
+
"val_ce_loss": 0.620458733074066,
|
| 1274 |
+
"val_dice_loss": 1.378459032436814
|
| 1275 |
+
},
|
| 1276 |
+
{
|
| 1277 |
+
"epoch": 50,
|
| 1278 |
+
"train_loss": 0.40785612624425155,
|
| 1279 |
+
"train_actor_loss": -0.3117211060837293,
|
| 1280 |
+
"train_critic_loss": 0.0002874374175893862,
|
| 1281 |
+
"train_mean_reward": -0.005808486011609114,
|
| 1282 |
+
"train_entropy": 0.578652333181638,
|
| 1283 |
+
"train_ce_loss": 0.38614423712715507,
|
| 1284 |
+
"train_dice_loss": 1.0527227765474563,
|
| 1285 |
+
"train_dice": 0.8228466028767619,
|
| 1286 |
+
"train_iou": 0.7193083793048144,
|
| 1287 |
+
"grad_norm": 3.9556965988415937,
|
| 1288 |
+
"lr": 0.0010050000000000003,
|
| 1289 |
+
"encoder_lr": 0.00010500000000000005,
|
| 1290 |
+
"alpha": 0.4011539816856384,
|
| 1291 |
+
"validated_this_epoch": true,
|
| 1292 |
+
"val_loss": 0.973936641052591,
|
| 1293 |
+
"val_dice": 0.598721435803563,
|
| 1294 |
+
"val_iou": 0.5054958597018323,
|
| 1295 |
+
"val_reward": -0.00015732800323284464,
|
| 1296 |
+
"val_entropy": 0.5884944338991184,
|
| 1297 |
+
"val_actor_loss": -0.007508175420537214,
|
| 1298 |
+
"val_critic_loss": 9.667180753662838e-05,
|
| 1299 |
+
"val_ce_loss": 0.6043368262812645,
|
| 1300 |
+
"val_dice_loss": 1.3584561116165583
|
| 1301 |
+
},
|
| 1302 |
+
{
|
| 1303 |
+
"epoch": 51,
|
| 1304 |
+
"train_loss": 0.3981555872238599,
|
| 1305 |
+
"train_actor_loss": -0.3339083688572432,
|
| 1306 |
+
"train_critic_loss": 0.00022680302117805695,
|
| 1307 |
+
"train_mean_reward": -0.004405126742242898,
|
| 1308 |
+
"train_entropy": 0.5950557226553941,
|
| 1309 |
+
"train_ce_loss": 0.38975393217510707,
|
| 1310 |
+
"train_dice_loss": 1.0741471636753817,
|
| 1311 |
+
"train_dice": 0.816609263420105,
|
| 1312 |
+
"train_iou": 0.7086980784056234,
|
| 1313 |
+
"grad_norm": 2.765407193165559,
|
| 1314 |
+
"lr": 0.0009737462947172626,
|
| 1315 |
+
"encoder_lr": 0.00010201597788757786,
|
| 1316 |
+
"alpha": 0.3995637595653534,
|
| 1317 |
+
"validated_this_epoch": true,
|
| 1318 |
+
"val_loss": 1.004517033286957,
|
| 1319 |
+
"val_dice": 0.5987905808439995,
|
| 1320 |
+
"val_iou": 0.5050340565926552,
|
| 1321 |
+
"val_reward": 0.00029258621279583126,
|
| 1322 |
+
"val_entropy": 0.5676330740403648,
|
| 1323 |
+
"val_actor_loss": -0.004747792157683183,
|
| 1324 |
+
"val_critic_loss": 8.265455923692295e-05,
|
| 1325 |
+
"val_ce_loss": 0.6607117628904454,
|
| 1326 |
+
"val_dice_loss": 1.3577352018669397
|
| 1327 |
+
},
|
| 1328 |
+
{
|
| 1329 |
+
"epoch": 52,
|
| 1330 |
+
"train_loss": 0.3773176986724138,
|
| 1331 |
+
"train_actor_loss": -0.32101157832986266,
|
| 1332 |
+
"train_critic_loss": 0.00022314436412274608,
|
| 1333 |
+
"train_mean_reward": -0.0039469530509869885,
|
| 1334 |
+
"train_entropy": 0.5824765687187513,
|
| 1335 |
+
"train_ce_loss": 0.36082526929198927,
|
| 1336 |
+
"train_dice_loss": 1.035610128289614,
|
| 1337 |
+
"train_dice": 0.8185752142458935,
|
| 1338 |
+
"train_iou": 0.7160072554725424,
|
| 1339 |
+
"grad_norm": 2.380001269854032,
|
| 1340 |
+
"lr": 0.0009425234330683334,
|
| 1341 |
+
"encoder_lr": 9.903490064471527e-05,
|
| 1342 |
+
"alpha": 0.39802736043930054,
|
| 1343 |
+
"validated_this_epoch": true,
|
| 1344 |
+
"val_loss": 0.9871147614057091,
|
| 1345 |
+
"val_dice": 0.5753095036146717,
|
| 1346 |
+
"val_iou": 0.4823572343419361,
|
| 1347 |
+
"val_reward": -0.0008637942364565778,
|
| 1348 |
+
"val_entropy": 0.47410729241491567,
|
| 1349 |
+
"val_actor_loss": 0.0003904263180859301,
|
| 1350 |
+
"val_critic_loss": 0.0001470433883540211,
|
| 1351 |
+
"val_ce_loss": 0.6220633809106668,
|
| 1352 |
+
"val_dice_loss": 1.351238220629066
|
| 1353 |
+
},
|
| 1354 |
+
{
|
| 1355 |
+
"epoch": 53,
|
| 1356 |
+
"train_loss": 0.40005503537563175,
|
| 1357 |
+
"train_actor_loss": -0.2850232432859066,
|
| 1358 |
+
"train_critic_loss": 0.00024363547527173068,
|
| 1359 |
+
"train_mean_reward": -0.003103199936235913,
|
| 1360 |
+
"train_entropy": 0.5461832661277208,
|
| 1361 |
+
"train_ce_loss": 0.3888430042455021,
|
| 1362 |
+
"train_dice_loss": 0.9810699014327466,
|
| 1363 |
+
"train_dice": 0.8008509681582043,
|
| 1364 |
+
"train_iou": 0.6992091197419175,
|
| 1365 |
+
"grad_norm": 3.2366771766772637,
|
| 1366 |
+
"lr": 0.0009113622282480784,
|
| 1367 |
+
"encoder_lr": 9.605971023474117e-05,
|
| 1368 |
+
"alpha": 0.3966984748840332,
|
| 1369 |
+
"validated_this_epoch": true,
|
| 1370 |
+
"val_loss": 0.9526657495563003,
|
| 1371 |
+
"val_dice": 0.5702537105202665,
|
| 1372 |
+
"val_iou": 0.478479631345317,
|
| 1373 |
+
"val_reward": 3.923963532827776e-05,
|
| 1374 |
+
"val_entropy": 0.5142723713258299,
|
| 1375 |
+
"val_actor_loss": -0.017187037593108894,
|
| 1376 |
+
"val_critic_loss": 0.00013542534167635209,
|
| 1377 |
+
"val_ce_loss": 0.658599174449766,
|
| 1378 |
+
"val_dice_loss": 1.280970943696571
|
| 1379 |
+
},
|
| 1380 |
+
{
|
| 1381 |
+
"epoch": 54,
|
| 1382 |
+
"train_loss": 0.3790936845426376,
|
| 1383 |
+
"train_actor_loss": -0.2657797268758981,
|
| 1384 |
+
"train_critic_loss": 0.0001957244224356075,
|
| 1385 |
+
"train_mean_reward": -0.002352446840562222,
|
| 1386 |
+
"train_entropy": 0.536931245563886,
|
| 1387 |
+
"train_ce_loss": 0.40124929451550806,
|
| 1388 |
+
"train_dice_loss": 0.8883017817368875,
|
| 1389 |
+
"train_dice": 0.8183611432711283,
|
| 1390 |
+
"train_iou": 0.713432524134131,
|
| 1391 |
+
"grad_norm": 4.772695988416672,
|
| 1392 |
+
"lr": 0.0008802934326035174,
|
| 1393 |
+
"encoder_lr": 9.309334281139113e-05,
|
| 1394 |
+
"alpha": 0.39535361528396606,
|
| 1395 |
+
"validated_this_epoch": true,
|
| 1396 |
+
"val_loss": 0.9569823158566247,
|
| 1397 |
+
"val_dice": 0.5979155156399514,
|
| 1398 |
+
"val_iou": 0.49529731666918886,
|
| 1399 |
+
"val_reward": 0.00026779470645675303,
|
| 1400 |
+
"val_entropy": 0.476477993257118,
|
| 1401 |
+
"val_actor_loss": -0.0190270969978148,
|
| 1402 |
+
"val_critic_loss": 9.659107486140325e-05,
|
| 1403 |
+
"val_ce_loss": 0.761239265684377,
|
| 1404 |
+
"val_dice_loss": 1.1906829349922412
|
| 1405 |
+
},
|
| 1406 |
+
{
|
| 1407 |
+
"epoch": 55,
|
| 1408 |
+
"train_loss": 0.3310044465873104,
|
| 1409 |
+
"train_actor_loss": -0.220159301868616,
|
| 1410 |
+
"train_critic_loss": 0.00014344809879054455,
|
| 1411 |
+
"train_mean_reward": -0.0005055578161283428,
|
| 1412 |
+
"train_entropy": 0.48876119529207546,
|
| 1413 |
+
"train_ce_loss": 0.4876343520382085,
|
| 1414 |
+
"train_dice_loss": 0.6145496853651145,
|
| 1415 |
+
"train_dice": 0.8250684901779773,
|
| 1416 |
+
"train_iou": 0.7268533464155945,
|
| 1417 |
+
"grad_norm": 5.684922461326305,
|
| 1418 |
+
"lr": 0.0008493477072849704,
|
| 1419 |
+
"encoder_lr": 9.013872582117809e-05,
|
| 1420 |
+
"alpha": 0.39424291253089905,
|
| 1421 |
+
"validated_this_epoch": true,
|
| 1422 |
+
"val_loss": 0.9095672179634372,
|
| 1423 |
+
"val_dice": 0.6010171087224292,
|
| 1424 |
+
"val_iou": 0.5011745247450986,
|
| 1425 |
+
"val_reward": 2.4689518940332113e-05,
|
| 1426 |
+
"val_entropy": 0.3994148960619262,
|
| 1427 |
+
"val_actor_loss": -0.013648396929387342,
|
| 1428 |
+
"val_critic_loss": 9.32219577637193e-05,
|
| 1429 |
+
"val_ce_loss": 0.8926971502639729,
|
| 1430 |
+
"val_dice_loss": 0.9536408291320608
|
| 1431 |
+
},
|
| 1432 |
+
{
|
| 1433 |
+
"epoch": 56,
|
| 1434 |
+
"train_loss": 0.360224764420239,
|
| 1435 |
+
"train_actor_loss": -0.20365880186168048,
|
| 1436 |
+
"train_critic_loss": 0.00014298350909787888,
|
| 1437 |
+
"train_mean_reward": -0.0001531212130127339,
|
| 1438 |
+
"train_entropy": 0.4768238806953796,
|
| 1439 |
+
"train_ce_loss": 0.5749119472904847,
|
| 1440 |
+
"train_dice_loss": 0.5527121867889013,
|
| 1441 |
+
"train_dice": 0.798070250956911,
|
| 1442 |
+
"train_iou": 0.6992428488677547,
|
| 1443 |
+
"grad_norm": 5.852275289022005,
|
| 1444 |
+
"lr": 0.0008185555919872042,
|
| 1445 |
+
"encoder_lr": 8.719877511435618e-05,
|
| 1446 |
+
"alpha": 0.39334413409233093,
|
| 1447 |
+
"validated_this_epoch": true,
|
| 1448 |
+
"val_loss": 0.9630573302168738,
|
| 1449 |
+
"val_dice": 0.6021569702503208,
|
| 1450 |
+
"val_iou": 0.5047192880469454,
|
| 1451 |
+
"val_reward": 0.00015348681374677993,
|
| 1452 |
+
"val_entropy": 0.5939082545463484,
|
| 1453 |
+
"val_actor_loss": -0.01032918768069434,
|
| 1454 |
+
"val_critic_loss": 4.252378887526167e-05,
|
| 1455 |
+
"val_ce_loss": 0.9349294388008239,
|
| 1456 |
+
"val_dice_loss": 1.0118010360183138
|
| 1457 |
+
},
|
| 1458 |
+
{
|
| 1459 |
+
"epoch": 57,
|
| 1460 |
+
"train_loss": 0.3155211993591645,
|
| 1461 |
+
"train_actor_loss": -0.2050319436268929,
|
| 1462 |
+
"train_critic_loss": 0.00013201636097502883,
|
| 1463 |
+
"train_mean_reward": 2.856413511136242e-05,
|
| 1464 |
+
"train_entropy": 0.4690752628331001,
|
| 1465 |
+
"train_ce_loss": 0.5186718319399425,
|
| 1466 |
+
"train_dice_loss": 0.5223024246784356,
|
| 1467 |
+
"train_dice": 0.7912606157122494,
|
| 1468 |
+
"train_iou": 0.6955179561785243,
|
| 1469 |
+
"grad_norm": 5.901923775672913,
|
| 1470 |
+
"lr": 0.0007879474748104406,
|
| 1471 |
+
"encoder_lr": 8.42763920673285e-05,
|
| 1472 |
+
"alpha": 0.39242708683013916,
|
| 1473 |
+
"validated_this_epoch": true,
|
| 1474 |
+
"val_loss": 0.9190666369077835,
|
| 1475 |
+
"val_dice": 0.5544260371263924,
|
| 1476 |
+
"val_iou": 0.4663340402232193,
|
| 1477 |
+
"val_reward": -0.0001875509490425677,
|
| 1478 |
+
"val_entropy": 0.432661645340197,
|
| 1479 |
+
"val_actor_loss": -0.022325715795629704,
|
| 1480 |
+
"val_critic_loss": 6.445403005272964e-05,
|
| 1481 |
+
"val_ce_loss": 0.7886494017685904,
|
| 1482 |
+
"val_dice_loss": 1.0940708247098057
|
| 1483 |
+
},
|
| 1484 |
+
{
|
| 1485 |
+
"epoch": 58,
|
| 1486 |
+
"train_loss": 0.2988742208556057,
|
| 1487 |
+
"train_actor_loss": -0.20120827648311088,
|
| 1488 |
+
"train_critic_loss": 0.0001257222616004583,
|
| 1489 |
+
"train_mean_reward": 0.00026400412617733353,
|
| 1490 |
+
"train_entropy": 0.4672800410443392,
|
| 1491 |
+
"train_ce_loss": 0.518354500294663,
|
| 1492 |
+
"train_dice_loss": 0.4816847649904397,
|
| 1493 |
+
"train_dice": 0.8307938673867605,
|
| 1494 |
+
"train_iou": 0.7330094355529135,
|
| 1495 |
+
"grad_norm": 6.328011012994326,
|
| 1496 |
+
"lr": 0.0007575535622709696,
|
| 1497 |
+
"encoder_lr": 8.13744607193388e-05,
|
| 1498 |
+
"alpha": 0.39149215817451477,
|
| 1499 |
+
"validated_this_epoch": true,
|
| 1500 |
+
"val_loss": 0.8397261832883074,
|
| 1501 |
+
"val_dice": 0.61054790651008,
|
| 1502 |
+
"val_iou": 0.5126376530679725,
|
| 1503 |
+
"val_reward": 0.0006512908411551884,
|
| 1504 |
+
"val_entropy": 0.3741108350681536,
|
| 1505 |
+
"val_actor_loss": -0.019535207602364768,
|
| 1506 |
+
"val_critic_loss": 7.24520628038172e-05,
|
| 1507 |
+
"val_ce_loss": 0.7501888982340166,
|
| 1508 |
+
"val_dice_loss": 0.968261407782333
|
| 1509 |
+
},
|
| 1510 |
+
{
|
| 1511 |
+
"epoch": 59,
|
| 1512 |
+
"train_loss": 0.31704950234863927,
|
| 1513 |
+
"train_actor_loss": -0.19207532192843077,
|
| 1514 |
+
"train_critic_loss": 0.00011681368721091974,
|
| 1515 |
+
"train_mean_reward": 0.00043603633620338,
|
| 1516 |
+
"train_entropy": 0.4563135536244282,
|
| 1517 |
+
"train_ce_loss": 0.5012142205706391,
|
| 1518 |
+
"train_dice_loss": 0.5169186057188573,
|
| 1519 |
+
"train_dice": 0.8207802192622765,
|
| 1520 |
+
"train_iou": 0.7224625371035632,
|
| 1521 |
+
"grad_norm": 6.640216405575092,
|
| 1522 |
+
"lr": 0.0007274038494909672,
|
| 1523 |
+
"encoder_lr": 7.849584492627325e-05,
|
| 1524 |
+
"alpha": 0.3907119929790497,
|
| 1525 |
+
"validated_this_epoch": true,
|
| 1526 |
+
"val_loss": 0.9806631163016639,
|
| 1527 |
+
"val_dice": 0.6035000998003927,
|
| 1528 |
+
"val_iou": 0.5044916446156028,
|
| 1529 |
+
"val_reward": 3.7590024007702707e-06,
|
| 1530 |
+
"val_entropy": 0.5903205377886993,
|
| 1531 |
+
"val_actor_loss": -0.00833842629094072,
|
| 1532 |
+
"val_critic_loss": 8.255485152454236e-05,
|
| 1533 |
+
"val_ce_loss": 1.0114480435302613,
|
| 1534 |
+
"val_dice_loss": 0.9664724511329574
|
| 1535 |
+
},
|
| 1536 |
+
{
|
| 1537 |
+
"epoch": 60,
|
| 1538 |
+
"train_loss": 0.3546457549763055,
|
| 1539 |
+
"train_actor_loss": -0.18791709076135588,
|
| 1540 |
+
"train_critic_loss": 0.00013003444632303377,
|
| 1541 |
+
"train_mean_reward": 0.00033643358027546084,
|
| 1542 |
+
"train_entropy": 0.463878160676895,
|
| 1543 |
+
"train_ce_loss": 0.5664026668080344,
|
| 1544 |
+
"train_dice_loss": 0.518592975078485,
|
| 1545 |
+
"train_dice": 0.7855852504863459,
|
| 1546 |
+
"train_iou": 0.6836964314913049,
|
| 1547 |
+
"grad_norm": 10.49738222360611,
|
| 1548 |
+
"lr": 0.0006975280905969278,
|
| 1549 |
+
"encoder_lr": 7.564338553438004e-05,
|
| 1550 |
+
"alpha": 0.38984256982803345,
|
| 1551 |
+
"validated_this_epoch": true,
|
| 1552 |
+
"val_loss": 0.8981024408007435,
|
| 1553 |
+
"val_dice": 0.5920096649828457,
|
| 1554 |
+
"val_iou": 0.4987811840036186,
|
| 1555 |
+
"val_reward": -0.000249157709942959,
|
| 1556 |
+
"val_entropy": 0.37998180377362956,
|
| 1557 |
+
"val_actor_loss": -0.02055685594031173,
|
| 1558 |
+
"val_critic_loss": 8.684048345292632e-05,
|
| 1559 |
+
"val_ce_loss": 0.8068438840540822,
|
| 1560 |
+
"val_dice_loss": 1.0303878338650019
|
| 1561 |
+
},
|
| 1562 |
+
{
|
| 1563 |
+
"epoch": 61,
|
| 1564 |
+
"train_loss": 0.35426985424083585,
|
| 1565 |
+
"train_actor_loss": -0.1813033310553202,
|
| 1566 |
+
"train_critic_loss": 0.0001353239086501372,
|
| 1567 |
+
"train_mean_reward": 0.00014979700544688222,
|
| 1568 |
+
"train_entropy": 0.4435393235240227,
|
| 1569 |
+
"train_ce_loss": 0.5475024840770624,
|
| 1570 |
+
"train_dice_loss": 0.5235085556140313,
|
| 1571 |
+
"train_dice": 0.824225618953216,
|
| 1572 |
+
"train_iou": 0.7225731860627167,
|
| 1573 |
+
"grad_norm": 5.893693547982436,
|
| 1574 |
+
"lr": 0.0006679557693559353,
|
| 1575 |
+
"encoder_lr": 7.281989757669733e-05,
|
| 1576 |
+
"alpha": 0.38900113105773926,
|
| 1577 |
+
"validated_this_epoch": true,
|
| 1578 |
+
"val_loss": 0.8689421100654837,
|
| 1579 |
+
"val_dice": 0.5991351567883311,
|
| 1580 |
+
"val_iou": 0.5086918193805082,
|
| 1581 |
+
"val_reward": 0.00022467425084198676,
|
| 1582 |
+
"val_entropy": 0.43538729951839256,
|
| 1583 |
+
"val_actor_loss": -0.010806561026115422,
|
| 1584 |
+
"val_critic_loss": 6.804124589994312e-05,
|
| 1585 |
+
"val_ce_loss": 0.7371449311536671,
|
| 1586 |
+
"val_dice_loss": 1.0222843418819738
|
| 1587 |
+
},
|
| 1588 |
+
{
|
| 1589 |
+
"epoch": 62,
|
| 1590 |
+
"train_loss": 0.29789374018302905,
|
| 1591 |
+
"train_actor_loss": -0.20822713963496378,
|
| 1592 |
+
"train_critic_loss": 0.00012262144792447754,
|
| 1593 |
+
"train_mean_reward": 2.12979830394033e-05,
|
| 1594 |
+
"train_entropy": 0.503169753230535,
|
| 1595 |
+
"train_ce_loss": 0.5239167032667842,
|
| 1596 |
+
"train_dice_loss": 0.4882024209468793,
|
| 1597 |
+
"train_dice": 0.8315736476112815,
|
| 1598 |
+
"train_iou": 0.7281825279488283,
|
| 1599 |
+
"grad_norm": 6.927320420742035,
|
| 1600 |
+
"lr": 0.000638716070078746,
|
| 1601 |
+
"encoder_lr": 7.002816749495564e-05,
|
| 1602 |
+
"alpha": 0.38805800676345825,
|
| 1603 |
+
"validated_this_epoch": true,
|
| 1604 |
+
"val_loss": 0.8798701828768984,
|
| 1605 |
+
"val_dice": 0.6012577298029174,
|
| 1606 |
+
"val_iou": 0.5047951410836997,
|
| 1607 |
+
"val_reward": -6.265571851650346e-05,
|
| 1608 |
+
"val_entropy": 0.4579823882591845,
|
| 1609 |
+
"val_actor_loss": -0.017741811279480282,
|
| 1610 |
+
"val_critic_loss": 0.00010230121054317149,
|
| 1611 |
+
"val_ce_loss": 0.7957013085616207,
|
| 1612 |
+
"val_dice_loss": 0.9994203514522976
|
| 1613 |
+
},
|
| 1614 |
+
{
|
| 1615 |
+
"epoch": 63,
|
| 1616 |
+
"train_loss": 0.2721634865066825,
|
| 1617 |
+
"train_actor_loss": -0.19396853303680056,
|
| 1618 |
+
"train_critic_loss": 0.00013623307617868145,
|
| 1619 |
+
"train_mean_reward": 0.0003019678705501204,
|
| 1620 |
+
"train_entropy": 0.47740578613220114,
|
| 1621 |
+
"train_ce_loss": 0.497816272247105,
|
| 1622 |
+
"train_dice_loss": 0.43431152632603276,
|
| 1623 |
+
"train_dice": 0.8496567630300335,
|
| 1624 |
+
"train_iou": 0.7481885751088461,
|
| 1625 |
+
"grad_norm": 5.170062395242544,
|
| 1626 |
+
"lr": 0.0006098378488183934,
|
| 1627 |
+
"encoder_lr": 6.727095038969585e-05,
|
| 1628 |
+
"alpha": 0.3870657980442047,
|
| 1629 |
+
"validated_this_epoch": true,
|
| 1630 |
+
"val_loss": 0.907638921345951,
|
| 1631 |
+
"val_dice": 0.5940856983439678,
|
| 1632 |
+
"val_iou": 0.4963501057751015,
|
| 1633 |
+
"val_reward": 0.0003048808802709067,
|
| 1634 |
+
"val_entropy": 0.47827045739901186,
|
| 1635 |
+
"val_actor_loss": -0.016212405160956105,
|
| 1636 |
+
"val_critic_loss": 0.00010640207507907635,
|
| 1637 |
+
"val_ce_loss": 0.8167261049866374,
|
| 1638 |
+
"val_dice_loss": 1.0308701158171953
|
| 1639 |
+
},
|
| 1640 |
+
{
|
| 1641 |
+
"epoch": 64,
|
| 1642 |
+
"train_loss": 0.2990101644088729,
|
| 1643 |
+
"train_actor_loss": -0.19718888889138514,
|
| 1644 |
+
"train_critic_loss": 0.0001311541775113512,
|
| 1645 |
+
"train_mean_reward": 0.00023054286255221366,
|
| 1646 |
+
"train_entropy": 0.49388196300237613,
|
| 1647 |
+
"train_ce_loss": 0.48686752580583864,
|
| 1648 |
+
"train_dice_loss": 0.5053994170366191,
|
| 1649 |
+
"train_dice": 0.8151196065729833,
|
| 1650 |
+
"train_iou": 0.7204434649275893,
|
| 1651 |
+
"grad_norm": 6.3060275086989765,
|
| 1652 |
+
"lr": 0.0005813496048927529,
|
| 1653 |
+
"encoder_lr": 6.455096730131812e-05,
|
| 1654 |
+
"alpha": 0.3859546184539795,
|
| 1655 |
+
"validated_this_epoch": true,
|
| 1656 |
+
"val_loss": 0.8688988572296997,
|
| 1657 |
+
"val_dice": 0.605714650229503,
|
| 1658 |
+
"val_iou": 0.5090319832709783,
|
| 1659 |
+
"val_reward": 0.00041856793836584505,
|
| 1660 |
+
"val_entropy": 0.42664617074258393,
|
| 1661 |
+
"val_actor_loss": -0.016066622752472854,
|
| 1662 |
+
"val_critic_loss": 9.757263071169474e-05,
|
| 1663 |
+
"val_ce_loss": 0.8054486300062502,
|
| 1664 |
+
"val_dice_loss": 0.9643847237933767
|
| 1665 |
+
},
|
| 1666 |
+
{
|
| 1667 |
+
"epoch": 65,
|
| 1668 |
+
"train_loss": 0.2745288407179312,
|
| 1669 |
+
"train_actor_loss": -0.17929921690852213,
|
| 1670 |
+
"train_critic_loss": 0.00013641673014064555,
|
| 1671 |
+
"train_mean_reward": 0.0005741906336541619,
|
| 1672 |
+
"train_entropy": 0.44949914963963705,
|
| 1673 |
+
"train_ce_loss": 0.4482910321929899,
|
| 1674 |
+
"train_dice_loss": 0.4592286592874772,
|
| 1675 |
+
"train_dice": 0.8526112483996972,
|
| 1676 |
+
"train_iou": 0.7533322631144056,
|
| 1677 |
+
"grad_norm": 6.073780610011174,
|
| 1678 |
+
"lr": 0.0005532794527591512,
|
| 1679 |
+
"encoder_lr": 6.187090252474308e-05,
|
| 1680 |
+
"alpha": 0.38495776057243347,
|
| 1681 |
+
"validated_this_epoch": true,
|
| 1682 |
+
"val_loss": 0.8922605873085558,
|
| 1683 |
+
"val_dice": 0.6029577262767564,
|
| 1684 |
+
"val_iou": 0.5057121563958837,
|
| 1685 |
+
"val_reward": 0.00014996218061067405,
|
| 1686 |
+
"val_entropy": 0.4466387471466353,
|
| 1687 |
+
"val_actor_loss": -0.013949708992231996,
|
| 1688 |
+
"val_critic_loss": 9.188769139754959e-05,
|
| 1689 |
+
"val_ce_loss": 0.8445297856579977,
|
| 1690 |
+
"val_dice_loss": 0.9677988889843526
|
| 1691 |
+
},
|
| 1692 |
+
{
|
| 1693 |
+
"epoch": 66,
|
| 1694 |
+
"train_loss": 0.28922464869594056,
|
| 1695 |
+
"train_actor_loss": -0.19522383355368406,
|
| 1696 |
+
"train_critic_loss": 0.0001344795403919983,
|
| 1697 |
+
"train_mean_reward": 0.00016107145717843453,
|
| 1698 |
+
"train_entropy": 0.49811874272731643,
|
| 1699 |
+
"train_ce_loss": 0.5179691114821112,
|
| 1700 |
+
"train_dice_loss": 0.45079336563746136,
|
| 1701 |
+
"train_dice": 0.8182452856919139,
|
| 1702 |
+
"train_iou": 0.7181444461448314,
|
| 1703 |
+
"grad_norm": 5.792482366928687,
|
| 1704 |
+
"lr": 0.0005256550942687934,
|
| 1705 |
+
"encoder_lr": 5.9233400960337056e-05,
|
| 1706 |
+
"alpha": 0.38398146629333496,
|
| 1707 |
+
"validated_this_epoch": true,
|
| 1708 |
+
"val_loss": 0.9030555196275766,
|
| 1709 |
+
"val_dice": 0.6064474311478678,
|
| 1710 |
+
"val_iou": 0.5103693877311523,
|
| 1711 |
+
"val_reward": -6.503151592304942e-05,
|
| 1712 |
+
"val_entropy": 0.5157438034963125,
|
| 1713 |
+
"val_actor_loss": -0.010442330489038594,
|
| 1714 |
+
"val_critic_loss": 7.160131190606519e-05,
|
| 1715 |
+
"val_ce_loss": 0.8759362922812048,
|
| 1716 |
+
"val_dice_loss": 0.950987774615336
|
| 1717 |
+
},
|
| 1718 |
+
{
|
| 1719 |
+
"epoch": 67,
|
| 1720 |
+
"train_loss": 0.3107026227069302,
|
| 1721 |
+
"train_actor_loss": -0.19075578670853222,
|
| 1722 |
+
"train_critic_loss": 0.0001530483284510317,
|
| 1723 |
+
"train_mean_reward": 0.00025004432190820435,
|
| 1724 |
+
"train_entropy": 0.488763628861843,
|
| 1725 |
+
"train_ce_loss": 0.5022753198057986,
|
| 1726 |
+
"train_dice_loss": 0.5004884413419626,
|
| 1727 |
+
"train_dice": 0.8059581387262413,
|
| 1728 |
+
"train_iou": 0.7085136932470913,
|
| 1729 |
+
"grad_norm": 9.728274372907785,
|
| 1730 |
+
"lr": 0.0004985037913283807,
|
| 1731 |
+
"encoder_lr": 5.6641065503714745e-05,
|
| 1732 |
+
"alpha": 0.38284817337989807,
|
| 1733 |
+
"validated_this_epoch": true,
|
| 1734 |
+
"val_loss": 0.8913962332193147,
|
| 1735 |
+
"val_dice": 0.6027492411901012,
|
| 1736 |
+
"val_iou": 0.5061110134774517,
|
| 1737 |
+
"val_reward": -7.252736799677629e-06,
|
| 1738 |
+
"val_entropy": 0.41084717229159196,
|
| 1739 |
+
"val_actor_loss": -0.02063259099741852,
|
| 1740 |
+
"val_critic_loss": 0.00012683981399598936,
|
| 1741 |
+
"val_ce_loss": 0.8199991476016513,
|
| 1742 |
+
"val_dice_loss": 1.0039316340528353
|
| 1743 |
+
},
|
| 1744 |
+
{
|
| 1745 |
+
"epoch": 68,
|
| 1746 |
+
"train_loss": 0.32868115906603634,
|
| 1747 |
+
"train_actor_loss": -0.18738090509596544,
|
| 1748 |
+
"train_critic_loss": 0.00014048918244906163,
|
| 1749 |
+
"train_mean_reward": 0.0003658626102291097,
|
| 1750 |
+
"train_entropy": 0.48206986868992835,
|
| 1751 |
+
"train_ce_loss": 0.49555252404071576,
|
| 1752 |
+
"train_dice_loss": 0.5364311050910217,
|
| 1753 |
+
"train_dice": 0.7835065184227434,
|
| 1754 |
+
"train_iou": 0.6937384607701952,
|
| 1755 |
+
"grad_norm": 5.309517626578991,
|
| 1756 |
+
"lr": 0.0004718523389958983,
|
| 1757 |
+
"encoder_lr": 5.4096454476995315e-05,
|
| 1758 |
+
"alpha": 0.38180801272392273,
|
| 1759 |
+
"validated_this_epoch": true,
|
| 1760 |
+
"val_loss": 0.8810200722972777,
|
| 1761 |
+
"val_dice": 0.5970910613609856,
|
| 1762 |
+
"val_iou": 0.5071858303877831,
|
| 1763 |
+
"val_reward": -0.0001621242210805778,
|
| 1764 |
+
"val_entropy": 0.3969633978123617,
|
| 1765 |
+
"val_actor_loss": -0.022147858157316506,
|
| 1766 |
+
"val_critic_loss": 0.00013076663070005266,
|
| 1767 |
+
"val_ce_loss": 0.7970191046426242,
|
| 1768 |
+
"val_dice_loss": 1.0091859574871833
|
| 1769 |
+
},
|
| 1770 |
+
{
|
| 1771 |
+
"epoch": 69,
|
| 1772 |
+
"train_loss": 0.28730765755216664,
|
| 1773 |
+
"train_actor_loss": -0.1929737371034347,
|
| 1774 |
+
"train_critic_loss": 0.00014561648407461456,
|
| 1775 |
+
"train_mean_reward": 0.0004443684541538227,
|
| 1776 |
+
"train_entropy": 0.4926067119798599,
|
| 1777 |
+
"train_ce_loss": 0.5086482349448862,
|
| 1778 |
+
"train_dice_loss": 0.4517689255567697,
|
| 1779 |
+
"train_dice": 0.8338557959410619,
|
| 1780 |
+
"train_iou": 0.7430367450727097,
|
| 1781 |
+
"grad_norm": 8.434496705348675,
|
| 1782 |
+
"lr": 0.0004457270390371302,
|
| 1783 |
+
"encoder_lr": 5.1602079104047604e-05,
|
| 1784 |
+
"alpha": 0.38070356845855713,
|
| 1785 |
+
"validated_this_epoch": true,
|
| 1786 |
+
"val_loss": 0.9066408646688091,
|
| 1787 |
+
"val_dice": 0.611139434679143,
|
| 1788 |
+
"val_iou": 0.5141351436794417,
|
| 1789 |
+
"val_reward": 9.719789765464763e-05,
|
| 1790 |
+
"val_entropy": 0.42087389242769496,
|
| 1791 |
+
"val_actor_loss": -0.018900024174298678,
|
| 1792 |
+
"val_critic_loss": 0.00012189041944561679,
|
| 1793 |
+
"val_ce_loss": 0.8825005121021108,
|
| 1794 |
+
"val_dice_loss": 0.9684593421642226
|
| 1795 |
+
},
|
| 1796 |
+
{
|
| 1797 |
+
"epoch": 70,
|
| 1798 |
+
"train_loss": 0.34099637091947865,
|
| 1799 |
+
"train_actor_loss": -0.1854064502299596,
|
| 1800 |
+
"train_critic_loss": 0.00014670636371122854,
|
| 1801 |
+
"train_mean_reward": 0.0006124423781353143,
|
| 1802 |
+
"train_entropy": 0.4824148356341399,
|
| 1803 |
+
"train_ce_loss": 0.5452016441342541,
|
| 1804 |
+
"train_dice_loss": 0.5074572807703263,
|
| 1805 |
+
"train_dice": 0.8030231604535206,
|
| 1806 |
+
"train_iou": 0.7004860176102203,
|
| 1807 |
+
"grad_norm": 10.249180711232698,
|
| 1808 |
+
"lr": 0.00042015367396898955,
|
| 1809 |
+
"encoder_lr": 4.916040103221508e-05,
|
| 1810 |
+
"alpha": 0.37965887784957886,
|
| 1811 |
+
"validated_this_epoch": true,
|
| 1812 |
+
"val_loss": 0.8702996345307452,
|
| 1813 |
+
"val_dice": 0.6219251386510134,
|
| 1814 |
+
"val_iou": 0.5283526989821701,
|
| 1815 |
+
"val_reward": 0.000216524192008438,
|
| 1816 |
+
"val_entropy": 0.44529393194901823,
|
| 1817 |
+
"val_actor_loss": -0.020135502586131823,
|
| 1818 |
+
"val_critic_loss": 0.00012282488386967952,
|
| 1819 |
+
"val_ce_loss": 0.7967280350352702,
|
| 1820 |
+
"val_dice_loss": 0.9840193875510281
|
| 1821 |
+
},
|
| 1822 |
+
{
|
| 1823 |
+
"epoch": 71,
|
| 1824 |
+
"train_loss": 0.30212150890236866,
|
| 1825 |
+
"train_actor_loss": -0.19124758214904716,
|
| 1826 |
+
"train_critic_loss": 0.00013821942818536534,
|
| 1827 |
+
"train_mean_reward": 0.00022915898811194126,
|
| 1828 |
+
"train_entropy": 0.48925272852946544,
|
| 1829 |
+
"train_ce_loss": 0.49435089609155863,
|
| 1830 |
+
"train_dice_loss": 0.4922490566968918,
|
| 1831 |
+
"train_dice": 0.8498296153311636,
|
| 1832 |
+
"train_iou": 0.7505032367566052,
|
| 1833 |
+
"grad_norm": 5.55736152942364,
|
| 1834 |
+
"lr": 0.0003951574816152885,
|
| 1835 |
+
"encoder_lr": 4.6773829902967244e-05,
|
| 1836 |
+
"alpha": 0.3785615861415863,
|
| 1837 |
+
"validated_this_epoch": true,
|
| 1838 |
+
"val_loss": 0.8599245181716414,
|
| 1839 |
+
"val_dice": 0.6212996052824842,
|
| 1840 |
+
"val_iou": 0.5272825920914537,
|
| 1841 |
+
"val_reward": 0.0002958468498256866,
|
| 1842 |
+
"val_entropy": 0.45900811571063416,
|
| 1843 |
+
"val_actor_loss": -0.02031575707892765,
|
| 1844 |
+
"val_critic_loss": 0.00011488897363428024,
|
| 1845 |
+
"val_ce_loss": 0.7881311783990398,
|
| 1846 |
+
"val_dice_loss": 0.9722344508074752
|
| 1847 |
+
},
|
| 1848 |
+
{
|
| 1849 |
+
"epoch": 72,
|
| 1850 |
+
"train_loss": 0.3004536844539241,
|
| 1851 |
+
"train_actor_loss": -0.1847940833809284,
|
| 1852 |
+
"train_critic_loss": 0.00014606587070048615,
|
| 1853 |
+
"train_mean_reward": 0.0002179395563381639,
|
| 1854 |
+
"train_entropy": 0.4770303479371926,
|
| 1855 |
+
"train_ce_loss": 0.47633102157702434,
|
| 1856 |
+
"train_dice_loss": 0.494018438534859,
|
| 1857 |
+
"train_dice": 0.8280124818431396,
|
| 1858 |
+
"train_iou": 0.7299795700419768,
|
| 1859 |
+
"grad_norm": 9.941169550785652,
|
| 1860 |
+
"lr": 0.0003707631302000539,
|
| 1861 |
+
"encoder_lr": 4.444472097387449e-05,
|
| 1862 |
+
"alpha": 0.37742236256599426,
|
| 1863 |
+
"validated_this_epoch": true,
|
| 1864 |
+
"val_loss": 0.8680593910993952,
|
| 1865 |
+
"val_dice": 0.6211641997746601,
|
| 1866 |
+
"val_iou": 0.5266808041231041,
|
| 1867 |
+
"val_reward": 7.150833911822183e-05,
|
| 1868 |
+
"val_entropy": 0.45297327788189207,
|
| 1869 |
+
"val_actor_loss": -0.020816786001869148,
|
| 1870 |
+
"val_critic_loss": 0.00012961177388647093,
|
| 1871 |
+
"val_ce_loss": 0.8060964255621939,
|
| 1872 |
+
"val_dice_loss": 0.9715262895280664
|
| 1873 |
+
},
|
| 1874 |
+
{
|
| 1875 |
+
"epoch": 73,
|
| 1876 |
+
"train_loss": 0.2841995174709994,
|
| 1877 |
+
"train_actor_loss": -0.19595208697212047,
|
| 1878 |
+
"train_critic_loss": 0.00014394682576051148,
|
| 1879 |
+
"train_mean_reward": 0.00015219458174005223,
|
| 1880 |
+
"train_entropy": 0.502110990958336,
|
| 1881 |
+
"train_ce_loss": 0.49953122411329204,
|
| 1882 |
+
"train_dice_loss": 0.46062802275021875,
|
| 1883 |
+
"train_dice": 0.8570822489027884,
|
| 1884 |
+
"train_iou": 0.7591080881801306,
|
| 1885 |
+
"grad_norm": 5.974417970730708,
|
| 1886 |
+
"lr": 0.00034699469400296674,
|
| 1887 |
+
"encoder_lr": 4.2175372794253104e-05,
|
| 1888 |
+
"alpha": 0.3762379586696625,
|
| 1889 |
+
"validated_this_epoch": true,
|
| 1890 |
+
"val_loss": 0.8855846142396331,
|
| 1891 |
+
"val_dice": 0.6103091710832977,
|
| 1892 |
+
"val_iou": 0.5120221808064016,
|
| 1893 |
+
"val_reward": -4.068112247470635e-05,
|
| 1894 |
+
"val_entropy": 0.4602475205455163,
|
| 1895 |
+
"val_actor_loss": -0.018500854101883377,
|
| 1896 |
+
"val_critic_loss": 0.00012601041155713033,
|
| 1897 |
+
"val_ce_loss": 0.8499220060466818,
|
| 1898 |
+
"val_dice_loss": 0.9581228852874101
|
| 1899 |
+
},
|
| 1900 |
+
{
|
| 1901 |
+
"epoch": 74,
|
| 1902 |
+
"train_loss": 0.31007336736477625,
|
| 1903 |
+
"train_actor_loss": -0.18613322193805987,
|
| 1904 |
+
"train_critic_loss": 0.00015803580458267764,
|
| 1905 |
+
"train_mean_reward": 0.0005916463468076053,
|
| 1906 |
+
"train_entropy": 0.4862243456718249,
|
| 1907 |
+
"train_ce_loss": 0.5056916614755605,
|
| 1908 |
+
"train_dice_loss": 0.48656347241157144,
|
| 1909 |
+
"train_dice": 0.8165767454162977,
|
| 1910 |
+
"train_iou": 0.7261843298401762,
|
| 1911 |
+
"grad_norm": 6.8189342480439405,
|
| 1912 |
+
"lr": 0.0003238756296009549,
|
| 1913 |
+
"encoder_lr": 3.996802493677459e-05,
|
| 1914 |
+
"alpha": 0.37507492303848267,
|
| 1915 |
+
"validated_this_epoch": true,
|
| 1916 |
+
"val_loss": 0.9208462698328675,
|
| 1917 |
+
"val_dice": 0.6029088126402966,
|
| 1918 |
+
"val_iou": 0.5078892600271895,
|
| 1919 |
+
"val_reward": 2.3689248718430413e-05,
|
| 1920 |
+
"val_entropy": 0.48940782790834253,
|
| 1921 |
+
"val_actor_loss": -0.017324071475704472,
|
| 1922 |
+
"val_critic_loss": 0.00011526664500918126,
|
| 1923 |
+
"val_ce_loss": 0.8885778282402139,
|
| 1924 |
+
"val_dice_loss": 0.9876475544890971
|
| 1925 |
+
},
|
| 1926 |
+
{
|
| 1927 |
+
"epoch": 75,
|
| 1928 |
+
"train_loss": 0.2814805630701952,
|
| 1929 |
+
"train_actor_loss": -0.18184969110939747,
|
| 1930 |
+
"train_critic_loss": 0.00014516125647554497,
|
| 1931 |
+
"train_mean_reward": 0.00029474913484451605,
|
| 1932 |
+
"train_entropy": 0.4810859522758386,
|
| 1933 |
+
"train_ce_loss": 0.46176618263603025,
|
| 1934 |
+
"train_dice_loss": 0.46474915704666037,
|
| 1935 |
+
"train_dice": 0.8488762067521319,
|
| 1936 |
+
"train_iou": 0.7544699536234725,
|
| 1937 |
+
"grad_norm": 5.232173529955057,
|
| 1938 |
+
"lr": 0.0003014287527193855,
|
| 1939 |
+
"encoder_lr": 3.782485578727801e-05,
|
| 1940 |
+
"alpha": 0.37393489480018616,
|
| 1941 |
+
"validated_this_epoch": true,
|
| 1942 |
+
"val_loss": 0.9202833509806431,
|
| 1943 |
+
"val_dice": 0.6119766296259063,
|
| 1944 |
+
"val_iou": 0.5180474791545857,
|
| 1945 |
+
"val_reward": -0.00020453242307143452,
|
| 1946 |
+
"val_entropy": 0.5237617375272692,
|
| 1947 |
+
"val_actor_loss": -0.017480364529124336,
|
| 1948 |
+
"val_critic_loss": 0.00010540165363645521,
|
| 1949 |
+
"val_ce_loss": 0.9160002517429265,
|
| 1950 |
+
"val_dice_loss": 0.9594217484647576
|
| 1951 |
+
},
|
| 1952 |
+
{
|
| 1953 |
+
"epoch": 76,
|
| 1954 |
+
"train_loss": 0.28183041254837565,
|
| 1955 |
+
"train_actor_loss": -0.18178620566733375,
|
| 1956 |
+
"train_critic_loss": 0.0001631878366984333,
|
| 1957 |
+
"train_mean_reward": 0.00054003503650025,
|
| 1958 |
+
"train_entropy": 0.4865949434729723,
|
| 1959 |
+
"train_ce_loss": 0.5124741354778123,
|
| 1960 |
+
"train_dice_loss": 0.4145959054048245,
|
| 1961 |
+
"train_dice": 0.8454685018343084,
|
| 1962 |
+
"train_iou": 0.7513091575865652,
|
| 1963 |
+
"grad_norm": 6.741730777116922,
|
| 1964 |
+
"lr": 0.0002796762157156959,
|
| 1965 |
+
"encoder_lr": 3.574798039496594e-05,
|
| 1966 |
+
"alpha": 0.3728664219379425,
|
| 1967 |
+
"validated_this_epoch": true,
|
| 1968 |
+
"val_loss": 0.9453801505380508,
|
| 1969 |
+
"val_dice": 0.6006386207594633,
|
| 1970 |
+
"val_iou": 0.5023192862513284,
|
| 1971 |
+
"val_reward": 0.00016976932316409678,
|
| 1972 |
+
"val_entropy": 0.5494604779012275,
|
| 1973 |
+
"val_actor_loss": -0.015140259852444734,
|
| 1974 |
+
"val_critic_loss": 0.0001216809782407482,
|
| 1975 |
+
"val_ce_loss": 0.9246900034598028,
|
| 1976 |
+
"val_dice_loss": 0.9962291067296808
|
| 1977 |
+
},
|
| 1978 |
+
{
|
| 1979 |
+
"epoch": 77,
|
| 1980 |
+
"train_loss": 0.24605555453924283,
|
| 1981 |
+
"train_actor_loss": -0.18466914387849662,
|
| 1982 |
+
"train_critic_loss": 0.00013230649207029134,
|
| 1983 |
+
"train_mean_reward": 0.0006642146627783428,
|
| 1984 |
+
"train_entropy": 0.48255815414281994,
|
| 1985 |
+
"train_ce_loss": 0.4454827116121753,
|
| 1986 |
+
"train_dice_loss": 0.41583436956772435,
|
| 1987 |
+
"train_dice": 0.8555910680808273,
|
| 1988 |
+
"train_iou": 0.7565827743679869,
|
| 1989 |
+
"grad_norm": 4.409337969926687,
|
| 1990 |
+
"lr": 0.00025863948571769286,
|
| 1991 |
+
"encoder_lr": 3.373944838510636e-05,
|
| 1992 |
+
"alpha": 0.3717004954814911,
|
| 1993 |
+
"validated_this_epoch": true,
|
| 1994 |
+
"val_loss": 0.9262658662698937,
|
| 1995 |
+
"val_dice": 0.6064658726600931,
|
| 1996 |
+
"val_iou": 0.5074053928219786,
|
| 1997 |
+
"val_reward": 0.0003167587459747572,
|
| 1998 |
+
"val_entropy": 0.5022999824899615,
|
| 1999 |
+
"val_actor_loss": -0.013811814804523806,
|
| 2000 |
+
"val_critic_loss": 0.0001108699472678556,
|
| 2001 |
+
"val_ce_loss": 0.8954231274511779,
|
| 2002 |
+
"val_dice_loss": 0.9846213333534471
|
| 2003 |
+
},
|
| 2004 |
+
{
|
| 2005 |
+
"epoch": 78,
|
| 2006 |
+
"train_loss": 0.26691848347679925,
|
| 2007 |
+
"train_actor_loss": -0.17751583051031983,
|
| 2008 |
+
"train_critic_loss": 0.00013943205368167793,
|
| 2009 |
+
"train_mean_reward": 0.0005254422461663266,
|
| 2010 |
+
"train_entropy": 0.4723521386965727,
|
| 2011 |
+
"train_ce_loss": 0.46296644443646073,
|
| 2012 |
+
"train_dice_loss": 0.425762741229473,
|
| 2013 |
+
"train_dice": 0.8485370044125651,
|
| 2014 |
+
"train_iou": 0.7561243847984314,
|
| 2015 |
+
"grad_norm": 5.1703636554571295,
|
| 2016 |
+
"lr": 0.00023833932343808993,
|
| 2017 |
+
"encoder_lr": 3.180124193630005e-05,
|
| 2018 |
+
"alpha": 0.3706134855747223,
|
| 2019 |
+
"validated_this_epoch": true,
|
| 2020 |
+
"val_loss": 0.9224502453368835,
|
| 2021 |
+
"val_dice": 0.6213934976736417,
|
| 2022 |
+
"val_iou": 0.527098100731283,
|
| 2023 |
+
"val_reward": -6.624878408205627e-05,
|
| 2024 |
+
"val_entropy": 0.5185162828426169,
|
| 2025 |
+
"val_actor_loss": -0.016354706689875313,
|
| 2026 |
+
"val_critic_loss": 0.00010780544884292253,
|
| 2027 |
+
"val_ce_loss": 0.9257690728303383,
|
| 2028 |
+
"val_dice_loss": 0.9517329946310833
|
| 2029 |
+
},
|
| 2030 |
+
{
|
| 2031 |
+
"epoch": 79,
|
| 2032 |
+
"train_loss": 0.297496394522918,
|
| 2033 |
+
"train_actor_loss": -0.18500445477473434,
|
| 2034 |
+
"train_critic_loss": 0.00014265464497862596,
|
| 2035 |
+
"train_mean_reward": 0.00018240018614680876,
|
| 2036 |
+
"train_entropy": 0.5008839649649768,
|
| 2037 |
+
"train_ce_loss": 0.5050131210412544,
|
| 2038 |
+
"train_dice_loss": 0.4598459119980151,
|
| 2039 |
+
"train_dice": 0.8363083449112527,
|
| 2040 |
+
"train_iou": 0.7415909851930982,
|
| 2041 |
+
"grad_norm": 9.3604137805792,
|
| 2042 |
+
"lr": 0.00021879576268618817,
|
| 2043 |
+
"encoder_lr": 2.993527382430943e-05,
|
| 2044 |
+
"alpha": 0.3693552613258362,
|
| 2045 |
+
"validated_this_epoch": true,
|
| 2046 |
+
"val_loss": 0.8920595440590247,
|
| 2047 |
+
"val_dice": 0.6238934243729093,
|
| 2048 |
+
"val_iou": 0.5306661044090598,
|
| 2049 |
+
"val_reward": -0.00019551313891148458,
|
| 2050 |
+
"val_entropy": 0.44737423820929095,
|
| 2051 |
+
"val_actor_loss": -0.0209494739452805,
|
| 2052 |
+
"val_critic_loss": 0.00012903150940878227,
|
| 2053 |
+
"val_ce_loss": 0.8655559535260604,
|
| 2054 |
+
"val_dice_loss": 0.9603330246125807
|
| 2055 |
+
},
|
| 2056 |
+
{
|
| 2057 |
+
"epoch": 80,
|
| 2058 |
+
"train_loss": 0.26588256865775645,
|
| 2059 |
+
"train_actor_loss": -0.1740036221841971,
|
| 2060 |
+
"train_critic_loss": 0.00014027008693609297,
|
| 2061 |
+
"train_mean_reward": 0.0006092136761067079,
|
| 2062 |
+
"train_entropy": 0.4640271275853499,
|
| 2063 |
+
"train_ce_loss": 0.45322618777983076,
|
| 2064 |
+
"train_dice_loss": 0.42640591775759673,
|
| 2065 |
+
"train_dice": 0.8570260633440578,
|
| 2066 |
+
"train_iou": 0.7646250000186995,
|
| 2067 |
+
"grad_norm": 5.054833015570273,
|
| 2068 |
+
"lr": 0.0002000280905969275,
|
| 2069 |
+
"encoder_lr": 2.814338553438002e-05,
|
| 2070 |
+
"alpha": 0.3683135211467743,
|
| 2071 |
+
"validated_this_epoch": true,
|
| 2072 |
+
"val_loss": 0.888139437432542,
|
| 2073 |
+
"val_dice": 0.6259517353671047,
|
| 2074 |
+
"val_iou": 0.5311825716078227,
|
| 2075 |
+
"val_reward": -2.698475608750331e-05,
|
| 2076 |
+
"val_entropy": 0.4602269205180081,
|
| 2077 |
+
"val_actor_loss": -0.018199615679079233,
|
| 2078 |
+
"val_critic_loss": 0.00011391557994772824,
|
| 2079 |
+
"val_ce_loss": 0.8400215685668617,
|
| 2080 |
+
"val_dice_loss": 0.9725425953816884
|
| 2081 |
+
},
|
| 2082 |
+
{
|
| 2083 |
+
"epoch": 81,
|
| 2084 |
+
"train_loss": 0.24823787364249045,
|
| 2085 |
+
"train_actor_loss": -0.17730279133105886,
|
| 2086 |
+
"train_critic_loss": 0.00014715902675678267,
|
| 2087 |
+
"train_mean_reward": 0.0001667045748595578,
|
| 2088 |
+
"train_entropy": 0.47732657041305154,
|
| 2089 |
+
"train_ce_loss": 0.4548338792012192,
|
| 2090 |
+
"train_dice_loss": 0.3961002849615537,
|
| 2091 |
+
"train_dice": 0.8571439104921678,
|
| 2092 |
+
"train_iou": 0.7633327672294542,
|
| 2093 |
+
"grad_norm": 4.981220158246847,
|
| 2094 |
+
"lr": 0.00018205482859681116,
|
| 2095 |
+
"encoder_lr": 2.6427345443916652e-05,
|
| 2096 |
+
"alpha": 0.3672678470611572,
|
| 2097 |
+
"validated_this_epoch": true,
|
| 2098 |
+
"val_loss": 0.9121875078417361,
|
| 2099 |
+
"val_dice": 0.628116376617652,
|
| 2100 |
+
"val_iou": 0.5325012422437716,
|
| 2101 |
+
"val_reward": 3.311433442142901e-05,
|
| 2102 |
+
"val_entropy": 0.5172066507917462,
|
| 2103 |
+
"val_actor_loss": -0.0154001694384285,
|
| 2104 |
+
"val_critic_loss": 9.967011668768371e-05,
|
| 2105 |
+
"val_ce_loss": 0.9080731739205392,
|
| 2106 |
+
"val_dice_loss": 0.9470024828356925
|
| 2107 |
+
},
|
| 2108 |
+
{
|
| 2109 |
+
"epoch": 82,
|
| 2110 |
+
"train_loss": 0.2940742807247891,
|
| 2111 |
+
"train_actor_loss": -0.18235400027762622,
|
| 2112 |
+
"train_critic_loss": 0.00014340668398150368,
|
| 2113 |
+
"train_mean_reward": 0.0003457837028457643,
|
| 2114 |
+
"train_entropy": 0.4921326803473326,
|
| 2115 |
+
"train_ce_loss": 0.5105866705162976,
|
| 2116 |
+
"train_dice_loss": 0.44212647584768444,
|
| 2117 |
+
"train_dice": 0.8629459218651641,
|
| 2118 |
+
"train_iou": 0.7718264933894662,
|
| 2119 |
+
"grad_norm": 8.240061037815535,
|
| 2120 |
+
"lr": 0.00016489371412549528,
|
| 2121 |
+
"encoder_lr": 2.4788847077308606e-05,
|
| 2122 |
+
"alpha": 0.3660430312156677,
|
| 2123 |
+
"validated_this_epoch": true,
|
| 2124 |
+
"val_loss": 0.9140039128745256,
|
| 2125 |
+
"val_dice": 0.6262328740091343,
|
| 2126 |
+
"val_iou": 0.5294085958711294,
|
| 2127 |
+
"val_reward": 9.253557100512832e-05,
|
| 2128 |
+
"val_entropy": 0.4962097415418336,
|
| 2129 |
+
"val_actor_loss": -0.01747522158977444,
|
| 2130 |
+
"val_critic_loss": 0.000114031184587089,
|
| 2131 |
+
"val_ce_loss": 0.9120565438125696,
|
| 2132 |
+
"val_dice_loss": 0.9507876652659792
|
| 2133 |
+
},
|
| 2134 |
+
{
|
| 2135 |
+
"epoch": 83,
|
| 2136 |
+
"train_loss": 0.27110321801872206,
|
| 2137 |
+
"train_actor_loss": -0.18032036363505397,
|
| 2138 |
+
"train_critic_loss": 0.00013988472077331745,
|
| 2139 |
+
"train_mean_reward": 0.0005736026509380165,
|
| 2140 |
+
"train_entropy": 0.48102405151495564,
|
| 2141 |
+
"train_ce_loss": 0.46033185648803526,
|
| 2142 |
+
"train_dice_loss": 0.44237541350034565,
|
| 2143 |
+
"train_dice": 0.8596369178712174,
|
| 2144 |
+
"train_iou": 0.7710822223304339,
|
| 2145 |
+
"grad_norm": 5.008076278062967,
|
| 2146 |
+
"lr": 0.00014856168313107603,
|
| 2147 |
+
"encoder_lr": 2.322950743462536e-05,
|
| 2148 |
+
"alpha": 0.3648666739463806,
|
| 2149 |
+
"validated_this_epoch": true,
|
| 2150 |
+
"val_loss": 0.8855890883838363,
|
| 2151 |
+
"val_dice": 0.625889737341694,
|
| 2152 |
+
"val_iou": 0.5311175002179157,
|
| 2153 |
+
"val_reward": -0.0001268098520159409,
|
| 2154 |
+
"val_entropy": 0.4768302592665258,
|
| 2155 |
+
"val_actor_loss": -0.022714404480983365,
|
| 2156 |
+
"val_critic_loss": 0.00016350325496673716,
|
| 2157 |
+
"val_ce_loss": 0.8590718081145727,
|
| 2158 |
+
"val_dice_loss": 0.9573716494170103
|
| 2159 |
+
},
|
| 2160 |
+
{
|
| 2161 |
+
"epoch": 84,
|
| 2162 |
+
"train_loss": 0.2676643088298778,
|
| 2163 |
+
"train_actor_loss": -0.17665937337546778,
|
| 2164 |
+
"train_critic_loss": 0.00014830253348918632,
|
| 2165 |
+
"train_mean_reward": 0.0007128644160146883,
|
| 2166 |
+
"train_entropy": 0.4825554761366967,
|
| 2167 |
+
"train_ce_loss": 0.463015327301736,
|
| 2168 |
+
"train_dice_loss": 0.4254837273023067,
|
| 2169 |
+
"train_dice": 0.8536968166921654,
|
| 2170 |
+
"train_iou": 0.7631068182926551,
|
| 2171 |
+
"grad_norm": 5.155165039576017,
|
| 2172 |
+
"lr": 0.0001330748533563558,
|
| 2173 |
+
"encoder_lr": 2.1750865395832973e-05,
|
| 2174 |
+
"alpha": 0.36373817920684814,
|
| 2175 |
+
"validated_this_epoch": true,
|
| 2176 |
+
"val_loss": 0.9004585269387021,
|
| 2177 |
+
"val_dice": 0.6220800710686369,
|
| 2178 |
+
"val_iou": 0.5296247949453172,
|
| 2179 |
+
"val_reward": -0.0001532350773718784,
|
| 2180 |
+
"val_entropy": 0.469834610518783,
|
| 2181 |
+
"val_actor_loss": -0.022014073442521907,
|
| 2182 |
+
"val_critic_loss": 0.00014843637802756587,
|
| 2183 |
+
"val_ce_loss": 0.887462952258912,
|
| 2184 |
+
"val_dice_loss": 0.9573337875231349
|
| 2185 |
+
},
|
| 2186 |
+
{
|
| 2187 |
+
"epoch": 85,
|
| 2188 |
+
"train_loss": 0.32850813274629986,
|
| 2189 |
+
"train_actor_loss": -0.16608662014970413,
|
| 2190 |
+
"train_critic_loss": 0.00015342057887233424,
|
| 2191 |
+
"train_mean_reward": 0.00040314422294147493,
|
| 2192 |
+
"train_entropy": 0.46191612564218354,
|
| 2193 |
+
"train_ce_loss": 0.5117159546472323,
|
| 2194 |
+
"train_dice_loss": 0.4773201201206599,
|
| 2195 |
+
"train_dice": 0.8304328407142677,
|
| 2196 |
+
"train_iou": 0.7338533262703933,
|
| 2197 |
+
"grad_norm": 7.113032263058883,
|
| 2198 |
+
"lr": 0.0001184485084325741,
|
| 2199 |
+
"encoder_lr": 2.035438020210507e-05,
|
| 2200 |
+
"alpha": 0.3626442849636078,
|
| 2201 |
+
"validated_this_epoch": true,
|
| 2202 |
+
"val_loss": 0.913818384647708,
|
| 2203 |
+
"val_dice": 0.616532847523817,
|
| 2204 |
+
"val_iou": 0.5234112563546448,
|
| 2205 |
+
"val_reward": -0.00048799517311764616,
|
| 2206 |
+
"val_entropy": 0.5095273745481415,
|
| 2207 |
+
"val_actor_loss": -0.022590550389423064,
|
| 2208 |
+
"val_critic_loss": 0.00014190132424118929,
|
| 2209 |
+
"val_ce_loss": 0.8960623152347075,
|
| 2210 |
+
"val_dice_loss": 0.9766136257335393
|
| 2211 |
+
},
|
| 2212 |
+
{
|
| 2213 |
+
"epoch": 86,
|
| 2214 |
+
"train_loss": 0.2801339369929897,
|
| 2215 |
+
"train_actor_loss": -0.18246090010954782,
|
| 2216 |
+
"train_critic_loss": 0.00015208929184714525,
|
| 2217 |
+
"train_mean_reward": 0.0004905434054937639,
|
| 2218 |
+
"train_entropy": 0.4969217872772461,
|
| 2219 |
+
"train_ce_loss": 0.48885619382445633,
|
| 2220 |
+
"train_dice_loss": 0.43618137943438995,
|
| 2221 |
+
"train_dice": 0.8447765043128366,
|
| 2222 |
+
"train_iou": 0.752234130061901,
|
| 2223 |
+
"grad_norm": 5.028124025234809,
|
| 2224 |
+
"lr": 0.00010469708279631079,
|
| 2225 |
+
"encoder_lr": 1.904143001572817e-05,
|
| 2226 |
+
"alpha": 0.36137738823890686,
|
| 2227 |
+
"validated_this_epoch": true,
|
| 2228 |
+
"val_loss": 0.9022679183999021,
|
| 2229 |
+
"val_dice": 0.620100275613823,
|
| 2230 |
+
"val_iou": 0.5262265482921982,
|
| 2231 |
+
"val_reward": -0.0002343872629996504,
|
| 2232 |
+
"val_entropy": 0.4711363485666237,
|
| 2233 |
+
"val_actor_loss": -0.02026250630627725,
|
| 2234 |
+
"val_critic_loss": 0.00013882769730980265,
|
| 2235 |
+
"val_ce_loss": 0.8692544558640533,
|
| 2236 |
+
"val_dice_loss": 0.9756675344524961
|
| 2237 |
+
},
|
| 2238 |
+
{
|
| 2239 |
+
"epoch": 87,
|
| 2240 |
+
"train_loss": 0.25412349432456094,
|
| 2241 |
+
"train_actor_loss": -0.16924200737132475,
|
| 2242 |
+
"train_critic_loss": 0.00013542888098947096,
|
| 2243 |
+
"train_mean_reward": 0.00036885723453358363,
|
| 2244 |
+
"train_entropy": 0.4665726570364757,
|
| 2245 |
+
"train_ce_loss": 0.45580293560543894,
|
| 2246 |
+
"train_dice_loss": 0.39079263538886333,
|
| 2247 |
+
"train_dice": 0.8428399269115253,
|
| 2248 |
+
"train_iou": 0.7565121730965289,
|
| 2249 |
+
"grad_norm": 4.22807672390571,
|
| 2250 |
+
"lr": 9.183414744443903e-05,
|
| 2251 |
+
"encoder_lr": 1.781331056002182e-05,
|
| 2252 |
+
"alpha": 0.3602869212627411,
|
| 2253 |
+
"validated_this_epoch": true,
|
| 2254 |
+
"val_loss": 0.917495120140357,
|
| 2255 |
+
"val_dice": 0.6232558047709208,
|
| 2256 |
+
"val_iou": 0.5304130403921805,
|
| 2257 |
+
"val_reward": -0.00020798711739189457,
|
| 2258 |
+
"val_entropy": 0.5088435399411906,
|
| 2259 |
+
"val_actor_loss": -0.017258266482097707,
|
| 2260 |
+
"val_critic_loss": 0.00010929314296872965,
|
| 2261 |
+
"val_ce_loss": 0.9092376711025758,
|
| 2262 |
+
"val_dice_loss": 0.9601597731763667
|
| 2263 |
+
},
|
| 2264 |
+
{
|
| 2265 |
+
"epoch": 88,
|
| 2266 |
+
"train_loss": 0.24987675250919822,
|
| 2267 |
+
"train_actor_loss": -0.1773699210622372,
|
| 2268 |
+
"train_critic_loss": 0.00014217863769935764,
|
| 2269 |
+
"train_mean_reward": 0.0004272598358418177,
|
| 2270 |
+
"train_entropy": 0.4890517294406891,
|
| 2271 |
+
"train_ce_loss": 0.4721373919373712,
|
| 2272 |
+
"train_dice_loss": 0.38221376866866386,
|
| 2273 |
+
"train_dice": 0.8801531651440788,
|
| 2274 |
+
"train_iou": 0.7920066252642987,
|
| 2275 |
+
"grad_norm": 4.4149839878082275,
|
| 2276 |
+
"lr": 7.987239654118994e-05,
|
| 2277 |
+
"encoder_lr": 1.6671233840616128e-05,
|
| 2278 |
+
"alpha": 0.3591207265853882,
|
| 2279 |
+
"validated_this_epoch": true,
|
| 2280 |
+
"val_loss": 0.9017015482496583,
|
| 2281 |
+
"val_dice": 0.620972061598477,
|
| 2282 |
+
"val_iou": 0.5268228133108221,
|
| 2283 |
+
"val_reward": 2.0260321678212468e-05,
|
| 2284 |
+
"val_entropy": 0.47483350381706696,
|
| 2285 |
+
"val_actor_loss": -0.02071889122769312,
|
| 2286 |
+
"val_critic_loss": 0.00014345834822483764,
|
| 2287 |
+
"val_ce_loss": 0.8763936660769912,
|
| 2288 |
+
"val_dice_loss": 0.968303723467721
|
| 2289 |
+
},
|
| 2290 |
+
{
|
| 2291 |
+
"epoch": 89,
|
| 2292 |
+
"train_loss": 0.25153620310056096,
|
| 2293 |
+
"train_actor_loss": -0.16806260969203254,
|
| 2294 |
+
"train_critic_loss": 0.0001429730819976472,
|
| 2295 |
+
"train_mean_reward": 0.0005251484444903699,
|
| 2296 |
+
"train_entropy": 0.46885893360162395,
|
| 2297 |
+
"train_ce_loss": 0.42593227754323143,
|
| 2298 |
+
"train_dice_loss": 0.4131223689287137,
|
| 2299 |
+
"train_dice": 0.8522295400801622,
|
| 2300 |
+
"train_iou": 0.7601515339294863,
|
| 2301 |
+
"grad_norm": 4.187319113658025,
|
| 2302 |
+
"lr": 6.88236348905457e-05,
|
| 2303 |
+
"encoder_lr": 1.5616326949348587e-05,
|
| 2304 |
+
"alpha": 0.35794714093208313,
|
| 2305 |
+
"validated_this_epoch": true,
|
| 2306 |
+
"val_loss": 0.9150632559533485,
|
| 2307 |
+
"val_dice": 0.6191683857012373,
|
| 2308 |
+
"val_iou": 0.5246207877519159,
|
| 2309 |
+
"val_reward": -0.00032087383426022086,
|
| 2310 |
+
"val_entropy": 0.4876164578729205,
|
| 2311 |
+
"val_actor_loss": -0.01864965163604027,
|
| 2312 |
+
"val_critic_loss": 0.00011723540551015917,
|
| 2313 |
+
"val_ce_loss": 0.8899840491052454,
|
| 2314 |
+
"val_dice_loss": 0.9773244972180839
|
| 2315 |
+
},
|
| 2316 |
+
{
|
| 2317 |
+
"epoch": 90,
|
| 2318 |
+
"train_loss": 0.24874367007914072,
|
| 2319 |
+
"train_actor_loss": -0.17659713332660687,
|
| 2320 |
+
"train_critic_loss": 0.00014121002270537877,
|
| 2321 |
+
"train_mean_reward": 0.00027989791847059497,
|
| 2322 |
+
"train_entropy": 0.49293287518696915,
|
| 2323 |
+
"train_ce_loss": 0.46498161166882473,
|
| 2324 |
+
"train_dice_loss": 0.3855587778947292,
|
| 2325 |
+
"train_dice": 0.8559814821563515,
|
| 2326 |
+
"train_iou": 0.7688565351361153,
|
| 2327 |
+
"grad_norm": 3.9449609838999233,
|
| 2328 |
+
"lr": 5.869876628632226e-05,
|
| 2329 |
+
"encoder_lr": 1.4649630951960419e-05,
|
| 2330 |
+
"alpha": 0.35671666264533997,
|
| 2331 |
+
"validated_this_epoch": true,
|
| 2332 |
+
"val_loss": 0.8983689219493306,
|
| 2333 |
+
"val_dice": 0.6223633261591301,
|
| 2334 |
+
"val_iou": 0.5269890712194824,
|
| 2335 |
+
"val_reward": -0.0004190053940957032,
|
| 2336 |
+
"val_entropy": 0.48030781053533456,
|
| 2337 |
+
"val_actor_loss": -0.02415347856297769,
|
| 2338 |
+
"val_critic_loss": 0.00016189747042737895,
|
| 2339 |
+
"val_ce_loss": 0.8936010239965686,
|
| 2340 |
+
"val_dice_loss": 0.9512818527944159
|
| 2341 |
+
},
|
| 2342 |
+
{
|
| 2343 |
+
"epoch": 91,
|
| 2344 |
+
"train_loss": 0.29253113590395796,
|
| 2345 |
+
"train_actor_loss": -0.1638823019292874,
|
| 2346 |
+
"train_critic_loss": 0.00015430227966088496,
|
| 2347 |
+
"train_mean_reward": 0.000467800049323276,
|
| 2348 |
+
"train_entropy": 0.4655343200533817,
|
| 2349 |
+
"train_ce_loss": 0.460670220402953,
|
| 2350 |
+
"train_dice_loss": 0.4520023423127639,
|
| 2351 |
+
"train_dice": 0.8488439481467641,
|
| 2352 |
+
"train_iou": 0.7608491053019133,
|
| 2353 |
+
"grad_norm": 5.247777067697966,
|
| 2354 |
+
"lr": 4.9507782751441774e-05,
|
| 2355 |
+
"encoder_lr": 1.3772099860690422e-05,
|
| 2356 |
+
"alpha": 0.35555821657180786,
|
| 2357 |
+
"validated_this_epoch": true,
|
| 2358 |
+
"val_loss": 0.9087012919286886,
|
| 2359 |
+
"val_dice": 0.6211960948109294,
|
| 2360 |
+
"val_iou": 0.5258964740923058,
|
| 2361 |
+
"val_reward": -0.000339452321831498,
|
| 2362 |
+
"val_entropy": 0.49580008890291666,
|
| 2363 |
+
"val_actor_loss": -0.020528437017058632,
|
| 2364 |
+
"val_critic_loss": 0.00013429034452477055,
|
| 2365 |
+
"val_ce_loss": 0.8962014911301208,
|
| 2366 |
+
"val_dice_loss": 0.9621236465796077
|
| 2367 |
+
},
|
| 2368 |
+
{
|
| 2369 |
+
"epoch": 92,
|
| 2370 |
+
"train_loss": 0.2889437225038329,
|
| 2371 |
+
"train_actor_loss": -0.1695738530311829,
|
| 2372 |
+
"train_critic_loss": 0.0001535360813063446,
|
| 2373 |
+
"train_mean_reward": 0.0005853210439990783,
|
| 2374 |
+
"train_entropy": 0.48432859453635335,
|
| 2375 |
+
"train_ce_loss": 0.46360836019261903,
|
| 2376 |
+
"train_dice_loss": 0.4532732413365291,
|
| 2377 |
+
"train_dice": 0.8377785514748652,
|
| 2378 |
+
"train_iou": 0.744639892941956,
|
| 2379 |
+
"grad_norm": 4.666837495106917,
|
| 2380 |
+
"lr": 4.12597546770122e-05,
|
| 2381 |
+
"encoder_lr": 1.298459969278006e-05,
|
| 2382 |
+
"alpha": 0.3543606996536255,
|
| 2383 |
+
"validated_this_epoch": true,
|
| 2384 |
+
"val_loss": 0.9132339483493883,
|
| 2385 |
+
"val_dice": 0.6219718835226969,
|
| 2386 |
+
"val_iou": 0.5261243123829716,
|
| 2387 |
+
"val_reward": -0.00034418005540801187,
|
| 2388 |
+
"val_entropy": 0.4817650560477767,
|
| 2389 |
+
"val_actor_loss": -0.02135762275958603,
|
| 2390 |
+
"val_critic_loss": 0.000157085660657226,
|
| 2391 |
+
"val_ce_loss": 0.9143161262394954,
|
| 2392 |
+
"val_dice_loss": 0.9547099061686583
|
| 2393 |
+
},
|
| 2394 |
+
{
|
| 2395 |
+
"epoch": 93,
|
| 2396 |
+
"train_loss": 0.32597759312305313,
|
| 2397 |
+
"train_actor_loss": -0.17009864226938823,
|
| 2398 |
+
"train_critic_loss": 0.0001761931712863794,
|
| 2399 |
+
"train_mean_reward": 0.00032547591601355514,
|
| 2400 |
+
"train_entropy": 0.4829088883140148,
|
| 2401 |
+
"train_ce_loss": 0.47624832000105816,
|
| 2402 |
+
"train_dice_loss": 0.5157279490660399,
|
| 2403 |
+
"train_dice": 0.8263353944754636,
|
| 2404 |
+
"train_iou": 0.7401983743915522,
|
| 2405 |
+
"grad_norm": 12.099165925612816,
|
| 2406 |
+
"lr": 3.396282187094642e-05,
|
| 2407 |
+
"encoder_lr": 1.2287907615819006e-05,
|
| 2408 |
+
"alpha": 0.35316231846809387,
|
| 2409 |
+
"validated_this_epoch": true,
|
| 2410 |
+
"val_loss": 0.9017766702666201,
|
| 2411 |
+
"val_dice": 0.6232273524179136,
|
| 2412 |
+
"val_iou": 0.5275790346266787,
|
| 2413 |
+
"val_reward": -0.00024246856900381224,
|
| 2414 |
+
"val_entropy": 0.4884287800752755,
|
| 2415 |
+
"val_actor_loss": -0.022333643211033007,
|
| 2416 |
+
"val_critic_loss": 0.00015568893016307177,
|
| 2417 |
+
"val_ce_loss": 0.8919454128622555,
|
| 2418 |
+
"val_dice_loss": 0.9561194937036495
|
| 2419 |
+
},
|
| 2420 |
+
{
|
| 2421 |
+
"epoch": 94,
|
| 2422 |
+
"train_loss": 0.27801704385246223,
|
| 2423 |
+
"train_actor_loss": -0.16949972591530055,
|
| 2424 |
+
"train_critic_loss": 0.00018177391320233044,
|
| 2425 |
+
"train_mean_reward": 0.0007979085432091383,
|
| 2426 |
+
"train_entropy": 0.4796367125251354,
|
| 2427 |
+
"train_ce_loss": 0.4563503716952908,
|
| 2428 |
+
"train_dice_loss": 0.4385013866883058,
|
| 2429 |
+
"train_dice": 0.8549340972129036,
|
| 2430 |
+
"train_iou": 0.7649991261024102,
|
| 2431 |
+
"grad_norm": 5.448335801179592,
|
| 2432 |
+
"lr": 2.762418552495473e-05,
|
| 2433 |
+
"encoder_lr": 1.1682711180774574e-05,
|
| 2434 |
+
"alpha": 0.3519377112388611,
|
| 2435 |
+
"validated_this_epoch": true,
|
| 2436 |
+
"val_loss": 0.9016437292536439,
|
| 2437 |
+
"val_dice": 0.6181031033605708,
|
| 2438 |
+
"val_iou": 0.5223249313558894,
|
| 2439 |
+
"val_reward": -0.00036827485914589926,
|
| 2440 |
+
"val_entropy": 0.4803736521439118,
|
| 2441 |
+
"val_actor_loss": -0.02343166409168543,
|
| 2442 |
+
"val_critic_loss": 0.00015151340910115967,
|
| 2443 |
+
"val_ce_loss": 0.8847880632560137,
|
| 2444 |
+
"val_dice_loss": 0.9652111786182481
|
| 2445 |
+
},
|
| 2446 |
+
{
|
| 2447 |
+
"epoch": 95,
|
| 2448 |
+
"train_loss": 0.2758521941880015,
|
| 2449 |
+
"train_actor_loss": -0.16630445282237652,
|
| 2450 |
+
"train_critic_loss": 0.00016334877358158102,
|
| 2451 |
+
"train_mean_reward": 0.00048493467471789935,
|
| 2452 |
+
"train_entropy": 0.4720621172052163,
|
| 2453 |
+
"train_ce_loss": 0.4678195087621227,
|
| 2454 |
+
"train_dice_loss": 0.41633042502097595,
|
| 2455 |
+
"train_dice": 0.8439491163281834,
|
| 2456 |
+
"train_iou": 0.750430192871421,
|
| 2457 |
+
"grad_norm": 7.460448833612295,
|
| 2458 |
+
"lr": 2.2250101107838036e-05,
|
| 2459 |
+
"encoder_lr": 1.1169607643461924e-05,
|
| 2460 |
+
"alpha": 0.35076025128364563,
|
| 2461 |
+
"validated_this_epoch": true,
|
| 2462 |
+
"val_loss": 0.9047963875345886,
|
| 2463 |
+
"val_dice": 0.6226404472512873,
|
| 2464 |
+
"val_iou": 0.5261662850200162,
|
| 2465 |
+
"val_reward": -0.00034857219515922196,
|
| 2466 |
+
"val_entropy": 0.48771001759803656,
|
| 2467 |
+
"val_actor_loss": -0.02373351020746714,
|
| 2468 |
+
"val_critic_loss": 0.000172863770181592,
|
| 2469 |
+
"val_ce_loss": 0.9056161659854379,
|
| 2470 |
+
"val_dice_loss": 0.9512707341198969
|
| 2471 |
+
},
|
| 2472 |
+
{
|
| 2473 |
+
"epoch": 96,
|
| 2474 |
+
"train_loss": 0.3537652181962935,
|
| 2475 |
+
"train_actor_loss": -0.1597088465037254,
|
| 2476 |
+
"train_critic_loss": 0.00015790711363595808,
|
| 2477 |
+
"train_mean_reward": 0.0002683689670984444,
|
| 2478 |
+
"train_entropy": 0.4696554938952128,
|
| 2479 |
+
"train_ce_loss": 0.5159649394022732,
|
| 2480 |
+
"train_dice_loss": 0.5108252737002495,
|
| 2481 |
+
"train_dice": 0.8125867229481224,
|
| 2482 |
+
"train_iou": 0.7189560783274888,
|
| 2483 |
+
"grad_norm": 7.1554928330274725,
|
| 2484 |
+
"lr": 1.784587219209463e-05,
|
| 2485 |
+
"encoder_lr": 1.0749103375124614e-05,
|
| 2486 |
+
"alpha": 0.349577397108078,
|
| 2487 |
+
"validated_this_epoch": true,
|
| 2488 |
+
"val_loss": 0.912448946584129,
|
| 2489 |
+
"val_dice": 0.6213410996597047,
|
| 2490 |
+
"val_iou": 0.5255512600959567,
|
| 2491 |
+
"val_reward": -0.00022617800439345752,
|
| 2492 |
+
"val_entropy": 0.48791983151676677,
|
| 2493 |
+
"val_actor_loss": -0.021414225823728063,
|
| 2494 |
+
"val_critic_loss": 0.0001534608482030087,
|
| 2495 |
+
"val_ce_loss": 0.9078516693164905,
|
| 2496 |
+
"val_dice_loss": 0.9597211808267265
|
| 2497 |
+
},
|
| 2498 |
+
{
|
| 2499 |
+
"epoch": 97,
|
| 2500 |
+
"train_loss": 0.2663845607939248,
|
| 2501 |
+
"train_actor_loss": -0.16546016582884848,
|
| 2502 |
+
"train_critic_loss": 0.00015433574538129062,
|
| 2503 |
+
"train_mean_reward": 0.0006431382953917656,
|
| 2504 |
+
"train_entropy": 0.477576905527176,
|
| 2505 |
+
"train_ce_loss": 0.44166308335768867,
|
| 2506 |
+
"train_dice_loss": 0.42187202511689603,
|
| 2507 |
+
"train_dice": 0.8600562320036047,
|
| 2508 |
+
"train_iou": 0.7638544375989952,
|
| 2509 |
+
"grad_norm": 4.410502273302812,
|
| 2510 |
+
"lr": 1.4415845219935406e-05,
|
| 2511 |
+
"encoder_lr": 1.0421613362707402e-05,
|
| 2512 |
+
"alpha": 0.34842491149902344,
|
| 2513 |
+
"validated_this_epoch": true,
|
| 2514 |
+
"val_loss": 0.901356447064741,
|
| 2515 |
+
"val_dice": 0.6198718321754444,
|
| 2516 |
+
"val_iou": 0.5257802922953267,
|
| 2517 |
+
"val_reward": -0.00019404460633240612,
|
| 2518 |
+
"val_entropy": 0.4758102659023169,
|
| 2519 |
+
"val_actor_loss": -0.021779958457413196,
|
| 2520 |
+
"val_critic_loss": 0.0001397187738435568,
|
| 2521 |
+
"val_ce_loss": 0.8775688060418224,
|
| 2522 |
+
"val_dice_loss": 0.9685642535638327
|
| 2523 |
+
},
|
| 2524 |
+
{
|
| 2525 |
+
"epoch": 98,
|
| 2526 |
+
"train_loss": 0.2550379827690239,
|
| 2527 |
+
"train_actor_loss": -0.16464740324478883,
|
| 2528 |
+
"train_critic_loss": 0.0001463998056025071,
|
| 2529 |
+
"train_mean_reward": 0.00034205407473564264,
|
| 2530 |
+
"train_entropy": 0.469524044639025,
|
| 2531 |
+
"train_ce_loss": 0.4372696820288323,
|
| 2532 |
+
"train_dice_loss": 0.4019546856482823,
|
| 2533 |
+
"train_dice": 0.859613426175772,
|
| 2534 |
+
"train_iou": 0.7663299479320937,
|
| 2535 |
+
"grad_norm": 5.338865073827597,
|
| 2536 |
+
"lr": 1.19634052138698e-05,
|
| 2537 |
+
"encoder_lr": 1.0187460799314204e-05,
|
| 2538 |
+
"alpha": 0.3473333418369293,
|
| 2539 |
+
"validated_this_epoch": true,
|
| 2540 |
+
"val_loss": 0.8939954183685282,
|
| 2541 |
+
"val_dice": 0.6188515578342881,
|
| 2542 |
+
"val_iou": 0.5258565427681711,
|
| 2543 |
+
"val_reward": -0.0001192783682640628,
|
| 2544 |
+
"val_entropy": 0.46963536513574194,
|
| 2545 |
+
"val_actor_loss": -0.022646716352983733,
|
| 2546 |
+
"val_critic_loss": 0.00015357221043887703,
|
| 2547 |
+
"val_ce_loss": 0.8651182216055918,
|
| 2548 |
+
"val_dice_loss": 0.9680124433955761
|
| 2549 |
+
},
|
| 2550 |
+
{
|
| 2551 |
+
"epoch": 99,
|
| 2552 |
+
"train_loss": 0.27821797961727357,
|
| 2553 |
+
"train_actor_loss": -0.15696958376046938,
|
| 2554 |
+
"train_critic_loss": 0.00014209586459820947,
|
| 2555 |
+
"train_mean_reward": 0.00021825241767659556,
|
| 2556 |
+
"train_entropy": 0.45981486370930313,
|
| 2557 |
+
"train_ce_loss": 0.4510410614908697,
|
| 2558 |
+
"train_dice_loss": 0.41919195995881,
|
| 2559 |
+
"train_dice": 0.8453442080348146,
|
| 2560 |
+
"train_iou": 0.7490223041352104,
|
| 2561 |
+
"grad_norm": 5.545630202843593,
|
| 2562 |
+
"lr": 1.049097243609706e-05,
|
| 2563 |
+
"encoder_lr": 1.0046876765255499e-05,
|
| 2564 |
+
"alpha": 0.3462064266204834,
|
| 2565 |
+
"validated_this_epoch": true,
|
| 2566 |
+
"val_loss": 0.8949461424525714,
|
| 2567 |
+
"val_dice": 0.6196340764437087,
|
| 2568 |
+
"val_iou": 0.5255957605597994,
|
| 2569 |
+
"val_reward": -0.0001519040910205498,
|
| 2570 |
+
"val_entropy": 0.47533140941099683,
|
| 2571 |
+
"val_actor_loss": -0.022427125848039537,
|
| 2572 |
+
"val_critic_loss": 0.00015188496594088394,
|
| 2573 |
+
"val_ce_loss": 0.8710821259371712,
|
| 2574 |
+
"val_dice_loss": 0.9635124941064855
|
| 2575 |
+
},
|
| 2576 |
+
{
|
| 2577 |
+
"epoch": 100,
|
| 2578 |
+
"train_loss": 0.29026232281019193,
|
| 2579 |
+
"train_actor_loss": -0.16318322627399212,
|
| 2580 |
+
"train_critic_loss": 0.00015787492871729622,
|
| 2581 |
+
"train_mean_reward": 0.0004404639182025559,
|
| 2582 |
+
"train_entropy": 0.46886270493268967,
|
| 2583 |
+
"train_ce_loss": 0.45911439769686413,
|
| 2584 |
+
"train_dice_loss": 0.4476188130867787,
|
| 2585 |
+
"train_dice": 0.8479103537727621,
|
| 2586 |
+
"train_iou": 0.7541419388593229,
|
| 2587 |
+
"grad_norm": 6.763477334609399,
|
| 2588 |
+
"lr": 1e-05,
|
| 2589 |
+
"encoder_lr": 1e-05,
|
| 2590 |
+
"alpha": 0.34506845474243164,
|
| 2591 |
+
"validated_this_epoch": true,
|
| 2592 |
+
"val_loss": 0.9081789445075573,
|
| 2593 |
+
"val_dice": 0.6166532046493403,
|
| 2594 |
+
"val_iou": 0.5223674392731117,
|
| 2595 |
+
"val_reward": -0.00024403816658175647,
|
| 2596 |
+
"val_entropy": 0.48237108762818154,
|
| 2597 |
+
"val_actor_loss": -0.022036860948616656,
|
| 2598 |
+
"val_critic_loss": 0.00015275699925868221,
|
| 2599 |
+
"val_ce_loss": 0.89091880452076,
|
| 2600 |
+
"val_dice_loss": 0.9693600258441886
|
| 2601 |
+
}
|
| 2602 |
+
]
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/low_iou_samples.csv
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
sample_id
|
| 2 |
+
BUSI_0202_B
|
| 3 |
+
BUSI_0194_M
|
| 4 |
+
BUSI_0251_B
|
| 5 |
+
BUSI_0163_B
|
| 6 |
+
BUSI_0116_M
|
| 7 |
+
BUSI_0104_B
|
| 8 |
+
BUSI_0270_B
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/run_config.json
ADDED
|
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"project_dir": "/workspace",
|
| 3 |
+
"data_root": "/workspace/BUSI_with_classes",
|
| 4 |
+
"run_type": "final",
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.1,
|
| 7 |
+
"dataset_name": "BUSI_with_classes",
|
| 8 |
+
"dataset_split_policy": "stratified",
|
| 9 |
+
"dataset_splits_path": "/workspace/dataset_splits_busi_with_classes_stratified.json",
|
| 10 |
+
"split_type": "80_10_10",
|
| 11 |
+
"train_subset_key": "10",
|
| 12 |
+
"normalization_cache_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/norm_stats_BUSI_with_classes_stratified_80_10_10_10pct.json",
|
| 13 |
+
"head_lr": 0.002,
|
| 14 |
+
"encoder_lr": 0.0002,
|
| 15 |
+
"weight_decay": 0.002,
|
| 16 |
+
"dropout_p": 0.3,
|
| 17 |
+
"tmax": 6,
|
| 18 |
+
"entropy_lr": 0.0002,
|
| 19 |
+
"max_epochs": 100,
|
| 20 |
+
"gamma": 0.95,
|
| 21 |
+
"grad_clip_norm": 6.0,
|
| 22 |
+
"scheduler_factor": 0.5,
|
| 23 |
+
"scheduler_patience": 5,
|
| 24 |
+
"scheduler_threshold": 0.001,
|
| 25 |
+
"scheduler_min_lr": 1e-05,
|
| 26 |
+
"execution_mode": "train_eval",
|
| 27 |
+
"evaluation_checkpoint_mode": "latest",
|
| 28 |
+
"strategy2_checkpoint_mode": "specific",
|
| 29 |
+
"train_resume_mode": "off",
|
| 30 |
+
"train_resume_specific_checkpoint": "",
|
| 31 |
+
"backbone_family": "smp",
|
| 32 |
+
"smp_encoder_name": "efficientnet-b0",
|
| 33 |
+
"smp_encoder_weights": "imagenet",
|
| 34 |
+
"smp_encoder_depth": 5,
|
| 35 |
+
"smp_encoder_proj_dim": 128,
|
| 36 |
+
"smp_decoder_type": "Unet",
|
| 37 |
+
"vgg_feature_scales": 4,
|
| 38 |
+
"vgg_feature_dilation": 1,
|
| 39 |
+
"strategy2_checkpoint_path": "/workspace/10%_Strategy2_latest.pt"
|
| 40 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_10/strategy_3/final/summary.json
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_val_iou": 0.5325012422437716,
|
| 3 |
+
"best_val_dice": 0.628116376617652,
|
| 4 |
+
"final_epoch": 100,
|
| 5 |
+
"elapsed_seconds": 645.5083975791931,
|
| 6 |
+
"seconds_per_epoch": 6.455083975791931,
|
| 7 |
+
"device_used": "cuda",
|
| 8 |
+
"strategy": 3,
|
| 9 |
+
"run_type": "final",
|
| 10 |
+
"resumed": false
|
| 11 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_100/norm_stats_BUSI_with_classes_stratified_80_10_10_100pct.json
ADDED
|
@@ -0,0 +1,525 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"global_mean": 84.1731207210797,
|
| 3 |
+
"global_std": 56.91848196862734,
|
| 4 |
+
"total_pixels": 479104653,
|
| 5 |
+
"num_images": 517,
|
| 6 |
+
"filenames": [
|
| 7 |
+
"BUSI_0133_B.png",
|
| 8 |
+
"BUSI_0154_M.png",
|
| 9 |
+
"BUSI_0293_B.png",
|
| 10 |
+
"BUSI_0366_B.png",
|
| 11 |
+
"BUSI_0007_B.png",
|
| 12 |
+
"BUSI_0138_M.png",
|
| 13 |
+
"BUSI_0359_B.png",
|
| 14 |
+
"BUSI_0244_B.png",
|
| 15 |
+
"BUSI_0418_B.png",
|
| 16 |
+
"BUSI_0170_M.png",
|
| 17 |
+
"BUSI_0119_M.png",
|
| 18 |
+
"BUSI_0048_B.png",
|
| 19 |
+
"BUSI_0037_M.png",
|
| 20 |
+
"BUSI_0312_B.png",
|
| 21 |
+
"BUSI_0109_B.png",
|
| 22 |
+
"BUSI_0274_B.png",
|
| 23 |
+
"BUSI_0100_B.png",
|
| 24 |
+
"BUSI_0175_M.png",
|
| 25 |
+
"BUSI_0044_B.png",
|
| 26 |
+
"BUSI_0413_B.png",
|
| 27 |
+
"BUSI_0024_B.png",
|
| 28 |
+
"BUSI_0365_B.png",
|
| 29 |
+
"BUSI_0279_B.png",
|
| 30 |
+
"BUSI_0167_B.png",
|
| 31 |
+
"BUSI_0072_B.png",
|
| 32 |
+
"BUSI_0303_B.png",
|
| 33 |
+
"BUSI_0286_B.png",
|
| 34 |
+
"BUSI_0266_B.png",
|
| 35 |
+
"BUSI_0159_B.png",
|
| 36 |
+
"BUSI_0432_B.png",
|
| 37 |
+
"BUSI_0122_B.png",
|
| 38 |
+
"BUSI_0267_B.png",
|
| 39 |
+
"BUSI_0208_M.png",
|
| 40 |
+
"BUSI_0220_B.png",
|
| 41 |
+
"BUSI_0431_B.png",
|
| 42 |
+
"BUSI_0333_B.png",
|
| 43 |
+
"BUSI_0394_B.png",
|
| 44 |
+
"BUSI_0297_B.png",
|
| 45 |
+
"BUSI_0081_M.png",
|
| 46 |
+
"BUSI_0255_B.png",
|
| 47 |
+
"BUSI_0380_B.png",
|
| 48 |
+
"BUSI_0273_B.png",
|
| 49 |
+
"BUSI_0341_B.png",
|
| 50 |
+
"BUSI_0301_B.png",
|
| 51 |
+
"BUSI_0192_B.png",
|
| 52 |
+
"BUSI_0132_B.png",
|
| 53 |
+
"BUSI_0117_B.png",
|
| 54 |
+
"BUSI_0075_M.png",
|
| 55 |
+
"BUSI_0280_B.png",
|
| 56 |
+
"BUSI_0103_M.png",
|
| 57 |
+
"BUSI_0087_B.png",
|
| 58 |
+
"BUSI_0028_M.png",
|
| 59 |
+
"BUSI_0083_B.png",
|
| 60 |
+
"BUSI_0093_B.png",
|
| 61 |
+
"BUSI_0284_B.png",
|
| 62 |
+
"BUSI_0203_B.png",
|
| 63 |
+
"BUSI_0033_M.png",
|
| 64 |
+
"BUSI_0315_B.png",
|
| 65 |
+
"BUSI_0186_B.png",
|
| 66 |
+
"BUSI_0017_B.png",
|
| 67 |
+
"BUSI_0134_B.png",
|
| 68 |
+
"BUSI_0092_M.png",
|
| 69 |
+
"BUSI_0136_M.png",
|
| 70 |
+
"BUSI_0181_B.png",
|
| 71 |
+
"BUSI_0218_B.png",
|
| 72 |
+
"BUSI_0049_B.png",
|
| 73 |
+
"BUSI_0420_B.png",
|
| 74 |
+
"BUSI_0101_B.png",
|
| 75 |
+
"BUSI_0425_B.png",
|
| 76 |
+
"BUSI_0265_B.png",
|
| 77 |
+
"BUSI_0163_M.png",
|
| 78 |
+
"BUSI_0156_M.png",
|
| 79 |
+
"BUSI_0192_M.png",
|
| 80 |
+
"BUSI_0046_M.png",
|
| 81 |
+
"BUSI_0166_B.png",
|
| 82 |
+
"BUSI_0019_B.png",
|
| 83 |
+
"BUSI_0436_B.png",
|
| 84 |
+
"BUSI_0028_B.png",
|
| 85 |
+
"BUSI_0077_B.png",
|
| 86 |
+
"BUSI_0242_B.png",
|
| 87 |
+
"BUSI_0018_B.png",
|
| 88 |
+
"BUSI_0070_B.png",
|
| 89 |
+
"BUSI_0135_B.png",
|
| 90 |
+
"BUSI_0019_M.png",
|
| 91 |
+
"BUSI_0176_B.png",
|
| 92 |
+
"BUSI_0403_B.png",
|
| 93 |
+
"BUSI_0147_M.png",
|
| 94 |
+
"BUSI_0424_B.png",
|
| 95 |
+
"BUSI_0122_M.png",
|
| 96 |
+
"BUSI_0126_M.png",
|
| 97 |
+
"BUSI_0162_B.png",
|
| 98 |
+
"BUSI_0206_B.png",
|
| 99 |
+
"BUSI_0200_B.png",
|
| 100 |
+
"BUSI_0349_B.png",
|
| 101 |
+
"BUSI_0371_B.png",
|
| 102 |
+
"BUSI_0150_M.png",
|
| 103 |
+
"BUSI_0348_B.png",
|
| 104 |
+
"BUSI_0031_B.png",
|
| 105 |
+
"BUSI_0155_M.png",
|
| 106 |
+
"BUSI_0253_B.png",
|
| 107 |
+
"BUSI_0036_M.png",
|
| 108 |
+
"BUSI_0249_B.png",
|
| 109 |
+
"BUSI_0188_B.png",
|
| 110 |
+
"BUSI_0193_M.png",
|
| 111 |
+
"BUSI_0038_M.png",
|
| 112 |
+
"BUSI_0096_M.png",
|
| 113 |
+
"BUSI_0009_B.png",
|
| 114 |
+
"BUSI_0145_B.png",
|
| 115 |
+
"BUSI_0152_B.png",
|
| 116 |
+
"BUSI_0012_B.png",
|
| 117 |
+
"BUSI_0113_B.png",
|
| 118 |
+
"BUSI_0338_B.png",
|
| 119 |
+
"BUSI_0313_B.png",
|
| 120 |
+
"BUSI_0409_B.png",
|
| 121 |
+
"BUSI_0087_M.png",
|
| 122 |
+
"BUSI_0164_M.png",
|
| 123 |
+
"BUSI_0063_B.png",
|
| 124 |
+
"BUSI_0385_B.png",
|
| 125 |
+
"BUSI_0288_B.png",
|
| 126 |
+
"BUSI_0332_B.png",
|
| 127 |
+
"BUSI_0134_M.png",
|
| 128 |
+
"BUSI_0003_B.png",
|
| 129 |
+
"BUSI_0161_B.png",
|
| 130 |
+
"BUSI_0084_B.png",
|
| 131 |
+
"BUSI_0081_B.png",
|
| 132 |
+
"BUSI_0310_B.png",
|
| 133 |
+
"BUSI_0035_M.png",
|
| 134 |
+
"BUSI_0322_B.png",
|
| 135 |
+
"BUSI_0098_M.png",
|
| 136 |
+
"BUSI_0058_B.png",
|
| 137 |
+
"BUSI_0069_M.png",
|
| 138 |
+
"BUSI_0046_B.png",
|
| 139 |
+
"BUSI_0040_M.png",
|
| 140 |
+
"BUSI_0360_B.png",
|
| 141 |
+
"BUSI_0105_B.png",
|
| 142 |
+
"BUSI_0435_B.png",
|
| 143 |
+
"BUSI_0175_B.png",
|
| 144 |
+
"BUSI_0160_B.png",
|
| 145 |
+
"BUSI_0168_B.png",
|
| 146 |
+
"BUSI_0137_B.png",
|
| 147 |
+
"BUSI_0305_B.png",
|
| 148 |
+
"BUSI_0101_M.png",
|
| 149 |
+
"BUSI_0275_B.png",
|
| 150 |
+
"BUSI_0262_B.png",
|
| 151 |
+
"BUSI_0382_B.png",
|
| 152 |
+
"BUSI_0130_M.png",
|
| 153 |
+
"BUSI_0027_B.png",
|
| 154 |
+
"BUSI_0126_B.png",
|
| 155 |
+
"BUSI_0074_B.png",
|
| 156 |
+
"BUSI_0271_B.png",
|
| 157 |
+
"BUSI_0120_B.png",
|
| 158 |
+
"BUSI_0169_M.png",
|
| 159 |
+
"BUSI_0077_M.png",
|
| 160 |
+
"BUSI_0190_B.png",
|
| 161 |
+
"BUSI_0113_M.png",
|
| 162 |
+
"BUSI_0417_B.png",
|
| 163 |
+
"BUSI_0076_M.png",
|
| 164 |
+
"BUSI_0183_M.png",
|
| 165 |
+
"BUSI_0183_B.png",
|
| 166 |
+
"BUSI_0032_B.png",
|
| 167 |
+
"BUSI_0155_B.png",
|
| 168 |
+
"BUSI_0363_B.png",
|
| 169 |
+
"BUSI_0198_M.png",
|
| 170 |
+
"BUSI_0031_M.png",
|
| 171 |
+
"BUSI_0160_M.png",
|
| 172 |
+
"BUSI_0396_B.png",
|
| 173 |
+
"BUSI_0093_M.png",
|
| 174 |
+
"BUSI_0140_B.png",
|
| 175 |
+
"BUSI_0022_B.png",
|
| 176 |
+
"BUSI_0189_B.png",
|
| 177 |
+
"BUSI_0264_B.png",
|
| 178 |
+
"BUSI_0036_B.png",
|
| 179 |
+
"BUSI_0012_M.png",
|
| 180 |
+
"BUSI_0088_M.png",
|
| 181 |
+
"BUSI_0094_B.png",
|
| 182 |
+
"BUSI_0074_M.png",
|
| 183 |
+
"BUSI_0120_M.png",
|
| 184 |
+
"BUSI_0405_B.png",
|
| 185 |
+
"BUSI_0174_M.png",
|
| 186 |
+
"BUSI_0225_B.png",
|
| 187 |
+
"BUSI_0191_B.png",
|
| 188 |
+
"BUSI_0070_M.png",
|
| 189 |
+
"BUSI_0373_B.png",
|
| 190 |
+
"BUSI_0223_B.png",
|
| 191 |
+
"BUSI_0061_M.png",
|
| 192 |
+
"BUSI_0296_B.png",
|
| 193 |
+
"BUSI_0300_B.png",
|
| 194 |
+
"BUSI_0047_B.png",
|
| 195 |
+
"BUSI_0060_B.png",
|
| 196 |
+
"BUSI_0168_M.png",
|
| 197 |
+
"BUSI_0243_B.png",
|
| 198 |
+
"BUSI_0159_M.png",
|
| 199 |
+
"BUSI_0072_M.png",
|
| 200 |
+
"BUSI_0006_B.png",
|
| 201 |
+
"BUSI_0182_M.png",
|
| 202 |
+
"BUSI_0107_M.png",
|
| 203 |
+
"BUSI_0381_B.png",
|
| 204 |
+
"BUSI_0042_M.png",
|
| 205 |
+
"BUSI_0011_B.png",
|
| 206 |
+
"BUSI_0330_B.png",
|
| 207 |
+
"BUSI_0323_B.png",
|
| 208 |
+
"BUSI_0339_B.png",
|
| 209 |
+
"BUSI_0158_B.png",
|
| 210 |
+
"BUSI_0091_M.png",
|
| 211 |
+
"BUSI_0030_B.png",
|
| 212 |
+
"BUSI_0056_B.png",
|
| 213 |
+
"BUSI_0090_M.png",
|
| 214 |
+
"BUSI_0148_B.png",
|
| 215 |
+
"BUSI_0343_B.png",
|
| 216 |
+
"BUSI_0186_M.png",
|
| 217 |
+
"BUSI_0153_B.png",
|
| 218 |
+
"BUSI_0127_B.png",
|
| 219 |
+
"BUSI_0213_B.png",
|
| 220 |
+
"BUSI_0367_B.png",
|
| 221 |
+
"BUSI_0200_M.png",
|
| 222 |
+
"BUSI_0258_B.png",
|
| 223 |
+
"BUSI_0317_B.png",
|
| 224 |
+
"BUSI_0065_M.png",
|
| 225 |
+
"BUSI_0433_B.png",
|
| 226 |
+
"BUSI_0426_B.png",
|
| 227 |
+
"BUSI_0421_B.png",
|
| 228 |
+
"BUSI_0131_B.png",
|
| 229 |
+
"BUSI_0389_B.png",
|
| 230 |
+
"BUSI_0171_B.png",
|
| 231 |
+
"BUSI_0096_B.png",
|
| 232 |
+
"BUSI_0079_M.png",
|
| 233 |
+
"BUSI_0404_B.png",
|
| 234 |
+
"BUSI_0222_B.png",
|
| 235 |
+
"BUSI_0022_M.png",
|
| 236 |
+
"BUSI_0145_M.png",
|
| 237 |
+
"BUSI_0118_M.png",
|
| 238 |
+
"BUSI_0203_M.png",
|
| 239 |
+
"BUSI_0388_B.png",
|
| 240 |
+
"BUSI_0384_B.png",
|
| 241 |
+
"BUSI_0086_M.png",
|
| 242 |
+
"BUSI_0241_B.png",
|
| 243 |
+
"BUSI_0053_M.png",
|
| 244 |
+
"BUSI_0103_B.png",
|
| 245 |
+
"BUSI_0068_M.png",
|
| 246 |
+
"BUSI_0199_B.png",
|
| 247 |
+
"BUSI_0076_B.png",
|
| 248 |
+
"BUSI_0196_B.png",
|
| 249 |
+
"BUSI_0344_B.png",
|
| 250 |
+
"BUSI_0257_B.png",
|
| 251 |
+
"BUSI_0428_B.png",
|
| 252 |
+
"BUSI_0043_B.png",
|
| 253 |
+
"BUSI_0136_B.png",
|
| 254 |
+
"BUSI_0194_B.png",
|
| 255 |
+
"BUSI_0197_B.png",
|
| 256 |
+
"BUSI_0390_B.png",
|
| 257 |
+
"BUSI_0106_M.png",
|
| 258 |
+
"BUSI_0094_M.png",
|
| 259 |
+
"BUSI_0309_B.png",
|
| 260 |
+
"BUSI_0085_B.png",
|
| 261 |
+
"BUSI_0182_B.png",
|
| 262 |
+
"BUSI_0167_M.png",
|
| 263 |
+
"BUSI_0035_B.png",
|
| 264 |
+
"BUSI_0232_B.png",
|
| 265 |
+
"BUSI_0170_B.png",
|
| 266 |
+
"BUSI_0139_B.png",
|
| 267 |
+
"BUSI_0052_B.png",
|
| 268 |
+
"BUSI_0248_B.png",
|
| 269 |
+
"BUSI_0260_B.png",
|
| 270 |
+
"BUSI_0008_M.png",
|
| 271 |
+
"BUSI_0017_M.png",
|
| 272 |
+
"BUSI_0060_M.png",
|
| 273 |
+
"BUSI_0111_M.png",
|
| 274 |
+
"BUSI_0146_M.png",
|
| 275 |
+
"BUSI_0135_M.png",
|
| 276 |
+
"BUSI_0252_B.png",
|
| 277 |
+
"BUSI_0247_B.png",
|
| 278 |
+
"BUSI_0066_M.png",
|
| 279 |
+
"BUSI_0234_B.png",
|
| 280 |
+
"BUSI_0276_B.png",
|
| 281 |
+
"BUSI_0246_B.png",
|
| 282 |
+
"BUSI_0277_B.png",
|
| 283 |
+
"BUSI_0227_B.png",
|
| 284 |
+
"BUSI_0195_B.png",
|
| 285 |
+
"BUSI_0221_B.png",
|
| 286 |
+
"BUSI_0158_M.png",
|
| 287 |
+
"BUSI_0177_M.png",
|
| 288 |
+
"BUSI_0346_B.png",
|
| 289 |
+
"BUSI_0085_M.png",
|
| 290 |
+
"BUSI_0347_B.png",
|
| 291 |
+
"BUSI_0238_B.png",
|
| 292 |
+
"BUSI_0178_B.png",
|
| 293 |
+
"BUSI_0079_B.png",
|
| 294 |
+
"BUSI_0185_M.png",
|
| 295 |
+
"BUSI_0173_M.png",
|
| 296 |
+
"BUSI_0369_B.png",
|
| 297 |
+
"BUSI_0398_B.png",
|
| 298 |
+
"BUSI_0179_B.png",
|
| 299 |
+
"BUSI_0171_M.png",
|
| 300 |
+
"BUSI_0353_B.png",
|
| 301 |
+
"BUSI_0089_B.png",
|
| 302 |
+
"BUSI_0427_B.png",
|
| 303 |
+
"BUSI_0051_M.png",
|
| 304 |
+
"BUSI_0368_B.png",
|
| 305 |
+
"BUSI_0005_B.png",
|
| 306 |
+
"BUSI_0231_B.png",
|
| 307 |
+
"BUSI_0039_M.png",
|
| 308 |
+
"BUSI_0038_B.png",
|
| 309 |
+
"BUSI_0166_M.png",
|
| 310 |
+
"BUSI_0188_M.png",
|
| 311 |
+
"BUSI_0023_B.png",
|
| 312 |
+
"BUSI_0169_B.png",
|
| 313 |
+
"BUSI_0259_B.png",
|
| 314 |
+
"BUSI_0191_M.png",
|
| 315 |
+
"BUSI_0174_B.png",
|
| 316 |
+
"BUSI_0049_M.png",
|
| 317 |
+
"BUSI_0165_B.png",
|
| 318 |
+
"BUSI_0151_B.png",
|
| 319 |
+
"BUSI_0112_M.png",
|
| 320 |
+
"BUSI_0001_M.png",
|
| 321 |
+
"BUSI_0419_B.png",
|
| 322 |
+
"BUSI_0015_M.png",
|
| 323 |
+
"BUSI_0335_B.png",
|
| 324 |
+
"BUSI_0083_M.png",
|
| 325 |
+
"BUSI_0129_B.png",
|
| 326 |
+
"BUSI_0329_B.png",
|
| 327 |
+
"BUSI_0025_M.png",
|
| 328 |
+
"BUSI_0015_B.png",
|
| 329 |
+
"BUSI_0006_M.png",
|
| 330 |
+
"BUSI_0071_M.png",
|
| 331 |
+
"BUSI_0205_M.png",
|
| 332 |
+
"BUSI_0143_B.png",
|
| 333 |
+
"BUSI_0327_B.png",
|
| 334 |
+
"BUSI_0033_B.png",
|
| 335 |
+
"BUSI_0263_B.png",
|
| 336 |
+
"BUSI_0044_M.png",
|
| 337 |
+
"BUSI_0295_B.png",
|
| 338 |
+
"BUSI_0429_B.png",
|
| 339 |
+
"BUSI_0237_B.png",
|
| 340 |
+
"BUSI_0088_B.png",
|
| 341 |
+
"BUSI_0053_B.png",
|
| 342 |
+
"BUSI_0377_B.png",
|
| 343 |
+
"BUSI_0350_B.png",
|
| 344 |
+
"BUSI_0146_B.png",
|
| 345 |
+
"BUSI_0082_M.png",
|
| 346 |
+
"BUSI_0014_B.png",
|
| 347 |
+
"BUSI_0005_M.png",
|
| 348 |
+
"BUSI_0410_B.png",
|
| 349 |
+
"BUSI_0111_B.png",
|
| 350 |
+
"BUSI_0304_B.png",
|
| 351 |
+
"BUSI_0073_B.png",
|
| 352 |
+
"BUSI_0291_B.png",
|
| 353 |
+
"BUSI_0030_M.png",
|
| 354 |
+
"BUSI_0287_B.png",
|
| 355 |
+
"BUSI_0294_B.png",
|
| 356 |
+
"BUSI_0356_B.png",
|
| 357 |
+
"BUSI_0408_B.png",
|
| 358 |
+
"BUSI_0014_M.png",
|
| 359 |
+
"BUSI_0434_B.png",
|
| 360 |
+
"BUSI_0292_B.png",
|
| 361 |
+
"BUSI_0177_B.png",
|
| 362 |
+
"BUSI_0001_B.png",
|
| 363 |
+
"BUSI_0162_M.png",
|
| 364 |
+
"BUSI_0261_B.png",
|
| 365 |
+
"BUSI_0104_M.png",
|
| 366 |
+
"BUSI_0102_M.png",
|
| 367 |
+
"BUSI_0029_M.png",
|
| 368 |
+
"BUSI_0064_M.png",
|
| 369 |
+
"BUSI_0026_M.png",
|
| 370 |
+
"BUSI_0029_B.png",
|
| 371 |
+
"BUSI_0406_B.png",
|
| 372 |
+
"BUSI_0210_B.png",
|
| 373 |
+
"BUSI_0051_B.png",
|
| 374 |
+
"BUSI_0176_M.png",
|
| 375 |
+
"BUSI_0370_B.png",
|
| 376 |
+
"BUSI_0065_B.png",
|
| 377 |
+
"BUSI_0050_M.png",
|
| 378 |
+
"BUSI_0151_M.png",
|
| 379 |
+
"BUSI_0149_B.png",
|
| 380 |
+
"BUSI_0108_B.png",
|
| 381 |
+
"BUSI_0180_B.png",
|
| 382 |
+
"BUSI_0407_B.png",
|
| 383 |
+
"BUSI_0054_M.png",
|
| 384 |
+
"BUSI_0195_M.png",
|
| 385 |
+
"BUSI_0364_B.png",
|
| 386 |
+
"BUSI_0137_M.png",
|
| 387 |
+
"BUSI_0289_B.png",
|
| 388 |
+
"BUSI_0042_B.png",
|
| 389 |
+
"BUSI_0034_M.png",
|
| 390 |
+
"BUSI_0281_B.png",
|
| 391 |
+
"BUSI_0066_B.png",
|
| 392 |
+
"BUSI_0355_B.png",
|
| 393 |
+
"BUSI_0207_B.png",
|
| 394 |
+
"BUSI_0173_B.png",
|
| 395 |
+
"BUSI_0386_B.png",
|
| 396 |
+
"BUSI_0430_B.png",
|
| 397 |
+
"BUSI_0319_B.png",
|
| 398 |
+
"BUSI_0105_M.png",
|
| 399 |
+
"BUSI_0157_B.png",
|
| 400 |
+
"BUSI_0376_B.png",
|
| 401 |
+
"BUSI_0228_B.png",
|
| 402 |
+
"BUSI_0147_B.png",
|
| 403 |
+
"BUSI_0197_M.png",
|
| 404 |
+
"BUSI_0193_B.png",
|
| 405 |
+
"BUSI_0027_M.png",
|
| 406 |
+
"BUSI_0116_B.png",
|
| 407 |
+
"BUSI_0133_M.png",
|
| 408 |
+
"BUSI_0345_B.png",
|
| 409 |
+
"BUSI_0215_B.png",
|
| 410 |
+
"BUSI_0254_B.png",
|
| 411 |
+
"BUSI_0112_B.png",
|
| 412 |
+
"BUSI_0235_B.png",
|
| 413 |
+
"BUSI_0239_B.png",
|
| 414 |
+
"BUSI_0209_B.png",
|
| 415 |
+
"BUSI_0010_M.png",
|
| 416 |
+
"BUSI_0128_M.png",
|
| 417 |
+
"BUSI_0422_B.png",
|
| 418 |
+
"BUSI_0334_B.png",
|
| 419 |
+
"BUSI_0054_B.png",
|
| 420 |
+
"BUSI_0149_M.png",
|
| 421 |
+
"BUSI_0144_M.png",
|
| 422 |
+
"BUSI_0208_B.png",
|
| 423 |
+
"BUSI_0109_M.png",
|
| 424 |
+
"BUSI_0209_M.png",
|
| 425 |
+
"BUSI_0011_M.png",
|
| 426 |
+
"BUSI_0393_B.png",
|
| 427 |
+
"BUSI_0108_M.png",
|
| 428 |
+
"BUSI_0020_M.png",
|
| 429 |
+
"BUSI_0416_B.png",
|
| 430 |
+
"BUSI_0024_M.png",
|
| 431 |
+
"BUSI_0308_B.png",
|
| 432 |
+
"BUSI_0185_B.png",
|
| 433 |
+
"BUSI_0098_B.png",
|
| 434 |
+
"BUSI_0201_M.png",
|
| 435 |
+
"BUSI_0045_M.png",
|
| 436 |
+
"BUSI_0351_B.png",
|
| 437 |
+
"BUSI_0414_B.png",
|
| 438 |
+
"BUSI_0395_B.png",
|
| 439 |
+
"BUSI_0018_M.png",
|
| 440 |
+
"BUSI_0062_B.png",
|
| 441 |
+
"BUSI_0314_B.png",
|
| 442 |
+
"BUSI_0002_M.png",
|
| 443 |
+
"BUSI_0306_B.png",
|
| 444 |
+
"BUSI_0118_B.png",
|
| 445 |
+
"BUSI_0142_B.png",
|
| 446 |
+
"BUSI_0240_B.png",
|
| 447 |
+
"BUSI_0318_B.png",
|
| 448 |
+
"BUSI_0181_M.png",
|
| 449 |
+
"BUSI_0236_B.png",
|
| 450 |
+
"BUSI_0148_M.png",
|
| 451 |
+
"BUSI_0127_M.png",
|
| 452 |
+
"BUSI_0089_M.png",
|
| 453 |
+
"BUSI_0091_B.png",
|
| 454 |
+
"BUSI_0268_B.png",
|
| 455 |
+
"BUSI_0123_B.png",
|
| 456 |
+
"BUSI_0204_M.png",
|
| 457 |
+
"BUSI_0324_B.png",
|
| 458 |
+
"BUSI_0400_B.png",
|
| 459 |
+
"BUSI_0144_B.png",
|
| 460 |
+
"BUSI_0196_M.png",
|
| 461 |
+
"BUSI_0375_B.png",
|
| 462 |
+
"BUSI_0004_M.png",
|
| 463 |
+
"BUSI_0187_B.png",
|
| 464 |
+
"BUSI_0013_B.png",
|
| 465 |
+
"BUSI_0047_M.png",
|
| 466 |
+
"BUSI_0328_B.png",
|
| 467 |
+
"BUSI_0016_B.png",
|
| 468 |
+
"BUSI_0283_B.png",
|
| 469 |
+
"BUSI_0056_M.png",
|
| 470 |
+
"BUSI_0058_M.png",
|
| 471 |
+
"BUSI_0069_B.png",
|
| 472 |
+
"BUSI_0071_B.png",
|
| 473 |
+
"BUSI_0302_B.png",
|
| 474 |
+
"BUSI_0057_M.png",
|
| 475 |
+
"BUSI_0090_B.png",
|
| 476 |
+
"BUSI_0092_B.png",
|
| 477 |
+
"BUSI_0331_B.png",
|
| 478 |
+
"BUSI_0082_B.png",
|
| 479 |
+
"BUSI_0034_B.png",
|
| 480 |
+
"BUSI_0107_B.png",
|
| 481 |
+
"BUSI_0114_B.png",
|
| 482 |
+
"BUSI_0078_M.png",
|
| 483 |
+
"BUSI_0041_M.png",
|
| 484 |
+
"BUSI_0204_B.png",
|
| 485 |
+
"BUSI_0199_M.png",
|
| 486 |
+
"BUSI_0055_B.png",
|
| 487 |
+
"BUSI_0336_B.png",
|
| 488 |
+
"BUSI_0387_B.png",
|
| 489 |
+
"BUSI_0190_M.png",
|
| 490 |
+
"BUSI_0064_B.png",
|
| 491 |
+
"BUSI_0125_M.png",
|
| 492 |
+
"BUSI_0152_M.png",
|
| 493 |
+
"BUSI_0391_B.png",
|
| 494 |
+
"BUSI_0050_B.png",
|
| 495 |
+
"BUSI_0002_B.png",
|
| 496 |
+
"BUSI_0212_B.png",
|
| 497 |
+
"BUSI_0272_B.png",
|
| 498 |
+
"BUSI_0045_B.png",
|
| 499 |
+
"BUSI_0245_B.png",
|
| 500 |
+
"BUSI_0374_B.png",
|
| 501 |
+
"BUSI_0139_M.png",
|
| 502 |
+
"BUSI_0184_M.png",
|
| 503 |
+
"BUSI_0043_M.png",
|
| 504 |
+
"BUSI_0402_B.png",
|
| 505 |
+
"BUSI_0180_M.png",
|
| 506 |
+
"BUSI_0078_B.png",
|
| 507 |
+
"BUSI_0025_B.png",
|
| 508 |
+
"BUSI_0206_M.png",
|
| 509 |
+
"BUSI_0140_M.png",
|
| 510 |
+
"BUSI_0205_B.png",
|
| 511 |
+
"BUSI_0004_B.png",
|
| 512 |
+
"BUSI_0202_M.png",
|
| 513 |
+
"BUSI_0164_B.png",
|
| 514 |
+
"BUSI_0124_M.png",
|
| 515 |
+
"BUSI_0150_B.png",
|
| 516 |
+
"BUSI_0179_M.png",
|
| 517 |
+
"BUSI_0032_M.png",
|
| 518 |
+
"BUSI_0187_M.png",
|
| 519 |
+
"BUSI_0086_B.png",
|
| 520 |
+
"BUSI_0132_M.png",
|
| 521 |
+
"BUSI_0110_M.png",
|
| 522 |
+
"BUSI_0397_B.png",
|
| 523 |
+
"BUSI_0321_B.png"
|
| 524 |
+
]
|
| 525 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/norm_stats_BUSI_with_classes_stratified_80_10_10_50pct.json
ADDED
|
@@ -0,0 +1,266 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"global_mean": 85.12605579417583,
|
| 3 |
+
"global_std": 56.68613974260357,
|
| 4 |
+
"total_pixels": 244630659,
|
| 5 |
+
"num_images": 258,
|
| 6 |
+
"filenames": [
|
| 7 |
+
"BUSI_0034_B.png",
|
| 8 |
+
"BUSI_0155_M.png",
|
| 9 |
+
"BUSI_0005_M.png",
|
| 10 |
+
"BUSI_0260_B.png",
|
| 11 |
+
"BUSI_0386_B.png",
|
| 12 |
+
"BUSI_0280_B.png",
|
| 13 |
+
"BUSI_0047_B.png",
|
| 14 |
+
"BUSI_0222_B.png",
|
| 15 |
+
"BUSI_0205_B.png",
|
| 16 |
+
"BUSI_0403_B.png",
|
| 17 |
+
"BUSI_0133_B.png",
|
| 18 |
+
"BUSI_0016_B.png",
|
| 19 |
+
"BUSI_0105_M.png",
|
| 20 |
+
"BUSI_0220_B.png",
|
| 21 |
+
"BUSI_0066_B.png",
|
| 22 |
+
"BUSI_0060_B.png",
|
| 23 |
+
"BUSI_0351_B.png",
|
| 24 |
+
"BUSI_0029_M.png",
|
| 25 |
+
"BUSI_0287_B.png",
|
| 26 |
+
"BUSI_0046_M.png",
|
| 27 |
+
"BUSI_0123_B.png",
|
| 28 |
+
"BUSI_0359_B.png",
|
| 29 |
+
"BUSI_0215_B.png",
|
| 30 |
+
"BUSI_0377_B.png",
|
| 31 |
+
"BUSI_0029_B.png",
|
| 32 |
+
"BUSI_0158_M.png",
|
| 33 |
+
"BUSI_0375_B.png",
|
| 34 |
+
"BUSI_0221_B.png",
|
| 35 |
+
"BUSI_0120_B.png",
|
| 36 |
+
"BUSI_0404_B.png",
|
| 37 |
+
"BUSI_0093_M.png",
|
| 38 |
+
"BUSI_0106_M.png",
|
| 39 |
+
"BUSI_0030_B.png",
|
| 40 |
+
"BUSI_0407_B.png",
|
| 41 |
+
"BUSI_0131_B.png",
|
| 42 |
+
"BUSI_0092_B.png",
|
| 43 |
+
"BUSI_0112_M.png",
|
| 44 |
+
"BUSI_0177_M.png",
|
| 45 |
+
"BUSI_0333_B.png",
|
| 46 |
+
"BUSI_0206_B.png",
|
| 47 |
+
"BUSI_0239_B.png",
|
| 48 |
+
"BUSI_0165_B.png",
|
| 49 |
+
"BUSI_0396_B.png",
|
| 50 |
+
"BUSI_0207_B.png",
|
| 51 |
+
"BUSI_0158_B.png",
|
| 52 |
+
"BUSI_0160_B.png",
|
| 53 |
+
"BUSI_0091_B.png",
|
| 54 |
+
"BUSI_0090_M.png",
|
| 55 |
+
"BUSI_0035_B.png",
|
| 56 |
+
"BUSI_0281_B.png",
|
| 57 |
+
"BUSI_0410_B.png",
|
| 58 |
+
"BUSI_0322_B.png",
|
| 59 |
+
"BUSI_0236_B.png",
|
| 60 |
+
"BUSI_0385_B.png",
|
| 61 |
+
"BUSI_0193_M.png",
|
| 62 |
+
"BUSI_0388_B.png",
|
| 63 |
+
"BUSI_0144_M.png",
|
| 64 |
+
"BUSI_0293_B.png",
|
| 65 |
+
"BUSI_0088_M.png",
|
| 66 |
+
"BUSI_0052_B.png",
|
| 67 |
+
"BUSI_0346_B.png",
|
| 68 |
+
"BUSI_0262_B.png",
|
| 69 |
+
"BUSI_0408_B.png",
|
| 70 |
+
"BUSI_0146_M.png",
|
| 71 |
+
"BUSI_0309_B.png",
|
| 72 |
+
"BUSI_0083_B.png",
|
| 73 |
+
"BUSI_0190_B.png",
|
| 74 |
+
"BUSI_0426_B.png",
|
| 75 |
+
"BUSI_0019_B.png",
|
| 76 |
+
"BUSI_0274_B.png",
|
| 77 |
+
"BUSI_0162_M.png",
|
| 78 |
+
"BUSI_0179_M.png",
|
| 79 |
+
"BUSI_0078_B.png",
|
| 80 |
+
"BUSI_0328_B.png",
|
| 81 |
+
"BUSI_0001_M.png",
|
| 82 |
+
"BUSI_0301_B.png",
|
| 83 |
+
"BUSI_0109_M.png",
|
| 84 |
+
"BUSI_0011_B.png",
|
| 85 |
+
"BUSI_0063_B.png",
|
| 86 |
+
"BUSI_0360_B.png",
|
| 87 |
+
"BUSI_0114_B.png",
|
| 88 |
+
"BUSI_0430_B.png",
|
| 89 |
+
"BUSI_0391_B.png",
|
| 90 |
+
"BUSI_0133_M.png",
|
| 91 |
+
"BUSI_0188_M.png",
|
| 92 |
+
"BUSI_0424_B.png",
|
| 93 |
+
"BUSI_0179_B.png",
|
| 94 |
+
"BUSI_0390_B.png",
|
| 95 |
+
"BUSI_0030_M.png",
|
| 96 |
+
"BUSI_0235_B.png",
|
| 97 |
+
"BUSI_0065_M.png",
|
| 98 |
+
"BUSI_0101_M.png",
|
| 99 |
+
"BUSI_0355_B.png",
|
| 100 |
+
"BUSI_0053_M.png",
|
| 101 |
+
"BUSI_0203_B.png",
|
| 102 |
+
"BUSI_0007_B.png",
|
| 103 |
+
"BUSI_0344_B.png",
|
| 104 |
+
"BUSI_0129_B.png",
|
| 105 |
+
"BUSI_0271_B.png",
|
| 106 |
+
"BUSI_0252_B.png",
|
| 107 |
+
"BUSI_0296_B.png",
|
| 108 |
+
"BUSI_0183_M.png",
|
| 109 |
+
"BUSI_0310_B.png",
|
| 110 |
+
"BUSI_0409_B.png",
|
| 111 |
+
"BUSI_0041_M.png",
|
| 112 |
+
"BUSI_0092_M.png",
|
| 113 |
+
"BUSI_0273_B.png",
|
| 114 |
+
"BUSI_0206_M.png",
|
| 115 |
+
"BUSI_0024_M.png",
|
| 116 |
+
"BUSI_0015_B.png",
|
| 117 |
+
"BUSI_0094_M.png",
|
| 118 |
+
"BUSI_0055_B.png",
|
| 119 |
+
"BUSI_0094_B.png",
|
| 120 |
+
"BUSI_0077_B.png",
|
| 121 |
+
"BUSI_0347_B.png",
|
| 122 |
+
"BUSI_0126_M.png",
|
| 123 |
+
"BUSI_0186_M.png",
|
| 124 |
+
"BUSI_0145_M.png",
|
| 125 |
+
"BUSI_0054_M.png",
|
| 126 |
+
"BUSI_0074_M.png",
|
| 127 |
+
"BUSI_0111_B.png",
|
| 128 |
+
"BUSI_0032_M.png",
|
| 129 |
+
"BUSI_0176_M.png",
|
| 130 |
+
"BUSI_0002_M.png",
|
| 131 |
+
"BUSI_0069_B.png",
|
| 132 |
+
"BUSI_0027_B.png",
|
| 133 |
+
"BUSI_0109_B.png",
|
| 134 |
+
"BUSI_0389_B.png",
|
| 135 |
+
"BUSI_0088_B.png",
|
| 136 |
+
"BUSI_0169_B.png",
|
| 137 |
+
"BUSI_0157_B.png",
|
| 138 |
+
"BUSI_0058_B.png",
|
| 139 |
+
"BUSI_0231_B.png",
|
| 140 |
+
"BUSI_0166_B.png",
|
| 141 |
+
"BUSI_0098_M.png",
|
| 142 |
+
"BUSI_0020_M.png",
|
| 143 |
+
"BUSI_0182_B.png",
|
| 144 |
+
"BUSI_0397_B.png",
|
| 145 |
+
"BUSI_0045_M.png",
|
| 146 |
+
"BUSI_0138_M.png",
|
| 147 |
+
"BUSI_0036_M.png",
|
| 148 |
+
"BUSI_0380_B.png",
|
| 149 |
+
"BUSI_0137_B.png",
|
| 150 |
+
"BUSI_0205_M.png",
|
| 151 |
+
"BUSI_0107_M.png",
|
| 152 |
+
"BUSI_0044_M.png",
|
| 153 |
+
"BUSI_0213_B.png",
|
| 154 |
+
"BUSI_0070_B.png",
|
| 155 |
+
"BUSI_0295_B.png",
|
| 156 |
+
"BUSI_0173_B.png",
|
| 157 |
+
"BUSI_0164_M.png",
|
| 158 |
+
"BUSI_0039_M.png",
|
| 159 |
+
"BUSI_0421_B.png",
|
| 160 |
+
"BUSI_0174_B.png",
|
| 161 |
+
"BUSI_0416_B.png",
|
| 162 |
+
"BUSI_0124_M.png",
|
| 163 |
+
"BUSI_0428_B.png",
|
| 164 |
+
"BUSI_0181_B.png",
|
| 165 |
+
"BUSI_0076_B.png",
|
| 166 |
+
"BUSI_0073_B.png",
|
| 167 |
+
"BUSI_0045_B.png",
|
| 168 |
+
"BUSI_0104_M.png",
|
| 169 |
+
"BUSI_0272_B.png",
|
| 170 |
+
"BUSI_0405_B.png",
|
| 171 |
+
"BUSI_0425_B.png",
|
| 172 |
+
"BUSI_0212_B.png",
|
| 173 |
+
"BUSI_0156_M.png",
|
| 174 |
+
"BUSI_0339_B.png",
|
| 175 |
+
"BUSI_0160_M.png",
|
| 176 |
+
"BUSI_0204_M.png",
|
| 177 |
+
"BUSI_0082_B.png",
|
| 178 |
+
"BUSI_0070_M.png",
|
| 179 |
+
"BUSI_0061_M.png",
|
| 180 |
+
"BUSI_0136_M.png",
|
| 181 |
+
"BUSI_0024_B.png",
|
| 182 |
+
"BUSI_0241_B.png",
|
| 183 |
+
"BUSI_0420_B.png",
|
| 184 |
+
"BUSI_0196_B.png",
|
| 185 |
+
"BUSI_0406_B.png",
|
| 186 |
+
"BUSI_0374_B.png",
|
| 187 |
+
"BUSI_0113_M.png",
|
| 188 |
+
"BUSI_0009_B.png",
|
| 189 |
+
"BUSI_0292_B.png",
|
| 190 |
+
"BUSI_0079_M.png",
|
| 191 |
+
"BUSI_0118_B.png",
|
| 192 |
+
"BUSI_0175_M.png",
|
| 193 |
+
"BUSI_0011_M.png",
|
| 194 |
+
"BUSI_0187_B.png",
|
| 195 |
+
"BUSI_0034_M.png",
|
| 196 |
+
"BUSI_0321_B.png",
|
| 197 |
+
"BUSI_0078_M.png",
|
| 198 |
+
"BUSI_0180_B.png",
|
| 199 |
+
"BUSI_0398_B.png",
|
| 200 |
+
"BUSI_0195_B.png",
|
| 201 |
+
"BUSI_0203_M.png",
|
| 202 |
+
"BUSI_0167_B.png",
|
| 203 |
+
"BUSI_0263_B.png",
|
| 204 |
+
"BUSI_0012_B.png",
|
| 205 |
+
"BUSI_0177_B.png",
|
| 206 |
+
"BUSI_0147_M.png",
|
| 207 |
+
"BUSI_0005_B.png",
|
| 208 |
+
"BUSI_0012_M.png",
|
| 209 |
+
"BUSI_0170_B.png",
|
| 210 |
+
"BUSI_0025_B.png",
|
| 211 |
+
"BUSI_0136_B.png",
|
| 212 |
+
"BUSI_0089_B.png",
|
| 213 |
+
"BUSI_0018_B.png",
|
| 214 |
+
"BUSI_0202_M.png",
|
| 215 |
+
"BUSI_0071_B.png",
|
| 216 |
+
"BUSI_0185_M.png",
|
| 217 |
+
"BUSI_0096_B.png",
|
| 218 |
+
"BUSI_0184_M.png",
|
| 219 |
+
"BUSI_0248_B.png",
|
| 220 |
+
"BUSI_0113_B.png",
|
| 221 |
+
"BUSI_0246_B.png",
|
| 222 |
+
"BUSI_0014_M.png",
|
| 223 |
+
"BUSI_0245_B.png",
|
| 224 |
+
"BUSI_0135_B.png",
|
| 225 |
+
"BUSI_0031_B.png",
|
| 226 |
+
"BUSI_0365_B.png",
|
| 227 |
+
"BUSI_0155_B.png",
|
| 228 |
+
"BUSI_0103_M.png",
|
| 229 |
+
"BUSI_0023_B.png",
|
| 230 |
+
"BUSI_0242_B.png",
|
| 231 |
+
"BUSI_0370_B.png",
|
| 232 |
+
"BUSI_0367_B.png",
|
| 233 |
+
"BUSI_0035_M.png",
|
| 234 |
+
"BUSI_0240_B.png",
|
| 235 |
+
"BUSI_0148_M.png",
|
| 236 |
+
"BUSI_0064_B.png",
|
| 237 |
+
"BUSI_0435_B.png",
|
| 238 |
+
"BUSI_0373_B.png",
|
| 239 |
+
"BUSI_0065_B.png",
|
| 240 |
+
"BUSI_0312_B.png",
|
| 241 |
+
"BUSI_0164_B.png",
|
| 242 |
+
"BUSI_0049_B.png",
|
| 243 |
+
"BUSI_0181_M.png",
|
| 244 |
+
"BUSI_0353_B.png",
|
| 245 |
+
"BUSI_0038_M.png",
|
| 246 |
+
"BUSI_0356_B.png",
|
| 247 |
+
"BUSI_0159_B.png",
|
| 248 |
+
"BUSI_0049_M.png",
|
| 249 |
+
"BUSI_0051_B.png",
|
| 250 |
+
"BUSI_0308_B.png",
|
| 251 |
+
"BUSI_0368_B.png",
|
| 252 |
+
"BUSI_0084_B.png",
|
| 253 |
+
"BUSI_0072_M.png",
|
| 254 |
+
"BUSI_0197_M.png",
|
| 255 |
+
"BUSI_0329_B.png",
|
| 256 |
+
"BUSI_0174_M.png",
|
| 257 |
+
"BUSI_0196_M.png",
|
| 258 |
+
"BUSI_0169_M.png",
|
| 259 |
+
"BUSI_0101_B.png",
|
| 260 |
+
"BUSI_0086_M.png",
|
| 261 |
+
"BUSI_0317_B.png",
|
| 262 |
+
"BUSI_0077_M.png",
|
| 263 |
+
"BUSI_0193_B.png",
|
| 264 |
+
"BUSI_0031_M.png"
|
| 265 |
+
]
|
| 266 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/best_params.json
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"head_lr": 0.002,
|
| 3 |
+
"encoder_lr": 0.0002,
|
| 4 |
+
"weight_decay": 0.002,
|
| 5 |
+
"dropout_p": 0.3,
|
| 6 |
+
"tmax": 6,
|
| 7 |
+
"entropy_lr": 0.0002
|
| 8 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/best.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/best.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 40,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0010.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0010.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 10,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0020.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0020.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 20,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0030.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0030.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 30,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0040.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0040.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 40,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0050.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0050.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 50,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0060.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0060.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 60,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0070.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0070.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 70,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0080.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0080.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 80,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0090.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0090.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 90,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0100.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/epoch_0100.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 100,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/latest.pt.meta.json
ADDED
|
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/latest.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 100,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
},
|
| 39 |
+
"last_restore": {
|
| 40 |
+
"restored_keys": {
|
| 41 |
+
"model_state_dict": true,
|
| 42 |
+
"optimizer_state_dict": true,
|
| 43 |
+
"scheduler_state_dict": true,
|
| 44 |
+
"scaler_state_dict": false,
|
| 45 |
+
"log_alpha": true,
|
| 46 |
+
"alpha_optimizer_state_dict": true,
|
| 47 |
+
"best_metric_name": true,
|
| 48 |
+
"best_metric_value": true,
|
| 49 |
+
"patience_counter": true,
|
| 50 |
+
"elapsed_seconds": true,
|
| 51 |
+
"run_config": true,
|
| 52 |
+
"epoch_metrics": true,
|
| 53 |
+
"resume_source": false
|
| 54 |
+
},
|
| 55 |
+
"restored_at_epoch": 100,
|
| 56 |
+
"expected_run_type": null
|
| 57 |
+
}
|
| 58 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/evaluation.json
ADDED
|
@@ -0,0 +1,619 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"strategy": 3,
|
| 3 |
+
"best_metric_name": "val_iou",
|
| 4 |
+
"metrics": {
|
| 5 |
+
"dice": {
|
| 6 |
+
"mean": 0.7515265941619873,
|
| 7 |
+
"std": 0.27273401618003845
|
| 8 |
+
},
|
| 9 |
+
"ppv": {
|
| 10 |
+
"mean": 0.823067307472229,
|
| 11 |
+
"std": 0.24877405166625977
|
| 12 |
+
},
|
| 13 |
+
"sen": {
|
| 14 |
+
"mean": 0.7724807858467102,
|
| 15 |
+
"std": 0.2787170708179474
|
| 16 |
+
},
|
| 17 |
+
"iou": {
|
| 18 |
+
"mean": 0.6618749499320984,
|
| 19 |
+
"std": 0.2781444787979126
|
| 20 |
+
},
|
| 21 |
+
"biou": {
|
| 22 |
+
"mean": 0.12041466683149338,
|
| 23 |
+
"std": 0.104156494140625
|
| 24 |
+
},
|
| 25 |
+
"hd95": {
|
| 26 |
+
"mean": Infinity,
|
| 27 |
+
"std": NaN
|
| 28 |
+
}
|
| 29 |
+
},
|
| 30 |
+
"per_sample": [
|
| 31 |
+
{
|
| 32 |
+
"sample_id": "BUSI_0097_B",
|
| 33 |
+
"dice": 0.7556420281003496,
|
| 34 |
+
"ppv": 0.8432479442792902,
|
| 35 |
+
"sen": 0.6845259187689137,
|
| 36 |
+
"iou": 0.6072545463647734,
|
| 37 |
+
"biou": 0.024725542657817953,
|
| 38 |
+
"hd95": 17.804492950439453
|
| 39 |
+
},
|
| 40 |
+
{
|
| 41 |
+
"sample_id": "BUSI_0415_B",
|
| 42 |
+
"dice": 0.9163019330933814,
|
| 43 |
+
"ppv": 0.9355149190772045,
|
| 44 |
+
"sen": 0.8978622341269169,
|
| 45 |
+
"iou": 0.8455324985021436,
|
| 46 |
+
"biou": 0.05000016379307521,
|
| 47 |
+
"hd95": 16.4559383392334
|
| 48 |
+
},
|
| 49 |
+
{
|
| 50 |
+
"sample_id": "BUSI_0290_B",
|
| 51 |
+
"dice": 0.9427293077689193,
|
| 52 |
+
"ppv": 0.9129116155584223,
|
| 53 |
+
"sen": 0.9745605932210641,
|
| 54 |
+
"iou": 0.891663144660891,
|
| 55 |
+
"biou": 0.07547198884528653,
|
| 56 |
+
"hd95": 2.8284270763397217
|
| 57 |
+
},
|
| 58 |
+
{
|
| 59 |
+
"sample_id": "BUSI_0354_B",
|
| 60 |
+
"dice": 0.8622117131220665,
|
| 61 |
+
"ppv": 0.7577962703848092,
|
| 62 |
+
"sen": 1.0,
|
| 63 |
+
"iou": 0.7577962703848092,
|
| 64 |
+
"biou": 0.037542990599661916,
|
| 65 |
+
"hd95": 10.630146026611328
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"sample_id": "BUSI_0067_B",
|
| 69 |
+
"dice": 0.9293525441875968,
|
| 70 |
+
"ppv": 0.9741630285966875,
|
| 71 |
+
"sen": 0.8884832429975691,
|
| 72 |
+
"iou": 0.8680285386501772,
|
| 73 |
+
"biou": 0.09621017603201866,
|
| 74 |
+
"hd95": 8.0
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"sample_id": "BUSI_0165_M",
|
| 78 |
+
"dice": 0.12392989368215934,
|
| 79 |
+
"ppv": 0.9913043497164458,
|
| 80 |
+
"sen": 0.06609654926661038,
|
| 81 |
+
"iou": 0.06605824907926627,
|
| 82 |
+
"biou": 0.010638508374785452,
|
| 83 |
+
"hd95": 50.72161865234375
|
| 84 |
+
},
|
| 85 |
+
{
|
| 86 |
+
"sample_id": "BUSI_0198_B",
|
| 87 |
+
"dice": 0.9579081646075203,
|
| 88 |
+
"ppv": 0.9482323265004844,
|
| 89 |
+
"sen": 0.9677835072304442,
|
| 90 |
+
"iou": 0.9192166512107313,
|
| 91 |
+
"biou": 0.16734727863376384,
|
| 92 |
+
"hd95": 2.0
|
| 93 |
+
},
|
| 94 |
+
{
|
| 95 |
+
"sample_id": "BUSI_0361_B",
|
| 96 |
+
"dice": 0.9299847819640494,
|
| 97 |
+
"ppv": 0.9119403050790817,
|
| 98 |
+
"sen": 0.9487577679535895,
|
| 99 |
+
"iou": 0.8691322994927241,
|
| 100 |
+
"biou": 0.10931210149307631,
|
| 101 |
+
"hd95": 2.2360680103302
|
| 102 |
+
},
|
| 103 |
+
{
|
| 104 |
+
"sample_id": "BUSI_0095_B",
|
| 105 |
+
"dice": 0.9352518140882939,
|
| 106 |
+
"ppv": 0.9701492685824534,
|
| 107 |
+
"sen": 0.902777822788045,
|
| 108 |
+
"iou": 0.8783784331628679,
|
| 109 |
+
"biou": 0.34285808163131193,
|
| 110 |
+
"hd95": 1.1449741125106812
|
| 111 |
+
},
|
| 112 |
+
{
|
| 113 |
+
"sample_id": "BUSI_0057_B",
|
| 114 |
+
"dice": 0.9237875464693888,
|
| 115 |
+
"ppv": 0.9478673232856287,
|
| 116 |
+
"sen": 0.9009009455401146,
|
| 117 |
+
"iou": 0.8583691594982148,
|
| 118 |
+
"biou": 0.23684310941696127,
|
| 119 |
+
"hd95": 1.0
|
| 120 |
+
},
|
| 121 |
+
{
|
| 122 |
+
"sample_id": "BUSI_0009_M",
|
| 123 |
+
"dice": 0.536507973293018,
|
| 124 |
+
"ppv": 0.8556962390642433,
|
| 125 |
+
"sen": 0.390751515520056,
|
| 126 |
+
"iou": 0.3665944287858537,
|
| 127 |
+
"biou": 0.08988815174822935,
|
| 128 |
+
"hd95": 15.817686080932617
|
| 129 |
+
},
|
| 130 |
+
{
|
| 131 |
+
"sample_id": "BUSI_0399_B",
|
| 132 |
+
"dice": 0.9555602816168344,
|
| 133 |
+
"ppv": 0.9296648767205264,
|
| 134 |
+
"sen": 0.9829396332922296,
|
| 135 |
+
"iou": 0.9149022835951839,
|
| 136 |
+
"biou": 0.17204330751135935,
|
| 137 |
+
"hd95": 3.605551242828369
|
| 138 |
+
},
|
| 139 |
+
{
|
| 140 |
+
"sample_id": "BUSI_0099_M",
|
| 141 |
+
"dice": 0.745387469533366,
|
| 142 |
+
"ppv": 0.8991098072537378,
|
| 143 |
+
"sen": 0.636554660025771,
|
| 144 |
+
"iou": 0.5941176868512071,
|
| 145 |
+
"biou": 0.07500046249976874,
|
| 146 |
+
"hd95": 12.110795974731445
|
| 147 |
+
},
|
| 148 |
+
{
|
| 149 |
+
"sample_id": "BUSI_0307_B",
|
| 150 |
+
"dice": 0.9214537020570103,
|
| 151 |
+
"ppv": 0.9290780309508201,
|
| 152 |
+
"sen": 0.913953508382905,
|
| 153 |
+
"iou": 0.8543478577504657,
|
| 154 |
+
"biou": 0.18897701655353027,
|
| 155 |
+
"hd95": 1.4142135381698608
|
| 156 |
+
},
|
| 157 |
+
{
|
| 158 |
+
"sample_id": "BUSI_0039_B",
|
| 159 |
+
"dice": 0.9075000115624986,
|
| 160 |
+
"ppv": 0.8461538820154121,
|
| 161 |
+
"sen": 0.9784366634941608,
|
| 162 |
+
"iou": 0.8306636543103766,
|
| 163 |
+
"biou": 0.17272802479270472,
|
| 164 |
+
"hd95": 2.2360680103302
|
| 165 |
+
},
|
| 166 |
+
{
|
| 167 |
+
"sample_id": "BUSI_0189_M",
|
| 168 |
+
"dice": 0.18084036366923145,
|
| 169 |
+
"ppv": 0.09944550464157098,
|
| 170 |
+
"sen": 0.996296297668038,
|
| 171 |
+
"iou": 0.09940875464121378,
|
| 172 |
+
"biou": 0.031496444292738464,
|
| 173 |
+
"hd95": 41.95243453979492
|
| 174 |
+
},
|
| 175 |
+
{
|
| 176 |
+
"sample_id": "BUSI_0021_B",
|
| 177 |
+
"dice": 0.8654545943801476,
|
| 178 |
+
"ppv": 0.7828948796744213,
|
| 179 |
+
"sen": 0.9674797012360152,
|
| 180 |
+
"iou": 0.7628206648585482,
|
| 181 |
+
"biou": 0.1641803519696239,
|
| 182 |
+
"hd95": 2.0
|
| 183 |
+
},
|
| 184 |
+
{
|
| 185 |
+
"sample_id": "BUSI_0311_B",
|
| 186 |
+
"dice": 0.3312102147484,
|
| 187 |
+
"ppv": 0.9473684317067952,
|
| 188 |
+
"sen": 0.20068614062237816,
|
| 189 |
+
"iou": 0.19847331643454974,
|
| 190 |
+
"biou": 0.010676508656046744,
|
| 191 |
+
"hd95": 63.140968322753906
|
| 192 |
+
},
|
| 193 |
+
{
|
| 194 |
+
"sample_id": "BUSI_0080_M",
|
| 195 |
+
"dice": 0.6588122226782346,
|
| 196 |
+
"ppv": 0.5359937583727011,
|
| 197 |
+
"sen": 0.8546475449377701,
|
| 198 |
+
"iou": 0.491215507665274,
|
| 199 |
+
"biou": 0.051075523904429064,
|
| 200 |
+
"hd95": 25.00807762145996
|
| 201 |
+
},
|
| 202 |
+
{
|
| 203 |
+
"sample_id": "BUSI_0130_B",
|
| 204 |
+
"dice": 0.9750927411185617,
|
| 205 |
+
"ppv": 0.9583333376736106,
|
| 206 |
+
"sen": 0.9924487602536397,
|
| 207 |
+
"iou": 0.9513960753468381,
|
| 208 |
+
"biou": 0.4109593075621181,
|
| 209 |
+
"hd95": 1.0
|
| 210 |
+
},
|
| 211 |
+
{
|
| 212 |
+
"sample_id": "BUSI_0401_B",
|
| 213 |
+
"dice": 0.8478527654032895,
|
| 214 |
+
"ppv": 0.9369491568170063,
|
| 215 |
+
"sen": 0.7742297045249465,
|
| 216 |
+
"iou": 0.7358892579398691,
|
| 217 |
+
"biou": 0.06827346655684075,
|
| 218 |
+
"hd95": 11.0
|
| 219 |
+
},
|
| 220 |
+
{
|
| 221 |
+
"sample_id": "BUSI_0326_B",
|
| 222 |
+
"dice": 0.9486510023599941,
|
| 223 |
+
"ppv": 0.9098497545993459,
|
| 224 |
+
"sen": 0.990909091460055,
|
| 225 |
+
"iou": 0.9023178861855471,
|
| 226 |
+
"biou": 0.08097203199512874,
|
| 227 |
+
"hd95": 2.2360680103302
|
| 228 |
+
},
|
| 229 |
+
{
|
| 230 |
+
"sample_id": "BUSI_0119_B",
|
| 231 |
+
"dice": 0.1842234598616629,
|
| 232 |
+
"ppv": 0.10145714509135753,
|
| 233 |
+
"sen": 1.0,
|
| 234 |
+
"iou": 0.10145714509135753,
|
| 235 |
+
"biou": 0.02597444763010925,
|
| 236 |
+
"hd95": 33.78740310668945
|
| 237 |
+
},
|
| 238 |
+
{
|
| 239 |
+
"sample_id": "BUSI_0372_B",
|
| 240 |
+
"dice": 0.9308078882563631,
|
| 241 |
+
"ppv": 0.8985074702606366,
|
| 242 |
+
"sen": 0.9655172441445674,
|
| 243 |
+
"iou": 0.8705712313397519,
|
| 244 |
+
"biou": 0.12444483358007397,
|
| 245 |
+
"hd95": 3.605551242828369
|
| 246 |
+
},
|
| 247 |
+
{
|
| 248 |
+
"sample_id": "BUSI_0003_M",
|
| 249 |
+
"dice": 0.7841044018917911,
|
| 250 |
+
"ppv": 0.9552023186123817,
|
| 251 |
+
"sen": 0.6649899733410489,
|
| 252 |
+
"iou": 0.6448780834265284,
|
| 253 |
+
"biou": 0.07462732605605668,
|
| 254 |
+
"hd95": 11.746427536010742
|
| 255 |
+
},
|
| 256 |
+
{
|
| 257 |
+
"sample_id": "BUSI_0007_M",
|
| 258 |
+
"dice": 0.6016666998611083,
|
| 259 |
+
"ppv": 0.9575596929549886,
|
| 260 |
+
"sen": 0.43863919336097285,
|
| 261 |
+
"iou": 0.43027420378138215,
|
| 262 |
+
"biou": 0.1270723054849141,
|
| 263 |
+
"hd95": 12.146934509277344
|
| 264 |
+
},
|
| 265 |
+
{
|
| 266 |
+
"sample_id": "BUSI_0411_B",
|
| 267 |
+
"dice": 0.9388932615029305,
|
| 268 |
+
"ppv": 0.9222050529518262,
|
| 269 |
+
"sen": 0.9561965835365073,
|
| 270 |
+
"iou": 0.884824523735812,
|
| 271 |
+
"biou": 0.09473715974134747,
|
| 272 |
+
"hd95": 3.0
|
| 273 |
+
},
|
| 274 |
+
{
|
| 275 |
+
"sample_id": "BUSI_0211_B",
|
| 276 |
+
"dice": 0.49238591563809253,
|
| 277 |
+
"ppv": 0.8818182892561007,
|
| 278 |
+
"sen": 0.3415495276234058,
|
| 279 |
+
"iou": 0.3265995533334838,
|
| 280 |
+
"biou": 0.10843480911468781,
|
| 281 |
+
"hd95": 12.653122901916504
|
| 282 |
+
},
|
| 283 |
+
{
|
| 284 |
+
"sample_id": "BUSI_0202_B",
|
| 285 |
+
"dice": 0.8864865478451093,
|
| 286 |
+
"ppv": 0.9111112098764335,
|
| 287 |
+
"sen": 0.8631580387810118,
|
| 288 |
+
"iou": 0.7961167027993177,
|
| 289 |
+
"biou": 0.2448995002051016,
|
| 290 |
+
"hd95": 1.0
|
| 291 |
+
},
|
| 292 |
+
{
|
| 293 |
+
"sample_id": "BUSI_0194_M",
|
| 294 |
+
"dice": 0.6612510246073375,
|
| 295 |
+
"ppv": 0.7772119810813507,
|
| 296 |
+
"sen": 0.5754005855136387,
|
| 297 |
+
"iou": 0.4939320593069555,
|
| 298 |
+
"biou": 0.018518821444808195,
|
| 299 |
+
"hd95": 14.270502090454102
|
| 300 |
+
},
|
| 301 |
+
{
|
| 302 |
+
"sample_id": "BUSI_0141_B",
|
| 303 |
+
"dice": 0.6217055528170591,
|
| 304 |
+
"ppv": 0.9393647752799621,
|
| 305 |
+
"sen": 0.4645959839986198,
|
| 306 |
+
"iou": 0.45106875273173025,
|
| 307 |
+
"biou": 0.025951725613888303,
|
| 308 |
+
"hd95": 50.322547912597656
|
| 309 |
+
},
|
| 310 |
+
{
|
| 311 |
+
"sample_id": "BUSI_0251_B",
|
| 312 |
+
"dice": 0.8585366083680311,
|
| 313 |
+
"ppv": 0.9635036629548676,
|
| 314 |
+
"sen": 0.7741936146059781,
|
| 315 |
+
"iou": 0.752136822753042,
|
| 316 |
+
"biou": 0.1717180083656481,
|
| 317 |
+
"hd95": 5.024755001068115
|
| 318 |
+
},
|
| 319 |
+
{
|
| 320 |
+
"sample_id": "BUSI_0161_M",
|
| 321 |
+
"dice": 0.7877535229588746,
|
| 322 |
+
"ppv": 0.8002624724340559,
|
| 323 |
+
"sen": 0.7756296164937773,
|
| 324 |
+
"iou": 0.6498295044793371,
|
| 325 |
+
"biou": 0.03803153511598767,
|
| 326 |
+
"hd95": 18.308090209960938
|
| 327 |
+
},
|
| 328 |
+
{
|
| 329 |
+
"sample_id": "BUSI_0055_M",
|
| 330 |
+
"dice": 0.8734253225299746,
|
| 331 |
+
"ppv": 0.8207441013817247,
|
| 332 |
+
"sen": 0.9333333418803408,
|
| 333 |
+
"iou": 0.7752928886802035,
|
| 334 |
+
"biou": 0.13402106493759539,
|
| 335 |
+
"hd95": 6.4031243324279785
|
| 336 |
+
},
|
| 337 |
+
{
|
| 338 |
+
"sample_id": "BUSI_0059_M",
|
| 339 |
+
"dice": 0.8882946409591026,
|
| 340 |
+
"ppv": 0.8315885336461417,
|
| 341 |
+
"sen": 0.9533002881998359,
|
| 342 |
+
"iou": 0.7990378351402179,
|
| 343 |
+
"biou": 0.03155837641846619,
|
| 344 |
+
"hd95": 7.2732086181640625
|
| 345 |
+
},
|
| 346 |
+
{
|
| 347 |
+
"sample_id": "BUSI_0285_B",
|
| 348 |
+
"dice": 0.9270563928061877,
|
| 349 |
+
"ppv": 0.9246646104577285,
|
| 350 |
+
"sen": 0.9294605882302295,
|
| 351 |
+
"iou": 0.8640308713567144,
|
| 352 |
+
"biou": 0.11891939517870531,
|
| 353 |
+
"hd95": 3.0
|
| 354 |
+
},
|
| 355 |
+
{
|
| 356 |
+
"sample_id": "BUSI_0163_B",
|
| 357 |
+
"dice": 0.22340435858984595,
|
| 358 |
+
"ppv": 0.125748633869965,
|
| 359 |
+
"sen": 1.0,
|
| 360 |
+
"iou": 0.125748633869965,
|
| 361 |
+
"biou": 0.05737782182145752,
|
| 362 |
+
"hd95": 26.97032928466797
|
| 363 |
+
},
|
| 364 |
+
{
|
| 365 |
+
"sample_id": "BUSI_0224_B",
|
| 366 |
+
"dice": 2.3719164522790216e-08,
|
| 367 |
+
"ppv": 4.6274870602736205e-08,
|
| 368 |
+
"sen": 4.866179811864729e-08,
|
| 369 |
+
"iou": 2.3719164522790216e-08,
|
| 370 |
+
"biou": 2.932550459662622e-07,
|
| 371 |
+
"hd95": 60.10823440551758
|
| 372 |
+
},
|
| 373 |
+
{
|
| 374 |
+
"sample_id": "BUSI_0316_B",
|
| 375 |
+
"dice": 0.843791723686767,
|
| 376 |
+
"ppv": 0.8689275917703466,
|
| 377 |
+
"sen": 0.8200692072652387,
|
| 378 |
+
"iou": 0.7297921519662486,
|
| 379 |
+
"biou": 0.0380953907785094,
|
| 380 |
+
"hd95": 16.278820037841797
|
| 381 |
+
},
|
| 382 |
+
{
|
| 383 |
+
"sample_id": "BUSI_0116_M",
|
| 384 |
+
"dice": 0.17142862074829637,
|
| 385 |
+
"ppv": 0.18067764357871474,
|
| 386 |
+
"sen": 0.1630805024823893,
|
| 387 |
+
"iou": 0.0937500590006472,
|
| 388 |
+
"biou": 4.132229697425745e-07,
|
| 389 |
+
"hd95": 89.24337768554688
|
| 390 |
+
},
|
| 391 |
+
{
|
| 392 |
+
"sample_id": "BUSI_0214_B",
|
| 393 |
+
"dice": 0.9214092183518039,
|
| 394 |
+
"ppv": 0.9052183274527873,
|
| 395 |
+
"sen": 0.9381898522969258,
|
| 396 |
+
"iou": 0.854271371430013,
|
| 397 |
+
"biou": 0.1620116413342786,
|
| 398 |
+
"hd95": 3.861940622329712
|
| 399 |
+
},
|
| 400 |
+
{
|
| 401 |
+
"sample_id": "BUSI_0207_M",
|
| 402 |
+
"dice": 5.2938059664475407e-08,
|
| 403 |
+
"ppv": 1.3513495252033443e-06,
|
| 404 |
+
"sen": 5.5096415697167185e-08,
|
| 405 |
+
"iou": 5.2938059664475407e-08,
|
| 406 |
+
"biou": 4.5248848303688553e-07,
|
| 407 |
+
"hd95": 65.11528015136719
|
| 408 |
+
},
|
| 409 |
+
{
|
| 410 |
+
"sample_id": "BUSI_0016_M",
|
| 411 |
+
"dice": 0.7007402238149414,
|
| 412 |
+
"ppv": 0.8434450639183256,
|
| 413 |
+
"sen": 0.5993367621544539,
|
| 414 |
+
"iou": 0.5393380483087886,
|
| 415 |
+
"biou": 0.03138095577804272,
|
| 416 |
+
"hd95": 17.0
|
| 417 |
+
},
|
| 418 |
+
{
|
| 419 |
+
"sample_id": "BUSI_0067_M",
|
| 420 |
+
"dice": 1.4245012215810227e-07,
|
| 421 |
+
"ppv": 1.0,
|
| 422 |
+
"sen": 1.4245012215810227e-07,
|
| 423 |
+
"iou": 1.4245012215810227e-07,
|
| 424 |
+
"biou": 7.692301775152481e-07,
|
| 425 |
+
"hd95": Infinity
|
| 426 |
+
},
|
| 427 |
+
{
|
| 428 |
+
"sample_id": "BUSI_0184_B",
|
| 429 |
+
"dice": 0.9601924768111881,
|
| 430 |
+
"ppv": 0.9878487854253472,
|
| 431 |
+
"sen": 0.9340425559981891,
|
| 432 |
+
"iou": 0.9234329018328606,
|
| 433 |
+
"biou": 0.2419827282849189,
|
| 434 |
+
"hd95": 2.0
|
| 435 |
+
},
|
| 436 |
+
{
|
| 437 |
+
"sample_id": "BUSI_0099_B",
|
| 438 |
+
"dice": 0.9645697784114404,
|
| 439 |
+
"ppv": 0.9794420000819383,
|
| 440 |
+
"sen": 0.9501424572446642,
|
| 441 |
+
"iou": 0.9315642553681207,
|
| 442 |
+
"biou": 0.32575808654690414,
|
| 443 |
+
"hd95": 1.0
|
| 444 |
+
},
|
| 445 |
+
{
|
| 446 |
+
"sample_id": "BUSI_0412_B",
|
| 447 |
+
"dice": 0.817295018742464,
|
| 448 |
+
"ppv": 0.9673392975381881,
|
| 449 |
+
"sen": 0.7075471742613029,
|
| 450 |
+
"iou": 0.6910387919298738,
|
| 451 |
+
"biou": 0.08203142929073648,
|
| 452 |
+
"hd95": 27.931867599487305
|
| 453 |
+
},
|
| 454 |
+
{
|
| 455 |
+
"sample_id": "BUSI_0100_M",
|
| 456 |
+
"dice": 0.7121387449630783,
|
| 457 |
+
"ppv": 0.8279570123713693,
|
| 458 |
+
"sen": 0.6247464883624251,
|
| 459 |
+
"iou": 0.5529623381541887,
|
| 460 |
+
"biou": 0.05726913776689966,
|
| 461 |
+
"hd95": 13.37458324432373
|
| 462 |
+
},
|
| 463 |
+
{
|
| 464 |
+
"sample_id": "BUSI_0073_M",
|
| 465 |
+
"dice": 0.8519061692150902,
|
| 466 |
+
"ppv": 0.775700964525906,
|
| 467 |
+
"sen": 0.9447154561438282,
|
| 468 |
+
"iou": 0.7420179128968183,
|
| 469 |
+
"biou": 0.11494303738905896,
|
| 470 |
+
"hd95": 4.322961807250977
|
| 471 |
+
},
|
| 472 |
+
{
|
| 473 |
+
"sample_id": "BUSI_0357_B",
|
| 474 |
+
"dice": 0.8443235999780013,
|
| 475 |
+
"ppv": 0.7340425846285361,
|
| 476 |
+
"sen": 0.9936000010239998,
|
| 477 |
+
"iou": 0.7305882669896157,
|
| 478 |
+
"biou": 0.21153896696220068,
|
| 479 |
+
"hd95": 9.13172435760498
|
| 480 |
+
},
|
| 481 |
+
{
|
| 482 |
+
"sample_id": "BUSI_0229_B",
|
| 483 |
+
"dice": 0.9608158904554004,
|
| 484 |
+
"ppv": 0.9686147220113938,
|
| 485 |
+
"sen": 0.9531416450328386,
|
| 486 |
+
"iou": 0.9245867846501256,
|
| 487 |
+
"biou": 0.21176516954990027,
|
| 488 |
+
"hd95": 1.0
|
| 489 |
+
},
|
| 490 |
+
{
|
| 491 |
+
"sample_id": "BUSI_0129_M",
|
| 492 |
+
"dice": 0.8487343448477328,
|
| 493 |
+
"ppv": 0.987163814442465,
|
| 494 |
+
"sen": 0.7443539753517595,
|
| 495 |
+
"iou": 0.7372185067373934,
|
| 496 |
+
"biou": 0.04363653752062954,
|
| 497 |
+
"hd95": 14.866068840026855
|
| 498 |
+
},
|
| 499 |
+
{
|
| 500 |
+
"sample_id": "BUSI_0156_B",
|
| 501 |
+
"dice": 0.9546218563660746,
|
| 502 |
+
"ppv": 0.9692832869340318,
|
| 503 |
+
"sen": 0.9403973707293475,
|
| 504 |
+
"iou": 0.9131833076581004,
|
| 505 |
+
"biou": 0.36708940874758383,
|
| 506 |
+
"hd95": 1.0
|
| 507 |
+
},
|
| 508 |
+
{
|
| 509 |
+
"sample_id": "BUSI_0320_B",
|
| 510 |
+
"dice": 0.930131011994431,
|
| 511 |
+
"ppv": 0.8783505405462803,
|
| 512 |
+
"sen": 0.9883990746173841,
|
| 513 |
+
"iou": 0.8693877817575956,
|
| 514 |
+
"biou": 0.22222288698898546,
|
| 515 |
+
"hd95": 2.0
|
| 516 |
+
},
|
| 517 |
+
{
|
| 518 |
+
"sample_id": "BUSI_0104_B",
|
| 519 |
+
"dice": 0.9071428903061107,
|
| 520 |
+
"ppv": 0.9338235780708984,
|
| 521 |
+
"sen": 0.8819445264274122,
|
| 522 |
+
"iou": 0.8300654705454441,
|
| 523 |
+
"biou": 0.19354968782308415,
|
| 524 |
+
"hd95": 1.0
|
| 525 |
+
},
|
| 526 |
+
{
|
| 527 |
+
"sample_id": "BUSI_0340_B",
|
| 528 |
+
"dice": 0.8879289078261058,
|
| 529 |
+
"ppv": 0.9304190395649499,
|
| 530 |
+
"sen": 0.8491501452041986,
|
| 531 |
+
"iou": 0.7984461753952015,
|
| 532 |
+
"biou": 0.05275802092133791,
|
| 533 |
+
"hd95": 13.349139213562012
|
| 534 |
+
},
|
| 535 |
+
{
|
| 536 |
+
"sample_id": "BUSI_0121_B",
|
| 537 |
+
"dice": 0.9428768105384163,
|
| 538 |
+
"ppv": 0.9448275938168836,
|
| 539 |
+
"sen": 0.9409340740475173,
|
| 540 |
+
"iou": 0.8919270974053258,
|
| 541 |
+
"biou": 0.2191786170009473,
|
| 542 |
+
"hd95": 2.0
|
| 543 |
+
},
|
| 544 |
+
{
|
| 545 |
+
"sample_id": "BUSI_0358_B",
|
| 546 |
+
"dice": 0.8846857910502873,
|
| 547 |
+
"ppv": 0.7993385006242005,
|
| 548 |
+
"sen": 0.9904371597763443,
|
| 549 |
+
"iou": 0.7932166528209351,
|
| 550 |
+
"biou": 0.07777829012317215,
|
| 551 |
+
"hd95": 4.123105525970459
|
| 552 |
+
},
|
| 553 |
+
{
|
| 554 |
+
"sample_id": "BUSI_0270_B",
|
| 555 |
+
"dice": 0.9265091959961685,
|
| 556 |
+
"ppv": 0.895939112705809,
|
| 557 |
+
"sen": 0.9592391415111029,
|
| 558 |
+
"iou": 0.8630807180731741,
|
| 559 |
+
"biou": 0.2549026912718713,
|
| 560 |
+
"hd95": 2.2360680103302
|
| 561 |
+
},
|
| 562 |
+
{
|
| 563 |
+
"sample_id": "BUSI_0023_M",
|
| 564 |
+
"dice": 0.7460753129204599,
|
| 565 |
+
"ppv": 0.9100160709690394,
|
| 566 |
+
"sen": 0.6321860533546781,
|
| 567 |
+
"iou": 0.5949921275609853,
|
| 568 |
+
"biou": 0.04339640690633832,
|
| 569 |
+
"hd95": 32.124385833740234
|
| 570 |
+
},
|
| 571 |
+
{
|
| 572 |
+
"sample_id": "BUSI_0020_B",
|
| 573 |
+
"dice": 0.899873263632873,
|
| 574 |
+
"ppv": 0.8179723712013397,
|
| 575 |
+
"sen": 1.0,
|
| 576 |
+
"iou": 0.8179723712013397,
|
| 577 |
+
"biou": 0.06965220415313227,
|
| 578 |
+
"hd95": 3.605551242828369
|
| 579 |
+
},
|
| 580 |
+
{
|
| 581 |
+
"sample_id": "BUSI_0153_M",
|
| 582 |
+
"dice": 0.8632911405716539,
|
| 583 |
+
"ppv": 0.9859875447956507,
|
| 584 |
+
"sen": 0.7677519957091792,
|
| 585 |
+
"iou": 0.7594654829627294,
|
| 586 |
+
"biou": 0.03696516790560936,
|
| 587 |
+
"hd95": 12.058175086975098
|
| 588 |
+
},
|
| 589 |
+
{
|
| 590 |
+
"sample_id": "BUSI_0154_B",
|
| 591 |
+
"dice": 0.9377289605388422,
|
| 592 |
+
"ppv": 0.9078014838287348,
|
| 593 |
+
"sen": 0.9696969926537935,
|
| 594 |
+
"iou": 0.8827587015457231,
|
| 595 |
+
"biou": 0.43137366397320787,
|
| 596 |
+
"hd95": 1.0
|
| 597 |
+
},
|
| 598 |
+
{
|
| 599 |
+
"sample_id": "BUSI_0062_M",
|
| 600 |
+
"dice": 0.6312817721264322,
|
| 601 |
+
"ppv": 0.8817034793609243,
|
| 602 |
+
"sen": 0.4916447013348855,
|
| 603 |
+
"iou": 0.46122114433906164,
|
| 604 |
+
"biou": 0.026163073789804134,
|
| 605 |
+
"hd95": 23.890357971191406
|
| 606 |
+
},
|
| 607 |
+
{
|
| 608 |
+
"sample_id": "BUSI_0075_B",
|
| 609 |
+
"dice": 0.9283154186097295,
|
| 610 |
+
"ppv": 0.9829222043790884,
|
| 611 |
+
"sen": 0.8794567267475846,
|
| 612 |
+
"iou": 0.8662207581570639,
|
| 613 |
+
"biou": 0.2480006015995187,
|
| 614 |
+
"hd95": 2.384157657623291
|
| 615 |
+
}
|
| 616 |
+
],
|
| 617 |
+
"checkpoint_mode": "latest",
|
| 618 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/checkpoints/latest.pt"
|
| 619 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/evaluation_results.xlsx
ADDED
|
Binary file (35.1 kB). View file
|
|
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/evaluation_summary.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"mean_iou": 0.6618749499320984,
|
| 3 |
+
"mean_dice": 0.7515265941619873
|
| 4 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/history.json
ADDED
|
@@ -0,0 +1,2242 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"epoch": 1,
|
| 4 |
+
"train_loss": 1.0741126812315842,
|
| 5 |
+
"train_actor_loss": -0.3411457851077598,
|
| 6 |
+
"train_critic_loss": 0.016774736517049568,
|
| 7 |
+
"train_mean_reward": -2.7981317769501415e-05,
|
| 8 |
+
"train_entropy": 0.6007360268588171,
|
| 9 |
+
"train_ce_loss": 1.2730903993493057,
|
| 10 |
+
"train_dice_loss": 1.5406517586658781,
|
| 11 |
+
"train_dice": 0.31233189710864495,
|
| 12 |
+
"train_iou": 0.21463618528776132,
|
| 13 |
+
"grad_norm": 16.149961311687793,
|
| 14 |
+
"lr": 0.001999509027563903,
|
| 15 |
+
"encoder_lr": 0.0001999531232347445,
|
| 16 |
+
"alpha": 0.49175167083740234,
|
| 17 |
+
"validated_this_epoch": false,
|
| 18 |
+
"val_loss": null,
|
| 19 |
+
"val_dice": null,
|
| 20 |
+
"val_iou": null,
|
| 21 |
+
"val_reward": null,
|
| 22 |
+
"val_entropy": null
|
| 23 |
+
},
|
| 24 |
+
{
|
| 25 |
+
"epoch": 2,
|
| 26 |
+
"train_loss": 0.6608776041061091,
|
| 27 |
+
"train_actor_loss": -0.3926134504214199,
|
| 28 |
+
"train_critic_loss": 0.0006972289073874497,
|
| 29 |
+
"train_mean_reward": -0.006643643656217472,
|
| 30 |
+
"train_entropy": 0.6736407152519053,
|
| 31 |
+
"train_ce_loss": 0.7191254076829483,
|
| 32 |
+
"train_dice_loss": 1.3871594479533744,
|
| 33 |
+
"train_dice": 0.4300876771372404,
|
| 34 |
+
"train_iou": 0.31759771983799806,
|
| 35 |
+
"grad_norm": 8.42764131551565,
|
| 36 |
+
"lr": 0.0019980365947861304,
|
| 37 |
+
"encoder_lr": 0.00019981253920068581,
|
| 38 |
+
"alpha": 0.48091697692871094,
|
| 39 |
+
"validated_this_epoch": false,
|
| 40 |
+
"val_loss": null,
|
| 41 |
+
"val_dice": null,
|
| 42 |
+
"val_iou": null,
|
| 43 |
+
"val_reward": null,
|
| 44 |
+
"val_entropy": null
|
| 45 |
+
},
|
| 46 |
+
{
|
| 47 |
+
"epoch": 3,
|
| 48 |
+
"train_loss": 0.5909319379576301,
|
| 49 |
+
"train_actor_loss": -0.40819361392957293,
|
| 50 |
+
"train_critic_loss": 0.0006155424364239717,
|
| 51 |
+
"train_mean_reward": -0.006901291133732882,
|
| 52 |
+
"train_entropy": 0.7078817028176877,
|
| 53 |
+
"train_ce_loss": 0.6519431095685839,
|
| 54 |
+
"train_dice_loss": 1.3456924352843018,
|
| 55 |
+
"train_dice": 0.48784107748687977,
|
| 56 |
+
"train_iou": 0.3654054562606787,
|
| 57 |
+
"grad_norm": 5.8614915507708405,
|
| 58 |
+
"lr": 0.001995584154780065,
|
| 59 |
+
"encoder_lr": 0.0001995783866372926,
|
| 60 |
+
"alpha": 0.4694448709487915,
|
| 61 |
+
"validated_this_epoch": false,
|
| 62 |
+
"val_loss": null,
|
| 63 |
+
"val_dice": null,
|
| 64 |
+
"val_iou": null,
|
| 65 |
+
"val_reward": null,
|
| 66 |
+
"val_entropy": null
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"epoch": 4,
|
| 70 |
+
"train_loss": 0.5775981763099833,
|
| 71 |
+
"train_actor_loss": -0.3896141921095454,
|
| 72 |
+
"train_critic_loss": 0.0005236226093721229,
|
| 73 |
+
"train_mean_reward": -0.006890872759438145,
|
| 74 |
+
"train_entropy": 0.6920760656666078,
|
| 75 |
+
"train_ce_loss": 0.6215853797445117,
|
| 76 |
+
"train_dice_loss": 1.3123157159918653,
|
| 77 |
+
"train_dice": 0.5254657108870483,
|
| 78 |
+
"train_iou": 0.4044469126712038,
|
| 79 |
+
"grad_norm": 5.1005431104076004,
|
| 80 |
+
"lr": 0.001992154127807906,
|
| 81 |
+
"encoder_lr": 0.0001992508966248754,
|
| 82 |
+
"alpha": 0.45865100622177124,
|
| 83 |
+
"validated_this_epoch": false,
|
| 84 |
+
"val_loss": null,
|
| 85 |
+
"val_dice": null,
|
| 86 |
+
"val_iou": null,
|
| 87 |
+
"val_reward": null,
|
| 88 |
+
"val_entropy": null
|
| 89 |
+
},
|
| 90 |
+
{
|
| 91 |
+
"epoch": 5,
|
| 92 |
+
"train_loss": 0.5468145863427384,
|
| 93 |
+
"train_actor_loss": -0.38164395167076004,
|
| 94 |
+
"train_critic_loss": 0.00048421882295532684,
|
| 95 |
+
"train_mean_reward": -0.007146549918554971,
|
| 96 |
+
"train_entropy": 0.6698939134330713,
|
| 97 |
+
"train_ce_loss": 0.5763124398278308,
|
| 98 |
+
"train_dice_loss": 1.2801204001102646,
|
| 99 |
+
"train_dice": 0.5299657847976745,
|
| 100 |
+
"train_iou": 0.4063318969356856,
|
| 101 |
+
"grad_norm": 4.392710255097973,
|
| 102 |
+
"lr": 0.0019877498988921626,
|
| 103 |
+
"encoder_lr": 0.0001988303923565381,
|
| 104 |
+
"alpha": 0.44845375418663025,
|
| 105 |
+
"validated_this_epoch": false,
|
| 106 |
+
"val_loss": null,
|
| 107 |
+
"val_dice": null,
|
| 108 |
+
"val_iou": null,
|
| 109 |
+
"val_reward": null,
|
| 110 |
+
"val_entropy": null
|
| 111 |
+
},
|
| 112 |
+
{
|
| 113 |
+
"epoch": 6,
|
| 114 |
+
"train_loss": 0.5299756855581158,
|
| 115 |
+
"train_actor_loss": -0.3532222462793962,
|
| 116 |
+
"train_critic_loss": 0.0004485518227795197,
|
| 117 |
+
"train_mean_reward": -0.006876298742529678,
|
| 118 |
+
"train_entropy": 0.6121507645500414,
|
| 119 |
+
"train_ce_loss": 0.5197163113267751,
|
| 120 |
+
"train_dice_loss": 1.2462309839189514,
|
| 121 |
+
"train_dice": 0.5856144546254541,
|
| 122 |
+
"train_iou": 0.4611405266922413,
|
| 123 |
+
"grad_norm": 4.431029622868974,
|
| 124 |
+
"lr": 0.0019823758144750458,
|
| 125 |
+
"encoder_lr": 0.00019831728881922545,
|
| 126 |
+
"alpha": 0.43994972109794617,
|
| 127 |
+
"validated_this_epoch": false,
|
| 128 |
+
"val_loss": null,
|
| 129 |
+
"val_dice": null,
|
| 130 |
+
"val_iou": null,
|
| 131 |
+
"val_reward": null,
|
| 132 |
+
"val_entropy": null
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"epoch": 7,
|
| 136 |
+
"train_loss": 0.5362654167819451,
|
| 137 |
+
"train_actor_loss": -0.3542120541164348,
|
| 138 |
+
"train_critic_loss": 0.00042636872290401843,
|
| 139 |
+
"train_mean_reward": -0.006879624508021448,
|
| 140 |
+
"train_entropy": 0.6349980977077508,
|
| 141 |
+
"train_ce_loss": 0.5355668826141812,
|
| 142 |
+
"train_dice_loss": 1.2449616744382743,
|
| 143 |
+
"train_dice": 0.5829077830163721,
|
| 144 |
+
"train_iou": 0.4571424068714386,
|
| 145 |
+
"grad_norm": 4.297189357668855,
|
| 146 |
+
"lr": 0.001976037178129054,
|
| 147 |
+
"encoder_lr": 0.00019771209238418101,
|
| 148 |
+
"alpha": 0.4306834638118744,
|
| 149 |
+
"validated_this_epoch": false,
|
| 150 |
+
"val_loss": null,
|
| 151 |
+
"val_dice": null,
|
| 152 |
+
"val_iou": null,
|
| 153 |
+
"val_reward": null,
|
| 154 |
+
"val_entropy": null
|
| 155 |
+
},
|
| 156 |
+
{
|
| 157 |
+
"epoch": 8,
|
| 158 |
+
"train_loss": 0.5009987314643208,
|
| 159 |
+
"train_actor_loss": -0.35848521459510774,
|
| 160 |
+
"train_critic_loss": 0.00037639240759716806,
|
| 161 |
+
"train_mean_reward": -0.006403236930590672,
|
| 162 |
+
"train_entropy": 0.6353201389466762,
|
| 163 |
+
"train_ce_loss": 0.4869496908125489,
|
| 164 |
+
"train_dice_loss": 1.231641791818678,
|
| 165 |
+
"train_dice": 0.605431755904745,
|
| 166 |
+
"train_iou": 0.4807629606286295,
|
| 167 |
+
"grad_norm": 3.9353794926820798,
|
| 168 |
+
"lr": 0.001968740245322988,
|
| 169 |
+
"encoder_lr": 0.00019701540030721995,
|
| 170 |
+
"alpha": 0.4216455817222595,
|
| 171 |
+
"validated_this_epoch": false,
|
| 172 |
+
"val_loss": null,
|
| 173 |
+
"val_dice": null,
|
| 174 |
+
"val_iou": null,
|
| 175 |
+
"val_reward": null,
|
| 176 |
+
"val_entropy": null
|
| 177 |
+
},
|
| 178 |
+
{
|
| 179 |
+
"epoch": 9,
|
| 180 |
+
"train_loss": 0.5318035671068716,
|
| 181 |
+
"train_actor_loss": -0.33093265714532955,
|
| 182 |
+
"train_critic_loss": 0.00040436069585626745,
|
| 183 |
+
"train_mean_reward": -0.006258622953487071,
|
| 184 |
+
"train_entropy": 0.6069376353723134,
|
| 185 |
+
"train_ce_loss": 0.5012540366182956,
|
| 186 |
+
"train_dice_loss": 1.223814037188079,
|
| 187 |
+
"train_dice": 0.6171918115066974,
|
| 188 |
+
"train_iou": 0.49165663630129236,
|
| 189 |
+
"grad_norm": 3.9776635521142056,
|
| 190 |
+
"lr": 0.0019604922172485584,
|
| 191 |
+
"encoder_lr": 0.0001962279001393096,
|
| 192 |
+
"alpha": 0.41327470541000366,
|
| 193 |
+
"validated_this_epoch": false,
|
| 194 |
+
"val_loss": null,
|
| 195 |
+
"val_dice": null,
|
| 196 |
+
"val_iou": null,
|
| 197 |
+
"val_reward": null,
|
| 198 |
+
"val_entropy": null
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"epoch": 10,
|
| 202 |
+
"train_loss": 0.5134964085566674,
|
| 203 |
+
"train_actor_loss": -0.31596932694970487,
|
| 204 |
+
"train_critic_loss": 0.00036497542624867483,
|
| 205 |
+
"train_mean_reward": -0.006346295277258827,
|
| 206 |
+
"train_entropy": 0.5870908589826691,
|
| 207 |
+
"train_ce_loss": 0.4594741934640887,
|
| 208 |
+
"train_dice_loss": 1.1990922873488383,
|
| 209 |
+
"train_dice": 0.6324887301466181,
|
| 210 |
+
"train_iou": 0.510605483476854,
|
| 211 |
+
"grad_norm": 3.6920520679433215,
|
| 212 |
+
"lr": 0.0019513012337136779,
|
| 213 |
+
"encoder_lr": 0.00019535036904803958,
|
| 214 |
+
"alpha": 0.4051162302494049,
|
| 215 |
+
"validated_this_epoch": true,
|
| 216 |
+
"val_loss": 0.8344497016955619,
|
| 217 |
+
"val_dice": 0.6385428847994628,
|
| 218 |
+
"val_iou": 0.5303423438900624,
|
| 219 |
+
"val_reward": 0.0001657725060519596,
|
| 220 |
+
"val_entropy": 0.36911067306393325,
|
| 221 |
+
"val_actor_loss": -0.006723731971402286,
|
| 222 |
+
"val_critic_loss": 0.0001622746943423702,
|
| 223 |
+
"val_ce_loss": 0.3789208377993694,
|
| 224 |
+
"val_dice_loss": 1.3032637238502502
|
| 225 |
+
},
|
| 226 |
+
{
|
| 227 |
+
"epoch": 11,
|
| 228 |
+
"train_loss": 0.5156139166007728,
|
| 229 |
+
"train_actor_loss": -0.3182427108441828,
|
| 230 |
+
"train_critic_loss": 0.0003438376285723012,
|
| 231 |
+
"train_mean_reward": -0.006143908537531162,
|
| 232 |
+
"train_entropy": 0.5916705061067906,
|
| 233 |
+
"train_ce_loss": 0.46640546291282053,
|
| 234 |
+
"train_dice_loss": 1.2009639364188338,
|
| 235 |
+
"train_dice": 0.6366947475886162,
|
| 236 |
+
"train_iou": 0.5159599563208375,
|
| 237 |
+
"grad_norm": 3.676837276580722,
|
| 238 |
+
"lr": 0.0019411763651094546,
|
| 239 |
+
"encoder_lr": 0.0001943836730506514,
|
| 240 |
+
"alpha": 0.39782804250717163,
|
| 241 |
+
"validated_this_epoch": false,
|
| 242 |
+
"val_loss": null,
|
| 243 |
+
"val_dice": null,
|
| 244 |
+
"val_iou": null,
|
| 245 |
+
"val_reward": null,
|
| 246 |
+
"val_entropy": null
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"epoch": 12,
|
| 250 |
+
"train_loss": 0.5145805713652034,
|
| 251 |
+
"train_actor_loss": -0.2632487918844673,
|
| 252 |
+
"train_critic_loss": 0.0002918824380516483,
|
| 253 |
+
"train_mean_reward": -0.0041082983453218395,
|
| 254 |
+
"train_entropy": 0.5236228445318627,
|
| 255 |
+
"train_ce_loss": 0.48962535416149083,
|
| 256 |
+
"train_dice_loss": 1.0657414714649356,
|
| 257 |
+
"train_dice": 0.6580954108772538,
|
| 258 |
+
"train_iou": 0.5413229830796763,
|
| 259 |
+
"grad_norm": 6.757707894772523,
|
| 260 |
+
"lr": 0.0019301276034588103,
|
| 261 |
+
"encoder_lr": 0.00019332876615938386,
|
| 262 |
+
"alpha": 0.3916478455066681,
|
| 263 |
+
"validated_this_epoch": false,
|
| 264 |
+
"val_loss": null,
|
| 265 |
+
"val_dice": null,
|
| 266 |
+
"val_iou": null,
|
| 267 |
+
"val_reward": null,
|
| 268 |
+
"val_entropy": null
|
| 269 |
+
},
|
| 270 |
+
{
|
| 271 |
+
"epoch": 13,
|
| 272 |
+
"train_loss": 0.5047390253076424,
|
| 273 |
+
"train_actor_loss": -0.16654624431382006,
|
| 274 |
+
"train_critic_loss": 0.00018036744469624163,
|
| 275 |
+
"train_mean_reward": -0.00013016532307413397,
|
| 276 |
+
"train_entropy": 0.429875893436507,
|
| 277 |
+
"train_ce_loss": 0.5942824843907559,
|
| 278 |
+
"train_dice_loss": 0.7481076699664734,
|
| 279 |
+
"train_dice": 0.7132504556492669,
|
| 280 |
+
"train_iou": 0.6111882301003592,
|
| 281 |
+
"grad_norm": 18.139413707940154,
|
| 282 |
+
"lr": 0.0019181658525555612,
|
| 283 |
+
"encoder_lr": 0.00019218668943997818,
|
| 284 |
+
"alpha": 0.38846951723098755,
|
| 285 |
+
"validated_this_epoch": false,
|
| 286 |
+
"val_loss": null,
|
| 287 |
+
"val_dice": null,
|
| 288 |
+
"val_iou": null,
|
| 289 |
+
"val_reward": null,
|
| 290 |
+
"val_entropy": null
|
| 291 |
+
},
|
| 292 |
+
{
|
| 293 |
+
"epoch": 14,
|
| 294 |
+
"train_loss": 0.4850627956785579,
|
| 295 |
+
"train_actor_loss": -0.17186715003844227,
|
| 296 |
+
"train_critic_loss": 0.0001725014724489937,
|
| 297 |
+
"train_mean_reward": -0.00044016669854436463,
|
| 298 |
+
"train_entropy": 0.4447379368200936,
|
| 299 |
+
"train_ce_loss": 0.5983741284993018,
|
| 300 |
+
"train_dice_loss": 0.7153132416943248,
|
| 301 |
+
"train_dice": 0.7160477751170792,
|
| 302 |
+
"train_iou": 0.610836328621273,
|
| 303 |
+
"grad_norm": 18.665997648423957,
|
| 304 |
+
"lr": 0.0019053029172036895,
|
| 305 |
+
"encoder_lr": 0.00019095856998427185,
|
| 306 |
+
"alpha": 0.38468822836875916,
|
| 307 |
+
"validated_this_epoch": false,
|
| 308 |
+
"val_loss": null,
|
| 309 |
+
"val_dice": null,
|
| 310 |
+
"val_iou": null,
|
| 311 |
+
"val_reward": null,
|
| 312 |
+
"val_entropy": null
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"epoch": 15,
|
| 316 |
+
"train_loss": 0.48023741434386014,
|
| 317 |
+
"train_actor_loss": -0.16573426168825894,
|
| 318 |
+
"train_critic_loss": 0.00016568681638993665,
|
| 319 |
+
"train_mean_reward": -9.000337410915838e-05,
|
| 320 |
+
"train_entropy": 0.44601188745417025,
|
| 321 |
+
"train_ce_loss": 0.5893520660390379,
|
| 322 |
+
"train_dice_loss": 0.7024255864687976,
|
| 323 |
+
"train_dice": 0.7230108907884883,
|
| 324 |
+
"train_iou": 0.6224435750851727,
|
| 325 |
+
"grad_norm": 31.583990398303484,
|
| 326 |
+
"lr": 0.001891551491567426,
|
| 327 |
+
"encoder_lr": 0.00018964561979789495,
|
| 328 |
+
"alpha": 0.38058730959892273,
|
| 329 |
+
"validated_this_epoch": false,
|
| 330 |
+
"val_loss": null,
|
| 331 |
+
"val_dice": null,
|
| 332 |
+
"val_iou": null,
|
| 333 |
+
"val_reward": null,
|
| 334 |
+
"val_entropy": null
|
| 335 |
+
},
|
| 336 |
+
{
|
| 337 |
+
"epoch": 16,
|
| 338 |
+
"train_loss": 0.46395173364196174,
|
| 339 |
+
"train_actor_loss": -0.17495667614832097,
|
| 340 |
+
"train_critic_loss": 0.00015448492764189538,
|
| 341 |
+
"train_mean_reward": -1.0957011177279541e-05,
|
| 342 |
+
"train_entropy": 0.4802844808930266,
|
| 343 |
+
"train_ce_loss": 0.6096916542309083,
|
| 344 |
+
"train_dice_loss": 0.6679706665126542,
|
| 345 |
+
"train_dice": 0.7416586365565054,
|
| 346 |
+
"train_iou": 0.6435737471071711,
|
| 347 |
+
"grad_norm": 25.312977646672447,
|
| 348 |
+
"lr": 0.0018769251466436443,
|
| 349 |
+
"encoder_lr": 0.00018824913460416704,
|
| 350 |
+
"alpha": 0.37529194355010986,
|
| 351 |
+
"validated_this_epoch": false,
|
| 352 |
+
"val_loss": null,
|
| 353 |
+
"val_dice": null,
|
| 354 |
+
"val_iou": null,
|
| 355 |
+
"val_reward": null,
|
| 356 |
+
"val_entropy": null
|
| 357 |
+
},
|
| 358 |
+
{
|
| 359 |
+
"epoch": 17,
|
| 360 |
+
"train_loss": 0.4351690307705367,
|
| 361 |
+
"train_actor_loss": -0.17212123990309178,
|
| 362 |
+
"train_critic_loss": 0.00016236738317277363,
|
| 363 |
+
"train_mean_reward": -5.417108564583721e-05,
|
| 364 |
+
"train_entropy": 0.4659447143051667,
|
| 365 |
+
"train_ce_loss": 0.5712283815302975,
|
| 366 |
+
"train_dice_loss": 0.6431897796247664,
|
| 367 |
+
"train_dice": 0.7381148313393834,
|
| 368 |
+
"train_iou": 0.639966572570273,
|
| 369 |
+
"grad_norm": 22.1101633005364,
|
| 370 |
+
"lr": 0.001861438316868924,
|
| 371 |
+
"encoder_lr": 0.00018677049256537467,
|
| 372 |
+
"alpha": 0.3701269328594208,
|
| 373 |
+
"validated_this_epoch": false,
|
| 374 |
+
"val_loss": null,
|
| 375 |
+
"val_dice": null,
|
| 376 |
+
"val_iou": null,
|
| 377 |
+
"val_reward": null,
|
| 378 |
+
"val_entropy": null
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"epoch": 18,
|
| 382 |
+
"train_loss": 0.4355678272437189,
|
| 383 |
+
"train_actor_loss": -0.1728742900244333,
|
| 384 |
+
"train_critic_loss": 0.00013966676925879573,
|
| 385 |
+
"train_mean_reward": -0.0001687893265439803,
|
| 386 |
+
"train_entropy": 0.4710097816755913,
|
| 387 |
+
"train_ce_loss": 0.5869353894356071,
|
| 388 |
+
"train_dice_loss": 0.629809167804028,
|
| 389 |
+
"train_dice": 0.7374954694755017,
|
| 390 |
+
"train_iou": 0.6398149874030777,
|
| 391 |
+
"grad_norm": 24.773016451865203,
|
| 392 |
+
"lr": 0.0018451062858745053,
|
| 393 |
+
"encoder_lr": 0.00018521115292269145,
|
| 394 |
+
"alpha": 0.3649460971355438,
|
| 395 |
+
"validated_this_epoch": false,
|
| 396 |
+
"val_loss": null,
|
| 397 |
+
"val_dice": null,
|
| 398 |
+
"val_iou": null,
|
| 399 |
+
"val_reward": null,
|
| 400 |
+
"val_entropy": null
|
| 401 |
+
},
|
| 402 |
+
{
|
| 403 |
+
"epoch": 19,
|
| 404 |
+
"train_loss": 0.41195738685866473,
|
| 405 |
+
"train_actor_loss": -0.1688926805397953,
|
| 406 |
+
"train_critic_loss": 0.000141800267675072,
|
| 407 |
+
"train_mean_reward": 7.073909302834854e-05,
|
| 408 |
+
"train_entropy": 0.4676288670309139,
|
| 409 |
+
"train_ce_loss": 0.5546417442458449,
|
| 410 |
+
"train_dice_loss": 0.6069165772399854,
|
| 411 |
+
"train_dice": 0.76020220987176,
|
| 412 |
+
"train_iou": 0.6643084470266617,
|
| 413 |
+
"grad_norm": 25.145223339860753,
|
| 414 |
+
"lr": 0.0018279451714031893,
|
| 415 |
+
"encoder_lr": 0.0001835726545560834,
|
| 416 |
+
"alpha": 0.3596271872520447,
|
| 417 |
+
"validated_this_epoch": false,
|
| 418 |
+
"val_loss": null,
|
| 419 |
+
"val_dice": null,
|
| 420 |
+
"val_iou": null,
|
| 421 |
+
"val_reward": null,
|
| 422 |
+
"val_entropy": null
|
| 423 |
+
},
|
| 424 |
+
{
|
| 425 |
+
"epoch": 20,
|
| 426 |
+
"train_loss": 0.43063712093253553,
|
| 427 |
+
"train_actor_loss": -0.15816791046131645,
|
| 428 |
+
"train_critic_loss": 0.00016382439744402362,
|
| 429 |
+
"train_mean_reward": 0.00015913972296854013,
|
| 430 |
+
"train_entropy": 0.4444377262347428,
|
| 431 |
+
"train_ce_loss": 0.5332112458127837,
|
| 432 |
+
"train_dice_loss": 0.6442349787214314,
|
| 433 |
+
"train_dice": 0.7445364356554776,
|
| 434 |
+
"train_iou": 0.6469754479546069,
|
| 435 |
+
"grad_norm": 27.981427448664526,
|
| 436 |
+
"lr": 0.001809971909403073,
|
| 437 |
+
"encoder_lr": 0.00018185661446562003,
|
| 438 |
+
"alpha": 0.35483115911483765,
|
| 439 |
+
"validated_this_epoch": true,
|
| 440 |
+
"val_loss": 0.7518900431248106,
|
| 441 |
+
"val_dice": 0.6884455044681402,
|
| 442 |
+
"val_iou": 0.5833100282514664,
|
| 443 |
+
"val_reward": 0.0010663757622677798,
|
| 444 |
+
"val_entropy": 0.5568547570946241,
|
| 445 |
+
"val_actor_loss": -0.0034513426909336085,
|
| 446 |
+
"val_critic_loss": 0.00016491248219025316,
|
| 447 |
+
"val_ce_loss": 0.678802298793494,
|
| 448 |
+
"val_dice_loss": 0.831715530819363
|
| 449 |
+
},
|
| 450 |
+
{
|
| 451 |
+
"epoch": 21,
|
| 452 |
+
"train_loss": 0.40197463798004585,
|
| 453 |
+
"train_actor_loss": -0.15420026087758038,
|
| 454 |
+
"train_critic_loss": 0.00013672358704811452,
|
| 455 |
+
"train_mean_reward": 0.00019549954704792467,
|
| 456 |
+
"train_entropy": 0.43547706815199955,
|
| 457 |
+
"train_ce_loss": 0.5186191646078991,
|
| 458 |
+
"train_dice_loss": 0.5935938948962732,
|
| 459 |
+
"train_dice": 0.7701566209325434,
|
| 460 |
+
"train_iou": 0.6728883171385585,
|
| 461 |
+
"grad_norm": 40.50493894991025,
|
| 462 |
+
"lr": 0.001791204237313812,
|
| 463 |
+
"encoder_lr": 0.0001800647261756906,
|
| 464 |
+
"alpha": 0.3506215512752533,
|
| 465 |
+
"validated_this_epoch": false,
|
| 466 |
+
"val_loss": null,
|
| 467 |
+
"val_dice": null,
|
| 468 |
+
"val_iou": null,
|
| 469 |
+
"val_reward": null,
|
| 470 |
+
"val_entropy": null
|
| 471 |
+
},
|
| 472 |
+
{
|
| 473 |
+
"epoch": 22,
|
| 474 |
+
"train_loss": 0.4244889148164454,
|
| 475 |
+
"train_actor_loss": -0.15961132275580742,
|
| 476 |
+
"train_critic_loss": 0.00015315577103826495,
|
| 477 |
+
"train_mean_reward": 0.0001422864273948552,
|
| 478 |
+
"train_entropy": 0.46247370683561617,
|
| 479 |
+
"train_ce_loss": 0.5451043724799554,
|
| 480 |
+
"train_dice_loss": 0.6229429338175506,
|
| 481 |
+
"train_dice": 0.755411448006958,
|
| 482 |
+
"train_iou": 0.66240800762379,
|
| 483 |
+
"grad_norm": 31.25583576878836,
|
| 484 |
+
"lr": 0.0017716606765619107,
|
| 485 |
+
"encoder_lr": 0.0001781987580637,
|
| 486 |
+
"alpha": 0.34514111280441284,
|
| 487 |
+
"validated_this_epoch": false,
|
| 488 |
+
"val_loss": null,
|
| 489 |
+
"val_dice": null,
|
| 490 |
+
"val_iou": null,
|
| 491 |
+
"val_reward": null,
|
| 492 |
+
"val_entropy": null
|
| 493 |
+
},
|
| 494 |
+
{
|
| 495 |
+
"epoch": 23,
|
| 496 |
+
"train_loss": 0.4078451221043518,
|
| 497 |
+
"train_actor_loss": -0.1503829905758455,
|
| 498 |
+
"train_critic_loss": 0.0001516331442599274,
|
| 499 |
+
"train_mean_reward": 0.000306969906566598,
|
| 500 |
+
"train_entropy": 0.445018073924136,
|
| 501 |
+
"train_ce_loss": 0.5208075326838524,
|
| 502 |
+
"train_dice_loss": 0.5954970474082987,
|
| 503 |
+
"train_dice": 0.7709046558529737,
|
| 504 |
+
"train_iou": 0.6741355455424975,
|
| 505 |
+
"grad_norm": 40.022170542515525,
|
| 506 |
+
"lr": 0.0017513605142823077,
|
| 507 |
+
"encoder_lr": 0.00017626055161489368,
|
| 508 |
+
"alpha": 0.3399776220321655,
|
| 509 |
+
"validated_this_epoch": false,
|
| 510 |
+
"val_loss": null,
|
| 511 |
+
"val_dice": null,
|
| 512 |
+
"val_iou": null,
|
| 513 |
+
"val_reward": null,
|
| 514 |
+
"val_entropy": null
|
| 515 |
+
},
|
| 516 |
+
{
|
| 517 |
+
"epoch": 24,
|
| 518 |
+
"train_loss": 0.39490715295163004,
|
| 519 |
+
"train_actor_loss": -0.1505500030552232,
|
| 520 |
+
"train_critic_loss": 0.00014347390645664247,
|
| 521 |
+
"train_mean_reward": 0.00014315298615254633,
|
| 522 |
+
"train_entropy": 0.45467633238057453,
|
| 523 |
+
"train_ce_loss": 0.5284508588358048,
|
| 524 |
+
"train_dice_loss": 0.5623199655566105,
|
| 525 |
+
"train_dice": 0.7720783138065057,
|
| 526 |
+
"train_iou": 0.6755921312407842,
|
| 527 |
+
"grad_norm": 27.77809438917988,
|
| 528 |
+
"lr": 0.001730323784284305,
|
| 529 |
+
"encoder_lr": 0.0001742520196050341,
|
| 530 |
+
"alpha": 0.33493903279304504,
|
| 531 |
+
"validated_this_epoch": false,
|
| 532 |
+
"val_loss": null,
|
| 533 |
+
"val_dice": null,
|
| 534 |
+
"val_iou": null,
|
| 535 |
+
"val_reward": null,
|
| 536 |
+
"val_entropy": null
|
| 537 |
+
},
|
| 538 |
+
{
|
| 539 |
+
"epoch": 25,
|
| 540 |
+
"train_loss": 0.40324637336410984,
|
| 541 |
+
"train_actor_loss": -0.1355495804172365,
|
| 542 |
+
"train_critic_loss": 0.000171354220912469,
|
| 543 |
+
"train_mean_reward": -8.121457401246175e-05,
|
| 544 |
+
"train_entropy": 0.40855037110065917,
|
| 545 |
+
"train_ce_loss": 0.504277900141695,
|
| 546 |
+
"train_dice_loss": 0.5731426391182637,
|
| 547 |
+
"train_dice": 0.7690523967955781,
|
| 548 |
+
"train_iou": 0.67507350921107,
|
| 549 |
+
"grad_norm": 27.55696317022161,
|
| 550 |
+
"lr": 0.0017085712472806151,
|
| 551 |
+
"encoder_lr": 0.000172175144212722,
|
| 552 |
+
"alpha": 0.3312670588493347,
|
| 553 |
+
"validated_this_epoch": false,
|
| 554 |
+
"val_loss": null,
|
| 555 |
+
"val_dice": null,
|
| 556 |
+
"val_iou": null,
|
| 557 |
+
"val_reward": null,
|
| 558 |
+
"val_entropy": null
|
| 559 |
+
},
|
| 560 |
+
{
|
| 561 |
+
"epoch": 26,
|
| 562 |
+
"train_loss": 0.37431245369750055,
|
| 563 |
+
"train_actor_loss": -0.14362630397273898,
|
| 564 |
+
"train_critic_loss": 0.00015612975146098198,
|
| 565 |
+
"train_mean_reward": 0.00020012743425703227,
|
| 566 |
+
"train_entropy": 0.4472269575070657,
|
| 567 |
+
"train_ce_loss": 0.5012843235496922,
|
| 568 |
+
"train_dice_loss": 0.5344370500369897,
|
| 569 |
+
"train_dice": 0.7905216598188005,
|
| 570 |
+
"train_iou": 0.6951829784110801,
|
| 571 |
+
"grad_norm": 26.276891073515248,
|
| 572 |
+
"lr": 0.0016861243703990456,
|
| 573 |
+
"encoder_lr": 0.00017003197506322543,
|
| 574 |
+
"alpha": 0.32607302069664,
|
| 575 |
+
"validated_this_epoch": false,
|
| 576 |
+
"val_loss": null,
|
| 577 |
+
"val_dice": null,
|
| 578 |
+
"val_iou": null,
|
| 579 |
+
"val_reward": null,
|
| 580 |
+
"val_entropy": null
|
| 581 |
+
},
|
| 582 |
+
{
|
| 583 |
+
"epoch": 27,
|
| 584 |
+
"train_loss": 0.37650126458665484,
|
| 585 |
+
"train_actor_loss": -0.14028941472698706,
|
| 586 |
+
"train_critic_loss": 0.00013231175536579926,
|
| 587 |
+
"train_mean_reward": 0.00013551521741827827,
|
| 588 |
+
"train_entropy": 0.4419891460959917,
|
| 589 |
+
"train_ce_loss": 0.5019474842739639,
|
| 590 |
+
"train_dice_loss": 0.5315015493899353,
|
| 591 |
+
"train_dice": 0.7890566176586681,
|
| 592 |
+
"train_iou": 0.6942398766809964,
|
| 593 |
+
"grad_norm": 34.396863994210264,
|
| 594 |
+
"lr": 0.001663005305997034,
|
| 595 |
+
"encoder_lr": 0.00016782462720574693,
|
| 596 |
+
"alpha": 0.32104143500328064,
|
| 597 |
+
"validated_this_epoch": false,
|
| 598 |
+
"val_loss": null,
|
| 599 |
+
"val_dice": null,
|
| 600 |
+
"val_iou": null,
|
| 601 |
+
"val_reward": null,
|
| 602 |
+
"val_entropy": null
|
| 603 |
+
},
|
| 604 |
+
{
|
| 605 |
+
"epoch": 28,
|
| 606 |
+
"train_loss": 0.342205090193771,
|
| 607 |
+
"train_actor_loss": -0.13663807302925166,
|
| 608 |
+
"train_critic_loss": 0.00016855171590227642,
|
| 609 |
+
"train_mean_reward": 0.00023048005183575178,
|
| 610 |
+
"train_entropy": 0.43396931108886266,
|
| 611 |
+
"train_ce_loss": 0.47401768678332407,
|
| 612 |
+
"train_dice_loss": 0.48350007651080146,
|
| 613 |
+
"train_dice": 0.8068621829417653,
|
| 614 |
+
"train_iou": 0.709923920172376,
|
| 615 |
+
"grad_norm": 26.462826338849325,
|
| 616 |
+
"lr": 0.0016392368697999468,
|
| 617 |
+
"encoder_lr": 0.00016555527902612554,
|
| 618 |
+
"alpha": 0.3162289261817932,
|
| 619 |
+
"validated_this_epoch": false,
|
| 620 |
+
"val_loss": null,
|
| 621 |
+
"val_dice": null,
|
| 622 |
+
"val_iou": null,
|
| 623 |
+
"val_reward": null,
|
| 624 |
+
"val_entropy": null
|
| 625 |
+
},
|
| 626 |
+
{
|
| 627 |
+
"epoch": 29,
|
| 628 |
+
"train_loss": 0.36697511156739365,
|
| 629 |
+
"train_actor_loss": -0.1320971890812311,
|
| 630 |
+
"train_critic_loss": 0.00015095479314553936,
|
| 631 |
+
"train_mean_reward": 0.00016696358618099898,
|
| 632 |
+
"train_entropy": 0.4294659935457762,
|
| 633 |
+
"train_ce_loss": 0.4829917902978816,
|
| 634 |
+
"train_dice_loss": 0.5150018467613583,
|
| 635 |
+
"train_dice": 0.7924489548300423,
|
| 636 |
+
"train_iou": 0.6975093194421245,
|
| 637 |
+
"grad_norm": 23.21526327151661,
|
| 638 |
+
"lr": 0.0016148425183847121,
|
| 639 |
+
"encoder_lr": 0.0001632261700970328,
|
| 640 |
+
"alpha": 0.31147441267967224,
|
| 641 |
+
"validated_this_epoch": false,
|
| 642 |
+
"val_loss": null,
|
| 643 |
+
"val_dice": null,
|
| 644 |
+
"val_iou": null,
|
| 645 |
+
"val_reward": null,
|
| 646 |
+
"val_entropy": null
|
| 647 |
+
},
|
| 648 |
+
{
|
| 649 |
+
"epoch": 30,
|
| 650 |
+
"train_loss": 0.3718450032427272,
|
| 651 |
+
"train_actor_loss": -0.1270626744457571,
|
| 652 |
+
"train_critic_loss": 0.00014299352383192925,
|
| 653 |
+
"train_mean_reward": -4.334966826285868e-06,
|
| 654 |
+
"train_entropy": 0.4187942345263268,
|
| 655 |
+
"train_ce_loss": 0.48375144478857746,
|
| 656 |
+
"train_dice_loss": 0.5139209054391206,
|
| 657 |
+
"train_dice": 0.7983629573922536,
|
| 658 |
+
"train_iou": 0.7065798277297906,
|
| 659 |
+
"grad_norm": 24.116265101488246,
|
| 660 |
+
"lr": 0.0015898463260310114,
|
| 661 |
+
"encoder_lr": 0.00016083959896778498,
|
| 662 |
+
"alpha": 0.3072684705257416,
|
| 663 |
+
"validated_this_epoch": true,
|
| 664 |
+
"val_loss": 0.6818797474543592,
|
| 665 |
+
"val_dice": 0.7282493638639558,
|
| 666 |
+
"val_iou": 0.6272079187278855,
|
| 667 |
+
"val_reward": 0.00025058676670375985,
|
| 668 |
+
"val_entropy": 0.49221798506650055,
|
| 669 |
+
"val_actor_loss": -0.003987543890487892,
|
| 670 |
+
"val_critic_loss": 0.0002466327131482691,
|
| 671 |
+
"val_ce_loss": 0.6161228309648884,
|
| 672 |
+
"val_dice_loss": 0.7553650890335892
|
| 673 |
+
},
|
| 674 |
+
{
|
| 675 |
+
"epoch": 31,
|
| 676 |
+
"train_loss": 0.34562945270788126,
|
| 677 |
+
"train_actor_loss": -0.12674483904754621,
|
| 678 |
+
"train_critic_loss": 0.0001485620874513945,
|
| 679 |
+
"train_mean_reward": 0.0002652644534772796,
|
| 680 |
+
"train_entropy": 0.4227157024912132,
|
| 681 |
+
"train_ce_loss": 0.4586306042561323,
|
| 682 |
+
"train_dice_loss": 0.4859694053806384,
|
| 683 |
+
"train_dice": 0.8037790677423979,
|
| 684 |
+
"train_iou": 0.7100024787500162,
|
| 685 |
+
"grad_norm": 23.475208647953448,
|
| 686 |
+
"lr": 0.0015642729609628705,
|
| 687 |
+
"encoder_lr": 0.00015839792089595244,
|
| 688 |
+
"alpha": 0.3029424548149109,
|
| 689 |
+
"validated_this_epoch": false,
|
| 690 |
+
"val_loss": null,
|
| 691 |
+
"val_dice": null,
|
| 692 |
+
"val_iou": null,
|
| 693 |
+
"val_reward": null,
|
| 694 |
+
"val_entropy": null
|
| 695 |
+
},
|
| 696 |
+
{
|
| 697 |
+
"epoch": 32,
|
| 698 |
+
"train_loss": 0.346733160493372,
|
| 699 |
+
"train_actor_loss": -0.12375980245974637,
|
| 700 |
+
"train_critic_loss": 0.00014669200047868453,
|
| 701 |
+
"train_mean_reward": 0.00020878418958636288,
|
| 702 |
+
"train_entropy": 0.42036574674669164,
|
| 703 |
+
"train_ce_loss": 0.4561642841911256,
|
| 704 |
+
"train_dice_loss": 0.48467493842738546,
|
| 705 |
+
"train_dice": 0.8082903978309892,
|
| 706 |
+
"train_iou": 0.715323538573202,
|
| 707 |
+
"grad_norm": 24.021718779275584,
|
| 708 |
+
"lr": 0.001538147661004102,
|
| 709 |
+
"encoder_lr": 0.0001559035455230047,
|
| 710 |
+
"alpha": 0.29842475056648254,
|
| 711 |
+
"validated_this_epoch": false,
|
| 712 |
+
"val_loss": null,
|
| 713 |
+
"val_dice": null,
|
| 714 |
+
"val_iou": null,
|
| 715 |
+
"val_reward": null,
|
| 716 |
+
"val_entropy": null
|
| 717 |
+
},
|
| 718 |
+
{
|
| 719 |
+
"epoch": 33,
|
| 720 |
+
"train_loss": 0.3549021499975072,
|
| 721 |
+
"train_actor_loss": -0.11947817770278116,
|
| 722 |
+
"train_critic_loss": 0.00013281131971084635,
|
| 723 |
+
"train_mean_reward": 0.00020401909182569033,
|
| 724 |
+
"train_entropy": 0.40790827424104203,
|
| 725 |
+
"train_ce_loss": 0.4520776262603103,
|
| 726 |
+
"train_dice_loss": 0.4965502072982394,
|
| 727 |
+
"train_dice": 0.7995483547117928,
|
| 728 |
+
"train_iou": 0.7089168485370477,
|
| 729 |
+
"grad_norm": 25.85705607159193,
|
| 730 |
+
"lr": 0.0015114962086716199,
|
| 731 |
+
"encoder_lr": 0.0001533589344962853,
|
| 732 |
+
"alpha": 0.2945246398448944,
|
| 733 |
+
"validated_this_epoch": false,
|
| 734 |
+
"val_loss": null,
|
| 735 |
+
"val_dice": null,
|
| 736 |
+
"val_iou": null,
|
| 737 |
+
"val_reward": null,
|
| 738 |
+
"val_entropy": null
|
| 739 |
+
},
|
| 740 |
+
{
|
| 741 |
+
"epoch": 34,
|
| 742 |
+
"train_loss": 0.3806022269535781,
|
| 743 |
+
"train_actor_loss": -0.117909142299909,
|
| 744 |
+
"train_critic_loss": 0.00015301836910269347,
|
| 745 |
+
"train_mean_reward": 5.937699507065553e-05,
|
| 746 |
+
"train_entropy": 0.41416192434015814,
|
| 747 |
+
"train_ce_loss": 0.4654722497690209,
|
| 748 |
+
"train_dice_loss": 0.5313974573476679,
|
| 749 |
+
"train_dice": 0.789701570989994,
|
| 750 |
+
"train_iou": 0.6965569582855109,
|
| 751 |
+
"grad_norm": 30.66891450521558,
|
| 752 |
+
"lr": 0.0014843449057312071,
|
| 753 |
+
"encoder_lr": 0.00015076659903966297,
|
| 754 |
+
"alpha": 0.2902570068836212,
|
| 755 |
+
"validated_this_epoch": false,
|
| 756 |
+
"val_loss": null,
|
| 757 |
+
"val_dice": null,
|
| 758 |
+
"val_iou": null,
|
| 759 |
+
"val_reward": null,
|
| 760 |
+
"val_entropy": null
|
| 761 |
+
},
|
| 762 |
+
{
|
| 763 |
+
"epoch": 35,
|
| 764 |
+
"train_loss": 0.3595721785397839,
|
| 765 |
+
"train_actor_loss": -0.1147364402762602,
|
| 766 |
+
"train_critic_loss": 0.00015738542347901126,
|
| 767 |
+
"train_mean_reward": 0.00014693824791943807,
|
| 768 |
+
"train_entropy": 0.40659246424294876,
|
| 769 |
+
"train_ce_loss": 0.45830702998985845,
|
| 770 |
+
"train_dice_loss": 0.4901528100480713,
|
| 771 |
+
"train_dice": 0.8036905463028813,
|
| 772 |
+
"train_iou": 0.7090619021739808,
|
| 773 |
+
"grad_norm": 22.89855726652367,
|
| 774 |
+
"lr": 0.0014567205472408495,
|
| 775 |
+
"encoder_lr": 0.00014812909747525697,
|
| 776 |
+
"alpha": 0.2862507104873657,
|
| 777 |
+
"validated_this_epoch": false,
|
| 778 |
+
"val_loss": null,
|
| 779 |
+
"val_dice": null,
|
| 780 |
+
"val_iou": null,
|
| 781 |
+
"val_reward": null,
|
| 782 |
+
"val_entropy": null
|
| 783 |
+
},
|
| 784 |
+
{
|
| 785 |
+
"epoch": 36,
|
| 786 |
+
"train_loss": 0.35618537070605066,
|
| 787 |
+
"train_actor_loss": -0.11205747378003382,
|
| 788 |
+
"train_critic_loss": 0.0001442429779989504,
|
| 789 |
+
"train_mean_reward": 0.00011645431835581883,
|
| 790 |
+
"train_entropy": 0.40403614841998386,
|
| 791 |
+
"train_ce_loss": 0.44645914905500006,
|
| 792 |
+
"train_dice_loss": 0.48988228451065935,
|
| 793 |
+
"train_dice": 0.8137279105324705,
|
| 794 |
+
"train_iou": 0.7202373197163956,
|
| 795 |
+
"grad_norm": 18.89398125068162,
|
| 796 |
+
"lr": 0.0014286503951072478,
|
| 797 |
+
"encoder_lr": 0.00014544903269868192,
|
| 798 |
+
"alpha": 0.2825221121311188,
|
| 799 |
+
"validated_this_epoch": false,
|
| 800 |
+
"val_loss": null,
|
| 801 |
+
"val_dice": null,
|
| 802 |
+
"val_iou": null,
|
| 803 |
+
"val_reward": null,
|
| 804 |
+
"val_entropy": null
|
| 805 |
+
},
|
| 806 |
+
{
|
| 807 |
+
"epoch": 37,
|
| 808 |
+
"train_loss": 0.36887315067628673,
|
| 809 |
+
"train_actor_loss": -0.10981759582754484,
|
| 810 |
+
"train_critic_loss": 0.00014274385064953762,
|
| 811 |
+
"train_mean_reward": 0.00011578047311583879,
|
| 812 |
+
"train_entropy": 0.3983500482423053,
|
| 813 |
+
"train_ce_loss": 0.4667677642989335,
|
| 814 |
+
"train_dice_loss": 0.49047097278811835,
|
| 815 |
+
"train_dice": 0.8042341401836179,
|
| 816 |
+
"train_iou": 0.714631068436458,
|
| 817 |
+
"grad_norm": 23.45157758672108,
|
| 818 |
+
"lr": 0.001400162151181607,
|
| 819 |
+
"encoder_lr": 0.00014272904961030417,
|
| 820 |
+
"alpha": 0.2789542078971863,
|
| 821 |
+
"validated_this_epoch": false,
|
| 822 |
+
"val_loss": null,
|
| 823 |
+
"val_dice": null,
|
| 824 |
+
"val_iou": null,
|
| 825 |
+
"val_reward": null,
|
| 826 |
+
"val_entropy": null
|
| 827 |
+
},
|
| 828 |
+
{
|
| 829 |
+
"epoch": 38,
|
| 830 |
+
"train_loss": 0.36314082727948016,
|
| 831 |
+
"train_actor_loss": -0.1088826876489845,
|
| 832 |
+
"train_critic_loss": 0.00015441986967746363,
|
| 833 |
+
"train_mean_reward": 9.395749091154523e-05,
|
| 834 |
+
"train_entropy": 0.4068067828698676,
|
| 835 |
+
"train_ce_loss": 0.4530124072832577,
|
| 836 |
+
"train_dice_loss": 0.49088018916990095,
|
| 837 |
+
"train_dice": 0.8032120628289456,
|
| 838 |
+
"train_iou": 0.7136228105632484,
|
| 839 |
+
"grad_norm": 33.02850089609161,
|
| 840 |
+
"lr": 0.0013712839299212553,
|
| 841 |
+
"encoder_lr": 0.00013997183250504444,
|
| 842 |
+
"alpha": 0.275070458650589,
|
| 843 |
+
"validated_this_epoch": false,
|
| 844 |
+
"val_loss": null,
|
| 845 |
+
"val_dice": null,
|
| 846 |
+
"val_iou": null,
|
| 847 |
+
"val_reward": null,
|
| 848 |
+
"val_entropy": null
|
| 849 |
+
},
|
| 850 |
+
{
|
| 851 |
+
"epoch": 39,
|
| 852 |
+
"train_loss": 0.3630618812198671,
|
| 853 |
+
"train_actor_loss": -0.10730836651160426,
|
| 854 |
+
"train_critic_loss": 0.0001339014913205233,
|
| 855 |
+
"train_mean_reward": 0.00015426125162409243,
|
| 856 |
+
"train_entropy": 0.3972595416644747,
|
| 857 |
+
"train_ce_loss": 0.4491404217020524,
|
| 858 |
+
"train_dice_loss": 0.49146615804319854,
|
| 859 |
+
"train_dice": 0.8137133171236035,
|
| 860 |
+
"train_iou": 0.724363623682219,
|
| 861 |
+
"grad_norm": 21.51526140935661,
|
| 862 |
+
"lr": 0.0013420442306440656,
|
| 863 |
+
"encoder_lr": 0.0001371801024233027,
|
| 864 |
+
"alpha": 0.27161359786987305,
|
| 865 |
+
"validated_this_epoch": false,
|
| 866 |
+
"val_loss": null,
|
| 867 |
+
"val_dice": null,
|
| 868 |
+
"val_iou": null,
|
| 869 |
+
"val_reward": null,
|
| 870 |
+
"val_entropy": null
|
| 871 |
+
},
|
| 872 |
+
{
|
| 873 |
+
"epoch": 40,
|
| 874 |
+
"train_loss": 0.34755232214725523,
|
| 875 |
+
"train_actor_loss": -0.1013733668551234,
|
| 876 |
+
"train_critic_loss": 0.00013690836689049074,
|
| 877 |
+
"train_mean_reward": 0.000142304112069284,
|
| 878 |
+
"train_entropy": 0.37945556334515873,
|
| 879 |
+
"train_ce_loss": 0.43229383392889986,
|
| 880 |
+
"train_dice_loss": 0.4654206228502653,
|
| 881 |
+
"train_dice": 0.8219442735483471,
|
| 882 |
+
"train_iou": 0.7326920701432271,
|
| 883 |
+
"grad_norm": 22.72048067492108,
|
| 884 |
+
"lr": 0.0013124719094030732,
|
| 885 |
+
"encoder_lr": 0.00013435661446562,
|
| 886 |
+
"alpha": 0.2688016891479492,
|
| 887 |
+
"validated_this_epoch": true,
|
| 888 |
+
"val_loss": 0.6280305419702816,
|
| 889 |
+
"val_dice": 0.7366531130935929,
|
| 890 |
+
"val_iou": 0.6427941490616009,
|
| 891 |
+
"val_reward": 7.572729919966243e-05,
|
| 892 |
+
"val_entropy": 0.30385184513800073,
|
| 893 |
+
"val_actor_loss": -0.00763518884989687,
|
| 894 |
+
"val_critic_loss": 0.000131644843735778,
|
| 895 |
+
"val_ce_loss": 0.5344147198207939,
|
| 896 |
+
"val_dice_loss": 0.7367850749781638
|
| 897 |
+
},
|
| 898 |
+
{
|
| 899 |
+
"epoch": 41,
|
| 900 |
+
"train_loss": 0.36695624943894817,
|
| 901 |
+
"train_actor_loss": -0.10163766559146019,
|
| 902 |
+
"train_critic_loss": 0.00013577004897142946,
|
| 903 |
+
"train_mean_reward": 0.0001936918581207486,
|
| 904 |
+
"train_entropy": 0.39098017560727216,
|
| 905 |
+
"train_ce_loss": 0.448513100820158,
|
| 906 |
+
"train_dice_loss": 0.4885389464462143,
|
| 907 |
+
"train_dice": 0.8074013005737727,
|
| 908 |
+
"train_iou": 0.7172632772047295,
|
| 909 |
+
"grad_norm": 20.533130052015764,
|
| 910 |
+
"lr": 0.0012825961505090336,
|
| 911 |
+
"encoder_lr": 0.00013150415507372682,
|
| 912 |
+
"alpha": 0.265962690114975,
|
| 913 |
+
"validated_this_epoch": false,
|
| 914 |
+
"val_loss": null,
|
| 915 |
+
"val_dice": null,
|
| 916 |
+
"val_iou": null,
|
| 917 |
+
"val_reward": null,
|
| 918 |
+
"val_entropy": null
|
| 919 |
+
},
|
| 920 |
+
{
|
| 921 |
+
"epoch": 42,
|
| 922 |
+
"train_loss": 0.3959062027983194,
|
| 923 |
+
"train_actor_loss": -0.09972424897574639,
|
| 924 |
+
"train_critic_loss": 0.00014400565343504484,
|
| 925 |
+
"train_mean_reward": 0.00018816262958545746,
|
| 926 |
+
"train_entropy": 0.37742192427163285,
|
| 927 |
+
"train_ce_loss": 0.4442636109154565,
|
| 928 |
+
"train_dice_loss": 0.5468532724127906,
|
| 929 |
+
"train_dice": 0.801511351849109,
|
| 930 |
+
"train_iou": 0.7106121973553181,
|
| 931 |
+
"grad_norm": 24.53680793629136,
|
| 932 |
+
"lr": 0.0012524464377290307,
|
| 933 |
+
"encoder_lr": 0.00012862553928066122,
|
| 934 |
+
"alpha": 0.26307323575019836,
|
| 935 |
+
"validated_this_epoch": false,
|
| 936 |
+
"val_loss": null,
|
| 937 |
+
"val_dice": null,
|
| 938 |
+
"val_iou": null,
|
| 939 |
+
"val_reward": null,
|
| 940 |
+
"val_entropy": null
|
| 941 |
+
},
|
| 942 |
+
{
|
| 943 |
+
"epoch": 43,
|
| 944 |
+
"train_loss": 0.3537987035287674,
|
| 945 |
+
"train_actor_loss": -0.09700544767601546,
|
| 946 |
+
"train_critic_loss": 0.00013367636373037543,
|
| 947 |
+
"train_mean_reward": 0.00020594160701001564,
|
| 948 |
+
"train_entropy": 0.37934615295524743,
|
| 949 |
+
"train_ce_loss": 0.42878445497899287,
|
| 950 |
+
"train_dice_loss": 0.47269016019133636,
|
| 951 |
+
"train_dice": 0.8202287179769641,
|
| 952 |
+
"train_iou": 0.7312704840856167,
|
| 953 |
+
"grad_norm": 18.355173633542172,
|
| 954 |
+
"lr": 0.0012220525251895604,
|
| 955 |
+
"encoder_lr": 0.0001257236079326716,
|
| 956 |
+
"alpha": 0.26052987575531006,
|
| 957 |
+
"validated_this_epoch": false,
|
| 958 |
+
"val_loss": null,
|
| 959 |
+
"val_dice": null,
|
| 960 |
+
"val_iou": null,
|
| 961 |
+
"val_reward": null,
|
| 962 |
+
"val_entropy": null
|
| 963 |
+
},
|
| 964 |
+
{
|
| 965 |
+
"epoch": 44,
|
| 966 |
+
"train_loss": 0.34118914987184396,
|
| 967 |
+
"train_actor_loss": -0.09695920137356417,
|
| 968 |
+
"train_critic_loss": 0.00013867552419274514,
|
| 969 |
+
"train_mean_reward": 0.00016291127502874813,
|
| 970 |
+
"train_entropy": 0.3807903510116483,
|
| 971 |
+
"train_ce_loss": 0.41309744910065094,
|
| 972 |
+
"train_dice_loss": 0.46306056816140506,
|
| 973 |
+
"train_dice": 0.8194147840803888,
|
| 974 |
+
"train_iou": 0.7298733038319848,
|
| 975 |
+
"grad_norm": 18.745351010514785,
|
| 976 |
+
"lr": 0.0011914444080127966,
|
| 977 |
+
"encoder_lr": 0.0001228012248856439,
|
| 978 |
+
"alpha": 0.25777170062065125,
|
| 979 |
+
"validated_this_epoch": false,
|
| 980 |
+
"val_loss": null,
|
| 981 |
+
"val_dice": null,
|
| 982 |
+
"val_iou": null,
|
| 983 |
+
"val_reward": null,
|
| 984 |
+
"val_entropy": null
|
| 985 |
+
},
|
| 986 |
+
{
|
| 987 |
+
"epoch": 45,
|
| 988 |
+
"train_loss": 0.32655436126876247,
|
| 989 |
+
"train_actor_loss": -0.09228693653807737,
|
| 990 |
+
"train_critic_loss": 0.00012424996059748862,
|
| 991 |
+
"train_mean_reward": 0.00020914869761205077,
|
| 992 |
+
"train_entropy": 0.3623525413759149,
|
| 993 |
+
"train_ce_loss": 0.3973827927640473,
|
| 994 |
+
"train_dice_loss": 0.44017554215066507,
|
| 995 |
+
"train_dice": 0.8323653453578531,
|
| 996 |
+
"train_iou": 0.7461590483647658,
|
| 997 |
+
"grad_norm": 19.149282676304956,
|
| 998 |
+
"lr": 0.0011606522927150302,
|
| 999 |
+
"encoder_lr": 0.00011986127417882199,
|
| 1000 |
+
"alpha": 0.2558322548866272,
|
| 1001 |
+
"validated_this_epoch": false,
|
| 1002 |
+
"val_loss": null,
|
| 1003 |
+
"val_dice": null,
|
| 1004 |
+
"val_iou": null,
|
| 1005 |
+
"val_reward": null,
|
| 1006 |
+
"val_entropy": null
|
| 1007 |
+
},
|
| 1008 |
+
{
|
| 1009 |
+
"epoch": 46,
|
| 1010 |
+
"train_loss": 0.34436306482643003,
|
| 1011 |
+
"train_actor_loss": -0.09292831365741039,
|
| 1012 |
+
"train_critic_loss": 0.00012774476153842117,
|
| 1013 |
+
"train_mean_reward": 0.00020433579859417824,
|
| 1014 |
+
"train_entropy": 0.36948704669487875,
|
| 1015 |
+
"train_ce_loss": 0.41186047286680816,
|
| 1016 |
+
"train_dice_loss": 0.4625945266057046,
|
| 1017 |
+
"train_dice": 0.8166965376021632,
|
| 1018 |
+
"train_iou": 0.7282293734060314,
|
| 1019 |
+
"grad_norm": 19.607700857543207,
|
| 1020 |
+
"lr": 0.0011297065673964834,
|
| 1021 |
+
"encoder_lr": 0.00011690665718860897,
|
| 1022 |
+
"alpha": 0.2535752058029175,
|
| 1023 |
+
"validated_this_epoch": false,
|
| 1024 |
+
"val_loss": null,
|
| 1025 |
+
"val_dice": null,
|
| 1026 |
+
"val_iou": null,
|
| 1027 |
+
"val_reward": null,
|
| 1028 |
+
"val_entropy": null
|
| 1029 |
+
},
|
| 1030 |
+
{
|
| 1031 |
+
"epoch": 47,
|
| 1032 |
+
"train_loss": 0.33155948482453823,
|
| 1033 |
+
"train_actor_loss": -0.0889451601939608,
|
| 1034 |
+
"train_critic_loss": 0.00013129253869736612,
|
| 1035 |
+
"train_mean_reward": 0.0001839348026263088,
|
| 1036 |
+
"train_entropy": 0.3611451057038566,
|
| 1037 |
+
"train_ce_loss": 0.40479391937695197,
|
| 1038 |
+
"train_dice_loss": 0.43608406825275076,
|
| 1039 |
+
"train_dice": 0.8361081037141498,
|
| 1040 |
+
"train_iou": 0.7472197321775983,
|
| 1041 |
+
"grad_norm": 21.615686439728552,
|
| 1042 |
+
"lr": 0.001098637771751922,
|
| 1043 |
+
"encoder_lr": 0.0001139402897652589,
|
| 1044 |
+
"alpha": 0.2517300546169281,
|
| 1045 |
+
"validated_this_epoch": false,
|
| 1046 |
+
"val_loss": null,
|
| 1047 |
+
"val_dice": null,
|
| 1048 |
+
"val_iou": null,
|
| 1049 |
+
"val_reward": null,
|
| 1050 |
+
"val_entropy": null
|
| 1051 |
+
},
|
| 1052 |
+
{
|
| 1053 |
+
"epoch": 48,
|
| 1054 |
+
"train_loss": 0.33897001978168767,
|
| 1055 |
+
"train_actor_loss": -0.08829710736764879,
|
| 1056 |
+
"train_critic_loss": 0.00013403985470715027,
|
| 1057 |
+
"train_mean_reward": 0.00016173901118466557,
|
| 1058 |
+
"train_entropy": 0.35892657028843267,
|
| 1059 |
+
"train_ce_loss": 0.40545150120826845,
|
| 1060 |
+
"train_dice_loss": 0.448948703378978,
|
| 1061 |
+
"train_dice": 0.8215692983824574,
|
| 1062 |
+
"train_iou": 0.7337636110408782,
|
| 1063 |
+
"grad_norm": 19.11684767127961,
|
| 1064 |
+
"lr": 0.0010674765669316673,
|
| 1065 |
+
"encoder_lr": 0.00011096509935528483,
|
| 1066 |
+
"alpha": 0.24997389316558838,
|
| 1067 |
+
"validated_this_epoch": false,
|
| 1068 |
+
"val_loss": null,
|
| 1069 |
+
"val_dice": null,
|
| 1070 |
+
"val_iou": null,
|
| 1071 |
+
"val_reward": null,
|
| 1072 |
+
"val_entropy": null
|
| 1073 |
+
},
|
| 1074 |
+
{
|
| 1075 |
+
"epoch": 49,
|
| 1076 |
+
"train_loss": 0.33810462748862863,
|
| 1077 |
+
"train_actor_loss": -0.08915085181141469,
|
| 1078 |
+
"train_critic_loss": 0.00011931227977901323,
|
| 1079 |
+
"train_mean_reward": 0.0001293580078976969,
|
| 1080 |
+
"train_entropy": 0.36489381562215717,
|
| 1081 |
+
"train_ce_loss": 0.4173848133339265,
|
| 1082 |
+
"train_dice_loss": 0.4370068222361326,
|
| 1083 |
+
"train_dice": 0.8210988745195886,
|
| 1084 |
+
"train_iou": 0.7353059221291577,
|
| 1085 |
+
"grad_norm": 20.108933841535286,
|
| 1086 |
+
"lr": 0.001036253705282738,
|
| 1087 |
+
"encoder_lr": 0.00010798402211242223,
|
| 1088 |
+
"alpha": 0.24809041619300842,
|
| 1089 |
+
"validated_this_epoch": false,
|
| 1090 |
+
"val_loss": null,
|
| 1091 |
+
"val_dice": null,
|
| 1092 |
+
"val_iou": null,
|
| 1093 |
+
"val_reward": null,
|
| 1094 |
+
"val_entropy": null
|
| 1095 |
+
},
|
| 1096 |
+
{
|
| 1097 |
+
"epoch": 50,
|
| 1098 |
+
"train_loss": 0.3210250280306552,
|
| 1099 |
+
"train_actor_loss": -0.08584695158030557,
|
| 1100 |
+
"train_critic_loss": 0.00011566773755170363,
|
| 1101 |
+
"train_mean_reward": 0.00019478979902604984,
|
| 1102 |
+
"train_entropy": 0.35414034732706473,
|
| 1103 |
+
"train_ce_loss": 0.3980501269234599,
|
| 1104 |
+
"train_dice_loss": 0.4155781517811216,
|
| 1105 |
+
"train_dice": 0.8380284284793391,
|
| 1106 |
+
"train_iou": 0.7526062759523447,
|
| 1107 |
+
"grad_norm": 22.926643622476,
|
| 1108 |
+
"lr": 0.0010050000000000003,
|
| 1109 |
+
"encoder_lr": 0.00010500000000000005,
|
| 1110 |
+
"alpha": 0.24650047719478607,
|
| 1111 |
+
"validated_this_epoch": true,
|
| 1112 |
+
"val_loss": 0.6601516836125291,
|
| 1113 |
+
"val_dice": 0.7110123335164173,
|
| 1114 |
+
"val_iou": 0.6171060720977757,
|
| 1115 |
+
"val_reward": -0.00011582099891757482,
|
| 1116 |
+
"val_entropy": 0.33679192535804986,
|
| 1117 |
+
"val_actor_loss": -0.002161156319110006,
|
| 1118 |
+
"val_critic_loss": 0.00019029130171153579,
|
| 1119 |
+
"val_ce_loss": 0.6272395633800765,
|
| 1120 |
+
"val_dice_loss": 0.6971958017710482
|
| 1121 |
+
},
|
| 1122 |
+
{
|
| 1123 |
+
"epoch": 51,
|
| 1124 |
+
"train_loss": 0.3392306830726795,
|
| 1125 |
+
"train_actor_loss": -0.08686230168978683,
|
| 1126 |
+
"train_critic_loss": 0.00012589489645010673,
|
| 1127 |
+
"train_mean_reward": 0.000219675933609554,
|
| 1128 |
+
"train_entropy": 0.3647166427985334,
|
| 1129 |
+
"train_ce_loss": 0.4164649098088787,
|
| 1130 |
+
"train_dice_loss": 0.43559515291406203,
|
| 1131 |
+
"train_dice": 0.8303098898440424,
|
| 1132 |
+
"train_iou": 0.7429548884765294,
|
| 1133 |
+
"grad_norm": 14.08036404894304,
|
| 1134 |
+
"lr": 0.0009737462947172626,
|
| 1135 |
+
"encoder_lr": 0.00010201597788757786,
|
| 1136 |
+
"alpha": 0.24443319439888,
|
| 1137 |
+
"validated_this_epoch": false,
|
| 1138 |
+
"val_loss": null,
|
| 1139 |
+
"val_dice": null,
|
| 1140 |
+
"val_iou": null,
|
| 1141 |
+
"val_reward": null,
|
| 1142 |
+
"val_entropy": null
|
| 1143 |
+
},
|
| 1144 |
+
{
|
| 1145 |
+
"epoch": 52,
|
| 1146 |
+
"train_loss": 0.31461827964124867,
|
| 1147 |
+
"train_actor_loss": -0.08632082716400617,
|
| 1148 |
+
"train_critic_loss": 0.00012792083003784638,
|
| 1149 |
+
"train_mean_reward": 0.00025670458521286603,
|
| 1150 |
+
"train_entropy": 0.3593870838213644,
|
| 1151 |
+
"train_ce_loss": 0.38497386346282364,
|
| 1152 |
+
"train_dice_loss": 0.4167764176077929,
|
| 1153 |
+
"train_dice": 0.8437300068592641,
|
| 1154 |
+
"train_iou": 0.7568631758187055,
|
| 1155 |
+
"grad_norm": 13.263733024745024,
|
| 1156 |
+
"lr": 0.0009425234330683334,
|
| 1157 |
+
"encoder_lr": 9.903490064471527e-05,
|
| 1158 |
+
"alpha": 0.24272878468036652,
|
| 1159 |
+
"validated_this_epoch": false,
|
| 1160 |
+
"val_loss": null,
|
| 1161 |
+
"val_dice": null,
|
| 1162 |
+
"val_iou": null,
|
| 1163 |
+
"val_reward": null,
|
| 1164 |
+
"val_entropy": null
|
| 1165 |
+
},
|
| 1166 |
+
{
|
| 1167 |
+
"epoch": 53,
|
| 1168 |
+
"train_loss": 0.3224379401473516,
|
| 1169 |
+
"train_actor_loss": -0.08387799084533093,
|
| 1170 |
+
"train_critic_loss": 0.00012826256318579628,
|
| 1171 |
+
"train_mean_reward": 0.0002479213550673582,
|
| 1172 |
+
"train_entropy": 0.35066043182206397,
|
| 1173 |
+
"train_ce_loss": 0.3793846436051859,
|
| 1174 |
+
"train_dice_loss": 0.4331189438662172,
|
| 1175 |
+
"train_dice": 0.8303272415602445,
|
| 1176 |
+
"train_iou": 0.7434805872207748,
|
| 1177 |
+
"grad_norm": 16.166020706642506,
|
| 1178 |
+
"lr": 0.0009113622282480784,
|
| 1179 |
+
"encoder_lr": 9.605971023474117e-05,
|
| 1180 |
+
"alpha": 0.24141384661197662,
|
| 1181 |
+
"validated_this_epoch": false,
|
| 1182 |
+
"val_loss": null,
|
| 1183 |
+
"val_dice": null,
|
| 1184 |
+
"val_iou": null,
|
| 1185 |
+
"val_reward": null,
|
| 1186 |
+
"val_entropy": null
|
| 1187 |
+
},
|
| 1188 |
+
{
|
| 1189 |
+
"epoch": 54,
|
| 1190 |
+
"train_loss": 0.3135488998910138,
|
| 1191 |
+
"train_actor_loss": -0.08246058253066023,
|
| 1192 |
+
"train_critic_loss": 0.0001202849061436649,
|
| 1193 |
+
"train_mean_reward": 0.00019403257594594533,
|
| 1194 |
+
"train_entropy": 0.34700823942974246,
|
| 1195 |
+
"train_ce_loss": 0.3842531600858756,
|
| 1196 |
+
"train_dice_loss": 0.4076455082234178,
|
| 1197 |
+
"train_dice": 0.836118005901925,
|
| 1198 |
+
"train_iou": 0.7487325420432952,
|
| 1199 |
+
"grad_norm": 16.820747276609257,
|
| 1200 |
+
"lr": 0.0008802934326035174,
|
| 1201 |
+
"encoder_lr": 9.309334281139113e-05,
|
| 1202 |
+
"alpha": 0.24027006328105927,
|
| 1203 |
+
"validated_this_epoch": false,
|
| 1204 |
+
"val_loss": null,
|
| 1205 |
+
"val_dice": null,
|
| 1206 |
+
"val_iou": null,
|
| 1207 |
+
"val_reward": null,
|
| 1208 |
+
"val_entropy": null
|
| 1209 |
+
},
|
| 1210 |
+
{
|
| 1211 |
+
"epoch": 55,
|
| 1212 |
+
"train_loss": 0.3294833752781499,
|
| 1213 |
+
"train_actor_loss": -0.08263185731248417,
|
| 1214 |
+
"train_critic_loss": 0.00011728787545491206,
|
| 1215 |
+
"train_mean_reward": 0.00017562639028110237,
|
| 1216 |
+
"train_entropy": 0.3472050580867501,
|
| 1217 |
+
"train_ce_loss": 0.3911710532733482,
|
| 1218 |
+
"train_dice_loss": 0.4329421119018426,
|
| 1219 |
+
"train_dice": 0.8317316169151782,
|
| 1220 |
+
"train_iou": 0.7478006424355002,
|
| 1221 |
+
"grad_norm": 12.179201228212017,
|
| 1222 |
+
"lr": 0.0008493477072849704,
|
| 1223 |
+
"encoder_lr": 9.013872582117809e-05,
|
| 1224 |
+
"alpha": 0.23919503390789032,
|
| 1225 |
+
"validated_this_epoch": false,
|
| 1226 |
+
"val_loss": null,
|
| 1227 |
+
"val_dice": null,
|
| 1228 |
+
"val_iou": null,
|
| 1229 |
+
"val_reward": null,
|
| 1230 |
+
"val_entropy": null
|
| 1231 |
+
},
|
| 1232 |
+
{
|
| 1233 |
+
"epoch": 56,
|
| 1234 |
+
"train_loss": 0.32447372708115874,
|
| 1235 |
+
"train_actor_loss": -0.08151914228418077,
|
| 1236 |
+
"train_critic_loss": 0.00012863667208960598,
|
| 1237 |
+
"train_mean_reward": 0.00024902315074609705,
|
| 1238 |
+
"train_entropy": 0.34606052567407447,
|
| 1239 |
+
"train_ce_loss": 0.3992842348696765,
|
| 1240 |
+
"train_dice_loss": 0.412572855789224,
|
| 1241 |
+
"train_dice": 0.8345937337508781,
|
| 1242 |
+
"train_iou": 0.749244606605986,
|
| 1243 |
+
"grad_norm": 13.610102970932806,
|
| 1244 |
+
"lr": 0.0008185555919872042,
|
| 1245 |
+
"encoder_lr": 8.719877511435618e-05,
|
| 1246 |
+
"alpha": 0.23800688982009888,
|
| 1247 |
+
"validated_this_epoch": false,
|
| 1248 |
+
"val_loss": null,
|
| 1249 |
+
"val_dice": null,
|
| 1250 |
+
"val_iou": null,
|
| 1251 |
+
"val_reward": null,
|
| 1252 |
+
"val_entropy": null
|
| 1253 |
+
},
|
| 1254 |
+
{
|
| 1255 |
+
"epoch": 57,
|
| 1256 |
+
"train_loss": 0.31574048952726547,
|
| 1257 |
+
"train_actor_loss": -0.08356423003392087,
|
| 1258 |
+
"train_critic_loss": 0.00012426459366455563,
|
| 1259 |
+
"train_mean_reward": 0.00030879209018839013,
|
| 1260 |
+
"train_entropy": 0.3513648657065645,
|
| 1261 |
+
"train_ce_loss": 0.3710295831729256,
|
| 1262 |
+
"train_dice_loss": 0.42745557874056106,
|
| 1263 |
+
"train_dice": 0.837906715864521,
|
| 1264 |
+
"train_iou": 0.7524746861526443,
|
| 1265 |
+
"grad_norm": 17.995266500831576,
|
| 1266 |
+
"lr": 0.0007879474748104406,
|
| 1267 |
+
"encoder_lr": 8.42763920673285e-05,
|
| 1268 |
+
"alpha": 0.236659437417984,
|
| 1269 |
+
"validated_this_epoch": false,
|
| 1270 |
+
"val_loss": null,
|
| 1271 |
+
"val_dice": null,
|
| 1272 |
+
"val_iou": null,
|
| 1273 |
+
"val_reward": null,
|
| 1274 |
+
"val_entropy": null
|
| 1275 |
+
},
|
| 1276 |
+
{
|
| 1277 |
+
"epoch": 58,
|
| 1278 |
+
"train_loss": 0.31065986316304567,
|
| 1279 |
+
"train_actor_loss": -0.08225364364125316,
|
| 1280 |
+
"train_critic_loss": 0.0001214118219133092,
|
| 1281 |
+
"train_mean_reward": 0.00034027539378064933,
|
| 1282 |
+
"train_entropy": 0.35118637636342404,
|
| 1283 |
+
"train_ce_loss": 0.3804563977809753,
|
| 1284 |
+
"train_dice_loss": 0.4052491935802676,
|
| 1285 |
+
"train_dice": 0.8422714683704885,
|
| 1286 |
+
"train_iou": 0.7585289649024871,
|
| 1287 |
+
"grad_norm": 12.93380750011104,
|
| 1288 |
+
"lr": 0.0007575535622709696,
|
| 1289 |
+
"encoder_lr": 8.13744607193388e-05,
|
| 1290 |
+
"alpha": 0.2354142963886261,
|
| 1291 |
+
"validated_this_epoch": false,
|
| 1292 |
+
"val_loss": null,
|
| 1293 |
+
"val_dice": null,
|
| 1294 |
+
"val_iou": null,
|
| 1295 |
+
"val_reward": null,
|
| 1296 |
+
"val_entropy": null
|
| 1297 |
+
},
|
| 1298 |
+
{
|
| 1299 |
+
"epoch": 59,
|
| 1300 |
+
"train_loss": 0.3265964566231814,
|
| 1301 |
+
"train_actor_loss": -0.08255374217189329,
|
| 1302 |
+
"train_critic_loss": 0.00011919140907941645,
|
| 1303 |
+
"train_mean_reward": 0.00026900076354638327,
|
| 1304 |
+
"train_entropy": 0.3502129768177829,
|
| 1305 |
+
"train_ce_loss": 0.384240411644325,
|
| 1306 |
+
"train_dice_loss": 0.43394078410565085,
|
| 1307 |
+
"train_dice": 0.8391680458007535,
|
| 1308 |
+
"train_iou": 0.7568418543016173,
|
| 1309 |
+
"grad_norm": 9.877583969933118,
|
| 1310 |
+
"lr": 0.0007274038494909672,
|
| 1311 |
+
"encoder_lr": 7.849584492627325e-05,
|
| 1312 |
+
"alpha": 0.23399090766906738,
|
| 1313 |
+
"validated_this_epoch": false,
|
| 1314 |
+
"val_loss": null,
|
| 1315 |
+
"val_dice": null,
|
| 1316 |
+
"val_iou": null,
|
| 1317 |
+
"val_reward": null,
|
| 1318 |
+
"val_entropy": null
|
| 1319 |
+
},
|
| 1320 |
+
{
|
| 1321 |
+
"epoch": 60,
|
| 1322 |
+
"train_loss": 0.3106941013676367,
|
| 1323 |
+
"train_actor_loss": -0.08035531176076072,
|
| 1324 |
+
"train_critic_loss": 0.00013021849194932801,
|
| 1325 |
+
"train_mean_reward": 0.00019047783340890115,
|
| 1326 |
+
"train_entropy": 0.3464120777578933,
|
| 1327 |
+
"train_ce_loss": 0.37322865934616545,
|
| 1328 |
+
"train_dice_loss": 0.4087399383694011,
|
| 1329 |
+
"train_dice": 0.8435935968778562,
|
| 1330 |
+
"train_iou": 0.7555974804572411,
|
| 1331 |
+
"grad_norm": 11.680240491101909,
|
| 1332 |
+
"lr": 0.0006975280905969278,
|
| 1333 |
+
"encoder_lr": 7.564338553438004e-05,
|
| 1334 |
+
"alpha": 0.23277175426483154,
|
| 1335 |
+
"validated_this_epoch": true,
|
| 1336 |
+
"val_loss": 0.6464337560764486,
|
| 1337 |
+
"val_dice": 0.7154152006286761,
|
| 1338 |
+
"val_iou": 0.6261228794536259,
|
| 1339 |
+
"val_reward": -0.00022144284527960488,
|
| 1340 |
+
"val_entropy": 0.3308497237587216,
|
| 1341 |
+
"val_actor_loss": -0.014067396582715741,
|
| 1342 |
+
"val_critic_loss": 0.00011004886400172322,
|
| 1343 |
+
"val_ce_loss": 0.6015639295730967,
|
| 1344 |
+
"val_dice_loss": 0.7193283038307922
|
| 1345 |
+
},
|
| 1346 |
+
{
|
| 1347 |
+
"epoch": 61,
|
| 1348 |
+
"train_loss": 0.32905462240193817,
|
| 1349 |
+
"train_actor_loss": -0.0813806894974114,
|
| 1350 |
+
"train_critic_loss": 0.00011936007433133367,
|
| 1351 |
+
"train_mean_reward": 0.0002992610009122418,
|
| 1352 |
+
"train_entropy": 0.3520125935398023,
|
| 1353 |
+
"train_ce_loss": 0.3912073390170912,
|
| 1354 |
+
"train_dice_loss": 0.42954391410492493,
|
| 1355 |
+
"train_dice": 0.8409348532070983,
|
| 1356 |
+
"train_iou": 0.755372920225622,
|
| 1357 |
+
"grad_norm": 11.910602663838587,
|
| 1358 |
+
"lr": 0.0006679557693559353,
|
| 1359 |
+
"encoder_lr": 7.281989757669733e-05,
|
| 1360 |
+
"alpha": 0.23143118619918823,
|
| 1361 |
+
"validated_this_epoch": false,
|
| 1362 |
+
"val_loss": null,
|
| 1363 |
+
"val_dice": null,
|
| 1364 |
+
"val_iou": null,
|
| 1365 |
+
"val_reward": null,
|
| 1366 |
+
"val_entropy": null
|
| 1367 |
+
},
|
| 1368 |
+
{
|
| 1369 |
+
"epoch": 62,
|
| 1370 |
+
"train_loss": 0.32315836369615886,
|
| 1371 |
+
"train_actor_loss": -0.07809418577517957,
|
| 1372 |
+
"train_critic_loss": 0.00012413299683566831,
|
| 1373 |
+
"train_mean_reward": 0.00026550721556307125,
|
| 1374 |
+
"train_entropy": 0.34025854549974743,
|
| 1375 |
+
"train_ce_loss": 0.381528151568112,
|
| 1376 |
+
"train_dice_loss": 0.42085280475382353,
|
| 1377 |
+
"train_dice": 0.8320374917682072,
|
| 1378 |
+
"train_iou": 0.7490136113908531,
|
| 1379 |
+
"grad_norm": 13.638553913249526,
|
| 1380 |
+
"lr": 0.000638716070078746,
|
| 1381 |
+
"encoder_lr": 7.002816749495564e-05,
|
| 1382 |
+
"alpha": 0.23079806566238403,
|
| 1383 |
+
"validated_this_epoch": false,
|
| 1384 |
+
"val_loss": null,
|
| 1385 |
+
"val_dice": null,
|
| 1386 |
+
"val_iou": null,
|
| 1387 |
+
"val_reward": null,
|
| 1388 |
+
"val_entropy": null
|
| 1389 |
+
},
|
| 1390 |
+
{
|
| 1391 |
+
"epoch": 63,
|
| 1392 |
+
"train_loss": 0.307995670778287,
|
| 1393 |
+
"train_actor_loss": -0.07988748453480482,
|
| 1394 |
+
"train_critic_loss": 0.00011830789746445283,
|
| 1395 |
+
"train_mean_reward": 0.00034647378531709415,
|
| 1396 |
+
"train_entropy": 0.34277106946675967,
|
| 1397 |
+
"train_ce_loss": 0.36699815596169877,
|
| 1398 |
+
"train_dice_loss": 0.40864983347343226,
|
| 1399 |
+
"train_dice": 0.8500556883136599,
|
| 1400 |
+
"train_iou": 0.7652164868896489,
|
| 1401 |
+
"grad_norm": 10.34581516694653,
|
| 1402 |
+
"lr": 0.0006098378488183934,
|
| 1403 |
+
"encoder_lr": 6.727095038969585e-05,
|
| 1404 |
+
"alpha": 0.22990544140338898,
|
| 1405 |
+
"validated_this_epoch": false,
|
| 1406 |
+
"val_loss": null,
|
| 1407 |
+
"val_dice": null,
|
| 1408 |
+
"val_iou": null,
|
| 1409 |
+
"val_reward": null,
|
| 1410 |
+
"val_entropy": null
|
| 1411 |
+
},
|
| 1412 |
+
{
|
| 1413 |
+
"epoch": 64,
|
| 1414 |
+
"train_loss": 0.31185069849512487,
|
| 1415 |
+
"train_actor_loss": -0.07672652699178426,
|
| 1416 |
+
"train_critic_loss": 0.00012167585124454479,
|
| 1417 |
+
"train_mean_reward": 0.00026455780426806806,
|
| 1418 |
+
"train_entropy": 0.3381984252915826,
|
| 1419 |
+
"train_ce_loss": 0.36074776535622716,
|
| 1420 |
+
"train_dice_loss": 0.4162849991820579,
|
| 1421 |
+
"train_dice": 0.8411847361906656,
|
| 1422 |
+
"train_iou": 0.7552094152644994,
|
| 1423 |
+
"grad_norm": 15.974313804345538,
|
| 1424 |
+
"lr": 0.0005813496048927529,
|
| 1425 |
+
"encoder_lr": 6.455096730131812e-05,
|
| 1426 |
+
"alpha": 0.22919757664203644,
|
| 1427 |
+
"validated_this_epoch": false,
|
| 1428 |
+
"val_loss": null,
|
| 1429 |
+
"val_dice": null,
|
| 1430 |
+
"val_iou": null,
|
| 1431 |
+
"val_reward": null,
|
| 1432 |
+
"val_entropy": null
|
| 1433 |
+
},
|
| 1434 |
+
{
|
| 1435 |
+
"epoch": 65,
|
| 1436 |
+
"train_loss": 0.30728749183273707,
|
| 1437 |
+
"train_actor_loss": -0.0791829751530128,
|
| 1438 |
+
"train_critic_loss": 0.0001279412133093741,
|
| 1439 |
+
"train_mean_reward": 0.0003256188193184312,
|
| 1440 |
+
"train_entropy": 0.3421229699708387,
|
| 1441 |
+
"train_ce_loss": 0.3648492643592798,
|
| 1442 |
+
"train_dice_loss": 0.4079637167835729,
|
| 1443 |
+
"train_dice": 0.8342330825644098,
|
| 1444 |
+
"train_iou": 0.750883375504501,
|
| 1445 |
+
"grad_norm": 9.802442484585814,
|
| 1446 |
+
"lr": 0.0005532794527591512,
|
| 1447 |
+
"encoder_lr": 6.187090252474308e-05,
|
| 1448 |
+
"alpha": 0.2281648963689804,
|
| 1449 |
+
"validated_this_epoch": false,
|
| 1450 |
+
"val_loss": null,
|
| 1451 |
+
"val_dice": null,
|
| 1452 |
+
"val_iou": null,
|
| 1453 |
+
"val_reward": null,
|
| 1454 |
+
"val_entropy": null
|
| 1455 |
+
},
|
| 1456 |
+
{
|
| 1457 |
+
"epoch": 66,
|
| 1458 |
+
"train_loss": 0.28307362356721316,
|
| 1459 |
+
"train_actor_loss": -0.07762010132374111,
|
| 1460 |
+
"train_critic_loss": 0.0001127899451723154,
|
| 1461 |
+
"train_mean_reward": 0.0004483267157893093,
|
| 1462 |
+
"train_entropy": 0.33327449623312444,
|
| 1463 |
+
"train_ce_loss": 0.34969130739357895,
|
| 1464 |
+
"train_dice_loss": 0.3715833415510734,
|
| 1465 |
+
"train_dice": 0.8605801417558504,
|
| 1466 |
+
"train_iou": 0.7771159271342621,
|
| 1467 |
+
"grad_norm": 6.445185325866522,
|
| 1468 |
+
"lr": 0.0005256550942687934,
|
| 1469 |
+
"encoder_lr": 5.9233400960337056e-05,
|
| 1470 |
+
"alpha": 0.22806203365325928,
|
| 1471 |
+
"validated_this_epoch": false,
|
| 1472 |
+
"val_loss": null,
|
| 1473 |
+
"val_dice": null,
|
| 1474 |
+
"val_iou": null,
|
| 1475 |
+
"val_reward": null,
|
| 1476 |
+
"val_entropy": null
|
| 1477 |
+
},
|
| 1478 |
+
{
|
| 1479 |
+
"epoch": 67,
|
| 1480 |
+
"train_loss": 0.32282443407764155,
|
| 1481 |
+
"train_actor_loss": -0.07754177872081905,
|
| 1482 |
+
"train_critic_loss": 0.00012542818690552702,
|
| 1483 |
+
"train_mean_reward": 0.000307677013193773,
|
| 1484 |
+
"train_entropy": 0.3366578403831452,
|
| 1485 |
+
"train_ce_loss": 0.36972970211786965,
|
| 1486 |
+
"train_dice_loss": 0.4308772831770184,
|
| 1487 |
+
"train_dice": 0.8380595273623204,
|
| 1488 |
+
"train_iou": 0.7535705130954963,
|
| 1489 |
+
"grad_norm": 13.897041998630346,
|
| 1490 |
+
"lr": 0.0004985037913283807,
|
| 1491 |
+
"encoder_lr": 5.6641065503714745e-05,
|
| 1492 |
+
"alpha": 0.22756028175354004,
|
| 1493 |
+
"validated_this_epoch": false,
|
| 1494 |
+
"val_loss": null,
|
| 1495 |
+
"val_dice": null,
|
| 1496 |
+
"val_iou": null,
|
| 1497 |
+
"val_reward": null,
|
| 1498 |
+
"val_entropy": null
|
| 1499 |
+
},
|
| 1500 |
+
{
|
| 1501 |
+
"epoch": 68,
|
| 1502 |
+
"train_loss": 0.2918738214544548,
|
| 1503 |
+
"train_actor_loss": -0.07928771803814927,
|
| 1504 |
+
"train_critic_loss": 0.00011836912177417246,
|
| 1505 |
+
"train_mean_reward": 0.00035127723011418746,
|
| 1506 |
+
"train_entropy": 0.3389960847198193,
|
| 1507 |
+
"train_ce_loss": 0.357273892734142,
|
| 1508 |
+
"train_dice_loss": 0.3849308053344411,
|
| 1509 |
+
"train_dice": 0.8528910183828574,
|
| 1510 |
+
"train_iou": 0.7695361721189322,
|
| 1511 |
+
"grad_norm": 6.8463406946307925,
|
| 1512 |
+
"lr": 0.0004718523389958983,
|
| 1513 |
+
"encoder_lr": 5.4096454476995315e-05,
|
| 1514 |
+
"alpha": 0.22678987681865692,
|
| 1515 |
+
"validated_this_epoch": false,
|
| 1516 |
+
"val_loss": null,
|
| 1517 |
+
"val_dice": null,
|
| 1518 |
+
"val_iou": null,
|
| 1519 |
+
"val_reward": null,
|
| 1520 |
+
"val_entropy": null
|
| 1521 |
+
},
|
| 1522 |
+
{
|
| 1523 |
+
"epoch": 69,
|
| 1524 |
+
"train_loss": 0.2924046531760011,
|
| 1525 |
+
"train_actor_loss": -0.07904473080441779,
|
| 1526 |
+
"train_critic_loss": 0.00012025145748780949,
|
| 1527 |
+
"train_mean_reward": 0.0003039297415898898,
|
| 1528 |
+
"train_entropy": 0.34279225812863934,
|
| 1529 |
+
"train_ce_loss": 0.3531149132586342,
|
| 1530 |
+
"train_dice_loss": 0.38966359171140413,
|
| 1531 |
+
"train_dice": 0.8535806189077878,
|
| 1532 |
+
"train_iou": 0.7692752574169467,
|
| 1533 |
+
"grad_norm": 6.892875376597855,
|
| 1534 |
+
"lr": 0.0004457270390371302,
|
| 1535 |
+
"encoder_lr": 5.1602079104047604e-05,
|
| 1536 |
+
"alpha": 0.22624582052230835,
|
| 1537 |
+
"validated_this_epoch": false,
|
| 1538 |
+
"val_loss": null,
|
| 1539 |
+
"val_dice": null,
|
| 1540 |
+
"val_iou": null,
|
| 1541 |
+
"val_reward": null,
|
| 1542 |
+
"val_entropy": null
|
| 1543 |
+
},
|
| 1544 |
+
{
|
| 1545 |
+
"epoch": 70,
|
| 1546 |
+
"train_loss": 0.3037114967844745,
|
| 1547 |
+
"train_actor_loss": -0.07704884413081084,
|
| 1548 |
+
"train_critic_loss": 0.0001265770168900826,
|
| 1549 |
+
"train_mean_reward": 0.00035038489694452156,
|
| 1550 |
+
"train_entropy": 0.3318693522142501,
|
| 1551 |
+
"train_ce_loss": 0.3555363452986114,
|
| 1552 |
+
"train_dice_loss": 0.4058577492255573,
|
| 1553 |
+
"train_dice": 0.8511551965867767,
|
| 1554 |
+
"train_iou": 0.7668857216512394,
|
| 1555 |
+
"grad_norm": 8.668443173401116,
|
| 1556 |
+
"lr": 0.00042015367396898955,
|
| 1557 |
+
"encoder_lr": 4.916040103221508e-05,
|
| 1558 |
+
"alpha": 0.22566784918308258,
|
| 1559 |
+
"validated_this_epoch": true,
|
| 1560 |
+
"val_loss": 0.6327815185032898,
|
| 1561 |
+
"val_dice": 0.7096048110366603,
|
| 1562 |
+
"val_iou": 0.6196106785322201,
|
| 1563 |
+
"val_reward": -0.00019969978329111593,
|
| 1564 |
+
"val_entropy": 0.3266731019724499,
|
| 1565 |
+
"val_actor_loss": -0.014286113298507503,
|
| 1566 |
+
"val_critic_loss": 0.00013539095589570053,
|
| 1567 |
+
"val_ce_loss": 0.5441237711253567,
|
| 1568 |
+
"val_dice_loss": 0.7498760768259414
|
| 1569 |
+
},
|
| 1570 |
+
{
|
| 1571 |
+
"epoch": 71,
|
| 1572 |
+
"train_loss": 0.30323557235849224,
|
| 1573 |
+
"train_actor_loss": -0.07765601396810948,
|
| 1574 |
+
"train_critic_loss": 0.00012552651641112143,
|
| 1575 |
+
"train_mean_reward": 0.0004216825807167717,
|
| 1576 |
+
"train_entropy": 0.3375805178392457,
|
| 1577 |
+
"train_ce_loss": 0.3634343034039114,
|
| 1578 |
+
"train_dice_loss": 0.39822333296448065,
|
| 1579 |
+
"train_dice": 0.850103477322427,
|
| 1580 |
+
"train_iou": 0.7675344913501694,
|
| 1581 |
+
"grad_norm": 15.946048521256262,
|
| 1582 |
+
"lr": 0.0003951574816152885,
|
| 1583 |
+
"encoder_lr": 4.6773829902967244e-05,
|
| 1584 |
+
"alpha": 0.22521552443504333,
|
| 1585 |
+
"validated_this_epoch": false,
|
| 1586 |
+
"val_loss": null,
|
| 1587 |
+
"val_dice": null,
|
| 1588 |
+
"val_iou": null,
|
| 1589 |
+
"val_reward": null,
|
| 1590 |
+
"val_entropy": null
|
| 1591 |
+
},
|
| 1592 |
+
{
|
| 1593 |
+
"epoch": 72,
|
| 1594 |
+
"train_loss": 0.3036112468581387,
|
| 1595 |
+
"train_actor_loss": -0.07664492336470027,
|
| 1596 |
+
"train_critic_loss": 0.00012340573032425765,
|
| 1597 |
+
"train_mean_reward": 0.00035961561413154583,
|
| 1598 |
+
"train_entropy": 0.33667391020737264,
|
| 1599 |
+
"train_ce_loss": 0.36112727232166697,
|
| 1600 |
+
"train_dice_loss": 0.39926165149809467,
|
| 1601 |
+
"train_dice": 0.8491855481088972,
|
| 1602 |
+
"train_iou": 0.7665005990648746,
|
| 1603 |
+
"grad_norm": 13.592286005962727,
|
| 1604 |
+
"lr": 0.0003707631302000539,
|
| 1605 |
+
"encoder_lr": 4.444472097387449e-05,
|
| 1606 |
+
"alpha": 0.22478757798671722,
|
| 1607 |
+
"validated_this_epoch": false,
|
| 1608 |
+
"val_loss": null,
|
| 1609 |
+
"val_dice": null,
|
| 1610 |
+
"val_iou": null,
|
| 1611 |
+
"val_reward": null,
|
| 1612 |
+
"val_entropy": null
|
| 1613 |
+
},
|
| 1614 |
+
{
|
| 1615 |
+
"epoch": 73,
|
| 1616 |
+
"train_loss": 0.304036049840091,
|
| 1617 |
+
"train_actor_loss": -0.07614740126499141,
|
| 1618 |
+
"train_critic_loss": 0.00012589064598713347,
|
| 1619 |
+
"train_mean_reward": 0.0003287162554802625,
|
| 1620 |
+
"train_entropy": 0.3360503486692135,
|
| 1621 |
+
"train_ce_loss": 0.3571828498363071,
|
| 1622 |
+
"train_dice_loss": 0.4030581517120973,
|
| 1623 |
+
"train_dice": 0.8440281183949556,
|
| 1624 |
+
"train_iou": 0.7613306533898776,
|
| 1625 |
+
"grad_norm": 6.94213994482691,
|
| 1626 |
+
"lr": 0.00034699469400296674,
|
| 1627 |
+
"encoder_lr": 4.2175372794253104e-05,
|
| 1628 |
+
"alpha": 0.2243524193763733,
|
| 1629 |
+
"validated_this_epoch": false,
|
| 1630 |
+
"val_loss": null,
|
| 1631 |
+
"val_dice": null,
|
| 1632 |
+
"val_iou": null,
|
| 1633 |
+
"val_reward": null,
|
| 1634 |
+
"val_entropy": null
|
| 1635 |
+
},
|
| 1636 |
+
{
|
| 1637 |
+
"epoch": 74,
|
| 1638 |
+
"train_loss": 0.31571957073354906,
|
| 1639 |
+
"train_actor_loss": -0.075948777420865,
|
| 1640 |
+
"train_critic_loss": 0.00012011940043110193,
|
| 1641 |
+
"train_mean_reward": 0.00042062418949702964,
|
| 1642 |
+
"train_entropy": 0.33300842238874084,
|
| 1643 |
+
"train_ce_loss": 0.35674971527268373,
|
| 1644 |
+
"train_dice_loss": 0.42646684780601385,
|
| 1645 |
+
"train_dice": 0.8441785887057908,
|
| 1646 |
+
"train_iou": 0.7618813772939412,
|
| 1647 |
+
"grad_norm": 8.49718480008517,
|
| 1648 |
+
"lr": 0.0003238756296009549,
|
| 1649 |
+
"encoder_lr": 3.996802493677459e-05,
|
| 1650 |
+
"alpha": 0.22393277287483215,
|
| 1651 |
+
"validated_this_epoch": false,
|
| 1652 |
+
"val_loss": null,
|
| 1653 |
+
"val_dice": null,
|
| 1654 |
+
"val_iou": null,
|
| 1655 |
+
"val_reward": null,
|
| 1656 |
+
"val_entropy": null
|
| 1657 |
+
},
|
| 1658 |
+
{
|
| 1659 |
+
"epoch": 75,
|
| 1660 |
+
"train_loss": 0.32766491686564314,
|
| 1661 |
+
"train_actor_loss": -0.07556292254631755,
|
| 1662 |
+
"train_critic_loss": 0.00012226070088422426,
|
| 1663 |
+
"train_mean_reward": 0.0003822259602698268,
|
| 1664 |
+
"train_entropy": 0.3351505511952925,
|
| 1665 |
+
"train_ce_loss": 0.3674851135237657,
|
| 1666 |
+
"train_dice_loss": 0.4388482941059487,
|
| 1667 |
+
"train_dice": 0.8408036791540902,
|
| 1668 |
+
"train_iou": 0.7566280168215942,
|
| 1669 |
+
"grad_norm": 10.791874277499295,
|
| 1670 |
+
"lr": 0.0003014287527193855,
|
| 1671 |
+
"encoder_lr": 3.782485578727801e-05,
|
| 1672 |
+
"alpha": 0.22362756729125977,
|
| 1673 |
+
"validated_this_epoch": false,
|
| 1674 |
+
"val_loss": null,
|
| 1675 |
+
"val_dice": null,
|
| 1676 |
+
"val_iou": null,
|
| 1677 |
+
"val_reward": null,
|
| 1678 |
+
"val_entropy": null
|
| 1679 |
+
},
|
| 1680 |
+
{
|
| 1681 |
+
"epoch": 76,
|
| 1682 |
+
"train_loss": 0.28053983976674635,
|
| 1683 |
+
"train_actor_loss": -0.07748339182723708,
|
| 1684 |
+
"train_critic_loss": 0.0001248351992576635,
|
| 1685 |
+
"train_mean_reward": 0.0004713248979653652,
|
| 1686 |
+
"train_entropy": 0.33636865993956877,
|
| 1687 |
+
"train_ce_loss": 0.3440032800574106,
|
| 1688 |
+
"train_dice_loss": 0.37191833744369424,
|
| 1689 |
+
"train_dice": 0.8558560824162333,
|
| 1690 |
+
"train_iou": 0.7710309460712091,
|
| 1691 |
+
"grad_norm": 11.470800769883533,
|
| 1692 |
+
"lr": 0.0002796762157156959,
|
| 1693 |
+
"encoder_lr": 3.574798039496594e-05,
|
| 1694 |
+
"alpha": 0.22300666570663452,
|
| 1695 |
+
"validated_this_epoch": false,
|
| 1696 |
+
"val_loss": null,
|
| 1697 |
+
"val_dice": null,
|
| 1698 |
+
"val_iou": null,
|
| 1699 |
+
"val_reward": null,
|
| 1700 |
+
"val_entropy": null
|
| 1701 |
+
},
|
| 1702 |
+
{
|
| 1703 |
+
"epoch": 77,
|
| 1704 |
+
"train_loss": 0.31113187001310577,
|
| 1705 |
+
"train_actor_loss": -0.0759144398679798,
|
| 1706 |
+
"train_critic_loss": 0.00012881555611172075,
|
| 1707 |
+
"train_mean_reward": 0.0003924087773386362,
|
| 1708 |
+
"train_entropy": 0.3340654087082052,
|
| 1709 |
+
"train_ce_loss": 0.35894228428847685,
|
| 1710 |
+
"train_dice_loss": 0.4150215103952768,
|
| 1711 |
+
"train_dice": 0.8411377759684717,
|
| 1712 |
+
"train_iou": 0.7567542722107657,
|
| 1713 |
+
"grad_norm": 13.39562037517858,
|
| 1714 |
+
"lr": 0.00025863948571769286,
|
| 1715 |
+
"encoder_lr": 3.373944838510636e-05,
|
| 1716 |
+
"alpha": 0.2226925641298294,
|
| 1717 |
+
"validated_this_epoch": false,
|
| 1718 |
+
"val_loss": null,
|
| 1719 |
+
"val_dice": null,
|
| 1720 |
+
"val_iou": null,
|
| 1721 |
+
"val_reward": null,
|
| 1722 |
+
"val_entropy": null
|
| 1723 |
+
},
|
| 1724 |
+
{
|
| 1725 |
+
"epoch": 78,
|
| 1726 |
+
"train_loss": 0.30900701728844365,
|
| 1727 |
+
"train_actor_loss": -0.07543811352937951,
|
| 1728 |
+
"train_critic_loss": 0.00012381937675412895,
|
| 1729 |
+
"train_mean_reward": 0.00032047648331534156,
|
| 1730 |
+
"train_entropy": 0.332611573316176,
|
| 1731 |
+
"train_ce_loss": 0.3594659344252266,
|
| 1732 |
+
"train_dice_loss": 0.4093004951027321,
|
| 1733 |
+
"train_dice": 0.8402555319647086,
|
| 1734 |
+
"train_iou": 0.7585298628207418,
|
| 1735 |
+
"grad_norm": 6.530554777892061,
|
| 1736 |
+
"lr": 0.00023833932343808993,
|
| 1737 |
+
"encoder_lr": 3.180124193630005e-05,
|
| 1738 |
+
"alpha": 0.22242261469364166,
|
| 1739 |
+
"validated_this_epoch": false,
|
| 1740 |
+
"val_loss": null,
|
| 1741 |
+
"val_dice": null,
|
| 1742 |
+
"val_iou": null,
|
| 1743 |
+
"val_reward": null,
|
| 1744 |
+
"val_entropy": null
|
| 1745 |
+
},
|
| 1746 |
+
{
|
| 1747 |
+
"epoch": 79,
|
| 1748 |
+
"train_loss": 0.2997042298165345,
|
| 1749 |
+
"train_actor_loss": -0.07521878020688699,
|
| 1750 |
+
"train_critic_loss": 0.00012052803169626476,
|
| 1751 |
+
"train_mean_reward": 0.00047374051269770455,
|
| 1752 |
+
"train_entropy": 0.3280081692746135,
|
| 1753 |
+
"train_ce_loss": 0.350841413430138,
|
| 1754 |
+
"train_dice_loss": 0.39888406816379046,
|
| 1755 |
+
"train_dice": 0.8515426802273581,
|
| 1756 |
+
"train_iou": 0.7653313625825643,
|
| 1757 |
+
"grad_norm": 9.509378351675448,
|
| 1758 |
+
"lr": 0.00021879576268618817,
|
| 1759 |
+
"encoder_lr": 2.993527382430943e-05,
|
| 1760 |
+
"alpha": 0.22231508791446686,
|
| 1761 |
+
"validated_this_epoch": false,
|
| 1762 |
+
"val_loss": null,
|
| 1763 |
+
"val_dice": null,
|
| 1764 |
+
"val_iou": null,
|
| 1765 |
+
"val_reward": null,
|
| 1766 |
+
"val_entropy": null
|
| 1767 |
+
},
|
| 1768 |
+
{
|
| 1769 |
+
"epoch": 80,
|
| 1770 |
+
"train_loss": 0.30812201697321595,
|
| 1771 |
+
"train_actor_loss": -0.07585890910730188,
|
| 1772 |
+
"train_critic_loss": 0.00012244161146587591,
|
| 1773 |
+
"train_mean_reward": 0.0003257392062320623,
|
| 1774 |
+
"train_entropy": 0.339325285158108,
|
| 1775 |
+
"train_ce_loss": 0.36114390663566787,
|
| 1776 |
+
"train_dice_loss": 0.40669549196881527,
|
| 1777 |
+
"train_dice": 0.8421935192453572,
|
| 1778 |
+
"train_iou": 0.7602231682119247,
|
| 1779 |
+
"grad_norm": 9.281670538954032,
|
| 1780 |
+
"lr": 0.0002000280905969275,
|
| 1781 |
+
"encoder_lr": 2.814338553438002e-05,
|
| 1782 |
+
"alpha": 0.22177168726921082,
|
| 1783 |
+
"validated_this_epoch": true,
|
| 1784 |
+
"val_loss": 0.6174627126293313,
|
| 1785 |
+
"val_dice": 0.7252799426950827,
|
| 1786 |
+
"val_iou": 0.6341242494090558,
|
| 1787 |
+
"val_reward": -0.0001555420118588378,
|
| 1788 |
+
"val_entropy": 0.32761264020445374,
|
| 1789 |
+
"val_actor_loss": -0.015833495271061735,
|
| 1790 |
+
"val_critic_loss": 0.0001275775916929897,
|
| 1791 |
+
"val_ce_loss": 0.5567514054580696,
|
| 1792 |
+
"val_dice_loss": 0.7097134090433217
|
| 1793 |
+
},
|
| 1794 |
+
{
|
| 1795 |
+
"epoch": 81,
|
| 1796 |
+
"train_loss": 0.3065570924005132,
|
| 1797 |
+
"train_actor_loss": -0.07545319280902392,
|
| 1798 |
+
"train_critic_loss": 0.00011476443820152048,
|
| 1799 |
+
"train_mean_reward": 0.00043770548838146703,
|
| 1800 |
+
"train_entropy": 0.32953977013050123,
|
| 1801 |
+
"train_ce_loss": 0.36189152589380663,
|
| 1802 |
+
"train_dice_loss": 0.4020142694622355,
|
| 1803 |
+
"train_dice": 0.8453430845456824,
|
| 1804 |
+
"train_iou": 0.7651260741689024,
|
| 1805 |
+
"grad_norm": 6.322403997413872,
|
| 1806 |
+
"lr": 0.00018205482859681116,
|
| 1807 |
+
"encoder_lr": 2.6427345443916652e-05,
|
| 1808 |
+
"alpha": 0.22174493968486786,
|
| 1809 |
+
"validated_this_epoch": false,
|
| 1810 |
+
"val_loss": null,
|
| 1811 |
+
"val_dice": null,
|
| 1812 |
+
"val_iou": null,
|
| 1813 |
+
"val_reward": null,
|
| 1814 |
+
"val_entropy": null
|
| 1815 |
+
},
|
| 1816 |
+
{
|
| 1817 |
+
"epoch": 82,
|
| 1818 |
+
"train_loss": 0.29437874478121945,
|
| 1819 |
+
"train_actor_loss": -0.07713441672501485,
|
| 1820 |
+
"train_critic_loss": 0.00011686544264113893,
|
| 1821 |
+
"train_mean_reward": 0.00041308785826086426,
|
| 1822 |
+
"train_entropy": 0.33611422145705505,
|
| 1823 |
+
"train_ce_loss": 0.3517065157548907,
|
| 1824 |
+
"train_dice_loss": 0.3912029314102744,
|
| 1825 |
+
"train_dice": 0.8507058320019429,
|
| 1826 |
+
"train_iou": 0.7677855135758551,
|
| 1827 |
+
"grad_norm": 9.044833423093309,
|
| 1828 |
+
"lr": 0.00016489371412549528,
|
| 1829 |
+
"encoder_lr": 2.4788847077308606e-05,
|
| 1830 |
+
"alpha": 0.22140787541866302,
|
| 1831 |
+
"validated_this_epoch": false,
|
| 1832 |
+
"val_loss": null,
|
| 1833 |
+
"val_dice": null,
|
| 1834 |
+
"val_iou": null,
|
| 1835 |
+
"val_reward": null,
|
| 1836 |
+
"val_entropy": null
|
| 1837 |
+
},
|
| 1838 |
+
{
|
| 1839 |
+
"epoch": 83,
|
| 1840 |
+
"train_loss": 0.3015456150408632,
|
| 1841 |
+
"train_actor_loss": -0.0758493058527856,
|
| 1842 |
+
"train_critic_loss": 0.00011325778518767037,
|
| 1843 |
+
"train_mean_reward": 0.0004789611194546735,
|
| 1844 |
+
"train_entropy": 0.3302456609500471,
|
| 1845 |
+
"train_ce_loss": 0.3644114487433868,
|
| 1846 |
+
"train_dice_loss": 0.3902651262529752,
|
| 1847 |
+
"train_dice": 0.8511293893597767,
|
| 1848 |
+
"train_iou": 0.7691080901732892,
|
| 1849 |
+
"grad_norm": 5.361961954785872,
|
| 1850 |
+
"lr": 0.00014856168313107603,
|
| 1851 |
+
"encoder_lr": 2.322950743462536e-05,
|
| 1852 |
+
"alpha": 0.22132723033428192,
|
| 1853 |
+
"validated_this_epoch": false,
|
| 1854 |
+
"val_loss": null,
|
| 1855 |
+
"val_dice": null,
|
| 1856 |
+
"val_iou": null,
|
| 1857 |
+
"val_reward": null,
|
| 1858 |
+
"val_entropy": null
|
| 1859 |
+
},
|
| 1860 |
+
{
|
| 1861 |
+
"epoch": 84,
|
| 1862 |
+
"train_loss": 0.31333459707251354,
|
| 1863 |
+
"train_actor_loss": -0.07552279957909455,
|
| 1864 |
+
"train_critic_loss": 0.0001252708413703752,
|
| 1865 |
+
"train_mean_reward": 0.00038974148468313443,
|
| 1866 |
+
"train_entropy": 0.3315104924007595,
|
| 1867 |
+
"train_ce_loss": 0.3604084323196448,
|
| 1868 |
+
"train_dice_loss": 0.4171810796248512,
|
| 1869 |
+
"train_dice": 0.8421329484505449,
|
| 1870 |
+
"train_iou": 0.7595187830008818,
|
| 1871 |
+
"grad_norm": 6.017300826634547,
|
| 1872 |
+
"lr": 0.0001330748533563558,
|
| 1873 |
+
"encoder_lr": 2.1750865395832973e-05,
|
| 1874 |
+
"alpha": 0.22097845375537872,
|
| 1875 |
+
"validated_this_epoch": false,
|
| 1876 |
+
"val_loss": null,
|
| 1877 |
+
"val_dice": null,
|
| 1878 |
+
"val_iou": null,
|
| 1879 |
+
"val_reward": null,
|
| 1880 |
+
"val_entropy": null
|
| 1881 |
+
},
|
| 1882 |
+
{
|
| 1883 |
+
"epoch": 85,
|
| 1884 |
+
"train_loss": 0.27991887096084594,
|
| 1885 |
+
"train_actor_loss": -0.0760813579925887,
|
| 1886 |
+
"train_critic_loss": 0.00011517167389289072,
|
| 1887 |
+
"train_mean_reward": 0.00041763428001372245,
|
| 1888 |
+
"train_entropy": 0.33277400317349176,
|
| 1889 |
+
"train_ce_loss": 0.3424206904077888,
|
| 1890 |
+
"train_dice_loss": 0.3694645863324789,
|
| 1891 |
+
"train_dice": 0.862755534055502,
|
| 1892 |
+
"train_iou": 0.7809131291425468,
|
| 1893 |
+
"grad_norm": 4.682364524796951,
|
| 1894 |
+
"lr": 0.0001184485084325741,
|
| 1895 |
+
"encoder_lr": 2.035438020210507e-05,
|
| 1896 |
+
"alpha": 0.22062133252620697,
|
| 1897 |
+
"validated_this_epoch": false,
|
| 1898 |
+
"val_loss": null,
|
| 1899 |
+
"val_dice": null,
|
| 1900 |
+
"val_iou": null,
|
| 1901 |
+
"val_reward": null,
|
| 1902 |
+
"val_entropy": null
|
| 1903 |
+
},
|
| 1904 |
+
{
|
| 1905 |
+
"epoch": 86,
|
| 1906 |
+
"train_loss": 0.30499647255597073,
|
| 1907 |
+
"train_actor_loss": -0.07530868115043146,
|
| 1908 |
+
"train_critic_loss": 0.0001156676482796284,
|
| 1909 |
+
"train_mean_reward": 0.0003803927420555082,
|
| 1910 |
+
"train_entropy": 0.333190485257511,
|
| 1911 |
+
"train_ce_loss": 0.36388044384517526,
|
| 1912 |
+
"train_dice_loss": 0.3966141849833249,
|
| 1913 |
+
"train_dice": 0.8528071957528088,
|
| 1914 |
+
"train_iou": 0.7728123354815425,
|
| 1915 |
+
"grad_norm": 3.9979898573816284,
|
| 1916 |
+
"lr": 0.00010469708279631079,
|
| 1917 |
+
"encoder_lr": 1.904143001572817e-05,
|
| 1918 |
+
"alpha": 0.22050800919532776,
|
| 1919 |
+
"validated_this_epoch": false,
|
| 1920 |
+
"val_loss": null,
|
| 1921 |
+
"val_dice": null,
|
| 1922 |
+
"val_iou": null,
|
| 1923 |
+
"val_reward": null,
|
| 1924 |
+
"val_entropy": null
|
| 1925 |
+
},
|
| 1926 |
+
{
|
| 1927 |
+
"epoch": 87,
|
| 1928 |
+
"train_loss": 0.28935780026408475,
|
| 1929 |
+
"train_actor_loss": -0.0748067925075305,
|
| 1930 |
+
"train_critic_loss": 0.00012033358279362806,
|
| 1931 |
+
"train_mean_reward": 0.0004136506742938893,
|
| 1932 |
+
"train_entropy": 0.3302284257356511,
|
| 1933 |
+
"train_ce_loss": 0.3471481174799717,
|
| 1934 |
+
"train_dice_loss": 0.3810607231094547,
|
| 1935 |
+
"train_dice": 0.8533148325002041,
|
| 1936 |
+
"train_iou": 0.7714358762950431,
|
| 1937 |
+
"grad_norm": 5.67940813672635,
|
| 1938 |
+
"lr": 9.183414744443903e-05,
|
| 1939 |
+
"encoder_lr": 1.781331056002182e-05,
|
| 1940 |
+
"alpha": 0.22034740447998047,
|
| 1941 |
+
"validated_this_epoch": false,
|
| 1942 |
+
"val_loss": null,
|
| 1943 |
+
"val_dice": null,
|
| 1944 |
+
"val_iou": null,
|
| 1945 |
+
"val_reward": null,
|
| 1946 |
+
"val_entropy": null
|
| 1947 |
+
},
|
| 1948 |
+
{
|
| 1949 |
+
"epoch": 88,
|
| 1950 |
+
"train_loss": 0.282804185019214,
|
| 1951 |
+
"train_actor_loss": -0.07428028449916715,
|
| 1952 |
+
"train_critic_loss": 0.0001225056968914072,
|
| 1953 |
+
"train_mean_reward": 0.000430432429729943,
|
| 1954 |
+
"train_entropy": 0.32950698416824487,
|
| 1955 |
+
"train_ce_loss": 0.3445821878523145,
|
| 1956 |
+
"train_dice_loss": 0.3694642359434172,
|
| 1957 |
+
"train_dice": 0.8555131171040044,
|
| 1958 |
+
"train_iou": 0.7744293875007415,
|
| 1959 |
+
"grad_norm": 7.281788512717846,
|
| 1960 |
+
"lr": 7.987239654118994e-05,
|
| 1961 |
+
"encoder_lr": 1.6671233840616128e-05,
|
| 1962 |
+
"alpha": 0.220547154545784,
|
| 1963 |
+
"validated_this_epoch": false,
|
| 1964 |
+
"val_loss": null,
|
| 1965 |
+
"val_dice": null,
|
| 1966 |
+
"val_iou": null,
|
| 1967 |
+
"val_reward": null,
|
| 1968 |
+
"val_entropy": null
|
| 1969 |
+
},
|
| 1970 |
+
{
|
| 1971 |
+
"epoch": 89,
|
| 1972 |
+
"train_loss": 0.30343953392273465,
|
| 1973 |
+
"train_actor_loss": -0.07557150258054567,
|
| 1974 |
+
"train_critic_loss": 0.00012203092769560718,
|
| 1975 |
+
"train_mean_reward": 0.000402743805909634,
|
| 1976 |
+
"train_entropy": 0.3318241170056415,
|
| 1977 |
+
"train_ce_loss": 0.3618190135669627,
|
| 1978 |
+
"train_dice_loss": 0.39608101712332827,
|
| 1979 |
+
"train_dice": 0.848079681284118,
|
| 1980 |
+
"train_iou": 0.7647072117836022,
|
| 1981 |
+
"grad_norm": 7.435718650965727,
|
| 1982 |
+
"lr": 6.88236348905457e-05,
|
| 1983 |
+
"encoder_lr": 1.5616326949348587e-05,
|
| 1984 |
+
"alpha": 0.22010201215744019,
|
| 1985 |
+
"validated_this_epoch": false,
|
| 1986 |
+
"val_loss": null,
|
| 1987 |
+
"val_dice": null,
|
| 1988 |
+
"val_iou": null,
|
| 1989 |
+
"val_reward": null,
|
| 1990 |
+
"val_entropy": null
|
| 1991 |
+
},
|
| 1992 |
+
{
|
| 1993 |
+
"epoch": 90,
|
| 1994 |
+
"train_loss": 0.29299391343352177,
|
| 1995 |
+
"train_actor_loss": -0.07527920065173345,
|
| 1996 |
+
"train_critic_loss": 0.00012124865210686879,
|
| 1997 |
+
"train_mean_reward": 0.0004890012844786305,
|
| 1998 |
+
"train_entropy": 0.330272088371198,
|
| 1999 |
+
"train_ce_loss": 0.341243314607916,
|
| 2000 |
+
"train_dice_loss": 0.39518165303447145,
|
| 2001 |
+
"train_dice": 0.8566729187038271,
|
| 2002 |
+
"train_iou": 0.7741321650236829,
|
| 2003 |
+
"grad_norm": 6.147999250842619,
|
| 2004 |
+
"lr": 5.869876628632226e-05,
|
| 2005 |
+
"encoder_lr": 1.4649630951960419e-05,
|
| 2006 |
+
"alpha": 0.22027471661567688,
|
| 2007 |
+
"validated_this_epoch": true,
|
| 2008 |
+
"val_loss": 0.6231435263938637,
|
| 2009 |
+
"val_dice": 0.7185076945194161,
|
| 2010 |
+
"val_iou": 0.6275294236356725,
|
| 2011 |
+
"val_reward": -0.00022649142653854786,
|
| 2012 |
+
"val_entropy": 0.34903986405844645,
|
| 2013 |
+
"val_actor_loss": -0.018371272722207423,
|
| 2014 |
+
"val_critic_loss": 0.00014913031373917058,
|
| 2015 |
+
"val_ce_loss": 0.5503802692677562,
|
| 2016 |
+
"val_dice_loss": 0.7325001786453555
|
| 2017 |
+
},
|
| 2018 |
+
{
|
| 2019 |
+
"epoch": 91,
|
| 2020 |
+
"train_loss": 0.2922775759217293,
|
| 2021 |
+
"train_actor_loss": -0.07469568186913657,
|
| 2022 |
+
"train_critic_loss": 0.00012143343768646457,
|
| 2023 |
+
"train_mean_reward": 0.0004205800890724761,
|
| 2024 |
+
"train_entropy": 0.33100477790878724,
|
| 2025 |
+
"train_ce_loss": 0.3469820861429102,
|
| 2026 |
+
"train_dice_loss": 0.3868429860249354,
|
| 2027 |
+
"train_dice": 0.860050558399192,
|
| 2028 |
+
"train_iou": 0.7775769864598991,
|
| 2029 |
+
"grad_norm": 9.022059423054836,
|
| 2030 |
+
"lr": 4.9507782751441774e-05,
|
| 2031 |
+
"encoder_lr": 1.3772099860690422e-05,
|
| 2032 |
+
"alpha": 0.22012366354465485,
|
| 2033 |
+
"validated_this_epoch": false,
|
| 2034 |
+
"val_loss": null,
|
| 2035 |
+
"val_dice": null,
|
| 2036 |
+
"val_iou": null,
|
| 2037 |
+
"val_reward": null,
|
| 2038 |
+
"val_entropy": null
|
| 2039 |
+
},
|
| 2040 |
+
{
|
| 2041 |
+
"epoch": 92,
|
| 2042 |
+
"train_loss": 0.28516306369312744,
|
| 2043 |
+
"train_actor_loss": -0.07682260387223358,
|
| 2044 |
+
"train_critic_loss": 0.00011868114738635595,
|
| 2045 |
+
"train_mean_reward": 0.00044725226394187306,
|
| 2046 |
+
"train_entropy": 0.33233393696004415,
|
| 2047 |
+
"train_ce_loss": 0.34437067152629064,
|
| 2048 |
+
"train_dice_loss": 0.37948197326610866,
|
| 2049 |
+
"train_dice": 0.86244508197457,
|
| 2050 |
+
"train_iou": 0.7799197974107795,
|
| 2051 |
+
"grad_norm": 7.276689930941707,
|
| 2052 |
+
"lr": 4.12597546770122e-05,
|
| 2053 |
+
"encoder_lr": 1.298459969278006e-05,
|
| 2054 |
+
"alpha": 0.2198239266872406,
|
| 2055 |
+
"validated_this_epoch": false,
|
| 2056 |
+
"val_loss": null,
|
| 2057 |
+
"val_dice": null,
|
| 2058 |
+
"val_iou": null,
|
| 2059 |
+
"val_reward": null,
|
| 2060 |
+
"val_entropy": null
|
| 2061 |
+
},
|
| 2062 |
+
{
|
| 2063 |
+
"epoch": 93,
|
| 2064 |
+
"train_loss": 0.29351433422127665,
|
| 2065 |
+
"train_actor_loss": -0.0745770329036762,
|
| 2066 |
+
"train_critic_loss": 0.00011629446902288009,
|
| 2067 |
+
"train_mean_reward": 0.0004551340727562151,
|
| 2068 |
+
"train_entropy": 0.3282424138151398,
|
| 2069 |
+
"train_ce_loss": 0.3466885743887725,
|
| 2070 |
+
"train_dice_loss": 0.38937785520726087,
|
| 2071 |
+
"train_dice": 0.8557019407874042,
|
| 2072 |
+
"train_iou": 0.7738658428342418,
|
| 2073 |
+
"grad_norm": 7.33312615867733,
|
| 2074 |
+
"lr": 3.396282187094642e-05,
|
| 2075 |
+
"encoder_lr": 1.2287907615819006e-05,
|
| 2076 |
+
"alpha": 0.22003594040870667,
|
| 2077 |
+
"validated_this_epoch": false,
|
| 2078 |
+
"val_loss": null,
|
| 2079 |
+
"val_dice": null,
|
| 2080 |
+
"val_iou": null,
|
| 2081 |
+
"val_reward": null,
|
| 2082 |
+
"val_entropy": null
|
| 2083 |
+
},
|
| 2084 |
+
{
|
| 2085 |
+
"epoch": 94,
|
| 2086 |
+
"train_loss": 0.28411788267061805,
|
| 2087 |
+
"train_actor_loss": -0.07442422117797289,
|
| 2088 |
+
"train_critic_loss": 0.00012036018253018328,
|
| 2089 |
+
"train_mean_reward": 0.00038961109346350294,
|
| 2090 |
+
"train_entropy": 0.3295715792003528,
|
| 2091 |
+
"train_ce_loss": 0.34260199886797243,
|
| 2092 |
+
"train_dice_loss": 0.3743618384042144,
|
| 2093 |
+
"train_dice": 0.8561752191020068,
|
| 2094 |
+
"train_iou": 0.7755373913409399,
|
| 2095 |
+
"grad_norm": 6.806314678155174,
|
| 2096 |
+
"lr": 2.762418552495473e-05,
|
| 2097 |
+
"encoder_lr": 1.1682711180774574e-05,
|
| 2098 |
+
"alpha": 0.219998300075531,
|
| 2099 |
+
"validated_this_epoch": false,
|
| 2100 |
+
"val_loss": null,
|
| 2101 |
+
"val_dice": null,
|
| 2102 |
+
"val_iou": null,
|
| 2103 |
+
"val_reward": null,
|
| 2104 |
+
"val_entropy": null
|
| 2105 |
+
},
|
| 2106 |
+
{
|
| 2107 |
+
"epoch": 95,
|
| 2108 |
+
"train_loss": 0.28163289465010166,
|
| 2109 |
+
"train_actor_loss": -0.07494764053563586,
|
| 2110 |
+
"train_critic_loss": 0.00011744915048542401,
|
| 2111 |
+
"train_mean_reward": 0.00045263889893079937,
|
| 2112 |
+
"train_entropy": 0.3298965358687926,
|
| 2113 |
+
"train_ce_loss": 0.3377505597677707,
|
| 2114 |
+
"train_dice_loss": 0.3752930517806563,
|
| 2115 |
+
"train_dice": 0.8615671392943681,
|
| 2116 |
+
"train_iou": 0.7802108262147726,
|
| 2117 |
+
"grad_norm": 8.11379330241403,
|
| 2118 |
+
"lr": 2.2250101107838036e-05,
|
| 2119 |
+
"encoder_lr": 1.1169607643461924e-05,
|
| 2120 |
+
"alpha": 0.22000868618488312,
|
| 2121 |
+
"validated_this_epoch": false,
|
| 2122 |
+
"val_loss": null,
|
| 2123 |
+
"val_dice": null,
|
| 2124 |
+
"val_iou": null,
|
| 2125 |
+
"val_reward": null,
|
| 2126 |
+
"val_entropy": null
|
| 2127 |
+
},
|
| 2128 |
+
{
|
| 2129 |
+
"epoch": 96,
|
| 2130 |
+
"train_loss": 0.2837922649616881,
|
| 2131 |
+
"train_actor_loss": -0.07566399072376333,
|
| 2132 |
+
"train_critic_loss": 0.0001165717263115026,
|
| 2133 |
+
"train_mean_reward": 0.0004425820033687345,
|
| 2134 |
+
"train_entropy": 0.3318348183881405,
|
| 2135 |
+
"train_ce_loss": 0.3435492101805261,
|
| 2136 |
+
"train_dice_loss": 0.37524672018157107,
|
| 2137 |
+
"train_dice": 0.8531550178510218,
|
| 2138 |
+
"train_iou": 0.7727129201776703,
|
| 2139 |
+
"grad_norm": 8.738612899946611,
|
| 2140 |
+
"lr": 1.784587219209463e-05,
|
| 2141 |
+
"encoder_lr": 1.0749103375124614e-05,
|
| 2142 |
+
"alpha": 0.21969462931156158,
|
| 2143 |
+
"validated_this_epoch": false,
|
| 2144 |
+
"val_loss": null,
|
| 2145 |
+
"val_dice": null,
|
| 2146 |
+
"val_iou": null,
|
| 2147 |
+
"val_reward": null,
|
| 2148 |
+
"val_entropy": null
|
| 2149 |
+
},
|
| 2150 |
+
{
|
| 2151 |
+
"epoch": 97,
|
| 2152 |
+
"train_loss": 0.2907992157933497,
|
| 2153 |
+
"train_actor_loss": -0.07545118864134763,
|
| 2154 |
+
"train_critic_loss": 0.00012038995201499448,
|
| 2155 |
+
"train_mean_reward": 0.0004139809480629969,
|
| 2156 |
+
"train_entropy": 0.33136177915858356,
|
| 2157 |
+
"train_ce_loss": 0.35024257426965955,
|
| 2158 |
+
"train_dice_loss": 0.382137835641856,
|
| 2159 |
+
"train_dice": 0.8566087859462782,
|
| 2160 |
+
"train_iou": 0.774030051314361,
|
| 2161 |
+
"grad_norm": 9.906422799409823,
|
| 2162 |
+
"lr": 1.4415845219935406e-05,
|
| 2163 |
+
"encoder_lr": 1.0421613362707402e-05,
|
| 2164 |
+
"alpha": 0.21944019198417664,
|
| 2165 |
+
"validated_this_epoch": false,
|
| 2166 |
+
"val_loss": null,
|
| 2167 |
+
"val_dice": null,
|
| 2168 |
+
"val_iou": null,
|
| 2169 |
+
"val_reward": null,
|
| 2170 |
+
"val_entropy": null
|
| 2171 |
+
},
|
| 2172 |
+
{
|
| 2173 |
+
"epoch": 98,
|
| 2174 |
+
"train_loss": 0.29480746672885827,
|
| 2175 |
+
"train_actor_loss": -0.0738000712364681,
|
| 2176 |
+
"train_critic_loss": 0.0001145676353197854,
|
| 2177 |
+
"train_mean_reward": 0.0004353652145469043,
|
| 2178 |
+
"train_entropy": 0.3256402525675389,
|
| 2179 |
+
"train_ce_loss": 0.3520909806241785,
|
| 2180 |
+
"train_dice_loss": 0.38500951765735636,
|
| 2181 |
+
"train_dice": 0.8544651219176642,
|
| 2182 |
+
"train_iou": 0.7751359553129819,
|
| 2183 |
+
"grad_norm": 7.403769441353258,
|
| 2184 |
+
"lr": 1.19634052138698e-05,
|
| 2185 |
+
"encoder_lr": 1.0187460799314204e-05,
|
| 2186 |
+
"alpha": 0.21985945105552673,
|
| 2187 |
+
"validated_this_epoch": false,
|
| 2188 |
+
"val_loss": null,
|
| 2189 |
+
"val_dice": null,
|
| 2190 |
+
"val_iou": null,
|
| 2191 |
+
"val_reward": null,
|
| 2192 |
+
"val_entropy": null
|
| 2193 |
+
},
|
| 2194 |
+
{
|
| 2195 |
+
"epoch": 99,
|
| 2196 |
+
"train_loss": 0.2910641819642958,
|
| 2197 |
+
"train_actor_loss": -0.07462473056767646,
|
| 2198 |
+
"train_critic_loss": 0.00011607745672117463,
|
| 2199 |
+
"train_mean_reward": 0.00040162837682952436,
|
| 2200 |
+
"train_entropy": 0.33086552474772896,
|
| 2201 |
+
"train_ce_loss": 0.34574993433507806,
|
| 2202 |
+
"train_dice_loss": 0.3855118016093892,
|
| 2203 |
+
"train_dice": 0.8540281955385888,
|
| 2204 |
+
"train_iou": 0.7728484504908818,
|
| 2205 |
+
"grad_norm": 6.449722734994666,
|
| 2206 |
+
"lr": 1.049097243609706e-05,
|
| 2207 |
+
"encoder_lr": 1.0046876765255499e-05,
|
| 2208 |
+
"alpha": 0.21979421377182007,
|
| 2209 |
+
"validated_this_epoch": false,
|
| 2210 |
+
"val_loss": null,
|
| 2211 |
+
"val_dice": null,
|
| 2212 |
+
"val_iou": null,
|
| 2213 |
+
"val_reward": null,
|
| 2214 |
+
"val_entropy": null
|
| 2215 |
+
},
|
| 2216 |
+
{
|
| 2217 |
+
"epoch": 100,
|
| 2218 |
+
"train_loss": 0.292507603797022,
|
| 2219 |
+
"train_actor_loss": -0.0748064739855839,
|
| 2220 |
+
"train_critic_loss": 0.00011732140204294785,
|
| 2221 |
+
"train_mean_reward": 0.0005162119463231639,
|
| 2222 |
+
"train_entropy": 0.32645806718319276,
|
| 2223 |
+
"train_ce_loss": 0.3505832417651491,
|
| 2224 |
+
"train_dice_loss": 0.38392758092214896,
|
| 2225 |
+
"train_dice": 0.8580238111442383,
|
| 2226 |
+
"train_iou": 0.7766493836311994,
|
| 2227 |
+
"grad_norm": 4.151609534441039,
|
| 2228 |
+
"lr": 1e-05,
|
| 2229 |
+
"encoder_lr": 1e-05,
|
| 2230 |
+
"alpha": 0.21999937295913696,
|
| 2231 |
+
"validated_this_epoch": true,
|
| 2232 |
+
"val_loss": 0.6268361883950561,
|
| 2233 |
+
"val_dice": 0.7187820154611619,
|
| 2234 |
+
"val_iou": 0.6281435286923981,
|
| 2235 |
+
"val_reward": -0.0002771024063427702,
|
| 2236 |
+
"val_entropy": 0.3362236471489223,
|
| 2237 |
+
"val_actor_loss": -0.016416844980076256,
|
| 2238 |
+
"val_critic_loss": 0.00013078153161090583,
|
| 2239 |
+
"val_ce_loss": 0.5630592065886829,
|
| 2240 |
+
"val_dice_loss": 0.7233160568608178
|
| 2241 |
+
}
|
| 2242 |
+
]
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/low_iou_samples.csv
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
sample_id
|
| 2 |
+
BUSI_0224_B
|
| 3 |
+
BUSI_0207_M
|
| 4 |
+
BUSI_0067_M
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/run_config.json
ADDED
|
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"project_dir": "/workspace",
|
| 3 |
+
"data_root": "/workspace/BUSI_with_classes",
|
| 4 |
+
"run_type": "final",
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"dataset_name": "BUSI_with_classes",
|
| 8 |
+
"dataset_split_policy": "stratified",
|
| 9 |
+
"dataset_splits_path": "/workspace/dataset_splits_busi_with_classes_stratified.json",
|
| 10 |
+
"split_type": "80_10_10",
|
| 11 |
+
"train_subset_key": "50",
|
| 12 |
+
"normalization_cache_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/norm_stats_BUSI_with_classes_stratified_80_10_10_50pct.json",
|
| 13 |
+
"head_lr": 0.002,
|
| 14 |
+
"encoder_lr": 0.0002,
|
| 15 |
+
"weight_decay": 0.002,
|
| 16 |
+
"dropout_p": 0.3,
|
| 17 |
+
"tmax": 6,
|
| 18 |
+
"entropy_lr": 0.0002,
|
| 19 |
+
"max_epochs": 100,
|
| 20 |
+
"gamma": 0.95,
|
| 21 |
+
"grad_clip_norm": 6.0,
|
| 22 |
+
"scheduler_factor": 0.5,
|
| 23 |
+
"scheduler_patience": 5,
|
| 24 |
+
"scheduler_threshold": 0.001,
|
| 25 |
+
"scheduler_min_lr": 1e-05,
|
| 26 |
+
"execution_mode": "train_eval",
|
| 27 |
+
"evaluation_checkpoint_mode": "latest",
|
| 28 |
+
"strategy2_checkpoint_mode": "specific",
|
| 29 |
+
"train_resume_mode": "off",
|
| 30 |
+
"train_resume_specific_checkpoint": "",
|
| 31 |
+
"backbone_family": "smp",
|
| 32 |
+
"smp_encoder_name": "efficientnet-b0",
|
| 33 |
+
"smp_encoder_weights": "imagenet",
|
| 34 |
+
"smp_encoder_depth": 5,
|
| 35 |
+
"smp_encoder_proj_dim": 96,
|
| 36 |
+
"smp_decoder_type": "Unet",
|
| 37 |
+
"vgg_feature_scales": 4,
|
| 38 |
+
"vgg_feature_dilation": 1,
|
| 39 |
+
"strategy2_checkpoint_path": "/workspace/50%_Strategt2_latest.pt"
|
| 40 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod/pct_50/strategy_3/final/summary.json
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_val_iou": 0.6427941490616009,
|
| 3 |
+
"best_val_dice": 0.7366531130935929,
|
| 4 |
+
"final_epoch": 100,
|
| 5 |
+
"elapsed_seconds": 1540.8181087970734,
|
| 6 |
+
"seconds_per_epoch": 15.408181087970734,
|
| 7 |
+
"device_used": "cuda",
|
| 8 |
+
"strategy": 3,
|
| 9 |
+
"run_type": "final",
|
| 10 |
+
"resumed": false
|
| 11 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/norm_stats_BUSI_with_classes_stratified_80_10_10_50pct.json
ADDED
|
@@ -0,0 +1,266 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"global_mean": 85.12605579417583,
|
| 3 |
+
"global_std": 56.68613974260357,
|
| 4 |
+
"total_pixels": 244630659,
|
| 5 |
+
"num_images": 258,
|
| 6 |
+
"filenames": [
|
| 7 |
+
"BUSI_0034_B.png",
|
| 8 |
+
"BUSI_0155_M.png",
|
| 9 |
+
"BUSI_0005_M.png",
|
| 10 |
+
"BUSI_0260_B.png",
|
| 11 |
+
"BUSI_0386_B.png",
|
| 12 |
+
"BUSI_0280_B.png",
|
| 13 |
+
"BUSI_0047_B.png",
|
| 14 |
+
"BUSI_0222_B.png",
|
| 15 |
+
"BUSI_0205_B.png",
|
| 16 |
+
"BUSI_0403_B.png",
|
| 17 |
+
"BUSI_0133_B.png",
|
| 18 |
+
"BUSI_0016_B.png",
|
| 19 |
+
"BUSI_0105_M.png",
|
| 20 |
+
"BUSI_0220_B.png",
|
| 21 |
+
"BUSI_0066_B.png",
|
| 22 |
+
"BUSI_0060_B.png",
|
| 23 |
+
"BUSI_0351_B.png",
|
| 24 |
+
"BUSI_0029_M.png",
|
| 25 |
+
"BUSI_0287_B.png",
|
| 26 |
+
"BUSI_0046_M.png",
|
| 27 |
+
"BUSI_0123_B.png",
|
| 28 |
+
"BUSI_0359_B.png",
|
| 29 |
+
"BUSI_0215_B.png",
|
| 30 |
+
"BUSI_0377_B.png",
|
| 31 |
+
"BUSI_0029_B.png",
|
| 32 |
+
"BUSI_0158_M.png",
|
| 33 |
+
"BUSI_0375_B.png",
|
| 34 |
+
"BUSI_0221_B.png",
|
| 35 |
+
"BUSI_0120_B.png",
|
| 36 |
+
"BUSI_0404_B.png",
|
| 37 |
+
"BUSI_0093_M.png",
|
| 38 |
+
"BUSI_0106_M.png",
|
| 39 |
+
"BUSI_0030_B.png",
|
| 40 |
+
"BUSI_0407_B.png",
|
| 41 |
+
"BUSI_0131_B.png",
|
| 42 |
+
"BUSI_0092_B.png",
|
| 43 |
+
"BUSI_0112_M.png",
|
| 44 |
+
"BUSI_0177_M.png",
|
| 45 |
+
"BUSI_0333_B.png",
|
| 46 |
+
"BUSI_0206_B.png",
|
| 47 |
+
"BUSI_0239_B.png",
|
| 48 |
+
"BUSI_0165_B.png",
|
| 49 |
+
"BUSI_0396_B.png",
|
| 50 |
+
"BUSI_0207_B.png",
|
| 51 |
+
"BUSI_0158_B.png",
|
| 52 |
+
"BUSI_0160_B.png",
|
| 53 |
+
"BUSI_0091_B.png",
|
| 54 |
+
"BUSI_0090_M.png",
|
| 55 |
+
"BUSI_0035_B.png",
|
| 56 |
+
"BUSI_0281_B.png",
|
| 57 |
+
"BUSI_0410_B.png",
|
| 58 |
+
"BUSI_0322_B.png",
|
| 59 |
+
"BUSI_0236_B.png",
|
| 60 |
+
"BUSI_0385_B.png",
|
| 61 |
+
"BUSI_0193_M.png",
|
| 62 |
+
"BUSI_0388_B.png",
|
| 63 |
+
"BUSI_0144_M.png",
|
| 64 |
+
"BUSI_0293_B.png",
|
| 65 |
+
"BUSI_0088_M.png",
|
| 66 |
+
"BUSI_0052_B.png",
|
| 67 |
+
"BUSI_0346_B.png",
|
| 68 |
+
"BUSI_0262_B.png",
|
| 69 |
+
"BUSI_0408_B.png",
|
| 70 |
+
"BUSI_0146_M.png",
|
| 71 |
+
"BUSI_0309_B.png",
|
| 72 |
+
"BUSI_0083_B.png",
|
| 73 |
+
"BUSI_0190_B.png",
|
| 74 |
+
"BUSI_0426_B.png",
|
| 75 |
+
"BUSI_0019_B.png",
|
| 76 |
+
"BUSI_0274_B.png",
|
| 77 |
+
"BUSI_0162_M.png",
|
| 78 |
+
"BUSI_0179_M.png",
|
| 79 |
+
"BUSI_0078_B.png",
|
| 80 |
+
"BUSI_0328_B.png",
|
| 81 |
+
"BUSI_0001_M.png",
|
| 82 |
+
"BUSI_0301_B.png",
|
| 83 |
+
"BUSI_0109_M.png",
|
| 84 |
+
"BUSI_0011_B.png",
|
| 85 |
+
"BUSI_0063_B.png",
|
| 86 |
+
"BUSI_0360_B.png",
|
| 87 |
+
"BUSI_0114_B.png",
|
| 88 |
+
"BUSI_0430_B.png",
|
| 89 |
+
"BUSI_0391_B.png",
|
| 90 |
+
"BUSI_0133_M.png",
|
| 91 |
+
"BUSI_0188_M.png",
|
| 92 |
+
"BUSI_0424_B.png",
|
| 93 |
+
"BUSI_0179_B.png",
|
| 94 |
+
"BUSI_0390_B.png",
|
| 95 |
+
"BUSI_0030_M.png",
|
| 96 |
+
"BUSI_0235_B.png",
|
| 97 |
+
"BUSI_0065_M.png",
|
| 98 |
+
"BUSI_0101_M.png",
|
| 99 |
+
"BUSI_0355_B.png",
|
| 100 |
+
"BUSI_0053_M.png",
|
| 101 |
+
"BUSI_0203_B.png",
|
| 102 |
+
"BUSI_0007_B.png",
|
| 103 |
+
"BUSI_0344_B.png",
|
| 104 |
+
"BUSI_0129_B.png",
|
| 105 |
+
"BUSI_0271_B.png",
|
| 106 |
+
"BUSI_0252_B.png",
|
| 107 |
+
"BUSI_0296_B.png",
|
| 108 |
+
"BUSI_0183_M.png",
|
| 109 |
+
"BUSI_0310_B.png",
|
| 110 |
+
"BUSI_0409_B.png",
|
| 111 |
+
"BUSI_0041_M.png",
|
| 112 |
+
"BUSI_0092_M.png",
|
| 113 |
+
"BUSI_0273_B.png",
|
| 114 |
+
"BUSI_0206_M.png",
|
| 115 |
+
"BUSI_0024_M.png",
|
| 116 |
+
"BUSI_0015_B.png",
|
| 117 |
+
"BUSI_0094_M.png",
|
| 118 |
+
"BUSI_0055_B.png",
|
| 119 |
+
"BUSI_0094_B.png",
|
| 120 |
+
"BUSI_0077_B.png",
|
| 121 |
+
"BUSI_0347_B.png",
|
| 122 |
+
"BUSI_0126_M.png",
|
| 123 |
+
"BUSI_0186_M.png",
|
| 124 |
+
"BUSI_0145_M.png",
|
| 125 |
+
"BUSI_0054_M.png",
|
| 126 |
+
"BUSI_0074_M.png",
|
| 127 |
+
"BUSI_0111_B.png",
|
| 128 |
+
"BUSI_0032_M.png",
|
| 129 |
+
"BUSI_0176_M.png",
|
| 130 |
+
"BUSI_0002_M.png",
|
| 131 |
+
"BUSI_0069_B.png",
|
| 132 |
+
"BUSI_0027_B.png",
|
| 133 |
+
"BUSI_0109_B.png",
|
| 134 |
+
"BUSI_0389_B.png",
|
| 135 |
+
"BUSI_0088_B.png",
|
| 136 |
+
"BUSI_0169_B.png",
|
| 137 |
+
"BUSI_0157_B.png",
|
| 138 |
+
"BUSI_0058_B.png",
|
| 139 |
+
"BUSI_0231_B.png",
|
| 140 |
+
"BUSI_0166_B.png",
|
| 141 |
+
"BUSI_0098_M.png",
|
| 142 |
+
"BUSI_0020_M.png",
|
| 143 |
+
"BUSI_0182_B.png",
|
| 144 |
+
"BUSI_0397_B.png",
|
| 145 |
+
"BUSI_0045_M.png",
|
| 146 |
+
"BUSI_0138_M.png",
|
| 147 |
+
"BUSI_0036_M.png",
|
| 148 |
+
"BUSI_0380_B.png",
|
| 149 |
+
"BUSI_0137_B.png",
|
| 150 |
+
"BUSI_0205_M.png",
|
| 151 |
+
"BUSI_0107_M.png",
|
| 152 |
+
"BUSI_0044_M.png",
|
| 153 |
+
"BUSI_0213_B.png",
|
| 154 |
+
"BUSI_0070_B.png",
|
| 155 |
+
"BUSI_0295_B.png",
|
| 156 |
+
"BUSI_0173_B.png",
|
| 157 |
+
"BUSI_0164_M.png",
|
| 158 |
+
"BUSI_0039_M.png",
|
| 159 |
+
"BUSI_0421_B.png",
|
| 160 |
+
"BUSI_0174_B.png",
|
| 161 |
+
"BUSI_0416_B.png",
|
| 162 |
+
"BUSI_0124_M.png",
|
| 163 |
+
"BUSI_0428_B.png",
|
| 164 |
+
"BUSI_0181_B.png",
|
| 165 |
+
"BUSI_0076_B.png",
|
| 166 |
+
"BUSI_0073_B.png",
|
| 167 |
+
"BUSI_0045_B.png",
|
| 168 |
+
"BUSI_0104_M.png",
|
| 169 |
+
"BUSI_0272_B.png",
|
| 170 |
+
"BUSI_0405_B.png",
|
| 171 |
+
"BUSI_0425_B.png",
|
| 172 |
+
"BUSI_0212_B.png",
|
| 173 |
+
"BUSI_0156_M.png",
|
| 174 |
+
"BUSI_0339_B.png",
|
| 175 |
+
"BUSI_0160_M.png",
|
| 176 |
+
"BUSI_0204_M.png",
|
| 177 |
+
"BUSI_0082_B.png",
|
| 178 |
+
"BUSI_0070_M.png",
|
| 179 |
+
"BUSI_0061_M.png",
|
| 180 |
+
"BUSI_0136_M.png",
|
| 181 |
+
"BUSI_0024_B.png",
|
| 182 |
+
"BUSI_0241_B.png",
|
| 183 |
+
"BUSI_0420_B.png",
|
| 184 |
+
"BUSI_0196_B.png",
|
| 185 |
+
"BUSI_0406_B.png",
|
| 186 |
+
"BUSI_0374_B.png",
|
| 187 |
+
"BUSI_0113_M.png",
|
| 188 |
+
"BUSI_0009_B.png",
|
| 189 |
+
"BUSI_0292_B.png",
|
| 190 |
+
"BUSI_0079_M.png",
|
| 191 |
+
"BUSI_0118_B.png",
|
| 192 |
+
"BUSI_0175_M.png",
|
| 193 |
+
"BUSI_0011_M.png",
|
| 194 |
+
"BUSI_0187_B.png",
|
| 195 |
+
"BUSI_0034_M.png",
|
| 196 |
+
"BUSI_0321_B.png",
|
| 197 |
+
"BUSI_0078_M.png",
|
| 198 |
+
"BUSI_0180_B.png",
|
| 199 |
+
"BUSI_0398_B.png",
|
| 200 |
+
"BUSI_0195_B.png",
|
| 201 |
+
"BUSI_0203_M.png",
|
| 202 |
+
"BUSI_0167_B.png",
|
| 203 |
+
"BUSI_0263_B.png",
|
| 204 |
+
"BUSI_0012_B.png",
|
| 205 |
+
"BUSI_0177_B.png",
|
| 206 |
+
"BUSI_0147_M.png",
|
| 207 |
+
"BUSI_0005_B.png",
|
| 208 |
+
"BUSI_0012_M.png",
|
| 209 |
+
"BUSI_0170_B.png",
|
| 210 |
+
"BUSI_0025_B.png",
|
| 211 |
+
"BUSI_0136_B.png",
|
| 212 |
+
"BUSI_0089_B.png",
|
| 213 |
+
"BUSI_0018_B.png",
|
| 214 |
+
"BUSI_0202_M.png",
|
| 215 |
+
"BUSI_0071_B.png",
|
| 216 |
+
"BUSI_0185_M.png",
|
| 217 |
+
"BUSI_0096_B.png",
|
| 218 |
+
"BUSI_0184_M.png",
|
| 219 |
+
"BUSI_0248_B.png",
|
| 220 |
+
"BUSI_0113_B.png",
|
| 221 |
+
"BUSI_0246_B.png",
|
| 222 |
+
"BUSI_0014_M.png",
|
| 223 |
+
"BUSI_0245_B.png",
|
| 224 |
+
"BUSI_0135_B.png",
|
| 225 |
+
"BUSI_0031_B.png",
|
| 226 |
+
"BUSI_0365_B.png",
|
| 227 |
+
"BUSI_0155_B.png",
|
| 228 |
+
"BUSI_0103_M.png",
|
| 229 |
+
"BUSI_0023_B.png",
|
| 230 |
+
"BUSI_0242_B.png",
|
| 231 |
+
"BUSI_0370_B.png",
|
| 232 |
+
"BUSI_0367_B.png",
|
| 233 |
+
"BUSI_0035_M.png",
|
| 234 |
+
"BUSI_0240_B.png",
|
| 235 |
+
"BUSI_0148_M.png",
|
| 236 |
+
"BUSI_0064_B.png",
|
| 237 |
+
"BUSI_0435_B.png",
|
| 238 |
+
"BUSI_0373_B.png",
|
| 239 |
+
"BUSI_0065_B.png",
|
| 240 |
+
"BUSI_0312_B.png",
|
| 241 |
+
"BUSI_0164_B.png",
|
| 242 |
+
"BUSI_0049_B.png",
|
| 243 |
+
"BUSI_0181_M.png",
|
| 244 |
+
"BUSI_0353_B.png",
|
| 245 |
+
"BUSI_0038_M.png",
|
| 246 |
+
"BUSI_0356_B.png",
|
| 247 |
+
"BUSI_0159_B.png",
|
| 248 |
+
"BUSI_0049_M.png",
|
| 249 |
+
"BUSI_0051_B.png",
|
| 250 |
+
"BUSI_0308_B.png",
|
| 251 |
+
"BUSI_0368_B.png",
|
| 252 |
+
"BUSI_0084_B.png",
|
| 253 |
+
"BUSI_0072_M.png",
|
| 254 |
+
"BUSI_0197_M.png",
|
| 255 |
+
"BUSI_0329_B.png",
|
| 256 |
+
"BUSI_0174_M.png",
|
| 257 |
+
"BUSI_0196_M.png",
|
| 258 |
+
"BUSI_0169_M.png",
|
| 259 |
+
"BUSI_0101_B.png",
|
| 260 |
+
"BUSI_0086_M.png",
|
| 261 |
+
"BUSI_0317_B.png",
|
| 262 |
+
"BUSI_0077_M.png",
|
| 263 |
+
"BUSI_0193_B.png",
|
| 264 |
+
"BUSI_0031_M.png"
|
| 265 |
+
]
|
| 266 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/best_params.json
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"head_lr": 0.002,
|
| 3 |
+
"encoder_lr": 0.0002,
|
| 4 |
+
"weight_decay": 0.002,
|
| 5 |
+
"dropout_p": 0.3,
|
| 6 |
+
"tmax": 6,
|
| 7 |
+
"entropy_lr": 0.0002
|
| 8 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/best.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/best.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 100,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0010.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0010.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 10,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0020.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0020.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 20,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0030.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0030.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 30,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0040.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0040.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 40,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0050.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0050.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 50,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0060.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0060.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 60,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0070.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0070.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 70,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0080.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0080.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 80,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0090.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0090.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 90,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0100.pt.meta.json
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/epoch_0100.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 100,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
}
|
| 39 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/latest.pt.meta.json
ADDED
|
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/latest.pt",
|
| 3 |
+
"run_type": "final",
|
| 4 |
+
"epoch": 100,
|
| 5 |
+
"strategy": 3,
|
| 6 |
+
"dataset_percent": 0.5,
|
| 7 |
+
"backbone_family": "smp",
|
| 8 |
+
"saved_keys": [
|
| 9 |
+
"alpha_optimizer_state_dict",
|
| 10 |
+
"best_metric_name",
|
| 11 |
+
"best_metric_value",
|
| 12 |
+
"config",
|
| 13 |
+
"elapsed_seconds",
|
| 14 |
+
"epoch",
|
| 15 |
+
"epoch_metrics",
|
| 16 |
+
"log_alpha",
|
| 17 |
+
"model_state_dict",
|
| 18 |
+
"optimizer_state_dict",
|
| 19 |
+
"patience_counter",
|
| 20 |
+
"run_config",
|
| 21 |
+
"run_type",
|
| 22 |
+
"scheduler_state_dict"
|
| 23 |
+
],
|
| 24 |
+
"state_presence": {
|
| 25 |
+
"model_state_dict": true,
|
| 26 |
+
"optimizer_state_dict": true,
|
| 27 |
+
"scheduler_state_dict": true,
|
| 28 |
+
"scaler_state_dict": false,
|
| 29 |
+
"log_alpha": true,
|
| 30 |
+
"alpha_optimizer_state_dict": true,
|
| 31 |
+
"best_metric_name": true,
|
| 32 |
+
"best_metric_value": true,
|
| 33 |
+
"patience_counter": true,
|
| 34 |
+
"elapsed_seconds": true,
|
| 35 |
+
"run_config": true,
|
| 36 |
+
"epoch_metrics": true,
|
| 37 |
+
"resume_source": false
|
| 38 |
+
},
|
| 39 |
+
"last_restore": {
|
| 40 |
+
"restored_keys": {
|
| 41 |
+
"model_state_dict": true,
|
| 42 |
+
"optimizer_state_dict": true,
|
| 43 |
+
"scheduler_state_dict": true,
|
| 44 |
+
"scaler_state_dict": false,
|
| 45 |
+
"log_alpha": true,
|
| 46 |
+
"alpha_optimizer_state_dict": true,
|
| 47 |
+
"best_metric_name": true,
|
| 48 |
+
"best_metric_value": true,
|
| 49 |
+
"patience_counter": true,
|
| 50 |
+
"elapsed_seconds": true,
|
| 51 |
+
"run_config": true,
|
| 52 |
+
"epoch_metrics": true,
|
| 53 |
+
"resume_source": false
|
| 54 |
+
},
|
| 55 |
+
"restored_at_epoch": 100,
|
| 56 |
+
"expected_run_type": null
|
| 57 |
+
}
|
| 58 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/evaluation.json
ADDED
|
@@ -0,0 +1,619 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"strategy": 3,
|
| 3 |
+
"best_metric_name": "val_iou",
|
| 4 |
+
"metrics": {
|
| 5 |
+
"dice": {
|
| 6 |
+
"mean": 0.755732536315918,
|
| 7 |
+
"std": 0.2690408229827881
|
| 8 |
+
},
|
| 9 |
+
"ppv": {
|
| 10 |
+
"mean": 0.8076343536376953,
|
| 11 |
+
"std": 0.265606552362442
|
| 12 |
+
},
|
| 13 |
+
"sen": {
|
| 14 |
+
"mean": 0.7759701013565063,
|
| 15 |
+
"std": 0.27538934350013733
|
| 16 |
+
},
|
| 17 |
+
"iou": {
|
| 18 |
+
"mean": 0.6657657027244568,
|
| 19 |
+
"std": 0.27382346987724304
|
| 20 |
+
},
|
| 21 |
+
"biou": {
|
| 22 |
+
"mean": 0.12322807312011719,
|
| 23 |
+
"std": 0.10781774669885635
|
| 24 |
+
},
|
| 25 |
+
"hd95": {
|
| 26 |
+
"mean": 14.990694999694824,
|
| 27 |
+
"std": 18.093849182128906
|
| 28 |
+
}
|
| 29 |
+
},
|
| 30 |
+
"per_sample": [
|
| 31 |
+
{
|
| 32 |
+
"sample_id": "BUSI_0097_B",
|
| 33 |
+
"dice": 0.6192219766768426,
|
| 34 |
+
"ppv": 0.882583177913687,
|
| 35 |
+
"sen": 0.47691224966823226,
|
| 36 |
+
"iou": 0.4484587521226797,
|
| 37 |
+
"biou": 0.024768103786964774,
|
| 38 |
+
"hd95": 32.0
|
| 39 |
+
},
|
| 40 |
+
{
|
| 41 |
+
"sample_id": "BUSI_0415_B",
|
| 42 |
+
"dice": 0.883868244060612,
|
| 43 |
+
"ppv": 0.9470588243279212,
|
| 44 |
+
"sen": 0.8285827417711435,
|
| 45 |
+
"iou": 0.791903142995294,
|
| 46 |
+
"biou": 0.07993980694723878,
|
| 47 |
+
"hd95": 33.162445068359375
|
| 48 |
+
},
|
| 49 |
+
{
|
| 50 |
+
"sample_id": "BUSI_0290_B",
|
| 51 |
+
"dice": 0.9405453745966613,
|
| 52 |
+
"ppv": 0.9100346059673613,
|
| 53 |
+
"sen": 0.9731729892149403,
|
| 54 |
+
"iou": 0.8877637178158769,
|
| 55 |
+
"biou": 0.052631872250194355,
|
| 56 |
+
"hd95": 3.0
|
| 57 |
+
},
|
| 58 |
+
{
|
| 59 |
+
"sample_id": "BUSI_0354_B",
|
| 60 |
+
"dice": 0.8622117131220665,
|
| 61 |
+
"ppv": 0.7577962703848092,
|
| 62 |
+
"sen": 1.0,
|
| 63 |
+
"iou": 0.7577962703848092,
|
| 64 |
+
"biou": 0.020067217368823623,
|
| 65 |
+
"hd95": 11.25096321105957
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"sample_id": "BUSI_0067_B",
|
| 69 |
+
"dice": 0.9283370760086845,
|
| 70 |
+
"ppv": 0.96760259295887,
|
| 71 |
+
"sen": 0.8921340892089582,
|
| 72 |
+
"iou": 0.8662584638653412,
|
| 73 |
+
"biou": 0.12684391538527573,
|
| 74 |
+
"hd95": 7.990509033203125
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"sample_id": "BUSI_0165_M",
|
| 78 |
+
"dice": 0.8526178223063133,
|
| 79 |
+
"ppv": 0.9078258037356614,
|
| 80 |
+
"sen": 0.803739675261057,
|
| 81 |
+
"iou": 0.7430983684923832,
|
| 82 |
+
"biou": 0.033925842661414754,
|
| 83 |
+
"hd95": 21.099750518798828
|
| 84 |
+
},
|
| 85 |
+
{
|
| 86 |
+
"sample_id": "BUSI_0198_B",
|
| 87 |
+
"dice": 0.9645480688422846,
|
| 88 |
+
"ppv": 0.9563014593855947,
|
| 89 |
+
"sen": 0.9729381460735731,
|
| 90 |
+
"iou": 0.93152375499545,
|
| 91 |
+
"biou": 0.22033931341554516,
|
| 92 |
+
"hd95": 1.4142135381698608
|
| 93 |
+
},
|
| 94 |
+
{
|
| 95 |
+
"sample_id": "BUSI_0361_B",
|
| 96 |
+
"dice": 0.9277016770280945,
|
| 97 |
+
"ppv": 0.9097014992760075,
|
| 98 |
+
"sen": 0.9464285755878435,
|
| 99 |
+
"iou": 0.8651526000601419,
|
| 100 |
+
"biou": 0.1183677067886911,
|
| 101 |
+
"hd95": 2.2360680103302
|
| 102 |
+
},
|
| 103 |
+
{
|
| 104 |
+
"sample_id": "BUSI_0095_B",
|
| 105 |
+
"dice": 0.9245742275975115,
|
| 106 |
+
"ppv": 0.9743589875082115,
|
| 107 |
+
"sen": 0.8796296853566271,
|
| 108 |
+
"iou": 0.8597285702585655,
|
| 109 |
+
"biou": 0.29166765046159654,
|
| 110 |
+
"hd95": 1.4142135381698608
|
| 111 |
+
},
|
| 112 |
+
{
|
| 113 |
+
"sample_id": "BUSI_0057_B",
|
| 114 |
+
"dice": 0.923433892474735,
|
| 115 |
+
"ppv": 0.9521531329410847,
|
| 116 |
+
"sen": 0.8963964430646653,
|
| 117 |
+
"iou": 0.8577586820005682,
|
| 118 |
+
"biou": 0.2567577611381606,
|
| 119 |
+
"hd95": 1.165686011314392
|
| 120 |
+
},
|
| 121 |
+
{
|
| 122 |
+
"sample_id": "BUSI_0009_M",
|
| 123 |
+
"dice": 0.6438661974820671,
|
| 124 |
+
"ppv": 0.9020833537326346,
|
| 125 |
+
"sen": 0.5005780924187176,
|
| 126 |
+
"iou": 0.474780759344215,
|
| 127 |
+
"biou": 0.1100004449997775,
|
| 128 |
+
"hd95": 13.121795654296875
|
| 129 |
+
},
|
| 130 |
+
{
|
| 131 |
+
"sample_id": "BUSI_0399_B",
|
| 132 |
+
"dice": 0.9583955630809566,
|
| 133 |
+
"ppv": 0.9354435678699056,
|
| 134 |
+
"sen": 0.9825021879920303,
|
| 135 |
+
"iou": 0.9201147103599053,
|
| 136 |
+
"biou": 0.1854548416527849,
|
| 137 |
+
"hd95": 3.605551242828369
|
| 138 |
+
},
|
| 139 |
+
{
|
| 140 |
+
"sample_id": "BUSI_0099_M",
|
| 141 |
+
"dice": 0.7345187164611455,
|
| 142 |
+
"ppv": 0.8821796933461423,
|
| 143 |
+
"sen": 0.629201719621668,
|
| 144 |
+
"iou": 0.5804263972455042,
|
| 145 |
+
"biou": 0.08121874049810127,
|
| 146 |
+
"hd95": 11.524836540222168
|
| 147 |
+
},
|
| 148 |
+
{
|
| 149 |
+
"sample_id": "BUSI_0307_B",
|
| 150 |
+
"dice": 0.9198113302109281,
|
| 151 |
+
"ppv": 0.933014370092256,
|
| 152 |
+
"sen": 0.9069767658193568,
|
| 153 |
+
"iou": 0.8515284166968522,
|
| 154 |
+
"biou": 0.18254033131719735,
|
| 155 |
+
"hd95": 1.4142135381698608
|
| 156 |
+
},
|
| 157 |
+
{
|
| 158 |
+
"sample_id": "BUSI_0039_B",
|
| 159 |
+
"dice": 0.9066002606973523,
|
| 160 |
+
"ppv": 0.8425926290294841,
|
| 161 |
+
"sen": 0.9811320805573907,
|
| 162 |
+
"iou": 0.8291572143149853,
|
| 163 |
+
"biou": 0.18348698762661686,
|
| 164 |
+
"hd95": 2.2360680103302
|
| 165 |
+
},
|
| 166 |
+
{
|
| 167 |
+
"sample_id": "BUSI_0189_M",
|
| 168 |
+
"dice": 0.19586509989607906,
|
| 169 |
+
"ppv": 0.10856457142884712,
|
| 170 |
+
"sen": 1.0,
|
| 171 |
+
"iou": 0.10856457142884712,
|
| 172 |
+
"biou": 0.03125033637141098,
|
| 173 |
+
"hd95": 43.0
|
| 174 |
+
},
|
| 175 |
+
{
|
| 176 |
+
"sample_id": "BUSI_0021_B",
|
| 177 |
+
"dice": 0.88235298442905,
|
| 178 |
+
"ppv": 0.8053692581414374,
|
| 179 |
+
"sen": 0.9756097759270115,
|
| 180 |
+
"iou": 0.7894738227145903,
|
| 181 |
+
"biou": 0.22222345678816383,
|
| 182 |
+
"hd95": 1.5313690900802612
|
| 183 |
+
},
|
| 184 |
+
{
|
| 185 |
+
"sample_id": "BUSI_0311_B",
|
| 186 |
+
"dice": 0.29858338908160587,
|
| 187 |
+
"ppv": 0.976247036520894,
|
| 188 |
+
"sen": 0.17624360307703243,
|
| 189 |
+
"iou": 0.175491068510202,
|
| 190 |
+
"biou": 0.02142892091824253,
|
| 191 |
+
"hd95": 66.98860931396484
|
| 192 |
+
},
|
| 193 |
+
{
|
| 194 |
+
"sample_id": "BUSI_0080_M",
|
| 195 |
+
"dice": 0.6806114325392483,
|
| 196 |
+
"ppv": 0.5968955975119662,
|
| 197 |
+
"sen": 0.7916406867348292,
|
| 198 |
+
"iou": 0.5158536782173301,
|
| 199 |
+
"biou": 0.020997632284086014,
|
| 200 |
+
"hd95": 19.28207778930664
|
| 201 |
+
},
|
| 202 |
+
{
|
| 203 |
+
"sample_id": "BUSI_0130_B",
|
| 204 |
+
"dice": 0.9699525582523691,
|
| 205 |
+
"ppv": 0.9484536135614832,
|
| 206 |
+
"sen": 0.9924487602536397,
|
| 207 |
+
"iou": 0.9416581431260856,
|
| 208 |
+
"biou": 0.36666708888860744,
|
| 209 |
+
"hd95": 1.4142135381698608
|
| 210 |
+
},
|
| 211 |
+
{
|
| 212 |
+
"sample_id": "BUSI_0401_B",
|
| 213 |
+
"dice": 0.8489472124215071,
|
| 214 |
+
"ppv": 0.9323056345639655,
|
| 215 |
+
"sen": 0.7792717210492033,
|
| 216 |
+
"iou": 0.7375397806182513,
|
| 217 |
+
"biou": 0.09016430730971012,
|
| 218 |
+
"hd95": 10.271232604980469
|
| 219 |
+
},
|
| 220 |
+
{
|
| 221 |
+
"sample_id": "BUSI_0326_B",
|
| 222 |
+
"dice": 0.9486211916220264,
|
| 223 |
+
"ppv": 0.910306411682094,
|
| 224 |
+
"sen": 0.9903030308907254,
|
| 225 |
+
"iou": 0.9022639479699642,
|
| 226 |
+
"biou": 0.08097203199512874,
|
| 227 |
+
"hd95": 2.2360680103302
|
| 228 |
+
},
|
| 229 |
+
{
|
| 230 |
+
"sample_id": "BUSI_0119_B",
|
| 231 |
+
"dice": 0.20401523581035072,
|
| 232 |
+
"ppv": 0.11359521972234322,
|
| 233 |
+
"sen": 1.0,
|
| 234 |
+
"iou": 0.11359521972234322,
|
| 235 |
+
"biou": 4.3103429696797547e-07,
|
| 236 |
+
"hd95": 30.38666343688965
|
| 237 |
+
},
|
| 238 |
+
{
|
| 239 |
+
"sample_id": "BUSI_0372_B",
|
| 240 |
+
"dice": 0.9283250314938661,
|
| 241 |
+
"ppv": 0.8891336351590576,
|
| 242 |
+
"sen": 0.9711307160280099,
|
| 243 |
+
"iou": 0.8662374916854441,
|
| 244 |
+
"biou": 0.11688349918463238,
|
| 245 |
+
"hd95": 3.605551242828369
|
| 246 |
+
},
|
| 247 |
+
{
|
| 248 |
+
"sample_id": "BUSI_0003_M",
|
| 249 |
+
"dice": 0.7960716466740816,
|
| 250 |
+
"ppv": 0.9348711078872309,
|
| 251 |
+
"sen": 0.6931589845916515,
|
| 252 |
+
"iou": 0.6612284394214549,
|
| 253 |
+
"biou": 0.09500045249977375,
|
| 254 |
+
"hd95": 11.058856964111328
|
| 255 |
+
},
|
| 256 |
+
{
|
| 257 |
+
"sample_id": "BUSI_0007_M",
|
| 258 |
+
"dice": 0.2507971040598189,
|
| 259 |
+
"ppv": 1.0,
|
| 260 |
+
"sen": 0.14337798986901704,
|
| 261 |
+
"iou": 0.14337798986901704,
|
| 262 |
+
"biou": 0.01863414991667707,
|
| 263 |
+
"hd95": 15.229679107666016
|
| 264 |
+
},
|
| 265 |
+
{
|
| 266 |
+
"sample_id": "BUSI_0411_B",
|
| 267 |
+
"dice": 0.9250831008672181,
|
| 268 |
+
"ppv": 0.8871996131829518,
|
| 269 |
+
"sen": 0.9663461556439019,
|
| 270 |
+
"iou": 0.8606089504943409,
|
| 271 |
+
"biou": 0.05769260971390715,
|
| 272 |
+
"hd95": 5.0
|
| 273 |
+
},
|
| 274 |
+
{
|
| 275 |
+
"sample_id": "BUSI_0211_B",
|
| 276 |
+
"dice": 0.4797981111620932,
|
| 277 |
+
"ppv": 0.8482144212371239,
|
| 278 |
+
"sen": 0.3345072765819449,
|
| 279 |
+
"iou": 0.3156148453106827,
|
| 280 |
+
"biou": 0.08333442460187547,
|
| 281 |
+
"hd95": 12.287936210632324
|
| 282 |
+
},
|
| 283 |
+
{
|
| 284 |
+
"sample_id": "BUSI_0202_B",
|
| 285 |
+
"dice": 0.8817204936986593,
|
| 286 |
+
"ppv": 0.901099009781308,
|
| 287 |
+
"sen": 0.8631580387810118,
|
| 288 |
+
"iou": 0.7884617418637098,
|
| 289 |
+
"biou": 0.20408325697294494,
|
| 290 |
+
"hd95": 1.0414206981658936
|
| 291 |
+
},
|
| 292 |
+
{
|
| 293 |
+
"sample_id": "BUSI_0194_M",
|
| 294 |
+
"dice": 0.7745518865298691,
|
| 295 |
+
"ppv": 0.8341490030370309,
|
| 296 |
+
"sen": 0.7229029348302104,
|
| 297 |
+
"iou": 0.6320560514191985,
|
| 298 |
+
"biou": 0.02366892790860121,
|
| 299 |
+
"hd95": 9.0
|
| 300 |
+
},
|
| 301 |
+
{
|
| 302 |
+
"sample_id": "BUSI_0141_B",
|
| 303 |
+
"dice": 0.6430252656282343,
|
| 304 |
+
"ppv": 0.9435873832929453,
|
| 305 |
+
"sen": 0.4876829764645606,
|
| 306 |
+
"iou": 0.4738667960931222,
|
| 307 |
+
"biou": 0.038526124199979195,
|
| 308 |
+
"hd95": 46.145957946777344
|
| 309 |
+
},
|
| 310 |
+
{
|
| 311 |
+
"sample_id": "BUSI_0251_B",
|
| 312 |
+
"dice": 0.8594507496848547,
|
| 313 |
+
"ppv": 0.9568345479012418,
|
| 314 |
+
"sen": 0.7800587155253034,
|
| 315 |
+
"iou": 0.7535411463056243,
|
| 316 |
+
"biou": 0.16000083999915998,
|
| 317 |
+
"hd95": 4.123105525970459
|
| 318 |
+
},
|
| 319 |
+
{
|
| 320 |
+
"sample_id": "BUSI_0161_M",
|
| 321 |
+
"dice": 0.7688544670627195,
|
| 322 |
+
"ppv": 0.8252625488546514,
|
| 323 |
+
"sen": 0.7196642147121798,
|
| 324 |
+
"iou": 0.6245033195473881,
|
| 325 |
+
"biou": 0.03870988414841201,
|
| 326 |
+
"hd95": 21.170658111572266
|
| 327 |
+
},
|
| 328 |
+
{
|
| 329 |
+
"sample_id": "BUSI_0055_M",
|
| 330 |
+
"dice": 0.8626760643969446,
|
| 331 |
+
"ppv": 0.795454567591497,
|
| 332 |
+
"sen": 0.9423076997041411,
|
| 333 |
+
"iou": 0.7585139568097051,
|
| 334 |
+
"biou": 0.11224535089522913,
|
| 335 |
+
"hd95": 7.071067810058594
|
| 336 |
+
},
|
| 337 |
+
{
|
| 338 |
+
"sample_id": "BUSI_0059_M",
|
| 339 |
+
"dice": 0.8909587693621613,
|
| 340 |
+
"ppv": 0.8502014730931151,
|
| 341 |
+
"sen": 0.9358205078053612,
|
| 342 |
+
"iou": 0.8033594668900456,
|
| 343 |
+
"biou": 0.03769860363122944,
|
| 344 |
+
"hd95": 8.9442720413208
|
| 345 |
+
},
|
| 346 |
+
{
|
| 347 |
+
"sample_id": "BUSI_0285_B",
|
| 348 |
+
"dice": 0.9375967025478751,
|
| 349 |
+
"ppv": 0.9323076992504924,
|
| 350 |
+
"sen": 0.9429460640097443,
|
| 351 |
+
"iou": 0.8825242832500696,
|
| 352 |
+
"biou": 0.174157767327097,
|
| 353 |
+
"hd95": 2.9313693046569824
|
| 354 |
+
},
|
| 355 |
+
{
|
| 356 |
+
"sample_id": "BUSI_0163_B",
|
| 357 |
+
"dice": 0.23013709176204225,
|
| 358 |
+
"ppv": 0.13003109442243121,
|
| 359 |
+
"sen": 1.0,
|
| 360 |
+
"iou": 0.13003109442243121,
|
| 361 |
+
"biou": 0.0666674444437963,
|
| 362 |
+
"hd95": 25.87345314025879
|
| 363 |
+
},
|
| 364 |
+
{
|
| 365 |
+
"sample_id": "BUSI_0224_B",
|
| 366 |
+
"dice": 2.323420020366635e-08,
|
| 367 |
+
"ppv": 4.446420433685174e-08,
|
| 368 |
+
"sen": 4.866179811864729e-08,
|
| 369 |
+
"iou": 2.323420020366635e-08,
|
| 370 |
+
"biou": 2.9154510450580044e-07,
|
| 371 |
+
"hd95": 59.95482635498047
|
| 372 |
+
},
|
| 373 |
+
{
|
| 374 |
+
"sample_id": "BUSI_0316_B",
|
| 375 |
+
"dice": 0.830335353608105,
|
| 376 |
+
"ppv": 0.8693923931593338,
|
| 377 |
+
"sen": 0.7946366817536906,
|
| 378 |
+
"iou": 0.709891812830111,
|
| 379 |
+
"biou": 0.0301509162226271,
|
| 380 |
+
"hd95": 17.0
|
| 381 |
+
},
|
| 382 |
+
{
|
| 383 |
+
"sample_id": "BUSI_0116_M",
|
| 384 |
+
"dice": 0.08680357262149134,
|
| 385 |
+
"ppv": 0.09002444160286599,
|
| 386 |
+
"sen": 0.08380531327233146,
|
| 387 |
+
"iou": 0.045370996605089114,
|
| 388 |
+
"biou": 4.0322564386062747e-07,
|
| 389 |
+
"hd95": 90.3084945678711
|
| 390 |
+
},
|
| 391 |
+
{
|
| 392 |
+
"sample_id": "BUSI_0214_B",
|
| 393 |
+
"dice": 0.9159482803907176,
|
| 394 |
+
"ppv": 0.8947368531855944,
|
| 395 |
+
"sen": 0.9381898522969258,
|
| 396 |
+
"iou": 0.8449304329094998,
|
| 397 |
+
"biou": 0.11290370273994477,
|
| 398 |
+
"hd95": 4.0861735343933105
|
| 399 |
+
},
|
| 400 |
+
{
|
| 401 |
+
"sample_id": "BUSI_0207_M",
|
| 402 |
+
"dice": 5.020080069273089e-08,
|
| 403 |
+
"ppv": 5.649714322195299e-07,
|
| 404 |
+
"sen": 5.5096415697167185e-08,
|
| 405 |
+
"iou": 5.020080069273089e-08,
|
| 406 |
+
"biou": 4.098358976082387e-07,
|
| 407 |
+
"hd95": 58.369380950927734
|
| 408 |
+
},
|
| 409 |
+
{
|
| 410 |
+
"sample_id": "BUSI_0016_M",
|
| 411 |
+
"dice": 0.5957446885998718,
|
| 412 |
+
"ppv": 0.8178947464265923,
|
| 413 |
+
"sen": 0.4684956446036887,
|
| 414 |
+
"iou": 0.42424243996062133,
|
| 415 |
+
"biou": 0.03703725994507872,
|
| 416 |
+
"hd95": 34.12311935424805
|
| 417 |
+
},
|
| 418 |
+
{
|
| 419 |
+
"sample_id": "BUSI_0067_M",
|
| 420 |
+
"dice": 0.013072024435029486,
|
| 421 |
+
"ppv": 0.07936654068803065,
|
| 422 |
+
"sen": 0.007122648558027271,
|
| 423 |
+
"iou": 0.006579078081700252,
|
| 424 |
+
"biou": 6.060602387513705e-07,
|
| 425 |
+
"hd95": 34.15515899658203
|
| 426 |
+
},
|
| 427 |
+
{
|
| 428 |
+
"sample_id": "BUSI_0184_B",
|
| 429 |
+
"dice": 0.9609357766290935,
|
| 430 |
+
"ppv": 0.9981659789928483,
|
| 431 |
+
"sen": 0.9263829818560433,
|
| 432 |
+
"iou": 0.9248088392179763,
|
| 433 |
+
"biou": 0.29663630072284386,
|
| 434 |
+
"hd95": 2.2360680103302
|
| 435 |
+
},
|
| 436 |
+
{
|
| 437 |
+
"sample_id": "BUSI_0099_B",
|
| 438 |
+
"dice": 0.9681620862400806,
|
| 439 |
+
"ppv": 0.9838235317906571,
|
| 440 |
+
"sen": 0.9529914596878263,
|
| 441 |
+
"iou": 0.9382889287112302,
|
| 442 |
+
"biou": 0.3779532457061058,
|
| 443 |
+
"hd95": 1.0
|
| 444 |
+
},
|
| 445 |
+
{
|
| 446 |
+
"sample_id": "BUSI_0412_B",
|
| 447 |
+
"dice": 0.8141451758030207,
|
| 448 |
+
"ppv": 0.9749522399712961,
|
| 449 |
+
"sen": 0.6988740155375226,
|
| 450 |
+
"iou": 0.6865470898871885,
|
| 451 |
+
"biou": 0.08918423355137885,
|
| 452 |
+
"hd95": 28.2966365814209
|
| 453 |
+
},
|
| 454 |
+
{
|
| 455 |
+
"sample_id": "BUSI_0100_M",
|
| 456 |
+
"dice": 0.6883561821714508,
|
| 457 |
+
"ppv": 0.787206294098395,
|
| 458 |
+
"sen": 0.6115619055211049,
|
| 459 |
+
"iou": 0.5248042189030271,
|
| 460 |
+
"biou": 0.0691248529378558,
|
| 461 |
+
"hd95": 14.083578109741211
|
| 462 |
+
},
|
| 463 |
+
{
|
| 464 |
+
"sample_id": "BUSI_0073_M",
|
| 465 |
+
"dice": 0.8588149335175616,
|
| 466 |
+
"ppv": 0.7805851355604873,
|
| 467 |
+
"sen": 0.9544715521184468,
|
| 468 |
+
"iou": 0.7525641342866495,
|
| 469 |
+
"biou": 0.1271681345848933,
|
| 470 |
+
"hd95": 5.0
|
| 471 |
+
},
|
| 472 |
+
{
|
| 473 |
+
"sample_id": "BUSI_0357_B",
|
| 474 |
+
"dice": 0.8422478102743527,
|
| 475 |
+
"ppv": 0.7300469800414343,
|
| 476 |
+
"sen": 0.9952000007679999,
|
| 477 |
+
"iou": 0.7274854119900103,
|
| 478 |
+
"biou": 0.17791461477630996,
|
| 479 |
+
"hd95": 8.96934986114502
|
| 480 |
+
},
|
| 481 |
+
{
|
| 482 |
+
"sample_id": "BUSI_0229_B",
|
| 483 |
+
"dice": 0.9562899810079968,
|
| 484 |
+
"ppv": 0.9573105701909743,
|
| 485 |
+
"sen": 0.9552715702586188,
|
| 486 |
+
"iou": 0.9162410708640377,
|
| 487 |
+
"biou": 0.19653225633973623,
|
| 488 |
+
"hd95": 1.4142135381698608
|
| 489 |
+
},
|
| 490 |
+
{
|
| 491 |
+
"sample_id": "BUSI_0129_M",
|
| 492 |
+
"dice": 0.8464295092524298,
|
| 493 |
+
"ppv": 0.9817518251875137,
|
| 494 |
+
"sen": 0.7438930750669369,
|
| 495 |
+
"iou": 0.7337475415404222,
|
| 496 |
+
"biou": 0.040776885286041696,
|
| 497 |
+
"hd95": 16.76305389404297
|
| 498 |
+
},
|
| 499 |
+
{
|
| 500 |
+
"sample_id": "BUSI_0156_B",
|
| 501 |
+
"dice": 0.9581239601132395,
|
| 502 |
+
"ppv": 0.9694915357655811,
|
| 503 |
+
"sen": 0.9470198850927533,
|
| 504 |
+
"iou": 0.9196141737575004,
|
| 505 |
+
"biou": 0.38461617356900824,
|
| 506 |
+
"hd95": 1.0
|
| 507 |
+
},
|
| 508 |
+
{
|
| 509 |
+
"sample_id": "BUSI_0320_B",
|
| 510 |
+
"dice": 0.9337748417467062,
|
| 511 |
+
"ppv": 0.8905263388365603,
|
| 512 |
+
"sen": 0.9814385193878146,
|
| 513 |
+
"iou": 0.875776423234695,
|
| 514 |
+
"biou": 0.2792799285766409,
|
| 515 |
+
"hd95": 2.0
|
| 516 |
+
},
|
| 517 |
+
{
|
| 518 |
+
"sample_id": "BUSI_0104_B",
|
| 519 |
+
"dice": 0.9219858432674315,
|
| 520 |
+
"ppv": 0.9420290275151975,
|
| 521 |
+
"sen": 0.902777845293163,
|
| 522 |
+
"iou": 0.8552632531162808,
|
| 523 |
+
"biou": 0.29824684518097333,
|
| 524 |
+
"hd95": 1.0
|
| 525 |
+
},
|
| 526 |
+
{
|
| 527 |
+
"sample_id": "BUSI_0340_B",
|
| 528 |
+
"dice": 0.8823602965075598,
|
| 529 |
+
"ppv": 0.9378155743338938,
|
| 530 |
+
"sen": 0.8330972655076189,
|
| 531 |
+
"iou": 0.7894854633224728,
|
| 532 |
+
"biou": 0.06067983964081562,
|
| 533 |
+
"hd95": 14.785070419311523
|
| 534 |
+
},
|
| 535 |
+
{
|
| 536 |
+
"sample_id": "BUSI_0121_B",
|
| 537 |
+
"dice": 0.9534246607243383,
|
| 538 |
+
"ppv": 0.9508196788497707,
|
| 539 |
+
"sen": 0.9560439620818734,
|
| 540 |
+
"iou": 0.9109947760478042,
|
| 541 |
+
"biou": 0.3308828449390846,
|
| 542 |
+
"hd95": 2.0
|
| 543 |
+
},
|
| 544 |
+
{
|
| 545 |
+
"sample_id": "BUSI_0358_B",
|
| 546 |
+
"dice": 0.8837492462721092,
|
| 547 |
+
"ppv": 0.79692647673694,
|
| 548 |
+
"sen": 0.9918032798082951,
|
| 549 |
+
"iou": 0.7917121274032577,
|
| 550 |
+
"biou": 0.0893859835832494,
|
| 551 |
+
"hd95": 4.4721360206604
|
| 552 |
+
},
|
| 553 |
+
{
|
| 554 |
+
"sample_id": "BUSI_0270_B",
|
| 555 |
+
"dice": 0.8345498984732483,
|
| 556 |
+
"ppv": 0.7555066617826736,
|
| 557 |
+
"sen": 0.9320652358518381,
|
| 558 |
+
"iou": 0.7160752158506856,
|
| 559 |
+
"biou": 0.1151085502816185,
|
| 560 |
+
"hd95": 12.188106536865234
|
| 561 |
+
},
|
| 562 |
+
{
|
| 563 |
+
"sample_id": "BUSI_0023_M",
|
| 564 |
+
"dice": 0.7543169425757712,
|
| 565 |
+
"ppv": 0.9141721877040481,
|
| 566 |
+
"sen": 0.6420465182875067,
|
| 567 |
+
"iou": 0.6055448393482218,
|
| 568 |
+
"biou": 0.06343301055354282,
|
| 569 |
+
"hd95": 31.170482635498047
|
| 570 |
+
},
|
| 571 |
+
{
|
| 572 |
+
"sample_id": "BUSI_0020_B",
|
| 573 |
+
"dice": 0.9274035364680486,
|
| 574 |
+
"ppv": 0.8656898820891474,
|
| 575 |
+
"sen": 0.998591549494148,
|
| 576 |
+
"iou": 0.8646341628494923,
|
| 577 |
+
"biou": 0.10052957643937754,
|
| 578 |
+
"hd95": 2.0
|
| 579 |
+
},
|
| 580 |
+
{
|
| 581 |
+
"sample_id": "BUSI_0153_M",
|
| 582 |
+
"dice": 0.8544416008032145,
|
| 583 |
+
"ppv": 0.9904109591230374,
|
| 584 |
+
"sen": 0.7512989305282485,
|
| 585 |
+
"iou": 0.7458734569141428,
|
| 586 |
+
"biou": 0.049900389241439275,
|
| 587 |
+
"hd95": 13.03840446472168
|
| 588 |
+
},
|
| 589 |
+
{
|
| 590 |
+
"sample_id": "BUSI_0154_B",
|
| 591 |
+
"dice": 0.9357143086734612,
|
| 592 |
+
"ppv": 0.8851352127464779,
|
| 593 |
+
"sen": 0.9924242481634483,
|
| 594 |
+
"iou": 0.8791947119498578,
|
| 595 |
+
"biou": 0.40384730029365323,
|
| 596 |
+
"hd95": 1.0
|
| 597 |
+
},
|
| 598 |
+
{
|
| 599 |
+
"sample_id": "BUSI_0062_M",
|
| 600 |
+
"dice": 0.7258022508269697,
|
| 601 |
+
"ppv": 0.8922386214086837,
|
| 602 |
+
"sen": 0.6116974665040691,
|
| 603 |
+
"iou": 0.5696150872393494,
|
| 604 |
+
"biou": 0.02116428011526981,
|
| 605 |
+
"hd95": 16.5150146484375
|
| 606 |
+
},
|
| 607 |
+
{
|
| 608 |
+
"sample_id": "BUSI_0075_B",
|
| 609 |
+
"dice": 0.9314336659453548,
|
| 610 |
+
"ppv": 0.9794007529212073,
|
| 611 |
+
"sen": 0.8879456896526843,
|
| 612 |
+
"iou": 0.871666688055552,
|
| 613 |
+
"biou": 0.22834706429365015,
|
| 614 |
+
"hd95": 2.2360680103302
|
| 615 |
+
}
|
| 616 |
+
],
|
| 617 |
+
"checkpoint_mode": "latest",
|
| 618 |
+
"checkpoint_path": "/workspace/runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/checkpoints/latest.pt"
|
| 619 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/evaluation_results.xlsx
ADDED
|
Binary file (35.2 kB). View file
|
|
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/evaluation_summary.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"mean_iou": 0.6657657027244568,
|
| 3 |
+
"mean_dice": 0.755732536315918
|
| 4 |
+
}
|
runs/EfficientNet_Strategy3_NewHyperParameters_RunPod_Fullpooling/pct_50/strategy_3/final/history.json
ADDED
|
@@ -0,0 +1,2242 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"epoch": 1,
|
| 4 |
+
"train_loss": 1.122809389656655,
|
| 5 |
+
"train_actor_loss": -0.3107988336094949,
|
| 6 |
+
"train_critic_loss": 0.0034548409573973856,
|
| 7 |
+
"train_mean_reward": -0.00021353209795945012,
|
| 8 |
+
"train_entropy": 0.5320818992578292,
|
| 9 |
+
"train_ce_loss": 1.318279987903798,
|
| 10 |
+
"train_dice_loss": 1.5454815886433424,
|
| 11 |
+
"train_dice": 0.31493909170089024,
|
| 12 |
+
"train_iou": 0.21633165172112426,
|
| 13 |
+
"grad_norm": 16.97805720798729,
|
| 14 |
+
"lr": 0.001999509027563903,
|
| 15 |
+
"encoder_lr": 0.0001999531232347445,
|
| 16 |
+
"alpha": 0.4933896064758301,
|
| 17 |
+
"validated_this_epoch": false,
|
| 18 |
+
"val_loss": null,
|
| 19 |
+
"val_dice": null,
|
| 20 |
+
"val_iou": null,
|
| 21 |
+
"val_reward": null,
|
| 22 |
+
"val_entropy": null
|
| 23 |
+
},
|
| 24 |
+
{
|
| 25 |
+
"epoch": 2,
|
| 26 |
+
"train_loss": 0.659615898768367,
|
| 27 |
+
"train_actor_loss": -0.38683646775839864,
|
| 28 |
+
"train_critic_loss": 0.0006725203085468735,
|
| 29 |
+
"train_mean_reward": -0.00638562779031253,
|
| 30 |
+
"train_entropy": 0.646799662258736,
|
| 31 |
+
"train_ce_loss": 0.7138427807263625,
|
| 32 |
+
"train_dice_loss": 1.378389415814895,
|
| 33 |
+
"train_dice": 0.44636741258536083,
|
| 34 |
+
"train_iou": 0.33164686304264407,
|
| 35 |
+
"grad_norm": 6.8028083591498145,
|
| 36 |
+
"lr": 0.0019980365947861304,
|
| 37 |
+
"encoder_lr": 0.00019981253920068581,
|
| 38 |
+
"alpha": 0.48246774077415466,
|
| 39 |
+
"validated_this_epoch": false,
|
| 40 |
+
"val_loss": null,
|
| 41 |
+
"val_dice": null,
|
| 42 |
+
"val_iou": null,
|
| 43 |
+
"val_reward": null,
|
| 44 |
+
"val_entropy": null
|
| 45 |
+
},
|
| 46 |
+
{
|
| 47 |
+
"epoch": 3,
|
| 48 |
+
"train_loss": 0.5888693920416079,
|
| 49 |
+
"train_actor_loss": -0.39422753339185734,
|
| 50 |
+
"train_critic_loss": 0.000567077534476768,
|
| 51 |
+
"train_mean_reward": -0.007123046205924295,
|
| 52 |
+
"train_entropy": 0.6754094219022944,
|
| 53 |
+
"train_ce_loss": 0.6279669601027433,
|
| 54 |
+
"train_dice_loss": 1.3376597953089138,
|
| 55 |
+
"train_dice": 0.4991640322708593,
|
| 56 |
+
"train_iou": 0.3748575024941418,
|
| 57 |
+
"grad_norm": 5.378530751827151,
|
| 58 |
+
"lr": 0.001995584154780065,
|
| 59 |
+
"encoder_lr": 0.0001995783866372926,
|
| 60 |
+
"alpha": 0.470917284488678,
|
| 61 |
+
"validated_this_epoch": false,
|
| 62 |
+
"val_loss": null,
|
| 63 |
+
"val_dice": null,
|
| 64 |
+
"val_iou": null,
|
| 65 |
+
"val_reward": null,
|
| 66 |
+
"val_entropy": null
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"epoch": 4,
|
| 70 |
+
"train_loss": 0.5727213552481559,
|
| 71 |
+
"train_actor_loss": -0.3923587586143529,
|
| 72 |
+
"train_critic_loss": 0.00047669555851708356,
|
| 73 |
+
"train_mean_reward": -0.006709725253815066,
|
| 74 |
+
"train_entropy": 0.696851239879002,
|
| 75 |
+
"train_ce_loss": 0.6159397132997942,
|
| 76 |
+
"train_dice_loss": 1.3137438036982714,
|
| 77 |
+
"train_dice": 0.5243197822805589,
|
| 78 |
+
"train_iou": 0.4008426038757825,
|
| 79 |
+
"grad_norm": 4.8283826584039735,
|
| 80 |
+
"lr": 0.001992154127807906,
|
| 81 |
+
"encoder_lr": 0.0001992508966248754,
|
| 82 |
+
"alpha": 0.45951777696609497,
|
| 83 |
+
"validated_this_epoch": false,
|
| 84 |
+
"val_loss": null,
|
| 85 |
+
"val_dice": null,
|
| 86 |
+
"val_iou": null,
|
| 87 |
+
"val_reward": null,
|
| 88 |
+
"val_entropy": null
|
| 89 |
+
},
|
| 90 |
+
{
|
| 91 |
+
"epoch": 5,
|
| 92 |
+
"train_loss": 0.5343380287441231,
|
| 93 |
+
"train_actor_loss": -0.39298394216508514,
|
| 94 |
+
"train_critic_loss": 0.0004594868102147546,
|
| 95 |
+
"train_mean_reward": -0.007040507830981735,
|
| 96 |
+
"train_entropy": 0.6930879724118136,
|
| 97 |
+
"train_ce_loss": 0.5805706514421013,
|
| 98 |
+
"train_dice_loss": 1.2736137964466745,
|
| 99 |
+
"train_dice": 0.5347368901202008,
|
| 100 |
+
"train_iou": 0.4115678787257616,
|
| 101 |
+
"grad_norm": 4.088428035724995,
|
| 102 |
+
"lr": 0.0019877498988921626,
|
| 103 |
+
"encoder_lr": 0.0001988303923565381,
|
| 104 |
+
"alpha": 0.44812825322151184,
|
| 105 |
+
"validated_this_epoch": false,
|
| 106 |
+
"val_loss": null,
|
| 107 |
+
"val_dice": null,
|
| 108 |
+
"val_iou": null,
|
| 109 |
+
"val_reward": null,
|
| 110 |
+
"val_entropy": null
|
| 111 |
+
},
|
| 112 |
+
{
|
| 113 |
+
"epoch": 6,
|
| 114 |
+
"train_loss": 0.5276677184954806,
|
| 115 |
+
"train_actor_loss": -0.36535874314278904,
|
| 116 |
+
"train_critic_loss": 0.00045152983492997156,
|
| 117 |
+
"train_mean_reward": -0.006463098140352662,
|
| 118 |
+
"train_entropy": 0.6429887968440389,
|
| 119 |
+
"train_ce_loss": 0.5240337582851353,
|
| 120 |
+
"train_dice_loss": 1.2615676209600089,
|
| 121 |
+
"train_dice": 0.579636548634965,
|
| 122 |
+
"train_iou": 0.45355634697326713,
|
| 123 |
+
"grad_norm": 4.1968280890191245,
|
| 124 |
+
"lr": 0.0019823758144750458,
|
| 125 |
+
"encoder_lr": 0.00019831728881922545,
|
| 126 |
+
"alpha": 0.438424289226532,
|
| 127 |
+
"validated_this_epoch": false,
|
| 128 |
+
"val_loss": null,
|
| 129 |
+
"val_dice": null,
|
| 130 |
+
"val_iou": null,
|
| 131 |
+
"val_reward": null,
|
| 132 |
+
"val_entropy": null
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"epoch": 7,
|
| 136 |
+
"train_loss": 0.5341179094208737,
|
| 137 |
+
"train_actor_loss": -0.3562825362291909,
|
| 138 |
+
"train_critic_loss": 0.00043695481056928815,
|
| 139 |
+
"train_mean_reward": -0.006400862039168074,
|
| 140 |
+
"train_entropy": 0.6352313085620409,
|
| 141 |
+
"train_ce_loss": 0.5317916763538509,
|
| 142 |
+
"train_dice_loss": 1.2485722442005955,
|
| 143 |
+
"train_dice": 0.5823677692036895,
|
| 144 |
+
"train_iou": 0.4572046968140397,
|
| 145 |
+
"grad_norm": 4.425033004246941,
|
| 146 |
+
"lr": 0.001976037178129054,
|
| 147 |
+
"encoder_lr": 0.00019771209238418101,
|
| 148 |
+
"alpha": 0.4290330708026886,
|
| 149 |
+
"validated_this_epoch": false,
|
| 150 |
+
"val_loss": null,
|
| 151 |
+
"val_dice": null,
|
| 152 |
+
"val_iou": null,
|
| 153 |
+
"val_reward": null,
|
| 154 |
+
"val_entropy": null
|
| 155 |
+
},
|
| 156 |
+
{
|
| 157 |
+
"epoch": 8,
|
| 158 |
+
"train_loss": 0.5017420548274246,
|
| 159 |
+
"train_actor_loss": -0.3661940216883363,
|
| 160 |
+
"train_critic_loss": 0.0003796231855515894,
|
| 161 |
+
"train_mean_reward": -0.006368049725621156,
|
| 162 |
+
"train_entropy": 0.6572325930422898,
|
| 163 |
+
"train_ce_loss": 0.5000163667175909,
|
| 164 |
+
"train_dice_loss": 1.2354761503001517,
|
| 165 |
+
"train_dice": 0.6074047675857387,
|
| 166 |
+
"train_iou": 0.4828729469481893,
|
| 167 |
+
"grad_norm": 3.5648219220398008,
|
| 168 |
+
"lr": 0.001968740245322988,
|
| 169 |
+
"encoder_lr": 0.00019701540030721995,
|
| 170 |
+
"alpha": 0.41927021741867065,
|
| 171 |
+
"validated_this_epoch": false,
|
| 172 |
+
"val_loss": null,
|
| 173 |
+
"val_dice": null,
|
| 174 |
+
"val_iou": null,
|
| 175 |
+
"val_reward": null,
|
| 176 |
+
"val_entropy": null
|
| 177 |
+
},
|
| 178 |
+
{
|
| 179 |
+
"epoch": 9,
|
| 180 |
+
"train_loss": 0.5298668614610337,
|
| 181 |
+
"train_actor_loss": -0.33232456169425667,
|
| 182 |
+
"train_critic_loss": 0.0003912863492549747,
|
| 183 |
+
"train_mean_reward": -0.006101896421049732,
|
| 184 |
+
"train_entropy": 0.6075103389970399,
|
| 185 |
+
"train_ce_loss": 0.4967910853197717,
|
| 186 |
+
"train_dice_loss": 1.2272004577540612,
|
| 187 |
+
"train_dice": 0.6214794504847432,
|
| 188 |
+
"train_iou": 0.4969447840739749,
|
| 189 |
+
"grad_norm": 3.8638702224391377,
|
| 190 |
+
"lr": 0.0019604922172485584,
|
| 191 |
+
"encoder_lr": 0.0001962279001393096,
|
| 192 |
+
"alpha": 0.4109015166759491,
|
| 193 |
+
"validated_this_epoch": false,
|
| 194 |
+
"val_loss": null,
|
| 195 |
+
"val_dice": null,
|
| 196 |
+
"val_iou": null,
|
| 197 |
+
"val_reward": null,
|
| 198 |
+
"val_entropy": null
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"epoch": 10,
|
| 202 |
+
"train_loss": 0.518087082333112,
|
| 203 |
+
"train_actor_loss": -0.3156055822025128,
|
| 204 |
+
"train_critic_loss": 0.00033443726392490904,
|
| 205 |
+
"train_mean_reward": -0.006073546188220412,
|
| 206 |
+
"train_entropy": 0.5958120547408281,
|
| 207 |
+
"train_ce_loss": 0.4600338188255115,
|
| 208 |
+
"train_dice_loss": 1.2070170555465904,
|
| 209 |
+
"train_dice": 0.63123976513376,
|
| 210 |
+
"train_iou": 0.5117284662461407,
|
| 211 |
+
"grad_norm": 3.6160758456518485,
|
| 212 |
+
"lr": 0.0019513012337136779,
|
| 213 |
+
"encoder_lr": 0.00019535036904803958,
|
| 214 |
+
"alpha": 0.4023796021938324,
|
| 215 |
+
"validated_this_epoch": true,
|
| 216 |
+
"val_loss": 0.8509404014259803,
|
| 217 |
+
"val_dice": 0.6255302644633656,
|
| 218 |
+
"val_iou": 0.5221329308557149,
|
| 219 |
+
"val_reward": 0.000625012848839179,
|
| 220 |
+
"val_entropy": 0.34133980852184875,
|
| 221 |
+
"val_actor_loss": -0.003340385114007413,
|
| 222 |
+
"val_critic_loss": 0.00016816472167982907,
|
| 223 |
+
"val_ce_loss": 0.381730621953429,
|
| 224 |
+
"val_dice_loss": 1.3266627616954572
|
| 225 |
+
},
|
| 226 |
+
{
|
| 227 |
+
"epoch": 11,
|
| 228 |
+
"train_loss": 0.5167460150388611,
|
| 229 |
+
"train_actor_loss": -0.3242803153507315,
|
| 230 |
+
"train_critic_loss": 0.00033826984984911365,
|
| 231 |
+
"train_mean_reward": -0.005761756748568314,
|
| 232 |
+
"train_entropy": 0.6119530427363492,
|
| 233 |
+
"train_ce_loss": 0.46831092955448334,
|
| 234 |
+
"train_dice_loss": 1.2134034453436386,
|
| 235 |
+
"train_dice": 0.637329815003702,
|
| 236 |
+
"train_iou": 0.5164420828153835,
|
| 237 |
+
"grad_norm": 3.7716032779493998,
|
| 238 |
+
"lr": 0.0019411763651094546,
|
| 239 |
+
"encoder_lr": 0.0001943836730506514,
|
| 240 |
+
"alpha": 0.39455240964889526,
|
| 241 |
+
"validated_this_epoch": false,
|
| 242 |
+
"val_loss": null,
|
| 243 |
+
"val_dice": null,
|
| 244 |
+
"val_iou": null,
|
| 245 |
+
"val_reward": null,
|
| 246 |
+
"val_entropy": null
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"epoch": 12,
|
| 250 |
+
"train_loss": 0.5149066645883081,
|
| 251 |
+
"train_actor_loss": -0.3023359139067258,
|
| 252 |
+
"train_critic_loss": 0.0003454410829414744,
|
| 253 |
+
"train_mean_reward": -0.005893218922234008,
|
| 254 |
+
"train_entropy": 0.5786789347554824,
|
| 255 |
+
"train_ce_loss": 0.44421532922075435,
|
| 256 |
+
"train_dice_loss": 1.1899243720280108,
|
| 257 |
+
"train_dice": 0.6346212186534125,
|
| 258 |
+
"train_iou": 0.5128833046584401,
|
| 259 |
+
"grad_norm": 3.72456025014552,
|
| 260 |
+
"lr": 0.0019301276034588103,
|
| 261 |
+
"encoder_lr": 0.00019332876615938386,
|
| 262 |
+
"alpha": 0.3871282935142517,
|
| 263 |
+
"validated_this_epoch": false,
|
| 264 |
+
"val_loss": null,
|
| 265 |
+
"val_dice": null,
|
| 266 |
+
"val_iou": null,
|
| 267 |
+
"val_reward": null,
|
| 268 |
+
"val_entropy": null
|
| 269 |
+
},
|
| 270 |
+
{
|
| 271 |
+
"epoch": 13,
|
| 272 |
+
"train_loss": 0.506700577120506,
|
| 273 |
+
"train_actor_loss": -0.21322156833268183,
|
| 274 |
+
"train_critic_loss": 0.0002509659909149727,
|
| 275 |
+
"train_mean_reward": -0.0020937638013027366,
|
| 276 |
+
"train_entropy": 0.4639017832313859,
|
| 277 |
+
"train_ce_loss": 0.4945101185343216,
|
| 278 |
+
"train_dice_loss": 0.9450831876860725,
|
| 279 |
+
"train_dice": 0.6890812154864222,
|
| 280 |
+
"train_iou": 0.581995713237155,
|
| 281 |
+
"grad_norm": 7.624448425548021,
|
| 282 |
+
"lr": 0.0019181658525555612,
|
| 283 |
+
"encoder_lr": 0.00019218668943997818,
|
| 284 |
+
"alpha": 0.3823469281196594,
|
| 285 |
+
"validated_this_epoch": false,
|
| 286 |
+
"val_loss": null,
|
| 287 |
+
"val_dice": null,
|
| 288 |
+
"val_iou": null,
|
| 289 |
+
"val_reward": null,
|
| 290 |
+
"val_entropy": null
|
| 291 |
+
},
|
| 292 |
+
{
|
| 293 |
+
"epoch": 14,
|
| 294 |
+
"train_loss": 0.4831936905691184,
|
| 295 |
+
"train_actor_loss": -0.18172498041094892,
|
| 296 |
+
"train_critic_loss": 0.0001759705503063836,
|
| 297 |
+
"train_mean_reward": -0.0004059999807985677,
|
| 298 |
+
"train_entropy": 0.4723863771681022,
|
| 299 |
+
"train_ce_loss": 0.5914532082030208,
|
| 300 |
+
"train_dice_loss": 0.7382081475214749,
|
| 301 |
+
"train_dice": 0.7205163327186863,
|
| 302 |
+
"train_iou": 0.6141165019856214,
|
| 303 |
+
"grad_norm": 15.514881435290787,
|
| 304 |
+
"lr": 0.0019053029172036895,
|
| 305 |
+
"encoder_lr": 0.00019095856998427185,
|
| 306 |
+
"alpha": 0.3782835900783539,
|
| 307 |
+
"validated_this_epoch": false,
|
| 308 |
+
"val_loss": null,
|
| 309 |
+
"val_dice": null,
|
| 310 |
+
"val_iou": null,
|
| 311 |
+
"val_reward": null,
|
| 312 |
+
"val_entropy": null
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"epoch": 15,
|
| 316 |
+
"train_loss": 0.4567362353554927,
|
| 317 |
+
"train_actor_loss": -0.1734470616958708,
|
| 318 |
+
"train_critic_loss": 0.00015864743494420008,
|
| 319 |
+
"train_mean_reward": -1.3926346570842864e-06,
|
| 320 |
+
"train_entropy": 0.46809116042476606,
|
| 321 |
+
"train_ce_loss": 0.5835047035937699,
|
| 322 |
+
"train_dice_loss": 0.6767032292770168,
|
| 323 |
+
"train_dice": 0.7341928305310697,
|
| 324 |
+
"train_iou": 0.6337336152007866,
|
| 325 |
+
"grad_norm": 23.462444724962694,
|
| 326 |
+
"lr": 0.001891551491567426,
|
| 327 |
+
"encoder_lr": 0.00018964561979789495,
|
| 328 |
+
"alpha": 0.37365731596946716,
|
| 329 |
+
"validated_this_epoch": false,
|
| 330 |
+
"val_loss": null,
|
| 331 |
+
"val_dice": null,
|
| 332 |
+
"val_iou": null,
|
| 333 |
+
"val_reward": null,
|
| 334 |
+
"val_entropy": null
|
| 335 |
+
},
|
| 336 |
+
{
|
| 337 |
+
"epoch": 16,
|
| 338 |
+
"train_loss": 0.45632827443128465,
|
| 339 |
+
"train_actor_loss": -0.17530018648551415,
|
| 340 |
+
"train_critic_loss": 0.00016799280645509804,
|
| 341 |
+
"train_mean_reward": -0.00014099961057680528,
|
| 342 |
+
"train_entropy": 0.48662347048135995,
|
| 343 |
+
"train_ce_loss": 0.5881788068118888,
|
| 344 |
+
"train_dice_loss": 0.6749101080173668,
|
| 345 |
+
"train_dice": 0.7398606067164106,
|
| 346 |
+
"train_iou": 0.6412177536366453,
|
| 347 |
+
"grad_norm": 23.895725244699523,
|
| 348 |
+
"lr": 0.0018769251466436443,
|
| 349 |
+
"encoder_lr": 0.00018824913460416704,
|
| 350 |
+
"alpha": 0.3683401942253113,
|
| 351 |
+
"validated_this_epoch": false,
|
| 352 |
+
"val_loss": null,
|
| 353 |
+
"val_dice": null,
|
| 354 |
+
"val_iou": null,
|
| 355 |
+
"val_reward": null,
|
| 356 |
+
"val_entropy": null
|
| 357 |
+
},
|
| 358 |
+
{
|
| 359 |
+
"epoch": 17,
|
| 360 |
+
"train_loss": 0.4239559060913807,
|
| 361 |
+
"train_actor_loss": -0.17023482211222946,
|
| 362 |
+
"train_critic_loss": 0.00018060995567953535,
|
| 363 |
+
"train_mean_reward": 4.863066662663244e-05,
|
| 364 |
+
"train_entropy": 0.4738664231924894,
|
| 365 |
+
"train_ce_loss": 0.5530880968620668,
|
| 366 |
+
"train_dice_loss": 0.6351127378084246,
|
| 367 |
+
"train_dice": 0.7492821696154174,
|
| 368 |
+
"train_iou": 0.6516460798863787,
|
| 369 |
+
"grad_norm": 23.573979053386424,
|
| 370 |
+
"lr": 0.001861438316868924,
|
| 371 |
+
"encoder_lr": 0.00018677049256537467,
|
| 372 |
+
"alpha": 0.3631233870983124,
|
| 373 |
+
"validated_this_epoch": false,
|
| 374 |
+
"val_loss": null,
|
| 375 |
+
"val_dice": null,
|
| 376 |
+
"val_iou": null,
|
| 377 |
+
"val_reward": null,
|
| 378 |
+
"val_entropy": null
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"epoch": 18,
|
| 382 |
+
"train_loss": 0.4458137009502208,
|
| 383 |
+
"train_actor_loss": -0.16781520722217338,
|
| 384 |
+
"train_critic_loss": 0.00017849586477681565,
|
| 385 |
+
"train_mean_reward": -4.841276411409797e-05,
|
| 386 |
+
"train_entropy": 0.4778901710336214,
|
| 387 |
+
"train_ce_loss": 0.57612441854614,
|
| 388 |
+
"train_dice_loss": 0.6509548890036206,
|
| 389 |
+
"train_dice": 0.7349027215696128,
|
| 390 |
+
"train_iou": 0.6367749625926888,
|
| 391 |
+
"grad_norm": 28.045268920964972,
|
| 392 |
+
"lr": 0.0018451062858745053,
|
| 393 |
+
"encoder_lr": 0.00018521115292269145,
|
| 394 |
+
"alpha": 0.35795825719833374,
|
| 395 |
+
"validated_this_epoch": false,
|
| 396 |
+
"val_loss": null,
|
| 397 |
+
"val_dice": null,
|
| 398 |
+
"val_iou": null,
|
| 399 |
+
"val_reward": null,
|
| 400 |
+
"val_entropy": null
|
| 401 |
+
},
|
| 402 |
+
{
|
| 403 |
+
"epoch": 19,
|
| 404 |
+
"train_loss": 0.42738864446679753,
|
| 405 |
+
"train_actor_loss": -0.15967348723814312,
|
| 406 |
+
"train_critic_loss": 0.00014469611048676615,
|
| 407 |
+
"train_mean_reward": 7.825111677716063e-05,
|
| 408 |
+
"train_entropy": 0.4575588467701769,
|
| 409 |
+
"train_ce_loss": 0.5552425827681826,
|
| 410 |
+
"train_dice_loss": 0.6187369698701903,
|
| 411 |
+
"train_dice": 0.7541415382171733,
|
| 412 |
+
"train_iou": 0.6563228469580749,
|
| 413 |
+
"grad_norm": 20.540933743003727,
|
| 414 |
+
"lr": 0.0018279451714031893,
|
| 415 |
+
"encoder_lr": 0.0001835726545560834,
|
| 416 |
+
"alpha": 0.35320475697517395,
|
| 417 |
+
"validated_this_epoch": false,
|
| 418 |
+
"val_loss": null,
|
| 419 |
+
"val_dice": null,
|
| 420 |
+
"val_iou": null,
|
| 421 |
+
"val_reward": null,
|
| 422 |
+
"val_entropy": null
|
| 423 |
+
},
|
| 424 |
+
{
|
| 425 |
+
"epoch": 20,
|
| 426 |
+
"train_loss": 0.41952777620000664,
|
| 427 |
+
"train_actor_loss": -0.15635209333257463,
|
| 428 |
+
"train_critic_loss": 0.000151969482174851,
|
| 429 |
+
"train_mean_reward": 0.00022175002820547544,
|
| 430 |
+
"train_entropy": 0.4549665328991197,
|
| 431 |
+
"train_ce_loss": 0.5219484158666462,
|
| 432 |
+
"train_dice_loss": 0.6296593416722863,
|
| 433 |
+
"train_dice": 0.7544853653236199,
|
| 434 |
+
"train_iou": 0.6569179955151743,
|
| 435 |
+
"grad_norm": 22.9208480359972,
|
| 436 |
+
"lr": 0.001809971909403073,
|
| 437 |
+
"encoder_lr": 0.00018185661446562003,
|
| 438 |
+
"alpha": 0.3484734892845154,
|
| 439 |
+
"validated_this_epoch": true,
|
| 440 |
+
"val_loss": 0.7632999594471502,
|
| 441 |
+
"val_dice": 0.6670111382993077,
|
| 442 |
+
"val_iou": 0.5687223110334083,
|
| 443 |
+
"val_reward": 0.0007674738781868729,
|
| 444 |
+
"val_entropy": 0.5539201008852083,
|
| 445 |
+
"val_actor_loss": -0.0018786448678719068,
|
| 446 |
+
"val_critic_loss": 0.00013182554930654536,
|
| 447 |
+
"val_ce_loss": 0.7151426424747427,
|
| 448 |
+
"val_dice_loss": 0.815082708693514
|
| 449 |
+
},
|
| 450 |
+
{
|
| 451 |
+
"epoch": 21,
|
| 452 |
+
"train_loss": 0.3951729669329501,
|
| 453 |
+
"train_actor_loss": -0.14558063055934908,
|
| 454 |
+
"train_critic_loss": 0.00014876328096110445,
|
| 455 |
+
"train_mean_reward": 0.00019544940333365518,
|
| 456 |
+
"train_entropy": 0.43898867317102364,
|
| 457 |
+
"train_ce_loss": 0.5049832546012596,
|
| 458 |
+
"train_dice_loss": 0.5763751636487876,
|
| 459 |
+
"train_dice": 0.7737505900778692,
|
| 460 |
+
"train_iou": 0.6778911324980018,
|
| 461 |
+
"grad_norm": 26.839744311432504,
|
| 462 |
+
"lr": 0.001791204237313812,
|
| 463 |
+
"encoder_lr": 0.0001800647261756906,
|
| 464 |
+
"alpha": 0.34424030780792236,
|
| 465 |
+
"validated_this_epoch": false,
|
| 466 |
+
"val_loss": null,
|
| 467 |
+
"val_dice": null,
|
| 468 |
+
"val_iou": null,
|
| 469 |
+
"val_reward": null,
|
| 470 |
+
"val_entropy": null
|
| 471 |
+
},
|
| 472 |
+
{
|
| 473 |
+
"epoch": 22,
|
| 474 |
+
"train_loss": 0.4139959968630109,
|
| 475 |
+
"train_actor_loss": -0.14985723126419756,
|
| 476 |
+
"train_critic_loss": 0.00018425816343840625,
|
| 477 |
+
"train_mean_reward": 0.0001969124619656862,
|
| 478 |
+
"train_entropy": 0.4623227658890939,
|
| 479 |
+
"train_ce_loss": 0.5221870039434483,
|
| 480 |
+
"train_dice_loss": 0.6053351824597795,
|
| 481 |
+
"train_dice": 0.7645198572524651,
|
| 482 |
+
"train_iou": 0.6692009470198282,
|
| 483 |
+
"grad_norm": 30.721105263214703,
|
| 484 |
+
"lr": 0.0017716606765619107,
|
| 485 |
+
"encoder_lr": 0.0001781987580637,
|
| 486 |
+
"alpha": 0.33899158239364624,
|
| 487 |
+
"validated_this_epoch": false,
|
| 488 |
+
"val_loss": null,
|
| 489 |
+
"val_dice": null,
|
| 490 |
+
"val_iou": null,
|
| 491 |
+
"val_reward": null,
|
| 492 |
+
"val_entropy": null
|
| 493 |
+
},
|
| 494 |
+
{
|
| 495 |
+
"epoch": 23,
|
| 496 |
+
"train_loss": 0.3932833785963671,
|
| 497 |
+
"train_actor_loss": -0.1461067467215603,
|
| 498 |
+
"train_critic_loss": 0.00016151701852346232,
|
| 499 |
+
"train_mean_reward": 0.00029588552263513245,
|
| 500 |
+
"train_entropy": 0.4528499181341447,
|
| 501 |
+
"train_ce_loss": 0.49375711111002896,
|
| 502 |
+
"train_dice_loss": 0.5848616103639283,
|
| 503 |
+
"train_dice": 0.7714194063221028,
|
| 504 |
+
"train_iou": 0.6737402006243907,
|
| 505 |
+
"grad_norm": 25.235608872979185,
|
| 506 |
+
"lr": 0.0017513605142823077,
|
| 507 |
+
"encoder_lr": 0.00017626055161489368,
|
| 508 |
+
"alpha": 0.333966463804245,
|
| 509 |
+
"validated_this_epoch": false,
|
| 510 |
+
"val_loss": null,
|
| 511 |
+
"val_dice": null,
|
| 512 |
+
"val_iou": null,
|
| 513 |
+
"val_reward": null,
|
| 514 |
+
"val_entropy": null
|
| 515 |
+
},
|
| 516 |
+
{
|
| 517 |
+
"epoch": 24,
|
| 518 |
+
"train_loss": 0.3842491174925321,
|
| 519 |
+
"train_actor_loss": -0.14339854943178576,
|
| 520 |
+
"train_critic_loss": 0.00017066113369724984,
|
| 521 |
+
"train_mean_reward": 0.0001806230724056615,
|
| 522 |
+
"train_entropy": 0.45681545538803703,
|
| 523 |
+
"train_ce_loss": 0.5042811185855525,
|
| 524 |
+
"train_dice_loss": 0.5508435399341336,
|
| 525 |
+
"train_dice": 0.7780393391177455,
|
| 526 |
+
"train_iou": 0.6805682424159337,
|
| 527 |
+
"grad_norm": 36.116379959176676,
|
| 528 |
+
"lr": 0.001730323784284305,
|
| 529 |
+
"encoder_lr": 0.0001742520196050341,
|
| 530 |
+
"alpha": 0.3288189470767975,
|
| 531 |
+
"validated_this_epoch": false,
|
| 532 |
+
"val_loss": null,
|
| 533 |
+
"val_dice": null,
|
| 534 |
+
"val_iou": null,
|
| 535 |
+
"val_reward": null,
|
| 536 |
+
"val_entropy": null
|
| 537 |
+
},
|
| 538 |
+
{
|
| 539 |
+
"epoch": 25,
|
| 540 |
+
"train_loss": 0.4041452443032244,
|
| 541 |
+
"train_actor_loss": -0.1349047557655385,
|
| 542 |
+
"train_critic_loss": 0.00018131303160362575,
|
| 543 |
+
"train_mean_reward": 5.5701572374917205e-05,
|
| 544 |
+
"train_entropy": 0.43942794635006316,
|
| 545 |
+
"train_ce_loss": 0.5041238257218663,
|
| 546 |
+
"train_dice_loss": 0.5737948489281558,
|
| 547 |
+
"train_dice": 0.7756987697658895,
|
| 548 |
+
"train_iou": 0.681066380021839,
|
| 549 |
+
"grad_norm": 23.01321125492569,
|
| 550 |
+
"lr": 0.0017085712472806151,
|
| 551 |
+
"encoder_lr": 0.000172175144212722,
|
| 552 |
+
"alpha": 0.324092835187912,
|
| 553 |
+
"validated_this_epoch": false,
|
| 554 |
+
"val_loss": null,
|
| 555 |
+
"val_dice": null,
|
| 556 |
+
"val_iou": null,
|
| 557 |
+
"val_reward": null,
|
| 558 |
+
"val_entropy": null
|
| 559 |
+
},
|
| 560 |
+
{
|
| 561 |
+
"epoch": 26,
|
| 562 |
+
"train_loss": 0.3794952503679854,
|
| 563 |
+
"train_actor_loss": -0.13500117688158222,
|
| 564 |
+
"train_critic_loss": 0.00014619813189884747,
|
| 565 |
+
"train_mean_reward": 0.00025165137418011683,
|
| 566 |
+
"train_entropy": 0.44366789536035645,
|
| 567 |
+
"train_ce_loss": 0.5046593210448147,
|
| 568 |
+
"train_dice_loss": 0.5241873225808452,
|
| 569 |
+
"train_dice": 0.7934330031867547,
|
| 570 |
+
"train_iou": 0.6985961932693258,
|
| 571 |
+
"grad_norm": 23.323872733485793,
|
| 572 |
+
"lr": 0.0016861243703990456,
|
| 573 |
+
"encoder_lr": 0.00017003197506322543,
|
| 574 |
+
"alpha": 0.3192978799343109,
|
| 575 |
+
"validated_this_epoch": false,
|
| 576 |
+
"val_loss": null,
|
| 577 |
+
"val_dice": null,
|
| 578 |
+
"val_iou": null,
|
| 579 |
+
"val_reward": null,
|
| 580 |
+
"val_entropy": null
|
| 581 |
+
},
|
| 582 |
+
{
|
| 583 |
+
"epoch": 27,
|
| 584 |
+
"train_loss": 0.3944645080924785,
|
| 585 |
+
"train_actor_loss": -0.1284808012469159,
|
| 586 |
+
"train_critic_loss": 0.00014440610243406083,
|
| 587 |
+
"train_mean_reward": 0.00017376433576407605,
|
| 588 |
+
"train_entropy": 0.42538873756963763,
|
| 589 |
+
"train_ce_loss": 0.4863128698225198,
|
| 590 |
+
"train_dice_loss": 0.5594333283198897,
|
| 591 |
+
"train_dice": 0.7875091953717596,
|
| 592 |
+
"train_iou": 0.6912481695438978,
|
| 593 |
+
"grad_norm": 34.744155206421546,
|
| 594 |
+
"lr": 0.001663005305997034,
|
| 595 |
+
"encoder_lr": 0.00016782462720574693,
|
| 596 |
+
"alpha": 0.31524407863616943,
|
| 597 |
+
"validated_this_epoch": false,
|
| 598 |
+
"val_loss": null,
|
| 599 |
+
"val_dice": null,
|
| 600 |
+
"val_iou": null,
|
| 601 |
+
"val_reward": null,
|
| 602 |
+
"val_entropy": null
|
| 603 |
+
},
|
| 604 |
+
{
|
| 605 |
+
"epoch": 28,
|
| 606 |
+
"train_loss": 0.34072865974230243,
|
| 607 |
+
"train_actor_loss": -0.12900038142941256,
|
| 608 |
+
"train_critic_loss": 0.00015297339519132036,
|
| 609 |
+
"train_mean_reward": 0.0004118765794230401,
|
| 610 |
+
"train_entropy": 0.4310346835073882,
|
| 611 |
+
"train_ce_loss": 0.45804114250023426,
|
| 612 |
+
"train_dice_loss": 0.4812639548180947,
|
| 613 |
+
"train_dice": 0.8069415401189542,
|
| 614 |
+
"train_iou": 0.7108221350525186,
|
| 615 |
+
"grad_norm": 24.49839576917101,
|
| 616 |
+
"lr": 0.0016392368697999468,
|
| 617 |
+
"encoder_lr": 0.00016555527902612554,
|
| 618 |
+
"alpha": 0.3107281029224396,
|
| 619 |
+
"validated_this_epoch": false,
|
| 620 |
+
"val_loss": null,
|
| 621 |
+
"val_dice": null,
|
| 622 |
+
"val_iou": null,
|
| 623 |
+
"val_reward": null,
|
| 624 |
+
"val_entropy": null
|
| 625 |
+
},
|
| 626 |
+
{
|
| 627 |
+
"epoch": 29,
|
| 628 |
+
"train_loss": 0.3766678067776122,
|
| 629 |
+
"train_actor_loss": -0.12661007960695167,
|
| 630 |
+
"train_critic_loss": 0.00015667974467743913,
|
| 631 |
+
"train_mean_reward": 0.00014682896077056168,
|
| 632 |
+
"train_entropy": 0.4281593105089141,
|
| 633 |
+
"train_ce_loss": 0.4717548740823159,
|
| 634 |
+
"train_dice_loss": 0.5346442060569153,
|
| 635 |
+
"train_dice": 0.7823287820779664,
|
| 636 |
+
"train_iou": 0.6881705477974973,
|
| 637 |
+
"grad_norm": 17.027618325495904,
|
| 638 |
+
"lr": 0.0016148425183847121,
|
| 639 |
+
"encoder_lr": 0.0001632261700970328,
|
| 640 |
+
"alpha": 0.30637499690055847,
|
| 641 |
+
"validated_this_epoch": false,
|
| 642 |
+
"val_loss": null,
|
| 643 |
+
"val_dice": null,
|
| 644 |
+
"val_iou": null,
|
| 645 |
+
"val_reward": null,
|
| 646 |
+
"val_entropy": null
|
| 647 |
+
},
|
| 648 |
+
{
|
| 649 |
+
"epoch": 30,
|
| 650 |
+
"train_loss": 0.38399921832317413,
|
| 651 |
+
"train_actor_loss": -0.11066250028016493,
|
| 652 |
+
"train_critic_loss": 0.00014921942248393965,
|
| 653 |
+
"train_mean_reward": 0.00021920748802818537,
|
| 654 |
+
"train_entropy": 0.38423825014449986,
|
| 655 |
+
"train_ce_loss": 0.46995057051914724,
|
| 656 |
+
"train_dice_loss": 0.5192236348178035,
|
| 657 |
+
"train_dice": 0.7996051701878744,
|
| 658 |
+
"train_iou": 0.7088806363462612,
|
| 659 |
+
"grad_norm": 17.94604222950085,
|
| 660 |
+
"lr": 0.0015898463260310114,
|
| 661 |
+
"encoder_lr": 0.00016083959896778498,
|
| 662 |
+
"alpha": 0.303342342376709,
|
| 663 |
+
"validated_this_epoch": true,
|
| 664 |
+
"val_loss": 0.6735828337506532,
|
| 665 |
+
"val_dice": 0.7149527117659106,
|
| 666 |
+
"val_iou": 0.6133315874097852,
|
| 667 |
+
"val_reward": 4.760155544970237e-05,
|
| 668 |
+
"val_entropy": 0.34986774638445695,
|
| 669 |
+
"val_actor_loss": 0.0005813433835459309,
|
| 670 |
+
"val_critic_loss": 0.0002916965805465537,
|
| 671 |
+
"val_ce_loss": 0.5930595405874868,
|
| 672 |
+
"val_dice_loss": 0.7526517182287543
|
| 673 |
+
},
|
| 674 |
+
{
|
| 675 |
+
"epoch": 31,
|
| 676 |
+
"train_loss": 0.34176113184671414,
|
| 677 |
+
"train_actor_loss": -0.11918333625704897,
|
| 678 |
+
"train_critic_loss": 0.00015182029256105125,
|
| 679 |
+
"train_mean_reward": 0.00024654864057920255,
|
| 680 |
+
"train_entropy": 0.42109907836261035,
|
| 681 |
+
"train_ce_loss": 0.4542518993629051,
|
| 682 |
+
"train_dice_loss": 0.4674852043775317,
|
| 683 |
+
"train_dice": 0.8054514507298842,
|
| 684 |
+
"train_iou": 0.7113861560244031,
|
| 685 |
+
"grad_norm": 21.193717238976973,
|
| 686 |
+
"lr": 0.0015642729609628705,
|
| 687 |
+
"encoder_lr": 0.00015839792089595244,
|
| 688 |
+
"alpha": 0.2998906373977661,
|
| 689 |
+
"validated_this_epoch": false,
|
| 690 |
+
"val_loss": null,
|
| 691 |
+
"val_dice": null,
|
| 692 |
+
"val_iou": null,
|
| 693 |
+
"val_reward": null,
|
| 694 |
+
"val_entropy": null
|
| 695 |
+
},
|
| 696 |
+
{
|
| 697 |
+
"epoch": 32,
|
| 698 |
+
"train_loss": 0.35512672610176627,
|
| 699 |
+
"train_actor_loss": -0.11901210213335138,
|
| 700 |
+
"train_critic_loss": 0.00015812984128028485,
|
| 701 |
+
"train_mean_reward": 0.0002931835111051439,
|
| 702 |
+
"train_entropy": 0.42644300873966484,
|
| 703 |
+
"train_ce_loss": 0.4459090589705701,
|
| 704 |
+
"train_dice_loss": 0.5022104562715043,
|
| 705 |
+
"train_dice": 0.8031166600330998,
|
| 706 |
+
"train_iou": 0.711644148181936,
|
| 707 |
+
"grad_norm": 23.39685513067615,
|
| 708 |
+
"lr": 0.001538147661004102,
|
| 709 |
+
"encoder_lr": 0.0001559035455230047,
|
| 710 |
+
"alpha": 0.29535916447639465,
|
| 711 |
+
"validated_this_epoch": false,
|
| 712 |
+
"val_loss": null,
|
| 713 |
+
"val_dice": null,
|
| 714 |
+
"val_iou": null,
|
| 715 |
+
"val_reward": null,
|
| 716 |
+
"val_entropy": null
|
| 717 |
+
},
|
| 718 |
+
{
|
| 719 |
+
"epoch": 33,
|
| 720 |
+
"train_loss": 0.35658658082171,
|
| 721 |
+
"train_actor_loss": -0.10893077447827548,
|
| 722 |
+
"train_critic_loss": 0.000143021270350824,
|
| 723 |
+
"train_mean_reward": 0.0003144957792076848,
|
| 724 |
+
"train_entropy": 0.39473499606053036,
|
| 725 |
+
"train_ce_loss": 0.43206190828051066,
|
| 726 |
+
"train_dice_loss": 0.4988297703808285,
|
| 727 |
+
"train_dice": 0.8034801231925097,
|
| 728 |
+
"train_iou": 0.7112604857635868,
|
| 729 |
+
"grad_norm": 17.036126112753106,
|
| 730 |
+
"lr": 0.0015114962086716199,
|
| 731 |
+
"encoder_lr": 0.0001533589344962853,
|
| 732 |
+
"alpha": 0.29221874475479126,
|
| 733 |
+
"validated_this_epoch": false,
|
| 734 |
+
"val_loss": null,
|
| 735 |
+
"val_dice": null,
|
| 736 |
+
"val_iou": null,
|
| 737 |
+
"val_reward": null,
|
| 738 |
+
"val_entropy": null
|
| 739 |
+
},
|
| 740 |
+
{
|
| 741 |
+
"epoch": 34,
|
| 742 |
+
"train_loss": 0.3728100263167086,
|
| 743 |
+
"train_actor_loss": -0.11287370920951362,
|
| 744 |
+
"train_critic_loss": 0.00014060894699050597,
|
| 745 |
+
"train_mean_reward": 0.00022903858154162758,
|
| 746 |
+
"train_entropy": 0.41475872949574344,
|
| 747 |
+
"train_ce_loss": 0.4519465886951648,
|
| 748 |
+
"train_dice_loss": 0.5192802607705119,
|
| 749 |
+
"train_dice": 0.7988190251733363,
|
| 750 |
+
"train_iou": 0.7068648738893577,
|
| 751 |
+
"grad_norm": 18.037854449693548,
|
| 752 |
+
"lr": 0.0014843449057312071,
|
| 753 |
+
"encoder_lr": 0.00015076659903966297,
|
| 754 |
+
"alpha": 0.28821128606796265,
|
| 755 |
+
"validated_this_epoch": false,
|
| 756 |
+
"val_loss": null,
|
| 757 |
+
"val_dice": null,
|
| 758 |
+
"val_iou": null,
|
| 759 |
+
"val_reward": null,
|
| 760 |
+
"val_entropy": null
|
| 761 |
+
},
|
| 762 |
+
{
|
| 763 |
+
"epoch": 35,
|
| 764 |
+
"train_loss": 0.374327719720557,
|
| 765 |
+
"train_actor_loss": -0.11043671389574844,
|
| 766 |
+
"train_critic_loss": 0.00014391915895310573,
|
| 767 |
+
"train_mean_reward": 0.00026625005156387426,
|
| 768 |
+
"train_entropy": 0.41190868223360344,
|
| 769 |
+
"train_ce_loss": 0.45913466406283415,
|
| 770 |
+
"train_dice_loss": 0.5102502692885484,
|
| 771 |
+
"train_dice": 0.7998975479934979,
|
| 772 |
+
"train_iou": 0.7091685576930477,
|
| 773 |
+
"grad_norm": 15.087900421878164,
|
| 774 |
+
"lr": 0.0014567205472408495,
|
| 775 |
+
"encoder_lr": 0.00014812909747525697,
|
| 776 |
+
"alpha": 0.28430619835853577,
|
| 777 |
+
"validated_this_epoch": false,
|
| 778 |
+
"val_loss": null,
|
| 779 |
+
"val_dice": null,
|
| 780 |
+
"val_iou": null,
|
| 781 |
+
"val_reward": null,
|
| 782 |
+
"val_entropy": null
|
| 783 |
+
},
|
| 784 |
+
{
|
| 785 |
+
"epoch": 36,
|
| 786 |
+
"train_loss": 0.3543958063273467,
|
| 787 |
+
"train_actor_loss": -0.10641251463991727,
|
| 788 |
+
"train_critic_loss": 0.00013948182224036523,
|
| 789 |
+
"train_mean_reward": 0.0002990052339113289,
|
| 790 |
+
"train_entropy": 0.40683788030338536,
|
| 791 |
+
"train_ce_loss": 0.43348631018397304,
|
| 792 |
+
"train_dice_loss": 0.48799083907474844,
|
| 793 |
+
"train_dice": 0.8152444728399345,
|
| 794 |
+
"train_iou": 0.7223868650931201,
|
| 795 |
+
"grad_norm": 16.074378915535387,
|
| 796 |
+
"lr": 0.0014286503951072478,
|
| 797 |
+
"encoder_lr": 0.00014544903269868192,
|
| 798 |
+
"alpha": 0.2805497944355011,
|
| 799 |
+
"validated_this_epoch": false,
|
| 800 |
+
"val_loss": null,
|
| 801 |
+
"val_dice": null,
|
| 802 |
+
"val_iou": null,
|
| 803 |
+
"val_reward": null,
|
| 804 |
+
"val_entropy": null
|
| 805 |
+
},
|
| 806 |
+
{
|
| 807 |
+
"epoch": 37,
|
| 808 |
+
"train_loss": 0.3579336386859821,
|
| 809 |
+
"train_actor_loss": -0.1035312391716496,
|
| 810 |
+
"train_critic_loss": 0.00014657508000011656,
|
| 811 |
+
"train_mean_reward": 0.00020749851775105153,
|
| 812 |
+
"train_entropy": 0.3968586622706063,
|
| 813 |
+
"train_ce_loss": 0.4337360269260095,
|
| 814 |
+
"train_dice_loss": 0.48904714188526455,
|
| 815 |
+
"train_dice": 0.8081231601318744,
|
| 816 |
+
"train_iou": 0.7177499719522671,
|
| 817 |
+
"grad_norm": 16.76060508203137,
|
| 818 |
+
"lr": 0.001400162151181607,
|
| 819 |
+
"encoder_lr": 0.00014272904961030417,
|
| 820 |
+
"alpha": 0.2772769331932068,
|
| 821 |
+
"validated_this_epoch": false,
|
| 822 |
+
"val_loss": null,
|
| 823 |
+
"val_dice": null,
|
| 824 |
+
"val_iou": null,
|
| 825 |
+
"val_reward": null,
|
| 826 |
+
"val_entropy": null
|
| 827 |
+
},
|
| 828 |
+
{
|
| 829 |
+
"epoch": 38,
|
| 830 |
+
"train_loss": 0.370790235846224,
|
| 831 |
+
"train_actor_loss": -0.10259911197524045,
|
| 832 |
+
"train_critic_loss": 0.00014528584258894424,
|
| 833 |
+
"train_mean_reward": 0.00020285158417592528,
|
| 834 |
+
"train_entropy": 0.40216287974393333,
|
| 835 |
+
"train_ce_loss": 0.4506687542305687,
|
| 836 |
+
"train_dice_loss": 0.4959646432590731,
|
| 837 |
+
"train_dice": 0.8068299546182535,
|
| 838 |
+
"train_iou": 0.719968302141822,
|
| 839 |
+
"grad_norm": 16.453235806882844,
|
| 840 |
+
"lr": 0.0013712839299212553,
|
| 841 |
+
"encoder_lr": 0.00013997183250504444,
|
| 842 |
+
"alpha": 0.2737799286842346,
|
| 843 |
+
"validated_this_epoch": false,
|
| 844 |
+
"val_loss": null,
|
| 845 |
+
"val_dice": null,
|
| 846 |
+
"val_iou": null,
|
| 847 |
+
"val_reward": null,
|
| 848 |
+
"val_entropy": null
|
| 849 |
+
},
|
| 850 |
+
{
|
| 851 |
+
"epoch": 39,
|
| 852 |
+
"train_loss": 0.3511020791890143,
|
| 853 |
+
"train_actor_loss": -0.10045221514123895,
|
| 854 |
+
"train_critic_loss": 0.00014023397357144275,
|
| 855 |
+
"train_mean_reward": 0.00016800072767100176,
|
| 856 |
+
"train_entropy": 0.4002091989152191,
|
| 857 |
+
"train_ce_loss": 0.4260108783174527,
|
| 858 |
+
"train_dice_loss": 0.47695746478800316,
|
| 859 |
+
"train_dice": 0.8214849619576512,
|
| 860 |
+
"train_iou": 0.7309336838080431,
|
| 861 |
+
"grad_norm": 20.690970424995864,
|
| 862 |
+
"lr": 0.0013420442306440656,
|
| 863 |
+
"encoder_lr": 0.0001371801024233027,
|
| 864 |
+
"alpha": 0.2704960107803345,
|
| 865 |
+
"validated_this_epoch": false,
|
| 866 |
+
"val_loss": null,
|
| 867 |
+
"val_dice": null,
|
| 868 |
+
"val_iou": null,
|
| 869 |
+
"val_reward": null,
|
| 870 |
+
"val_entropy": null
|
| 871 |
+
},
|
| 872 |
+
{
|
| 873 |
+
"epoch": 40,
|
| 874 |
+
"train_loss": 0.34654889714203835,
|
| 875 |
+
"train_actor_loss": -0.09787431744574654,
|
| 876 |
+
"train_critic_loss": 0.0001326317844256032,
|
| 877 |
+
"train_mean_reward": 0.0001993380423152493,
|
| 878 |
+
"train_entropy": 0.38943512745451864,
|
| 879 |
+
"train_ce_loss": 0.41918712376932143,
|
| 880 |
+
"train_dice_loss": 0.4695266614588656,
|
| 881 |
+
"train_dice": 0.8206628674265705,
|
| 882 |
+
"train_iou": 0.732380223630864,
|
| 883 |
+
"grad_norm": 13.000041358692702,
|
| 884 |
+
"lr": 0.0013124719094030732,
|
| 885 |
+
"encoder_lr": 0.00013435661446562,
|
| 886 |
+
"alpha": 0.2673097550868988,
|
| 887 |
+
"validated_this_epoch": true,
|
| 888 |
+
"val_loss": 0.6415375441976005,
|
| 889 |
+
"val_dice": 0.701783773693781,
|
| 890 |
+
"val_iou": 0.6123823421974542,
|
| 891 |
+
"val_reward": -6.908963089267901e-07,
|
| 892 |
+
"val_entropy": 0.3444194959269629,
|
| 893 |
+
"val_actor_loss": -0.008142723066019682,
|
| 894 |
+
"val_critic_loss": 0.00018916144988153716,
|
| 895 |
+
"val_ce_loss": 0.5305643291682983,
|
| 896 |
+
"val_dice_loss": 0.7686070245925826
|
| 897 |
+
},
|
| 898 |
+
{
|
| 899 |
+
"epoch": 41,
|
| 900 |
+
"train_loss": 0.3486620614362895,
|
| 901 |
+
"train_actor_loss": -0.09702625348971487,
|
| 902 |
+
"train_critic_loss": 0.00014040683384769593,
|
| 903 |
+
"train_mean_reward": 0.00023582053645899466,
|
| 904 |
+
"train_entropy": 0.38906036088247936,
|
| 905 |
+
"train_ce_loss": 0.41063696843660996,
|
| 906 |
+
"train_dice_loss": 0.4805992416633192,
|
| 907 |
+
"train_dice": 0.8152734130862598,
|
| 908 |
+
"train_iou": 0.7250225351014964,
|
| 909 |
+
"grad_norm": 12.043104198089866,
|
| 910 |
+
"lr": 0.0012825961505090336,
|
| 911 |
+
"encoder_lr": 0.00013150415507372682,
|
| 912 |
+
"alpha": 0.2646113336086273,
|
| 913 |
+
"validated_this_epoch": false,
|
| 914 |
+
"val_loss": null,
|
| 915 |
+
"val_dice": null,
|
| 916 |
+
"val_iou": null,
|
| 917 |
+
"val_reward": null,
|
| 918 |
+
"val_entropy": null
|
| 919 |
+
},
|
| 920 |
+
{
|
| 921 |
+
"epoch": 42,
|
| 922 |
+
"train_loss": 0.3888709625416948,
|
| 923 |
+
"train_actor_loss": -0.09222593834715143,
|
| 924 |
+
"train_critic_loss": 0.00014433348661213453,
|
| 925 |
+
"train_mean_reward": 0.0003279169764427276,
|
| 926 |
+
"train_entropy": 0.38326878614665927,
|
| 927 |
+
"train_ce_loss": 0.43157595801058896,
|
| 928 |
+
"train_dice_loss": 0.5304734947145447,
|
| 929 |
+
"train_dice": 0.8025415084819764,
|
| 930 |
+
"train_iou": 0.7104108407105724,
|
| 931 |
+
"grad_norm": 25.159271628819695,
|
| 932 |
+
"lr": 0.0012524464377290307,
|
| 933 |
+
"encoder_lr": 0.00012862553928066122,
|
| 934 |
+
"alpha": 0.2615024149417877,
|
| 935 |
+
"validated_this_epoch": false,
|
| 936 |
+
"val_loss": null,
|
| 937 |
+
"val_dice": null,
|
| 938 |
+
"val_iou": null,
|
| 939 |
+
"val_reward": null,
|
| 940 |
+
"val_entropy": null
|
| 941 |
+
},
|
| 942 |
+
{
|
| 943 |
+
"epoch": 43,
|
| 944 |
+
"train_loss": 0.36074973677479943,
|
| 945 |
+
"train_actor_loss": -0.07968840547247719,
|
| 946 |
+
"train_critic_loss": 0.00013460830302406524,
|
| 947 |
+
"train_mean_reward": 0.00037672200938776033,
|
| 948 |
+
"train_entropy": 0.3378434822141393,
|
| 949 |
+
"train_ce_loss": 0.40141497130532205,
|
| 950 |
+
"train_dice_loss": 0.47932669181540344,
|
| 951 |
+
"train_dice": 0.8201022134554604,
|
| 952 |
+
"train_iou": 0.7338341316503348,
|
| 953 |
+
"grad_norm": 20.336131717807564,
|
| 954 |
+
"lr": 0.0012220525251895604,
|
| 955 |
+
"encoder_lr": 0.0001257236079326716,
|
| 956 |
+
"alpha": 0.2612733244895935,
|
| 957 |
+
"validated_this_epoch": false,
|
| 958 |
+
"val_loss": null,
|
| 959 |
+
"val_dice": null,
|
| 960 |
+
"val_iou": null,
|
| 961 |
+
"val_reward": null,
|
| 962 |
+
"val_entropy": null
|
| 963 |
+
},
|
| 964 |
+
{
|
| 965 |
+
"epoch": 44,
|
| 966 |
+
"train_loss": 0.3683151816696564,
|
| 967 |
+
"train_actor_loss": -0.09121021863015354,
|
| 968 |
+
"train_critic_loss": 0.0001508469661222533,
|
| 969 |
+
"train_mean_reward": 0.00028739183889974005,
|
| 970 |
+
"train_entropy": 0.3833474852240979,
|
| 971 |
+
"train_ce_loss": 0.42116220554129946,
|
| 972 |
+
"train_dice_loss": 0.49773773520183817,
|
| 973 |
+
"train_dice": 0.8155273205288429,
|
| 974 |
+
"train_iou": 0.7266095360782527,
|
| 975 |
+
"grad_norm": 32.703362630319226,
|
| 976 |
+
"lr": 0.0011914444080127966,
|
| 977 |
+
"encoder_lr": 0.0001228012248856439,
|
| 978 |
+
"alpha": 0.25858134031295776,
|
| 979 |
+
"validated_this_epoch": false,
|
| 980 |
+
"val_loss": null,
|
| 981 |
+
"val_dice": null,
|
| 982 |
+
"val_iou": null,
|
| 983 |
+
"val_reward": null,
|
| 984 |
+
"val_entropy": null
|
| 985 |
+
},
|
| 986 |
+
{
|
| 987 |
+
"epoch": 45,
|
| 988 |
+
"train_loss": 0.32464760384890573,
|
| 989 |
+
"train_actor_loss": -0.08892081199257179,
|
| 990 |
+
"train_critic_loss": 0.00013528678060286908,
|
| 991 |
+
"train_mean_reward": 0.00034308635907800354,
|
| 992 |
+
"train_entropy": 0.3637862527112474,
|
| 993 |
+
"train_ce_loss": 0.3792107364539423,
|
| 994 |
+
"train_dice_loss": 0.4477907965965665,
|
| 995 |
+
"train_dice": 0.8311413130010965,
|
| 996 |
+
"train_iou": 0.7441400931219537,
|
| 997 |
+
"grad_norm": 10.287710869035056,
|
| 998 |
+
"lr": 0.0011606522927150302,
|
| 999 |
+
"encoder_lr": 0.00011986127417882199,
|
| 1000 |
+
"alpha": 0.2565890848636627,
|
| 1001 |
+
"validated_this_epoch": false,
|
| 1002 |
+
"val_loss": null,
|
| 1003 |
+
"val_dice": null,
|
| 1004 |
+
"val_iou": null,
|
| 1005 |
+
"val_reward": null,
|
| 1006 |
+
"val_entropy": null
|
| 1007 |
+
},
|
| 1008 |
+
{
|
| 1009 |
+
"epoch": 46,
|
| 1010 |
+
"train_loss": 0.3317328339745832,
|
| 1011 |
+
"train_actor_loss": -0.09035823478794222,
|
| 1012 |
+
"train_critic_loss": 0.0001378832271680145,
|
| 1013 |
+
"train_mean_reward": 0.00027988374535837325,
|
| 1014 |
+
"train_entropy": 0.3775562456488918,
|
| 1015 |
+
"train_ce_loss": 0.3926363725393279,
|
| 1016 |
+
"train_dice_loss": 0.4514078695798721,
|
| 1017 |
+
"train_dice": 0.8211396246793582,
|
| 1018 |
+
"train_iou": 0.7330374619864602,
|
| 1019 |
+
"grad_norm": 13.453849950040034,
|
| 1020 |
+
"lr": 0.0011297065673964834,
|
| 1021 |
+
"encoder_lr": 0.00011690665718860897,
|
| 1022 |
+
"alpha": 0.2542921304702759,
|
| 1023 |
+
"validated_this_epoch": false,
|
| 1024 |
+
"val_loss": null,
|
| 1025 |
+
"val_dice": null,
|
| 1026 |
+
"val_iou": null,
|
| 1027 |
+
"val_reward": null,
|
| 1028 |
+
"val_entropy": null
|
| 1029 |
+
},
|
| 1030 |
+
{
|
| 1031 |
+
"epoch": 47,
|
| 1032 |
+
"train_loss": 0.3216659417917389,
|
| 1033 |
+
"train_actor_loss": -0.08770794207735579,
|
| 1034 |
+
"train_critic_loss": 0.00012730641285995642,
|
| 1035 |
+
"train_mean_reward": 0.00031240313191042084,
|
| 1036 |
+
"train_entropy": 0.37395655257911337,
|
| 1037 |
+
"train_ce_loss": 0.3880618838045881,
|
| 1038 |
+
"train_dice_loss": 0.4305585666682369,
|
| 1039 |
+
"train_dice": 0.8390136553062969,
|
| 1040 |
+
"train_iou": 0.7505303253861844,
|
| 1041 |
+
"grad_norm": 8.464987816736679,
|
| 1042 |
+
"lr": 0.001098637771751922,
|
| 1043 |
+
"encoder_lr": 0.0001139402897652589,
|
| 1044 |
+
"alpha": 0.25193166732788086,
|
| 1045 |
+
"validated_this_epoch": false,
|
| 1046 |
+
"val_loss": null,
|
| 1047 |
+
"val_dice": null,
|
| 1048 |
+
"val_iou": null,
|
| 1049 |
+
"val_reward": null,
|
| 1050 |
+
"val_entropy": null
|
| 1051 |
+
},
|
| 1052 |
+
{
|
| 1053 |
+
"epoch": 48,
|
| 1054 |
+
"train_loss": 0.3357040968600863,
|
| 1055 |
+
"train_actor_loss": -0.08461173549198212,
|
| 1056 |
+
"train_critic_loss": 0.00014108629719596292,
|
| 1057 |
+
"train_mean_reward": 0.0003504721940379091,
|
| 1058 |
+
"train_entropy": 0.3610355612131206,
|
| 1059 |
+
"train_ce_loss": 0.39261157803468466,
|
| 1060 |
+
"train_dice_loss": 0.44787898929236164,
|
| 1061 |
+
"train_dice": 0.8258875743734718,
|
| 1062 |
+
"train_iou": 0.7394617712256595,
|
| 1063 |
+
"grad_norm": 12.553831160530564,
|
| 1064 |
+
"lr": 0.0010674765669316673,
|
| 1065 |
+
"encoder_lr": 0.00011096509935528483,
|
| 1066 |
+
"alpha": 0.250160813331604,
|
| 1067 |
+
"validated_this_epoch": false,
|
| 1068 |
+
"val_loss": null,
|
| 1069 |
+
"val_dice": null,
|
| 1070 |
+
"val_iou": null,
|
| 1071 |
+
"val_reward": null,
|
| 1072 |
+
"val_entropy": null
|
| 1073 |
+
},
|
| 1074 |
+
{
|
| 1075 |
+
"epoch": 49,
|
| 1076 |
+
"train_loss": 0.334587997439238,
|
| 1077 |
+
"train_actor_loss": -0.08755494870765265,
|
| 1078 |
+
"train_critic_loss": 0.00013336657283304093,
|
| 1079 |
+
"train_mean_reward": 0.00022615180738797157,
|
| 1080 |
+
"train_entropy": 0.37572924376920214,
|
| 1081 |
+
"train_ce_loss": 0.4037856269394377,
|
| 1082 |
+
"train_dice_loss": 0.44036688884715397,
|
| 1083 |
+
"train_dice": 0.8276519184956054,
|
| 1084 |
+
"train_iou": 0.7429016542578117,
|
| 1085 |
+
"grad_norm": 10.797130710856859,
|
| 1086 |
+
"lr": 0.001036253705282738,
|
| 1087 |
+
"encoder_lr": 0.00010798402211242223,
|
| 1088 |
+
"alpha": 0.24771109223365784,
|
| 1089 |
+
"validated_this_epoch": false,
|
| 1090 |
+
"val_loss": null,
|
| 1091 |
+
"val_dice": null,
|
| 1092 |
+
"val_iou": null,
|
| 1093 |
+
"val_reward": null,
|
| 1094 |
+
"val_entropy": null
|
| 1095 |
+
},
|
| 1096 |
+
{
|
| 1097 |
+
"epoch": 50,
|
| 1098 |
+
"train_loss": 0.3369048665487027,
|
| 1099 |
+
"train_actor_loss": -0.08308573590365337,
|
| 1100 |
+
"train_critic_loss": 0.00012938498585655426,
|
| 1101 |
+
"train_mean_reward": 0.00035792575732505747,
|
| 1102 |
+
"train_entropy": 0.35968148754548657,
|
| 1103 |
+
"train_ce_loss": 0.3923364242187458,
|
| 1104 |
+
"train_dice_loss": 0.44751538124503404,
|
| 1105 |
+
"train_dice": 0.8327864464237467,
|
| 1106 |
+
"train_iou": 0.7477988788412686,
|
| 1107 |
+
"grad_norm": 9.523176794828371,
|
| 1108 |
+
"lr": 0.0010050000000000003,
|
| 1109 |
+
"encoder_lr": 0.00010500000000000005,
|
| 1110 |
+
"alpha": 0.24595437943935394,
|
| 1111 |
+
"validated_this_epoch": true,
|
| 1112 |
+
"val_loss": 0.6354440125552091,
|
| 1113 |
+
"val_dice": 0.710371547901053,
|
| 1114 |
+
"val_iou": 0.6166608473245011,
|
| 1115 |
+
"val_reward": -0.0001858324570764465,
|
| 1116 |
+
"val_entropy": 0.34401636209451797,
|
| 1117 |
+
"val_actor_loss": -0.005128610147177384,
|
| 1118 |
+
"val_critic_loss": 0.00018911030518422488,
|
| 1119 |
+
"val_ce_loss": 0.5728504054925658,
|
| 1120 |
+
"val_dice_loss": 0.7081057104197416
|
| 1121 |
+
},
|
| 1122 |
+
{
|
| 1123 |
+
"epoch": 51,
|
| 1124 |
+
"train_loss": 0.3307339201190386,
|
| 1125 |
+
"train_actor_loss": -0.08414840875843237,
|
| 1126 |
+
"train_critic_loss": 0.00013228695734001085,
|
| 1127 |
+
"train_mean_reward": 0.00036351559962709745,
|
| 1128 |
+
"train_entropy": 0.3655515008772066,
|
| 1129 |
+
"train_ce_loss": 0.3954008232094636,
|
| 1130 |
+
"train_dice_loss": 0.4342315382581657,
|
| 1131 |
+
"train_dice": 0.8360835794264525,
|
| 1132 |
+
"train_iou": 0.7496404728584023,
|
| 1133 |
+
"grad_norm": 8.562869288662608,
|
| 1134 |
+
"lr": 0.0009737462947172626,
|
| 1135 |
+
"encoder_lr": 0.00010201597788757786,
|
| 1136 |
+
"alpha": 0.24400851130485535,
|
| 1137 |
+
"validated_this_epoch": false,
|
| 1138 |
+
"val_loss": null,
|
| 1139 |
+
"val_dice": null,
|
| 1140 |
+
"val_iou": null,
|
| 1141 |
+
"val_reward": null,
|
| 1142 |
+
"val_entropy": null
|
| 1143 |
+
},
|
| 1144 |
+
{
|
| 1145 |
+
"epoch": 52,
|
| 1146 |
+
"train_loss": 0.31837643642527186,
|
| 1147 |
+
"train_actor_loss": -0.08241229838371586,
|
| 1148 |
+
"train_critic_loss": 0.00013319373558498212,
|
| 1149 |
+
"train_mean_reward": 0.0003757772751046328,
|
| 1150 |
+
"train_entropy": 0.3587823763143185,
|
| 1151 |
+
"train_ce_loss": 0.3790775892969692,
|
| 1152 |
+
"train_dice_loss": 0.42236667340115985,
|
| 1153 |
+
"train_dice": 0.8429392108298223,
|
| 1154 |
+
"train_iou": 0.7576524358847162,
|
| 1155 |
+
"grad_norm": 16.2514938388684,
|
| 1156 |
+
"lr": 0.0009425234330683334,
|
| 1157 |
+
"encoder_lr": 9.903490064471527e-05,
|
| 1158 |
+
"alpha": 0.24228397011756897,
|
| 1159 |
+
"validated_this_epoch": false,
|
| 1160 |
+
"val_loss": null,
|
| 1161 |
+
"val_dice": null,
|
| 1162 |
+
"val_iou": null,
|
| 1163 |
+
"val_reward": null,
|
| 1164 |
+
"val_entropy": null
|
| 1165 |
+
},
|
| 1166 |
+
{
|
| 1167 |
+
"epoch": 53,
|
| 1168 |
+
"train_loss": 0.32197661226788576,
|
| 1169 |
+
"train_actor_loss": -0.0824909445827283,
|
| 1170 |
+
"train_critic_loss": 0.00013975331806659243,
|
| 1171 |
+
"train_mean_reward": 0.00037474774681793235,
|
| 1172 |
+
"train_entropy": 0.36426120524797634,
|
| 1173 |
+
"train_ce_loss": 0.37147354612064376,
|
| 1174 |
+
"train_dice_loss": 0.43732180251939656,
|
| 1175 |
+
"train_dice": 0.8323800961294479,
|
| 1176 |
+
"train_iou": 0.7450570186533914,
|
| 1177 |
+
"grad_norm": 9.89003196612809,
|
| 1178 |
+
"lr": 0.0009113622282480784,
|
| 1179 |
+
"encoder_lr": 9.605971023474117e-05,
|
| 1180 |
+
"alpha": 0.2404448240995407,
|
| 1181 |
+
"validated_this_epoch": false,
|
| 1182 |
+
"val_loss": null,
|
| 1183 |
+
"val_dice": null,
|
| 1184 |
+
"val_iou": null,
|
| 1185 |
+
"val_reward": null,
|
| 1186 |
+
"val_entropy": null
|
| 1187 |
+
},
|
| 1188 |
+
{
|
| 1189 |
+
"epoch": 54,
|
| 1190 |
+
"train_loss": 0.3082753579211674,
|
| 1191 |
+
"train_actor_loss": -0.07823313300278836,
|
| 1192 |
+
"train_critic_loss": 0.0001282339567179225,
|
| 1193 |
+
"train_mean_reward": 0.0003064791481243371,
|
| 1194 |
+
"train_entropy": 0.35167766328737404,
|
| 1195 |
+
"train_ce_loss": 0.369310392315911,
|
| 1196 |
+
"train_dice_loss": 0.40357834439561036,
|
| 1197 |
+
"train_dice": 0.8422776023234118,
|
| 1198 |
+
"train_iou": 0.7545646257905606,
|
| 1199 |
+
"grad_norm": 8.769259092650673,
|
| 1200 |
+
"lr": 0.0008802934326035174,
|
| 1201 |
+
"encoder_lr": 9.309334281139113e-05,
|
| 1202 |
+
"alpha": 0.23913083970546722,
|
| 1203 |
+
"validated_this_epoch": false,
|
| 1204 |
+
"val_loss": null,
|
| 1205 |
+
"val_dice": null,
|
| 1206 |
+
"val_iou": null,
|
| 1207 |
+
"val_reward": null,
|
| 1208 |
+
"val_entropy": null
|
| 1209 |
+
},
|
| 1210 |
+
{
|
| 1211 |
+
"epoch": 55,
|
| 1212 |
+
"train_loss": 0.32538383428815953,
|
| 1213 |
+
"train_actor_loss": -0.07933734324387726,
|
| 1214 |
+
"train_critic_loss": 0.00012820682812018618,
|
| 1215 |
+
"train_mean_reward": 0.0003618560274384018,
|
| 1216 |
+
"train_entropy": 0.35111820077865313,
|
| 1217 |
+
"train_ce_loss": 0.37904192940560616,
|
| 1218 |
+
"train_dice_loss": 0.43027220978293307,
|
| 1219 |
+
"train_dice": 0.8378236286084461,
|
| 1220 |
+
"train_iou": 0.752134913587515,
|
| 1221 |
+
"grad_norm": 11.320292324520821,
|
| 1222 |
+
"lr": 0.0008493477072849704,
|
| 1223 |
+
"encoder_lr": 9.013872582117809e-05,
|
| 1224 |
+
"alpha": 0.2378426194190979,
|
| 1225 |
+
"validated_this_epoch": false,
|
| 1226 |
+
"val_loss": null,
|
| 1227 |
+
"val_dice": null,
|
| 1228 |
+
"val_iou": null,
|
| 1229 |
+
"val_reward": null,
|
| 1230 |
+
"val_entropy": null
|
| 1231 |
+
},
|
| 1232 |
+
{
|
| 1233 |
+
"epoch": 56,
|
| 1234 |
+
"train_loss": 0.32043155256745426,
|
| 1235 |
+
"train_actor_loss": -0.08012726316252573,
|
| 1236 |
+
"train_critic_loss": 0.00014749675955022626,
|
| 1237 |
+
"train_mean_reward": 0.0004917921659802637,
|
| 1238 |
+
"train_entropy": 0.3520416223879624,
|
| 1239 |
+
"train_ce_loss": 0.37978302069653297,
|
| 1240 |
+
"train_dice_loss": 0.42118710210156995,
|
| 1241 |
+
"train_dice": 0.8348648219388372,
|
| 1242 |
+
"train_iou": 0.7497471468347237,
|
| 1243 |
+
"grad_norm": 11.375983123169389,
|
| 1244 |
+
"lr": 0.0008185555919872042,
|
| 1245 |
+
"encoder_lr": 8.719877511435618e-05,
|
| 1246 |
+
"alpha": 0.23636719584465027,
|
| 1247 |
+
"validated_this_epoch": false,
|
| 1248 |
+
"val_loss": null,
|
| 1249 |
+
"val_dice": null,
|
| 1250 |
+
"val_iou": null,
|
| 1251 |
+
"val_reward": null,
|
| 1252 |
+
"val_entropy": null
|
| 1253 |
+
},
|
| 1254 |
+
{
|
| 1255 |
+
"epoch": 57,
|
| 1256 |
+
"train_loss": 0.3136355709465033,
|
| 1257 |
+
"train_actor_loss": -0.07834348190113787,
|
| 1258 |
+
"train_critic_loss": 0.00012907433981740194,
|
| 1259 |
+
"train_mean_reward": 0.0004341555929923233,
|
| 1260 |
+
"train_entropy": 0.35038877640506705,
|
| 1261 |
+
"train_ce_loss": 0.35941387500219263,
|
| 1262 |
+
"train_dice_loss": 0.42441514434740524,
|
| 1263 |
+
"train_dice": 0.8436250884573696,
|
| 1264 |
+
"train_iou": 0.7597331877029442,
|
| 1265 |
+
"grad_norm": 6.198838898608851,
|
| 1266 |
+
"lr": 0.0007879474748104406,
|
| 1267 |
+
"encoder_lr": 8.42763920673285e-05,
|
| 1268 |
+
"alpha": 0.23513460159301758,
|
| 1269 |
+
"validated_this_epoch": false,
|
| 1270 |
+
"val_loss": null,
|
| 1271 |
+
"val_dice": null,
|
| 1272 |
+
"val_iou": null,
|
| 1273 |
+
"val_reward": null,
|
| 1274 |
+
"val_entropy": null
|
| 1275 |
+
},
|
| 1276 |
+
{
|
| 1277 |
+
"epoch": 58,
|
| 1278 |
+
"train_loss": 0.30257535634865595,
|
| 1279 |
+
"train_actor_loss": -0.07731811599733755,
|
| 1280 |
+
"train_critic_loss": 0.0001238810108540773,
|
| 1281 |
+
"train_mean_reward": 0.0003711728832152729,
|
| 1282 |
+
"train_entropy": 0.34737028801472,
|
| 1283 |
+
"train_ce_loss": 0.3630384509842746,
|
| 1284 |
+
"train_dice_loss": 0.39662460185760673,
|
| 1285 |
+
"train_dice": 0.8443553555896789,
|
| 1286 |
+
"train_iou": 0.7604982336653796,
|
| 1287 |
+
"grad_norm": 5.456226746703303,
|
| 1288 |
+
"lr": 0.0007575535622709696,
|
| 1289 |
+
"encoder_lr": 8.13744607193388e-05,
|
| 1290 |
+
"alpha": 0.23428326845169067,
|
| 1291 |
+
"validated_this_epoch": false,
|
| 1292 |
+
"val_loss": null,
|
| 1293 |
+
"val_dice": null,
|
| 1294 |
+
"val_iou": null,
|
| 1295 |
+
"val_reward": null,
|
| 1296 |
+
"val_entropy": null
|
| 1297 |
+
},
|
| 1298 |
+
{
|
| 1299 |
+
"epoch": 59,
|
| 1300 |
+
"train_loss": 0.320874744423398,
|
| 1301 |
+
"train_actor_loss": -0.07865402885310795,
|
| 1302 |
+
"train_critic_loss": 0.00012862170222102266,
|
| 1303 |
+
"train_mean_reward": 0.0003906931508546276,
|
| 1304 |
+
"train_entropy": 0.3473684414836172,
|
| 1305 |
+
"train_ce_loss": 0.3601892029256524,
|
| 1306 |
+
"train_dice_loss": 0.4387397103839451,
|
| 1307 |
+
"train_dice": 0.84268284852974,
|
| 1308 |
+
"train_iou": 0.758812186256334,
|
| 1309 |
+
"grad_norm": 11.089829599672509,
|
| 1310 |
+
"lr": 0.0007274038494909672,
|
| 1311 |
+
"encoder_lr": 7.849584492627325e-05,
|
| 1312 |
+
"alpha": 0.23299352824687958,
|
| 1313 |
+
"validated_this_epoch": false,
|
| 1314 |
+
"val_loss": null,
|
| 1315 |
+
"val_dice": null,
|
| 1316 |
+
"val_iou": null,
|
| 1317 |
+
"val_reward": null,
|
| 1318 |
+
"val_entropy": null
|
| 1319 |
+
},
|
| 1320 |
+
{
|
| 1321 |
+
"epoch": 60,
|
| 1322 |
+
"train_loss": 0.3105282830146625,
|
| 1323 |
+
"train_actor_loss": -0.07681781812077652,
|
| 1324 |
+
"train_critic_loss": 0.00012296370882182312,
|
| 1325 |
+
"train_mean_reward": 0.0003677816997278381,
|
| 1326 |
+
"train_entropy": 0.34588067549297663,
|
| 1327 |
+
"train_ce_loss": 0.36672850531220286,
|
| 1328 |
+
"train_dice_loss": 0.4078407214439501,
|
| 1329 |
+
"train_dice": 0.8468779468829136,
|
| 1330 |
+
"train_iou": 0.7609672114476496,
|
| 1331 |
+
"grad_norm": 7.908274856186653,
|
| 1332 |
+
"lr": 0.0006975280905969278,
|
| 1333 |
+
"encoder_lr": 7.564338553438004e-05,
|
| 1334 |
+
"alpha": 0.23200729489326477,
|
| 1335 |
+
"validated_this_epoch": true,
|
| 1336 |
+
"val_loss": 0.6649563134767849,
|
| 1337 |
+
"val_dice": 0.6958528190812739,
|
| 1338 |
+
"val_iou": 0.6138639254904781,
|
| 1339 |
+
"val_reward": -1.8240613596821964e-05,
|
| 1340 |
+
"val_entropy": 0.34913111967269816,
|
| 1341 |
+
"val_actor_loss": -0.006444395366630452,
|
| 1342 |
+
"val_critic_loss": 9.751615970679893e-05,
|
| 1343 |
+
"val_ce_loss": 0.5944674070430638,
|
| 1344 |
+
"val_dice_loss": 0.74823647225746
|
| 1345 |
+
},
|
| 1346 |
+
{
|
| 1347 |
+
"epoch": 61,
|
| 1348 |
+
"train_loss": 0.3301074927701622,
|
| 1349 |
+
"train_actor_loss": -0.07813184656857367,
|
| 1350 |
+
"train_critic_loss": 0.00011892455378069673,
|
| 1351 |
+
"train_mean_reward": 0.0003756327501569644,
|
| 1352 |
+
"train_entropy": 0.353619627981839,
|
| 1353 |
+
"train_ce_loss": 0.38330065801460533,
|
| 1354 |
+
"train_dice_loss": 0.43305908455405123,
|
| 1355 |
+
"train_dice": 0.837464534699256,
|
| 1356 |
+
"train_iou": 0.7537084327493899,
|
| 1357 |
+
"grad_norm": 6.959432761798533,
|
| 1358 |
+
"lr": 0.0006679557693559353,
|
| 1359 |
+
"encoder_lr": 7.281989757669733e-05,
|
| 1360 |
+
"alpha": 0.23060384392738342,
|
| 1361 |
+
"validated_this_epoch": false,
|
| 1362 |
+
"val_loss": null,
|
| 1363 |
+
"val_dice": null,
|
| 1364 |
+
"val_iou": null,
|
| 1365 |
+
"val_reward": null,
|
| 1366 |
+
"val_entropy": null
|
| 1367 |
+
},
|
| 1368 |
+
{
|
| 1369 |
+
"epoch": 62,
|
| 1370 |
+
"train_loss": 0.31614200634044387,
|
| 1371 |
+
"train_actor_loss": -0.07481188490717293,
|
| 1372 |
+
"train_critic_loss": 0.00012997443304659138,
|
| 1373 |
+
"train_mean_reward": 0.00035827715709149785,
|
| 1374 |
+
"train_entropy": 0.3429424884206873,
|
| 1375 |
+
"train_ce_loss": 0.3665655380179889,
|
| 1376 |
+
"train_dice_loss": 0.4152122576569402,
|
| 1377 |
+
"train_dice": 0.8340546897767646,
|
| 1378 |
+
"train_iou": 0.7493192868554416,
|
| 1379 |
+
"grad_norm": 7.151455770629321,
|
| 1380 |
+
"lr": 0.000638716070078746,
|
| 1381 |
+
"encoder_lr": 7.002816749495564e-05,
|
| 1382 |
+
"alpha": 0.22981132566928864,
|
| 1383 |
+
"validated_this_epoch": false,
|
| 1384 |
+
"val_loss": null,
|
| 1385 |
+
"val_dice": null,
|
| 1386 |
+
"val_iou": null,
|
| 1387 |
+
"val_reward": null,
|
| 1388 |
+
"val_entropy": null
|
| 1389 |
+
},
|
| 1390 |
+
{
|
| 1391 |
+
"epoch": 63,
|
| 1392 |
+
"train_loss": 0.30474833079592084,
|
| 1393 |
+
"train_actor_loss": -0.07550278343789724,
|
| 1394 |
+
"train_critic_loss": 0.0001240398346665241,
|
| 1395 |
+
"train_mean_reward": 0.0004029202844719014,
|
| 1396 |
+
"train_entropy": 0.3461758896939514,
|
| 1397 |
+
"train_ce_loss": 0.3557888093712507,
|
| 1398 |
+
"train_dice_loss": 0.40458936754431224,
|
| 1399 |
+
"train_dice": 0.8510938875218844,
|
| 1400 |
+
"train_iou": 0.7674811212822945,
|
| 1401 |
+
"grad_norm": 5.469105230745419,
|
| 1402 |
+
"lr": 0.0006098378488183934,
|
| 1403 |
+
"encoder_lr": 6.727095038969585e-05,
|
| 1404 |
+
"alpha": 0.2287554144859314,
|
| 1405 |
+
"validated_this_epoch": false,
|
| 1406 |
+
"val_loss": null,
|
| 1407 |
+
"val_dice": null,
|
| 1408 |
+
"val_iou": null,
|
| 1409 |
+
"val_reward": null,
|
| 1410 |
+
"val_entropy": null
|
| 1411 |
+
},
|
| 1412 |
+
{
|
| 1413 |
+
"epoch": 64,
|
| 1414 |
+
"train_loss": 0.31098741356846554,
|
| 1415 |
+
"train_actor_loss": -0.0741845433015463,
|
| 1416 |
+
"train_critic_loss": 0.00012258184143409552,
|
| 1417 |
+
"train_mean_reward": 0.00041871562106320136,
|
| 1418 |
+
"train_entropy": 0.34271789532761243,
|
| 1419 |
+
"train_ce_loss": 0.3555328384732457,
|
| 1420 |
+
"train_dice_loss": 0.4146884819179851,
|
| 1421 |
+
"train_dice": 0.8468261339730058,
|
| 1422 |
+
"train_iou": 0.7618381483191923,
|
| 1423 |
+
"grad_norm": 5.863205309524092,
|
| 1424 |
+
"lr": 0.0005813496048927529,
|
| 1425 |
+
"encoder_lr": 6.455096730131812e-05,
|
| 1426 |
+
"alpha": 0.22778737545013428,
|
| 1427 |
+
"validated_this_epoch": false,
|
| 1428 |
+
"val_loss": null,
|
| 1429 |
+
"val_dice": null,
|
| 1430 |
+
"val_iou": null,
|
| 1431 |
+
"val_reward": null,
|
| 1432 |
+
"val_entropy": null
|
| 1433 |
+
},
|
| 1434 |
+
{
|
| 1435 |
+
"epoch": 65,
|
| 1436 |
+
"train_loss": 0.29980203252675575,
|
| 1437 |
+
"train_actor_loss": -0.0764638027176261,
|
| 1438 |
+
"train_critic_loss": 0.00013137878641130153,
|
| 1439 |
+
"train_mean_reward": 0.0004909961741867722,
|
| 1440 |
+
"train_entropy": 0.34714439830884763,
|
| 1441 |
+
"train_ce_loss": 0.35501439714785044,
|
| 1442 |
+
"train_dice_loss": 0.39738588346991427,
|
| 1443 |
+
"train_dice": 0.8396979201480519,
|
| 1444 |
+
"train_iou": 0.7559530514796582,
|
| 1445 |
+
"grad_norm": 5.454546915699345,
|
| 1446 |
+
"lr": 0.0005532794527591512,
|
| 1447 |
+
"encoder_lr": 6.187090252474308e-05,
|
| 1448 |
+
"alpha": 0.22657504677772522,
|
| 1449 |
+
"validated_this_epoch": false,
|
| 1450 |
+
"val_loss": null,
|
| 1451 |
+
"val_dice": null,
|
| 1452 |
+
"val_iou": null,
|
| 1453 |
+
"val_reward": null,
|
| 1454 |
+
"val_entropy": null
|
| 1455 |
+
},
|
| 1456 |
+
{
|
| 1457 |
+
"epoch": 66,
|
| 1458 |
+
"train_loss": 0.29371613644193423,
|
| 1459 |
+
"train_actor_loss": -0.07414838349451468,
|
| 1460 |
+
"train_critic_loss": 0.0001205393308248433,
|
| 1461 |
+
"train_mean_reward": 0.00048073157674036,
|
| 1462 |
+
"train_entropy": 0.3377886277067569,
|
| 1463 |
+
"train_ce_loss": 0.34613867473213406,
|
| 1464 |
+
"train_dice_loss": 0.389469816318162,
|
| 1465 |
+
"train_dice": 0.8538910952080454,
|
| 1466 |
+
"train_iou": 0.7699731266117984,
|
| 1467 |
+
"grad_norm": 8.586026681023975,
|
| 1468 |
+
"lr": 0.0005256550942687934,
|
| 1469 |
+
"encoder_lr": 5.9233400960337056e-05,
|
| 1470 |
+
"alpha": 0.22605356574058533,
|
| 1471 |
+
"validated_this_epoch": false,
|
| 1472 |
+
"val_loss": null,
|
| 1473 |
+
"val_dice": null,
|
| 1474 |
+
"val_iou": null,
|
| 1475 |
+
"val_reward": null,
|
| 1476 |
+
"val_entropy": null
|
| 1477 |
+
},
|
| 1478 |
+
{
|
| 1479 |
+
"epoch": 67,
|
| 1480 |
+
"train_loss": 0.329106808603965,
|
| 1481 |
+
"train_actor_loss": -0.07454048351261967,
|
| 1482 |
+
"train_critic_loss": 0.00013679793005805731,
|
| 1483 |
+
"train_mean_reward": 0.00045310201175696247,
|
| 1484 |
+
"train_entropy": 0.34272057655476784,
|
| 1485 |
+
"train_ce_loss": 0.3664403586667657,
|
| 1486 |
+
"train_dice_loss": 0.44071741529213365,
|
| 1487 |
+
"train_dice": 0.8368334628670275,
|
| 1488 |
+
"train_iou": 0.7519591024692609,
|
| 1489 |
+
"grad_norm": 7.676452435957369,
|
| 1490 |
+
"lr": 0.0004985037913283807,
|
| 1491 |
+
"encoder_lr": 5.6641065503714745e-05,
|
| 1492 |
+
"alpha": 0.22515802085399628,
|
| 1493 |
+
"validated_this_epoch": false,
|
| 1494 |
+
"val_loss": null,
|
| 1495 |
+
"val_dice": null,
|
| 1496 |
+
"val_iou": null,
|
| 1497 |
+
"val_reward": null,
|
| 1498 |
+
"val_entropy": null
|
| 1499 |
+
},
|
| 1500 |
+
{
|
| 1501 |
+
"epoch": 68,
|
| 1502 |
+
"train_loss": 0.29258083760045295,
|
| 1503 |
+
"train_actor_loss": -0.07457079501095823,
|
| 1504 |
+
"train_critic_loss": 0.00012341178801371879,
|
| 1505 |
+
"train_mean_reward": 0.0004348753051417424,
|
| 1506 |
+
"train_entropy": 0.3423526900067193,
|
| 1507 |
+
"train_ce_loss": 0.3466221644959439,
|
| 1508 |
+
"train_dice_loss": 0.38755767929153545,
|
| 1509 |
+
"train_dice": 0.8491057322818648,
|
| 1510 |
+
"train_iou": 0.7672465711322699,
|
| 1511 |
+
"grad_norm": 4.713813254075457,
|
| 1512 |
+
"lr": 0.0004718523389958983,
|
| 1513 |
+
"encoder_lr": 5.4096454476995315e-05,
|
| 1514 |
+
"alpha": 0.22431327402591705,
|
| 1515 |
+
"validated_this_epoch": false,
|
| 1516 |
+
"val_loss": null,
|
| 1517 |
+
"val_dice": null,
|
| 1518 |
+
"val_iou": null,
|
| 1519 |
+
"val_reward": null,
|
| 1520 |
+
"val_entropy": null
|
| 1521 |
+
},
|
| 1522 |
+
{
|
| 1523 |
+
"epoch": 69,
|
| 1524 |
+
"train_loss": 0.2899718071810555,
|
| 1525 |
+
"train_actor_loss": -0.07490196581496748,
|
| 1526 |
+
"train_critic_loss": 0.00012304655573589632,
|
| 1527 |
+
"train_mean_reward": 0.0005400658274894771,
|
| 1528 |
+
"train_entropy": 0.3426005875147899,
|
| 1529 |
+
"train_ce_loss": 0.3466990191215115,
|
| 1530 |
+
"train_dice_loss": 0.38292547097809865,
|
| 1531 |
+
"train_dice": 0.8568332990466453,
|
| 1532 |
+
"train_iou": 0.7745671559981744,
|
| 1533 |
+
"grad_norm": 4.423254855843478,
|
| 1534 |
+
"lr": 0.0004457270390371302,
|
| 1535 |
+
"encoder_lr": 5.1602079104047604e-05,
|
| 1536 |
+
"alpha": 0.22371403872966766,
|
| 1537 |
+
"validated_this_epoch": false,
|
| 1538 |
+
"val_loss": null,
|
| 1539 |
+
"val_dice": null,
|
| 1540 |
+
"val_iou": null,
|
| 1541 |
+
"val_reward": null,
|
| 1542 |
+
"val_entropy": null
|
| 1543 |
+
},
|
| 1544 |
+
{
|
| 1545 |
+
"epoch": 70,
|
| 1546 |
+
"train_loss": 0.305683577375472,
|
| 1547 |
+
"train_actor_loss": -0.07387165723305955,
|
| 1548 |
+
"train_critic_loss": 0.00012970691598942116,
|
| 1549 |
+
"train_mean_reward": 0.0004675847167686639,
|
| 1550 |
+
"train_entropy": 0.33587084764504005,
|
| 1551 |
+
"train_ce_loss": 0.3494242132143226,
|
| 1552 |
+
"train_dice_loss": 0.4095565396377898,
|
| 1553 |
+
"train_dice": 0.8465224365174171,
|
| 1554 |
+
"train_iou": 0.7623571347045611,
|
| 1555 |
+
"grad_norm": 5.9620601534843445,
|
| 1556 |
+
"lr": 0.00042015367396898955,
|
| 1557 |
+
"encoder_lr": 4.916040103221508e-05,
|
| 1558 |
+
"alpha": 0.22300413250923157,
|
| 1559 |
+
"validated_this_epoch": true,
|
| 1560 |
+
"val_loss": 0.6379372473509813,
|
| 1561 |
+
"val_dice": 0.7050720054154305,
|
| 1562 |
+
"val_iou": 0.6170998262921583,
|
| 1563 |
+
"val_reward": 7.479539349195402e-05,
|
| 1564 |
+
"val_entropy": 0.34862531074369796,
|
| 1565 |
+
"val_actor_loss": -0.007996960452991782,
|
| 1566 |
+
"val_critic_loss": 0.00013556680095338047,
|
| 1567 |
+
"val_ce_loss": 0.5276325090901193,
|
| 1568 |
+
"val_dice_loss": 0.7641003195083502
|
| 1569 |
+
},
|
| 1570 |
+
{
|
| 1571 |
+
"epoch": 71,
|
| 1572 |
+
"train_loss": 0.302607683136395,
|
| 1573 |
+
"train_actor_loss": -0.07420629979104035,
|
| 1574 |
+
"train_critic_loss": 0.0001304845529771592,
|
| 1575 |
+
"train_mean_reward": 0.0005424884828230003,
|
| 1576 |
+
"train_entropy": 0.33843888987327425,
|
| 1577 |
+
"train_ce_loss": 0.35047898109333264,
|
| 1578 |
+
"train_dice_loss": 0.40301849067365164,
|
| 1579 |
+
"train_dice": 0.8512519690890744,
|
| 1580 |
+
"train_iou": 0.7682522469497006,
|
| 1581 |
+
"grad_norm": 6.331226801918459,
|
| 1582 |
+
"lr": 0.0003951574816152885,
|
| 1583 |
+
"encoder_lr": 4.6773829902967244e-05,
|
| 1584 |
+
"alpha": 0.22242465615272522,
|
| 1585 |
+
"validated_this_epoch": false,
|
| 1586 |
+
"val_loss": null,
|
| 1587 |
+
"val_dice": null,
|
| 1588 |
+
"val_iou": null,
|
| 1589 |
+
"val_reward": null,
|
| 1590 |
+
"val_entropy": null
|
| 1591 |
+
},
|
| 1592 |
+
{
|
| 1593 |
+
"epoch": 72,
|
| 1594 |
+
"train_loss": 0.3075731982650279,
|
| 1595 |
+
"train_actor_loss": -0.07134573754568124,
|
| 1596 |
+
"train_critic_loss": 0.0001290610435448011,
|
| 1597 |
+
"train_mean_reward": 0.00047534032294369375,
|
| 1598 |
+
"train_entropy": 0.3377253075517734,
|
| 1599 |
+
"train_ce_loss": 0.35495021727203224,
|
| 1600 |
+
"train_dice_loss": 0.402758582309851,
|
| 1601 |
+
"train_dice": 0.8463728467991206,
|
| 1602 |
+
"train_iou": 0.764645717069721,
|
| 1603 |
+
"grad_norm": 6.850196821745052,
|
| 1604 |
+
"lr": 0.0003707631302000539,
|
| 1605 |
+
"encoder_lr": 4.444472097387449e-05,
|
| 1606 |
+
"alpha": 0.22197401523590088,
|
| 1607 |
+
"validated_this_epoch": false,
|
| 1608 |
+
"val_loss": null,
|
| 1609 |
+
"val_dice": null,
|
| 1610 |
+
"val_iou": null,
|
| 1611 |
+
"val_reward": null,
|
| 1612 |
+
"val_entropy": null
|
| 1613 |
+
},
|
| 1614 |
+
{
|
| 1615 |
+
"epoch": 73,
|
| 1616 |
+
"train_loss": 0.2975923801743418,
|
| 1617 |
+
"train_actor_loss": -0.07246502377189099,
|
| 1618 |
+
"train_critic_loss": 0.00012726665574556452,
|
| 1619 |
+
"train_mean_reward": 0.0004314572466396162,
|
| 1620 |
+
"train_entropy": 0.33378850096880003,
|
| 1621 |
+
"train_ce_loss": 0.3434440972643257,
|
| 1622 |
+
"train_dice_loss": 0.39654343342288206,
|
| 1623 |
+
"train_dice": 0.8469582528769581,
|
| 1624 |
+
"train_iou": 0.7639696928234165,
|
| 1625 |
+
"grad_norm": 4.767858616372411,
|
| 1626 |
+
"lr": 0.00034699469400296674,
|
| 1627 |
+
"encoder_lr": 4.2175372794253104e-05,
|
| 1628 |
+
"alpha": 0.22163283824920654,
|
| 1629 |
+
"validated_this_epoch": false,
|
| 1630 |
+
"val_loss": null,
|
| 1631 |
+
"val_dice": null,
|
| 1632 |
+
"val_iou": null,
|
| 1633 |
+
"val_reward": null,
|
| 1634 |
+
"val_entropy": null
|
| 1635 |
+
},
|
| 1636 |
+
{
|
| 1637 |
+
"epoch": 74,
|
| 1638 |
+
"train_loss": 0.3076653238819089,
|
| 1639 |
+
"train_actor_loss": -0.07338658093116235,
|
| 1640 |
+
"train_critic_loss": 0.00012848509119919968,
|
| 1641 |
+
"train_mean_reward": 0.0006016681011137135,
|
| 1642 |
+
"train_entropy": 0.3369606134772917,
|
| 1643 |
+
"train_ce_loss": 0.3439181402135515,
|
| 1644 |
+
"train_dice_loss": 0.41805717325949854,
|
| 1645 |
+
"train_dice": 0.8493592940463257,
|
| 1646 |
+
"train_iou": 0.7668287099033275,
|
| 1647 |
+
"grad_norm": 7.26604178663372,
|
| 1648 |
+
"lr": 0.0003238756296009549,
|
| 1649 |
+
"encoder_lr": 3.996802493677459e-05,
|
| 1650 |
+
"alpha": 0.22104693949222565,
|
| 1651 |
+
"validated_this_epoch": false,
|
| 1652 |
+
"val_loss": null,
|
| 1653 |
+
"val_dice": null,
|
| 1654 |
+
"val_iou": null,
|
| 1655 |
+
"val_reward": null,
|
| 1656 |
+
"val_entropy": null
|
| 1657 |
+
},
|
| 1658 |
+
{
|
| 1659 |
+
"epoch": 75,
|
| 1660 |
+
"train_loss": 0.3228878648915552,
|
| 1661 |
+
"train_actor_loss": -0.0733624054528099,
|
| 1662 |
+
"train_critic_loss": 0.00012589249067798772,
|
| 1663 |
+
"train_mean_reward": 0.0006108267801936062,
|
| 1664 |
+
"train_entropy": 0.3334116924178693,
|
| 1665 |
+
"train_ce_loss": 0.3691949419609875,
|
| 1666 |
+
"train_dice_loss": 0.4231796933082955,
|
| 1667 |
+
"train_dice": 0.8466939581862797,
|
| 1668 |
+
"train_iou": 0.7621407584407365,
|
| 1669 |
+
"grad_norm": 5.312329442002053,
|
| 1670 |
+
"lr": 0.0003014287527193855,
|
| 1671 |
+
"encoder_lr": 3.782485578727801e-05,
|
| 1672 |
+
"alpha": 0.22076164186000824,
|
| 1673 |
+
"validated_this_epoch": false,
|
| 1674 |
+
"val_loss": null,
|
| 1675 |
+
"val_dice": null,
|
| 1676 |
+
"val_iou": null,
|
| 1677 |
+
"val_reward": null,
|
| 1678 |
+
"val_entropy": null
|
| 1679 |
+
},
|
| 1680 |
+
{
|
| 1681 |
+
"epoch": 76,
|
| 1682 |
+
"train_loss": 0.290782397991956,
|
| 1683 |
+
"train_actor_loss": -0.07353188385537197,
|
| 1684 |
+
"train_critic_loss": 0.0001260726710141227,
|
| 1685 |
+
"train_mean_reward": 0.0006490182575724187,
|
| 1686 |
+
"train_entropy": 0.3373131515597804,
|
| 1687 |
+
"train_ce_loss": 0.34604928147430397,
|
| 1688 |
+
"train_dice_loss": 0.38245319973590763,
|
| 1689 |
+
"train_dice": 0.8528491707374566,
|
| 1690 |
+
"train_iou": 0.771310685411331,
|
| 1691 |
+
"grad_norm": 5.503487736679787,
|
| 1692 |
+
"lr": 0.0002796762157156959,
|
| 1693 |
+
"encoder_lr": 3.574798039496594e-05,
|
| 1694 |
+
"alpha": 0.2202751785516739,
|
| 1695 |
+
"validated_this_epoch": false,
|
| 1696 |
+
"val_loss": null,
|
| 1697 |
+
"val_dice": null,
|
| 1698 |
+
"val_iou": null,
|
| 1699 |
+
"val_reward": null,
|
| 1700 |
+
"val_entropy": null
|
| 1701 |
+
},
|
| 1702 |
+
{
|
| 1703 |
+
"epoch": 77,
|
| 1704 |
+
"train_loss": 0.31818086738416623,
|
| 1705 |
+
"train_actor_loss": -0.07224201335684814,
|
| 1706 |
+
"train_critic_loss": 0.00012977430109492586,
|
| 1707 |
+
"train_mean_reward": 0.00048386759957563086,
|
| 1708 |
+
"train_entropy": 0.3344568842757272,
|
| 1709 |
+
"train_ce_loss": 0.3542836576433087,
|
| 1710 |
+
"train_dice_loss": 0.426432318311637,
|
| 1711 |
+
"train_dice": 0.83912447215906,
|
| 1712 |
+
"train_iou": 0.7565115097710524,
|
| 1713 |
+
"grad_norm": 7.540519112533377,
|
| 1714 |
+
"lr": 0.00025863948571769286,
|
| 1715 |
+
"encoder_lr": 3.373944838510636e-05,
|
| 1716 |
+
"alpha": 0.21985827386379242,
|
| 1717 |
+
"validated_this_epoch": false,
|
| 1718 |
+
"val_loss": null,
|
| 1719 |
+
"val_dice": null,
|
| 1720 |
+
"val_iou": null,
|
| 1721 |
+
"val_reward": null,
|
| 1722 |
+
"val_entropy": null
|
| 1723 |
+
},
|
| 1724 |
+
{
|
| 1725 |
+
"epoch": 78,
|
| 1726 |
+
"train_loss": 0.29576459631338253,
|
| 1727 |
+
"train_actor_loss": -0.07245105145937097,
|
| 1728 |
+
"train_critic_loss": 0.00012679694167994969,
|
| 1729 |
+
"train_mean_reward": 0.0005185490083988056,
|
| 1730 |
+
"train_entropy": 0.3339699495593399,
|
| 1731 |
+
"train_ce_loss": 0.34016143481995553,
|
| 1732 |
+
"train_dice_loss": 0.3961430534681916,
|
| 1733 |
+
"train_dice": 0.8449982278366371,
|
| 1734 |
+
"train_iou": 0.7641634822257897,
|
| 1735 |
+
"grad_norm": 5.001694010209667,
|
| 1736 |
+
"lr": 0.00023833932343808993,
|
| 1737 |
+
"encoder_lr": 3.180124193630005e-05,
|
| 1738 |
+
"alpha": 0.2195238620042801,
|
| 1739 |
+
"validated_this_epoch": false,
|
| 1740 |
+
"val_loss": null,
|
| 1741 |
+
"val_dice": null,
|
| 1742 |
+
"val_iou": null,
|
| 1743 |
+
"val_reward": null,
|
| 1744 |
+
"val_entropy": null
|
| 1745 |
+
},
|
| 1746 |
+
{
|
| 1747 |
+
"epoch": 79,
|
| 1748 |
+
"train_loss": 0.30173337690228985,
|
| 1749 |
+
"train_actor_loss": -0.07268108915807125,
|
| 1750 |
+
"train_critic_loss": 0.0001269460498205326,
|
| 1751 |
+
"train_mean_reward": 0.0005702076259506868,
|
| 1752 |
+
"train_entropy": 0.3311257389318727,
|
| 1753 |
+
"train_ce_loss": 0.34784492430245884,
|
| 1754 |
+
"train_dice_loss": 0.40085704990443644,
|
| 1755 |
+
"train_dice": 0.854640596598362,
|
| 1756 |
+
"train_iou": 0.7711418630650612,
|
| 1757 |
+
"grad_norm": 4.701322214779004,
|
| 1758 |
+
"lr": 0.00021879576268618817,
|
| 1759 |
+
"encoder_lr": 2.993527382430943e-05,
|
| 1760 |
+
"alpha": 0.21928933262825012,
|
| 1761 |
+
"validated_this_epoch": false,
|
| 1762 |
+
"val_loss": null,
|
| 1763 |
+
"val_dice": null,
|
| 1764 |
+
"val_iou": null,
|
| 1765 |
+
"val_reward": null,
|
| 1766 |
+
"val_entropy": null
|
| 1767 |
+
},
|
| 1768 |
+
{
|
| 1769 |
+
"epoch": 80,
|
| 1770 |
+
"train_loss": 0.2950800986556235,
|
| 1771 |
+
"train_actor_loss": -0.07325161035949948,
|
| 1772 |
+
"train_critic_loss": 0.00012739866826934576,
|
| 1773 |
+
"train_mean_reward": 0.0005077752536596894,
|
| 1774 |
+
"train_entropy": 0.3354396403644436,
|
| 1775 |
+
"train_ce_loss": 0.3472645461910647,
|
| 1776 |
+
"train_dice_loss": 0.3892714626105257,
|
| 1777 |
+
"train_dice": 0.8520999605508252,
|
| 1778 |
+
"train_iou": 0.7686250539354682,
|
| 1779 |
+
"grad_norm": 4.815572581780973,
|
| 1780 |
+
"lr": 0.0002000280905969275,
|
| 1781 |
+
"encoder_lr": 2.814338553438002e-05,
|
| 1782 |
+
"alpha": 0.2189188301563263,
|
| 1783 |
+
"validated_this_epoch": true,
|
| 1784 |
+
"val_loss": 0.6329006587903043,
|
| 1785 |
+
"val_dice": 0.7035739955410152,
|
| 1786 |
+
"val_iou": 0.6207132033097706,
|
| 1787 |
+
"val_reward": -0.00014857503323252206,
|
| 1788 |
+
"val_entropy": 0.32817754886969175,
|
| 1789 |
+
"val_actor_loss": -0.009106944154035488,
|
| 1790 |
+
"val_critic_loss": 0.0001434129035570467,
|
| 1791 |
+
"val_ce_loss": 0.5413949469320778,
|
| 1792 |
+
"val_dice_loss": 0.7424768203436726
|
| 1793 |
+
},
|
| 1794 |
+
{
|
| 1795 |
+
"epoch": 81,
|
| 1796 |
+
"train_loss": 0.3003477848156594,
|
| 1797 |
+
"train_actor_loss": -0.0726600034927675,
|
| 1798 |
+
"train_critic_loss": 0.00011839045538070925,
|
| 1799 |
+
"train_mean_reward": 0.0005749791793611554,
|
| 1800 |
+
"train_entropy": 0.3297243164683806,
|
| 1801 |
+
"train_ce_loss": 0.3492033497673443,
|
| 1802 |
+
"train_dice_loss": 0.3966938235655003,
|
| 1803 |
+
"train_dice": 0.8511939499762166,
|
| 1804 |
+
"train_iou": 0.7707929403525045,
|
| 1805 |
+
"grad_norm": 3.875607972459276,
|
| 1806 |
+
"lr": 0.00018205482859681116,
|
| 1807 |
+
"encoder_lr": 2.6427345443916652e-05,
|
| 1808 |
+
"alpha": 0.2189090996980667,
|
| 1809 |
+
"validated_this_epoch": false,
|
| 1810 |
+
"val_loss": null,
|
| 1811 |
+
"val_dice": null,
|
| 1812 |
+
"val_iou": null,
|
| 1813 |
+
"val_reward": null,
|
| 1814 |
+
"val_entropy": null
|
| 1815 |
+
},
|
| 1816 |
+
{
|
| 1817 |
+
"epoch": 82,
|
| 1818 |
+
"train_loss": 0.2957473771155689,
|
| 1819 |
+
"train_actor_loss": -0.07189562755518952,
|
| 1820 |
+
"train_critic_loss": 0.00011833291176755256,
|
| 1821 |
+
"train_mean_reward": 0.0005660512295139068,
|
| 1822 |
+
"train_entropy": 0.3330608050143996,
|
| 1823 |
+
"train_ce_loss": 0.34475735014693165,
|
| 1824 |
+
"train_dice_loss": 0.39041031669892695,
|
| 1825 |
+
"train_dice": 0.8531728218796693,
|
| 1826 |
+
"train_iou": 0.7714247744385839,
|
| 1827 |
+
"grad_norm": 3.8687666007714676,
|
| 1828 |
+
"lr": 0.00016489371412549528,
|
| 1829 |
+
"encoder_lr": 2.4788847077308606e-05,
|
| 1830 |
+
"alpha": 0.21881625056266785,
|
| 1831 |
+
"validated_this_epoch": false,
|
| 1832 |
+
"val_loss": null,
|
| 1833 |
+
"val_dice": null,
|
| 1834 |
+
"val_iou": null,
|
| 1835 |
+
"val_reward": null,
|
| 1836 |
+
"val_entropy": null
|
| 1837 |
+
},
|
| 1838 |
+
{
|
| 1839 |
+
"epoch": 83,
|
| 1840 |
+
"train_loss": 0.2942674369948317,
|
| 1841 |
+
"train_actor_loss": -0.07288596809430178,
|
| 1842 |
+
"train_critic_loss": 0.00011483322810942766,
|
| 1843 |
+
"train_mean_reward": 0.0006078585665266259,
|
| 1844 |
+
"train_entropy": 0.3294820947818054,
|
| 1845 |
+
"train_ce_loss": 0.3541158644220183,
|
| 1846 |
+
"train_dice_loss": 0.3800761016610365,
|
| 1847 |
+
"train_dice": 0.8553531212041255,
|
| 1848 |
+
"train_iou": 0.7739833114555351,
|
| 1849 |
+
"grad_norm": 3.613588293393453,
|
| 1850 |
+
"lr": 0.00014856168313107603,
|
| 1851 |
+
"encoder_lr": 2.322950743462536e-05,
|
| 1852 |
+
"alpha": 0.21876578032970428,
|
| 1853 |
+
"validated_this_epoch": false,
|
| 1854 |
+
"val_loss": null,
|
| 1855 |
+
"val_dice": null,
|
| 1856 |
+
"val_iou": null,
|
| 1857 |
+
"val_reward": null,
|
| 1858 |
+
"val_entropy": null
|
| 1859 |
+
},
|
| 1860 |
+
{
|
| 1861 |
+
"epoch": 84,
|
| 1862 |
+
"train_loss": 0.30299433295492284,
|
| 1863 |
+
"train_actor_loss": -0.07273269801350064,
|
| 1864 |
+
"train_critic_loss": 0.0001281139475859469,
|
| 1865 |
+
"train_mean_reward": 0.0006103556736651643,
|
| 1866 |
+
"train_entropy": 0.33006053392893275,
|
| 1867 |
+
"train_ce_loss": 0.3439405761600186,
|
| 1868 |
+
"train_dice_loss": 0.407385361826081,
|
| 1869 |
+
"train_dice": 0.8419462800496388,
|
| 1870 |
+
"train_iou": 0.7585171846710167,
|
| 1871 |
+
"grad_norm": 6.251431583434112,
|
| 1872 |
+
"lr": 0.0001330748533563558,
|
| 1873 |
+
"encoder_lr": 2.1750865395832973e-05,
|
| 1874 |
+
"alpha": 0.21858546137809753,
|
| 1875 |
+
"validated_this_epoch": false,
|
| 1876 |
+
"val_loss": null,
|
| 1877 |
+
"val_dice": null,
|
| 1878 |
+
"val_iou": null,
|
| 1879 |
+
"val_reward": null,
|
| 1880 |
+
"val_entropy": null
|
| 1881 |
+
},
|
| 1882 |
+
{
|
| 1883 |
+
"epoch": 85,
|
| 1884 |
+
"train_loss": 0.27844606626547813,
|
| 1885 |
+
"train_actor_loss": -0.0726372158005641,
|
| 1886 |
+
"train_critic_loss": 0.00012456709807693353,
|
| 1887 |
+
"train_mean_reward": 0.0005866477289987474,
|
| 1888 |
+
"train_entropy": 0.3331384026958037,
|
| 1889 |
+
"train_ce_loss": 0.3355021872256885,
|
| 1890 |
+
"train_dice_loss": 0.36653979865771547,
|
| 1891 |
+
"train_dice": 0.8636427388895668,
|
| 1892 |
+
"train_iou": 0.781636528442747,
|
| 1893 |
+
"grad_norm": 3.052272923695025,
|
| 1894 |
+
"lr": 0.0001184485084325741,
|
| 1895 |
+
"encoder_lr": 2.035438020210507e-05,
|
| 1896 |
+
"alpha": 0.21820293366909027,
|
| 1897 |
+
"validated_this_epoch": false,
|
| 1898 |
+
"val_loss": null,
|
| 1899 |
+
"val_dice": null,
|
| 1900 |
+
"val_iou": null,
|
| 1901 |
+
"val_reward": null,
|
| 1902 |
+
"val_entropy": null
|
| 1903 |
+
},
|
| 1904 |
+
{
|
| 1905 |
+
"epoch": 86,
|
| 1906 |
+
"train_loss": 0.31039473258117717,
|
| 1907 |
+
"train_actor_loss": -0.07158213671421973,
|
| 1908 |
+
"train_critic_loss": 0.00012783202084665156,
|
| 1909 |
+
"train_mean_reward": 0.0004846688327800383,
|
| 1910 |
+
"train_entropy": 0.3312787178913444,
|
| 1911 |
+
"train_ce_loss": 0.36330115666206286,
|
| 1912 |
+
"train_dice_loss": 0.40052473814604506,
|
| 1913 |
+
"train_dice": 0.8512062368310445,
|
| 1914 |
+
"train_iou": 0.7711882558483109,
|
| 1915 |
+
"grad_norm": 3.344195562739705,
|
| 1916 |
+
"lr": 0.00010469708279631079,
|
| 1917 |
+
"encoder_lr": 1.904143001572817e-05,
|
| 1918 |
+
"alpha": 0.21822436153888702,
|
| 1919 |
+
"validated_this_epoch": false,
|
| 1920 |
+
"val_loss": null,
|
| 1921 |
+
"val_dice": null,
|
| 1922 |
+
"val_iou": null,
|
| 1923 |
+
"val_reward": null,
|
| 1924 |
+
"val_entropy": null
|
| 1925 |
+
},
|
| 1926 |
+
{
|
| 1927 |
+
"epoch": 87,
|
| 1928 |
+
"train_loss": 0.29110966717766584,
|
| 1929 |
+
"train_actor_loss": -0.07334934561741983,
|
| 1930 |
+
"train_critic_loss": 0.00012235310142910956,
|
| 1931 |
+
"train_mean_reward": 0.0005854650860663439,
|
| 1932 |
+
"train_entropy": 0.33028301381017505,
|
| 1933 |
+
"train_ce_loss": 0.3440100855310748,
|
| 1934 |
+
"train_dice_loss": 0.3847855763841969,
|
| 1935 |
+
"train_dice": 0.8503085265979164,
|
| 1936 |
+
"train_iou": 0.7699025515159946,
|
| 1937 |
+
"grad_norm": 5.067409404488497,
|
| 1938 |
+
"lr": 9.183414744443903e-05,
|
| 1939 |
+
"encoder_lr": 1.781331056002182e-05,
|
| 1940 |
+
"alpha": 0.21813806891441345,
|
| 1941 |
+
"validated_this_epoch": false,
|
| 1942 |
+
"val_loss": null,
|
| 1943 |
+
"val_dice": null,
|
| 1944 |
+
"val_iou": null,
|
| 1945 |
+
"val_reward": null,
|
| 1946 |
+
"val_entropy": null
|
| 1947 |
+
},
|
| 1948 |
+
{
|
| 1949 |
+
"epoch": 88,
|
| 1950 |
+
"train_loss": 0.2794101889779459,
|
| 1951 |
+
"train_actor_loss": -0.07298987358808517,
|
| 1952 |
+
"train_critic_loss": 0.00012673062629679072,
|
| 1953 |
+
"train_mean_reward": 0.0006920671223801019,
|
| 1954 |
+
"train_entropy": 0.3303463742213964,
|
| 1955 |
+
"train_ce_loss": 0.33826364233586886,
|
| 1956 |
+
"train_dice_loss": 0.3664097426319615,
|
| 1957 |
+
"train_dice": 0.8579464613755736,
|
| 1958 |
+
"train_iou": 0.7765472521918535,
|
| 1959 |
+
"grad_norm": 3.761576713517655,
|
| 1960 |
+
"lr": 7.987239654118994e-05,
|
| 1961 |
+
"encoder_lr": 1.6671233840616128e-05,
|
| 1962 |
+
"alpha": 0.21825584769248962,
|
| 1963 |
+
"validated_this_epoch": false,
|
| 1964 |
+
"val_loss": null,
|
| 1965 |
+
"val_dice": null,
|
| 1966 |
+
"val_iou": null,
|
| 1967 |
+
"val_reward": null,
|
| 1968 |
+
"val_entropy": null
|
| 1969 |
+
},
|
| 1970 |
+
{
|
| 1971 |
+
"epoch": 89,
|
| 1972 |
+
"train_loss": 0.3108916397661317,
|
| 1973 |
+
"train_actor_loss": -0.0721594071983136,
|
| 1974 |
+
"train_critic_loss": 0.00012775127425900966,
|
| 1975 |
+
"train_mean_reward": 0.0004806854820389249,
|
| 1976 |
+
"train_entropy": 0.3311913824936216,
|
| 1977 |
+
"train_ce_loss": 0.3578439022315719,
|
| 1978 |
+
"train_dice_loss": 0.4081304284337263,
|
| 1979 |
+
"train_dice": 0.850257847886959,
|
| 1980 |
+
"train_iou": 0.7678737385572921,
|
| 1981 |
+
"grad_norm": 3.85180858840314,
|
| 1982 |
+
"lr": 6.88236348905457e-05,
|
| 1983 |
+
"encoder_lr": 1.5616326949348587e-05,
|
| 1984 |
+
"alpha": 0.21780183911323547,
|
| 1985 |
+
"validated_this_epoch": false,
|
| 1986 |
+
"val_loss": null,
|
| 1987 |
+
"val_dice": null,
|
| 1988 |
+
"val_iou": null,
|
| 1989 |
+
"val_reward": null,
|
| 1990 |
+
"val_entropy": null
|
| 1991 |
+
},
|
| 1992 |
+
{
|
| 1993 |
+
"epoch": 90,
|
| 1994 |
+
"train_loss": 0.29786181929990535,
|
| 1995 |
+
"train_actor_loss": -0.07285923995462654,
|
| 1996 |
+
"train_critic_loss": 0.00012484254698545518,
|
| 1997 |
+
"train_mean_reward": 0.000593109968966874,
|
| 1998 |
+
"train_entropy": 0.33138341025403617,
|
| 1999 |
+
"train_ce_loss": 0.33793467814848793,
|
| 2000 |
+
"train_dice_loss": 0.40338258841857116,
|
| 2001 |
+
"train_dice": 0.855368249937704,
|
| 2002 |
+
"train_iou": 0.7740828322662261,
|
| 2003 |
+
"grad_norm": 4.141527655512788,
|
| 2004 |
+
"lr": 5.869876628632226e-05,
|
| 2005 |
+
"encoder_lr": 1.4649630951960419e-05,
|
| 2006 |
+
"alpha": 0.2179034948348999,
|
| 2007 |
+
"validated_this_epoch": true,
|
| 2008 |
+
"val_loss": 0.6305578547939094,
|
| 2009 |
+
"val_dice": 0.7103913092200269,
|
| 2010 |
+
"val_iou": 0.6242352911416997,
|
| 2011 |
+
"val_reward": 9.485594560975749e-05,
|
| 2012 |
+
"val_entropy": 0.35002723893131876,
|
| 2013 |
+
"val_actor_loss": -0.009579253584355811,
|
| 2014 |
+
"val_critic_loss": 0.0001476603156862264,
|
| 2015 |
+
"val_ce_loss": 0.5297447526404127,
|
| 2016 |
+
"val_dice_loss": 0.7503817849086992
|
| 2017 |
+
},
|
| 2018 |
+
{
|
| 2019 |
+
"epoch": 91,
|
| 2020 |
+
"train_loss": 0.29263472519526185,
|
| 2021 |
+
"train_actor_loss": -0.07188433761532763,
|
| 2022 |
+
"train_critic_loss": 0.0001279856262998779,
|
| 2023 |
+
"train_mean_reward": 0.0005758833933040282,
|
| 2024 |
+
"train_entropy": 0.3308441816066278,
|
| 2025 |
+
"train_ce_loss": 0.3418424096564914,
|
| 2026 |
+
"train_dice_loss": 0.38706771848429694,
|
| 2027 |
+
"train_dice": 0.8560342560049871,
|
| 2028 |
+
"train_iou": 0.7730549597503144,
|
| 2029 |
+
"grad_norm": 5.002177060574525,
|
| 2030 |
+
"lr": 4.9507782751441774e-05,
|
| 2031 |
+
"encoder_lr": 1.3772099860690422e-05,
|
| 2032 |
+
"alpha": 0.2177559733390808,
|
| 2033 |
+
"validated_this_epoch": false,
|
| 2034 |
+
"val_loss": null,
|
| 2035 |
+
"val_dice": null,
|
| 2036 |
+
"val_iou": null,
|
| 2037 |
+
"val_reward": null,
|
| 2038 |
+
"val_entropy": null
|
| 2039 |
+
},
|
| 2040 |
+
{
|
| 2041 |
+
"epoch": 92,
|
| 2042 |
+
"train_loss": 0.2971994457549827,
|
| 2043 |
+
"train_actor_loss": -0.07344508408364246,
|
| 2044 |
+
"train_critic_loss": 0.00012346376983616908,
|
| 2045 |
+
"train_mean_reward": 0.0005349894414036712,
|
| 2046 |
+
"train_entropy": 0.330479318973938,
|
| 2047 |
+
"train_ce_loss": 0.34777834941211905,
|
| 2048 |
+
"train_dice_loss": 0.3933872364904222,
|
| 2049 |
+
"train_dice": 0.854655266964506,
|
| 2050 |
+
"train_iou": 0.7727474647247524,
|
| 2051 |
+
"grad_norm": 5.181883335344551,
|
| 2052 |
+
"lr": 4.12597546770122e-05,
|
| 2053 |
+
"encoder_lr": 1.298459969278006e-05,
|
| 2054 |
+
"alpha": 0.21763105690479279,
|
| 2055 |
+
"validated_this_epoch": false,
|
| 2056 |
+
"val_loss": null,
|
| 2057 |
+
"val_dice": null,
|
| 2058 |
+
"val_iou": null,
|
| 2059 |
+
"val_reward": null,
|
| 2060 |
+
"val_entropy": null
|
| 2061 |
+
},
|
| 2062 |
+
{
|
| 2063 |
+
"epoch": 93,
|
| 2064 |
+
"train_loss": 0.2884946270094997,
|
| 2065 |
+
"train_actor_loss": -0.07248325619451913,
|
| 2066 |
+
"train_critic_loss": 0.00012222201523214487,
|
| 2067 |
+
"train_mean_reward": 0.0006053991236611661,
|
| 2068 |
+
"train_entropy": 0.3279680321883478,
|
| 2069 |
+
"train_ce_loss": 0.3343253576453024,
|
| 2070 |
+
"train_dice_loss": 0.3875081747077232,
|
| 2071 |
+
"train_dice": 0.8620714876657734,
|
| 2072 |
+
"train_iou": 0.7795042783348156,
|
| 2073 |
+
"grad_norm": 4.025809176208437,
|
| 2074 |
+
"lr": 3.396282187094642e-05,
|
| 2075 |
+
"encoder_lr": 1.2287907615819006e-05,
|
| 2076 |
+
"alpha": 0.2178627997636795,
|
| 2077 |
+
"validated_this_epoch": false,
|
| 2078 |
+
"val_loss": null,
|
| 2079 |
+
"val_dice": null,
|
| 2080 |
+
"val_iou": null,
|
| 2081 |
+
"val_reward": null,
|
| 2082 |
+
"val_entropy": null
|
| 2083 |
+
},
|
| 2084 |
+
{
|
| 2085 |
+
"epoch": 94,
|
| 2086 |
+
"train_loss": 0.2755984605196141,
|
| 2087 |
+
"train_actor_loss": -0.07287909354754658,
|
| 2088 |
+
"train_critic_loss": 0.00012752900273838794,
|
| 2089 |
+
"train_mean_reward": 0.0005341678359809169,
|
| 2090 |
+
"train_entropy": 0.3300027222365372,
|
| 2091 |
+
"train_ce_loss": 0.33157425382215466,
|
| 2092 |
+
"train_dice_loss": 0.3652533154925019,
|
| 2093 |
+
"train_dice": 0.8627643439476325,
|
| 2094 |
+
"train_iou": 0.7813043197968654,
|
| 2095 |
+
"grad_norm": 3.359746343405672,
|
| 2096 |
+
"lr": 2.762418552495473e-05,
|
| 2097 |
+
"encoder_lr": 1.1682711180774574e-05,
|
| 2098 |
+
"alpha": 0.21779978275299072,
|
| 2099 |
+
"validated_this_epoch": false,
|
| 2100 |
+
"val_loss": null,
|
| 2101 |
+
"val_dice": null,
|
| 2102 |
+
"val_iou": null,
|
| 2103 |
+
"val_reward": null,
|
| 2104 |
+
"val_entropy": null
|
| 2105 |
+
},
|
| 2106 |
+
{
|
| 2107 |
+
"epoch": 95,
|
| 2108 |
+
"train_loss": 0.28890359234209206,
|
| 2109 |
+
"train_actor_loss": -0.07208383568978896,
|
| 2110 |
+
"train_critic_loss": 0.00012546290716364435,
|
| 2111 |
+
"train_mean_reward": 0.0005960843635657229,
|
| 2112 |
+
"train_entropy": 0.33177007557655797,
|
| 2113 |
+
"train_ce_loss": 0.3341652474195265,
|
| 2114 |
+
"train_dice_loss": 0.387684136013036,
|
| 2115 |
+
"train_dice": 0.8587547602725253,
|
| 2116 |
+
"train_iou": 0.7774633744039755,
|
| 2117 |
+
"grad_norm": 3.333088038965713,
|
| 2118 |
+
"lr": 2.2250101107838036e-05,
|
| 2119 |
+
"encoder_lr": 1.1169607643461924e-05,
|
| 2120 |
+
"alpha": 0.21766352653503418,
|
| 2121 |
+
"validated_this_epoch": false,
|
| 2122 |
+
"val_loss": null,
|
| 2123 |
+
"val_dice": null,
|
| 2124 |
+
"val_iou": null,
|
| 2125 |
+
"val_reward": null,
|
| 2126 |
+
"val_entropy": null
|
| 2127 |
+
},
|
| 2128 |
+
{
|
| 2129 |
+
"epoch": 96,
|
| 2130 |
+
"train_loss": 0.279721588091275,
|
| 2131 |
+
"train_actor_loss": -0.07310429053713186,
|
| 2132 |
+
"train_critic_loss": 0.00012310332280185437,
|
| 2133 |
+
"train_mean_reward": 0.0005285839674668641,
|
| 2134 |
+
"train_entropy": 0.3295354085375172,
|
| 2135 |
+
"train_ce_loss": 0.3328894537768883,
|
| 2136 |
+
"train_dice_loss": 0.3726391903189726,
|
| 2137 |
+
"train_dice": 0.8565933318114415,
|
| 2138 |
+
"train_iou": 0.7755106831343687,
|
| 2139 |
+
"grad_norm": 2.7869174799253775,
|
| 2140 |
+
"lr": 1.784587219209463e-05,
|
| 2141 |
+
"encoder_lr": 1.0749103375124614e-05,
|
| 2142 |
+
"alpha": 0.21750156581401825,
|
| 2143 |
+
"validated_this_epoch": false,
|
| 2144 |
+
"val_loss": null,
|
| 2145 |
+
"val_dice": null,
|
| 2146 |
+
"val_iou": null,
|
| 2147 |
+
"val_reward": null,
|
| 2148 |
+
"val_entropy": null
|
| 2149 |
+
},
|
| 2150 |
+
{
|
| 2151 |
+
"epoch": 97,
|
| 2152 |
+
"train_loss": 0.29764250348603655,
|
| 2153 |
+
"train_actor_loss": -0.07192213034849296,
|
| 2154 |
+
"train_critic_loss": 0.00012884712457910842,
|
| 2155 |
+
"train_mean_reward": 0.0003743555443407464,
|
| 2156 |
+
"train_entropy": 0.3317035970299743,
|
| 2157 |
+
"train_ce_loss": 0.3412867104201469,
|
| 2158 |
+
"train_dice_loss": 0.397713700468226,
|
| 2159 |
+
"train_dice": 0.8494988814389015,
|
| 2160 |
+
"train_iou": 0.7671673728170393,
|
| 2161 |
+
"grad_norm": 5.330061546822851,
|
| 2162 |
+
"lr": 1.4415845219935406e-05,
|
| 2163 |
+
"encoder_lr": 1.0421613362707402e-05,
|
| 2164 |
+
"alpha": 0.21721813082695007,
|
| 2165 |
+
"validated_this_epoch": false,
|
| 2166 |
+
"val_loss": null,
|
| 2167 |
+
"val_dice": null,
|
| 2168 |
+
"val_iou": null,
|
| 2169 |
+
"val_reward": null,
|
| 2170 |
+
"val_entropy": null
|
| 2171 |
+
},
|
| 2172 |
+
{
|
| 2173 |
+
"epoch": 98,
|
| 2174 |
+
"train_loss": 0.29285643487273494,
|
| 2175 |
+
"train_actor_loss": -0.0714600747138339,
|
| 2176 |
+
"train_critic_loss": 0.00012328618878712684,
|
| 2177 |
+
"train_mean_reward": 0.0005974431581484571,
|
| 2178 |
+
"train_entropy": 0.3283141048803194,
|
| 2179 |
+
"train_ce_loss": 0.34252902938086704,
|
| 2180 |
+
"train_dice_loss": 0.38598069340683694,
|
| 2181 |
+
"train_dice": 0.8530926981937806,
|
| 2182 |
+
"train_iou": 0.7751076994763964,
|
| 2183 |
+
"grad_norm": 2.9286488585693893,
|
| 2184 |
+
"lr": 1.19634052138698e-05,
|
| 2185 |
+
"encoder_lr": 1.0187460799314204e-05,
|
| 2186 |
+
"alpha": 0.2174145132303238,
|
| 2187 |
+
"validated_this_epoch": false,
|
| 2188 |
+
"val_loss": null,
|
| 2189 |
+
"val_dice": null,
|
| 2190 |
+
"val_iou": null,
|
| 2191 |
+
"val_reward": null,
|
| 2192 |
+
"val_entropy": null
|
| 2193 |
+
},
|
| 2194 |
+
{
|
| 2195 |
+
"epoch": 99,
|
| 2196 |
+
"train_loss": 0.2924680574039041,
|
| 2197 |
+
"train_actor_loss": -0.07318189177536996,
|
| 2198 |
+
"train_critic_loss": 0.00012567996098895322,
|
| 2199 |
+
"train_mean_reward": 0.0005259196532561034,
|
| 2200 |
+
"train_entropy": 0.33116844200194656,
|
| 2201 |
+
"train_ce_loss": 0.3427411302110753,
|
| 2202 |
+
"train_dice_loss": 0.3884330779237033,
|
| 2203 |
+
"train_dice": 0.8540698995510698,
|
| 2204 |
+
"train_iou": 0.7729114993130836,
|
| 2205 |
+
"grad_norm": 3.142038368901541,
|
| 2206 |
+
"lr": 1.049097243609706e-05,
|
| 2207 |
+
"encoder_lr": 1.0046876765255499e-05,
|
| 2208 |
+
"alpha": 0.217305526137352,
|
| 2209 |
+
"validated_this_epoch": false,
|
| 2210 |
+
"val_loss": null,
|
| 2211 |
+
"val_dice": null,
|
| 2212 |
+
"val_iou": null,
|
| 2213 |
+
"val_reward": null,
|
| 2214 |
+
"val_entropy": null
|
| 2215 |
+
},
|
| 2216 |
+
{
|
| 2217 |
+
"epoch": 100,
|
| 2218 |
+
"train_loss": 0.2949535176843809,
|
| 2219 |
+
"train_actor_loss": -0.07229916730688676,
|
| 2220 |
+
"train_critic_loss": 0.0001259548047668201,
|
| 2221 |
+
"train_mean_reward": 0.0006596473387501996,
|
| 2222 |
+
"train_entropy": 0.3281508139515108,
|
| 2223 |
+
"train_ce_loss": 0.34345720693991266,
|
| 2224 |
+
"train_dice_loss": 0.3909221970757773,
|
| 2225 |
+
"train_dice": 0.8563173362460633,
|
| 2226 |
+
"train_iou": 0.7756491191360922,
|
| 2227 |
+
"grad_norm": 3.9696367817331653,
|
| 2228 |
+
"lr": 1e-05,
|
| 2229 |
+
"encoder_lr": 1e-05,
|
| 2230 |
+
"alpha": 0.2174254059791565,
|
| 2231 |
+
"validated_this_epoch": true,
|
| 2232 |
+
"val_loss": 0.6225139721063897,
|
| 2233 |
+
"val_dice": 0.7154367071264605,
|
| 2234 |
+
"val_iou": 0.6284476085254171,
|
| 2235 |
+
"val_reward": 9.7928330368898e-05,
|
| 2236 |
+
"val_entropy": 0.3410583541850851,
|
| 2237 |
+
"val_actor_loss": -0.008168109109968588,
|
| 2238 |
+
"val_critic_loss": 0.00013335745697159754,
|
| 2239 |
+
"val_ce_loss": 0.5315143069689812,
|
| 2240 |
+
"val_dice_loss": 0.7297164722524508
|
| 2241 |
+
}
|
| 2242 |
+
]
|