yvokeller commited on
Commit
860e12e
·
verified ·
1 Parent(s): c566d26

Messis | W&B Run exp-7-label-hierarchy-zueri_crop-0 (https://wandb.ai/crop-classification/messis/runs/0nk545h7)

Browse files
Files changed (2) hide show
  1. config.json +24 -14
  2. model.safetensors +2 -2
config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "hparams": {
3
- "accumulate_grad_batches": 4,
4
  "backbone_weights_path": "./prithvi/models/Prithvi_100M.pt",
5
  "bands": [
6
  0,
@@ -10,7 +10,7 @@
10
  4,
11
  5
12
  ],
13
- "batch_size": 4,
14
  "bottleneck_reduction_factor": 8,
15
  "data_augmentation": {
16
  "enabled": false,
@@ -19,34 +19,44 @@
19
  },
20
  "debug": false,
21
  "dropout_p": 0.1,
22
- "early_stopping_metric": "val_f1_tier2_majority",
23
  "early_stopping_mode": "max",
24
  "early_stopping_patience": 10,
25
  "experiment_group": "exp-7-label-hierarchy",
26
- "experiment_name": "exp-7-label-hierarchy-seasonality-reduced",
27
  "freeze_backbone": true,
28
  "heads_spec": {
29
  "tier1": {
30
  "is_metrics_tier": true,
 
31
  "loss_weight": 1,
32
- "num_classes_to_predict": 4,
33
  "target_idx": 0,
34
  "type": "HierarchicalFCNHead"
35
  },
36
  "tier2": {
37
- "is_last_tier": true,
38
  "is_metrics_tier": true,
 
39
  "loss_weight": 1,
40
- "num_classes_to_predict": 25,
41
  "target_idx": 1,
42
  "type": "HierarchicalFCNHead"
43
  },
44
- "tier2_refinement_head": {
 
 
 
 
 
 
 
 
 
45
  "is_final_head": true,
46
  "is_metrics_tier": false,
47
  "loss_weight": 1,
48
- "num_classes_to_predict": 25,
49
- "target_idx": 1,
50
  "type": "LabelRefinementHead"
51
  }
52
  },
@@ -54,7 +64,7 @@
54
  "loss_ignore_background": false,
55
  "lr": 0.001,
56
  "max_epochs": 400,
57
- "name": "seasonality-reduced",
58
  "num_frames": 3,
59
  "optimizer": "Adam",
60
  "optimizer_momentum": null,
@@ -64,14 +74,14 @@
64
  5
65
  ],
66
  "train_folds": [
67
- 0,
68
  1,
69
  2,
70
- 3
 
71
  ],
72
  "use_bottleneck_neck": false,
73
  "val_folds": [
74
- 4
75
  ]
76
  }
77
  }
 
1
  {
2
  "hparams": {
3
+ "accumulate_grad_batches": 1,
4
  "backbone_weights_path": "./prithvi/models/Prithvi_100M.pt",
5
  "bands": [
6
  0,
 
10
  4,
11
  5
12
  ],
13
+ "batch_size": 16,
14
  "bottleneck_reduction_factor": 8,
15
  "data_augmentation": {
16
  "enabled": false,
 
19
  },
20
  "debug": false,
21
  "dropout_p": 0.1,
22
+ "early_stopping_metric": "val_f1_tier3_majority",
23
  "early_stopping_mode": "max",
24
  "early_stopping_patience": 10,
25
  "experiment_group": "exp-7-label-hierarchy",
26
+ "experiment_name": "exp-7-label-hierarchy-zueri_crop",
27
  "freeze_backbone": true,
28
  "heads_spec": {
29
  "tier1": {
30
  "is_metrics_tier": true,
31
+ "kernel_size": 3,
32
  "loss_weight": 1,
33
+ "num_classes_to_predict": 6,
34
  "target_idx": 0,
35
  "type": "HierarchicalFCNHead"
36
  },
37
  "tier2": {
 
38
  "is_metrics_tier": true,
39
+ "kernel_size": 3,
40
  "loss_weight": 1,
41
+ "num_classes_to_predict": 17,
42
  "target_idx": 1,
43
  "type": "HierarchicalFCNHead"
44
  },
45
+ "tier3": {
46
+ "is_last_tier": true,
47
+ "is_metrics_tier": true,
48
+ "kernel_size": 3,
49
+ "loss_weight": 1,
50
+ "num_classes_to_predict": 49,
51
+ "target_idx": 2,
52
+ "type": "HierarchicalFCNHead"
53
+ },
54
+ "tier3_refinement_head": {
55
  "is_final_head": true,
56
  "is_metrics_tier": false,
57
  "loss_weight": 1,
58
+ "num_classes_to_predict": 49,
59
+ "target_idx": 2,
60
  "type": "LabelRefinementHead"
61
  }
62
  },
 
64
  "loss_ignore_background": false,
65
  "lr": 0.001,
66
  "max_epochs": 400,
67
+ "name": "zueri_crop",
68
  "num_frames": 3,
69
  "optimizer": "Adam",
70
  "optimizer_momentum": null,
 
74
  5
75
  ],
76
  "train_folds": [
 
77
  1,
78
  2,
79
+ 3,
80
+ 4
81
  ],
82
  "use_bottleneck_neck": false,
83
  "val_folds": [
84
+ 0
85
  ]
86
  }
87
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d2c74695915aa55395b7a04b746bc45ad20a4c3d24023099341a796cadf2122c
3
- size 711438096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32c2e9762a2b38dff765239888521a5f9bab1f20807b3441a86476c95b8ce214
3
+ size 713880940