lewaldm commited on
Commit
f66034b
·
verified ·
1 Parent(s): c070ba9

Upload 2 files

Browse files
Files changed (2) hide show
  1. config.yaml +236 -0
  2. model_final.pth +3 -0
config.yaml ADDED
@@ -0,0 +1,236 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ MODEL:
2
+ WEIGHTS: /hkfs/work/workspace/scratch/tum_mhj8661-panopticon/dino_logs/abl_a3/v4_16gpu/lr=5e-4_e70b-1_warmup=5/lr=5e-4_warmup=5_hold_epochs=0_min_lr=1e-6_frz=last_layer=1_break_at_epochs=-1/model_final.pth
3
+ pretrained_weights: []
4
+ WEIGHTS_ONLY: ''
5
+ compute_precision:
6
+ grad_scaler: true
7
+ teacher:
8
+ backbone:
9
+ sharding_strategy: NO_SHARD
10
+ mixed_precision:
11
+ param_dtype: fp16
12
+ reduce_dtype: fp16
13
+ buffer_dtype: fp32
14
+ dino_head:
15
+ sharding_strategy: NO_SHARD
16
+ mixed_precision:
17
+ param_dtype: fp16
18
+ reduce_dtype: fp16
19
+ buffer_dtype: fp32
20
+ ibot_head:
21
+ sharding_strategy: NO_SHARD
22
+ mixed_precision:
23
+ param_dtype: fp16
24
+ reduce_dtype: fp16
25
+ buffer_dtype: fp32
26
+ student:
27
+ backbone:
28
+ sharding_strategy: NO_SHARD
29
+ mixed_precision:
30
+ param_dtype: fp16
31
+ reduce_dtype: fp16
32
+ buffer_dtype: fp32
33
+ dino_head:
34
+ sharding_strategy: NO_SHARD
35
+ mixed_precision:
36
+ param_dtype: fp16
37
+ reduce_dtype: fp32
38
+ buffer_dtype: fp32
39
+ ibot_head:
40
+ sharding_strategy: NO_SHARD
41
+ mixed_precision:
42
+ param_dtype: fp16
43
+ reduce_dtype: fp32
44
+ buffer_dtype: fp32
45
+ dino:
46
+ loss_weight: 1.0
47
+ head_n_prototypes: 65536
48
+ head_bottleneck_dim: 256
49
+ head_nlayers: 3
50
+ head_hidden_dim: 2048
51
+ koleo_loss_weight: 0.0
52
+ ibot:
53
+ loss_weight: 0.1
54
+ mask_sample_probability: 0.5
55
+ mask_ratio_min_max:
56
+ - 0.1
57
+ - 0.5
58
+ separate_head: false
59
+ head_n_prototypes: 65536
60
+ head_bottleneck_dim: 256
61
+ head_nlayers: 3
62
+ head_hidden_dim: 2048
63
+ pe_distil:
64
+ loss_weight: 0.0
65
+ aux_loss:
66
+ enable: false
67
+ gating_loss_weight: 0.0
68
+ diversity_loss_weight: 0.0
69
+ sparsity_loss_weight: 0.0
70
+ gradient_accumulation_steps: 1
71
+ train:
72
+ dataset:
73
+ id: ConcatDataset
74
+ datasets:
75
+ - id: FmowDataset
76
+ root: /hkfs/work/workspace/scratch/tum_mhj8661-panopticon/datasets/
77
+ split: /hkfs/work/workspace/scratch/tum_mhj8661-panopticon/datasets/fmow/metadata_v2/fmow_iwm_onid_train_val_savioclean_min42max1024.parquet
78
+ num_sens: 2
79
+ full_spectra: false
80
+ subset: 800000
81
+ - id: SatlasDataset
82
+ root: /hkfs/work/workspace/scratch/tum_mhj8661-panopticon/datasets/satlas
83
+ num_sens: 2
84
+ full_spectra: false
85
+ metadata_path: /hkfs/work/workspace/scratch/tum_mhj8661-panopticon/datasets/satlas/metadata_v2/fmow_iwm_onid_3sensors_all_clean.parquet
86
+ - id: MMEarth
87
+ base_path: /hkfs/work/workspace/scratch/tum_mhj8661-panopticon/datasets/mmearth/data_1M_v001
88
+ split: train
89
+ modalities: MODALITY_MINIMAL_SET1
90
+ full_spectra: false
91
+ - id: SpectralEarth
92
+ split: train
93
+ root: /hkfs/work/workspace/scratch/tum_mhj8661-panopticon/datasets/
94
+ full_spectra: false
95
+ dino_augm:
96
+ id: ChnSpatialAugmentationV4
97
+ global_crops_number: 2
98
+ global_crops_size: 224
99
+ global_crops_scale:
100
+ - 0.32
101
+ - 1.0
102
+ global_crops_spectral_size:
103
+ - 1.0
104
+ - 4
105
+ - 13
106
+ global_modes_probs:
107
+ - 1
108
+ - 0
109
+ - 0
110
+ - 0
111
+ global_multi_select_view_in_single_sensor_modes: false
112
+ global_hs_modes_probs:
113
+ - 1
114
+ - 0
115
+ - 0
116
+ - 0
117
+ - 0
118
+ global_hs_multi_select_view_in_single_sensor_modes: true
119
+ local_crops_number: 4
120
+ local_crops_size: 98
121
+ local_crops_scale:
122
+ - 0.05
123
+ - 0.32
124
+ local_crops_spectral_size:
125
+ - 1.0
126
+ - 1
127
+ - 4
128
+ local_modes_probs:
129
+ - 1
130
+ - 0
131
+ - 0
132
+ - 0
133
+ local_multi_select_view_in_single_sensor_modes: true
134
+ local_hs_modes_probs:
135
+ - 1
136
+ - 0
137
+ - 0
138
+ - 0
139
+ - 0
140
+ local_hs_multi_select_view_in_single_sensor_modes: true
141
+ batch_size_per_gpu: 75
142
+ num_workers: 16
143
+ drop_last: true
144
+ pin_memory: true
145
+ persistent_workers: true
146
+ use_wandb: true
147
+ output_dir: /hkfs/work/workspace/scratch/tum_mhj8661-panopticon/dino_logs/abl_a3_s2/v9_16gpu/ds=fmow-800k^satl^mme^spee_lr=5e-4_augm=v4_c13_e=70_model=stg1_16gpu/lr=5e-4_warmup=5_hold_epochs=0_min_lr=1e-6_frz=last_layer=1_break_at_epochs=-1_bsz=75
148
+ saveckp_freq: 5
149
+ seed: 21
150
+ OFFICIAL_EPOCH_LENGTH: 1250
151
+ centering: centering
152
+ log_every_n_steps: 10
153
+ eff_bsz: 75
154
+ short_exp_name: abl_a3_s2/v9_16gpu/ds=fmow-800k^satl^mme^spee_lr=5e-4_augm=v4_c13_e=70_model=stg1_16gpu
155
+ student:
156
+ arch: vit_base
157
+ embed_layer: ChnFusionPE2
158
+ patch_size: 14
159
+ pos_emb_img_size: 518
160
+ drop_path_rate: 0.0
161
+ layerscale: 1.0e-05
162
+ drop_path_uniform: true
163
+ pretrained_weights: []
164
+ ffn_layer: mlp
165
+ block_chunks: 0
166
+ qkv_bias: true
167
+ proj_bias: true
168
+ ffn_bias: true
169
+ num_register_tokens: 0
170
+ interpolate_antialias: false
171
+ interpolate_offset: 0.1
172
+ in_chans: 3
173
+ pe_args:
174
+ tokenizer_type: conv3d
175
+ before_proj_id: 0
176
+ after_proj_id: -1
177
+ in_dim: 2304
178
+ chnfus_cfg:
179
+ id: ChnAttn
180
+ query_mode: standard
181
+ layer_norm: false
182
+ chnemb_cfg:
183
+ id: ChnEmbV3
184
+ attn_cfg:
185
+ num_heads: 16
186
+ s_attn_cfg:
187
+ num_blocks: 0
188
+ mode: mix_chns
189
+ num_heads: 4
190
+ teacher:
191
+ pretrained_weights: []
192
+ momentum_teacher: 0.994
193
+ final_momentum_teacher: 1.0
194
+ warmup_teacher_temp: 0.04
195
+ teacher_temp: 0.07
196
+ warmup_teacher_temp_epochs: 5
197
+ optim:
198
+ epochs: 70
199
+ weight_decay: 0.04
200
+ weight_decay_end: 0.2
201
+ base_lr: 0.0005
202
+ lr: 0.0005412658773652743
203
+ warmup_epochs: 5
204
+ min_lr: 1.0e-06
205
+ lr_multiplier: blocks=0.2
206
+ freeze_weights: last_layer=1
207
+ scaling_rule: sqrt_wrt_1024
208
+ clip_grad: 3.0
209
+ layerwise_decay: 0.9
210
+ adamw_beta1: 0.9
211
+ adamw_beta2: 0.999
212
+ break_at_epochs: -1
213
+ online_lr_batch_scaling: false
214
+ hold_epochs: 0
215
+ eval:
216
+ only_eval: false
217
+ skip: false
218
+ config_obj: /home/hk-project-pai00028/tum_mhj8661/code/PanOpticOn/dinov2/configs/eval/oe
219
+ overwrite: false
220
+ overwrites:
221
+ optim:
222
+ dl:
223
+ batch_size: 200
224
+ num_workers: 8
225
+ eval_period_epoch: 2
226
+ eval_period_iterations: -1
227
+ remove_ckpts: false
228
+ include_final_ckpt: true
229
+ final_model:
230
+ config_obj: /home/hk-project-pai00028/tum_mhj8661/code/PanOpticOn/dinov2/configs/eval/oe;/home/hk-project-pai00028/tum_mhj8661/code/PanOpticOn/dinov2/configs/eval/fm/
231
+ overwrite: false
232
+ overwrites: null
233
+ use_online_eval: false
234
+ add_args: true
235
+ tmpfile: null
236
+ slurm_job_id: '2957504'
model_final.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a99b3899c5d3f22ea58dc3e52669ff0a828e62a97f86b759cb8634eeaaf4ae53
3
+ size 1954696007