dgcnz commited on
Commit
f1d1ce2
·
verified ·
1 Parent(s): de12f15

Upload folder using huggingface_hub

Browse files
outputs/2025-06-22/19-16-53/.hydra/config.yaml ADDED
@@ -0,0 +1,187 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ extras:
2
+ ignore_warnings: false
3
+ enforce_tags: true
4
+ print_config: true
5
+ theme: ansi_light
6
+ task_name: pretrain
7
+ tags:
8
+ - pretrain
9
+ - imagenet1k
10
+ - partmae_v6
11
+ - vit_b_16
12
+ ckpt_path: null
13
+ seed: 12345
14
+ trainer:
15
+ accelerator: gpu
16
+ devices: 4
17
+ num_nodes: 1
18
+ precision: bf16-mixed
19
+ strategy: ddp
20
+ max_epochs: 200
21
+ min_epochs: null
22
+ accumulate_grad_batches: ${eval:${total_batch_size_per_gpu} // ${train_dataloader.batch_size}}
23
+ gradient_clip_val: 0.3
24
+ paths:
25
+ root_dir: ${oc.env:PROJECT_ROOT}
26
+ log_dir: ${paths.root_dir}/logs/
27
+ output_dir: ${hydra:runtime.output_dir}
28
+ work_dir: ${hydra:runtime.cwd}
29
+ run:
30
+ dir: ${paths.log_dir}/${task_name}/runs/${now:%Y-%m-%d}_${now:%H-%M-%S}
31
+ sweep:
32
+ dir: ${paths.log_dir}/${task_name}/multiruns/${now:%Y-%m-%d}_${now:%H-%M-%S}
33
+ subdir: ${hydra:job.num}
34
+ job_logging:
35
+ handlers:
36
+ file:
37
+ filename: ${hydra:runtime.output_dir}/${task_name}.log
38
+ data:
39
+ transform:
40
+ _target_: src.data.components.transforms.multi_crop_v4.ParametrizedMultiCropV4
41
+ canonical_size: 512
42
+ canonical_crop_scale:
43
+ - 0.9
44
+ - 1.0
45
+ global_crops_scale:
46
+ - 0.3
47
+ - 1.0
48
+ local_crops_scale:
49
+ - 0.05
50
+ - 0.3
51
+ n_global_crops: 2
52
+ n_local_crops: 10
53
+ distort_color: true
54
+ _target_: src.data.components.image_folder.ImageFolderNoLabels
55
+ root: /scratch-nvme/ml-datasets/imagenet/torchvision_ImageFolder/train
56
+ model:
57
+ _target_: src.models.components.partmae_v6.PARTMaskedAutoEncoderViT
58
+ img_size: 224
59
+ canonical_img_size: 512
60
+ local_img_size: 96
61
+ max_scale_ratio: 6.0
62
+ patch_size: 16
63
+ in_chans: 3
64
+ embed_dim: 768
65
+ depth: 12
66
+ num_heads: 12
67
+ mlp_ratio: 4
68
+ norm_layer:
69
+ _target_: torch.nn.LayerNorm
70
+ eps: 1.0e-06
71
+ _partial_: true
72
+ mask_ratio: 0.75
73
+ pos_mask_ratio: 0.75
74
+ num_views: ${eval:${data.transform.n_global_crops} + ${data.transform.n_local_crops}}
75
+ decoder_embed_dim: 512
76
+ decoder_depth: 8
77
+ decoder_num_heads: 16
78
+ proj_bottleneck_dim: 256
79
+ apply_tanh: false
80
+ sampler: stratified_jittered
81
+ lambda_pose: 0.111
82
+ lambda_psmooth: 0
83
+ lambda_pstress: 0
84
+ lambda_pcr: 0
85
+ lambda_ccr: 0.333
86
+ lambda_cinv: 0.556
87
+ criterion: l1
88
+ alpha_t: 0.5
89
+ alpha_ts: 0.8
90
+ alpha_s: 1.0
91
+ sigma_yx: 0.2
92
+ sigma_hw: 1.0
93
+ cr_eps: 0.1
94
+ cos_eps: 1.0e-08
95
+ pos_embed_mode: learn
96
+ ls_init_values: null
97
+ num_register_tokens: 0
98
+ lambda_pmatch: 0
99
+ beta_f: 0.1
100
+ beta_w: 3.0
101
+ bypass_loss: true
102
+ uncertainty_mode: correlated_proj
103
+ gate_dim: 32
104
+ tau: 0.07
105
+ decoder_from_proj: true
106
+ callbacks:
107
+ checkpoint:
108
+ _target_: src.callbacks.common.checkpoint.ModelCheckpoint
109
+ dirpath: ${paths.output_dir}
110
+ every_n_epochs: 25
111
+ save_last: true
112
+ verbose: true
113
+ metric_logger:
114
+ _target_: src.callbacks.common.metric_logger.MetricLogger
115
+ every_n_steps: 25
116
+ iter_timer:
117
+ _target_: src.callbacks.common.iter_timer.IterTimer
118
+ every_n_steps: 25
119
+ logger:
120
+ wandb:
121
+ _target_: wandb.integration.lightning.fabric.WandbLogger
122
+ save_dir: ${paths.output_dir}
123
+ offline: false
124
+ id: null
125
+ anonymous: null
126
+ project: PART-pretrain
127
+ log_model: false
128
+ prefix: ''
129
+ group: imagenet1k/partmae_v6/vit_b_16
130
+ tags: ${tags}
131
+ job_type: ''
132
+ scheduler:
133
+ _target_: timm.scheduler.cosine_lr.CosineLRScheduler
134
+ _partial_: true
135
+ t_initial: ${eval:${trainer.max_epochs} * ${steps_per_epoch}}
136
+ cycle_mul: 1.0
137
+ lr_min: 1.0e-07
138
+ cycle_decay: 0.1
139
+ warmup_lr_init: 1.0e-07
140
+ warmup_t: ${eval:${warmup_epochs} * ${steps_per_epoch}}
141
+ cycle_limit: 1
142
+ t_in_epochs: false
143
+ noise_range_t: null
144
+ noise_pct: 0.67
145
+ noise_std: 1.0
146
+ noise_seed: 42
147
+ optimizer:
148
+ _target_: timm.optim.create_optimizer_v2
149
+ _partial_: true
150
+ opt: adamw
151
+ lr: ${eval:${blr} * ${total_batch_size} / 256}
152
+ weight_decay: 0.05
153
+ filter_bias_and_bn: true
154
+ param_group_fn:
155
+ _target_: src.utils.optimizer.make_param_group_fn
156
+ base_lr: ${..lr}
157
+ weight_decay: ${..weight_decay}
158
+ filter_bias_and_bn: ${..filter_bias_and_bn}
159
+ param_groups:
160
+ dino_head: ${eval:${blr} * ${total_batch_size} / 256}
161
+ decoder: ${eval:${blr} * ${total_batch_size} / 256}
162
+ pose_head: ${eval:${blr} * ${total_batch_size} / 256}
163
+ float32_matmul_precision: high
164
+ cudnn_benchmark: true
165
+ total_images: 1281167
166
+ total_batch_size: 2048
167
+ total_batch_size_per_gpu: ${eval:${total_batch_size} // ${trainer.devices}}
168
+ steps_per_epoch: ${eval:${total_images} // ${total_batch_size}}
169
+ blr: 5.0e-05
170
+ train_dataloader:
171
+ pin_memory: true
172
+ batch_size: 512
173
+ num_workers: 16
174
+ drop_last: true
175
+ shuffle: true
176
+ metric_collection:
177
+ _target_: src.models.components.metrics.partmae_v6.V6Metrics
178
+ nan_strategy: disable
179
+ compile_kwargs:
180
+ fullgraph: true
181
+ mode: default
182
+ compile_expr:
183
+ torch._subclasses.fake_tensor.CONSTANT_NUMEL_LIMIT: 100000
184
+ torch._dynamo.config.optimize_ddp: python_reducer
185
+ torch._dynamo.config.compiled_autograd: true
186
+ compile: false
187
+ warmup_epochs: 10
outputs/2025-06-22/19-16-53/.hydra/hydra.yaml ADDED
@@ -0,0 +1,166 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - experiment=pretrain/in1k/partmae_v6/vit_b_16/h100
116
+ job:
117
+ name: main_pretrain
118
+ chdir: null
119
+ override_dirname: experiment=pretrain/in1k/partmae_v6/vit_b_16/h100
120
+ id: ???
121
+ num: ???
122
+ config_name: pretrain.yaml
123
+ env_set: {}
124
+ env_copy: []
125
+ config:
126
+ override_dirname:
127
+ kv_sep: '='
128
+ item_sep: ','
129
+ exclude_keys: []
130
+ runtime:
131
+ version: 1.4.0.dev1
132
+ version_base: '1.3'
133
+ cwd: /gpfs/home2/dcanez/development/PART
134
+ config_sources:
135
+ - path: hydra.conf
136
+ schema: pkg
137
+ provider: hydra
138
+ - path: /gpfs/home2/dcanez/development/PART/fabric_configs
139
+ schema: file
140
+ provider: main
141
+ - path: hydra_plugins.hydra_colorlog.conf
142
+ schema: pkg
143
+ provider: hydra-colorlog
144
+ - path: ''
145
+ schema: structured
146
+ provider: schema
147
+ output_dir: /gpfs/home2/dcanez/development/PART/outputs/2025-06-22/19-16-53
148
+ choices:
149
+ metric_logger: null
150
+ experiment: pretrain/in1k/partmae_v6/vit_b_16/h100
151
+ optimizer: adamw
152
+ scheduler: cosine
153
+ logger: wandb
154
+ model: pretrain/partmae_v6/vit_b_16_224
155
+ data: in1k_snellius
156
+ data/transform: multi_crop_v4
157
+ hydra/env: default
158
+ hydra/callbacks: null
159
+ hydra/job_logging: default
160
+ hydra/hydra_logging: default
161
+ hydra/hydra_help: default
162
+ hydra/help: default
163
+ hydra/sweeper: basic
164
+ hydra/launcher: basic
165
+ hydra/output: default
166
+ verbose: false
outputs/2025-06-22/19-16-53/.hydra/overrides.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ - experiment=pretrain/in1k/partmae_v6/vit_b_16/h100