Knowing commited on
Commit
0304ae5
·
verified ·
1 Parent(s): 3047aba

Upload folder using huggingface_hub

Browse files
Files changed (41) hide show
  1. checkpoints/1028_DL3DV_FULL_U/.hydra/config.yaml +209 -0
  2. checkpoints/1028_DL3DV_FULL_U/.hydra/hydra.yaml +168 -0
  3. checkpoints/1028_DL3DV_FULL_U/.hydra/overrides.yaml +8 -0
  4. checkpoints/1028_DL3DV_FULL_U/main.log +46 -0
  5. checkpoints/1028_DL3DV_FULL_U/train_ddp_process_1.log +18 -0
  6. checkpoints/1028_DL3DV_FULL_U/train_ddp_process_2.log +18 -0
  7. checkpoints/1028_DL3DV_FULL_U/train_ddp_process_3.log +18 -0
  8. checkpoints/1028_DL3DV_FULL_U/train_ddp_process_4.log +18 -0
  9. checkpoints/1028_DL3DV_FULL_U/train_ddp_process_5.log +18 -0
  10. checkpoints/1028_DL3DV_FULL_U/train_ddp_process_6.log +18 -0
  11. checkpoints/1028_DL3DV_FULL_U/train_ddp_process_7.log +18 -0
  12. checkpoints/1028_context/.hydra/config.yaml +209 -0
  13. checkpoints/1028_context/.hydra/hydra.yaml +169 -0
  14. checkpoints/1028_context/.hydra/overrides.yaml +9 -0
  15. checkpoints/1028_context/checkpoints/epoch_4-step_18750.ckpt +3 -0
  16. checkpoints/1028_context/main.log +233 -0
  17. checkpoints/1028_context/train_ddp_process_1.log +93 -0
  18. checkpoints/1028_target/.hydra/config.yaml +209 -0
  19. checkpoints/1028_target/.hydra/hydra.yaml +169 -0
  20. checkpoints/1028_target/.hydra/overrides.yaml +9 -0
  21. checkpoints/1028_target/checkpoints/epoch_4-step_18750.ckpt +3 -0
  22. checkpoints/1028_target/main.log +167 -0
  23. checkpoints/1028_target/train_ddp_process_1.log +83 -0
  24. checkpoints/1028_target_segment1/.hydra/config.yaml +209 -0
  25. checkpoints/1028_target_segment1/.hydra/hydra.yaml +170 -0
  26. checkpoints/1028_target_segment1/.hydra/overrides.yaml +10 -0
  27. checkpoints/1028_target_segment1/checkpoints/epoch_4-step_18750.ckpt +3 -0
  28. checkpoints/1028_target_segment1/main.log +640 -0
  29. checkpoints/1028_target_segment1/train_ddp_process_1.log +143 -0
  30. checkpoints/1028_target_segment1/train_ddp_process_2.log +10 -0
  31. checkpoints/1028_target_segment1/train_ddp_process_3.log +10 -0
  32. checkpoints/1028_target_segment1/train_ddp_process_4.log +10 -0
  33. checkpoints/1028_target_segment1/train_ddp_process_5.log +10 -0
  34. checkpoints/1028_target_segment1/train_ddp_process_6.log +10 -0
  35. checkpoints/1028_target_segment1/train_ddp_process_7.log +10 -0
  36. checkpoints/1028_target_segment2/.hydra/config.yaml +209 -0
  37. checkpoints/1028_target_segment2/.hydra/hydra.yaml +170 -0
  38. checkpoints/1028_target_segment2/.hydra/overrides.yaml +10 -0
  39. checkpoints/1028_target_segment2/checkpoints/epoch_4-step_18750.ckpt +3 -0
  40. checkpoints/1028_target_segment2/main.log +167 -0
  41. checkpoints/1028_target_segment2/train_ddp_process_1.log +83 -0
checkpoints/1028_DL3DV_FULL_U/.hydra/config.yaml ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ encoder:
3
+ name: dcsplat
4
+ input_image_shape:
5
+ - 518
6
+ - 518
7
+ head_mode: pcd
8
+ same_head: false
9
+ num_level: 3
10
+ gs_param_dim: 256
11
+ camera_prompt: false
12
+ num_visual_prompt: 0
13
+ use_extrinsic_token: false
14
+ use_intrinsic_token: false
15
+ extrinsic_drop: 0.1
16
+ intrinsic_drop: 0.1
17
+ use_pnp_solver: false
18
+ align_corners: false
19
+ image_scale: 1.0
20
+ use_uncertainty_token: true
21
+ gs_bf16: true
22
+ aggregator_bf16: false
23
+ img_num_frame_token_N: 0
24
+ img_num_global_token_N: 0
25
+ use_gs_param_img_num_embed: true
26
+ use_mean_param_img_num_embed: true
27
+ decoder:
28
+ name: splatting_cuda
29
+ background_color:
30
+ - 0.0
31
+ - 0.0
32
+ - 0.0
33
+ make_scale_invariant: false
34
+ use_gsplat: true
35
+ density_control:
36
+ name: density_control_module
37
+ mean_dim: 32
38
+ gs_param_dim: 256
39
+ refinement_layer_num: 1
40
+ num_level: 3
41
+ grad_mode: absgrad
42
+ use_mean_features: true
43
+ refinement_type: quadtree
44
+ refinement_hidden_dim: 32
45
+ aggregation_mode: mean
46
+ num_heads: 1
47
+ score_mode: 3
48
+ sigmoid_score: true
49
+ latent_dim: 128
50
+ num_latents: 64
51
+ num_self_attn_per_block: 2
52
+ render_loss:
53
+ mse:
54
+ weight: 1.0
55
+ lpips:
56
+ weight: 0.05
57
+ apply_after_step: 0
58
+ density_control_loss:
59
+ complex_score:
60
+ weight: 0.01
61
+ log_scale: false
62
+ grad_scale: 5000.0
63
+ mode: 1
64
+ direct_loss:
65
+ l1:
66
+ weight: 0.8
67
+ ssim:
68
+ weight: 0.2
69
+ wandb:
70
+ project: DCSplat
71
+ entity: scene-representation-group
72
+ name: 1028_DL3DV_FULL_U
73
+ mode: online
74
+ tags:
75
+ - dl3dv
76
+ - 224x448
77
+ - full
78
+ mode: train
79
+ data_loader:
80
+ train:
81
+ num_workers: 16
82
+ persistent_workers: true
83
+ batch_size: 4
84
+ seed: 1234
85
+ test:
86
+ num_workers: 4
87
+ persistent_workers: false
88
+ batch_size: 1
89
+ seed: 2345
90
+ val:
91
+ num_workers: 1
92
+ persistent_workers: true
93
+ batch_size: 1
94
+ seed: 3456
95
+ optimizer:
96
+ lr: 0.0002
97
+ warm_up_steps: 125
98
+ backbone_lr_multiplier: 0.1
99
+ backbone_trainable: GB+H
100
+ accumulate: 1
101
+ checkpointing:
102
+ load: null
103
+ every_n_train_steps: 9375
104
+ save_top_k: 1
105
+ save_weights_only: true
106
+ train:
107
+ extended_visualization: false
108
+ print_log_every_n_steps: 10
109
+ camera_loss: 1.0
110
+ one_sample_validation: datasets/dl3dv/room_batch.pt
111
+ align_corners: false
112
+ intrinsic_scaling: true
113
+ train_context: true
114
+ train_hybrid: false
115
+ use_pred_point_focal: false
116
+ scaling_detach: true
117
+ complex_threshold_range:
118
+ - 0.0
119
+ - 3.0
120
+ use_umeyama: false
121
+ pred_focal_cam_loss: false
122
+ direct_loss_weight: 0.0
123
+ direct_iteration_steps: 0
124
+ twice_backward: false
125
+ use_uncertainty_camera_weight: true
126
+ use_target_uncertainty: point_similarity_mult
127
+ verbose: false
128
+ complex_random_mode: num_gs_mode3
129
+ start_density_control: 0.0
130
+ base_complex_exp_rate: 4.0
131
+ use_refine_aux: false
132
+ train_extrinsic: true
133
+ test:
134
+ output_path: test/full/dl3dv
135
+ align_pose: false
136
+ pose_align_steps: 100
137
+ rot_opt_lr: 0.005
138
+ trans_opt_lr: 0.005
139
+ compute_scores: true
140
+ save_image: false
141
+ save_video: false
142
+ save_active_mask_image: false
143
+ save_complex_score_image: false
144
+ save_compare: false
145
+ pred_intrinsic: true
146
+ complex_threshold: 0.4
147
+ complex_threshold_list:
148
+ - 0.0
149
+ - 0.2
150
+ - 0.4
151
+ - 0.6
152
+ - 0.8
153
+ threshold_mode: ratio
154
+ nvs_view_N_list:
155
+ - 3
156
+ - 6
157
+ - 16
158
+ - 32
159
+ - 64
160
+ seed: 111123
161
+ trainer:
162
+ max_steps: 18751
163
+ val_check_interval: 500
164
+ gradient_clip_val: 0.5
165
+ num_nodes: 1
166
+ dataset:
167
+ dl3dv:
168
+ make_baseline_1: false
169
+ relative_pose: true
170
+ augment: true
171
+ background_color:
172
+ - 0.0
173
+ - 0.0
174
+ - 0.0
175
+ overfit_to_scene: null
176
+ skip_bad_shape: true
177
+ view_sampler:
178
+ name: bounded_dl3dv
179
+ min_target_views: 4
180
+ num_context_views: 16
181
+ warm_up_steps: 0
182
+ max_img_per_gpu: 16
183
+ min_distance_to_context_views: 0
184
+ initial_min_distance_between_context_views: 5
185
+ initial_max_distance_between_context_views: 7
186
+ min_distance_between_context_views: 32
187
+ max_distance_between_context_views: 256
188
+ min_gap_multiplier: 3
189
+ max_gap_multiplier: 5
190
+ initial_min_gap_multiplier: 2
191
+ initial_max_gap_multiplier: 3
192
+ target_view_sampling_mode: random
193
+ name: dl3dv
194
+ roots:
195
+ - datasets/dl3dv
196
+ input_image_shape:
197
+ - 224
198
+ - 448
199
+ original_image_shape:
200
+ - 270
201
+ - 480
202
+ cameras_are_circular: false
203
+ baseline_min: 0.001
204
+ baseline_max: 100.0
205
+ max_fov: 100.0
206
+ avg_pose: false
207
+ rescale_to_1cube: false
208
+ intr_augment: false
209
+ normalize_by_pts3d: false
checkpoints/1028_DL3DV_FULL_U/.hydra/hydra.yaml ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/full/dl3dv/${wandb.name}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - +experiment=dl3dv
116
+ - wandb.mode=online
117
+ - wandb.name=1028_DL3DV_FULL_U
118
+ - train.one_sample_validation=datasets/dl3dv/room_batch.pt
119
+ - optimizer.backbone_trainable=GB+H
120
+ - model.encoder.use_uncertainty_token=true
121
+ - train.use_uncertainty_camera_weight=true
122
+ - train.use_target_uncertainty=point_similarity_mult
123
+ job:
124
+ name: main
125
+ chdir: null
126
+ override_dirname: +experiment=dl3dv,model.encoder.use_uncertainty_token=true,optimizer.backbone_trainable=GB+H,train.one_sample_validation=datasets/dl3dv/room_batch.pt,train.use_target_uncertainty=point_similarity_mult,train.use_uncertainty_camera_weight=true,wandb.mode=online,wandb.name=1028_DL3DV_FULL_U
127
+ id: ???
128
+ num: ???
129
+ config_name: main
130
+ env_set: {}
131
+ env_copy: []
132
+ config:
133
+ override_dirname:
134
+ kv_sep: '='
135
+ item_sep: ','
136
+ exclude_keys: []
137
+ runtime:
138
+ version: 1.3.2
139
+ version_base: '1.3'
140
+ cwd: /root/highspeedstorage/injae_dl3dv/code/CVPR2026
141
+ config_sources:
142
+ - path: hydra.conf
143
+ schema: pkg
144
+ provider: hydra
145
+ - path: /root/highspeedstorage/injae_dl3dv/code/CVPR2026/config
146
+ schema: file
147
+ provider: main
148
+ - path: ''
149
+ schema: structured
150
+ provider: schema
151
+ output_dir: /root/highspeedstorage/injae_dl3dv/code/CVPR2026/outputs/full/dl3dv/1028_DL3DV_FULL_U
152
+ choices:
153
+ experiment: dl3dv
154
+ dataset@dataset.dl3dv: dl3dv
155
+ dataset/view_sampler@dataset.dl3dv.view_sampler: bounded_dl3dv
156
+ model/density_control: density_control_module
157
+ model/decoder: splatting_cuda
158
+ model/encoder: dcsplat
159
+ hydra/env: default
160
+ hydra/callbacks: null
161
+ hydra/job_logging: default
162
+ hydra/hydra_logging: default
163
+ hydra/hydra_help: default
164
+ hydra/help: default
165
+ hydra/sweeper: basic
166
+ hydra/launcher: basic
167
+ hydra/output: default
168
+ verbose: false
checkpoints/1028_DL3DV_FULL_U/.hydra/overrides.yaml ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ - +experiment=dl3dv
2
+ - wandb.mode=online
3
+ - wandb.name=1028_DL3DV_FULL_U
4
+ - train.one_sample_validation=datasets/dl3dv/room_batch.pt
5
+ - optimizer.backbone_trainable=GB+H
6
+ - model.encoder.use_uncertainty_token=true
7
+ - train.use_uncertainty_camera_weight=true
8
+ - train.use_target_uncertainty=point_similarity_mult
checkpoints/1028_DL3DV_FULL_U/main.log ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 07:26:55,923][dinov2][INFO] - using MLP layer as FFN
2
+ [2025-10-27 07:27:26,906][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2025-10-27 07:27:26,906][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2025-10-27 07:29:23,024][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=31` in the `DataLoader` to improve performance.
9
+
10
+ [2025-10-27 07:29:23,028][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
11
+ warnings.warn( # warn only once
12
+
13
+ [2025-10-27 07:29:29,025][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
14
+
15
+ [2025-10-27 07:29:29,027][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
16
+ warnings.warn(
17
+
18
+ [2025-10-27 07:29:29,027][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
19
+ warnings.warn(msg)
20
+
21
+ [2025-10-27 07:29:33,364][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
22
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
23
+
24
+ [2025-10-27 07:29:42,193][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
25
+
26
+ [2025-10-27 07:29:42,197][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
27
+
28
+ [2025-10-27 07:29:42,197][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
29
+
30
+ [2025-10-27 07:29:42,197][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
31
+
32
+ [2025-10-27 07:29:42,197][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/uncertainty_weight_mean', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
33
+
34
+ [2025-10-27 07:29:42,198][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
35
+
36
+ [2025-10-27 07:31:47,805][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
37
+ warnings.warn( # warn only once
38
+
39
+ [2025-10-27 07:31:57,578][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
40
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
41
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
42
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
43
+
44
+ [2025-10-27 07:36:57,296][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
45
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
46
+
checkpoints/1028_DL3DV_FULL_U/train_ddp_process_1.log ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 07:27:42,457][dinov2][INFO] - using MLP layer as FFN
2
+ [2025-10-27 07:28:27,692][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2025-10-27 07:28:27,693][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2025-10-27 07:29:23,028][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
9
+ warnings.warn( # warn only once
10
+
11
+ [2025-10-27 07:31:57,578][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
12
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
13
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
14
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
15
+
16
+ [2025-10-27 07:36:57,296][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
17
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
18
+
checkpoints/1028_DL3DV_FULL_U/train_ddp_process_2.log ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 07:27:42,299][dinov2][INFO] - using MLP layer as FFN
2
+ [2025-10-27 07:28:27,844][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2025-10-27 07:28:27,845][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2025-10-27 07:29:23,028][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
9
+ warnings.warn( # warn only once
10
+
11
+ [2025-10-27 07:31:57,578][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
12
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
13
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
14
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
15
+
16
+ [2025-10-27 07:36:57,296][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
17
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
18
+
checkpoints/1028_DL3DV_FULL_U/train_ddp_process_3.log ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 07:27:41,721][dinov2][INFO] - using MLP layer as FFN
2
+ [2025-10-27 07:28:27,742][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2025-10-27 07:28:27,749][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2025-10-27 07:29:23,028][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
9
+ warnings.warn( # warn only once
10
+
11
+ [2025-10-27 07:31:57,578][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
12
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
13
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
14
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
15
+
16
+ [2025-10-27 07:36:57,296][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
17
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
18
+
checkpoints/1028_DL3DV_FULL_U/train_ddp_process_4.log ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 07:27:41,361][dinov2][INFO] - using MLP layer as FFN
2
+ [2025-10-27 07:28:26,056][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2025-10-27 07:28:26,057][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2025-10-27 07:29:23,028][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
9
+ warnings.warn( # warn only once
10
+
11
+ [2025-10-27 07:31:57,578][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
12
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
13
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
14
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
15
+
16
+ [2025-10-27 07:36:57,297][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
17
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
18
+
checkpoints/1028_DL3DV_FULL_U/train_ddp_process_5.log ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 07:27:40,766][dinov2][INFO] - using MLP layer as FFN
2
+ [2025-10-27 07:27:59,762][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2025-10-27 07:27:59,762][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2025-10-27 07:29:23,028][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
9
+ warnings.warn( # warn only once
10
+
11
+ [2025-10-27 07:31:57,578][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
12
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
13
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
14
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
15
+
16
+ [2025-10-27 07:36:57,296][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
17
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
18
+
checkpoints/1028_DL3DV_FULL_U/train_ddp_process_6.log ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 07:27:42,085][dinov2][INFO] - using MLP layer as FFN
2
+ [2025-10-27 07:28:28,088][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2025-10-27 07:28:28,088][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2025-10-27 07:29:23,028][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
9
+ warnings.warn( # warn only once
10
+
11
+ [2025-10-27 07:31:57,578][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
12
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
13
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
14
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
15
+
16
+ [2025-10-27 07:36:57,296][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
17
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
18
+
checkpoints/1028_DL3DV_FULL_U/train_ddp_process_7.log ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 07:27:41,445][dinov2][INFO] - using MLP layer as FFN
2
+ [2025-10-27 07:28:27,624][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2025-10-27 07:28:27,624][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2025-10-27 07:29:23,028][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
9
+ warnings.warn( # warn only once
10
+
11
+ [2025-10-27 07:31:57,578][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
12
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
13
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
14
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
15
+
16
+ [2025-10-27 07:36:57,296][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
17
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
18
+
checkpoints/1028_context/.hydra/config.yaml ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ encoder:
3
+ name: dcsplat
4
+ input_image_shape:
5
+ - 518
6
+ - 518
7
+ head_mode: pcd
8
+ same_head: false
9
+ num_level: 3
10
+ gs_param_dim: 256
11
+ camera_prompt: false
12
+ num_visual_prompt: 0
13
+ use_extrinsic_token: false
14
+ use_intrinsic_token: false
15
+ extrinsic_drop: 0.1
16
+ intrinsic_drop: 0.1
17
+ use_pnp_solver: false
18
+ align_corners: false
19
+ image_scale: 1.0
20
+ use_uncertainty_token: true
21
+ gs_bf16: true
22
+ aggregator_bf16: false
23
+ img_num_frame_token_N: 0
24
+ img_num_global_token_N: 0
25
+ use_gs_param_img_num_embed: true
26
+ use_mean_param_img_num_embed: true
27
+ decoder:
28
+ name: splatting_cuda
29
+ background_color:
30
+ - 0.0
31
+ - 0.0
32
+ - 0.0
33
+ make_scale_invariant: false
34
+ use_gsplat: true
35
+ density_control:
36
+ name: density_control_module
37
+ mean_dim: 32
38
+ gs_param_dim: 256
39
+ refinement_layer_num: 1
40
+ num_level: 3
41
+ grad_mode: absgrad
42
+ use_mean_features: true
43
+ refinement_type: quadtree
44
+ refinement_hidden_dim: 32
45
+ aggregation_mode: mean
46
+ num_heads: 1
47
+ score_mode: 3
48
+ sigmoid_score: true
49
+ latent_dim: 128
50
+ num_latents: 64
51
+ num_self_attn_per_block: 2
52
+ render_loss:
53
+ mse:
54
+ weight: 1.0
55
+ lpips:
56
+ weight: 0.05
57
+ apply_after_step: 0
58
+ density_control_loss:
59
+ complex_score:
60
+ weight: 0.01
61
+ log_scale: false
62
+ grad_scale: 5000.0
63
+ mode: 1
64
+ direct_loss:
65
+ l1:
66
+ weight: 0.8
67
+ ssim:
68
+ weight: 0.2
69
+ wandb:
70
+ project: DCSplat
71
+ entity: scene-representation-group
72
+ name: 1028_context
73
+ mode: online
74
+ tags:
75
+ - dl3dv
76
+ - 224x448
77
+ - full
78
+ mode: train
79
+ data_loader:
80
+ train:
81
+ num_workers: 16
82
+ persistent_workers: true
83
+ batch_size: 4
84
+ seed: 1234
85
+ test:
86
+ num_workers: 4
87
+ persistent_workers: false
88
+ batch_size: 1
89
+ seed: 2345
90
+ val:
91
+ num_workers: 1
92
+ persistent_workers: true
93
+ batch_size: 1
94
+ seed: 3456
95
+ optimizer:
96
+ lr: 0.0002
97
+ warm_up_steps: 125
98
+ backbone_lr_multiplier: 0.1
99
+ backbone_trainable: GB+H
100
+ accumulate: 1
101
+ checkpointing:
102
+ load: null
103
+ every_n_train_steps: 9375
104
+ save_top_k: 1
105
+ save_weights_only: true
106
+ train:
107
+ extended_visualization: false
108
+ print_log_every_n_steps: 10
109
+ camera_loss: 1.0
110
+ one_sample_validation: datasets/dl3dv/room_batch.pt
111
+ align_corners: false
112
+ intrinsic_scaling: true
113
+ train_context: true
114
+ train_hybrid: false
115
+ use_pred_point_focal: false
116
+ scaling_detach: true
117
+ complex_threshold_range:
118
+ - 0.0
119
+ - 3.0
120
+ use_umeyama: false
121
+ pred_focal_cam_loss: false
122
+ direct_loss_weight: 0.0
123
+ direct_iteration_steps: 0
124
+ twice_backward: false
125
+ use_uncertainty_camera_weight: true
126
+ use_target_uncertainty: point_similarity_mult
127
+ verbose: false
128
+ complex_random_mode: num_gs_mode3
129
+ start_density_control: 0.0
130
+ base_complex_exp_rate: 4.0
131
+ use_refine_aux: false
132
+ train_extrinsic: true
133
+ test:
134
+ output_path: test/full/dl3dv
135
+ align_pose: false
136
+ pose_align_steps: 100
137
+ rot_opt_lr: 0.005
138
+ trans_opt_lr: 0.005
139
+ compute_scores: true
140
+ save_image: false
141
+ save_video: false
142
+ save_active_mask_image: false
143
+ save_complex_score_image: false
144
+ save_compare: false
145
+ pred_intrinsic: true
146
+ complex_threshold: 0.4
147
+ complex_threshold_list:
148
+ - 0.0
149
+ - 0.2
150
+ - 0.4
151
+ - 0.6
152
+ - 0.8
153
+ threshold_mode: ratio
154
+ nvs_view_N_list:
155
+ - 3
156
+ - 6
157
+ - 16
158
+ - 32
159
+ - 64
160
+ seed: 111123
161
+ trainer:
162
+ max_steps: 18751
163
+ val_check_interval: 500
164
+ gradient_clip_val: 0.5
165
+ num_nodes: 1
166
+ dataset:
167
+ dl3dv:
168
+ make_baseline_1: false
169
+ relative_pose: true
170
+ augment: true
171
+ background_color:
172
+ - 0.0
173
+ - 0.0
174
+ - 0.0
175
+ overfit_to_scene: null
176
+ skip_bad_shape: true
177
+ view_sampler:
178
+ name: bounded_dl3dv
179
+ min_target_views: 4
180
+ num_context_views: 16
181
+ warm_up_steps: 0
182
+ max_img_per_gpu: 16
183
+ min_distance_to_context_views: 0
184
+ initial_min_distance_between_context_views: 5
185
+ initial_max_distance_between_context_views: 7
186
+ min_distance_between_context_views: 32
187
+ max_distance_between_context_views: 256
188
+ min_gap_multiplier: 3
189
+ max_gap_multiplier: 5
190
+ initial_min_gap_multiplier: 2
191
+ initial_max_gap_multiplier: 3
192
+ target_view_sampling_mode: random
193
+ name: dl3dv
194
+ roots:
195
+ - datasets/dl3dv
196
+ input_image_shape:
197
+ - 224
198
+ - 448
199
+ original_image_shape:
200
+ - 270
201
+ - 480
202
+ cameras_are_circular: false
203
+ baseline_min: 0.001
204
+ baseline_max: 100.0
205
+ max_fov: 100.0
206
+ avg_pose: false
207
+ rescale_to_1cube: false
208
+ intr_augment: false
209
+ normalize_by_pts3d: false
checkpoints/1028_context/.hydra/hydra.yaml ADDED
@@ -0,0 +1,169 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/full/dl3dv/${wandb.name}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - +experiment=dl3dv
116
+ - wandb.mode=online
117
+ - wandb.name=1028_context
118
+ - train.one_sample_validation=datasets/dl3dv/room_batch.pt
119
+ - optimizer.backbone_trainable=GB+H
120
+ - model.encoder.use_uncertainty_token=true
121
+ - train.use_uncertainty_camera_weight=true
122
+ - train.use_target_uncertainty=point_similarity_mult
123
+ - train.train_context=true
124
+ job:
125
+ name: main
126
+ chdir: null
127
+ override_dirname: +experiment=dl3dv,model.encoder.use_uncertainty_token=true,optimizer.backbone_trainable=GB+H,train.one_sample_validation=datasets/dl3dv/room_batch.pt,train.train_context=true,train.use_target_uncertainty=point_similarity_mult,train.use_uncertainty_camera_weight=true,wandb.mode=online,wandb.name=1028_context
128
+ id: ???
129
+ num: ???
130
+ config_name: main
131
+ env_set: {}
132
+ env_copy: []
133
+ config:
134
+ override_dirname:
135
+ kv_sep: '='
136
+ item_sep: ','
137
+ exclude_keys: []
138
+ runtime:
139
+ version: 1.3.2
140
+ version_base: '1.3'
141
+ cwd: /root/highspeedstorage/injae_dl3dv/code/CVPR2026
142
+ config_sources:
143
+ - path: hydra.conf
144
+ schema: pkg
145
+ provider: hydra
146
+ - path: /root/highspeedstorage/injae_dl3dv/code/CVPR2026/config
147
+ schema: file
148
+ provider: main
149
+ - path: ''
150
+ schema: structured
151
+ provider: schema
152
+ output_dir: /root/highspeedstorage/injae_dl3dv/code/CVPR2026/outputs/full/dl3dv/1028_context
153
+ choices:
154
+ experiment: dl3dv
155
+ dataset@dataset.dl3dv: dl3dv
156
+ dataset/view_sampler@dataset.dl3dv.view_sampler: bounded_dl3dv
157
+ model/density_control: density_control_module
158
+ model/decoder: splatting_cuda
159
+ model/encoder: dcsplat
160
+ hydra/env: default
161
+ hydra/callbacks: null
162
+ hydra/job_logging: default
163
+ hydra/hydra_logging: default
164
+ hydra/hydra_help: default
165
+ hydra/help: default
166
+ hydra/sweeper: basic
167
+ hydra/launcher: basic
168
+ hydra/output: default
169
+ verbose: false
checkpoints/1028_context/.hydra/overrides.yaml ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ - +experiment=dl3dv
2
+ - wandb.mode=online
3
+ - wandb.name=1028_context
4
+ - train.one_sample_validation=datasets/dl3dv/room_batch.pt
5
+ - optimizer.backbone_trainable=GB+H
6
+ - model.encoder.use_uncertainty_token=true
7
+ - train.use_uncertainty_camera_weight=true
8
+ - train.use_target_uncertainty=point_similarity_mult
9
+ - train.train_context=true
checkpoints/1028_context/checkpoints/epoch_4-step_18750.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:689337d2ede93204d8ec67b788979f556c79561544e66361c18d9b940c56eb2d
3
+ size 4790867401
checkpoints/1028_context/main.log ADDED
@@ -0,0 +1,233 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 08:09:07,876][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
2
+ [2025-10-27 08:09:07,884][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
3
+ [2025-10-27 08:09:08,989][dinov2][INFO] - using MLP layer as FFN
4
+ [2025-10-27 08:09:13,931][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
5
+ [2025-10-27 08:09:18,430][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
6
+ warnings.warn(
7
+
8
+ [2025-10-27 08:09:18,430][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
9
+ warnings.warn(msg)
10
+
11
+ [2025-10-27 08:10:23,325][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=127` in the `DataLoader` to improve performance.
12
+
13
+ [2025-10-27 08:10:23,327][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
14
+ warnings.warn( # warn only once
15
+
16
+ [2025-10-27 08:10:27,209][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
17
+
18
+ [2025-10-27 08:10:27,210][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
19
+ warnings.warn(
20
+
21
+ [2025-10-27 08:10:27,211][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
22
+ warnings.warn(msg)
23
+
24
+ [2025-10-27 08:10:31,150][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
25
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
26
+
27
+ [2025-10-27 08:10:39,796][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
28
+
29
+ [2025-10-27 08:10:39,800][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
30
+
31
+ [2025-10-27 08:10:39,801][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
32
+
33
+ [2025-10-27 08:10:39,802][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
34
+
35
+ [2025-10-27 08:10:39,802][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/uncertainty_weight_mean', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
36
+
37
+ [2025-10-27 08:10:39,802][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
38
+
39
+ [2025-10-27 08:12:51,162][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
40
+ warnings.warn( # warn only once
41
+
42
+ [2025-10-27 08:12:56,908][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
43
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
44
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
45
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
46
+
47
+ [2025-10-27 08:16:12,614][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
48
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
49
+
50
+ [2025-10-27 08:49:39,590][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
51
+ [2025-10-27 08:49:39,598][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
52
+ [2025-10-27 08:49:40,725][dinov2][INFO] - using MLP layer as FFN
53
+ [2025-10-27 08:49:49,233][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
54
+ [2025-10-27 08:50:01,113][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
55
+ warnings.warn(
56
+
57
+ [2025-10-27 08:50:01,115][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
58
+ warnings.warn(msg)
59
+
60
+ [2025-10-27 08:53:33,914][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
61
+ [2025-10-27 08:53:33,922][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
62
+ [2025-10-27 08:53:35,086][dinov2][INFO] - using MLP layer as FFN
63
+ [2025-10-27 08:53:40,041][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
64
+ [2025-10-27 08:53:43,949][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
65
+ warnings.warn(
66
+
67
+ [2025-10-27 08:53:43,950][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
68
+ warnings.warn(msg)
69
+
70
+ [2025-10-27 08:55:45,763][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
71
+ [2025-10-27 08:55:45,771][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
72
+ [2025-10-27 08:55:46,890][dinov2][INFO] - using MLP layer as FFN
73
+ [2025-10-27 08:55:51,717][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
74
+ [2025-10-27 08:55:53,736][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
75
+ warnings.warn(
76
+
77
+ [2025-10-27 08:55:53,736][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
78
+ warnings.warn(msg)
79
+
80
+ [2025-10-27 08:55:58,135][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=255` in the `DataLoader` to improve performance.
81
+
82
+ [2025-10-27 08:56:01,803][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
83
+
84
+ [2025-10-27 08:56:01,804][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
85
+ warnings.warn(
86
+
87
+ [2025-10-27 08:56:01,805][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
88
+ warnings.warn(msg)
89
+
90
+ [2025-10-27 08:56:05,847][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
91
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
92
+
93
+ [2025-10-27 08:59:57,645][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
94
+ [2025-10-27 08:59:57,653][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
95
+ [2025-10-27 08:59:58,768][dinov2][INFO] - using MLP layer as FFN
96
+ [2025-10-27 09:00:03,499][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
97
+ [2025-10-27 09:00:05,786][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
98
+ warnings.warn(
99
+
100
+ [2025-10-27 09:00:05,787][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
101
+ warnings.warn(msg)
102
+
103
+ [2025-10-27 09:00:10,723][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=255` in the `DataLoader` to improve performance.
104
+
105
+ [2025-10-27 09:00:14,887][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
106
+
107
+ [2025-10-27 09:00:14,889][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
108
+ warnings.warn(
109
+
110
+ [2025-10-27 09:00:14,889][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
111
+ warnings.warn(msg)
112
+
113
+ [2025-10-27 09:00:18,900][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
114
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
115
+
116
+ [2025-10-27 09:00:59,779][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
117
+ [2025-10-27 09:00:59,785][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
118
+ [2025-10-27 09:01:00,958][dinov2][INFO] - using MLP layer as FFN
119
+ [2025-10-27 09:01:05,875][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
120
+ [2025-10-27 09:01:08,150][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
121
+ warnings.warn(
122
+
123
+ [2025-10-27 09:01:08,150][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
124
+ warnings.warn(msg)
125
+
126
+ [2025-10-27 09:04:49,510][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
127
+ [2025-10-27 09:04:49,518][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
128
+ [2025-10-27 09:04:50,765][dinov2][INFO] - using MLP layer as FFN
129
+ [2025-10-27 09:04:55,579][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
130
+ [2025-10-27 09:04:58,179][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
131
+ warnings.warn(
132
+
133
+ [2025-10-27 09:04:58,179][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
134
+ warnings.warn(msg)
135
+
136
+ [2025-10-27 09:23:20,766][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
137
+ [2025-10-27 09:23:20,774][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
138
+ [2025-10-27 09:23:21,877][dinov2][INFO] - using MLP layer as FFN
139
+ [2025-10-27 09:23:26,832][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
140
+ [2025-10-27 09:23:27,652][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
141
+ warnings.warn(
142
+
143
+ [2025-10-27 09:23:27,652][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
144
+ warnings.warn(msg)
145
+
146
+ [2025-10-27 09:24:26,595][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=127` in the `DataLoader` to improve performance.
147
+
148
+ [2025-10-27 09:24:26,597][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
149
+ warnings.warn( # warn only once
150
+
151
+ [2025-10-27 09:24:30,593][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
152
+
153
+ [2025-10-27 09:24:30,594][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
154
+ warnings.warn(
155
+
156
+ [2025-10-27 09:24:30,594][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
157
+ warnings.warn(msg)
158
+
159
+ [2025-10-27 09:24:34,797][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
160
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
161
+
162
+ [2025-10-27 09:24:43,576][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
163
+
164
+ [2025-10-27 09:24:43,580][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
165
+
166
+ [2025-10-27 09:24:43,580][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
167
+
168
+ [2025-10-27 09:24:43,581][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
169
+
170
+ [2025-10-27 09:24:43,581][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/uncertainty_weight_mean', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
171
+
172
+ [2025-10-27 09:24:43,581][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
173
+
174
+ [2025-10-27 09:26:55,294][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
175
+ warnings.warn( # warn only once
176
+
177
+ [2025-10-27 09:27:01,017][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
178
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
179
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
180
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
181
+
182
+ [2025-10-27 09:46:48,420][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
183
+ [2025-10-27 09:46:48,432][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
184
+ [2025-10-27 09:46:49,544][dinov2][INFO] - using MLP layer as FFN
185
+ [2025-10-27 09:46:54,402][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
186
+ [2025-10-27 09:46:58,263][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
187
+ warnings.warn(
188
+
189
+ [2025-10-27 09:46:58,264][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
190
+ warnings.warn(msg)
191
+
192
+ [2025-10-27 09:48:07,484][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=127` in the `DataLoader` to improve performance.
193
+
194
+ [2025-10-27 09:48:07,486][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
195
+ warnings.warn( # warn only once
196
+
197
+ [2025-10-27 09:48:11,263][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
198
+
199
+ [2025-10-27 09:48:11,265][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
200
+ warnings.warn(
201
+
202
+ [2025-10-27 09:48:11,265][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
203
+ warnings.warn(msg)
204
+
205
+ [2025-10-27 09:48:15,353][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
206
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
207
+
208
+ [2025-10-27 09:48:24,088][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
209
+
210
+ [2025-10-27 09:48:24,092][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
211
+
212
+ [2025-10-27 09:48:24,092][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
213
+
214
+ [2025-10-27 09:48:24,092][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
215
+
216
+ [2025-10-27 09:48:24,093][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/uncertainty_weight_mean', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
217
+
218
+ [2025-10-27 09:48:24,093][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
219
+
220
+ [2025-10-27 09:50:43,846][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
221
+ warnings.warn( # warn only once
222
+
223
+ [2025-10-27 09:50:49,478][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
224
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
225
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
226
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
227
+
228
+ [2025-10-27 09:54:04,835][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
229
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
230
+
231
+ [2025-10-27 18:23:59,100][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
232
+ warnings.warn( # warn only once
233
+
checkpoints/1028_context/train_ddp_process_1.log ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 08:09:29,938][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
2
+ [2025-10-27 08:09:29,946][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
3
+ [2025-10-27 08:09:31,046][dinov2][INFO] - using MLP layer as FFN
4
+ [2025-10-27 08:09:35,932][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
5
+ [2025-10-27 08:09:40,527][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
6
+ warnings.warn(
7
+
8
+ [2025-10-27 08:09:40,528][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
9
+ warnings.warn(msg)
10
+
11
+ [2025-10-27 08:10:23,326][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
12
+ warnings.warn( # warn only once
13
+
14
+ [2025-10-27 08:12:56,906][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
15
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
16
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
17
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
18
+
19
+ [2025-10-27 08:16:12,612][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
20
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
21
+
22
+ [2025-10-27 08:50:13,091][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
23
+ [2025-10-27 08:50:13,100][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
24
+ [2025-10-27 08:50:14,279][dinov2][INFO] - using MLP layer as FFN
25
+ [2025-10-27 08:50:19,361][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
26
+ [2025-10-27 08:50:21,606][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
27
+ warnings.warn(
28
+
29
+ [2025-10-27 08:50:21,606][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
30
+ warnings.warn(msg)
31
+
32
+ [2025-10-27 08:53:58,375][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
33
+ [2025-10-27 08:53:58,384][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
34
+ [2025-10-27 08:53:59,654][dinov2][INFO] - using MLP layer as FFN
35
+ [2025-10-27 08:54:04,637][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
36
+ [2025-10-27 08:54:08,208][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
37
+ warnings.warn(
38
+
39
+ [2025-10-27 08:54:08,209][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
40
+ warnings.warn(msg)
41
+
42
+ [2025-10-27 09:05:07,770][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
43
+ [2025-10-27 09:05:07,779][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
44
+ [2025-10-27 09:05:08,905][dinov2][INFO] - using MLP layer as FFN
45
+ [2025-10-27 09:05:13,765][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
46
+ [2025-10-27 09:05:16,340][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
47
+ warnings.warn(
48
+
49
+ [2025-10-27 09:05:16,340][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
50
+ warnings.warn(msg)
51
+
52
+ [2025-10-27 09:23:37,628][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
53
+ [2025-10-27 09:23:37,636][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
54
+ [2025-10-27 09:23:38,753][dinov2][INFO] - using MLP layer as FFN
55
+ [2025-10-27 09:23:43,588][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
56
+ [2025-10-27 09:23:44,598][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
57
+ warnings.warn(
58
+
59
+ [2025-10-27 09:23:44,599][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
60
+ warnings.warn(msg)
61
+
62
+ [2025-10-27 09:24:26,596][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
63
+ warnings.warn( # warn only once
64
+
65
+ [2025-10-27 09:27:01,015][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
66
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
67
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
68
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
69
+
70
+ [2025-10-27 09:47:13,577][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
71
+ [2025-10-27 09:47:13,586][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
72
+ [2025-10-27 09:47:14,692][dinov2][INFO] - using MLP layer as FFN
73
+ [2025-10-27 09:47:19,709][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
74
+ [2025-10-27 09:47:23,849][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
75
+ warnings.warn(
76
+
77
+ [2025-10-27 09:47:23,850][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
78
+ warnings.warn(msg)
79
+
80
+ [2025-10-27 09:48:07,486][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
81
+ warnings.warn( # warn only once
82
+
83
+ [2025-10-27 09:50:49,477][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
84
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
85
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
86
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
87
+
88
+ [2025-10-27 09:54:04,834][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
89
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
90
+
91
+ [2025-10-27 18:23:59,100][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
92
+ warnings.warn( # warn only once
93
+
checkpoints/1028_target/.hydra/config.yaml ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ encoder:
3
+ name: dcsplat
4
+ input_image_shape:
5
+ - 518
6
+ - 518
7
+ head_mode: pcd
8
+ same_head: false
9
+ num_level: 3
10
+ gs_param_dim: 256
11
+ camera_prompt: false
12
+ num_visual_prompt: 0
13
+ use_extrinsic_token: false
14
+ use_intrinsic_token: false
15
+ extrinsic_drop: 0.1
16
+ intrinsic_drop: 0.1
17
+ use_pnp_solver: false
18
+ align_corners: false
19
+ image_scale: 1.0
20
+ use_uncertainty_token: true
21
+ gs_bf16: true
22
+ aggregator_bf16: false
23
+ img_num_frame_token_N: 0
24
+ img_num_global_token_N: 0
25
+ use_gs_param_img_num_embed: true
26
+ use_mean_param_img_num_embed: true
27
+ decoder:
28
+ name: splatting_cuda
29
+ background_color:
30
+ - 0.0
31
+ - 0.0
32
+ - 0.0
33
+ make_scale_invariant: false
34
+ use_gsplat: true
35
+ density_control:
36
+ name: density_control_module
37
+ mean_dim: 32
38
+ gs_param_dim: 256
39
+ refinement_layer_num: 1
40
+ num_level: 3
41
+ grad_mode: absgrad
42
+ use_mean_features: true
43
+ refinement_type: quadtree
44
+ refinement_hidden_dim: 32
45
+ aggregation_mode: mean
46
+ num_heads: 1
47
+ score_mode: 3
48
+ sigmoid_score: true
49
+ latent_dim: 128
50
+ num_latents: 64
51
+ num_self_attn_per_block: 2
52
+ render_loss:
53
+ mse:
54
+ weight: 1.0
55
+ lpips:
56
+ weight: 0.05
57
+ apply_after_step: 0
58
+ density_control_loss:
59
+ complex_score:
60
+ weight: 0.01
61
+ log_scale: false
62
+ grad_scale: 5000.0
63
+ mode: 1
64
+ direct_loss:
65
+ l1:
66
+ weight: 0.8
67
+ ssim:
68
+ weight: 0.2
69
+ wandb:
70
+ project: DCSplat
71
+ entity: scene-representation-group
72
+ name: 1028_target
73
+ mode: online
74
+ tags:
75
+ - dl3dv
76
+ - 224x448
77
+ - full
78
+ mode: train
79
+ data_loader:
80
+ train:
81
+ num_workers: 16
82
+ persistent_workers: true
83
+ batch_size: 4
84
+ seed: 1234
85
+ test:
86
+ num_workers: 4
87
+ persistent_workers: false
88
+ batch_size: 1
89
+ seed: 2345
90
+ val:
91
+ num_workers: 1
92
+ persistent_workers: true
93
+ batch_size: 1
94
+ seed: 3456
95
+ optimizer:
96
+ lr: 0.0002
97
+ warm_up_steps: 125
98
+ backbone_lr_multiplier: 0.1
99
+ backbone_trainable: GB+H
100
+ accumulate: 1
101
+ checkpointing:
102
+ load: null
103
+ every_n_train_steps: 9375
104
+ save_top_k: 1
105
+ save_weights_only: true
106
+ train:
107
+ extended_visualization: false
108
+ print_log_every_n_steps: 10
109
+ camera_loss: 1.0
110
+ one_sample_validation: datasets/dl3dv/room_batch.pt
111
+ align_corners: false
112
+ intrinsic_scaling: true
113
+ train_context: false
114
+ train_hybrid: false
115
+ use_pred_point_focal: false
116
+ scaling_detach: true
117
+ complex_threshold_range:
118
+ - 0.0
119
+ - 3.0
120
+ use_umeyama: false
121
+ pred_focal_cam_loss: false
122
+ direct_loss_weight: 0.0
123
+ direct_iteration_steps: 0
124
+ twice_backward: false
125
+ use_uncertainty_camera_weight: true
126
+ use_target_uncertainty: point_similarity_mult
127
+ verbose: false
128
+ complex_random_mode: num_gs_mode3
129
+ start_density_control: 0.0
130
+ base_complex_exp_rate: 4.0
131
+ use_refine_aux: false
132
+ train_extrinsic: true
133
+ test:
134
+ output_path: test/full/dl3dv
135
+ align_pose: false
136
+ pose_align_steps: 100
137
+ rot_opt_lr: 0.005
138
+ trans_opt_lr: 0.005
139
+ compute_scores: true
140
+ save_image: false
141
+ save_video: false
142
+ save_active_mask_image: false
143
+ save_complex_score_image: false
144
+ save_compare: false
145
+ pred_intrinsic: true
146
+ complex_threshold: 0.4
147
+ complex_threshold_list:
148
+ - 0.0
149
+ - 0.2
150
+ - 0.4
151
+ - 0.6
152
+ - 0.8
153
+ threshold_mode: ratio
154
+ nvs_view_N_list:
155
+ - 3
156
+ - 6
157
+ - 16
158
+ - 32
159
+ - 64
160
+ seed: 111123
161
+ trainer:
162
+ max_steps: 18751
163
+ val_check_interval: 500
164
+ gradient_clip_val: 0.5
165
+ num_nodes: 1
166
+ dataset:
167
+ dl3dv:
168
+ make_baseline_1: false
169
+ relative_pose: true
170
+ augment: true
171
+ background_color:
172
+ - 0.0
173
+ - 0.0
174
+ - 0.0
175
+ overfit_to_scene: null
176
+ skip_bad_shape: true
177
+ view_sampler:
178
+ name: bounded_dl3dv
179
+ min_target_views: 4
180
+ num_context_views: 16
181
+ warm_up_steps: 0
182
+ max_img_per_gpu: 16
183
+ min_distance_to_context_views: 0
184
+ initial_min_distance_between_context_views: 5
185
+ initial_max_distance_between_context_views: 7
186
+ min_distance_between_context_views: 32
187
+ max_distance_between_context_views: 256
188
+ min_gap_multiplier: 3
189
+ max_gap_multiplier: 5
190
+ initial_min_gap_multiplier: 2
191
+ initial_max_gap_multiplier: 3
192
+ target_view_sampling_mode: random
193
+ name: dl3dv
194
+ roots:
195
+ - datasets/dl3dv
196
+ input_image_shape:
197
+ - 224
198
+ - 448
199
+ original_image_shape:
200
+ - 270
201
+ - 480
202
+ cameras_are_circular: false
203
+ baseline_min: 0.001
204
+ baseline_max: 100.0
205
+ max_fov: 100.0
206
+ avg_pose: false
207
+ rescale_to_1cube: false
208
+ intr_augment: false
209
+ normalize_by_pts3d: false
checkpoints/1028_target/.hydra/hydra.yaml ADDED
@@ -0,0 +1,169 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/full/dl3dv/${wandb.name}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - +experiment=dl3dv
116
+ - wandb.mode=online
117
+ - wandb.name=1028_target
118
+ - train.one_sample_validation=datasets/dl3dv/room_batch.pt
119
+ - optimizer.backbone_trainable=GB+H
120
+ - model.encoder.use_uncertainty_token=true
121
+ - train.use_uncertainty_camera_weight=true
122
+ - train.use_target_uncertainty=point_similarity_mult
123
+ - train.train_context=false
124
+ job:
125
+ name: main
126
+ chdir: null
127
+ override_dirname: +experiment=dl3dv,model.encoder.use_uncertainty_token=true,optimizer.backbone_trainable=GB+H,train.one_sample_validation=datasets/dl3dv/room_batch.pt,train.train_context=false,train.use_target_uncertainty=point_similarity_mult,train.use_uncertainty_camera_weight=true,wandb.mode=online,wandb.name=1028_target
128
+ id: ???
129
+ num: ???
130
+ config_name: main
131
+ env_set: {}
132
+ env_copy: []
133
+ config:
134
+ override_dirname:
135
+ kv_sep: '='
136
+ item_sep: ','
137
+ exclude_keys: []
138
+ runtime:
139
+ version: 1.3.2
140
+ version_base: '1.3'
141
+ cwd: /root/highspeedstorage/injae_dl3dv/code/CVPR2026
142
+ config_sources:
143
+ - path: hydra.conf
144
+ schema: pkg
145
+ provider: hydra
146
+ - path: /root/highspeedstorage/injae_dl3dv/code/CVPR2026/config
147
+ schema: file
148
+ provider: main
149
+ - path: ''
150
+ schema: structured
151
+ provider: schema
152
+ output_dir: /root/highspeedstorage/injae_dl3dv/code/CVPR2026/outputs/full/dl3dv/1028_target
153
+ choices:
154
+ experiment: dl3dv
155
+ dataset@dataset.dl3dv: dl3dv
156
+ dataset/view_sampler@dataset.dl3dv.view_sampler: bounded_dl3dv
157
+ model/density_control: density_control_module
158
+ model/decoder: splatting_cuda
159
+ model/encoder: dcsplat
160
+ hydra/env: default
161
+ hydra/callbacks: null
162
+ hydra/job_logging: default
163
+ hydra/hydra_logging: default
164
+ hydra/hydra_help: default
165
+ hydra/help: default
166
+ hydra/sweeper: basic
167
+ hydra/launcher: basic
168
+ hydra/output: default
169
+ verbose: false
checkpoints/1028_target/.hydra/overrides.yaml ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ - +experiment=dl3dv
2
+ - wandb.mode=online
3
+ - wandb.name=1028_target
4
+ - train.one_sample_validation=datasets/dl3dv/room_batch.pt
5
+ - optimizer.backbone_trainable=GB+H
6
+ - model.encoder.use_uncertainty_token=true
7
+ - train.use_uncertainty_camera_weight=true
8
+ - train.use_target_uncertainty=point_similarity_mult
9
+ - train.train_context=false
checkpoints/1028_target/checkpoints/epoch_4-step_18750.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:386a6301d036cce710e8a92a36db3a601345eac3c304f26f3758fdca06cdb387
3
+ size 4790867401
checkpoints/1028_target/main.log ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 08:09:09,805][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
2
+ [2025-10-27 08:09:09,814][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
3
+ [2025-10-27 08:09:11,040][dinov2][INFO] - using MLP layer as FFN
4
+ [2025-10-27 08:09:15,980][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
5
+ [2025-10-27 08:09:20,888][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
6
+ warnings.warn(
7
+
8
+ [2025-10-27 08:09:20,889][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
9
+ warnings.warn(msg)
10
+
11
+ [2025-10-27 08:10:24,249][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=127` in the `DataLoader` to improve performance.
12
+
13
+ [2025-10-27 08:10:24,251][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
14
+ warnings.warn( # warn only once
15
+
16
+ [2025-10-27 08:10:28,021][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
17
+
18
+ [2025-10-27 08:10:28,022][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
19
+ warnings.warn(
20
+
21
+ [2025-10-27 08:10:28,022][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
22
+ warnings.warn(msg)
23
+
24
+ [2025-10-27 08:10:31,963][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
25
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
26
+
27
+ [2025-10-27 08:10:40,647][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
28
+
29
+ [2025-10-27 08:10:40,650][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
30
+
31
+ [2025-10-27 08:10:40,651][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
32
+
33
+ [2025-10-27 08:10:40,651][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
34
+
35
+ [2025-10-27 08:10:40,652][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/uncertainty_weight_mean', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
36
+
37
+ [2025-10-27 08:10:40,652][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
38
+
39
+ [2025-10-27 08:12:50,555][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
40
+ warnings.warn( # warn only once
41
+
42
+ [2025-10-27 08:12:56,129][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
43
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
44
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
45
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
46
+
47
+ [2025-10-27 08:16:09,698][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
48
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
49
+
50
+ [2025-10-27 08:49:40,474][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
51
+ [2025-10-27 08:49:40,482][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
52
+ [2025-10-27 08:49:41,658][dinov2][INFO] - using MLP layer as FFN
53
+ [2025-10-27 08:49:49,734][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
54
+ [2025-10-27 08:50:02,996][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
55
+ warnings.warn(
56
+
57
+ [2025-10-27 08:50:02,997][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
58
+ warnings.warn(msg)
59
+
60
+ [2025-10-27 08:53:42,625][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
61
+ [2025-10-27 08:53:42,635][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
62
+ [2025-10-27 08:53:44,041][dinov2][INFO] - using MLP layer as FFN
63
+ [2025-10-27 08:53:49,145][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
64
+ [2025-10-27 08:53:54,889][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
65
+ warnings.warn(
66
+
67
+ [2025-10-27 08:53:54,890][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
68
+ warnings.warn(msg)
69
+
70
+ [2025-10-27 09:24:07,364][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
71
+ [2025-10-27 09:24:07,372][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
72
+ [2025-10-27 09:24:08,475][dinov2][INFO] - using MLP layer as FFN
73
+ [2025-10-27 09:24:13,257][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
74
+ [2025-10-27 09:24:14,197][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
75
+ warnings.warn(
76
+
77
+ [2025-10-27 09:24:14,198][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
78
+ warnings.warn(msg)
79
+
80
+ [2025-10-27 09:25:12,138][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=127` in the `DataLoader` to improve performance.
81
+
82
+ [2025-10-27 09:25:12,139][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
83
+ warnings.warn( # warn only once
84
+
85
+ [2025-10-27 09:25:15,928][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
86
+
87
+ [2025-10-27 09:25:15,929][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
88
+ warnings.warn(
89
+
90
+ [2025-10-27 09:25:15,929][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
91
+ warnings.warn(msg)
92
+
93
+ [2025-10-27 09:25:20,073][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
94
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
95
+
96
+ [2025-10-27 09:25:28,757][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
97
+
98
+ [2025-10-27 09:25:28,759][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
99
+
100
+ [2025-10-27 09:25:28,759][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
101
+
102
+ [2025-10-27 09:25:28,760][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
103
+
104
+ [2025-10-27 09:25:28,760][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/uncertainty_weight_mean', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
105
+
106
+ [2025-10-27 09:25:28,760][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
107
+
108
+ [2025-10-27 09:27:45,295][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
109
+ warnings.warn( # warn only once
110
+
111
+ [2025-10-27 09:27:51,579][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
112
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
113
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
114
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
115
+
116
+ [2025-10-27 09:46:50,673][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
117
+ [2025-10-27 09:46:50,680][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
118
+ [2025-10-27 09:46:51,815][dinov2][INFO] - using MLP layer as FFN
119
+ [2025-10-27 09:46:57,102][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
120
+ [2025-10-27 09:47:04,397][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
121
+ warnings.warn(
122
+
123
+ [2025-10-27 09:47:04,397][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
124
+ warnings.warn(msg)
125
+
126
+ [2025-10-27 09:48:09,535][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=127` in the `DataLoader` to improve performance.
127
+
128
+ [2025-10-27 09:48:09,537][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
129
+ warnings.warn( # warn only once
130
+
131
+ [2025-10-27 09:48:13,373][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
132
+
133
+ [2025-10-27 09:48:13,374][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
134
+ warnings.warn(
135
+
136
+ [2025-10-27 09:48:13,375][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
137
+ warnings.warn(msg)
138
+
139
+ [2025-10-27 09:48:17,358][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
140
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
141
+
142
+ [2025-10-27 09:48:26,097][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
143
+
144
+ [2025-10-27 09:48:26,100][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
145
+
146
+ [2025-10-27 09:48:26,101][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
147
+
148
+ [2025-10-27 09:48:26,101][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
149
+
150
+ [2025-10-27 09:48:26,101][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/uncertainty_weight_mean', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
151
+
152
+ [2025-10-27 09:48:26,101][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
153
+
154
+ [2025-10-27 09:50:35,465][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
155
+ warnings.warn( # warn only once
156
+
157
+ [2025-10-27 09:50:41,810][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
158
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
159
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
160
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
161
+
162
+ [2025-10-27 09:53:54,884][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
163
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
164
+
165
+ [2025-10-27 18:20:21,607][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
166
+ warnings.warn( # warn only once
167
+
checkpoints/1028_target/train_ddp_process_1.log ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 08:09:31,306][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
2
+ [2025-10-27 08:09:31,315][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
3
+ [2025-10-27 08:09:32,463][dinov2][INFO] - using MLP layer as FFN
4
+ [2025-10-27 08:09:37,458][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
5
+ [2025-10-27 08:09:43,508][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
6
+ warnings.warn(
7
+
8
+ [2025-10-27 08:09:43,509][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
9
+ warnings.warn(msg)
10
+
11
+ [2025-10-27 08:10:24,251][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
12
+ warnings.warn( # warn only once
13
+
14
+ [2025-10-27 08:12:56,127][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
15
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
16
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
17
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
18
+
19
+ [2025-10-27 08:16:09,697][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
20
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
21
+
22
+ [2025-10-27 08:50:14,347][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
23
+ [2025-10-27 08:50:14,356][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
24
+ [2025-10-27 08:50:15,540][dinov2][INFO] - using MLP layer as FFN
25
+ [2025-10-27 08:50:21,222][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
26
+ [2025-10-27 08:50:30,067][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
27
+ warnings.warn(
28
+
29
+ [2025-10-27 08:50:30,068][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
30
+ warnings.warn(msg)
31
+
32
+ [2025-10-27 08:54:05,911][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
33
+ [2025-10-27 08:54:05,919][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
34
+ [2025-10-27 08:54:07,297][dinov2][INFO] - using MLP layer as FFN
35
+ [2025-10-27 08:54:12,767][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
36
+ [2025-10-27 08:54:21,030][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
37
+ warnings.warn(
38
+
39
+ [2025-10-27 08:54:21,030][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
40
+ warnings.warn(msg)
41
+
42
+ [2025-10-27 09:24:24,823][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
43
+ [2025-10-27 09:24:24,831][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
44
+ [2025-10-27 09:24:25,899][dinov2][INFO] - using MLP layer as FFN
45
+ [2025-10-27 09:24:30,693][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
46
+ [2025-10-27 09:24:31,534][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
47
+ warnings.warn(
48
+
49
+ [2025-10-27 09:24:31,534][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
50
+ warnings.warn(msg)
51
+
52
+ [2025-10-27 09:25:12,139][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
53
+ warnings.warn( # warn only once
54
+
55
+ [2025-10-27 09:27:51,577][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
56
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
57
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
58
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
59
+
60
+ [2025-10-27 09:47:15,191][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
61
+ [2025-10-27 09:47:15,199][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
62
+ [2025-10-27 09:47:16,434][dinov2][INFO] - using MLP layer as FFN
63
+ [2025-10-27 09:47:21,875][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
64
+ [2025-10-27 09:47:27,360][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
65
+ warnings.warn(
66
+
67
+ [2025-10-27 09:47:27,360][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
68
+ warnings.warn(msg)
69
+
70
+ [2025-10-27 09:48:09,536][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
71
+ warnings.warn( # warn only once
72
+
73
+ [2025-10-27 09:50:41,809][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
74
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
75
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
76
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
77
+
78
+ [2025-10-27 09:53:54,884][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
79
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
80
+
81
+ [2025-10-27 18:20:21,607][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
82
+ warnings.warn( # warn only once
83
+
checkpoints/1028_target_segment1/.hydra/config.yaml ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ encoder:
3
+ name: dcsplat
4
+ input_image_shape:
5
+ - 518
6
+ - 518
7
+ head_mode: pcd
8
+ same_head: false
9
+ num_level: 3
10
+ gs_param_dim: 256
11
+ camera_prompt: false
12
+ num_visual_prompt: 0
13
+ use_extrinsic_token: false
14
+ use_intrinsic_token: false
15
+ extrinsic_drop: 0.1
16
+ intrinsic_drop: 0.1
17
+ use_pnp_solver: false
18
+ align_corners: false
19
+ image_scale: 1.0
20
+ use_uncertainty_token: true
21
+ gs_bf16: true
22
+ aggregator_bf16: false
23
+ img_num_frame_token_N: 0
24
+ img_num_global_token_N: 0
25
+ use_gs_param_img_num_embed: true
26
+ use_mean_param_img_num_embed: true
27
+ decoder:
28
+ name: splatting_cuda
29
+ background_color:
30
+ - 0.0
31
+ - 0.0
32
+ - 0.0
33
+ make_scale_invariant: false
34
+ use_gsplat: true
35
+ density_control:
36
+ name: density_control_module
37
+ mean_dim: 32
38
+ gs_param_dim: 256
39
+ refinement_layer_num: 1
40
+ num_level: 3
41
+ grad_mode: absgrad
42
+ use_mean_features: true
43
+ refinement_type: quadtree
44
+ refinement_hidden_dim: 32
45
+ aggregation_mode: mean
46
+ num_heads: 1
47
+ score_mode: 3
48
+ sigmoid_score: true
49
+ latent_dim: 128
50
+ num_latents: 64
51
+ num_self_attn_per_block: 2
52
+ render_loss:
53
+ mse:
54
+ weight: 1.0
55
+ lpips:
56
+ weight: 0.05
57
+ apply_after_step: 0
58
+ density_control_loss:
59
+ complex_score:
60
+ weight: 0.01
61
+ log_scale: false
62
+ grad_scale: 5000.0
63
+ mode: 1
64
+ direct_loss:
65
+ l1:
66
+ weight: 0.8
67
+ ssim:
68
+ weight: 0.2
69
+ wandb:
70
+ project: DCSplat
71
+ entity: scene-representation-group
72
+ name: 1028_target_segment1
73
+ mode: online
74
+ tags:
75
+ - dl3dv
76
+ - 224x448
77
+ - full
78
+ mode: train
79
+ data_loader:
80
+ train:
81
+ num_workers: 16
82
+ persistent_workers: true
83
+ batch_size: 4
84
+ seed: 1234
85
+ test:
86
+ num_workers: 4
87
+ persistent_workers: false
88
+ batch_size: 1
89
+ seed: 2345
90
+ val:
91
+ num_workers: 1
92
+ persistent_workers: true
93
+ batch_size: 1
94
+ seed: 3456
95
+ optimizer:
96
+ lr: 0.0002
97
+ warm_up_steps: 125
98
+ backbone_lr_multiplier: 0.1
99
+ backbone_trainable: GB+H
100
+ accumulate: 1
101
+ checkpointing:
102
+ load: null
103
+ every_n_train_steps: 9375
104
+ save_top_k: 1
105
+ save_weights_only: true
106
+ train:
107
+ extended_visualization: false
108
+ print_log_every_n_steps: 10
109
+ camera_loss: 1.0
110
+ one_sample_validation: datasets/dl3dv/room_batch.pt
111
+ align_corners: false
112
+ intrinsic_scaling: true
113
+ train_context: false
114
+ train_hybrid: false
115
+ use_pred_point_focal: false
116
+ scaling_detach: true
117
+ complex_threshold_range:
118
+ - 0.0
119
+ - 3.0
120
+ use_umeyama: false
121
+ pred_focal_cam_loss: false
122
+ direct_loss_weight: 0.0
123
+ direct_iteration_steps: 0
124
+ twice_backward: false
125
+ use_uncertainty_camera_weight: true
126
+ use_target_uncertainty: point_similarity_mult
127
+ verbose: false
128
+ complex_random_mode: num_gs_mode3
129
+ start_density_control: 0.0
130
+ base_complex_exp_rate: 4.0
131
+ use_refine_aux: false
132
+ train_extrinsic: true
133
+ test:
134
+ output_path: test/full/dl3dv
135
+ align_pose: false
136
+ pose_align_steps: 100
137
+ rot_opt_lr: 0.005
138
+ trans_opt_lr: 0.005
139
+ compute_scores: true
140
+ save_image: false
141
+ save_video: false
142
+ save_active_mask_image: false
143
+ save_complex_score_image: false
144
+ save_compare: false
145
+ pred_intrinsic: true
146
+ complex_threshold: 0.4
147
+ complex_threshold_list:
148
+ - 0.0
149
+ - 0.2
150
+ - 0.4
151
+ - 0.6
152
+ - 0.8
153
+ threshold_mode: ratio
154
+ nvs_view_N_list:
155
+ - 3
156
+ - 6
157
+ - 16
158
+ - 32
159
+ - 64
160
+ seed: 111123
161
+ trainer:
162
+ max_steps: 18751
163
+ val_check_interval: 500
164
+ gradient_clip_val: 0.5
165
+ num_nodes: 1
166
+ dataset:
167
+ dl3dv:
168
+ make_baseline_1: false
169
+ relative_pose: true
170
+ augment: true
171
+ background_color:
172
+ - 0.0
173
+ - 0.0
174
+ - 0.0
175
+ overfit_to_scene: null
176
+ skip_bad_shape: true
177
+ view_sampler:
178
+ name: bounded_dl3dv
179
+ min_target_views: 4
180
+ num_context_views: 16
181
+ warm_up_steps: 0
182
+ max_img_per_gpu: 16
183
+ min_distance_to_context_views: 0
184
+ initial_min_distance_between_context_views: 5
185
+ initial_max_distance_between_context_views: 7
186
+ min_distance_between_context_views: 32
187
+ max_distance_between_context_views: 256
188
+ min_gap_multiplier: 3
189
+ max_gap_multiplier: 5
190
+ initial_min_gap_multiplier: 2
191
+ initial_max_gap_multiplier: 3
192
+ target_view_sampling_mode: segment1
193
+ name: dl3dv
194
+ roots:
195
+ - datasets/dl3dv
196
+ input_image_shape:
197
+ - 224
198
+ - 448
199
+ original_image_shape:
200
+ - 270
201
+ - 480
202
+ cameras_are_circular: false
203
+ baseline_min: 0.001
204
+ baseline_max: 100.0
205
+ max_fov: 100.0
206
+ avg_pose: false
207
+ rescale_to_1cube: false
208
+ intr_augment: false
209
+ normalize_by_pts3d: false
checkpoints/1028_target_segment1/.hydra/hydra.yaml ADDED
@@ -0,0 +1,170 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/full/dl3dv/${wandb.name}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - +experiment=dl3dv
116
+ - wandb.mode=online
117
+ - wandb.name=1028_target_segment1
118
+ - train.one_sample_validation=datasets/dl3dv/room_batch.pt
119
+ - optimizer.backbone_trainable=GB+H
120
+ - model.encoder.use_uncertainty_token=true
121
+ - train.use_uncertainty_camera_weight=true
122
+ - train.use_target_uncertainty=point_similarity_mult
123
+ - train.train_context=false
124
+ - dataset.dl3dv.view_sampler.target_view_sampling_mode=segment1
125
+ job:
126
+ name: main
127
+ chdir: null
128
+ override_dirname: +experiment=dl3dv,dataset.dl3dv.view_sampler.target_view_sampling_mode=segment1,model.encoder.use_uncertainty_token=true,optimizer.backbone_trainable=GB+H,train.one_sample_validation=datasets/dl3dv/room_batch.pt,train.train_context=false,train.use_target_uncertainty=point_similarity_mult,train.use_uncertainty_camera_weight=true,wandb.mode=online,wandb.name=1028_target_segment1
129
+ id: ???
130
+ num: ???
131
+ config_name: main
132
+ env_set: {}
133
+ env_copy: []
134
+ config:
135
+ override_dirname:
136
+ kv_sep: '='
137
+ item_sep: ','
138
+ exclude_keys: []
139
+ runtime:
140
+ version: 1.3.2
141
+ version_base: '1.3'
142
+ cwd: /root/highspeedstorage/injae_dl3dv/code/CVPR2026
143
+ config_sources:
144
+ - path: hydra.conf
145
+ schema: pkg
146
+ provider: hydra
147
+ - path: /root/highspeedstorage/injae_dl3dv/code/CVPR2026/config
148
+ schema: file
149
+ provider: main
150
+ - path: ''
151
+ schema: structured
152
+ provider: schema
153
+ output_dir: /root/highspeedstorage/injae_dl3dv/code/CVPR2026/outputs/full/dl3dv/1028_target_segment1
154
+ choices:
155
+ experiment: dl3dv
156
+ dataset@dataset.dl3dv: dl3dv
157
+ dataset/view_sampler@dataset.dl3dv.view_sampler: bounded_dl3dv
158
+ model/density_control: density_control_module
159
+ model/decoder: splatting_cuda
160
+ model/encoder: dcsplat
161
+ hydra/env: default
162
+ hydra/callbacks: null
163
+ hydra/job_logging: default
164
+ hydra/hydra_logging: default
165
+ hydra/hydra_help: default
166
+ hydra/help: default
167
+ hydra/sweeper: basic
168
+ hydra/launcher: basic
169
+ hydra/output: default
170
+ verbose: false
checkpoints/1028_target_segment1/.hydra/overrides.yaml ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ - +experiment=dl3dv
2
+ - wandb.mode=online
3
+ - wandb.name=1028_target_segment1
4
+ - train.one_sample_validation=datasets/dl3dv/room_batch.pt
5
+ - optimizer.backbone_trainable=GB+H
6
+ - model.encoder.use_uncertainty_token=true
7
+ - train.use_uncertainty_camera_weight=true
8
+ - train.use_target_uncertainty=point_similarity_mult
9
+ - train.train_context=false
10
+ - dataset.dl3dv.view_sampler.target_view_sampling_mode=segment1
checkpoints/1028_target_segment1/checkpoints/epoch_4-step_18750.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35b3abf246c43f9c380cec1bb46f279c292e7c0ef797213ab9a3b8fab1759d91
3
+ size 4790867401
checkpoints/1028_target_segment1/main.log ADDED
@@ -0,0 +1,640 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 07:58:42,711][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
2
+ [2025-10-27 07:58:42,719][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
3
+ [2025-10-27 07:58:43,854][dinov2][INFO] - using MLP layer as FFN
4
+ [2025-10-27 07:58:48,730][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
5
+ [2025-10-27 07:59:15,532][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
6
+ warnings.warn(
7
+
8
+ [2025-10-27 07:59:15,532][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
9
+ warnings.warn(msg)
10
+
11
+ [2025-10-27 07:59:24,712][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=255` in the `DataLoader` to improve performance.
12
+
13
+ [2025-10-27 07:59:26,993][asyncio][WARNING] - socket.send() raised exception.
14
+ [2025-10-27 07:59:26,993][asyncio][WARNING] - socket.send() raised exception.
15
+ [2025-10-27 07:59:27,002][asyncio][WARNING] - socket.send() raised exception.
16
+ [2025-10-27 07:59:27,002][asyncio][WARNING] - socket.send() raised exception.
17
+ [2025-10-27 07:59:27,002][asyncio][WARNING] - socket.send() raised exception.
18
+ [2025-10-27 07:59:27,002][asyncio][WARNING] - socket.send() raised exception.
19
+ [2025-10-27 07:59:27,003][asyncio][WARNING] - socket.send() raised exception.
20
+ [2025-10-27 07:59:27,003][asyncio][WARNING] - socket.send() raised exception.
21
+ [2025-10-27 07:59:27,005][asyncio][WARNING] - socket.send() raised exception.
22
+ [2025-10-27 07:59:27,005][asyncio][WARNING] - socket.send() raised exception.
23
+ [2025-10-27 07:59:27,005][asyncio][WARNING] - socket.send() raised exception.
24
+ [2025-10-27 07:59:27,005][asyncio][WARNING] - socket.send() raised exception.
25
+ [2025-10-27 07:59:27,012][asyncio][WARNING] - socket.send() raised exception.
26
+ [2025-10-27 07:59:27,012][asyncio][WARNING] - socket.send() raised exception.
27
+ [2025-10-27 07:59:27,012][asyncio][WARNING] - socket.send() raised exception.
28
+ [2025-10-27 07:59:27,012][asyncio][WARNING] - socket.send() raised exception.
29
+ [2025-10-27 07:59:27,014][asyncio][WARNING] - socket.send() raised exception.
30
+ [2025-10-27 07:59:27,014][asyncio][WARNING] - socket.send() raised exception.
31
+ [2025-10-27 07:59:27,014][asyncio][WARNING] - socket.send() raised exception.
32
+ [2025-10-27 07:59:27,014][asyncio][WARNING] - socket.send() raised exception.
33
+ [2025-10-27 07:59:27,015][asyncio][WARNING] - socket.send() raised exception.
34
+ [2025-10-27 07:59:27,015][asyncio][WARNING] - socket.send() raised exception.
35
+ [2025-10-27 07:59:27,015][asyncio][WARNING] - socket.send() raised exception.
36
+ [2025-10-27 07:59:27,015][asyncio][WARNING] - socket.send() raised exception.
37
+ [2025-10-27 07:59:27,015][asyncio][WARNING] - socket.send() raised exception.
38
+ [2025-10-27 07:59:27,020][asyncio][WARNING] - socket.send() raised exception.
39
+ [2025-10-27 07:59:27,021][asyncio][WARNING] - socket.send() raised exception.
40
+ [2025-10-27 07:59:27,021][asyncio][WARNING] - socket.send() raised exception.
41
+ [2025-10-27 07:59:27,021][asyncio][WARNING] - socket.send() raised exception.
42
+ [2025-10-27 07:59:27,021][asyncio][WARNING] - socket.send() raised exception.
43
+ [2025-10-27 07:59:27,021][asyncio][WARNING] - socket.send() raised exception.
44
+ [2025-10-27 07:59:27,021][asyncio][WARNING] - socket.send() raised exception.
45
+ [2025-10-27 07:59:27,021][asyncio][WARNING] - socket.send() raised exception.
46
+ [2025-10-27 07:59:27,021][asyncio][WARNING] - socket.send() raised exception.
47
+ [2025-10-27 07:59:27,021][asyncio][WARNING] - socket.send() raised exception.
48
+ [2025-10-27 07:59:27,022][asyncio][WARNING] - socket.send() raised exception.
49
+ [2025-10-27 07:59:27,022][asyncio][WARNING] - socket.send() raised exception.
50
+ [2025-10-27 07:59:27,022][asyncio][WARNING] - socket.send() raised exception.
51
+ [2025-10-27 07:59:27,022][asyncio][WARNING] - socket.send() raised exception.
52
+ [2025-10-27 07:59:27,022][asyncio][WARNING] - socket.send() raised exception.
53
+ [2025-10-27 07:59:27,022][asyncio][WARNING] - socket.send() raised exception.
54
+ [2025-10-27 07:59:27,022][asyncio][WARNING] - socket.send() raised exception.
55
+ [2025-10-27 07:59:27,022][asyncio][WARNING] - socket.send() raised exception.
56
+ [2025-10-27 07:59:27,022][asyncio][WARNING] - socket.send() raised exception.
57
+ [2025-10-27 07:59:27,022][asyncio][WARNING] - socket.send() raised exception.
58
+ [2025-10-27 07:59:27,022][asyncio][WARNING] - socket.send() raised exception.
59
+ [2025-10-27 07:59:27,022][asyncio][WARNING] - socket.send() raised exception.
60
+ [2025-10-27 07:59:27,022][asyncio][WARNING] - socket.send() raised exception.
61
+ [2025-10-27 07:59:27,022][asyncio][WARNING] - socket.send() raised exception.
62
+ [2025-10-27 07:59:27,023][asyncio][WARNING] - socket.send() raised exception.
63
+ [2025-10-27 07:59:27,023][asyncio][WARNING] - socket.send() raised exception.
64
+ [2025-10-27 07:59:27,023][asyncio][WARNING] - socket.send() raised exception.
65
+ [2025-10-27 07:59:27,023][asyncio][WARNING] - socket.send() raised exception.
66
+ [2025-10-27 07:59:27,023][asyncio][WARNING] - socket.send() raised exception.
67
+ [2025-10-27 07:59:27,023][asyncio][WARNING] - socket.send() raised exception.
68
+ [2025-10-27 07:59:27,023][asyncio][WARNING] - socket.send() raised exception.
69
+ [2025-10-27 07:59:27,023][asyncio][WARNING] - socket.send() raised exception.
70
+ [2025-10-27 07:59:27,023][asyncio][WARNING] - socket.send() raised exception.
71
+ [2025-10-27 07:59:27,023][asyncio][WARNING] - socket.send() raised exception.
72
+ [2025-10-27 07:59:27,023][asyncio][WARNING] - socket.send() raised exception.
73
+ [2025-10-27 07:59:27,023][asyncio][WARNING] - socket.send() raised exception.
74
+ [2025-10-27 07:59:27,023][asyncio][WARNING] - socket.send() raised exception.
75
+ [2025-10-27 07:59:27,023][asyncio][WARNING] - socket.send() raised exception.
76
+ [2025-10-27 07:59:27,023][asyncio][WARNING] - socket.send() raised exception.
77
+ [2025-10-27 07:59:27,023][asyncio][WARNING] - socket.send() raised exception.
78
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
79
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
80
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
81
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
82
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
83
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
84
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
85
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
86
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
87
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
88
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
89
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
90
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
91
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
92
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
93
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
94
+ [2025-10-27 07:59:27,024][asyncio][WARNING] - socket.send() raised exception.
95
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
96
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
97
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
98
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
99
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
100
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
101
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
102
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
103
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
104
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
105
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
106
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
107
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
108
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
109
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
110
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
111
+ [2025-10-27 07:59:27,025][asyncio][WARNING] - socket.send() raised exception.
112
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
113
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
114
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
115
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
116
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
117
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
118
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
119
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
120
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
121
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
122
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
123
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
124
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
125
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
126
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
127
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
128
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
129
+ [2025-10-27 07:59:27,026][asyncio][WARNING] - socket.send() raised exception.
130
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
131
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
132
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
133
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
134
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
135
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
136
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
137
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
138
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
139
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
140
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
141
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
142
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
143
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
144
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
145
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
146
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
147
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
148
+ [2025-10-27 07:59:27,027][asyncio][WARNING] - socket.send() raised exception.
149
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
150
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
151
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
152
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
153
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
154
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
155
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
156
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
157
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
158
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
159
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
160
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
161
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
162
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
163
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
164
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
165
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
166
+ [2025-10-27 07:59:27,028][asyncio][WARNING] - socket.send() raised exception.
167
+ [2025-10-27 07:59:27,029][asyncio][WARNING] - socket.send() raised exception.
168
+ [2025-10-27 07:59:27,029][asyncio][WARNING] - socket.send() raised exception.
169
+ [2025-10-27 07:59:27,029][asyncio][WARNING] - socket.send() raised exception.
170
+ [2025-10-27 07:59:27,029][asyncio][WARNING] - socket.send() raised exception.
171
+ [2025-10-27 07:59:27,029][asyncio][WARNING] - socket.send() raised exception.
172
+ [2025-10-27 07:59:27,029][asyncio][WARNING] - socket.send() raised exception.
173
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
174
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
175
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
176
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
177
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
178
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
179
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
180
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
181
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
182
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
183
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
184
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
185
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
186
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
187
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
188
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
189
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
190
+ [2025-10-27 07:59:27,030][asyncio][WARNING] - socket.send() raised exception.
191
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
192
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
193
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
194
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
195
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
196
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
197
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
198
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
199
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
200
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
201
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
202
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
203
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
204
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
205
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
206
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
207
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
208
+ [2025-10-27 07:59:27,031][asyncio][WARNING] - socket.send() raised exception.
209
+ [2025-10-27 07:59:27,032][asyncio][WARNING] - socket.send() raised exception.
210
+ [2025-10-27 07:59:27,032][asyncio][WARNING] - socket.send() raised exception.
211
+ [2025-10-27 07:59:27,032][asyncio][WARNING] - socket.send() raised exception.
212
+ [2025-10-27 07:59:27,032][asyncio][WARNING] - socket.send() raised exception.
213
+ [2025-10-27 07:59:27,032][asyncio][WARNING] - socket.send() raised exception.
214
+ [2025-10-27 07:59:27,032][asyncio][WARNING] - socket.send() raised exception.
215
+ [2025-10-27 07:59:27,032][asyncio][WARNING] - socket.send() raised exception.
216
+ [2025-10-27 07:59:27,033][asyncio][WARNING] - socket.send() raised exception.
217
+ [2025-10-27 07:59:27,033][asyncio][WARNING] - socket.send() raised exception.
218
+ [2025-10-27 07:59:27,033][asyncio][WARNING] - socket.send() raised exception.
219
+ [2025-10-27 07:59:27,033][asyncio][WARNING] - socket.send() raised exception.
220
+ [2025-10-27 07:59:27,033][asyncio][WARNING] - socket.send() raised exception.
221
+ [2025-10-27 07:59:27,033][asyncio][WARNING] - socket.send() raised exception.
222
+ [2025-10-27 07:59:27,033][asyncio][WARNING] - socket.send() raised exception.
223
+ [2025-10-27 07:59:27,033][asyncio][WARNING] - socket.send() raised exception.
224
+ [2025-10-27 07:59:27,033][asyncio][WARNING] - socket.send() raised exception.
225
+ [2025-10-27 07:59:27,033][asyncio][WARNING] - socket.send() raised exception.
226
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
227
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
228
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
229
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
230
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
231
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
232
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
233
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
234
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
235
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
236
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
237
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
238
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
239
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
240
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
241
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
242
+ [2025-10-27 07:59:27,034][asyncio][WARNING] - socket.send() raised exception.
243
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
244
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
245
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
246
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
247
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
248
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
249
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
250
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
251
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
252
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
253
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
254
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
255
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
256
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
257
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
258
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
259
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
260
+ [2025-10-27 07:59:27,035][asyncio][WARNING] - socket.send() raised exception.
261
+ [2025-10-27 07:59:27,036][asyncio][WARNING] - socket.send() raised exception.
262
+ [2025-10-27 07:59:27,036][asyncio][WARNING] - socket.send() raised exception.
263
+ [2025-10-27 07:59:27,036][asyncio][WARNING] - socket.send() raised exception.
264
+ [2025-10-27 07:59:27,036][asyncio][WARNING] - socket.send() raised exception.
265
+ [2025-10-27 07:59:27,036][asyncio][WARNING] - socket.send() raised exception.
266
+ [2025-10-27 07:59:27,036][asyncio][WARNING] - socket.send() raised exception.
267
+ [2025-10-27 07:59:27,036][asyncio][WARNING] - socket.send() raised exception.
268
+ [2025-10-27 07:59:27,036][asyncio][WARNING] - socket.send() raised exception.
269
+ [2025-10-27 07:59:27,036][asyncio][WARNING] - socket.send() raised exception.
270
+ [2025-10-27 07:59:27,036][asyncio][WARNING] - socket.send() raised exception.
271
+ [2025-10-27 07:59:27,036][asyncio][WARNING] - socket.send() raised exception.
272
+ [2025-10-27 07:59:27,036][asyncio][WARNING] - socket.send() raised exception.
273
+ [2025-10-27 07:59:27,037][asyncio][WARNING] - socket.send() raised exception.
274
+ [2025-10-27 07:59:27,037][asyncio][WARNING] - socket.send() raised exception.
275
+ [2025-10-27 07:59:27,037][asyncio][WARNING] - socket.send() raised exception.
276
+ [2025-10-27 07:59:27,037][asyncio][WARNING] - socket.send() raised exception.
277
+ [2025-10-27 07:59:27,037][asyncio][WARNING] - socket.send() raised exception.
278
+ [2025-10-27 07:59:27,037][asyncio][WARNING] - socket.send() raised exception.
279
+ [2025-10-27 07:59:27,037][asyncio][WARNING] - socket.send() raised exception.
280
+ [2025-10-27 07:59:27,037][asyncio][WARNING] - socket.send() raised exception.
281
+ [2025-10-27 07:59:27,037][asyncio][WARNING] - socket.send() raised exception.
282
+ [2025-10-27 07:59:27,037][asyncio][WARNING] - socket.send() raised exception.
283
+ [2025-10-27 07:59:27,037][asyncio][WARNING] - socket.send() raised exception.
284
+ [2025-10-27 07:59:27,037][asyncio][WARNING] - socket.send() raised exception.
285
+ [2025-10-27 07:59:27,037][asyncio][WARNING] - socket.send() raised exception.
286
+ [2025-10-27 07:59:27,037][asyncio][WARNING] - socket.send() raised exception.
287
+ [2025-10-27 07:59:27,037][asyncio][WARNING] - socket.send() raised exception.
288
+ [2025-10-27 07:59:27,037][asyncio][WARNING] - socket.send() raised exception.
289
+ [2025-10-27 07:59:27,038][asyncio][WARNING] - socket.send() raised exception.
290
+ [2025-10-27 07:59:27,038][asyncio][WARNING] - socket.send() raised exception.
291
+ [2025-10-27 07:59:27,038][asyncio][WARNING] - socket.send() raised exception.
292
+ [2025-10-27 07:59:27,038][asyncio][WARNING] - socket.send() raised exception.
293
+ [2025-10-27 07:59:41,705][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
294
+ [2025-10-27 07:59:41,713][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
295
+ [2025-10-27 07:59:42,840][dinov2][INFO] - using MLP layer as FFN
296
+ [2025-10-27 07:59:47,696][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
297
+ [2025-10-27 07:59:50,349][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
298
+ warnings.warn(
299
+
300
+ [2025-10-27 07:59:50,350][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
301
+ warnings.warn(msg)
302
+
303
+ [2025-10-27 08:03:33,542][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
304
+ [2025-10-27 08:03:33,550][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
305
+ [2025-10-27 08:03:34,615][dinov2][INFO] - using MLP layer as FFN
306
+ [2025-10-27 08:03:39,147][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
307
+ [2025-10-27 08:03:41,307][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
308
+ warnings.warn(
309
+
310
+ [2025-10-27 08:03:41,307][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
311
+ warnings.warn(msg)
312
+
313
+ [2025-10-27 08:03:45,266][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=255` in the `DataLoader` to improve performance.
314
+
315
+ [2025-10-27 08:03:50,506][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
316
+
317
+ [2025-10-27 08:03:50,507][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
318
+ warnings.warn(
319
+
320
+ [2025-10-27 08:03:50,507][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
321
+ warnings.warn(msg)
322
+
323
+ [2025-10-27 08:03:54,451][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
324
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
325
+
326
+ [2025-10-27 08:07:02,882][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
327
+ [2025-10-27 08:07:02,890][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
328
+ [2025-10-27 08:07:03,994][dinov2][INFO] - using MLP layer as FFN
329
+ [2025-10-27 08:07:08,862][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
330
+ [2025-10-27 08:07:10,921][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
331
+ warnings.warn(
332
+
333
+ [2025-10-27 08:07:10,921][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
334
+ warnings.warn(msg)
335
+
336
+ [2025-10-27 08:10:10,377][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=127` in the `DataLoader` to improve performance.
337
+
338
+ [2025-10-27 08:10:10,379][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
339
+ warnings.warn( # warn only once
340
+
341
+ [2025-10-27 08:10:14,164][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
342
+
343
+ [2025-10-27 08:10:14,165][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
344
+ warnings.warn(
345
+
346
+ [2025-10-27 08:10:14,165][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
347
+ warnings.warn(msg)
348
+
349
+ [2025-10-27 08:10:18,187][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
350
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
351
+
352
+ [2025-10-27 08:10:26,791][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
353
+
354
+ [2025-10-27 08:10:26,794][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
355
+
356
+ [2025-10-27 08:10:26,795][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
357
+
358
+ [2025-10-27 08:10:26,795][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
359
+
360
+ [2025-10-27 08:10:26,795][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/uncertainty_weight_mean', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
361
+
362
+ [2025-10-27 08:10:26,795][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
363
+
364
+ [2025-10-27 08:12:32,141][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
365
+ warnings.warn( # warn only once
366
+
367
+ [2025-10-27 08:12:37,891][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
368
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
369
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
370
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
371
+
372
+ [2025-10-27 08:15:53,030][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
373
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
374
+
375
+ [2025-10-27 08:48:09,339][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
376
+ [2025-10-27 08:48:09,344][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
377
+ [2025-10-27 08:48:10,460][dinov2][INFO] - using MLP layer as FFN
378
+ [2025-10-27 08:48:15,357][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
379
+ [2025-10-27 08:48:17,864][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
380
+ warnings.warn(
381
+
382
+ [2025-10-27 08:48:17,865][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
383
+ warnings.warn(msg)
384
+
385
+ [2025-10-27 08:48:30,750][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=255` in the `DataLoader` to improve performance.
386
+
387
+ [2025-10-27 08:48:34,409][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
388
+
389
+ [2025-10-27 08:48:34,410][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
390
+ warnings.warn(
391
+
392
+ [2025-10-27 08:48:34,410][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
393
+ warnings.warn(msg)
394
+
395
+ [2025-10-27 08:48:38,481][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
396
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
397
+
398
+ [2025-10-27 08:49:34,562][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
399
+ [2025-10-27 08:49:34,571][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
400
+ [2025-10-27 08:49:35,710][dinov2][INFO] - using MLP layer as FFN
401
+ [2025-10-27 08:49:40,660][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
402
+ [2025-10-27 08:49:46,628][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
403
+ warnings.warn(
404
+
405
+ [2025-10-27 08:49:46,629][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
406
+ warnings.warn(msg)
407
+
408
+ [2025-10-27 08:53:38,522][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
409
+ [2025-10-27 08:53:38,530][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
410
+ [2025-10-27 08:53:39,655][dinov2][INFO] - using MLP layer as FFN
411
+ [2025-10-27 08:53:46,659][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
412
+ [2025-10-27 08:53:51,771][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
413
+ warnings.warn(
414
+
415
+ [2025-10-27 08:53:51,772][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
416
+ warnings.warn(msg)
417
+
418
+ [2025-10-27 08:55:41,043][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
419
+ [2025-10-27 08:55:41,051][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
420
+ [2025-10-27 08:55:42,181][dinov2][INFO] - using MLP layer as FFN
421
+ [2025-10-27 08:59:28,194][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
422
+ [2025-10-27 08:59:28,201][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
423
+ [2025-10-27 08:59:29,263][dinov2][INFO] - using MLP layer as FFN
424
+ [2025-10-27 08:59:33,805][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
425
+ [2025-10-27 08:59:36,271][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
426
+ warnings.warn(
427
+
428
+ [2025-10-27 08:59:36,272][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
429
+ warnings.warn(msg)
430
+
431
+ [2025-10-27 09:04:05,758][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
432
+ [2025-10-27 09:04:05,766][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
433
+ [2025-10-27 09:04:06,894][dinov2][INFO] - using MLP layer as FFN
434
+ [2025-10-27 09:04:11,803][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
435
+ [2025-10-27 09:04:13,888][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
436
+ warnings.warn(
437
+
438
+ [2025-10-27 09:04:13,889][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
439
+ warnings.warn(msg)
440
+
441
+ [2025-10-27 09:04:28,212][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=255` in the `DataLoader` to improve performance.
442
+
443
+ [2025-10-27 09:04:31,808][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
444
+
445
+ [2025-10-27 09:04:31,809][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
446
+ warnings.warn(
447
+
448
+ [2025-10-27 09:04:31,809][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
449
+ warnings.warn(msg)
450
+
451
+ [2025-10-27 09:04:35,811][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
452
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
453
+
454
+ [2025-10-27 09:08:56,344][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
455
+ [2025-10-27 09:08:56,352][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
456
+ [2025-10-27 09:08:57,459][dinov2][INFO] - using MLP layer as FFN
457
+ [2025-10-27 09:09:02,454][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
458
+ [2025-10-27 09:09:04,947][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
459
+ warnings.warn(
460
+
461
+ [2025-10-27 09:09:04,948][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
462
+ warnings.warn(msg)
463
+
464
+ [2025-10-27 09:11:10,604][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
465
+ [2025-10-27 09:11:10,612][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
466
+ [2025-10-27 09:11:11,719][dinov2][INFO] - using MLP layer as FFN
467
+ [2025-10-27 09:11:16,577][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
468
+ [2025-10-27 09:11:19,010][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
469
+ warnings.warn(
470
+
471
+ [2025-10-27 09:11:19,010][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
472
+ warnings.warn(msg)
473
+
474
+ [2025-10-27 09:14:30,142][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
475
+ [2025-10-27 09:14:30,150][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
476
+ [2025-10-27 09:14:31,260][dinov2][INFO] - using MLP layer as FFN
477
+ [2025-10-27 09:14:36,100][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
478
+ [2025-10-27 09:14:38,226][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
479
+ warnings.warn(
480
+
481
+ [2025-10-27 09:14:38,227][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
482
+ warnings.warn(msg)
483
+
484
+ [2025-10-27 09:16:08,214][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
485
+ [2025-10-27 09:16:08,222][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
486
+ [2025-10-27 09:16:09,339][dinov2][INFO] - using MLP layer as FFN
487
+ [2025-10-27 09:16:14,110][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
488
+ [2025-10-27 09:16:16,287][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
489
+ warnings.warn(
490
+
491
+ [2025-10-27 09:16:16,287][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
492
+ warnings.warn(msg)
493
+
494
+ [2025-10-27 09:16:21,317][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=255` in the `DataLoader` to improve performance.
495
+
496
+ [2025-10-27 09:16:24,877][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
497
+
498
+ [2025-10-27 09:16:24,878][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
499
+ warnings.warn(
500
+
501
+ [2025-10-27 09:16:24,878][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
502
+ warnings.warn(msg)
503
+
504
+ [2025-10-27 09:16:28,873][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
505
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
506
+
507
+ [2025-10-27 09:18:05,621][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
508
+ [2025-10-27 09:18:05,629][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
509
+ [2025-10-27 09:18:05,648][httpx][INFO] - HTTP Request: GET https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
510
+ [2025-10-27 09:18:06,815][dinov2][INFO] - using MLP layer as FFN
511
+ [2025-10-27 09:18:11,792][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
512
+ [2025-10-27 09:18:12,003][httpx][INFO] - HTTP Request: GET https://huggingface.co/api/models/facebook/VGGT-1B/xet-read-token/860abec7937da0a4c03c41d3c269c366e82abdf9 "HTTP/1.1 200 OK"
513
+ [2025-10-27 09:21:03,866][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
514
+ warnings.warn(
515
+
516
+ [2025-10-27 09:21:03,867][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
517
+ warnings.warn(msg)
518
+
519
+ [2025-10-27 09:22:53,081][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=127` in the `DataLoader` to improve performance.
520
+
521
+ [2025-10-27 09:22:53,082][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
522
+ warnings.warn( # warn only once
523
+
524
+ [2025-10-27 09:22:56,785][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
525
+
526
+ [2025-10-27 09:22:56,786][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
527
+ warnings.warn(
528
+
529
+ [2025-10-27 09:22:56,787][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
530
+ warnings.warn(msg)
531
+
532
+ [2025-10-27 09:23:00,846][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
533
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
534
+
535
+ [2025-10-27 09:23:09,473][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
536
+
537
+ [2025-10-27 09:23:09,476][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
538
+
539
+ [2025-10-27 09:23:09,477][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
540
+
541
+ [2025-10-27 09:23:09,477][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
542
+
543
+ [2025-10-27 09:23:09,477][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/uncertainty_weight_mean', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
544
+
545
+ [2025-10-27 09:23:09,477][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
546
+
547
+ [2025-10-27 09:25:36,657][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
548
+ warnings.warn( # warn only once
549
+
550
+ [2025-10-27 09:25:42,892][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
551
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
552
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
553
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
554
+
555
+ [2025-10-27 09:29:31,736][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
556
+ [2025-10-27 09:29:31,743][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
557
+ [2025-10-27 09:29:32,845][dinov2][INFO] - using MLP layer as FFN
558
+ [2025-10-27 09:29:37,669][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
559
+ [2025-10-27 09:29:38,687][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
560
+ warnings.warn(
561
+
562
+ [2025-10-27 09:29:38,687][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
563
+ warnings.warn(msg)
564
+
565
+ [2025-10-27 09:29:43,116][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=255` in the `DataLoader` to improve performance.
566
+
567
+ [2025-10-27 09:33:37,383][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
568
+ [2025-10-27 09:33:37,392][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
569
+ [2025-10-27 09:33:38,501][dinov2][INFO] - using MLP layer as FFN
570
+ [2025-10-27 09:33:43,256][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
571
+ [2025-10-27 09:33:45,359][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
572
+ warnings.warn(
573
+
574
+ [2025-10-27 09:33:45,363][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
575
+ warnings.warn(msg)
576
+
577
+ [2025-10-27 09:34:12,917][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
578
+ [2025-10-27 09:34:12,926][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
579
+ [2025-10-27 09:34:14,195][dinov2][INFO] - using MLP layer as FFN
580
+ [2025-10-27 09:34:19,040][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
581
+ [2025-10-27 09:34:20,903][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
582
+ warnings.warn(
583
+
584
+ [2025-10-27 09:34:20,903][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
585
+ warnings.warn(msg)
586
+
587
+ [2025-10-27 09:40:06,894][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
588
+ [2025-10-27 09:40:06,899][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
589
+ [2025-10-27 09:40:06,914][httpx][INFO] - HTTP Request: GET https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
590
+ [2025-10-27 09:40:08,063][dinov2][INFO] - using MLP layer as FFN
591
+ [2025-10-27 09:40:12,930][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
592
+ [2025-10-27 09:40:13,065][httpx][INFO] - HTTP Request: GET https://huggingface.co/api/models/facebook/VGGT-1B/xet-read-token/860abec7937da0a4c03c41d3c269c366e82abdf9 "HTTP/1.1 200 OK"
593
+ [2025-10-27 09:43:07,219][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
594
+ warnings.warn(
595
+
596
+ [2025-10-27 09:43:07,219][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
597
+ warnings.warn(msg)
598
+
599
+ [2025-10-27 09:47:29,410][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=127` in the `DataLoader` to improve performance.
600
+
601
+ [2025-10-27 09:47:29,412][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
602
+ warnings.warn( # warn only once
603
+
604
+ [2025-10-27 09:47:33,902][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
605
+
606
+ [2025-10-27 09:47:33,904][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
607
+ warnings.warn(
608
+
609
+ [2025-10-27 09:47:33,904][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
610
+ warnings.warn(msg)
611
+
612
+ [2025-10-27 09:47:38,066][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
613
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
614
+
615
+ [2025-10-27 09:47:46,796][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
616
+
617
+ [2025-10-27 09:47:46,800][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
618
+
619
+ [2025-10-27 09:47:46,800][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
620
+
621
+ [2025-10-27 09:47:46,801][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
622
+
623
+ [2025-10-27 09:47:46,801][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/uncertainty_weight_mean', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
624
+
625
+ [2025-10-27 09:47:46,801][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
626
+
627
+ [2025-10-27 09:50:33,877][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
628
+ warnings.warn( # warn only once
629
+
630
+ [2025-10-27 09:50:40,986][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
631
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
632
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
633
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
634
+
635
+ [2025-10-27 09:54:23,312][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
636
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
637
+
638
+ [2025-10-27 18:20:41,259][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
639
+ warnings.warn( # warn only once
640
+
checkpoints/1028_target_segment1/train_ddp_process_1.log ADDED
@@ -0,0 +1,143 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 08:00:02,289][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
2
+ [2025-10-27 08:00:02,298][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
3
+ [2025-10-27 08:00:03,419][dinov2][INFO] - using MLP layer as FFN
4
+ [2025-10-27 08:00:19,453][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
5
+ [2025-10-27 08:00:49,139][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
6
+ warnings.warn(
7
+
8
+ [2025-10-27 08:00:49,139][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
9
+ warnings.warn(msg)
10
+
11
+ [2025-10-27 08:07:20,782][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
12
+ [2025-10-27 08:07:20,790][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
13
+ [2025-10-27 08:07:21,921][dinov2][INFO] - using MLP layer as FFN
14
+ [2025-10-27 08:07:26,748][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
15
+ [2025-10-27 08:07:29,164][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
16
+ warnings.warn(
17
+
18
+ [2025-10-27 08:07:29,165][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
19
+ warnings.warn(msg)
20
+
21
+ [2025-10-27 08:10:10,379][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
22
+ warnings.warn( # warn only once
23
+
24
+ [2025-10-27 08:12:37,890][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
25
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
26
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
27
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
28
+
29
+ [2025-10-27 08:15:53,035][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
30
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
31
+
32
+ [2025-10-27 08:50:06,315][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
33
+ [2025-10-27 08:50:06,322][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
34
+ [2025-10-27 08:50:07,435][dinov2][INFO] - using MLP layer as FFN
35
+ [2025-10-27 08:50:12,229][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
36
+ [2025-10-27 08:50:14,411][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
37
+ warnings.warn(
38
+
39
+ [2025-10-27 08:50:14,411][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
40
+ warnings.warn(msg)
41
+
42
+ [2025-10-27 08:54:05,029][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
43
+ [2025-10-27 08:54:05,038][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
44
+ [2025-10-27 08:54:06,239][dinov2][INFO] - using MLP layer as FFN
45
+ [2025-10-27 08:54:11,736][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
46
+ [2025-10-27 08:54:20,113][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
47
+ warnings.warn(
48
+
49
+ [2025-10-27 08:54:20,114][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
50
+ warnings.warn(msg)
51
+
52
+ [2025-10-27 08:59:45,336][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
53
+ [2025-10-27 08:59:45,344][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
54
+ [2025-10-27 08:59:46,453][dinov2][INFO] - using MLP layer as FFN
55
+ [2025-10-27 08:59:51,280][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
56
+ [2025-10-27 08:59:53,605][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
57
+ warnings.warn(
58
+
59
+ [2025-10-27 08:59:53,605][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
60
+ warnings.warn(msg)
61
+
62
+ [2025-10-27 09:09:14,778][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
63
+ [2025-10-27 09:09:14,786][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
64
+ [2025-10-27 09:09:15,899][dinov2][INFO] - using MLP layer as FFN
65
+ [2025-10-27 09:09:20,757][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
66
+ [2025-10-27 09:09:22,760][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
67
+ warnings.warn(
68
+
69
+ [2025-10-27 09:09:22,760][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
70
+ warnings.warn(msg)
71
+
72
+ [2025-10-27 09:11:28,773][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
73
+ [2025-10-27 09:11:28,781][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
74
+ [2025-10-27 09:11:29,902][dinov2][INFO] - using MLP layer as FFN
75
+ [2025-10-27 09:11:34,849][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
76
+ [2025-10-27 09:11:37,375][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
77
+ warnings.warn(
78
+
79
+ [2025-10-27 09:11:37,376][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
80
+ warnings.warn(msg)
81
+
82
+ [2025-10-27 09:14:47,892][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
83
+ [2025-10-27 09:14:47,900][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
84
+ [2025-10-27 09:14:49,023][dinov2][INFO] - using MLP layer as FFN
85
+ [2025-10-27 09:14:53,793][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
86
+ [2025-10-27 09:14:55,906][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
87
+ warnings.warn(
88
+
89
+ [2025-10-27 09:14:55,907][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
90
+ warnings.warn(msg)
91
+
92
+ [2025-10-27 09:21:33,463][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
93
+ [2025-10-27 09:21:33,472][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
94
+ [2025-10-27 09:21:34,639][dinov2][INFO] - using MLP layer as FFN
95
+ [2025-10-27 09:21:39,704][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
96
+ [2025-10-27 09:21:40,784][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
97
+ warnings.warn(
98
+
99
+ [2025-10-27 09:21:40,785][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
100
+ warnings.warn(msg)
101
+
102
+ [2025-10-27 09:22:53,082][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
103
+ warnings.warn( # warn only once
104
+
105
+ [2025-10-27 09:25:42,889][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
106
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
107
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
108
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
109
+
110
+ [2025-10-27 09:34:30,552][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
111
+ [2025-10-27 09:34:30,560][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
112
+ [2025-10-27 09:34:31,668][dinov2][INFO] - using MLP layer as FFN
113
+ [2025-10-27 09:34:36,467][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
114
+ [2025-10-27 09:34:38,842][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
115
+ warnings.warn(
116
+
117
+ [2025-10-27 09:34:38,842][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
118
+ warnings.warn(msg)
119
+
120
+ [2025-10-27 09:43:35,093][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
121
+ [2025-10-27 09:43:35,101][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
122
+ [2025-10-27 09:43:36,219][dinov2][INFO] - using MLP layer as FFN
123
+ [2025-10-27 09:43:41,264][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
124
+ [2025-10-27 09:43:43,839][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
125
+ warnings.warn(
126
+
127
+ [2025-10-27 09:43:43,840][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
128
+ warnings.warn(msg)
129
+
130
+ [2025-10-27 09:47:29,412][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
131
+ warnings.warn( # warn only once
132
+
133
+ [2025-10-27 09:50:40,985][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
134
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
135
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
136
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
137
+
138
+ [2025-10-27 09:54:23,312][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
139
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
140
+
141
+ [2025-10-27 18:20:41,247][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
142
+ warnings.warn( # warn only once
143
+
checkpoints/1028_target_segment1/train_ddp_process_2.log ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 08:00:02,149][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
2
+ [2025-10-27 08:00:02,159][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
3
+ [2025-10-27 08:00:03,337][dinov2][INFO] - using MLP layer as FFN
4
+ [2025-10-27 08:00:18,854][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
5
+ [2025-10-27 08:00:47,459][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
6
+ warnings.warn(
7
+
8
+ [2025-10-27 08:00:47,461][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
9
+ warnings.warn(msg)
10
+
checkpoints/1028_target_segment1/train_ddp_process_3.log ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 08:00:02,583][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
2
+ [2025-10-27 08:00:02,593][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
3
+ [2025-10-27 08:00:03,944][dinov2][INFO] - using MLP layer as FFN
4
+ [2025-10-27 08:00:20,416][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
5
+ [2025-10-27 08:00:49,833][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
6
+ warnings.warn(
7
+
8
+ [2025-10-27 08:00:49,833][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
9
+ warnings.warn(msg)
10
+
checkpoints/1028_target_segment1/train_ddp_process_4.log ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 08:00:02,506][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
2
+ [2025-10-27 08:00:02,516][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
3
+ [2025-10-27 08:00:03,852][dinov2][INFO] - using MLP layer as FFN
4
+ [2025-10-27 08:00:20,537][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
5
+ [2025-10-27 08:00:49,662][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
6
+ warnings.warn(
7
+
8
+ [2025-10-27 08:00:49,663][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
9
+ warnings.warn(msg)
10
+
checkpoints/1028_target_segment1/train_ddp_process_5.log ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 08:00:02,967][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
2
+ [2025-10-27 08:00:02,977][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
3
+ [2025-10-27 08:00:04,159][dinov2][INFO] - using MLP layer as FFN
4
+ [2025-10-27 08:00:20,483][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
5
+ [2025-10-27 08:00:49,557][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
6
+ warnings.warn(
7
+
8
+ [2025-10-27 08:00:49,558][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
9
+ warnings.warn(msg)
10
+
checkpoints/1028_target_segment1/train_ddp_process_6.log ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 08:00:02,978][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
2
+ [2025-10-27 08:00:02,988][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
3
+ [2025-10-27 08:00:04,106][dinov2][INFO] - using MLP layer as FFN
4
+ [2025-10-27 08:00:20,384][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
5
+ [2025-10-27 08:00:49,547][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
6
+ warnings.warn(
7
+
8
+ [2025-10-27 08:00:49,548][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
9
+ warnings.warn(msg)
10
+
checkpoints/1028_target_segment1/train_ddp_process_7.log ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 08:00:02,145][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
2
+ [2025-10-27 08:00:02,155][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
3
+ [2025-10-27 08:00:03,267][dinov2][INFO] - using MLP layer as FFN
4
+ [2025-10-27 08:00:17,309][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
5
+ [2025-10-27 08:00:32,875][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
6
+ warnings.warn(
7
+
8
+ [2025-10-27 08:00:32,875][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
9
+ warnings.warn(msg)
10
+
checkpoints/1028_target_segment2/.hydra/config.yaml ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ encoder:
3
+ name: dcsplat
4
+ input_image_shape:
5
+ - 518
6
+ - 518
7
+ head_mode: pcd
8
+ same_head: false
9
+ num_level: 3
10
+ gs_param_dim: 256
11
+ camera_prompt: false
12
+ num_visual_prompt: 0
13
+ use_extrinsic_token: false
14
+ use_intrinsic_token: false
15
+ extrinsic_drop: 0.1
16
+ intrinsic_drop: 0.1
17
+ use_pnp_solver: false
18
+ align_corners: false
19
+ image_scale: 1.0
20
+ use_uncertainty_token: true
21
+ gs_bf16: true
22
+ aggregator_bf16: false
23
+ img_num_frame_token_N: 0
24
+ img_num_global_token_N: 0
25
+ use_gs_param_img_num_embed: true
26
+ use_mean_param_img_num_embed: true
27
+ decoder:
28
+ name: splatting_cuda
29
+ background_color:
30
+ - 0.0
31
+ - 0.0
32
+ - 0.0
33
+ make_scale_invariant: false
34
+ use_gsplat: true
35
+ density_control:
36
+ name: density_control_module
37
+ mean_dim: 32
38
+ gs_param_dim: 256
39
+ refinement_layer_num: 1
40
+ num_level: 3
41
+ grad_mode: absgrad
42
+ use_mean_features: true
43
+ refinement_type: quadtree
44
+ refinement_hidden_dim: 32
45
+ aggregation_mode: mean
46
+ num_heads: 1
47
+ score_mode: 3
48
+ sigmoid_score: true
49
+ latent_dim: 128
50
+ num_latents: 64
51
+ num_self_attn_per_block: 2
52
+ render_loss:
53
+ mse:
54
+ weight: 1.0
55
+ lpips:
56
+ weight: 0.05
57
+ apply_after_step: 0
58
+ density_control_loss:
59
+ complex_score:
60
+ weight: 0.01
61
+ log_scale: false
62
+ grad_scale: 5000.0
63
+ mode: 1
64
+ direct_loss:
65
+ l1:
66
+ weight: 0.8
67
+ ssim:
68
+ weight: 0.2
69
+ wandb:
70
+ project: DCSplat
71
+ entity: scene-representation-group
72
+ name: 1028_target_segment2
73
+ mode: online
74
+ tags:
75
+ - dl3dv
76
+ - 224x448
77
+ - full
78
+ mode: train
79
+ data_loader:
80
+ train:
81
+ num_workers: 16
82
+ persistent_workers: true
83
+ batch_size: 4
84
+ seed: 1234
85
+ test:
86
+ num_workers: 4
87
+ persistent_workers: false
88
+ batch_size: 1
89
+ seed: 2345
90
+ val:
91
+ num_workers: 1
92
+ persistent_workers: true
93
+ batch_size: 1
94
+ seed: 3456
95
+ optimizer:
96
+ lr: 0.0002
97
+ warm_up_steps: 125
98
+ backbone_lr_multiplier: 0.1
99
+ backbone_trainable: GB+H
100
+ accumulate: 1
101
+ checkpointing:
102
+ load: null
103
+ every_n_train_steps: 9375
104
+ save_top_k: 1
105
+ save_weights_only: true
106
+ train:
107
+ extended_visualization: false
108
+ print_log_every_n_steps: 10
109
+ camera_loss: 1.0
110
+ one_sample_validation: datasets/dl3dv/room_batch.pt
111
+ align_corners: false
112
+ intrinsic_scaling: true
113
+ train_context: false
114
+ train_hybrid: false
115
+ use_pred_point_focal: false
116
+ scaling_detach: true
117
+ complex_threshold_range:
118
+ - 0.0
119
+ - 3.0
120
+ use_umeyama: false
121
+ pred_focal_cam_loss: false
122
+ direct_loss_weight: 0.0
123
+ direct_iteration_steps: 0
124
+ twice_backward: false
125
+ use_uncertainty_camera_weight: true
126
+ use_target_uncertainty: point_similarity_mult
127
+ verbose: false
128
+ complex_random_mode: num_gs_mode3
129
+ start_density_control: 0.0
130
+ base_complex_exp_rate: 4.0
131
+ use_refine_aux: false
132
+ train_extrinsic: true
133
+ test:
134
+ output_path: test/full/dl3dv
135
+ align_pose: false
136
+ pose_align_steps: 100
137
+ rot_opt_lr: 0.005
138
+ trans_opt_lr: 0.005
139
+ compute_scores: true
140
+ save_image: false
141
+ save_video: false
142
+ save_active_mask_image: false
143
+ save_complex_score_image: false
144
+ save_compare: false
145
+ pred_intrinsic: true
146
+ complex_threshold: 0.4
147
+ complex_threshold_list:
148
+ - 0.0
149
+ - 0.2
150
+ - 0.4
151
+ - 0.6
152
+ - 0.8
153
+ threshold_mode: ratio
154
+ nvs_view_N_list:
155
+ - 3
156
+ - 6
157
+ - 16
158
+ - 32
159
+ - 64
160
+ seed: 111123
161
+ trainer:
162
+ max_steps: 18751
163
+ val_check_interval: 500
164
+ gradient_clip_val: 0.5
165
+ num_nodes: 1
166
+ dataset:
167
+ dl3dv:
168
+ make_baseline_1: false
169
+ relative_pose: true
170
+ augment: true
171
+ background_color:
172
+ - 0.0
173
+ - 0.0
174
+ - 0.0
175
+ overfit_to_scene: null
176
+ skip_bad_shape: true
177
+ view_sampler:
178
+ name: bounded_dl3dv
179
+ min_target_views: 4
180
+ num_context_views: 16
181
+ warm_up_steps: 0
182
+ max_img_per_gpu: 16
183
+ min_distance_to_context_views: 0
184
+ initial_min_distance_between_context_views: 5
185
+ initial_max_distance_between_context_views: 7
186
+ min_distance_between_context_views: 32
187
+ max_distance_between_context_views: 256
188
+ min_gap_multiplier: 3
189
+ max_gap_multiplier: 5
190
+ initial_min_gap_multiplier: 2
191
+ initial_max_gap_multiplier: 3
192
+ target_view_sampling_mode: segment2
193
+ name: dl3dv
194
+ roots:
195
+ - datasets/dl3dv
196
+ input_image_shape:
197
+ - 224
198
+ - 448
199
+ original_image_shape:
200
+ - 270
201
+ - 480
202
+ cameras_are_circular: false
203
+ baseline_min: 0.001
204
+ baseline_max: 100.0
205
+ max_fov: 100.0
206
+ avg_pose: false
207
+ rescale_to_1cube: false
208
+ intr_augment: false
209
+ normalize_by_pts3d: false
checkpoints/1028_target_segment2/.hydra/hydra.yaml ADDED
@@ -0,0 +1,170 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/full/dl3dv/${wandb.name}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - +experiment=dl3dv
116
+ - wandb.mode=online
117
+ - wandb.name=1028_target_segment2
118
+ - train.one_sample_validation=datasets/dl3dv/room_batch.pt
119
+ - optimizer.backbone_trainable=GB+H
120
+ - model.encoder.use_uncertainty_token=true
121
+ - train.use_uncertainty_camera_weight=true
122
+ - train.use_target_uncertainty=point_similarity_mult
123
+ - train.train_context=false
124
+ - dataset.dl3dv.view_sampler.target_view_sampling_mode=segment2
125
+ job:
126
+ name: main
127
+ chdir: null
128
+ override_dirname: +experiment=dl3dv,dataset.dl3dv.view_sampler.target_view_sampling_mode=segment2,model.encoder.use_uncertainty_token=true,optimizer.backbone_trainable=GB+H,train.one_sample_validation=datasets/dl3dv/room_batch.pt,train.train_context=false,train.use_target_uncertainty=point_similarity_mult,train.use_uncertainty_camera_weight=true,wandb.mode=online,wandb.name=1028_target_segment2
129
+ id: ???
130
+ num: ???
131
+ config_name: main
132
+ env_set: {}
133
+ env_copy: []
134
+ config:
135
+ override_dirname:
136
+ kv_sep: '='
137
+ item_sep: ','
138
+ exclude_keys: []
139
+ runtime:
140
+ version: 1.3.2
141
+ version_base: '1.3'
142
+ cwd: /root/highspeedstorage/injae_dl3dv/code/CVPR2026
143
+ config_sources:
144
+ - path: hydra.conf
145
+ schema: pkg
146
+ provider: hydra
147
+ - path: /root/highspeedstorage/injae_dl3dv/code/CVPR2026/config
148
+ schema: file
149
+ provider: main
150
+ - path: ''
151
+ schema: structured
152
+ provider: schema
153
+ output_dir: /root/highspeedstorage/injae_dl3dv/code/CVPR2026/outputs/full/dl3dv/1028_target_segment2
154
+ choices:
155
+ experiment: dl3dv
156
+ dataset@dataset.dl3dv: dl3dv
157
+ dataset/view_sampler@dataset.dl3dv.view_sampler: bounded_dl3dv
158
+ model/density_control: density_control_module
159
+ model/decoder: splatting_cuda
160
+ model/encoder: dcsplat
161
+ hydra/env: default
162
+ hydra/callbacks: null
163
+ hydra/job_logging: default
164
+ hydra/hydra_logging: default
165
+ hydra/hydra_help: default
166
+ hydra/help: default
167
+ hydra/sweeper: basic
168
+ hydra/launcher: basic
169
+ hydra/output: default
170
+ verbose: false
checkpoints/1028_target_segment2/.hydra/overrides.yaml ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ - +experiment=dl3dv
2
+ - wandb.mode=online
3
+ - wandb.name=1028_target_segment2
4
+ - train.one_sample_validation=datasets/dl3dv/room_batch.pt
5
+ - optimizer.backbone_trainable=GB+H
6
+ - model.encoder.use_uncertainty_token=true
7
+ - train.use_uncertainty_camera_weight=true
8
+ - train.use_target_uncertainty=point_similarity_mult
9
+ - train.train_context=false
10
+ - dataset.dl3dv.view_sampler.target_view_sampling_mode=segment2
checkpoints/1028_target_segment2/checkpoints/epoch_4-step_18750.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee61b905200cd9ac6329b398eb75b0715c750ab2f8daec812aff65fe6a162989
3
+ size 4790867401
checkpoints/1028_target_segment2/main.log ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 08:09:13,068][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
2
+ [2025-10-27 08:09:13,078][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
3
+ [2025-10-27 08:09:14,190][dinov2][INFO] - using MLP layer as FFN
4
+ [2025-10-27 08:09:22,249][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
5
+ [2025-10-27 08:09:24,348][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
6
+ warnings.warn(
7
+
8
+ [2025-10-27 08:09:24,349][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
9
+ warnings.warn(msg)
10
+
11
+ [2025-10-27 08:10:27,720][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=127` in the `DataLoader` to improve performance.
12
+
13
+ [2025-10-27 08:10:27,722][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
14
+ warnings.warn( # warn only once
15
+
16
+ [2025-10-27 08:10:31,409][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
17
+
18
+ [2025-10-27 08:10:31,410][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
19
+ warnings.warn(
20
+
21
+ [2025-10-27 08:10:31,411][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
22
+ warnings.warn(msg)
23
+
24
+ [2025-10-27 08:10:35,435][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
25
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
26
+
27
+ [2025-10-27 08:10:44,090][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
28
+
29
+ [2025-10-27 08:10:44,092][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
30
+
31
+ [2025-10-27 08:10:44,093][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
32
+
33
+ [2025-10-27 08:10:44,093][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
34
+
35
+ [2025-10-27 08:10:44,093][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/uncertainty_weight_mean', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
36
+
37
+ [2025-10-27 08:10:44,093][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
38
+
39
+ [2025-10-27 08:12:52,167][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
40
+ warnings.warn( # warn only once
41
+
42
+ [2025-10-27 08:12:57,610][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
43
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
44
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
45
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
46
+
47
+ [2025-10-27 08:16:07,803][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
48
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
49
+
50
+ [2025-10-27 08:49:41,621][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
51
+ [2025-10-27 08:49:41,631][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
52
+ [2025-10-27 08:49:43,009][dinov2][INFO] - using MLP layer as FFN
53
+ [2025-10-27 08:49:50,054][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
54
+ [2025-10-27 08:50:02,934][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
55
+ warnings.warn(
56
+
57
+ [2025-10-27 08:50:02,935][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
58
+ warnings.warn(msg)
59
+
60
+ [2025-10-27 08:53:46,226][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
61
+ [2025-10-27 08:53:46,234][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
62
+ [2025-10-27 08:53:47,400][dinov2][INFO] - using MLP layer as FFN
63
+ [2025-10-27 08:53:56,700][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
64
+ [2025-10-27 08:53:59,624][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
65
+ warnings.warn(
66
+
67
+ [2025-10-27 08:53:59,625][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
68
+ warnings.warn(msg)
69
+
70
+ [2025-10-27 09:24:14,446][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
71
+ [2025-10-27 09:24:14,456][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
72
+ [2025-10-27 09:24:15,615][dinov2][INFO] - using MLP layer as FFN
73
+ [2025-10-27 09:24:20,414][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
74
+ [2025-10-27 09:24:21,307][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
75
+ warnings.warn(
76
+
77
+ [2025-10-27 09:24:21,307][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
78
+ warnings.warn(msg)
79
+
80
+ [2025-10-27 09:25:18,600][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=127` in the `DataLoader` to improve performance.
81
+
82
+ [2025-10-27 09:25:18,603][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
83
+ warnings.warn( # warn only once
84
+
85
+ [2025-10-27 09:25:22,720][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
86
+
87
+ [2025-10-27 09:25:22,722][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
88
+ warnings.warn(
89
+
90
+ [2025-10-27 09:25:22,722][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
91
+ warnings.warn(msg)
92
+
93
+ [2025-10-27 09:25:26,706][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
94
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
95
+
96
+ [2025-10-27 09:25:35,321][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
97
+
98
+ [2025-10-27 09:25:35,324][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
99
+
100
+ [2025-10-27 09:25:35,324][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
101
+
102
+ [2025-10-27 09:25:35,325][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
103
+
104
+ [2025-10-27 09:25:35,325][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/uncertainty_weight_mean', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
105
+
106
+ [2025-10-27 09:25:35,325][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
107
+
108
+ [2025-10-27 09:28:15,573][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
109
+ warnings.warn( # warn only once
110
+
111
+ [2025-10-27 09:28:21,506][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
112
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
113
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
114
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
115
+
116
+ [2025-10-27 09:46:51,613][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
117
+ [2025-10-27 09:46:51,621][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
118
+ [2025-10-27 09:46:52,763][dinov2][INFO] - using MLP layer as FFN
119
+ [2025-10-27 09:46:58,575][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
120
+ [2025-10-27 09:47:06,139][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
121
+ warnings.warn(
122
+
123
+ [2025-10-27 09:47:06,139][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
124
+ warnings.warn(msg)
125
+
126
+ [2025-10-27 09:48:09,375][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=127` in the `DataLoader` to improve performance.
127
+
128
+ [2025-10-27 09:48:09,376][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
129
+ warnings.warn( # warn only once
130
+
131
+ [2025-10-27 09:48:13,126][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
132
+
133
+ [2025-10-27 09:48:13,128][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
134
+ warnings.warn(
135
+
136
+ [2025-10-27 09:48:13,128][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
137
+ warnings.warn(msg)
138
+
139
+ [2025-10-27 09:48:17,129][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
140
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
141
+
142
+ [2025-10-27 09:48:25,877][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
143
+
144
+ [2025-10-27 09:48:25,880][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
145
+
146
+ [2025-10-27 09:48:25,880][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
147
+
148
+ [2025-10-27 09:48:25,881][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
149
+
150
+ [2025-10-27 09:48:25,881][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/uncertainty_weight_mean', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
151
+
152
+ [2025-10-27 09:48:25,881][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
153
+
154
+ [2025-10-27 09:50:34,426][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
155
+ warnings.warn( # warn only once
156
+
157
+ [2025-10-27 09:50:40,979][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
158
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
159
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
160
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
161
+
162
+ [2025-10-27 09:53:53,305][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
163
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
164
+
165
+ [2025-10-27 18:23:04,985][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
166
+ warnings.warn( # warn only once
167
+
checkpoints/1028_target_segment2/train_ddp_process_1.log ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-10-27 08:09:34,800][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
2
+ [2025-10-27 08:09:34,811][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
3
+ [2025-10-27 08:09:35,920][dinov2][INFO] - using MLP layer as FFN
4
+ [2025-10-27 08:09:45,472][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
5
+ [2025-10-27 08:09:47,960][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
6
+ warnings.warn(
7
+
8
+ [2025-10-27 08:09:47,961][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
9
+ warnings.warn(msg)
10
+
11
+ [2025-10-27 08:10:27,721][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
12
+ warnings.warn( # warn only once
13
+
14
+ [2025-10-27 08:12:57,610][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
15
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
16
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
17
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
18
+
19
+ [2025-10-27 08:16:07,803][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
20
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
21
+
22
+ [2025-10-27 08:50:14,327][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
23
+ [2025-10-27 08:50:14,337][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
24
+ [2025-10-27 08:50:15,530][dinov2][INFO] - using MLP layer as FFN
25
+ [2025-10-27 08:50:21,217][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
26
+ [2025-10-27 08:50:30,039][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
27
+ warnings.warn(
28
+
29
+ [2025-10-27 08:50:30,040][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
30
+ warnings.warn(msg)
31
+
32
+ [2025-10-27 08:54:11,200][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
33
+ [2025-10-27 08:54:11,207][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
34
+ [2025-10-27 08:54:12,343][dinov2][INFO] - using MLP layer as FFN
35
+ [2025-10-27 08:54:22,815][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
36
+ [2025-10-27 08:54:25,544][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
37
+ warnings.warn(
38
+
39
+ [2025-10-27 08:54:25,545][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
40
+ warnings.warn(msg)
41
+
42
+ [2025-10-27 09:24:31,482][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
43
+ [2025-10-27 09:24:31,490][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
44
+ [2025-10-27 09:24:32,588][dinov2][INFO] - using MLP layer as FFN
45
+ [2025-10-27 09:24:37,420][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
46
+ [2025-10-27 09:24:38,159][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
47
+ warnings.warn(
48
+
49
+ [2025-10-27 09:24:38,159][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
50
+ warnings.warn(msg)
51
+
52
+ [2025-10-27 09:25:18,616][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
53
+ warnings.warn( # warn only once
54
+
55
+ [2025-10-27 09:28:21,499][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
56
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
57
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
58
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
59
+
60
+ [2025-10-27 09:47:17,007][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/config.json "HTTP/1.1 307 Temporary Redirect"
61
+ [2025-10-27 09:47:17,020][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/api/resolve-cache/models/facebook/VGGT-1B/860abec7937da0a4c03c41d3c269c366e82abdf9/config.json "HTTP/1.1 200 OK"
62
+ [2025-10-27 09:47:18,185][dinov2][INFO] - using MLP layer as FFN
63
+ [2025-10-27 09:47:24,820][httpx][INFO] - HTTP Request: HEAD https://huggingface.co/facebook/VGGT-1B/resolve/main/model.safetensors "HTTP/1.1 302 Found"
64
+ [2025-10-27 09:47:28,901][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
65
+ warnings.warn(
66
+
67
+ [2025-10-27 09:47:28,901][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
68
+ warnings.warn(msg)
69
+
70
+ [2025-10-27 09:48:09,377][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
71
+ warnings.warn( # warn only once
72
+
73
+ [2025-10-27 09:50:40,978][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
74
+ grad.sizes() = [1, 288, 1, 1], strides() = [288, 1, 288, 288]
75
+ bucket_view.sizes() = [1, 288, 1, 1], strides() = [288, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
76
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
77
+
78
+ [2025-10-27 09:53:53,304][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
79
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
80
+
81
+ [2025-10-27 18:23:04,989][py.warnings][WARNING] - /opt/conda/envs/dcsplat/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
82
+ warnings.warn( # warn only once
83
+