Knowing commited on
Commit
9929962
·
verified ·
1 Parent(s): 351117d

Add files using upload-large-folder tool

Browse files
Files changed (50) hide show
  1. ABLATION_0225_ctxTrain_depth_vggtDistl/.hydra/config.yaml +189 -0
  2. ABLATION_0225_ctxTrain_depth_vggtDistl/.hydra/hydra.yaml +168 -0
  3. ABLATION_0225_ctxTrain_depth_vggtDistl/.hydra/overrides.yaml +7 -0
  4. ABLATION_0225_ctxTrain_depth_vggtDistl/main.log +45 -0
  5. ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_1.log +19 -0
  6. ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_2.log +19 -0
  7. ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_3.log +19 -0
  8. ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_4.log +19 -0
  9. ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_5.log +19 -0
  10. ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_6.log +19 -0
  11. ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_7.log +19 -0
  12. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/debug-internal.log +11 -0
  13. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/debug.log +21 -0
  14. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/config.yaml +310 -0
  15. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/media/images/active_mask_imgs_1_690a5f29fe9230f3ead3.png +3 -0
  16. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/media/images/comparison_0_e330894a93360b5335a1.png +3 -0
  17. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/media/images/error_scores_2_bf819ed247f7546e0637.png +3 -0
  18. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/output.log +140 -0
  19. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/requirements.txt +172 -0
  20. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/wandb-metadata.json +96 -0
  21. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/wandb-summary.json +1 -0
  22. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug-core.log +15 -0
  23. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug-internal.log +11 -0
  24. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug.log +21 -0
  25. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/run-g2qj8ry7.wandb +0 -0
  26. ABLATION_0225_randomSelect/main.log +12 -0
  27. ABLATION_0225_randomSelect/peak_vram_memory.json +6 -0
  28. ABLATION_0225_randomSelect/train_ddp_process_1.log +6 -0
  29. ABLATION_0225_randomSelect/train_ddp_process_2.log +6 -0
  30. ABLATION_0225_randomSelect/train_ddp_process_3.log +6 -0
  31. ABLATION_0225_randomSelect/train_ddp_process_4.log +6 -0
  32. ABLATION_0225_randomSelect/train_ddp_process_5.log +6 -0
  33. ABLATION_0225_randomSelect/train_ddp_process_6.log +6 -0
  34. ABLATION_0225_randomSelect/train_ddp_process_7.log +6 -0
  35. ABLATION_0225_randomSelect/wandb/debug-internal.log +5 -0
  36. ABLATION_0225_randomSelect/wandb/debug.log +2 -0
  37. ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/config.yaml +307 -0
  38. ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/active_mask_imgs_182_2edf4f17c23388f66672.png +3 -0
  39. ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/active_mask_imgs_198_399ec9b2f678832e0977.png +3 -0
  40. ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/comparison_197_ddc312f116aae122416e.png +3 -0
  41. ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/error_scores_183_e47e70d75361cb369a35.png +3 -0
  42. ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/error_scores_199_b1c8e80e7cf22aab1c67.png +3 -0
  43. ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/train/comparison_188_a456ce8a38a4d0dc65b1.png +3 -0
  44. ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/train/comparison_202_b207767914aae1b684b5.png +3 -0
  45. ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/train/error_scores_187_347cd0a01cf553ca4f4d.png +3 -0
  46. ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/train/error_scores_201_05456dc706694ddf5d92.png +3 -0
  47. ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/wandb-summary.json +1 -0
  48. ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/logs/debug-core.log +8 -0
  49. ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/logs/debug-internal.log +5 -0
  50. ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/logs/debug.log +2 -0
ABLATION_0225_ctxTrain_depth_vggtDistl/.hydra/config.yaml ADDED
@@ -0,0 +1,189 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ encoder:
3
+ name: dcsplat
4
+ input_image_shape:
5
+ - 518
6
+ - 518
7
+ head_mode: depth
8
+ num_level: 3
9
+ gs_param_dim: 256
10
+ align_corners: false
11
+ use_voxelize: true
12
+ decoder:
13
+ name: splatting_cuda
14
+ background_color:
15
+ - 0.0
16
+ - 0.0
17
+ - 0.0
18
+ make_scale_invariant: false
19
+ density_control:
20
+ name: density_control_module
21
+ mean_dim: 32
22
+ gs_param_dim: 256
23
+ refinement_layer_num: 1
24
+ num_level: 3
25
+ grad_mode: absgrad
26
+ use_mean_features: true
27
+ refinement_type: voxelize
28
+ refinement_hidden_dim: 32
29
+ aggregation_mode: mean
30
+ num_heads: 1
31
+ score_mode: absgrad
32
+ latent_dim: 128
33
+ num_latents: 64
34
+ num_self_attn_per_block: 2
35
+ voxel_size: 0.001
36
+ aux_refine: false
37
+ refine_error: false
38
+ use_refine_module: true
39
+ voxelize_activate: true
40
+ use_depth: true
41
+ render_loss:
42
+ mse:
43
+ weight: 1.0
44
+ lpips:
45
+ weight: 0.05
46
+ apply_after_step: 0
47
+ depth_consis:
48
+ weight: 1.0
49
+ sigma_image: null
50
+ use_second_derivative: false
51
+ density_control_loss:
52
+ error_score:
53
+ weight: 0.01
54
+ log_scale: false
55
+ grad_scale: 10000.0
56
+ mode: original
57
+ direct_loss:
58
+ l1:
59
+ weight: 0.8
60
+ ssim:
61
+ weight: 0.2
62
+ wandb:
63
+ project: DCSplat
64
+ entity: scene-representation-group
65
+ name: ABLATION_0225_ctxTrain_depth_vggtDistl
66
+ mode: online
67
+ tags:
68
+ - re10k
69
+ - 256x256
70
+ mode: train
71
+ data_loader:
72
+ train:
73
+ num_workers: 16
74
+ persistent_workers: true
75
+ batch_size: 16
76
+ seed: 1234
77
+ test:
78
+ num_workers: 4
79
+ persistent_workers: false
80
+ batch_size: 1
81
+ seed: 2345
82
+ val:
83
+ num_workers: 1
84
+ persistent_workers: true
85
+ batch_size: 1
86
+ seed: 3456
87
+ optimizer:
88
+ lr: 0.0002
89
+ warm_up_steps: 25
90
+ backbone_lr_multiplier: 0.1
91
+ backbone_trainable: T+H
92
+ accumulate: 1
93
+ checkpointing:
94
+ load: null
95
+ every_n_train_steps: 1500
96
+ save_top_k: 2
97
+ save_weights_only: false
98
+ train:
99
+ extended_visualization: false
100
+ print_log_every_n_steps: 10
101
+ camera_loss: 10.0
102
+ one_sample_validation: null
103
+ align_corners: false
104
+ intrinsic_scaling: false
105
+ verbose: false
106
+ beta_dist_param:
107
+ - 0.5
108
+ - 4.0
109
+ use_refine_aux: false
110
+ train_target_set: true
111
+ train_gs_num: 1
112
+ ext_scale_detach: false
113
+ cam_scale_mode: sum
114
+ scene_scale_reg_loss: 0.01
115
+ train_aux: true
116
+ vggt_cam_loss: true
117
+ vggt_distil: true
118
+ context_view_train: true
119
+ test:
120
+ output_path: test/ablation/re10k
121
+ align_pose: false
122
+ pose_align_steps: 100
123
+ rot_opt_lr: 0.005
124
+ trans_opt_lr: 0.005
125
+ compute_scores: true
126
+ save_image: false
127
+ save_video: false
128
+ save_active_mask_image: false
129
+ save_error_score_image: false
130
+ save_compare: false
131
+ pred_intrinsic: false
132
+ error_threshold: 0.4
133
+ error_threshold_list:
134
+ - 0.2
135
+ - 0.4
136
+ - 0.6
137
+ - 0.8
138
+ - 1.0
139
+ threshold_mode: ratio
140
+ nvs_view_N_list:
141
+ - 3
142
+ - 6
143
+ - 16
144
+ - 32
145
+ - 64
146
+ seed: 111123
147
+ trainer:
148
+ max_steps: 3001
149
+ val_check_interval: 250
150
+ gradient_clip_val: 0.5
151
+ num_nodes: 1
152
+ dataset:
153
+ re10k:
154
+ make_baseline_1: true
155
+ relative_pose: true
156
+ augment: true
157
+ background_color:
158
+ - 0.0
159
+ - 0.0
160
+ - 0.0
161
+ overfit_to_scene: null
162
+ skip_bad_shape: true
163
+ view_sampler:
164
+ name: bounded
165
+ num_target_views: 4
166
+ num_context_views: 2
167
+ min_distance_between_context_views: 45
168
+ max_distance_between_context_views: 90
169
+ min_distance_to_context_views: 0
170
+ warm_up_steps: 1000
171
+ initial_min_distance_between_context_views: 25
172
+ initial_max_distance_between_context_views: 25
173
+ same_target_gap: false
174
+ num_target_set: 3
175
+ name: re10k
176
+ roots:
177
+ - datasets/re10k
178
+ input_image_shape:
179
+ - 256
180
+ - 256
181
+ original_image_shape:
182
+ - 360
183
+ - 640
184
+ cameras_are_circular: false
185
+ baseline_min: 0.001
186
+ baseline_max: 10000000000.0
187
+ max_fov: 100.0
188
+ dynamic_context_views: true
189
+ max_context_views_per_gpu: 24
ABLATION_0225_ctxTrain_depth_vggtDistl/.hydra/hydra.yaml ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/ablation/re10k/${wandb.name}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - +experiment=re10k_ablation_24v_depthloss
116
+ - wandb.mode=online
117
+ - wandb.name=ABLATION_0225_ctxTrain_depth_vggtDistl
118
+ - model.encoder.head_mode=depth
119
+ - model.density_control.use_depth=true
120
+ - train.context_view_train=true
121
+ - train.vggt_distil=true
122
+ job:
123
+ name: main
124
+ chdir: null
125
+ override_dirname: +experiment=re10k_ablation_24v_depthloss,model.density_control.use_depth=true,model.encoder.head_mode=depth,train.context_view_train=true,train.vggt_distil=true,wandb.mode=online,wandb.name=ABLATION_0225_ctxTrain_depth_vggtDistl
126
+ id: ???
127
+ num: ???
128
+ config_name: main
129
+ env_set: {}
130
+ env_copy: []
131
+ config:
132
+ override_dirname:
133
+ kv_sep: '='
134
+ item_sep: ','
135
+ exclude_keys: []
136
+ runtime:
137
+ version: 1.3.2
138
+ version_base: '1.3'
139
+ cwd: /workspace/code/CVPR2026
140
+ config_sources:
141
+ - path: hydra.conf
142
+ schema: pkg
143
+ provider: hydra
144
+ - path: /workspace/code/CVPR2026/config
145
+ schema: file
146
+ provider: main
147
+ - path: ''
148
+ schema: structured
149
+ provider: schema
150
+ output_dir: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl
151
+ choices:
152
+ experiment: re10k_ablation_24v_depthloss
153
+ dataset@dataset.re10k: re10k
154
+ dataset/view_sampler_dataset_specific_config@dataset.re10k.view_sampler: bounded_re10k
155
+ dataset/view_sampler@dataset.re10k.view_sampler: bounded
156
+ model/density_control: density_control_module
157
+ model/decoder: splatting_cuda
158
+ model/encoder: dcsplat
159
+ hydra/env: default
160
+ hydra/callbacks: null
161
+ hydra/job_logging: default
162
+ hydra/hydra_logging: default
163
+ hydra/hydra_help: default
164
+ hydra/help: default
165
+ hydra/sweeper: basic
166
+ hydra/launcher: basic
167
+ hydra/output: default
168
+ verbose: false
ABLATION_0225_ctxTrain_depth_vggtDistl/.hydra/overrides.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ - +experiment=re10k_ablation_24v_depthloss
2
+ - wandb.mode=online
3
+ - wandb.name=ABLATION_0225_ctxTrain_depth_vggtDistl
4
+ - model.encoder.head_mode=depth
5
+ - model.density_control.use_depth=true
6
+ - train.context_view_train=true
7
+ - train.vggt_distil=true
ABLATION_0225_ctxTrain_depth_vggtDistl/main.log ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-25 13:52:21,786][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-02-25 13:52:28,031][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-02-25 13:52:28,032][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-02-25 13:52:29,480][dinov2][INFO] - using MLP layer as FFN
9
+ [2026-02-25 13:54:48,808][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=31` in the `DataLoader` to improve performance.
10
+
11
+ [2026-02-25 13:54:48,809][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
12
+ warnings.warn( # warn only once
13
+
14
+ [2026-02-25 13:54:51,686][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
15
+ result[selector] = overlay
16
+
17
+ [2026-02-25 13:54:51,695][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
18
+
19
+ [2026-02-25 13:54:51,696][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
20
+ warnings.warn(
21
+
22
+ [2026-02-25 13:54:51,696][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
23
+ warnings.warn(msg)
24
+
25
+ [2026-02-25 13:54:53,564][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4322.)
26
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
27
+
28
+ [2026-02-25 13:54:53,845][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
29
+
30
+ [2026-02-25 13:54:53,847][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
31
+
32
+ [2026-02-25 13:54:53,847][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
33
+
34
+ [2026-02-25 13:54:53,848][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
35
+
36
+ [2026-02-25 13:54:53,848][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
37
+
38
+ [2026-02-25 14:29:16,843][dinov2][INFO] - using MLP layer as FFN
39
+ [2026-02-25 14:29:23,026][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
40
+ warnings.warn(
41
+
42
+ [2026-02-25 14:29:23,026][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
43
+ warnings.warn(msg)
44
+
45
+ [2026-02-25 14:29:24,307][dinov2][INFO] - using MLP layer as FFN
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_1.log ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-25 13:52:46,873][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-02-25 13:53:20,123][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-02-25 13:53:20,123][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-02-25 13:53:24,223][dinov2][INFO] - using MLP layer as FFN
9
+ [2026-02-25 13:54:48,809][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
10
+ warnings.warn( # warn only once
11
+
12
+ [2026-02-25 14:29:41,340][dinov2][INFO] - using MLP layer as FFN
13
+ [2026-02-25 14:29:59,908][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
14
+ warnings.warn(
15
+
16
+ [2026-02-25 14:29:59,909][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
17
+ warnings.warn(msg)
18
+
19
+ [2026-02-25 14:30:04,698][dinov2][INFO] - using MLP layer as FFN
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_2.log ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-25 13:52:46,752][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-02-25 13:53:19,432][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-02-25 13:53:19,433][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-02-25 13:53:22,571][dinov2][INFO] - using MLP layer as FFN
9
+ [2026-02-25 13:54:48,809][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
10
+ warnings.warn( # warn only once
11
+
12
+ [2026-02-25 14:29:41,333][dinov2][INFO] - using MLP layer as FFN
13
+ [2026-02-25 14:30:12,121][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
14
+ warnings.warn(
15
+
16
+ [2026-02-25 14:30:12,123][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
17
+ warnings.warn(msg)
18
+
19
+ [2026-02-25 14:30:17,034][dinov2][INFO] - using MLP layer as FFN
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_3.log ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-25 13:52:46,815][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-02-25 13:53:19,826][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-02-25 13:53:19,827][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-02-25 13:53:23,871][dinov2][INFO] - using MLP layer as FFN
9
+ [2026-02-25 13:54:48,809][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
10
+ warnings.warn( # warn only once
11
+
12
+ [2026-02-25 14:29:41,170][dinov2][INFO] - using MLP layer as FFN
13
+ [2026-02-25 14:30:12,406][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
14
+ warnings.warn(
15
+
16
+ [2026-02-25 14:30:12,406][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
17
+ warnings.warn(msg)
18
+
19
+ [2026-02-25 14:30:17,604][dinov2][INFO] - using MLP layer as FFN
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_4.log ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-25 13:52:46,821][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-02-25 13:53:15,687][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-02-25 13:53:15,687][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-02-25 13:53:20,106][dinov2][INFO] - using MLP layer as FFN
9
+ [2026-02-25 13:54:48,809][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
10
+ warnings.warn( # warn only once
11
+
12
+ [2026-02-25 14:29:41,163][dinov2][INFO] - using MLP layer as FFN
13
+ [2026-02-25 14:30:12,743][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
14
+ warnings.warn(
15
+
16
+ [2026-02-25 14:30:12,745][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
17
+ warnings.warn(msg)
18
+
19
+ [2026-02-25 14:30:17,884][dinov2][INFO] - using MLP layer as FFN
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_5.log ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-25 13:52:46,848][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-02-25 13:53:17,936][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-02-25 13:53:17,940][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-02-25 13:53:20,315][dinov2][INFO] - using MLP layer as FFN
9
+ [2026-02-25 13:54:48,813][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
10
+ warnings.warn( # warn only once
11
+
12
+ [2026-02-25 14:29:41,164][dinov2][INFO] - using MLP layer as FFN
13
+ [2026-02-25 14:30:09,668][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
14
+ warnings.warn(
15
+
16
+ [2026-02-25 14:30:09,670][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
17
+ warnings.warn(msg)
18
+
19
+ [2026-02-25 14:30:13,767][dinov2][INFO] - using MLP layer as FFN
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_6.log ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-25 13:52:46,807][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-02-25 13:53:19,201][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-02-25 13:53:19,203][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-02-25 13:53:21,001][dinov2][INFO] - using MLP layer as FFN
9
+ [2026-02-25 13:54:48,809][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
10
+ warnings.warn( # warn only once
11
+
12
+ [2026-02-25 14:29:41,135][dinov2][INFO] - using MLP layer as FFN
13
+ [2026-02-25 14:30:01,944][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
14
+ warnings.warn(
15
+
16
+ [2026-02-25 14:30:01,944][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
17
+ warnings.warn(msg)
18
+
19
+ [2026-02-25 14:30:07,824][dinov2][INFO] - using MLP layer as FFN
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_7.log ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-25 13:52:46,799][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-02-25 13:53:19,450][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-02-25 13:53:19,450][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-02-25 13:53:22,975][dinov2][INFO] - using MLP layer as FFN
9
+ [2026-02-25 13:54:48,809][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
10
+ warnings.warn( # warn only once
11
+
12
+ [2026-02-25 14:29:41,375][dinov2][INFO] - using MLP layer as FFN
13
+ [2026-02-25 14:30:00,005][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
14
+ warnings.warn(
15
+
16
+ [2026-02-25 14:30:00,006][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
17
+ warnings.warn(msg)
18
+
19
+ [2026-02-25 14:30:04,401][dinov2][INFO] - using MLP layer as FFN
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/debug-internal.log ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-02-25T13:54:42.16430492Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
2
+ {"time":"2026-02-25T13:54:42.614803801Z","level":"INFO","msg":"stream: created new stream","id":"g2qj8ry7"}
3
+ {"time":"2026-02-25T13:54:42.615183417Z","level":"INFO","msg":"handler: started","stream_id":"g2qj8ry7"}
4
+ {"time":"2026-02-25T13:54:42.615386361Z","level":"INFO","msg":"stream: started","id":"g2qj8ry7"}
5
+ {"time":"2026-02-25T13:54:42.615431072Z","level":"INFO","msg":"sender: started","stream_id":"g2qj8ry7"}
6
+ {"time":"2026-02-25T13:54:42.615463052Z","level":"INFO","msg":"writer: started","stream_id":"g2qj8ry7"}
7
+ {"time":"2026-02-25T13:55:10.309513894Z","level":"INFO","msg":"stream: closing","id":"g2qj8ry7"}
8
+ {"time":"2026-02-25T13:55:10.794171575Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
9
+ {"time":"2026-02-25T13:55:11.099789022Z","level":"INFO","msg":"handler: closed","stream_id":"g2qj8ry7"}
10
+ {"time":"2026-02-25T13:55:11.099961565Z","level":"INFO","msg":"sender: closed","stream_id":"g2qj8ry7"}
11
+ {"time":"2026-02-25T13:55:11.099983596Z","level":"INFO","msg":"stream: closed","id":"g2qj8ry7"}
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/debug.log ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
2
+ 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_setup.py:_flush():81] Configure stats pid to 168367
3
+ 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
+ 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug.log
5
+ 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug-internal.log
6
+ 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:init():844] calling init triggers
7
+ 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
+ config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'depth', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'absgrad', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': True, 'voxelize_activate': True, 'use_depth': True}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'density_control_loss': {'error_score': {'weight': 0.01, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0225_ctxTrain_depth_vggtDistl', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': True, 'vggt_cam_loss': True, 'vggt_distil': True, 'context_view_train': True}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
9
+ 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:init():892] starting backend
10
+ 2026-02-25 13:54:42,153 INFO MainThread:168367 [wandb_init.py:init():895] sending inform_init request
11
+ 2026-02-25 13:54:42,160 INFO MainThread:168367 [wandb_init.py:init():903] backend started and connected
12
+ 2026-02-25 13:54:42,166 INFO MainThread:168367 [wandb_init.py:init():973] updated telemetry
13
+ 2026-02-25 13:54:42,175 INFO MainThread:168367 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
+ 2026-02-25 13:54:43,768 INFO MainThread:168367 [wandb_init.py:init():1042] starting run threads in backend
15
+ 2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_console_start():2524] atexit reg
16
+ 2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_redirect():2373] redirect: wrap_raw
17
+ 2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
+ 2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_redirect():2465] Redirects installed.
19
+ 2026-02-25 13:54:43,896 INFO MainThread:168367 [wandb_init.py:init():1082] run started, returning control to user process
20
+ 2026-02-25 13:55:10,309 INFO wandb-AsyncioManager-main:168367 [service_client.py:_forward_responses():134] Reached EOF.
21
+ 2026-02-25 13:55:10,309 INFO wandb-AsyncioManager-main:168367 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/config.yaml ADDED
@@ -0,0 +1,310 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.25.0
4
+ e:
5
+ q96rx8ljkgqkb464effa7lv06bj6w0fc:
6
+ args:
7
+ - +experiment=re10k_ablation_24v
8
+ - wandb.mode=online
9
+ - wandb.name=ABLATION_0225_ctxTrain_depth_vggtDistl
10
+ - model.encoder.head_mode=depth
11
+ - model.density_control.use_depth=true
12
+ - train.context_view_train=true
13
+ - train.vggt_distil=true
14
+ cpu_count: 128
15
+ cpu_count_logical: 256
16
+ cudaVersion: "13.1"
17
+ disk:
18
+ /:
19
+ total: "1170378588160"
20
+ used: "756693458944"
21
+ email: dna9041@korea.ac.kr
22
+ executable: /venv/main/bin/python
23
+ git:
24
+ commit: 488f8178913cf0d68a295e744f2709f488f576cd
25
+ remote: git@github.com:K-nowing/CVPR2026.git
26
+ gpu: NVIDIA H200
27
+ gpu_count: 8
28
+ gpu_nvidia:
29
+ - architecture: Hopper
30
+ cudaCores: 16896
31
+ memoryTotal: "150754820096"
32
+ name: NVIDIA H200
33
+ uuid: GPU-2649ab80-a3a6-5a1c-0fa5-12bc11bd75e9
34
+ - architecture: Hopper
35
+ cudaCores: 16896
36
+ memoryTotal: "150754820096"
37
+ name: NVIDIA H200
38
+ uuid: GPU-e92921d9-c681-246f-af93-637e0dc938ca
39
+ - architecture: Hopper
40
+ cudaCores: 16896
41
+ memoryTotal: "150754820096"
42
+ name: NVIDIA H200
43
+ uuid: GPU-ffe12ffc-9bb7-82de-5692-1ec0ee2e68d8
44
+ - architecture: Hopper
45
+ cudaCores: 16896
46
+ memoryTotal: "150754820096"
47
+ name: NVIDIA H200
48
+ uuid: GPU-499e5acd-b6ab-2010-c51b-ee9b5aa65825
49
+ - architecture: Hopper
50
+ cudaCores: 16896
51
+ memoryTotal: "150754820096"
52
+ name: NVIDIA H200
53
+ uuid: GPU-3b2522d9-1c72-e49b-2c30-96165680b74a
54
+ - architecture: Hopper
55
+ cudaCores: 16896
56
+ memoryTotal: "150754820096"
57
+ name: NVIDIA H200
58
+ uuid: GPU-a9a280c5-b2f9-dc1e-a8a9-7326a74001ff
59
+ - architecture: Hopper
60
+ cudaCores: 16896
61
+ memoryTotal: "150754820096"
62
+ name: NVIDIA H200
63
+ uuid: GPU-07d0167b-a6a1-1900-2d27-7c6c11598409
64
+ - architecture: Hopper
65
+ cudaCores: 16896
66
+ memoryTotal: "150754820096"
67
+ name: NVIDIA H200
68
+ uuid: GPU-8362a999-20d1-c27b-5d18-032d23f859ab
69
+ host: 27d18dedec6d
70
+ memory:
71
+ total: "1622948257792"
72
+ os: Linux-6.8.0-90-generic-x86_64-with-glibc2.39
73
+ program: -m src.main
74
+ python: CPython 3.12.12
75
+ root: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl
76
+ startedAt: "2026-02-25T13:54:41.861013Z"
77
+ writerId: q96rx8ljkgqkb464effa7lv06bj6w0fc
78
+ m:
79
+ - "1": trainer/global_step
80
+ "6":
81
+ - 3
82
+ "7": []
83
+ - "2": '*'
84
+ "5": 1
85
+ "6":
86
+ - 1
87
+ "7": []
88
+ python_version: 3.12.12
89
+ t:
90
+ "1":
91
+ - 1
92
+ - 41
93
+ - 49
94
+ - 50
95
+ - 106
96
+ "2":
97
+ - 1
98
+ - 41
99
+ - 49
100
+ - 50
101
+ - 106
102
+ "3":
103
+ - 7
104
+ - 13
105
+ - 15
106
+ - 16
107
+ - 66
108
+ "4": 3.12.12
109
+ "5": 0.25.0
110
+ "12": 0.25.0
111
+ "13": linux-x86_64
112
+ checkpointing:
113
+ value:
114
+ every_n_train_steps: 1500
115
+ load: null
116
+ save_top_k: 2
117
+ save_weights_only: false
118
+ data_loader:
119
+ value:
120
+ test:
121
+ batch_size: 1
122
+ num_workers: 4
123
+ persistent_workers: false
124
+ seed: 2345
125
+ train:
126
+ batch_size: 16
127
+ num_workers: 16
128
+ persistent_workers: true
129
+ seed: 1234
130
+ val:
131
+ batch_size: 1
132
+ num_workers: 1
133
+ persistent_workers: true
134
+ seed: 3456
135
+ dataset:
136
+ value:
137
+ re10k:
138
+ augment: true
139
+ background_color:
140
+ - 0
141
+ - 0
142
+ - 0
143
+ baseline_max: 1e+10
144
+ baseline_min: 0.001
145
+ cameras_are_circular: false
146
+ dynamic_context_views: true
147
+ input_image_shape:
148
+ - 256
149
+ - 256
150
+ make_baseline_1: true
151
+ max_context_views_per_gpu: 24
152
+ max_fov: 100
153
+ name: re10k
154
+ original_image_shape:
155
+ - 360
156
+ - 640
157
+ overfit_to_scene: null
158
+ relative_pose: true
159
+ roots:
160
+ - datasets/re10k
161
+ skip_bad_shape: true
162
+ view_sampler:
163
+ initial_max_distance_between_context_views: 25
164
+ initial_min_distance_between_context_views: 25
165
+ max_distance_between_context_views: 90
166
+ min_distance_between_context_views: 45
167
+ min_distance_to_context_views: 0
168
+ name: bounded
169
+ num_context_views: 2
170
+ num_target_set: 3
171
+ num_target_views: 4
172
+ same_target_gap: false
173
+ warm_up_steps: 1000
174
+ density_control_loss:
175
+ value:
176
+ error_score:
177
+ grad_scale: 10000
178
+ log_scale: false
179
+ mode: original
180
+ weight: 0.01
181
+ direct_loss:
182
+ value:
183
+ l1:
184
+ weight: 0.8
185
+ ssim:
186
+ weight: 0.2
187
+ mode:
188
+ value: train
189
+ model:
190
+ value:
191
+ decoder:
192
+ background_color:
193
+ - 0
194
+ - 0
195
+ - 0
196
+ make_scale_invariant: false
197
+ name: splatting_cuda
198
+ density_control:
199
+ aggregation_mode: mean
200
+ aux_refine: false
201
+ grad_mode: absgrad
202
+ gs_param_dim: 256
203
+ latent_dim: 128
204
+ mean_dim: 32
205
+ name: density_control_module
206
+ num_heads: 1
207
+ num_latents: 64
208
+ num_level: 3
209
+ num_self_attn_per_block: 2
210
+ refine_error: false
211
+ refinement_hidden_dim: 32
212
+ refinement_layer_num: 1
213
+ refinement_type: voxelize
214
+ score_mode: absgrad
215
+ use_depth: true
216
+ use_mean_features: true
217
+ use_refine_module: true
218
+ voxel_size: 0.001
219
+ voxelize_activate: true
220
+ encoder:
221
+ align_corners: false
222
+ gs_param_dim: 256
223
+ head_mode: depth
224
+ input_image_shape:
225
+ - 518
226
+ - 518
227
+ name: dcsplat
228
+ num_level: 3
229
+ use_voxelize: true
230
+ optimizer:
231
+ value:
232
+ accumulate: 1
233
+ backbone_lr_multiplier: 0.1
234
+ backbone_trainable: T+H
235
+ lr: 0.0002
236
+ warm_up_steps: 25
237
+ render_loss:
238
+ value:
239
+ lpips:
240
+ apply_after_step: 0
241
+ weight: 0.05
242
+ mse:
243
+ weight: 1
244
+ seed:
245
+ value: 111123
246
+ test:
247
+ value:
248
+ align_pose: false
249
+ compute_scores: true
250
+ error_threshold: 0.4
251
+ error_threshold_list:
252
+ - 0.2
253
+ - 0.4
254
+ - 0.6
255
+ - 0.8
256
+ - 1
257
+ nvs_view_N_list:
258
+ - 3
259
+ - 6
260
+ - 16
261
+ - 32
262
+ - 64
263
+ output_path: test/ablation/re10k
264
+ pose_align_steps: 100
265
+ pred_intrinsic: false
266
+ rot_opt_lr: 0.005
267
+ save_active_mask_image: false
268
+ save_compare: false
269
+ save_error_score_image: false
270
+ save_image: false
271
+ save_video: false
272
+ threshold_mode: ratio
273
+ trans_opt_lr: 0.005
274
+ train:
275
+ value:
276
+ align_corners: false
277
+ beta_dist_param:
278
+ - 0.5
279
+ - 4
280
+ cam_scale_mode: sum
281
+ camera_loss: 10
282
+ context_view_train: true
283
+ ext_scale_detach: false
284
+ extended_visualization: false
285
+ intrinsic_scaling: false
286
+ one_sample_validation: null
287
+ print_log_every_n_steps: 10
288
+ scene_scale_reg_loss: 0.01
289
+ train_aux: true
290
+ train_gs_num: 1
291
+ train_target_set: true
292
+ use_refine_aux: false
293
+ verbose: false
294
+ vggt_cam_loss: true
295
+ vggt_distil: true
296
+ trainer:
297
+ value:
298
+ gradient_clip_val: 0.5
299
+ max_steps: 3001
300
+ num_nodes: 1
301
+ val_check_interval: 250
302
+ wandb:
303
+ value:
304
+ entity: scene-representation-group
305
+ mode: online
306
+ name: ABLATION_0225_ctxTrain_depth_vggtDistl
307
+ project: DCSplat
308
+ tags:
309
+ - re10k
310
+ - 256x256
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/media/images/active_mask_imgs_1_690a5f29fe9230f3ead3.png ADDED

Git LFS Details

  • SHA256: 690a5f29fe9230f3ead382844289ad1899dfe18260121766c39a025ce2d4d666
  • Pointer size: 131 Bytes
  • Size of remote file: 459 kB
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/media/images/comparison_0_e330894a93360b5335a1.png ADDED

Git LFS Details

  • SHA256: e330894a93360b5335a1b717dcffcd8052fe0dd86b7e75c58e226e943720bca4
  • Pointer size: 132 Bytes
  • Size of remote file: 1.66 MB
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/media/images/error_scores_2_bf819ed247f7546e0637.png ADDED

Git LFS Details

  • SHA256: bf819ed247f7546e0637f42c82a216400ea84ba86b3bd04140fd72d3da4c0adb
  • Pointer size: 131 Bytes
  • Size of remote file: 387 kB
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/output.log ADDED
@@ -0,0 +1,140 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7]
2
+
3
+ | Name | Type | Params | Mode
4
+ ------------------------------------------------------------------------
5
+ 0 | encoder | OurSplat | 888 M | train
6
+ 1 | density_control_module | DensityControlModule | 2.6 M | train
7
+ 2 | decoder | DecoderSplattingCUDA | 0 | train
8
+ 3 | render_losses | ModuleList | 0 | train
9
+ 4 | density_control_losses | ModuleList | 0 | train
10
+ 5 | direct_losses | ModuleList | 0 | train
11
+ 6 | distill_aggregator | Aggregator | 909 M | train
12
+ 7 | distill_camera_head | CameraHead | 216 M | train
13
+ 8 | distill_depth_head | DPTHead | 32.7 M | train
14
+ 9 | loss_distill | DistillLoss | 0 | train
15
+ ------------------------------------------------------------------------
16
+ 891 M Trainable params
17
+ 1.2 B Non-trainable params
18
+ 2.0 B Total params
19
+ 8,196.093 Total estimated model params size (MB)
20
+ 2778 Modules in train mode
21
+ 522 Modules in eval mode
22
+ Sanity Checking: | | 0/? [00:00<?, ?it/s][2026-02-25 13:54:48,808][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=31` in the `DataLoader` to improve performance.
23
+
24
+ [2026-02-25 13:54:48,809][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
25
+ warnings.warn( # warn only once
26
+
27
+ Validation epoch start on rank 0
28
+ Sanity Checking DataLoader 0: 0%| | 0/1 [00:00<?, ?it/s]validation step 0; scene = ['306e2b7785657539'];
29
+ target intrinsic: tensor(0.8595, device='cuda:0') tensor(0.8597, device='cuda:0')
30
+ pred intrinsic: tensor(0.8779, device='cuda:0') tensor(0.8773, device='cuda:0')
31
+ [rank0]:W0225 13:54:51.619000 168367 site-packages/torch/utils/cpp_extension.py:2425] TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
32
+ [rank0]:W0225 13:54:51.619000 168367 site-packages/torch/utils/cpp_extension.py:2425] If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'] to specific architectures.
33
+ [2026-02-25 13:54:51,686][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
34
+ result[selector] = overlay
35
+
36
+ [2026-02-25 13:54:51,695][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
37
+
38
+ Setting up [LPIPS] perceptual loss: trunk [vgg], v[0.1], spatial [off]
39
+ [2026-02-25 13:54:51,696][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
40
+ warnings.warn(
41
+
42
+ [2026-02-25 13:54:51,696][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
43
+ warnings.warn(msg)
44
+
45
+ Loading model from: /venv/main/lib/python3.12/site-packages/lpips/weights/v0.1/vgg.pth
46
+ [2026-02-25 13:54:53,564][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4322.)
47
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
48
+
49
+ Sanity Checking DataLoader 0: 100%|████████████████████████████████████████████████████████████████████| 1/1 [00:04<00:00, 0.23it/s][2026-02-25 13:54:53,845][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
50
+
51
+ [2026-02-25 13:54:53,847][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
52
+
53
+ [2026-02-25 13:54:53,847][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
54
+
55
+ [2026-02-25 13:54:53,848][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
56
+
57
+ [2026-02-25 13:54:53,848][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
58
+
59
+ Epoch 0: | | 0/? [00:00<?, ?it/s]context = [[34, 36, 50, 53, 54, 60, 63, 70, 76, 78, 79, 80, 81, 88, 92, 94, 102, 110, 112, 114, 122, 125, 126, 131]]target = [[126, 96, 109, 55, 99, 116, 43, 60, 113, 85, 103, 90, 130, 62, 76, 123, 35, 102, 125, 128, 98, 67, 129, 79]]
60
+ Error executing job with overrides: ['+experiment=re10k_ablation_24v', 'wandb.mode=online', 'wandb.name=ABLATION_0225_ctxTrain_depth_vggtDistl', 'model.encoder.head_mode=depth', 'model.density_control.use_depth=true', 'train.context_view_train=true', 'train.vggt_distil=true']
61
+ Traceback (most recent call last):
62
+ File "/workspace/code/CVPR2026/src/main.py", line 226, in train
63
+ trainer.fit(model_wrapper, datamodule=data_module)#, ckpt_path=checkpoint_path)
64
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
65
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 561, in fit
66
+ call._call_and_handle_interrupt(
67
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/call.py", line 47, in _call_and_handle_interrupt
68
+ return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs)
69
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
70
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch
71
+ return function(*args, **kwargs)
72
+ ^^^^^^^^^^^^^^^^^^^^^^^^^
73
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 599, in _fit_impl
74
+ self._run(model, ckpt_path=ckpt_path)
75
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 1012, in _run
76
+ results = self._run_stage()
77
+ ^^^^^^^^^^^^^^^^^
78
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 1056, in _run_stage
79
+ self.fit_loop.run()
80
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run
81
+ self.advance()
82
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/fit_loop.py", line 455, in advance
83
+ self.epoch_loop.run(self._data_fetcher)
84
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 150, in run
85
+ self.advance(data_fetcher)
86
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 322, in advance
87
+ batch_output = self.manual_optimization.run(kwargs)
88
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
89
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/optimization/manual.py", line 94, in run
90
+ self.advance(kwargs)
91
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/optimization/manual.py", line 114, in advance
92
+ training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values())
93
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
94
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/call.py", line 328, in _call_strategy_hook
95
+ output = fn(*args, **kwargs)
96
+ ^^^^^^^^^^^^^^^^^^^
97
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step
98
+ return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs)
99
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
100
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__
101
+ wrapper_output = wrapper_module(*args, **kwargs)
102
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
103
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl
104
+ return self._call_impl(*args, **kwargs)
105
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
106
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl
107
+ return forward_call(*args, **kwargs)
108
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
109
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/parallel/distributed.py", line 1648, in forward
110
+ else self._run_ddp_forward(*inputs, **kwargs)
111
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
112
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/parallel/distributed.py", line 1474, in _run_ddp_forward
113
+ return self.module(*inputs, **kwargs) # type: ignore[index]
114
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
115
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl
116
+ return self._call_impl(*args, **kwargs)
117
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
118
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl
119
+ return forward_call(*args, **kwargs)
120
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
121
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward
122
+ out = method(*_args, **_kwargs)
123
+ ^^^^^^^^^^^^^^^^^^^^^^^^^
124
+ File "/venv/main/lib/python3.12/site-packages/jaxtyping/_decorator.py", line 562, in wrapped_fn
125
+ return wrapped_fn_impl(args, kwargs, bound, memos)
126
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
127
+ File "/venv/main/lib/python3.12/site-packages/jaxtyping/_decorator.py", line 486, in wrapped_fn_impl
128
+ out = fn(*args, **kwargs)
129
+ ^^^^^^^^^^^^^^^^^^^
130
+ File "/workspace/code/CVPR2026/src/model/model_wrapper.py", line 563, in training_step
131
+ depth_consis_loss = self.render_losses[2].forward(output, batch, depth_dict, self.global_step)
132
+ ~~~~~~~~~~~~~~~~~~^^^
133
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/container.py", line 377, in __getitem__
134
+ return self._modules[self._get_abs_string_index(idx)]
135
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
136
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/container.py", line 361, in _get_abs_string_index
137
+ raise IndexError(f"index {idx} is out of range")
138
+ IndexError: index 2 is out of range
139
+
140
+ Set the environment variable HYDRA_FULL_ERROR=1 for a complete stack trace.
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/requirements.txt ADDED
@@ -0,0 +1,172 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wheel==0.45.1
2
+ pytz==2025.2
3
+ easydict==1.13
4
+ antlr4-python3-runtime==4.9.3
5
+ wadler_lindig==0.1.7
6
+ urllib3==2.5.0
7
+ tzdata==2025.2
8
+ typing-inspection==0.4.1
9
+ tabulate==0.9.0
10
+ smmap==5.0.2
11
+ kornia_rs==0.1.9
12
+ setuptools==78.1.1
13
+ safetensors==0.5.3
14
+ PyYAML==6.0.2
15
+ PySocks==1.7.1
16
+ pyparsing==3.2.5
17
+ pydantic_core==2.33.2
18
+ pycparser==2.23
19
+ protobuf==6.32.1
20
+ propcache==0.3.2
21
+ proglog==0.1.12
22
+ fsspec==2024.6.1
23
+ platformdirs==4.4.0
24
+ pip==25.2
25
+ pillow==10.4.0
26
+ frozenlist==1.7.0
27
+ packaging==24.2
28
+ opt_einsum==3.4.0
29
+ numpy==1.26.4
30
+ ninja==1.13.0
31
+ fonttools==4.60.0
32
+ networkx==3.4.2
33
+ multidict==6.6.4
34
+ mdurl==0.1.2
35
+ MarkupSafe==3.0.2
36
+ kiwisolver==1.4.9
37
+ imageio-ffmpeg==0.6.0
38
+ idna==3.7
39
+ hf-xet==1.1.10
40
+ gmpy2==2.2.1
41
+ einops==0.8.1
42
+ filelock==3.17.0
43
+ decorator==4.4.2
44
+ dacite==1.9.2
45
+ cycler==0.12.1
46
+ colorama==0.4.6
47
+ click==8.3.0
48
+ nvidia-nvtx-cu12==12.8.90
49
+ charset-normalizer==3.3.2
50
+ certifi==2025.8.3
51
+ beartype==0.19.0
52
+ attrs==25.3.0
53
+ async-timeout==5.0.1
54
+ annotated-types==0.7.0
55
+ aiohappyeyeballs==2.6.1
56
+ yarl==1.20.1
57
+ tifffile==2025.5.10
58
+ sentry-sdk==2.39.0
59
+ scipy==1.15.3
60
+ pydantic==2.11.9
61
+ pandas==2.3.2
62
+ opencv-python==4.11.0.86
63
+ omegaconf==2.3.0
64
+ markdown-it-py==4.0.0
65
+ lightning-utilities==0.14.3
66
+ lazy_loader==0.4
67
+ jaxtyping==0.2.37
68
+ imageio==2.37.0
69
+ gitdb==4.0.12
70
+ contourpy==1.3.2
71
+ colorspacious==1.1.2
72
+ cffi==1.17.1
73
+ aiosignal==1.4.0
74
+ scikit-video==1.1.11
75
+ scikit-image==0.25.2
76
+ rich==14.1.0
77
+ moviepy==1.0.3
78
+ matplotlib==3.10.6
79
+ hydra-core==1.3.2
80
+ nvidia-nccl-cu12==2.27.3
81
+ huggingface-hub==0.35.1
82
+ GitPython==3.1.45
83
+ brotlicffi==1.0.9.2
84
+ aiohttp==3.12.15
85
+ torchmetrics==1.8.2
86
+ opt-einsum-fx==0.1.4
87
+ kornia==0.8.1
88
+ pytorch-lightning==2.5.1
89
+ lpips==0.1.4
90
+ e3nn==0.6.0
91
+ lightning==2.5.1
92
+ nvidia-cusparselt-cu12==0.7.1
93
+ triton==3.4.0
94
+ nvidia-nvjitlink-cu12==12.8.93
95
+ nvidia-curand-cu12==10.3.9.90
96
+ nvidia-cufile-cu12==1.13.1.3
97
+ nvidia-cuda-runtime-cu12==12.8.90
98
+ nvidia-cuda-nvrtc-cu12==12.8.93
99
+ nvidia-cuda-cupti-cu12==12.8.90
100
+ nvidia-cublas-cu12==12.8.4.1
101
+ nvidia-cusparse-cu12==12.5.8.93
102
+ nvidia-cufft-cu12==11.3.3.83
103
+ nvidia-cudnn-cu12==9.10.2.21
104
+ nvidia-cusolver-cu12==11.7.3.90
105
+ torch==2.8.0+cu128
106
+ torchvision==0.23.0+cu128
107
+ torchaudio==2.8.0+cu128
108
+ torch_scatter==2.1.2+pt28cu128
109
+ gsplat==1.5.3
110
+ wandb==0.25.0
111
+ cuda-bindings==13.0.3
112
+ cuda-pathfinder==1.3.3
113
+ Jinja2==3.1.6
114
+ mpmath==1.3.0
115
+ nvidia-cublas==13.1.0.3
116
+ nvidia-cuda-cupti==13.0.85
117
+ nvidia-cuda-nvrtc==13.0.88
118
+ nvidia-cuda-runtime==13.0.96
119
+ nvidia-cudnn-cu13==9.15.1.9
120
+ nvidia-cufft==12.0.0.61
121
+ nvidia-cufile==1.15.1.6
122
+ nvidia-curand==10.4.0.35
123
+ nvidia-cusolver==12.0.4.66
124
+ nvidia-cusparse==12.6.3.3
125
+ nvidia-cusparselt-cu13==0.8.0
126
+ nvidia-nccl-cu13==2.28.9
127
+ nvidia-nvjitlink==13.0.88
128
+ nvidia-nvshmem-cu13==3.4.5
129
+ nvidia-nvtx==13.0.85
130
+ requests==2.32.5
131
+ sentencepiece==0.2.1
132
+ sympy==1.14.0
133
+ torchcodec==0.10.0
134
+ torchdata==0.10.0
135
+ torchtext==0.6.0
136
+ anyio==4.12.0
137
+ asttokens==3.0.1
138
+ comm==0.2.3
139
+ debugpy==1.8.19
140
+ executing==2.2.1
141
+ h11==0.16.0
142
+ httpcore==1.0.9
143
+ httpx==0.28.1
144
+ ipykernel==7.1.0
145
+ ipython==9.8.0
146
+ ipython_pygments_lexers==1.1.1
147
+ ipywidgets==8.1.8
148
+ jedi==0.19.2
149
+ jupyter_client==8.7.0
150
+ jupyter_core==5.9.1
151
+ jupyterlab_widgets==3.0.16
152
+ matplotlib-inline==0.2.1
153
+ nest-asyncio==1.6.0
154
+ parso==0.8.5
155
+ pexpect==4.9.0
156
+ prompt_toolkit==3.0.52
157
+ psutil==7.2.1
158
+ ptyprocess==0.7.0
159
+ pure_eval==0.2.3
160
+ Pygments==2.19.2
161
+ python-dateutil==2.9.0.post0
162
+ pyzmq==27.1.0
163
+ shellingham==1.5.4
164
+ six==1.17.0
165
+ stack-data==0.6.3
166
+ tornado==6.5.4
167
+ tqdm==4.67.1
168
+ traitlets==5.14.3
169
+ typer-slim==0.21.0
170
+ typing_extensions==4.15.0
171
+ wcwidth==0.2.14
172
+ widgetsnbextension==4.0.15
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/wandb-metadata.json ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.8.0-90-generic-x86_64-with-glibc2.39",
3
+ "python": "CPython 3.12.12",
4
+ "startedAt": "2026-02-25T13:54:41.861013Z",
5
+ "args": [
6
+ "+experiment=re10k_ablation_24v",
7
+ "wandb.mode=online",
8
+ "wandb.name=ABLATION_0225_ctxTrain_depth_vggtDistl",
9
+ "model.encoder.head_mode=depth",
10
+ "model.density_control.use_depth=true",
11
+ "train.context_view_train=true",
12
+ "train.vggt_distil=true"
13
+ ],
14
+ "program": "-m src.main",
15
+ "git": {
16
+ "remote": "git@github.com:K-nowing/CVPR2026.git",
17
+ "commit": "488f8178913cf0d68a295e744f2709f488f576cd"
18
+ },
19
+ "email": "dna9041@korea.ac.kr",
20
+ "root": "/workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl",
21
+ "host": "27d18dedec6d",
22
+ "executable": "/venv/main/bin/python",
23
+ "cpu_count": 128,
24
+ "cpu_count_logical": 256,
25
+ "gpu": "NVIDIA H200",
26
+ "gpu_count": 8,
27
+ "disk": {
28
+ "/": {
29
+ "total": "1170378588160",
30
+ "used": "756693458944"
31
+ }
32
+ },
33
+ "memory": {
34
+ "total": "1622948257792"
35
+ },
36
+ "gpu_nvidia": [
37
+ {
38
+ "name": "NVIDIA H200",
39
+ "memoryTotal": "150754820096",
40
+ "cudaCores": 16896,
41
+ "architecture": "Hopper",
42
+ "uuid": "GPU-2649ab80-a3a6-5a1c-0fa5-12bc11bd75e9"
43
+ },
44
+ {
45
+ "name": "NVIDIA H200",
46
+ "memoryTotal": "150754820096",
47
+ "cudaCores": 16896,
48
+ "architecture": "Hopper",
49
+ "uuid": "GPU-e92921d9-c681-246f-af93-637e0dc938ca"
50
+ },
51
+ {
52
+ "name": "NVIDIA H200",
53
+ "memoryTotal": "150754820096",
54
+ "cudaCores": 16896,
55
+ "architecture": "Hopper",
56
+ "uuid": "GPU-ffe12ffc-9bb7-82de-5692-1ec0ee2e68d8"
57
+ },
58
+ {
59
+ "name": "NVIDIA H200",
60
+ "memoryTotal": "150754820096",
61
+ "cudaCores": 16896,
62
+ "architecture": "Hopper",
63
+ "uuid": "GPU-499e5acd-b6ab-2010-c51b-ee9b5aa65825"
64
+ },
65
+ {
66
+ "name": "NVIDIA H200",
67
+ "memoryTotal": "150754820096",
68
+ "cudaCores": 16896,
69
+ "architecture": "Hopper",
70
+ "uuid": "GPU-3b2522d9-1c72-e49b-2c30-96165680b74a"
71
+ },
72
+ {
73
+ "name": "NVIDIA H200",
74
+ "memoryTotal": "150754820096",
75
+ "cudaCores": 16896,
76
+ "architecture": "Hopper",
77
+ "uuid": "GPU-a9a280c5-b2f9-dc1e-a8a9-7326a74001ff"
78
+ },
79
+ {
80
+ "name": "NVIDIA H200",
81
+ "memoryTotal": "150754820096",
82
+ "cudaCores": 16896,
83
+ "architecture": "Hopper",
84
+ "uuid": "GPU-07d0167b-a6a1-1900-2d27-7c6c11598409"
85
+ },
86
+ {
87
+ "name": "NVIDIA H200",
88
+ "memoryTotal": "150754820096",
89
+ "cudaCores": 16896,
90
+ "architecture": "Hopper",
91
+ "uuid": "GPU-8362a999-20d1-c27b-5d18-032d23f859ab"
92
+ }
93
+ ],
94
+ "cudaVersion": "13.1",
95
+ "writerId": "q96rx8ljkgqkb464effa7lv06bj6w0fc"
96
+ }
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"trainer/global_step":0,"active_mask_imgs":{"width":536,"height":800,"format":"png","count":1,"filenames":["media/images/active_mask_imgs_1_690a5f29fe9230f3ead3.png"],"captions":["306e2b7785657539"],"_type":"images/separated"},"error_scores":{"_type":"images/separated","width":800,"height":536,"format":"png","count":1,"filenames":["media/images/error_scores_2_bf819ed247f7546e0637.png"],"captions":["306e2b7785657539"]},"_wandb":{"runtime":26},"_runtime":26,"_timestamp":1.7720276938425457e+09,"_step":2,"comparison":{"format":"png","count":1,"filenames":["media/images/comparison_0_e330894a93360b5335a1.png"],"captions":["306e2b7785657539"],"_type":"images/separated","width":1064,"height":1098}}
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug-core.log ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-02-25T13:54:41.978700745Z","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpw2b5bx5a/port-168367.txt","pid":168367,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2026-02-25T13:54:41.980052668Z","level":"INFO","msg":"server: will exit if parent process dies","ppid":168367}
3
+ {"time":"2026-02-25T13:54:41.980043458Z","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-168367-170905-176849560/socket","Net":"unix"}}
4
+ {"time":"2026-02-25T13:54:42.152857208Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2026-02-25T13:54:42.164042466Z","level":"INFO","msg":"handleInformInit: received","streamId":"g2qj8ry7","id":"1(@)"}
6
+ {"time":"2026-02-25T13:54:42.615403591Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"g2qj8ry7","id":"1(@)"}
7
+ {"time":"2026-02-25T13:54:49.068751173Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"1riv14vupx9g"}
8
+ {"time":"2026-02-25T13:55:10.309398102Z","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
9
+ {"time":"2026-02-25T13:55:10.309487963Z","level":"INFO","msg":"connection: closing","id":"1(@)"}
10
+ {"time":"2026-02-25T13:55:10.309511483Z","level":"INFO","msg":"server is shutting down"}
11
+ {"time":"2026-02-25T13:55:10.309607405Z","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
12
+ {"time":"2026-02-25T13:55:10.309714557Z","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-168367-170905-176849560/socket","Net":"unix"}}
13
+ {"time":"2026-02-25T13:55:11.10141868Z","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
14
+ {"time":"2026-02-25T13:55:11.10145892Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
15
+ {"time":"2026-02-25T13:55:11.101482131Z","level":"INFO","msg":"server is closed"}
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug-internal.log ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-02-25T13:54:42.16430492Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
2
+ {"time":"2026-02-25T13:54:42.614803801Z","level":"INFO","msg":"stream: created new stream","id":"g2qj8ry7"}
3
+ {"time":"2026-02-25T13:54:42.615183417Z","level":"INFO","msg":"handler: started","stream_id":"g2qj8ry7"}
4
+ {"time":"2026-02-25T13:54:42.615386361Z","level":"INFO","msg":"stream: started","id":"g2qj8ry7"}
5
+ {"time":"2026-02-25T13:54:42.615431072Z","level":"INFO","msg":"sender: started","stream_id":"g2qj8ry7"}
6
+ {"time":"2026-02-25T13:54:42.615463052Z","level":"INFO","msg":"writer: started","stream_id":"g2qj8ry7"}
7
+ {"time":"2026-02-25T13:55:10.309513894Z","level":"INFO","msg":"stream: closing","id":"g2qj8ry7"}
8
+ {"time":"2026-02-25T13:55:10.794171575Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
9
+ {"time":"2026-02-25T13:55:11.099789022Z","level":"INFO","msg":"handler: closed","stream_id":"g2qj8ry7"}
10
+ {"time":"2026-02-25T13:55:11.099961565Z","level":"INFO","msg":"sender: closed","stream_id":"g2qj8ry7"}
11
+ {"time":"2026-02-25T13:55:11.099983596Z","level":"INFO","msg":"stream: closed","id":"g2qj8ry7"}
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug.log ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
2
+ 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_setup.py:_flush():81] Configure stats pid to 168367
3
+ 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
+ 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug.log
5
+ 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug-internal.log
6
+ 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:init():844] calling init triggers
7
+ 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
+ config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'depth', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'absgrad', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': True, 'voxelize_activate': True, 'use_depth': True}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'density_control_loss': {'error_score': {'weight': 0.01, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0225_ctxTrain_depth_vggtDistl', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': True, 'vggt_cam_loss': True, 'vggt_distil': True, 'context_view_train': True}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
9
+ 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:init():892] starting backend
10
+ 2026-02-25 13:54:42,153 INFO MainThread:168367 [wandb_init.py:init():895] sending inform_init request
11
+ 2026-02-25 13:54:42,160 INFO MainThread:168367 [wandb_init.py:init():903] backend started and connected
12
+ 2026-02-25 13:54:42,166 INFO MainThread:168367 [wandb_init.py:init():973] updated telemetry
13
+ 2026-02-25 13:54:42,175 INFO MainThread:168367 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
+ 2026-02-25 13:54:43,768 INFO MainThread:168367 [wandb_init.py:init():1042] starting run threads in backend
15
+ 2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_console_start():2524] atexit reg
16
+ 2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_redirect():2373] redirect: wrap_raw
17
+ 2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
+ 2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_redirect():2465] Redirects installed.
19
+ 2026-02-25 13:54:43,896 INFO MainThread:168367 [wandb_init.py:init():1082] run started, returning control to user process
20
+ 2026-02-25 13:55:10,309 INFO wandb-AsyncioManager-main:168367 [service_client.py:_forward_responses():134] Reached EOF.
21
+ 2026-02-25 13:55:10,309 INFO wandb-AsyncioManager-main:168367 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/run-g2qj8ry7.wandb ADDED
Binary file (32.9 kB). View file
 
ABLATION_0225_randomSelect/main.log CHANGED
@@ -114,3 +114,15 @@ bucket_view.sizes() = [256, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered in
114
  [2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
115
  result[selector] = overlay
116
 
 
 
 
 
 
 
 
 
 
 
 
 
 
114
  [2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
115
  result[selector] = overlay
116
 
117
+ [2026-02-25 13:35:42,292][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
118
+ result[selector] = overlay
119
+
120
+ [2026-02-25 13:38:57,687][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
121
+ result[selector] = overlay
122
+
123
+ [2026-02-25 13:51:58,139][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
124
+ result[selector] = overlay
125
+
126
+ [2026-02-25 13:52:01,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
127
+ result[selector] = overlay
128
+
ABLATION_0225_randomSelect/peak_vram_memory.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "peak_memory_allocated_gb": 103.167,
3
+ "peak_memory_reserved_gb": 137.896,
4
+ "total_elapsed_hours": 3.21,
5
+ "mode": "train"
6
+ }
ABLATION_0225_randomSelect/train_ddp_process_1.log CHANGED
@@ -58,3 +58,9 @@ bucket_view.sizes() = [256, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered in
58
  [2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
59
  result[selector] = overlay
60
 
 
 
 
 
 
 
 
58
  [2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
59
  result[selector] = overlay
60
 
61
+ [2026-02-25 13:38:57,685][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
62
+ result[selector] = overlay
63
+
64
+ [2026-02-25 13:52:01,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
65
+ result[selector] = overlay
66
+
ABLATION_0225_randomSelect/train_ddp_process_2.log CHANGED
@@ -58,3 +58,9 @@ bucket_view.sizes() = [256, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered in
58
  [2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
59
  result[selector] = overlay
60
 
 
 
 
 
 
 
 
58
  [2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
59
  result[selector] = overlay
60
 
61
+ [2026-02-25 13:38:57,685][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
62
+ result[selector] = overlay
63
+
64
+ [2026-02-25 13:52:01,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
65
+ result[selector] = overlay
66
+
ABLATION_0225_randomSelect/train_ddp_process_3.log CHANGED
@@ -58,3 +58,9 @@ bucket_view.sizes() = [256, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered in
58
  [2026-02-25 13:26:12,526][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
59
  result[selector] = overlay
60
 
 
 
 
 
 
 
 
58
  [2026-02-25 13:26:12,526][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
59
  result[selector] = overlay
60
 
61
+ [2026-02-25 13:38:57,685][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
62
+ result[selector] = overlay
63
+
64
+ [2026-02-25 13:52:01,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
65
+ result[selector] = overlay
66
+
ABLATION_0225_randomSelect/train_ddp_process_4.log CHANGED
@@ -58,3 +58,9 @@ bucket_view.sizes() = [256, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered in
58
  [2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
59
  result[selector] = overlay
60
 
 
 
 
 
 
 
 
58
  [2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
59
  result[selector] = overlay
60
 
61
+ [2026-02-25 13:38:57,685][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
62
+ result[selector] = overlay
63
+
64
+ [2026-02-25 13:52:01,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
65
+ result[selector] = overlay
66
+
ABLATION_0225_randomSelect/train_ddp_process_5.log CHANGED
@@ -58,3 +58,9 @@ bucket_view.sizes() = [256, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered in
58
  [2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
59
  result[selector] = overlay
60
 
 
 
 
 
 
 
 
58
  [2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
59
  result[selector] = overlay
60
 
61
+ [2026-02-25 13:38:57,685][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
62
+ result[selector] = overlay
63
+
64
+ [2026-02-25 13:52:01,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
65
+ result[selector] = overlay
66
+
ABLATION_0225_randomSelect/train_ddp_process_6.log CHANGED
@@ -58,3 +58,9 @@ bucket_view.sizes() = [256, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered in
58
  [2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
59
  result[selector] = overlay
60
 
 
 
 
 
 
 
 
58
  [2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
59
  result[selector] = overlay
60
 
61
+ [2026-02-25 13:38:57,685][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
62
+ result[selector] = overlay
63
+
64
+ [2026-02-25 13:52:01,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
65
+ result[selector] = overlay
66
+
ABLATION_0225_randomSelect/train_ddp_process_7.log CHANGED
@@ -58,3 +58,9 @@ bucket_view.sizes() = [256, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered in
58
  [2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
59
  result[selector] = overlay
60
 
 
 
 
 
 
 
 
58
  [2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
59
  result[selector] = overlay
60
 
61
+ [2026-02-25 13:38:57,685][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
62
+ result[selector] = overlay
63
+
64
+ [2026-02-25 13:52:01,501][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
65
+ result[selector] = overlay
66
+
ABLATION_0225_randomSelect/wandb/debug-internal.log CHANGED
@@ -4,3 +4,8 @@
4
  {"time":"2026-02-25T10:39:56.464733133Z","level":"INFO","msg":"stream: started","id":"n76tgbfj"}
5
  {"time":"2026-02-25T10:39:56.464769993Z","level":"INFO","msg":"sender: started","stream_id":"n76tgbfj"}
6
  {"time":"2026-02-25T10:39:56.464783063Z","level":"INFO","msg":"writer: started","stream_id":"n76tgbfj"}
 
 
 
 
 
 
4
  {"time":"2026-02-25T10:39:56.464733133Z","level":"INFO","msg":"stream: started","id":"n76tgbfj"}
5
  {"time":"2026-02-25T10:39:56.464769993Z","level":"INFO","msg":"sender: started","stream_id":"n76tgbfj"}
6
  {"time":"2026-02-25T10:39:56.464783063Z","level":"INFO","msg":"writer: started","stream_id":"n76tgbfj"}
7
+ {"time":"2026-02-25T13:52:10.426476781Z","level":"INFO","msg":"stream: closing","id":"n76tgbfj"}
8
+ {"time":"2026-02-25T13:52:11.22551075Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
9
+ {"time":"2026-02-25T13:52:11.503165049Z","level":"INFO","msg":"handler: closed","stream_id":"n76tgbfj"}
10
+ {"time":"2026-02-25T13:52:11.503344582Z","level":"INFO","msg":"sender: closed","stream_id":"n76tgbfj"}
11
+ {"time":"2026-02-25T13:52:11.503360133Z","level":"INFO","msg":"stream: closed","id":"n76tgbfj"}
ABLATION_0225_randomSelect/wandb/debug.log CHANGED
@@ -17,3 +17,5 @@ config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518
17
  2026-02-25 10:39:57,797 INFO MainThread:145562 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
  2026-02-25 10:39:57,797 INFO MainThread:145562 [wandb_run.py:_redirect():2465] Redirects installed.
19
  2026-02-25 10:39:57,800 INFO MainThread:145562 [wandb_init.py:init():1082] run started, returning control to user process
 
 
 
17
  2026-02-25 10:39:57,797 INFO MainThread:145562 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
  2026-02-25 10:39:57,797 INFO MainThread:145562 [wandb_run.py:_redirect():2465] Redirects installed.
19
  2026-02-25 10:39:57,800 INFO MainThread:145562 [wandb_init.py:init():1082] run started, returning control to user process
20
+ 2026-02-25 13:52:10,426 INFO wandb-AsyncioManager-main:145562 [service_client.py:_forward_responses():134] Reached EOF.
21
+ 2026-02-25 13:52:10,426 INFO wandb-AsyncioManager-main:145562 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/config.yaml ADDED
@@ -0,0 +1,307 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.25.0
4
+ e:
5
+ 81z9i9czdg5aqgfjdth6b57yriddztva:
6
+ args:
7
+ - +experiment=re10k_ablation_24v
8
+ - wandb.mode=online
9
+ - wandb.name=ABLATION_0225_randomSelect
10
+ - model.density_control.score_mode=random
11
+ cpu_count: 128
12
+ cpu_count_logical: 256
13
+ cudaVersion: "13.1"
14
+ disk:
15
+ /:
16
+ total: "1170378588160"
17
+ used: "732546121728"
18
+ email: dna9041@korea.ac.kr
19
+ executable: /venv/main/bin/python
20
+ git:
21
+ commit: 2512754c6c27ca5150bf17fbcbdde3f192fd53cc
22
+ remote: git@github.com:K-nowing/CVPR2026.git
23
+ gpu: NVIDIA H200
24
+ gpu_count: 8
25
+ gpu_nvidia:
26
+ - architecture: Hopper
27
+ cudaCores: 16896
28
+ memoryTotal: "150754820096"
29
+ name: NVIDIA H200
30
+ uuid: GPU-2649ab80-a3a6-5a1c-0fa5-12bc11bd75e9
31
+ - architecture: Hopper
32
+ cudaCores: 16896
33
+ memoryTotal: "150754820096"
34
+ name: NVIDIA H200
35
+ uuid: GPU-e92921d9-c681-246f-af93-637e0dc938ca
36
+ - architecture: Hopper
37
+ cudaCores: 16896
38
+ memoryTotal: "150754820096"
39
+ name: NVIDIA H200
40
+ uuid: GPU-ffe12ffc-9bb7-82de-5692-1ec0ee2e68d8
41
+ - architecture: Hopper
42
+ cudaCores: 16896
43
+ memoryTotal: "150754820096"
44
+ name: NVIDIA H200
45
+ uuid: GPU-499e5acd-b6ab-2010-c51b-ee9b5aa65825
46
+ - architecture: Hopper
47
+ cudaCores: 16896
48
+ memoryTotal: "150754820096"
49
+ name: NVIDIA H200
50
+ uuid: GPU-3b2522d9-1c72-e49b-2c30-96165680b74a
51
+ - architecture: Hopper
52
+ cudaCores: 16896
53
+ memoryTotal: "150754820096"
54
+ name: NVIDIA H200
55
+ uuid: GPU-a9a280c5-b2f9-dc1e-a8a9-7326a74001ff
56
+ - architecture: Hopper
57
+ cudaCores: 16896
58
+ memoryTotal: "150754820096"
59
+ name: NVIDIA H200
60
+ uuid: GPU-07d0167b-a6a1-1900-2d27-7c6c11598409
61
+ - architecture: Hopper
62
+ cudaCores: 16896
63
+ memoryTotal: "150754820096"
64
+ name: NVIDIA H200
65
+ uuid: GPU-8362a999-20d1-c27b-5d18-032d23f859ab
66
+ host: 27d18dedec6d
67
+ memory:
68
+ total: "1622948257792"
69
+ os: Linux-6.8.0-90-generic-x86_64-with-glibc2.39
70
+ program: -m src.main
71
+ python: CPython 3.12.12
72
+ root: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_randomSelect
73
+ startedAt: "2026-02-25T10:39:55.441642Z"
74
+ writerId: 81z9i9czdg5aqgfjdth6b57yriddztva
75
+ m:
76
+ - "1": trainer/global_step
77
+ "6":
78
+ - 3
79
+ "7": []
80
+ - "2": '*'
81
+ "5": 1
82
+ "6":
83
+ - 1
84
+ "7": []
85
+ python_version: 3.12.12
86
+ t:
87
+ "1":
88
+ - 1
89
+ - 41
90
+ - 49
91
+ - 50
92
+ - 106
93
+ "2":
94
+ - 1
95
+ - 41
96
+ - 49
97
+ - 50
98
+ - 106
99
+ "3":
100
+ - 7
101
+ - 13
102
+ - 15
103
+ - 16
104
+ - 66
105
+ "4": 3.12.12
106
+ "5": 0.25.0
107
+ "12": 0.25.0
108
+ "13": linux-x86_64
109
+ checkpointing:
110
+ value:
111
+ every_n_train_steps: 1500
112
+ load: null
113
+ save_top_k: 2
114
+ save_weights_only: false
115
+ data_loader:
116
+ value:
117
+ test:
118
+ batch_size: 1
119
+ num_workers: 4
120
+ persistent_workers: false
121
+ seed: 2345
122
+ train:
123
+ batch_size: 16
124
+ num_workers: 16
125
+ persistent_workers: true
126
+ seed: 1234
127
+ val:
128
+ batch_size: 1
129
+ num_workers: 1
130
+ persistent_workers: true
131
+ seed: 3456
132
+ dataset:
133
+ value:
134
+ re10k:
135
+ augment: true
136
+ background_color:
137
+ - 0
138
+ - 0
139
+ - 0
140
+ baseline_max: 1e+10
141
+ baseline_min: 0.001
142
+ cameras_are_circular: false
143
+ dynamic_context_views: true
144
+ input_image_shape:
145
+ - 256
146
+ - 256
147
+ make_baseline_1: true
148
+ max_context_views_per_gpu: 24
149
+ max_fov: 100
150
+ name: re10k
151
+ original_image_shape:
152
+ - 360
153
+ - 640
154
+ overfit_to_scene: null
155
+ relative_pose: true
156
+ roots:
157
+ - datasets/re10k
158
+ skip_bad_shape: true
159
+ view_sampler:
160
+ initial_max_distance_between_context_views: 25
161
+ initial_min_distance_between_context_views: 25
162
+ max_distance_between_context_views: 90
163
+ min_distance_between_context_views: 45
164
+ min_distance_to_context_views: 0
165
+ name: bounded
166
+ num_context_views: 2
167
+ num_target_set: 3
168
+ num_target_views: 4
169
+ same_target_gap: false
170
+ warm_up_steps: 1000
171
+ density_control_loss:
172
+ value:
173
+ error_score:
174
+ grad_scale: 10000
175
+ log_scale: false
176
+ mode: original
177
+ weight: 0.01
178
+ direct_loss:
179
+ value:
180
+ l1:
181
+ weight: 0.8
182
+ ssim:
183
+ weight: 0.2
184
+ mode:
185
+ value: train
186
+ model:
187
+ value:
188
+ decoder:
189
+ background_color:
190
+ - 0
191
+ - 0
192
+ - 0
193
+ make_scale_invariant: false
194
+ name: splatting_cuda
195
+ density_control:
196
+ aggregation_mode: mean
197
+ aux_refine: false
198
+ grad_mode: absgrad
199
+ gs_param_dim: 256
200
+ latent_dim: 128
201
+ mean_dim: 32
202
+ name: density_control_module
203
+ num_heads: 1
204
+ num_latents: 64
205
+ num_level: 3
206
+ num_self_attn_per_block: 2
207
+ refine_error: false
208
+ refinement_hidden_dim: 32
209
+ refinement_layer_num: 1
210
+ refinement_type: voxelize
211
+ score_mode: random
212
+ use_depth: false
213
+ use_mean_features: true
214
+ use_refine_module: true
215
+ voxel_size: 0.001
216
+ voxelize_activate: true
217
+ encoder:
218
+ align_corners: false
219
+ gs_param_dim: 256
220
+ head_mode: pcd
221
+ input_image_shape:
222
+ - 518
223
+ - 518
224
+ name: dcsplat
225
+ num_level: 3
226
+ use_voxelize: true
227
+ optimizer:
228
+ value:
229
+ accumulate: 1
230
+ backbone_lr_multiplier: 0.1
231
+ backbone_trainable: T+H
232
+ lr: 0.0002
233
+ warm_up_steps: 25
234
+ render_loss:
235
+ value:
236
+ lpips:
237
+ apply_after_step: 0
238
+ weight: 0.05
239
+ mse:
240
+ weight: 1
241
+ seed:
242
+ value: 111123
243
+ test:
244
+ value:
245
+ align_pose: false
246
+ compute_scores: true
247
+ error_threshold: 0.4
248
+ error_threshold_list:
249
+ - 0.2
250
+ - 0.4
251
+ - 0.6
252
+ - 0.8
253
+ - 1
254
+ nvs_view_N_list:
255
+ - 3
256
+ - 6
257
+ - 16
258
+ - 32
259
+ - 64
260
+ output_path: test/ablation/re10k
261
+ pose_align_steps: 100
262
+ pred_intrinsic: false
263
+ rot_opt_lr: 0.005
264
+ save_active_mask_image: false
265
+ save_compare: false
266
+ save_error_score_image: false
267
+ save_image: false
268
+ save_video: false
269
+ threshold_mode: ratio
270
+ trans_opt_lr: 0.005
271
+ train:
272
+ value:
273
+ align_corners: false
274
+ beta_dist_param:
275
+ - 0.5
276
+ - 4
277
+ cam_scale_mode: sum
278
+ camera_loss: 10
279
+ context_view_train: false
280
+ ext_scale_detach: false
281
+ extended_visualization: false
282
+ intrinsic_scaling: false
283
+ one_sample_validation: null
284
+ print_log_every_n_steps: 10
285
+ scene_scale_reg_loss: 0.01
286
+ train_aux: true
287
+ train_gs_num: 1
288
+ train_target_set: true
289
+ use_refine_aux: false
290
+ verbose: false
291
+ vggt_cam_loss: true
292
+ vggt_distil: false
293
+ trainer:
294
+ value:
295
+ gradient_clip_val: 0.5
296
+ max_steps: 3001
297
+ num_nodes: 1
298
+ val_check_interval: 250
299
+ wandb:
300
+ value:
301
+ entity: scene-representation-group
302
+ mode: online
303
+ name: ABLATION_0225_randomSelect
304
+ project: DCSplat
305
+ tags:
306
+ - re10k
307
+ - 256x256
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/active_mask_imgs_182_2edf4f17c23388f66672.png ADDED

Git LFS Details

  • SHA256: 2edf4f17c23388f666722bd8fd3a5cd49cfdae5902841810cc427f8d8a60c28c
  • Pointer size: 131 Bytes
  • Size of remote file: 356 kB
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/active_mask_imgs_198_399ec9b2f678832e0977.png ADDED

Git LFS Details

  • SHA256: 399ec9b2f678832e097700ad133f6200572a37cd8f0766a1d997d5d25a125ba2
  • Pointer size: 131 Bytes
  • Size of remote file: 500 kB
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/comparison_197_ddc312f116aae122416e.png ADDED

Git LFS Details

  • SHA256: ddc312f116aae122416e781016d99eb73e9e7ad4c0949fe468787667b8b26485
  • Pointer size: 132 Bytes
  • Size of remote file: 1.07 MB
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/error_scores_183_e47e70d75361cb369a35.png ADDED

Git LFS Details

  • SHA256: e47e70d75361cb369a35ed08d2bb5e392e3e5cad8d3c1675d24bc99b4ed51f7e
  • Pointer size: 131 Bytes
  • Size of remote file: 341 kB
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/error_scores_199_b1c8e80e7cf22aab1c67.png ADDED

Git LFS Details

  • SHA256: b1c8e80e7cf22aab1c6711f70391693bc29c3edc04e6c12c856c217f871f528a
  • Pointer size: 131 Bytes
  • Size of remote file: 480 kB
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/train/comparison_188_a456ce8a38a4d0dc65b1.png ADDED

Git LFS Details

  • SHA256: a456ce8a38a4d0dc65b105ba537a4f85cfdc86470ed5eda2f3d0f21dfcb57c47
  • Pointer size: 132 Bytes
  • Size of remote file: 9.34 MB
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/train/comparison_202_b207767914aae1b684b5.png ADDED

Git LFS Details

  • SHA256: b207767914aae1b684b5fba30dc5bf025dea206036803d2bf90cb0941ad404bd
  • Pointer size: 132 Bytes
  • Size of remote file: 3.24 MB
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/train/error_scores_187_347cd0a01cf553ca4f4d.png ADDED

Git LFS Details

  • SHA256: 347cd0a01cf553ca4f4d55850214581ad62a06fd8daedc7b36dbd7a9cc8d7114
  • Pointer size: 132 Bytes
  • Size of remote file: 8.26 MB
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/train/error_scores_201_05456dc706694ddf5d92.png ADDED

Git LFS Details

  • SHA256: 05456dc706694ddf5d9268431dae67d3a6389bf0ddfc6f0e9d8a503568fe9d0a
  • Pointer size: 132 Bytes
  • Size of remote file: 2.87 MB
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"loss/total":0.08273646980524063,"info/global_step":3000,"loss/aux_2/lpips":0.009566728956997395,"loss/final_3dgs/lpips":0.009549135342240334,"train/error_scores":{"width":1328,"height":2120,"format":"png","count":1,"filenames":["media/images/train/error_scores_201_05456dc706694ddf5d92.png"],"captions":[["0621c7675fab1418"]],"_type":"images/separated"},"_runtime":11532,"lr-AdamW/pg1-momentum":0.9,"lr-AdamW/pg2-momentum":0.9,"train/psnr_probabilistic":20.13189697265625,"loss/aux_1/error_score":0.4180681109428406,"loss/aux_1/lpips":0.009899786673486233,"val/ssim":0.7995258569717407,"loss/scene_scale_reg":0.00020948528253939003,"val/lpips":0.23768818378448486,"comparison":{"filenames":["media/images/comparison_197_ddc312f116aae122416e.png"],"captions":["a76028640ffa1ef9"],"_type":"images/separated","width":1064,"height":1098,"format":"png","count":1},"trainer/global_step":3001,"loss/final_3dgs/mse":0.010063939727842808,"loss/aux_0/error_score":0.7389501333236694,"train/scene_scale":1.003704309463501,"val/psnr":21.38949203491211,"loss/camera":0.0002018376690102741,"active_mask_imgs":{"captions":["a76028640ffa1ef9"],"_type":"images/separated","width":536,"height":800,"format":"png","count":1,"filenames":["media/images/active_mask_imgs_198_399ec9b2f678832e0977.png"]},"_timestamp":1.7720275236574388e+09,"loss/aux_1/mse":0.010359433479607105,"loss/aux_0/mse":0.00995652750134468,"_wandb":{"runtime":11532},"loss/aux_2/mse":0.010155964642763138,"lr-AdamW/pg2":2e-05,"error_scores":{"width":800,"height":536,"format":"png","count":1,"filenames":["media/images/error_scores_199_b1c8e80e7cf22aab1c67.png"],"captions":["a76028640ffa1ef9"],"_type":"images/separated"},"loss/aux_0/lpips":0.010957089252769947,"epoch":0,"val/gaussian_num_ratio":0.39984893798828125,"train/comparison":{"format":"png","count":1,"filenames":["media/images/train/comparison_202_b207767914aae1b684b5.png"],"captions":[["0621c7675fab1418"]],"_type":"images/separated","width":1328,"height":2154},"_step":202,"lr-AdamW/pg1":2.003594834351718e-05}
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/logs/debug-core.log CHANGED
@@ -5,3 +5,11 @@
5
  {"time":"2026-02-25T10:39:55.743235764Z","level":"INFO","msg":"handleInformInit: received","streamId":"n76tgbfj","id":"1(@)"}
6
  {"time":"2026-02-25T10:39:56.464752793Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"n76tgbfj","id":"1(@)"}
7
  {"time":"2026-02-25T10:40:02.800543333Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"rz66kc5pj9qh"}
 
 
 
 
 
 
 
 
 
5
  {"time":"2026-02-25T10:39:55.743235764Z","level":"INFO","msg":"handleInformInit: received","streamId":"n76tgbfj","id":"1(@)"}
6
  {"time":"2026-02-25T10:39:56.464752793Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"n76tgbfj","id":"1(@)"}
7
  {"time":"2026-02-25T10:40:02.800543333Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"rz66kc5pj9qh"}
8
+ {"time":"2026-02-25T13:52:10.426350249Z","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
9
+ {"time":"2026-02-25T13:52:10.426467411Z","level":"INFO","msg":"connection: closing","id":"1(@)"}
10
+ {"time":"2026-02-25T13:52:10.426509722Z","level":"INFO","msg":"server is shutting down"}
11
+ {"time":"2026-02-25T13:52:10.426527162Z","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
12
+ {"time":"2026-02-25T13:52:10.426686075Z","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-145562-147994-3073551312/socket","Net":"unix"}}
13
+ {"time":"2026-02-25T13:52:11.504270148Z","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
14
+ {"time":"2026-02-25T13:52:11.504299818Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
15
+ {"time":"2026-02-25T13:52:11.504311038Z","level":"INFO","msg":"server is closed"}
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/logs/debug-internal.log CHANGED
@@ -4,3 +4,8 @@
4
  {"time":"2026-02-25T10:39:56.464733133Z","level":"INFO","msg":"stream: started","id":"n76tgbfj"}
5
  {"time":"2026-02-25T10:39:56.464769993Z","level":"INFO","msg":"sender: started","stream_id":"n76tgbfj"}
6
  {"time":"2026-02-25T10:39:56.464783063Z","level":"INFO","msg":"writer: started","stream_id":"n76tgbfj"}
 
 
 
 
 
 
4
  {"time":"2026-02-25T10:39:56.464733133Z","level":"INFO","msg":"stream: started","id":"n76tgbfj"}
5
  {"time":"2026-02-25T10:39:56.464769993Z","level":"INFO","msg":"sender: started","stream_id":"n76tgbfj"}
6
  {"time":"2026-02-25T10:39:56.464783063Z","level":"INFO","msg":"writer: started","stream_id":"n76tgbfj"}
7
+ {"time":"2026-02-25T13:52:10.426476781Z","level":"INFO","msg":"stream: closing","id":"n76tgbfj"}
8
+ {"time":"2026-02-25T13:52:11.22551075Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
9
+ {"time":"2026-02-25T13:52:11.503165049Z","level":"INFO","msg":"handler: closed","stream_id":"n76tgbfj"}
10
+ {"time":"2026-02-25T13:52:11.503344582Z","level":"INFO","msg":"sender: closed","stream_id":"n76tgbfj"}
11
+ {"time":"2026-02-25T13:52:11.503360133Z","level":"INFO","msg":"stream: closed","id":"n76tgbfj"}
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/logs/debug.log CHANGED
@@ -17,3 +17,5 @@ config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518
17
  2026-02-25 10:39:57,797 INFO MainThread:145562 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
  2026-02-25 10:39:57,797 INFO MainThread:145562 [wandb_run.py:_redirect():2465] Redirects installed.
19
  2026-02-25 10:39:57,800 INFO MainThread:145562 [wandb_init.py:init():1082] run started, returning control to user process
 
 
 
17
  2026-02-25 10:39:57,797 INFO MainThread:145562 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
  2026-02-25 10:39:57,797 INFO MainThread:145562 [wandb_run.py:_redirect():2465] Redirects installed.
19
  2026-02-25 10:39:57,800 INFO MainThread:145562 [wandb_init.py:init():1082] run started, returning control to user process
20
+ 2026-02-25 13:52:10,426 INFO wandb-AsyncioManager-main:145562 [service_client.py:_forward_responses():134] Reached EOF.
21
+ 2026-02-25 13:52:10,426 INFO wandb-AsyncioManager-main:145562 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.