Knowing commited on
Commit
2370512
·
verified ·
1 Parent(s): a3c539a

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. ABLATION_0302_FreqSelect/checkpoints/epoch_0-step_1500.ckpt +3 -0
  3. ABLATION_0302_FreqSelect/wandb/run-20260302_164522-8h1p5i95/files/media/images/comparison_0_f9d64709298973e4dd8b.png +3 -0
  4. ABLATION_0302_FreqSelect/wandb/run-20260302_164522-8h1p5i95/run-8h1p5i95.wandb +3 -0
  5. ABLATION_0302_noAux/.hydra/config.yaml +188 -0
  6. ABLATION_0302_noAux/.hydra/hydra.yaml +165 -0
  7. ABLATION_0302_noAux/.hydra/overrides.yaml +4 -0
  8. ABLATION_0302_noAux/wandb/debug-internal.log +11 -0
  9. ABLATION_0302_noAux/wandb/debug.log +21 -0
  10. ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/config.yaml +310 -0
  11. ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/active_mask_imgs_168_43fa18ddb9c77705edd2.png +3 -0
  12. ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/active_mask_imgs_92_3031367680d80379ffab.png +3 -0
  13. ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/comparison_167_3dfd8c9e7e3d980df015.png +3 -0
  14. ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/comparison_182_5c2b90f763b7bff88204.png +3 -0
  15. ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/error_scores_169_b17e3324e4e0f9205834.png +3 -0
  16. ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/output.log +0 -0
  17. ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/requirements.txt +173 -0
  18. ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/wandb-metadata.json +93 -0
  19. ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/wandb-summary.json +1 -0
  20. ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug-core.log +15 -0
  21. ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug-internal.log +11 -0
  22. ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug.log +21 -0
  23. ABLATION_0302_noTgtAlign/.hydra/config.yaml +188 -0
  24. ABLATION_0302_noTgtAlign/.hydra/hydra.yaml +165 -0
  25. ABLATION_0302_noTgtAlign/.hydra/overrides.yaml +4 -0
  26. ABLATION_0302_noTgtAlign/main.log +83 -0
  27. ABLATION_0302_noTgtAlign/train_ddp_process_1.log +44 -0
  28. ABLATION_0302_noTgtAlign/train_ddp_process_2.log +44 -0
  29. ABLATION_0302_noTgtAlign/train_ddp_process_3.log +44 -0
  30. ABLATION_0302_noTgtAlign/train_ddp_process_4.log +44 -0
  31. ABLATION_0302_noTgtAlign/train_ddp_process_5.log +44 -0
  32. ABLATION_0302_noTgtAlign/train_ddp_process_6.log +44 -0
  33. ABLATION_0302_noTgtAlign/train_ddp_process_7.log +44 -0
  34. ABLATION_0302_noTgtAlign/wandb/debug-internal.log +6 -0
  35. ABLATION_0302_noTgtAlign/wandb/debug.log +19 -0
  36. ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/config.yaml +311 -0
  37. ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/output.log +72 -0
  38. ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/requirements.txt +173 -0
  39. ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/wandb-metadata.json +93 -0
  40. ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/wandb-summary.json +1 -0
  41. ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/logs/debug-core.log +19 -0
  42. ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/logs/debug-internal.log +12 -0
  43. ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/logs/debug.log +24 -0
  44. ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/run-beokhgg2.wandb +0 -0
  45. ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/config.yaml +311 -0
  46. ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/output.log +117 -0
  47. ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/requirements.txt +173 -0
  48. ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/wandb-metadata.json +93 -0
  49. ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/wandb-summary.json +1 -0
  50. ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/logs/debug-core.log +15 -0
.gitattributes CHANGED
@@ -58,3 +58,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
58
  # Video files - compressed
59
  *.mp4 filter=lfs diff=lfs merge=lfs -text
60
  *.webm filter=lfs diff=lfs merge=lfs -text
 
 
58
  # Video files - compressed
59
  *.mp4 filter=lfs diff=lfs merge=lfs -text
60
  *.webm filter=lfs diff=lfs merge=lfs -text
61
+ ABLATION_0302_FreqSelect/wandb/run-20260302_164522-8h1p5i95/run-8h1p5i95.wandb filter=lfs diff=lfs merge=lfs -text
ABLATION_0302_FreqSelect/checkpoints/epoch_0-step_1500.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9905f9951634363d9e0582b2282eaacefe4f1b93637184b1fd0b8984ee9d77fd
3
+ size 11880857960
ABLATION_0302_FreqSelect/wandb/run-20260302_164522-8h1p5i95/files/media/images/comparison_0_f9d64709298973e4dd8b.png ADDED

Git LFS Details

  • SHA256: f9d64709298973e4dd8be410726c3ba97c5aecddf6fb06bc327c3c12d7494f0f
  • Pointer size: 132 Bytes
  • Size of remote file: 1.7 MB
ABLATION_0302_FreqSelect/wandb/run-20260302_164522-8h1p5i95/run-8h1p5i95.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97304ca2cc941f3f04e6359c6f0e44f3f0af7ad6c82c5fda66a78a456d1bd836
3
+ size 4279158
ABLATION_0302_noAux/.hydra/config.yaml ADDED
@@ -0,0 +1,188 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ encoder:
3
+ name: dcsplat
4
+ input_image_shape:
5
+ - 518
6
+ - 518
7
+ head_mode: pcd
8
+ num_level: 3
9
+ gs_param_dim: 256
10
+ align_corners: false
11
+ use_voxelize: true
12
+ decoder:
13
+ name: splatting_cuda
14
+ background_color:
15
+ - 0.0
16
+ - 0.0
17
+ - 0.0
18
+ make_scale_invariant: false
19
+ density_control:
20
+ name: density_control_module
21
+ mean_dim: 32
22
+ gs_param_dim: 256
23
+ refinement_layer_num: 1
24
+ num_level: 3
25
+ grad_mode: absgrad
26
+ use_mean_features: true
27
+ refinement_type: voxelize
28
+ refinement_hidden_dim: 32
29
+ aggregation_mode: mean
30
+ num_heads: 1
31
+ score_mode: absgrad
32
+ latent_dim: 128
33
+ num_latents: 64
34
+ num_self_attn_per_block: 2
35
+ voxel_size: 0.001
36
+ aux_refine: false
37
+ refine_error: false
38
+ use_refine_module: false
39
+ voxelize_activate: false
40
+ use_depth: false
41
+ render_loss:
42
+ mse:
43
+ weight: 1.0
44
+ lpips:
45
+ weight: 0.05
46
+ apply_after_step: 0
47
+ density_control_loss:
48
+ error_score:
49
+ weight: 0.0001
50
+ log_scale: false
51
+ grad_scale: 10000.0
52
+ mode: original
53
+ direct_loss:
54
+ l1:
55
+ weight: 0.8
56
+ ssim:
57
+ weight: 0.2
58
+ wandb:
59
+ project: DCSplat
60
+ entity: scene-representation-group
61
+ name: ABLATION_0302_noAux
62
+ mode: online
63
+ tags:
64
+ - re10k
65
+ - 256x256
66
+ mode: train
67
+ data_loader:
68
+ train:
69
+ num_workers: 16
70
+ persistent_workers: true
71
+ batch_size: 16
72
+ seed: 1234
73
+ test:
74
+ num_workers: 4
75
+ persistent_workers: false
76
+ batch_size: 1
77
+ seed: 2345
78
+ val:
79
+ num_workers: 1
80
+ persistent_workers: true
81
+ batch_size: 1
82
+ seed: 3456
83
+ optimizer:
84
+ lr: 0.0002
85
+ warm_up_steps: 25
86
+ backbone_lr_multiplier: 0.1
87
+ backbone_trainable: T+H
88
+ accumulate: 1
89
+ checkpointing:
90
+ load: null
91
+ every_n_train_steps: 1500
92
+ save_top_k: 2
93
+ save_weights_only: false
94
+ train:
95
+ extended_visualization: false
96
+ print_log_every_n_steps: 10
97
+ camera_loss: 10.0
98
+ one_sample_validation: null
99
+ align_corners: false
100
+ intrinsic_scaling: false
101
+ verbose: false
102
+ beta_dist_param:
103
+ - 0.5
104
+ - 4.0
105
+ use_refine_aux: false
106
+ train_target_set: true
107
+ train_gs_num: 1
108
+ ext_scale_detach: false
109
+ cam_scale_mode: sum
110
+ scene_scale_reg_loss: 0.01
111
+ train_aux: false
112
+ vggt_cam_loss: true
113
+ vggt_distil: false
114
+ context_view_train: false
115
+ test:
116
+ output_path: test/ablation/re10k
117
+ align_pose: false
118
+ pose_align_steps: 100
119
+ rot_opt_lr: 0.005
120
+ trans_opt_lr: 0.005
121
+ compute_scores: true
122
+ save_image: false
123
+ save_video: false
124
+ save_active_mask_image: false
125
+ save_error_score_image: false
126
+ save_compare: false
127
+ save_gs: false
128
+ save_sample_wise_metrics: true
129
+ pred_intrinsic: false
130
+ error_threshold: 0.4
131
+ error_threshold_list:
132
+ - 0.2
133
+ - 0.4
134
+ - 0.6
135
+ - 0.8
136
+ - 1.0
137
+ threshold_mode: ratio
138
+ nvs_view_N_list:
139
+ - 3
140
+ - 6
141
+ - 16
142
+ - 32
143
+ - 64
144
+ seed: 111123
145
+ trainer:
146
+ max_steps: 3001
147
+ val_check_interval: 250
148
+ gradient_clip_val: 0.5
149
+ num_nodes: 1
150
+ dataset:
151
+ re10k:
152
+ make_baseline_1: true
153
+ relative_pose: true
154
+ augment: true
155
+ background_color:
156
+ - 0.0
157
+ - 0.0
158
+ - 0.0
159
+ overfit_to_scene: null
160
+ skip_bad_shape: true
161
+ view_sampler:
162
+ name: bounded
163
+ num_target_views: 4
164
+ num_context_views: 2
165
+ min_distance_between_context_views: 45
166
+ max_distance_between_context_views: 90
167
+ min_distance_to_context_views: 0
168
+ warm_up_steps: 1000
169
+ initial_min_distance_between_context_views: 25
170
+ initial_max_distance_between_context_views: 25
171
+ same_target_gap: false
172
+ num_target_set: 3
173
+ target_align: true
174
+ name: re10k
175
+ roots:
176
+ - datasets/re10k
177
+ input_image_shape:
178
+ - 256
179
+ - 256
180
+ original_image_shape:
181
+ - 360
182
+ - 640
183
+ cameras_are_circular: false
184
+ baseline_min: 0.001
185
+ baseline_max: 10000000000.0
186
+ max_fov: 100.0
187
+ dynamic_context_views: true
188
+ max_context_views_per_gpu: 24
ABLATION_0302_noAux/.hydra/hydra.yaml ADDED
@@ -0,0 +1,165 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/ablation/re10k/${wandb.name}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - +experiment=re10k_ablation_24v
116
+ - wandb.mode=online
117
+ - wandb.name=ABLATION_0302_noAux
118
+ - train.train_aux=false
119
+ job:
120
+ name: main
121
+ chdir: null
122
+ override_dirname: +experiment=re10k_ablation_24v,train.train_aux=false,wandb.mode=online,wandb.name=ABLATION_0302_noAux
123
+ id: ???
124
+ num: ???
125
+ config_name: main
126
+ env_set: {}
127
+ env_copy: []
128
+ config:
129
+ override_dirname:
130
+ kv_sep: '='
131
+ item_sep: ','
132
+ exclude_keys: []
133
+ runtime:
134
+ version: 1.3.2
135
+ version_base: '1.3'
136
+ cwd: /workspace/code/CVPR2026
137
+ config_sources:
138
+ - path: hydra.conf
139
+ schema: pkg
140
+ provider: hydra
141
+ - path: /workspace/code/CVPR2026/config
142
+ schema: file
143
+ provider: main
144
+ - path: ''
145
+ schema: structured
146
+ provider: schema
147
+ output_dir: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noAux
148
+ choices:
149
+ experiment: re10k_ablation_24v
150
+ dataset@dataset.re10k: re10k
151
+ dataset/view_sampler_dataset_specific_config@dataset.re10k.view_sampler: bounded_re10k
152
+ dataset/view_sampler@dataset.re10k.view_sampler: bounded
153
+ model/density_control: density_control_module
154
+ model/decoder: splatting_cuda
155
+ model/encoder: dcsplat
156
+ hydra/env: default
157
+ hydra/callbacks: null
158
+ hydra/job_logging: default
159
+ hydra/hydra_logging: default
160
+ hydra/hydra_help: default
161
+ hydra/help: default
162
+ hydra/sweeper: basic
163
+ hydra/launcher: basic
164
+ hydra/output: default
165
+ verbose: false
ABLATION_0302_noAux/.hydra/overrides.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ - +experiment=re10k_ablation_24v
2
+ - wandb.mode=online
3
+ - wandb.name=ABLATION_0302_noAux
4
+ - train.train_aux=false
ABLATION_0302_noAux/wandb/debug-internal.log ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-03-03T14:34:53.54334025Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
2
+ {"time":"2026-03-03T14:34:53.958050168Z","level":"INFO","msg":"stream: created new stream","id":"l8o7jl7r"}
3
+ {"time":"2026-03-03T14:34:53.958155019Z","level":"INFO","msg":"handler: started","stream_id":"l8o7jl7r"}
4
+ {"time":"2026-03-03T14:34:53.9583015Z","level":"INFO","msg":"stream: started","id":"l8o7jl7r"}
5
+ {"time":"2026-03-03T14:34:53.95834932Z","level":"INFO","msg":"sender: started","stream_id":"l8o7jl7r"}
6
+ {"time":"2026-03-03T14:34:53.95835085Z","level":"INFO","msg":"writer: started","stream_id":"l8o7jl7r"}
7
+ {"time":"2026-03-03T17:25:37.368615437Z","level":"INFO","msg":"stream: closing","id":"l8o7jl7r"}
8
+ {"time":"2026-03-03T17:25:38.278996749Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
9
+ {"time":"2026-03-03T17:25:38.528169562Z","level":"INFO","msg":"handler: closed","stream_id":"l8o7jl7r"}
10
+ {"time":"2026-03-03T17:25:38.528336264Z","level":"INFO","msg":"sender: closed","stream_id":"l8o7jl7r"}
11
+ {"time":"2026-03-03T17:25:38.528358534Z","level":"INFO","msg":"stream: closed","id":"l8o7jl7r"}
ABLATION_0302_noAux/wandb/debug.log ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
2
+ 2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_setup.py:_flush():81] Configure stats pid to 854556
3
+ 2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
+ 2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug.log
5
+ 2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug-internal.log
6
+ 2026-03-03 14:34:53,294 INFO MainThread:854556 [wandb_init.py:init():844] calling init triggers
7
+ 2026-03-03 14:34:53,294 INFO MainThread:854556 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
+ config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'pcd', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'absgrad', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': False, 'voxelize_activate': False, 'use_depth': False}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'density_control_loss': {'error_score': {'weight': 0.0001, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0302_noAux', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': False, 'vggt_cam_loss': True, 'vggt_distil': False, 'context_view_train': False}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'save_gs': False, 'save_sample_wise_metrics': True, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3, 'target_align': True}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
9
+ 2026-03-03 14:34:53,294 INFO MainThread:854556 [wandb_init.py:init():892] starting backend
10
+ 2026-03-03 14:34:53,535 INFO MainThread:854556 [wandb_init.py:init():895] sending inform_init request
11
+ 2026-03-03 14:34:53,540 INFO MainThread:854556 [wandb_init.py:init():903] backend started and connected
12
+ 2026-03-03 14:34:53,542 INFO MainThread:854556 [wandb_init.py:init():973] updated telemetry
13
+ 2026-03-03 14:34:53,548 INFO MainThread:854556 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
+ 2026-03-03 14:34:54,767 INFO MainThread:854556 [wandb_init.py:init():1042] starting run threads in backend
15
+ 2026-03-03 14:34:54,895 INFO MainThread:854556 [wandb_run.py:_console_start():2524] atexit reg
16
+ 2026-03-03 14:34:54,895 INFO MainThread:854556 [wandb_run.py:_redirect():2373] redirect: wrap_raw
17
+ 2026-03-03 14:34:54,895 INFO MainThread:854556 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
+ 2026-03-03 14:34:54,896 INFO MainThread:854556 [wandb_run.py:_redirect():2465] Redirects installed.
19
+ 2026-03-03 14:34:54,899 INFO MainThread:854556 [wandb_init.py:init():1082] run started, returning control to user process
20
+ 2026-03-03 17:25:37,368 INFO wandb-AsyncioManager-main:854556 [service_client.py:_forward_responses():134] Reached EOF.
21
+ 2026-03-03 17:25:37,368 INFO wandb-AsyncioManager-main:854556 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/config.yaml ADDED
@@ -0,0 +1,310 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.25.0
4
+ e:
5
+ ycd5umfcq5p6rih4lqfm265f2qblrtrn:
6
+ args:
7
+ - +experiment=re10k_ablation_24v
8
+ - wandb.mode=online
9
+ - wandb.name=ABLATION_0302_noAux
10
+ - train.train_aux=false
11
+ cpu_count: 128
12
+ cpu_count_logical: 256
13
+ cudaVersion: "13.0"
14
+ disk:
15
+ /:
16
+ total: "735513149440"
17
+ used: "676972425216"
18
+ email: dna9041@korea.ac.kr
19
+ executable: /venv/main/bin/python
20
+ git:
21
+ commit: 9dfce172a0f8c7ce85e763899f7ef741ecffc454
22
+ remote: git@github.com:K-nowing/CVPR2026.git
23
+ gpu: NVIDIA H200
24
+ gpu_count: 8
25
+ gpu_nvidia:
26
+ - architecture: Hopper
27
+ cudaCores: 16896
28
+ memoryTotal: "150754820096"
29
+ name: NVIDIA H200
30
+ uuid: GPU-9a20101e-d876-facd-5f05-805081aede41
31
+ - architecture: Hopper
32
+ cudaCores: 16896
33
+ memoryTotal: "150754820096"
34
+ name: NVIDIA H200
35
+ uuid: GPU-84736a77-ee75-3324-e4e1-99cc15bfb5e9
36
+ - architecture: Hopper
37
+ cudaCores: 16896
38
+ memoryTotal: "150754820096"
39
+ name: NVIDIA H200
40
+ uuid: GPU-423d3161-cdc4-3fc0-caee-d15cfaa83ca6
41
+ - architecture: Hopper
42
+ cudaCores: 16896
43
+ memoryTotal: "150754820096"
44
+ name: NVIDIA H200
45
+ uuid: GPU-5b0058b2-cdb9-c952-04f9-87dcaa7ea742
46
+ - architecture: Hopper
47
+ cudaCores: 16896
48
+ memoryTotal: "150754820096"
49
+ name: NVIDIA H200
50
+ uuid: GPU-08b37f98-4603-d483-2f2b-fe5311aa42f2
51
+ - architecture: Hopper
52
+ cudaCores: 16896
53
+ memoryTotal: "150754820096"
54
+ name: NVIDIA H200
55
+ uuid: GPU-03273b5b-2fdd-a5fe-4460-c897334ae464
56
+ - architecture: Hopper
57
+ cudaCores: 16896
58
+ memoryTotal: "150754820096"
59
+ name: NVIDIA H200
60
+ uuid: GPU-292d466c-d00d-25a4-28b6-e6c978d3e70c
61
+ - architecture: Hopper
62
+ cudaCores: 16896
63
+ memoryTotal: "150754820096"
64
+ name: NVIDIA H200
65
+ uuid: GPU-46f38561-3148-e442-7f7f-bfe447bab7fe
66
+ host: e9d3310a05da
67
+ memory:
68
+ total: "1622950240256"
69
+ os: Linux-6.8.0-94-generic-x86_64-with-glibc2.39
70
+ program: -m src.main
71
+ python: CPython 3.12.12
72
+ root: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noAux
73
+ startedAt: "2026-03-03T14:34:53.292304Z"
74
+ writerId: ycd5umfcq5p6rih4lqfm265f2qblrtrn
75
+ m:
76
+ - "1": trainer/global_step
77
+ "6":
78
+ - 3
79
+ "7": []
80
+ - "2": '*'
81
+ "5": 1
82
+ "6":
83
+ - 1
84
+ "7": []
85
+ python_version: 3.12.12
86
+ t:
87
+ "1":
88
+ - 1
89
+ - 41
90
+ - 49
91
+ - 50
92
+ - 106
93
+ "2":
94
+ - 1
95
+ - 41
96
+ - 49
97
+ - 50
98
+ - 106
99
+ "3":
100
+ - 7
101
+ - 13
102
+ - 15
103
+ - 16
104
+ - 66
105
+ "4": 3.12.12
106
+ "5": 0.25.0
107
+ "12": 0.25.0
108
+ "13": linux-x86_64
109
+ checkpointing:
110
+ value:
111
+ every_n_train_steps: 1500
112
+ load: null
113
+ save_top_k: 2
114
+ save_weights_only: false
115
+ data_loader:
116
+ value:
117
+ test:
118
+ batch_size: 1
119
+ num_workers: 4
120
+ persistent_workers: false
121
+ seed: 2345
122
+ train:
123
+ batch_size: 16
124
+ num_workers: 16
125
+ persistent_workers: true
126
+ seed: 1234
127
+ val:
128
+ batch_size: 1
129
+ num_workers: 1
130
+ persistent_workers: true
131
+ seed: 3456
132
+ dataset:
133
+ value:
134
+ re10k:
135
+ augment: true
136
+ background_color:
137
+ - 0
138
+ - 0
139
+ - 0
140
+ baseline_max: 1e+10
141
+ baseline_min: 0.001
142
+ cameras_are_circular: false
143
+ dynamic_context_views: true
144
+ input_image_shape:
145
+ - 256
146
+ - 256
147
+ make_baseline_1: true
148
+ max_context_views_per_gpu: 24
149
+ max_fov: 100
150
+ name: re10k
151
+ original_image_shape:
152
+ - 360
153
+ - 640
154
+ overfit_to_scene: null
155
+ relative_pose: true
156
+ roots:
157
+ - datasets/re10k
158
+ skip_bad_shape: true
159
+ view_sampler:
160
+ initial_max_distance_between_context_views: 25
161
+ initial_min_distance_between_context_views: 25
162
+ max_distance_between_context_views: 90
163
+ min_distance_between_context_views: 45
164
+ min_distance_to_context_views: 0
165
+ name: bounded
166
+ num_context_views: 2
167
+ num_target_set: 3
168
+ num_target_views: 4
169
+ same_target_gap: false
170
+ target_align: true
171
+ warm_up_steps: 1000
172
+ density_control_loss:
173
+ value:
174
+ error_score:
175
+ grad_scale: 10000
176
+ log_scale: false
177
+ mode: original
178
+ weight: 0.0001
179
+ direct_loss:
180
+ value:
181
+ l1:
182
+ weight: 0.8
183
+ ssim:
184
+ weight: 0.2
185
+ mode:
186
+ value: train
187
+ model:
188
+ value:
189
+ decoder:
190
+ background_color:
191
+ - 0
192
+ - 0
193
+ - 0
194
+ make_scale_invariant: false
195
+ name: splatting_cuda
196
+ density_control:
197
+ aggregation_mode: mean
198
+ aux_refine: false
199
+ grad_mode: absgrad
200
+ gs_param_dim: 256
201
+ latent_dim: 128
202
+ mean_dim: 32
203
+ name: density_control_module
204
+ num_heads: 1
205
+ num_latents: 64
206
+ num_level: 3
207
+ num_self_attn_per_block: 2
208
+ refine_error: false
209
+ refinement_hidden_dim: 32
210
+ refinement_layer_num: 1
211
+ refinement_type: voxelize
212
+ score_mode: absgrad
213
+ use_depth: false
214
+ use_mean_features: true
215
+ use_refine_module: false
216
+ voxel_size: 0.001
217
+ voxelize_activate: false
218
+ encoder:
219
+ align_corners: false
220
+ gs_param_dim: 256
221
+ head_mode: pcd
222
+ input_image_shape:
223
+ - 518
224
+ - 518
225
+ name: dcsplat
226
+ num_level: 3
227
+ use_voxelize: true
228
+ optimizer:
229
+ value:
230
+ accumulate: 1
231
+ backbone_lr_multiplier: 0.1
232
+ backbone_trainable: T+H
233
+ lr: 0.0002
234
+ warm_up_steps: 25
235
+ render_loss:
236
+ value:
237
+ lpips:
238
+ apply_after_step: 0
239
+ weight: 0.05
240
+ mse:
241
+ weight: 1
242
+ seed:
243
+ value: 111123
244
+ test:
245
+ value:
246
+ align_pose: false
247
+ compute_scores: true
248
+ error_threshold: 0.4
249
+ error_threshold_list:
250
+ - 0.2
251
+ - 0.4
252
+ - 0.6
253
+ - 0.8
254
+ - 1
255
+ nvs_view_N_list:
256
+ - 3
257
+ - 6
258
+ - 16
259
+ - 32
260
+ - 64
261
+ output_path: test/ablation/re10k
262
+ pose_align_steps: 100
263
+ pred_intrinsic: false
264
+ rot_opt_lr: 0.005
265
+ save_active_mask_image: false
266
+ save_compare: false
267
+ save_error_score_image: false
268
+ save_gs: false
269
+ save_image: false
270
+ save_sample_wise_metrics: true
271
+ save_video: false
272
+ threshold_mode: ratio
273
+ trans_opt_lr: 0.005
274
+ train:
275
+ value:
276
+ align_corners: false
277
+ beta_dist_param:
278
+ - 0.5
279
+ - 4
280
+ cam_scale_mode: sum
281
+ camera_loss: 10
282
+ context_view_train: false
283
+ ext_scale_detach: false
284
+ extended_visualization: false
285
+ intrinsic_scaling: false
286
+ one_sample_validation: null
287
+ print_log_every_n_steps: 10
288
+ scene_scale_reg_loss: 0.01
289
+ train_aux: false
290
+ train_gs_num: 1
291
+ train_target_set: true
292
+ use_refine_aux: false
293
+ verbose: false
294
+ vggt_cam_loss: true
295
+ vggt_distil: false
296
+ trainer:
297
+ value:
298
+ gradient_clip_val: 0.5
299
+ max_steps: 3001
300
+ num_nodes: 1
301
+ val_check_interval: 250
302
+ wandb:
303
+ value:
304
+ entity: scene-representation-group
305
+ mode: online
306
+ name: ABLATION_0302_noAux
307
+ project: DCSplat
308
+ tags:
309
+ - re10k
310
+ - 256x256
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/active_mask_imgs_168_43fa18ddb9c77705edd2.png ADDED

Git LFS Details

  • SHA256: 43fa18ddb9c77705edd2814e9de063c8aa4efb817065649d13efc0648486b66d
  • Pointer size: 131 Bytes
  • Size of remote file: 355 kB
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/active_mask_imgs_92_3031367680d80379ffab.png ADDED

Git LFS Details

  • SHA256: 3031367680d80379ffab5719ff677f2b837fd4e785cc0490f1ee1dec1afda210
  • Pointer size: 131 Bytes
  • Size of remote file: 470 kB
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/comparison_167_3dfd8c9e7e3d980df015.png ADDED

Git LFS Details

  • SHA256: 3dfd8c9e7e3d980df0158d695108bdf1ce36442b2efec350eb8984d52b486fed
  • Pointer size: 131 Bytes
  • Size of remote file: 760 kB
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/comparison_182_5c2b90f763b7bff88204.png ADDED

Git LFS Details

  • SHA256: 5c2b90f763b7bff882043d44a5121b8ccb59844c7e9dacbc26b86e5aba06fd31
  • Pointer size: 132 Bytes
  • Size of remote file: 1.05 MB
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/error_scores_169_b17e3324e4e0f9205834.png ADDED

Git LFS Details

  • SHA256: b17e3324e4e0f92058345508c8e0f8cac5c0f1001b91d91ef6600603bbc095fe
  • Pointer size: 131 Bytes
  • Size of remote file: 300 kB
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/output.log ADDED
The diff for this file is too large to render. See raw diff
 
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/requirements.txt ADDED
@@ -0,0 +1,173 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wheel==0.45.1
2
+ pytz==2025.2
3
+ easydict==1.13
4
+ antlr4-python3-runtime==4.9.3
5
+ wadler_lindig==0.1.7
6
+ networkx==3.4.2
7
+ urllib3==2.5.0
8
+ tzdata==2025.2
9
+ typing-inspection==0.4.1
10
+ tabulate==0.9.0
11
+ smmap==5.0.2
12
+ setuptools==78.1.1
13
+ safetensors==0.5.3
14
+ multidict==6.6.4
15
+ PyYAML==6.0.2
16
+ PySocks==1.7.1
17
+ pyparsing==3.2.5
18
+ pydantic_core==2.33.2
19
+ pycparser==2.23
20
+ protobuf==6.32.1
21
+ propcache==0.3.2
22
+ proglog==0.1.12
23
+ platformdirs==4.4.0
24
+ pip==25.2
25
+ mdurl==0.1.2
26
+ pillow==10.4.0
27
+ packaging==24.2
28
+ opt_einsum==3.4.0
29
+ frozenlist==1.7.0
30
+ numpy==1.26.4
31
+ ninja==1.13.0
32
+ MarkupSafe==3.0.2
33
+ kornia_rs==0.1.9
34
+ kiwisolver==1.4.9
35
+ imageio-ffmpeg==0.6.0
36
+ idna==3.7
37
+ fsspec==2024.6.1
38
+ hf-xet==1.1.10
39
+ gmpy2==2.2.1
40
+ fonttools==4.60.0
41
+ triton==3.4.0
42
+ filelock==3.17.0
43
+ einops==0.8.1
44
+ decorator==4.4.2
45
+ dacite==1.9.2
46
+ cycler==0.12.1
47
+ colorama==0.4.6
48
+ click==8.3.0
49
+ nvidia-nvtx-cu12==12.8.90
50
+ charset-normalizer==3.3.2
51
+ certifi==2025.8.3
52
+ beartype==0.19.0
53
+ attrs==25.3.0
54
+ async-timeout==5.0.1
55
+ annotated-types==0.7.0
56
+ aiohappyeyeballs==2.6.1
57
+ yarl==1.20.1
58
+ tifffile==2025.5.10
59
+ sentry-sdk==2.39.0
60
+ scipy==1.15.3
61
+ pydantic==2.11.9
62
+ pandas==2.3.2
63
+ opencv-python==4.11.0.86
64
+ omegaconf==2.3.0
65
+ markdown-it-py==4.0.0
66
+ lightning-utilities==0.14.3
67
+ lazy_loader==0.4
68
+ jaxtyping==0.2.37
69
+ imageio==2.37.0
70
+ gitdb==4.0.12
71
+ contourpy==1.3.2
72
+ colorspacious==1.1.2
73
+ cffi==1.17.1
74
+ aiosignal==1.4.0
75
+ scikit-video==1.1.11
76
+ scikit-image==0.25.2
77
+ rich==14.1.0
78
+ moviepy==1.0.3
79
+ matplotlib==3.10.6
80
+ hydra-core==1.3.2
81
+ huggingface-hub==0.35.1
82
+ GitPython==3.1.45
83
+ brotlicffi==1.0.9.2
84
+ aiohttp==3.12.15
85
+ torchmetrics==1.8.2
86
+ opt-einsum-fx==0.1.4
87
+ kornia==0.8.1
88
+ pytorch-lightning==2.5.1
89
+ lpips==0.1.4
90
+ e3nn==0.6.0
91
+ lightning==2.5.1
92
+ gsplat==1.5.3
93
+ nvidia-cusparselt-cu12==0.7.1
94
+ nvidia-nvjitlink-cu12==12.8.93
95
+ nvidia-nccl-cu12==2.27.3
96
+ nvidia-curand-cu12==10.3.9.90
97
+ nvidia-cufile-cu12==1.13.1.3
98
+ nvidia-cuda-runtime-cu12==12.8.90
99
+ nvidia-cuda-nvrtc-cu12==12.8.93
100
+ nvidia-cuda-cupti-cu12==12.8.90
101
+ nvidia-cublas-cu12==12.8.4.1
102
+ nvidia-cusparse-cu12==12.5.8.93
103
+ nvidia-cufft-cu12==11.3.3.83
104
+ nvidia-cudnn-cu12==9.10.2.21
105
+ nvidia-cusolver-cu12==11.7.3.90
106
+ torch==2.8.0+cu128
107
+ torchvision==0.23.0+cu128
108
+ torchaudio==2.8.0+cu128
109
+ torch_scatter==2.1.2+pt28cu128
110
+ plyfile==1.1.3
111
+ wandb==0.25.0
112
+ cuda-bindings==13.0.3
113
+ cuda-pathfinder==1.3.3
114
+ Jinja2==3.1.6
115
+ mpmath==1.3.0
116
+ nvidia-cublas==13.1.0.3
117
+ nvidia-cuda-cupti==13.0.85
118
+ nvidia-cuda-nvrtc==13.0.88
119
+ nvidia-cuda-runtime==13.0.96
120
+ nvidia-cudnn-cu13==9.15.1.9
121
+ nvidia-cufft==12.0.0.61
122
+ nvidia-cufile==1.15.1.6
123
+ nvidia-curand==10.4.0.35
124
+ nvidia-cusolver==12.0.4.66
125
+ nvidia-cusparse==12.6.3.3
126
+ nvidia-cusparselt-cu13==0.8.0
127
+ nvidia-nccl-cu13==2.28.9
128
+ nvidia-nvjitlink==13.0.88
129
+ nvidia-nvshmem-cu13==3.4.5
130
+ nvidia-nvtx==13.0.85
131
+ requests==2.32.5
132
+ sentencepiece==0.2.1
133
+ sympy==1.14.0
134
+ torchcodec==0.10.0
135
+ torchdata==0.10.0
136
+ torchtext==0.6.0
137
+ anyio==4.12.0
138
+ asttokens==3.0.1
139
+ comm==0.2.3
140
+ debugpy==1.8.19
141
+ executing==2.2.1
142
+ h11==0.16.0
143
+ httpcore==1.0.9
144
+ httpx==0.28.1
145
+ ipykernel==7.1.0
146
+ ipython==9.8.0
147
+ ipython_pygments_lexers==1.1.1
148
+ ipywidgets==8.1.8
149
+ jedi==0.19.2
150
+ jupyter_client==8.7.0
151
+ jupyter_core==5.9.1
152
+ jupyterlab_widgets==3.0.16
153
+ matplotlib-inline==0.2.1
154
+ nest-asyncio==1.6.0
155
+ parso==0.8.5
156
+ pexpect==4.9.0
157
+ prompt_toolkit==3.0.52
158
+ psutil==7.2.1
159
+ ptyprocess==0.7.0
160
+ pure_eval==0.2.3
161
+ Pygments==2.19.2
162
+ python-dateutil==2.9.0.post0
163
+ pyzmq==27.1.0
164
+ shellingham==1.5.4
165
+ six==1.17.0
166
+ stack-data==0.6.3
167
+ tornado==6.5.4
168
+ tqdm==4.67.1
169
+ traitlets==5.14.3
170
+ typer-slim==0.21.0
171
+ typing_extensions==4.15.0
172
+ wcwidth==0.2.14
173
+ widgetsnbextension==4.0.15
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/wandb-metadata.json ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.8.0-94-generic-x86_64-with-glibc2.39",
3
+ "python": "CPython 3.12.12",
4
+ "startedAt": "2026-03-03T14:34:53.292304Z",
5
+ "args": [
6
+ "+experiment=re10k_ablation_24v",
7
+ "wandb.mode=online",
8
+ "wandb.name=ABLATION_0302_noAux",
9
+ "train.train_aux=false"
10
+ ],
11
+ "program": "-m src.main",
12
+ "git": {
13
+ "remote": "git@github.com:K-nowing/CVPR2026.git",
14
+ "commit": "9dfce172a0f8c7ce85e763899f7ef741ecffc454"
15
+ },
16
+ "email": "dna9041@korea.ac.kr",
17
+ "root": "/workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noAux",
18
+ "host": "e9d3310a05da",
19
+ "executable": "/venv/main/bin/python",
20
+ "cpu_count": 128,
21
+ "cpu_count_logical": 256,
22
+ "gpu": "NVIDIA H200",
23
+ "gpu_count": 8,
24
+ "disk": {
25
+ "/": {
26
+ "total": "735513149440",
27
+ "used": "676972425216"
28
+ }
29
+ },
30
+ "memory": {
31
+ "total": "1622950240256"
32
+ },
33
+ "gpu_nvidia": [
34
+ {
35
+ "name": "NVIDIA H200",
36
+ "memoryTotal": "150754820096",
37
+ "cudaCores": 16896,
38
+ "architecture": "Hopper",
39
+ "uuid": "GPU-9a20101e-d876-facd-5f05-805081aede41"
40
+ },
41
+ {
42
+ "name": "NVIDIA H200",
43
+ "memoryTotal": "150754820096",
44
+ "cudaCores": 16896,
45
+ "architecture": "Hopper",
46
+ "uuid": "GPU-84736a77-ee75-3324-e4e1-99cc15bfb5e9"
47
+ },
48
+ {
49
+ "name": "NVIDIA H200",
50
+ "memoryTotal": "150754820096",
51
+ "cudaCores": 16896,
52
+ "architecture": "Hopper",
53
+ "uuid": "GPU-423d3161-cdc4-3fc0-caee-d15cfaa83ca6"
54
+ },
55
+ {
56
+ "name": "NVIDIA H200",
57
+ "memoryTotal": "150754820096",
58
+ "cudaCores": 16896,
59
+ "architecture": "Hopper",
60
+ "uuid": "GPU-5b0058b2-cdb9-c952-04f9-87dcaa7ea742"
61
+ },
62
+ {
63
+ "name": "NVIDIA H200",
64
+ "memoryTotal": "150754820096",
65
+ "cudaCores": 16896,
66
+ "architecture": "Hopper",
67
+ "uuid": "GPU-08b37f98-4603-d483-2f2b-fe5311aa42f2"
68
+ },
69
+ {
70
+ "name": "NVIDIA H200",
71
+ "memoryTotal": "150754820096",
72
+ "cudaCores": 16896,
73
+ "architecture": "Hopper",
74
+ "uuid": "GPU-03273b5b-2fdd-a5fe-4460-c897334ae464"
75
+ },
76
+ {
77
+ "name": "NVIDIA H200",
78
+ "memoryTotal": "150754820096",
79
+ "cudaCores": 16896,
80
+ "architecture": "Hopper",
81
+ "uuid": "GPU-292d466c-d00d-25a4-28b6-e6c978d3e70c"
82
+ },
83
+ {
84
+ "name": "NVIDIA H200",
85
+ "memoryTotal": "150754820096",
86
+ "cudaCores": 16896,
87
+ "architecture": "Hopper",
88
+ "uuid": "GPU-46f38561-3148-e442-7f7f-bfe447bab7fe"
89
+ }
90
+ ],
91
+ "cudaVersion": "13.0",
92
+ "writerId": "ycd5umfcq5p6rih4lqfm265f2qblrtrn"
93
+ }
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"active_mask_imgs":{"filenames":["media/images/active_mask_imgs_183_27237392d1ba36fdb4c3.png"],"captions":["1072aae07584e091"],"_type":"images/separated","width":536,"height":800,"format":"png","count":1},"lr-AdamW/pg2":2e-05,"train/scene_scale":1.0012001991271973,"val/ssim":0.5464253425598145,"val/psnr":17.485671997070312,"loss/final_3dgs/mse":0.007215200457721949,"epoch":0,"val/lpips":0.3175295889377594,"loss/scene_scale_reg":3.102540722466074e-05,"trainer/global_step":3001,"loss/total":0.01728440262377262,"_step":186,"_timestamp":1.77255873057175e+09,"_runtime":10242,"train/psnr_probabilistic":21.999189376831055,"comparison":{"filenames":["media/images/comparison_182_5c2b90f763b7bff88204.png"],"captions":["1072aae07584e091"],"_type":"images/separated","width":1064,"height":1098,"format":"png","count":1},"loss/final_3dgs/lpips":0.0082782544195652,"train/comparison":{"filenames":["media/images/train/comparison_186_c80084b1f8dbfc2a13d7.png"],"captions":[["c270572a7f5ea828"]],"_type":"images/separated","width":536,"height":6378,"format":"png","count":1},"lr-AdamW/pg1-momentum":0.9,"_wandb":{"runtime":10242},"error_scores":{"height":536,"format":"png","count":1,"filenames":["media/images/error_scores_184_deb971ba58ebc4d5fe90.png"],"captions":["1072aae07584e091"],"_type":"images/separated","width":800},"loss/camera":0.00017599233251530677,"info/global_step":3000,"val/gaussian_num_ratio":0.40003204345703125,"lr-AdamW/pg1":2.003594834351718e-05,"lr-AdamW/pg2-momentum":0.9}
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug-core.log ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-03-03T14:34:53.358083416Z","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmppl6b06_b/port-854556.txt","pid":854556,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2026-03-03T14:34:53.358933202Z","level":"INFO","msg":"server: will exit if parent process dies","ppid":854556}
3
+ {"time":"2026-03-03T14:34:53.358910572Z","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-854556-856980-924415834/socket","Net":"unix"}}
4
+ {"time":"2026-03-03T14:34:53.535292678Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2026-03-03T14:34:53.543148589Z","level":"INFO","msg":"handleInformInit: received","streamId":"l8o7jl7r","id":"1(@)"}
6
+ {"time":"2026-03-03T14:34:53.95830905Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"l8o7jl7r","id":"1(@)"}
7
+ {"time":"2026-03-03T14:34:59.899286538Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"vehmru59g0oz"}
8
+ {"time":"2026-03-03T17:25:37.368559306Z","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
9
+ {"time":"2026-03-03T17:25:37.368609447Z","level":"INFO","msg":"connection: closing","id":"1(@)"}
10
+ {"time":"2026-03-03T17:25:37.368630207Z","level":"INFO","msg":"server is shutting down"}
11
+ {"time":"2026-03-03T17:25:37.368663987Z","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
12
+ {"time":"2026-03-03T17:25:37.368838899Z","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-854556-856980-924415834/socket","Net":"unix"}}
13
+ {"time":"2026-03-03T17:25:38.528751408Z","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
14
+ {"time":"2026-03-03T17:25:38.528775028Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
15
+ {"time":"2026-03-03T17:25:38.528786418Z","level":"INFO","msg":"server is closed"}
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug-internal.log ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-03-03T14:34:53.54334025Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
2
+ {"time":"2026-03-03T14:34:53.958050168Z","level":"INFO","msg":"stream: created new stream","id":"l8o7jl7r"}
3
+ {"time":"2026-03-03T14:34:53.958155019Z","level":"INFO","msg":"handler: started","stream_id":"l8o7jl7r"}
4
+ {"time":"2026-03-03T14:34:53.9583015Z","level":"INFO","msg":"stream: started","id":"l8o7jl7r"}
5
+ {"time":"2026-03-03T14:34:53.95834932Z","level":"INFO","msg":"sender: started","stream_id":"l8o7jl7r"}
6
+ {"time":"2026-03-03T14:34:53.95835085Z","level":"INFO","msg":"writer: started","stream_id":"l8o7jl7r"}
7
+ {"time":"2026-03-03T17:25:37.368615437Z","level":"INFO","msg":"stream: closing","id":"l8o7jl7r"}
8
+ {"time":"2026-03-03T17:25:38.278996749Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
9
+ {"time":"2026-03-03T17:25:38.528169562Z","level":"INFO","msg":"handler: closed","stream_id":"l8o7jl7r"}
10
+ {"time":"2026-03-03T17:25:38.528336264Z","level":"INFO","msg":"sender: closed","stream_id":"l8o7jl7r"}
11
+ {"time":"2026-03-03T17:25:38.528358534Z","level":"INFO","msg":"stream: closed","id":"l8o7jl7r"}
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug.log ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
2
+ 2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_setup.py:_flush():81] Configure stats pid to 854556
3
+ 2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
+ 2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug.log
5
+ 2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug-internal.log
6
+ 2026-03-03 14:34:53,294 INFO MainThread:854556 [wandb_init.py:init():844] calling init triggers
7
+ 2026-03-03 14:34:53,294 INFO MainThread:854556 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
+ config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'pcd', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'absgrad', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': False, 'voxelize_activate': False, 'use_depth': False}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'density_control_loss': {'error_score': {'weight': 0.0001, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0302_noAux', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': False, 'vggt_cam_loss': True, 'vggt_distil': False, 'context_view_train': False}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'save_gs': False, 'save_sample_wise_metrics': True, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3, 'target_align': True}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
9
+ 2026-03-03 14:34:53,294 INFO MainThread:854556 [wandb_init.py:init():892] starting backend
10
+ 2026-03-03 14:34:53,535 INFO MainThread:854556 [wandb_init.py:init():895] sending inform_init request
11
+ 2026-03-03 14:34:53,540 INFO MainThread:854556 [wandb_init.py:init():903] backend started and connected
12
+ 2026-03-03 14:34:53,542 INFO MainThread:854556 [wandb_init.py:init():973] updated telemetry
13
+ 2026-03-03 14:34:53,548 INFO MainThread:854556 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
+ 2026-03-03 14:34:54,767 INFO MainThread:854556 [wandb_init.py:init():1042] starting run threads in backend
15
+ 2026-03-03 14:34:54,895 INFO MainThread:854556 [wandb_run.py:_console_start():2524] atexit reg
16
+ 2026-03-03 14:34:54,895 INFO MainThread:854556 [wandb_run.py:_redirect():2373] redirect: wrap_raw
17
+ 2026-03-03 14:34:54,895 INFO MainThread:854556 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
+ 2026-03-03 14:34:54,896 INFO MainThread:854556 [wandb_run.py:_redirect():2465] Redirects installed.
19
+ 2026-03-03 14:34:54,899 INFO MainThread:854556 [wandb_init.py:init():1082] run started, returning control to user process
20
+ 2026-03-03 17:25:37,368 INFO wandb-AsyncioManager-main:854556 [service_client.py:_forward_responses():134] Reached EOF.
21
+ 2026-03-03 17:25:37,368 INFO wandb-AsyncioManager-main:854556 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.
ABLATION_0302_noTgtAlign/.hydra/config.yaml ADDED
@@ -0,0 +1,188 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ encoder:
3
+ name: dcsplat
4
+ input_image_shape:
5
+ - 518
6
+ - 518
7
+ head_mode: pcd
8
+ num_level: 3
9
+ gs_param_dim: 256
10
+ align_corners: false
11
+ use_voxelize: true
12
+ decoder:
13
+ name: splatting_cuda
14
+ background_color:
15
+ - 0.0
16
+ - 0.0
17
+ - 0.0
18
+ make_scale_invariant: false
19
+ density_control:
20
+ name: density_control_module
21
+ mean_dim: 32
22
+ gs_param_dim: 256
23
+ refinement_layer_num: 1
24
+ num_level: 3
25
+ grad_mode: absgrad
26
+ use_mean_features: true
27
+ refinement_type: voxelize
28
+ refinement_hidden_dim: 32
29
+ aggregation_mode: mean
30
+ num_heads: 1
31
+ score_mode: random
32
+ latent_dim: 128
33
+ num_latents: 64
34
+ num_self_attn_per_block: 2
35
+ voxel_size: 0.001
36
+ aux_refine: false
37
+ refine_error: false
38
+ use_refine_module: false
39
+ voxelize_activate: false
40
+ use_depth: false
41
+ render_loss:
42
+ mse:
43
+ weight: 1.0
44
+ lpips:
45
+ weight: 0.05
46
+ apply_after_step: 0
47
+ density_control_loss:
48
+ error_score:
49
+ weight: 0.0001
50
+ log_scale: false
51
+ grad_scale: 10000.0
52
+ mode: original
53
+ direct_loss:
54
+ l1:
55
+ weight: 0.8
56
+ ssim:
57
+ weight: 0.2
58
+ wandb:
59
+ project: DCSplat
60
+ entity: scene-representation-group
61
+ name: ABLATION_0302_noTgtAlign
62
+ mode: online
63
+ tags:
64
+ - re10k
65
+ - 256x256
66
+ mode: train
67
+ data_loader:
68
+ train:
69
+ num_workers: 16
70
+ persistent_workers: true
71
+ batch_size: 16
72
+ seed: 1234
73
+ test:
74
+ num_workers: 4
75
+ persistent_workers: false
76
+ batch_size: 1
77
+ seed: 2345
78
+ val:
79
+ num_workers: 1
80
+ persistent_workers: true
81
+ batch_size: 1
82
+ seed: 3456
83
+ optimizer:
84
+ lr: 0.0002
85
+ warm_up_steps: 25
86
+ backbone_lr_multiplier: 0.1
87
+ backbone_trainable: T+H
88
+ accumulate: 1
89
+ checkpointing:
90
+ load: null
91
+ every_n_train_steps: 1500
92
+ save_top_k: 2
93
+ save_weights_only: false
94
+ train:
95
+ extended_visualization: false
96
+ print_log_every_n_steps: 10
97
+ camera_loss: 10.0
98
+ one_sample_validation: null
99
+ align_corners: false
100
+ intrinsic_scaling: false
101
+ verbose: false
102
+ beta_dist_param:
103
+ - 0.5
104
+ - 4.0
105
+ use_refine_aux: false
106
+ train_target_set: true
107
+ train_gs_num: 1
108
+ ext_scale_detach: false
109
+ cam_scale_mode: sum
110
+ scene_scale_reg_loss: 0.01
111
+ train_aux: true
112
+ vggt_cam_loss: true
113
+ vggt_distil: false
114
+ context_view_train: false
115
+ test:
116
+ output_path: test/ablation/re10k
117
+ align_pose: false
118
+ pose_align_steps: 100
119
+ rot_opt_lr: 0.005
120
+ trans_opt_lr: 0.005
121
+ compute_scores: true
122
+ save_image: false
123
+ save_video: false
124
+ save_active_mask_image: false
125
+ save_error_score_image: false
126
+ save_compare: false
127
+ save_gs: false
128
+ save_sample_wise_metrics: true
129
+ pred_intrinsic: false
130
+ error_threshold: 0.4
131
+ error_threshold_list:
132
+ - 0.2
133
+ - 0.4
134
+ - 0.6
135
+ - 0.8
136
+ - 1.0
137
+ threshold_mode: ratio
138
+ nvs_view_N_list:
139
+ - 3
140
+ - 6
141
+ - 16
142
+ - 32
143
+ - 64
144
+ seed: 111123
145
+ trainer:
146
+ max_steps: 3001
147
+ val_check_interval: 250
148
+ gradient_clip_val: 0.5
149
+ num_nodes: 1
150
+ dataset:
151
+ re10k:
152
+ make_baseline_1: true
153
+ relative_pose: true
154
+ augment: true
155
+ background_color:
156
+ - 0.0
157
+ - 0.0
158
+ - 0.0
159
+ overfit_to_scene: null
160
+ skip_bad_shape: true
161
+ view_sampler:
162
+ name: bounded
163
+ num_target_views: 4
164
+ num_context_views: 2
165
+ min_distance_between_context_views: 45
166
+ max_distance_between_context_views: 90
167
+ min_distance_to_context_views: 0
168
+ warm_up_steps: 1000
169
+ initial_min_distance_between_context_views: 25
170
+ initial_max_distance_between_context_views: 25
171
+ same_target_gap: false
172
+ num_target_set: 3
173
+ target_align: true
174
+ name: re10k
175
+ roots:
176
+ - datasets/re10k
177
+ input_image_shape:
178
+ - 256
179
+ - 256
180
+ original_image_shape:
181
+ - 360
182
+ - 640
183
+ cameras_are_circular: false
184
+ baseline_min: 0.001
185
+ baseline_max: 10000000000.0
186
+ max_fov: 100.0
187
+ dynamic_context_views: true
188
+ max_context_views_per_gpu: 24
ABLATION_0302_noTgtAlign/.hydra/hydra.yaml ADDED
@@ -0,0 +1,165 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/ablation/re10k/${wandb.name}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - +experiment=re10k_ablation_24v
116
+ - wandb.mode=online
117
+ - wandb.name=ABLATION_0302_noTgtAlign
118
+ - model.density_control.score_mode=random
119
+ job:
120
+ name: main
121
+ chdir: null
122
+ override_dirname: +experiment=re10k_ablation_24v,model.density_control.score_mode=random,wandb.mode=online,wandb.name=ABLATION_0302_noTgtAlign
123
+ id: ???
124
+ num: ???
125
+ config_name: main
126
+ env_set: {}
127
+ env_copy: []
128
+ config:
129
+ override_dirname:
130
+ kv_sep: '='
131
+ item_sep: ','
132
+ exclude_keys: []
133
+ runtime:
134
+ version: 1.3.2
135
+ version_base: '1.3'
136
+ cwd: /workspace/code/CVPR2026
137
+ config_sources:
138
+ - path: hydra.conf
139
+ schema: pkg
140
+ provider: hydra
141
+ - path: /workspace/code/CVPR2026/config
142
+ schema: file
143
+ provider: main
144
+ - path: ''
145
+ schema: structured
146
+ provider: schema
147
+ output_dir: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign
148
+ choices:
149
+ experiment: re10k_ablation_24v
150
+ dataset@dataset.re10k: re10k
151
+ dataset/view_sampler_dataset_specific_config@dataset.re10k.view_sampler: bounded_re10k
152
+ dataset/view_sampler@dataset.re10k.view_sampler: bounded
153
+ model/density_control: density_control_module
154
+ model/decoder: splatting_cuda
155
+ model/encoder: dcsplat
156
+ hydra/env: default
157
+ hydra/callbacks: null
158
+ hydra/job_logging: default
159
+ hydra/hydra_logging: default
160
+ hydra/hydra_help: default
161
+ hydra/help: default
162
+ hydra/sweeper: basic
163
+ hydra/launcher: basic
164
+ hydra/output: default
165
+ verbose: false
ABLATION_0302_noTgtAlign/.hydra/overrides.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ - +experiment=re10k_ablation_24v
2
+ - wandb.mode=online
3
+ - wandb.name=ABLATION_0302_noTgtAlign
4
+ - model.density_control.score_mode=random
ABLATION_0302_noTgtAlign/main.log ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-03-03 17:25:57,101][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-03-03 17:26:03,305][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-03-03 17:26:03,305][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-03-03 17:26:08,283][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=255` in the `DataLoader` to improve performance.
9
+
10
+ [2026-03-03 17:29:31,696][dinov2][INFO] - using MLP layer as FFN
11
+ [2026-03-03 17:29:37,687][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
12
+ warnings.warn(
13
+
14
+ [2026-03-03 17:29:37,687][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
15
+ warnings.warn(msg)
16
+
17
+ [2026-03-03 17:30:28,360][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=31` in the `DataLoader` to improve performance.
18
+
19
+ [2026-03-03 17:30:28,362][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
20
+ warnings.warn( # warn only once
21
+
22
+ [2026-03-03 17:32:05,291][dinov2][INFO] - using MLP layer as FFN
23
+ [2026-03-03 17:32:11,332][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
24
+ warnings.warn(
25
+
26
+ [2026-03-03 17:32:11,332][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
27
+ warnings.warn(msg)
28
+
29
+ [2026-03-03 17:33:01,336][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=31` in the `DataLoader` to improve performance.
30
+
31
+ [2026-03-03 17:33:01,337][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
32
+ warnings.warn( # warn only once
33
+
34
+ [2026-03-03 17:34:59,889][dinov2][INFO] - using MLP layer as FFN
35
+ [2026-03-03 17:35:05,864][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
36
+ warnings.warn(
37
+
38
+ [2026-03-03 17:35:05,864][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
39
+ warnings.warn(msg)
40
+
41
+ [2026-03-03 17:35:58,952][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=31` in the `DataLoader` to improve performance.
42
+
43
+ [2026-03-03 17:35:58,954][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
44
+ warnings.warn( # warn only once
45
+
46
+ [2026-03-03 17:36:01,340][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
47
+ result[selector] = overlay
48
+
49
+ [2026-03-03 17:36:01,350][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
50
+
51
+ [2026-03-03 17:36:01,350][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
52
+ warnings.warn(
53
+
54
+ [2026-03-03 17:36:01,351][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
55
+ warnings.warn(msg)
56
+
57
+ [2026-03-03 17:36:03,034][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4322.)
58
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
59
+
60
+ [2026-03-03 17:36:03,326][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
61
+
62
+ [2026-03-03 17:36:03,327][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
63
+
64
+ [2026-03-03 17:36:03,328][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
65
+
66
+ [2026-03-03 17:36:03,328][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
67
+
68
+ [2026-03-03 17:36:03,328][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
69
+
70
+ [2026-03-03 17:36:13,075][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
71
+ grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
72
+ bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
73
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
74
+
75
+ [2026-03-03 17:36:13,149][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
76
+ result[selector] = overlay
77
+
78
+ [2026-03-03 17:37:46,051][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
79
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
80
+
81
+ [2026-03-03 17:48:34,774][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
82
+ result[selector] = overlay
83
+
ABLATION_0302_noTgtAlign/train_ddp_process_1.log ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-03-03 17:29:48,213][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-03-03 17:30:11,285][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-03-03 17:30:11,285][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-03-03 17:30:28,361][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
9
+ warnings.warn( # warn only once
10
+
11
+ [2026-03-03 17:32:21,762][dinov2][INFO] - using MLP layer as FFN
12
+ [2026-03-03 17:32:48,615][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
13
+ warnings.warn(
14
+
15
+ [2026-03-03 17:32:48,616][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
16
+ warnings.warn(msg)
17
+
18
+ [2026-03-03 17:33:01,337][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
19
+ warnings.warn( # warn only once
20
+
21
+ [2026-03-03 17:35:16,082][dinov2][INFO] - using MLP layer as FFN
22
+ [2026-03-03 17:35:42,645][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
23
+ warnings.warn(
24
+
25
+ [2026-03-03 17:35:42,646][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
26
+ warnings.warn(msg)
27
+
28
+ [2026-03-03 17:35:58,953][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
29
+ warnings.warn( # warn only once
30
+
31
+ [2026-03-03 17:36:13,070][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
32
+ grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
33
+ bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
34
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
35
+
36
+ [2026-03-03 17:36:13,172][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
37
+ result[selector] = overlay
38
+
39
+ [2026-03-03 17:37:46,051][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
40
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
41
+
42
+ [2026-03-03 17:48:34,774][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
43
+ result[selector] = overlay
44
+
ABLATION_0302_noTgtAlign/train_ddp_process_2.log ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-03-03 17:29:48,269][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-03-03 17:30:16,899][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-03-03 17:30:16,900][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-03-03 17:30:28,361][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
9
+ warnings.warn( # warn only once
10
+
11
+ [2026-03-03 17:32:21,758][dinov2][INFO] - using MLP layer as FFN
12
+ [2026-03-03 17:32:48,877][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
13
+ warnings.warn(
14
+
15
+ [2026-03-03 17:32:48,877][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
16
+ warnings.warn(msg)
17
+
18
+ [2026-03-03 17:33:01,337][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
19
+ warnings.warn( # warn only once
20
+
21
+ [2026-03-03 17:35:16,091][dinov2][INFO] - using MLP layer as FFN
22
+ [2026-03-03 17:35:47,851][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
23
+ warnings.warn(
24
+
25
+ [2026-03-03 17:35:47,851][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
26
+ warnings.warn(msg)
27
+
28
+ [2026-03-03 17:35:58,953][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
29
+ warnings.warn( # warn only once
30
+
31
+ [2026-03-03 17:36:13,078][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
32
+ grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
33
+ bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
34
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
35
+
36
+ [2026-03-03 17:36:13,182][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
37
+ result[selector] = overlay
38
+
39
+ [2026-03-03 17:37:46,051][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
40
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
41
+
42
+ [2026-03-03 17:48:34,774][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
43
+ result[selector] = overlay
44
+
ABLATION_0302_noTgtAlign/train_ddp_process_3.log ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-03-03 17:29:48,206][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-03-03 17:30:16,857][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-03-03 17:30:16,857][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-03-03 17:30:28,361][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
9
+ warnings.warn( # warn only once
10
+
11
+ [2026-03-03 17:32:21,762][dinov2][INFO] - using MLP layer as FFN
12
+ [2026-03-03 17:32:48,656][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
13
+ warnings.warn(
14
+
15
+ [2026-03-03 17:32:48,657][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
16
+ warnings.warn(msg)
17
+
18
+ [2026-03-03 17:33:01,336][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
19
+ warnings.warn( # warn only once
20
+
21
+ [2026-03-03 17:35:16,096][dinov2][INFO] - using MLP layer as FFN
22
+ [2026-03-03 17:35:47,666][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
23
+ warnings.warn(
24
+
25
+ [2026-03-03 17:35:47,666][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
26
+ warnings.warn(msg)
27
+
28
+ [2026-03-03 17:35:58,953][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
29
+ warnings.warn( # warn only once
30
+
31
+ [2026-03-03 17:36:13,076][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
32
+ grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
33
+ bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
34
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
35
+
36
+ [2026-03-03 17:36:13,191][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
37
+ result[selector] = overlay
38
+
39
+ [2026-03-03 17:37:46,079][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
40
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
41
+
42
+ [2026-03-03 17:48:34,775][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
43
+ result[selector] = overlay
44
+
ABLATION_0302_noTgtAlign/train_ddp_process_4.log ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-03-03 17:29:48,312][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-03-03 17:30:03,981][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-03-03 17:30:03,981][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-03-03 17:30:28,361][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
9
+ warnings.warn( # warn only once
10
+
11
+ [2026-03-03 17:32:21,721][dinov2][INFO] - using MLP layer as FFN
12
+ [2026-03-03 17:32:48,708][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
13
+ warnings.warn(
14
+
15
+ [2026-03-03 17:32:48,708][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
16
+ warnings.warn(msg)
17
+
18
+ [2026-03-03 17:33:01,337][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
19
+ warnings.warn( # warn only once
20
+
21
+ [2026-03-03 17:35:16,084][dinov2][INFO] - using MLP layer as FFN
22
+ [2026-03-03 17:35:47,371][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
23
+ warnings.warn(
24
+
25
+ [2026-03-03 17:35:47,371][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
26
+ warnings.warn(msg)
27
+
28
+ [2026-03-03 17:35:58,953][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
29
+ warnings.warn( # warn only once
30
+
31
+ [2026-03-03 17:36:13,073][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
32
+ grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
33
+ bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
34
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
35
+
36
+ [2026-03-03 17:36:13,182][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
37
+ result[selector] = overlay
38
+
39
+ [2026-03-03 17:37:46,076][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
40
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
41
+
42
+ [2026-03-03 17:48:34,775][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
43
+ result[selector] = overlay
44
+
ABLATION_0302_noTgtAlign/train_ddp_process_5.log ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-03-03 17:29:48,314][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-03-03 17:30:17,098][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-03-03 17:30:17,098][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-03-03 17:30:28,361][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
9
+ warnings.warn( # warn only once
10
+
11
+ [2026-03-03 17:32:21,789][dinov2][INFO] - using MLP layer as FFN
12
+ [2026-03-03 17:32:37,905][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
13
+ warnings.warn(
14
+
15
+ [2026-03-03 17:32:37,906][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
16
+ warnings.warn(msg)
17
+
18
+ [2026-03-03 17:33:01,336][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
19
+ warnings.warn( # warn only once
20
+
21
+ [2026-03-03 17:35:16,102][dinov2][INFO] - using MLP layer as FFN
22
+ [2026-03-03 17:35:45,673][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
23
+ warnings.warn(
24
+
25
+ [2026-03-03 17:35:45,676][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
26
+ warnings.warn(msg)
27
+
28
+ [2026-03-03 17:35:58,953][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
29
+ warnings.warn( # warn only once
30
+
31
+ [2026-03-03 17:36:12,327][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
32
+ grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
33
+ bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
34
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
35
+
36
+ [2026-03-03 17:36:13,191][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
37
+ result[selector] = overlay
38
+
39
+ [2026-03-03 17:37:46,079][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
40
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
41
+
42
+ [2026-03-03 17:48:34,774][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
43
+ result[selector] = overlay
44
+
ABLATION_0302_noTgtAlign/train_ddp_process_6.log ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-03-03 17:29:48,277][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-03-03 17:30:16,775][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-03-03 17:30:16,775][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-03-03 17:30:28,361][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
9
+ warnings.warn( # warn only once
10
+
11
+ [2026-03-03 17:32:21,773][dinov2][INFO] - using MLP layer as FFN
12
+ [2026-03-03 17:32:38,406][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
13
+ warnings.warn(
14
+
15
+ [2026-03-03 17:32:38,406][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
16
+ warnings.warn(msg)
17
+
18
+ [2026-03-03 17:33:01,337][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
19
+ warnings.warn( # warn only once
20
+
21
+ [2026-03-03 17:35:16,116][dinov2][INFO] - using MLP layer as FFN
22
+ [2026-03-03 17:35:47,749][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
23
+ warnings.warn(
24
+
25
+ [2026-03-03 17:35:47,749][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
26
+ warnings.warn(msg)
27
+
28
+ [2026-03-03 17:35:58,953][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
29
+ warnings.warn( # warn only once
30
+
31
+ [2026-03-03 17:36:12,560][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
32
+ grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
33
+ bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
34
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
35
+
36
+ [2026-03-03 17:36:13,180][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
37
+ result[selector] = overlay
38
+
39
+ [2026-03-03 17:37:46,058][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
40
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
41
+
42
+ [2026-03-03 17:48:34,774][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
43
+ result[selector] = overlay
44
+
ABLATION_0302_noTgtAlign/train_ddp_process_7.log ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-03-03 17:29:48,209][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-03-03 17:30:15,376][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-03-03 17:30:15,376][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-03-03 17:30:28,361][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
9
+ warnings.warn( # warn only once
10
+
11
+ [2026-03-03 17:32:21,793][dinov2][INFO] - using MLP layer as FFN
12
+ [2026-03-03 17:32:48,771][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
13
+ warnings.warn(
14
+
15
+ [2026-03-03 17:32:48,771][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
16
+ warnings.warn(msg)
17
+
18
+ [2026-03-03 17:33:01,336][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
19
+ warnings.warn( # warn only once
20
+
21
+ [2026-03-03 17:35:16,110][dinov2][INFO] - using MLP layer as FFN
22
+ [2026-03-03 17:35:47,614][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
23
+ warnings.warn(
24
+
25
+ [2026-03-03 17:35:47,614][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
26
+ warnings.warn(msg)
27
+
28
+ [2026-03-03 17:35:58,954][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
29
+ warnings.warn( # warn only once
30
+
31
+ [2026-03-03 17:36:12,568][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
32
+ grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
33
+ bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
34
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
35
+
36
+ [2026-03-03 17:36:13,185][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
37
+ result[selector] = overlay
38
+
39
+ [2026-03-03 17:37:46,051][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
40
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
41
+
42
+ [2026-03-03 17:48:34,774][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
43
+ result[selector] = overlay
44
+
ABLATION_0302_noTgtAlign/wandb/debug-internal.log ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {"time":"2026-03-03T17:35:55.708502915Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
2
+ {"time":"2026-03-03T17:35:56.156384015Z","level":"INFO","msg":"stream: created new stream","id":"et94mfhp"}
3
+ {"time":"2026-03-03T17:35:56.156522256Z","level":"INFO","msg":"handler: started","stream_id":"et94mfhp"}
4
+ {"time":"2026-03-03T17:35:56.156785449Z","level":"INFO","msg":"stream: started","id":"et94mfhp"}
5
+ {"time":"2026-03-03T17:35:56.156813659Z","level":"INFO","msg":"sender: started","stream_id":"et94mfhp"}
6
+ {"time":"2026-03-03T17:35:56.156835559Z","level":"INFO","msg":"writer: started","stream_id":"et94mfhp"}
ABLATION_0302_noTgtAlign/wandb/debug.log ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-03-03 17:35:55,409 INFO MainThread:870906 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
2
+ 2026-03-03 17:35:55,409 INFO MainThread:870906 [wandb_setup.py:_flush():81] Configure stats pid to 870906
3
+ 2026-03-03 17:35:55,409 INFO MainThread:870906 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
+ 2026-03-03 17:35:55,409 INFO MainThread:870906 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign/wandb/run-20260303_173555-et94mfhp/logs/debug.log
5
+ 2026-03-03 17:35:55,409 INFO MainThread:870906 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign/wandb/run-20260303_173555-et94mfhp/logs/debug-internal.log
6
+ 2026-03-03 17:35:55,410 INFO MainThread:870906 [wandb_init.py:init():844] calling init triggers
7
+ 2026-03-03 17:35:55,410 INFO MainThread:870906 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
+ config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'pcd', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'random', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': False, 'voxelize_activate': False, 'use_depth': False}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'density_control_loss': {'error_score': {'weight': 0.0001, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0302_noTgtAlign', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': True, 'vggt_cam_loss': True, 'vggt_distil': False, 'context_view_train': False}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'save_gs': False, 'save_sample_wise_metrics': True, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3, 'target_align': True}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
9
+ 2026-03-03 17:35:55,410 INFO MainThread:870906 [wandb_init.py:init():892] starting backend
10
+ 2026-03-03 17:35:55,698 INFO MainThread:870906 [wandb_init.py:init():895] sending inform_init request
11
+ 2026-03-03 17:35:55,705 INFO MainThread:870906 [wandb_init.py:init():903] backend started and connected
12
+ 2026-03-03 17:35:55,709 INFO MainThread:870906 [wandb_init.py:init():973] updated telemetry
13
+ 2026-03-03 17:35:55,716 INFO MainThread:870906 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
+ 2026-03-03 17:35:56,664 INFO MainThread:870906 [wandb_init.py:init():1042] starting run threads in backend
15
+ 2026-03-03 17:35:56,791 INFO MainThread:870906 [wandb_run.py:_console_start():2524] atexit reg
16
+ 2026-03-03 17:35:56,791 INFO MainThread:870906 [wandb_run.py:_redirect():2373] redirect: wrap_raw
17
+ 2026-03-03 17:35:56,791 INFO MainThread:870906 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
+ 2026-03-03 17:35:56,791 INFO MainThread:870906 [wandb_run.py:_redirect():2465] Redirects installed.
19
+ 2026-03-03 17:35:56,794 INFO MainThread:870906 [wandb_init.py:init():1082] run started, returning control to user process
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/config.yaml ADDED
@@ -0,0 +1,311 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.25.0
4
+ e:
5
+ vpcdp2qfp7cgu3zf4ilibu30q5on1nko:
6
+ args:
7
+ - +experiment=re10k_ablation_24v
8
+ - wandb.mode=online
9
+ - wandb.name=ABLATION_0302_noTgtAlign
10
+ - model.density_control.score_mode=random
11
+ cpu_count: 128
12
+ cpu_count_logical: 256
13
+ cudaVersion: "13.0"
14
+ disk:
15
+ /:
16
+ total: "735513149440"
17
+ used: "700800790528"
18
+ email: dna9041@korea.ac.kr
19
+ executable: /venv/main/bin/python
20
+ git:
21
+ commit: 9dfce172a0f8c7ce85e763899f7ef741ecffc454
22
+ remote: git@github.com:K-nowing/CVPR2026.git
23
+ gpu: NVIDIA H200
24
+ gpu_count: 8
25
+ gpu_nvidia:
26
+ - architecture: Hopper
27
+ cudaCores: 16896
28
+ memoryTotal: "150754820096"
29
+ name: NVIDIA H200
30
+ uuid: GPU-9a20101e-d876-facd-5f05-805081aede41
31
+ - architecture: Hopper
32
+ cudaCores: 16896
33
+ memoryTotal: "150754820096"
34
+ name: NVIDIA H200
35
+ uuid: GPU-84736a77-ee75-3324-e4e1-99cc15bfb5e9
36
+ - architecture: Hopper
37
+ cudaCores: 16896
38
+ memoryTotal: "150754820096"
39
+ name: NVIDIA H200
40
+ uuid: GPU-423d3161-cdc4-3fc0-caee-d15cfaa83ca6
41
+ - architecture: Hopper
42
+ cudaCores: 16896
43
+ memoryTotal: "150754820096"
44
+ name: NVIDIA H200
45
+ uuid: GPU-5b0058b2-cdb9-c952-04f9-87dcaa7ea742
46
+ - architecture: Hopper
47
+ cudaCores: 16896
48
+ memoryTotal: "150754820096"
49
+ name: NVIDIA H200
50
+ uuid: GPU-08b37f98-4603-d483-2f2b-fe5311aa42f2
51
+ - architecture: Hopper
52
+ cudaCores: 16896
53
+ memoryTotal: "150754820096"
54
+ name: NVIDIA H200
55
+ uuid: GPU-03273b5b-2fdd-a5fe-4460-c897334ae464
56
+ - architecture: Hopper
57
+ cudaCores: 16896
58
+ memoryTotal: "150754820096"
59
+ name: NVIDIA H200
60
+ uuid: GPU-292d466c-d00d-25a4-28b6-e6c978d3e70c
61
+ - architecture: Hopper
62
+ cudaCores: 16896
63
+ memoryTotal: "150754820096"
64
+ name: NVIDIA H200
65
+ uuid: GPU-46f38561-3148-e442-7f7f-bfe447bab7fe
66
+ host: e9d3310a05da
67
+ memory:
68
+ total: "1622950240256"
69
+ os: Linux-6.8.0-94-generic-x86_64-with-glibc2.39
70
+ program: -m src.main
71
+ python: CPython 3.12.12
72
+ root: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign
73
+ startedAt: "2026-03-03T17:26:05.471917Z"
74
+ writerId: vpcdp2qfp7cgu3zf4ilibu30q5on1nko
75
+ m:
76
+ - "1": trainer/global_step
77
+ "6":
78
+ - 3
79
+ "7": []
80
+ - "2": '*'
81
+ "5": 1
82
+ "6":
83
+ - 1
84
+ "7": []
85
+ python_version: 3.12.12
86
+ t:
87
+ "1":
88
+ - 1
89
+ - 41
90
+ - 49
91
+ - 50
92
+ - 106
93
+ "2":
94
+ - 1
95
+ - 41
96
+ - 49
97
+ - 50
98
+ - 106
99
+ "3":
100
+ - 2
101
+ - 7
102
+ - 13
103
+ - 15
104
+ - 16
105
+ - 66
106
+ "4": 3.12.12
107
+ "5": 0.25.0
108
+ "12": 0.25.0
109
+ "13": linux-x86_64
110
+ checkpointing:
111
+ value:
112
+ every_n_train_steps: 1500
113
+ load: null
114
+ save_top_k: 2
115
+ save_weights_only: false
116
+ data_loader:
117
+ value:
118
+ test:
119
+ batch_size: 1
120
+ num_workers: 4
121
+ persistent_workers: false
122
+ seed: 2345
123
+ train:
124
+ batch_size: 16
125
+ num_workers: 16
126
+ persistent_workers: true
127
+ seed: 1234
128
+ val:
129
+ batch_size: 1
130
+ num_workers: 1
131
+ persistent_workers: true
132
+ seed: 3456
133
+ dataset:
134
+ value:
135
+ re10k:
136
+ augment: true
137
+ background_color:
138
+ - 0
139
+ - 0
140
+ - 0
141
+ baseline_max: 1e+10
142
+ baseline_min: 0.001
143
+ cameras_are_circular: false
144
+ dynamic_context_views: true
145
+ input_image_shape:
146
+ - 256
147
+ - 256
148
+ make_baseline_1: true
149
+ max_context_views_per_gpu: 24
150
+ max_fov: 100
151
+ name: re10k
152
+ original_image_shape:
153
+ - 360
154
+ - 640
155
+ overfit_to_scene: null
156
+ relative_pose: true
157
+ roots:
158
+ - datasets/re10k
159
+ skip_bad_shape: true
160
+ view_sampler:
161
+ initial_max_distance_between_context_views: 25
162
+ initial_min_distance_between_context_views: 25
163
+ max_distance_between_context_views: 90
164
+ min_distance_between_context_views: 45
165
+ min_distance_to_context_views: 0
166
+ name: bounded
167
+ num_context_views: 2
168
+ num_target_set: 3
169
+ num_target_views: 4
170
+ same_target_gap: false
171
+ target_align: true
172
+ warm_up_steps: 1000
173
+ density_control_loss:
174
+ value:
175
+ error_score:
176
+ grad_scale: 10000
177
+ log_scale: false
178
+ mode: original
179
+ weight: 0.0001
180
+ direct_loss:
181
+ value:
182
+ l1:
183
+ weight: 0.8
184
+ ssim:
185
+ weight: 0.2
186
+ mode:
187
+ value: train
188
+ model:
189
+ value:
190
+ decoder:
191
+ background_color:
192
+ - 0
193
+ - 0
194
+ - 0
195
+ make_scale_invariant: false
196
+ name: splatting_cuda
197
+ density_control:
198
+ aggregation_mode: mean
199
+ aux_refine: false
200
+ grad_mode: absgrad
201
+ gs_param_dim: 256
202
+ latent_dim: 128
203
+ mean_dim: 32
204
+ name: density_control_module
205
+ num_heads: 1
206
+ num_latents: 64
207
+ num_level: 3
208
+ num_self_attn_per_block: 2
209
+ refine_error: false
210
+ refinement_hidden_dim: 32
211
+ refinement_layer_num: 1
212
+ refinement_type: voxelize
213
+ score_mode: random
214
+ use_depth: false
215
+ use_mean_features: true
216
+ use_refine_module: false
217
+ voxel_size: 0.001
218
+ voxelize_activate: false
219
+ encoder:
220
+ align_corners: false
221
+ gs_param_dim: 256
222
+ head_mode: pcd
223
+ input_image_shape:
224
+ - 518
225
+ - 518
226
+ name: dcsplat
227
+ num_level: 3
228
+ use_voxelize: true
229
+ optimizer:
230
+ value:
231
+ accumulate: 1
232
+ backbone_lr_multiplier: 0.1
233
+ backbone_trainable: T+H
234
+ lr: 0.0002
235
+ warm_up_steps: 25
236
+ render_loss:
237
+ value:
238
+ lpips:
239
+ apply_after_step: 0
240
+ weight: 0.05
241
+ mse:
242
+ weight: 1
243
+ seed:
244
+ value: 111123
245
+ test:
246
+ value:
247
+ align_pose: false
248
+ compute_scores: true
249
+ error_threshold: 0.4
250
+ error_threshold_list:
251
+ - 0.2
252
+ - 0.4
253
+ - 0.6
254
+ - 0.8
255
+ - 1
256
+ nvs_view_N_list:
257
+ - 3
258
+ - 6
259
+ - 16
260
+ - 32
261
+ - 64
262
+ output_path: test/ablation/re10k
263
+ pose_align_steps: 100
264
+ pred_intrinsic: false
265
+ rot_opt_lr: 0.005
266
+ save_active_mask_image: false
267
+ save_compare: false
268
+ save_error_score_image: false
269
+ save_gs: false
270
+ save_image: false
271
+ save_sample_wise_metrics: true
272
+ save_video: false
273
+ threshold_mode: ratio
274
+ trans_opt_lr: 0.005
275
+ train:
276
+ value:
277
+ align_corners: false
278
+ beta_dist_param:
279
+ - 0.5
280
+ - 4
281
+ cam_scale_mode: sum
282
+ camera_loss: 10
283
+ context_view_train: false
284
+ ext_scale_detach: false
285
+ extended_visualization: false
286
+ intrinsic_scaling: false
287
+ one_sample_validation: null
288
+ print_log_every_n_steps: 10
289
+ scene_scale_reg_loss: 0.01
290
+ train_aux: true
291
+ train_gs_num: 1
292
+ train_target_set: true
293
+ use_refine_aux: false
294
+ verbose: false
295
+ vggt_cam_loss: true
296
+ vggt_distil: false
297
+ trainer:
298
+ value:
299
+ gradient_clip_val: 0.5
300
+ max_steps: 3001
301
+ num_nodes: 1
302
+ val_check_interval: 250
303
+ wandb:
304
+ value:
305
+ entity: scene-representation-group
306
+ mode: online
307
+ name: ABLATION_0302_noTgtAlign
308
+ project: DCSplat
309
+ tags:
310
+ - re10k
311
+ - 256x256
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/output.log ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0]
2
+
3
+ | Name | Type | Params | Mode
4
+ ------------------------------------------------------------------------
5
+ 0 | encoder | OurSplat | 888 M | train
6
+ 1 | density_control_module | DensityControlModule | 0 | train
7
+ 2 | decoder | DecoderSplattingCUDA | 0 | train
8
+ 3 | render_losses | ModuleList | 0 | train
9
+ 4 | density_control_losses | ModuleList | 0 | train
10
+ 5 | direct_losses | ModuleList | 0 | train
11
+ ------------------------------------------------------------------------
12
+ 888 M Trainable params
13
+ 0 Non-trainable params
14
+ 888 M Total params
15
+ 3,553.933 Total estimated model params size (MB)
16
+ 1202 Modules in train mode
17
+ 522 Modules in eval mode
18
+ Sanity Checking: | | 0/? [00:00<?, ?it/s][2026-03-03 17:26:08,283][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=255` in the `DataLoader` to improve performance.
19
+
20
+ Validation epoch start on rank 0
21
+ Sanity Checking DataLoader 0: 0%| | 0/1 [00:00<?, ?it/s]validation step 0; scene = ['306e2b7785657539'];
22
+ > /workspace/code/CVPR2026/src/model/model_wrapper.py(1570)get_normalized_target_cam()
23
+ -> gt_left_c2w = gt_context_c2w[:, 0:1, :, :]
24
+ tensor([[0.0369]], device='cuda:0')
25
+ tensor([[0.0369, 0.0369, 0.0369, 0.0369]], device='cuda:0')
26
+ tensor([[[1, 0],
27
+ [0, 1],
28
+ [0, 1],
29
+ [0, 1]]], device='cuda:0')
30
+ tensor([[[[0.8595, 0.0000, 0.5000],
31
+ [0.0000, 0.8597, 0.5000],
32
+ [0.0000, 0.0000, 1.0000]],
33
+
34
+ [[0.8595, 0.0000, 0.5000],
35
+ [0.0000, 0.8597, 0.5000],
36
+ [0.0000, 0.0000, 1.0000]],
37
+
38
+ [[0.8595, 0.0000, 0.5000],
39
+ [0.0000, 0.8597, 0.5000],
40
+ [0.0000, 0.0000, 1.0000]],
41
+
42
+ [[0.8595, 0.0000, 0.5000],
43
+ [0.0000, 0.8597, 0.5000],
44
+ [0.0000, 0.0000, 1.0000]]]], device='cuda:0')
45
+ tensor([[[[0.8837, 0.0000, 0.5000],
46
+ [0.0000, 0.8834, 0.5000],
47
+ [0.0000, 0.0000, 1.0000]],
48
+
49
+ [[0.8721, 0.0000, 0.5000],
50
+ [0.0000, 0.8713, 0.5000],
51
+ [0.0000, 0.0000, 1.0000]]]], device='cuda:0')
52
+ torch.Size([1, 4, 3, 3])
53
+ torch.Size([1, 2, 3, 3])
54
+ tensor([[[[0.8779, 0.0000, 0.5000],
55
+ [0.0000, 0.8773, 0.5000],
56
+ [0.0000, 0.0000, 1.0000]],
57
+
58
+ [[0.8779, 0.0000, 0.5000],
59
+ [0.0000, 0.8773, 0.5000],
60
+ [0.0000, 0.0000, 1.0000]],
61
+
62
+ [[0.8779, 0.0000, 0.5000],
63
+ [0.0000, 0.8773, 0.5000],
64
+ [0.0000, 0.0000, 1.0000]],
65
+
66
+ [[0.8779, 0.0000, 0.5000],
67
+ [0.0000, 0.8773, 0.5000],
68
+ [0.0000, 0.0000, 1.0000]]]], device='cuda:0')
69
+ Error executing job with overrides: ['+experiment=re10k_ablation_24v', 'wandb.mode=online', 'wandb.name=ABLATION_0302_noTgtAlign', 'model.density_control.score_mode=random']
70
+
71
+
72
+ Received interrupt signal. Cleaning up...
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/requirements.txt ADDED
@@ -0,0 +1,173 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wheel==0.45.1
2
+ pytz==2025.2
3
+ easydict==1.13
4
+ antlr4-python3-runtime==4.9.3
5
+ wadler_lindig==0.1.7
6
+ networkx==3.4.2
7
+ urllib3==2.5.0
8
+ tzdata==2025.2
9
+ typing-inspection==0.4.1
10
+ tabulate==0.9.0
11
+ smmap==5.0.2
12
+ setuptools==78.1.1
13
+ safetensors==0.5.3
14
+ multidict==6.6.4
15
+ PyYAML==6.0.2
16
+ PySocks==1.7.1
17
+ pyparsing==3.2.5
18
+ pydantic_core==2.33.2
19
+ pycparser==2.23
20
+ protobuf==6.32.1
21
+ propcache==0.3.2
22
+ proglog==0.1.12
23
+ platformdirs==4.4.0
24
+ pip==25.2
25
+ mdurl==0.1.2
26
+ pillow==10.4.0
27
+ packaging==24.2
28
+ opt_einsum==3.4.0
29
+ frozenlist==1.7.0
30
+ numpy==1.26.4
31
+ ninja==1.13.0
32
+ MarkupSafe==3.0.2
33
+ kornia_rs==0.1.9
34
+ kiwisolver==1.4.9
35
+ imageio-ffmpeg==0.6.0
36
+ idna==3.7
37
+ fsspec==2024.6.1
38
+ hf-xet==1.1.10
39
+ gmpy2==2.2.1
40
+ fonttools==4.60.0
41
+ triton==3.4.0
42
+ filelock==3.17.0
43
+ einops==0.8.1
44
+ decorator==4.4.2
45
+ dacite==1.9.2
46
+ cycler==0.12.1
47
+ colorama==0.4.6
48
+ click==8.3.0
49
+ nvidia-nvtx-cu12==12.8.90
50
+ charset-normalizer==3.3.2
51
+ certifi==2025.8.3
52
+ beartype==0.19.0
53
+ attrs==25.3.0
54
+ async-timeout==5.0.1
55
+ annotated-types==0.7.0
56
+ aiohappyeyeballs==2.6.1
57
+ yarl==1.20.1
58
+ tifffile==2025.5.10
59
+ sentry-sdk==2.39.0
60
+ scipy==1.15.3
61
+ pydantic==2.11.9
62
+ pandas==2.3.2
63
+ opencv-python==4.11.0.86
64
+ omegaconf==2.3.0
65
+ markdown-it-py==4.0.0
66
+ lightning-utilities==0.14.3
67
+ lazy_loader==0.4
68
+ jaxtyping==0.2.37
69
+ imageio==2.37.0
70
+ gitdb==4.0.12
71
+ contourpy==1.3.2
72
+ colorspacious==1.1.2
73
+ cffi==1.17.1
74
+ aiosignal==1.4.0
75
+ scikit-video==1.1.11
76
+ scikit-image==0.25.2
77
+ rich==14.1.0
78
+ moviepy==1.0.3
79
+ matplotlib==3.10.6
80
+ hydra-core==1.3.2
81
+ huggingface-hub==0.35.1
82
+ GitPython==3.1.45
83
+ brotlicffi==1.0.9.2
84
+ aiohttp==3.12.15
85
+ torchmetrics==1.8.2
86
+ opt-einsum-fx==0.1.4
87
+ kornia==0.8.1
88
+ pytorch-lightning==2.5.1
89
+ lpips==0.1.4
90
+ e3nn==0.6.0
91
+ lightning==2.5.1
92
+ gsplat==1.5.3
93
+ nvidia-cusparselt-cu12==0.7.1
94
+ nvidia-nvjitlink-cu12==12.8.93
95
+ nvidia-nccl-cu12==2.27.3
96
+ nvidia-curand-cu12==10.3.9.90
97
+ nvidia-cufile-cu12==1.13.1.3
98
+ nvidia-cuda-runtime-cu12==12.8.90
99
+ nvidia-cuda-nvrtc-cu12==12.8.93
100
+ nvidia-cuda-cupti-cu12==12.8.90
101
+ nvidia-cublas-cu12==12.8.4.1
102
+ nvidia-cusparse-cu12==12.5.8.93
103
+ nvidia-cufft-cu12==11.3.3.83
104
+ nvidia-cudnn-cu12==9.10.2.21
105
+ nvidia-cusolver-cu12==11.7.3.90
106
+ torch==2.8.0+cu128
107
+ torchvision==0.23.0+cu128
108
+ torchaudio==2.8.0+cu128
109
+ torch_scatter==2.1.2+pt28cu128
110
+ plyfile==1.1.3
111
+ wandb==0.25.0
112
+ cuda-bindings==13.0.3
113
+ cuda-pathfinder==1.3.3
114
+ Jinja2==3.1.6
115
+ mpmath==1.3.0
116
+ nvidia-cublas==13.1.0.3
117
+ nvidia-cuda-cupti==13.0.85
118
+ nvidia-cuda-nvrtc==13.0.88
119
+ nvidia-cuda-runtime==13.0.96
120
+ nvidia-cudnn-cu13==9.15.1.9
121
+ nvidia-cufft==12.0.0.61
122
+ nvidia-cufile==1.15.1.6
123
+ nvidia-curand==10.4.0.35
124
+ nvidia-cusolver==12.0.4.66
125
+ nvidia-cusparse==12.6.3.3
126
+ nvidia-cusparselt-cu13==0.8.0
127
+ nvidia-nccl-cu13==2.28.9
128
+ nvidia-nvjitlink==13.0.88
129
+ nvidia-nvshmem-cu13==3.4.5
130
+ nvidia-nvtx==13.0.85
131
+ requests==2.32.5
132
+ sentencepiece==0.2.1
133
+ sympy==1.14.0
134
+ torchcodec==0.10.0
135
+ torchdata==0.10.0
136
+ torchtext==0.6.0
137
+ anyio==4.12.0
138
+ asttokens==3.0.1
139
+ comm==0.2.3
140
+ debugpy==1.8.19
141
+ executing==2.2.1
142
+ h11==0.16.0
143
+ httpcore==1.0.9
144
+ httpx==0.28.1
145
+ ipykernel==7.1.0
146
+ ipython==9.8.0
147
+ ipython_pygments_lexers==1.1.1
148
+ ipywidgets==8.1.8
149
+ jedi==0.19.2
150
+ jupyter_client==8.7.0
151
+ jupyter_core==5.9.1
152
+ jupyterlab_widgets==3.0.16
153
+ matplotlib-inline==0.2.1
154
+ nest-asyncio==1.6.0
155
+ parso==0.8.5
156
+ pexpect==4.9.0
157
+ prompt_toolkit==3.0.52
158
+ psutil==7.2.1
159
+ ptyprocess==0.7.0
160
+ pure_eval==0.2.3
161
+ Pygments==2.19.2
162
+ python-dateutil==2.9.0.post0
163
+ pyzmq==27.1.0
164
+ shellingham==1.5.4
165
+ six==1.17.0
166
+ stack-data==0.6.3
167
+ tornado==6.5.4
168
+ tqdm==4.67.1
169
+ traitlets==5.14.3
170
+ typer-slim==0.21.0
171
+ typing_extensions==4.15.0
172
+ wcwidth==0.2.14
173
+ widgetsnbextension==4.0.15
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/wandb-metadata.json ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.8.0-94-generic-x86_64-with-glibc2.39",
3
+ "python": "CPython 3.12.12",
4
+ "startedAt": "2026-03-03T17:26:05.471917Z",
5
+ "args": [
6
+ "+experiment=re10k_ablation_24v",
7
+ "wandb.mode=online",
8
+ "wandb.name=ABLATION_0302_noTgtAlign",
9
+ "model.density_control.score_mode=random"
10
+ ],
11
+ "program": "-m src.main",
12
+ "git": {
13
+ "remote": "git@github.com:K-nowing/CVPR2026.git",
14
+ "commit": "9dfce172a0f8c7ce85e763899f7ef741ecffc454"
15
+ },
16
+ "email": "dna9041@korea.ac.kr",
17
+ "root": "/workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign",
18
+ "host": "e9d3310a05da",
19
+ "executable": "/venv/main/bin/python",
20
+ "cpu_count": 128,
21
+ "cpu_count_logical": 256,
22
+ "gpu": "NVIDIA H200",
23
+ "gpu_count": 8,
24
+ "disk": {
25
+ "/": {
26
+ "total": "735513149440",
27
+ "used": "700800790528"
28
+ }
29
+ },
30
+ "memory": {
31
+ "total": "1622950240256"
32
+ },
33
+ "gpu_nvidia": [
34
+ {
35
+ "name": "NVIDIA H200",
36
+ "memoryTotal": "150754820096",
37
+ "cudaCores": 16896,
38
+ "architecture": "Hopper",
39
+ "uuid": "GPU-9a20101e-d876-facd-5f05-805081aede41"
40
+ },
41
+ {
42
+ "name": "NVIDIA H200",
43
+ "memoryTotal": "150754820096",
44
+ "cudaCores": 16896,
45
+ "architecture": "Hopper",
46
+ "uuid": "GPU-84736a77-ee75-3324-e4e1-99cc15bfb5e9"
47
+ },
48
+ {
49
+ "name": "NVIDIA H200",
50
+ "memoryTotal": "150754820096",
51
+ "cudaCores": 16896,
52
+ "architecture": "Hopper",
53
+ "uuid": "GPU-423d3161-cdc4-3fc0-caee-d15cfaa83ca6"
54
+ },
55
+ {
56
+ "name": "NVIDIA H200",
57
+ "memoryTotal": "150754820096",
58
+ "cudaCores": 16896,
59
+ "architecture": "Hopper",
60
+ "uuid": "GPU-5b0058b2-cdb9-c952-04f9-87dcaa7ea742"
61
+ },
62
+ {
63
+ "name": "NVIDIA H200",
64
+ "memoryTotal": "150754820096",
65
+ "cudaCores": 16896,
66
+ "architecture": "Hopper",
67
+ "uuid": "GPU-08b37f98-4603-d483-2f2b-fe5311aa42f2"
68
+ },
69
+ {
70
+ "name": "NVIDIA H200",
71
+ "memoryTotal": "150754820096",
72
+ "cudaCores": 16896,
73
+ "architecture": "Hopper",
74
+ "uuid": "GPU-03273b5b-2fdd-a5fe-4460-c897334ae464"
75
+ },
76
+ {
77
+ "name": "NVIDIA H200",
78
+ "memoryTotal": "150754820096",
79
+ "cudaCores": 16896,
80
+ "architecture": "Hopper",
81
+ "uuid": "GPU-292d466c-d00d-25a4-28b6-e6c978d3e70c"
82
+ },
83
+ {
84
+ "name": "NVIDIA H200",
85
+ "memoryTotal": "150754820096",
86
+ "cudaCores": 16896,
87
+ "architecture": "Hopper",
88
+ "uuid": "GPU-46f38561-3148-e442-7f7f-bfe447bab7fe"
89
+ }
90
+ ],
91
+ "cudaVersion": "13.0",
92
+ "writerId": "vpcdp2qfp7cgu3zf4ilibu30q5on1nko"
93
+ }
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb":{"runtime":190},"_runtime":190}
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/logs/debug-core.log ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-03-03T17:26:05.544279566Z","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpnepv657f/port-862779.txt","pid":862779,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2026-03-03T17:26:05.545034072Z","level":"INFO","msg":"server: will exit if parent process dies","ppid":862779}
3
+ {"time":"2026-03-03T17:26:05.545012722Z","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-862779-863051-2889925930/socket","Net":"unix"}}
4
+ {"time":"2026-03-03T17:26:05.717282063Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2026-03-03T17:26:05.726827389Z","level":"INFO","msg":"handleInformInit: received","streamId":"beokhgg2","id":"1(@)"}
6
+ {"time":"2026-03-03T17:26:06.186040503Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"beokhgg2","id":"1(@)"}
7
+ {"time":"2026-03-03T17:26:12.470381356Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"ahylq6e9uxwi"}
8
+ {"time":"2026-03-03T17:29:17.914358058Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"ahylq6e9uxwi"}
9
+ {"time":"2026-03-03T17:29:19.452510575Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"ahylq6e9uxwi"}
10
+ {"time":"2026-03-03T17:29:19.453368693Z","level":"INFO","msg":"handleInformFinish: finish message received","streamId":"beokhgg2","id":"1(@)"}
11
+ {"time":"2026-03-03T17:29:19.453861608Z","level":"INFO","msg":"handleInformFinish: stream closed","streamId":"beokhgg2","id":"1(@)"}
12
+ {"time":"2026-03-03T17:29:19.454484434Z","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
13
+ {"time":"2026-03-03T17:29:19.454514804Z","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
14
+ {"time":"2026-03-03T17:29:19.454527464Z","level":"INFO","msg":"server is shutting down"}
15
+ {"time":"2026-03-03T17:29:19.454569715Z","level":"INFO","msg":"connection: closing","id":"1(@)"}
16
+ {"time":"2026-03-03T17:29:19.454680076Z","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
17
+ {"time":"2026-03-03T17:29:19.454709016Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
18
+ {"time":"2026-03-03T17:29:19.454665865Z","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-862779-863051-2889925930/socket","Net":"unix"}}
19
+ {"time":"2026-03-03T17:29:19.454730766Z","level":"INFO","msg":"server is closed"}
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/logs/debug-internal.log ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-03-03T17:26:05.727056141Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
2
+ {"time":"2026-03-03T17:26:06.185659999Z","level":"INFO","msg":"stream: created new stream","id":"beokhgg2"}
3
+ {"time":"2026-03-03T17:26:06.185780721Z","level":"INFO","msg":"handler: started","stream_id":"beokhgg2"}
4
+ {"time":"2026-03-03T17:26:06.186025753Z","level":"INFO","msg":"stream: started","id":"beokhgg2"}
5
+ {"time":"2026-03-03T17:26:06.186049093Z","level":"INFO","msg":"writer: started","stream_id":"beokhgg2"}
6
+ {"time":"2026-03-03T17:26:06.186076323Z","level":"INFO","msg":"sender: started","stream_id":"beokhgg2"}
7
+ {"time":"2026-03-03T17:29:19.172521278Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
8
+ {"time":"2026-03-03T17:29:19.449459346Z","level":"INFO","msg":"handler: operation stats","stats":{}}
9
+ {"time":"2026-03-03T17:29:19.453393683Z","level":"INFO","msg":"stream: closing","id":"beokhgg2"}
10
+ {"time":"2026-03-03T17:29:19.453413883Z","level":"INFO","msg":"handler: closed","stream_id":"beokhgg2"}
11
+ {"time":"2026-03-03T17:29:19.453531935Z","level":"INFO","msg":"sender: closed","stream_id":"beokhgg2"}
12
+ {"time":"2026-03-03T17:29:19.453555075Z","level":"INFO","msg":"stream: closed","id":"beokhgg2"}
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/logs/debug.log ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-03-03 17:26:05,473 INFO MainThread:862779 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
2
+ 2026-03-03 17:26:05,473 INFO MainThread:862779 [wandb_setup.py:_flush():81] Configure stats pid to 862779
3
+ 2026-03-03 17:26:05,473 INFO MainThread:862779 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
+ 2026-03-03 17:26:05,473 INFO MainThread:862779 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/logs/debug.log
5
+ 2026-03-03 17:26:05,473 INFO MainThread:862779 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/logs/debug-internal.log
6
+ 2026-03-03 17:26:05,473 INFO MainThread:862779 [wandb_init.py:init():844] calling init triggers
7
+ 2026-03-03 17:26:05,473 INFO MainThread:862779 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
+ config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'pcd', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'random', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': False, 'voxelize_activate': False, 'use_depth': False}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'density_control_loss': {'error_score': {'weight': 0.0001, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0302_noTgtAlign', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': True, 'vggt_cam_loss': True, 'vggt_distil': False, 'context_view_train': False}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'save_gs': False, 'save_sample_wise_metrics': True, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3, 'target_align': True}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
9
+ 2026-03-03 17:26:05,473 INFO MainThread:862779 [wandb_init.py:init():892] starting backend
10
+ 2026-03-03 17:26:05,717 INFO MainThread:862779 [wandb_init.py:init():895] sending inform_init request
11
+ 2026-03-03 17:26:05,723 INFO MainThread:862779 [wandb_init.py:init():903] backend started and connected
12
+ 2026-03-03 17:26:05,730 INFO MainThread:862779 [wandb_init.py:init():973] updated telemetry
13
+ 2026-03-03 17:26:05,737 INFO MainThread:862779 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
+ 2026-03-03 17:26:07,330 INFO MainThread:862779 [wandb_init.py:init():1042] starting run threads in backend
15
+ 2026-03-03 17:26:07,463 INFO MainThread:862779 [wandb_run.py:_console_start():2524] atexit reg
16
+ 2026-03-03 17:26:07,463 INFO MainThread:862779 [wandb_run.py:_redirect():2373] redirect: wrap_raw
17
+ 2026-03-03 17:26:07,464 INFO MainThread:862779 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
+ 2026-03-03 17:26:07,464 INFO MainThread:862779 [wandb_run.py:_redirect():2465] Redirects installed.
19
+ 2026-03-03 17:26:07,468 INFO MainThread:862779 [wandb_init.py:init():1082] run started, returning control to user process
20
+ 2026-03-03 17:29:17,912 INFO MainThread:862779 [wandb_run.py:_finish():2291] finishing run know/DCSplat/beokhgg2
21
+ 2026-03-03 17:29:17,913 INFO MainThread:862779 [wandb_run.py:_atexit_cleanup():2490] got exitcode: 0
22
+ 2026-03-03 17:29:17,913 INFO MainThread:862779 [wandb_run.py:_restore():2472] restore
23
+ 2026-03-03 17:29:17,913 INFO MainThread:862779 [wandb_run.py:_restore():2478] restore done
24
+ 2026-03-03 17:29:19,452 INFO MainThread:862779 [wandb_run.py:_footer_sync_info():3868] logging synced files
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/run-beokhgg2.wandb ADDED
Binary file (51.8 kB). View file
 
ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/config.yaml ADDED
@@ -0,0 +1,311 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.25.0
4
+ e:
5
+ mytbwtaba7k7jw5rce0hn1lj8k7t5dm8:
6
+ args:
7
+ - +experiment=re10k_ablation_24v
8
+ - wandb.mode=online
9
+ - wandb.name=ABLATION_0302_noTgtAlign
10
+ - model.density_control.score_mode=random
11
+ cpu_count: 128
12
+ cpu_count_logical: 256
13
+ cudaVersion: "13.0"
14
+ disk:
15
+ /:
16
+ total: "735513149440"
17
+ used: "700800118784"
18
+ email: dna9041@korea.ac.kr
19
+ executable: /venv/main/bin/python
20
+ git:
21
+ commit: 9dfce172a0f8c7ce85e763899f7ef741ecffc454
22
+ remote: git@github.com:K-nowing/CVPR2026.git
23
+ gpu: NVIDIA H200
24
+ gpu_count: 8
25
+ gpu_nvidia:
26
+ - architecture: Hopper
27
+ cudaCores: 16896
28
+ memoryTotal: "150754820096"
29
+ name: NVIDIA H200
30
+ uuid: GPU-9a20101e-d876-facd-5f05-805081aede41
31
+ - architecture: Hopper
32
+ cudaCores: 16896
33
+ memoryTotal: "150754820096"
34
+ name: NVIDIA H200
35
+ uuid: GPU-84736a77-ee75-3324-e4e1-99cc15bfb5e9
36
+ - architecture: Hopper
37
+ cudaCores: 16896
38
+ memoryTotal: "150754820096"
39
+ name: NVIDIA H200
40
+ uuid: GPU-423d3161-cdc4-3fc0-caee-d15cfaa83ca6
41
+ - architecture: Hopper
42
+ cudaCores: 16896
43
+ memoryTotal: "150754820096"
44
+ name: NVIDIA H200
45
+ uuid: GPU-5b0058b2-cdb9-c952-04f9-87dcaa7ea742
46
+ - architecture: Hopper
47
+ cudaCores: 16896
48
+ memoryTotal: "150754820096"
49
+ name: NVIDIA H200
50
+ uuid: GPU-08b37f98-4603-d483-2f2b-fe5311aa42f2
51
+ - architecture: Hopper
52
+ cudaCores: 16896
53
+ memoryTotal: "150754820096"
54
+ name: NVIDIA H200
55
+ uuid: GPU-03273b5b-2fdd-a5fe-4460-c897334ae464
56
+ - architecture: Hopper
57
+ cudaCores: 16896
58
+ memoryTotal: "150754820096"
59
+ name: NVIDIA H200
60
+ uuid: GPU-292d466c-d00d-25a4-28b6-e6c978d3e70c
61
+ - architecture: Hopper
62
+ cudaCores: 16896
63
+ memoryTotal: "150754820096"
64
+ name: NVIDIA H200
65
+ uuid: GPU-46f38561-3148-e442-7f7f-bfe447bab7fe
66
+ host: e9d3310a05da
67
+ memory:
68
+ total: "1622950240256"
69
+ os: Linux-6.8.0-94-generic-x86_64-with-glibc2.39
70
+ program: -m src.main
71
+ python: CPython 3.12.12
72
+ root: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign
73
+ startedAt: "2026-03-03T17:30:24.387912Z"
74
+ writerId: mytbwtaba7k7jw5rce0hn1lj8k7t5dm8
75
+ m:
76
+ - "1": trainer/global_step
77
+ "6":
78
+ - 3
79
+ "7": []
80
+ - "2": '*'
81
+ "5": 1
82
+ "6":
83
+ - 1
84
+ "7": []
85
+ python_version: 3.12.12
86
+ t:
87
+ "1":
88
+ - 1
89
+ - 41
90
+ - 49
91
+ - 50
92
+ - 106
93
+ "2":
94
+ - 1
95
+ - 41
96
+ - 49
97
+ - 50
98
+ - 106
99
+ "3":
100
+ - 7
101
+ - 13
102
+ - 15
103
+ - 16
104
+ - 41
105
+ - 66
106
+ "4": 3.12.12
107
+ "5": 0.25.0
108
+ "12": 0.25.0
109
+ "13": linux-x86_64
110
+ checkpointing:
111
+ value:
112
+ every_n_train_steps: 1500
113
+ load: null
114
+ save_top_k: 2
115
+ save_weights_only: false
116
+ data_loader:
117
+ value:
118
+ test:
119
+ batch_size: 1
120
+ num_workers: 4
121
+ persistent_workers: false
122
+ seed: 2345
123
+ train:
124
+ batch_size: 16
125
+ num_workers: 16
126
+ persistent_workers: true
127
+ seed: 1234
128
+ val:
129
+ batch_size: 1
130
+ num_workers: 1
131
+ persistent_workers: true
132
+ seed: 3456
133
+ dataset:
134
+ value:
135
+ re10k:
136
+ augment: true
137
+ background_color:
138
+ - 0
139
+ - 0
140
+ - 0
141
+ baseline_max: 1e+10
142
+ baseline_min: 0.001
143
+ cameras_are_circular: false
144
+ dynamic_context_views: true
145
+ input_image_shape:
146
+ - 256
147
+ - 256
148
+ make_baseline_1: true
149
+ max_context_views_per_gpu: 24
150
+ max_fov: 100
151
+ name: re10k
152
+ original_image_shape:
153
+ - 360
154
+ - 640
155
+ overfit_to_scene: null
156
+ relative_pose: true
157
+ roots:
158
+ - datasets/re10k
159
+ skip_bad_shape: true
160
+ view_sampler:
161
+ initial_max_distance_between_context_views: 25
162
+ initial_min_distance_between_context_views: 25
163
+ max_distance_between_context_views: 90
164
+ min_distance_between_context_views: 45
165
+ min_distance_to_context_views: 0
166
+ name: bounded
167
+ num_context_views: 2
168
+ num_target_set: 3
169
+ num_target_views: 4
170
+ same_target_gap: false
171
+ target_align: true
172
+ warm_up_steps: 1000
173
+ density_control_loss:
174
+ value:
175
+ error_score:
176
+ grad_scale: 10000
177
+ log_scale: false
178
+ mode: original
179
+ weight: 0.0001
180
+ direct_loss:
181
+ value:
182
+ l1:
183
+ weight: 0.8
184
+ ssim:
185
+ weight: 0.2
186
+ mode:
187
+ value: train
188
+ model:
189
+ value:
190
+ decoder:
191
+ background_color:
192
+ - 0
193
+ - 0
194
+ - 0
195
+ make_scale_invariant: false
196
+ name: splatting_cuda
197
+ density_control:
198
+ aggregation_mode: mean
199
+ aux_refine: false
200
+ grad_mode: absgrad
201
+ gs_param_dim: 256
202
+ latent_dim: 128
203
+ mean_dim: 32
204
+ name: density_control_module
205
+ num_heads: 1
206
+ num_latents: 64
207
+ num_level: 3
208
+ num_self_attn_per_block: 2
209
+ refine_error: false
210
+ refinement_hidden_dim: 32
211
+ refinement_layer_num: 1
212
+ refinement_type: voxelize
213
+ score_mode: random
214
+ use_depth: false
215
+ use_mean_features: true
216
+ use_refine_module: false
217
+ voxel_size: 0.001
218
+ voxelize_activate: false
219
+ encoder:
220
+ align_corners: false
221
+ gs_param_dim: 256
222
+ head_mode: pcd
223
+ input_image_shape:
224
+ - 518
225
+ - 518
226
+ name: dcsplat
227
+ num_level: 3
228
+ use_voxelize: true
229
+ optimizer:
230
+ value:
231
+ accumulate: 1
232
+ backbone_lr_multiplier: 0.1
233
+ backbone_trainable: T+H
234
+ lr: 0.0002
235
+ warm_up_steps: 25
236
+ render_loss:
237
+ value:
238
+ lpips:
239
+ apply_after_step: 0
240
+ weight: 0.05
241
+ mse:
242
+ weight: 1
243
+ seed:
244
+ value: 111123
245
+ test:
246
+ value:
247
+ align_pose: false
248
+ compute_scores: true
249
+ error_threshold: 0.4
250
+ error_threshold_list:
251
+ - 0.2
252
+ - 0.4
253
+ - 0.6
254
+ - 0.8
255
+ - 1
256
+ nvs_view_N_list:
257
+ - 3
258
+ - 6
259
+ - 16
260
+ - 32
261
+ - 64
262
+ output_path: test/ablation/re10k
263
+ pose_align_steps: 100
264
+ pred_intrinsic: false
265
+ rot_opt_lr: 0.005
266
+ save_active_mask_image: false
267
+ save_compare: false
268
+ save_error_score_image: false
269
+ save_gs: false
270
+ save_image: false
271
+ save_sample_wise_metrics: true
272
+ save_video: false
273
+ threshold_mode: ratio
274
+ trans_opt_lr: 0.005
275
+ train:
276
+ value:
277
+ align_corners: false
278
+ beta_dist_param:
279
+ - 0.5
280
+ - 4
281
+ cam_scale_mode: sum
282
+ camera_loss: 10
283
+ context_view_train: false
284
+ ext_scale_detach: false
285
+ extended_visualization: false
286
+ intrinsic_scaling: false
287
+ one_sample_validation: null
288
+ print_log_every_n_steps: 10
289
+ scene_scale_reg_loss: 0.01
290
+ train_aux: true
291
+ train_gs_num: 1
292
+ train_target_set: true
293
+ use_refine_aux: false
294
+ verbose: false
295
+ vggt_cam_loss: true
296
+ vggt_distil: false
297
+ trainer:
298
+ value:
299
+ gradient_clip_val: 0.5
300
+ max_steps: 3001
301
+ num_nodes: 1
302
+ val_check_interval: 250
303
+ wandb:
304
+ value:
305
+ entity: scene-representation-group
306
+ mode: online
307
+ name: ABLATION_0302_noTgtAlign
308
+ project: DCSplat
309
+ tags:
310
+ - re10k
311
+ - 256x256
ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/output.log ADDED
@@ -0,0 +1,117 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7]
2
+
3
+ | Name | Type | Params | Mode
4
+ ------------------------------------------------------------------------
5
+ 0 | encoder | OurSplat | 888 M | train
6
+ 1 | density_control_module | DensityControlModule | 0 | train
7
+ 2 | decoder | DecoderSplattingCUDA | 0 | train
8
+ 3 | render_losses | ModuleList | 0 | train
9
+ 4 | density_control_losses | ModuleList | 0 | train
10
+ 5 | direct_losses | ModuleList | 0 | train
11
+ ------------------------------------------------------------------------
12
+ 888 M Trainable params
13
+ 0 Non-trainable params
14
+ 888 M Total params
15
+ 3,553.933 Total estimated model params size (MB)
16
+ 1202 Modules in train mode
17
+ 522 Modules in eval mode
18
+ Sanity Checking: | | 0/? [00:00<?, ?it/s][2026-03-03 17:30:28,360][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=31` in the `DataLoader` to improve performance.
19
+
20
+ [2026-03-03 17:30:28,362][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
21
+ warnings.warn( # warn only once
22
+
23
+ Validation epoch start on rank 0
24
+ Sanity Checking DataLoader 0: 0%| | 0/1 [00:00<?, ?it/s]validation step 0; scene = ['306e2b7785657539'];
25
+ Error executing job with overrides: ['+experiment=re10k_ablation_24v', 'wandb.mode=online', 'wandb.name=ABLATION_0302_noTgtAlign', 'model.density_control.score_mode=random']
26
+ Traceback (most recent call last):
27
+ File "/workspace/code/CVPR2026/src/main.py", line 226, in train
28
+ trainer.fit(model_wrapper, datamodule=data_module)#, ckpt_path=checkpoint_path)
29
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
30
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 561, in fit
31
+ call._call_and_handle_interrupt(
32
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/call.py", line 47, in _call_and_handle_interrupt
33
+ return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs)
34
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
35
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch
36
+ return function(*args, **kwargs)
37
+ ^^^^^^^^^^^^^^^^^^^^^^^^^
38
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 599, in _fit_impl
39
+ self._run(model, ckpt_path=ckpt_path)
40
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 1012, in _run
41
+ results = self._run_stage()
42
+ ^^^^^^^^^^^^^^^^^
43
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 1054, in _run_stage
44
+ self._run_sanity_check()
45
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 1083, in _run_sanity_check
46
+ val_loop.run()
47
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/utilities.py", line 179, in _decorator
48
+ return loop_run(self, *args, **kwargs)
49
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
50
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/evaluation_loop.py", line 145, in run
51
+ self._evaluation_step(batch, batch_idx, dataloader_idx, dataloader_iter)
52
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/evaluation_loop.py", line 437, in _evaluation_step
53
+ output = call._call_strategy_hook(trainer, hook_name, *step_args)
54
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
55
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/call.py", line 328, in _call_strategy_hook
56
+ output = fn(*args, **kwargs)
57
+ ^^^^^^^^^^^^^^^^^^^
58
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/strategies/strategy.py", line 411, in validation_step
59
+ return self._forward_redirection(self.model, self.lightning_module, "validation_step", *args, **kwargs)
60
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
61
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__
62
+ wrapper_output = wrapper_module(*args, **kwargs)
63
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
64
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl
65
+ return self._call_impl(*args, **kwargs)
66
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
67
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl
68
+ return forward_call(*args, **kwargs)
69
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
70
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/parallel/distributed.py", line 1648, in forward
71
+ else self._run_ddp_forward(*inputs, **kwargs)
72
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
73
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/parallel/distributed.py", line 1474, in _run_ddp_forward
74
+ return self.module(*inputs, **kwargs) # type: ignore[index]
75
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
76
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl
77
+ return self._call_impl(*args, **kwargs)
78
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
79
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl
80
+ return forward_call(*args, **kwargs)
81
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
82
+ File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward
83
+ out = method(*_args, **_kwargs)
84
+ ^^^^^^^^^^^^^^^^^^^^^^^^^
85
+ File "/venv/main/lib/python3.12/site-packages/lightning_utilities/core/rank_zero.py", line 41, in wrapped_fn
86
+ return fn(*args, **kwargs)
87
+ ^^^^^^^^^^^^^^^^^^^
88
+ File "/venv/main/lib/python3.12/site-packages/jaxtyping/_decorator.py", line 562, in wrapped_fn
89
+ return wrapped_fn_impl(args, kwargs, bound, memos)
90
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
91
+ File "/venv/main/lib/python3.12/site-packages/jaxtyping/_decorator.py", line 486, in wrapped_fn_impl
92
+ out = fn(*args, **kwargs)
93
+ ^^^^^^^^^^^^^^^^^^^
94
+ File "/workspace/code/CVPR2026/src/model/model_wrapper.py", line 1093, in validation_step
95
+ target_extrinsic_c2w, target_intrinsic, near_plane, far_plane = self.get_target_projections(batch, backbone_outputs, target_key=target_key_list[target_key_idx])
96
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
97
+ File "/venv/main/lib/python3.12/site-packages/jaxtyping/_decorator.py", line 562, in wrapped_fn
98
+ return wrapped_fn_impl(args, kwargs, bound, memos)
99
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
100
+ File "/venv/main/lib/python3.12/site-packages/jaxtyping/_decorator.py", line 486, in wrapped_fn_impl
101
+ out = fn(*args, **kwargs)
102
+ ^^^^^^^^^^^^^^^^^^^
103
+ File "/workspace/code/CVPR2026/src/model/model_wrapper.py", line 1718, in get_target_projections
104
+ target_extrinsic_c2w, target_intrinsic, target_nearest_context_idx, near_plane, far_plane = self.get_normalized_target_cam(batch, gs_outputs, target_key, intr_scale)
105
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
106
+ File "/venv/main/lib/python3.12/site-packages/jaxtyping/_decorator.py", line 562, in wrapped_fn
107
+ return wrapped_fn_impl(args, kwargs, bound, memos)
108
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
109
+ File "/venv/main/lib/python3.12/site-packages/jaxtyping/_decorator.py", line 486, in wrapped_fn_impl
110
+ out = fn(*args, **kwargs)
111
+ ^^^^^^^^^^^^^^^^^^^
112
+ File "/workspace/code/CVPR2026/src/model/model_wrapper.py", line 1595, in get_normalized_target_cam
113
+ target_intrinsic = pred_intrinsics.mean(dim=1, keepdim=True).repeat(1, TVN, 1, 1)
114
+ ^^^^^^^^^^^^^^^
115
+ NameError: name 'pred_intrinsics' is not defined
116
+
117
+ Set the environment variable HYDRA_FULL_ERROR=1 for a complete stack trace.
ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/requirements.txt ADDED
@@ -0,0 +1,173 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wheel==0.45.1
2
+ pytz==2025.2
3
+ easydict==1.13
4
+ antlr4-python3-runtime==4.9.3
5
+ wadler_lindig==0.1.7
6
+ networkx==3.4.2
7
+ urllib3==2.5.0
8
+ tzdata==2025.2
9
+ typing-inspection==0.4.1
10
+ tabulate==0.9.0
11
+ smmap==5.0.2
12
+ setuptools==78.1.1
13
+ safetensors==0.5.3
14
+ multidict==6.6.4
15
+ PyYAML==6.0.2
16
+ PySocks==1.7.1
17
+ pyparsing==3.2.5
18
+ pydantic_core==2.33.2
19
+ pycparser==2.23
20
+ protobuf==6.32.1
21
+ propcache==0.3.2
22
+ proglog==0.1.12
23
+ platformdirs==4.4.0
24
+ pip==25.2
25
+ mdurl==0.1.2
26
+ pillow==10.4.0
27
+ packaging==24.2
28
+ opt_einsum==3.4.0
29
+ frozenlist==1.7.0
30
+ numpy==1.26.4
31
+ ninja==1.13.0
32
+ MarkupSafe==3.0.2
33
+ kornia_rs==0.1.9
34
+ kiwisolver==1.4.9
35
+ imageio-ffmpeg==0.6.0
36
+ idna==3.7
37
+ fsspec==2024.6.1
38
+ hf-xet==1.1.10
39
+ gmpy2==2.2.1
40
+ fonttools==4.60.0
41
+ triton==3.4.0
42
+ filelock==3.17.0
43
+ einops==0.8.1
44
+ decorator==4.4.2
45
+ dacite==1.9.2
46
+ cycler==0.12.1
47
+ colorama==0.4.6
48
+ click==8.3.0
49
+ nvidia-nvtx-cu12==12.8.90
50
+ charset-normalizer==3.3.2
51
+ certifi==2025.8.3
52
+ beartype==0.19.0
53
+ attrs==25.3.0
54
+ async-timeout==5.0.1
55
+ annotated-types==0.7.0
56
+ aiohappyeyeballs==2.6.1
57
+ yarl==1.20.1
58
+ tifffile==2025.5.10
59
+ sentry-sdk==2.39.0
60
+ scipy==1.15.3
61
+ pydantic==2.11.9
62
+ pandas==2.3.2
63
+ opencv-python==4.11.0.86
64
+ omegaconf==2.3.0
65
+ markdown-it-py==4.0.0
66
+ lightning-utilities==0.14.3
67
+ lazy_loader==0.4
68
+ jaxtyping==0.2.37
69
+ imageio==2.37.0
70
+ gitdb==4.0.12
71
+ contourpy==1.3.2
72
+ colorspacious==1.1.2
73
+ cffi==1.17.1
74
+ aiosignal==1.4.0
75
+ scikit-video==1.1.11
76
+ scikit-image==0.25.2
77
+ rich==14.1.0
78
+ moviepy==1.0.3
79
+ matplotlib==3.10.6
80
+ hydra-core==1.3.2
81
+ huggingface-hub==0.35.1
82
+ GitPython==3.1.45
83
+ brotlicffi==1.0.9.2
84
+ aiohttp==3.12.15
85
+ torchmetrics==1.8.2
86
+ opt-einsum-fx==0.1.4
87
+ kornia==0.8.1
88
+ pytorch-lightning==2.5.1
89
+ lpips==0.1.4
90
+ e3nn==0.6.0
91
+ lightning==2.5.1
92
+ gsplat==1.5.3
93
+ nvidia-cusparselt-cu12==0.7.1
94
+ nvidia-nvjitlink-cu12==12.8.93
95
+ nvidia-nccl-cu12==2.27.3
96
+ nvidia-curand-cu12==10.3.9.90
97
+ nvidia-cufile-cu12==1.13.1.3
98
+ nvidia-cuda-runtime-cu12==12.8.90
99
+ nvidia-cuda-nvrtc-cu12==12.8.93
100
+ nvidia-cuda-cupti-cu12==12.8.90
101
+ nvidia-cublas-cu12==12.8.4.1
102
+ nvidia-cusparse-cu12==12.5.8.93
103
+ nvidia-cufft-cu12==11.3.3.83
104
+ nvidia-cudnn-cu12==9.10.2.21
105
+ nvidia-cusolver-cu12==11.7.3.90
106
+ torch==2.8.0+cu128
107
+ torchvision==0.23.0+cu128
108
+ torchaudio==2.8.0+cu128
109
+ torch_scatter==2.1.2+pt28cu128
110
+ plyfile==1.1.3
111
+ wandb==0.25.0
112
+ cuda-bindings==13.0.3
113
+ cuda-pathfinder==1.3.3
114
+ Jinja2==3.1.6
115
+ mpmath==1.3.0
116
+ nvidia-cublas==13.1.0.3
117
+ nvidia-cuda-cupti==13.0.85
118
+ nvidia-cuda-nvrtc==13.0.88
119
+ nvidia-cuda-runtime==13.0.96
120
+ nvidia-cudnn-cu13==9.15.1.9
121
+ nvidia-cufft==12.0.0.61
122
+ nvidia-cufile==1.15.1.6
123
+ nvidia-curand==10.4.0.35
124
+ nvidia-cusolver==12.0.4.66
125
+ nvidia-cusparse==12.6.3.3
126
+ nvidia-cusparselt-cu13==0.8.0
127
+ nvidia-nccl-cu13==2.28.9
128
+ nvidia-nvjitlink==13.0.88
129
+ nvidia-nvshmem-cu13==3.4.5
130
+ nvidia-nvtx==13.0.85
131
+ requests==2.32.5
132
+ sentencepiece==0.2.1
133
+ sympy==1.14.0
134
+ torchcodec==0.10.0
135
+ torchdata==0.10.0
136
+ torchtext==0.6.0
137
+ anyio==4.12.0
138
+ asttokens==3.0.1
139
+ comm==0.2.3
140
+ debugpy==1.8.19
141
+ executing==2.2.1
142
+ h11==0.16.0
143
+ httpcore==1.0.9
144
+ httpx==0.28.1
145
+ ipykernel==7.1.0
146
+ ipython==9.8.0
147
+ ipython_pygments_lexers==1.1.1
148
+ ipywidgets==8.1.8
149
+ jedi==0.19.2
150
+ jupyter_client==8.7.0
151
+ jupyter_core==5.9.1
152
+ jupyterlab_widgets==3.0.16
153
+ matplotlib-inline==0.2.1
154
+ nest-asyncio==1.6.0
155
+ parso==0.8.5
156
+ pexpect==4.9.0
157
+ prompt_toolkit==3.0.52
158
+ psutil==7.2.1
159
+ ptyprocess==0.7.0
160
+ pure_eval==0.2.3
161
+ Pygments==2.19.2
162
+ python-dateutil==2.9.0.post0
163
+ pyzmq==27.1.0
164
+ shellingham==1.5.4
165
+ six==1.17.0
166
+ stack-data==0.6.3
167
+ tornado==6.5.4
168
+ tqdm==4.67.1
169
+ traitlets==5.14.3
170
+ typer-slim==0.21.0
171
+ typing_extensions==4.15.0
172
+ wcwidth==0.2.14
173
+ widgetsnbextension==4.0.15
ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/wandb-metadata.json ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.8.0-94-generic-x86_64-with-glibc2.39",
3
+ "python": "CPython 3.12.12",
4
+ "startedAt": "2026-03-03T17:30:24.387912Z",
5
+ "args": [
6
+ "+experiment=re10k_ablation_24v",
7
+ "wandb.mode=online",
8
+ "wandb.name=ABLATION_0302_noTgtAlign",
9
+ "model.density_control.score_mode=random"
10
+ ],
11
+ "program": "-m src.main",
12
+ "git": {
13
+ "remote": "git@github.com:K-nowing/CVPR2026.git",
14
+ "commit": "9dfce172a0f8c7ce85e763899f7ef741ecffc454"
15
+ },
16
+ "email": "dna9041@korea.ac.kr",
17
+ "root": "/workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign",
18
+ "host": "e9d3310a05da",
19
+ "executable": "/venv/main/bin/python",
20
+ "cpu_count": 128,
21
+ "cpu_count_logical": 256,
22
+ "gpu": "NVIDIA H200",
23
+ "gpu_count": 8,
24
+ "disk": {
25
+ "/": {
26
+ "total": "735513149440",
27
+ "used": "700800118784"
28
+ }
29
+ },
30
+ "memory": {
31
+ "total": "1622950240256"
32
+ },
33
+ "gpu_nvidia": [
34
+ {
35
+ "name": "NVIDIA H200",
36
+ "memoryTotal": "150754820096",
37
+ "cudaCores": 16896,
38
+ "architecture": "Hopper",
39
+ "uuid": "GPU-9a20101e-d876-facd-5f05-805081aede41"
40
+ },
41
+ {
42
+ "name": "NVIDIA H200",
43
+ "memoryTotal": "150754820096",
44
+ "cudaCores": 16896,
45
+ "architecture": "Hopper",
46
+ "uuid": "GPU-84736a77-ee75-3324-e4e1-99cc15bfb5e9"
47
+ },
48
+ {
49
+ "name": "NVIDIA H200",
50
+ "memoryTotal": "150754820096",
51
+ "cudaCores": 16896,
52
+ "architecture": "Hopper",
53
+ "uuid": "GPU-423d3161-cdc4-3fc0-caee-d15cfaa83ca6"
54
+ },
55
+ {
56
+ "name": "NVIDIA H200",
57
+ "memoryTotal": "150754820096",
58
+ "cudaCores": 16896,
59
+ "architecture": "Hopper",
60
+ "uuid": "GPU-5b0058b2-cdb9-c952-04f9-87dcaa7ea742"
61
+ },
62
+ {
63
+ "name": "NVIDIA H200",
64
+ "memoryTotal": "150754820096",
65
+ "cudaCores": 16896,
66
+ "architecture": "Hopper",
67
+ "uuid": "GPU-08b37f98-4603-d483-2f2b-fe5311aa42f2"
68
+ },
69
+ {
70
+ "name": "NVIDIA H200",
71
+ "memoryTotal": "150754820096",
72
+ "cudaCores": 16896,
73
+ "architecture": "Hopper",
74
+ "uuid": "GPU-03273b5b-2fdd-a5fe-4460-c897334ae464"
75
+ },
76
+ {
77
+ "name": "NVIDIA H200",
78
+ "memoryTotal": "150754820096",
79
+ "cudaCores": 16896,
80
+ "architecture": "Hopper",
81
+ "uuid": "GPU-292d466c-d00d-25a4-28b6-e6c978d3e70c"
82
+ },
83
+ {
84
+ "name": "NVIDIA H200",
85
+ "memoryTotal": "150754820096",
86
+ "cudaCores": 16896,
87
+ "architecture": "Hopper",
88
+ "uuid": "GPU-46f38561-3148-e442-7f7f-bfe447bab7fe"
89
+ }
90
+ ],
91
+ "cudaVersion": "13.0",
92
+ "writerId": "mytbwtaba7k7jw5rce0hn1lj8k7t5dm8"
93
+ }
ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_runtime":7,"_wandb":{"runtime":7}}
ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/logs/debug-core.log ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-03-03T17:30:24.492490052Z","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpztx1j514/port-863948.txt","pid":863948,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2026-03-03T17:30:24.493158838Z","level":"INFO","msg":"server: will exit if parent process dies","ppid":863948}
3
+ {"time":"2026-03-03T17:30:24.493143718Z","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-863948-866375-1399352181/socket","Net":"unix"}}
4
+ {"time":"2026-03-03T17:30:24.673880578Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2026-03-03T17:30:24.684765982Z","level":"INFO","msg":"handleInformInit: received","streamId":"phck7alc","id":"1(@)"}
6
+ {"time":"2026-03-03T17:30:25.148320888Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"phck7alc","id":"1(@)"}
7
+ {"time":"2026-03-03T17:30:31.082435508Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"i6tdoiz1oxwi"}
8
+ {"time":"2026-03-03T17:30:33.040285798Z","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
9
+ {"time":"2026-03-03T17:30:33.040358339Z","level":"INFO","msg":"connection: closing","id":"1(@)"}
10
+ {"time":"2026-03-03T17:30:33.040416149Z","level":"INFO","msg":"server is shutting down"}
11
+ {"time":"2026-03-03T17:30:33.040443929Z","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
12
+ {"time":"2026-03-03T17:30:33.040579961Z","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-863948-866375-1399352181/socket","Net":"unix"}}
13
+ {"time":"2026-03-03T17:30:34.141387811Z","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
14
+ {"time":"2026-03-03T17:30:34.141419972Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
15
+ {"time":"2026-03-03T17:30:34.141433602Z","level":"INFO","msg":"server is closed"}