Knowing commited on
Commit
c1e9c80
·
verified ·
1 Parent(s): 24f3723

Add files using upload-large-folder tool

Browse files
Files changed (32) hide show
  1. .gitattributes +1 -0
  2. ABLATION_0225_ctxTrain_depth_vggtDistl/main.log +46 -0
  3. ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_1.log +17 -0
  4. ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_2.log +17 -0
  5. ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_3.log +17 -0
  6. ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_4.log +17 -0
  7. ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_5.log +17 -0
  8. ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_6.log +17 -0
  9. ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_7.log +17 -0
  10. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/debug-internal.log +6 -11
  11. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/debug.log +19 -21
  12. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/active_mask_imgs_18_e74266ca2c24ca638ee2.png +3 -0
  13. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/active_mask_imgs_1_690a5f29fe9230f3ead3.png +3 -0
  14. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/comparison_0_e330894a93360b5335a1.png +3 -0
  15. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/comparison_17_6e317bfdf78b6c03c329.png +3 -0
  16. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/error_scores_19_cebe8e2cd2233435a7a5.png +3 -0
  17. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/error_scores_2_bf819ed247f7546e0637.png +3 -0
  18. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/train/comparison_14_2e2099ac37d79d14f71c.png +3 -0
  19. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/train/comparison_4_69221fdfe7cd493b747b.png +3 -0
  20. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/train/error_scores_13_a60c0930fe11800b83fd.png +3 -0
  21. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/train/error_scores_3_0a2cb05f0b9c7f260e6d.png +3 -0
  22. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/output.log +132 -0
  23. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/requirements.txt +172 -0
  24. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/wandb-metadata.json +96 -0
  25. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/logs/debug-core.log +7 -0
  26. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/logs/debug-internal.log +6 -0
  27. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/logs/debug.log +19 -0
  28. ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/run-pa77a1wq.wandb +3 -0
  29. DEBUG/.hydra/config.yaml +185 -0
  30. DEBUG/.hydra/hydra.yaml +170 -0
  31. DEBUG/.hydra/overrides.yaml +9 -0
  32. DEBUG/main.log +24 -0
.gitattributes CHANGED
@@ -64,3 +64,4 @@ ABLATION_0225_OURS/wandb/run-20260224_191508-0b125b6z/run-0b125b6z.wandb filter=
64
  ABLATION_0225_FreqSelect/wandb/run-20260224_222739-y7wvpmyy/run-y7wvpmyy.wandb filter=lfs diff=lfs merge=lfs -text
65
  ABLATION_0225_targetTrain/wandb/run-20260225_014059-qetzseh9/run-qetzseh9.wandb filter=lfs diff=lfs merge=lfs -text
66
  ABLATION_0225_targetTrain_SSR/wandb/run-20260225_043637-hluaxp6d/run-hluaxp6d.wandb filter=lfs diff=lfs merge=lfs -text
 
 
64
  ABLATION_0225_FreqSelect/wandb/run-20260224_222739-y7wvpmyy/run-y7wvpmyy.wandb filter=lfs diff=lfs merge=lfs -text
65
  ABLATION_0225_targetTrain/wandb/run-20260225_014059-qetzseh9/run-qetzseh9.wandb filter=lfs diff=lfs merge=lfs -text
66
  ABLATION_0225_targetTrain_SSR/wandb/run-20260225_043637-hluaxp6d/run-hluaxp6d.wandb filter=lfs diff=lfs merge=lfs -text
67
+ ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/run-pa77a1wq.wandb filter=lfs diff=lfs merge=lfs -text
ABLATION_0225_ctxTrain_depth_vggtDistl/main.log CHANGED
@@ -43,3 +43,49 @@
43
  warnings.warn(msg)
44
 
45
  [2026-02-25 14:29:24,307][dinov2][INFO] - using MLP layer as FFN
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
43
  warnings.warn(msg)
44
 
45
  [2026-02-25 14:29:24,307][dinov2][INFO] - using MLP layer as FFN
46
+ [2026-02-25 14:31:48,045][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=31` in the `DataLoader` to improve performance.
47
+
48
+ [2026-02-25 14:31:48,046][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
49
+ warnings.warn( # warn only once
50
+
51
+ [2026-02-25 14:31:50,895][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
52
+ result[selector] = overlay
53
+
54
+ [2026-02-25 14:31:50,904][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
55
+
56
+ [2026-02-25 14:31:50,905][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
57
+ warnings.warn(
58
+
59
+ [2026-02-25 14:31:50,906][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
60
+ warnings.warn(msg)
61
+
62
+ [2026-02-25 14:31:52,827][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4322.)
63
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
64
+
65
+ [2026-02-25 14:31:53,133][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
66
+
67
+ [2026-02-25 14:31:53,134][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
68
+
69
+ [2026-02-25 14:31:53,134][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
70
+
71
+ [2026-02-25 14:31:53,135][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
72
+
73
+ [2026-02-25 14:31:53,135][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
74
+
75
+ [2026-02-25 14:32:07,271][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
76
+ grad.sizes() = [2, 32, 1, 1], strides() = [32, 1, 32, 32]
77
+ bucket_view.sizes() = [2, 32, 1, 1], strides() = [32, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
78
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
79
+
80
+ [2026-02-25 14:32:07,422][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
81
+ result[selector] = overlay
82
+
83
+ [2026-02-25 14:34:04,202][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
84
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
85
+
86
+ [2026-02-25 14:47:38,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
87
+ result[selector] = overlay
88
+
89
+ [2026-02-25 14:51:37,911][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
90
+ result[selector] = overlay
91
+
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_1.log CHANGED
@@ -17,3 +17,20 @@
17
  warnings.warn(msg)
18
 
19
  [2026-02-25 14:30:04,698][dinov2][INFO] - using MLP layer as FFN
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  warnings.warn(msg)
18
 
19
  [2026-02-25 14:30:04,698][dinov2][INFO] - using MLP layer as FFN
20
+ [2026-02-25 14:31:48,046][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
21
+ warnings.warn( # warn only once
22
+
23
+ [2026-02-25 14:32:07,271][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
24
+ grad.sizes() = [2, 32, 1, 1], strides() = [32, 1, 32, 32]
25
+ bucket_view.sizes() = [2, 32, 1, 1], strides() = [32, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
26
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
27
+
28
+ [2026-02-25 14:32:07,384][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
29
+ result[selector] = overlay
30
+
31
+ [2026-02-25 14:34:04,202][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
32
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
33
+
34
+ [2026-02-25 14:47:38,498][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
35
+ result[selector] = overlay
36
+
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_2.log CHANGED
@@ -17,3 +17,20 @@
17
  warnings.warn(msg)
18
 
19
  [2026-02-25 14:30:17,034][dinov2][INFO] - using MLP layer as FFN
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  warnings.warn(msg)
18
 
19
  [2026-02-25 14:30:17,034][dinov2][INFO] - using MLP layer as FFN
20
+ [2026-02-25 14:31:48,046][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
21
+ warnings.warn( # warn only once
22
+
23
+ [2026-02-25 14:32:07,277][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
24
+ grad.sizes() = [2, 32, 1, 1], strides() = [32, 1, 32, 32]
25
+ bucket_view.sizes() = [2, 32, 1, 1], strides() = [32, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
26
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
27
+
28
+ [2026-02-25 14:32:07,386][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
29
+ result[selector] = overlay
30
+
31
+ [2026-02-25 14:34:04,202][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
32
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
33
+
34
+ [2026-02-25 14:47:38,498][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
35
+ result[selector] = overlay
36
+
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_3.log CHANGED
@@ -17,3 +17,20 @@
17
  warnings.warn(msg)
18
 
19
  [2026-02-25 14:30:17,604][dinov2][INFO] - using MLP layer as FFN
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  warnings.warn(msg)
18
 
19
  [2026-02-25 14:30:17,604][dinov2][INFO] - using MLP layer as FFN
20
+ [2026-02-25 14:31:48,046][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
21
+ warnings.warn( # warn only once
22
+
23
+ [2026-02-25 14:32:07,273][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
24
+ grad.sizes() = [2, 32, 1, 1], strides() = [32, 1, 32, 32]
25
+ bucket_view.sizes() = [2, 32, 1, 1], strides() = [32, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
26
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
27
+
28
+ [2026-02-25 14:32:07,396][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
29
+ result[selector] = overlay
30
+
31
+ [2026-02-25 14:34:04,233][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
32
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
33
+
34
+ [2026-02-25 14:47:38,498][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
35
+ result[selector] = overlay
36
+
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_4.log CHANGED
@@ -17,3 +17,20 @@
17
  warnings.warn(msg)
18
 
19
  [2026-02-25 14:30:17,884][dinov2][INFO] - using MLP layer as FFN
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  warnings.warn(msg)
18
 
19
  [2026-02-25 14:30:17,884][dinov2][INFO] - using MLP layer as FFN
20
+ [2026-02-25 14:31:48,047][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
21
+ warnings.warn( # warn only once
22
+
23
+ [2026-02-25 14:32:07,276][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
24
+ grad.sizes() = [2, 32, 1, 1], strides() = [32, 1, 32, 32]
25
+ bucket_view.sizes() = [2, 32, 1, 1], strides() = [32, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
26
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
27
+
28
+ [2026-02-25 14:32:07,389][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
29
+ result[selector] = overlay
30
+
31
+ [2026-02-25 14:34:04,222][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
32
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
33
+
34
+ [2026-02-25 14:47:38,498][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
35
+ result[selector] = overlay
36
+
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_5.log CHANGED
@@ -17,3 +17,20 @@
17
  warnings.warn(msg)
18
 
19
  [2026-02-25 14:30:13,767][dinov2][INFO] - using MLP layer as FFN
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  warnings.warn(msg)
18
 
19
  [2026-02-25 14:30:13,767][dinov2][INFO] - using MLP layer as FFN
20
+ [2026-02-25 14:31:48,046][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
21
+ warnings.warn( # warn only once
22
+
23
+ [2026-02-25 14:32:06,540][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
24
+ grad.sizes() = [2, 32, 1, 1], strides() = [32, 1, 32, 32]
25
+ bucket_view.sizes() = [2, 32, 1, 1], strides() = [32, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
26
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
27
+
28
+ [2026-02-25 14:32:07,384][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
29
+ result[selector] = overlay
30
+
31
+ [2026-02-25 14:34:04,218][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
32
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
33
+
34
+ [2026-02-25 14:47:38,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
35
+ result[selector] = overlay
36
+
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_6.log CHANGED
@@ -17,3 +17,20 @@
17
  warnings.warn(msg)
18
 
19
  [2026-02-25 14:30:07,824][dinov2][INFO] - using MLP layer as FFN
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  warnings.warn(msg)
18
 
19
  [2026-02-25 14:30:07,824][dinov2][INFO] - using MLP layer as FFN
20
+ [2026-02-25 14:31:48,046][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
21
+ warnings.warn( # warn only once
22
+
23
+ [2026-02-25 14:32:06,759][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
24
+ grad.sizes() = [2, 32, 1, 1], strides() = [32, 1, 32, 32]
25
+ bucket_view.sizes() = [2, 32, 1, 1], strides() = [32, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
26
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
27
+
28
+ [2026-02-25 14:32:07,391][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
29
+ result[selector] = overlay
30
+
31
+ [2026-02-25 14:34:04,202][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
32
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
33
+
34
+ [2026-02-25 14:47:38,498][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
35
+ result[selector] = overlay
36
+
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_7.log CHANGED
@@ -17,3 +17,20 @@
17
  warnings.warn(msg)
18
 
19
  [2026-02-25 14:30:04,401][dinov2][INFO] - using MLP layer as FFN
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  warnings.warn(msg)
18
 
19
  [2026-02-25 14:30:04,401][dinov2][INFO] - using MLP layer as FFN
20
+ [2026-02-25 14:31:48,046][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
21
+ warnings.warn( # warn only once
22
+
23
+ [2026-02-25 14:32:06,759][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
24
+ grad.sizes() = [2, 32, 1, 1], strides() = [32, 1, 32, 32]
25
+ bucket_view.sizes() = [2, 32, 1, 1], strides() = [32, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
26
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
27
+
28
+ [2026-02-25 14:32:07,388][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
29
+ result[selector] = overlay
30
+
31
+ [2026-02-25 14:34:04,202][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
32
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
33
+
34
+ [2026-02-25 14:47:38,498][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
35
+ result[selector] = overlay
36
+
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/debug-internal.log CHANGED
@@ -1,11 +1,6 @@
1
- {"time":"2026-02-25T13:54:42.16430492Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
2
- {"time":"2026-02-25T13:54:42.614803801Z","level":"INFO","msg":"stream: created new stream","id":"g2qj8ry7"}
3
- {"time":"2026-02-25T13:54:42.615183417Z","level":"INFO","msg":"handler: started","stream_id":"g2qj8ry7"}
4
- {"time":"2026-02-25T13:54:42.615386361Z","level":"INFO","msg":"stream: started","id":"g2qj8ry7"}
5
- {"time":"2026-02-25T13:54:42.615431072Z","level":"INFO","msg":"sender: started","stream_id":"g2qj8ry7"}
6
- {"time":"2026-02-25T13:54:42.615463052Z","level":"INFO","msg":"writer: started","stream_id":"g2qj8ry7"}
7
- {"time":"2026-02-25T13:55:10.309513894Z","level":"INFO","msg":"stream: closing","id":"g2qj8ry7"}
8
- {"time":"2026-02-25T13:55:10.794171575Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
9
- {"time":"2026-02-25T13:55:11.099789022Z","level":"INFO","msg":"handler: closed","stream_id":"g2qj8ry7"}
10
- {"time":"2026-02-25T13:55:11.099961565Z","level":"INFO","msg":"sender: closed","stream_id":"g2qj8ry7"}
11
- {"time":"2026-02-25T13:55:11.099983596Z","level":"INFO","msg":"stream: closed","id":"g2qj8ry7"}
 
1
+ {"time":"2026-02-25T14:31:41.445785511Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
2
+ {"time":"2026-02-25T14:31:41.986197956Z","level":"INFO","msg":"stream: created new stream","id":"pa77a1wq"}
3
+ {"time":"2026-02-25T14:31:41.986357928Z","level":"INFO","msg":"handler: started","stream_id":"pa77a1wq"}
4
+ {"time":"2026-02-25T14:31:41.986655173Z","level":"INFO","msg":"stream: started","id":"pa77a1wq"}
5
+ {"time":"2026-02-25T14:31:41.986647023Z","level":"INFO","msg":"writer: started","stream_id":"pa77a1wq"}
6
+ {"time":"2026-02-25T14:31:41.986720105Z","level":"INFO","msg":"sender: started","stream_id":"pa77a1wq"}
 
 
 
 
 
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/debug.log CHANGED
@@ -1,21 +1,19 @@
1
- 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
2
- 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_setup.py:_flush():81] Configure stats pid to 168367
3
- 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
- 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug.log
5
- 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug-internal.log
6
- 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:init():844] calling init triggers
7
- 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
- config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'depth', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'absgrad', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': True, 'voxelize_activate': True, 'use_depth': True}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'density_control_loss': {'error_score': {'weight': 0.01, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0225_ctxTrain_depth_vggtDistl', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': True, 'vggt_cam_loss': True, 'vggt_distil': True, 'context_view_train': True}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
9
- 2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:init():892] starting backend
10
- 2026-02-25 13:54:42,153 INFO MainThread:168367 [wandb_init.py:init():895] sending inform_init request
11
- 2026-02-25 13:54:42,160 INFO MainThread:168367 [wandb_init.py:init():903] backend started and connected
12
- 2026-02-25 13:54:42,166 INFO MainThread:168367 [wandb_init.py:init():973] updated telemetry
13
- 2026-02-25 13:54:42,175 INFO MainThread:168367 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
- 2026-02-25 13:54:43,768 INFO MainThread:168367 [wandb_init.py:init():1042] starting run threads in backend
15
- 2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_console_start():2524] atexit reg
16
- 2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_redirect():2373] redirect: wrap_raw
17
- 2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
- 2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_redirect():2465] Redirects installed.
19
- 2026-02-25 13:54:43,896 INFO MainThread:168367 [wandb_init.py:init():1082] run started, returning control to user process
20
- 2026-02-25 13:55:10,309 INFO wandb-AsyncioManager-main:168367 [service_client.py:_forward_responses():134] Reached EOF.
21
- 2026-02-25 13:55:10,309 INFO wandb-AsyncioManager-main:168367 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.
 
1
+ 2026-02-25 14:31:41,152 INFO MainThread:180369 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
2
+ 2026-02-25 14:31:41,152 INFO MainThread:180369 [wandb_setup.py:_flush():81] Configure stats pid to 180369
3
+ 2026-02-25 14:31:41,152 INFO MainThread:180369 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
+ 2026-02-25 14:31:41,152 INFO MainThread:180369 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/logs/debug.log
5
+ 2026-02-25 14:31:41,152 INFO MainThread:180369 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/logs/debug-internal.log
6
+ 2026-02-25 14:31:41,152 INFO MainThread:180369 [wandb_init.py:init():844] calling init triggers
7
+ 2026-02-25 14:31:41,153 INFO MainThread:180369 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
+ config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'depth', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'absgrad', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': True, 'voxelize_activate': True, 'use_depth': True}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}, 'depth_consis': {'weight': 1.0, 'sigma_image': None, 'use_second_derivative': False}}, 'density_control_loss': {'error_score': {'weight': 0.01, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0225_ctxTrain_depth_vggtDistl', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': True, 'vggt_cam_loss': True, 'vggt_distil': True, 'context_view_train': True}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
9
+ 2026-02-25 14:31:41,153 INFO MainThread:180369 [wandb_init.py:init():892] starting backend
10
+ 2026-02-25 14:31:41,435 INFO MainThread:180369 [wandb_init.py:init():895] sending inform_init request
11
+ 2026-02-25 14:31:41,442 INFO MainThread:180369 [wandb_init.py:init():903] backend started and connected
12
+ 2026-02-25 14:31:41,447 INFO MainThread:180369 [wandb_init.py:init():973] updated telemetry
13
+ 2026-02-25 14:31:41,455 INFO MainThread:180369 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
+ 2026-02-25 14:31:43,288 INFO MainThread:180369 [wandb_init.py:init():1042] starting run threads in backend
15
+ 2026-02-25 14:31:43,418 INFO MainThread:180369 [wandb_run.py:_console_start():2524] atexit reg
16
+ 2026-02-25 14:31:43,418 INFO MainThread:180369 [wandb_run.py:_redirect():2373] redirect: wrap_raw
17
+ 2026-02-25 14:31:43,418 INFO MainThread:180369 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
+ 2026-02-25 14:31:43,418 INFO MainThread:180369 [wandb_run.py:_redirect():2465] Redirects installed.
19
+ 2026-02-25 14:31:43,420 INFO MainThread:180369 [wandb_init.py:init():1082] run started, returning control to user process
 
 
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/active_mask_imgs_18_e74266ca2c24ca638ee2.png ADDED

Git LFS Details

  • SHA256: e74266ca2c24ca638ee2d36e14d016a3e341557bda56e88a35413e67221f9fdd
  • Pointer size: 131 Bytes
  • Size of remote file: 457 kB
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/active_mask_imgs_1_690a5f29fe9230f3ead3.png ADDED

Git LFS Details

  • SHA256: 690a5f29fe9230f3ead382844289ad1899dfe18260121766c39a025ce2d4d666
  • Pointer size: 131 Bytes
  • Size of remote file: 459 kB
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/comparison_0_e330894a93360b5335a1.png ADDED

Git LFS Details

  • SHA256: e330894a93360b5335a1b717dcffcd8052fe0dd86b7e75c58e226e943720bca4
  • Pointer size: 132 Bytes
  • Size of remote file: 1.66 MB
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/comparison_17_6e317bfdf78b6c03c329.png ADDED

Git LFS Details

  • SHA256: 6e317bfdf78b6c03c329bd10f19e7b53bd51e6bb11b2f87af9cfbb96e4cd5712
  • Pointer size: 131 Bytes
  • Size of remote file: 966 kB
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/error_scores_19_cebe8e2cd2233435a7a5.png ADDED

Git LFS Details

  • SHA256: cebe8e2cd2233435a7a5f8af5dd17807ef20a4b2d1539f887dd985b71bc06a00
  • Pointer size: 131 Bytes
  • Size of remote file: 409 kB
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/error_scores_2_bf819ed247f7546e0637.png ADDED

Git LFS Details

  • SHA256: bf819ed247f7546e0637f42c82a216400ea84ba86b3bd04140fd72d3da4c0adb
  • Pointer size: 131 Bytes
  • Size of remote file: 387 kB
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/train/comparison_14_2e2099ac37d79d14f71c.png ADDED

Git LFS Details

  • SHA256: 2e2099ac37d79d14f71cf79febbe876e6c99532a9f411dee4fe64fdbbd426641
  • Pointer size: 132 Bytes
  • Size of remote file: 8.79 MB
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/train/comparison_4_69221fdfe7cd493b747b.png ADDED

Git LFS Details

  • SHA256: 69221fdfe7cd493b747bf79d74d2f01bc0cbd4c35d0bc14b3e1403cc38b543ad
  • Pointer size: 133 Bytes
  • Size of remote file: 10.5 MB
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/train/error_scores_13_a60c0930fe11800b83fd.png ADDED

Git LFS Details

  • SHA256: a60c0930fe11800b83fd87c38939bd81743eee47708957cfd3ec9d99c2001330
  • Pointer size: 132 Bytes
  • Size of remote file: 6.98 MB
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/media/images/train/error_scores_3_0a2cb05f0b9c7f260e6d.png ADDED

Git LFS Details

  • SHA256: 0a2cb05f0b9c7f260e6d901c0bf4ca4530a3e1ff23f12d66e1d0f9f3e5684781
  • Pointer size: 132 Bytes
  • Size of remote file: 8.13 MB
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/output.log ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7]
2
+
3
+ | Name | Type | Params | Mode
4
+ ------------------------------------------------------------------------
5
+ 0 | encoder | OurSplat | 888 M | train
6
+ 1 | density_control_module | DensityControlModule | 2.6 M | train
7
+ 2 | decoder | DecoderSplattingCUDA | 0 | train
8
+ 3 | render_losses | ModuleList | 0 | train
9
+ 4 | density_control_losses | ModuleList | 0 | train
10
+ 5 | direct_losses | ModuleList | 0 | train
11
+ 6 | distill_aggregator | Aggregator | 909 M | train
12
+ 7 | distill_camera_head | CameraHead | 216 M | train
13
+ 8 | distill_depth_head | DPTHead | 32.7 M | train
14
+ 9 | loss_distill | DistillLoss | 0 | train
15
+ ------------------------------------------------------------------------
16
+ 891 M Trainable params
17
+ 1.2 B Non-trainable params
18
+ 2.0 B Total params
19
+ 8,196.093 Total estimated model params size (MB)
20
+ 2779 Modules in train mode
21
+ 522 Modules in eval mode
22
+ Sanity Checking: | | 0/? [00:00<?, ?it/s][2026-02-25 14:31:48,045][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=31` in the `DataLoader` to improve performance.
23
+
24
+ [2026-02-25 14:31:48,046][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
25
+ warnings.warn( # warn only once
26
+
27
+ Validation epoch start on rank 0
28
+ Sanity Checking DataLoader 0: 0%| | 0/1 [00:00<?, ?it/s]validation step 0; scene = ['306e2b7785657539'];
29
+ target intrinsic: tensor(0.8595, device='cuda:0') tensor(0.8597, device='cuda:0')
30
+ pred intrinsic: tensor(0.8779, device='cuda:0') tensor(0.8773, device='cuda:0')
31
+ [rank0]:W0225 14:31:50.831000 180369 site-packages/torch/utils/cpp_extension.py:2425] TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
32
+ [rank0]:W0225 14:31:50.831000 180369 site-packages/torch/utils/cpp_extension.py:2425] If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'] to specific architectures.
33
+ [2026-02-25 14:31:50,895][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
34
+ result[selector] = overlay
35
+
36
+ [2026-02-25 14:31:50,904][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
37
+
38
+ Setting up [LPIPS] perceptual loss: trunk [vgg], v[0.1], spatial [off]
39
+ [2026-02-25 14:31:50,905][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
40
+ warnings.warn(
41
+
42
+ [2026-02-25 14:31:50,906][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
43
+ warnings.warn(msg)
44
+
45
+ Loading model from: /venv/main/lib/python3.12/site-packages/lpips/weights/v0.1/vgg.pth
46
+ [2026-02-25 14:31:52,827][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4322.)
47
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
48
+
49
+ Sanity Checking DataLoader 0: 100%|████████████████████████████████████████████████████████████████████| 1/1 [00:04<00:00, 0.23it/s][2026-02-25 14:31:53,133][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
50
+
51
+ [2026-02-25 14:31:53,134][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
52
+
53
+ [2026-02-25 14:31:53,134][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
54
+
55
+ [2026-02-25 14:31:53,135][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
56
+
57
+ [2026-02-25 14:31:53,135][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
58
+
59
+ Epoch 0: | | 0/? [00:00<?, ?it/s]context = [[34, 36, 50, 53, 54, 60, 63, 70, 76, 78, 79, 80, 81, 88, 92, 94, 102, 110, 112, 114, 122, 125, 126, 131]]target = [[126, 96, 109, 55, 99, 116, 43, 60, 113, 85, 103, 90, 130, 62, 76, 123, 35, 102, 125, 128, 98, 67, 129, 79]]
60
+ [2026-02-25 14:32:07,271][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
61
+ grad.sizes() = [2, 32, 1, 1], strides() = [32, 1, 32, 32]
62
+ bucket_view.sizes() = [2, 32, 1, 1], strides() = [32, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
63
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
64
+
65
+ [2026-02-25 14:32:07,422][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
66
+ result[selector] = overlay
67
+
68
+ Epoch 0: | | 9/? [00:51<00:00, 0.17it/s, v_num=a1wq]train step 10; scene = [['08c26703c4987851']]; loss = 0.911580
69
+ Epoch 0: | | 10/? [00:56<00:00, 0.18it/s, v_num=a1wq]context = [[98, 107, 112, 119, 122, 123], [21, 22, 27, 36, 38, 46], [63, 66, 67, 77, 84, 88], [56, 62, 65, 68, 78, 81]]target = [[105, 110, 116, 112, 111, 101], [29, 37, 36, 22, 40, 45], [73, 66, 75, 87, 83, 64], [79, 69, 75, 58, 61, 62]]
70
+ Epoch 0: | | 19/? [01:38<00:00, 0.19it/s, v_num=a1wq]train step 20; scene = [['4012c15c8381568b'], ['af08406c5a9a43a0'], ['9f9f9beffb86fad7'], ['fc8d08df6c875cb0']]; loss = 0.309024
71
+ Epoch 0: | | 20/? [01:43<00:00, 0.19it/s, v_num=a1wq]context = [[144, 152, 157, 164, 166, 169, 171, 177], [201, 211, 216, 221, 228, 230, 233, 234], [11, 15, 16, 23, 30, 38, 43, 44]]target = [[153, 170, 149, 169, 145, 174, 165, 157], [229, 216, 205, 206, 203, 213, 233, 215], [37, 38, 39, 35, 15, 24, 19, 25]]
72
+ Epoch 0: | | 24/? [02:01<00:00, 0.20it/s, v_num=a1wq][2026-02-25 14:34:04,202][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
73
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
74
+
75
+ Epoch 0: | | 29/? [02:23<00:00, 0.20it/s, v_num=a1wq]train step 30; scene = [['00980329a3221f1c'], ['1e7c432d2207b6f2'], ['af2748330e5243d0']]; loss = 0.235074
76
+ Epoch 0: | | 30/? [02:27<00:00, 0.20it/s, v_num=a1wq]context = [[2, 7, 10, 15, 20, 24, 25, 29, 30, 32, 33, 37, 40, 44, 49, 60, 64, 74, 76, 81, 89, 91, 95, 99]]target = [[79, 84, 35, 43, 89, 44, 63, 58, 48, 13, 65, 7, 96, 27, 51, 20, 60, 15, 85, 59, 22, 66, 17, 62]]
77
+ Epoch 0: | | 39/? [03:09<00:00, 0.21it/s, v_num=a1wq]train step 40; scene = [['79a9385753d426bc'], ['593538382d2dc847'], ['c9c67636b9d521be']]; loss = 0.200153
78
+ Epoch 0: | | 40/? [03:13<00:00, 0.21it/s, v_num=a1wq]context = [[200, 208, 223, 225], [6, 8, 12, 31], [104, 107, 111, 129], [69, 80, 86, 95], [11, 15, 18, 37], [52, 60, 70, 78]]target = [[213, 203, 214, 201], [29, 11, 15, 28], [111, 122, 120, 115], [77, 90, 89, 78], [18, 24, 29, 14], [64, 67, 63, 74]]
79
+ Epoch 0: | | 49/? [03:53<00:00, 0.21it/s, v_num=a1wq]train step 50; scene = [['579a11551b3315d9'], ['c9dd64b7415e788e'], ['6f3fb517d1798d03']]; loss = 0.216190
80
+ Epoch 0: | | 50/? [03:58<00:00, 0.21it/s, v_num=a1wq]context = [[20, 22, 25, 28, 36, 47], [10, 13, 15, 22, 26, 35], [31, 47, 49, 51, 52, 56], [109, 110, 115, 116, 125, 135]]target = [[40, 24, 23, 30, 37, 33], [19, 32, 25, 24, 26, 28], [35, 52, 50, 55, 45, 36], [131, 123, 130, 110, 115, 134]]
81
+ Epoch 0: | | 59/? [04:38<00:00, 0.21it/s, v_num=a1wq]train step 60; scene = [['07916b8004a8e336'], ['e51ef9945ae527c4'], ['db84f84b1d775bb8'], ['92ed61f8e16b7e67']]; loss = 0.384119
82
+ Epoch 0: | | 60/? [04:43<00:00, 0.21it/s, v_num=a1wq]context = [[96, 97, 101, 102, 105, 106, 107, 114, 124, 134, 138, 145], [76, 78, 82, 83, 96, 98, 100, 111, 112, 118, 119, 125]]target = [[144, 124, 137, 102, 107, 130, 119, 129, 118, 123, 133, 126], [98, 108, 96, 100, 113, 123, 77, 78, 112, 109, 80, 102]]
83
+ Epoch 0: | | 69/? [05:25<00:00, 0.21it/s, v_num=a1wq]train step 70; scene = [['c34efa1505a0cfaa'], ['a3d0cca9fb57fd85'], ['43d0e6dce7bb1e95'], ['d8c2f0a3734cb493']]; loss = 0.215408
84
+ Epoch 0: | | 70/? [05:29<00:00, 0.21it/s, v_num=a1wq]context = [[204, 208, 211, 218, 232, 234, 237, 239, 240, 241, 247, 253], [6, 7, 10, 12, 35, 36, 37, 38, 40, 45, 48, 55]]target = [[230, 236, 222, 227, 220, 252, 216, 221, 213, 217, 249, 240], [18, 24, 25, 12, 9, 42, 34, 52, 41, 33, 43, 48]]
85
+ Epoch 0: | | 79/? [06:10<00:00, 0.21it/s, v_num=a1wq]train step 80; scene = [['24d756c820744e31'], ['cd6c21656a85e9b9'], ['f3b24cf238154fc0']]; loss = 0.172991
86
+ Epoch 0: | | 80/? [06:14<00:00, 0.21it/s, v_num=a1wq]context = [[4, 30], [52, 79], [61, 87], [12, 40], [83, 109], [3, 29], [221, 249], [198, 227], [9, 38], [46, 72], [0, 26], [123, 150]]target = [[27, 26], [58, 57], [70, 77], [26, 20], [87, 98], [26, 14], [223, 224], [221, 200], [23, 33], [67, 64], [4, 25], [140, 148]]
87
+ Epoch 0: | | 89/? [06:56<00:00, 0.21it/s, v_num=a1wq]train step 90; scene = [['617b4bc98d7e0bb6'], ['666e4a9aba27bb64']]; loss = 0.156331
88
+ Epoch 0: | | 90/? [07:01<00:00, 0.21it/s, v_num=a1wq]context = [[134, 135, 138, 146, 147, 151, 156, 161, 162, 164, 166, 167, 168, 169, 187, 189, 197, 210, 215, 224, 225, 228, 230, 231]]target = [[143, 149, 191, 182, 151, 226, 165, 140, 208, 171, 179, 223, 168, 136, 194, 207, 227, 144, 187, 185, 145, 218, 139, 170]]
89
+ Epoch 0: | | 99/? [07:44<00:00, 0.21it/s, v_num=a1wq]train step 100; scene = [['12fee7f1978d52f1'], ['c963bb60939e2d81']]; loss = 0.152084
90
+ Epoch 0: | | 100/? [07:48<00:00, 0.21it/s, v_num=a1wq]context = [[40, 44, 47, 48, 49, 50, 56, 63, 77, 79, 88, 89], [21, 33, 34, 35, 36, 40, 43, 46, 47, 60, 62, 70]]target = [[58, 81, 76, 64, 68, 72, 51, 87, 77, 65, 88, 45], [31, 47, 25, 35, 55, 22, 48, 65, 29, 40, 63, 67]]
91
+ Epoch 0: | | 109/? [08:30<00:00, 0.21it/s, v_num=a1wq]train step 110; scene = [['47396d5a5299873e']]; loss = 0.200728
92
+ Epoch 0: | | 110/? [08:34<00:00, 0.21it/s, v_num=a1wq]context = [[20, 22, 29, 30, 31, 34, 37, 45, 47, 51, 56, 69], [18, 22, 24, 27, 31, 37, 44, 48, 49, 60, 61, 67]]target = [[49, 22, 36, 59, 63, 60, 45, 66, 38, 28, 26, 64], [61, 65, 25, 20, 63, 60, 26, 22, 33, 45, 37, 35]]
93
+ Epoch 0: | | 119/? [09:17<00:00, 0.21it/s, v_num=a1wq]train step 120; scene = [['9bd7044e7cbf8e60'], ['76e44cf6b5658b26']]; loss = 0.090040
94
+ Epoch 0: | | 120/? [09:21<00:00, 0.21it/s, v_num=a1wq]context = [[7, 14, 26, 28, 31, 34, 37, 40], [10, 20, 26, 32, 35, 39, 41, 43], [17, 23, 26, 28, 44, 46, 49, 50]]target = [[14, 26, 10, 22, 30, 13, 31, 11], [24, 31, 36, 14, 40, 34, 41, 13], [24, 44, 34, 48, 46, 36, 26, 38]]
95
+ Epoch 0: | | 129/? [10:03<00:00, 0.21it/s, v_num=a1wq]train step 130; scene = [['a8cef6a851fbea3c'], ['b6699f4d039a5b06'], ['55cf2bbe9e017ea4'], ['6b0dd861e1ab1fec'], ['14db202c335af709'], ['8b6ff6c5153a7794'], ['b75f3820760d835c'], ['f7dbc855fd2a7669'], ['cfb20f8971e6a591'], ['95f2be7bb8303f50'], ['ff422469e034ae11'], ['5a2ad43377e9d18d']]; loss = 0.172321
96
+ Epoch 0: | | 130/? [10:08<00:00, 0.21it/s, v_num=a1wq]context = [[8, 15, 26, 30, 32, 35, 40, 46, 47, 48, 50, 53, 54, 62, 69, 70, 72, 76, 80, 86, 90, 99, 100, 105]]target = [[26, 12, 49, 100, 89, 24, 10, 81, 37, 63, 52, 17, 39, 70, 16, 56, 40, 55, 43, 34, 72, 28, 48, 45]]
97
+ Epoch 0: | | 139/? [10:49<00:00, 0.21it/s, v_num=a1wq]train step 140; scene = [['f62a962df5c26a1a'], ['b076420679a04731']]; loss = 0.112418
98
+ Epoch 0: | | 140/? [10:53<00:00, 0.21it/s, v_num=a1wq]context = [[90, 95, 97, 104, 118, 121], [14, 29, 34, 35, 36, 44], [12, 14, 22, 27, 34, 41], [6, 10, 16, 18, 34, 35]]target = [[99, 104, 93, 113, 110, 96], [29, 37, 22, 21, 39, 41], [19, 25, 39, 27, 29, 13], [26, 28, 9, 7, 13, 34]]
99
+ Epoch 0: | | 149/? [11:36<00:00, 0.21it/s, v_num=a1wq]train step 150; scene = [['a52d26a78b04aebd']]; loss = 0.094999
100
+ Epoch 0: | | 150/? [11:41<00:00, 0.21it/s, v_num=a1wq]context = [[115, 132, 134, 145], [16, 32, 39, 44], [57, 63, 78, 88], [9, 11, 23, 36], [15, 26, 28, 44], [1, 8, 22, 33]]target = [[124, 132, 116, 119], [27, 24, 33, 29], [81, 61, 85, 79], [28, 26, 29, 16], [28, 29, 37, 22], [9, 4, 32, 27]]
101
+ Epoch 0: | | 159/? [12:22<00:00, 0.21it/s, v_num=a1wq]train step 160; scene = [['268fbffc6c479d5b']]; loss = 0.088229
102
+ Epoch 0: | | 160/? [12:27<00:00, 0.21it/s, v_num=a1wq]context = [[18, 25, 26, 37, 42, 46, 49, 51, 53, 64, 65, 67], [69, 75, 78, 79, 82, 84, 94, 95, 104, 108, 117, 118]]target = [[53, 27, 22, 32, 41, 38, 50, 43, 47, 48, 23, 19], [74, 70, 114, 115, 90, 89, 88, 92, 94, 110, 107, 101]]
103
+ Epoch 0: | | 169/? [13:10<00:00, 0.21it/s, v_num=a1wq]train step 170; scene = [['719e2e8912e4eed3'], ['a3e51565a737569f']]; loss = 0.169240
104
+ Epoch 0: | | 170/? [13:15<00:00, 0.21it/s, v_num=a1wq]context = [[14, 18, 20, 21, 24, 27, 33, 40, 45, 47, 48, 51, 52, 60, 64, 70, 75, 77, 80, 85, 90, 98, 102, 111]]target = [[29, 93, 32, 39, 81, 108, 72, 107, 51, 35, 16, 36, 70, 18, 34, 92, 94, 47, 23, 74, 50, 77, 19, 37]]
105
+ Epoch 0: | | 179/? [13:56<00:00, 0.21it/s, v_num=a1wq]train step 180; scene = [['f44b9aa76a94a0a3']]; loss = 0.083026
106
+ Epoch 0: | | 180/? [14:01<00:00, 0.21it/s, v_num=a1wq]context = [[0, 6, 17, 22, 26, 28, 33, 41, 50, 55, 56, 57, 71, 76, 79, 81, 84, 85, 86, 87, 89, 95, 96, 97]]target = [[37, 49, 12, 78, 9, 16, 84, 13, 5, 4, 6, 38, 80, 51, 43, 68, 64, 46, 56, 24, 25, 72, 36, 21]]
107
+ Epoch 0: | | 189/? [14:44<00:00, 0.21it/s, v_num=a1wq]train step 190; scene = [['71bb669d936a5718'], ['a47203cfd5e0a478'], ['4b009f82cf5c7098']]; loss = 0.124435
108
+ Epoch 0: | | 190/? [14:49<00:00, 0.21it/s, v_num=a1wq]context = [[9, 11, 19, 25, 33, 43, 46, 47, 48, 54, 57, 58, 63, 70, 72, 75, 79, 80, 83, 84, 85, 96, 99, 106]]target = [[30, 53, 82, 31, 55, 12, 20, 72, 104, 70, 24, 52, 21, 32, 102, 71, 35, 11, 46, 10, 15, 74, 33, 26]]
109
+ Epoch 0: | | 199/? [15:30<00:00, 0.21it/s, v_num=a1wq]train step 200; scene = [['dd5ec950a01c42a0'], ['6d0db0358f7e051e'], ['983fe650a925ec1b']]; loss = 0.127213
110
+ Epoch 0: | | 200/? [15:35<00:00, 0.21it/s, v_num=a1wq]context = [[8, 10, 14, 15, 22, 23, 27, 30, 36, 38, 39, 47, 64, 65, 67, 78, 79, 80, 83, 86, 93, 96, 98, 105]]target = [[73, 63, 10, 27, 89, 35, 44, 58, 97, 71, 17, 24, 66, 87, 50, 12, 23, 11, 31, 45, 69, 96, 98, 94]]
111
+ [2026-02-25 14:47:38,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
112
+ result[selector] = overlay
113
+
114
+ Epoch 0: | | 209/? [16:24<00:00, 0.21it/s, v_num=a1wq]train step 210; scene = [['9be9b273b3c22c61'], ['4b5883872c9b860c']]; loss = 0.084080
115
+ Epoch 0: | | 210/? [16:29<00:00, 0.21it/s, v_num=a1wq]context = [[34, 35, 37, 54, 56, 59, 61, 68, 69, 76, 90, 96, 98, 101, 103, 107, 116, 119, 120, 121, 122, 125, 126, 131]]target = [[95, 50, 119, 121, 84, 107, 72, 52, 80, 42, 127, 94, 79, 98, 46, 128, 73, 75, 106, 92, 37, 110, 96, 56]]
116
+ Epoch 0: | | 219/? [17:12<00:00, 0.21it/s, v_num=a1wq]train step 220; scene = [['a3b6faa8d238d993'], ['df9ba36fbe753843']]; loss = 0.142685
117
+ Epoch 0: | | 220/? [17:16<00:00, 0.21it/s, v_num=a1wq]context = [[39, 71, 74], [21, 41, 51], [28, 55, 59], [2, 31, 37], [15, 25, 48], [57, 64, 93], [76, 87, 105], [31, 53, 64]]target = [[52, 43, 59], [28, 25, 49], [48, 58, 51], [26, 9, 6], [30, 32, 47], [84, 88, 59], [104, 87, 88], [42, 52, 59]]
118
+ Epoch 0: | | 229/? [17:59<00:00, 0.21it/s, v_num=a1wq]train step 230; scene = [['ca04de3c55cd1ca0'], ['3d90d586b33daa63'], ['d1772c09b4b6d95f'], ['03d05f69a1cab4f8'], ['60d296908f43a97a'], ['37c400e282bc481e']]; loss = 0.123628
119
+ Epoch 0: | | 230/? [18:03<00:00, 0.21it/s, v_num=a1wq]context = [[203, 204, 208, 209, 210, 233], [41, 42, 43, 52, 60, 76], [2, 8, 9, 23, 29, 31], [70, 74, 92, 97, 98, 100]]target = [[226, 223, 210, 217, 228, 222], [46, 70, 71, 75, 43, 56], [24, 23, 29, 25, 3, 21], [95, 81, 74, 73, 98, 92]]
120
+ Epoch 0: | | 239/? [18:46<00:00, 0.21it/s, v_num=a1wq]train step 240; scene = [['9794641b7e015578']]; loss = 0.228188
121
+ Epoch 0: | | 240/? [18:51<00:00, 0.21it/s, v_num=a1wq]context = [[62, 65, 68, 71, 80, 86, 87, 96, 99, 101, 103, 111], [134, 138, 139, 142, 151, 163, 171, 172, 173, 174, 181, 183]]target = [[87, 89, 96, 93, 103, 71, 65, 77, 63, 98, 102, 105], [165, 177, 159, 147, 138, 152, 171, 141, 181, 146, 161, 176]]
122
+ Epoch 0: | | 249/? [19:33<00:00, 0.21it/s, v_num=a1wq]train step 250; scene = [['93dff1b985f2c7f9']]; loss = 0.084510
123
+ Epoch 0: | | 250/? [19:38<00:00, 0.21it/s, v_num=a1wq]Validation epoch start on rank 0
124
+ Validation: | | 0/? [00:00<?, ?it/s]validation step 250; scene = ['49b8f80c849dc341'];
125
+ target intrinsic: tensor(0.8891, device='cuda:0') tensor(0.8894, device='cuda:0') | 0/1 [00:00<?, ?it/s]
126
+ pred intrinsic: tensor(0.8701, device='cuda:0') tensor(0.8341, device='cuda:0')
127
+ [2026-02-25 14:51:37,911][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
128
+ result[selector] = overlay
129
+
130
+ Epoch 0: | | 250/? [19:40<00:00, 0.21it/s, v_num=a1wq]context = [[113, 115, 120, 131, 135, 137, 144, 152], [16, 20, 23, 28, 32, 35, 39, 52], [14, 18, 25, 26, 28, 29, 51, 53]]target = [[149, 139, 118, 151, 137, 141, 121, 130], [45, 51, 28, 50, 35, 37, 27, 23], [29, 26, 22, 51, 24, 34, 47, 45]]
131
+ Epoch 0: | | 259/? [20:20<00:00, 0.21it/s, v_num=a1wq]train step 260; scene = [['b2288bf7003d5d4d']]; loss = 0.117824
132
+ Epoch 0: | | 260/? [20:25<00:00, 0.21it/s, v_num=a1wq]context = [[206, 208, 212, 216, 221, 236], [58, 60, 68, 69, 70, 93], [15, 19, 26, 34, 41, 44], [15, 18, 23, 34, 42, 44]]target = [[225, 216, 234, 214, 207, 232], [88, 76, 91, 67, 64, 70], [20, 29, 22, 36, 23, 40], [42, 28, 31, 30, 16, 37]]
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/requirements.txt ADDED
@@ -0,0 +1,172 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wheel==0.45.1
2
+ pytz==2025.2
3
+ easydict==1.13
4
+ antlr4-python3-runtime==4.9.3
5
+ wadler_lindig==0.1.7
6
+ urllib3==2.5.0
7
+ tzdata==2025.2
8
+ typing-inspection==0.4.1
9
+ tabulate==0.9.0
10
+ smmap==5.0.2
11
+ kornia_rs==0.1.9
12
+ setuptools==78.1.1
13
+ safetensors==0.5.3
14
+ PyYAML==6.0.2
15
+ PySocks==1.7.1
16
+ pyparsing==3.2.5
17
+ pydantic_core==2.33.2
18
+ pycparser==2.23
19
+ protobuf==6.32.1
20
+ propcache==0.3.2
21
+ proglog==0.1.12
22
+ fsspec==2024.6.1
23
+ platformdirs==4.4.0
24
+ pip==25.2
25
+ pillow==10.4.0
26
+ frozenlist==1.7.0
27
+ packaging==24.2
28
+ opt_einsum==3.4.0
29
+ numpy==1.26.4
30
+ ninja==1.13.0
31
+ fonttools==4.60.0
32
+ networkx==3.4.2
33
+ multidict==6.6.4
34
+ mdurl==0.1.2
35
+ MarkupSafe==3.0.2
36
+ kiwisolver==1.4.9
37
+ imageio-ffmpeg==0.6.0
38
+ idna==3.7
39
+ hf-xet==1.1.10
40
+ gmpy2==2.2.1
41
+ einops==0.8.1
42
+ filelock==3.17.0
43
+ decorator==4.4.2
44
+ dacite==1.9.2
45
+ cycler==0.12.1
46
+ colorama==0.4.6
47
+ click==8.3.0
48
+ nvidia-nvtx-cu12==12.8.90
49
+ charset-normalizer==3.3.2
50
+ certifi==2025.8.3
51
+ beartype==0.19.0
52
+ attrs==25.3.0
53
+ async-timeout==5.0.1
54
+ annotated-types==0.7.0
55
+ aiohappyeyeballs==2.6.1
56
+ yarl==1.20.1
57
+ tifffile==2025.5.10
58
+ sentry-sdk==2.39.0
59
+ scipy==1.15.3
60
+ pydantic==2.11.9
61
+ pandas==2.3.2
62
+ opencv-python==4.11.0.86
63
+ omegaconf==2.3.0
64
+ markdown-it-py==4.0.0
65
+ lightning-utilities==0.14.3
66
+ lazy_loader==0.4
67
+ jaxtyping==0.2.37
68
+ imageio==2.37.0
69
+ gitdb==4.0.12
70
+ contourpy==1.3.2
71
+ colorspacious==1.1.2
72
+ cffi==1.17.1
73
+ aiosignal==1.4.0
74
+ scikit-video==1.1.11
75
+ scikit-image==0.25.2
76
+ rich==14.1.0
77
+ moviepy==1.0.3
78
+ matplotlib==3.10.6
79
+ hydra-core==1.3.2
80
+ nvidia-nccl-cu12==2.27.3
81
+ huggingface-hub==0.35.1
82
+ GitPython==3.1.45
83
+ brotlicffi==1.0.9.2
84
+ aiohttp==3.12.15
85
+ torchmetrics==1.8.2
86
+ opt-einsum-fx==0.1.4
87
+ kornia==0.8.1
88
+ pytorch-lightning==2.5.1
89
+ lpips==0.1.4
90
+ e3nn==0.6.0
91
+ lightning==2.5.1
92
+ nvidia-cusparselt-cu12==0.7.1
93
+ triton==3.4.0
94
+ nvidia-nvjitlink-cu12==12.8.93
95
+ nvidia-curand-cu12==10.3.9.90
96
+ nvidia-cufile-cu12==1.13.1.3
97
+ nvidia-cuda-runtime-cu12==12.8.90
98
+ nvidia-cuda-nvrtc-cu12==12.8.93
99
+ nvidia-cuda-cupti-cu12==12.8.90
100
+ nvidia-cublas-cu12==12.8.4.1
101
+ nvidia-cusparse-cu12==12.5.8.93
102
+ nvidia-cufft-cu12==11.3.3.83
103
+ nvidia-cudnn-cu12==9.10.2.21
104
+ nvidia-cusolver-cu12==11.7.3.90
105
+ torch==2.8.0+cu128
106
+ torchvision==0.23.0+cu128
107
+ torchaudio==2.8.0+cu128
108
+ torch_scatter==2.1.2+pt28cu128
109
+ gsplat==1.5.3
110
+ wandb==0.25.0
111
+ cuda-bindings==13.0.3
112
+ cuda-pathfinder==1.3.3
113
+ Jinja2==3.1.6
114
+ mpmath==1.3.0
115
+ nvidia-cublas==13.1.0.3
116
+ nvidia-cuda-cupti==13.0.85
117
+ nvidia-cuda-nvrtc==13.0.88
118
+ nvidia-cuda-runtime==13.0.96
119
+ nvidia-cudnn-cu13==9.15.1.9
120
+ nvidia-cufft==12.0.0.61
121
+ nvidia-cufile==1.15.1.6
122
+ nvidia-curand==10.4.0.35
123
+ nvidia-cusolver==12.0.4.66
124
+ nvidia-cusparse==12.6.3.3
125
+ nvidia-cusparselt-cu13==0.8.0
126
+ nvidia-nccl-cu13==2.28.9
127
+ nvidia-nvjitlink==13.0.88
128
+ nvidia-nvshmem-cu13==3.4.5
129
+ nvidia-nvtx==13.0.85
130
+ requests==2.32.5
131
+ sentencepiece==0.2.1
132
+ sympy==1.14.0
133
+ torchcodec==0.10.0
134
+ torchdata==0.10.0
135
+ torchtext==0.6.0
136
+ anyio==4.12.0
137
+ asttokens==3.0.1
138
+ comm==0.2.3
139
+ debugpy==1.8.19
140
+ executing==2.2.1
141
+ h11==0.16.0
142
+ httpcore==1.0.9
143
+ httpx==0.28.1
144
+ ipykernel==7.1.0
145
+ ipython==9.8.0
146
+ ipython_pygments_lexers==1.1.1
147
+ ipywidgets==8.1.8
148
+ jedi==0.19.2
149
+ jupyter_client==8.7.0
150
+ jupyter_core==5.9.1
151
+ jupyterlab_widgets==3.0.16
152
+ matplotlib-inline==0.2.1
153
+ nest-asyncio==1.6.0
154
+ parso==0.8.5
155
+ pexpect==4.9.0
156
+ prompt_toolkit==3.0.52
157
+ psutil==7.2.1
158
+ ptyprocess==0.7.0
159
+ pure_eval==0.2.3
160
+ Pygments==2.19.2
161
+ python-dateutil==2.9.0.post0
162
+ pyzmq==27.1.0
163
+ shellingham==1.5.4
164
+ six==1.17.0
165
+ stack-data==0.6.3
166
+ tornado==6.5.4
167
+ tqdm==4.67.1
168
+ traitlets==5.14.3
169
+ typer-slim==0.21.0
170
+ typing_extensions==4.15.0
171
+ wcwidth==0.2.14
172
+ widgetsnbextension==4.0.15
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/files/wandb-metadata.json ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.8.0-90-generic-x86_64-with-glibc2.39",
3
+ "python": "CPython 3.12.12",
4
+ "startedAt": "2026-02-25T14:31:41.148858Z",
5
+ "args": [
6
+ "+experiment=re10k_ablation_24v_depthloss",
7
+ "wandb.mode=online",
8
+ "wandb.name=ABLATION_0225_ctxTrain_depth_vggtDistl",
9
+ "model.encoder.head_mode=depth",
10
+ "model.density_control.use_depth=true",
11
+ "train.context_view_train=true",
12
+ "train.vggt_distil=true"
13
+ ],
14
+ "program": "-m src.main",
15
+ "git": {
16
+ "remote": "git@github.com:K-nowing/CVPR2026.git",
17
+ "commit": "488f8178913cf0d68a295e744f2709f488f576cd"
18
+ },
19
+ "email": "dna9041@korea.ac.kr",
20
+ "root": "/workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl",
21
+ "host": "27d18dedec6d",
22
+ "executable": "/venv/main/bin/python",
23
+ "cpu_count": 128,
24
+ "cpu_count_logical": 256,
25
+ "gpu": "NVIDIA H200",
26
+ "gpu_count": 8,
27
+ "disk": {
28
+ "/": {
29
+ "total": "1170378588160",
30
+ "used": "756877160448"
31
+ }
32
+ },
33
+ "memory": {
34
+ "total": "1622948257792"
35
+ },
36
+ "gpu_nvidia": [
37
+ {
38
+ "name": "NVIDIA H200",
39
+ "memoryTotal": "150754820096",
40
+ "cudaCores": 16896,
41
+ "architecture": "Hopper",
42
+ "uuid": "GPU-2649ab80-a3a6-5a1c-0fa5-12bc11bd75e9"
43
+ },
44
+ {
45
+ "name": "NVIDIA H200",
46
+ "memoryTotal": "150754820096",
47
+ "cudaCores": 16896,
48
+ "architecture": "Hopper",
49
+ "uuid": "GPU-e92921d9-c681-246f-af93-637e0dc938ca"
50
+ },
51
+ {
52
+ "name": "NVIDIA H200",
53
+ "memoryTotal": "150754820096",
54
+ "cudaCores": 16896,
55
+ "architecture": "Hopper",
56
+ "uuid": "GPU-ffe12ffc-9bb7-82de-5692-1ec0ee2e68d8"
57
+ },
58
+ {
59
+ "name": "NVIDIA H200",
60
+ "memoryTotal": "150754820096",
61
+ "cudaCores": 16896,
62
+ "architecture": "Hopper",
63
+ "uuid": "GPU-499e5acd-b6ab-2010-c51b-ee9b5aa65825"
64
+ },
65
+ {
66
+ "name": "NVIDIA H200",
67
+ "memoryTotal": "150754820096",
68
+ "cudaCores": 16896,
69
+ "architecture": "Hopper",
70
+ "uuid": "GPU-3b2522d9-1c72-e49b-2c30-96165680b74a"
71
+ },
72
+ {
73
+ "name": "NVIDIA H200",
74
+ "memoryTotal": "150754820096",
75
+ "cudaCores": 16896,
76
+ "architecture": "Hopper",
77
+ "uuid": "GPU-a9a280c5-b2f9-dc1e-a8a9-7326a74001ff"
78
+ },
79
+ {
80
+ "name": "NVIDIA H200",
81
+ "memoryTotal": "150754820096",
82
+ "cudaCores": 16896,
83
+ "architecture": "Hopper",
84
+ "uuid": "GPU-07d0167b-a6a1-1900-2d27-7c6c11598409"
85
+ },
86
+ {
87
+ "name": "NVIDIA H200",
88
+ "memoryTotal": "150754820096",
89
+ "cudaCores": 16896,
90
+ "architecture": "Hopper",
91
+ "uuid": "GPU-8362a999-20d1-c27b-5d18-032d23f859ab"
92
+ }
93
+ ],
94
+ "cudaVersion": "13.1",
95
+ "writerId": "3uyx3143kqbde4nu8smcy1nl2irx3zfk"
96
+ }
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/logs/debug-core.log ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {"time":"2026-02-25T14:31:41.262957491Z","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpay480ore/port-180369.txt","pid":180369,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2026-02-25T14:31:41.263882206Z","level":"INFO","msg":"server: will exit if parent process dies","ppid":180369}
3
+ {"time":"2026-02-25T14:31:41.263868086Z","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-180369-183577-2731200046/socket","Net":"unix"}}
4
+ {"time":"2026-02-25T14:31:41.435038311Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2026-02-25T14:31:41.445530166Z","level":"INFO","msg":"handleInformInit: received","streamId":"pa77a1wq","id":"1(@)"}
6
+ {"time":"2026-02-25T14:31:41.986678614Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"pa77a1wq","id":"1(@)"}
7
+ {"time":"2026-02-25T14:31:48.420898309Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"il99tnokokrm"}
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/logs/debug-internal.log ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {"time":"2026-02-25T14:31:41.445785511Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
2
+ {"time":"2026-02-25T14:31:41.986197956Z","level":"INFO","msg":"stream: created new stream","id":"pa77a1wq"}
3
+ {"time":"2026-02-25T14:31:41.986357928Z","level":"INFO","msg":"handler: started","stream_id":"pa77a1wq"}
4
+ {"time":"2026-02-25T14:31:41.986655173Z","level":"INFO","msg":"stream: started","id":"pa77a1wq"}
5
+ {"time":"2026-02-25T14:31:41.986647023Z","level":"INFO","msg":"writer: started","stream_id":"pa77a1wq"}
6
+ {"time":"2026-02-25T14:31:41.986720105Z","level":"INFO","msg":"sender: started","stream_id":"pa77a1wq"}
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/logs/debug.log ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-02-25 14:31:41,152 INFO MainThread:180369 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
2
+ 2026-02-25 14:31:41,152 INFO MainThread:180369 [wandb_setup.py:_flush():81] Configure stats pid to 180369
3
+ 2026-02-25 14:31:41,152 INFO MainThread:180369 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
+ 2026-02-25 14:31:41,152 INFO MainThread:180369 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/logs/debug.log
5
+ 2026-02-25 14:31:41,152 INFO MainThread:180369 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/logs/debug-internal.log
6
+ 2026-02-25 14:31:41,152 INFO MainThread:180369 [wandb_init.py:init():844] calling init triggers
7
+ 2026-02-25 14:31:41,153 INFO MainThread:180369 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
+ config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'depth', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'absgrad', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': True, 'voxelize_activate': True, 'use_depth': True}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}, 'depth_consis': {'weight': 1.0, 'sigma_image': None, 'use_second_derivative': False}}, 'density_control_loss': {'error_score': {'weight': 0.01, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0225_ctxTrain_depth_vggtDistl', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': True, 'vggt_cam_loss': True, 'vggt_distil': True, 'context_view_train': True}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
9
+ 2026-02-25 14:31:41,153 INFO MainThread:180369 [wandb_init.py:init():892] starting backend
10
+ 2026-02-25 14:31:41,435 INFO MainThread:180369 [wandb_init.py:init():895] sending inform_init request
11
+ 2026-02-25 14:31:41,442 INFO MainThread:180369 [wandb_init.py:init():903] backend started and connected
12
+ 2026-02-25 14:31:41,447 INFO MainThread:180369 [wandb_init.py:init():973] updated telemetry
13
+ 2026-02-25 14:31:41,455 INFO MainThread:180369 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
+ 2026-02-25 14:31:43,288 INFO MainThread:180369 [wandb_init.py:init():1042] starting run threads in backend
15
+ 2026-02-25 14:31:43,418 INFO MainThread:180369 [wandb_run.py:_console_start():2524] atexit reg
16
+ 2026-02-25 14:31:43,418 INFO MainThread:180369 [wandb_run.py:_redirect():2373] redirect: wrap_raw
17
+ 2026-02-25 14:31:43,418 INFO MainThread:180369 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
+ 2026-02-25 14:31:43,418 INFO MainThread:180369 [wandb_run.py:_redirect():2465] Redirects installed.
19
+ 2026-02-25 14:31:43,420 INFO MainThread:180369 [wandb_init.py:init():1082] run started, returning control to user process
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_143141-pa77a1wq/run-pa77a1wq.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:394f2203dbb0ff42ca1548127b7fd3252e35179a1668e52a5564465d08167d9b
3
+ size 425984
DEBUG/.hydra/config.yaml ADDED
@@ -0,0 +1,185 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ encoder:
3
+ name: dcsplat
4
+ input_image_shape:
5
+ - 518
6
+ - 518
7
+ head_mode: depth
8
+ num_level: 3
9
+ gs_param_dim: 256
10
+ align_corners: false
11
+ use_voxelize: true
12
+ decoder:
13
+ name: splatting_cuda
14
+ background_color:
15
+ - 0.0
16
+ - 0.0
17
+ - 0.0
18
+ make_scale_invariant: false
19
+ density_control:
20
+ name: density_control_module
21
+ mean_dim: 32
22
+ gs_param_dim: 256
23
+ refinement_layer_num: 1
24
+ num_level: 3
25
+ grad_mode: absgrad
26
+ use_mean_features: true
27
+ refinement_type: voxelize
28
+ refinement_hidden_dim: 32
29
+ aggregation_mode: mean
30
+ num_heads: 1
31
+ score_mode: absgrad
32
+ latent_dim: 128
33
+ num_latents: 64
34
+ num_self_attn_per_block: 2
35
+ voxel_size: 0.001
36
+ aux_refine: false
37
+ refine_error: false
38
+ use_refine_module: true
39
+ voxelize_activate: true
40
+ use_depth: true
41
+ render_loss:
42
+ mse:
43
+ weight: 1.0
44
+ lpips:
45
+ weight: 0.05
46
+ apply_after_step: 0
47
+ density_control_loss:
48
+ error_score:
49
+ weight: 0.01
50
+ log_scale: false
51
+ grad_scale: 10000.0
52
+ mode: original
53
+ direct_loss:
54
+ l1:
55
+ weight: 0.8
56
+ ssim:
57
+ weight: 0.2
58
+ wandb:
59
+ project: DCSplat
60
+ entity: scene-representation-group
61
+ name: DEBUG
62
+ mode: disabled
63
+ tags:
64
+ - re10k
65
+ - 256x256
66
+ mode: train
67
+ data_loader:
68
+ train:
69
+ num_workers: 16
70
+ persistent_workers: true
71
+ batch_size: 16
72
+ seed: 1234
73
+ test:
74
+ num_workers: 4
75
+ persistent_workers: false
76
+ batch_size: 1
77
+ seed: 2345
78
+ val:
79
+ num_workers: 1
80
+ persistent_workers: true
81
+ batch_size: 1
82
+ seed: 3456
83
+ optimizer:
84
+ lr: 0.0002
85
+ warm_up_steps: 25
86
+ backbone_lr_multiplier: 0.1
87
+ backbone_trainable: T+H
88
+ accumulate: 1
89
+ checkpointing:
90
+ load: null
91
+ every_n_train_steps: 1500
92
+ save_top_k: 2
93
+ save_weights_only: false
94
+ train:
95
+ extended_visualization: false
96
+ print_log_every_n_steps: 10
97
+ camera_loss: 10.0
98
+ one_sample_validation: null
99
+ align_corners: false
100
+ intrinsic_scaling: false
101
+ verbose: false
102
+ beta_dist_param:
103
+ - 0.5
104
+ - 4.0
105
+ use_refine_aux: false
106
+ train_target_set: true
107
+ train_gs_num: 1
108
+ ext_scale_detach: false
109
+ cam_scale_mode: sum
110
+ scene_scale_reg_loss: 0.01
111
+ train_aux: true
112
+ vggt_cam_loss: true
113
+ vggt_distil: true
114
+ context_view_train: true
115
+ test:
116
+ output_path: test/ablation/re10k
117
+ align_pose: false
118
+ pose_align_steps: 100
119
+ rot_opt_lr: 0.005
120
+ trans_opt_lr: 0.005
121
+ compute_scores: true
122
+ save_image: false
123
+ save_video: false
124
+ save_active_mask_image: false
125
+ save_error_score_image: false
126
+ save_compare: false
127
+ pred_intrinsic: false
128
+ error_threshold: 0.4
129
+ error_threshold_list:
130
+ - 0.2
131
+ - 0.4
132
+ - 0.6
133
+ - 0.8
134
+ - 1.0
135
+ threshold_mode: ratio
136
+ nvs_view_N_list:
137
+ - 3
138
+ - 6
139
+ - 16
140
+ - 32
141
+ - 64
142
+ seed: 111123
143
+ trainer:
144
+ max_steps: 3001
145
+ val_check_interval: 250
146
+ gradient_clip_val: 0.5
147
+ num_nodes: 1
148
+ dataset:
149
+ re10k:
150
+ make_baseline_1: true
151
+ relative_pose: true
152
+ augment: true
153
+ background_color:
154
+ - 0.0
155
+ - 0.0
156
+ - 0.0
157
+ overfit_to_scene: null
158
+ skip_bad_shape: true
159
+ view_sampler:
160
+ name: bounded
161
+ num_target_views: 4
162
+ num_context_views: 2
163
+ min_distance_between_context_views: 45
164
+ max_distance_between_context_views: 90
165
+ min_distance_to_context_views: 0
166
+ warm_up_steps: 1000
167
+ initial_min_distance_between_context_views: 25
168
+ initial_max_distance_between_context_views: 25
169
+ same_target_gap: false
170
+ num_target_set: 3
171
+ name: re10k
172
+ roots:
173
+ - datasets/re10k
174
+ input_image_shape:
175
+ - 256
176
+ - 256
177
+ original_image_shape:
178
+ - 360
179
+ - 640
180
+ cameras_are_circular: false
181
+ baseline_min: 0.001
182
+ baseline_max: 10000000000.0
183
+ max_fov: 100.0
184
+ dynamic_context_views: true
185
+ max_context_views_per_gpu: 24
DEBUG/.hydra/hydra.yaml ADDED
@@ -0,0 +1,170 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/ablation/re10k/${wandb.name}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - +experiment=re10k_ablation_24v
116
+ - dataset.re10k.roots=[datasets/re10k]
117
+ - wandb.mode=disabled
118
+ - wandb.name=DEBUG
119
+ - model.encoder.head_mode=depth
120
+ - model.density_control.use_depth=true
121
+ - train.context_view_train=true
122
+ - dataset.re10k.dynamic_context_views=true
123
+ - train.vggt_distil=true
124
+ job:
125
+ name: main
126
+ chdir: null
127
+ override_dirname: +experiment=re10k_ablation_24v,dataset.re10k.dynamic_context_views=true,dataset.re10k.roots=[datasets/re10k],model.density_control.use_depth=true,model.encoder.head_mode=depth,train.context_view_train=true,train.vggt_distil=true,wandb.mode=disabled,wandb.name=DEBUG
128
+ id: ???
129
+ num: ???
130
+ config_name: main
131
+ env_set: {}
132
+ env_copy: []
133
+ config:
134
+ override_dirname:
135
+ kv_sep: '='
136
+ item_sep: ','
137
+ exclude_keys: []
138
+ runtime:
139
+ version: 1.3.2
140
+ version_base: '1.3'
141
+ cwd: /workspace/code/CVPR2026
142
+ config_sources:
143
+ - path: hydra.conf
144
+ schema: pkg
145
+ provider: hydra
146
+ - path: /workspace/code/CVPR2026/config
147
+ schema: file
148
+ provider: main
149
+ - path: ''
150
+ schema: structured
151
+ provider: schema
152
+ output_dir: /workspace/code/CVPR2026/outputs/ablation/re10k/DEBUG
153
+ choices:
154
+ experiment: re10k_ablation_24v
155
+ dataset@dataset.re10k: re10k
156
+ dataset/view_sampler_dataset_specific_config@dataset.re10k.view_sampler: bounded_re10k
157
+ dataset/view_sampler@dataset.re10k.view_sampler: bounded
158
+ model/density_control: density_control_module
159
+ model/decoder: splatting_cuda
160
+ model/encoder: dcsplat
161
+ hydra/env: default
162
+ hydra/callbacks: null
163
+ hydra/job_logging: default
164
+ hydra/hydra_logging: default
165
+ hydra/hydra_help: default
166
+ hydra/help: default
167
+ hydra/sweeper: basic
168
+ hydra/launcher: basic
169
+ hydra/output: default
170
+ verbose: false
DEBUG/.hydra/overrides.yaml ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ - +experiment=re10k_ablation_24v
2
+ - dataset.re10k.roots=[datasets/re10k]
3
+ - wandb.mode=disabled
4
+ - wandb.name=DEBUG
5
+ - model.encoder.head_mode=depth
6
+ - model.density_control.use_depth=true
7
+ - train.context_view_train=true
8
+ - dataset.re10k.dynamic_context_views=true
9
+ - train.vggt_distil=true
DEBUG/main.log ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-25 14:26:18,200][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-02-25 14:26:24,465][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-02-25 14:26:24,465][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2026-02-25 14:26:25,845][dinov2][INFO] - using MLP layer as FFN
9
+ [2026-02-25 14:26:35,621][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=255` in the `DataLoader` to improve performance.
10
+
11
+ [2026-02-25 14:26:56,525][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
12
+ result[selector] = overlay
13
+
14
+ [2026-02-25 14:26:56,541][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
15
+
16
+ [2026-02-25 14:26:56,542][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
17
+ warnings.warn(
18
+
19
+ [2026-02-25 14:26:56,542][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
20
+ warnings.warn(msg)
21
+
22
+ [2026-02-25 14:26:58,171][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4322.)
23
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
24
+