Knowing commited on
Commit
57b6191
·
verified ·
1 Parent(s): a9bf74b

Add files using upload-large-folder tool

Browse files
ABLATION_0302_LapFreqSelect/main.log ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ [2026-03-04 01:26:41,277][dinov2][INFO] - using MLP layer as FFN
2
+ [2026-03-04 01:26:47,317][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2026-03-04 01:26:47,317][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
ABLATION_0302_LapFreqSelect/train_ddp_process_1.log ADDED
@@ -0,0 +1 @@
 
 
1
+ [2026-03-04 01:26:54,799][dinov2][INFO] - using MLP layer as FFN
ABLATION_0302_LapFreqSelect/train_ddp_process_2.log ADDED
@@ -0,0 +1 @@
 
 
1
+ [2026-03-04 01:26:54,812][dinov2][INFO] - using MLP layer as FFN
ABLATION_0302_LapFreqSelect/train_ddp_process_3.log ADDED
@@ -0,0 +1 @@
 
 
1
+ [2026-03-04 01:26:54,820][dinov2][INFO] - using MLP layer as FFN
ABLATION_0302_LapFreqSelect/train_ddp_process_4.log ADDED
@@ -0,0 +1 @@
 
 
1
+ [2026-03-04 01:26:54,913][dinov2][INFO] - using MLP layer as FFN
ABLATION_0302_LapFreqSelect/train_ddp_process_6.log ADDED
@@ -0,0 +1 @@
 
 
1
+ [2026-03-04 01:26:54,939][dinov2][INFO] - using MLP layer as FFN
ABLATION_0302_LapFreqSelect/train_ddp_process_7.log ADDED
@@ -0,0 +1 @@
 
 
1
+ [2026-03-04 01:26:54,765][dinov2][INFO] - using MLP layer as FFN
ABLATION_0302_noTgtAlign/.hydra/config.yaml CHANGED
@@ -28,7 +28,7 @@ model:
28
  refinement_hidden_dim: 32
29
  aggregation_mode: mean
30
  num_heads: 1
31
- score_mode: random
32
  latent_dim: 128
33
  num_latents: 64
34
  num_self_attn_per_block: 2
 
28
  refinement_hidden_dim: 32
29
  aggregation_mode: mean
30
  num_heads: 1
31
+ score_mode: absgrad
32
  latent_dim: 128
33
  num_latents: 64
34
  num_self_attn_per_block: 2
ABLATION_0302_noTgtAlign/.hydra/hydra.yaml CHANGED
@@ -115,11 +115,10 @@ hydra:
115
  - +experiment=re10k_ablation_24v
116
  - wandb.mode=online
117
  - wandb.name=ABLATION_0302_noTgtAlign
118
- - model.density_control.score_mode=random
119
  job:
120
  name: main
121
  chdir: null
122
- override_dirname: +experiment=re10k_ablation_24v,model.density_control.score_mode=random,wandb.mode=online,wandb.name=ABLATION_0302_noTgtAlign
123
  id: ???
124
  num: ???
125
  config_name: main
 
115
  - +experiment=re10k_ablation_24v
116
  - wandb.mode=online
117
  - wandb.name=ABLATION_0302_noTgtAlign
 
118
  job:
119
  name: main
120
  chdir: null
121
+ override_dirname: +experiment=re10k_ablation_24v,wandb.mode=online,wandb.name=ABLATION_0302_noTgtAlign
122
  id: ???
123
  num: ???
124
  config_name: main
ABLATION_0302_noTgtAlign/.hydra/overrides.yaml CHANGED
@@ -1,4 +1,3 @@
1
  - +experiment=re10k_ablation_24v
2
  - wandb.mode=online
3
  - wandb.name=ABLATION_0302_noTgtAlign
4
- - model.density_control.score_mode=random
 
1
  - +experiment=re10k_ablation_24v
2
  - wandb.mode=online
3
  - wandb.name=ABLATION_0302_noTgtAlign
 
ABLATION_0302_noTgtAlign/peak_vram_memory.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "peak_memory_allocated_gb": 95.862,
3
+ "peak_memory_reserved_gb": 110.58,
4
+ "total_elapsed_hours": 3.11,
5
+ "mode": "train"
6
+ }
ABLATION_0302_noTgtAlign/train_ddp_process_4.log CHANGED
@@ -42,3 +42,78 @@ bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered int
42
  [2026-03-03 17:48:34,775][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
43
  result[selector] = overlay
44
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
42
  [2026-03-03 17:48:34,775][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
43
  result[selector] = overlay
44
 
45
+ [2026-03-03 18:01:02,021][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
46
+ result[selector] = overlay
47
+
48
+ [2026-03-03 18:13:25,680][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
49
+ result[selector] = overlay
50
+
51
+ [2026-03-03 18:25:47,338][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
52
+ result[selector] = overlay
53
+
54
+ [2026-03-03 18:31:57,697][dinov2][INFO] - using MLP layer as FFN
55
+ [2026-03-03 18:32:27,497][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
56
+ warnings.warn(
57
+
58
+ [2026-03-03 18:32:27,501][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
59
+ warnings.warn(msg)
60
+
61
+ [2026-03-03 18:32:42,134][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
62
+ warnings.warn( # warn only once
63
+
64
+ [2026-03-03 18:33:01,313][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
65
+ grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
66
+ bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
67
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
68
+
69
+ [2026-03-03 18:33:01,426][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
70
+ result[selector] = overlay
71
+
72
+ [2026-03-03 18:34:34,275][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
73
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
74
+
75
+ [2026-03-03 18:45:20,241][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
76
+ result[selector] = overlay
77
+
78
+ [2026-03-03 18:57:44,444][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
79
+ result[selector] = overlay
80
+
81
+ [2026-03-03 19:10:06,141][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
82
+ result[selector] = overlay
83
+
84
+ [2026-03-03 19:22:26,646][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
85
+ result[selector] = overlay
86
+
87
+ [2026-03-03 19:34:44,817][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
88
+ result[selector] = overlay
89
+
90
+ [2026-03-03 19:47:04,316][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
91
+ result[selector] = overlay
92
+
93
+ [2026-03-03 19:59:23,283][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
94
+ result[selector] = overlay
95
+
96
+ [2026-03-03 20:11:55,578][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
97
+ result[selector] = overlay
98
+
99
+ [2026-03-03 20:24:20,974][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
100
+ result[selector] = overlay
101
+
102
+ [2026-03-03 20:36:39,577][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
103
+ result[selector] = overlay
104
+
105
+ [2026-03-03 20:49:01,141][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
106
+ result[selector] = overlay
107
+
108
+ [2026-03-03 21:01:27,559][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
109
+ result[selector] = overlay
110
+
111
+ [2026-03-03 21:13:42,613][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
112
+ result[selector] = overlay
113
+
114
+ [2026-03-03 21:26:01,866][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
115
+ result[selector] = overlay
116
+
117
+ [2026-03-03 21:38:35,396][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
118
+ result[selector] = overlay
119
+
ABLATION_0302_noTgtAlign/wandb/debug-internal.log CHANGED
@@ -1,6 +1,11 @@
1
- {"time":"2026-03-03T17:35:55.708502915Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
2
- {"time":"2026-03-03T17:35:56.156384015Z","level":"INFO","msg":"stream: created new stream","id":"et94mfhp"}
3
- {"time":"2026-03-03T17:35:56.156522256Z","level":"INFO","msg":"handler: started","stream_id":"et94mfhp"}
4
- {"time":"2026-03-03T17:35:56.156785449Z","level":"INFO","msg":"stream: started","id":"et94mfhp"}
5
- {"time":"2026-03-03T17:35:56.156813659Z","level":"INFO","msg":"sender: started","stream_id":"et94mfhp"}
6
- {"time":"2026-03-03T17:35:56.156835559Z","level":"INFO","msg":"writer: started","stream_id":"et94mfhp"}
 
 
 
 
 
 
1
+ {"time":"2026-03-03T18:32:36.416950124Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
2
+ {"time":"2026-03-03T18:32:36.896736702Z","level":"INFO","msg":"stream: created new stream","id":"bjvibjti"}
3
+ {"time":"2026-03-03T18:32:36.896847584Z","level":"INFO","msg":"handler: started","stream_id":"bjvibjti"}
4
+ {"time":"2026-03-03T18:32:36.897054806Z","level":"INFO","msg":"stream: started","id":"bjvibjti"}
5
+ {"time":"2026-03-03T18:32:36.897086516Z","level":"INFO","msg":"sender: started","stream_id":"bjvibjti"}
6
+ {"time":"2026-03-03T18:32:36.897099556Z","level":"INFO","msg":"writer: started","stream_id":"bjvibjti"}
7
+ {"time":"2026-03-03T21:38:44.775871388Z","level":"INFO","msg":"stream: closing","id":"bjvibjti"}
8
+ {"time":"2026-03-03T21:38:45.836130953Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
9
+ {"time":"2026-03-03T21:38:46.109942009Z","level":"INFO","msg":"handler: closed","stream_id":"bjvibjti"}
10
+ {"time":"2026-03-03T21:38:46.1101932Z","level":"INFO","msg":"sender: closed","stream_id":"bjvibjti"}
11
+ {"time":"2026-03-03T21:38:46.110223611Z","level":"INFO","msg":"stream: closed","id":"bjvibjti"}
ABLATION_0302_noTgtAlign/wandb/debug.log CHANGED
@@ -1,19 +1,21 @@
1
- 2026-03-03 17:35:55,409 INFO MainThread:870906 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
2
- 2026-03-03 17:35:55,409 INFO MainThread:870906 [wandb_setup.py:_flush():81] Configure stats pid to 870906
3
- 2026-03-03 17:35:55,409 INFO MainThread:870906 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
- 2026-03-03 17:35:55,409 INFO MainThread:870906 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign/wandb/run-20260303_173555-et94mfhp/logs/debug.log
5
- 2026-03-03 17:35:55,409 INFO MainThread:870906 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign/wandb/run-20260303_173555-et94mfhp/logs/debug-internal.log
6
- 2026-03-03 17:35:55,410 INFO MainThread:870906 [wandb_init.py:init():844] calling init triggers
7
- 2026-03-03 17:35:55,410 INFO MainThread:870906 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
- config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'pcd', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'random', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': False, 'voxelize_activate': False, 'use_depth': False}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'density_control_loss': {'error_score': {'weight': 0.0001, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0302_noTgtAlign', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': True, 'vggt_cam_loss': True, 'vggt_distil': False, 'context_view_train': False}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'save_gs': False, 'save_sample_wise_metrics': True, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3, 'target_align': True}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
9
- 2026-03-03 17:35:55,410 INFO MainThread:870906 [wandb_init.py:init():892] starting backend
10
- 2026-03-03 17:35:55,698 INFO MainThread:870906 [wandb_init.py:init():895] sending inform_init request
11
- 2026-03-03 17:35:55,705 INFO MainThread:870906 [wandb_init.py:init():903] backend started and connected
12
- 2026-03-03 17:35:55,709 INFO MainThread:870906 [wandb_init.py:init():973] updated telemetry
13
- 2026-03-03 17:35:55,716 INFO MainThread:870906 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
- 2026-03-03 17:35:56,664 INFO MainThread:870906 [wandb_init.py:init():1042] starting run threads in backend
15
- 2026-03-03 17:35:56,791 INFO MainThread:870906 [wandb_run.py:_console_start():2524] atexit reg
16
- 2026-03-03 17:35:56,791 INFO MainThread:870906 [wandb_run.py:_redirect():2373] redirect: wrap_raw
17
- 2026-03-03 17:35:56,791 INFO MainThread:870906 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
- 2026-03-03 17:35:56,791 INFO MainThread:870906 [wandb_run.py:_redirect():2465] Redirects installed.
19
- 2026-03-03 17:35:56,794 INFO MainThread:870906 [wandb_init.py:init():1082] run started, returning control to user process
 
 
 
1
+ 2026-03-03 18:32:36,159 INFO MainThread:880223 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
2
+ 2026-03-03 18:32:36,159 INFO MainThread:880223 [wandb_setup.py:_flush():81] Configure stats pid to 880223
3
+ 2026-03-03 18:32:36,159 INFO MainThread:880223 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
+ 2026-03-03 18:32:36,159 INFO MainThread:880223 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign/wandb/run-20260303_183236-bjvibjti/logs/debug.log
5
+ 2026-03-03 18:32:36,159 INFO MainThread:880223 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign/wandb/run-20260303_183236-bjvibjti/logs/debug-internal.log
6
+ 2026-03-03 18:32:36,159 INFO MainThread:880223 [wandb_init.py:init():844] calling init triggers
7
+ 2026-03-03 18:32:36,159 INFO MainThread:880223 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
+ config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'pcd', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'absgrad', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': False, 'voxelize_activate': False, 'use_depth': False}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'density_control_loss': {'error_score': {'weight': 0.0001, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0302_noTgtAlign', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': True, 'vggt_cam_loss': True, 'vggt_distil': False, 'context_view_train': False}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'save_gs': False, 'save_sample_wise_metrics': True, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3, 'target_align': True}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
9
+ 2026-03-03 18:32:36,159 INFO MainThread:880223 [wandb_init.py:init():892] starting backend
10
+ 2026-03-03 18:32:36,410 INFO MainThread:880223 [wandb_init.py:init():895] sending inform_init request
11
+ 2026-03-03 18:32:36,415 INFO MainThread:880223 [wandb_init.py:init():903] backend started and connected
12
+ 2026-03-03 18:32:36,419 INFO MainThread:880223 [wandb_init.py:init():973] updated telemetry
13
+ 2026-03-03 18:32:36,426 INFO MainThread:880223 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
+ 2026-03-03 18:32:37,451 INFO MainThread:880223 [wandb_init.py:init():1042] starting run threads in backend
15
+ 2026-03-03 18:32:37,532 INFO MainThread:880223 [wandb_run.py:_console_start():2524] atexit reg
16
+ 2026-03-03 18:32:37,533 INFO MainThread:880223 [wandb_run.py:_redirect():2373] redirect: wrap_raw
17
+ 2026-03-03 18:32:37,533 INFO MainThread:880223 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
+ 2026-03-03 18:32:37,533 INFO MainThread:880223 [wandb_run.py:_redirect():2465] Redirects installed.
19
+ 2026-03-03 18:32:37,536 INFO MainThread:880223 [wandb_init.py:init():1082] run started, returning control to user process
20
+ 2026-03-03 21:38:44,775 INFO wandb-AsyncioManager-main:880223 [service_client.py:_forward_responses():134] Reached EOF.
21
+ 2026-03-03 21:38:44,776 INFO wandb-AsyncioManager-main:880223 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.
ABLATION_0302_noTgtAlign/wandb/run-20260303_173555-et94mfhp/files/config.yaml ADDED
@@ -0,0 +1,313 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.25.0
4
+ e:
5
+ 0fg4ri1obzq9oqbw2tt47dd2pwoxs7dt:
6
+ args:
7
+ - +experiment=re10k_ablation_24v
8
+ - wandb.mode=online
9
+ - wandb.name=ABLATION_0302_noTgtAlign
10
+ - model.density_control.score_mode=random
11
+ cpu_count: 128
12
+ cpu_count_logical: 256
13
+ cudaVersion: "13.0"
14
+ disk:
15
+ /:
16
+ total: "735513149440"
17
+ used: "700808368128"
18
+ email: dna9041@korea.ac.kr
19
+ executable: /venv/main/bin/python
20
+ git:
21
+ commit: 9dfce172a0f8c7ce85e763899f7ef741ecffc454
22
+ remote: git@github.com:K-nowing/CVPR2026.git
23
+ gpu: NVIDIA H200
24
+ gpu_count: 8
25
+ gpu_nvidia:
26
+ - architecture: Hopper
27
+ cudaCores: 16896
28
+ memoryTotal: "150754820096"
29
+ name: NVIDIA H200
30
+ uuid: GPU-9a20101e-d876-facd-5f05-805081aede41
31
+ - architecture: Hopper
32
+ cudaCores: 16896
33
+ memoryTotal: "150754820096"
34
+ name: NVIDIA H200
35
+ uuid: GPU-84736a77-ee75-3324-e4e1-99cc15bfb5e9
36
+ - architecture: Hopper
37
+ cudaCores: 16896
38
+ memoryTotal: "150754820096"
39
+ name: NVIDIA H200
40
+ uuid: GPU-423d3161-cdc4-3fc0-caee-d15cfaa83ca6
41
+ - architecture: Hopper
42
+ cudaCores: 16896
43
+ memoryTotal: "150754820096"
44
+ name: NVIDIA H200
45
+ uuid: GPU-5b0058b2-cdb9-c952-04f9-87dcaa7ea742
46
+ - architecture: Hopper
47
+ cudaCores: 16896
48
+ memoryTotal: "150754820096"
49
+ name: NVIDIA H200
50
+ uuid: GPU-08b37f98-4603-d483-2f2b-fe5311aa42f2
51
+ - architecture: Hopper
52
+ cudaCores: 16896
53
+ memoryTotal: "150754820096"
54
+ name: NVIDIA H200
55
+ uuid: GPU-03273b5b-2fdd-a5fe-4460-c897334ae464
56
+ - architecture: Hopper
57
+ cudaCores: 16896
58
+ memoryTotal: "150754820096"
59
+ name: NVIDIA H200
60
+ uuid: GPU-292d466c-d00d-25a4-28b6-e6c978d3e70c
61
+ - architecture: Hopper
62
+ cudaCores: 16896
63
+ memoryTotal: "150754820096"
64
+ name: NVIDIA H200
65
+ uuid: GPU-46f38561-3148-e442-7f7f-bfe447bab7fe
66
+ host: e9d3310a05da
67
+ memory:
68
+ total: "1622950240256"
69
+ os: Linux-6.8.0-94-generic-x86_64-with-glibc2.39
70
+ program: -m src.main
71
+ python: CPython 3.12.12
72
+ root: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign
73
+ startedAt: "2026-03-03T17:35:55.406237Z"
74
+ writerId: 0fg4ri1obzq9oqbw2tt47dd2pwoxs7dt
75
+ m:
76
+ - "1": trainer/global_step
77
+ "6":
78
+ - 3
79
+ "7": []
80
+ - "2": '*'
81
+ "5": 1
82
+ "6":
83
+ - 1
84
+ "7": []
85
+ python_version: 3.12.12
86
+ t:
87
+ "1":
88
+ - 1
89
+ - 41
90
+ - 49
91
+ - 50
92
+ - 106
93
+ "2":
94
+ - 1
95
+ - 41
96
+ - 49
97
+ - 50
98
+ - 106
99
+ "3":
100
+ - 2
101
+ - 7
102
+ - 13
103
+ - 15
104
+ - 16
105
+ - 33
106
+ - 41
107
+ - 66
108
+ "4": 3.12.12
109
+ "5": 0.25.0
110
+ "12": 0.25.0
111
+ "13": linux-x86_64
112
+ checkpointing:
113
+ value:
114
+ every_n_train_steps: 1500
115
+ load: null
116
+ save_top_k: 2
117
+ save_weights_only: false
118
+ data_loader:
119
+ value:
120
+ test:
121
+ batch_size: 1
122
+ num_workers: 4
123
+ persistent_workers: false
124
+ seed: 2345
125
+ train:
126
+ batch_size: 16
127
+ num_workers: 16
128
+ persistent_workers: true
129
+ seed: 1234
130
+ val:
131
+ batch_size: 1
132
+ num_workers: 1
133
+ persistent_workers: true
134
+ seed: 3456
135
+ dataset:
136
+ value:
137
+ re10k:
138
+ augment: true
139
+ background_color:
140
+ - 0
141
+ - 0
142
+ - 0
143
+ baseline_max: 1e+10
144
+ baseline_min: 0.001
145
+ cameras_are_circular: false
146
+ dynamic_context_views: true
147
+ input_image_shape:
148
+ - 256
149
+ - 256
150
+ make_baseline_1: true
151
+ max_context_views_per_gpu: 24
152
+ max_fov: 100
153
+ name: re10k
154
+ original_image_shape:
155
+ - 360
156
+ - 640
157
+ overfit_to_scene: null
158
+ relative_pose: true
159
+ roots:
160
+ - datasets/re10k
161
+ skip_bad_shape: true
162
+ view_sampler:
163
+ initial_max_distance_between_context_views: 25
164
+ initial_min_distance_between_context_views: 25
165
+ max_distance_between_context_views: 90
166
+ min_distance_between_context_views: 45
167
+ min_distance_to_context_views: 0
168
+ name: bounded
169
+ num_context_views: 2
170
+ num_target_set: 3
171
+ num_target_views: 4
172
+ same_target_gap: false
173
+ target_align: true
174
+ warm_up_steps: 1000
175
+ density_control_loss:
176
+ value:
177
+ error_score:
178
+ grad_scale: 10000
179
+ log_scale: false
180
+ mode: original
181
+ weight: 0.0001
182
+ direct_loss:
183
+ value:
184
+ l1:
185
+ weight: 0.8
186
+ ssim:
187
+ weight: 0.2
188
+ mode:
189
+ value: train
190
+ model:
191
+ value:
192
+ decoder:
193
+ background_color:
194
+ - 0
195
+ - 0
196
+ - 0
197
+ make_scale_invariant: false
198
+ name: splatting_cuda
199
+ density_control:
200
+ aggregation_mode: mean
201
+ aux_refine: false
202
+ grad_mode: absgrad
203
+ gs_param_dim: 256
204
+ latent_dim: 128
205
+ mean_dim: 32
206
+ name: density_control_module
207
+ num_heads: 1
208
+ num_latents: 64
209
+ num_level: 3
210
+ num_self_attn_per_block: 2
211
+ refine_error: false
212
+ refinement_hidden_dim: 32
213
+ refinement_layer_num: 1
214
+ refinement_type: voxelize
215
+ score_mode: random
216
+ use_depth: false
217
+ use_mean_features: true
218
+ use_refine_module: false
219
+ voxel_size: 0.001
220
+ voxelize_activate: false
221
+ encoder:
222
+ align_corners: false
223
+ gs_param_dim: 256
224
+ head_mode: pcd
225
+ input_image_shape:
226
+ - 518
227
+ - 518
228
+ name: dcsplat
229
+ num_level: 3
230
+ use_voxelize: true
231
+ optimizer:
232
+ value:
233
+ accumulate: 1
234
+ backbone_lr_multiplier: 0.1
235
+ backbone_trainable: T+H
236
+ lr: 0.0002
237
+ warm_up_steps: 25
238
+ render_loss:
239
+ value:
240
+ lpips:
241
+ apply_after_step: 0
242
+ weight: 0.05
243
+ mse:
244
+ weight: 1
245
+ seed:
246
+ value: 111123
247
+ test:
248
+ value:
249
+ align_pose: false
250
+ compute_scores: true
251
+ error_threshold: 0.4
252
+ error_threshold_list:
253
+ - 0.2
254
+ - 0.4
255
+ - 0.6
256
+ - 0.8
257
+ - 1
258
+ nvs_view_N_list:
259
+ - 3
260
+ - 6
261
+ - 16
262
+ - 32
263
+ - 64
264
+ output_path: test/ablation/re10k
265
+ pose_align_steps: 100
266
+ pred_intrinsic: false
267
+ rot_opt_lr: 0.005
268
+ save_active_mask_image: false
269
+ save_compare: false
270
+ save_error_score_image: false
271
+ save_gs: false
272
+ save_image: false
273
+ save_sample_wise_metrics: true
274
+ save_video: false
275
+ threshold_mode: ratio
276
+ trans_opt_lr: 0.005
277
+ train:
278
+ value:
279
+ align_corners: false
280
+ beta_dist_param:
281
+ - 0.5
282
+ - 4
283
+ cam_scale_mode: sum
284
+ camera_loss: 10
285
+ context_view_train: false
286
+ ext_scale_detach: false
287
+ extended_visualization: false
288
+ intrinsic_scaling: false
289
+ one_sample_validation: null
290
+ print_log_every_n_steps: 10
291
+ scene_scale_reg_loss: 0.01
292
+ train_aux: true
293
+ train_gs_num: 1
294
+ train_target_set: true
295
+ use_refine_aux: false
296
+ verbose: false
297
+ vggt_cam_loss: true
298
+ vggt_distil: false
299
+ trainer:
300
+ value:
301
+ gradient_clip_val: 0.5
302
+ max_steps: 3001
303
+ num_nodes: 1
304
+ val_check_interval: 250
305
+ wandb:
306
+ value:
307
+ entity: scene-representation-group
308
+ mode: online
309
+ name: ABLATION_0302_noTgtAlign
310
+ project: DCSplat
311
+ tags:
312
+ - re10k
313
+ - 256x256
ABLATION_0302_noTgtAlign/wandb/run-20260303_173555-et94mfhp/files/output.log ADDED
@@ -0,0 +1,278 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7]
2
+
3
+ | Name | Type | Params | Mode
4
+ ------------------------------------------------------------------------
5
+ 0 | encoder | OurSplat | 888 M | train
6
+ 1 | density_control_module | DensityControlModule | 0 | train
7
+ 2 | decoder | DecoderSplattingCUDA | 0 | train
8
+ 3 | render_losses | ModuleList | 0 | train
9
+ 4 | density_control_losses | ModuleList | 0 | train
10
+ 5 | direct_losses | ModuleList | 0 | train
11
+ ------------------------------------------------------------------------
12
+ 888 M Trainable params
13
+ 0 Non-trainable params
14
+ 888 M Total params
15
+ 3,553.933 Total estimated model params size (MB)
16
+ 1202 Modules in train mode
17
+ 522 Modules in eval mode
18
+ Sanity Checking: | | 0/? [00:00<?, ?it/s][2026-03-03 17:35:58,952][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=31` in the `DataLoader` to improve performance.
19
+
20
+ [2026-03-03 17:35:58,954][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
21
+ warnings.warn( # warn only once
22
+
23
+ Validation epoch start on rank 0
24
+ Sanity Checking DataLoader 0: 0%| | 0/1 [00:00<?, ?it/s]validation step 0; scene = ['306e2b7785657539'];
25
+ target intrinsic: tensor(0.8595, device='cuda:0') tensor(0.8597, device='cuda:0')
26
+ pred intrinsic: tensor(0.8779, device='cuda:0') tensor(0.8773, device='cuda:0')
27
+ [rank0]:W0303 17:36:01.277000 870906 site-packages/torch/utils/cpp_extension.py:2425] TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
28
+ [rank0]:W0303 17:36:01.277000 870906 site-packages/torch/utils/cpp_extension.py:2425] If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'] to specific architectures.
29
+ [2026-03-03 17:36:01,340][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
30
+ result[selector] = overlay
31
+
32
+ [2026-03-03 17:36:01,350][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
33
+
34
+ Setting up [LPIPS] perceptual loss: trunk [vgg], v[0.1], spatial [off]
35
+ [2026-03-03 17:36:01,350][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
36
+ warnings.warn(
37
+
38
+ [2026-03-03 17:36:01,351][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
39
+ warnings.warn(msg)
40
+
41
+ Loading model from: /venv/main/lib/python3.12/site-packages/lpips/weights/v0.1/vgg.pth
42
+ [2026-03-03 17:36:03,034][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4322.)
43
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
44
+
45
+ Sanity Checking DataLoader 0: 100%|████████████████████████████████████████████████████████████████████| 1/1 [00:03<00:00, 0.27it/s][2026-03-03 17:36:03,326][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
46
+
47
+ [2026-03-03 17:36:03,327][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
48
+
49
+ [2026-03-03 17:36:03,328][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
50
+
51
+ [2026-03-03 17:36:03,328][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
52
+
53
+ [2026-03-03 17:36:03,328][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
54
+
55
+ Epoch 0: | | 0/? [00:00<?, ?it/s]context = [[34, 36, 50, 53, 54, 60, 63, 70, 76, 78, 79, 80, 81, 88, 92, 94, 102, 110, 112, 114, 122, 125, 126, 131]]target = [[126, 96, 109, 55, 99, 116, 43, 60, 113, 85, 103, 90, 130, 62, 76, 123, 35, 102, 125, 128, 98, 67, 129, 79]]
56
+ [2026-03-03 17:36:13,075][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
57
+ grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
58
+ bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
59
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
60
+
61
+ [2026-03-03 17:36:13,149][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
62
+ result[selector] = overlay
63
+
64
+ Epoch 0: | | 9/? [00:41<00:00, 0.22it/s, v_num=mfhp]train step 10; scene = [['08c26703c4987851']]; loss = 0.814219
65
+ Epoch 0: | | 10/? [00:45<00:00, 0.22it/s, v_num=mfhp]context = [[98, 107, 112, 119, 122, 123], [21, 22, 27, 36, 38, 46], [63, 66, 67, 77, 84, 88], [56, 62, 65, 68, 78, 81]]target = [[105, 110, 116, 112, 111, 101], [29, 37, 36, 22, 40, 45], [73, 66, 75, 87, 83, 64], [79, 69, 75, 58, 61, 62]]
66
+ Epoch 0: | | 19/? [01:18<00:00, 0.24it/s, v_num=mfhp]train step 20; scene = [['4012c15c8381568b'], ['af08406c5a9a43a0'], ['9f9f9beffb86fad7'], ['fc8d08df6c875cb0']]; loss = 0.247813
67
+ Epoch 0: | | 20/? [01:21<00:00, 0.24it/s, v_num=mfhp]context = [[144, 152, 157, 164, 166, 169, 171, 177], [201, 211, 216, 221, 228, 230, 233, 234], [11, 15, 16, 23, 30, 38, 43, 44]]target = [[153, 170, 149, 169, 145, 174, 165, 157], [229, 216, 205, 206, 203, 213, 233, 215], [37, 38, 39, 35, 15, 24, 19, 25]]
68
+ Epoch 0: | | 24/? [01:35<00:00, 0.25it/s, v_num=mfhp][2026-03-03 17:37:46,051][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
69
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
70
+
71
+ Epoch 0: | | 29/? [01:53<00:00, 0.25it/s, v_num=mfhp]train step 30; scene = [['00980329a3221f1c'], ['1e7c432d2207b6f2'], ['af2748330e5243d0']]; loss = 0.187076
72
+ Epoch 0: | | 30/? [01:57<00:00, 0.26it/s, v_num=mfhp]context = [[2, 7, 10, 15, 20, 24, 25, 29, 30, 32, 33, 37, 40, 44, 49, 60, 64, 74, 76, 81, 89, 91, 95, 99]]target = [[79, 84, 35, 43, 89, 44, 63, 58, 48, 13, 65, 7, 96, 27, 51, 20, 60, 15, 85, 59, 22, 66, 17, 62]]
73
+ Epoch 0: | | 39/? [02:30<00:00, 0.26it/s, v_num=mfhp]train step 40; scene = [['79a9385753d426bc'], ['593538382d2dc847'], ['c9c67636b9d521be']]; loss = 0.155894
74
+ Epoch 0: | | 40/? [02:34<00:00, 0.26it/s, v_num=mfhp]context = [[200, 208, 223, 225], [6, 8, 12, 31], [104, 107, 111, 129], [69, 80, 86, 95], [11, 15, 18, 37], [52, 60, 70, 78]]target = [[213, 203, 214, 201], [29, 11, 15, 28], [111, 122, 120, 115], [77, 90, 89, 78], [18, 24, 29, 14], [64, 67, 63, 74]]
75
+ Epoch 0: | | 49/? [03:06<00:00, 0.26it/s, v_num=mfhp]train step 50; scene = [['579a11551b3315d9'], ['c9dd64b7415e788e'], ['6f3fb517d1798d03']]; loss = 0.144112
76
+ Epoch 0: | | 50/? [03:09<00:00, 0.26it/s, v_num=mfhp]context = [[20, 22, 25, 28, 36, 47], [10, 13, 15, 22, 26, 35], [31, 47, 49, 51, 52, 56], [109, 110, 115, 116, 125, 135]]target = [[40, 24, 23, 30, 37, 33], [19, 32, 25, 24, 26, 28], [35, 52, 50, 55, 45, 36], [131, 123, 130, 110, 115, 134]]
77
+ Epoch 0: | | 59/? [03:43<00:00, 0.26it/s, v_num=mfhp]train step 60; scene = [['07916b8004a8e336'], ['e51ef9945ae527c4'], ['db84f84b1d775bb8'], ['92ed61f8e16b7e67']]; loss = 0.142657
78
+ Epoch 0: | | 60/? [03:46<00:00, 0.26it/s, v_num=mfhp]context = [[96, 97, 101, 102, 105, 106, 107, 114, 124, 134, 138, 145], [76, 78, 82, 83, 96, 98, 100, 111, 112, 118, 119, 125]]target = [[144, 124, 137, 102, 107, 130, 119, 129, 118, 123, 133, 126], [98, 108, 96, 100, 113, 123, 77, 78, 112, 109, 80, 102]]
79
+ Epoch 0: | | 69/? [04:20<00:00, 0.26it/s, v_num=mfhp]train step 70; scene = [['c34efa1505a0cfaa'], ['a3d0cca9fb57fd85'], ['43d0e6dce7bb1e95'], ['d8c2f0a3734cb493']]; loss = 0.100209
80
+ Epoch 0: | | 70/? [04:23<00:00, 0.27it/s, v_num=mfhp]context = [[204, 208, 211, 218, 232, 234, 237, 239, 240, 241, 247, 253], [6, 7, 10, 12, 35, 36, 37, 38, 40, 45, 48, 55]]target = [[230, 236, 222, 227, 220, 252, 216, 221, 213, 217, 249, 240], [18, 24, 25, 12, 9, 42, 34, 52, 41, 33, 43, 48]]
81
+ Epoch 0: | | 79/? [04:56<00:00, 0.27it/s, v_num=mfhp]train step 80; scene = [['24d756c820744e31'], ['cd6c21656a85e9b9'], ['f3b24cf238154fc0']]; loss = 0.096154
82
+ Epoch 0: | | 80/? [05:00<00:00, 0.27it/s, v_num=mfhp]context = [[4, 30], [52, 79], [61, 87], [12, 40], [83, 109], [3, 29], [221, 249], [198, 227], [9, 38], [46, 72], [0, 26], [123, 150]]target = [[27, 26], [58, 57], [70, 77], [26, 20], [87, 98], [26, 14], [223, 224], [221, 200], [23, 33], [67, 64], [4, 25], [140, 148]]
83
+ Epoch 0: | | 89/? [05:34<00:00, 0.27it/s, v_num=mfhp]train step 90; scene = [['617b4bc98d7e0bb6'], ['666e4a9aba27bb64']]; loss = 0.099716
84
+ Epoch 0: | | 90/? [05:37<00:00, 0.27it/s, v_num=mfhp]context = [[134, 135, 138, 146, 147, 151, 156, 161, 162, 164, 166, 167, 168, 169, 187, 189, 197, 210, 215, 224, 225, 228, 230, 231]]target = [[143, 149, 191, 182, 151, 226, 165, 140, 208, 171, 179, 223, 168, 136, 194, 207, 227, 144, 187, 185, 145, 218, 139, 170]]
85
+ Epoch 0: | | 99/? [06:11<00:00, 0.27it/s, v_num=mfhp]train step 100; scene = [['12fee7f1978d52f1'], ['c963bb60939e2d81']]; loss = 0.108745
86
+ Epoch 0: | | 100/? [06:15<00:00, 0.27it/s, v_num=mfhp]context = [[40, 44, 47, 48, 49, 50, 56, 63, 77, 79, 88, 89], [21, 33, 34, 35, 36, 40, 43, 46, 47, 60, 62, 70]]target = [[58, 81, 76, 64, 68, 72, 51, 87, 77, 65, 88, 45], [31, 47, 25, 35, 55, 22, 48, 65, 29, 40, 63, 67]]
87
+ Epoch 0: | | 109/? [06:48<00:00, 0.27it/s, v_num=mfhp]train step 110; scene = [['47396d5a5299873e']]; loss = 0.127887
88
+ Epoch 0: | | 110/? [06:52<00:00, 0.27it/s, v_num=mfhp]context = [[20, 22, 29, 30, 31, 34, 37, 45, 47, 51, 56, 69], [18, 22, 24, 27, 31, 37, 44, 48, 49, 60, 61, 67]]target = [[49, 22, 36, 59, 63, 60, 45, 66, 38, 28, 26, 64], [61, 65, 25, 20, 63, 60, 26, 22, 33, 45, 37, 35]]
89
+ Epoch 0: | | 119/? [07:25<00:00, 0.27it/s, v_num=mfhp]train step 120; scene = [['9bd7044e7cbf8e60'], ['76e44cf6b5658b26']]; loss = 0.085511
90
+ Epoch 0: | | 120/? [07:29<00:00, 0.27it/s, v_num=mfhp]context = [[7, 14, 26, 28, 31, 34, 37, 40], [10, 20, 26, 32, 35, 39, 41, 43], [17, 23, 26, 28, 44, 46, 49, 50]]target = [[14, 26, 10, 22, 30, 13, 31, 11], [24, 31, 36, 14, 40, 34, 41, 13], [24, 44, 34, 48, 46, 36, 26, 38]]
91
+ Epoch 0: | | 129/? [08:02<00:00, 0.27it/s, v_num=mfhp]train step 130; scene = [['a8cef6a851fbea3c'], ['b6699f4d039a5b06'], ['55cf2bbe9e017ea4'], ['6b0dd861e1ab1fec'], ['14db202c335af709'], ['8b6ff6c5153a7794'], ['b75f3820760d835c'], ['f7dbc855fd2a7669'], ['cfb20f8971e6a591'], ['95f2be7bb8303f50'], ['ff422469e034ae11'], ['5a2ad43377e9d18d']]; loss = 0.110932
92
+ Epoch 0: | | 130/? [08:06<00:00, 0.27it/s, v_num=mfhp]context = [[8, 15, 26, 30, 32, 35, 40, 46, 47, 48, 50, 53, 54, 62, 69, 70, 72, 76, 80, 86, 90, 99, 100, 105]]target = [[26, 12, 49, 100, 89, 24, 10, 81, 37, 63, 52, 17, 39, 70, 16, 56, 40, 55, 43, 34, 72, 28, 48, 45]]
93
+ Epoch 0: | | 139/? [08:38<00:00, 0.27it/s, v_num=mfhp]train step 140; scene = [['f62a962df5c26a1a'], ['b076420679a04731']]; loss = 0.078446
94
+ Epoch 0: | | 140/? [08:42<00:00, 0.27it/s, v_num=mfhp]context = [[90, 95, 97, 104, 118, 121], [14, 29, 34, 35, 36, 44], [12, 14, 22, 27, 34, 41], [6, 10, 16, 18, 34, 35]]target = [[99, 104, 93, 113, 110, 96], [29, 37, 22, 21, 39, 41], [19, 25, 39, 27, 29, 13], [26, 28, 9, 7, 13, 34]]
95
+ Epoch 0: | | 149/? [09:16<00:00, 0.27it/s, v_num=mfhp]train step 150; scene = [['a52d26a78b04aebd']]; loss = 0.071345
96
+ Epoch 0: | | 150/? [09:19<00:00, 0.27it/s, v_num=mfhp]context = [[115, 132, 134, 145], [16, 32, 39, 44], [57, 63, 78, 88], [9, 11, 23, 36], [15, 26, 28, 44], [1, 8, 22, 33]]target = [[124, 132, 116, 119], [27, 24, 33, 29], [81, 61, 85, 79], [28, 26, 29, 16], [28, 29, 37, 22], [9, 4, 32, 27]]
97
+ Epoch 0: | | 159/? [09:52<00:00, 0.27it/s, v_num=mfhp]train step 160; scene = [['268fbffc6c479d5b']]; loss = 0.069589
98
+ Epoch 0: | | 160/? [09:56<00:00, 0.27it/s, v_num=mfhp]context = [[18, 25, 26, 37, 42, 46, 49, 51, 53, 64, 65, 67], [69, 75, 78, 79, 82, 84, 94, 95, 104, 108, 117, 118]]target = [[53, 27, 22, 32, 41, 38, 50, 43, 47, 48, 23, 19], [74, 70, 114, 115, 90, 89, 88, 92, 94, 110, 107, 101]]
99
+ Epoch 0: | | 169/? [10:30<00:00, 0.27it/s, v_num=mfhp]train step 170; scene = [['719e2e8912e4eed3'], ['a3e51565a737569f']]; loss = 0.156401
100
+ Epoch 0: | | 170/? [10:34<00:00, 0.27it/s, v_num=mfhp]context = [[14, 18, 20, 21, 24, 27, 33, 40, 45, 47, 48, 51, 52, 60, 64, 70, 75, 77, 80, 85, 90, 98, 102, 111]]target = [[29, 93, 32, 39, 81, 108, 72, 107, 51, 35, 16, 36, 70, 18, 34, 92, 94, 47, 23, 74, 50, 77, 19, 37]]
101
+ Epoch 0: | | 179/? [11:07<00:00, 0.27it/s, v_num=mfhp]train step 180; scene = [['f44b9aa76a94a0a3']]; loss = 0.112078
102
+ Epoch 0: | | 180/? [11:10<00:00, 0.27it/s, v_num=mfhp]context = [[0, 6, 17, 22, 26, 28, 33, 41, 50, 55, 56, 57, 71, 76, 79, 81, 84, 85, 86, 87, 89, 95, 96, 97]]target = [[37, 49, 12, 78, 9, 16, 84, 13, 5, 4, 6, 38, 80, 51, 43, 68, 64, 46, 56, 24, 25, 72, 36, 21]]
103
+ Epoch 0: | | 189/? [11:44<00:00, 0.27it/s, v_num=mfhp]train step 190; scene = [['71bb669d936a5718'], ['a47203cfd5e0a478'], ['4b009f82cf5c7098']]; loss = 0.108068
104
+ Epoch 0: | | 190/? [11:48<00:00, 0.27it/s, v_num=mfhp]context = [[9, 11, 19, 25, 33, 43, 46, 47, 48, 54, 57, 58, 63, 70, 72, 75, 79, 80, 83, 84, 85, 96, 99, 106]]target = [[30, 53, 82, 31, 55, 12, 20, 72, 104, 70, 24, 52, 21, 32, 102, 71, 35, 11, 46, 10, 15, 74, 33, 26]]
105
+ Epoch 0: | | 199/? [12:20<00:00, 0.27it/s, v_num=mfhp]train step 200; scene = [['dd5ec950a01c42a0'], ['6d0db0358f7e051e'], ['983fe650a925ec1b']]; loss = 0.115138
106
+ Epoch 0: | | 200/? [12:24<00:00, 0.27it/s, v_num=mfhp]context = [[8, 10, 14, 15, 22, 23, 27, 30, 36, 38, 39, 47, 64, 65, 67, 78, 79, 80, 83, 86, 93, 96, 98, 105]]target = [[73, 63, 10, 27, 89, 35, 44, 58, 97, 71, 17, 24, 66, 87, 50, 12, 23, 11, 31, 45, 69, 96, 98, 94]]
107
+ [2026-03-03 17:48:34,774][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
108
+ result[selector] = overlay
109
+
110
+ Epoch 0: | | 209/? [13:04<00:00, 0.27it/s, v_num=mfhp]train step 210; scene = [['9be9b273b3c22c61'], ['4b5883872c9b860c']]; loss = 0.090815
111
+ Epoch 0: | | 210/? [13:07<00:00, 0.27it/s, v_num=mfhp]context = [[34, 35, 37, 54, 56, 59, 61, 68, 69, 76, 90, 96, 98, 101, 103, 107, 116, 119, 120, 121, 122, 125, 126, 131]]target = [[95, 50, 119, 121, 84, 107, 72, 52, 80, 42, 127, 94, 79, 98, 46, 128, 73, 75, 106, 92, 37, 110, 96, 56]]
112
+ Epoch 0: | | 219/? [13:41<00:00, 0.27it/s, v_num=mfhp]train step 220; scene = [['a3b6faa8d238d993'], ['df9ba36fbe753843']]; loss = 0.064624
113
+ Epoch 0: | | 220/? [13:45<00:00, 0.27it/s, v_num=mfhp]context = [[39, 71, 74], [21, 41, 51], [28, 55, 59], [2, 31, 37], [15, 25, 48], [57, 64, 93], [76, 87, 105], [31, 53, 64]]target = [[52, 43, 59], [28, 25, 49], [48, 58, 51], [26, 9, 6], [30, 32, 47], [84, 88, 59], [104, 87, 88], [42, 52, 59]]
114
+ Epoch 0: | | 229/? [14:18<00:00, 0.27it/s, v_num=mfhp]train step 230; scene = [['ca04de3c55cd1ca0'], ['3d90d586b33daa63'], ['d1772c09b4b6d95f'], ['03d05f69a1cab4f8'], ['60d296908f43a97a'], ['37c400e282bc481e']]; loss = 0.075824
115
+ Epoch 0: | | 230/? [14:22<00:00, 0.27it/s, v_num=mfhp]context = [[203, 204, 208, 209, 210, 233], [41, 42, 43, 52, 60, 76], [2, 8, 9, 23, 29, 31], [70, 74, 92, 97, 98, 100]]target = [[226, 223, 210, 217, 228, 222], [46, 70, 71, 75, 43, 56], [24, 23, 29, 25, 3, 21], [95, 81, 74, 73, 98, 92]]
116
+ Epoch 0: | | 239/? [14:56<00:00, 0.27it/s, v_num=mfhp]train step 240; scene = [['9794641b7e015578']]; loss = 0.116614
117
+ Epoch 0: | | 240/? [14:59<00:00, 0.27it/s, v_num=mfhp]context = [[62, 65, 68, 71, 80, 86, 87, 96, 99, 101, 103, 111], [134, 138, 139, 142, 151, 163, 171, 172, 173, 174, 181, 183]]target = [[87, 89, 96, 93, 103, 71, 65, 77, 63, 98, 102, 105], [165, 177, 159, 147, 138, 152, 171, 141, 181, 146, 161, 176]]
118
+ Epoch 0: | | 249/? [15:33<00:00, 0.27it/s, v_num=mfhp]train step 250; scene = [['93dff1b985f2c7f9']]; loss = 0.094535
119
+ Epoch 0: | | 250/? [15:37<00:00, 0.27it/s, v_num=mfhp]Validation epoch start on rank 0
120
+ Validation: | | 0/? [00:00<?, ?it/s]validation step 250; scene = ['49b8f80c849dc341'];
121
+ target intrinsic: tensor(0.8891, device='cuda:0') tensor(0.8894, device='cuda:0') | 0/1 [00:00<?, ?it/s]
122
+ pred intrinsic: tensor(0.8626, device='cuda:0') tensor(0.8625, device='cuda:0')
123
+ [2026-03-03 17:51:44,568][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
124
+ result[selector] = overlay
125
+
126
+ Epoch 0: | | 250/? [15:38<00:00, 0.27it/s, v_num=mfhp]context = [[113, 115, 120, 131, 135, 137, 144, 152], [16, 20, 23, 28, 32, 35, 39, 52], [14, 18, 25, 26, 28, 29, 51, 53]]target = [[149, 139, 118, 151, 137, 141, 121, 130], [45, 51, 28, 50, 35, 37, 27, 23], [29, 26, 22, 51, 24, 34, 47, 45]]
127
+ Epoch 0: | | 259/? [16:10<00:00, 0.27it/s, v_num=mfhp]train step 260; scene = [['b2288bf7003d5d4d']]; loss = 0.079294
128
+ Epoch 0: | | 260/? [16:13<00:00, 0.27it/s, v_num=mfhp]context = [[206, 208, 212, 216, 221, 236], [58, 60, 68, 69, 70, 93], [15, 19, 26, 34, 41, 44], [15, 18, 23, 34, 42, 44]]target = [[225, 216, 234, 214, 207, 232], [88, 76, 91, 67, 64, 70], [20, 29, 22, 36, 23, 40], [42, 28, 31, 30, 16, 37]]
129
+ Epoch 0: | | 269/? [16:47<00:00, 0.27it/s, v_num=mfhp]train step 270; scene = [['013ec74a4fde6737'], ['78e816776b064fc4'], ['1b778f72bbee1f27'], ['c71549de92ecb2e4'], ['8e16c8644efeec52'], ['35c5fc80e85db7cd'], ['34c8c62d878eca66'], ['203a5fd3a45ac4a7']]; loss = 0.062206
130
+ Epoch 0: | | 270/? [16:51<00:00, 0.27it/s, v_num=mfhp]context = [[160, 174, 177, 178, 192, 200], [34, 52, 54, 56, 58, 67], [78, 85, 101, 102, 109, 118], [123, 135, 136, 144, 150, 155]]target = [[169, 195, 175, 174, 194, 189], [55, 53, 57, 48, 40, 66], [90, 108, 110, 84, 109, 83], [131, 130, 127, 124, 139, 133]]
131
+ Epoch 0: | | 279/? [17:25<00:00, 0.27it/s, v_num=mfhp]train step 280; scene = [['75335793f866b96d'], ['e9d9dc952f5bbd83']]; loss = 0.040364
132
+ Epoch 0: | | 280/? [17:29<00:00, 0.27it/s, v_num=mfhp]context = [[2, 31], [39, 73], [69, 98], [58, 92], [109, 146], [27, 59], [83, 114], [199, 231], [13, 50], [77, 109], [72, 105], [7, 45]]target = [[20, 26], [71, 57], [92, 77], [61, 72], [125, 129], [44, 51], [110, 101], [211, 202], [27, 17], [98, 104], [80, 89], [36, 37]]
133
+ Epoch 0: | | 289/? [18:00<00:00, 0.27it/s, v_num=mfhp]train step 290; scene = [['51252022ddf74fb9'], ['8dd73309b133b8bf'], ['9e8db62a9b3cbd5e'], ['d41e59ee023e977b'], ['ce1a9465dc08ef4c'], ['e7887dec76685627']]; loss = 0.071847
134
+ Epoch 0: | | 290/? [18:04<00:00, 0.27it/s, v_num=mfhp]context = [[74, 83, 85, 88, 89, 93, 100, 101, 108, 119, 122, 123], [17, 20, 23, 24, 28, 30, 32, 35, 48, 52, 64, 66]]target = [[110, 79, 95, 92, 111, 109, 99, 89, 77, 98, 81, 102], [28, 29, 52, 36, 54, 64, 49, 23, 40, 58, 31, 27]]
135
+ Epoch 0: | | 299/? [18:38<00:00, 0.27it/s, v_num=mfhp]train step 300; scene = [['0c5d83212982c0ec'], ['00793a8a3b268d7c'], ['47a9b1e96499a466'], ['a1fb990016d7b3af']]; loss = 0.051479
136
+ Epoch 0: | | 300/? [18:42<00:00, 0.27it/s, v_num=mfhp]context = [[83, 88, 89, 91, 92, 100, 102, 104, 107, 111, 113, 115, 129, 132, 134, 135, 140, 142, 145, 150, 154, 173, 175, 180]]target = [[115, 140, 130, 169, 114, 163, 175, 177, 104, 152, 151, 89, 154, 123, 132, 85, 122, 119, 155, 91, 117, 171, 158, 157]]
137
+ Epoch 0: | | 309/? [19:14<00:00, 0.27it/s, v_num=mfhp]train step 310; scene = [['9b73ab94b5c43711'], ['8c845b940aa8244c'], ['b2789c1a5c127a02'], ['3db6c0e172d18826']]; loss = 0.066993
138
+ Epoch 0: | | 310/? [19:18<00:00, 0.27it/s, v_num=mfhp]context = [[30, 47, 66], [16, 31, 46], [10, 28, 40], [43, 68, 75], [21, 43, 63], [15, 29, 46], [36, 44, 66], [17, 39, 59]]target = [[51, 41, 31], [29, 43, 44], [23, 35, 33], [64, 71, 63], [48, 59, 27], [45, 29, 27], [60, 42, 50], [34, 26, 47]]
139
+ Epoch 0: | | 319/? [19:52<00:00, 0.27it/s, v_num=mfhp]train step 320; scene = [['591cd9d079cd7842'], ['3dd7802a2c93a865']]; loss = 0.084790
140
+ Epoch 0: | | 320/? [19:56<00:00, 0.27it/s, v_num=mfhp]context = [[4, 22, 28, 31, 34, 37, 45, 46], [26, 28, 38, 41, 46, 55, 57, 59], [129, 139, 145, 150, 151, 158, 167, 169]]target = [[7, 41, 13, 5, 20, 27, 32, 44], [29, 32, 54, 28, 40, 52, 43, 33], [137, 143, 146, 152, 130, 131, 145, 133]]
141
+ Epoch 0: | | 329/? [20:30<00:00, 0.27it/s, v_num=mfhp]train step 330; scene = [['30d9f6321281dade'], ['2a08fac923c9e50d']]; loss = 0.063144
142
+ Epoch 0: | | 330/? [20:34<00:00, 0.27it/s, v_num=mfhp]context = [[4, 6, 12, 14, 24, 41, 42, 44, 45, 48, 49, 53], [102, 106, 112, 113, 118, 122, 125, 130, 134, 140, 149, 151]]target = [[50, 42, 7, 28, 11, 33, 20, 45, 25, 8, 22, 16], [135, 134, 113, 128, 139, 112, 117, 108, 122, 127, 138, 120]]
143
+ Epoch 0: | | 339/? [21:06<00:00, 0.27it/s, v_num=mfhp]train step 340; scene = [['bd9f2096d355b1b8'], ['07d3325178e7a790'], ['8204d757ce43dda8']]; loss = 0.062880
144
+ Epoch 0: | | 340/? [21:10<00:00, 0.27it/s, v_num=mfhp]context = [[87, 88, 93, 105, 107, 111, 116, 118, 123, 131, 132, 136], [104, 111, 116, 121, 122, 127, 131, 132, 136, 137, 142, 153]]target = [[114, 108, 107, 130, 95, 129, 118, 104, 116, 96, 109, 99], [127, 106, 145, 129, 114, 109, 143, 150, 111, 133, 146, 122]]
145
+ Epoch 0: | | 349/? [21:44<00:00, 0.27it/s, v_num=mfhp]train step 350; scene = [['9d0bfbe5b7f98545'], ['06a16655c8e8ad9c']]; loss = 0.105639
146
+ Epoch 0: | | 350/? [21:48<00:00, 0.27it/s, v_num=mfhp]context = [[33, 52, 53, 56, 58, 64, 73, 75], [100, 102, 115, 117, 130, 134, 139, 142], [223, 225, 240, 242, 255, 256, 263, 267]]target = [[41, 60, 53, 63, 72, 38, 37, 73], [122, 141, 116, 138, 110, 119, 118, 135], [260, 237, 261, 264, 238, 255, 243, 253]]
147
+ Epoch 0: | | 359/? [22:21<00:00, 0.27it/s, v_num=mfhp]train step 360; scene = [['73b27f4f150327af'], ['169aaaf51ef3849c'], ['068a8406f1a383d8'], ['a9936b77895f33b3']]; loss = 0.078213
148
+ Epoch 0: | | 360/? [22:25<00:00, 0.27it/s, v_num=mfhp]context = [[0, 14, 18, 23, 25, 32, 42, 45, 46, 49, 53, 55, 59, 60, 61, 74, 78, 80, 84, 85, 89, 90, 92, 97]]target = [[50, 81, 69, 64, 38, 86, 91, 16, 27, 33, 44, 90, 48, 23, 8, 79, 6, 39, 42, 36, 82, 78, 59, 54]]
149
+ Epoch 0: | | 369/? [22:57<00:00, 0.27it/s, v_num=mfhp]train step 370; scene = [['8673faf0a9d48165'], ['99a0790d72e6c2af'], ['6cbbe9075b0d2138']]; loss = 0.060302
150
+ Epoch 0: | | 370/? [23:01<00:00, 0.27it/s, v_num=mfhp]context = [[61, 63, 65, 72, 73, 74, 76, 89, 92, 94, 100, 101, 120, 124, 126, 127, 136, 140, 144, 145, 150, 152, 156, 158]]target = [[121, 100, 147, 98, 148, 143, 126, 63, 73, 141, 79, 119, 115, 106, 153, 101, 120, 71, 91, 62, 105, 84, 151, 145]]
151
+ Epoch 0: | | 379/? [23:35<00:00, 0.27it/s, v_num=mfhp]train step 380; scene = [['656330f47c5df010'], ['6dfb89a98e14ca66']]; loss = 0.060539
152
+ Epoch 0: | | 380/? [23:38<00:00, 0.27it/s, v_num=mfhp]context = [[210, 215, 223, 225, 231, 234, 235, 249], [28, 32, 35, 53, 54, 56, 57, 68], [149, 150, 154, 155, 174, 177, 181, 183]]target = [[241, 236, 218, 230, 223, 228, 239, 227], [55, 46, 44, 49, 30, 33, 60, 45], [157, 160, 161, 164, 170, 181, 168, 166]]
153
+ Epoch 0: | | 389/? [24:12<00:00, 0.27it/s, v_num=mfhp]train step 390; scene = [['723f94d150ab09f2'], ['393cdfb7e832d285'], ['14900b71ac66b7bd'], ['452625cd6b071b87'], ['281599bbab3e73dd'], ['0a2b42e240751d33']]; loss = 0.071231
154
+ Epoch 0: | | 390/? [24:15<00:00, 0.27it/s, v_num=mfhp]context = [[48, 52, 80], [14, 47, 60], [142, 159, 187], [78, 120, 121], [6, 24, 45], [47, 68, 90], [15, 19, 48], [9, 20, 52]]target = [[51, 50, 70], [45, 15, 56], [172, 148, 157], [120, 98, 109], [20, 33, 11], [83, 84, 73], [45, 30, 32], [16, 36, 45]]
155
+ Epoch 0: | | 399/? [24:47<00:00, 0.27it/s, v_num=mfhp]train step 400; scene = [['4303746d8f23f16b'], ['0fe8246bb7e2fe40'], ['b7d77240852d6a52'], ['6e5505414fd63528'], ['44985936f68c3a36'], ['1550f1b4fff1f2a4'], ['cea3d842c3285c65'], ['b34bb5f53856d34f']]; loss = 0.088220
156
+ Epoch 0: | | 400/? [24:51<00:00, 0.27it/s, v_num=mfhp]context = [[9, 20, 28, 31, 34, 53], [189, 196, 216, 223, 230, 232], [74, 77, 86, 88, 98, 118], [132, 145, 157, 168, 179, 181]]target = [[51, 32, 14, 50, 45, 34], [212, 226, 201, 205, 215, 222], [93, 113, 107, 94, 83, 79], [161, 139, 177, 151, 156, 172]]
157
+ [2026-03-03 18:01:02,020][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
158
+ result[selector] = overlay
159
+
160
+ Epoch 0: | | 409/? [25:25<00:00, 0.27it/s, v_num=mfhp]train step 410; scene = [['144e1ec915e46d29'], ['b290b6a0afa1dac7'], ['b3d84dba6581c3d9']]; loss = 0.065141
161
+ Epoch 0: | | 410/? [25:29<00:00, 0.27it/s, v_num=mfhp]context = [[26, 40, 41, 50, 55, 58], [159, 174, 179, 192, 198, 201], [2, 3, 5, 7, 24, 40], [187, 204, 205, 208, 217, 221]]target = [[38, 32, 44, 52, 42, 27], [198, 191, 196, 162, 163, 169], [33, 17, 23, 37, 15, 39], [214, 211, 195, 220, 210, 192]]
162
+ Epoch 0: | | 419/? [26:02<00:00, 0.27it/s, v_num=mfhp]train step 420; scene = [['a1dff9c50d92dc9c']]; loss = 0.056085
163
+ Epoch 0: | | 420/? [26:06<00:00, 0.27it/s, v_num=mfhp]context = [[11, 13, 14, 26, 27, 41, 46, 51], [151, 155, 161, 167, 172, 176, 191, 195], [49, 52, 55, 57, 59, 74, 79, 90]]target = [[25, 44, 17, 33, 37, 24, 38, 40], [193, 191, 194, 162, 158, 176, 163, 171], [87, 85, 67, 58, 86, 68, 83, 78]]
164
+ Epoch 0: | | 429/? [26:39<00:00, 0.27it/s, v_num=mfhp]train step 430; scene = [['36664e22fd10a141'], ['0474328f4cefd619']]; loss = 0.051113
165
+ Epoch 0: | | 430/? [26:42<00:00, 0.27it/s, v_num=mfhp]context = [[123, 130, 131, 132, 136, 154, 169, 173], [35, 36, 47, 49, 65, 67, 76, 86], [33, 34, 39, 41, 59, 64, 65, 73]]target = [[145, 147, 167, 130, 146, 170, 138, 137], [59, 38, 49, 42, 70, 76, 39, 40], [56, 44, 36, 35, 38, 45, 48, 58]]
166
+ Epoch 0: | | 439/? [27:17<00:00, 0.27it/s, v_num=mfhp]train step 440; scene = [['342099a48847f4f6'], ['5ad0327426e3718b'], ['c25b314716aa6b10'], ['c91e2b5399b14430'], ['e1d9ade67e615bd8'], ['46df912c9748215b']]; loss = 0.066512
167
+ Epoch 0: | | 440/? [27:20<00:00, 0.27it/s, v_num=mfhp]context = [[0, 33], [10, 57], [36, 82], [46, 79], [62, 99], [33, 79], [11, 52], [1, 34], [0, 37], [26, 62], [18, 67], [96, 138]]target = [[22, 18], [37, 23], [44, 54], [69, 50], [68, 69], [69, 56], [49, 41], [27, 30], [31, 17], [42, 32], [59, 54], [108, 115]]
168
+ Epoch 0: | | 449/? [27:54<00:00, 0.27it/s, v_num=mfhp]train step 450; scene = [['e19c6facac1c9624'], ['5244830b7357365b'], ['b80c2522b1070e2f'], ['6ea0ff32c8ea695c'], ['2f311b2bbbeb5940'], ['3f7992e72a096099']]; loss = 0.069044
169
+ Epoch 0: | | 450/? [27:57<00:00, 0.27it/s, v_num=mfhp]context = [[41, 42, 44, 49, 57, 66, 68, 73, 75, 76, 83, 91], [4, 5, 15, 24, 32, 35, 37, 38, 45, 47, 49, 56]]target = [[79, 77, 90, 46, 81, 83, 56, 68, 89, 70, 84, 57], [7, 8, 40, 37, 32, 10, 25, 35, 23, 43, 48, 52]]
170
+ Epoch 0: | | 459/? [28:31<00:00, 0.27it/s, v_num=mfhp]train step 460; scene = [['46fb6702ed1b9967'], ['bdc3f978b0d3aa8f']]; loss = 0.056139
171
+ Epoch 0: | | 460/? [28:35<00:00, 0.27it/s, v_num=mfhp]context = [[65, 66, 69, 78, 84, 85, 90, 91, 99, 103, 112, 116], [39, 40, 52, 63, 64, 68, 72, 73, 83, 87, 90, 92]]target = [[99, 86, 101, 76, 84, 77, 105, 66, 89, 75, 100, 106], [84, 61, 68, 82, 59, 47, 51, 71, 85, 56, 57, 50]]
172
+ Epoch 0: | | 469/? [29:07<00:00, 0.27it/s, v_num=mfhp]train step 470; scene = [['2c88995e05a17d17'], ['2b1f47da224557a3'], ['62216d162b71b5b4'], ['61d39a97cb69d99f'], ['42000d5a83b48ee4'], ['cc8480640599f9f3']]; loss = 0.065505
173
+ Epoch 0: | | 470/? [29:10<00:00, 0.27it/s, v_num=mfhp]context = [[1, 6, 15, 26, 28, 41, 46, 52], [0, 1, 3, 5, 14, 26, 32, 37], [52, 56, 58, 79, 87, 88, 93, 97]]target = [[24, 22, 32, 14, 42, 21, 12, 37], [14, 10, 11, 25, 12, 34, 7, 32], [62, 60, 80, 87, 63, 55, 78, 88]]
174
+ Epoch 0: | | 479/? [29:44<00:00, 0.27it/s, v_num=mfhp]train step 480; scene = [['2e3bb7fb33e1ed30'], ['7460f503eb18fa6a'], ['bde49071d2088850'], ['e80016be3043dfa4']]; loss = 0.097861
175
+ Epoch 0: | | 480/? [29:47<00:00, 0.27it/s, v_num=mfhp]context = [[9, 15, 37, 47], [0, 5, 7, 35], [101, 113, 134, 153], [64, 88, 93, 107], [18, 44, 56, 62], [31, 46, 58, 75]]target = [[45, 27, 28, 18], [10, 8, 14, 3], [128, 143, 110, 129], [87, 84, 90, 94], [29, 31, 40, 36], [52, 72, 33, 56]]
176
+ Epoch 0: | | 489/? [30:21<00:00, 0.27it/s, v_num=mfhp]train step 490; scene = [['83085493f4bc18d2']]; loss = 0.092899
177
+ Epoch 0: | | 490/? [30:25<00:00, 0.27it/s, v_num=mfhp]context = [[0, 3, 7, 12, 38, 40, 45, 48], [25, 28, 57, 62, 69, 70, 76, 78], [25, 34, 37, 39, 41, 53, 65, 67]]target = [[15, 7, 39, 47, 42, 26, 25, 5], [63, 40, 59, 31, 75, 34, 47, 39], [58, 35, 53, 33, 65, 44, 38, 31]]
178
+ Epoch 0: | | 499/? [30:58<00:00, 0.27it/s, v_num=mfhp]train step 500; scene = [['1241bcb5732a9502'], ['d33a9e90e1416efb']]; loss = 0.044272
179
+ Epoch 0: | | 500/? [31:02<00:00, 0.27it/s, v_num=mfhp]Validation epoch start on rank 0
180
+ Validation: | | 0/? [00:00<?, ?it/s]validation step 500; scene = ['73d6f935f31b3fd4'];
181
+ target intrinsic: tensor(0.8576, device='cuda:0') tensor(0.8579, device='cuda:0') | 0/1 [00:00<?, ?it/s]
182
+ pred intrinsic: tensor(0.8589, device='cuda:0') tensor(0.8561, device='cuda:0')
183
+ [2026-03-03 18:07:09,779][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
184
+ result[selector] = overlay
185
+
186
+ Epoch 0: | | 500/? [31:03<00:00, 0.27it/s, v_num=mfhp]context = [[1, 21, 23, 29, 39, 48, 49, 52], [17, 18, 20, 28, 34, 39, 45, 51], [99, 103, 106, 110, 116, 136, 138, 140]]target = [[4, 46, 41, 43, 40, 37, 39, 25], [30, 48, 22, 42, 44, 31, 38, 37], [112, 108, 133, 116, 111, 125, 123, 127]]
187
+ Epoch 0: | | 509/? [31:35<00:00, 0.27it/s, v_num=mfhp]train step 510; scene = [['0eed4548041bea8e'], ['277a96ce456580f4']]; loss = 0.051501
188
+ Epoch 0: | | 510/? [31:39<00:00, 0.27it/s, v_num=mfhp]context = [[3, 13, 21, 24, 27, 28, 30, 37, 42, 44, 46, 55, 62, 65, 69, 71, 73, 80, 83, 84, 89, 95, 97, 100]]target = [[51, 81, 54, 89, 55, 45, 10, 19, 21, 68, 48, 67, 33, 69, 8, 25, 64, 83, 37, 29, 24, 82, 99, 46]]
189
+ Epoch 0: | | 519/? [32:13<00:00, 0.27it/s, v_num=mfhp]train step 520; scene = [['625e3aa0ff734714'], ['395802511d26f32e'], ['39343936591c28de']]; loss = 0.137702
190
+ Epoch 0: | | 520/? [32:17<00:00, 0.27it/s, v_num=mfhp]context = [[13, 16, 17, 22, 25, 30, 35, 39, 53, 56, 62, 68, 78, 87, 90, 92, 93, 95, 99, 100, 103, 106, 107, 110]]target = [[39, 96, 58, 28, 16, 59, 44, 17, 26, 83, 103, 31, 57, 35, 107, 51, 27, 77, 46, 30, 100, 91, 93, 97]]
191
+ Epoch 0: | | 529/? [32:49<00:00, 0.27it/s, v_num=mfhp]train step 530; scene = [['0c199c575b699444'], ['70d878da47f984e4'], ['15f77c76ea744f99'], ['e54b5eec8cc47776'], ['1969ed97e68d83d9'], ['c7cf9b63dc3e5830'], ['bcef3076b93012b1'], ['ab2680bf91942e23']]; loss = 0.082492
192
+ Epoch 0: | | 530/? [32:53<00:00, 0.27it/s, v_num=mfhp]context = [[78, 84, 85, 92, 124, 127], [14, 15, 24, 52, 64, 68], [1, 6, 16, 20, 22, 40], [33, 51, 61, 65, 66, 70]]target = [[104, 103, 94, 96, 88, 82], [15, 53, 58, 47, 56, 60], [31, 10, 18, 26, 23, 24], [37, 69, 62, 65, 49, 64]]
193
+ Epoch 0: | | 539/? [33:26<00:00, 0.27it/s, v_num=mfhp]train step 540; scene = [['a071d9276f6a9272']]; loss = 0.114086
194
+ Epoch 0: | | 540/? [33:30<00:00, 0.27it/s, v_num=mfhp]context = [[35, 46, 47, 49, 52, 54, 55, 56, 57, 65, 66, 85], [51, 54, 55, 59, 68, 77, 82, 86, 87, 93, 105, 106]]target = [[69, 68, 75, 48, 44, 79, 53, 74, 60, 81, 72, 71], [77, 73, 88, 59, 67, 61, 102, 62, 93, 75, 95, 105]]
195
+ Epoch 0: | | 549/? [34:04<00:00, 0.27it/s, v_num=mfhp]train step 550; scene = [['836250796ea45b6c']]; loss = 0.085734
196
+ Epoch 0: | | 550/? [34:08<00:00, 0.27it/s, v_num=mfhp]context = [[30, 34, 39, 45, 47, 79, 80, 83], [16, 21, 22, 24, 33, 42, 51, 60], [115, 128, 129, 138, 141, 150, 156, 163]]target = [[79, 51, 62, 77, 46, 54, 33, 49], [27, 59, 20, 41, 21, 48, 58, 30], [133, 125, 124, 159, 122, 129, 155, 119]]
197
+ Epoch 0: | | 559/? [34:42<00:00, 0.27it/s, v_num=mfhp]train step 560; scene = [['d70ca840b3c5aec9'], ['65c3f29c43dd1e63'], ['d3917d0a1eda2a1f'], ['5c83dfc8f9ab44fa']]; loss = 0.060085
198
+ Epoch 0: | | 560/? [34:45<00:00, 0.27it/s, v_num=mfhp]context = [[62, 66, 67, 68, 71, 87, 88, 92, 100, 108, 110, 122], [153, 160, 165, 168, 171, 177, 178, 181, 183, 185, 200, 205]]target = [[95, 104, 83, 68, 115, 81, 74, 106, 119, 79, 90, 72], [204, 165, 196, 185, 197, 198, 195, 187, 155, 188, 154, 201]]
199
+ Epoch 0: | | 569/? [35:19<00:00, 0.27it/s, v_num=mfhp]train step 570; scene = [['9d8ddcdbe1f7ac42'], ['721df0f45094ca34'], ['fdbfe35f5940d3ad']]; loss = 0.045826
200
+ Epoch 0: | | 570/? [35:23<00:00, 0.27it/s, v_num=mfhp]context = [[23, 34, 41, 42, 48, 54, 56, 59, 60, 63, 66, 71, 74, 79, 88, 90, 92, 93, 95, 97, 110, 111, 119, 120]]target = [[30, 27, 28, 53, 31, 75, 85, 54, 77, 111, 76, 25, 44, 52, 33, 41, 69, 89, 73, 68, 26, 93, 83, 119]]
201
+ Epoch 0: | | 579/? [35:57<00:00, 0.27it/s, v_num=mfhp]train step 580; scene = [['88a0267e41b851f0'], ['df71fbb70b19cbc3'], ['1c713c10ecf5a0c9']]; loss = 0.066170
202
+ Epoch 0: | | 580/? [36:00<00:00, 0.27it/s, v_num=mfhp]context = [[9, 22, 36, 39, 43, 68], [70, 85, 95, 98, 116, 121], [10, 28, 45, 47, 50, 57], [132, 144, 154, 158, 168, 188]]target = [[67, 45, 11, 46, 50, 30], [114, 94, 71, 104, 117, 76], [22, 12, 24, 51, 50, 13], [180, 169, 183, 173, 182, 171]]
203
+ Epoch 0: | | 589/? [36:34<00:00, 0.27it/s, v_num=mfhp]train step 590; scene = [['3f732b63cdd0729e'], ['9be3165beb073d95'], ['42a6c835ff830674'], ['f928d960cbfae15a'], ['140b10a4f6bb5aa5'], ['cc8e19c8ad1846f4']]; loss = 0.089595
204
+ Epoch 0: | | 590/? [36:37<00:00, 0.27it/s, v_num=mfhp]context = [[198, 200, 201, 204, 209, 210, 224, 234], [6, 14, 18, 23, 25, 27, 58, 59], [13, 29, 40, 51, 55, 66, 70, 71]]target = [[203, 208, 227, 217, 222, 223, 200, 211], [11, 46, 45, 19, 20, 18, 47, 7], [16, 51, 40, 15, 55, 44, 23, 61]]
205
+ Epoch 0: | | 599/? [37:11<00:00, 0.27it/s, v_num=mfhp]train step 600; scene = [['cb734fdc69e9900e']]; loss = 0.055887
206
+ Epoch 0: | | 600/? [37:15<00:00, 0.27it/s, v_num=mfhp]context = [[10, 16, 24, 31, 33, 34, 36, 41, 49, 59, 63, 67], [160, 162, 167, 172, 173, 180, 183, 187, 189, 205, 212, 214]]target = [[32, 54, 42, 14, 39, 60, 66, 33, 62, 22, 47, 49], [182, 199, 163, 179, 203, 200, 174, 209, 164, 172, 194, 167]]
207
+ [2026-03-03 18:13:25,680][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
208
+ result[selector] = overlay
209
+
210
+ Epoch 0: | | 609/? [37:49<00:00, 0.27it/s, v_num=mfhp]train step 610; scene = [['ed9409fa128e193b'], ['a5c03b0c5fb7203e']]; loss = 0.042144
211
+ Epoch 0: | | 610/? [37:53<00:00, 0.27it/s, v_num=mfhp]context = [[36, 37, 39, 50, 51, 52, 57, 59, 72, 73, 86, 94], [45, 46, 54, 57, 66, 75, 77, 81, 83, 89, 97, 101]]target = [[47, 37, 89, 79, 92, 78, 81, 90, 38, 56, 84, 69], [84, 91, 85, 69, 79, 53, 55, 48, 87, 49, 92, 63]]
212
+ Epoch 0: | | 619/? [38:26<00:00, 0.27it/s, v_num=mfhp]train step 620; scene = [['7898a828b7203ca4'], ['9c269fce78f0dd27'], ['e1e317857deb7afc'], ['30124191dafb3383'], ['c39f1a9a73797efe'], ['a640a55439a43108']]; loss = 0.048999
213
+ Epoch 0: | | 620/? [38:30<00:00, 0.27it/s, v_num=mfhp]context = [[2, 15, 16, 21, 26, 29, 47, 56], [6, 12, 15, 37, 45, 47, 50, 68], [9, 18, 23, 26, 39, 48, 50, 60]]target = [[38, 7, 32, 36, 16, 21, 53, 22], [50, 67, 28, 43, 15, 41, 44, 23], [12, 54, 16, 24, 58, 32, 30, 47]]
214
+ Epoch 0: | | 629/? [39:03<00:00, 0.27it/s, v_num=mfhp]train step 630; scene = [['dd3bbf1f7f832e83'], ['0ff7277275e55096'], ['5f45c360d76a3b12']]; loss = 0.040934
215
+ Epoch 0: | | 630/? [39:07<00:00, 0.27it/s, v_num=mfhp]context = [[29, 42, 44, 45, 76, 81, 87, 93], [77, 84, 94, 96, 100, 102, 121, 125], [67, 68, 70, 89, 90, 91, 96, 107]]target = [[83, 55, 36, 89, 45, 62, 90, 63], [106, 92, 114, 99, 117, 82, 119, 107], [99, 68, 75, 92, 79, 85, 101, 90]]
216
+ Epoch 0: | | 639/? [39:41<00:00, 0.27it/s, v_num=mfhp]train step 640; scene = [['867edbda9bb8ef59'], ['1d83764e77e159d8'], ['e318dafa4071cef9'], ['169f09c33ee35289']]; loss = 0.091318
217
+ Epoch 0: | | 640/? [39:45<00:00, 0.27it/s, v_num=mfhp]context = [[52, 57, 67, 75, 77, 80, 85, 86, 99, 105, 112, 117], [6, 11, 22, 23, 32, 39, 42, 47, 48, 53, 63, 64]]target = [[114, 60, 111, 110, 57, 69, 68, 80, 62, 99, 77, 90], [27, 15, 46, 56, 51, 60, 41, 63, 49, 16, 9, 43]]
218
+ Epoch 0: | | 649/? [40:18<00:00, 0.27it/s, v_num=mfhp]train step 650; scene = [['23668135f32e0126'], ['daca15248046e480'], ['174ebd189316bd92']]; loss = 0.049692
219
+ Epoch 0: | | 650/? [40:22<00:00, 0.27it/s, v_num=mfhp]context = [[6, 23, 25, 26, 28, 32, 37, 38, 39, 40, 50, 57], [0, 2, 3, 11, 17, 19, 24, 25, 38, 40, 45, 53]]target = [[15, 47, 23, 31, 12, 30, 20, 55, 33, 11, 22, 10], [51, 9, 13, 34, 52, 30, 26, 45, 38, 27, 11, 14]]
220
+ Epoch 0: | | 659/? [40:55<00:00, 0.27it/s, v_num=mfhp]train step 660; scene = [['60499200285c9abe']]; loss = 0.073265
221
+ Epoch 0: | | 660/? [40:58<00:00, 0.27it/s, v_num=mfhp]context = [[57, 65, 76, 86, 93, 99, 100, 122], [31, 34, 42, 47, 60, 80, 81, 83], [13, 16, 19, 30, 33, 49, 50, 55]]target = [[85, 75, 88, 114, 107, 70, 116, 79], [82, 35, 67, 56, 68, 61, 74, 58], [54, 44, 23, 36, 31, 39, 45, 28]]
222
+ Epoch 0: | | 669/? [41:31<00:00, 0.27it/s, v_num=mfhp]train step 670; scene = [['7665ff641f430aa5']]; loss = 0.038339
223
+ Epoch 0: | | 670/? [41:34<00:00, 0.27it/s, v_num=mfhp]context = [[50, 57, 61, 69, 82, 87, 98, 101], [26, 33, 44, 47, 53, 56, 62, 66], [11, 33, 39, 48, 56, 67, 75, 78]]target = [[67, 96, 71, 75, 73, 57, 55, 89], [65, 39, 63, 59, 60, 47, 55, 41], [18, 56, 21, 64, 32, 59, 33, 25]]
224
+ Epoch 0: | | 679/? [42:08<00:00, 0.27it/s, v_num=mfhp]train step 680; scene = [['43c939b11c5fed4a']]; loss = 0.084846
225
+ Epoch 0: | | 680/? [42:12<00:00, 0.27it/s, v_num=mfhp]context = [[47, 54, 60, 64, 79, 80, 83, 93], [12, 13, 21, 30, 34, 45, 53, 55], [57, 75, 78, 82, 96, 103, 113, 117]]target = [[80, 62, 75, 54, 57, 92, 53, 58], [38, 32, 15, 44, 42, 51, 27, 39], [81, 87, 106, 94, 99, 103, 73, 79]]
226
+ Epoch 0: | | 689/? [42:44<00:00, 0.27it/s, v_num=mfhp]train step 690; scene = [['1848b8b363d0d2b9'], ['afe6b05d0554a880']]; loss = 0.068950
227
+ Epoch 0: | | 690/? [42:48<00:00, 0.27it/s, v_num=mfhp]context = [[12, 18, 20, 27, 36, 37, 38, 43, 48, 49, 52, 58, 59, 62, 67, 81, 83, 86, 93, 95, 101, 104, 108, 109]]target = [[34, 24, 43, 31, 87, 30, 51, 54, 52, 94, 86, 21, 44, 97, 61, 95, 38, 60, 49, 73, 41, 19, 65, 67]]
228
+ Epoch 0: | | 699/? [43:20<00:00, 0.27it/s, v_num=mfhp]train step 700; scene = [['674ef9fb9cf20f9f'], ['8624ee0839cb6e4c'], ['caed302f388b799f']]; loss = 0.052812
229
+ Epoch 0: | | 700/? [43:23<00:00, 0.27it/s, v_num=mfhp]context = [[31, 32, 41, 46, 53, 54, 55, 57, 65, 68, 73, 74, 80, 85, 100, 105, 108, 109, 113, 114, 116, 118, 126, 128]]target = [[117, 52, 85, 57, 37, 45, 78, 100, 125, 35, 113, 66, 105, 103, 61, 83, 88, 40, 116, 60, 79, 32, 102, 107]]
230
+ Epoch 0: | | 709/? [43:57<00:00, 0.27it/s, v_num=mfhp]train step 710; scene = [['db6cd90de8fee2ff'], ['7a20ba81fb778529'], ['970350268b239272']]; loss = 0.047345
231
+ Epoch 0: | | 710/? [44:01<00:00, 0.27it/s, v_num=mfhp]context = [[3, 12, 16, 31, 41, 45, 49, 53, 54, 55, 67, 69], [208, 211, 217, 221, 222, 227, 230, 231, 239, 247, 252, 267]]target = [[38, 34, 68, 25, 44, 63, 48, 7, 42, 6, 64, 28], [228, 253, 259, 266, 244, 238, 249, 234, 232, 241, 220, 213]]
232
+ Epoch 0: | | 719/? [44:34<00:00, 0.27it/s, v_num=mfhp]train step 720; scene = [['f63d2df8871ce70c'], ['0fdeda15097ed4a4']]; loss = 0.043177
233
+ Epoch 0: | | 720/? [44:38<00:00, 0.27it/s, v_num=mfhp]context = [[55, 57, 71, 72, 75, 85, 90, 92, 98, 99, 102, 108, 112, 116, 125, 126, 130, 131, 135, 141, 144, 145, 146, 152]]target = [[61, 145, 86, 66, 62, 119, 100, 139, 105, 125, 58, 101, 140, 132, 118, 128, 65, 141, 151, 78, 104, 107, 150, 138]]
234
+ Epoch 0: | | 729/? [45:11<00:00, 0.27it/s, v_num=mfhp]train step 730; scene = [['232abb354c423e81'], ['d34926c73ae1277e']]; loss = 0.032255
235
+ Epoch 0: | | 730/? [45:15<00:00, 0.27it/s, v_num=mfhp]context = [[45, 50, 54, 58, 66, 70, 72, 78, 85, 88, 94, 105, 109, 110, 111, 117, 120, 122, 126, 127, 128, 133, 135, 142]]target = [[95, 96, 50, 114, 135, 124, 104, 100, 49, 119, 139, 62, 92, 123, 58, 46, 57, 112, 116, 90, 54, 101, 85, 81]]
236
+ Epoch 0: | | 739/? [45:48<00:00, 0.27it/s, v_num=mfhp]train step 740; scene = [['19f7966006ad778d'], ['dde0212418df7ca9'], ['ad75e36b74f6b033'], ['ea97e5ae55e56208'], ['9d29b0289133ab4e'], ['282938f90821bdef']]; loss = 0.119270
237
+ Epoch 0: | | 740/? [45:52<00:00, 0.27it/s, v_num=mfhp]context = [[8, 13, 19, 22, 30, 31, 36, 45, 46, 48, 54, 57, 58, 59, 60, 61, 65, 73, 77, 83, 86, 88, 93, 105]]target = [[46, 101, 90, 61, 31, 23, 37, 95, 18, 67, 32, 100, 93, 35, 89, 45, 10, 70, 60, 97, 85, 81, 66, 79]]
238
+ Epoch 0: | | 749/? [46:26<00:00, 0.27it/s, v_num=mfhp]train step 750; scene = [['f85921f42c5d98d7'], ['a95dacbd3ea3db36']]; loss = 0.045635
239
+ Epoch 0: | | 750/? [46:30<00:00, 0.27it/s, v_num=mfhp]Validation epoch start on rank 0
240
+ Validation: | | 0/? [00:00<?, ?it/s]validation step 750; scene = ['91fda69e1cda4602'];
241
+ target intrinsic: tensor(0.8937, device='cuda:0') tensor(0.8939, device='cuda:0') | 0/1 [00:00<?, ?it/s]
242
+ pred intrinsic: tensor(0.9293, device='cuda:0') tensor(0.9274, device='cuda:0')
243
+ [2026-03-03 18:22:37,515][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
244
+ result[selector] = overlay
245
+
246
+ Epoch 0: | | 750/? [46:31<00:00, 0.27it/s, v_num=mfhp]context = [[1, 38, 46, 49, 52, 59], [213, 218, 227, 233, 252, 257], [99, 105, 109, 124, 125, 153], [8, 17, 20, 37, 62, 65]]target = [[54, 19, 39, 57, 7, 4], [252, 219, 223, 216, 248, 237], [148, 136, 102, 103, 149, 132], [12, 52, 13, 44, 63, 39]]
247
+ Epoch 0: | | 759/? [47:04<00:00, 0.27it/s, v_num=mfhp]train step 760; scene = [['75617c97bff1e873'], ['ff02f88545dfa566']]; loss = 0.033483
248
+ Epoch 0: | | 760/? [47:08<00:00, 0.27it/s, v_num=mfhp]context = [[71, 73, 111], [2, 14, 56], [29, 82, 88], [120, 144, 168], [181, 214, 250], [0, 10, 71], [12, 53, 54], [14, 27, 87]]target = [[101, 74, 84], [43, 51, 38], [82, 55, 72], [156, 139, 147], [198, 214, 207], [22, 5, 40], [36, 13, 39], [68, 35, 70]]
249
+ Epoch 0: | | 769/? [47:42<00:00, 0.27it/s, v_num=mfhp]train step 770; scene = [['62b0d4ee613af70f'], ['f7926eb1096de201'], ['c63b37ec347f0d0e'], ['b43d9f7c70f5caa0']]; loss = 0.074525
250
+ Epoch 0: | | 770/? [47:45<00:00, 0.27it/s, v_num=mfhp]context = [[13, 29, 30, 31, 32, 34, 41, 42, 46, 52, 63, 66], [146, 149, 157, 166, 177, 180, 181, 193, 197, 204, 211, 213]]target = [[35, 55, 52, 59, 24, 51, 63, 18, 42, 44, 33, 61], [183, 158, 171, 175, 163, 196, 177, 181, 210, 167, 187, 174]]
251
+ Epoch 0: | | 779/? [48:19<00:00, 0.27it/s, v_num=mfhp]train step 780; scene = [['b41f4db8b8a42a71']]; loss = 0.089240
252
+ Epoch 0: | | 780/? [48:23<00:00, 0.27it/s, v_num=mfhp]context = [[43, 50, 53, 55, 57, 58, 59, 60, 68, 81, 94, 96, 102, 106, 108, 110, 112, 121, 125, 126, 130, 131, 136, 140]]target = [[74, 122, 93, 98, 70, 84, 49, 136, 77, 117, 135, 138, 123, 89, 119, 45, 129, 105, 50, 58, 63, 103, 82, 121]]
253
+ Epoch 0: | | 789/? [48:56<00:00, 0.27it/s, v_num=mfhp]train step 790; scene = [['d79666d294813d8e']]; loss = 0.259840
254
+ Epoch 0: | | 790/? [48:59<00:00, 0.27it/s, v_num=mfhp]context = [[59, 79, 88, 93, 100, 107], [3, 16, 19, 31, 46, 51], [41, 49, 54, 56, 72, 83], [226, 229, 235, 244, 255, 272]]target = [[103, 98, 81, 75, 74, 72], [48, 44, 20, 31, 15, 46], [64, 82, 70, 50, 47, 60], [251, 249, 271, 250, 259, 260]]
255
+ Epoch 0: | | 799/? [49:33<00:00, 0.27it/s, v_num=mfhp]train step 800; scene = [['cb797cd30542e55c']]; loss = 0.091464
256
+ Epoch 0: | | 800/? [49:36<00:00, 0.27it/s, v_num=mfhp]context = [[10, 19, 20, 39, 42, 55], [1, 6, 11, 20, 30, 42], [0, 3, 4, 19, 42, 45], [21, 35, 52, 82, 85, 88]]target = [[38, 39, 26, 21, 37, 40], [11, 40, 18, 27, 3, 24], [43, 33, 8, 24, 5, 19], [83, 71, 66, 32, 63, 47]]
257
+ [2026-03-03 18:25:47,338][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
258
+ result[selector] = overlay
259
+
260
+ Epoch 0: | | 809/? [50:10<00:00, 0.27it/s, v_num=mfhp]train step 810; scene = [['9bd08fc9288bef8b']]; loss = 0.096803
261
+ Epoch 0: | | 810/? [50:14<00:00, 0.27it/s, v_num=mfhp]context = [[5, 6, 14, 17, 19, 23, 30, 37, 40, 49, 53, 55], [3, 6, 15, 17, 27, 34, 38, 40, 42, 46, 56, 65]]target = [[16, 8, 48, 23, 44, 49, 10, 7, 18, 50, 22, 15], [34, 18, 9, 42, 31, 38, 45, 26, 43, 6, 49, 44]]
262
+ Epoch 0: | | 819/? [50:47<00:00, 0.27it/s, v_num=mfhp]train step 820; scene = [['49952f737be91dd2']]; loss = 0.073990
263
+ Epoch 0: | | 820/? [50:51<00:00, 0.27it/s, v_num=mfhp]context = [[22, 23, 26, 28, 31, 43, 51, 53, 57, 65, 67, 68, 71, 76, 77, 78, 80, 84, 88, 90, 107, 112, 116, 119]]target = [[35, 45, 31, 72, 24, 71, 89, 54, 25, 61, 44, 97, 56, 47, 82, 87, 50, 110, 108, 75, 103, 67, 86, 117]]
264
+ Epoch 0: | | 829/? [51:23<00:00, 0.27it/s, v_num=mfhp]train step 830; scene = [['2f9c9d1b56eb7f75'], ['1c392cc98b3a7642']]; loss = 0.063626
265
+ Epoch 0: | | 830/? [51:27<00:00, 0.27it/s, v_num=mfhp]context = [[42, 67, 90, 119], [18, 54, 72, 90], [43, 67, 73, 102], [2, 22, 40, 43], [23, 40, 52, 67], [19, 37, 75, 78]]target = [[86, 112, 87, 94], [59, 40, 20, 47], [58, 97, 66, 77], [34, 19, 32, 17], [62, 37, 35, 54], [54, 22, 31, 50]]
266
+ Epoch 0: | | 839/? [51:59<00:00, 0.27it/s, v_num=mfhp]train step 840; scene = [['c51c7bc0c8151abb'], ['a0d16e79ab441c4f']]; loss = 0.133728
267
+ Epoch 0: | | 840/? [52:03<00:00, 0.27it/s, v_num=mfhp]context = [[1, 18, 22, 27, 29, 35, 38, 49], [3, 12, 30, 36, 43, 47, 54, 55], [17, 36, 55, 65, 72, 77, 86, 88]]target = [[5, 36, 48, 11, 2, 39, 35, 37], [6, 54, 43, 18, 30, 15, 10, 11], [52, 59, 56, 77, 25, 68, 22, 27]]
268
+ Epoch 0: | | 849/? [52:36<00:00, 0.27it/s, v_num=mfhp]train step 850; scene = [['818df63ddd1cf294'], ['3002f9cbe7f00e6c'], ['8ec42c5dfea6823b'], ['64ae75e57c6aa0a4']]; loss = 0.140717
269
+ Epoch 0: | | 850/? [52:40<00:00, 0.27it/s, v_num=mfhp]context = [[31, 75, 86], [20, 22, 73], [7, 55, 63], [3, 62, 68], [10, 43, 54], [52, 92, 111], [0, 34, 42], [51, 90, 102]]target = [[40, 57, 36], [39, 67, 34], [12, 9, 52], [29, 12, 39], [49, 13, 17], [106, 93, 85], [39, 20, 9], [79, 60, 53]]
270
+ Epoch 0: | | 859/? [53:14<00:00, 0.27it/s, v_num=mfhp]train step 860; scene = [['eb0aa1a4fb58c50c']]; loss = 0.039084
271
+ Epoch 0: | | 860/? [53:17<00:00, 0.27it/s, v_num=mfhp]context = [[147, 155, 165, 189, 194, 195], [9, 15, 44, 47, 49, 82], [5, 22, 23, 51, 61, 64], [97, 98, 140, 149, 160, 170]]target = [[194, 192, 155, 151, 174, 158], [22, 27, 54, 18, 81, 31], [25, 27, 29, 23, 6, 59], [137, 162, 98, 111, 152, 164]]
272
+ Epoch 0: | | 869/? [53:51<00:00, 0.27it/s, v_num=mfhp]train step 870; scene = [['c357fbd8aca05570']]; loss = 0.051317
273
+ Epoch 0: | | 870/? [53:55<00:00, 0.27it/s, v_num=mfhp]context = [[44, 45, 82, 100], [70, 93, 96, 147], [144, 176, 184, 205], [0, 51, 65, 67], [105, 106, 130, 152], [5, 6, 55, 60]]target = [[81, 66, 67, 97], [116, 118, 137, 91], [200, 193, 197, 176], [11, 13, 22, 32], [120, 117, 127, 130], [17, 14, 58, 27]]
274
+ Epoch 0: | | 879/? [54:28<00:00, 0.27it/s, v_num=mfhp]train step 880; scene = [['c672fa3960b73528'], ['a60e4127f167ac93'], ['a9739ec3a34012af']]; loss = 0.079480
275
+ Epoch 0: | | 880/? [54:32<00:00, 0.27it/s, v_num=mfhp]context = [[15, 85, 91], [0, 68, 74], [73, 85, 121], [4, 53, 80], [15, 36, 63], [42, 81, 92], [0, 45, 49], [65, 97, 132]]target = [[49, 21, 35], [47, 44, 69], [119, 89, 106], [57, 76, 6], [28, 32, 55], [62, 43, 49], [18, 44, 24], [90, 104, 71]]
276
+ Epoch 0: | | 889/? [55:04<00:00, 0.27it/s, v_num=mfhp]train step 890; scene = [['40a3f4f9389dd20c'], ['b14ec6f019932d8d'], ['4b9ed7532c875dab'], ['10c36bd5ef5f5a6b'], ['bc9a64096787007d'], ['d58a26d24f2776b2'], ['46f2228076e6f3f7'], ['399668567ff33ad7']]; loss = 0.065856
277
+ Epoch 0: | | 890/? [55:08<00:00, 0.27it/s, v_num=mfhp]context = [[32, 34, 35, 37, 39, 44, 52, 59, 73, 84, 87, 93], [20, 21, 22, 34, 43, 47, 52, 57, 64, 69, 72, 96]]target = [[87, 80, 41, 73, 35, 37, 86, 66, 72, 39, 50, 83], [28, 22, 50, 33, 26, 79, 75, 73, 29, 77, 60, 37]]
278
+ Epoch 0: | | 893/? [55:19<00:00, 0.27it/s, v_num=mfhp]
ABLATION_0302_noTgtAlign/wandb/run-20260303_173555-et94mfhp/files/requirements.txt ADDED
@@ -0,0 +1,173 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wheel==0.45.1
2
+ pytz==2025.2
3
+ easydict==1.13
4
+ antlr4-python3-runtime==4.9.3
5
+ wadler_lindig==0.1.7
6
+ networkx==3.4.2
7
+ urllib3==2.5.0
8
+ tzdata==2025.2
9
+ typing-inspection==0.4.1
10
+ tabulate==0.9.0
11
+ smmap==5.0.2
12
+ setuptools==78.1.1
13
+ safetensors==0.5.3
14
+ multidict==6.6.4
15
+ PyYAML==6.0.2
16
+ PySocks==1.7.1
17
+ pyparsing==3.2.5
18
+ pydantic_core==2.33.2
19
+ pycparser==2.23
20
+ protobuf==6.32.1
21
+ propcache==0.3.2
22
+ proglog==0.1.12
23
+ platformdirs==4.4.0
24
+ pip==25.2
25
+ mdurl==0.1.2
26
+ pillow==10.4.0
27
+ packaging==24.2
28
+ opt_einsum==3.4.0
29
+ frozenlist==1.7.0
30
+ numpy==1.26.4
31
+ ninja==1.13.0
32
+ MarkupSafe==3.0.2
33
+ kornia_rs==0.1.9
34
+ kiwisolver==1.4.9
35
+ imageio-ffmpeg==0.6.0
36
+ idna==3.7
37
+ fsspec==2024.6.1
38
+ hf-xet==1.1.10
39
+ gmpy2==2.2.1
40
+ fonttools==4.60.0
41
+ triton==3.4.0
42
+ filelock==3.17.0
43
+ einops==0.8.1
44
+ decorator==4.4.2
45
+ dacite==1.9.2
46
+ cycler==0.12.1
47
+ colorama==0.4.6
48
+ click==8.3.0
49
+ nvidia-nvtx-cu12==12.8.90
50
+ charset-normalizer==3.3.2
51
+ certifi==2025.8.3
52
+ beartype==0.19.0
53
+ attrs==25.3.0
54
+ async-timeout==5.0.1
55
+ annotated-types==0.7.0
56
+ aiohappyeyeballs==2.6.1
57
+ yarl==1.20.1
58
+ tifffile==2025.5.10
59
+ sentry-sdk==2.39.0
60
+ scipy==1.15.3
61
+ pydantic==2.11.9
62
+ pandas==2.3.2
63
+ opencv-python==4.11.0.86
64
+ omegaconf==2.3.0
65
+ markdown-it-py==4.0.0
66
+ lightning-utilities==0.14.3
67
+ lazy_loader==0.4
68
+ jaxtyping==0.2.37
69
+ imageio==2.37.0
70
+ gitdb==4.0.12
71
+ contourpy==1.3.2
72
+ colorspacious==1.1.2
73
+ cffi==1.17.1
74
+ aiosignal==1.4.0
75
+ scikit-video==1.1.11
76
+ scikit-image==0.25.2
77
+ rich==14.1.0
78
+ moviepy==1.0.3
79
+ matplotlib==3.10.6
80
+ hydra-core==1.3.2
81
+ huggingface-hub==0.35.1
82
+ GitPython==3.1.45
83
+ brotlicffi==1.0.9.2
84
+ aiohttp==3.12.15
85
+ torchmetrics==1.8.2
86
+ opt-einsum-fx==0.1.4
87
+ kornia==0.8.1
88
+ pytorch-lightning==2.5.1
89
+ lpips==0.1.4
90
+ e3nn==0.6.0
91
+ lightning==2.5.1
92
+ gsplat==1.5.3
93
+ nvidia-cusparselt-cu12==0.7.1
94
+ nvidia-nvjitlink-cu12==12.8.93
95
+ nvidia-nccl-cu12==2.27.3
96
+ nvidia-curand-cu12==10.3.9.90
97
+ nvidia-cufile-cu12==1.13.1.3
98
+ nvidia-cuda-runtime-cu12==12.8.90
99
+ nvidia-cuda-nvrtc-cu12==12.8.93
100
+ nvidia-cuda-cupti-cu12==12.8.90
101
+ nvidia-cublas-cu12==12.8.4.1
102
+ nvidia-cusparse-cu12==12.5.8.93
103
+ nvidia-cufft-cu12==11.3.3.83
104
+ nvidia-cudnn-cu12==9.10.2.21
105
+ nvidia-cusolver-cu12==11.7.3.90
106
+ torch==2.8.0+cu128
107
+ torchvision==0.23.0+cu128
108
+ torchaudio==2.8.0+cu128
109
+ torch_scatter==2.1.2+pt28cu128
110
+ plyfile==1.1.3
111
+ wandb==0.25.0
112
+ cuda-bindings==13.0.3
113
+ cuda-pathfinder==1.3.3
114
+ Jinja2==3.1.6
115
+ mpmath==1.3.0
116
+ nvidia-cublas==13.1.0.3
117
+ nvidia-cuda-cupti==13.0.85
118
+ nvidia-cuda-nvrtc==13.0.88
119
+ nvidia-cuda-runtime==13.0.96
120
+ nvidia-cudnn-cu13==9.15.1.9
121
+ nvidia-cufft==12.0.0.61
122
+ nvidia-cufile==1.15.1.6
123
+ nvidia-curand==10.4.0.35
124
+ nvidia-cusolver==12.0.4.66
125
+ nvidia-cusparse==12.6.3.3
126
+ nvidia-cusparselt-cu13==0.8.0
127
+ nvidia-nccl-cu13==2.28.9
128
+ nvidia-nvjitlink==13.0.88
129
+ nvidia-nvshmem-cu13==3.4.5
130
+ nvidia-nvtx==13.0.85
131
+ requests==2.32.5
132
+ sentencepiece==0.2.1
133
+ sympy==1.14.0
134
+ torchcodec==0.10.0
135
+ torchdata==0.10.0
136
+ torchtext==0.6.0
137
+ anyio==4.12.0
138
+ asttokens==3.0.1
139
+ comm==0.2.3
140
+ debugpy==1.8.19
141
+ executing==2.2.1
142
+ h11==0.16.0
143
+ httpcore==1.0.9
144
+ httpx==0.28.1
145
+ ipykernel==7.1.0
146
+ ipython==9.8.0
147
+ ipython_pygments_lexers==1.1.1
148
+ ipywidgets==8.1.8
149
+ jedi==0.19.2
150
+ jupyter_client==8.7.0
151
+ jupyter_core==5.9.1
152
+ jupyterlab_widgets==3.0.16
153
+ matplotlib-inline==0.2.1
154
+ nest-asyncio==1.6.0
155
+ parso==0.8.5
156
+ pexpect==4.9.0
157
+ prompt_toolkit==3.0.52
158
+ psutil==7.2.1
159
+ ptyprocess==0.7.0
160
+ pure_eval==0.2.3
161
+ Pygments==2.19.2
162
+ python-dateutil==2.9.0.post0
163
+ pyzmq==27.1.0
164
+ shellingham==1.5.4
165
+ six==1.17.0
166
+ stack-data==0.6.3
167
+ tornado==6.5.4
168
+ tqdm==4.67.1
169
+ traitlets==5.14.3
170
+ typer-slim==0.21.0
171
+ typing_extensions==4.15.0
172
+ wcwidth==0.2.14
173
+ widgetsnbextension==4.0.15
ABLATION_0302_noTgtAlign/wandb/run-20260303_173555-et94mfhp/files/wandb-metadata.json ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.8.0-94-generic-x86_64-with-glibc2.39",
3
+ "python": "CPython 3.12.12",
4
+ "startedAt": "2026-03-03T17:35:55.406237Z",
5
+ "args": [
6
+ "+experiment=re10k_ablation_24v",
7
+ "wandb.mode=online",
8
+ "wandb.name=ABLATION_0302_noTgtAlign",
9
+ "model.density_control.score_mode=random"
10
+ ],
11
+ "program": "-m src.main",
12
+ "git": {
13
+ "remote": "git@github.com:K-nowing/CVPR2026.git",
14
+ "commit": "9dfce172a0f8c7ce85e763899f7ef741ecffc454"
15
+ },
16
+ "email": "dna9041@korea.ac.kr",
17
+ "root": "/workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign",
18
+ "host": "e9d3310a05da",
19
+ "executable": "/venv/main/bin/python",
20
+ "cpu_count": 128,
21
+ "cpu_count_logical": 256,
22
+ "gpu": "NVIDIA H200",
23
+ "gpu_count": 8,
24
+ "disk": {
25
+ "/": {
26
+ "total": "735513149440",
27
+ "used": "700808368128"
28
+ }
29
+ },
30
+ "memory": {
31
+ "total": "1622950240256"
32
+ },
33
+ "gpu_nvidia": [
34
+ {
35
+ "name": "NVIDIA H200",
36
+ "memoryTotal": "150754820096",
37
+ "cudaCores": 16896,
38
+ "architecture": "Hopper",
39
+ "uuid": "GPU-9a20101e-d876-facd-5f05-805081aede41"
40
+ },
41
+ {
42
+ "name": "NVIDIA H200",
43
+ "memoryTotal": "150754820096",
44
+ "cudaCores": 16896,
45
+ "architecture": "Hopper",
46
+ "uuid": "GPU-84736a77-ee75-3324-e4e1-99cc15bfb5e9"
47
+ },
48
+ {
49
+ "name": "NVIDIA H200",
50
+ "memoryTotal": "150754820096",
51
+ "cudaCores": 16896,
52
+ "architecture": "Hopper",
53
+ "uuid": "GPU-423d3161-cdc4-3fc0-caee-d15cfaa83ca6"
54
+ },
55
+ {
56
+ "name": "NVIDIA H200",
57
+ "memoryTotal": "150754820096",
58
+ "cudaCores": 16896,
59
+ "architecture": "Hopper",
60
+ "uuid": "GPU-5b0058b2-cdb9-c952-04f9-87dcaa7ea742"
61
+ },
62
+ {
63
+ "name": "NVIDIA H200",
64
+ "memoryTotal": "150754820096",
65
+ "cudaCores": 16896,
66
+ "architecture": "Hopper",
67
+ "uuid": "GPU-08b37f98-4603-d483-2f2b-fe5311aa42f2"
68
+ },
69
+ {
70
+ "name": "NVIDIA H200",
71
+ "memoryTotal": "150754820096",
72
+ "cudaCores": 16896,
73
+ "architecture": "Hopper",
74
+ "uuid": "GPU-03273b5b-2fdd-a5fe-4460-c897334ae464"
75
+ },
76
+ {
77
+ "name": "NVIDIA H200",
78
+ "memoryTotal": "150754820096",
79
+ "cudaCores": 16896,
80
+ "architecture": "Hopper",
81
+ "uuid": "GPU-292d466c-d00d-25a4-28b6-e6c978d3e70c"
82
+ },
83
+ {
84
+ "name": "NVIDIA H200",
85
+ "memoryTotal": "150754820096",
86
+ "cudaCores": 16896,
87
+ "architecture": "Hopper",
88
+ "uuid": "GPU-46f38561-3148-e442-7f7f-bfe447bab7fe"
89
+ }
90
+ ],
91
+ "cudaVersion": "13.0",
92
+ "writerId": "0fg4ri1obzq9oqbw2tt47dd2pwoxs7dt"
93
+ }
ABLATION_0302_noTgtAlign/wandb/run-20260303_173555-et94mfhp/logs/debug-core.log ADDED
@@ -0,0 +1,84 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-03-03T17:35:55.52462423Z","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpsb6khdpj/port-870906.txt","pid":870906,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2026-03-03T17:35:55.525474789Z","level":"INFO","msg":"server: will exit if parent process dies","ppid":870906}
3
+ {"time":"2026-03-03T17:35:55.525442248Z","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-870906-873679-3545240817/socket","Net":"unix"}}
4
+ {"time":"2026-03-03T17:35:55.698072141Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2026-03-03T17:35:55.708276033Z","level":"INFO","msg":"handleInformInit: received","streamId":"et94mfhp","id":"1(@)"}
6
+ {"time":"2026-03-03T17:35:56.156799869Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"et94mfhp","id":"1(@)"}
7
+ {"time":"2026-03-03T17:36:01.795021213Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"f0yqferh89g5"}
8
+ {"time":"2026-03-03T18:31:28.268666189Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"2(@)"}
9
+ {"time":"2026-03-03T18:31:28.268906962Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"3(@)"}
10
+ {"time":"2026-03-03T18:31:28.269258716Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"4(@)"}
11
+ {"time":"2026-03-03T18:31:28.269886062Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"5(@)"}
12
+ {"time":"2026-03-03T18:31:28.269887682Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"6(@)"}
13
+ {"time":"2026-03-03T18:31:28.269957393Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"7(@)"}
14
+ {"time":"2026-03-03T18:31:28.271280458Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"8(@)"}
15
+ {"time":"2026-03-03T18:31:28.271412419Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"9(@)"}
16
+ {"time":"2026-03-03T18:31:28.271435899Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"10(@)"}
17
+ {"time":"2026-03-03T18:31:28.271430169Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"11(@)"}
18
+ {"time":"2026-03-03T18:31:28.272139377Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"12(@)"}
19
+ {"time":"2026-03-03T18:31:28.272176878Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"13(@)"}
20
+ {"time":"2026-03-03T18:31:28.274330311Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"15(@)"}
21
+ {"time":"2026-03-03T18:31:28.274330501Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"14(@)"}
22
+ {"time":"2026-03-03T18:31:28.2760356Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"16(@)"}
23
+ {"time":"2026-03-03T18:31:28.278878101Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"17(@)"}
24
+ {"time":"2026-03-03T18:31:28.279916192Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"18(@)"}
25
+ {"time":"2026-03-03T18:31:30.04483325Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"f0yqferh89g5"}
26
+ {"time":"2026-03-03T18:31:30.505041834Z","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
27
+ {"time":"2026-03-03T18:31:30.505122885Z","level":"INFO","msg":"connection: closing","id":"4(@)"}
28
+ {"time":"2026-03-03T18:31:30.505124115Z","level":"INFO","msg":"connection: closing","id":"9(@)"}
29
+ {"time":"2026-03-03T18:31:30.505198555Z","level":"INFO","msg":"connection: closing","id":"5(@)"}
30
+ {"time":"2026-03-03T18:31:30.505185325Z","level":"INFO","msg":"connection: closing","id":"15(@)"}
31
+ {"time":"2026-03-03T18:31:30.505246616Z","level":"INFO","msg":"server is shutting down"}
32
+ {"time":"2026-03-03T18:31:30.505241016Z","level":"INFO","msg":"connection: closed successfully","id":"4(@)"}
33
+ {"time":"2026-03-03T18:31:30.505309627Z","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-870906-873679-3545240817/socket","Net":"unix"}}
34
+ {"time":"2026-03-03T18:31:30.505236746Z","level":"INFO","msg":"connection: closed successfully","id":"5(@)"}
35
+ {"time":"2026-03-03T18:31:30.505332537Z","level":"INFO","msg":"connection: closing","id":"3(@)"}
36
+ {"time":"2026-03-03T18:31:30.505339877Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"5(@)"}
37
+ {"time":"2026-03-03T18:31:30.505333437Z","level":"INFO","msg":"connection: closing","id":"11(@)"}
38
+ {"time":"2026-03-03T18:31:30.505353487Z","level":"INFO","msg":"connection: closing","id":"1(@)"}
39
+ {"time":"2026-03-03T18:31:30.505331007Z","level":"INFO","msg":"connection: closing","id":"16(@)"}
40
+ {"time":"2026-03-03T18:31:30.505438568Z","level":"INFO","msg":"connection: closed successfully","id":"16(@)"}
41
+ {"time":"2026-03-03T18:31:30.505444188Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"16(@)"}
42
+ {"time":"2026-03-03T18:31:30.505324557Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"4(@)"}
43
+ {"time":"2026-03-03T18:31:30.505336057Z","level":"INFO","msg":"connection: closing","id":"10(@)"}
44
+ {"time":"2026-03-03T18:31:30.505257996Z","level":"INFO","msg":"connection: closed successfully","id":"9(@)"}
45
+ {"time":"2026-03-03T18:31:30.505478358Z","level":"INFO","msg":"connection: closed successfully","id":"10(@)"}
46
+ {"time":"2026-03-03T18:31:30.505483569Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"10(@)"}
47
+ {"time":"2026-03-03T18:31:30.505482539Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"9(@)"}
48
+ {"time":"2026-03-03T18:31:30.505192215Z","level":"INFO","msg":"connection: closing","id":"6(@)"}
49
+ {"time":"2026-03-03T18:31:30.505517929Z","level":"INFO","msg":"connection: closed successfully","id":"6(@)"}
50
+ {"time":"2026-03-03T18:31:30.505522309Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"6(@)"}
51
+ {"time":"2026-03-03T18:31:30.505179335Z","level":"INFO","msg":"connection: closing","id":"8(@)"}
52
+ {"time":"2026-03-03T18:31:30.50558072Z","level":"INFO","msg":"connection: closed successfully","id":"8(@)"}
53
+ {"time":"2026-03-03T18:31:30.50559922Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"8(@)"}
54
+ {"time":"2026-03-03T18:31:30.505286956Z","level":"INFO","msg":"connection: closed successfully","id":"15(@)"}
55
+ {"time":"2026-03-03T18:31:30.50560804Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"15(@)"}
56
+ {"time":"2026-03-03T18:31:30.505268996Z","level":"INFO","msg":"connection: closing","id":"13(@)"}
57
+ {"time":"2026-03-03T18:31:30.505292706Z","level":"INFO","msg":"connection: closing","id":"18(@)"}
58
+ {"time":"2026-03-03T18:31:30.505681601Z","level":"INFO","msg":"connection: closed successfully","id":"18(@)"}
59
+ {"time":"2026-03-03T18:31:30.505692061Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"18(@)"}
60
+ {"time":"2026-03-03T18:31:30.505302067Z","level":"INFO","msg":"connection: closing","id":"17(@)"}
61
+ {"time":"2026-03-03T18:31:30.505310757Z","level":"INFO","msg":"connection: closing","id":"14(@)"}
62
+ {"time":"2026-03-03T18:31:30.505726091Z","level":"INFO","msg":"connection: closed successfully","id":"17(@)"}
63
+ {"time":"2026-03-03T18:31:30.505736691Z","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
64
+ {"time":"2026-03-03T18:31:30.505745891Z","level":"INFO","msg":"connection: closed successfully","id":"14(@)"}
65
+ {"time":"2026-03-03T18:31:30.505738021Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"17(@)"}
66
+ {"time":"2026-03-03T18:31:30.505355827Z","level":"INFO","msg":"connection: closed successfully","id":"3(@)"}
67
+ {"time":"2026-03-03T18:31:30.505769532Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"3(@)"}
68
+ {"time":"2026-03-03T18:31:30.505359617Z","level":"INFO","msg":"connection: closing","id":"2(@)"}
69
+ {"time":"2026-03-03T18:31:30.505818722Z","level":"INFO","msg":"connection: closed successfully","id":"2(@)"}
70
+ {"time":"2026-03-03T18:31:30.505823912Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"2(@)"}
71
+ {"time":"2026-03-03T18:31:30.505393308Z","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
72
+ {"time":"2026-03-03T18:31:30.505832472Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
73
+ {"time":"2026-03-03T18:31:30.505389887Z","level":"INFO","msg":"connection: closed successfully","id":"11(@)"}
74
+ {"time":"2026-03-03T18:31:30.505841602Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"11(@)"}
75
+ {"time":"2026-03-03T18:31:30.505186255Z","level":"INFO","msg":"connection: closing","id":"7(@)"}
76
+ {"time":"2026-03-03T18:31:30.50563443Z","level":"INFO","msg":"connection: closed successfully","id":"13(@)"}
77
+ {"time":"2026-03-03T18:31:30.505871433Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"13(@)"}
78
+ {"time":"2026-03-03T18:31:30.505295396Z","level":"INFO","msg":"connection: closing","id":"12(@)"}
79
+ {"time":"2026-03-03T18:31:30.505752631Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"14(@)"}
80
+ {"time":"2026-03-03T18:31:30.505865483Z","level":"INFO","msg":"connection: closed successfully","id":"7(@)"}
81
+ {"time":"2026-03-03T18:31:30.505911503Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"7(@)"}
82
+ {"time":"2026-03-03T18:31:30.505896973Z","level":"INFO","msg":"connection: closed successfully","id":"12(@)"}
83
+ {"time":"2026-03-03T18:31:30.505921183Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"12(@)"}
84
+ {"time":"2026-03-03T18:31:30.505930643Z","level":"INFO","msg":"server is closed"}
ABLATION_0302_noTgtAlign/wandb/run-20260303_173555-et94mfhp/logs/debug.log ADDED
The diff for this file is too large to render. See raw diff