Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes. See raw diff
- .gitattributes +1 -0
- ABLATION_0302_FreqSelect/checkpoints/epoch_0-step_1500.ckpt +3 -0
- ABLATION_0302_FreqSelect/wandb/run-20260302_164522-8h1p5i95/files/media/images/comparison_0_f9d64709298973e4dd8b.png +3 -0
- ABLATION_0302_FreqSelect/wandb/run-20260302_164522-8h1p5i95/run-8h1p5i95.wandb +3 -0
- ABLATION_0302_noAux/.hydra/config.yaml +188 -0
- ABLATION_0302_noAux/.hydra/hydra.yaml +165 -0
- ABLATION_0302_noAux/.hydra/overrides.yaml +4 -0
- ABLATION_0302_noAux/wandb/debug-internal.log +11 -0
- ABLATION_0302_noAux/wandb/debug.log +21 -0
- ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/config.yaml +310 -0
- ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/active_mask_imgs_168_43fa18ddb9c77705edd2.png +3 -0
- ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/active_mask_imgs_92_3031367680d80379ffab.png +3 -0
- ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/comparison_167_3dfd8c9e7e3d980df015.png +3 -0
- ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/comparison_182_5c2b90f763b7bff88204.png +3 -0
- ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/error_scores_169_b17e3324e4e0f9205834.png +3 -0
- ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/output.log +0 -0
- ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/requirements.txt +173 -0
- ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/wandb-metadata.json +93 -0
- ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/wandb-summary.json +1 -0
- ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug-core.log +15 -0
- ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug-internal.log +11 -0
- ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug.log +21 -0
- ABLATION_0302_noTgtAlign/.hydra/config.yaml +188 -0
- ABLATION_0302_noTgtAlign/.hydra/hydra.yaml +165 -0
- ABLATION_0302_noTgtAlign/.hydra/overrides.yaml +4 -0
- ABLATION_0302_noTgtAlign/main.log +83 -0
- ABLATION_0302_noTgtAlign/train_ddp_process_1.log +44 -0
- ABLATION_0302_noTgtAlign/train_ddp_process_2.log +44 -0
- ABLATION_0302_noTgtAlign/train_ddp_process_3.log +44 -0
- ABLATION_0302_noTgtAlign/train_ddp_process_4.log +44 -0
- ABLATION_0302_noTgtAlign/train_ddp_process_5.log +44 -0
- ABLATION_0302_noTgtAlign/train_ddp_process_6.log +44 -0
- ABLATION_0302_noTgtAlign/train_ddp_process_7.log +44 -0
- ABLATION_0302_noTgtAlign/wandb/debug-internal.log +6 -0
- ABLATION_0302_noTgtAlign/wandb/debug.log +19 -0
- ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/config.yaml +311 -0
- ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/output.log +72 -0
- ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/requirements.txt +173 -0
- ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/wandb-metadata.json +93 -0
- ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/wandb-summary.json +1 -0
- ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/logs/debug-core.log +19 -0
- ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/logs/debug-internal.log +12 -0
- ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/logs/debug.log +24 -0
- ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/run-beokhgg2.wandb +0 -0
- ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/config.yaml +311 -0
- ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/output.log +117 -0
- ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/requirements.txt +173 -0
- ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/wandb-metadata.json +93 -0
- ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/wandb-summary.json +1 -0
- ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/logs/debug-core.log +15 -0
.gitattributes
CHANGED
|
@@ -58,3 +58,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 58 |
# Video files - compressed
|
| 59 |
*.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 60 |
*.webm filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 58 |
# Video files - compressed
|
| 59 |
*.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 60 |
*.webm filter=lfs diff=lfs merge=lfs -text
|
| 61 |
+
ABLATION_0302_FreqSelect/wandb/run-20260302_164522-8h1p5i95/run-8h1p5i95.wandb filter=lfs diff=lfs merge=lfs -text
|
ABLATION_0302_FreqSelect/checkpoints/epoch_0-step_1500.ckpt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9905f9951634363d9e0582b2282eaacefe4f1b93637184b1fd0b8984ee9d77fd
|
| 3 |
+
size 11880857960
|
ABLATION_0302_FreqSelect/wandb/run-20260302_164522-8h1p5i95/files/media/images/comparison_0_f9d64709298973e4dd8b.png
ADDED
|
Git LFS Details
|
ABLATION_0302_FreqSelect/wandb/run-20260302_164522-8h1p5i95/run-8h1p5i95.wandb
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:97304ca2cc941f3f04e6359c6f0e44f3f0af7ad6c82c5fda66a78a456d1bd836
|
| 3 |
+
size 4279158
|
ABLATION_0302_noAux/.hydra/config.yaml
ADDED
|
@@ -0,0 +1,188 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
encoder:
|
| 3 |
+
name: dcsplat
|
| 4 |
+
input_image_shape:
|
| 5 |
+
- 518
|
| 6 |
+
- 518
|
| 7 |
+
head_mode: pcd
|
| 8 |
+
num_level: 3
|
| 9 |
+
gs_param_dim: 256
|
| 10 |
+
align_corners: false
|
| 11 |
+
use_voxelize: true
|
| 12 |
+
decoder:
|
| 13 |
+
name: splatting_cuda
|
| 14 |
+
background_color:
|
| 15 |
+
- 0.0
|
| 16 |
+
- 0.0
|
| 17 |
+
- 0.0
|
| 18 |
+
make_scale_invariant: false
|
| 19 |
+
density_control:
|
| 20 |
+
name: density_control_module
|
| 21 |
+
mean_dim: 32
|
| 22 |
+
gs_param_dim: 256
|
| 23 |
+
refinement_layer_num: 1
|
| 24 |
+
num_level: 3
|
| 25 |
+
grad_mode: absgrad
|
| 26 |
+
use_mean_features: true
|
| 27 |
+
refinement_type: voxelize
|
| 28 |
+
refinement_hidden_dim: 32
|
| 29 |
+
aggregation_mode: mean
|
| 30 |
+
num_heads: 1
|
| 31 |
+
score_mode: absgrad
|
| 32 |
+
latent_dim: 128
|
| 33 |
+
num_latents: 64
|
| 34 |
+
num_self_attn_per_block: 2
|
| 35 |
+
voxel_size: 0.001
|
| 36 |
+
aux_refine: false
|
| 37 |
+
refine_error: false
|
| 38 |
+
use_refine_module: false
|
| 39 |
+
voxelize_activate: false
|
| 40 |
+
use_depth: false
|
| 41 |
+
render_loss:
|
| 42 |
+
mse:
|
| 43 |
+
weight: 1.0
|
| 44 |
+
lpips:
|
| 45 |
+
weight: 0.05
|
| 46 |
+
apply_after_step: 0
|
| 47 |
+
density_control_loss:
|
| 48 |
+
error_score:
|
| 49 |
+
weight: 0.0001
|
| 50 |
+
log_scale: false
|
| 51 |
+
grad_scale: 10000.0
|
| 52 |
+
mode: original
|
| 53 |
+
direct_loss:
|
| 54 |
+
l1:
|
| 55 |
+
weight: 0.8
|
| 56 |
+
ssim:
|
| 57 |
+
weight: 0.2
|
| 58 |
+
wandb:
|
| 59 |
+
project: DCSplat
|
| 60 |
+
entity: scene-representation-group
|
| 61 |
+
name: ABLATION_0302_noAux
|
| 62 |
+
mode: online
|
| 63 |
+
tags:
|
| 64 |
+
- re10k
|
| 65 |
+
- 256x256
|
| 66 |
+
mode: train
|
| 67 |
+
data_loader:
|
| 68 |
+
train:
|
| 69 |
+
num_workers: 16
|
| 70 |
+
persistent_workers: true
|
| 71 |
+
batch_size: 16
|
| 72 |
+
seed: 1234
|
| 73 |
+
test:
|
| 74 |
+
num_workers: 4
|
| 75 |
+
persistent_workers: false
|
| 76 |
+
batch_size: 1
|
| 77 |
+
seed: 2345
|
| 78 |
+
val:
|
| 79 |
+
num_workers: 1
|
| 80 |
+
persistent_workers: true
|
| 81 |
+
batch_size: 1
|
| 82 |
+
seed: 3456
|
| 83 |
+
optimizer:
|
| 84 |
+
lr: 0.0002
|
| 85 |
+
warm_up_steps: 25
|
| 86 |
+
backbone_lr_multiplier: 0.1
|
| 87 |
+
backbone_trainable: T+H
|
| 88 |
+
accumulate: 1
|
| 89 |
+
checkpointing:
|
| 90 |
+
load: null
|
| 91 |
+
every_n_train_steps: 1500
|
| 92 |
+
save_top_k: 2
|
| 93 |
+
save_weights_only: false
|
| 94 |
+
train:
|
| 95 |
+
extended_visualization: false
|
| 96 |
+
print_log_every_n_steps: 10
|
| 97 |
+
camera_loss: 10.0
|
| 98 |
+
one_sample_validation: null
|
| 99 |
+
align_corners: false
|
| 100 |
+
intrinsic_scaling: false
|
| 101 |
+
verbose: false
|
| 102 |
+
beta_dist_param:
|
| 103 |
+
- 0.5
|
| 104 |
+
- 4.0
|
| 105 |
+
use_refine_aux: false
|
| 106 |
+
train_target_set: true
|
| 107 |
+
train_gs_num: 1
|
| 108 |
+
ext_scale_detach: false
|
| 109 |
+
cam_scale_mode: sum
|
| 110 |
+
scene_scale_reg_loss: 0.01
|
| 111 |
+
train_aux: false
|
| 112 |
+
vggt_cam_loss: true
|
| 113 |
+
vggt_distil: false
|
| 114 |
+
context_view_train: false
|
| 115 |
+
test:
|
| 116 |
+
output_path: test/ablation/re10k
|
| 117 |
+
align_pose: false
|
| 118 |
+
pose_align_steps: 100
|
| 119 |
+
rot_opt_lr: 0.005
|
| 120 |
+
trans_opt_lr: 0.005
|
| 121 |
+
compute_scores: true
|
| 122 |
+
save_image: false
|
| 123 |
+
save_video: false
|
| 124 |
+
save_active_mask_image: false
|
| 125 |
+
save_error_score_image: false
|
| 126 |
+
save_compare: false
|
| 127 |
+
save_gs: false
|
| 128 |
+
save_sample_wise_metrics: true
|
| 129 |
+
pred_intrinsic: false
|
| 130 |
+
error_threshold: 0.4
|
| 131 |
+
error_threshold_list:
|
| 132 |
+
- 0.2
|
| 133 |
+
- 0.4
|
| 134 |
+
- 0.6
|
| 135 |
+
- 0.8
|
| 136 |
+
- 1.0
|
| 137 |
+
threshold_mode: ratio
|
| 138 |
+
nvs_view_N_list:
|
| 139 |
+
- 3
|
| 140 |
+
- 6
|
| 141 |
+
- 16
|
| 142 |
+
- 32
|
| 143 |
+
- 64
|
| 144 |
+
seed: 111123
|
| 145 |
+
trainer:
|
| 146 |
+
max_steps: 3001
|
| 147 |
+
val_check_interval: 250
|
| 148 |
+
gradient_clip_val: 0.5
|
| 149 |
+
num_nodes: 1
|
| 150 |
+
dataset:
|
| 151 |
+
re10k:
|
| 152 |
+
make_baseline_1: true
|
| 153 |
+
relative_pose: true
|
| 154 |
+
augment: true
|
| 155 |
+
background_color:
|
| 156 |
+
- 0.0
|
| 157 |
+
- 0.0
|
| 158 |
+
- 0.0
|
| 159 |
+
overfit_to_scene: null
|
| 160 |
+
skip_bad_shape: true
|
| 161 |
+
view_sampler:
|
| 162 |
+
name: bounded
|
| 163 |
+
num_target_views: 4
|
| 164 |
+
num_context_views: 2
|
| 165 |
+
min_distance_between_context_views: 45
|
| 166 |
+
max_distance_between_context_views: 90
|
| 167 |
+
min_distance_to_context_views: 0
|
| 168 |
+
warm_up_steps: 1000
|
| 169 |
+
initial_min_distance_between_context_views: 25
|
| 170 |
+
initial_max_distance_between_context_views: 25
|
| 171 |
+
same_target_gap: false
|
| 172 |
+
num_target_set: 3
|
| 173 |
+
target_align: true
|
| 174 |
+
name: re10k
|
| 175 |
+
roots:
|
| 176 |
+
- datasets/re10k
|
| 177 |
+
input_image_shape:
|
| 178 |
+
- 256
|
| 179 |
+
- 256
|
| 180 |
+
original_image_shape:
|
| 181 |
+
- 360
|
| 182 |
+
- 640
|
| 183 |
+
cameras_are_circular: false
|
| 184 |
+
baseline_min: 0.001
|
| 185 |
+
baseline_max: 10000000000.0
|
| 186 |
+
max_fov: 100.0
|
| 187 |
+
dynamic_context_views: true
|
| 188 |
+
max_context_views_per_gpu: 24
|
ABLATION_0302_noAux/.hydra/hydra.yaml
ADDED
|
@@ -0,0 +1,165 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
hydra:
|
| 2 |
+
run:
|
| 3 |
+
dir: outputs/ablation/re10k/${wandb.name}
|
| 4 |
+
sweep:
|
| 5 |
+
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
|
| 6 |
+
subdir: ${hydra.job.num}
|
| 7 |
+
launcher:
|
| 8 |
+
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
|
| 9 |
+
sweeper:
|
| 10 |
+
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
|
| 11 |
+
max_batch_size: null
|
| 12 |
+
params: null
|
| 13 |
+
help:
|
| 14 |
+
app_name: ${hydra.job.name}
|
| 15 |
+
header: '${hydra.help.app_name} is powered by Hydra.
|
| 16 |
+
|
| 17 |
+
'
|
| 18 |
+
footer: 'Powered by Hydra (https://hydra.cc)
|
| 19 |
+
|
| 20 |
+
Use --hydra-help to view Hydra specific help
|
| 21 |
+
|
| 22 |
+
'
|
| 23 |
+
template: '${hydra.help.header}
|
| 24 |
+
|
| 25 |
+
== Configuration groups ==
|
| 26 |
+
|
| 27 |
+
Compose your configuration from those groups (group=option)
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
$APP_CONFIG_GROUPS
|
| 31 |
+
|
| 32 |
+
|
| 33 |
+
== Config ==
|
| 34 |
+
|
| 35 |
+
Override anything in the config (foo.bar=value)
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
$CONFIG
|
| 39 |
+
|
| 40 |
+
|
| 41 |
+
${hydra.help.footer}
|
| 42 |
+
|
| 43 |
+
'
|
| 44 |
+
hydra_help:
|
| 45 |
+
template: 'Hydra (${hydra.runtime.version})
|
| 46 |
+
|
| 47 |
+
See https://hydra.cc for more info.
|
| 48 |
+
|
| 49 |
+
|
| 50 |
+
== Flags ==
|
| 51 |
+
|
| 52 |
+
$FLAGS_HELP
|
| 53 |
+
|
| 54 |
+
|
| 55 |
+
== Configuration groups ==
|
| 56 |
+
|
| 57 |
+
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
|
| 58 |
+
to command line)
|
| 59 |
+
|
| 60 |
+
|
| 61 |
+
$HYDRA_CONFIG_GROUPS
|
| 62 |
+
|
| 63 |
+
|
| 64 |
+
Use ''--cfg hydra'' to Show the Hydra config.
|
| 65 |
+
|
| 66 |
+
'
|
| 67 |
+
hydra_help: ???
|
| 68 |
+
hydra_logging:
|
| 69 |
+
version: 1
|
| 70 |
+
formatters:
|
| 71 |
+
simple:
|
| 72 |
+
format: '[%(asctime)s][HYDRA] %(message)s'
|
| 73 |
+
handlers:
|
| 74 |
+
console:
|
| 75 |
+
class: logging.StreamHandler
|
| 76 |
+
formatter: simple
|
| 77 |
+
stream: ext://sys.stdout
|
| 78 |
+
root:
|
| 79 |
+
level: INFO
|
| 80 |
+
handlers:
|
| 81 |
+
- console
|
| 82 |
+
loggers:
|
| 83 |
+
logging_example:
|
| 84 |
+
level: DEBUG
|
| 85 |
+
disable_existing_loggers: false
|
| 86 |
+
job_logging:
|
| 87 |
+
version: 1
|
| 88 |
+
formatters:
|
| 89 |
+
simple:
|
| 90 |
+
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
|
| 91 |
+
handlers:
|
| 92 |
+
console:
|
| 93 |
+
class: logging.StreamHandler
|
| 94 |
+
formatter: simple
|
| 95 |
+
stream: ext://sys.stdout
|
| 96 |
+
file:
|
| 97 |
+
class: logging.FileHandler
|
| 98 |
+
formatter: simple
|
| 99 |
+
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
|
| 100 |
+
root:
|
| 101 |
+
level: INFO
|
| 102 |
+
handlers:
|
| 103 |
+
- console
|
| 104 |
+
- file
|
| 105 |
+
disable_existing_loggers: false
|
| 106 |
+
env: {}
|
| 107 |
+
mode: RUN
|
| 108 |
+
searchpath: []
|
| 109 |
+
callbacks: {}
|
| 110 |
+
output_subdir: .hydra
|
| 111 |
+
overrides:
|
| 112 |
+
hydra:
|
| 113 |
+
- hydra.mode=RUN
|
| 114 |
+
task:
|
| 115 |
+
- +experiment=re10k_ablation_24v
|
| 116 |
+
- wandb.mode=online
|
| 117 |
+
- wandb.name=ABLATION_0302_noAux
|
| 118 |
+
- train.train_aux=false
|
| 119 |
+
job:
|
| 120 |
+
name: main
|
| 121 |
+
chdir: null
|
| 122 |
+
override_dirname: +experiment=re10k_ablation_24v,train.train_aux=false,wandb.mode=online,wandb.name=ABLATION_0302_noAux
|
| 123 |
+
id: ???
|
| 124 |
+
num: ???
|
| 125 |
+
config_name: main
|
| 126 |
+
env_set: {}
|
| 127 |
+
env_copy: []
|
| 128 |
+
config:
|
| 129 |
+
override_dirname:
|
| 130 |
+
kv_sep: '='
|
| 131 |
+
item_sep: ','
|
| 132 |
+
exclude_keys: []
|
| 133 |
+
runtime:
|
| 134 |
+
version: 1.3.2
|
| 135 |
+
version_base: '1.3'
|
| 136 |
+
cwd: /workspace/code/CVPR2026
|
| 137 |
+
config_sources:
|
| 138 |
+
- path: hydra.conf
|
| 139 |
+
schema: pkg
|
| 140 |
+
provider: hydra
|
| 141 |
+
- path: /workspace/code/CVPR2026/config
|
| 142 |
+
schema: file
|
| 143 |
+
provider: main
|
| 144 |
+
- path: ''
|
| 145 |
+
schema: structured
|
| 146 |
+
provider: schema
|
| 147 |
+
output_dir: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noAux
|
| 148 |
+
choices:
|
| 149 |
+
experiment: re10k_ablation_24v
|
| 150 |
+
dataset@dataset.re10k: re10k
|
| 151 |
+
dataset/view_sampler_dataset_specific_config@dataset.re10k.view_sampler: bounded_re10k
|
| 152 |
+
dataset/view_sampler@dataset.re10k.view_sampler: bounded
|
| 153 |
+
model/density_control: density_control_module
|
| 154 |
+
model/decoder: splatting_cuda
|
| 155 |
+
model/encoder: dcsplat
|
| 156 |
+
hydra/env: default
|
| 157 |
+
hydra/callbacks: null
|
| 158 |
+
hydra/job_logging: default
|
| 159 |
+
hydra/hydra_logging: default
|
| 160 |
+
hydra/hydra_help: default
|
| 161 |
+
hydra/help: default
|
| 162 |
+
hydra/sweeper: basic
|
| 163 |
+
hydra/launcher: basic
|
| 164 |
+
hydra/output: default
|
| 165 |
+
verbose: false
|
ABLATION_0302_noAux/.hydra/overrides.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
- +experiment=re10k_ablation_24v
|
| 2 |
+
- wandb.mode=online
|
| 3 |
+
- wandb.name=ABLATION_0302_noAux
|
| 4 |
+
- train.train_aux=false
|
ABLATION_0302_noAux/wandb/debug-internal.log
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2026-03-03T14:34:53.54334025Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
|
| 2 |
+
{"time":"2026-03-03T14:34:53.958050168Z","level":"INFO","msg":"stream: created new stream","id":"l8o7jl7r"}
|
| 3 |
+
{"time":"2026-03-03T14:34:53.958155019Z","level":"INFO","msg":"handler: started","stream_id":"l8o7jl7r"}
|
| 4 |
+
{"time":"2026-03-03T14:34:53.9583015Z","level":"INFO","msg":"stream: started","id":"l8o7jl7r"}
|
| 5 |
+
{"time":"2026-03-03T14:34:53.95834932Z","level":"INFO","msg":"sender: started","stream_id":"l8o7jl7r"}
|
| 6 |
+
{"time":"2026-03-03T14:34:53.95835085Z","level":"INFO","msg":"writer: started","stream_id":"l8o7jl7r"}
|
| 7 |
+
{"time":"2026-03-03T17:25:37.368615437Z","level":"INFO","msg":"stream: closing","id":"l8o7jl7r"}
|
| 8 |
+
{"time":"2026-03-03T17:25:38.278996749Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
|
| 9 |
+
{"time":"2026-03-03T17:25:38.528169562Z","level":"INFO","msg":"handler: closed","stream_id":"l8o7jl7r"}
|
| 10 |
+
{"time":"2026-03-03T17:25:38.528336264Z","level":"INFO","msg":"sender: closed","stream_id":"l8o7jl7r"}
|
| 11 |
+
{"time":"2026-03-03T17:25:38.528358534Z","level":"INFO","msg":"stream: closed","id":"l8o7jl7r"}
|
ABLATION_0302_noAux/wandb/debug.log
ADDED
|
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
|
| 2 |
+
2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_setup.py:_flush():81] Configure stats pid to 854556
|
| 3 |
+
2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_setup.py:_flush():81] Loading settings from environment variables
|
| 4 |
+
2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug.log
|
| 5 |
+
2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug-internal.log
|
| 6 |
+
2026-03-03 14:34:53,294 INFO MainThread:854556 [wandb_init.py:init():844] calling init triggers
|
| 7 |
+
2026-03-03 14:34:53,294 INFO MainThread:854556 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
|
| 8 |
+
config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'pcd', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'absgrad', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': False, 'voxelize_activate': False, 'use_depth': False}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'density_control_loss': {'error_score': {'weight': 0.0001, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0302_noAux', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': False, 'vggt_cam_loss': True, 'vggt_distil': False, 'context_view_train': False}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'save_gs': False, 'save_sample_wise_metrics': True, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3, 'target_align': True}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
|
| 9 |
+
2026-03-03 14:34:53,294 INFO MainThread:854556 [wandb_init.py:init():892] starting backend
|
| 10 |
+
2026-03-03 14:34:53,535 INFO MainThread:854556 [wandb_init.py:init():895] sending inform_init request
|
| 11 |
+
2026-03-03 14:34:53,540 INFO MainThread:854556 [wandb_init.py:init():903] backend started and connected
|
| 12 |
+
2026-03-03 14:34:53,542 INFO MainThread:854556 [wandb_init.py:init():973] updated telemetry
|
| 13 |
+
2026-03-03 14:34:53,548 INFO MainThread:854556 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
|
| 14 |
+
2026-03-03 14:34:54,767 INFO MainThread:854556 [wandb_init.py:init():1042] starting run threads in backend
|
| 15 |
+
2026-03-03 14:34:54,895 INFO MainThread:854556 [wandb_run.py:_console_start():2524] atexit reg
|
| 16 |
+
2026-03-03 14:34:54,895 INFO MainThread:854556 [wandb_run.py:_redirect():2373] redirect: wrap_raw
|
| 17 |
+
2026-03-03 14:34:54,895 INFO MainThread:854556 [wandb_run.py:_redirect():2442] Wrapping output streams.
|
| 18 |
+
2026-03-03 14:34:54,896 INFO MainThread:854556 [wandb_run.py:_redirect():2465] Redirects installed.
|
| 19 |
+
2026-03-03 14:34:54,899 INFO MainThread:854556 [wandb_init.py:init():1082] run started, returning control to user process
|
| 20 |
+
2026-03-03 17:25:37,368 INFO wandb-AsyncioManager-main:854556 [service_client.py:_forward_responses():134] Reached EOF.
|
| 21 |
+
2026-03-03 17:25:37,368 INFO wandb-AsyncioManager-main:854556 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.
|
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/config.yaml
ADDED
|
@@ -0,0 +1,310 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
_wandb:
|
| 2 |
+
value:
|
| 3 |
+
cli_version: 0.25.0
|
| 4 |
+
e:
|
| 5 |
+
ycd5umfcq5p6rih4lqfm265f2qblrtrn:
|
| 6 |
+
args:
|
| 7 |
+
- +experiment=re10k_ablation_24v
|
| 8 |
+
- wandb.mode=online
|
| 9 |
+
- wandb.name=ABLATION_0302_noAux
|
| 10 |
+
- train.train_aux=false
|
| 11 |
+
cpu_count: 128
|
| 12 |
+
cpu_count_logical: 256
|
| 13 |
+
cudaVersion: "13.0"
|
| 14 |
+
disk:
|
| 15 |
+
/:
|
| 16 |
+
total: "735513149440"
|
| 17 |
+
used: "676972425216"
|
| 18 |
+
email: dna9041@korea.ac.kr
|
| 19 |
+
executable: /venv/main/bin/python
|
| 20 |
+
git:
|
| 21 |
+
commit: 9dfce172a0f8c7ce85e763899f7ef741ecffc454
|
| 22 |
+
remote: git@github.com:K-nowing/CVPR2026.git
|
| 23 |
+
gpu: NVIDIA H200
|
| 24 |
+
gpu_count: 8
|
| 25 |
+
gpu_nvidia:
|
| 26 |
+
- architecture: Hopper
|
| 27 |
+
cudaCores: 16896
|
| 28 |
+
memoryTotal: "150754820096"
|
| 29 |
+
name: NVIDIA H200
|
| 30 |
+
uuid: GPU-9a20101e-d876-facd-5f05-805081aede41
|
| 31 |
+
- architecture: Hopper
|
| 32 |
+
cudaCores: 16896
|
| 33 |
+
memoryTotal: "150754820096"
|
| 34 |
+
name: NVIDIA H200
|
| 35 |
+
uuid: GPU-84736a77-ee75-3324-e4e1-99cc15bfb5e9
|
| 36 |
+
- architecture: Hopper
|
| 37 |
+
cudaCores: 16896
|
| 38 |
+
memoryTotal: "150754820096"
|
| 39 |
+
name: NVIDIA H200
|
| 40 |
+
uuid: GPU-423d3161-cdc4-3fc0-caee-d15cfaa83ca6
|
| 41 |
+
- architecture: Hopper
|
| 42 |
+
cudaCores: 16896
|
| 43 |
+
memoryTotal: "150754820096"
|
| 44 |
+
name: NVIDIA H200
|
| 45 |
+
uuid: GPU-5b0058b2-cdb9-c952-04f9-87dcaa7ea742
|
| 46 |
+
- architecture: Hopper
|
| 47 |
+
cudaCores: 16896
|
| 48 |
+
memoryTotal: "150754820096"
|
| 49 |
+
name: NVIDIA H200
|
| 50 |
+
uuid: GPU-08b37f98-4603-d483-2f2b-fe5311aa42f2
|
| 51 |
+
- architecture: Hopper
|
| 52 |
+
cudaCores: 16896
|
| 53 |
+
memoryTotal: "150754820096"
|
| 54 |
+
name: NVIDIA H200
|
| 55 |
+
uuid: GPU-03273b5b-2fdd-a5fe-4460-c897334ae464
|
| 56 |
+
- architecture: Hopper
|
| 57 |
+
cudaCores: 16896
|
| 58 |
+
memoryTotal: "150754820096"
|
| 59 |
+
name: NVIDIA H200
|
| 60 |
+
uuid: GPU-292d466c-d00d-25a4-28b6-e6c978d3e70c
|
| 61 |
+
- architecture: Hopper
|
| 62 |
+
cudaCores: 16896
|
| 63 |
+
memoryTotal: "150754820096"
|
| 64 |
+
name: NVIDIA H200
|
| 65 |
+
uuid: GPU-46f38561-3148-e442-7f7f-bfe447bab7fe
|
| 66 |
+
host: e9d3310a05da
|
| 67 |
+
memory:
|
| 68 |
+
total: "1622950240256"
|
| 69 |
+
os: Linux-6.8.0-94-generic-x86_64-with-glibc2.39
|
| 70 |
+
program: -m src.main
|
| 71 |
+
python: CPython 3.12.12
|
| 72 |
+
root: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noAux
|
| 73 |
+
startedAt: "2026-03-03T14:34:53.292304Z"
|
| 74 |
+
writerId: ycd5umfcq5p6rih4lqfm265f2qblrtrn
|
| 75 |
+
m:
|
| 76 |
+
- "1": trainer/global_step
|
| 77 |
+
"6":
|
| 78 |
+
- 3
|
| 79 |
+
"7": []
|
| 80 |
+
- "2": '*'
|
| 81 |
+
"5": 1
|
| 82 |
+
"6":
|
| 83 |
+
- 1
|
| 84 |
+
"7": []
|
| 85 |
+
python_version: 3.12.12
|
| 86 |
+
t:
|
| 87 |
+
"1":
|
| 88 |
+
- 1
|
| 89 |
+
- 41
|
| 90 |
+
- 49
|
| 91 |
+
- 50
|
| 92 |
+
- 106
|
| 93 |
+
"2":
|
| 94 |
+
- 1
|
| 95 |
+
- 41
|
| 96 |
+
- 49
|
| 97 |
+
- 50
|
| 98 |
+
- 106
|
| 99 |
+
"3":
|
| 100 |
+
- 7
|
| 101 |
+
- 13
|
| 102 |
+
- 15
|
| 103 |
+
- 16
|
| 104 |
+
- 66
|
| 105 |
+
"4": 3.12.12
|
| 106 |
+
"5": 0.25.0
|
| 107 |
+
"12": 0.25.0
|
| 108 |
+
"13": linux-x86_64
|
| 109 |
+
checkpointing:
|
| 110 |
+
value:
|
| 111 |
+
every_n_train_steps: 1500
|
| 112 |
+
load: null
|
| 113 |
+
save_top_k: 2
|
| 114 |
+
save_weights_only: false
|
| 115 |
+
data_loader:
|
| 116 |
+
value:
|
| 117 |
+
test:
|
| 118 |
+
batch_size: 1
|
| 119 |
+
num_workers: 4
|
| 120 |
+
persistent_workers: false
|
| 121 |
+
seed: 2345
|
| 122 |
+
train:
|
| 123 |
+
batch_size: 16
|
| 124 |
+
num_workers: 16
|
| 125 |
+
persistent_workers: true
|
| 126 |
+
seed: 1234
|
| 127 |
+
val:
|
| 128 |
+
batch_size: 1
|
| 129 |
+
num_workers: 1
|
| 130 |
+
persistent_workers: true
|
| 131 |
+
seed: 3456
|
| 132 |
+
dataset:
|
| 133 |
+
value:
|
| 134 |
+
re10k:
|
| 135 |
+
augment: true
|
| 136 |
+
background_color:
|
| 137 |
+
- 0
|
| 138 |
+
- 0
|
| 139 |
+
- 0
|
| 140 |
+
baseline_max: 1e+10
|
| 141 |
+
baseline_min: 0.001
|
| 142 |
+
cameras_are_circular: false
|
| 143 |
+
dynamic_context_views: true
|
| 144 |
+
input_image_shape:
|
| 145 |
+
- 256
|
| 146 |
+
- 256
|
| 147 |
+
make_baseline_1: true
|
| 148 |
+
max_context_views_per_gpu: 24
|
| 149 |
+
max_fov: 100
|
| 150 |
+
name: re10k
|
| 151 |
+
original_image_shape:
|
| 152 |
+
- 360
|
| 153 |
+
- 640
|
| 154 |
+
overfit_to_scene: null
|
| 155 |
+
relative_pose: true
|
| 156 |
+
roots:
|
| 157 |
+
- datasets/re10k
|
| 158 |
+
skip_bad_shape: true
|
| 159 |
+
view_sampler:
|
| 160 |
+
initial_max_distance_between_context_views: 25
|
| 161 |
+
initial_min_distance_between_context_views: 25
|
| 162 |
+
max_distance_between_context_views: 90
|
| 163 |
+
min_distance_between_context_views: 45
|
| 164 |
+
min_distance_to_context_views: 0
|
| 165 |
+
name: bounded
|
| 166 |
+
num_context_views: 2
|
| 167 |
+
num_target_set: 3
|
| 168 |
+
num_target_views: 4
|
| 169 |
+
same_target_gap: false
|
| 170 |
+
target_align: true
|
| 171 |
+
warm_up_steps: 1000
|
| 172 |
+
density_control_loss:
|
| 173 |
+
value:
|
| 174 |
+
error_score:
|
| 175 |
+
grad_scale: 10000
|
| 176 |
+
log_scale: false
|
| 177 |
+
mode: original
|
| 178 |
+
weight: 0.0001
|
| 179 |
+
direct_loss:
|
| 180 |
+
value:
|
| 181 |
+
l1:
|
| 182 |
+
weight: 0.8
|
| 183 |
+
ssim:
|
| 184 |
+
weight: 0.2
|
| 185 |
+
mode:
|
| 186 |
+
value: train
|
| 187 |
+
model:
|
| 188 |
+
value:
|
| 189 |
+
decoder:
|
| 190 |
+
background_color:
|
| 191 |
+
- 0
|
| 192 |
+
- 0
|
| 193 |
+
- 0
|
| 194 |
+
make_scale_invariant: false
|
| 195 |
+
name: splatting_cuda
|
| 196 |
+
density_control:
|
| 197 |
+
aggregation_mode: mean
|
| 198 |
+
aux_refine: false
|
| 199 |
+
grad_mode: absgrad
|
| 200 |
+
gs_param_dim: 256
|
| 201 |
+
latent_dim: 128
|
| 202 |
+
mean_dim: 32
|
| 203 |
+
name: density_control_module
|
| 204 |
+
num_heads: 1
|
| 205 |
+
num_latents: 64
|
| 206 |
+
num_level: 3
|
| 207 |
+
num_self_attn_per_block: 2
|
| 208 |
+
refine_error: false
|
| 209 |
+
refinement_hidden_dim: 32
|
| 210 |
+
refinement_layer_num: 1
|
| 211 |
+
refinement_type: voxelize
|
| 212 |
+
score_mode: absgrad
|
| 213 |
+
use_depth: false
|
| 214 |
+
use_mean_features: true
|
| 215 |
+
use_refine_module: false
|
| 216 |
+
voxel_size: 0.001
|
| 217 |
+
voxelize_activate: false
|
| 218 |
+
encoder:
|
| 219 |
+
align_corners: false
|
| 220 |
+
gs_param_dim: 256
|
| 221 |
+
head_mode: pcd
|
| 222 |
+
input_image_shape:
|
| 223 |
+
- 518
|
| 224 |
+
- 518
|
| 225 |
+
name: dcsplat
|
| 226 |
+
num_level: 3
|
| 227 |
+
use_voxelize: true
|
| 228 |
+
optimizer:
|
| 229 |
+
value:
|
| 230 |
+
accumulate: 1
|
| 231 |
+
backbone_lr_multiplier: 0.1
|
| 232 |
+
backbone_trainable: T+H
|
| 233 |
+
lr: 0.0002
|
| 234 |
+
warm_up_steps: 25
|
| 235 |
+
render_loss:
|
| 236 |
+
value:
|
| 237 |
+
lpips:
|
| 238 |
+
apply_after_step: 0
|
| 239 |
+
weight: 0.05
|
| 240 |
+
mse:
|
| 241 |
+
weight: 1
|
| 242 |
+
seed:
|
| 243 |
+
value: 111123
|
| 244 |
+
test:
|
| 245 |
+
value:
|
| 246 |
+
align_pose: false
|
| 247 |
+
compute_scores: true
|
| 248 |
+
error_threshold: 0.4
|
| 249 |
+
error_threshold_list:
|
| 250 |
+
- 0.2
|
| 251 |
+
- 0.4
|
| 252 |
+
- 0.6
|
| 253 |
+
- 0.8
|
| 254 |
+
- 1
|
| 255 |
+
nvs_view_N_list:
|
| 256 |
+
- 3
|
| 257 |
+
- 6
|
| 258 |
+
- 16
|
| 259 |
+
- 32
|
| 260 |
+
- 64
|
| 261 |
+
output_path: test/ablation/re10k
|
| 262 |
+
pose_align_steps: 100
|
| 263 |
+
pred_intrinsic: false
|
| 264 |
+
rot_opt_lr: 0.005
|
| 265 |
+
save_active_mask_image: false
|
| 266 |
+
save_compare: false
|
| 267 |
+
save_error_score_image: false
|
| 268 |
+
save_gs: false
|
| 269 |
+
save_image: false
|
| 270 |
+
save_sample_wise_metrics: true
|
| 271 |
+
save_video: false
|
| 272 |
+
threshold_mode: ratio
|
| 273 |
+
trans_opt_lr: 0.005
|
| 274 |
+
train:
|
| 275 |
+
value:
|
| 276 |
+
align_corners: false
|
| 277 |
+
beta_dist_param:
|
| 278 |
+
- 0.5
|
| 279 |
+
- 4
|
| 280 |
+
cam_scale_mode: sum
|
| 281 |
+
camera_loss: 10
|
| 282 |
+
context_view_train: false
|
| 283 |
+
ext_scale_detach: false
|
| 284 |
+
extended_visualization: false
|
| 285 |
+
intrinsic_scaling: false
|
| 286 |
+
one_sample_validation: null
|
| 287 |
+
print_log_every_n_steps: 10
|
| 288 |
+
scene_scale_reg_loss: 0.01
|
| 289 |
+
train_aux: false
|
| 290 |
+
train_gs_num: 1
|
| 291 |
+
train_target_set: true
|
| 292 |
+
use_refine_aux: false
|
| 293 |
+
verbose: false
|
| 294 |
+
vggt_cam_loss: true
|
| 295 |
+
vggt_distil: false
|
| 296 |
+
trainer:
|
| 297 |
+
value:
|
| 298 |
+
gradient_clip_val: 0.5
|
| 299 |
+
max_steps: 3001
|
| 300 |
+
num_nodes: 1
|
| 301 |
+
val_check_interval: 250
|
| 302 |
+
wandb:
|
| 303 |
+
value:
|
| 304 |
+
entity: scene-representation-group
|
| 305 |
+
mode: online
|
| 306 |
+
name: ABLATION_0302_noAux
|
| 307 |
+
project: DCSplat
|
| 308 |
+
tags:
|
| 309 |
+
- re10k
|
| 310 |
+
- 256x256
|
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/active_mask_imgs_168_43fa18ddb9c77705edd2.png
ADDED
|
Git LFS Details
|
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/active_mask_imgs_92_3031367680d80379ffab.png
ADDED
|
Git LFS Details
|
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/comparison_167_3dfd8c9e7e3d980df015.png
ADDED
|
Git LFS Details
|
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/comparison_182_5c2b90f763b7bff88204.png
ADDED
|
Git LFS Details
|
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/media/images/error_scores_169_b17e3324e4e0f9205834.png
ADDED
|
Git LFS Details
|
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/output.log
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/requirements.txt
ADDED
|
@@ -0,0 +1,173 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
wheel==0.45.1
|
| 2 |
+
pytz==2025.2
|
| 3 |
+
easydict==1.13
|
| 4 |
+
antlr4-python3-runtime==4.9.3
|
| 5 |
+
wadler_lindig==0.1.7
|
| 6 |
+
networkx==3.4.2
|
| 7 |
+
urllib3==2.5.0
|
| 8 |
+
tzdata==2025.2
|
| 9 |
+
typing-inspection==0.4.1
|
| 10 |
+
tabulate==0.9.0
|
| 11 |
+
smmap==5.0.2
|
| 12 |
+
setuptools==78.1.1
|
| 13 |
+
safetensors==0.5.3
|
| 14 |
+
multidict==6.6.4
|
| 15 |
+
PyYAML==6.0.2
|
| 16 |
+
PySocks==1.7.1
|
| 17 |
+
pyparsing==3.2.5
|
| 18 |
+
pydantic_core==2.33.2
|
| 19 |
+
pycparser==2.23
|
| 20 |
+
protobuf==6.32.1
|
| 21 |
+
propcache==0.3.2
|
| 22 |
+
proglog==0.1.12
|
| 23 |
+
platformdirs==4.4.0
|
| 24 |
+
pip==25.2
|
| 25 |
+
mdurl==0.1.2
|
| 26 |
+
pillow==10.4.0
|
| 27 |
+
packaging==24.2
|
| 28 |
+
opt_einsum==3.4.0
|
| 29 |
+
frozenlist==1.7.0
|
| 30 |
+
numpy==1.26.4
|
| 31 |
+
ninja==1.13.0
|
| 32 |
+
MarkupSafe==3.0.2
|
| 33 |
+
kornia_rs==0.1.9
|
| 34 |
+
kiwisolver==1.4.9
|
| 35 |
+
imageio-ffmpeg==0.6.0
|
| 36 |
+
idna==3.7
|
| 37 |
+
fsspec==2024.6.1
|
| 38 |
+
hf-xet==1.1.10
|
| 39 |
+
gmpy2==2.2.1
|
| 40 |
+
fonttools==4.60.0
|
| 41 |
+
triton==3.4.0
|
| 42 |
+
filelock==3.17.0
|
| 43 |
+
einops==0.8.1
|
| 44 |
+
decorator==4.4.2
|
| 45 |
+
dacite==1.9.2
|
| 46 |
+
cycler==0.12.1
|
| 47 |
+
colorama==0.4.6
|
| 48 |
+
click==8.3.0
|
| 49 |
+
nvidia-nvtx-cu12==12.8.90
|
| 50 |
+
charset-normalizer==3.3.2
|
| 51 |
+
certifi==2025.8.3
|
| 52 |
+
beartype==0.19.0
|
| 53 |
+
attrs==25.3.0
|
| 54 |
+
async-timeout==5.0.1
|
| 55 |
+
annotated-types==0.7.0
|
| 56 |
+
aiohappyeyeballs==2.6.1
|
| 57 |
+
yarl==1.20.1
|
| 58 |
+
tifffile==2025.5.10
|
| 59 |
+
sentry-sdk==2.39.0
|
| 60 |
+
scipy==1.15.3
|
| 61 |
+
pydantic==2.11.9
|
| 62 |
+
pandas==2.3.2
|
| 63 |
+
opencv-python==4.11.0.86
|
| 64 |
+
omegaconf==2.3.0
|
| 65 |
+
markdown-it-py==4.0.0
|
| 66 |
+
lightning-utilities==0.14.3
|
| 67 |
+
lazy_loader==0.4
|
| 68 |
+
jaxtyping==0.2.37
|
| 69 |
+
imageio==2.37.0
|
| 70 |
+
gitdb==4.0.12
|
| 71 |
+
contourpy==1.3.2
|
| 72 |
+
colorspacious==1.1.2
|
| 73 |
+
cffi==1.17.1
|
| 74 |
+
aiosignal==1.4.0
|
| 75 |
+
scikit-video==1.1.11
|
| 76 |
+
scikit-image==0.25.2
|
| 77 |
+
rich==14.1.0
|
| 78 |
+
moviepy==1.0.3
|
| 79 |
+
matplotlib==3.10.6
|
| 80 |
+
hydra-core==1.3.2
|
| 81 |
+
huggingface-hub==0.35.1
|
| 82 |
+
GitPython==3.1.45
|
| 83 |
+
brotlicffi==1.0.9.2
|
| 84 |
+
aiohttp==3.12.15
|
| 85 |
+
torchmetrics==1.8.2
|
| 86 |
+
opt-einsum-fx==0.1.4
|
| 87 |
+
kornia==0.8.1
|
| 88 |
+
pytorch-lightning==2.5.1
|
| 89 |
+
lpips==0.1.4
|
| 90 |
+
e3nn==0.6.0
|
| 91 |
+
lightning==2.5.1
|
| 92 |
+
gsplat==1.5.3
|
| 93 |
+
nvidia-cusparselt-cu12==0.7.1
|
| 94 |
+
nvidia-nvjitlink-cu12==12.8.93
|
| 95 |
+
nvidia-nccl-cu12==2.27.3
|
| 96 |
+
nvidia-curand-cu12==10.3.9.90
|
| 97 |
+
nvidia-cufile-cu12==1.13.1.3
|
| 98 |
+
nvidia-cuda-runtime-cu12==12.8.90
|
| 99 |
+
nvidia-cuda-nvrtc-cu12==12.8.93
|
| 100 |
+
nvidia-cuda-cupti-cu12==12.8.90
|
| 101 |
+
nvidia-cublas-cu12==12.8.4.1
|
| 102 |
+
nvidia-cusparse-cu12==12.5.8.93
|
| 103 |
+
nvidia-cufft-cu12==11.3.3.83
|
| 104 |
+
nvidia-cudnn-cu12==9.10.2.21
|
| 105 |
+
nvidia-cusolver-cu12==11.7.3.90
|
| 106 |
+
torch==2.8.0+cu128
|
| 107 |
+
torchvision==0.23.0+cu128
|
| 108 |
+
torchaudio==2.8.0+cu128
|
| 109 |
+
torch_scatter==2.1.2+pt28cu128
|
| 110 |
+
plyfile==1.1.3
|
| 111 |
+
wandb==0.25.0
|
| 112 |
+
cuda-bindings==13.0.3
|
| 113 |
+
cuda-pathfinder==1.3.3
|
| 114 |
+
Jinja2==3.1.6
|
| 115 |
+
mpmath==1.3.0
|
| 116 |
+
nvidia-cublas==13.1.0.3
|
| 117 |
+
nvidia-cuda-cupti==13.0.85
|
| 118 |
+
nvidia-cuda-nvrtc==13.0.88
|
| 119 |
+
nvidia-cuda-runtime==13.0.96
|
| 120 |
+
nvidia-cudnn-cu13==9.15.1.9
|
| 121 |
+
nvidia-cufft==12.0.0.61
|
| 122 |
+
nvidia-cufile==1.15.1.6
|
| 123 |
+
nvidia-curand==10.4.0.35
|
| 124 |
+
nvidia-cusolver==12.0.4.66
|
| 125 |
+
nvidia-cusparse==12.6.3.3
|
| 126 |
+
nvidia-cusparselt-cu13==0.8.0
|
| 127 |
+
nvidia-nccl-cu13==2.28.9
|
| 128 |
+
nvidia-nvjitlink==13.0.88
|
| 129 |
+
nvidia-nvshmem-cu13==3.4.5
|
| 130 |
+
nvidia-nvtx==13.0.85
|
| 131 |
+
requests==2.32.5
|
| 132 |
+
sentencepiece==0.2.1
|
| 133 |
+
sympy==1.14.0
|
| 134 |
+
torchcodec==0.10.0
|
| 135 |
+
torchdata==0.10.0
|
| 136 |
+
torchtext==0.6.0
|
| 137 |
+
anyio==4.12.0
|
| 138 |
+
asttokens==3.0.1
|
| 139 |
+
comm==0.2.3
|
| 140 |
+
debugpy==1.8.19
|
| 141 |
+
executing==2.2.1
|
| 142 |
+
h11==0.16.0
|
| 143 |
+
httpcore==1.0.9
|
| 144 |
+
httpx==0.28.1
|
| 145 |
+
ipykernel==7.1.0
|
| 146 |
+
ipython==9.8.0
|
| 147 |
+
ipython_pygments_lexers==1.1.1
|
| 148 |
+
ipywidgets==8.1.8
|
| 149 |
+
jedi==0.19.2
|
| 150 |
+
jupyter_client==8.7.0
|
| 151 |
+
jupyter_core==5.9.1
|
| 152 |
+
jupyterlab_widgets==3.0.16
|
| 153 |
+
matplotlib-inline==0.2.1
|
| 154 |
+
nest-asyncio==1.6.0
|
| 155 |
+
parso==0.8.5
|
| 156 |
+
pexpect==4.9.0
|
| 157 |
+
prompt_toolkit==3.0.52
|
| 158 |
+
psutil==7.2.1
|
| 159 |
+
ptyprocess==0.7.0
|
| 160 |
+
pure_eval==0.2.3
|
| 161 |
+
Pygments==2.19.2
|
| 162 |
+
python-dateutil==2.9.0.post0
|
| 163 |
+
pyzmq==27.1.0
|
| 164 |
+
shellingham==1.5.4
|
| 165 |
+
six==1.17.0
|
| 166 |
+
stack-data==0.6.3
|
| 167 |
+
tornado==6.5.4
|
| 168 |
+
tqdm==4.67.1
|
| 169 |
+
traitlets==5.14.3
|
| 170 |
+
typer-slim==0.21.0
|
| 171 |
+
typing_extensions==4.15.0
|
| 172 |
+
wcwidth==0.2.14
|
| 173 |
+
widgetsnbextension==4.0.15
|
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/wandb-metadata.json
ADDED
|
@@ -0,0 +1,93 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"os": "Linux-6.8.0-94-generic-x86_64-with-glibc2.39",
|
| 3 |
+
"python": "CPython 3.12.12",
|
| 4 |
+
"startedAt": "2026-03-03T14:34:53.292304Z",
|
| 5 |
+
"args": [
|
| 6 |
+
"+experiment=re10k_ablation_24v",
|
| 7 |
+
"wandb.mode=online",
|
| 8 |
+
"wandb.name=ABLATION_0302_noAux",
|
| 9 |
+
"train.train_aux=false"
|
| 10 |
+
],
|
| 11 |
+
"program": "-m src.main",
|
| 12 |
+
"git": {
|
| 13 |
+
"remote": "git@github.com:K-nowing/CVPR2026.git",
|
| 14 |
+
"commit": "9dfce172a0f8c7ce85e763899f7ef741ecffc454"
|
| 15 |
+
},
|
| 16 |
+
"email": "dna9041@korea.ac.kr",
|
| 17 |
+
"root": "/workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noAux",
|
| 18 |
+
"host": "e9d3310a05da",
|
| 19 |
+
"executable": "/venv/main/bin/python",
|
| 20 |
+
"cpu_count": 128,
|
| 21 |
+
"cpu_count_logical": 256,
|
| 22 |
+
"gpu": "NVIDIA H200",
|
| 23 |
+
"gpu_count": 8,
|
| 24 |
+
"disk": {
|
| 25 |
+
"/": {
|
| 26 |
+
"total": "735513149440",
|
| 27 |
+
"used": "676972425216"
|
| 28 |
+
}
|
| 29 |
+
},
|
| 30 |
+
"memory": {
|
| 31 |
+
"total": "1622950240256"
|
| 32 |
+
},
|
| 33 |
+
"gpu_nvidia": [
|
| 34 |
+
{
|
| 35 |
+
"name": "NVIDIA H200",
|
| 36 |
+
"memoryTotal": "150754820096",
|
| 37 |
+
"cudaCores": 16896,
|
| 38 |
+
"architecture": "Hopper",
|
| 39 |
+
"uuid": "GPU-9a20101e-d876-facd-5f05-805081aede41"
|
| 40 |
+
},
|
| 41 |
+
{
|
| 42 |
+
"name": "NVIDIA H200",
|
| 43 |
+
"memoryTotal": "150754820096",
|
| 44 |
+
"cudaCores": 16896,
|
| 45 |
+
"architecture": "Hopper",
|
| 46 |
+
"uuid": "GPU-84736a77-ee75-3324-e4e1-99cc15bfb5e9"
|
| 47 |
+
},
|
| 48 |
+
{
|
| 49 |
+
"name": "NVIDIA H200",
|
| 50 |
+
"memoryTotal": "150754820096",
|
| 51 |
+
"cudaCores": 16896,
|
| 52 |
+
"architecture": "Hopper",
|
| 53 |
+
"uuid": "GPU-423d3161-cdc4-3fc0-caee-d15cfaa83ca6"
|
| 54 |
+
},
|
| 55 |
+
{
|
| 56 |
+
"name": "NVIDIA H200",
|
| 57 |
+
"memoryTotal": "150754820096",
|
| 58 |
+
"cudaCores": 16896,
|
| 59 |
+
"architecture": "Hopper",
|
| 60 |
+
"uuid": "GPU-5b0058b2-cdb9-c952-04f9-87dcaa7ea742"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"name": "NVIDIA H200",
|
| 64 |
+
"memoryTotal": "150754820096",
|
| 65 |
+
"cudaCores": 16896,
|
| 66 |
+
"architecture": "Hopper",
|
| 67 |
+
"uuid": "GPU-08b37f98-4603-d483-2f2b-fe5311aa42f2"
|
| 68 |
+
},
|
| 69 |
+
{
|
| 70 |
+
"name": "NVIDIA H200",
|
| 71 |
+
"memoryTotal": "150754820096",
|
| 72 |
+
"cudaCores": 16896,
|
| 73 |
+
"architecture": "Hopper",
|
| 74 |
+
"uuid": "GPU-03273b5b-2fdd-a5fe-4460-c897334ae464"
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"name": "NVIDIA H200",
|
| 78 |
+
"memoryTotal": "150754820096",
|
| 79 |
+
"cudaCores": 16896,
|
| 80 |
+
"architecture": "Hopper",
|
| 81 |
+
"uuid": "GPU-292d466c-d00d-25a4-28b6-e6c978d3e70c"
|
| 82 |
+
},
|
| 83 |
+
{
|
| 84 |
+
"name": "NVIDIA H200",
|
| 85 |
+
"memoryTotal": "150754820096",
|
| 86 |
+
"cudaCores": 16896,
|
| 87 |
+
"architecture": "Hopper",
|
| 88 |
+
"uuid": "GPU-46f38561-3148-e442-7f7f-bfe447bab7fe"
|
| 89 |
+
}
|
| 90 |
+
],
|
| 91 |
+
"cudaVersion": "13.0",
|
| 92 |
+
"writerId": "ycd5umfcq5p6rih4lqfm265f2qblrtrn"
|
| 93 |
+
}
|
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/files/wandb-summary.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"active_mask_imgs":{"filenames":["media/images/active_mask_imgs_183_27237392d1ba36fdb4c3.png"],"captions":["1072aae07584e091"],"_type":"images/separated","width":536,"height":800,"format":"png","count":1},"lr-AdamW/pg2":2e-05,"train/scene_scale":1.0012001991271973,"val/ssim":0.5464253425598145,"val/psnr":17.485671997070312,"loss/final_3dgs/mse":0.007215200457721949,"epoch":0,"val/lpips":0.3175295889377594,"loss/scene_scale_reg":3.102540722466074e-05,"trainer/global_step":3001,"loss/total":0.01728440262377262,"_step":186,"_timestamp":1.77255873057175e+09,"_runtime":10242,"train/psnr_probabilistic":21.999189376831055,"comparison":{"filenames":["media/images/comparison_182_5c2b90f763b7bff88204.png"],"captions":["1072aae07584e091"],"_type":"images/separated","width":1064,"height":1098,"format":"png","count":1},"loss/final_3dgs/lpips":0.0082782544195652,"train/comparison":{"filenames":["media/images/train/comparison_186_c80084b1f8dbfc2a13d7.png"],"captions":[["c270572a7f5ea828"]],"_type":"images/separated","width":536,"height":6378,"format":"png","count":1},"lr-AdamW/pg1-momentum":0.9,"_wandb":{"runtime":10242},"error_scores":{"height":536,"format":"png","count":1,"filenames":["media/images/error_scores_184_deb971ba58ebc4d5fe90.png"],"captions":["1072aae07584e091"],"_type":"images/separated","width":800},"loss/camera":0.00017599233251530677,"info/global_step":3000,"val/gaussian_num_ratio":0.40003204345703125,"lr-AdamW/pg1":2.003594834351718e-05,"lr-AdamW/pg2-momentum":0.9}
|
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug-core.log
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2026-03-03T14:34:53.358083416Z","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmppl6b06_b/port-854556.txt","pid":854556,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
|
| 2 |
+
{"time":"2026-03-03T14:34:53.358933202Z","level":"INFO","msg":"server: will exit if parent process dies","ppid":854556}
|
| 3 |
+
{"time":"2026-03-03T14:34:53.358910572Z","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-854556-856980-924415834/socket","Net":"unix"}}
|
| 4 |
+
{"time":"2026-03-03T14:34:53.535292678Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
|
| 5 |
+
{"time":"2026-03-03T14:34:53.543148589Z","level":"INFO","msg":"handleInformInit: received","streamId":"l8o7jl7r","id":"1(@)"}
|
| 6 |
+
{"time":"2026-03-03T14:34:53.95830905Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"l8o7jl7r","id":"1(@)"}
|
| 7 |
+
{"time":"2026-03-03T14:34:59.899286538Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"vehmru59g0oz"}
|
| 8 |
+
{"time":"2026-03-03T17:25:37.368559306Z","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
|
| 9 |
+
{"time":"2026-03-03T17:25:37.368609447Z","level":"INFO","msg":"connection: closing","id":"1(@)"}
|
| 10 |
+
{"time":"2026-03-03T17:25:37.368630207Z","level":"INFO","msg":"server is shutting down"}
|
| 11 |
+
{"time":"2026-03-03T17:25:37.368663987Z","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
|
| 12 |
+
{"time":"2026-03-03T17:25:37.368838899Z","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-854556-856980-924415834/socket","Net":"unix"}}
|
| 13 |
+
{"time":"2026-03-03T17:25:38.528751408Z","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
|
| 14 |
+
{"time":"2026-03-03T17:25:38.528775028Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
|
| 15 |
+
{"time":"2026-03-03T17:25:38.528786418Z","level":"INFO","msg":"server is closed"}
|
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug-internal.log
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2026-03-03T14:34:53.54334025Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
|
| 2 |
+
{"time":"2026-03-03T14:34:53.958050168Z","level":"INFO","msg":"stream: created new stream","id":"l8o7jl7r"}
|
| 3 |
+
{"time":"2026-03-03T14:34:53.958155019Z","level":"INFO","msg":"handler: started","stream_id":"l8o7jl7r"}
|
| 4 |
+
{"time":"2026-03-03T14:34:53.9583015Z","level":"INFO","msg":"stream: started","id":"l8o7jl7r"}
|
| 5 |
+
{"time":"2026-03-03T14:34:53.95834932Z","level":"INFO","msg":"sender: started","stream_id":"l8o7jl7r"}
|
| 6 |
+
{"time":"2026-03-03T14:34:53.95835085Z","level":"INFO","msg":"writer: started","stream_id":"l8o7jl7r"}
|
| 7 |
+
{"time":"2026-03-03T17:25:37.368615437Z","level":"INFO","msg":"stream: closing","id":"l8o7jl7r"}
|
| 8 |
+
{"time":"2026-03-03T17:25:38.278996749Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
|
| 9 |
+
{"time":"2026-03-03T17:25:38.528169562Z","level":"INFO","msg":"handler: closed","stream_id":"l8o7jl7r"}
|
| 10 |
+
{"time":"2026-03-03T17:25:38.528336264Z","level":"INFO","msg":"sender: closed","stream_id":"l8o7jl7r"}
|
| 11 |
+
{"time":"2026-03-03T17:25:38.528358534Z","level":"INFO","msg":"stream: closed","id":"l8o7jl7r"}
|
ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug.log
ADDED
|
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
|
| 2 |
+
2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_setup.py:_flush():81] Configure stats pid to 854556
|
| 3 |
+
2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_setup.py:_flush():81] Loading settings from environment variables
|
| 4 |
+
2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug.log
|
| 5 |
+
2026-03-03 14:34:53,293 INFO MainThread:854556 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noAux/wandb/run-20260303_143453-l8o7jl7r/logs/debug-internal.log
|
| 6 |
+
2026-03-03 14:34:53,294 INFO MainThread:854556 [wandb_init.py:init():844] calling init triggers
|
| 7 |
+
2026-03-03 14:34:53,294 INFO MainThread:854556 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
|
| 8 |
+
config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'pcd', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'absgrad', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': False, 'voxelize_activate': False, 'use_depth': False}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'density_control_loss': {'error_score': {'weight': 0.0001, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0302_noAux', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': False, 'vggt_cam_loss': True, 'vggt_distil': False, 'context_view_train': False}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'save_gs': False, 'save_sample_wise_metrics': True, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3, 'target_align': True}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
|
| 9 |
+
2026-03-03 14:34:53,294 INFO MainThread:854556 [wandb_init.py:init():892] starting backend
|
| 10 |
+
2026-03-03 14:34:53,535 INFO MainThread:854556 [wandb_init.py:init():895] sending inform_init request
|
| 11 |
+
2026-03-03 14:34:53,540 INFO MainThread:854556 [wandb_init.py:init():903] backend started and connected
|
| 12 |
+
2026-03-03 14:34:53,542 INFO MainThread:854556 [wandb_init.py:init():973] updated telemetry
|
| 13 |
+
2026-03-03 14:34:53,548 INFO MainThread:854556 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
|
| 14 |
+
2026-03-03 14:34:54,767 INFO MainThread:854556 [wandb_init.py:init():1042] starting run threads in backend
|
| 15 |
+
2026-03-03 14:34:54,895 INFO MainThread:854556 [wandb_run.py:_console_start():2524] atexit reg
|
| 16 |
+
2026-03-03 14:34:54,895 INFO MainThread:854556 [wandb_run.py:_redirect():2373] redirect: wrap_raw
|
| 17 |
+
2026-03-03 14:34:54,895 INFO MainThread:854556 [wandb_run.py:_redirect():2442] Wrapping output streams.
|
| 18 |
+
2026-03-03 14:34:54,896 INFO MainThread:854556 [wandb_run.py:_redirect():2465] Redirects installed.
|
| 19 |
+
2026-03-03 14:34:54,899 INFO MainThread:854556 [wandb_init.py:init():1082] run started, returning control to user process
|
| 20 |
+
2026-03-03 17:25:37,368 INFO wandb-AsyncioManager-main:854556 [service_client.py:_forward_responses():134] Reached EOF.
|
| 21 |
+
2026-03-03 17:25:37,368 INFO wandb-AsyncioManager-main:854556 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.
|
ABLATION_0302_noTgtAlign/.hydra/config.yaml
ADDED
|
@@ -0,0 +1,188 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
encoder:
|
| 3 |
+
name: dcsplat
|
| 4 |
+
input_image_shape:
|
| 5 |
+
- 518
|
| 6 |
+
- 518
|
| 7 |
+
head_mode: pcd
|
| 8 |
+
num_level: 3
|
| 9 |
+
gs_param_dim: 256
|
| 10 |
+
align_corners: false
|
| 11 |
+
use_voxelize: true
|
| 12 |
+
decoder:
|
| 13 |
+
name: splatting_cuda
|
| 14 |
+
background_color:
|
| 15 |
+
- 0.0
|
| 16 |
+
- 0.0
|
| 17 |
+
- 0.0
|
| 18 |
+
make_scale_invariant: false
|
| 19 |
+
density_control:
|
| 20 |
+
name: density_control_module
|
| 21 |
+
mean_dim: 32
|
| 22 |
+
gs_param_dim: 256
|
| 23 |
+
refinement_layer_num: 1
|
| 24 |
+
num_level: 3
|
| 25 |
+
grad_mode: absgrad
|
| 26 |
+
use_mean_features: true
|
| 27 |
+
refinement_type: voxelize
|
| 28 |
+
refinement_hidden_dim: 32
|
| 29 |
+
aggregation_mode: mean
|
| 30 |
+
num_heads: 1
|
| 31 |
+
score_mode: random
|
| 32 |
+
latent_dim: 128
|
| 33 |
+
num_latents: 64
|
| 34 |
+
num_self_attn_per_block: 2
|
| 35 |
+
voxel_size: 0.001
|
| 36 |
+
aux_refine: false
|
| 37 |
+
refine_error: false
|
| 38 |
+
use_refine_module: false
|
| 39 |
+
voxelize_activate: false
|
| 40 |
+
use_depth: false
|
| 41 |
+
render_loss:
|
| 42 |
+
mse:
|
| 43 |
+
weight: 1.0
|
| 44 |
+
lpips:
|
| 45 |
+
weight: 0.05
|
| 46 |
+
apply_after_step: 0
|
| 47 |
+
density_control_loss:
|
| 48 |
+
error_score:
|
| 49 |
+
weight: 0.0001
|
| 50 |
+
log_scale: false
|
| 51 |
+
grad_scale: 10000.0
|
| 52 |
+
mode: original
|
| 53 |
+
direct_loss:
|
| 54 |
+
l1:
|
| 55 |
+
weight: 0.8
|
| 56 |
+
ssim:
|
| 57 |
+
weight: 0.2
|
| 58 |
+
wandb:
|
| 59 |
+
project: DCSplat
|
| 60 |
+
entity: scene-representation-group
|
| 61 |
+
name: ABLATION_0302_noTgtAlign
|
| 62 |
+
mode: online
|
| 63 |
+
tags:
|
| 64 |
+
- re10k
|
| 65 |
+
- 256x256
|
| 66 |
+
mode: train
|
| 67 |
+
data_loader:
|
| 68 |
+
train:
|
| 69 |
+
num_workers: 16
|
| 70 |
+
persistent_workers: true
|
| 71 |
+
batch_size: 16
|
| 72 |
+
seed: 1234
|
| 73 |
+
test:
|
| 74 |
+
num_workers: 4
|
| 75 |
+
persistent_workers: false
|
| 76 |
+
batch_size: 1
|
| 77 |
+
seed: 2345
|
| 78 |
+
val:
|
| 79 |
+
num_workers: 1
|
| 80 |
+
persistent_workers: true
|
| 81 |
+
batch_size: 1
|
| 82 |
+
seed: 3456
|
| 83 |
+
optimizer:
|
| 84 |
+
lr: 0.0002
|
| 85 |
+
warm_up_steps: 25
|
| 86 |
+
backbone_lr_multiplier: 0.1
|
| 87 |
+
backbone_trainable: T+H
|
| 88 |
+
accumulate: 1
|
| 89 |
+
checkpointing:
|
| 90 |
+
load: null
|
| 91 |
+
every_n_train_steps: 1500
|
| 92 |
+
save_top_k: 2
|
| 93 |
+
save_weights_only: false
|
| 94 |
+
train:
|
| 95 |
+
extended_visualization: false
|
| 96 |
+
print_log_every_n_steps: 10
|
| 97 |
+
camera_loss: 10.0
|
| 98 |
+
one_sample_validation: null
|
| 99 |
+
align_corners: false
|
| 100 |
+
intrinsic_scaling: false
|
| 101 |
+
verbose: false
|
| 102 |
+
beta_dist_param:
|
| 103 |
+
- 0.5
|
| 104 |
+
- 4.0
|
| 105 |
+
use_refine_aux: false
|
| 106 |
+
train_target_set: true
|
| 107 |
+
train_gs_num: 1
|
| 108 |
+
ext_scale_detach: false
|
| 109 |
+
cam_scale_mode: sum
|
| 110 |
+
scene_scale_reg_loss: 0.01
|
| 111 |
+
train_aux: true
|
| 112 |
+
vggt_cam_loss: true
|
| 113 |
+
vggt_distil: false
|
| 114 |
+
context_view_train: false
|
| 115 |
+
test:
|
| 116 |
+
output_path: test/ablation/re10k
|
| 117 |
+
align_pose: false
|
| 118 |
+
pose_align_steps: 100
|
| 119 |
+
rot_opt_lr: 0.005
|
| 120 |
+
trans_opt_lr: 0.005
|
| 121 |
+
compute_scores: true
|
| 122 |
+
save_image: false
|
| 123 |
+
save_video: false
|
| 124 |
+
save_active_mask_image: false
|
| 125 |
+
save_error_score_image: false
|
| 126 |
+
save_compare: false
|
| 127 |
+
save_gs: false
|
| 128 |
+
save_sample_wise_metrics: true
|
| 129 |
+
pred_intrinsic: false
|
| 130 |
+
error_threshold: 0.4
|
| 131 |
+
error_threshold_list:
|
| 132 |
+
- 0.2
|
| 133 |
+
- 0.4
|
| 134 |
+
- 0.6
|
| 135 |
+
- 0.8
|
| 136 |
+
- 1.0
|
| 137 |
+
threshold_mode: ratio
|
| 138 |
+
nvs_view_N_list:
|
| 139 |
+
- 3
|
| 140 |
+
- 6
|
| 141 |
+
- 16
|
| 142 |
+
- 32
|
| 143 |
+
- 64
|
| 144 |
+
seed: 111123
|
| 145 |
+
trainer:
|
| 146 |
+
max_steps: 3001
|
| 147 |
+
val_check_interval: 250
|
| 148 |
+
gradient_clip_val: 0.5
|
| 149 |
+
num_nodes: 1
|
| 150 |
+
dataset:
|
| 151 |
+
re10k:
|
| 152 |
+
make_baseline_1: true
|
| 153 |
+
relative_pose: true
|
| 154 |
+
augment: true
|
| 155 |
+
background_color:
|
| 156 |
+
- 0.0
|
| 157 |
+
- 0.0
|
| 158 |
+
- 0.0
|
| 159 |
+
overfit_to_scene: null
|
| 160 |
+
skip_bad_shape: true
|
| 161 |
+
view_sampler:
|
| 162 |
+
name: bounded
|
| 163 |
+
num_target_views: 4
|
| 164 |
+
num_context_views: 2
|
| 165 |
+
min_distance_between_context_views: 45
|
| 166 |
+
max_distance_between_context_views: 90
|
| 167 |
+
min_distance_to_context_views: 0
|
| 168 |
+
warm_up_steps: 1000
|
| 169 |
+
initial_min_distance_between_context_views: 25
|
| 170 |
+
initial_max_distance_between_context_views: 25
|
| 171 |
+
same_target_gap: false
|
| 172 |
+
num_target_set: 3
|
| 173 |
+
target_align: true
|
| 174 |
+
name: re10k
|
| 175 |
+
roots:
|
| 176 |
+
- datasets/re10k
|
| 177 |
+
input_image_shape:
|
| 178 |
+
- 256
|
| 179 |
+
- 256
|
| 180 |
+
original_image_shape:
|
| 181 |
+
- 360
|
| 182 |
+
- 640
|
| 183 |
+
cameras_are_circular: false
|
| 184 |
+
baseline_min: 0.001
|
| 185 |
+
baseline_max: 10000000000.0
|
| 186 |
+
max_fov: 100.0
|
| 187 |
+
dynamic_context_views: true
|
| 188 |
+
max_context_views_per_gpu: 24
|
ABLATION_0302_noTgtAlign/.hydra/hydra.yaml
ADDED
|
@@ -0,0 +1,165 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
hydra:
|
| 2 |
+
run:
|
| 3 |
+
dir: outputs/ablation/re10k/${wandb.name}
|
| 4 |
+
sweep:
|
| 5 |
+
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
|
| 6 |
+
subdir: ${hydra.job.num}
|
| 7 |
+
launcher:
|
| 8 |
+
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
|
| 9 |
+
sweeper:
|
| 10 |
+
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
|
| 11 |
+
max_batch_size: null
|
| 12 |
+
params: null
|
| 13 |
+
help:
|
| 14 |
+
app_name: ${hydra.job.name}
|
| 15 |
+
header: '${hydra.help.app_name} is powered by Hydra.
|
| 16 |
+
|
| 17 |
+
'
|
| 18 |
+
footer: 'Powered by Hydra (https://hydra.cc)
|
| 19 |
+
|
| 20 |
+
Use --hydra-help to view Hydra specific help
|
| 21 |
+
|
| 22 |
+
'
|
| 23 |
+
template: '${hydra.help.header}
|
| 24 |
+
|
| 25 |
+
== Configuration groups ==
|
| 26 |
+
|
| 27 |
+
Compose your configuration from those groups (group=option)
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
$APP_CONFIG_GROUPS
|
| 31 |
+
|
| 32 |
+
|
| 33 |
+
== Config ==
|
| 34 |
+
|
| 35 |
+
Override anything in the config (foo.bar=value)
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
$CONFIG
|
| 39 |
+
|
| 40 |
+
|
| 41 |
+
${hydra.help.footer}
|
| 42 |
+
|
| 43 |
+
'
|
| 44 |
+
hydra_help:
|
| 45 |
+
template: 'Hydra (${hydra.runtime.version})
|
| 46 |
+
|
| 47 |
+
See https://hydra.cc for more info.
|
| 48 |
+
|
| 49 |
+
|
| 50 |
+
== Flags ==
|
| 51 |
+
|
| 52 |
+
$FLAGS_HELP
|
| 53 |
+
|
| 54 |
+
|
| 55 |
+
== Configuration groups ==
|
| 56 |
+
|
| 57 |
+
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
|
| 58 |
+
to command line)
|
| 59 |
+
|
| 60 |
+
|
| 61 |
+
$HYDRA_CONFIG_GROUPS
|
| 62 |
+
|
| 63 |
+
|
| 64 |
+
Use ''--cfg hydra'' to Show the Hydra config.
|
| 65 |
+
|
| 66 |
+
'
|
| 67 |
+
hydra_help: ???
|
| 68 |
+
hydra_logging:
|
| 69 |
+
version: 1
|
| 70 |
+
formatters:
|
| 71 |
+
simple:
|
| 72 |
+
format: '[%(asctime)s][HYDRA] %(message)s'
|
| 73 |
+
handlers:
|
| 74 |
+
console:
|
| 75 |
+
class: logging.StreamHandler
|
| 76 |
+
formatter: simple
|
| 77 |
+
stream: ext://sys.stdout
|
| 78 |
+
root:
|
| 79 |
+
level: INFO
|
| 80 |
+
handlers:
|
| 81 |
+
- console
|
| 82 |
+
loggers:
|
| 83 |
+
logging_example:
|
| 84 |
+
level: DEBUG
|
| 85 |
+
disable_existing_loggers: false
|
| 86 |
+
job_logging:
|
| 87 |
+
version: 1
|
| 88 |
+
formatters:
|
| 89 |
+
simple:
|
| 90 |
+
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
|
| 91 |
+
handlers:
|
| 92 |
+
console:
|
| 93 |
+
class: logging.StreamHandler
|
| 94 |
+
formatter: simple
|
| 95 |
+
stream: ext://sys.stdout
|
| 96 |
+
file:
|
| 97 |
+
class: logging.FileHandler
|
| 98 |
+
formatter: simple
|
| 99 |
+
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
|
| 100 |
+
root:
|
| 101 |
+
level: INFO
|
| 102 |
+
handlers:
|
| 103 |
+
- console
|
| 104 |
+
- file
|
| 105 |
+
disable_existing_loggers: false
|
| 106 |
+
env: {}
|
| 107 |
+
mode: RUN
|
| 108 |
+
searchpath: []
|
| 109 |
+
callbacks: {}
|
| 110 |
+
output_subdir: .hydra
|
| 111 |
+
overrides:
|
| 112 |
+
hydra:
|
| 113 |
+
- hydra.mode=RUN
|
| 114 |
+
task:
|
| 115 |
+
- +experiment=re10k_ablation_24v
|
| 116 |
+
- wandb.mode=online
|
| 117 |
+
- wandb.name=ABLATION_0302_noTgtAlign
|
| 118 |
+
- model.density_control.score_mode=random
|
| 119 |
+
job:
|
| 120 |
+
name: main
|
| 121 |
+
chdir: null
|
| 122 |
+
override_dirname: +experiment=re10k_ablation_24v,model.density_control.score_mode=random,wandb.mode=online,wandb.name=ABLATION_0302_noTgtAlign
|
| 123 |
+
id: ???
|
| 124 |
+
num: ???
|
| 125 |
+
config_name: main
|
| 126 |
+
env_set: {}
|
| 127 |
+
env_copy: []
|
| 128 |
+
config:
|
| 129 |
+
override_dirname:
|
| 130 |
+
kv_sep: '='
|
| 131 |
+
item_sep: ','
|
| 132 |
+
exclude_keys: []
|
| 133 |
+
runtime:
|
| 134 |
+
version: 1.3.2
|
| 135 |
+
version_base: '1.3'
|
| 136 |
+
cwd: /workspace/code/CVPR2026
|
| 137 |
+
config_sources:
|
| 138 |
+
- path: hydra.conf
|
| 139 |
+
schema: pkg
|
| 140 |
+
provider: hydra
|
| 141 |
+
- path: /workspace/code/CVPR2026/config
|
| 142 |
+
schema: file
|
| 143 |
+
provider: main
|
| 144 |
+
- path: ''
|
| 145 |
+
schema: structured
|
| 146 |
+
provider: schema
|
| 147 |
+
output_dir: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign
|
| 148 |
+
choices:
|
| 149 |
+
experiment: re10k_ablation_24v
|
| 150 |
+
dataset@dataset.re10k: re10k
|
| 151 |
+
dataset/view_sampler_dataset_specific_config@dataset.re10k.view_sampler: bounded_re10k
|
| 152 |
+
dataset/view_sampler@dataset.re10k.view_sampler: bounded
|
| 153 |
+
model/density_control: density_control_module
|
| 154 |
+
model/decoder: splatting_cuda
|
| 155 |
+
model/encoder: dcsplat
|
| 156 |
+
hydra/env: default
|
| 157 |
+
hydra/callbacks: null
|
| 158 |
+
hydra/job_logging: default
|
| 159 |
+
hydra/hydra_logging: default
|
| 160 |
+
hydra/hydra_help: default
|
| 161 |
+
hydra/help: default
|
| 162 |
+
hydra/sweeper: basic
|
| 163 |
+
hydra/launcher: basic
|
| 164 |
+
hydra/output: default
|
| 165 |
+
verbose: false
|
ABLATION_0302_noTgtAlign/.hydra/overrides.yaml
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
- +experiment=re10k_ablation_24v
|
| 2 |
+
- wandb.mode=online
|
| 3 |
+
- wandb.name=ABLATION_0302_noTgtAlign
|
| 4 |
+
- model.density_control.score_mode=random
|
ABLATION_0302_noTgtAlign/main.log
ADDED
|
@@ -0,0 +1,83 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2026-03-03 17:25:57,101][dinov2][INFO] - using MLP layer as FFN
|
| 2 |
+
[2026-03-03 17:26:03,305][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 3 |
+
warnings.warn(
|
| 4 |
+
|
| 5 |
+
[2026-03-03 17:26:03,305][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 6 |
+
warnings.warn(msg)
|
| 7 |
+
|
| 8 |
+
[2026-03-03 17:26:08,283][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=255` in the `DataLoader` to improve performance.
|
| 9 |
+
|
| 10 |
+
[2026-03-03 17:29:31,696][dinov2][INFO] - using MLP layer as FFN
|
| 11 |
+
[2026-03-03 17:29:37,687][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 12 |
+
warnings.warn(
|
| 13 |
+
|
| 14 |
+
[2026-03-03 17:29:37,687][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 15 |
+
warnings.warn(msg)
|
| 16 |
+
|
| 17 |
+
[2026-03-03 17:30:28,360][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=31` in the `DataLoader` to improve performance.
|
| 18 |
+
|
| 19 |
+
[2026-03-03 17:30:28,362][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 20 |
+
warnings.warn( # warn only once
|
| 21 |
+
|
| 22 |
+
[2026-03-03 17:32:05,291][dinov2][INFO] - using MLP layer as FFN
|
| 23 |
+
[2026-03-03 17:32:11,332][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 24 |
+
warnings.warn(
|
| 25 |
+
|
| 26 |
+
[2026-03-03 17:32:11,332][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 27 |
+
warnings.warn(msg)
|
| 28 |
+
|
| 29 |
+
[2026-03-03 17:33:01,336][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=31` in the `DataLoader` to improve performance.
|
| 30 |
+
|
| 31 |
+
[2026-03-03 17:33:01,337][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 32 |
+
warnings.warn( # warn only once
|
| 33 |
+
|
| 34 |
+
[2026-03-03 17:34:59,889][dinov2][INFO] - using MLP layer as FFN
|
| 35 |
+
[2026-03-03 17:35:05,864][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 36 |
+
warnings.warn(
|
| 37 |
+
|
| 38 |
+
[2026-03-03 17:35:05,864][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 39 |
+
warnings.warn(msg)
|
| 40 |
+
|
| 41 |
+
[2026-03-03 17:35:58,952][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=31` in the `DataLoader` to improve performance.
|
| 42 |
+
|
| 43 |
+
[2026-03-03 17:35:58,954][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 44 |
+
warnings.warn( # warn only once
|
| 45 |
+
|
| 46 |
+
[2026-03-03 17:36:01,340][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 47 |
+
result[selector] = overlay
|
| 48 |
+
|
| 49 |
+
[2026-03-03 17:36:01,350][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
|
| 50 |
+
|
| 51 |
+
[2026-03-03 17:36:01,350][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 52 |
+
warnings.warn(
|
| 53 |
+
|
| 54 |
+
[2026-03-03 17:36:01,351][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 55 |
+
warnings.warn(msg)
|
| 56 |
+
|
| 57 |
+
[2026-03-03 17:36:03,034][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4322.)
|
| 58 |
+
return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
|
| 59 |
+
|
| 60 |
+
[2026-03-03 17:36:03,326][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
|
| 61 |
+
|
| 62 |
+
[2026-03-03 17:36:03,327][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
|
| 63 |
+
|
| 64 |
+
[2026-03-03 17:36:03,328][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
|
| 65 |
+
|
| 66 |
+
[2026-03-03 17:36:03,328][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
|
| 67 |
+
|
| 68 |
+
[2026-03-03 17:36:03,328][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
|
| 69 |
+
|
| 70 |
+
[2026-03-03 17:36:13,075][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
|
| 71 |
+
grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
|
| 72 |
+
bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
|
| 73 |
+
return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
|
| 74 |
+
|
| 75 |
+
[2026-03-03 17:36:13,149][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 76 |
+
result[selector] = overlay
|
| 77 |
+
|
| 78 |
+
[2026-03-03 17:37:46,051][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
|
| 79 |
+
warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
|
| 80 |
+
|
| 81 |
+
[2026-03-03 17:48:34,774][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 82 |
+
result[selector] = overlay
|
| 83 |
+
|
ABLATION_0302_noTgtAlign/train_ddp_process_1.log
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2026-03-03 17:29:48,213][dinov2][INFO] - using MLP layer as FFN
|
| 2 |
+
[2026-03-03 17:30:11,285][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 3 |
+
warnings.warn(
|
| 4 |
+
|
| 5 |
+
[2026-03-03 17:30:11,285][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 6 |
+
warnings.warn(msg)
|
| 7 |
+
|
| 8 |
+
[2026-03-03 17:30:28,361][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 9 |
+
warnings.warn( # warn only once
|
| 10 |
+
|
| 11 |
+
[2026-03-03 17:32:21,762][dinov2][INFO] - using MLP layer as FFN
|
| 12 |
+
[2026-03-03 17:32:48,615][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 13 |
+
warnings.warn(
|
| 14 |
+
|
| 15 |
+
[2026-03-03 17:32:48,616][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 16 |
+
warnings.warn(msg)
|
| 17 |
+
|
| 18 |
+
[2026-03-03 17:33:01,337][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 19 |
+
warnings.warn( # warn only once
|
| 20 |
+
|
| 21 |
+
[2026-03-03 17:35:16,082][dinov2][INFO] - using MLP layer as FFN
|
| 22 |
+
[2026-03-03 17:35:42,645][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 23 |
+
warnings.warn(
|
| 24 |
+
|
| 25 |
+
[2026-03-03 17:35:42,646][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 26 |
+
warnings.warn(msg)
|
| 27 |
+
|
| 28 |
+
[2026-03-03 17:35:58,953][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 29 |
+
warnings.warn( # warn only once
|
| 30 |
+
|
| 31 |
+
[2026-03-03 17:36:13,070][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
|
| 32 |
+
grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
|
| 33 |
+
bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
|
| 34 |
+
return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
|
| 35 |
+
|
| 36 |
+
[2026-03-03 17:36:13,172][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 37 |
+
result[selector] = overlay
|
| 38 |
+
|
| 39 |
+
[2026-03-03 17:37:46,051][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
|
| 40 |
+
warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
|
| 41 |
+
|
| 42 |
+
[2026-03-03 17:48:34,774][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 43 |
+
result[selector] = overlay
|
| 44 |
+
|
ABLATION_0302_noTgtAlign/train_ddp_process_2.log
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2026-03-03 17:29:48,269][dinov2][INFO] - using MLP layer as FFN
|
| 2 |
+
[2026-03-03 17:30:16,899][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 3 |
+
warnings.warn(
|
| 4 |
+
|
| 5 |
+
[2026-03-03 17:30:16,900][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 6 |
+
warnings.warn(msg)
|
| 7 |
+
|
| 8 |
+
[2026-03-03 17:30:28,361][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 9 |
+
warnings.warn( # warn only once
|
| 10 |
+
|
| 11 |
+
[2026-03-03 17:32:21,758][dinov2][INFO] - using MLP layer as FFN
|
| 12 |
+
[2026-03-03 17:32:48,877][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 13 |
+
warnings.warn(
|
| 14 |
+
|
| 15 |
+
[2026-03-03 17:32:48,877][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 16 |
+
warnings.warn(msg)
|
| 17 |
+
|
| 18 |
+
[2026-03-03 17:33:01,337][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 19 |
+
warnings.warn( # warn only once
|
| 20 |
+
|
| 21 |
+
[2026-03-03 17:35:16,091][dinov2][INFO] - using MLP layer as FFN
|
| 22 |
+
[2026-03-03 17:35:47,851][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 23 |
+
warnings.warn(
|
| 24 |
+
|
| 25 |
+
[2026-03-03 17:35:47,851][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 26 |
+
warnings.warn(msg)
|
| 27 |
+
|
| 28 |
+
[2026-03-03 17:35:58,953][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 29 |
+
warnings.warn( # warn only once
|
| 30 |
+
|
| 31 |
+
[2026-03-03 17:36:13,078][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
|
| 32 |
+
grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
|
| 33 |
+
bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
|
| 34 |
+
return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
|
| 35 |
+
|
| 36 |
+
[2026-03-03 17:36:13,182][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 37 |
+
result[selector] = overlay
|
| 38 |
+
|
| 39 |
+
[2026-03-03 17:37:46,051][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
|
| 40 |
+
warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
|
| 41 |
+
|
| 42 |
+
[2026-03-03 17:48:34,774][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 43 |
+
result[selector] = overlay
|
| 44 |
+
|
ABLATION_0302_noTgtAlign/train_ddp_process_3.log
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2026-03-03 17:29:48,206][dinov2][INFO] - using MLP layer as FFN
|
| 2 |
+
[2026-03-03 17:30:16,857][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 3 |
+
warnings.warn(
|
| 4 |
+
|
| 5 |
+
[2026-03-03 17:30:16,857][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 6 |
+
warnings.warn(msg)
|
| 7 |
+
|
| 8 |
+
[2026-03-03 17:30:28,361][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 9 |
+
warnings.warn( # warn only once
|
| 10 |
+
|
| 11 |
+
[2026-03-03 17:32:21,762][dinov2][INFO] - using MLP layer as FFN
|
| 12 |
+
[2026-03-03 17:32:48,656][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 13 |
+
warnings.warn(
|
| 14 |
+
|
| 15 |
+
[2026-03-03 17:32:48,657][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 16 |
+
warnings.warn(msg)
|
| 17 |
+
|
| 18 |
+
[2026-03-03 17:33:01,336][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 19 |
+
warnings.warn( # warn only once
|
| 20 |
+
|
| 21 |
+
[2026-03-03 17:35:16,096][dinov2][INFO] - using MLP layer as FFN
|
| 22 |
+
[2026-03-03 17:35:47,666][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 23 |
+
warnings.warn(
|
| 24 |
+
|
| 25 |
+
[2026-03-03 17:35:47,666][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 26 |
+
warnings.warn(msg)
|
| 27 |
+
|
| 28 |
+
[2026-03-03 17:35:58,953][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 29 |
+
warnings.warn( # warn only once
|
| 30 |
+
|
| 31 |
+
[2026-03-03 17:36:13,076][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
|
| 32 |
+
grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
|
| 33 |
+
bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
|
| 34 |
+
return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
|
| 35 |
+
|
| 36 |
+
[2026-03-03 17:36:13,191][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 37 |
+
result[selector] = overlay
|
| 38 |
+
|
| 39 |
+
[2026-03-03 17:37:46,079][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
|
| 40 |
+
warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
|
| 41 |
+
|
| 42 |
+
[2026-03-03 17:48:34,775][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 43 |
+
result[selector] = overlay
|
| 44 |
+
|
ABLATION_0302_noTgtAlign/train_ddp_process_4.log
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2026-03-03 17:29:48,312][dinov2][INFO] - using MLP layer as FFN
|
| 2 |
+
[2026-03-03 17:30:03,981][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 3 |
+
warnings.warn(
|
| 4 |
+
|
| 5 |
+
[2026-03-03 17:30:03,981][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 6 |
+
warnings.warn(msg)
|
| 7 |
+
|
| 8 |
+
[2026-03-03 17:30:28,361][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 9 |
+
warnings.warn( # warn only once
|
| 10 |
+
|
| 11 |
+
[2026-03-03 17:32:21,721][dinov2][INFO] - using MLP layer as FFN
|
| 12 |
+
[2026-03-03 17:32:48,708][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 13 |
+
warnings.warn(
|
| 14 |
+
|
| 15 |
+
[2026-03-03 17:32:48,708][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 16 |
+
warnings.warn(msg)
|
| 17 |
+
|
| 18 |
+
[2026-03-03 17:33:01,337][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 19 |
+
warnings.warn( # warn only once
|
| 20 |
+
|
| 21 |
+
[2026-03-03 17:35:16,084][dinov2][INFO] - using MLP layer as FFN
|
| 22 |
+
[2026-03-03 17:35:47,371][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 23 |
+
warnings.warn(
|
| 24 |
+
|
| 25 |
+
[2026-03-03 17:35:47,371][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 26 |
+
warnings.warn(msg)
|
| 27 |
+
|
| 28 |
+
[2026-03-03 17:35:58,953][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 29 |
+
warnings.warn( # warn only once
|
| 30 |
+
|
| 31 |
+
[2026-03-03 17:36:13,073][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
|
| 32 |
+
grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
|
| 33 |
+
bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
|
| 34 |
+
return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
|
| 35 |
+
|
| 36 |
+
[2026-03-03 17:36:13,182][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 37 |
+
result[selector] = overlay
|
| 38 |
+
|
| 39 |
+
[2026-03-03 17:37:46,076][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
|
| 40 |
+
warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
|
| 41 |
+
|
| 42 |
+
[2026-03-03 17:48:34,775][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 43 |
+
result[selector] = overlay
|
| 44 |
+
|
ABLATION_0302_noTgtAlign/train_ddp_process_5.log
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2026-03-03 17:29:48,314][dinov2][INFO] - using MLP layer as FFN
|
| 2 |
+
[2026-03-03 17:30:17,098][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 3 |
+
warnings.warn(
|
| 4 |
+
|
| 5 |
+
[2026-03-03 17:30:17,098][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 6 |
+
warnings.warn(msg)
|
| 7 |
+
|
| 8 |
+
[2026-03-03 17:30:28,361][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 9 |
+
warnings.warn( # warn only once
|
| 10 |
+
|
| 11 |
+
[2026-03-03 17:32:21,789][dinov2][INFO] - using MLP layer as FFN
|
| 12 |
+
[2026-03-03 17:32:37,905][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 13 |
+
warnings.warn(
|
| 14 |
+
|
| 15 |
+
[2026-03-03 17:32:37,906][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 16 |
+
warnings.warn(msg)
|
| 17 |
+
|
| 18 |
+
[2026-03-03 17:33:01,336][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 19 |
+
warnings.warn( # warn only once
|
| 20 |
+
|
| 21 |
+
[2026-03-03 17:35:16,102][dinov2][INFO] - using MLP layer as FFN
|
| 22 |
+
[2026-03-03 17:35:45,673][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 23 |
+
warnings.warn(
|
| 24 |
+
|
| 25 |
+
[2026-03-03 17:35:45,676][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 26 |
+
warnings.warn(msg)
|
| 27 |
+
|
| 28 |
+
[2026-03-03 17:35:58,953][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 29 |
+
warnings.warn( # warn only once
|
| 30 |
+
|
| 31 |
+
[2026-03-03 17:36:12,327][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
|
| 32 |
+
grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
|
| 33 |
+
bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
|
| 34 |
+
return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
|
| 35 |
+
|
| 36 |
+
[2026-03-03 17:36:13,191][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 37 |
+
result[selector] = overlay
|
| 38 |
+
|
| 39 |
+
[2026-03-03 17:37:46,079][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
|
| 40 |
+
warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
|
| 41 |
+
|
| 42 |
+
[2026-03-03 17:48:34,774][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 43 |
+
result[selector] = overlay
|
| 44 |
+
|
ABLATION_0302_noTgtAlign/train_ddp_process_6.log
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2026-03-03 17:29:48,277][dinov2][INFO] - using MLP layer as FFN
|
| 2 |
+
[2026-03-03 17:30:16,775][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 3 |
+
warnings.warn(
|
| 4 |
+
|
| 5 |
+
[2026-03-03 17:30:16,775][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 6 |
+
warnings.warn(msg)
|
| 7 |
+
|
| 8 |
+
[2026-03-03 17:30:28,361][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 9 |
+
warnings.warn( # warn only once
|
| 10 |
+
|
| 11 |
+
[2026-03-03 17:32:21,773][dinov2][INFO] - using MLP layer as FFN
|
| 12 |
+
[2026-03-03 17:32:38,406][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 13 |
+
warnings.warn(
|
| 14 |
+
|
| 15 |
+
[2026-03-03 17:32:38,406][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 16 |
+
warnings.warn(msg)
|
| 17 |
+
|
| 18 |
+
[2026-03-03 17:33:01,337][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 19 |
+
warnings.warn( # warn only once
|
| 20 |
+
|
| 21 |
+
[2026-03-03 17:35:16,116][dinov2][INFO] - using MLP layer as FFN
|
| 22 |
+
[2026-03-03 17:35:47,749][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 23 |
+
warnings.warn(
|
| 24 |
+
|
| 25 |
+
[2026-03-03 17:35:47,749][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 26 |
+
warnings.warn(msg)
|
| 27 |
+
|
| 28 |
+
[2026-03-03 17:35:58,953][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 29 |
+
warnings.warn( # warn only once
|
| 30 |
+
|
| 31 |
+
[2026-03-03 17:36:12,560][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
|
| 32 |
+
grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
|
| 33 |
+
bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
|
| 34 |
+
return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
|
| 35 |
+
|
| 36 |
+
[2026-03-03 17:36:13,180][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 37 |
+
result[selector] = overlay
|
| 38 |
+
|
| 39 |
+
[2026-03-03 17:37:46,058][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
|
| 40 |
+
warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
|
| 41 |
+
|
| 42 |
+
[2026-03-03 17:48:34,774][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 43 |
+
result[selector] = overlay
|
| 44 |
+
|
ABLATION_0302_noTgtAlign/train_ddp_process_7.log
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2026-03-03 17:29:48,209][dinov2][INFO] - using MLP layer as FFN
|
| 2 |
+
[2026-03-03 17:30:15,376][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 3 |
+
warnings.warn(
|
| 4 |
+
|
| 5 |
+
[2026-03-03 17:30:15,376][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 6 |
+
warnings.warn(msg)
|
| 7 |
+
|
| 8 |
+
[2026-03-03 17:30:28,361][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 9 |
+
warnings.warn( # warn only once
|
| 10 |
+
|
| 11 |
+
[2026-03-03 17:32:21,793][dinov2][INFO] - using MLP layer as FFN
|
| 12 |
+
[2026-03-03 17:32:48,771][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 13 |
+
warnings.warn(
|
| 14 |
+
|
| 15 |
+
[2026-03-03 17:32:48,771][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 16 |
+
warnings.warn(msg)
|
| 17 |
+
|
| 18 |
+
[2026-03-03 17:33:01,336][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 19 |
+
warnings.warn( # warn only once
|
| 20 |
+
|
| 21 |
+
[2026-03-03 17:35:16,110][dinov2][INFO] - using MLP layer as FFN
|
| 22 |
+
[2026-03-03 17:35:47,614][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 23 |
+
warnings.warn(
|
| 24 |
+
|
| 25 |
+
[2026-03-03 17:35:47,614][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 26 |
+
warnings.warn(msg)
|
| 27 |
+
|
| 28 |
+
[2026-03-03 17:35:58,954][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 29 |
+
warnings.warn( # warn only once
|
| 30 |
+
|
| 31 |
+
[2026-03-03 17:36:12,568][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/autograd/graph.py:829: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
|
| 32 |
+
grad.sizes() = [57, 256, 1, 1], strides() = [256, 1, 256, 256]
|
| 33 |
+
bucket_view.sizes() = [57, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:334.)
|
| 34 |
+
return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
|
| 35 |
+
|
| 36 |
+
[2026-03-03 17:36:13,185][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 37 |
+
result[selector] = overlay
|
| 38 |
+
|
| 39 |
+
[2026-03-03 17:37:46,051][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:209: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
|
| 40 |
+
warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
|
| 41 |
+
|
| 42 |
+
[2026-03-03 17:48:34,774][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 43 |
+
result[selector] = overlay
|
| 44 |
+
|
ABLATION_0302_noTgtAlign/wandb/debug-internal.log
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2026-03-03T17:35:55.708502915Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
|
| 2 |
+
{"time":"2026-03-03T17:35:56.156384015Z","level":"INFO","msg":"stream: created new stream","id":"et94mfhp"}
|
| 3 |
+
{"time":"2026-03-03T17:35:56.156522256Z","level":"INFO","msg":"handler: started","stream_id":"et94mfhp"}
|
| 4 |
+
{"time":"2026-03-03T17:35:56.156785449Z","level":"INFO","msg":"stream: started","id":"et94mfhp"}
|
| 5 |
+
{"time":"2026-03-03T17:35:56.156813659Z","level":"INFO","msg":"sender: started","stream_id":"et94mfhp"}
|
| 6 |
+
{"time":"2026-03-03T17:35:56.156835559Z","level":"INFO","msg":"writer: started","stream_id":"et94mfhp"}
|
ABLATION_0302_noTgtAlign/wandb/debug.log
ADDED
|
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2026-03-03 17:35:55,409 INFO MainThread:870906 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
|
| 2 |
+
2026-03-03 17:35:55,409 INFO MainThread:870906 [wandb_setup.py:_flush():81] Configure stats pid to 870906
|
| 3 |
+
2026-03-03 17:35:55,409 INFO MainThread:870906 [wandb_setup.py:_flush():81] Loading settings from environment variables
|
| 4 |
+
2026-03-03 17:35:55,409 INFO MainThread:870906 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign/wandb/run-20260303_173555-et94mfhp/logs/debug.log
|
| 5 |
+
2026-03-03 17:35:55,409 INFO MainThread:870906 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign/wandb/run-20260303_173555-et94mfhp/logs/debug-internal.log
|
| 6 |
+
2026-03-03 17:35:55,410 INFO MainThread:870906 [wandb_init.py:init():844] calling init triggers
|
| 7 |
+
2026-03-03 17:35:55,410 INFO MainThread:870906 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
|
| 8 |
+
config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'pcd', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'random', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': False, 'voxelize_activate': False, 'use_depth': False}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'density_control_loss': {'error_score': {'weight': 0.0001, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0302_noTgtAlign', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': True, 'vggt_cam_loss': True, 'vggt_distil': False, 'context_view_train': False}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'save_gs': False, 'save_sample_wise_metrics': True, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3, 'target_align': True}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
|
| 9 |
+
2026-03-03 17:35:55,410 INFO MainThread:870906 [wandb_init.py:init():892] starting backend
|
| 10 |
+
2026-03-03 17:35:55,698 INFO MainThread:870906 [wandb_init.py:init():895] sending inform_init request
|
| 11 |
+
2026-03-03 17:35:55,705 INFO MainThread:870906 [wandb_init.py:init():903] backend started and connected
|
| 12 |
+
2026-03-03 17:35:55,709 INFO MainThread:870906 [wandb_init.py:init():973] updated telemetry
|
| 13 |
+
2026-03-03 17:35:55,716 INFO MainThread:870906 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
|
| 14 |
+
2026-03-03 17:35:56,664 INFO MainThread:870906 [wandb_init.py:init():1042] starting run threads in backend
|
| 15 |
+
2026-03-03 17:35:56,791 INFO MainThread:870906 [wandb_run.py:_console_start():2524] atexit reg
|
| 16 |
+
2026-03-03 17:35:56,791 INFO MainThread:870906 [wandb_run.py:_redirect():2373] redirect: wrap_raw
|
| 17 |
+
2026-03-03 17:35:56,791 INFO MainThread:870906 [wandb_run.py:_redirect():2442] Wrapping output streams.
|
| 18 |
+
2026-03-03 17:35:56,791 INFO MainThread:870906 [wandb_run.py:_redirect():2465] Redirects installed.
|
| 19 |
+
2026-03-03 17:35:56,794 INFO MainThread:870906 [wandb_init.py:init():1082] run started, returning control to user process
|
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/config.yaml
ADDED
|
@@ -0,0 +1,311 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
_wandb:
|
| 2 |
+
value:
|
| 3 |
+
cli_version: 0.25.0
|
| 4 |
+
e:
|
| 5 |
+
vpcdp2qfp7cgu3zf4ilibu30q5on1nko:
|
| 6 |
+
args:
|
| 7 |
+
- +experiment=re10k_ablation_24v
|
| 8 |
+
- wandb.mode=online
|
| 9 |
+
- wandb.name=ABLATION_0302_noTgtAlign
|
| 10 |
+
- model.density_control.score_mode=random
|
| 11 |
+
cpu_count: 128
|
| 12 |
+
cpu_count_logical: 256
|
| 13 |
+
cudaVersion: "13.0"
|
| 14 |
+
disk:
|
| 15 |
+
/:
|
| 16 |
+
total: "735513149440"
|
| 17 |
+
used: "700800790528"
|
| 18 |
+
email: dna9041@korea.ac.kr
|
| 19 |
+
executable: /venv/main/bin/python
|
| 20 |
+
git:
|
| 21 |
+
commit: 9dfce172a0f8c7ce85e763899f7ef741ecffc454
|
| 22 |
+
remote: git@github.com:K-nowing/CVPR2026.git
|
| 23 |
+
gpu: NVIDIA H200
|
| 24 |
+
gpu_count: 8
|
| 25 |
+
gpu_nvidia:
|
| 26 |
+
- architecture: Hopper
|
| 27 |
+
cudaCores: 16896
|
| 28 |
+
memoryTotal: "150754820096"
|
| 29 |
+
name: NVIDIA H200
|
| 30 |
+
uuid: GPU-9a20101e-d876-facd-5f05-805081aede41
|
| 31 |
+
- architecture: Hopper
|
| 32 |
+
cudaCores: 16896
|
| 33 |
+
memoryTotal: "150754820096"
|
| 34 |
+
name: NVIDIA H200
|
| 35 |
+
uuid: GPU-84736a77-ee75-3324-e4e1-99cc15bfb5e9
|
| 36 |
+
- architecture: Hopper
|
| 37 |
+
cudaCores: 16896
|
| 38 |
+
memoryTotal: "150754820096"
|
| 39 |
+
name: NVIDIA H200
|
| 40 |
+
uuid: GPU-423d3161-cdc4-3fc0-caee-d15cfaa83ca6
|
| 41 |
+
- architecture: Hopper
|
| 42 |
+
cudaCores: 16896
|
| 43 |
+
memoryTotal: "150754820096"
|
| 44 |
+
name: NVIDIA H200
|
| 45 |
+
uuid: GPU-5b0058b2-cdb9-c952-04f9-87dcaa7ea742
|
| 46 |
+
- architecture: Hopper
|
| 47 |
+
cudaCores: 16896
|
| 48 |
+
memoryTotal: "150754820096"
|
| 49 |
+
name: NVIDIA H200
|
| 50 |
+
uuid: GPU-08b37f98-4603-d483-2f2b-fe5311aa42f2
|
| 51 |
+
- architecture: Hopper
|
| 52 |
+
cudaCores: 16896
|
| 53 |
+
memoryTotal: "150754820096"
|
| 54 |
+
name: NVIDIA H200
|
| 55 |
+
uuid: GPU-03273b5b-2fdd-a5fe-4460-c897334ae464
|
| 56 |
+
- architecture: Hopper
|
| 57 |
+
cudaCores: 16896
|
| 58 |
+
memoryTotal: "150754820096"
|
| 59 |
+
name: NVIDIA H200
|
| 60 |
+
uuid: GPU-292d466c-d00d-25a4-28b6-e6c978d3e70c
|
| 61 |
+
- architecture: Hopper
|
| 62 |
+
cudaCores: 16896
|
| 63 |
+
memoryTotal: "150754820096"
|
| 64 |
+
name: NVIDIA H200
|
| 65 |
+
uuid: GPU-46f38561-3148-e442-7f7f-bfe447bab7fe
|
| 66 |
+
host: e9d3310a05da
|
| 67 |
+
memory:
|
| 68 |
+
total: "1622950240256"
|
| 69 |
+
os: Linux-6.8.0-94-generic-x86_64-with-glibc2.39
|
| 70 |
+
program: -m src.main
|
| 71 |
+
python: CPython 3.12.12
|
| 72 |
+
root: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign
|
| 73 |
+
startedAt: "2026-03-03T17:26:05.471917Z"
|
| 74 |
+
writerId: vpcdp2qfp7cgu3zf4ilibu30q5on1nko
|
| 75 |
+
m:
|
| 76 |
+
- "1": trainer/global_step
|
| 77 |
+
"6":
|
| 78 |
+
- 3
|
| 79 |
+
"7": []
|
| 80 |
+
- "2": '*'
|
| 81 |
+
"5": 1
|
| 82 |
+
"6":
|
| 83 |
+
- 1
|
| 84 |
+
"7": []
|
| 85 |
+
python_version: 3.12.12
|
| 86 |
+
t:
|
| 87 |
+
"1":
|
| 88 |
+
- 1
|
| 89 |
+
- 41
|
| 90 |
+
- 49
|
| 91 |
+
- 50
|
| 92 |
+
- 106
|
| 93 |
+
"2":
|
| 94 |
+
- 1
|
| 95 |
+
- 41
|
| 96 |
+
- 49
|
| 97 |
+
- 50
|
| 98 |
+
- 106
|
| 99 |
+
"3":
|
| 100 |
+
- 2
|
| 101 |
+
- 7
|
| 102 |
+
- 13
|
| 103 |
+
- 15
|
| 104 |
+
- 16
|
| 105 |
+
- 66
|
| 106 |
+
"4": 3.12.12
|
| 107 |
+
"5": 0.25.0
|
| 108 |
+
"12": 0.25.0
|
| 109 |
+
"13": linux-x86_64
|
| 110 |
+
checkpointing:
|
| 111 |
+
value:
|
| 112 |
+
every_n_train_steps: 1500
|
| 113 |
+
load: null
|
| 114 |
+
save_top_k: 2
|
| 115 |
+
save_weights_only: false
|
| 116 |
+
data_loader:
|
| 117 |
+
value:
|
| 118 |
+
test:
|
| 119 |
+
batch_size: 1
|
| 120 |
+
num_workers: 4
|
| 121 |
+
persistent_workers: false
|
| 122 |
+
seed: 2345
|
| 123 |
+
train:
|
| 124 |
+
batch_size: 16
|
| 125 |
+
num_workers: 16
|
| 126 |
+
persistent_workers: true
|
| 127 |
+
seed: 1234
|
| 128 |
+
val:
|
| 129 |
+
batch_size: 1
|
| 130 |
+
num_workers: 1
|
| 131 |
+
persistent_workers: true
|
| 132 |
+
seed: 3456
|
| 133 |
+
dataset:
|
| 134 |
+
value:
|
| 135 |
+
re10k:
|
| 136 |
+
augment: true
|
| 137 |
+
background_color:
|
| 138 |
+
- 0
|
| 139 |
+
- 0
|
| 140 |
+
- 0
|
| 141 |
+
baseline_max: 1e+10
|
| 142 |
+
baseline_min: 0.001
|
| 143 |
+
cameras_are_circular: false
|
| 144 |
+
dynamic_context_views: true
|
| 145 |
+
input_image_shape:
|
| 146 |
+
- 256
|
| 147 |
+
- 256
|
| 148 |
+
make_baseline_1: true
|
| 149 |
+
max_context_views_per_gpu: 24
|
| 150 |
+
max_fov: 100
|
| 151 |
+
name: re10k
|
| 152 |
+
original_image_shape:
|
| 153 |
+
- 360
|
| 154 |
+
- 640
|
| 155 |
+
overfit_to_scene: null
|
| 156 |
+
relative_pose: true
|
| 157 |
+
roots:
|
| 158 |
+
- datasets/re10k
|
| 159 |
+
skip_bad_shape: true
|
| 160 |
+
view_sampler:
|
| 161 |
+
initial_max_distance_between_context_views: 25
|
| 162 |
+
initial_min_distance_between_context_views: 25
|
| 163 |
+
max_distance_between_context_views: 90
|
| 164 |
+
min_distance_between_context_views: 45
|
| 165 |
+
min_distance_to_context_views: 0
|
| 166 |
+
name: bounded
|
| 167 |
+
num_context_views: 2
|
| 168 |
+
num_target_set: 3
|
| 169 |
+
num_target_views: 4
|
| 170 |
+
same_target_gap: false
|
| 171 |
+
target_align: true
|
| 172 |
+
warm_up_steps: 1000
|
| 173 |
+
density_control_loss:
|
| 174 |
+
value:
|
| 175 |
+
error_score:
|
| 176 |
+
grad_scale: 10000
|
| 177 |
+
log_scale: false
|
| 178 |
+
mode: original
|
| 179 |
+
weight: 0.0001
|
| 180 |
+
direct_loss:
|
| 181 |
+
value:
|
| 182 |
+
l1:
|
| 183 |
+
weight: 0.8
|
| 184 |
+
ssim:
|
| 185 |
+
weight: 0.2
|
| 186 |
+
mode:
|
| 187 |
+
value: train
|
| 188 |
+
model:
|
| 189 |
+
value:
|
| 190 |
+
decoder:
|
| 191 |
+
background_color:
|
| 192 |
+
- 0
|
| 193 |
+
- 0
|
| 194 |
+
- 0
|
| 195 |
+
make_scale_invariant: false
|
| 196 |
+
name: splatting_cuda
|
| 197 |
+
density_control:
|
| 198 |
+
aggregation_mode: mean
|
| 199 |
+
aux_refine: false
|
| 200 |
+
grad_mode: absgrad
|
| 201 |
+
gs_param_dim: 256
|
| 202 |
+
latent_dim: 128
|
| 203 |
+
mean_dim: 32
|
| 204 |
+
name: density_control_module
|
| 205 |
+
num_heads: 1
|
| 206 |
+
num_latents: 64
|
| 207 |
+
num_level: 3
|
| 208 |
+
num_self_attn_per_block: 2
|
| 209 |
+
refine_error: false
|
| 210 |
+
refinement_hidden_dim: 32
|
| 211 |
+
refinement_layer_num: 1
|
| 212 |
+
refinement_type: voxelize
|
| 213 |
+
score_mode: random
|
| 214 |
+
use_depth: false
|
| 215 |
+
use_mean_features: true
|
| 216 |
+
use_refine_module: false
|
| 217 |
+
voxel_size: 0.001
|
| 218 |
+
voxelize_activate: false
|
| 219 |
+
encoder:
|
| 220 |
+
align_corners: false
|
| 221 |
+
gs_param_dim: 256
|
| 222 |
+
head_mode: pcd
|
| 223 |
+
input_image_shape:
|
| 224 |
+
- 518
|
| 225 |
+
- 518
|
| 226 |
+
name: dcsplat
|
| 227 |
+
num_level: 3
|
| 228 |
+
use_voxelize: true
|
| 229 |
+
optimizer:
|
| 230 |
+
value:
|
| 231 |
+
accumulate: 1
|
| 232 |
+
backbone_lr_multiplier: 0.1
|
| 233 |
+
backbone_trainable: T+H
|
| 234 |
+
lr: 0.0002
|
| 235 |
+
warm_up_steps: 25
|
| 236 |
+
render_loss:
|
| 237 |
+
value:
|
| 238 |
+
lpips:
|
| 239 |
+
apply_after_step: 0
|
| 240 |
+
weight: 0.05
|
| 241 |
+
mse:
|
| 242 |
+
weight: 1
|
| 243 |
+
seed:
|
| 244 |
+
value: 111123
|
| 245 |
+
test:
|
| 246 |
+
value:
|
| 247 |
+
align_pose: false
|
| 248 |
+
compute_scores: true
|
| 249 |
+
error_threshold: 0.4
|
| 250 |
+
error_threshold_list:
|
| 251 |
+
- 0.2
|
| 252 |
+
- 0.4
|
| 253 |
+
- 0.6
|
| 254 |
+
- 0.8
|
| 255 |
+
- 1
|
| 256 |
+
nvs_view_N_list:
|
| 257 |
+
- 3
|
| 258 |
+
- 6
|
| 259 |
+
- 16
|
| 260 |
+
- 32
|
| 261 |
+
- 64
|
| 262 |
+
output_path: test/ablation/re10k
|
| 263 |
+
pose_align_steps: 100
|
| 264 |
+
pred_intrinsic: false
|
| 265 |
+
rot_opt_lr: 0.005
|
| 266 |
+
save_active_mask_image: false
|
| 267 |
+
save_compare: false
|
| 268 |
+
save_error_score_image: false
|
| 269 |
+
save_gs: false
|
| 270 |
+
save_image: false
|
| 271 |
+
save_sample_wise_metrics: true
|
| 272 |
+
save_video: false
|
| 273 |
+
threshold_mode: ratio
|
| 274 |
+
trans_opt_lr: 0.005
|
| 275 |
+
train:
|
| 276 |
+
value:
|
| 277 |
+
align_corners: false
|
| 278 |
+
beta_dist_param:
|
| 279 |
+
- 0.5
|
| 280 |
+
- 4
|
| 281 |
+
cam_scale_mode: sum
|
| 282 |
+
camera_loss: 10
|
| 283 |
+
context_view_train: false
|
| 284 |
+
ext_scale_detach: false
|
| 285 |
+
extended_visualization: false
|
| 286 |
+
intrinsic_scaling: false
|
| 287 |
+
one_sample_validation: null
|
| 288 |
+
print_log_every_n_steps: 10
|
| 289 |
+
scene_scale_reg_loss: 0.01
|
| 290 |
+
train_aux: true
|
| 291 |
+
train_gs_num: 1
|
| 292 |
+
train_target_set: true
|
| 293 |
+
use_refine_aux: false
|
| 294 |
+
verbose: false
|
| 295 |
+
vggt_cam_loss: true
|
| 296 |
+
vggt_distil: false
|
| 297 |
+
trainer:
|
| 298 |
+
value:
|
| 299 |
+
gradient_clip_val: 0.5
|
| 300 |
+
max_steps: 3001
|
| 301 |
+
num_nodes: 1
|
| 302 |
+
val_check_interval: 250
|
| 303 |
+
wandb:
|
| 304 |
+
value:
|
| 305 |
+
entity: scene-representation-group
|
| 306 |
+
mode: online
|
| 307 |
+
name: ABLATION_0302_noTgtAlign
|
| 308 |
+
project: DCSplat
|
| 309 |
+
tags:
|
| 310 |
+
- re10k
|
| 311 |
+
- 256x256
|
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/output.log
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0]
|
| 2 |
+
|
| 3 |
+
| Name | Type | Params | Mode
|
| 4 |
+
------------------------------------------------------------------------
|
| 5 |
+
0 | encoder | OurSplat | 888 M | train
|
| 6 |
+
1 | density_control_module | DensityControlModule | 0 | train
|
| 7 |
+
2 | decoder | DecoderSplattingCUDA | 0 | train
|
| 8 |
+
3 | render_losses | ModuleList | 0 | train
|
| 9 |
+
4 | density_control_losses | ModuleList | 0 | train
|
| 10 |
+
5 | direct_losses | ModuleList | 0 | train
|
| 11 |
+
------------------------------------------------------------------------
|
| 12 |
+
888 M Trainable params
|
| 13 |
+
0 Non-trainable params
|
| 14 |
+
888 M Total params
|
| 15 |
+
3,553.933 Total estimated model params size (MB)
|
| 16 |
+
1202 Modules in train mode
|
| 17 |
+
522 Modules in eval mode
|
| 18 |
+
Sanity Checking: | | 0/? [00:00<?, ?it/s][2026-03-03 17:26:08,283][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=255` in the `DataLoader` to improve performance.
|
| 19 |
+
|
| 20 |
+
Validation epoch start on rank 0
|
| 21 |
+
Sanity Checking DataLoader 0: 0%| | 0/1 [00:00<?, ?it/s]validation step 0; scene = ['306e2b7785657539'];
|
| 22 |
+
> /workspace/code/CVPR2026/src/model/model_wrapper.py(1570)get_normalized_target_cam()
|
| 23 |
+
-> gt_left_c2w = gt_context_c2w[:, 0:1, :, :]
|
| 24 |
+
tensor([[0.0369]], device='cuda:0')
|
| 25 |
+
tensor([[0.0369, 0.0369, 0.0369, 0.0369]], device='cuda:0')
|
| 26 |
+
tensor([[[1, 0],
|
| 27 |
+
[0, 1],
|
| 28 |
+
[0, 1],
|
| 29 |
+
[0, 1]]], device='cuda:0')
|
| 30 |
+
tensor([[[[0.8595, 0.0000, 0.5000],
|
| 31 |
+
[0.0000, 0.8597, 0.5000],
|
| 32 |
+
[0.0000, 0.0000, 1.0000]],
|
| 33 |
+
|
| 34 |
+
[[0.8595, 0.0000, 0.5000],
|
| 35 |
+
[0.0000, 0.8597, 0.5000],
|
| 36 |
+
[0.0000, 0.0000, 1.0000]],
|
| 37 |
+
|
| 38 |
+
[[0.8595, 0.0000, 0.5000],
|
| 39 |
+
[0.0000, 0.8597, 0.5000],
|
| 40 |
+
[0.0000, 0.0000, 1.0000]],
|
| 41 |
+
|
| 42 |
+
[[0.8595, 0.0000, 0.5000],
|
| 43 |
+
[0.0000, 0.8597, 0.5000],
|
| 44 |
+
[0.0000, 0.0000, 1.0000]]]], device='cuda:0')
|
| 45 |
+
tensor([[[[0.8837, 0.0000, 0.5000],
|
| 46 |
+
[0.0000, 0.8834, 0.5000],
|
| 47 |
+
[0.0000, 0.0000, 1.0000]],
|
| 48 |
+
|
| 49 |
+
[[0.8721, 0.0000, 0.5000],
|
| 50 |
+
[0.0000, 0.8713, 0.5000],
|
| 51 |
+
[0.0000, 0.0000, 1.0000]]]], device='cuda:0')
|
| 52 |
+
torch.Size([1, 4, 3, 3])
|
| 53 |
+
torch.Size([1, 2, 3, 3])
|
| 54 |
+
tensor([[[[0.8779, 0.0000, 0.5000],
|
| 55 |
+
[0.0000, 0.8773, 0.5000],
|
| 56 |
+
[0.0000, 0.0000, 1.0000]],
|
| 57 |
+
|
| 58 |
+
[[0.8779, 0.0000, 0.5000],
|
| 59 |
+
[0.0000, 0.8773, 0.5000],
|
| 60 |
+
[0.0000, 0.0000, 1.0000]],
|
| 61 |
+
|
| 62 |
+
[[0.8779, 0.0000, 0.5000],
|
| 63 |
+
[0.0000, 0.8773, 0.5000],
|
| 64 |
+
[0.0000, 0.0000, 1.0000]],
|
| 65 |
+
|
| 66 |
+
[[0.8779, 0.0000, 0.5000],
|
| 67 |
+
[0.0000, 0.8773, 0.5000],
|
| 68 |
+
[0.0000, 0.0000, 1.0000]]]], device='cuda:0')
|
| 69 |
+
Error executing job with overrides: ['+experiment=re10k_ablation_24v', 'wandb.mode=online', 'wandb.name=ABLATION_0302_noTgtAlign', 'model.density_control.score_mode=random']
|
| 70 |
+
|
| 71 |
+
|
| 72 |
+
Received interrupt signal. Cleaning up...
|
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/requirements.txt
ADDED
|
@@ -0,0 +1,173 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
wheel==0.45.1
|
| 2 |
+
pytz==2025.2
|
| 3 |
+
easydict==1.13
|
| 4 |
+
antlr4-python3-runtime==4.9.3
|
| 5 |
+
wadler_lindig==0.1.7
|
| 6 |
+
networkx==3.4.2
|
| 7 |
+
urllib3==2.5.0
|
| 8 |
+
tzdata==2025.2
|
| 9 |
+
typing-inspection==0.4.1
|
| 10 |
+
tabulate==0.9.0
|
| 11 |
+
smmap==5.0.2
|
| 12 |
+
setuptools==78.1.1
|
| 13 |
+
safetensors==0.5.3
|
| 14 |
+
multidict==6.6.4
|
| 15 |
+
PyYAML==6.0.2
|
| 16 |
+
PySocks==1.7.1
|
| 17 |
+
pyparsing==3.2.5
|
| 18 |
+
pydantic_core==2.33.2
|
| 19 |
+
pycparser==2.23
|
| 20 |
+
protobuf==6.32.1
|
| 21 |
+
propcache==0.3.2
|
| 22 |
+
proglog==0.1.12
|
| 23 |
+
platformdirs==4.4.0
|
| 24 |
+
pip==25.2
|
| 25 |
+
mdurl==0.1.2
|
| 26 |
+
pillow==10.4.0
|
| 27 |
+
packaging==24.2
|
| 28 |
+
opt_einsum==3.4.0
|
| 29 |
+
frozenlist==1.7.0
|
| 30 |
+
numpy==1.26.4
|
| 31 |
+
ninja==1.13.0
|
| 32 |
+
MarkupSafe==3.0.2
|
| 33 |
+
kornia_rs==0.1.9
|
| 34 |
+
kiwisolver==1.4.9
|
| 35 |
+
imageio-ffmpeg==0.6.0
|
| 36 |
+
idna==3.7
|
| 37 |
+
fsspec==2024.6.1
|
| 38 |
+
hf-xet==1.1.10
|
| 39 |
+
gmpy2==2.2.1
|
| 40 |
+
fonttools==4.60.0
|
| 41 |
+
triton==3.4.0
|
| 42 |
+
filelock==3.17.0
|
| 43 |
+
einops==0.8.1
|
| 44 |
+
decorator==4.4.2
|
| 45 |
+
dacite==1.9.2
|
| 46 |
+
cycler==0.12.1
|
| 47 |
+
colorama==0.4.6
|
| 48 |
+
click==8.3.0
|
| 49 |
+
nvidia-nvtx-cu12==12.8.90
|
| 50 |
+
charset-normalizer==3.3.2
|
| 51 |
+
certifi==2025.8.3
|
| 52 |
+
beartype==0.19.0
|
| 53 |
+
attrs==25.3.0
|
| 54 |
+
async-timeout==5.0.1
|
| 55 |
+
annotated-types==0.7.0
|
| 56 |
+
aiohappyeyeballs==2.6.1
|
| 57 |
+
yarl==1.20.1
|
| 58 |
+
tifffile==2025.5.10
|
| 59 |
+
sentry-sdk==2.39.0
|
| 60 |
+
scipy==1.15.3
|
| 61 |
+
pydantic==2.11.9
|
| 62 |
+
pandas==2.3.2
|
| 63 |
+
opencv-python==4.11.0.86
|
| 64 |
+
omegaconf==2.3.0
|
| 65 |
+
markdown-it-py==4.0.0
|
| 66 |
+
lightning-utilities==0.14.3
|
| 67 |
+
lazy_loader==0.4
|
| 68 |
+
jaxtyping==0.2.37
|
| 69 |
+
imageio==2.37.0
|
| 70 |
+
gitdb==4.0.12
|
| 71 |
+
contourpy==1.3.2
|
| 72 |
+
colorspacious==1.1.2
|
| 73 |
+
cffi==1.17.1
|
| 74 |
+
aiosignal==1.4.0
|
| 75 |
+
scikit-video==1.1.11
|
| 76 |
+
scikit-image==0.25.2
|
| 77 |
+
rich==14.1.0
|
| 78 |
+
moviepy==1.0.3
|
| 79 |
+
matplotlib==3.10.6
|
| 80 |
+
hydra-core==1.3.2
|
| 81 |
+
huggingface-hub==0.35.1
|
| 82 |
+
GitPython==3.1.45
|
| 83 |
+
brotlicffi==1.0.9.2
|
| 84 |
+
aiohttp==3.12.15
|
| 85 |
+
torchmetrics==1.8.2
|
| 86 |
+
opt-einsum-fx==0.1.4
|
| 87 |
+
kornia==0.8.1
|
| 88 |
+
pytorch-lightning==2.5.1
|
| 89 |
+
lpips==0.1.4
|
| 90 |
+
e3nn==0.6.0
|
| 91 |
+
lightning==2.5.1
|
| 92 |
+
gsplat==1.5.3
|
| 93 |
+
nvidia-cusparselt-cu12==0.7.1
|
| 94 |
+
nvidia-nvjitlink-cu12==12.8.93
|
| 95 |
+
nvidia-nccl-cu12==2.27.3
|
| 96 |
+
nvidia-curand-cu12==10.3.9.90
|
| 97 |
+
nvidia-cufile-cu12==1.13.1.3
|
| 98 |
+
nvidia-cuda-runtime-cu12==12.8.90
|
| 99 |
+
nvidia-cuda-nvrtc-cu12==12.8.93
|
| 100 |
+
nvidia-cuda-cupti-cu12==12.8.90
|
| 101 |
+
nvidia-cublas-cu12==12.8.4.1
|
| 102 |
+
nvidia-cusparse-cu12==12.5.8.93
|
| 103 |
+
nvidia-cufft-cu12==11.3.3.83
|
| 104 |
+
nvidia-cudnn-cu12==9.10.2.21
|
| 105 |
+
nvidia-cusolver-cu12==11.7.3.90
|
| 106 |
+
torch==2.8.0+cu128
|
| 107 |
+
torchvision==0.23.0+cu128
|
| 108 |
+
torchaudio==2.8.0+cu128
|
| 109 |
+
torch_scatter==2.1.2+pt28cu128
|
| 110 |
+
plyfile==1.1.3
|
| 111 |
+
wandb==0.25.0
|
| 112 |
+
cuda-bindings==13.0.3
|
| 113 |
+
cuda-pathfinder==1.3.3
|
| 114 |
+
Jinja2==3.1.6
|
| 115 |
+
mpmath==1.3.0
|
| 116 |
+
nvidia-cublas==13.1.0.3
|
| 117 |
+
nvidia-cuda-cupti==13.0.85
|
| 118 |
+
nvidia-cuda-nvrtc==13.0.88
|
| 119 |
+
nvidia-cuda-runtime==13.0.96
|
| 120 |
+
nvidia-cudnn-cu13==9.15.1.9
|
| 121 |
+
nvidia-cufft==12.0.0.61
|
| 122 |
+
nvidia-cufile==1.15.1.6
|
| 123 |
+
nvidia-curand==10.4.0.35
|
| 124 |
+
nvidia-cusolver==12.0.4.66
|
| 125 |
+
nvidia-cusparse==12.6.3.3
|
| 126 |
+
nvidia-cusparselt-cu13==0.8.0
|
| 127 |
+
nvidia-nccl-cu13==2.28.9
|
| 128 |
+
nvidia-nvjitlink==13.0.88
|
| 129 |
+
nvidia-nvshmem-cu13==3.4.5
|
| 130 |
+
nvidia-nvtx==13.0.85
|
| 131 |
+
requests==2.32.5
|
| 132 |
+
sentencepiece==0.2.1
|
| 133 |
+
sympy==1.14.0
|
| 134 |
+
torchcodec==0.10.0
|
| 135 |
+
torchdata==0.10.0
|
| 136 |
+
torchtext==0.6.0
|
| 137 |
+
anyio==4.12.0
|
| 138 |
+
asttokens==3.0.1
|
| 139 |
+
comm==0.2.3
|
| 140 |
+
debugpy==1.8.19
|
| 141 |
+
executing==2.2.1
|
| 142 |
+
h11==0.16.0
|
| 143 |
+
httpcore==1.0.9
|
| 144 |
+
httpx==0.28.1
|
| 145 |
+
ipykernel==7.1.0
|
| 146 |
+
ipython==9.8.0
|
| 147 |
+
ipython_pygments_lexers==1.1.1
|
| 148 |
+
ipywidgets==8.1.8
|
| 149 |
+
jedi==0.19.2
|
| 150 |
+
jupyter_client==8.7.0
|
| 151 |
+
jupyter_core==5.9.1
|
| 152 |
+
jupyterlab_widgets==3.0.16
|
| 153 |
+
matplotlib-inline==0.2.1
|
| 154 |
+
nest-asyncio==1.6.0
|
| 155 |
+
parso==0.8.5
|
| 156 |
+
pexpect==4.9.0
|
| 157 |
+
prompt_toolkit==3.0.52
|
| 158 |
+
psutil==7.2.1
|
| 159 |
+
ptyprocess==0.7.0
|
| 160 |
+
pure_eval==0.2.3
|
| 161 |
+
Pygments==2.19.2
|
| 162 |
+
python-dateutil==2.9.0.post0
|
| 163 |
+
pyzmq==27.1.0
|
| 164 |
+
shellingham==1.5.4
|
| 165 |
+
six==1.17.0
|
| 166 |
+
stack-data==0.6.3
|
| 167 |
+
tornado==6.5.4
|
| 168 |
+
tqdm==4.67.1
|
| 169 |
+
traitlets==5.14.3
|
| 170 |
+
typer-slim==0.21.0
|
| 171 |
+
typing_extensions==4.15.0
|
| 172 |
+
wcwidth==0.2.14
|
| 173 |
+
widgetsnbextension==4.0.15
|
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/wandb-metadata.json
ADDED
|
@@ -0,0 +1,93 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"os": "Linux-6.8.0-94-generic-x86_64-with-glibc2.39",
|
| 3 |
+
"python": "CPython 3.12.12",
|
| 4 |
+
"startedAt": "2026-03-03T17:26:05.471917Z",
|
| 5 |
+
"args": [
|
| 6 |
+
"+experiment=re10k_ablation_24v",
|
| 7 |
+
"wandb.mode=online",
|
| 8 |
+
"wandb.name=ABLATION_0302_noTgtAlign",
|
| 9 |
+
"model.density_control.score_mode=random"
|
| 10 |
+
],
|
| 11 |
+
"program": "-m src.main",
|
| 12 |
+
"git": {
|
| 13 |
+
"remote": "git@github.com:K-nowing/CVPR2026.git",
|
| 14 |
+
"commit": "9dfce172a0f8c7ce85e763899f7ef741ecffc454"
|
| 15 |
+
},
|
| 16 |
+
"email": "dna9041@korea.ac.kr",
|
| 17 |
+
"root": "/workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign",
|
| 18 |
+
"host": "e9d3310a05da",
|
| 19 |
+
"executable": "/venv/main/bin/python",
|
| 20 |
+
"cpu_count": 128,
|
| 21 |
+
"cpu_count_logical": 256,
|
| 22 |
+
"gpu": "NVIDIA H200",
|
| 23 |
+
"gpu_count": 8,
|
| 24 |
+
"disk": {
|
| 25 |
+
"/": {
|
| 26 |
+
"total": "735513149440",
|
| 27 |
+
"used": "700800790528"
|
| 28 |
+
}
|
| 29 |
+
},
|
| 30 |
+
"memory": {
|
| 31 |
+
"total": "1622950240256"
|
| 32 |
+
},
|
| 33 |
+
"gpu_nvidia": [
|
| 34 |
+
{
|
| 35 |
+
"name": "NVIDIA H200",
|
| 36 |
+
"memoryTotal": "150754820096",
|
| 37 |
+
"cudaCores": 16896,
|
| 38 |
+
"architecture": "Hopper",
|
| 39 |
+
"uuid": "GPU-9a20101e-d876-facd-5f05-805081aede41"
|
| 40 |
+
},
|
| 41 |
+
{
|
| 42 |
+
"name": "NVIDIA H200",
|
| 43 |
+
"memoryTotal": "150754820096",
|
| 44 |
+
"cudaCores": 16896,
|
| 45 |
+
"architecture": "Hopper",
|
| 46 |
+
"uuid": "GPU-84736a77-ee75-3324-e4e1-99cc15bfb5e9"
|
| 47 |
+
},
|
| 48 |
+
{
|
| 49 |
+
"name": "NVIDIA H200",
|
| 50 |
+
"memoryTotal": "150754820096",
|
| 51 |
+
"cudaCores": 16896,
|
| 52 |
+
"architecture": "Hopper",
|
| 53 |
+
"uuid": "GPU-423d3161-cdc4-3fc0-caee-d15cfaa83ca6"
|
| 54 |
+
},
|
| 55 |
+
{
|
| 56 |
+
"name": "NVIDIA H200",
|
| 57 |
+
"memoryTotal": "150754820096",
|
| 58 |
+
"cudaCores": 16896,
|
| 59 |
+
"architecture": "Hopper",
|
| 60 |
+
"uuid": "GPU-5b0058b2-cdb9-c952-04f9-87dcaa7ea742"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"name": "NVIDIA H200",
|
| 64 |
+
"memoryTotal": "150754820096",
|
| 65 |
+
"cudaCores": 16896,
|
| 66 |
+
"architecture": "Hopper",
|
| 67 |
+
"uuid": "GPU-08b37f98-4603-d483-2f2b-fe5311aa42f2"
|
| 68 |
+
},
|
| 69 |
+
{
|
| 70 |
+
"name": "NVIDIA H200",
|
| 71 |
+
"memoryTotal": "150754820096",
|
| 72 |
+
"cudaCores": 16896,
|
| 73 |
+
"architecture": "Hopper",
|
| 74 |
+
"uuid": "GPU-03273b5b-2fdd-a5fe-4460-c897334ae464"
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"name": "NVIDIA H200",
|
| 78 |
+
"memoryTotal": "150754820096",
|
| 79 |
+
"cudaCores": 16896,
|
| 80 |
+
"architecture": "Hopper",
|
| 81 |
+
"uuid": "GPU-292d466c-d00d-25a4-28b6-e6c978d3e70c"
|
| 82 |
+
},
|
| 83 |
+
{
|
| 84 |
+
"name": "NVIDIA H200",
|
| 85 |
+
"memoryTotal": "150754820096",
|
| 86 |
+
"cudaCores": 16896,
|
| 87 |
+
"architecture": "Hopper",
|
| 88 |
+
"uuid": "GPU-46f38561-3148-e442-7f7f-bfe447bab7fe"
|
| 89 |
+
}
|
| 90 |
+
],
|
| 91 |
+
"cudaVersion": "13.0",
|
| 92 |
+
"writerId": "vpcdp2qfp7cgu3zf4ilibu30q5on1nko"
|
| 93 |
+
}
|
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/files/wandb-summary.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"_wandb":{"runtime":190},"_runtime":190}
|
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/logs/debug-core.log
ADDED
|
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2026-03-03T17:26:05.544279566Z","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpnepv657f/port-862779.txt","pid":862779,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
|
| 2 |
+
{"time":"2026-03-03T17:26:05.545034072Z","level":"INFO","msg":"server: will exit if parent process dies","ppid":862779}
|
| 3 |
+
{"time":"2026-03-03T17:26:05.545012722Z","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-862779-863051-2889925930/socket","Net":"unix"}}
|
| 4 |
+
{"time":"2026-03-03T17:26:05.717282063Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
|
| 5 |
+
{"time":"2026-03-03T17:26:05.726827389Z","level":"INFO","msg":"handleInformInit: received","streamId":"beokhgg2","id":"1(@)"}
|
| 6 |
+
{"time":"2026-03-03T17:26:06.186040503Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"beokhgg2","id":"1(@)"}
|
| 7 |
+
{"time":"2026-03-03T17:26:12.470381356Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"ahylq6e9uxwi"}
|
| 8 |
+
{"time":"2026-03-03T17:29:17.914358058Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"ahylq6e9uxwi"}
|
| 9 |
+
{"time":"2026-03-03T17:29:19.452510575Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"ahylq6e9uxwi"}
|
| 10 |
+
{"time":"2026-03-03T17:29:19.453368693Z","level":"INFO","msg":"handleInformFinish: finish message received","streamId":"beokhgg2","id":"1(@)"}
|
| 11 |
+
{"time":"2026-03-03T17:29:19.453861608Z","level":"INFO","msg":"handleInformFinish: stream closed","streamId":"beokhgg2","id":"1(@)"}
|
| 12 |
+
{"time":"2026-03-03T17:29:19.454484434Z","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
|
| 13 |
+
{"time":"2026-03-03T17:29:19.454514804Z","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
|
| 14 |
+
{"time":"2026-03-03T17:29:19.454527464Z","level":"INFO","msg":"server is shutting down"}
|
| 15 |
+
{"time":"2026-03-03T17:29:19.454569715Z","level":"INFO","msg":"connection: closing","id":"1(@)"}
|
| 16 |
+
{"time":"2026-03-03T17:29:19.454680076Z","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
|
| 17 |
+
{"time":"2026-03-03T17:29:19.454709016Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
|
| 18 |
+
{"time":"2026-03-03T17:29:19.454665865Z","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-862779-863051-2889925930/socket","Net":"unix"}}
|
| 19 |
+
{"time":"2026-03-03T17:29:19.454730766Z","level":"INFO","msg":"server is closed"}
|
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/logs/debug-internal.log
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2026-03-03T17:26:05.727056141Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
|
| 2 |
+
{"time":"2026-03-03T17:26:06.185659999Z","level":"INFO","msg":"stream: created new stream","id":"beokhgg2"}
|
| 3 |
+
{"time":"2026-03-03T17:26:06.185780721Z","level":"INFO","msg":"handler: started","stream_id":"beokhgg2"}
|
| 4 |
+
{"time":"2026-03-03T17:26:06.186025753Z","level":"INFO","msg":"stream: started","id":"beokhgg2"}
|
| 5 |
+
{"time":"2026-03-03T17:26:06.186049093Z","level":"INFO","msg":"writer: started","stream_id":"beokhgg2"}
|
| 6 |
+
{"time":"2026-03-03T17:26:06.186076323Z","level":"INFO","msg":"sender: started","stream_id":"beokhgg2"}
|
| 7 |
+
{"time":"2026-03-03T17:29:19.172521278Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
|
| 8 |
+
{"time":"2026-03-03T17:29:19.449459346Z","level":"INFO","msg":"handler: operation stats","stats":{}}
|
| 9 |
+
{"time":"2026-03-03T17:29:19.453393683Z","level":"INFO","msg":"stream: closing","id":"beokhgg2"}
|
| 10 |
+
{"time":"2026-03-03T17:29:19.453413883Z","level":"INFO","msg":"handler: closed","stream_id":"beokhgg2"}
|
| 11 |
+
{"time":"2026-03-03T17:29:19.453531935Z","level":"INFO","msg":"sender: closed","stream_id":"beokhgg2"}
|
| 12 |
+
{"time":"2026-03-03T17:29:19.453555075Z","level":"INFO","msg":"stream: closed","id":"beokhgg2"}
|
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/logs/debug.log
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2026-03-03 17:26:05,473 INFO MainThread:862779 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
|
| 2 |
+
2026-03-03 17:26:05,473 INFO MainThread:862779 [wandb_setup.py:_flush():81] Configure stats pid to 862779
|
| 3 |
+
2026-03-03 17:26:05,473 INFO MainThread:862779 [wandb_setup.py:_flush():81] Loading settings from environment variables
|
| 4 |
+
2026-03-03 17:26:05,473 INFO MainThread:862779 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/logs/debug.log
|
| 5 |
+
2026-03-03 17:26:05,473 INFO MainThread:862779 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/logs/debug-internal.log
|
| 6 |
+
2026-03-03 17:26:05,473 INFO MainThread:862779 [wandb_init.py:init():844] calling init triggers
|
| 7 |
+
2026-03-03 17:26:05,473 INFO MainThread:862779 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
|
| 8 |
+
config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'pcd', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'random', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': False, 'voxelize_activate': False, 'use_depth': False}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'density_control_loss': {'error_score': {'weight': 0.0001, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0302_noTgtAlign', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': True, 'vggt_cam_loss': True, 'vggt_distil': False, 'context_view_train': False}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'save_gs': False, 'save_sample_wise_metrics': True, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3, 'target_align': True}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
|
| 9 |
+
2026-03-03 17:26:05,473 INFO MainThread:862779 [wandb_init.py:init():892] starting backend
|
| 10 |
+
2026-03-03 17:26:05,717 INFO MainThread:862779 [wandb_init.py:init():895] sending inform_init request
|
| 11 |
+
2026-03-03 17:26:05,723 INFO MainThread:862779 [wandb_init.py:init():903] backend started and connected
|
| 12 |
+
2026-03-03 17:26:05,730 INFO MainThread:862779 [wandb_init.py:init():973] updated telemetry
|
| 13 |
+
2026-03-03 17:26:05,737 INFO MainThread:862779 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
|
| 14 |
+
2026-03-03 17:26:07,330 INFO MainThread:862779 [wandb_init.py:init():1042] starting run threads in backend
|
| 15 |
+
2026-03-03 17:26:07,463 INFO MainThread:862779 [wandb_run.py:_console_start():2524] atexit reg
|
| 16 |
+
2026-03-03 17:26:07,463 INFO MainThread:862779 [wandb_run.py:_redirect():2373] redirect: wrap_raw
|
| 17 |
+
2026-03-03 17:26:07,464 INFO MainThread:862779 [wandb_run.py:_redirect():2442] Wrapping output streams.
|
| 18 |
+
2026-03-03 17:26:07,464 INFO MainThread:862779 [wandb_run.py:_redirect():2465] Redirects installed.
|
| 19 |
+
2026-03-03 17:26:07,468 INFO MainThread:862779 [wandb_init.py:init():1082] run started, returning control to user process
|
| 20 |
+
2026-03-03 17:29:17,912 INFO MainThread:862779 [wandb_run.py:_finish():2291] finishing run know/DCSplat/beokhgg2
|
| 21 |
+
2026-03-03 17:29:17,913 INFO MainThread:862779 [wandb_run.py:_atexit_cleanup():2490] got exitcode: 0
|
| 22 |
+
2026-03-03 17:29:17,913 INFO MainThread:862779 [wandb_run.py:_restore():2472] restore
|
| 23 |
+
2026-03-03 17:29:17,913 INFO MainThread:862779 [wandb_run.py:_restore():2478] restore done
|
| 24 |
+
2026-03-03 17:29:19,452 INFO MainThread:862779 [wandb_run.py:_footer_sync_info():3868] logging synced files
|
ABLATION_0302_noTgtAlign/wandb/run-20260303_172605-beokhgg2/run-beokhgg2.wandb
ADDED
|
Binary file (51.8 kB). View file
|
|
|
ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/config.yaml
ADDED
|
@@ -0,0 +1,311 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
_wandb:
|
| 2 |
+
value:
|
| 3 |
+
cli_version: 0.25.0
|
| 4 |
+
e:
|
| 5 |
+
mytbwtaba7k7jw5rce0hn1lj8k7t5dm8:
|
| 6 |
+
args:
|
| 7 |
+
- +experiment=re10k_ablation_24v
|
| 8 |
+
- wandb.mode=online
|
| 9 |
+
- wandb.name=ABLATION_0302_noTgtAlign
|
| 10 |
+
- model.density_control.score_mode=random
|
| 11 |
+
cpu_count: 128
|
| 12 |
+
cpu_count_logical: 256
|
| 13 |
+
cudaVersion: "13.0"
|
| 14 |
+
disk:
|
| 15 |
+
/:
|
| 16 |
+
total: "735513149440"
|
| 17 |
+
used: "700800118784"
|
| 18 |
+
email: dna9041@korea.ac.kr
|
| 19 |
+
executable: /venv/main/bin/python
|
| 20 |
+
git:
|
| 21 |
+
commit: 9dfce172a0f8c7ce85e763899f7ef741ecffc454
|
| 22 |
+
remote: git@github.com:K-nowing/CVPR2026.git
|
| 23 |
+
gpu: NVIDIA H200
|
| 24 |
+
gpu_count: 8
|
| 25 |
+
gpu_nvidia:
|
| 26 |
+
- architecture: Hopper
|
| 27 |
+
cudaCores: 16896
|
| 28 |
+
memoryTotal: "150754820096"
|
| 29 |
+
name: NVIDIA H200
|
| 30 |
+
uuid: GPU-9a20101e-d876-facd-5f05-805081aede41
|
| 31 |
+
- architecture: Hopper
|
| 32 |
+
cudaCores: 16896
|
| 33 |
+
memoryTotal: "150754820096"
|
| 34 |
+
name: NVIDIA H200
|
| 35 |
+
uuid: GPU-84736a77-ee75-3324-e4e1-99cc15bfb5e9
|
| 36 |
+
- architecture: Hopper
|
| 37 |
+
cudaCores: 16896
|
| 38 |
+
memoryTotal: "150754820096"
|
| 39 |
+
name: NVIDIA H200
|
| 40 |
+
uuid: GPU-423d3161-cdc4-3fc0-caee-d15cfaa83ca6
|
| 41 |
+
- architecture: Hopper
|
| 42 |
+
cudaCores: 16896
|
| 43 |
+
memoryTotal: "150754820096"
|
| 44 |
+
name: NVIDIA H200
|
| 45 |
+
uuid: GPU-5b0058b2-cdb9-c952-04f9-87dcaa7ea742
|
| 46 |
+
- architecture: Hopper
|
| 47 |
+
cudaCores: 16896
|
| 48 |
+
memoryTotal: "150754820096"
|
| 49 |
+
name: NVIDIA H200
|
| 50 |
+
uuid: GPU-08b37f98-4603-d483-2f2b-fe5311aa42f2
|
| 51 |
+
- architecture: Hopper
|
| 52 |
+
cudaCores: 16896
|
| 53 |
+
memoryTotal: "150754820096"
|
| 54 |
+
name: NVIDIA H200
|
| 55 |
+
uuid: GPU-03273b5b-2fdd-a5fe-4460-c897334ae464
|
| 56 |
+
- architecture: Hopper
|
| 57 |
+
cudaCores: 16896
|
| 58 |
+
memoryTotal: "150754820096"
|
| 59 |
+
name: NVIDIA H200
|
| 60 |
+
uuid: GPU-292d466c-d00d-25a4-28b6-e6c978d3e70c
|
| 61 |
+
- architecture: Hopper
|
| 62 |
+
cudaCores: 16896
|
| 63 |
+
memoryTotal: "150754820096"
|
| 64 |
+
name: NVIDIA H200
|
| 65 |
+
uuid: GPU-46f38561-3148-e442-7f7f-bfe447bab7fe
|
| 66 |
+
host: e9d3310a05da
|
| 67 |
+
memory:
|
| 68 |
+
total: "1622950240256"
|
| 69 |
+
os: Linux-6.8.0-94-generic-x86_64-with-glibc2.39
|
| 70 |
+
program: -m src.main
|
| 71 |
+
python: CPython 3.12.12
|
| 72 |
+
root: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign
|
| 73 |
+
startedAt: "2026-03-03T17:30:24.387912Z"
|
| 74 |
+
writerId: mytbwtaba7k7jw5rce0hn1lj8k7t5dm8
|
| 75 |
+
m:
|
| 76 |
+
- "1": trainer/global_step
|
| 77 |
+
"6":
|
| 78 |
+
- 3
|
| 79 |
+
"7": []
|
| 80 |
+
- "2": '*'
|
| 81 |
+
"5": 1
|
| 82 |
+
"6":
|
| 83 |
+
- 1
|
| 84 |
+
"7": []
|
| 85 |
+
python_version: 3.12.12
|
| 86 |
+
t:
|
| 87 |
+
"1":
|
| 88 |
+
- 1
|
| 89 |
+
- 41
|
| 90 |
+
- 49
|
| 91 |
+
- 50
|
| 92 |
+
- 106
|
| 93 |
+
"2":
|
| 94 |
+
- 1
|
| 95 |
+
- 41
|
| 96 |
+
- 49
|
| 97 |
+
- 50
|
| 98 |
+
- 106
|
| 99 |
+
"3":
|
| 100 |
+
- 7
|
| 101 |
+
- 13
|
| 102 |
+
- 15
|
| 103 |
+
- 16
|
| 104 |
+
- 41
|
| 105 |
+
- 66
|
| 106 |
+
"4": 3.12.12
|
| 107 |
+
"5": 0.25.0
|
| 108 |
+
"12": 0.25.0
|
| 109 |
+
"13": linux-x86_64
|
| 110 |
+
checkpointing:
|
| 111 |
+
value:
|
| 112 |
+
every_n_train_steps: 1500
|
| 113 |
+
load: null
|
| 114 |
+
save_top_k: 2
|
| 115 |
+
save_weights_only: false
|
| 116 |
+
data_loader:
|
| 117 |
+
value:
|
| 118 |
+
test:
|
| 119 |
+
batch_size: 1
|
| 120 |
+
num_workers: 4
|
| 121 |
+
persistent_workers: false
|
| 122 |
+
seed: 2345
|
| 123 |
+
train:
|
| 124 |
+
batch_size: 16
|
| 125 |
+
num_workers: 16
|
| 126 |
+
persistent_workers: true
|
| 127 |
+
seed: 1234
|
| 128 |
+
val:
|
| 129 |
+
batch_size: 1
|
| 130 |
+
num_workers: 1
|
| 131 |
+
persistent_workers: true
|
| 132 |
+
seed: 3456
|
| 133 |
+
dataset:
|
| 134 |
+
value:
|
| 135 |
+
re10k:
|
| 136 |
+
augment: true
|
| 137 |
+
background_color:
|
| 138 |
+
- 0
|
| 139 |
+
- 0
|
| 140 |
+
- 0
|
| 141 |
+
baseline_max: 1e+10
|
| 142 |
+
baseline_min: 0.001
|
| 143 |
+
cameras_are_circular: false
|
| 144 |
+
dynamic_context_views: true
|
| 145 |
+
input_image_shape:
|
| 146 |
+
- 256
|
| 147 |
+
- 256
|
| 148 |
+
make_baseline_1: true
|
| 149 |
+
max_context_views_per_gpu: 24
|
| 150 |
+
max_fov: 100
|
| 151 |
+
name: re10k
|
| 152 |
+
original_image_shape:
|
| 153 |
+
- 360
|
| 154 |
+
- 640
|
| 155 |
+
overfit_to_scene: null
|
| 156 |
+
relative_pose: true
|
| 157 |
+
roots:
|
| 158 |
+
- datasets/re10k
|
| 159 |
+
skip_bad_shape: true
|
| 160 |
+
view_sampler:
|
| 161 |
+
initial_max_distance_between_context_views: 25
|
| 162 |
+
initial_min_distance_between_context_views: 25
|
| 163 |
+
max_distance_between_context_views: 90
|
| 164 |
+
min_distance_between_context_views: 45
|
| 165 |
+
min_distance_to_context_views: 0
|
| 166 |
+
name: bounded
|
| 167 |
+
num_context_views: 2
|
| 168 |
+
num_target_set: 3
|
| 169 |
+
num_target_views: 4
|
| 170 |
+
same_target_gap: false
|
| 171 |
+
target_align: true
|
| 172 |
+
warm_up_steps: 1000
|
| 173 |
+
density_control_loss:
|
| 174 |
+
value:
|
| 175 |
+
error_score:
|
| 176 |
+
grad_scale: 10000
|
| 177 |
+
log_scale: false
|
| 178 |
+
mode: original
|
| 179 |
+
weight: 0.0001
|
| 180 |
+
direct_loss:
|
| 181 |
+
value:
|
| 182 |
+
l1:
|
| 183 |
+
weight: 0.8
|
| 184 |
+
ssim:
|
| 185 |
+
weight: 0.2
|
| 186 |
+
mode:
|
| 187 |
+
value: train
|
| 188 |
+
model:
|
| 189 |
+
value:
|
| 190 |
+
decoder:
|
| 191 |
+
background_color:
|
| 192 |
+
- 0
|
| 193 |
+
- 0
|
| 194 |
+
- 0
|
| 195 |
+
make_scale_invariant: false
|
| 196 |
+
name: splatting_cuda
|
| 197 |
+
density_control:
|
| 198 |
+
aggregation_mode: mean
|
| 199 |
+
aux_refine: false
|
| 200 |
+
grad_mode: absgrad
|
| 201 |
+
gs_param_dim: 256
|
| 202 |
+
latent_dim: 128
|
| 203 |
+
mean_dim: 32
|
| 204 |
+
name: density_control_module
|
| 205 |
+
num_heads: 1
|
| 206 |
+
num_latents: 64
|
| 207 |
+
num_level: 3
|
| 208 |
+
num_self_attn_per_block: 2
|
| 209 |
+
refine_error: false
|
| 210 |
+
refinement_hidden_dim: 32
|
| 211 |
+
refinement_layer_num: 1
|
| 212 |
+
refinement_type: voxelize
|
| 213 |
+
score_mode: random
|
| 214 |
+
use_depth: false
|
| 215 |
+
use_mean_features: true
|
| 216 |
+
use_refine_module: false
|
| 217 |
+
voxel_size: 0.001
|
| 218 |
+
voxelize_activate: false
|
| 219 |
+
encoder:
|
| 220 |
+
align_corners: false
|
| 221 |
+
gs_param_dim: 256
|
| 222 |
+
head_mode: pcd
|
| 223 |
+
input_image_shape:
|
| 224 |
+
- 518
|
| 225 |
+
- 518
|
| 226 |
+
name: dcsplat
|
| 227 |
+
num_level: 3
|
| 228 |
+
use_voxelize: true
|
| 229 |
+
optimizer:
|
| 230 |
+
value:
|
| 231 |
+
accumulate: 1
|
| 232 |
+
backbone_lr_multiplier: 0.1
|
| 233 |
+
backbone_trainable: T+H
|
| 234 |
+
lr: 0.0002
|
| 235 |
+
warm_up_steps: 25
|
| 236 |
+
render_loss:
|
| 237 |
+
value:
|
| 238 |
+
lpips:
|
| 239 |
+
apply_after_step: 0
|
| 240 |
+
weight: 0.05
|
| 241 |
+
mse:
|
| 242 |
+
weight: 1
|
| 243 |
+
seed:
|
| 244 |
+
value: 111123
|
| 245 |
+
test:
|
| 246 |
+
value:
|
| 247 |
+
align_pose: false
|
| 248 |
+
compute_scores: true
|
| 249 |
+
error_threshold: 0.4
|
| 250 |
+
error_threshold_list:
|
| 251 |
+
- 0.2
|
| 252 |
+
- 0.4
|
| 253 |
+
- 0.6
|
| 254 |
+
- 0.8
|
| 255 |
+
- 1
|
| 256 |
+
nvs_view_N_list:
|
| 257 |
+
- 3
|
| 258 |
+
- 6
|
| 259 |
+
- 16
|
| 260 |
+
- 32
|
| 261 |
+
- 64
|
| 262 |
+
output_path: test/ablation/re10k
|
| 263 |
+
pose_align_steps: 100
|
| 264 |
+
pred_intrinsic: false
|
| 265 |
+
rot_opt_lr: 0.005
|
| 266 |
+
save_active_mask_image: false
|
| 267 |
+
save_compare: false
|
| 268 |
+
save_error_score_image: false
|
| 269 |
+
save_gs: false
|
| 270 |
+
save_image: false
|
| 271 |
+
save_sample_wise_metrics: true
|
| 272 |
+
save_video: false
|
| 273 |
+
threshold_mode: ratio
|
| 274 |
+
trans_opt_lr: 0.005
|
| 275 |
+
train:
|
| 276 |
+
value:
|
| 277 |
+
align_corners: false
|
| 278 |
+
beta_dist_param:
|
| 279 |
+
- 0.5
|
| 280 |
+
- 4
|
| 281 |
+
cam_scale_mode: sum
|
| 282 |
+
camera_loss: 10
|
| 283 |
+
context_view_train: false
|
| 284 |
+
ext_scale_detach: false
|
| 285 |
+
extended_visualization: false
|
| 286 |
+
intrinsic_scaling: false
|
| 287 |
+
one_sample_validation: null
|
| 288 |
+
print_log_every_n_steps: 10
|
| 289 |
+
scene_scale_reg_loss: 0.01
|
| 290 |
+
train_aux: true
|
| 291 |
+
train_gs_num: 1
|
| 292 |
+
train_target_set: true
|
| 293 |
+
use_refine_aux: false
|
| 294 |
+
verbose: false
|
| 295 |
+
vggt_cam_loss: true
|
| 296 |
+
vggt_distil: false
|
| 297 |
+
trainer:
|
| 298 |
+
value:
|
| 299 |
+
gradient_clip_val: 0.5
|
| 300 |
+
max_steps: 3001
|
| 301 |
+
num_nodes: 1
|
| 302 |
+
val_check_interval: 250
|
| 303 |
+
wandb:
|
| 304 |
+
value:
|
| 305 |
+
entity: scene-representation-group
|
| 306 |
+
mode: online
|
| 307 |
+
name: ABLATION_0302_noTgtAlign
|
| 308 |
+
project: DCSplat
|
| 309 |
+
tags:
|
| 310 |
+
- re10k
|
| 311 |
+
- 256x256
|
ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/output.log
ADDED
|
@@ -0,0 +1,117 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7]
|
| 2 |
+
|
| 3 |
+
| Name | Type | Params | Mode
|
| 4 |
+
------------------------------------------------------------------------
|
| 5 |
+
0 | encoder | OurSplat | 888 M | train
|
| 6 |
+
1 | density_control_module | DensityControlModule | 0 | train
|
| 7 |
+
2 | decoder | DecoderSplattingCUDA | 0 | train
|
| 8 |
+
3 | render_losses | ModuleList | 0 | train
|
| 9 |
+
4 | density_control_losses | ModuleList | 0 | train
|
| 10 |
+
5 | direct_losses | ModuleList | 0 | train
|
| 11 |
+
------------------------------------------------------------------------
|
| 12 |
+
888 M Trainable params
|
| 13 |
+
0 Non-trainable params
|
| 14 |
+
888 M Total params
|
| 15 |
+
3,553.933 Total estimated model params size (MB)
|
| 16 |
+
1202 Modules in train mode
|
| 17 |
+
522 Modules in eval mode
|
| 18 |
+
Sanity Checking: | | 0/? [00:00<?, ?it/s][2026-03-03 17:30:28,360][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=31` in the `DataLoader` to improve performance.
|
| 19 |
+
|
| 20 |
+
[2026-03-03 17:30:28,362][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 21 |
+
warnings.warn( # warn only once
|
| 22 |
+
|
| 23 |
+
Validation epoch start on rank 0
|
| 24 |
+
Sanity Checking DataLoader 0: 0%| | 0/1 [00:00<?, ?it/s]validation step 0; scene = ['306e2b7785657539'];
|
| 25 |
+
Error executing job with overrides: ['+experiment=re10k_ablation_24v', 'wandb.mode=online', 'wandb.name=ABLATION_0302_noTgtAlign', 'model.density_control.score_mode=random']
|
| 26 |
+
Traceback (most recent call last):
|
| 27 |
+
File "/workspace/code/CVPR2026/src/main.py", line 226, in train
|
| 28 |
+
trainer.fit(model_wrapper, datamodule=data_module)#, ckpt_path=checkpoint_path)
|
| 29 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 30 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 561, in fit
|
| 31 |
+
call._call_and_handle_interrupt(
|
| 32 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/call.py", line 47, in _call_and_handle_interrupt
|
| 33 |
+
return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs)
|
| 34 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 35 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch
|
| 36 |
+
return function(*args, **kwargs)
|
| 37 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 38 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 599, in _fit_impl
|
| 39 |
+
self._run(model, ckpt_path=ckpt_path)
|
| 40 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 1012, in _run
|
| 41 |
+
results = self._run_stage()
|
| 42 |
+
^^^^^^^^^^^^^^^^^
|
| 43 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 1054, in _run_stage
|
| 44 |
+
self._run_sanity_check()
|
| 45 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 1083, in _run_sanity_check
|
| 46 |
+
val_loop.run()
|
| 47 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/utilities.py", line 179, in _decorator
|
| 48 |
+
return loop_run(self, *args, **kwargs)
|
| 49 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 50 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/evaluation_loop.py", line 145, in run
|
| 51 |
+
self._evaluation_step(batch, batch_idx, dataloader_idx, dataloader_iter)
|
| 52 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/evaluation_loop.py", line 437, in _evaluation_step
|
| 53 |
+
output = call._call_strategy_hook(trainer, hook_name, *step_args)
|
| 54 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 55 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/call.py", line 328, in _call_strategy_hook
|
| 56 |
+
output = fn(*args, **kwargs)
|
| 57 |
+
^^^^^^^^^^^^^^^^^^^
|
| 58 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/strategies/strategy.py", line 411, in validation_step
|
| 59 |
+
return self._forward_redirection(self.model, self.lightning_module, "validation_step", *args, **kwargs)
|
| 60 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 61 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__
|
| 62 |
+
wrapper_output = wrapper_module(*args, **kwargs)
|
| 63 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 64 |
+
File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl
|
| 65 |
+
return self._call_impl(*args, **kwargs)
|
| 66 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 67 |
+
File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl
|
| 68 |
+
return forward_call(*args, **kwargs)
|
| 69 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 70 |
+
File "/venv/main/lib/python3.12/site-packages/torch/nn/parallel/distributed.py", line 1648, in forward
|
| 71 |
+
else self._run_ddp_forward(*inputs, **kwargs)
|
| 72 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 73 |
+
File "/venv/main/lib/python3.12/site-packages/torch/nn/parallel/distributed.py", line 1474, in _run_ddp_forward
|
| 74 |
+
return self.module(*inputs, **kwargs) # type: ignore[index]
|
| 75 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 76 |
+
File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl
|
| 77 |
+
return self._call_impl(*args, **kwargs)
|
| 78 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 79 |
+
File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl
|
| 80 |
+
return forward_call(*args, **kwargs)
|
| 81 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 82 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward
|
| 83 |
+
out = method(*_args, **_kwargs)
|
| 84 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 85 |
+
File "/venv/main/lib/python3.12/site-packages/lightning_utilities/core/rank_zero.py", line 41, in wrapped_fn
|
| 86 |
+
return fn(*args, **kwargs)
|
| 87 |
+
^^^^^^^^^^^^^^^^^^^
|
| 88 |
+
File "/venv/main/lib/python3.12/site-packages/jaxtyping/_decorator.py", line 562, in wrapped_fn
|
| 89 |
+
return wrapped_fn_impl(args, kwargs, bound, memos)
|
| 90 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 91 |
+
File "/venv/main/lib/python3.12/site-packages/jaxtyping/_decorator.py", line 486, in wrapped_fn_impl
|
| 92 |
+
out = fn(*args, **kwargs)
|
| 93 |
+
^^^^^^^^^^^^^^^^^^^
|
| 94 |
+
File "/workspace/code/CVPR2026/src/model/model_wrapper.py", line 1093, in validation_step
|
| 95 |
+
target_extrinsic_c2w, target_intrinsic, near_plane, far_plane = self.get_target_projections(batch, backbone_outputs, target_key=target_key_list[target_key_idx])
|
| 96 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 97 |
+
File "/venv/main/lib/python3.12/site-packages/jaxtyping/_decorator.py", line 562, in wrapped_fn
|
| 98 |
+
return wrapped_fn_impl(args, kwargs, bound, memos)
|
| 99 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 100 |
+
File "/venv/main/lib/python3.12/site-packages/jaxtyping/_decorator.py", line 486, in wrapped_fn_impl
|
| 101 |
+
out = fn(*args, **kwargs)
|
| 102 |
+
^^^^^^^^^^^^^^^^^^^
|
| 103 |
+
File "/workspace/code/CVPR2026/src/model/model_wrapper.py", line 1718, in get_target_projections
|
| 104 |
+
target_extrinsic_c2w, target_intrinsic, target_nearest_context_idx, near_plane, far_plane = self.get_normalized_target_cam(batch, gs_outputs, target_key, intr_scale)
|
| 105 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 106 |
+
File "/venv/main/lib/python3.12/site-packages/jaxtyping/_decorator.py", line 562, in wrapped_fn
|
| 107 |
+
return wrapped_fn_impl(args, kwargs, bound, memos)
|
| 108 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 109 |
+
File "/venv/main/lib/python3.12/site-packages/jaxtyping/_decorator.py", line 486, in wrapped_fn_impl
|
| 110 |
+
out = fn(*args, **kwargs)
|
| 111 |
+
^^^^^^^^^^^^^^^^^^^
|
| 112 |
+
File "/workspace/code/CVPR2026/src/model/model_wrapper.py", line 1595, in get_normalized_target_cam
|
| 113 |
+
target_intrinsic = pred_intrinsics.mean(dim=1, keepdim=True).repeat(1, TVN, 1, 1)
|
| 114 |
+
^^^^^^^^^^^^^^^
|
| 115 |
+
NameError: name 'pred_intrinsics' is not defined
|
| 116 |
+
|
| 117 |
+
Set the environment variable HYDRA_FULL_ERROR=1 for a complete stack trace.
|
ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/requirements.txt
ADDED
|
@@ -0,0 +1,173 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
wheel==0.45.1
|
| 2 |
+
pytz==2025.2
|
| 3 |
+
easydict==1.13
|
| 4 |
+
antlr4-python3-runtime==4.9.3
|
| 5 |
+
wadler_lindig==0.1.7
|
| 6 |
+
networkx==3.4.2
|
| 7 |
+
urllib3==2.5.0
|
| 8 |
+
tzdata==2025.2
|
| 9 |
+
typing-inspection==0.4.1
|
| 10 |
+
tabulate==0.9.0
|
| 11 |
+
smmap==5.0.2
|
| 12 |
+
setuptools==78.1.1
|
| 13 |
+
safetensors==0.5.3
|
| 14 |
+
multidict==6.6.4
|
| 15 |
+
PyYAML==6.0.2
|
| 16 |
+
PySocks==1.7.1
|
| 17 |
+
pyparsing==3.2.5
|
| 18 |
+
pydantic_core==2.33.2
|
| 19 |
+
pycparser==2.23
|
| 20 |
+
protobuf==6.32.1
|
| 21 |
+
propcache==0.3.2
|
| 22 |
+
proglog==0.1.12
|
| 23 |
+
platformdirs==4.4.0
|
| 24 |
+
pip==25.2
|
| 25 |
+
mdurl==0.1.2
|
| 26 |
+
pillow==10.4.0
|
| 27 |
+
packaging==24.2
|
| 28 |
+
opt_einsum==3.4.0
|
| 29 |
+
frozenlist==1.7.0
|
| 30 |
+
numpy==1.26.4
|
| 31 |
+
ninja==1.13.0
|
| 32 |
+
MarkupSafe==3.0.2
|
| 33 |
+
kornia_rs==0.1.9
|
| 34 |
+
kiwisolver==1.4.9
|
| 35 |
+
imageio-ffmpeg==0.6.0
|
| 36 |
+
idna==3.7
|
| 37 |
+
fsspec==2024.6.1
|
| 38 |
+
hf-xet==1.1.10
|
| 39 |
+
gmpy2==2.2.1
|
| 40 |
+
fonttools==4.60.0
|
| 41 |
+
triton==3.4.0
|
| 42 |
+
filelock==3.17.0
|
| 43 |
+
einops==0.8.1
|
| 44 |
+
decorator==4.4.2
|
| 45 |
+
dacite==1.9.2
|
| 46 |
+
cycler==0.12.1
|
| 47 |
+
colorama==0.4.6
|
| 48 |
+
click==8.3.0
|
| 49 |
+
nvidia-nvtx-cu12==12.8.90
|
| 50 |
+
charset-normalizer==3.3.2
|
| 51 |
+
certifi==2025.8.3
|
| 52 |
+
beartype==0.19.0
|
| 53 |
+
attrs==25.3.0
|
| 54 |
+
async-timeout==5.0.1
|
| 55 |
+
annotated-types==0.7.0
|
| 56 |
+
aiohappyeyeballs==2.6.1
|
| 57 |
+
yarl==1.20.1
|
| 58 |
+
tifffile==2025.5.10
|
| 59 |
+
sentry-sdk==2.39.0
|
| 60 |
+
scipy==1.15.3
|
| 61 |
+
pydantic==2.11.9
|
| 62 |
+
pandas==2.3.2
|
| 63 |
+
opencv-python==4.11.0.86
|
| 64 |
+
omegaconf==2.3.0
|
| 65 |
+
markdown-it-py==4.0.0
|
| 66 |
+
lightning-utilities==0.14.3
|
| 67 |
+
lazy_loader==0.4
|
| 68 |
+
jaxtyping==0.2.37
|
| 69 |
+
imageio==2.37.0
|
| 70 |
+
gitdb==4.0.12
|
| 71 |
+
contourpy==1.3.2
|
| 72 |
+
colorspacious==1.1.2
|
| 73 |
+
cffi==1.17.1
|
| 74 |
+
aiosignal==1.4.0
|
| 75 |
+
scikit-video==1.1.11
|
| 76 |
+
scikit-image==0.25.2
|
| 77 |
+
rich==14.1.0
|
| 78 |
+
moviepy==1.0.3
|
| 79 |
+
matplotlib==3.10.6
|
| 80 |
+
hydra-core==1.3.2
|
| 81 |
+
huggingface-hub==0.35.1
|
| 82 |
+
GitPython==3.1.45
|
| 83 |
+
brotlicffi==1.0.9.2
|
| 84 |
+
aiohttp==3.12.15
|
| 85 |
+
torchmetrics==1.8.2
|
| 86 |
+
opt-einsum-fx==0.1.4
|
| 87 |
+
kornia==0.8.1
|
| 88 |
+
pytorch-lightning==2.5.1
|
| 89 |
+
lpips==0.1.4
|
| 90 |
+
e3nn==0.6.0
|
| 91 |
+
lightning==2.5.1
|
| 92 |
+
gsplat==1.5.3
|
| 93 |
+
nvidia-cusparselt-cu12==0.7.1
|
| 94 |
+
nvidia-nvjitlink-cu12==12.8.93
|
| 95 |
+
nvidia-nccl-cu12==2.27.3
|
| 96 |
+
nvidia-curand-cu12==10.3.9.90
|
| 97 |
+
nvidia-cufile-cu12==1.13.1.3
|
| 98 |
+
nvidia-cuda-runtime-cu12==12.8.90
|
| 99 |
+
nvidia-cuda-nvrtc-cu12==12.8.93
|
| 100 |
+
nvidia-cuda-cupti-cu12==12.8.90
|
| 101 |
+
nvidia-cublas-cu12==12.8.4.1
|
| 102 |
+
nvidia-cusparse-cu12==12.5.8.93
|
| 103 |
+
nvidia-cufft-cu12==11.3.3.83
|
| 104 |
+
nvidia-cudnn-cu12==9.10.2.21
|
| 105 |
+
nvidia-cusolver-cu12==11.7.3.90
|
| 106 |
+
torch==2.8.0+cu128
|
| 107 |
+
torchvision==0.23.0+cu128
|
| 108 |
+
torchaudio==2.8.0+cu128
|
| 109 |
+
torch_scatter==2.1.2+pt28cu128
|
| 110 |
+
plyfile==1.1.3
|
| 111 |
+
wandb==0.25.0
|
| 112 |
+
cuda-bindings==13.0.3
|
| 113 |
+
cuda-pathfinder==1.3.3
|
| 114 |
+
Jinja2==3.1.6
|
| 115 |
+
mpmath==1.3.0
|
| 116 |
+
nvidia-cublas==13.1.0.3
|
| 117 |
+
nvidia-cuda-cupti==13.0.85
|
| 118 |
+
nvidia-cuda-nvrtc==13.0.88
|
| 119 |
+
nvidia-cuda-runtime==13.0.96
|
| 120 |
+
nvidia-cudnn-cu13==9.15.1.9
|
| 121 |
+
nvidia-cufft==12.0.0.61
|
| 122 |
+
nvidia-cufile==1.15.1.6
|
| 123 |
+
nvidia-curand==10.4.0.35
|
| 124 |
+
nvidia-cusolver==12.0.4.66
|
| 125 |
+
nvidia-cusparse==12.6.3.3
|
| 126 |
+
nvidia-cusparselt-cu13==0.8.0
|
| 127 |
+
nvidia-nccl-cu13==2.28.9
|
| 128 |
+
nvidia-nvjitlink==13.0.88
|
| 129 |
+
nvidia-nvshmem-cu13==3.4.5
|
| 130 |
+
nvidia-nvtx==13.0.85
|
| 131 |
+
requests==2.32.5
|
| 132 |
+
sentencepiece==0.2.1
|
| 133 |
+
sympy==1.14.0
|
| 134 |
+
torchcodec==0.10.0
|
| 135 |
+
torchdata==0.10.0
|
| 136 |
+
torchtext==0.6.0
|
| 137 |
+
anyio==4.12.0
|
| 138 |
+
asttokens==3.0.1
|
| 139 |
+
comm==0.2.3
|
| 140 |
+
debugpy==1.8.19
|
| 141 |
+
executing==2.2.1
|
| 142 |
+
h11==0.16.0
|
| 143 |
+
httpcore==1.0.9
|
| 144 |
+
httpx==0.28.1
|
| 145 |
+
ipykernel==7.1.0
|
| 146 |
+
ipython==9.8.0
|
| 147 |
+
ipython_pygments_lexers==1.1.1
|
| 148 |
+
ipywidgets==8.1.8
|
| 149 |
+
jedi==0.19.2
|
| 150 |
+
jupyter_client==8.7.0
|
| 151 |
+
jupyter_core==5.9.1
|
| 152 |
+
jupyterlab_widgets==3.0.16
|
| 153 |
+
matplotlib-inline==0.2.1
|
| 154 |
+
nest-asyncio==1.6.0
|
| 155 |
+
parso==0.8.5
|
| 156 |
+
pexpect==4.9.0
|
| 157 |
+
prompt_toolkit==3.0.52
|
| 158 |
+
psutil==7.2.1
|
| 159 |
+
ptyprocess==0.7.0
|
| 160 |
+
pure_eval==0.2.3
|
| 161 |
+
Pygments==2.19.2
|
| 162 |
+
python-dateutil==2.9.0.post0
|
| 163 |
+
pyzmq==27.1.0
|
| 164 |
+
shellingham==1.5.4
|
| 165 |
+
six==1.17.0
|
| 166 |
+
stack-data==0.6.3
|
| 167 |
+
tornado==6.5.4
|
| 168 |
+
tqdm==4.67.1
|
| 169 |
+
traitlets==5.14.3
|
| 170 |
+
typer-slim==0.21.0
|
| 171 |
+
typing_extensions==4.15.0
|
| 172 |
+
wcwidth==0.2.14
|
| 173 |
+
widgetsnbextension==4.0.15
|
ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/wandb-metadata.json
ADDED
|
@@ -0,0 +1,93 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"os": "Linux-6.8.0-94-generic-x86_64-with-glibc2.39",
|
| 3 |
+
"python": "CPython 3.12.12",
|
| 4 |
+
"startedAt": "2026-03-03T17:30:24.387912Z",
|
| 5 |
+
"args": [
|
| 6 |
+
"+experiment=re10k_ablation_24v",
|
| 7 |
+
"wandb.mode=online",
|
| 8 |
+
"wandb.name=ABLATION_0302_noTgtAlign",
|
| 9 |
+
"model.density_control.score_mode=random"
|
| 10 |
+
],
|
| 11 |
+
"program": "-m src.main",
|
| 12 |
+
"git": {
|
| 13 |
+
"remote": "git@github.com:K-nowing/CVPR2026.git",
|
| 14 |
+
"commit": "9dfce172a0f8c7ce85e763899f7ef741ecffc454"
|
| 15 |
+
},
|
| 16 |
+
"email": "dna9041@korea.ac.kr",
|
| 17 |
+
"root": "/workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0302_noTgtAlign",
|
| 18 |
+
"host": "e9d3310a05da",
|
| 19 |
+
"executable": "/venv/main/bin/python",
|
| 20 |
+
"cpu_count": 128,
|
| 21 |
+
"cpu_count_logical": 256,
|
| 22 |
+
"gpu": "NVIDIA H200",
|
| 23 |
+
"gpu_count": 8,
|
| 24 |
+
"disk": {
|
| 25 |
+
"/": {
|
| 26 |
+
"total": "735513149440",
|
| 27 |
+
"used": "700800118784"
|
| 28 |
+
}
|
| 29 |
+
},
|
| 30 |
+
"memory": {
|
| 31 |
+
"total": "1622950240256"
|
| 32 |
+
},
|
| 33 |
+
"gpu_nvidia": [
|
| 34 |
+
{
|
| 35 |
+
"name": "NVIDIA H200",
|
| 36 |
+
"memoryTotal": "150754820096",
|
| 37 |
+
"cudaCores": 16896,
|
| 38 |
+
"architecture": "Hopper",
|
| 39 |
+
"uuid": "GPU-9a20101e-d876-facd-5f05-805081aede41"
|
| 40 |
+
},
|
| 41 |
+
{
|
| 42 |
+
"name": "NVIDIA H200",
|
| 43 |
+
"memoryTotal": "150754820096",
|
| 44 |
+
"cudaCores": 16896,
|
| 45 |
+
"architecture": "Hopper",
|
| 46 |
+
"uuid": "GPU-84736a77-ee75-3324-e4e1-99cc15bfb5e9"
|
| 47 |
+
},
|
| 48 |
+
{
|
| 49 |
+
"name": "NVIDIA H200",
|
| 50 |
+
"memoryTotal": "150754820096",
|
| 51 |
+
"cudaCores": 16896,
|
| 52 |
+
"architecture": "Hopper",
|
| 53 |
+
"uuid": "GPU-423d3161-cdc4-3fc0-caee-d15cfaa83ca6"
|
| 54 |
+
},
|
| 55 |
+
{
|
| 56 |
+
"name": "NVIDIA H200",
|
| 57 |
+
"memoryTotal": "150754820096",
|
| 58 |
+
"cudaCores": 16896,
|
| 59 |
+
"architecture": "Hopper",
|
| 60 |
+
"uuid": "GPU-5b0058b2-cdb9-c952-04f9-87dcaa7ea742"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"name": "NVIDIA H200",
|
| 64 |
+
"memoryTotal": "150754820096",
|
| 65 |
+
"cudaCores": 16896,
|
| 66 |
+
"architecture": "Hopper",
|
| 67 |
+
"uuid": "GPU-08b37f98-4603-d483-2f2b-fe5311aa42f2"
|
| 68 |
+
},
|
| 69 |
+
{
|
| 70 |
+
"name": "NVIDIA H200",
|
| 71 |
+
"memoryTotal": "150754820096",
|
| 72 |
+
"cudaCores": 16896,
|
| 73 |
+
"architecture": "Hopper",
|
| 74 |
+
"uuid": "GPU-03273b5b-2fdd-a5fe-4460-c897334ae464"
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"name": "NVIDIA H200",
|
| 78 |
+
"memoryTotal": "150754820096",
|
| 79 |
+
"cudaCores": 16896,
|
| 80 |
+
"architecture": "Hopper",
|
| 81 |
+
"uuid": "GPU-292d466c-d00d-25a4-28b6-e6c978d3e70c"
|
| 82 |
+
},
|
| 83 |
+
{
|
| 84 |
+
"name": "NVIDIA H200",
|
| 85 |
+
"memoryTotal": "150754820096",
|
| 86 |
+
"cudaCores": 16896,
|
| 87 |
+
"architecture": "Hopper",
|
| 88 |
+
"uuid": "GPU-46f38561-3148-e442-7f7f-bfe447bab7fe"
|
| 89 |
+
}
|
| 90 |
+
],
|
| 91 |
+
"cudaVersion": "13.0",
|
| 92 |
+
"writerId": "mytbwtaba7k7jw5rce0hn1lj8k7t5dm8"
|
| 93 |
+
}
|
ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/files/wandb-summary.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"_runtime":7,"_wandb":{"runtime":7}}
|
ABLATION_0302_noTgtAlign/wandb/run-20260303_173024-phck7alc/logs/debug-core.log
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2026-03-03T17:30:24.492490052Z","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpztx1j514/port-863948.txt","pid":863948,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
|
| 2 |
+
{"time":"2026-03-03T17:30:24.493158838Z","level":"INFO","msg":"server: will exit if parent process dies","ppid":863948}
|
| 3 |
+
{"time":"2026-03-03T17:30:24.493143718Z","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-863948-866375-1399352181/socket","Net":"unix"}}
|
| 4 |
+
{"time":"2026-03-03T17:30:24.673880578Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
|
| 5 |
+
{"time":"2026-03-03T17:30:24.684765982Z","level":"INFO","msg":"handleInformInit: received","streamId":"phck7alc","id":"1(@)"}
|
| 6 |
+
{"time":"2026-03-03T17:30:25.148320888Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"phck7alc","id":"1(@)"}
|
| 7 |
+
{"time":"2026-03-03T17:30:31.082435508Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"i6tdoiz1oxwi"}
|
| 8 |
+
{"time":"2026-03-03T17:30:33.040285798Z","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
|
| 9 |
+
{"time":"2026-03-03T17:30:33.040358339Z","level":"INFO","msg":"connection: closing","id":"1(@)"}
|
| 10 |
+
{"time":"2026-03-03T17:30:33.040416149Z","level":"INFO","msg":"server is shutting down"}
|
| 11 |
+
{"time":"2026-03-03T17:30:33.040443929Z","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
|
| 12 |
+
{"time":"2026-03-03T17:30:33.040579961Z","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-863948-866375-1399352181/socket","Net":"unix"}}
|
| 13 |
+
{"time":"2026-03-03T17:30:34.141387811Z","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
|
| 14 |
+
{"time":"2026-03-03T17:30:34.141419972Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
|
| 15 |
+
{"time":"2026-03-03T17:30:34.141433602Z","level":"INFO","msg":"server is closed"}
|