Add files using upload-large-folder tool
Browse files- ABLATION_0225_ctxTrain_depth_vggtDistl/.hydra/config.yaml +189 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/.hydra/hydra.yaml +168 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/.hydra/overrides.yaml +7 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/main.log +45 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_1.log +19 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_2.log +19 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_3.log +19 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_4.log +19 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_5.log +19 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_6.log +19 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_7.log +19 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/debug-internal.log +11 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/debug.log +21 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/config.yaml +310 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/media/images/active_mask_imgs_1_690a5f29fe9230f3ead3.png +3 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/media/images/comparison_0_e330894a93360b5335a1.png +3 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/media/images/error_scores_2_bf819ed247f7546e0637.png +3 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/output.log +140 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/requirements.txt +172 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/wandb-metadata.json +96 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/wandb-summary.json +1 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug-core.log +15 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug-internal.log +11 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug.log +21 -0
- ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/run-g2qj8ry7.wandb +0 -0
- ABLATION_0225_randomSelect/main.log +12 -0
- ABLATION_0225_randomSelect/peak_vram_memory.json +6 -0
- ABLATION_0225_randomSelect/train_ddp_process_1.log +6 -0
- ABLATION_0225_randomSelect/train_ddp_process_2.log +6 -0
- ABLATION_0225_randomSelect/train_ddp_process_3.log +6 -0
- ABLATION_0225_randomSelect/train_ddp_process_4.log +6 -0
- ABLATION_0225_randomSelect/train_ddp_process_5.log +6 -0
- ABLATION_0225_randomSelect/train_ddp_process_6.log +6 -0
- ABLATION_0225_randomSelect/train_ddp_process_7.log +6 -0
- ABLATION_0225_randomSelect/wandb/debug-internal.log +5 -0
- ABLATION_0225_randomSelect/wandb/debug.log +2 -0
- ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/config.yaml +307 -0
- ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/active_mask_imgs_182_2edf4f17c23388f66672.png +3 -0
- ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/active_mask_imgs_198_399ec9b2f678832e0977.png +3 -0
- ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/comparison_197_ddc312f116aae122416e.png +3 -0
- ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/error_scores_183_e47e70d75361cb369a35.png +3 -0
- ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/error_scores_199_b1c8e80e7cf22aab1c67.png +3 -0
- ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/train/comparison_188_a456ce8a38a4d0dc65b1.png +3 -0
- ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/train/comparison_202_b207767914aae1b684b5.png +3 -0
- ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/train/error_scores_187_347cd0a01cf553ca4f4d.png +3 -0
- ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/train/error_scores_201_05456dc706694ddf5d92.png +3 -0
- ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/wandb-summary.json +1 -0
- ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/logs/debug-core.log +8 -0
- ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/logs/debug-internal.log +5 -0
- ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/logs/debug.log +2 -0
ABLATION_0225_ctxTrain_depth_vggtDistl/.hydra/config.yaml
ADDED
|
@@ -0,0 +1,189 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
encoder:
|
| 3 |
+
name: dcsplat
|
| 4 |
+
input_image_shape:
|
| 5 |
+
- 518
|
| 6 |
+
- 518
|
| 7 |
+
head_mode: depth
|
| 8 |
+
num_level: 3
|
| 9 |
+
gs_param_dim: 256
|
| 10 |
+
align_corners: false
|
| 11 |
+
use_voxelize: true
|
| 12 |
+
decoder:
|
| 13 |
+
name: splatting_cuda
|
| 14 |
+
background_color:
|
| 15 |
+
- 0.0
|
| 16 |
+
- 0.0
|
| 17 |
+
- 0.0
|
| 18 |
+
make_scale_invariant: false
|
| 19 |
+
density_control:
|
| 20 |
+
name: density_control_module
|
| 21 |
+
mean_dim: 32
|
| 22 |
+
gs_param_dim: 256
|
| 23 |
+
refinement_layer_num: 1
|
| 24 |
+
num_level: 3
|
| 25 |
+
grad_mode: absgrad
|
| 26 |
+
use_mean_features: true
|
| 27 |
+
refinement_type: voxelize
|
| 28 |
+
refinement_hidden_dim: 32
|
| 29 |
+
aggregation_mode: mean
|
| 30 |
+
num_heads: 1
|
| 31 |
+
score_mode: absgrad
|
| 32 |
+
latent_dim: 128
|
| 33 |
+
num_latents: 64
|
| 34 |
+
num_self_attn_per_block: 2
|
| 35 |
+
voxel_size: 0.001
|
| 36 |
+
aux_refine: false
|
| 37 |
+
refine_error: false
|
| 38 |
+
use_refine_module: true
|
| 39 |
+
voxelize_activate: true
|
| 40 |
+
use_depth: true
|
| 41 |
+
render_loss:
|
| 42 |
+
mse:
|
| 43 |
+
weight: 1.0
|
| 44 |
+
lpips:
|
| 45 |
+
weight: 0.05
|
| 46 |
+
apply_after_step: 0
|
| 47 |
+
depth_consis:
|
| 48 |
+
weight: 1.0
|
| 49 |
+
sigma_image: null
|
| 50 |
+
use_second_derivative: false
|
| 51 |
+
density_control_loss:
|
| 52 |
+
error_score:
|
| 53 |
+
weight: 0.01
|
| 54 |
+
log_scale: false
|
| 55 |
+
grad_scale: 10000.0
|
| 56 |
+
mode: original
|
| 57 |
+
direct_loss:
|
| 58 |
+
l1:
|
| 59 |
+
weight: 0.8
|
| 60 |
+
ssim:
|
| 61 |
+
weight: 0.2
|
| 62 |
+
wandb:
|
| 63 |
+
project: DCSplat
|
| 64 |
+
entity: scene-representation-group
|
| 65 |
+
name: ABLATION_0225_ctxTrain_depth_vggtDistl
|
| 66 |
+
mode: online
|
| 67 |
+
tags:
|
| 68 |
+
- re10k
|
| 69 |
+
- 256x256
|
| 70 |
+
mode: train
|
| 71 |
+
data_loader:
|
| 72 |
+
train:
|
| 73 |
+
num_workers: 16
|
| 74 |
+
persistent_workers: true
|
| 75 |
+
batch_size: 16
|
| 76 |
+
seed: 1234
|
| 77 |
+
test:
|
| 78 |
+
num_workers: 4
|
| 79 |
+
persistent_workers: false
|
| 80 |
+
batch_size: 1
|
| 81 |
+
seed: 2345
|
| 82 |
+
val:
|
| 83 |
+
num_workers: 1
|
| 84 |
+
persistent_workers: true
|
| 85 |
+
batch_size: 1
|
| 86 |
+
seed: 3456
|
| 87 |
+
optimizer:
|
| 88 |
+
lr: 0.0002
|
| 89 |
+
warm_up_steps: 25
|
| 90 |
+
backbone_lr_multiplier: 0.1
|
| 91 |
+
backbone_trainable: T+H
|
| 92 |
+
accumulate: 1
|
| 93 |
+
checkpointing:
|
| 94 |
+
load: null
|
| 95 |
+
every_n_train_steps: 1500
|
| 96 |
+
save_top_k: 2
|
| 97 |
+
save_weights_only: false
|
| 98 |
+
train:
|
| 99 |
+
extended_visualization: false
|
| 100 |
+
print_log_every_n_steps: 10
|
| 101 |
+
camera_loss: 10.0
|
| 102 |
+
one_sample_validation: null
|
| 103 |
+
align_corners: false
|
| 104 |
+
intrinsic_scaling: false
|
| 105 |
+
verbose: false
|
| 106 |
+
beta_dist_param:
|
| 107 |
+
- 0.5
|
| 108 |
+
- 4.0
|
| 109 |
+
use_refine_aux: false
|
| 110 |
+
train_target_set: true
|
| 111 |
+
train_gs_num: 1
|
| 112 |
+
ext_scale_detach: false
|
| 113 |
+
cam_scale_mode: sum
|
| 114 |
+
scene_scale_reg_loss: 0.01
|
| 115 |
+
train_aux: true
|
| 116 |
+
vggt_cam_loss: true
|
| 117 |
+
vggt_distil: true
|
| 118 |
+
context_view_train: true
|
| 119 |
+
test:
|
| 120 |
+
output_path: test/ablation/re10k
|
| 121 |
+
align_pose: false
|
| 122 |
+
pose_align_steps: 100
|
| 123 |
+
rot_opt_lr: 0.005
|
| 124 |
+
trans_opt_lr: 0.005
|
| 125 |
+
compute_scores: true
|
| 126 |
+
save_image: false
|
| 127 |
+
save_video: false
|
| 128 |
+
save_active_mask_image: false
|
| 129 |
+
save_error_score_image: false
|
| 130 |
+
save_compare: false
|
| 131 |
+
pred_intrinsic: false
|
| 132 |
+
error_threshold: 0.4
|
| 133 |
+
error_threshold_list:
|
| 134 |
+
- 0.2
|
| 135 |
+
- 0.4
|
| 136 |
+
- 0.6
|
| 137 |
+
- 0.8
|
| 138 |
+
- 1.0
|
| 139 |
+
threshold_mode: ratio
|
| 140 |
+
nvs_view_N_list:
|
| 141 |
+
- 3
|
| 142 |
+
- 6
|
| 143 |
+
- 16
|
| 144 |
+
- 32
|
| 145 |
+
- 64
|
| 146 |
+
seed: 111123
|
| 147 |
+
trainer:
|
| 148 |
+
max_steps: 3001
|
| 149 |
+
val_check_interval: 250
|
| 150 |
+
gradient_clip_val: 0.5
|
| 151 |
+
num_nodes: 1
|
| 152 |
+
dataset:
|
| 153 |
+
re10k:
|
| 154 |
+
make_baseline_1: true
|
| 155 |
+
relative_pose: true
|
| 156 |
+
augment: true
|
| 157 |
+
background_color:
|
| 158 |
+
- 0.0
|
| 159 |
+
- 0.0
|
| 160 |
+
- 0.0
|
| 161 |
+
overfit_to_scene: null
|
| 162 |
+
skip_bad_shape: true
|
| 163 |
+
view_sampler:
|
| 164 |
+
name: bounded
|
| 165 |
+
num_target_views: 4
|
| 166 |
+
num_context_views: 2
|
| 167 |
+
min_distance_between_context_views: 45
|
| 168 |
+
max_distance_between_context_views: 90
|
| 169 |
+
min_distance_to_context_views: 0
|
| 170 |
+
warm_up_steps: 1000
|
| 171 |
+
initial_min_distance_between_context_views: 25
|
| 172 |
+
initial_max_distance_between_context_views: 25
|
| 173 |
+
same_target_gap: false
|
| 174 |
+
num_target_set: 3
|
| 175 |
+
name: re10k
|
| 176 |
+
roots:
|
| 177 |
+
- datasets/re10k
|
| 178 |
+
input_image_shape:
|
| 179 |
+
- 256
|
| 180 |
+
- 256
|
| 181 |
+
original_image_shape:
|
| 182 |
+
- 360
|
| 183 |
+
- 640
|
| 184 |
+
cameras_are_circular: false
|
| 185 |
+
baseline_min: 0.001
|
| 186 |
+
baseline_max: 10000000000.0
|
| 187 |
+
max_fov: 100.0
|
| 188 |
+
dynamic_context_views: true
|
| 189 |
+
max_context_views_per_gpu: 24
|
ABLATION_0225_ctxTrain_depth_vggtDistl/.hydra/hydra.yaml
ADDED
|
@@ -0,0 +1,168 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
hydra:
|
| 2 |
+
run:
|
| 3 |
+
dir: outputs/ablation/re10k/${wandb.name}
|
| 4 |
+
sweep:
|
| 5 |
+
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
|
| 6 |
+
subdir: ${hydra.job.num}
|
| 7 |
+
launcher:
|
| 8 |
+
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
|
| 9 |
+
sweeper:
|
| 10 |
+
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
|
| 11 |
+
max_batch_size: null
|
| 12 |
+
params: null
|
| 13 |
+
help:
|
| 14 |
+
app_name: ${hydra.job.name}
|
| 15 |
+
header: '${hydra.help.app_name} is powered by Hydra.
|
| 16 |
+
|
| 17 |
+
'
|
| 18 |
+
footer: 'Powered by Hydra (https://hydra.cc)
|
| 19 |
+
|
| 20 |
+
Use --hydra-help to view Hydra specific help
|
| 21 |
+
|
| 22 |
+
'
|
| 23 |
+
template: '${hydra.help.header}
|
| 24 |
+
|
| 25 |
+
== Configuration groups ==
|
| 26 |
+
|
| 27 |
+
Compose your configuration from those groups (group=option)
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
$APP_CONFIG_GROUPS
|
| 31 |
+
|
| 32 |
+
|
| 33 |
+
== Config ==
|
| 34 |
+
|
| 35 |
+
Override anything in the config (foo.bar=value)
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
$CONFIG
|
| 39 |
+
|
| 40 |
+
|
| 41 |
+
${hydra.help.footer}
|
| 42 |
+
|
| 43 |
+
'
|
| 44 |
+
hydra_help:
|
| 45 |
+
template: 'Hydra (${hydra.runtime.version})
|
| 46 |
+
|
| 47 |
+
See https://hydra.cc for more info.
|
| 48 |
+
|
| 49 |
+
|
| 50 |
+
== Flags ==
|
| 51 |
+
|
| 52 |
+
$FLAGS_HELP
|
| 53 |
+
|
| 54 |
+
|
| 55 |
+
== Configuration groups ==
|
| 56 |
+
|
| 57 |
+
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
|
| 58 |
+
to command line)
|
| 59 |
+
|
| 60 |
+
|
| 61 |
+
$HYDRA_CONFIG_GROUPS
|
| 62 |
+
|
| 63 |
+
|
| 64 |
+
Use ''--cfg hydra'' to Show the Hydra config.
|
| 65 |
+
|
| 66 |
+
'
|
| 67 |
+
hydra_help: ???
|
| 68 |
+
hydra_logging:
|
| 69 |
+
version: 1
|
| 70 |
+
formatters:
|
| 71 |
+
simple:
|
| 72 |
+
format: '[%(asctime)s][HYDRA] %(message)s'
|
| 73 |
+
handlers:
|
| 74 |
+
console:
|
| 75 |
+
class: logging.StreamHandler
|
| 76 |
+
formatter: simple
|
| 77 |
+
stream: ext://sys.stdout
|
| 78 |
+
root:
|
| 79 |
+
level: INFO
|
| 80 |
+
handlers:
|
| 81 |
+
- console
|
| 82 |
+
loggers:
|
| 83 |
+
logging_example:
|
| 84 |
+
level: DEBUG
|
| 85 |
+
disable_existing_loggers: false
|
| 86 |
+
job_logging:
|
| 87 |
+
version: 1
|
| 88 |
+
formatters:
|
| 89 |
+
simple:
|
| 90 |
+
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
|
| 91 |
+
handlers:
|
| 92 |
+
console:
|
| 93 |
+
class: logging.StreamHandler
|
| 94 |
+
formatter: simple
|
| 95 |
+
stream: ext://sys.stdout
|
| 96 |
+
file:
|
| 97 |
+
class: logging.FileHandler
|
| 98 |
+
formatter: simple
|
| 99 |
+
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
|
| 100 |
+
root:
|
| 101 |
+
level: INFO
|
| 102 |
+
handlers:
|
| 103 |
+
- console
|
| 104 |
+
- file
|
| 105 |
+
disable_existing_loggers: false
|
| 106 |
+
env: {}
|
| 107 |
+
mode: RUN
|
| 108 |
+
searchpath: []
|
| 109 |
+
callbacks: {}
|
| 110 |
+
output_subdir: .hydra
|
| 111 |
+
overrides:
|
| 112 |
+
hydra:
|
| 113 |
+
- hydra.mode=RUN
|
| 114 |
+
task:
|
| 115 |
+
- +experiment=re10k_ablation_24v_depthloss
|
| 116 |
+
- wandb.mode=online
|
| 117 |
+
- wandb.name=ABLATION_0225_ctxTrain_depth_vggtDistl
|
| 118 |
+
- model.encoder.head_mode=depth
|
| 119 |
+
- model.density_control.use_depth=true
|
| 120 |
+
- train.context_view_train=true
|
| 121 |
+
- train.vggt_distil=true
|
| 122 |
+
job:
|
| 123 |
+
name: main
|
| 124 |
+
chdir: null
|
| 125 |
+
override_dirname: +experiment=re10k_ablation_24v_depthloss,model.density_control.use_depth=true,model.encoder.head_mode=depth,train.context_view_train=true,train.vggt_distil=true,wandb.mode=online,wandb.name=ABLATION_0225_ctxTrain_depth_vggtDistl
|
| 126 |
+
id: ???
|
| 127 |
+
num: ???
|
| 128 |
+
config_name: main
|
| 129 |
+
env_set: {}
|
| 130 |
+
env_copy: []
|
| 131 |
+
config:
|
| 132 |
+
override_dirname:
|
| 133 |
+
kv_sep: '='
|
| 134 |
+
item_sep: ','
|
| 135 |
+
exclude_keys: []
|
| 136 |
+
runtime:
|
| 137 |
+
version: 1.3.2
|
| 138 |
+
version_base: '1.3'
|
| 139 |
+
cwd: /workspace/code/CVPR2026
|
| 140 |
+
config_sources:
|
| 141 |
+
- path: hydra.conf
|
| 142 |
+
schema: pkg
|
| 143 |
+
provider: hydra
|
| 144 |
+
- path: /workspace/code/CVPR2026/config
|
| 145 |
+
schema: file
|
| 146 |
+
provider: main
|
| 147 |
+
- path: ''
|
| 148 |
+
schema: structured
|
| 149 |
+
provider: schema
|
| 150 |
+
output_dir: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl
|
| 151 |
+
choices:
|
| 152 |
+
experiment: re10k_ablation_24v_depthloss
|
| 153 |
+
dataset@dataset.re10k: re10k
|
| 154 |
+
dataset/view_sampler_dataset_specific_config@dataset.re10k.view_sampler: bounded_re10k
|
| 155 |
+
dataset/view_sampler@dataset.re10k.view_sampler: bounded
|
| 156 |
+
model/density_control: density_control_module
|
| 157 |
+
model/decoder: splatting_cuda
|
| 158 |
+
model/encoder: dcsplat
|
| 159 |
+
hydra/env: default
|
| 160 |
+
hydra/callbacks: null
|
| 161 |
+
hydra/job_logging: default
|
| 162 |
+
hydra/hydra_logging: default
|
| 163 |
+
hydra/hydra_help: default
|
| 164 |
+
hydra/help: default
|
| 165 |
+
hydra/sweeper: basic
|
| 166 |
+
hydra/launcher: basic
|
| 167 |
+
hydra/output: default
|
| 168 |
+
verbose: false
|
ABLATION_0225_ctxTrain_depth_vggtDistl/.hydra/overrides.yaml
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
- +experiment=re10k_ablation_24v_depthloss
|
| 2 |
+
- wandb.mode=online
|
| 3 |
+
- wandb.name=ABLATION_0225_ctxTrain_depth_vggtDistl
|
| 4 |
+
- model.encoder.head_mode=depth
|
| 5 |
+
- model.density_control.use_depth=true
|
| 6 |
+
- train.context_view_train=true
|
| 7 |
+
- train.vggt_distil=true
|
ABLATION_0225_ctxTrain_depth_vggtDistl/main.log
ADDED
|
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2026-02-25 13:52:21,786][dinov2][INFO] - using MLP layer as FFN
|
| 2 |
+
[2026-02-25 13:52:28,031][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 3 |
+
warnings.warn(
|
| 4 |
+
|
| 5 |
+
[2026-02-25 13:52:28,032][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 6 |
+
warnings.warn(msg)
|
| 7 |
+
|
| 8 |
+
[2026-02-25 13:52:29,480][dinov2][INFO] - using MLP layer as FFN
|
| 9 |
+
[2026-02-25 13:54:48,808][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=31` in the `DataLoader` to improve performance.
|
| 10 |
+
|
| 11 |
+
[2026-02-25 13:54:48,809][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 12 |
+
warnings.warn( # warn only once
|
| 13 |
+
|
| 14 |
+
[2026-02-25 13:54:51,686][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 15 |
+
result[selector] = overlay
|
| 16 |
+
|
| 17 |
+
[2026-02-25 13:54:51,695][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
|
| 18 |
+
|
| 19 |
+
[2026-02-25 13:54:51,696][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 20 |
+
warnings.warn(
|
| 21 |
+
|
| 22 |
+
[2026-02-25 13:54:51,696][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 23 |
+
warnings.warn(msg)
|
| 24 |
+
|
| 25 |
+
[2026-02-25 13:54:53,564][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4322.)
|
| 26 |
+
return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
|
| 27 |
+
|
| 28 |
+
[2026-02-25 13:54:53,845][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
|
| 29 |
+
|
| 30 |
+
[2026-02-25 13:54:53,847][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
|
| 31 |
+
|
| 32 |
+
[2026-02-25 13:54:53,847][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
|
| 33 |
+
|
| 34 |
+
[2026-02-25 13:54:53,848][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
|
| 35 |
+
|
| 36 |
+
[2026-02-25 13:54:53,848][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
|
| 37 |
+
|
| 38 |
+
[2026-02-25 14:29:16,843][dinov2][INFO] - using MLP layer as FFN
|
| 39 |
+
[2026-02-25 14:29:23,026][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 40 |
+
warnings.warn(
|
| 41 |
+
|
| 42 |
+
[2026-02-25 14:29:23,026][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 43 |
+
warnings.warn(msg)
|
| 44 |
+
|
| 45 |
+
[2026-02-25 14:29:24,307][dinov2][INFO] - using MLP layer as FFN
|
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_1.log
ADDED
|
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2026-02-25 13:52:46,873][dinov2][INFO] - using MLP layer as FFN
|
| 2 |
+
[2026-02-25 13:53:20,123][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 3 |
+
warnings.warn(
|
| 4 |
+
|
| 5 |
+
[2026-02-25 13:53:20,123][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 6 |
+
warnings.warn(msg)
|
| 7 |
+
|
| 8 |
+
[2026-02-25 13:53:24,223][dinov2][INFO] - using MLP layer as FFN
|
| 9 |
+
[2026-02-25 13:54:48,809][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 10 |
+
warnings.warn( # warn only once
|
| 11 |
+
|
| 12 |
+
[2026-02-25 14:29:41,340][dinov2][INFO] - using MLP layer as FFN
|
| 13 |
+
[2026-02-25 14:29:59,908][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 14 |
+
warnings.warn(
|
| 15 |
+
|
| 16 |
+
[2026-02-25 14:29:59,909][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 17 |
+
warnings.warn(msg)
|
| 18 |
+
|
| 19 |
+
[2026-02-25 14:30:04,698][dinov2][INFO] - using MLP layer as FFN
|
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_2.log
ADDED
|
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2026-02-25 13:52:46,752][dinov2][INFO] - using MLP layer as FFN
|
| 2 |
+
[2026-02-25 13:53:19,432][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 3 |
+
warnings.warn(
|
| 4 |
+
|
| 5 |
+
[2026-02-25 13:53:19,433][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 6 |
+
warnings.warn(msg)
|
| 7 |
+
|
| 8 |
+
[2026-02-25 13:53:22,571][dinov2][INFO] - using MLP layer as FFN
|
| 9 |
+
[2026-02-25 13:54:48,809][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 10 |
+
warnings.warn( # warn only once
|
| 11 |
+
|
| 12 |
+
[2026-02-25 14:29:41,333][dinov2][INFO] - using MLP layer as FFN
|
| 13 |
+
[2026-02-25 14:30:12,121][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 14 |
+
warnings.warn(
|
| 15 |
+
|
| 16 |
+
[2026-02-25 14:30:12,123][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 17 |
+
warnings.warn(msg)
|
| 18 |
+
|
| 19 |
+
[2026-02-25 14:30:17,034][dinov2][INFO] - using MLP layer as FFN
|
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_3.log
ADDED
|
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2026-02-25 13:52:46,815][dinov2][INFO] - using MLP layer as FFN
|
| 2 |
+
[2026-02-25 13:53:19,826][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 3 |
+
warnings.warn(
|
| 4 |
+
|
| 5 |
+
[2026-02-25 13:53:19,827][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 6 |
+
warnings.warn(msg)
|
| 7 |
+
|
| 8 |
+
[2026-02-25 13:53:23,871][dinov2][INFO] - using MLP layer as FFN
|
| 9 |
+
[2026-02-25 13:54:48,809][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 10 |
+
warnings.warn( # warn only once
|
| 11 |
+
|
| 12 |
+
[2026-02-25 14:29:41,170][dinov2][INFO] - using MLP layer as FFN
|
| 13 |
+
[2026-02-25 14:30:12,406][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 14 |
+
warnings.warn(
|
| 15 |
+
|
| 16 |
+
[2026-02-25 14:30:12,406][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 17 |
+
warnings.warn(msg)
|
| 18 |
+
|
| 19 |
+
[2026-02-25 14:30:17,604][dinov2][INFO] - using MLP layer as FFN
|
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_4.log
ADDED
|
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2026-02-25 13:52:46,821][dinov2][INFO] - using MLP layer as FFN
|
| 2 |
+
[2026-02-25 13:53:15,687][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 3 |
+
warnings.warn(
|
| 4 |
+
|
| 5 |
+
[2026-02-25 13:53:15,687][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 6 |
+
warnings.warn(msg)
|
| 7 |
+
|
| 8 |
+
[2026-02-25 13:53:20,106][dinov2][INFO] - using MLP layer as FFN
|
| 9 |
+
[2026-02-25 13:54:48,809][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 10 |
+
warnings.warn( # warn only once
|
| 11 |
+
|
| 12 |
+
[2026-02-25 14:29:41,163][dinov2][INFO] - using MLP layer as FFN
|
| 13 |
+
[2026-02-25 14:30:12,743][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 14 |
+
warnings.warn(
|
| 15 |
+
|
| 16 |
+
[2026-02-25 14:30:12,745][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 17 |
+
warnings.warn(msg)
|
| 18 |
+
|
| 19 |
+
[2026-02-25 14:30:17,884][dinov2][INFO] - using MLP layer as FFN
|
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_5.log
ADDED
|
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2026-02-25 13:52:46,848][dinov2][INFO] - using MLP layer as FFN
|
| 2 |
+
[2026-02-25 13:53:17,936][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 3 |
+
warnings.warn(
|
| 4 |
+
|
| 5 |
+
[2026-02-25 13:53:17,940][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 6 |
+
warnings.warn(msg)
|
| 7 |
+
|
| 8 |
+
[2026-02-25 13:53:20,315][dinov2][INFO] - using MLP layer as FFN
|
| 9 |
+
[2026-02-25 13:54:48,813][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 10 |
+
warnings.warn( # warn only once
|
| 11 |
+
|
| 12 |
+
[2026-02-25 14:29:41,164][dinov2][INFO] - using MLP layer as FFN
|
| 13 |
+
[2026-02-25 14:30:09,668][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 14 |
+
warnings.warn(
|
| 15 |
+
|
| 16 |
+
[2026-02-25 14:30:09,670][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 17 |
+
warnings.warn(msg)
|
| 18 |
+
|
| 19 |
+
[2026-02-25 14:30:13,767][dinov2][INFO] - using MLP layer as FFN
|
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_6.log
ADDED
|
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2026-02-25 13:52:46,807][dinov2][INFO] - using MLP layer as FFN
|
| 2 |
+
[2026-02-25 13:53:19,201][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 3 |
+
warnings.warn(
|
| 4 |
+
|
| 5 |
+
[2026-02-25 13:53:19,203][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 6 |
+
warnings.warn(msg)
|
| 7 |
+
|
| 8 |
+
[2026-02-25 13:53:21,001][dinov2][INFO] - using MLP layer as FFN
|
| 9 |
+
[2026-02-25 13:54:48,809][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 10 |
+
warnings.warn( # warn only once
|
| 11 |
+
|
| 12 |
+
[2026-02-25 14:29:41,135][dinov2][INFO] - using MLP layer as FFN
|
| 13 |
+
[2026-02-25 14:30:01,944][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 14 |
+
warnings.warn(
|
| 15 |
+
|
| 16 |
+
[2026-02-25 14:30:01,944][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 17 |
+
warnings.warn(msg)
|
| 18 |
+
|
| 19 |
+
[2026-02-25 14:30:07,824][dinov2][INFO] - using MLP layer as FFN
|
ABLATION_0225_ctxTrain_depth_vggtDistl/train_ddp_process_7.log
ADDED
|
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2026-02-25 13:52:46,799][dinov2][INFO] - using MLP layer as FFN
|
| 2 |
+
[2026-02-25 13:53:19,450][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 3 |
+
warnings.warn(
|
| 4 |
+
|
| 5 |
+
[2026-02-25 13:53:19,450][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 6 |
+
warnings.warn(msg)
|
| 7 |
+
|
| 8 |
+
[2026-02-25 13:53:22,975][dinov2][INFO] - using MLP layer as FFN
|
| 9 |
+
[2026-02-25 13:54:48,809][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 10 |
+
warnings.warn( # warn only once
|
| 11 |
+
|
| 12 |
+
[2026-02-25 14:29:41,375][dinov2][INFO] - using MLP layer as FFN
|
| 13 |
+
[2026-02-25 14:30:00,005][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 14 |
+
warnings.warn(
|
| 15 |
+
|
| 16 |
+
[2026-02-25 14:30:00,006][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 17 |
+
warnings.warn(msg)
|
| 18 |
+
|
| 19 |
+
[2026-02-25 14:30:04,401][dinov2][INFO] - using MLP layer as FFN
|
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/debug-internal.log
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2026-02-25T13:54:42.16430492Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
|
| 2 |
+
{"time":"2026-02-25T13:54:42.614803801Z","level":"INFO","msg":"stream: created new stream","id":"g2qj8ry7"}
|
| 3 |
+
{"time":"2026-02-25T13:54:42.615183417Z","level":"INFO","msg":"handler: started","stream_id":"g2qj8ry7"}
|
| 4 |
+
{"time":"2026-02-25T13:54:42.615386361Z","level":"INFO","msg":"stream: started","id":"g2qj8ry7"}
|
| 5 |
+
{"time":"2026-02-25T13:54:42.615431072Z","level":"INFO","msg":"sender: started","stream_id":"g2qj8ry7"}
|
| 6 |
+
{"time":"2026-02-25T13:54:42.615463052Z","level":"INFO","msg":"writer: started","stream_id":"g2qj8ry7"}
|
| 7 |
+
{"time":"2026-02-25T13:55:10.309513894Z","level":"INFO","msg":"stream: closing","id":"g2qj8ry7"}
|
| 8 |
+
{"time":"2026-02-25T13:55:10.794171575Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
|
| 9 |
+
{"time":"2026-02-25T13:55:11.099789022Z","level":"INFO","msg":"handler: closed","stream_id":"g2qj8ry7"}
|
| 10 |
+
{"time":"2026-02-25T13:55:11.099961565Z","level":"INFO","msg":"sender: closed","stream_id":"g2qj8ry7"}
|
| 11 |
+
{"time":"2026-02-25T13:55:11.099983596Z","level":"INFO","msg":"stream: closed","id":"g2qj8ry7"}
|
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/debug.log
ADDED
|
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
|
| 2 |
+
2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_setup.py:_flush():81] Configure stats pid to 168367
|
| 3 |
+
2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_setup.py:_flush():81] Loading settings from environment variables
|
| 4 |
+
2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug.log
|
| 5 |
+
2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug-internal.log
|
| 6 |
+
2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:init():844] calling init triggers
|
| 7 |
+
2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
|
| 8 |
+
config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'depth', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'absgrad', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': True, 'voxelize_activate': True, 'use_depth': True}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'density_control_loss': {'error_score': {'weight': 0.01, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0225_ctxTrain_depth_vggtDistl', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': True, 'vggt_cam_loss': True, 'vggt_distil': True, 'context_view_train': True}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
|
| 9 |
+
2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:init():892] starting backend
|
| 10 |
+
2026-02-25 13:54:42,153 INFO MainThread:168367 [wandb_init.py:init():895] sending inform_init request
|
| 11 |
+
2026-02-25 13:54:42,160 INFO MainThread:168367 [wandb_init.py:init():903] backend started and connected
|
| 12 |
+
2026-02-25 13:54:42,166 INFO MainThread:168367 [wandb_init.py:init():973] updated telemetry
|
| 13 |
+
2026-02-25 13:54:42,175 INFO MainThread:168367 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
|
| 14 |
+
2026-02-25 13:54:43,768 INFO MainThread:168367 [wandb_init.py:init():1042] starting run threads in backend
|
| 15 |
+
2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_console_start():2524] atexit reg
|
| 16 |
+
2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_redirect():2373] redirect: wrap_raw
|
| 17 |
+
2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_redirect():2442] Wrapping output streams.
|
| 18 |
+
2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_redirect():2465] Redirects installed.
|
| 19 |
+
2026-02-25 13:54:43,896 INFO MainThread:168367 [wandb_init.py:init():1082] run started, returning control to user process
|
| 20 |
+
2026-02-25 13:55:10,309 INFO wandb-AsyncioManager-main:168367 [service_client.py:_forward_responses():134] Reached EOF.
|
| 21 |
+
2026-02-25 13:55:10,309 INFO wandb-AsyncioManager-main:168367 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.
|
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/config.yaml
ADDED
|
@@ -0,0 +1,310 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
_wandb:
|
| 2 |
+
value:
|
| 3 |
+
cli_version: 0.25.0
|
| 4 |
+
e:
|
| 5 |
+
q96rx8ljkgqkb464effa7lv06bj6w0fc:
|
| 6 |
+
args:
|
| 7 |
+
- +experiment=re10k_ablation_24v
|
| 8 |
+
- wandb.mode=online
|
| 9 |
+
- wandb.name=ABLATION_0225_ctxTrain_depth_vggtDistl
|
| 10 |
+
- model.encoder.head_mode=depth
|
| 11 |
+
- model.density_control.use_depth=true
|
| 12 |
+
- train.context_view_train=true
|
| 13 |
+
- train.vggt_distil=true
|
| 14 |
+
cpu_count: 128
|
| 15 |
+
cpu_count_logical: 256
|
| 16 |
+
cudaVersion: "13.1"
|
| 17 |
+
disk:
|
| 18 |
+
/:
|
| 19 |
+
total: "1170378588160"
|
| 20 |
+
used: "756693458944"
|
| 21 |
+
email: dna9041@korea.ac.kr
|
| 22 |
+
executable: /venv/main/bin/python
|
| 23 |
+
git:
|
| 24 |
+
commit: 488f8178913cf0d68a295e744f2709f488f576cd
|
| 25 |
+
remote: git@github.com:K-nowing/CVPR2026.git
|
| 26 |
+
gpu: NVIDIA H200
|
| 27 |
+
gpu_count: 8
|
| 28 |
+
gpu_nvidia:
|
| 29 |
+
- architecture: Hopper
|
| 30 |
+
cudaCores: 16896
|
| 31 |
+
memoryTotal: "150754820096"
|
| 32 |
+
name: NVIDIA H200
|
| 33 |
+
uuid: GPU-2649ab80-a3a6-5a1c-0fa5-12bc11bd75e9
|
| 34 |
+
- architecture: Hopper
|
| 35 |
+
cudaCores: 16896
|
| 36 |
+
memoryTotal: "150754820096"
|
| 37 |
+
name: NVIDIA H200
|
| 38 |
+
uuid: GPU-e92921d9-c681-246f-af93-637e0dc938ca
|
| 39 |
+
- architecture: Hopper
|
| 40 |
+
cudaCores: 16896
|
| 41 |
+
memoryTotal: "150754820096"
|
| 42 |
+
name: NVIDIA H200
|
| 43 |
+
uuid: GPU-ffe12ffc-9bb7-82de-5692-1ec0ee2e68d8
|
| 44 |
+
- architecture: Hopper
|
| 45 |
+
cudaCores: 16896
|
| 46 |
+
memoryTotal: "150754820096"
|
| 47 |
+
name: NVIDIA H200
|
| 48 |
+
uuid: GPU-499e5acd-b6ab-2010-c51b-ee9b5aa65825
|
| 49 |
+
- architecture: Hopper
|
| 50 |
+
cudaCores: 16896
|
| 51 |
+
memoryTotal: "150754820096"
|
| 52 |
+
name: NVIDIA H200
|
| 53 |
+
uuid: GPU-3b2522d9-1c72-e49b-2c30-96165680b74a
|
| 54 |
+
- architecture: Hopper
|
| 55 |
+
cudaCores: 16896
|
| 56 |
+
memoryTotal: "150754820096"
|
| 57 |
+
name: NVIDIA H200
|
| 58 |
+
uuid: GPU-a9a280c5-b2f9-dc1e-a8a9-7326a74001ff
|
| 59 |
+
- architecture: Hopper
|
| 60 |
+
cudaCores: 16896
|
| 61 |
+
memoryTotal: "150754820096"
|
| 62 |
+
name: NVIDIA H200
|
| 63 |
+
uuid: GPU-07d0167b-a6a1-1900-2d27-7c6c11598409
|
| 64 |
+
- architecture: Hopper
|
| 65 |
+
cudaCores: 16896
|
| 66 |
+
memoryTotal: "150754820096"
|
| 67 |
+
name: NVIDIA H200
|
| 68 |
+
uuid: GPU-8362a999-20d1-c27b-5d18-032d23f859ab
|
| 69 |
+
host: 27d18dedec6d
|
| 70 |
+
memory:
|
| 71 |
+
total: "1622948257792"
|
| 72 |
+
os: Linux-6.8.0-90-generic-x86_64-with-glibc2.39
|
| 73 |
+
program: -m src.main
|
| 74 |
+
python: CPython 3.12.12
|
| 75 |
+
root: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl
|
| 76 |
+
startedAt: "2026-02-25T13:54:41.861013Z"
|
| 77 |
+
writerId: q96rx8ljkgqkb464effa7lv06bj6w0fc
|
| 78 |
+
m:
|
| 79 |
+
- "1": trainer/global_step
|
| 80 |
+
"6":
|
| 81 |
+
- 3
|
| 82 |
+
"7": []
|
| 83 |
+
- "2": '*'
|
| 84 |
+
"5": 1
|
| 85 |
+
"6":
|
| 86 |
+
- 1
|
| 87 |
+
"7": []
|
| 88 |
+
python_version: 3.12.12
|
| 89 |
+
t:
|
| 90 |
+
"1":
|
| 91 |
+
- 1
|
| 92 |
+
- 41
|
| 93 |
+
- 49
|
| 94 |
+
- 50
|
| 95 |
+
- 106
|
| 96 |
+
"2":
|
| 97 |
+
- 1
|
| 98 |
+
- 41
|
| 99 |
+
- 49
|
| 100 |
+
- 50
|
| 101 |
+
- 106
|
| 102 |
+
"3":
|
| 103 |
+
- 7
|
| 104 |
+
- 13
|
| 105 |
+
- 15
|
| 106 |
+
- 16
|
| 107 |
+
- 66
|
| 108 |
+
"4": 3.12.12
|
| 109 |
+
"5": 0.25.0
|
| 110 |
+
"12": 0.25.0
|
| 111 |
+
"13": linux-x86_64
|
| 112 |
+
checkpointing:
|
| 113 |
+
value:
|
| 114 |
+
every_n_train_steps: 1500
|
| 115 |
+
load: null
|
| 116 |
+
save_top_k: 2
|
| 117 |
+
save_weights_only: false
|
| 118 |
+
data_loader:
|
| 119 |
+
value:
|
| 120 |
+
test:
|
| 121 |
+
batch_size: 1
|
| 122 |
+
num_workers: 4
|
| 123 |
+
persistent_workers: false
|
| 124 |
+
seed: 2345
|
| 125 |
+
train:
|
| 126 |
+
batch_size: 16
|
| 127 |
+
num_workers: 16
|
| 128 |
+
persistent_workers: true
|
| 129 |
+
seed: 1234
|
| 130 |
+
val:
|
| 131 |
+
batch_size: 1
|
| 132 |
+
num_workers: 1
|
| 133 |
+
persistent_workers: true
|
| 134 |
+
seed: 3456
|
| 135 |
+
dataset:
|
| 136 |
+
value:
|
| 137 |
+
re10k:
|
| 138 |
+
augment: true
|
| 139 |
+
background_color:
|
| 140 |
+
- 0
|
| 141 |
+
- 0
|
| 142 |
+
- 0
|
| 143 |
+
baseline_max: 1e+10
|
| 144 |
+
baseline_min: 0.001
|
| 145 |
+
cameras_are_circular: false
|
| 146 |
+
dynamic_context_views: true
|
| 147 |
+
input_image_shape:
|
| 148 |
+
- 256
|
| 149 |
+
- 256
|
| 150 |
+
make_baseline_1: true
|
| 151 |
+
max_context_views_per_gpu: 24
|
| 152 |
+
max_fov: 100
|
| 153 |
+
name: re10k
|
| 154 |
+
original_image_shape:
|
| 155 |
+
- 360
|
| 156 |
+
- 640
|
| 157 |
+
overfit_to_scene: null
|
| 158 |
+
relative_pose: true
|
| 159 |
+
roots:
|
| 160 |
+
- datasets/re10k
|
| 161 |
+
skip_bad_shape: true
|
| 162 |
+
view_sampler:
|
| 163 |
+
initial_max_distance_between_context_views: 25
|
| 164 |
+
initial_min_distance_between_context_views: 25
|
| 165 |
+
max_distance_between_context_views: 90
|
| 166 |
+
min_distance_between_context_views: 45
|
| 167 |
+
min_distance_to_context_views: 0
|
| 168 |
+
name: bounded
|
| 169 |
+
num_context_views: 2
|
| 170 |
+
num_target_set: 3
|
| 171 |
+
num_target_views: 4
|
| 172 |
+
same_target_gap: false
|
| 173 |
+
warm_up_steps: 1000
|
| 174 |
+
density_control_loss:
|
| 175 |
+
value:
|
| 176 |
+
error_score:
|
| 177 |
+
grad_scale: 10000
|
| 178 |
+
log_scale: false
|
| 179 |
+
mode: original
|
| 180 |
+
weight: 0.01
|
| 181 |
+
direct_loss:
|
| 182 |
+
value:
|
| 183 |
+
l1:
|
| 184 |
+
weight: 0.8
|
| 185 |
+
ssim:
|
| 186 |
+
weight: 0.2
|
| 187 |
+
mode:
|
| 188 |
+
value: train
|
| 189 |
+
model:
|
| 190 |
+
value:
|
| 191 |
+
decoder:
|
| 192 |
+
background_color:
|
| 193 |
+
- 0
|
| 194 |
+
- 0
|
| 195 |
+
- 0
|
| 196 |
+
make_scale_invariant: false
|
| 197 |
+
name: splatting_cuda
|
| 198 |
+
density_control:
|
| 199 |
+
aggregation_mode: mean
|
| 200 |
+
aux_refine: false
|
| 201 |
+
grad_mode: absgrad
|
| 202 |
+
gs_param_dim: 256
|
| 203 |
+
latent_dim: 128
|
| 204 |
+
mean_dim: 32
|
| 205 |
+
name: density_control_module
|
| 206 |
+
num_heads: 1
|
| 207 |
+
num_latents: 64
|
| 208 |
+
num_level: 3
|
| 209 |
+
num_self_attn_per_block: 2
|
| 210 |
+
refine_error: false
|
| 211 |
+
refinement_hidden_dim: 32
|
| 212 |
+
refinement_layer_num: 1
|
| 213 |
+
refinement_type: voxelize
|
| 214 |
+
score_mode: absgrad
|
| 215 |
+
use_depth: true
|
| 216 |
+
use_mean_features: true
|
| 217 |
+
use_refine_module: true
|
| 218 |
+
voxel_size: 0.001
|
| 219 |
+
voxelize_activate: true
|
| 220 |
+
encoder:
|
| 221 |
+
align_corners: false
|
| 222 |
+
gs_param_dim: 256
|
| 223 |
+
head_mode: depth
|
| 224 |
+
input_image_shape:
|
| 225 |
+
- 518
|
| 226 |
+
- 518
|
| 227 |
+
name: dcsplat
|
| 228 |
+
num_level: 3
|
| 229 |
+
use_voxelize: true
|
| 230 |
+
optimizer:
|
| 231 |
+
value:
|
| 232 |
+
accumulate: 1
|
| 233 |
+
backbone_lr_multiplier: 0.1
|
| 234 |
+
backbone_trainable: T+H
|
| 235 |
+
lr: 0.0002
|
| 236 |
+
warm_up_steps: 25
|
| 237 |
+
render_loss:
|
| 238 |
+
value:
|
| 239 |
+
lpips:
|
| 240 |
+
apply_after_step: 0
|
| 241 |
+
weight: 0.05
|
| 242 |
+
mse:
|
| 243 |
+
weight: 1
|
| 244 |
+
seed:
|
| 245 |
+
value: 111123
|
| 246 |
+
test:
|
| 247 |
+
value:
|
| 248 |
+
align_pose: false
|
| 249 |
+
compute_scores: true
|
| 250 |
+
error_threshold: 0.4
|
| 251 |
+
error_threshold_list:
|
| 252 |
+
- 0.2
|
| 253 |
+
- 0.4
|
| 254 |
+
- 0.6
|
| 255 |
+
- 0.8
|
| 256 |
+
- 1
|
| 257 |
+
nvs_view_N_list:
|
| 258 |
+
- 3
|
| 259 |
+
- 6
|
| 260 |
+
- 16
|
| 261 |
+
- 32
|
| 262 |
+
- 64
|
| 263 |
+
output_path: test/ablation/re10k
|
| 264 |
+
pose_align_steps: 100
|
| 265 |
+
pred_intrinsic: false
|
| 266 |
+
rot_opt_lr: 0.005
|
| 267 |
+
save_active_mask_image: false
|
| 268 |
+
save_compare: false
|
| 269 |
+
save_error_score_image: false
|
| 270 |
+
save_image: false
|
| 271 |
+
save_video: false
|
| 272 |
+
threshold_mode: ratio
|
| 273 |
+
trans_opt_lr: 0.005
|
| 274 |
+
train:
|
| 275 |
+
value:
|
| 276 |
+
align_corners: false
|
| 277 |
+
beta_dist_param:
|
| 278 |
+
- 0.5
|
| 279 |
+
- 4
|
| 280 |
+
cam_scale_mode: sum
|
| 281 |
+
camera_loss: 10
|
| 282 |
+
context_view_train: true
|
| 283 |
+
ext_scale_detach: false
|
| 284 |
+
extended_visualization: false
|
| 285 |
+
intrinsic_scaling: false
|
| 286 |
+
one_sample_validation: null
|
| 287 |
+
print_log_every_n_steps: 10
|
| 288 |
+
scene_scale_reg_loss: 0.01
|
| 289 |
+
train_aux: true
|
| 290 |
+
train_gs_num: 1
|
| 291 |
+
train_target_set: true
|
| 292 |
+
use_refine_aux: false
|
| 293 |
+
verbose: false
|
| 294 |
+
vggt_cam_loss: true
|
| 295 |
+
vggt_distil: true
|
| 296 |
+
trainer:
|
| 297 |
+
value:
|
| 298 |
+
gradient_clip_val: 0.5
|
| 299 |
+
max_steps: 3001
|
| 300 |
+
num_nodes: 1
|
| 301 |
+
val_check_interval: 250
|
| 302 |
+
wandb:
|
| 303 |
+
value:
|
| 304 |
+
entity: scene-representation-group
|
| 305 |
+
mode: online
|
| 306 |
+
name: ABLATION_0225_ctxTrain_depth_vggtDistl
|
| 307 |
+
project: DCSplat
|
| 308 |
+
tags:
|
| 309 |
+
- re10k
|
| 310 |
+
- 256x256
|
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/media/images/active_mask_imgs_1_690a5f29fe9230f3ead3.png
ADDED
|
Git LFS Details
|
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/media/images/comparison_0_e330894a93360b5335a1.png
ADDED
|
Git LFS Details
|
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/media/images/error_scores_2_bf819ed247f7546e0637.png
ADDED
|
Git LFS Details
|
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/output.log
ADDED
|
@@ -0,0 +1,140 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7]
|
| 2 |
+
|
| 3 |
+
| Name | Type | Params | Mode
|
| 4 |
+
------------------------------------------------------------------------
|
| 5 |
+
0 | encoder | OurSplat | 888 M | train
|
| 6 |
+
1 | density_control_module | DensityControlModule | 2.6 M | train
|
| 7 |
+
2 | decoder | DecoderSplattingCUDA | 0 | train
|
| 8 |
+
3 | render_losses | ModuleList | 0 | train
|
| 9 |
+
4 | density_control_losses | ModuleList | 0 | train
|
| 10 |
+
5 | direct_losses | ModuleList | 0 | train
|
| 11 |
+
6 | distill_aggregator | Aggregator | 909 M | train
|
| 12 |
+
7 | distill_camera_head | CameraHead | 216 M | train
|
| 13 |
+
8 | distill_depth_head | DPTHead | 32.7 M | train
|
| 14 |
+
9 | loss_distill | DistillLoss | 0 | train
|
| 15 |
+
------------------------------------------------------------------------
|
| 16 |
+
891 M Trainable params
|
| 17 |
+
1.2 B Non-trainable params
|
| 18 |
+
2.0 B Total params
|
| 19 |
+
8,196.093 Total estimated model params size (MB)
|
| 20 |
+
2778 Modules in train mode
|
| 21 |
+
522 Modules in eval mode
|
| 22 |
+
Sanity Checking: | | 0/? [00:00<?, ?it/s][2026-02-25 13:54:48,808][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=31` in the `DataLoader` to improve performance.
|
| 23 |
+
|
| 24 |
+
[2026-02-25 13:54:48,809][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4807: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
|
| 25 |
+
warnings.warn( # warn only once
|
| 26 |
+
|
| 27 |
+
Validation epoch start on rank 0
|
| 28 |
+
Sanity Checking DataLoader 0: 0%| | 0/1 [00:00<?, ?it/s]validation step 0; scene = ['306e2b7785657539'];
|
| 29 |
+
target intrinsic: tensor(0.8595, device='cuda:0') tensor(0.8597, device='cuda:0')
|
| 30 |
+
pred intrinsic: tensor(0.8779, device='cuda:0') tensor(0.8773, device='cuda:0')
|
| 31 |
+
[rank0]:W0225 13:54:51.619000 168367 site-packages/torch/utils/cpp_extension.py:2425] TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
|
| 32 |
+
[rank0]:W0225 13:54:51.619000 168367 site-packages/torch/utils/cpp_extension.py:2425] If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'] to specific architectures.
|
| 33 |
+
[2026-02-25 13:54:51,686][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 34 |
+
result[selector] = overlay
|
| 35 |
+
|
| 36 |
+
[2026-02-25 13:54:51,695][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
|
| 37 |
+
|
| 38 |
+
Setting up [LPIPS] perceptual loss: trunk [vgg], v[0.1], spatial [off]
|
| 39 |
+
[2026-02-25 13:54:51,696][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
|
| 40 |
+
warnings.warn(
|
| 41 |
+
|
| 42 |
+
[2026-02-25 13:54:51,696][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
|
| 43 |
+
warnings.warn(msg)
|
| 44 |
+
|
| 45 |
+
Loading model from: /venv/main/lib/python3.12/site-packages/lpips/weights/v0.1/vgg.pth
|
| 46 |
+
[2026-02-25 13:54:53,564][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4322.)
|
| 47 |
+
return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
|
| 48 |
+
|
| 49 |
+
Sanity Checking DataLoader 0: 100%|████████████████████████████████████████████████████████████████████| 1/1 [00:04<00:00, 0.23it/s][2026-02-25 13:54:53,845][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
|
| 50 |
+
|
| 51 |
+
[2026-02-25 13:54:53,847][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
|
| 52 |
+
|
| 53 |
+
[2026-02-25 13:54:53,847][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
|
| 54 |
+
|
| 55 |
+
[2026-02-25 13:54:53,848][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
|
| 56 |
+
|
| 57 |
+
[2026-02-25 13:54:53,848][py.warnings][WARNING] - /venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('info/global_step', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
|
| 58 |
+
|
| 59 |
+
Epoch 0: | | 0/? [00:00<?, ?it/s]context = [[34, 36, 50, 53, 54, 60, 63, 70, 76, 78, 79, 80, 81, 88, 92, 94, 102, 110, 112, 114, 122, 125, 126, 131]]target = [[126, 96, 109, 55, 99, 116, 43, 60, 113, 85, 103, 90, 130, 62, 76, 123, 35, 102, 125, 128, 98, 67, 129, 79]]
|
| 60 |
+
Error executing job with overrides: ['+experiment=re10k_ablation_24v', 'wandb.mode=online', 'wandb.name=ABLATION_0225_ctxTrain_depth_vggtDistl', 'model.encoder.head_mode=depth', 'model.density_control.use_depth=true', 'train.context_view_train=true', 'train.vggt_distil=true']
|
| 61 |
+
Traceback (most recent call last):
|
| 62 |
+
File "/workspace/code/CVPR2026/src/main.py", line 226, in train
|
| 63 |
+
trainer.fit(model_wrapper, datamodule=data_module)#, ckpt_path=checkpoint_path)
|
| 64 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 65 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 561, in fit
|
| 66 |
+
call._call_and_handle_interrupt(
|
| 67 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/call.py", line 47, in _call_and_handle_interrupt
|
| 68 |
+
return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs)
|
| 69 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 70 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch
|
| 71 |
+
return function(*args, **kwargs)
|
| 72 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 73 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 599, in _fit_impl
|
| 74 |
+
self._run(model, ckpt_path=ckpt_path)
|
| 75 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 1012, in _run
|
| 76 |
+
results = self._run_stage()
|
| 77 |
+
^^^^^^^^^^^^^^^^^
|
| 78 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/trainer.py", line 1056, in _run_stage
|
| 79 |
+
self.fit_loop.run()
|
| 80 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run
|
| 81 |
+
self.advance()
|
| 82 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/fit_loop.py", line 455, in advance
|
| 83 |
+
self.epoch_loop.run(self._data_fetcher)
|
| 84 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 150, in run
|
| 85 |
+
self.advance(data_fetcher)
|
| 86 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 322, in advance
|
| 87 |
+
batch_output = self.manual_optimization.run(kwargs)
|
| 88 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 89 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/optimization/manual.py", line 94, in run
|
| 90 |
+
self.advance(kwargs)
|
| 91 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/loops/optimization/manual.py", line 114, in advance
|
| 92 |
+
training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values())
|
| 93 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 94 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/trainer/call.py", line 328, in _call_strategy_hook
|
| 95 |
+
output = fn(*args, **kwargs)
|
| 96 |
+
^^^^^^^^^^^^^^^^^^^
|
| 97 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step
|
| 98 |
+
return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs)
|
| 99 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 100 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__
|
| 101 |
+
wrapper_output = wrapper_module(*args, **kwargs)
|
| 102 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 103 |
+
File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl
|
| 104 |
+
return self._call_impl(*args, **kwargs)
|
| 105 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 106 |
+
File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl
|
| 107 |
+
return forward_call(*args, **kwargs)
|
| 108 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 109 |
+
File "/venv/main/lib/python3.12/site-packages/torch/nn/parallel/distributed.py", line 1648, in forward
|
| 110 |
+
else self._run_ddp_forward(*inputs, **kwargs)
|
| 111 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 112 |
+
File "/venv/main/lib/python3.12/site-packages/torch/nn/parallel/distributed.py", line 1474, in _run_ddp_forward
|
| 113 |
+
return self.module(*inputs, **kwargs) # type: ignore[index]
|
| 114 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 115 |
+
File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl
|
| 116 |
+
return self._call_impl(*args, **kwargs)
|
| 117 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 118 |
+
File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl
|
| 119 |
+
return forward_call(*args, **kwargs)
|
| 120 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 121 |
+
File "/venv/main/lib/python3.12/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward
|
| 122 |
+
out = method(*_args, **_kwargs)
|
| 123 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 124 |
+
File "/venv/main/lib/python3.12/site-packages/jaxtyping/_decorator.py", line 562, in wrapped_fn
|
| 125 |
+
return wrapped_fn_impl(args, kwargs, bound, memos)
|
| 126 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 127 |
+
File "/venv/main/lib/python3.12/site-packages/jaxtyping/_decorator.py", line 486, in wrapped_fn_impl
|
| 128 |
+
out = fn(*args, **kwargs)
|
| 129 |
+
^^^^^^^^^^^^^^^^^^^
|
| 130 |
+
File "/workspace/code/CVPR2026/src/model/model_wrapper.py", line 563, in training_step
|
| 131 |
+
depth_consis_loss = self.render_losses[2].forward(output, batch, depth_dict, self.global_step)
|
| 132 |
+
~~~~~~~~~~~~~~~~~~^^^
|
| 133 |
+
File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/container.py", line 377, in __getitem__
|
| 134 |
+
return self._modules[self._get_abs_string_index(idx)]
|
| 135 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 136 |
+
File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/container.py", line 361, in _get_abs_string_index
|
| 137 |
+
raise IndexError(f"index {idx} is out of range")
|
| 138 |
+
IndexError: index 2 is out of range
|
| 139 |
+
|
| 140 |
+
Set the environment variable HYDRA_FULL_ERROR=1 for a complete stack trace.
|
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/requirements.txt
ADDED
|
@@ -0,0 +1,172 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
wheel==0.45.1
|
| 2 |
+
pytz==2025.2
|
| 3 |
+
easydict==1.13
|
| 4 |
+
antlr4-python3-runtime==4.9.3
|
| 5 |
+
wadler_lindig==0.1.7
|
| 6 |
+
urllib3==2.5.0
|
| 7 |
+
tzdata==2025.2
|
| 8 |
+
typing-inspection==0.4.1
|
| 9 |
+
tabulate==0.9.0
|
| 10 |
+
smmap==5.0.2
|
| 11 |
+
kornia_rs==0.1.9
|
| 12 |
+
setuptools==78.1.1
|
| 13 |
+
safetensors==0.5.3
|
| 14 |
+
PyYAML==6.0.2
|
| 15 |
+
PySocks==1.7.1
|
| 16 |
+
pyparsing==3.2.5
|
| 17 |
+
pydantic_core==2.33.2
|
| 18 |
+
pycparser==2.23
|
| 19 |
+
protobuf==6.32.1
|
| 20 |
+
propcache==0.3.2
|
| 21 |
+
proglog==0.1.12
|
| 22 |
+
fsspec==2024.6.1
|
| 23 |
+
platformdirs==4.4.0
|
| 24 |
+
pip==25.2
|
| 25 |
+
pillow==10.4.0
|
| 26 |
+
frozenlist==1.7.0
|
| 27 |
+
packaging==24.2
|
| 28 |
+
opt_einsum==3.4.0
|
| 29 |
+
numpy==1.26.4
|
| 30 |
+
ninja==1.13.0
|
| 31 |
+
fonttools==4.60.0
|
| 32 |
+
networkx==3.4.2
|
| 33 |
+
multidict==6.6.4
|
| 34 |
+
mdurl==0.1.2
|
| 35 |
+
MarkupSafe==3.0.2
|
| 36 |
+
kiwisolver==1.4.9
|
| 37 |
+
imageio-ffmpeg==0.6.0
|
| 38 |
+
idna==3.7
|
| 39 |
+
hf-xet==1.1.10
|
| 40 |
+
gmpy2==2.2.1
|
| 41 |
+
einops==0.8.1
|
| 42 |
+
filelock==3.17.0
|
| 43 |
+
decorator==4.4.2
|
| 44 |
+
dacite==1.9.2
|
| 45 |
+
cycler==0.12.1
|
| 46 |
+
colorama==0.4.6
|
| 47 |
+
click==8.3.0
|
| 48 |
+
nvidia-nvtx-cu12==12.8.90
|
| 49 |
+
charset-normalizer==3.3.2
|
| 50 |
+
certifi==2025.8.3
|
| 51 |
+
beartype==0.19.0
|
| 52 |
+
attrs==25.3.0
|
| 53 |
+
async-timeout==5.0.1
|
| 54 |
+
annotated-types==0.7.0
|
| 55 |
+
aiohappyeyeballs==2.6.1
|
| 56 |
+
yarl==1.20.1
|
| 57 |
+
tifffile==2025.5.10
|
| 58 |
+
sentry-sdk==2.39.0
|
| 59 |
+
scipy==1.15.3
|
| 60 |
+
pydantic==2.11.9
|
| 61 |
+
pandas==2.3.2
|
| 62 |
+
opencv-python==4.11.0.86
|
| 63 |
+
omegaconf==2.3.0
|
| 64 |
+
markdown-it-py==4.0.0
|
| 65 |
+
lightning-utilities==0.14.3
|
| 66 |
+
lazy_loader==0.4
|
| 67 |
+
jaxtyping==0.2.37
|
| 68 |
+
imageio==2.37.0
|
| 69 |
+
gitdb==4.0.12
|
| 70 |
+
contourpy==1.3.2
|
| 71 |
+
colorspacious==1.1.2
|
| 72 |
+
cffi==1.17.1
|
| 73 |
+
aiosignal==1.4.0
|
| 74 |
+
scikit-video==1.1.11
|
| 75 |
+
scikit-image==0.25.2
|
| 76 |
+
rich==14.1.0
|
| 77 |
+
moviepy==1.0.3
|
| 78 |
+
matplotlib==3.10.6
|
| 79 |
+
hydra-core==1.3.2
|
| 80 |
+
nvidia-nccl-cu12==2.27.3
|
| 81 |
+
huggingface-hub==0.35.1
|
| 82 |
+
GitPython==3.1.45
|
| 83 |
+
brotlicffi==1.0.9.2
|
| 84 |
+
aiohttp==3.12.15
|
| 85 |
+
torchmetrics==1.8.2
|
| 86 |
+
opt-einsum-fx==0.1.4
|
| 87 |
+
kornia==0.8.1
|
| 88 |
+
pytorch-lightning==2.5.1
|
| 89 |
+
lpips==0.1.4
|
| 90 |
+
e3nn==0.6.0
|
| 91 |
+
lightning==2.5.1
|
| 92 |
+
nvidia-cusparselt-cu12==0.7.1
|
| 93 |
+
triton==3.4.0
|
| 94 |
+
nvidia-nvjitlink-cu12==12.8.93
|
| 95 |
+
nvidia-curand-cu12==10.3.9.90
|
| 96 |
+
nvidia-cufile-cu12==1.13.1.3
|
| 97 |
+
nvidia-cuda-runtime-cu12==12.8.90
|
| 98 |
+
nvidia-cuda-nvrtc-cu12==12.8.93
|
| 99 |
+
nvidia-cuda-cupti-cu12==12.8.90
|
| 100 |
+
nvidia-cublas-cu12==12.8.4.1
|
| 101 |
+
nvidia-cusparse-cu12==12.5.8.93
|
| 102 |
+
nvidia-cufft-cu12==11.3.3.83
|
| 103 |
+
nvidia-cudnn-cu12==9.10.2.21
|
| 104 |
+
nvidia-cusolver-cu12==11.7.3.90
|
| 105 |
+
torch==2.8.0+cu128
|
| 106 |
+
torchvision==0.23.0+cu128
|
| 107 |
+
torchaudio==2.8.0+cu128
|
| 108 |
+
torch_scatter==2.1.2+pt28cu128
|
| 109 |
+
gsplat==1.5.3
|
| 110 |
+
wandb==0.25.0
|
| 111 |
+
cuda-bindings==13.0.3
|
| 112 |
+
cuda-pathfinder==1.3.3
|
| 113 |
+
Jinja2==3.1.6
|
| 114 |
+
mpmath==1.3.0
|
| 115 |
+
nvidia-cublas==13.1.0.3
|
| 116 |
+
nvidia-cuda-cupti==13.0.85
|
| 117 |
+
nvidia-cuda-nvrtc==13.0.88
|
| 118 |
+
nvidia-cuda-runtime==13.0.96
|
| 119 |
+
nvidia-cudnn-cu13==9.15.1.9
|
| 120 |
+
nvidia-cufft==12.0.0.61
|
| 121 |
+
nvidia-cufile==1.15.1.6
|
| 122 |
+
nvidia-curand==10.4.0.35
|
| 123 |
+
nvidia-cusolver==12.0.4.66
|
| 124 |
+
nvidia-cusparse==12.6.3.3
|
| 125 |
+
nvidia-cusparselt-cu13==0.8.0
|
| 126 |
+
nvidia-nccl-cu13==2.28.9
|
| 127 |
+
nvidia-nvjitlink==13.0.88
|
| 128 |
+
nvidia-nvshmem-cu13==3.4.5
|
| 129 |
+
nvidia-nvtx==13.0.85
|
| 130 |
+
requests==2.32.5
|
| 131 |
+
sentencepiece==0.2.1
|
| 132 |
+
sympy==1.14.0
|
| 133 |
+
torchcodec==0.10.0
|
| 134 |
+
torchdata==0.10.0
|
| 135 |
+
torchtext==0.6.0
|
| 136 |
+
anyio==4.12.0
|
| 137 |
+
asttokens==3.0.1
|
| 138 |
+
comm==0.2.3
|
| 139 |
+
debugpy==1.8.19
|
| 140 |
+
executing==2.2.1
|
| 141 |
+
h11==0.16.0
|
| 142 |
+
httpcore==1.0.9
|
| 143 |
+
httpx==0.28.1
|
| 144 |
+
ipykernel==7.1.0
|
| 145 |
+
ipython==9.8.0
|
| 146 |
+
ipython_pygments_lexers==1.1.1
|
| 147 |
+
ipywidgets==8.1.8
|
| 148 |
+
jedi==0.19.2
|
| 149 |
+
jupyter_client==8.7.0
|
| 150 |
+
jupyter_core==5.9.1
|
| 151 |
+
jupyterlab_widgets==3.0.16
|
| 152 |
+
matplotlib-inline==0.2.1
|
| 153 |
+
nest-asyncio==1.6.0
|
| 154 |
+
parso==0.8.5
|
| 155 |
+
pexpect==4.9.0
|
| 156 |
+
prompt_toolkit==3.0.52
|
| 157 |
+
psutil==7.2.1
|
| 158 |
+
ptyprocess==0.7.0
|
| 159 |
+
pure_eval==0.2.3
|
| 160 |
+
Pygments==2.19.2
|
| 161 |
+
python-dateutil==2.9.0.post0
|
| 162 |
+
pyzmq==27.1.0
|
| 163 |
+
shellingham==1.5.4
|
| 164 |
+
six==1.17.0
|
| 165 |
+
stack-data==0.6.3
|
| 166 |
+
tornado==6.5.4
|
| 167 |
+
tqdm==4.67.1
|
| 168 |
+
traitlets==5.14.3
|
| 169 |
+
typer-slim==0.21.0
|
| 170 |
+
typing_extensions==4.15.0
|
| 171 |
+
wcwidth==0.2.14
|
| 172 |
+
widgetsnbextension==4.0.15
|
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/wandb-metadata.json
ADDED
|
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"os": "Linux-6.8.0-90-generic-x86_64-with-glibc2.39",
|
| 3 |
+
"python": "CPython 3.12.12",
|
| 4 |
+
"startedAt": "2026-02-25T13:54:41.861013Z",
|
| 5 |
+
"args": [
|
| 6 |
+
"+experiment=re10k_ablation_24v",
|
| 7 |
+
"wandb.mode=online",
|
| 8 |
+
"wandb.name=ABLATION_0225_ctxTrain_depth_vggtDistl",
|
| 9 |
+
"model.encoder.head_mode=depth",
|
| 10 |
+
"model.density_control.use_depth=true",
|
| 11 |
+
"train.context_view_train=true",
|
| 12 |
+
"train.vggt_distil=true"
|
| 13 |
+
],
|
| 14 |
+
"program": "-m src.main",
|
| 15 |
+
"git": {
|
| 16 |
+
"remote": "git@github.com:K-nowing/CVPR2026.git",
|
| 17 |
+
"commit": "488f8178913cf0d68a295e744f2709f488f576cd"
|
| 18 |
+
},
|
| 19 |
+
"email": "dna9041@korea.ac.kr",
|
| 20 |
+
"root": "/workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl",
|
| 21 |
+
"host": "27d18dedec6d",
|
| 22 |
+
"executable": "/venv/main/bin/python",
|
| 23 |
+
"cpu_count": 128,
|
| 24 |
+
"cpu_count_logical": 256,
|
| 25 |
+
"gpu": "NVIDIA H200",
|
| 26 |
+
"gpu_count": 8,
|
| 27 |
+
"disk": {
|
| 28 |
+
"/": {
|
| 29 |
+
"total": "1170378588160",
|
| 30 |
+
"used": "756693458944"
|
| 31 |
+
}
|
| 32 |
+
},
|
| 33 |
+
"memory": {
|
| 34 |
+
"total": "1622948257792"
|
| 35 |
+
},
|
| 36 |
+
"gpu_nvidia": [
|
| 37 |
+
{
|
| 38 |
+
"name": "NVIDIA H200",
|
| 39 |
+
"memoryTotal": "150754820096",
|
| 40 |
+
"cudaCores": 16896,
|
| 41 |
+
"architecture": "Hopper",
|
| 42 |
+
"uuid": "GPU-2649ab80-a3a6-5a1c-0fa5-12bc11bd75e9"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"name": "NVIDIA H200",
|
| 46 |
+
"memoryTotal": "150754820096",
|
| 47 |
+
"cudaCores": 16896,
|
| 48 |
+
"architecture": "Hopper",
|
| 49 |
+
"uuid": "GPU-e92921d9-c681-246f-af93-637e0dc938ca"
|
| 50 |
+
},
|
| 51 |
+
{
|
| 52 |
+
"name": "NVIDIA H200",
|
| 53 |
+
"memoryTotal": "150754820096",
|
| 54 |
+
"cudaCores": 16896,
|
| 55 |
+
"architecture": "Hopper",
|
| 56 |
+
"uuid": "GPU-ffe12ffc-9bb7-82de-5692-1ec0ee2e68d8"
|
| 57 |
+
},
|
| 58 |
+
{
|
| 59 |
+
"name": "NVIDIA H200",
|
| 60 |
+
"memoryTotal": "150754820096",
|
| 61 |
+
"cudaCores": 16896,
|
| 62 |
+
"architecture": "Hopper",
|
| 63 |
+
"uuid": "GPU-499e5acd-b6ab-2010-c51b-ee9b5aa65825"
|
| 64 |
+
},
|
| 65 |
+
{
|
| 66 |
+
"name": "NVIDIA H200",
|
| 67 |
+
"memoryTotal": "150754820096",
|
| 68 |
+
"cudaCores": 16896,
|
| 69 |
+
"architecture": "Hopper",
|
| 70 |
+
"uuid": "GPU-3b2522d9-1c72-e49b-2c30-96165680b74a"
|
| 71 |
+
},
|
| 72 |
+
{
|
| 73 |
+
"name": "NVIDIA H200",
|
| 74 |
+
"memoryTotal": "150754820096",
|
| 75 |
+
"cudaCores": 16896,
|
| 76 |
+
"architecture": "Hopper",
|
| 77 |
+
"uuid": "GPU-a9a280c5-b2f9-dc1e-a8a9-7326a74001ff"
|
| 78 |
+
},
|
| 79 |
+
{
|
| 80 |
+
"name": "NVIDIA H200",
|
| 81 |
+
"memoryTotal": "150754820096",
|
| 82 |
+
"cudaCores": 16896,
|
| 83 |
+
"architecture": "Hopper",
|
| 84 |
+
"uuid": "GPU-07d0167b-a6a1-1900-2d27-7c6c11598409"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"name": "NVIDIA H200",
|
| 88 |
+
"memoryTotal": "150754820096",
|
| 89 |
+
"cudaCores": 16896,
|
| 90 |
+
"architecture": "Hopper",
|
| 91 |
+
"uuid": "GPU-8362a999-20d1-c27b-5d18-032d23f859ab"
|
| 92 |
+
}
|
| 93 |
+
],
|
| 94 |
+
"cudaVersion": "13.1",
|
| 95 |
+
"writerId": "q96rx8ljkgqkb464effa7lv06bj6w0fc"
|
| 96 |
+
}
|
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/files/wandb-summary.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"trainer/global_step":0,"active_mask_imgs":{"width":536,"height":800,"format":"png","count":1,"filenames":["media/images/active_mask_imgs_1_690a5f29fe9230f3ead3.png"],"captions":["306e2b7785657539"],"_type":"images/separated"},"error_scores":{"_type":"images/separated","width":800,"height":536,"format":"png","count":1,"filenames":["media/images/error_scores_2_bf819ed247f7546e0637.png"],"captions":["306e2b7785657539"]},"_wandb":{"runtime":26},"_runtime":26,"_timestamp":1.7720276938425457e+09,"_step":2,"comparison":{"format":"png","count":1,"filenames":["media/images/comparison_0_e330894a93360b5335a1.png"],"captions":["306e2b7785657539"],"_type":"images/separated","width":1064,"height":1098}}
|
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug-core.log
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2026-02-25T13:54:41.978700745Z","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpw2b5bx5a/port-168367.txt","pid":168367,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
|
| 2 |
+
{"time":"2026-02-25T13:54:41.980052668Z","level":"INFO","msg":"server: will exit if parent process dies","ppid":168367}
|
| 3 |
+
{"time":"2026-02-25T13:54:41.980043458Z","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-168367-170905-176849560/socket","Net":"unix"}}
|
| 4 |
+
{"time":"2026-02-25T13:54:42.152857208Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
|
| 5 |
+
{"time":"2026-02-25T13:54:42.164042466Z","level":"INFO","msg":"handleInformInit: received","streamId":"g2qj8ry7","id":"1(@)"}
|
| 6 |
+
{"time":"2026-02-25T13:54:42.615403591Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"g2qj8ry7","id":"1(@)"}
|
| 7 |
+
{"time":"2026-02-25T13:54:49.068751173Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"1riv14vupx9g"}
|
| 8 |
+
{"time":"2026-02-25T13:55:10.309398102Z","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
|
| 9 |
+
{"time":"2026-02-25T13:55:10.309487963Z","level":"INFO","msg":"connection: closing","id":"1(@)"}
|
| 10 |
+
{"time":"2026-02-25T13:55:10.309511483Z","level":"INFO","msg":"server is shutting down"}
|
| 11 |
+
{"time":"2026-02-25T13:55:10.309607405Z","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
|
| 12 |
+
{"time":"2026-02-25T13:55:10.309714557Z","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-168367-170905-176849560/socket","Net":"unix"}}
|
| 13 |
+
{"time":"2026-02-25T13:55:11.10141868Z","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
|
| 14 |
+
{"time":"2026-02-25T13:55:11.10145892Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
|
| 15 |
+
{"time":"2026-02-25T13:55:11.101482131Z","level":"INFO","msg":"server is closed"}
|
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug-internal.log
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2026-02-25T13:54:42.16430492Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
|
| 2 |
+
{"time":"2026-02-25T13:54:42.614803801Z","level":"INFO","msg":"stream: created new stream","id":"g2qj8ry7"}
|
| 3 |
+
{"time":"2026-02-25T13:54:42.615183417Z","level":"INFO","msg":"handler: started","stream_id":"g2qj8ry7"}
|
| 4 |
+
{"time":"2026-02-25T13:54:42.615386361Z","level":"INFO","msg":"stream: started","id":"g2qj8ry7"}
|
| 5 |
+
{"time":"2026-02-25T13:54:42.615431072Z","level":"INFO","msg":"sender: started","stream_id":"g2qj8ry7"}
|
| 6 |
+
{"time":"2026-02-25T13:54:42.615463052Z","level":"INFO","msg":"writer: started","stream_id":"g2qj8ry7"}
|
| 7 |
+
{"time":"2026-02-25T13:55:10.309513894Z","level":"INFO","msg":"stream: closing","id":"g2qj8ry7"}
|
| 8 |
+
{"time":"2026-02-25T13:55:10.794171575Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
|
| 9 |
+
{"time":"2026-02-25T13:55:11.099789022Z","level":"INFO","msg":"handler: closed","stream_id":"g2qj8ry7"}
|
| 10 |
+
{"time":"2026-02-25T13:55:11.099961565Z","level":"INFO","msg":"sender: closed","stream_id":"g2qj8ry7"}
|
| 11 |
+
{"time":"2026-02-25T13:55:11.099983596Z","level":"INFO","msg":"stream: closed","id":"g2qj8ry7"}
|
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug.log
ADDED
|
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
|
| 2 |
+
2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_setup.py:_flush():81] Configure stats pid to 168367
|
| 3 |
+
2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_setup.py:_flush():81] Loading settings from environment variables
|
| 4 |
+
2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug.log
|
| 5 |
+
2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/logs/debug-internal.log
|
| 6 |
+
2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:init():844] calling init triggers
|
| 7 |
+
2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
|
| 8 |
+
config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518], 'head_mode': 'depth', 'num_level': 3, 'gs_param_dim': 256, 'align_corners': False, 'use_voxelize': True}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False}, 'density_control': {'name': 'density_control_module', 'mean_dim': 32, 'gs_param_dim': 256, 'refinement_layer_num': 1, 'num_level': 3, 'grad_mode': 'absgrad', 'use_mean_features': True, 'refinement_type': 'voxelize', 'refinement_hidden_dim': 32, 'aggregation_mode': 'mean', 'num_heads': 1, 'score_mode': 'absgrad', 'latent_dim': 128, 'num_latents': 64, 'num_self_attn_per_block': 2, 'voxel_size': 0.001, 'aux_refine': False, 'refine_error': False, 'use_refine_module': True, 'voxelize_activate': True, 'use_depth': True}}, 'render_loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'density_control_loss': {'error_score': {'weight': 0.01, 'log_scale': False, 'grad_scale': 10000.0, 'mode': 'original'}}, 'direct_loss': {'l1': {'weight': 0.8}, 'ssim': {'weight': 0.2}}, 'wandb': {'project': 'DCSplat', 'entity': 'scene-representation-group', 'name': 'ABLATION_0225_ctxTrain_depth_vggtDistl', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 25, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'T+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 1500, 'save_top_k': 2, 'save_weights_only': False}, 'train': {'extended_visualization': False, 'print_log_every_n_steps': 10, 'camera_loss': 10.0, 'one_sample_validation': None, 'align_corners': False, 'intrinsic_scaling': False, 'verbose': False, 'beta_dist_param': [0.5, 4.0], 'use_refine_aux': False, 'train_target_set': True, 'train_gs_num': 1, 'ext_scale_detach': False, 'cam_scale_mode': 'sum', 'scene_scale_reg_loss': 0.01, 'train_aux': True, 'vggt_cam_loss': True, 'vggt_distil': True, 'context_view_train': True}, 'test': {'output_path': 'test/ablation/re10k', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_active_mask_image': False, 'save_error_score_image': False, 'save_compare': False, 'pred_intrinsic': False, 'error_threshold': 0.4, 'error_threshold_list': [0.2, 0.4, 0.6, 0.8, 1.0], 'threshold_mode': 'ratio', 'nvs_view_N_list': [3, 6, 16, 32, 64]}, 'seed': 111123, 'trainer': {'max_steps': 3001, 'val_check_interval': 250, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 1000, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'same_target_gap': False, 'num_target_set': 3}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0, 'dynamic_context_views': True, 'max_context_views_per_gpu': 24}}, '_wandb': {}}
|
| 9 |
+
2026-02-25 13:54:41,864 INFO MainThread:168367 [wandb_init.py:init():892] starting backend
|
| 10 |
+
2026-02-25 13:54:42,153 INFO MainThread:168367 [wandb_init.py:init():895] sending inform_init request
|
| 11 |
+
2026-02-25 13:54:42,160 INFO MainThread:168367 [wandb_init.py:init():903] backend started and connected
|
| 12 |
+
2026-02-25 13:54:42,166 INFO MainThread:168367 [wandb_init.py:init():973] updated telemetry
|
| 13 |
+
2026-02-25 13:54:42,175 INFO MainThread:168367 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
|
| 14 |
+
2026-02-25 13:54:43,768 INFO MainThread:168367 [wandb_init.py:init():1042] starting run threads in backend
|
| 15 |
+
2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_console_start():2524] atexit reg
|
| 16 |
+
2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_redirect():2373] redirect: wrap_raw
|
| 17 |
+
2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_redirect():2442] Wrapping output streams.
|
| 18 |
+
2026-02-25 13:54:43,893 INFO MainThread:168367 [wandb_run.py:_redirect():2465] Redirects installed.
|
| 19 |
+
2026-02-25 13:54:43,896 INFO MainThread:168367 [wandb_init.py:init():1082] run started, returning control to user process
|
| 20 |
+
2026-02-25 13:55:10,309 INFO wandb-AsyncioManager-main:168367 [service_client.py:_forward_responses():134] Reached EOF.
|
| 21 |
+
2026-02-25 13:55:10,309 INFO wandb-AsyncioManager-main:168367 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.
|
ABLATION_0225_ctxTrain_depth_vggtDistl/wandb/run-20260225_135441-g2qj8ry7/run-g2qj8ry7.wandb
ADDED
|
Binary file (32.9 kB). View file
|
|
|
ABLATION_0225_randomSelect/main.log
CHANGED
|
@@ -114,3 +114,15 @@ bucket_view.sizes() = [256, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered in
|
|
| 114 |
[2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 115 |
result[selector] = overlay
|
| 116 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 114 |
[2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 115 |
result[selector] = overlay
|
| 116 |
|
| 117 |
+
[2026-02-25 13:35:42,292][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 118 |
+
result[selector] = overlay
|
| 119 |
+
|
| 120 |
+
[2026-02-25 13:38:57,687][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 121 |
+
result[selector] = overlay
|
| 122 |
+
|
| 123 |
+
[2026-02-25 13:51:58,139][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 124 |
+
result[selector] = overlay
|
| 125 |
+
|
| 126 |
+
[2026-02-25 13:52:01,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 127 |
+
result[selector] = overlay
|
| 128 |
+
|
ABLATION_0225_randomSelect/peak_vram_memory.json
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"peak_memory_allocated_gb": 103.167,
|
| 3 |
+
"peak_memory_reserved_gb": 137.896,
|
| 4 |
+
"total_elapsed_hours": 3.21,
|
| 5 |
+
"mode": "train"
|
| 6 |
+
}
|
ABLATION_0225_randomSelect/train_ddp_process_1.log
CHANGED
|
@@ -58,3 +58,9 @@ bucket_view.sizes() = [256, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered in
|
|
| 58 |
[2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 59 |
result[selector] = overlay
|
| 60 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 58 |
[2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 59 |
result[selector] = overlay
|
| 60 |
|
| 61 |
+
[2026-02-25 13:38:57,685][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 62 |
+
result[selector] = overlay
|
| 63 |
+
|
| 64 |
+
[2026-02-25 13:52:01,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 65 |
+
result[selector] = overlay
|
| 66 |
+
|
ABLATION_0225_randomSelect/train_ddp_process_2.log
CHANGED
|
@@ -58,3 +58,9 @@ bucket_view.sizes() = [256, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered in
|
|
| 58 |
[2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 59 |
result[selector] = overlay
|
| 60 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 58 |
[2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 59 |
result[selector] = overlay
|
| 60 |
|
| 61 |
+
[2026-02-25 13:38:57,685][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 62 |
+
result[selector] = overlay
|
| 63 |
+
|
| 64 |
+
[2026-02-25 13:52:01,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 65 |
+
result[selector] = overlay
|
| 66 |
+
|
ABLATION_0225_randomSelect/train_ddp_process_3.log
CHANGED
|
@@ -58,3 +58,9 @@ bucket_view.sizes() = [256, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered in
|
|
| 58 |
[2026-02-25 13:26:12,526][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 59 |
result[selector] = overlay
|
| 60 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 58 |
[2026-02-25 13:26:12,526][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 59 |
result[selector] = overlay
|
| 60 |
|
| 61 |
+
[2026-02-25 13:38:57,685][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 62 |
+
result[selector] = overlay
|
| 63 |
+
|
| 64 |
+
[2026-02-25 13:52:01,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 65 |
+
result[selector] = overlay
|
| 66 |
+
|
ABLATION_0225_randomSelect/train_ddp_process_4.log
CHANGED
|
@@ -58,3 +58,9 @@ bucket_view.sizes() = [256, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered in
|
|
| 58 |
[2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 59 |
result[selector] = overlay
|
| 60 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 58 |
[2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 59 |
result[selector] = overlay
|
| 60 |
|
| 61 |
+
[2026-02-25 13:38:57,685][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 62 |
+
result[selector] = overlay
|
| 63 |
+
|
| 64 |
+
[2026-02-25 13:52:01,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 65 |
+
result[selector] = overlay
|
| 66 |
+
|
ABLATION_0225_randomSelect/train_ddp_process_5.log
CHANGED
|
@@ -58,3 +58,9 @@ bucket_view.sizes() = [256, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered in
|
|
| 58 |
[2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 59 |
result[selector] = overlay
|
| 60 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 58 |
[2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 59 |
result[selector] = overlay
|
| 60 |
|
| 61 |
+
[2026-02-25 13:38:57,685][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 62 |
+
result[selector] = overlay
|
| 63 |
+
|
| 64 |
+
[2026-02-25 13:52:01,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 65 |
+
result[selector] = overlay
|
| 66 |
+
|
ABLATION_0225_randomSelect/train_ddp_process_6.log
CHANGED
|
@@ -58,3 +58,9 @@ bucket_view.sizes() = [256, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered in
|
|
| 58 |
[2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 59 |
result[selector] = overlay
|
| 60 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 58 |
[2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 59 |
result[selector] = overlay
|
| 60 |
|
| 61 |
+
[2026-02-25 13:38:57,685][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 62 |
+
result[selector] = overlay
|
| 63 |
+
|
| 64 |
+
[2026-02-25 13:52:01,499][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 65 |
+
result[selector] = overlay
|
| 66 |
+
|
ABLATION_0225_randomSelect/train_ddp_process_7.log
CHANGED
|
@@ -58,3 +58,9 @@ bucket_view.sizes() = [256, 256, 1, 1], strides() = [256, 1, 1, 1] (Triggered in
|
|
| 58 |
[2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 59 |
result[selector] = overlay
|
| 60 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 58 |
[2026-02-25 13:26:12,524][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 59 |
result[selector] = overlay
|
| 60 |
|
| 61 |
+
[2026-02-25 13:38:57,685][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 62 |
+
result[selector] = overlay
|
| 63 |
+
|
| 64 |
+
[2026-02-25 13:52:01,501][py.warnings][WARNING] - /workspace/code/CVPR2026/src/visualization/layout.py:105: UserWarning: Using a non-tuple sequence for multidimensional indexing is deprecated and will be changed in pytorch 2.9; use x[tuple(seq)] instead of x[seq]. In pytorch 2.9 this will be interpreted as tensor index, x[torch.tensor(seq)], which will result either in an error or a different result (Triggered internally at /pytorch/torch/csrc/autograd/python_variable_indexing.cpp:316.)
|
| 65 |
+
result[selector] = overlay
|
| 66 |
+
|
ABLATION_0225_randomSelect/wandb/debug-internal.log
CHANGED
|
@@ -4,3 +4,8 @@
|
|
| 4 |
{"time":"2026-02-25T10:39:56.464733133Z","level":"INFO","msg":"stream: started","id":"n76tgbfj"}
|
| 5 |
{"time":"2026-02-25T10:39:56.464769993Z","level":"INFO","msg":"sender: started","stream_id":"n76tgbfj"}
|
| 6 |
{"time":"2026-02-25T10:39:56.464783063Z","level":"INFO","msg":"writer: started","stream_id":"n76tgbfj"}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
{"time":"2026-02-25T10:39:56.464733133Z","level":"INFO","msg":"stream: started","id":"n76tgbfj"}
|
| 5 |
{"time":"2026-02-25T10:39:56.464769993Z","level":"INFO","msg":"sender: started","stream_id":"n76tgbfj"}
|
| 6 |
{"time":"2026-02-25T10:39:56.464783063Z","level":"INFO","msg":"writer: started","stream_id":"n76tgbfj"}
|
| 7 |
+
{"time":"2026-02-25T13:52:10.426476781Z","level":"INFO","msg":"stream: closing","id":"n76tgbfj"}
|
| 8 |
+
{"time":"2026-02-25T13:52:11.22551075Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
|
| 9 |
+
{"time":"2026-02-25T13:52:11.503165049Z","level":"INFO","msg":"handler: closed","stream_id":"n76tgbfj"}
|
| 10 |
+
{"time":"2026-02-25T13:52:11.503344582Z","level":"INFO","msg":"sender: closed","stream_id":"n76tgbfj"}
|
| 11 |
+
{"time":"2026-02-25T13:52:11.503360133Z","level":"INFO","msg":"stream: closed","id":"n76tgbfj"}
|
ABLATION_0225_randomSelect/wandb/debug.log
CHANGED
|
@@ -17,3 +17,5 @@ config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518
|
|
| 17 |
2026-02-25 10:39:57,797 INFO MainThread:145562 [wandb_run.py:_redirect():2442] Wrapping output streams.
|
| 18 |
2026-02-25 10:39:57,797 INFO MainThread:145562 [wandb_run.py:_redirect():2465] Redirects installed.
|
| 19 |
2026-02-25 10:39:57,800 INFO MainThread:145562 [wandb_init.py:init():1082] run started, returning control to user process
|
|
|
|
|
|
|
|
|
| 17 |
2026-02-25 10:39:57,797 INFO MainThread:145562 [wandb_run.py:_redirect():2442] Wrapping output streams.
|
| 18 |
2026-02-25 10:39:57,797 INFO MainThread:145562 [wandb_run.py:_redirect():2465] Redirects installed.
|
| 19 |
2026-02-25 10:39:57,800 INFO MainThread:145562 [wandb_init.py:init():1082] run started, returning control to user process
|
| 20 |
+
2026-02-25 13:52:10,426 INFO wandb-AsyncioManager-main:145562 [service_client.py:_forward_responses():134] Reached EOF.
|
| 21 |
+
2026-02-25 13:52:10,426 INFO wandb-AsyncioManager-main:145562 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.
|
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/config.yaml
ADDED
|
@@ -0,0 +1,307 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
_wandb:
|
| 2 |
+
value:
|
| 3 |
+
cli_version: 0.25.0
|
| 4 |
+
e:
|
| 5 |
+
81z9i9czdg5aqgfjdth6b57yriddztva:
|
| 6 |
+
args:
|
| 7 |
+
- +experiment=re10k_ablation_24v
|
| 8 |
+
- wandb.mode=online
|
| 9 |
+
- wandb.name=ABLATION_0225_randomSelect
|
| 10 |
+
- model.density_control.score_mode=random
|
| 11 |
+
cpu_count: 128
|
| 12 |
+
cpu_count_logical: 256
|
| 13 |
+
cudaVersion: "13.1"
|
| 14 |
+
disk:
|
| 15 |
+
/:
|
| 16 |
+
total: "1170378588160"
|
| 17 |
+
used: "732546121728"
|
| 18 |
+
email: dna9041@korea.ac.kr
|
| 19 |
+
executable: /venv/main/bin/python
|
| 20 |
+
git:
|
| 21 |
+
commit: 2512754c6c27ca5150bf17fbcbdde3f192fd53cc
|
| 22 |
+
remote: git@github.com:K-nowing/CVPR2026.git
|
| 23 |
+
gpu: NVIDIA H200
|
| 24 |
+
gpu_count: 8
|
| 25 |
+
gpu_nvidia:
|
| 26 |
+
- architecture: Hopper
|
| 27 |
+
cudaCores: 16896
|
| 28 |
+
memoryTotal: "150754820096"
|
| 29 |
+
name: NVIDIA H200
|
| 30 |
+
uuid: GPU-2649ab80-a3a6-5a1c-0fa5-12bc11bd75e9
|
| 31 |
+
- architecture: Hopper
|
| 32 |
+
cudaCores: 16896
|
| 33 |
+
memoryTotal: "150754820096"
|
| 34 |
+
name: NVIDIA H200
|
| 35 |
+
uuid: GPU-e92921d9-c681-246f-af93-637e0dc938ca
|
| 36 |
+
- architecture: Hopper
|
| 37 |
+
cudaCores: 16896
|
| 38 |
+
memoryTotal: "150754820096"
|
| 39 |
+
name: NVIDIA H200
|
| 40 |
+
uuid: GPU-ffe12ffc-9bb7-82de-5692-1ec0ee2e68d8
|
| 41 |
+
- architecture: Hopper
|
| 42 |
+
cudaCores: 16896
|
| 43 |
+
memoryTotal: "150754820096"
|
| 44 |
+
name: NVIDIA H200
|
| 45 |
+
uuid: GPU-499e5acd-b6ab-2010-c51b-ee9b5aa65825
|
| 46 |
+
- architecture: Hopper
|
| 47 |
+
cudaCores: 16896
|
| 48 |
+
memoryTotal: "150754820096"
|
| 49 |
+
name: NVIDIA H200
|
| 50 |
+
uuid: GPU-3b2522d9-1c72-e49b-2c30-96165680b74a
|
| 51 |
+
- architecture: Hopper
|
| 52 |
+
cudaCores: 16896
|
| 53 |
+
memoryTotal: "150754820096"
|
| 54 |
+
name: NVIDIA H200
|
| 55 |
+
uuid: GPU-a9a280c5-b2f9-dc1e-a8a9-7326a74001ff
|
| 56 |
+
- architecture: Hopper
|
| 57 |
+
cudaCores: 16896
|
| 58 |
+
memoryTotal: "150754820096"
|
| 59 |
+
name: NVIDIA H200
|
| 60 |
+
uuid: GPU-07d0167b-a6a1-1900-2d27-7c6c11598409
|
| 61 |
+
- architecture: Hopper
|
| 62 |
+
cudaCores: 16896
|
| 63 |
+
memoryTotal: "150754820096"
|
| 64 |
+
name: NVIDIA H200
|
| 65 |
+
uuid: GPU-8362a999-20d1-c27b-5d18-032d23f859ab
|
| 66 |
+
host: 27d18dedec6d
|
| 67 |
+
memory:
|
| 68 |
+
total: "1622948257792"
|
| 69 |
+
os: Linux-6.8.0-90-generic-x86_64-with-glibc2.39
|
| 70 |
+
program: -m src.main
|
| 71 |
+
python: CPython 3.12.12
|
| 72 |
+
root: /workspace/code/CVPR2026/outputs/ablation/re10k/ABLATION_0225_randomSelect
|
| 73 |
+
startedAt: "2026-02-25T10:39:55.441642Z"
|
| 74 |
+
writerId: 81z9i9czdg5aqgfjdth6b57yriddztva
|
| 75 |
+
m:
|
| 76 |
+
- "1": trainer/global_step
|
| 77 |
+
"6":
|
| 78 |
+
- 3
|
| 79 |
+
"7": []
|
| 80 |
+
- "2": '*'
|
| 81 |
+
"5": 1
|
| 82 |
+
"6":
|
| 83 |
+
- 1
|
| 84 |
+
"7": []
|
| 85 |
+
python_version: 3.12.12
|
| 86 |
+
t:
|
| 87 |
+
"1":
|
| 88 |
+
- 1
|
| 89 |
+
- 41
|
| 90 |
+
- 49
|
| 91 |
+
- 50
|
| 92 |
+
- 106
|
| 93 |
+
"2":
|
| 94 |
+
- 1
|
| 95 |
+
- 41
|
| 96 |
+
- 49
|
| 97 |
+
- 50
|
| 98 |
+
- 106
|
| 99 |
+
"3":
|
| 100 |
+
- 7
|
| 101 |
+
- 13
|
| 102 |
+
- 15
|
| 103 |
+
- 16
|
| 104 |
+
- 66
|
| 105 |
+
"4": 3.12.12
|
| 106 |
+
"5": 0.25.0
|
| 107 |
+
"12": 0.25.0
|
| 108 |
+
"13": linux-x86_64
|
| 109 |
+
checkpointing:
|
| 110 |
+
value:
|
| 111 |
+
every_n_train_steps: 1500
|
| 112 |
+
load: null
|
| 113 |
+
save_top_k: 2
|
| 114 |
+
save_weights_only: false
|
| 115 |
+
data_loader:
|
| 116 |
+
value:
|
| 117 |
+
test:
|
| 118 |
+
batch_size: 1
|
| 119 |
+
num_workers: 4
|
| 120 |
+
persistent_workers: false
|
| 121 |
+
seed: 2345
|
| 122 |
+
train:
|
| 123 |
+
batch_size: 16
|
| 124 |
+
num_workers: 16
|
| 125 |
+
persistent_workers: true
|
| 126 |
+
seed: 1234
|
| 127 |
+
val:
|
| 128 |
+
batch_size: 1
|
| 129 |
+
num_workers: 1
|
| 130 |
+
persistent_workers: true
|
| 131 |
+
seed: 3456
|
| 132 |
+
dataset:
|
| 133 |
+
value:
|
| 134 |
+
re10k:
|
| 135 |
+
augment: true
|
| 136 |
+
background_color:
|
| 137 |
+
- 0
|
| 138 |
+
- 0
|
| 139 |
+
- 0
|
| 140 |
+
baseline_max: 1e+10
|
| 141 |
+
baseline_min: 0.001
|
| 142 |
+
cameras_are_circular: false
|
| 143 |
+
dynamic_context_views: true
|
| 144 |
+
input_image_shape:
|
| 145 |
+
- 256
|
| 146 |
+
- 256
|
| 147 |
+
make_baseline_1: true
|
| 148 |
+
max_context_views_per_gpu: 24
|
| 149 |
+
max_fov: 100
|
| 150 |
+
name: re10k
|
| 151 |
+
original_image_shape:
|
| 152 |
+
- 360
|
| 153 |
+
- 640
|
| 154 |
+
overfit_to_scene: null
|
| 155 |
+
relative_pose: true
|
| 156 |
+
roots:
|
| 157 |
+
- datasets/re10k
|
| 158 |
+
skip_bad_shape: true
|
| 159 |
+
view_sampler:
|
| 160 |
+
initial_max_distance_between_context_views: 25
|
| 161 |
+
initial_min_distance_between_context_views: 25
|
| 162 |
+
max_distance_between_context_views: 90
|
| 163 |
+
min_distance_between_context_views: 45
|
| 164 |
+
min_distance_to_context_views: 0
|
| 165 |
+
name: bounded
|
| 166 |
+
num_context_views: 2
|
| 167 |
+
num_target_set: 3
|
| 168 |
+
num_target_views: 4
|
| 169 |
+
same_target_gap: false
|
| 170 |
+
warm_up_steps: 1000
|
| 171 |
+
density_control_loss:
|
| 172 |
+
value:
|
| 173 |
+
error_score:
|
| 174 |
+
grad_scale: 10000
|
| 175 |
+
log_scale: false
|
| 176 |
+
mode: original
|
| 177 |
+
weight: 0.01
|
| 178 |
+
direct_loss:
|
| 179 |
+
value:
|
| 180 |
+
l1:
|
| 181 |
+
weight: 0.8
|
| 182 |
+
ssim:
|
| 183 |
+
weight: 0.2
|
| 184 |
+
mode:
|
| 185 |
+
value: train
|
| 186 |
+
model:
|
| 187 |
+
value:
|
| 188 |
+
decoder:
|
| 189 |
+
background_color:
|
| 190 |
+
- 0
|
| 191 |
+
- 0
|
| 192 |
+
- 0
|
| 193 |
+
make_scale_invariant: false
|
| 194 |
+
name: splatting_cuda
|
| 195 |
+
density_control:
|
| 196 |
+
aggregation_mode: mean
|
| 197 |
+
aux_refine: false
|
| 198 |
+
grad_mode: absgrad
|
| 199 |
+
gs_param_dim: 256
|
| 200 |
+
latent_dim: 128
|
| 201 |
+
mean_dim: 32
|
| 202 |
+
name: density_control_module
|
| 203 |
+
num_heads: 1
|
| 204 |
+
num_latents: 64
|
| 205 |
+
num_level: 3
|
| 206 |
+
num_self_attn_per_block: 2
|
| 207 |
+
refine_error: false
|
| 208 |
+
refinement_hidden_dim: 32
|
| 209 |
+
refinement_layer_num: 1
|
| 210 |
+
refinement_type: voxelize
|
| 211 |
+
score_mode: random
|
| 212 |
+
use_depth: false
|
| 213 |
+
use_mean_features: true
|
| 214 |
+
use_refine_module: true
|
| 215 |
+
voxel_size: 0.001
|
| 216 |
+
voxelize_activate: true
|
| 217 |
+
encoder:
|
| 218 |
+
align_corners: false
|
| 219 |
+
gs_param_dim: 256
|
| 220 |
+
head_mode: pcd
|
| 221 |
+
input_image_shape:
|
| 222 |
+
- 518
|
| 223 |
+
- 518
|
| 224 |
+
name: dcsplat
|
| 225 |
+
num_level: 3
|
| 226 |
+
use_voxelize: true
|
| 227 |
+
optimizer:
|
| 228 |
+
value:
|
| 229 |
+
accumulate: 1
|
| 230 |
+
backbone_lr_multiplier: 0.1
|
| 231 |
+
backbone_trainable: T+H
|
| 232 |
+
lr: 0.0002
|
| 233 |
+
warm_up_steps: 25
|
| 234 |
+
render_loss:
|
| 235 |
+
value:
|
| 236 |
+
lpips:
|
| 237 |
+
apply_after_step: 0
|
| 238 |
+
weight: 0.05
|
| 239 |
+
mse:
|
| 240 |
+
weight: 1
|
| 241 |
+
seed:
|
| 242 |
+
value: 111123
|
| 243 |
+
test:
|
| 244 |
+
value:
|
| 245 |
+
align_pose: false
|
| 246 |
+
compute_scores: true
|
| 247 |
+
error_threshold: 0.4
|
| 248 |
+
error_threshold_list:
|
| 249 |
+
- 0.2
|
| 250 |
+
- 0.4
|
| 251 |
+
- 0.6
|
| 252 |
+
- 0.8
|
| 253 |
+
- 1
|
| 254 |
+
nvs_view_N_list:
|
| 255 |
+
- 3
|
| 256 |
+
- 6
|
| 257 |
+
- 16
|
| 258 |
+
- 32
|
| 259 |
+
- 64
|
| 260 |
+
output_path: test/ablation/re10k
|
| 261 |
+
pose_align_steps: 100
|
| 262 |
+
pred_intrinsic: false
|
| 263 |
+
rot_opt_lr: 0.005
|
| 264 |
+
save_active_mask_image: false
|
| 265 |
+
save_compare: false
|
| 266 |
+
save_error_score_image: false
|
| 267 |
+
save_image: false
|
| 268 |
+
save_video: false
|
| 269 |
+
threshold_mode: ratio
|
| 270 |
+
trans_opt_lr: 0.005
|
| 271 |
+
train:
|
| 272 |
+
value:
|
| 273 |
+
align_corners: false
|
| 274 |
+
beta_dist_param:
|
| 275 |
+
- 0.5
|
| 276 |
+
- 4
|
| 277 |
+
cam_scale_mode: sum
|
| 278 |
+
camera_loss: 10
|
| 279 |
+
context_view_train: false
|
| 280 |
+
ext_scale_detach: false
|
| 281 |
+
extended_visualization: false
|
| 282 |
+
intrinsic_scaling: false
|
| 283 |
+
one_sample_validation: null
|
| 284 |
+
print_log_every_n_steps: 10
|
| 285 |
+
scene_scale_reg_loss: 0.01
|
| 286 |
+
train_aux: true
|
| 287 |
+
train_gs_num: 1
|
| 288 |
+
train_target_set: true
|
| 289 |
+
use_refine_aux: false
|
| 290 |
+
verbose: false
|
| 291 |
+
vggt_cam_loss: true
|
| 292 |
+
vggt_distil: false
|
| 293 |
+
trainer:
|
| 294 |
+
value:
|
| 295 |
+
gradient_clip_val: 0.5
|
| 296 |
+
max_steps: 3001
|
| 297 |
+
num_nodes: 1
|
| 298 |
+
val_check_interval: 250
|
| 299 |
+
wandb:
|
| 300 |
+
value:
|
| 301 |
+
entity: scene-representation-group
|
| 302 |
+
mode: online
|
| 303 |
+
name: ABLATION_0225_randomSelect
|
| 304 |
+
project: DCSplat
|
| 305 |
+
tags:
|
| 306 |
+
- re10k
|
| 307 |
+
- 256x256
|
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/active_mask_imgs_182_2edf4f17c23388f66672.png
ADDED
|
Git LFS Details
|
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/active_mask_imgs_198_399ec9b2f678832e0977.png
ADDED
|
Git LFS Details
|
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/comparison_197_ddc312f116aae122416e.png
ADDED
|
Git LFS Details
|
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/error_scores_183_e47e70d75361cb369a35.png
ADDED
|
Git LFS Details
|
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/error_scores_199_b1c8e80e7cf22aab1c67.png
ADDED
|
Git LFS Details
|
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/train/comparison_188_a456ce8a38a4d0dc65b1.png
ADDED
|
Git LFS Details
|
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/train/comparison_202_b207767914aae1b684b5.png
ADDED
|
Git LFS Details
|
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/train/error_scores_187_347cd0a01cf553ca4f4d.png
ADDED
|
Git LFS Details
|
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/media/images/train/error_scores_201_05456dc706694ddf5d92.png
ADDED
|
Git LFS Details
|
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/files/wandb-summary.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"loss/total":0.08273646980524063,"info/global_step":3000,"loss/aux_2/lpips":0.009566728956997395,"loss/final_3dgs/lpips":0.009549135342240334,"train/error_scores":{"width":1328,"height":2120,"format":"png","count":1,"filenames":["media/images/train/error_scores_201_05456dc706694ddf5d92.png"],"captions":[["0621c7675fab1418"]],"_type":"images/separated"},"_runtime":11532,"lr-AdamW/pg1-momentum":0.9,"lr-AdamW/pg2-momentum":0.9,"train/psnr_probabilistic":20.13189697265625,"loss/aux_1/error_score":0.4180681109428406,"loss/aux_1/lpips":0.009899786673486233,"val/ssim":0.7995258569717407,"loss/scene_scale_reg":0.00020948528253939003,"val/lpips":0.23768818378448486,"comparison":{"filenames":["media/images/comparison_197_ddc312f116aae122416e.png"],"captions":["a76028640ffa1ef9"],"_type":"images/separated","width":1064,"height":1098,"format":"png","count":1},"trainer/global_step":3001,"loss/final_3dgs/mse":0.010063939727842808,"loss/aux_0/error_score":0.7389501333236694,"train/scene_scale":1.003704309463501,"val/psnr":21.38949203491211,"loss/camera":0.0002018376690102741,"active_mask_imgs":{"captions":["a76028640ffa1ef9"],"_type":"images/separated","width":536,"height":800,"format":"png","count":1,"filenames":["media/images/active_mask_imgs_198_399ec9b2f678832e0977.png"]},"_timestamp":1.7720275236574388e+09,"loss/aux_1/mse":0.010359433479607105,"loss/aux_0/mse":0.00995652750134468,"_wandb":{"runtime":11532},"loss/aux_2/mse":0.010155964642763138,"lr-AdamW/pg2":2e-05,"error_scores":{"width":800,"height":536,"format":"png","count":1,"filenames":["media/images/error_scores_199_b1c8e80e7cf22aab1c67.png"],"captions":["a76028640ffa1ef9"],"_type":"images/separated"},"loss/aux_0/lpips":0.010957089252769947,"epoch":0,"val/gaussian_num_ratio":0.39984893798828125,"train/comparison":{"format":"png","count":1,"filenames":["media/images/train/comparison_202_b207767914aae1b684b5.png"],"captions":[["0621c7675fab1418"]],"_type":"images/separated","width":1328,"height":2154},"_step":202,"lr-AdamW/pg1":2.003594834351718e-05}
|
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/logs/debug-core.log
CHANGED
|
@@ -5,3 +5,11 @@
|
|
| 5 |
{"time":"2026-02-25T10:39:55.743235764Z","level":"INFO","msg":"handleInformInit: received","streamId":"n76tgbfj","id":"1(@)"}
|
| 6 |
{"time":"2026-02-25T10:39:56.464752793Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"n76tgbfj","id":"1(@)"}
|
| 7 |
{"time":"2026-02-25T10:40:02.800543333Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"rz66kc5pj9qh"}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5 |
{"time":"2026-02-25T10:39:55.743235764Z","level":"INFO","msg":"handleInformInit: received","streamId":"n76tgbfj","id":"1(@)"}
|
| 6 |
{"time":"2026-02-25T10:39:56.464752793Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"n76tgbfj","id":"1(@)"}
|
| 7 |
{"time":"2026-02-25T10:40:02.800543333Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"rz66kc5pj9qh"}
|
| 8 |
+
{"time":"2026-02-25T13:52:10.426350249Z","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
|
| 9 |
+
{"time":"2026-02-25T13:52:10.426467411Z","level":"INFO","msg":"connection: closing","id":"1(@)"}
|
| 10 |
+
{"time":"2026-02-25T13:52:10.426509722Z","level":"INFO","msg":"server is shutting down"}
|
| 11 |
+
{"time":"2026-02-25T13:52:10.426527162Z","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
|
| 12 |
+
{"time":"2026-02-25T13:52:10.426686075Z","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-145562-147994-3073551312/socket","Net":"unix"}}
|
| 13 |
+
{"time":"2026-02-25T13:52:11.504270148Z","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
|
| 14 |
+
{"time":"2026-02-25T13:52:11.504299818Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
|
| 15 |
+
{"time":"2026-02-25T13:52:11.504311038Z","level":"INFO","msg":"server is closed"}
|
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/logs/debug-internal.log
CHANGED
|
@@ -4,3 +4,8 @@
|
|
| 4 |
{"time":"2026-02-25T10:39:56.464733133Z","level":"INFO","msg":"stream: started","id":"n76tgbfj"}
|
| 5 |
{"time":"2026-02-25T10:39:56.464769993Z","level":"INFO","msg":"sender: started","stream_id":"n76tgbfj"}
|
| 6 |
{"time":"2026-02-25T10:39:56.464783063Z","level":"INFO","msg":"writer: started","stream_id":"n76tgbfj"}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
{"time":"2026-02-25T10:39:56.464733133Z","level":"INFO","msg":"stream: started","id":"n76tgbfj"}
|
| 5 |
{"time":"2026-02-25T10:39:56.464769993Z","level":"INFO","msg":"sender: started","stream_id":"n76tgbfj"}
|
| 6 |
{"time":"2026-02-25T10:39:56.464783063Z","level":"INFO","msg":"writer: started","stream_id":"n76tgbfj"}
|
| 7 |
+
{"time":"2026-02-25T13:52:10.426476781Z","level":"INFO","msg":"stream: closing","id":"n76tgbfj"}
|
| 8 |
+
{"time":"2026-02-25T13:52:11.22551075Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
|
| 9 |
+
{"time":"2026-02-25T13:52:11.503165049Z","level":"INFO","msg":"handler: closed","stream_id":"n76tgbfj"}
|
| 10 |
+
{"time":"2026-02-25T13:52:11.503344582Z","level":"INFO","msg":"sender: closed","stream_id":"n76tgbfj"}
|
| 11 |
+
{"time":"2026-02-25T13:52:11.503360133Z","level":"INFO","msg":"stream: closed","id":"n76tgbfj"}
|
ABLATION_0225_randomSelect/wandb/run-20260225_103955-n76tgbfj/logs/debug.log
CHANGED
|
@@ -17,3 +17,5 @@ config: {'model': {'encoder': {'name': 'dcsplat', 'input_image_shape': [518, 518
|
|
| 17 |
2026-02-25 10:39:57,797 INFO MainThread:145562 [wandb_run.py:_redirect():2442] Wrapping output streams.
|
| 18 |
2026-02-25 10:39:57,797 INFO MainThread:145562 [wandb_run.py:_redirect():2465] Redirects installed.
|
| 19 |
2026-02-25 10:39:57,800 INFO MainThread:145562 [wandb_init.py:init():1082] run started, returning control to user process
|
|
|
|
|
|
|
|
|
| 17 |
2026-02-25 10:39:57,797 INFO MainThread:145562 [wandb_run.py:_redirect():2442] Wrapping output streams.
|
| 18 |
2026-02-25 10:39:57,797 INFO MainThread:145562 [wandb_run.py:_redirect():2465] Redirects installed.
|
| 19 |
2026-02-25 10:39:57,800 INFO MainThread:145562 [wandb_init.py:init():1082] run started, returning control to user process
|
| 20 |
+
2026-02-25 13:52:10,426 INFO wandb-AsyncioManager-main:145562 [service_client.py:_forward_responses():134] Reached EOF.
|
| 21 |
+
2026-02-25 13:52:10,426 INFO wandb-AsyncioManager-main:145562 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.
|