CHYang25 commited on
Commit
7161dcb
·
verified ·
1 Parent(s): 84b8958

Upload folder using huggingface_hub

Browse files
Files changed (23) hide show
  1. .gitattributes +1 -0
  2. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/.hydra/config.yaml +148 -0
  3. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/.hydra/hydra.yaml +155 -0
  4. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/.hydra/overrides.yaml +1 -0
  5. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/checkpoints/epoch=0500-val_loss=0.007.ckpt +3 -0
  6. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/checkpoints/epoch=0700-val_loss=0.007.ckpt +3 -0
  7. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/checkpoints/epoch=0750-val_loss=0.005.ckpt +3 -0
  8. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/checkpoints/epoch=0850-val_loss=0.004.ckpt +3 -0
  9. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/checkpoints/epoch=0950-val_loss=0.003.ckpt +3 -0
  10. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/checkpoints/latest.ckpt +3 -0
  11. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/logs.json.txt +0 -0
  12. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/train.log +8 -0
  13. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/debug-internal.log +16 -0
  14. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/debug.log +27 -0
  15. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/files/config.yaml +229 -0
  16. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/files/output.log +5 -0
  17. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/files/wandb-metadata.json +56 -0
  18. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/files/wandb-summary.json +1 -0
  19. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/logs/debug-core.log +14 -0
  20. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/logs/debug-internal.log +16 -0
  21. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/logs/debug.log +27 -0
  22. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/run-0ietglbs.wandb +3 -0
  23. 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/wandb-resume.json +1 -0
.gitattributes CHANGED
@@ -48,3 +48,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
48
  2026.04.03/15.45.53_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_154603-fqwyxpia/run-fqwyxpia.wandb filter=lfs diff=lfs merge=lfs -text
49
  2026.04.03/09.56.26_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_095636-ih5orjh7/run-ih5orjh7.wandb filter=lfs diff=lfs merge=lfs -text
50
  2026.04.03/12.18.56_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_121906-e4lmo5nr/run-e4lmo5nr.wandb filter=lfs diff=lfs merge=lfs -text
 
 
48
  2026.04.03/15.45.53_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_154603-fqwyxpia/run-fqwyxpia.wandb filter=lfs diff=lfs merge=lfs -text
49
  2026.04.03/09.56.26_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_095636-ih5orjh7/run-ih5orjh7.wandb filter=lfs diff=lfs merge=lfs -text
50
  2026.04.03/12.18.56_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_121906-e4lmo5nr/run-e4lmo5nr.wandb filter=lfs diff=lfs merge=lfs -text
51
+ 2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/run-0ietglbs.wandb filter=lfs diff=lfs merge=lfs -text
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/.hydra/config.yaml ADDED
@@ -0,0 +1,148 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: train_diffusion_unet_lowdim
2
+ _target_: llmbc.workspace.train_diffusion_unet_lowdim_workspace.TrainDiffusionUnetLowdimWorkspace
3
+ obs_dim: ${task.obs_dim}
4
+ action_dim: ${task.action_dim}
5
+ task_name: ${task.name}
6
+ exp_name: default
7
+ horizon: 1
8
+ n_obs_steps: 1
9
+ n_action_steps: 1
10
+ n_latency_steps: 0
11
+ past_action_visible: false
12
+ keypoint_visible_rate: 1.0
13
+ obs_as_local_cond: false
14
+ obs_as_global_cond: true
15
+ pred_action_steps_only: false
16
+ policy:
17
+ _target_: llmbc.policy.diffusion_unet_lowdim_policy.DiffusionUnetLowdimPolicy
18
+ model:
19
+ _target_: llmbc.model.diffusion.conditional_unet1d.ConditionalUnet1D
20
+ input_dim: '${eval: ${task.action_dim} if ${obs_as_local_cond} or ${obs_as_global_cond}
21
+ else ${task.obs_dim} + ${task.action_dim}}'
22
+ local_cond_dim: '${eval: ${task.obs_dim} if ${obs_as_local_cond} else None}'
23
+ global_cond_dim: '${eval: ${task.obs_dim}*${n_obs_steps} if ${obs_as_global_cond}
24
+ else None}'
25
+ diffusion_step_embed_dim: 256
26
+ down_dims:
27
+ - 256
28
+ - 512
29
+ - 1024
30
+ kernel_size: 5
31
+ n_groups: 8
32
+ cond_predict_scale: true
33
+ noise_scheduler:
34
+ _target_: diffusers.schedulers.scheduling_ddpm.DDPMScheduler
35
+ num_train_timesteps: 100
36
+ beta_start: 0.0001
37
+ beta_end: 0.02
38
+ beta_schedule: squaredcos_cap_v2
39
+ variance_type: fixed_small
40
+ clip_sample: true
41
+ prediction_type: epsilon
42
+ horizon: ${horizon}
43
+ obs_dim: ${obs_dim}
44
+ action_dim: ${action_dim}
45
+ n_action_steps: ${eval:'${n_action_steps}+${n_latency_steps}'}
46
+ n_obs_steps: ${n_obs_steps}
47
+ num_inference_steps: 100
48
+ obs_as_local_cond: ${obs_as_local_cond}
49
+ obs_as_global_cond: ${obs_as_global_cond}
50
+ pred_action_steps_only: ${pred_action_steps_only}
51
+ oa_step_convention: true
52
+ ema:
53
+ _target_: llmbc.model.diffusion.ema_model.EMAModel
54
+ update_after_step: 0
55
+ inv_gamma: 1.0
56
+ power: 0.75
57
+ min_value: 0.0
58
+ max_value: 0.9999
59
+ dataloader:
60
+ batch_size: 256
61
+ num_workers: 0
62
+ shuffle: true
63
+ pin_memory: false
64
+ persistent_workers: false
65
+ val_dataloader:
66
+ batch_size: 256
67
+ num_workers: 0
68
+ shuffle: false
69
+ pin_memory: false
70
+ persistent_workers: false
71
+ optimizer:
72
+ _target_: torch.optim.AdamW
73
+ lr: 0.0001
74
+ betas:
75
+ - 0.95
76
+ - 0.999
77
+ eps: 1.0e-08
78
+ weight_decay: 1.0e-06
79
+ training:
80
+ device: cuda:0
81
+ seed: 45
82
+ debug: false
83
+ resume: true
84
+ lr_scheduler: cosine
85
+ lr_warmup_steps: 500
86
+ num_epochs: 1001
87
+ gradient_accumulate_every: 1
88
+ use_ema: true
89
+ rollout_every: 1000
90
+ checkpoint_every: 50
91
+ val_every: 1
92
+ sample_every: 1000
93
+ max_train_steps: null
94
+ max_val_steps: null
95
+ tqdm_interval_sec: 1.0
96
+ logging:
97
+ project: ${task.name}-training
98
+ resume: true
99
+ mode: online
100
+ name: ${now:%Y.%m.%d-%H.%M.%S}_${name}_${task_name}
101
+ tags:
102
+ - ${name}
103
+ - ${task_name}
104
+ - ${exp_name}
105
+ id: null
106
+ group: null
107
+ checkpoint:
108
+ topk:
109
+ monitor_key: val_loss
110
+ mode: min
111
+ k: 5
112
+ format_str: epoch={epoch:04d}-val_loss={val_loss:.3f}.ckpt
113
+ save_last_ckpt: true
114
+ save_last_snapshot: false
115
+ multi_run:
116
+ run_dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
117
+ wandb_name_base: ${now:%Y.%m.%d-%H.%M.%S}_${name}_${task_name}
118
+ task:
119
+ name: PegInsertionSide-v1
120
+ obs_dim: 43
121
+ action_dim: 8
122
+ env_runner:
123
+ _target_: llmbc.env_runner.maniskill_lowdim_runner.ManiskillLowdimRunner
124
+ env_name: llf-maniskill-PegInsertionSide-v1
125
+ instruction_type: b
126
+ feedback_type:
127
+ - hp
128
+ - hn
129
+ visual: false
130
+ n_train: 10
131
+ n_test: 50
132
+ n_envs: 10
133
+ max_steps: 100
134
+ n_obs_steps: ${n_obs_steps}
135
+ n_action_steps: ${n_action_steps}
136
+ discount: 0.99
137
+ dataset:
138
+ _target_: llmbc.dataset.maniskill_lowdim_dataset.ManiskillLowdimDataset
139
+ data_path: datasets/PegInsertionSide-v1-reward-trans.pt
140
+ data_path2: datasets/PegInsertionSide-v1-reward-trans.pt
141
+ horizon: ${horizon}
142
+ pad_before: ${eval:'${n_obs_steps}-1'}
143
+ pad_after: ${eval:'${n_action_steps}-1'}
144
+ use_manual_normalizer: false
145
+ val_ratio: 0.02
146
+ dummy_normalizer: false
147
+ instructor:
148
+ _target_: llmbc.translator.instructor.maniskill_instructor.PegInsertionSide_v1_instructor.PegInsertionSideV1Instructor
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/.hydra/hydra.yaml ADDED
@@ -0,0 +1,155 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
4
+ sweep:
5
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - training.seed=45
116
+ job:
117
+ name: train
118
+ chdir: null
119
+ override_dirname: training.seed=45
120
+ id: ???
121
+ num: ???
122
+ config_name: bdpo_behavior_PegInsertionSide-v1_exp.yaml
123
+ env_set: {}
124
+ env_copy: []
125
+ config:
126
+ override_dirname:
127
+ kv_sep: '='
128
+ item_sep: ','
129
+ exclude_keys: []
130
+ runtime:
131
+ version: 1.2.0
132
+ version_base: '1.2'
133
+ cwd: /tmp2/chyang/workspace/LLM-BC
134
+ config_sources:
135
+ - path: hydra.conf
136
+ schema: pkg
137
+ provider: hydra
138
+ - path: /tmp2/chyang/workspace/LLM-BC/config/main_table
139
+ schema: file
140
+ provider: main
141
+ - path: ''
142
+ schema: structured
143
+ provider: schema
144
+ output_dir: /tmp2/chyang/workspace/LLM-BC/data/outputs/2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1
145
+ choices:
146
+ hydra/env: default
147
+ hydra/callbacks: null
148
+ hydra/job_logging: default
149
+ hydra/hydra_logging: default
150
+ hydra/hydra_help: default
151
+ hydra/help: default
152
+ hydra/sweeper: basic
153
+ hydra/launcher: basic
154
+ hydra/output: default
155
+ verbose: false
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/.hydra/overrides.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ - training.seed=45
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/checkpoints/epoch=0500-val_loss=0.007.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f198d8b62c4e7dce1338d45366fc02ffb0725d585c75e5e426224ee0601f06a
3
+ size 1048405554
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/checkpoints/epoch=0700-val_loss=0.007.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20069e31ef588b6b3e2f71944fb1e734f8a7415bf0e4350b15ca13a945ccbd38
3
+ size 1048405554
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/checkpoints/epoch=0750-val_loss=0.005.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4aeef42f1e731826fe703ac60f61b0370031220b48af4c72e10bb7183eebd09b
3
+ size 1048405554
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/checkpoints/epoch=0850-val_loss=0.004.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33a4efb4e986c2c70f54f62f12b39f8c7244c5ba484862a9069ac6a6d35514bb
3
+ size 1048405554
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/checkpoints/epoch=0950-val_loss=0.003.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5be98cd219945966bcd78942dcb4ea04a3760c436381cfb48e75bd3dabdab826
3
+ size 1048405554
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/checkpoints/latest.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:008c1ddd4abf22335752598eb631ce231b7af640d52cece3f46589ccd02ae35c
3
+ size 1048405554
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/logs.json.txt ADDED
The diff for this file is too large to render. See raw diff
 
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/train.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ [2026-04-03 13:29:03,228][numexpr.utils][INFO] - Note: NumExpr detected 24 cores but "NUMEXPR_MAX_THREADS" not set, so enforcing safe limit of 8.
2
+ [2026-04-03 13:29:03,228][numexpr.utils][INFO] - NumExpr defaulting to 8 threads.
3
+ [2026-04-03 13:29:05,810][llmbc.model.diffusion.conditional_unet1d][INFO] - number of parameters: 6.550938e+07
4
+ [2026-04-03 13:29:10,161][absl][INFO] - MUJOCO_GL=osmesa, attempting to import specified OpenGL backend.
5
+ [2026-04-03 13:29:10,164][absl][INFO] - MuJoCo library version is: 2.3.7
6
+ [2026-04-03 13:29:10,232][datasets][INFO] - PyTorch version 2.2.2 available.
7
+ [2026-04-03 13:29:10,232][datasets][INFO] - TensorFlow version 2.15.1 available.
8
+ [2026-04-03 13:29:10,233][datasets][INFO] - JAX version 0.4.30 available.
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/debug-internal.log ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-04-03T13:29:11.183039659+08:00","level":"INFO","msg":"using version","core version":"0.18.6"}
2
+ {"time":"2026-04-03T13:29:11.18304756+08:00","level":"INFO","msg":"created symlink","path":"/tmp2/chyang/workspace/LLM-BC/data/outputs/2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/logs/debug-core.log"}
3
+ {"time":"2026-04-03T13:29:11.290645001+08:00","level":"INFO","msg":"created new stream","id":"0ietglbs"}
4
+ {"time":"2026-04-03T13:29:11.290709294+08:00","level":"INFO","msg":"stream: started","id":"0ietglbs"}
5
+ {"time":"2026-04-03T13:29:11.290717333+08:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"0ietglbs"}}
6
+ {"time":"2026-04-03T13:29:11.290758229+08:00","level":"INFO","msg":"sender: started","stream_id":"0ietglbs"}
7
+ {"time":"2026-04-03T13:29:11.290756734+08:00","level":"INFO","msg":"handler: started","stream_id":{"value":"0ietglbs"}}
8
+ {"time":"2026-04-03T13:29:11.922864379+08:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2026-04-03T14:38:23.941643874+08:00","level":"INFO","msg":"stream: closing","id":"0ietglbs"}
10
+ {"time":"2026-04-03T14:38:23.941726051+08:00","level":"INFO","msg":"Stopping system monitor"}
11
+ {"time":"2026-04-03T14:38:23.942686523+08:00","level":"INFO","msg":"Stopped system monitor"}
12
+ {"time":"2026-04-03T14:38:25.088055309+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
13
+ {"time":"2026-04-03T14:38:25.520228302+08:00","level":"INFO","msg":"handler: closed","stream_id":{"value":"0ietglbs"}}
14
+ {"time":"2026-04-03T14:38:25.520305908+08:00","level":"INFO","msg":"sender: closed","stream_id":"0ietglbs"}
15
+ {"time":"2026-04-03T14:38:25.520298235+08:00","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"0ietglbs"}}
16
+ {"time":"2026-04-03T14:38:25.520468398+08:00","level":"INFO","msg":"stream: closed","id":"0ietglbs"}
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/debug.log ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-04-03 13:29:11,178 INFO MainThread:3260278 [wandb_setup.py:_flush():79] Current SDK version is 0.18.6
2
+ 2026-04-03 13:29:11,178 INFO MainThread:3260278 [wandb_setup.py:_flush():79] Configure stats pid to 3260278
3
+ 2026-04-03 13:29:11,178 INFO MainThread:3260278 [wandb_setup.py:_flush():79] Loading settings from /home/chyang/.config/wandb/settings
4
+ 2026-04-03 13:29:11,178 INFO MainThread:3260278 [wandb_setup.py:_flush():79] Loading settings from /tmp2/chyang/workspace/LLM-BC/wandb/settings
5
+ 2026-04-03 13:29:11,178 INFO MainThread:3260278 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': 'online', '_disable_service': None}
7
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train.py', 'program_abspath': '/tmp2/chyang/workspace/LLM-BC/train.py', 'program': '/tmp2/chyang/workspace/LLM-BC/./train.py'}
8
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_init.py:_log_setup():533] Logging user logs to /tmp2/chyang/workspace/LLM-BC/data/outputs/2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/logs/debug.log
10
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_init.py:_log_setup():534] Logging internal logs to /tmp2/chyang/workspace/LLM-BC/data/outputs/2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/logs/debug-internal.log
11
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_init.py:init():619] calling init triggers
12
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_init.py:init():626] wandb.init called with sweep_config: {}
13
+ config: {'name': 'train_diffusion_unet_lowdim', '_target_': 'llmbc.workspace.train_diffusion_unet_lowdim_workspace.TrainDiffusionUnetLowdimWorkspace', 'obs_dim': 43, 'action_dim': 8, 'task_name': 'PegInsertionSide-v1', 'exp_name': 'default', 'horizon': 1, 'n_obs_steps': 1, 'n_action_steps': 1, 'n_latency_steps': 0, 'past_action_visible': False, 'keypoint_visible_rate': 1.0, 'obs_as_local_cond': False, 'obs_as_global_cond': True, 'pred_action_steps_only': False, 'policy': {'_target_': 'llmbc.policy.diffusion_unet_lowdim_policy.DiffusionUnetLowdimPolicy', 'model': {'_target_': 'llmbc.model.diffusion.conditional_unet1d.ConditionalUnet1D', 'input_dim': 8, 'local_cond_dim': None, 'global_cond_dim': 43, 'diffusion_step_embed_dim': 256, 'down_dims': [256, 512, 1024], 'kernel_size': 5, 'n_groups': 8, 'cond_predict_scale': True}, 'noise_scheduler': {'_target_': 'diffusers.schedulers.scheduling_ddpm.DDPMScheduler', 'num_train_timesteps': 100, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'squaredcos_cap_v2', 'variance_type': 'fixed_small', 'clip_sample': True, 'prediction_type': 'epsilon'}, 'horizon': 1, 'obs_dim': 43, 'action_dim': 8, 'n_action_steps': 1, 'n_obs_steps': 1, 'num_inference_steps': 100, 'obs_as_local_cond': False, 'obs_as_global_cond': True, 'pred_action_steps_only': False, 'oa_step_convention': True}, 'ema': {'_target_': 'llmbc.model.diffusion.ema_model.EMAModel', 'update_after_step': 0, 'inv_gamma': 1.0, 'power': 0.75, 'min_value': 0.0, 'max_value': 0.9999}, 'dataloader': {'batch_size': 256, 'num_workers': 0, 'shuffle': True, 'pin_memory': False, 'persistent_workers': False}, 'val_dataloader': {'batch_size': 256, 'num_workers': 0, 'shuffle': False, 'pin_memory': False, 'persistent_workers': False}, 'optimizer': {'_target_': 'torch.optim.AdamW', 'lr': 0.0001, 'betas': [0.95, 0.999], 'eps': 1e-08, 'weight_decay': 1e-06}, 'training': {'device': 'cuda:0', 'seed': 45, 'debug': False, 'resume': True, 'lr_scheduler': 'cosine', 'lr_warmup_steps': 500, 'num_epochs': 1001, 'gradient_accumulate_every': 1, 'use_ema': True, 'rollout_every': 1000, 'checkpoint_every': 50, 'val_every': 1, 'sample_every': 1000, 'max_train_steps': None, 'max_val_steps': None, 'tqdm_interval_sec': 1.0}, 'logging': {'project': 'PegInsertionSide-v1-training', 'resume': True, 'mode': 'online', 'name': '2026.04.03-13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1', 'tags': ['train_diffusion_unet_lowdim', 'PegInsertionSide-v1', 'default'], 'id': None, 'group': None}, 'checkpoint': {'topk': {'monitor_key': 'val_loss', 'mode': 'min', 'k': 5, 'format_str': 'epoch={epoch:04d}-val_loss={val_loss:.3f}.ckpt'}, 'save_last_ckpt': True, 'save_last_snapshot': False}, 'multi_run': {'run_dir': 'data/outputs/2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1', 'wandb_name_base': '2026.04.03-13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1'}, 'task': {'name': 'PegInsertionSide-v1', 'obs_dim': 43, 'action_dim': 8, 'env_runner': {'_target_': 'llmbc.env_runner.maniskill_lowdim_runner.ManiskillLowdimRunner', 'env_name': 'llf-maniskill-PegInsertionSide-v1', 'instruction_type': 'b', 'feedback_type': ['hp', 'hn'], 'visual': False, 'n_train': 10, 'n_test': 50, 'n_envs': 10, 'max_steps': 100, 'n_obs_steps': 1, 'n_action_steps': 1, 'discount': 0.99}, 'dataset': {'_target_': 'llmbc.dataset.maniskill_lowdim_dataset.ManiskillLowdimDataset', 'data_path': 'datasets/PegInsertionSide-v1-reward-trans.pt', 'data_path2': 'datasets/PegInsertionSide-v1-reward-trans.pt', 'horizon': 1, 'pad_before': 0, 'pad_after': 0, 'use_manual_normalizer': False, 'val_ratio': 0.02, 'dummy_normalizer': False}, 'instructor': {'_target_': 'llmbc.translator.instructor.maniskill_instructor.PegInsertionSide_v1_instructor.PegInsertionSideV1Instructor'}}}
14
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_init.py:init():669] starting backend
15
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_init.py:init():673] sending inform_init request
16
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2026-04-03 13:29:11,180 INFO MainThread:3260278 [wandb_init.py:init():686] backend started and connected
18
+ 2026-04-03 13:29:11,185 INFO MainThread:3260278 [wandb_init.py:init():781] updated telemetry
19
+ 2026-04-03 13:29:11,240 INFO MainThread:3260278 [wandb_init.py:init():814] communicating run to backend with 90.0 second timeout
20
+ 2026-04-03 13:29:11,920 INFO MainThread:3260278 [wandb_init.py:init():867] starting run threads in backend
21
+ 2026-04-03 13:29:12,222 INFO MainThread:3260278 [wandb_run.py:_console_start():2451] atexit reg
22
+ 2026-04-03 13:29:12,222 INFO MainThread:3260278 [wandb_run.py:_redirect():2299] redirect: wrap_raw
23
+ 2026-04-03 13:29:12,222 INFO MainThread:3260278 [wandb_run.py:_redirect():2364] Wrapping output streams.
24
+ 2026-04-03 13:29:12,222 INFO MainThread:3260278 [wandb_run.py:_redirect():2389] Redirects installed.
25
+ 2026-04-03 13:29:12,224 INFO MainThread:3260278 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2026-04-03 13:29:12,224 INFO MainThread:3260278 [wandb_run.py:_config_callback():1389] config_cb None None {'output_dir': '/tmp2/chyang/workspace/LLM-BC/data/outputs/2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1'}
27
+ 2026-04-03 14:38:23,941 WARNING MsgRouterThr:3260278 [router.py:message_loop():75] message_loop has been closed
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/files/config.yaml ADDED
@@ -0,0 +1,229 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _target_:
2
+ value: llmbc.workspace.train_diffusion_unet_lowdim_workspace.TrainDiffusionUnetLowdimWorkspace
3
+ _wandb:
4
+ value:
5
+ cli_version: 0.18.6
6
+ m: []
7
+ python_version: 3.9.20
8
+ t:
9
+ "1":
10
+ - 1
11
+ - 2
12
+ - 3
13
+ - 5
14
+ - 11
15
+ - 12
16
+ - 41
17
+ - 49
18
+ - 50
19
+ - 51
20
+ - 53
21
+ - 55
22
+ - 71
23
+ - 83
24
+ - 95
25
+ - 98
26
+ - 100
27
+ "2":
28
+ - 1
29
+ - 2
30
+ - 3
31
+ - 5
32
+ - 11
33
+ - 12
34
+ - 41
35
+ - 49
36
+ - 50
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 83
42
+ - 95
43
+ - 98
44
+ - 100
45
+ "3":
46
+ - 13
47
+ - 15
48
+ - 16
49
+ - 23
50
+ - 55
51
+ - 61
52
+ "4": 3.9.20
53
+ "5": 0.18.6
54
+ "6": 4.47.1
55
+ "8":
56
+ - 5
57
+ "12": 0.18.6
58
+ "13": linux-x86_64
59
+ action_dim:
60
+ value: 8
61
+ checkpoint:
62
+ value:
63
+ save_last_ckpt: true
64
+ save_last_snapshot: false
65
+ topk:
66
+ format_str: epoch={epoch:04d}-val_loss={val_loss:.3f}.ckpt
67
+ k: 5
68
+ mode: min
69
+ monitor_key: val_loss
70
+ dataloader:
71
+ value:
72
+ batch_size: 256
73
+ num_workers: 0
74
+ persistent_workers: false
75
+ pin_memory: false
76
+ shuffle: true
77
+ ema:
78
+ value:
79
+ _target_: llmbc.model.diffusion.ema_model.EMAModel
80
+ inv_gamma: 1
81
+ max_value: 0.9999
82
+ min_value: 0
83
+ power: 0.75
84
+ update_after_step: 0
85
+ exp_name:
86
+ value: default
87
+ horizon:
88
+ value: 1
89
+ keypoint_visible_rate:
90
+ value: 1
91
+ logging:
92
+ value:
93
+ group: null
94
+ id: null
95
+ mode: online
96
+ name: 2026.04.03-13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1
97
+ project: PegInsertionSide-v1-training
98
+ resume: true
99
+ tags:
100
+ - train_diffusion_unet_lowdim
101
+ - PegInsertionSide-v1
102
+ - default
103
+ multi_run:
104
+ value:
105
+ run_dir: data/outputs/2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1
106
+ wandb_name_base: 2026.04.03-13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1
107
+ n_action_steps:
108
+ value: 1
109
+ n_latency_steps:
110
+ value: 0
111
+ n_obs_steps:
112
+ value: 1
113
+ name:
114
+ value: train_diffusion_unet_lowdim
115
+ obs_as_global_cond:
116
+ value: true
117
+ obs_as_local_cond:
118
+ value: false
119
+ obs_dim:
120
+ value: 43
121
+ optimizer:
122
+ value:
123
+ _target_: torch.optim.AdamW
124
+ betas:
125
+ - 0.95
126
+ - 0.999
127
+ eps: 1e-08
128
+ lr: 0.0001
129
+ weight_decay: 1e-06
130
+ output_dir:
131
+ value: /tmp2/chyang/workspace/LLM-BC/data/outputs/2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1
132
+ past_action_visible:
133
+ value: false
134
+ policy:
135
+ value:
136
+ _target_: llmbc.policy.diffusion_unet_lowdim_policy.DiffusionUnetLowdimPolicy
137
+ action_dim: 8
138
+ horizon: 1
139
+ model:
140
+ _target_: llmbc.model.diffusion.conditional_unet1d.ConditionalUnet1D
141
+ cond_predict_scale: true
142
+ diffusion_step_embed_dim: 256
143
+ down_dims:
144
+ - 256
145
+ - 512
146
+ - 1024
147
+ global_cond_dim: 43
148
+ input_dim: 8
149
+ kernel_size: 5
150
+ local_cond_dim: null
151
+ n_groups: 8
152
+ n_action_steps: 1
153
+ n_obs_steps: 1
154
+ noise_scheduler:
155
+ _target_: diffusers.schedulers.scheduling_ddpm.DDPMScheduler
156
+ beta_end: 0.02
157
+ beta_schedule: squaredcos_cap_v2
158
+ beta_start: 0.0001
159
+ clip_sample: true
160
+ num_train_timesteps: 100
161
+ prediction_type: epsilon
162
+ variance_type: fixed_small
163
+ num_inference_steps: 100
164
+ oa_step_convention: true
165
+ obs_as_global_cond: true
166
+ obs_as_local_cond: false
167
+ obs_dim: 43
168
+ pred_action_steps_only: false
169
+ pred_action_steps_only:
170
+ value: false
171
+ task:
172
+ value:
173
+ action_dim: 8
174
+ dataset:
175
+ _target_: llmbc.dataset.maniskill_lowdim_dataset.ManiskillLowdimDataset
176
+ data_path: datasets/PegInsertionSide-v1-reward-trans.pt
177
+ data_path2: datasets/PegInsertionSide-v1-reward-trans.pt
178
+ dummy_normalizer: false
179
+ horizon: 1
180
+ pad_after: 0
181
+ pad_before: 0
182
+ use_manual_normalizer: false
183
+ val_ratio: 0.02
184
+ env_runner:
185
+ _target_: llmbc.env_runner.maniskill_lowdim_runner.ManiskillLowdimRunner
186
+ discount: 0.99
187
+ env_name: llf-maniskill-PegInsertionSide-v1
188
+ feedback_type:
189
+ - hp
190
+ - hn
191
+ instruction_type: b
192
+ max_steps: 100
193
+ n_action_steps: 1
194
+ n_envs: 10
195
+ n_obs_steps: 1
196
+ n_test: 50
197
+ n_train: 10
198
+ visual: false
199
+ instructor:
200
+ _target_: llmbc.translator.instructor.maniskill_instructor.PegInsertionSide_v1_instructor.PegInsertionSideV1Instructor
201
+ name: PegInsertionSide-v1
202
+ obs_dim: 43
203
+ task_name:
204
+ value: PegInsertionSide-v1
205
+ training:
206
+ value:
207
+ checkpoint_every: 50
208
+ debug: false
209
+ device: cuda:0
210
+ gradient_accumulate_every: 1
211
+ lr_scheduler: cosine
212
+ lr_warmup_steps: 500
213
+ max_train_steps: null
214
+ max_val_steps: null
215
+ num_epochs: 1001
216
+ resume: true
217
+ rollout_every: 1000
218
+ sample_every: 1000
219
+ seed: 45
220
+ tqdm_interval_sec: 1
221
+ use_ema: true
222
+ val_every: 1
223
+ val_dataloader:
224
+ value:
225
+ batch_size: 256
226
+ num_workers: 0
227
+ persistent_workers: false
228
+ pin_memory: false
229
+ shuffle: false
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/files/output.log ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ /home/chyang/miniconda3/envs/llm-bc/lib/python3.9/site-packages/torch/random.py:157: UserWarning: CUDA reports that you have 2 available devices, and you have used fork_rng without explicitly specifying which devices are being used. For safety, we initialize *every* CUDA device by default, which can be quite slow if you have a lot of CUDAs. If you know that you are only making use of a few CUDA devices, set the environment variable CUDA_VISIBLE_DEVICES or the 'devices' keyword argument of fork_rng with the set of devices you are actually using. For example, if you are using CPU only, set device.upper()_VISIBLE_DEVICES= or devices=[]; if you are using device 0 only, set CUDA_VISIBLE_DEVICES=0 or devices=[0]. To initialize all devices and suppress this warning, set the 'devices' keyword argument to `range(torch.cuda.device_count())`.
2
+ warnings.warn(message)
3
+ Eval ManiskilldLowdimRunner 1/6: 0%| | 0/100 [00:00<?, ?it/s]/tmp2/chyang/workspace/LLM-BC/llmbc/common/llfbench_util.py:39: UserWarning: Creating a tensor from a list of numpy.ndarrays is extremely slow. Please consider converting the list to a single numpy.ndarray with numpy.array() before converting to a tensor. (Triggered internally at ../torch/csrc/utils/tensor_new.cpp:275.)
4
+ obs = torch.tensor(obs, dtype=torch.float32).unsqueeze(dim=0).to(device)
5
+
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/files/wandb-metadata.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.8.0-101-generic-x86_64-with-glibc2.35",
3
+ "python": "3.9.20",
4
+ "startedAt": "2026-04-03T05:29:11.180286Z",
5
+ "args": [
6
+ "--config-path",
7
+ "config/main_table",
8
+ "--config-name",
9
+ "bdpo_behavior_PegInsertionSide-v1_exp.yaml",
10
+ "training.seed=45"
11
+ ],
12
+ "program": "/tmp2/chyang/workspace/LLM-BC/./train.py",
13
+ "codePath": "train.py",
14
+ "git": {
15
+ "remote": "https://github.com/CHYang25/LLM-BC.git",
16
+ "commit": "53919ebcf1196f1d199e9e87cc0cfe08e06b97ce"
17
+ },
18
+ "email": "chris920325@gmail.com",
19
+ "root": "/tmp2/chyang/workspace/LLM-BC/data/outputs/2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1",
20
+ "host": "A6000-2",
21
+ "username": "chyang",
22
+ "executable": "/home/chyang/miniconda3/envs/llm-bc/bin/python3",
23
+ "codePathLocal": "train.py",
24
+ "cpu_count": 12,
25
+ "cpu_count_logical": 24,
26
+ "gpu": "NVIDIA RTX A6000",
27
+ "gpu_count": 2,
28
+ "disk": {
29
+ "/": {
30
+ "total": "1967317549056",
31
+ "used": "1010299678720"
32
+ }
33
+ },
34
+ "memory": {
35
+ "total": "134538502144"
36
+ },
37
+ "cpu": {
38
+ "count": 12,
39
+ "countLogical": 24
40
+ },
41
+ "gpu_nvidia": [
42
+ {
43
+ "name": "NVIDIA RTX A6000",
44
+ "memoryTotal": "51527024640",
45
+ "cudaCores": 10752,
46
+ "architecture": "Ampere"
47
+ },
48
+ {
49
+ "name": "NVIDIA RTX A6000",
50
+ "memoryTotal": "51527024640",
51
+ "cudaCores": 10752,
52
+ "architecture": "Ampere"
53
+ }
54
+ ],
55
+ "cudaVersion": "12.6"
56
+ }
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/success_rate":0.3,"_step":72071,"train/cumulative_reward":25.12234242059938,"test/cumulative_reward":22.414451283032026,"test/mean_score":0.8277486884593963,"lr":0,"train_action_mse_error":6.132695853011683e-05,"_runtime":4152.761406703,"global_step":72071,"epoch":1000,"train/mean_score":0.7515433937311172,"test/success_rate":0.48,"_wandb":{"runtime":4152},"_timestamp":1.775198297653251e+09,"val_loss":0.01334930770099163,"train_loss":0.00043470339409800037}
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/logs/debug-core.log ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-04-03T13:29:10.547033785+08:00","level":"INFO","msg":"started logging, with flags","port-filename":"/tmp/tmpwwn0l4u7/port-3260278.txt","pid":3260278,"debug":false,"disable-analytics":false}
2
+ {"time":"2026-04-03T13:29:10.54705783+08:00","level":"INFO","msg":"FeatureState","shutdownOnParentExitEnabled":false}
3
+ {"time":"2026-04-03T13:29:10.548061284+08:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":3260278}
4
+ {"time":"2026-04-03T13:29:10.548080108+08:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":41995,"Zone":""}}
5
+ {"time":"2026-04-03T13:29:10.743882614+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:58688"}
6
+ {"time":"2026-04-03T13:29:11.182938502+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"0ietglbs","id":"127.0.0.1:58688"}
7
+ {"time":"2026-04-03T13:29:11.290720006+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"0ietglbs","id":"127.0.0.1:58688"}
8
+ {"time":"2026-04-03T14:38:23.941510522+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"127.0.0.1:58688"}
9
+ {"time":"2026-04-03T14:38:23.941638228+08:00","level":"INFO","msg":"connection: Close: initiating connection closure","id":"127.0.0.1:58688"}
10
+ {"time":"2026-04-03T14:38:23.941673177+08:00","level":"INFO","msg":"server is shutting down"}
11
+ {"time":"2026-04-03T14:38:23.941847669+08:00","level":"INFO","msg":"connection: Close: connection successfully closed","id":"127.0.0.1:58688"}
12
+ {"time":"2026-04-03T14:38:25.520598219+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"127.0.0.1:58688"}
13
+ {"time":"2026-04-03T14:38:25.520647389+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"127.0.0.1:58688"}
14
+ {"time":"2026-04-03T14:38:25.520676055+08:00","level":"INFO","msg":"server is closed"}
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/logs/debug-internal.log ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-04-03T13:29:11.183039659+08:00","level":"INFO","msg":"using version","core version":"0.18.6"}
2
+ {"time":"2026-04-03T13:29:11.18304756+08:00","level":"INFO","msg":"created symlink","path":"/tmp2/chyang/workspace/LLM-BC/data/outputs/2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/logs/debug-core.log"}
3
+ {"time":"2026-04-03T13:29:11.290645001+08:00","level":"INFO","msg":"created new stream","id":"0ietglbs"}
4
+ {"time":"2026-04-03T13:29:11.290709294+08:00","level":"INFO","msg":"stream: started","id":"0ietglbs"}
5
+ {"time":"2026-04-03T13:29:11.290717333+08:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"0ietglbs"}}
6
+ {"time":"2026-04-03T13:29:11.290758229+08:00","level":"INFO","msg":"sender: started","stream_id":"0ietglbs"}
7
+ {"time":"2026-04-03T13:29:11.290756734+08:00","level":"INFO","msg":"handler: started","stream_id":{"value":"0ietglbs"}}
8
+ {"time":"2026-04-03T13:29:11.922864379+08:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2026-04-03T14:38:23.941643874+08:00","level":"INFO","msg":"stream: closing","id":"0ietglbs"}
10
+ {"time":"2026-04-03T14:38:23.941726051+08:00","level":"INFO","msg":"Stopping system monitor"}
11
+ {"time":"2026-04-03T14:38:23.942686523+08:00","level":"INFO","msg":"Stopped system monitor"}
12
+ {"time":"2026-04-03T14:38:25.088055309+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
13
+ {"time":"2026-04-03T14:38:25.520228302+08:00","level":"INFO","msg":"handler: closed","stream_id":{"value":"0ietglbs"}}
14
+ {"time":"2026-04-03T14:38:25.520305908+08:00","level":"INFO","msg":"sender: closed","stream_id":"0ietglbs"}
15
+ {"time":"2026-04-03T14:38:25.520298235+08:00","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"0ietglbs"}}
16
+ {"time":"2026-04-03T14:38:25.520468398+08:00","level":"INFO","msg":"stream: closed","id":"0ietglbs"}
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/logs/debug.log ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-04-03 13:29:11,178 INFO MainThread:3260278 [wandb_setup.py:_flush():79] Current SDK version is 0.18.6
2
+ 2026-04-03 13:29:11,178 INFO MainThread:3260278 [wandb_setup.py:_flush():79] Configure stats pid to 3260278
3
+ 2026-04-03 13:29:11,178 INFO MainThread:3260278 [wandb_setup.py:_flush():79] Loading settings from /home/chyang/.config/wandb/settings
4
+ 2026-04-03 13:29:11,178 INFO MainThread:3260278 [wandb_setup.py:_flush():79] Loading settings from /tmp2/chyang/workspace/LLM-BC/wandb/settings
5
+ 2026-04-03 13:29:11,178 INFO MainThread:3260278 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': 'online', '_disable_service': None}
7
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train.py', 'program_abspath': '/tmp2/chyang/workspace/LLM-BC/train.py', 'program': '/tmp2/chyang/workspace/LLM-BC/./train.py'}
8
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_init.py:_log_setup():533] Logging user logs to /tmp2/chyang/workspace/LLM-BC/data/outputs/2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/logs/debug.log
10
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_init.py:_log_setup():534] Logging internal logs to /tmp2/chyang/workspace/LLM-BC/data/outputs/2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/logs/debug-internal.log
11
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_init.py:init():619] calling init triggers
12
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_init.py:init():626] wandb.init called with sweep_config: {}
13
+ config: {'name': 'train_diffusion_unet_lowdim', '_target_': 'llmbc.workspace.train_diffusion_unet_lowdim_workspace.TrainDiffusionUnetLowdimWorkspace', 'obs_dim': 43, 'action_dim': 8, 'task_name': 'PegInsertionSide-v1', 'exp_name': 'default', 'horizon': 1, 'n_obs_steps': 1, 'n_action_steps': 1, 'n_latency_steps': 0, 'past_action_visible': False, 'keypoint_visible_rate': 1.0, 'obs_as_local_cond': False, 'obs_as_global_cond': True, 'pred_action_steps_only': False, 'policy': {'_target_': 'llmbc.policy.diffusion_unet_lowdim_policy.DiffusionUnetLowdimPolicy', 'model': {'_target_': 'llmbc.model.diffusion.conditional_unet1d.ConditionalUnet1D', 'input_dim': 8, 'local_cond_dim': None, 'global_cond_dim': 43, 'diffusion_step_embed_dim': 256, 'down_dims': [256, 512, 1024], 'kernel_size': 5, 'n_groups': 8, 'cond_predict_scale': True}, 'noise_scheduler': {'_target_': 'diffusers.schedulers.scheduling_ddpm.DDPMScheduler', 'num_train_timesteps': 100, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'squaredcos_cap_v2', 'variance_type': 'fixed_small', 'clip_sample': True, 'prediction_type': 'epsilon'}, 'horizon': 1, 'obs_dim': 43, 'action_dim': 8, 'n_action_steps': 1, 'n_obs_steps': 1, 'num_inference_steps': 100, 'obs_as_local_cond': False, 'obs_as_global_cond': True, 'pred_action_steps_only': False, 'oa_step_convention': True}, 'ema': {'_target_': 'llmbc.model.diffusion.ema_model.EMAModel', 'update_after_step': 0, 'inv_gamma': 1.0, 'power': 0.75, 'min_value': 0.0, 'max_value': 0.9999}, 'dataloader': {'batch_size': 256, 'num_workers': 0, 'shuffle': True, 'pin_memory': False, 'persistent_workers': False}, 'val_dataloader': {'batch_size': 256, 'num_workers': 0, 'shuffle': False, 'pin_memory': False, 'persistent_workers': False}, 'optimizer': {'_target_': 'torch.optim.AdamW', 'lr': 0.0001, 'betas': [0.95, 0.999], 'eps': 1e-08, 'weight_decay': 1e-06}, 'training': {'device': 'cuda:0', 'seed': 45, 'debug': False, 'resume': True, 'lr_scheduler': 'cosine', 'lr_warmup_steps': 500, 'num_epochs': 1001, 'gradient_accumulate_every': 1, 'use_ema': True, 'rollout_every': 1000, 'checkpoint_every': 50, 'val_every': 1, 'sample_every': 1000, 'max_train_steps': None, 'max_val_steps': None, 'tqdm_interval_sec': 1.0}, 'logging': {'project': 'PegInsertionSide-v1-training', 'resume': True, 'mode': 'online', 'name': '2026.04.03-13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1', 'tags': ['train_diffusion_unet_lowdim', 'PegInsertionSide-v1', 'default'], 'id': None, 'group': None}, 'checkpoint': {'topk': {'monitor_key': 'val_loss', 'mode': 'min', 'k': 5, 'format_str': 'epoch={epoch:04d}-val_loss={val_loss:.3f}.ckpt'}, 'save_last_ckpt': True, 'save_last_snapshot': False}, 'multi_run': {'run_dir': 'data/outputs/2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1', 'wandb_name_base': '2026.04.03-13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1'}, 'task': {'name': 'PegInsertionSide-v1', 'obs_dim': 43, 'action_dim': 8, 'env_runner': {'_target_': 'llmbc.env_runner.maniskill_lowdim_runner.ManiskillLowdimRunner', 'env_name': 'llf-maniskill-PegInsertionSide-v1', 'instruction_type': 'b', 'feedback_type': ['hp', 'hn'], 'visual': False, 'n_train': 10, 'n_test': 50, 'n_envs': 10, 'max_steps': 100, 'n_obs_steps': 1, 'n_action_steps': 1, 'discount': 0.99}, 'dataset': {'_target_': 'llmbc.dataset.maniskill_lowdim_dataset.ManiskillLowdimDataset', 'data_path': 'datasets/PegInsertionSide-v1-reward-trans.pt', 'data_path2': 'datasets/PegInsertionSide-v1-reward-trans.pt', 'horizon': 1, 'pad_before': 0, 'pad_after': 0, 'use_manual_normalizer': False, 'val_ratio': 0.02, 'dummy_normalizer': False}, 'instructor': {'_target_': 'llmbc.translator.instructor.maniskill_instructor.PegInsertionSide_v1_instructor.PegInsertionSideV1Instructor'}}}
14
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_init.py:init():669] starting backend
15
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [wandb_init.py:init():673] sending inform_init request
16
+ 2026-04-03 13:29:11,179 INFO MainThread:3260278 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2026-04-03 13:29:11,180 INFO MainThread:3260278 [wandb_init.py:init():686] backend started and connected
18
+ 2026-04-03 13:29:11,185 INFO MainThread:3260278 [wandb_init.py:init():781] updated telemetry
19
+ 2026-04-03 13:29:11,240 INFO MainThread:3260278 [wandb_init.py:init():814] communicating run to backend with 90.0 second timeout
20
+ 2026-04-03 13:29:11,920 INFO MainThread:3260278 [wandb_init.py:init():867] starting run threads in backend
21
+ 2026-04-03 13:29:12,222 INFO MainThread:3260278 [wandb_run.py:_console_start():2451] atexit reg
22
+ 2026-04-03 13:29:12,222 INFO MainThread:3260278 [wandb_run.py:_redirect():2299] redirect: wrap_raw
23
+ 2026-04-03 13:29:12,222 INFO MainThread:3260278 [wandb_run.py:_redirect():2364] Wrapping output streams.
24
+ 2026-04-03 13:29:12,222 INFO MainThread:3260278 [wandb_run.py:_redirect():2389] Redirects installed.
25
+ 2026-04-03 13:29:12,224 INFO MainThread:3260278 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2026-04-03 13:29:12,224 INFO MainThread:3260278 [wandb_run.py:_config_callback():1389] config_cb None None {'output_dir': '/tmp2/chyang/workspace/LLM-BC/data/outputs/2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1'}
27
+ 2026-04-03 14:38:23,941 WARNING MsgRouterThr:3260278 [router.py:message_loop():75] message_loop has been closed
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/run-20260403_132911-0ietglbs/run-0ietglbs.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa8fddd8b9efc9ae49c4e088a6808b80ba409e333c59efc58d0723c7eda0d23c
3
+ size 35970313
2026.04.03/13.29.01_train_diffusion_unet_lowdim_PegInsertionSide-v1/wandb/wandb-resume.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"run_id": "0ietglbs"}