CHYang25 commited on
Commit
d060d7e
·
verified ·
1 Parent(s): 786f449

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +5 -0
  2. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/.hydra/config.yaml +163 -0
  3. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/.hydra/hydra.yaml +155 -0
  4. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/.hydra/overrides.yaml +1 -0
  5. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=0100-test_success_rate=0.160.ckpt +3 -0
  6. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=0500-test_success_rate=0.780.ckpt +3 -0
  7. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=0700-test_success_rate=0.760.ckpt +3 -0
  8. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=1200-test_success_rate=0.760.ckpt +3 -0
  9. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=1600-test_success_rate=0.780.ckpt +3 -0
  10. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=7900-test_success_rate=0.760.ckpt +3 -0
  11. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/checkpoints/latest.ckpt +3 -0
  12. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/logs.json.txt +3 -0
  13. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/train.log +16 -0
  14. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/debug-internal.log +16 -0
  15. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/debug.log +28 -0
  16. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/files/output.log +3 -0
  17. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/files/requirements.txt +858 -0
  18. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/files/wandb-metadata.json +107 -0
  19. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/logs/debug-core.log +7 -0
  20. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/logs/debug-internal.log +8 -0
  21. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/logs/debug.log +26 -0
  22. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/run-raachjxu.wandb +3 -0
  23. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/files/config.yaml +272 -0
  24. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/files/output.log +3 -0
  25. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/files/wandb-metadata.json +107 -0
  26. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/files/wandb-summary.json +1 -0
  27. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/logs/debug-core.log +14 -0
  28. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/logs/debug-internal.log +16 -0
  29. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/logs/debug.log +28 -0
  30. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/run-raachjxu.wandb +3 -0
  31. 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/wandb-resume.json +1 -0
  32. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/.hydra/config.yaml +163 -0
  33. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/.hydra/hydra.yaml +155 -0
  34. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/.hydra/overrides.yaml +1 -0
  35. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=10200-test_success_rate=0.780.ckpt +3 -0
  36. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=11400-test_success_rate=0.800.ckpt +3 -0
  37. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=12200-test_success_rate=0.800.ckpt +3 -0
  38. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=12300-test_success_rate=0.780.ckpt +3 -0
  39. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=7400-test_success_rate=0.780.ckpt +3 -0
  40. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/checkpoints/latest.ckpt +3 -0
  41. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/logs.json.txt +3 -0
  42. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/train.log +8 -0
  43. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/debug-internal.log +16 -0
  44. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/debug.log +27 -0
  45. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/files/config.yaml +271 -0
  46. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/files/output.log +3 -0
  47. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/files/wandb-metadata.json +107 -0
  48. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/files/wandb-summary.json +1 -0
  49. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/logs/debug-core.log +14 -0
  50. 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/logs/debug-internal.log +16 -0
.gitattributes CHANGED
@@ -42,3 +42,8 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
42
  2026.01.26/11.48.33_train_llm_lowdim_box-close-v2/wandb/run-20260126_114842-zut8rsd7/run-zut8rsd7.wandb filter=lfs diff=lfs merge=lfs -text
43
  2026.01.26/11.49.03_train_llm_lowdim_box-close-v2/wandb/run-20260126_114912-eqoznxjg/run-eqoznxjg.wandb filter=lfs diff=lfs merge=lfs -text
44
  2026.01.26/12.30.16_train_llm_lowdim_box-close-v2/wandb/run-20260126_123023-egujxdao/run-egujxdao.wandb filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
42
  2026.01.26/11.48.33_train_llm_lowdim_box-close-v2/wandb/run-20260126_114842-zut8rsd7/run-zut8rsd7.wandb filter=lfs diff=lfs merge=lfs -text
43
  2026.01.26/11.49.03_train_llm_lowdim_box-close-v2/wandb/run-20260126_114912-eqoznxjg/run-eqoznxjg.wandb filter=lfs diff=lfs merge=lfs -text
44
  2026.01.26/12.30.16_train_llm_lowdim_box-close-v2/wandb/run-20260126_123023-egujxdao/run-egujxdao.wandb filter=lfs diff=lfs merge=lfs -text
45
+ 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/logs.json.txt filter=lfs diff=lfs merge=lfs -text
46
+ 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/run-raachjxu.wandb filter=lfs diff=lfs merge=lfs -text
47
+ 2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/run-raachjxu.wandb filter=lfs diff=lfs merge=lfs -text
48
+ 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/logs.json.txt filter=lfs diff=lfs merge=lfs -text
49
+ 2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/run-ccfd9xxl.wandb filter=lfs diff=lfs merge=lfs -text
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/.hydra/config.yaml ADDED
@@ -0,0 +1,163 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: train_llmbc_lowdim
2
+ _target_: llmbc.workspace.train_llmbc_lowdim_workspace.TrainLLMBCLowdimWorkspace
3
+ obs_dim: ${task.obs_dim}
4
+ action_dim: ${task.action_dim}
5
+ task_name: ${task.name}
6
+ exp_name: default
7
+ model_name: ${llm.name}
8
+ horizon: 1
9
+ n_obs_steps: 1
10
+ n_action_steps: 1
11
+ n_latency_steps: 0
12
+ past_action_visible: false
13
+ llm_orig_expert_feedback: true
14
+ llm_do_sample: false
15
+ policy:
16
+ _target_: llmbc.policy.llmbc_lowdim_policy.LLMBCLowdimPolicy
17
+ model:
18
+ _target_: llmbc.model.policy.policy_mlp.PolicyMLP
19
+ input_size: ${eval:'${n_obs_steps}*${obs_dim}'}
20
+ hidden_size:
21
+ - 256
22
+ - 256
23
+ output_size: ${eval:'${n_action_steps}*${action_dim}'}
24
+ activation: relu
25
+ n_obs_steps: ${n_obs_steps}
26
+ n_action_steps: ${n_action_steps}
27
+ obs_dim: ${obs_dim}
28
+ action_dim: ${action_dim}
29
+ llm_discriminator:
30
+ _target_: llmbc.discriminator.llm_ce_discriminator.LLMCEDiscriminator
31
+ task_id: ${task_name}
32
+ llm_translator:
33
+ _target_: llmbc.translator.llm_translator.LLMTranslator
34
+ cfg: ${llm}
35
+ obs_dim: ${task.obs_dim}
36
+ action_dim: ${task.action_dim}
37
+ horizon: ${horizon}
38
+ n_obs_steps: ${n_obs_steps}
39
+ n_action_steps: ${n_action_steps}
40
+ loss_bc_weight: 1.0
41
+ loss_llm_weight: 0.001
42
+ horizon: ${horizon}
43
+ n_obs_steps: ${n_obs_steps}
44
+ n_action_steps: ${n_action_steps}
45
+ normalize_llm_loss: true
46
+ dataloader:
47
+ batch_size: 16
48
+ num_workers: 0
49
+ shuffle: true
50
+ pin_memory: false
51
+ persistent_workers: false
52
+ val_dataloader:
53
+ batch_size: 16
54
+ num_workers: 0
55
+ shuffle: true
56
+ pin_memory: false
57
+ persistent_workers: false
58
+ optimizer:
59
+ _target_: torch.optim.AdamW
60
+ lr: 0.01
61
+ betas:
62
+ - 0.95
63
+ - 0.999
64
+ eps: 1.0e-08
65
+ weight_decay: 1.0e-06
66
+ training:
67
+ device: cuda:0
68
+ seed: 42
69
+ debug: false
70
+ resume: false
71
+ lr_scheduler: cosine
72
+ lr_warmup_steps: 10
73
+ num_epochs: 20001
74
+ gradient_accumulate_every: 8
75
+ grad_norm_clip: 0.5
76
+ rollout_every: 100
77
+ checkpoint_every: 100
78
+ val_every: 1
79
+ sample_every: 100
80
+ sample_max_batch: 128
81
+ max_train_steps: null
82
+ max_val_steps: null
83
+ tqdm_interval_sec: 1.0
84
+ logging:
85
+ project: ${task.name}-training-split-ablation
86
+ resume: true
87
+ mode: online
88
+ name: ${now:%Y.%m.%d-%H.%M.%S}_${name}_${task_name}
89
+ tags:
90
+ - ${name}
91
+ - ${task_name}
92
+ - ${exp_name}
93
+ id: null
94
+ group: null
95
+ checkpoint:
96
+ topk:
97
+ monitor_key: test_success_rate
98
+ mode: max
99
+ k: 5
100
+ format_str: epoch={epoch:04d}-test_success_rate={test_success_rate:.3f}.ckpt
101
+ save_last_ckpt: true
102
+ save_last_snapshot: false
103
+ multi_run:
104
+ run_dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
105
+ wandb_name_base: ${now:%Y.%m.%d-%H.%M.%S}_${name}_${task_name}
106
+ task:
107
+ name: box-close-v2
108
+ obs_dim: 9
109
+ action_dim: 4
110
+ env_runner:
111
+ _target_: llmbc.env_runner.metaworld_lowdim_runner.MetaworldLowdimRunner
112
+ env_name: llf-metaworld-box-close-v2
113
+ n_train: 10
114
+ n_test: 50
115
+ n_envs: 10
116
+ max_steps: 30
117
+ n_obs_steps: ${n_obs_steps}
118
+ n_action_steps: ${n_action_steps}
119
+ instruction_type: b
120
+ feedback_type:
121
+ - hp
122
+ - hn
123
+ - fp
124
+ visual: false
125
+ discount: 0.9
126
+ dataset:
127
+ _target_: llmbc.dataset.metaworld_lowdim_dataset.MetaworldLowdimDataset
128
+ data_path: datasets/box-close-v2-o4-mini-split0.05.pt
129
+ data_path2: datasets/box-close-v2-o4-mini-split0.05.pt
130
+ horizon: ${horizon}
131
+ pad_before: ${eval:'${n_obs_steps}-1'}
132
+ pad_after: ${eval:'${n_action_steps}-1'}
133
+ obs_eef_target: true
134
+ use_manual_normalizer: false
135
+ val_ratio: 0.1
136
+ dummy_normalizer: true
137
+ instructor:
138
+ _target_: llmbc.translator.instructor.metaworld_instructor.box_close_v2_instructor.BoxCloseV2Instructor
139
+ llm:
140
+ name: HuggingFaceTB/SmolLM2-135M-Instruct
141
+ model_name: SmolLM2-135M-Instruct
142
+ config_target: llmbc.model.llm.llama_lowdim_model.LowdimLlamaConfig
143
+ causal_lm_target: llmbc.model.llm.llama_lowdim_model.LowdimLlamaForCausalLM
144
+ use_quantization: false
145
+ use_joint_mlp_projector: true
146
+ llm_mode: ete-finetuned
147
+ finetune_mode: orig
148
+ checkpoint: data/outputs/2026.01.26/10.26.52_train_llm_lowdim_box-close-v2/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-box-close-v2/checkpoint-6000
149
+ max_length: 100
150
+ lora_config:
151
+ r: 32
152
+ lora_alpha: 64
153
+ lora_dropout: 0.05
154
+ bias: none
155
+ task_type: CAUSAL_LM
156
+ prompter:
157
+ _target_: llmbc.translator.prompter.smollm2_prompter.SmolLM2Prompter
158
+ use_joint_mlp_projector: true
159
+ hydra:
160
+ job:
161
+ override_dirname: ${model_name}
162
+ run:
163
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${model_name}
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/.hydra/hydra.yaml ADDED
@@ -0,0 +1,155 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
4
+ sweep:
5
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - training.seed=42
116
+ job:
117
+ name: train
118
+ chdir: null
119
+ override_dirname: training.seed=42
120
+ id: ???
121
+ num: ???
122
+ config_name: llmbc_box-close-v2_o4-mini_split0.05.yaml
123
+ env_set: {}
124
+ env_copy: []
125
+ config:
126
+ override_dirname:
127
+ kv_sep: '='
128
+ item_sep: ','
129
+ exclude_keys: []
130
+ runtime:
131
+ version: 1.2.0
132
+ version_base: '1.2'
133
+ cwd: /work/u1131674/LLM-BC
134
+ config_sources:
135
+ - path: hydra.conf
136
+ schema: pkg
137
+ provider: hydra
138
+ - path: /work/u1131674/LLM-BC/config/data_efficiency
139
+ schema: file
140
+ provider: main
141
+ - path: ''
142
+ schema: structured
143
+ provider: schema
144
+ output_dir: /work/u1131674/LLM-BC/data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2
145
+ choices:
146
+ hydra/env: default
147
+ hydra/callbacks: null
148
+ hydra/job_logging: default
149
+ hydra/hydra_logging: default
150
+ hydra/hydra_help: default
151
+ hydra/help: default
152
+ hydra/sweeper: basic
153
+ hydra/launcher: basic
154
+ hydra/output: default
155
+ verbose: false
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/.hydra/overrides.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ - training.seed=42
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=0100-test_success_rate=0.160.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22f161c891d32a16e1f25c104b2b9aa3a57c120633055ebb854b3ecd6dae61ec
3
+ size 864584
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=0500-test_success_rate=0.780.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2812880df907aa389c47e00e965dc13dfc39c20829a359b733daae4e5d47c24e
3
+ size 864584
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=0700-test_success_rate=0.760.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cf5ea1887ccf4fbe54ee0f62a3c42c8636858e5d7cab73a41db1886461230f1
3
+ size 864584
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=1200-test_success_rate=0.760.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07061c3c037995b5755f7f10d60cd9842941c79ebe1b414fa338525f19827c0b
3
+ size 864584
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=1600-test_success_rate=0.780.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7bf4a2477a2cac9ed1ad4be7517475fc7dc7d7581322acb3f22e003e7648cde
3
+ size 864584
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=7900-test_success_rate=0.760.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70bdf629f036405c4028c1cdb1788aa9dc7af9c9798911176ed24681a70194c9
3
+ size 864584
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/checkpoints/latest.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e32da9502909cec8e4929595d3503929356d81b482509099339493650f33ce3
3
+ size 864584
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/logs.json.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c07fd38f5637dc4005b7cbb80a9b9f7811c967b19a896a276b2c5c328249d07
3
+ size 55693005
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/train.log ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-01-27 11:52:52,548][numexpr.utils][INFO] - Note: detected 112 virtual cores but NumExpr set to maximum of 64, check "NUMEXPR_MAX_THREADS" environment variable.
2
+ [2026-01-27 11:52:52,548][numexpr.utils][INFO] - Note: NumExpr detected 112 cores but "NUMEXPR_MAX_THREADS" not set, so enforcing safe limit of 16.
3
+ [2026-01-27 11:52:52,549][numexpr.utils][INFO] - NumExpr defaulting to 16 threads.
4
+ [2026-01-27 11:52:52,558][numexpr.utils][INFO] - Note: detected 112 virtual cores but NumExpr set to maximum of 64, check "NUMEXPR_MAX_THREADS" environment variable.
5
+ [2026-01-27 11:52:52,558][numexpr.utils][INFO] - Note: NumExpr detected 112 cores but "NUMEXPR_MAX_THREADS" not set, so enforcing safe limit of 16.
6
+ [2026-01-27 11:52:52,559][numexpr.utils][INFO] - NumExpr defaulting to 16 threads.
7
+ [2026-01-27 11:52:58,492][datasets][INFO] - PyTorch version 2.2.2 available.
8
+ [2026-01-27 11:52:58,494][datasets][INFO] - TensorFlow version 2.15.1 available.
9
+ [2026-01-27 11:52:58,495][datasets][INFO] - JAX version 0.4.30 available.
10
+ [2026-01-27 11:52:58,756][datasets][INFO] - PyTorch version 2.2.2 available.
11
+ [2026-01-27 11:52:58,757][datasets][INFO] - TensorFlow version 2.15.1 available.
12
+ [2026-01-27 11:52:58,760][datasets][INFO] - JAX version 0.4.30 available.
13
+ [2026-01-27 11:53:06,547][absl][INFO] - MUJOCO_GL=osmesa, attempting to import specified OpenGL backend.
14
+ [2026-01-27 11:53:06,558][absl][INFO] - MuJoCo library version is: 2.3.7
15
+ [2026-01-27 11:53:07,104][absl][INFO] - MUJOCO_GL=osmesa, attempting to import specified OpenGL backend.
16
+ [2026-01-27 11:53:07,115][absl][INFO] - MuJoCo library version is: 2.3.7
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/debug-internal.log ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-01-27T11:53:08.074868883+08:00","level":"INFO","msg":"using version","core version":"0.18.6"}
2
+ {"time":"2026-01-27T11:53:08.074878267+08:00","level":"INFO","msg":"created symlink","path":"/work/u1131674/LLM-BC/data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/logs/debug-core.log"}
3
+ {"time":"2026-01-27T11:53:08.190280513+08:00","level":"INFO","msg":"created new stream","id":"raachjxu"}
4
+ {"time":"2026-01-27T11:53:08.190312975+08:00","level":"INFO","msg":"stream: started","id":"raachjxu"}
5
+ {"time":"2026-01-27T11:53:08.190338185+08:00","level":"INFO","msg":"sender: started","stream_id":"raachjxu"}
6
+ {"time":"2026-01-27T11:53:08.190329168+08:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"raachjxu"}}
7
+ {"time":"2026-01-27T11:53:08.1903328+08:00","level":"INFO","msg":"handler: started","stream_id":{"value":"raachjxu"}}
8
+ {"time":"2026-01-27T11:53:08.732831596+08:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2026-01-27T22:37:36.57283309+08:00","level":"INFO","msg":"stream: closing","id":"raachjxu"}
10
+ {"time":"2026-01-27T22:37:36.572875878+08:00","level":"INFO","msg":"Stopping system monitor"}
11
+ {"time":"2026-01-27T22:37:36.573489042+08:00","level":"INFO","msg":"Stopped system monitor"}
12
+ {"time":"2026-01-27T22:37:40.850119587+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
13
+ {"time":"2026-01-27T22:37:41.125467393+08:00","level":"INFO","msg":"handler: closed","stream_id":{"value":"raachjxu"}}
14
+ {"time":"2026-01-27T22:37:41.125498907+08:00","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"raachjxu"}}
15
+ {"time":"2026-01-27T22:37:41.125540962+08:00","level":"INFO","msg":"sender: closed","stream_id":"raachjxu"}
16
+ {"time":"2026-01-27T22:37:41.125562117+08:00","level":"INFO","msg":"stream: closed","id":"raachjxu"}
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/debug.log ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-01-27 11:53:08,070 INFO MainThread:2419965 [wandb_setup.py:_flush():79] Current SDK version is 0.18.6
2
+ 2026-01-27 11:53:08,070 INFO MainThread:2419965 [wandb_setup.py:_flush():79] Configure stats pid to 2419965
3
+ 2026-01-27 11:53:08,070 INFO MainThread:2419965 [wandb_setup.py:_flush():79] Loading settings from /home/u1131674/.config/wandb/settings
4
+ 2026-01-27 11:53:08,070 INFO MainThread:2419965 [wandb_setup.py:_flush():79] Loading settings from /work/u1131674/LLM-BC/wandb/settings
5
+ 2026-01-27 11:53:08,070 INFO MainThread:2419965 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2026-01-27 11:53:08,070 INFO MainThread:2419965 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': 'online', '_disable_service': None}
7
+ 2026-01-27 11:53:08,071 INFO MainThread:2419965 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train.py', 'program_abspath': '/work/u1131674/LLM-BC/train.py', 'program': '/work/u1131674/LLM-BC/./train.py'}
8
+ 2026-01-27 11:53:08,071 INFO MainThread:2419965 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2026-01-27 11:53:08,071 INFO MainThread:2419965 [wandb_init.py:_log_setup():533] Logging user logs to /work/u1131674/LLM-BC/data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/logs/debug.log
10
+ 2026-01-27 11:53:08,071 INFO MainThread:2419965 [wandb_init.py:_log_setup():534] Logging internal logs to /work/u1131674/LLM-BC/data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/logs/debug-internal.log
11
+ 2026-01-27 11:53:08,071 INFO MainThread:2419965 [wandb_init.py:init():619] calling init triggers
12
+ 2026-01-27 11:53:08,071 INFO MainThread:2419965 [wandb_init.py:init():626] wandb.init called with sweep_config: {}
13
+ config: {'name': 'train_llmbc_lowdim', '_target_': 'llmbc.workspace.train_llmbc_lowdim_workspace.TrainLLMBCLowdimWorkspace', 'obs_dim': 9, 'action_dim': 4, 'task_name': 'box-close-v2', 'exp_name': 'default', 'model_name': 'HuggingFaceTB/SmolLM2-135M-Instruct', 'horizon': 1, 'n_obs_steps': 1, 'n_action_steps': 1, 'n_latency_steps': 0, 'past_action_visible': False, 'llm_orig_expert_feedback': True, 'llm_do_sample': False, 'policy': {'_target_': 'llmbc.policy.llmbc_lowdim_policy.LLMBCLowdimPolicy', 'model': {'_target_': 'llmbc.model.policy.policy_mlp.PolicyMLP', 'input_size': 9, 'hidden_size': [256, 256], 'output_size': 4, 'activation': 'relu', 'n_obs_steps': 1, 'n_action_steps': 1}, 'obs_dim': 9, 'action_dim': 4, 'llm_discriminator': {'_target_': 'llmbc.discriminator.llm_ce_discriminator.LLMCEDiscriminator', 'task_id': 'box-close-v2', 'llm_translator': {'_target_': 'llmbc.translator.llm_translator.LLMTranslator', 'cfg': {'name': 'HuggingFaceTB/SmolLM2-135M-Instruct', 'model_name': 'SmolLM2-135M-Instruct', 'config_target': 'llmbc.model.llm.llama_lowdim_model.LowdimLlamaConfig', 'causal_lm_target': 'llmbc.model.llm.llama_lowdim_model.LowdimLlamaForCausalLM', 'use_quantization': False, 'use_joint_mlp_projector': True, 'llm_mode': 'ete-finetuned', 'finetune_mode': 'orig', 'checkpoint': 'data/outputs/2026.01.26/10.26.52_train_llm_lowdim_box-close-v2/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-box-close-v2/checkpoint-6000', 'max_length': 100, 'lora_config': {'r': 32, 'lora_alpha': 64, 'lora_dropout': 0.05, 'bias': 'none', 'task_type': 'CAUSAL_LM'}, 'prompter': {'_target_': 'llmbc.translator.prompter.smollm2_prompter.SmolLM2Prompter', 'use_joint_mlp_projector': True}, 'hydra': {'job': {'override_dirname': 'HuggingFaceTB/SmolLM2-135M-Instruct'}, 'run': {'dir': 'data/outputs/2026.01.27/11.52.50_HuggingFaceTB/SmolLM2-135M-Instruct'}}}, 'obs_dim': 9, 'action_dim': 4, 'horizon': 1, 'n_obs_steps': 1, 'n_action_steps': 1}}, 'loss_bc_weight': 1.0, 'loss_llm_weight': 0.001, 'horizon': 1, 'n_obs_steps': 1, 'n_action_steps': 1, 'normalize_llm_loss': True}, 'dataloader': {'batch_size': 16, 'num_workers': 0, 'shuffle': True, 'pin_memory': False, 'persistent_workers': False}, 'val_dataloader': {'batch_size': 16, 'num_workers': 0, 'shuffle': True, 'pin_memory': False, 'persistent_workers': False}, 'optimizer': {'_target_': 'torch.optim.AdamW', 'lr': 0.01, 'betas': [0.95, 0.999], 'eps': 1e-08, 'weight_decay': 1e-06}, 'training': {'device': 'cuda:0', 'seed': 43, 'debug': False, 'resume': False, 'lr_scheduler': 'cosine', 'lr_warmup_steps': 10, 'num_epochs': 20001, 'gradient_accumulate_every': 8, 'grad_norm_clip': 0.5, 'rollout_every': 100, 'checkpoint_every': 100, 'val_every': 1, 'sample_every': 100, 'sample_max_batch': 128, 'max_train_steps': None, 'max_val_steps': None, 'tqdm_interval_sec': 1.0}, 'logging': {'project': 'box-close-v2-training-split-ablation', 'resume': True, 'mode': 'online', 'name': '2026.01.27-11.52.50_train_llmbc_lowdim_box-close-v2', 'tags': ['train_llmbc_lowdim', 'box-close-v2', 'default'], 'id': None, 'group': None}, 'checkpoint': {'topk': {'monitor_key': 'test_success_rate', 'mode': 'max', 'k': 5, 'format_str': 'epoch={epoch:04d}-test_success_rate={test_success_rate:.3f}.ckpt'}, 'save_last_ckpt': True, 'save_last_snapshot': False}, 'multi_run': {'run_dir': 'data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2', 'wandb_name_base': '2026.01.27-11.52.50_train_llmbc_lowdim_box-close-v2'}, 'task': {'name': 'box-close-v2', 'obs_dim': 9, 'action_dim': 4, 'env_runner': {'_target_': 'llmbc.env_runner.metaworld_lowdim_runner.MetaworldLowdimRunner', 'env_name': 'llf-metaworld-box-close-v2', 'n_train': 10, 'n_test': 50, 'n_envs': 10, 'max_steps': 30, 'n_obs_steps': 1, 'n_action_steps': 1, 'instruction_type': 'b', 'feedback_type': ['hp', 'hn', 'fp'], 'visual': False, 'discount': 0.9}, 'dataset': {'_target_': 'llmbc.dataset.metaworld_lowdim_dataset.MetaworldLowdimDataset', 'data_path': 'datasets/box-close-v2-o4-mini-split0.05.pt', 'data_path2': 'datasets/box-close-v2-o4-mini-split0.05.pt', 'horizon': 1, 'pad_before': 0, 'pad_after': 0, 'obs_eef_target': True, 'use_manual_normalizer': False, 'val_ratio': 0.1, 'dummy_normalizer': True}, 'instructor': {'_target_': 'llmbc.translator.instructor.metaworld_instructor.box_close_v2_instructor.BoxCloseV2Instructor'}}, 'llm': {'name': 'HuggingFaceTB/SmolLM2-135M-Instruct', 'model_name': 'SmolLM2-135M-Instruct', 'config_target': 'llmbc.model.llm.llama_lowdim_model.LowdimLlamaConfig', 'causal_lm_target': 'llmbc.model.llm.llama_lowdim_model.LowdimLlamaForCausalLM', 'use_quantization': False, 'use_joint_mlp_projector': True, 'llm_mode': 'ete-finetuned', 'finetune_mode': 'orig', 'checkpoint': 'data/outputs/2026.01.26/10.26.52_train_llm_lowdim_box-close-v2/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-box-close-v2/checkpoint-6000', 'max_length': 100, 'lora_config': {'r': 32, 'lora_alpha': 64, 'lora_dropout': 0.05, 'bias': 'none', 'task_type': 'CAUSAL_LM'}, 'prompter': {'_target_': 'llmbc.translator.prompter.smollm2_prompter.SmolLM2Prompter', 'use_joint_mlp_projector': True}, 'hydra': {'job': {'override_dirname': 'HuggingFaceTB/SmolLM2-135M-Instruct'}, 'run': {'dir': 'data/outputs/2026.01.27/11.52.50_HuggingFaceTB/SmolLM2-135M-Instruct'}}}}
14
+ 2026-01-27 11:53:08,071 INFO MainThread:2419965 [wandb_init.py:init():669] starting backend
15
+ 2026-01-27 11:53:08,071 INFO MainThread:2419965 [wandb_init.py:init():673] sending inform_init request
16
+ 2026-01-27 11:53:08,072 INFO MainThread:2419965 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2026-01-27 11:53:08,073 INFO MainThread:2419965 [wandb_init.py:init():686] backend started and connected
18
+ 2026-01-27 11:53:08,082 INFO MainThread:2419965 [wandb_init.py:init():781] updated telemetry
19
+ 2026-01-27 11:53:08,142 INFO MainThread:2419965 [wandb_init.py:init():814] communicating run to backend with 90.0 second timeout
20
+ 2026-01-27 11:53:08,717 INFO MainThread:2419965 [wandb_init.py:init():859] run resumed
21
+ 2026-01-27 11:53:08,728 INFO MainThread:2419965 [wandb_init.py:init():867] starting run threads in backend
22
+ 2026-01-27 11:53:09,276 INFO MainThread:2419965 [wandb_run.py:_console_start():2451] atexit reg
23
+ 2026-01-27 11:53:09,276 INFO MainThread:2419965 [wandb_run.py:_redirect():2299] redirect: wrap_raw
24
+ 2026-01-27 11:53:09,276 INFO MainThread:2419965 [wandb_run.py:_redirect():2364] Wrapping output streams.
25
+ 2026-01-27 11:53:09,276 INFO MainThread:2419965 [wandb_run.py:_redirect():2389] Redirects installed.
26
+ 2026-01-27 11:53:09,294 INFO MainThread:2419965 [wandb_init.py:init():911] run started, returning control to user process
27
+ 2026-01-27 11:53:09,295 INFO MainThread:2419965 [wandb_run.py:_config_callback():1389] config_cb None None {'output_dir': '/work/u1131674/LLM-BC/data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2'}
28
+ 2026-01-27 22:37:36,572 WARNING MsgRouterThr:2419965 [router.py:message_loop():75] message_loop has been closed
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/files/output.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ Eval MetaworldLowdimRunner 1/6: 0%| | 0/30 [00:00<?, ?it/s]/work/u1131674/LLM-BC/llmbc/common/llfbench_util.py:39: UserWarning: Creating a tensor from a list of numpy.ndarrays is extremely slow. Please consider converting the list to a single numpy.ndarray with numpy.array() before converting to a tensor. (Triggered internally at ../torch/csrc/utils/tensor_new.cpp:275.)
2
+ obs = torch.tensor(obs, dtype=torch.float32).unsqueeze(dim=0).to(device)
3
+
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/files/requirements.txt ADDED
@@ -0,0 +1,858 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ rpds-py==0.27.1
2
+ typeguard==4.4.4
3
+ flatbuffers==25.12.19
4
+ toppra==0.6.3
5
+ sympy==1.14.0
6
+ tiktoken==0.8.0
7
+ nvidia-cuda-cupti-cu12==12.1.105
8
+ arm_pytorch_utilities==0.4.3
9
+ pynndescent==0.6.0
10
+ multidict==6.7.0
11
+ fonttools==4.60.2
12
+ numexpr==2.10.1
13
+ cmudict==1.0.13
14
+ PyOpenGL-accelerate==3.1.10
15
+ gmpy2==2.2.1
16
+ peft==0.14.0
17
+ metaworld==2.0.0
18
+ nvidia-cufft-cu12==11.0.2.54
19
+ python-dateutil==2.9.0.post0
20
+ aiosignal==1.4.0
21
+ pexpect==4.9.0
22
+ protobuf==4.25.8
23
+ typing_extensions==4.15.0
24
+ mujoco==2.3.7
25
+ tokenizers==0.21.0
26
+ pytorch-kinematics==0.7.5
27
+ sniffio==1.3.1
28
+ aiofiles==25.1.0
29
+ mplib==0.1.1
30
+ wcwidth==0.2.14
31
+ Pygments==2.19.1
32
+ anyio==4.12.1
33
+ tensorflow-estimator==2.15.0
34
+ filelock==3.17.0
35
+ numpy==1.23.5
36
+ attrs==25.4.0
37
+ Markdown==3.9
38
+ fsspec==2024.3.1
39
+ libclang==18.1.1
40
+ umap-learn==0.5.9.post2
41
+ dill==0.3.8
42
+ narwhals==2.15.0
43
+ tensorboard==2.15.2
44
+ dacite==1.9.2
45
+ termcolor==3.1.0
46
+ llmbc==0.0.0
47
+ python-multipart==0.0.20
48
+ exceptiongroup==1.3.1
49
+ sapien==3.0.0b1
50
+ pygame==2.6.1
51
+ nvidia-curand-cu12==10.3.2.106
52
+ evaluate==0.4.3
53
+ msgpack==1.1.1
54
+ tensorflow-probability==0.23.0
55
+ diffusers==0.31.0
56
+ certifi==2025.10.5
57
+ d4rl==1.1
58
+ pydub==0.25.1
59
+ annotated-doc==0.0.4
60
+ gitdb==4.0.12
61
+ gradio_client==0.2.9
62
+ Shapely==1.8.4
63
+ mani_skill==3.0.0b20
64
+ tensorflow-io-gcs-filesystem==0.37.1
65
+ fasteners==0.20
66
+ hjson==3.1.0
67
+ ninja==1.13.0
68
+ stack-data==0.6.3
69
+ pyarrow==21.0.0
70
+ networkx==3.2.1
71
+ nvidia-cusparse-cu12==12.1.0.106
72
+ pyparsing==3.3.1
73
+ timm==1.0.22
74
+ typing-inspection==0.4.2
75
+ openai==2.8.1
76
+ pybullet==3.2.6
77
+ hydra-core==1.2.0
78
+ gradio==3.36.1
79
+ tensorflow==2.15.1
80
+ asttokens==3.0.1
81
+ importlib-metadata==5.2.0
82
+ astunparse==1.6.3
83
+ tifffile==2024.8.30
84
+ annotated-types==0.7.0
85
+ Bottleneck==1.4.2
86
+ accelerate==1.0.1
87
+ pytz==2025.2
88
+ urllib3==2.5.0
89
+ frozenlist==1.8.0
90
+ sentry-sdk==2.50.0
91
+ jsonschema==4.25.1
92
+ tyro==0.9.1
93
+ Farama-Notifications==0.0.4
94
+ ffmpy==1.0.0
95
+ httpx==0.28.1
96
+ pymunk==6.2.1
97
+ shtab==1.7.2
98
+ glfw==2.0.0
99
+ hf-xet==1.1.8
100
+ omegaconf==2.2.1
101
+ blobfile==3.0.0
102
+ decorator==5.2.1
103
+ cffi==1.17.1
104
+ matplotlib-inline==0.2.1
105
+ eval_type_backport==0.2.2
106
+ torchaudio==2.2.2
107
+ colorama==0.4.6
108
+ click==8.1.8
109
+ Cython==0.29.37
110
+ orjson==3.11.5
111
+ gym_bandits==0.0.2
112
+ traitlets==5.14.3
113
+ docker-pycreds==0.4.0
114
+ multiprocess==0.70.15
115
+ zipp==3.21.0
116
+ antlr4-python3-runtime==4.9.3
117
+ uc-micro-py==1.0.3
118
+ mpmath==1.3.0
119
+ idna==3.11
120
+ aiodns==3.5.0
121
+ charset-normalizer==3.4.4
122
+ nvidia-nvjitlink-cu12==12.9.86
123
+ nvidia-cuda-nvrtc-cu12==12.1.105
124
+ seaborn==0.13.2
125
+ pyarrow-hotfix==0.7
126
+ pillow==11.3.0
127
+ pyautogen==0.1.0
128
+ requests==2.32.0
129
+ MarkupSafe==3.0.2
130
+ websockets==15.0.1
131
+ nvidia-nccl-cu12==2.19.3
132
+ pure_eval==0.2.3
133
+ parso==0.8.5
134
+ huggingface-hub==0.26.2
135
+ syllables==1.0.9
136
+ tf-agents==0.19.0
137
+ six==1.17.0
138
+ referencing==0.36.2
139
+ ptyprocess==0.7.0
140
+ platformdirs==4.4.0
141
+ fastapi==0.128.0
142
+ stable-baselines3==2.2.1
143
+ av==10.0.0
144
+ diskcache==5.6.3
145
+ pynvml==13.0.1
146
+ pytorch-seed==0.2.0
147
+ zarr==2.12.0
148
+ mdurl==0.1.2
149
+ docstring-parser==0.16
150
+ packaging==25.0
151
+ numcodecs==0.12.1
152
+ opt_einsum==3.4.0
153
+ markdown-it-py==2.2.0
154
+ nvidia-cuda-runtime-cu12==12.1.105
155
+ PyWavelets==1.6.0
156
+ datasets==2.19.0
157
+ contourpy==1.3.0
158
+ aiohappyeyeballs==2.6.1
159
+ jaxlib==0.4.30
160
+ ImageIO==2.37.2
161
+ wandb==0.18.6
162
+ jiter==0.12.0
163
+ gymnasium==0.29.1
164
+ pycryptodomex==3.23.0
165
+ google-pasta==0.2.0
166
+ ipython==8.18.1
167
+ threadpoolctl==3.6.0
168
+ py-cpuinfo==9.0.0
169
+ bitsandbytes==0.45.0
170
+ xxhash==3.5.0
171
+ google-auth-oauthlib==1.2.4
172
+ rsa==4.9.1
173
+ rouge_score==0.1.2
174
+ dm-control==1.0.14
175
+ oauthlib==3.3.1
176
+ pandas==2.3.3
177
+ tenacity==9.1.2
178
+ asciitree==0.3.3
179
+ scipy==1.13.1
180
+ jedi==0.19.2
181
+ gast==0.7.0
182
+ google-auth==2.47.0
183
+ transforms3d==0.4.2
184
+ kiwisolver==1.4.7
185
+ matplotlib==3.7.5
186
+ aiohttp==3.12.15
187
+ pip==23.3.2
188
+ imageio-ffmpeg==0.6.0
189
+ deepspeed==0.16.1
190
+ yarl==1.18.0
191
+ nvidia-nvtx-cu12==12.1.105
192
+ llfbench==0.1.0
193
+ wheel==0.45.1
194
+ PySocks==1.7.1
195
+ ml-dtypes==0.3.2
196
+ PyYAML==6.0.2
197
+ fast_kinematics==0.2.2
198
+ gin-config==0.5.0
199
+ setproctitle==1.3.7
200
+ safetensors==0.5.3
201
+ torchvision==0.17.2
202
+ semantic-version==2.10.0
203
+ PyOpenGL==3.1.10
204
+ nltk==3.9.2
205
+ lxml==6.0.2
206
+ pydantic==2.12.5
207
+ tqdm==4.67.1
208
+ keras==2.15.0
209
+ parse==1.19.1
210
+ linkify-it-py==2.0.3
211
+ dm-tree==0.1.8
212
+ requests-oauthlib==2.0.0
213
+ scikit-learn==1.6.1
214
+ altair==6.0.0
215
+ Werkzeug==3.1.5
216
+ sentencepiece==0.2.0
217
+ uvicorn==0.39.0
218
+ cycler==0.12.1
219
+ transformers==4.47.1
220
+ uvloop==0.22.1
221
+ mkl_random==1.2.8
222
+ GitPython==3.1.46
223
+ regex==2025.9.1
224
+ jax==0.4.30
225
+ llvmlite==0.39.1
226
+ pyasn1_modules==0.4.2
227
+ nvidia-cudnn-cu12==8.9.2.26
228
+ pydantic_core==2.41.5
229
+ google-genai==1.47.0
230
+ propcache==0.3.1
231
+ pycares==4.10.0
232
+ pyperclip==1.11.0
233
+ pyasn1==0.6.2
234
+ async-timeout==5.0.1
235
+ psutil==7.0.0
236
+ gym==0.23.1
237
+ dm-env==1.6
238
+ Jinja2==3.1.6
239
+ sentence-transformers==3.2.1
240
+ einops==0.4.1
241
+ triton==2.2.0
242
+ grpcio==1.76.0
243
+ labmaze==1.0.6
244
+ nvidia-ml-py==13.590.44
245
+ brotlicffi==1.0.9.2
246
+ smmap==5.0.2
247
+ cloudpickle==3.1.2
248
+ setuptools==80.9.0
249
+ starlette==0.49.3
250
+ prompt_toolkit==3.0.52
251
+ wrapt==1.14.2
252
+ h5py==3.14.0
253
+ scikit-image==0.19.3
254
+ joblib==1.5.3
255
+ opencv-python==4.11.0.86
256
+ rich==14.2.0
257
+ trl==0.11.4
258
+ gym-notices==0.1.0
259
+ trimesh==4.11.1
260
+ mdit-py-plugins==0.3.3
261
+ distro==1.9.0
262
+ executing==2.2.1
263
+ mkl-service==2.4.0
264
+ nvidia-cusolver-cu12==11.4.5.107
265
+ FLAML==2.3.6
266
+ mujoco-py==2.1.2.14
267
+ h11==0.16.0
268
+ highway-env==1.9.1
269
+ httpcore==1.0.9
270
+ tensorboard-data-server==0.7.2
271
+ tzdata==2025.3
272
+ absl-py==2.3.1
273
+ jsonschema-specifications==2025.9.1
274
+ numba==0.56.4
275
+ tabulate==0.9.0
276
+ importlib-resources==5.13.0
277
+ pycparser==2.23
278
+ mkl_fft==1.3.11
279
+ torch==2.2.2
280
+ nvidia-cublas-cu12==12.1.3.1
281
+ rpds-py==0.27.1
282
+ typeguard==4.4.4
283
+ flatbuffers==25.12.19
284
+ toppra==0.6.3
285
+ sympy==1.14.0
286
+ tiktoken==0.8.0
287
+ nvidia-cuda-cupti-cu12==12.1.105
288
+ arm_pytorch_utilities==0.4.3
289
+ pynndescent==0.6.0
290
+ multidict==6.7.0
291
+ fonttools==4.60.2
292
+ numexpr==2.10.1
293
+ cmudict==1.0.13
294
+ PyOpenGL-accelerate==3.1.10
295
+ gmpy2==2.2.1
296
+ peft==0.14.0
297
+ metaworld==2.0.0
298
+ nvidia-cufft-cu12==11.0.2.54
299
+ python-dateutil==2.9.0.post0
300
+ aiosignal==1.4.0
301
+ pexpect==4.9.0
302
+ protobuf==4.25.8
303
+ typing_extensions==4.15.0
304
+ mujoco==2.3.7
305
+ tokenizers==0.21.0
306
+ pytorch-kinematics==0.7.5
307
+ sniffio==1.3.1
308
+ aiofiles==25.1.0
309
+ mplib==0.1.1
310
+ wcwidth==0.2.14
311
+ Pygments==2.19.1
312
+ anyio==4.12.1
313
+ tensorflow-estimator==2.15.0
314
+ filelock==3.17.0
315
+ numpy==1.23.5
316
+ attrs==25.4.0
317
+ Markdown==3.9
318
+ fsspec==2024.3.1
319
+ libclang==18.1.1
320
+ umap-learn==0.5.9.post2
321
+ dill==0.3.8
322
+ narwhals==2.15.0
323
+ tensorboard==2.15.2
324
+ dacite==1.9.2
325
+ termcolor==3.1.0
326
+ llmbc==0.0.0
327
+ python-multipart==0.0.20
328
+ exceptiongroup==1.3.1
329
+ sapien==3.0.0b1
330
+ pygame==2.6.1
331
+ nvidia-curand-cu12==10.3.2.106
332
+ evaluate==0.4.3
333
+ msgpack==1.1.1
334
+ tensorflow-probability==0.23.0
335
+ diffusers==0.31.0
336
+ certifi==2025.10.5
337
+ d4rl==1.1
338
+ pydub==0.25.1
339
+ annotated-doc==0.0.4
340
+ gitdb==4.0.12
341
+ gradio_client==0.2.9
342
+ Shapely==1.8.4
343
+ mani_skill==3.0.0b20
344
+ tensorflow-io-gcs-filesystem==0.37.1
345
+ fasteners==0.20
346
+ hjson==3.1.0
347
+ ninja==1.13.0
348
+ stack-data==0.6.3
349
+ pyarrow==21.0.0
350
+ networkx==3.2.1
351
+ nvidia-cusparse-cu12==12.1.0.106
352
+ pyparsing==3.3.1
353
+ timm==1.0.22
354
+ typing-inspection==0.4.2
355
+ openai==2.8.1
356
+ pybullet==3.2.6
357
+ hydra-core==1.2.0
358
+ gradio==3.36.1
359
+ tensorflow==2.15.1
360
+ asttokens==3.0.1
361
+ importlib-metadata==5.2.0
362
+ astunparse==1.6.3
363
+ tifffile==2024.8.30
364
+ annotated-types==0.7.0
365
+ Bottleneck==1.4.2
366
+ accelerate==1.0.1
367
+ pytz==2025.2
368
+ urllib3==2.5.0
369
+ frozenlist==1.8.0
370
+ sentry-sdk==2.50.0
371
+ jsonschema==4.25.1
372
+ tyro==0.9.1
373
+ Farama-Notifications==0.0.4
374
+ ffmpy==1.0.0
375
+ httpx==0.28.1
376
+ pymunk==6.2.1
377
+ shtab==1.7.2
378
+ glfw==2.0.0
379
+ hf-xet==1.1.8
380
+ omegaconf==2.2.1
381
+ blobfile==3.0.0
382
+ decorator==5.2.1
383
+ cffi==1.17.1
384
+ matplotlib-inline==0.2.1
385
+ eval_type_backport==0.2.2
386
+ torchaudio==2.2.2
387
+ colorama==0.4.6
388
+ click==8.1.8
389
+ Cython==0.29.37
390
+ orjson==3.11.5
391
+ gym_bandits==0.0.2
392
+ traitlets==5.14.3
393
+ docker-pycreds==0.4.0
394
+ multiprocess==0.70.15
395
+ zipp==3.21.0
396
+ antlr4-python3-runtime==4.9.3
397
+ uc-micro-py==1.0.3
398
+ mpmath==1.3.0
399
+ idna==3.11
400
+ aiodns==3.5.0
401
+ charset-normalizer==3.4.4
402
+ nvidia-nvjitlink-cu12==12.9.86
403
+ nvidia-cuda-nvrtc-cu12==12.1.105
404
+ seaborn==0.13.2
405
+ pyarrow-hotfix==0.7
406
+ pillow==11.3.0
407
+ pyautogen==0.1.0
408
+ requests==2.32.0
409
+ MarkupSafe==3.0.2
410
+ websockets==15.0.1
411
+ nvidia-nccl-cu12==2.19.3
412
+ pure_eval==0.2.3
413
+ parso==0.8.5
414
+ huggingface-hub==0.26.2
415
+ syllables==1.0.9
416
+ tf-agents==0.19.0
417
+ six==1.17.0
418
+ referencing==0.36.2
419
+ ptyprocess==0.7.0
420
+ platformdirs==4.4.0
421
+ fastapi==0.128.0
422
+ stable-baselines3==2.2.1
423
+ av==10.0.0
424
+ diskcache==5.6.3
425
+ pynvml==13.0.1
426
+ pytorch-seed==0.2.0
427
+ zarr==2.12.0
428
+ mdurl==0.1.2
429
+ docstring-parser==0.16
430
+ packaging==25.0
431
+ numcodecs==0.12.1
432
+ opt_einsum==3.4.0
433
+ markdown-it-py==2.2.0
434
+ nvidia-cuda-runtime-cu12==12.1.105
435
+ PyWavelets==1.6.0
436
+ datasets==2.19.0
437
+ contourpy==1.3.0
438
+ aiohappyeyeballs==2.6.1
439
+ jaxlib==0.4.30
440
+ ImageIO==2.37.2
441
+ wandb==0.18.6
442
+ jiter==0.12.0
443
+ gymnasium==0.29.1
444
+ pycryptodomex==3.23.0
445
+ google-pasta==0.2.0
446
+ ipython==8.18.1
447
+ threadpoolctl==3.6.0
448
+ py-cpuinfo==9.0.0
449
+ bitsandbytes==0.45.0
450
+ xxhash==3.5.0
451
+ google-auth-oauthlib==1.2.4
452
+ rsa==4.9.1
453
+ rouge_score==0.1.2
454
+ dm-control==1.0.14
455
+ oauthlib==3.3.1
456
+ pandas==2.3.3
457
+ tenacity==9.1.2
458
+ asciitree==0.3.3
459
+ scipy==1.13.1
460
+ jedi==0.19.2
461
+ gast==0.7.0
462
+ google-auth==2.47.0
463
+ transforms3d==0.4.2
464
+ kiwisolver==1.4.7
465
+ matplotlib==3.7.5
466
+ aiohttp==3.12.15
467
+ pip==23.3.2
468
+ imageio-ffmpeg==0.6.0
469
+ deepspeed==0.16.1
470
+ yarl==1.18.0
471
+ nvidia-nvtx-cu12==12.1.105
472
+ llfbench==0.1.0
473
+ wheel==0.45.1
474
+ PySocks==1.7.1
475
+ ml-dtypes==0.3.2
476
+ PyYAML==6.0.2
477
+ fast_kinematics==0.2.2
478
+ gin-config==0.5.0
479
+ setproctitle==1.3.7
480
+ safetensors==0.5.3
481
+ torchvision==0.17.2
482
+ semantic-version==2.10.0
483
+ PyOpenGL==3.1.10
484
+ nltk==3.9.2
485
+ lxml==6.0.2
486
+ pydantic==2.12.5
487
+ tqdm==4.67.1
488
+ keras==2.15.0
489
+ parse==1.19.1
490
+ linkify-it-py==2.0.3
491
+ dm-tree==0.1.8
492
+ requests-oauthlib==2.0.0
493
+ scikit-learn==1.6.1
494
+ altair==6.0.0
495
+ Werkzeug==3.1.5
496
+ sentencepiece==0.2.0
497
+ uvicorn==0.39.0
498
+ cycler==0.12.1
499
+ transformers==4.47.1
500
+ uvloop==0.22.1
501
+ mkl_random==1.2.8
502
+ GitPython==3.1.46
503
+ regex==2025.9.1
504
+ jax==0.4.30
505
+ llvmlite==0.39.1
506
+ pyasn1_modules==0.4.2
507
+ nvidia-cudnn-cu12==8.9.2.26
508
+ pydantic_core==2.41.5
509
+ google-genai==1.47.0
510
+ propcache==0.3.1
511
+ pycares==4.10.0
512
+ pyperclip==1.11.0
513
+ pyasn1==0.6.2
514
+ async-timeout==5.0.1
515
+ psutil==7.0.0
516
+ gym==0.23.1
517
+ dm-env==1.6
518
+ Jinja2==3.1.6
519
+ sentence-transformers==3.2.1
520
+ einops==0.4.1
521
+ triton==2.2.0
522
+ grpcio==1.76.0
523
+ labmaze==1.0.6
524
+ nvidia-ml-py==13.590.44
525
+ brotlicffi==1.0.9.2
526
+ smmap==5.0.2
527
+ cloudpickle==3.1.2
528
+ setuptools==80.9.0
529
+ starlette==0.49.3
530
+ prompt_toolkit==3.0.52
531
+ wrapt==1.14.2
532
+ h5py==3.14.0
533
+ scikit-image==0.19.3
534
+ joblib==1.5.3
535
+ opencv-python==4.11.0.86
536
+ rich==14.2.0
537
+ trl==0.11.4
538
+ gym-notices==0.1.0
539
+ trimesh==4.11.1
540
+ mdit-py-plugins==0.3.3
541
+ distro==1.9.0
542
+ executing==2.2.1
543
+ mkl-service==2.4.0
544
+ nvidia-cusolver-cu12==11.4.5.107
545
+ FLAML==2.3.6
546
+ mujoco-py==2.1.2.14
547
+ h11==0.16.0
548
+ highway-env==1.9.1
549
+ httpcore==1.0.9
550
+ tensorboard-data-server==0.7.2
551
+ tzdata==2025.3
552
+ absl-py==2.3.1
553
+ jsonschema-specifications==2025.9.1
554
+ numba==0.56.4
555
+ tabulate==0.9.0
556
+ importlib-resources==5.13.0
557
+ pycparser==2.23
558
+ mkl_fft==1.3.11
559
+ torch==2.2.2
560
+ nvidia-cublas-cu12==12.1.3.1
561
+ llmbc==0.0.0
562
+ llmbc==0.0.0
563
+ rpds-py==0.27.1
564
+ typeguard==4.4.4
565
+ flatbuffers==25.12.19
566
+ toppra==0.6.3
567
+ sympy==1.14.0
568
+ tiktoken==0.8.0
569
+ nvidia-cuda-cupti-cu12==12.1.105
570
+ arm_pytorch_utilities==0.4.3
571
+ pynndescent==0.6.0
572
+ multidict==6.7.0
573
+ fonttools==4.60.2
574
+ numexpr==2.10.1
575
+ cmudict==1.0.13
576
+ PyOpenGL-accelerate==3.1.10
577
+ gmpy2==2.2.1
578
+ peft==0.14.0
579
+ metaworld==2.0.0
580
+ nvidia-cufft-cu12==11.0.2.54
581
+ python-dateutil==2.9.0.post0
582
+ aiosignal==1.4.0
583
+ pexpect==4.9.0
584
+ protobuf==4.25.8
585
+ typing_extensions==4.15.0
586
+ mujoco==2.3.7
587
+ tokenizers==0.21.0
588
+ pytorch-kinematics==0.7.5
589
+ sniffio==1.3.1
590
+ aiofiles==25.1.0
591
+ mplib==0.1.1
592
+ wcwidth==0.2.14
593
+ Pygments==2.19.1
594
+ anyio==4.12.1
595
+ tensorflow-estimator==2.15.0
596
+ filelock==3.17.0
597
+ numpy==1.23.5
598
+ attrs==25.4.0
599
+ Markdown==3.9
600
+ fsspec==2024.3.1
601
+ libclang==18.1.1
602
+ umap-learn==0.5.9.post2
603
+ dill==0.3.8
604
+ narwhals==2.15.0
605
+ tensorboard==2.15.2
606
+ dacite==1.9.2
607
+ termcolor==3.1.0
608
+ llmbc==0.0.0
609
+ python-multipart==0.0.20
610
+ exceptiongroup==1.3.1
611
+ sapien==3.0.0b1
612
+ pygame==2.6.1
613
+ nvidia-curand-cu12==10.3.2.106
614
+ evaluate==0.4.3
615
+ msgpack==1.1.1
616
+ tensorflow-probability==0.23.0
617
+ diffusers==0.31.0
618
+ certifi==2025.10.5
619
+ d4rl==1.1
620
+ pydub==0.25.1
621
+ annotated-doc==0.0.4
622
+ gitdb==4.0.12
623
+ gradio_client==0.2.9
624
+ Shapely==1.8.4
625
+ mani_skill==3.0.0b20
626
+ tensorflow-io-gcs-filesystem==0.37.1
627
+ fasteners==0.20
628
+ hjson==3.1.0
629
+ ninja==1.13.0
630
+ stack-data==0.6.3
631
+ pyarrow==21.0.0
632
+ networkx==3.2.1
633
+ nvidia-cusparse-cu12==12.1.0.106
634
+ pyparsing==3.3.1
635
+ timm==1.0.22
636
+ typing-inspection==0.4.2
637
+ openai==2.8.1
638
+ pybullet==3.2.6
639
+ hydra-core==1.2.0
640
+ gradio==3.36.1
641
+ tensorflow==2.15.1
642
+ asttokens==3.0.1
643
+ importlib-metadata==5.2.0
644
+ astunparse==1.6.3
645
+ tifffile==2024.8.30
646
+ annotated-types==0.7.0
647
+ Bottleneck==1.4.2
648
+ accelerate==1.0.1
649
+ pytz==2025.2
650
+ urllib3==2.5.0
651
+ frozenlist==1.8.0
652
+ sentry-sdk==2.50.0
653
+ jsonschema==4.25.1
654
+ tyro==0.9.1
655
+ Farama-Notifications==0.0.4
656
+ ffmpy==1.0.0
657
+ httpx==0.28.1
658
+ pymunk==6.2.1
659
+ shtab==1.7.2
660
+ glfw==2.0.0
661
+ hf-xet==1.1.8
662
+ omegaconf==2.2.1
663
+ blobfile==3.0.0
664
+ decorator==5.2.1
665
+ cffi==1.17.1
666
+ matplotlib-inline==0.2.1
667
+ eval_type_backport==0.2.2
668
+ torchaudio==2.2.2
669
+ colorama==0.4.6
670
+ click==8.1.8
671
+ Cython==0.29.37
672
+ orjson==3.11.5
673
+ gym_bandits==0.0.2
674
+ traitlets==5.14.3
675
+ docker-pycreds==0.4.0
676
+ multiprocess==0.70.15
677
+ zipp==3.21.0
678
+ antlr4-python3-runtime==4.9.3
679
+ uc-micro-py==1.0.3
680
+ mpmath==1.3.0
681
+ idna==3.11
682
+ aiodns==3.5.0
683
+ charset-normalizer==3.4.4
684
+ nvidia-nvjitlink-cu12==12.9.86
685
+ nvidia-cuda-nvrtc-cu12==12.1.105
686
+ seaborn==0.13.2
687
+ pyarrow-hotfix==0.7
688
+ pillow==11.3.0
689
+ pyautogen==0.1.0
690
+ requests==2.32.0
691
+ MarkupSafe==3.0.2
692
+ websockets==15.0.1
693
+ nvidia-nccl-cu12==2.19.3
694
+ pure_eval==0.2.3
695
+ parso==0.8.5
696
+ huggingface-hub==0.26.2
697
+ syllables==1.0.9
698
+ tf-agents==0.19.0
699
+ six==1.17.0
700
+ referencing==0.36.2
701
+ ptyprocess==0.7.0
702
+ platformdirs==4.4.0
703
+ fastapi==0.128.0
704
+ stable-baselines3==2.2.1
705
+ av==10.0.0
706
+ diskcache==5.6.3
707
+ pynvml==13.0.1
708
+ pytorch-seed==0.2.0
709
+ zarr==2.12.0
710
+ mdurl==0.1.2
711
+ docstring-parser==0.16
712
+ packaging==25.0
713
+ numcodecs==0.12.1
714
+ opt_einsum==3.4.0
715
+ markdown-it-py==2.2.0
716
+ nvidia-cuda-runtime-cu12==12.1.105
717
+ PyWavelets==1.6.0
718
+ datasets==2.19.0
719
+ contourpy==1.3.0
720
+ aiohappyeyeballs==2.6.1
721
+ jaxlib==0.4.30
722
+ ImageIO==2.37.2
723
+ wandb==0.18.6
724
+ jiter==0.12.0
725
+ gymnasium==0.29.1
726
+ pycryptodomex==3.23.0
727
+ google-pasta==0.2.0
728
+ ipython==8.18.1
729
+ threadpoolctl==3.6.0
730
+ py-cpuinfo==9.0.0
731
+ bitsandbytes==0.45.0
732
+ xxhash==3.5.0
733
+ google-auth-oauthlib==1.2.4
734
+ rsa==4.9.1
735
+ rouge_score==0.1.2
736
+ dm-control==1.0.14
737
+ oauthlib==3.3.1
738
+ pandas==2.3.3
739
+ tenacity==9.1.2
740
+ asciitree==0.3.3
741
+ scipy==1.13.1
742
+ jedi==0.19.2
743
+ gast==0.7.0
744
+ google-auth==2.47.0
745
+ transforms3d==0.4.2
746
+ kiwisolver==1.4.7
747
+ matplotlib==3.7.5
748
+ aiohttp==3.12.15
749
+ pip==23.3.2
750
+ imageio-ffmpeg==0.6.0
751
+ deepspeed==0.16.1
752
+ yarl==1.18.0
753
+ nvidia-nvtx-cu12==12.1.105
754
+ llfbench==0.1.0
755
+ wheel==0.45.1
756
+ PySocks==1.7.1
757
+ ml-dtypes==0.3.2
758
+ PyYAML==6.0.2
759
+ fast_kinematics==0.2.2
760
+ gin-config==0.5.0
761
+ setproctitle==1.3.7
762
+ safetensors==0.5.3
763
+ torchvision==0.17.2
764
+ semantic-version==2.10.0
765
+ PyOpenGL==3.1.10
766
+ nltk==3.9.2
767
+ lxml==6.0.2
768
+ pydantic==2.12.5
769
+ tqdm==4.67.1
770
+ keras==2.15.0
771
+ parse==1.19.1
772
+ linkify-it-py==2.0.3
773
+ dm-tree==0.1.8
774
+ requests-oauthlib==2.0.0
775
+ scikit-learn==1.6.1
776
+ altair==6.0.0
777
+ Werkzeug==3.1.5
778
+ sentencepiece==0.2.0
779
+ uvicorn==0.39.0
780
+ cycler==0.12.1
781
+ transformers==4.47.1
782
+ uvloop==0.22.1
783
+ mkl_random==1.2.8
784
+ GitPython==3.1.46
785
+ regex==2025.9.1
786
+ jax==0.4.30
787
+ llvmlite==0.39.1
788
+ pyasn1_modules==0.4.2
789
+ nvidia-cudnn-cu12==8.9.2.26
790
+ pydantic_core==2.41.5
791
+ google-genai==1.47.0
792
+ propcache==0.3.1
793
+ pycares==4.10.0
794
+ pyperclip==1.11.0
795
+ pyasn1==0.6.2
796
+ async-timeout==5.0.1
797
+ psutil==7.0.0
798
+ gym==0.23.1
799
+ dm-env==1.6
800
+ Jinja2==3.1.6
801
+ sentence-transformers==3.2.1
802
+ einops==0.4.1
803
+ triton==2.2.0
804
+ grpcio==1.76.0
805
+ labmaze==1.0.6
806
+ nvidia-ml-py==13.590.44
807
+ brotlicffi==1.0.9.2
808
+ smmap==5.0.2
809
+ cloudpickle==3.1.2
810
+ setuptools==80.9.0
811
+ starlette==0.49.3
812
+ prompt_toolkit==3.0.52
813
+ wrapt==1.14.2
814
+ h5py==3.14.0
815
+ scikit-image==0.19.3
816
+ joblib==1.5.3
817
+ opencv-python==4.11.0.86
818
+ rich==14.2.0
819
+ trl==0.11.4
820
+ gym-notices==0.1.0
821
+ trimesh==4.11.1
822
+ mdit-py-plugins==0.3.3
823
+ distro==1.9.0
824
+ executing==2.2.1
825
+ mkl-service==2.4.0
826
+ nvidia-cusolver-cu12==11.4.5.107
827
+ FLAML==2.3.6
828
+ mujoco-py==2.1.2.14
829
+ h11==0.16.0
830
+ highway-env==1.9.1
831
+ httpcore==1.0.9
832
+ tensorboard-data-server==0.7.2
833
+ tzdata==2025.3
834
+ absl-py==2.3.1
835
+ jsonschema-specifications==2025.9.1
836
+ numba==0.56.4
837
+ tabulate==0.9.0
838
+ importlib-resources==5.13.0
839
+ pycparser==2.23
840
+ mkl_fft==1.3.11
841
+ torch==2.2.2
842
+ nvidia-cublas-cu12==12.1.3.1
843
+ zipp==3.19.2
844
+ jaraco.text==3.12.1
845
+ jaraco.context==5.3.0
846
+ importlib_metadata==8.0.0
847
+ typeguard==4.3.0
848
+ inflect==7.3.1
849
+ more-itertools==10.3.0
850
+ wheel==0.45.1
851
+ packaging==24.2
852
+ backports.tarfile==1.2.0
853
+ autocommand==2.2.2
854
+ jaraco.collections==5.1.0
855
+ tomli==2.0.1
856
+ platformdirs==4.2.2
857
+ jaraco.functools==4.0.1
858
+ typing_extensions==4.12.2
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/files/wandb-metadata.json ADDED
@@ -0,0 +1,107 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-4.18.0-513.24.1.el8_9.x86_64-x86_64-with-glibc2.28",
3
+ "python": "3.9.25",
4
+ "startedAt": "2026-01-27T03:53:07.506395Z",
5
+ "args": [
6
+ "--config-path",
7
+ "config/data_efficiency",
8
+ "--config-name",
9
+ "llmbc_box-close-v2_o4-mini_split0.05.yaml",
10
+ "training.seed=42"
11
+ ],
12
+ "program": "/work/u1131674/LLM-BC/./train.py",
13
+ "codePath": "train.py",
14
+ "git": {
15
+ "remote": "https://github.com/CHYang25/LLM-BC.git",
16
+ "commit": "206b9c57fa687707dfa1ee0ebb33b507135a59c6"
17
+ },
18
+ "email": "chris920325@gmail.com",
19
+ "root": "/work/u1131674/LLM-BC/data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2",
20
+ "host": "hgpn05",
21
+ "username": "u1131674",
22
+ "executable": "/home/u1131674/.conda/envs/llm-bc/bin/python3",
23
+ "codePathLocal": "train.py",
24
+ "cpu_count": 112,
25
+ "cpu_count_logical": 112,
26
+ "gpu": "NVIDIA H100 80GB HBM3",
27
+ "gpu_count": 1,
28
+ "disk": {
29
+ "/": {
30
+ "total": "1918024196096",
31
+ "used": "437322797056"
32
+ }
33
+ },
34
+ "memory": {
35
+ "total": "2163617771520"
36
+ },
37
+ "cpu": {
38
+ "count": 112,
39
+ "countLogical": 112
40
+ },
41
+ "gpu_nvidia": [
42
+ {
43
+ "name": "NVIDIA H100 80GB HBM3",
44
+ "memoryTotal": "85520809984",
45
+ "cudaCores": 16896,
46
+ "architecture": "Hopper"
47
+ }
48
+ ],
49
+ "slurm": {
50
+ "cluster_name": "hpc",
51
+ "conf": "/etc/slurm/slurm.conf",
52
+ "cpu_bind": "quiet,mask_cpu:0x000000000FE04000000000000000",
53
+ "cpu_bind_list": "0x000000000FE04000000000000000",
54
+ "cpu_bind_type": "mask_cpu:",
55
+ "cpu_bind_verbose": "quiet",
56
+ "cpus_on_node": "8",
57
+ "cpus_per_task": "8",
58
+ "distribution": "cyclic,pack",
59
+ "gpus_on_node": "1",
60
+ "gpus_per_node": "1",
61
+ "gtids": "0",
62
+ "job_account": "mst114558",
63
+ "job_cpus_per_node": "8",
64
+ "job_end_time": "1769658768",
65
+ "job_gid": "106773",
66
+ "job_group": "MST114558",
67
+ "job_id": "104242",
68
+ "job_name": "python3",
69
+ "job_nodelist": "hgpn05",
70
+ "job_num_nodes": "1",
71
+ "job_partition": "normal",
72
+ "job_qos": "normal",
73
+ "job_start_time": "1769485968",
74
+ "job_uid": "41408",
75
+ "job_user": "u1131674",
76
+ "jobid": "104242",
77
+ "launch_node_ipaddr": "172.21.101.1",
78
+ "localid": "0",
79
+ "mem_per_node": "204800",
80
+ "nnodes": "1",
81
+ "nodeid": "0",
82
+ "nodelist": "hgpn05",
83
+ "nprocs": "1",
84
+ "ntasks": "1",
85
+ "prio_process": "0",
86
+ "procid": "0",
87
+ "srun_comm_host": "172.21.101.1",
88
+ "srun_comm_port": "42069",
89
+ "step_gpus": "5",
90
+ "step_id": "0",
91
+ "step_launcher_port": "42069",
92
+ "step_nodelist": "hgpn05",
93
+ "step_num_nodes": "1",
94
+ "step_num_tasks": "1",
95
+ "step_tasks_per_node": "1",
96
+ "stepid": "0",
97
+ "submit_dir": "/work/u1131674/LLM-BC",
98
+ "submit_host": "cbi-lgn01",
99
+ "task_pid": "754949",
100
+ "tasks_per_node": "1",
101
+ "topology_addr": "ibsw1.hgpn05",
102
+ "topology_addr_pattern": "switch.node",
103
+ "tres_per_task": "cpu:8",
104
+ "umask": "0022"
105
+ },
106
+ "cudaVersion": "12.4"
107
+ }
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/logs/debug-core.log ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {"time":"2026-01-27T11:53:06.897507519+08:00","level":"INFO","msg":"started logging, with flags","port-filename":"/tmp/tmpi1osecmd/port-754949.txt","pid":754949,"debug":false,"disable-analytics":false}
2
+ {"time":"2026-01-27T11:53:06.897527345+08:00","level":"INFO","msg":"FeatureState","shutdownOnParentExitEnabled":false}
3
+ {"time":"2026-01-27T11:53:06.897861154+08:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":754949}
4
+ {"time":"2026-01-27T11:53:06.897850599+08:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":32979,"Zone":""}}
5
+ {"time":"2026-01-27T11:53:07.091367358+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:53454"}
6
+ {"time":"2026-01-27T11:53:07.506927211+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"raachjxu","id":"127.0.0.1:53454"}
7
+ {"time":"2026-01-27T11:53:07.624448731+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"raachjxu","id":"127.0.0.1:53454"}
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/logs/debug-internal.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-01-27T11:53:07.507948839+08:00","level":"INFO","msg":"using version","core version":"0.18.6"}
2
+ {"time":"2026-01-27T11:53:07.507957394+08:00","level":"INFO","msg":"created symlink","path":"/work/u1131674/LLM-BC/data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/logs/debug-core.log"}
3
+ {"time":"2026-01-27T11:53:07.624409739+08:00","level":"INFO","msg":"created new stream","id":"raachjxu"}
4
+ {"time":"2026-01-27T11:53:07.624443283+08:00","level":"INFO","msg":"stream: started","id":"raachjxu"}
5
+ {"time":"2026-01-27T11:53:07.624479516+08:00","level":"INFO","msg":"sender: started","stream_id":"raachjxu"}
6
+ {"time":"2026-01-27T11:53:07.624459006+08:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"raachjxu"}}
7
+ {"time":"2026-01-27T11:53:07.624475123+08:00","level":"INFO","msg":"handler: started","stream_id":{"value":"raachjxu"}}
8
+ {"time":"2026-01-27T11:53:08.339033565+08:00","level":"INFO","msg":"Starting system monitor"}
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/logs/debug.log ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-01-27 11:53:07,503 INFO MainThread:754949 [wandb_setup.py:_flush():79] Current SDK version is 0.18.6
2
+ 2026-01-27 11:53:07,504 INFO MainThread:754949 [wandb_setup.py:_flush():79] Configure stats pid to 754949
3
+ 2026-01-27 11:53:07,504 INFO MainThread:754949 [wandb_setup.py:_flush():79] Loading settings from /home/u1131674/.config/wandb/settings
4
+ 2026-01-27 11:53:07,504 INFO MainThread:754949 [wandb_setup.py:_flush():79] Loading settings from /work/u1131674/LLM-BC/wandb/settings
5
+ 2026-01-27 11:53:07,504 INFO MainThread:754949 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2026-01-27 11:53:07,504 INFO MainThread:754949 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': 'online', '_disable_service': None}
7
+ 2026-01-27 11:53:07,504 INFO MainThread:754949 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train.py', 'program_abspath': '/work/u1131674/LLM-BC/train.py', 'program': '/work/u1131674/LLM-BC/./train.py'}
8
+ 2026-01-27 11:53:07,504 INFO MainThread:754949 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2026-01-27 11:53:07,504 INFO MainThread:754949 [wandb_init.py:_log_setup():533] Logging user logs to /work/u1131674/LLM-BC/data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/logs/debug.log
10
+ 2026-01-27 11:53:07,504 INFO MainThread:754949 [wandb_init.py:_log_setup():534] Logging internal logs to /work/u1131674/LLM-BC/data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/logs/debug-internal.log
11
+ 2026-01-27 11:53:07,504 INFO MainThread:754949 [wandb_init.py:init():619] calling init triggers
12
+ 2026-01-27 11:53:07,504 INFO MainThread:754949 [wandb_init.py:init():626] wandb.init called with sweep_config: {}
13
+ config: {'name': 'train_llmbc_lowdim', '_target_': 'llmbc.workspace.train_llmbc_lowdim_workspace.TrainLLMBCLowdimWorkspace', 'obs_dim': 9, 'action_dim': 4, 'task_name': 'box-close-v2', 'exp_name': 'default', 'model_name': 'HuggingFaceTB/SmolLM2-135M-Instruct', 'horizon': 1, 'n_obs_steps': 1, 'n_action_steps': 1, 'n_latency_steps': 0, 'past_action_visible': False, 'llm_orig_expert_feedback': True, 'llm_do_sample': False, 'policy': {'_target_': 'llmbc.policy.llmbc_lowdim_policy.LLMBCLowdimPolicy', 'model': {'_target_': 'llmbc.model.policy.policy_mlp.PolicyMLP', 'input_size': 9, 'hidden_size': [256, 256], 'output_size': 4, 'activation': 'relu', 'n_obs_steps': 1, 'n_action_steps': 1}, 'obs_dim': 9, 'action_dim': 4, 'llm_discriminator': {'_target_': 'llmbc.discriminator.llm_ce_discriminator.LLMCEDiscriminator', 'task_id': 'box-close-v2', 'llm_translator': {'_target_': 'llmbc.translator.llm_translator.LLMTranslator', 'cfg': {'name': 'HuggingFaceTB/SmolLM2-135M-Instruct', 'model_name': 'SmolLM2-135M-Instruct', 'config_target': 'llmbc.model.llm.llama_lowdim_model.LowdimLlamaConfig', 'causal_lm_target': 'llmbc.model.llm.llama_lowdim_model.LowdimLlamaForCausalLM', 'use_quantization': False, 'use_joint_mlp_projector': True, 'llm_mode': 'ete-finetuned', 'finetune_mode': 'orig', 'checkpoint': 'data/outputs/2026.01.26/10.26.52_train_llm_lowdim_box-close-v2/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-box-close-v2/checkpoint-6000', 'max_length': 100, 'lora_config': {'r': 32, 'lora_alpha': 64, 'lora_dropout': 0.05, 'bias': 'none', 'task_type': 'CAUSAL_LM'}, 'prompter': {'_target_': 'llmbc.translator.prompter.smollm2_prompter.SmolLM2Prompter', 'use_joint_mlp_projector': True}, 'hydra': {'job': {'override_dirname': 'HuggingFaceTB/SmolLM2-135M-Instruct'}, 'run': {'dir': 'data/outputs/2026.01.27/11.52.50_HuggingFaceTB/SmolLM2-135M-Instruct'}}}, 'obs_dim': 9, 'action_dim': 4, 'horizon': 1, 'n_obs_steps': 1, 'n_action_steps': 1}}, 'loss_bc_weight': 1.0, 'loss_llm_weight': 0.001, 'horizon': 1, 'n_obs_steps': 1, 'n_action_steps': 1, 'normalize_llm_loss': True}, 'dataloader': {'batch_size': 16, 'num_workers': 0, 'shuffle': True, 'pin_memory': False, 'persistent_workers': False}, 'val_dataloader': {'batch_size': 16, 'num_workers': 0, 'shuffle': True, 'pin_memory': False, 'persistent_workers': False}, 'optimizer': {'_target_': 'torch.optim.AdamW', 'lr': 0.01, 'betas': [0.95, 0.999], 'eps': 1e-08, 'weight_decay': 1e-06}, 'training': {'device': 'cuda:0', 'seed': 42, 'debug': False, 'resume': False, 'lr_scheduler': 'cosine', 'lr_warmup_steps': 10, 'num_epochs': 20001, 'gradient_accumulate_every': 8, 'grad_norm_clip': 0.5, 'rollout_every': 100, 'checkpoint_every': 100, 'val_every': 1, 'sample_every': 100, 'sample_max_batch': 128, 'max_train_steps': None, 'max_val_steps': None, 'tqdm_interval_sec': 1.0}, 'logging': {'project': 'box-close-v2-training-split-ablation', 'resume': True, 'mode': 'online', 'name': '2026.01.27-11.52.50_train_llmbc_lowdim_box-close-v2', 'tags': ['train_llmbc_lowdim', 'box-close-v2', 'default'], 'id': None, 'group': None}, 'checkpoint': {'topk': {'monitor_key': 'test_success_rate', 'mode': 'max', 'k': 5, 'format_str': 'epoch={epoch:04d}-test_success_rate={test_success_rate:.3f}.ckpt'}, 'save_last_ckpt': True, 'save_last_snapshot': False}, 'multi_run': {'run_dir': 'data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2', 'wandb_name_base': '2026.01.27-11.52.50_train_llmbc_lowdim_box-close-v2'}, 'task': {'name': 'box-close-v2', 'obs_dim': 9, 'action_dim': 4, 'env_runner': {'_target_': 'llmbc.env_runner.metaworld_lowdim_runner.MetaworldLowdimRunner', 'env_name': 'llf-metaworld-box-close-v2', 'n_train': 10, 'n_test': 50, 'n_envs': 10, 'max_steps': 30, 'n_obs_steps': 1, 'n_action_steps': 1, 'instruction_type': 'b', 'feedback_type': ['hp', 'hn', 'fp'], 'visual': False, 'discount': 0.9}, 'dataset': {'_target_': 'llmbc.dataset.metaworld_lowdim_dataset.MetaworldLowdimDataset', 'data_path': 'datasets/box-close-v2-o4-mini-split0.05.pt', 'data_path2': 'datasets/box-close-v2-o4-mini-split0.05.pt', 'horizon': 1, 'pad_before': 0, 'pad_after': 0, 'obs_eef_target': True, 'use_manual_normalizer': False, 'val_ratio': 0.1, 'dummy_normalizer': True}, 'instructor': {'_target_': 'llmbc.translator.instructor.metaworld_instructor.box_close_v2_instructor.BoxCloseV2Instructor'}}, 'llm': {'name': 'HuggingFaceTB/SmolLM2-135M-Instruct', 'model_name': 'SmolLM2-135M-Instruct', 'config_target': 'llmbc.model.llm.llama_lowdim_model.LowdimLlamaConfig', 'causal_lm_target': 'llmbc.model.llm.llama_lowdim_model.LowdimLlamaForCausalLM', 'use_quantization': False, 'use_joint_mlp_projector': True, 'llm_mode': 'ete-finetuned', 'finetune_mode': 'orig', 'checkpoint': 'data/outputs/2026.01.26/10.26.52_train_llm_lowdim_box-close-v2/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-box-close-v2/checkpoint-6000', 'max_length': 100, 'lora_config': {'r': 32, 'lora_alpha': 64, 'lora_dropout': 0.05, 'bias': 'none', 'task_type': 'CAUSAL_LM'}, 'prompter': {'_target_': 'llmbc.translator.prompter.smollm2_prompter.SmolLM2Prompter', 'use_joint_mlp_projector': True}, 'hydra': {'job': {'override_dirname': 'HuggingFaceTB/SmolLM2-135M-Instruct'}, 'run': {'dir': 'data/outputs/2026.01.27/11.52.50_HuggingFaceTB/SmolLM2-135M-Instruct'}}}}
14
+ 2026-01-27 11:53:07,504 INFO MainThread:754949 [wandb_init.py:init():669] starting backend
15
+ 2026-01-27 11:53:07,504 INFO MainThread:754949 [wandb_init.py:init():673] sending inform_init request
16
+ 2026-01-27 11:53:07,505 INFO MainThread:754949 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2026-01-27 11:53:07,506 INFO MainThread:754949 [wandb_init.py:init():686] backend started and connected
18
+ 2026-01-27 11:53:07,515 INFO MainThread:754949 [wandb_init.py:init():781] updated telemetry
19
+ 2026-01-27 11:53:07,570 INFO MainThread:754949 [wandb_init.py:init():814] communicating run to backend with 90.0 second timeout
20
+ 2026-01-27 11:53:08,334 INFO MainThread:754949 [wandb_init.py:init():867] starting run threads in backend
21
+ 2026-01-27 11:53:08,858 INFO MainThread:754949 [wandb_run.py:_console_start():2451] atexit reg
22
+ 2026-01-27 11:53:08,858 INFO MainThread:754949 [wandb_run.py:_redirect():2299] redirect: wrap_raw
23
+ 2026-01-27 11:53:08,858 INFO MainThread:754949 [wandb_run.py:_redirect():2364] Wrapping output streams.
24
+ 2026-01-27 11:53:08,858 INFO MainThread:754949 [wandb_run.py:_redirect():2389] Redirects installed.
25
+ 2026-01-27 11:53:08,861 INFO MainThread:754949 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2026-01-27 11:53:08,862 INFO MainThread:754949 [wandb_run.py:_config_callback():1389] config_cb None None {'output_dir': '/work/u1131674/LLM-BC/data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2'}
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115307-raachjxu/run-raachjxu.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77644bf1a49da807f34879bd1902e767c1fad49735292efcd9bb182c392eb3fb
3
+ size 1769472
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/files/config.yaml ADDED
@@ -0,0 +1,272 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _target_:
2
+ value: llmbc.workspace.train_llmbc_lowdim_workspace.TrainLLMBCLowdimWorkspace
3
+ _wandb:
4
+ value:
5
+ cli_version: 0.18.6
6
+ m: []
7
+ python_version: 3.9.25
8
+ t:
9
+ "1":
10
+ - 1
11
+ - 2
12
+ - 3
13
+ - 5
14
+ - 11
15
+ - 12
16
+ - 41
17
+ - 49
18
+ - 50
19
+ - 51
20
+ - 53
21
+ - 55
22
+ - 71
23
+ - 83
24
+ - 95
25
+ - 98
26
+ - 100
27
+ - 105
28
+ "2":
29
+ - 1
30
+ - 2
31
+ - 3
32
+ - 5
33
+ - 11
34
+ - 12
35
+ - 41
36
+ - 49
37
+ - 50
38
+ - 51
39
+ - 53
40
+ - 55
41
+ - 71
42
+ - 83
43
+ - 95
44
+ - 98
45
+ - 100
46
+ - 105
47
+ "3":
48
+ - 5
49
+ - 13
50
+ - 15
51
+ - 16
52
+ - 23
53
+ - 55
54
+ - 61
55
+ "4": 3.9.25
56
+ "5": 0.18.6
57
+ "6": 4.47.1
58
+ "8":
59
+ - 5
60
+ "12": 0.18.6
61
+ "13": linux-x86_64
62
+ action_dim:
63
+ value: 4
64
+ checkpoint:
65
+ value:
66
+ save_last_ckpt: true
67
+ save_last_snapshot: false
68
+ topk:
69
+ format_str: epoch={epoch:04d}-test_success_rate={test_success_rate:.3f}.ckpt
70
+ k: 5
71
+ mode: max
72
+ monitor_key: test_success_rate
73
+ dataloader:
74
+ value:
75
+ batch_size: 16
76
+ num_workers: 0
77
+ persistent_workers: false
78
+ pin_memory: false
79
+ shuffle: true
80
+ exp_name:
81
+ value: default
82
+ horizon:
83
+ value: 1
84
+ llm:
85
+ value:
86
+ causal_lm_target: llmbc.model.llm.llama_lowdim_model.LowdimLlamaForCausalLM
87
+ checkpoint: data/outputs/2026.01.26/10.26.52_train_llm_lowdim_box-close-v2/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-box-close-v2/checkpoint-6000
88
+ config_target: llmbc.model.llm.llama_lowdim_model.LowdimLlamaConfig
89
+ finetune_mode: orig
90
+ hydra:
91
+ job:
92
+ override_dirname: HuggingFaceTB/SmolLM2-135M-Instruct
93
+ run:
94
+ dir: data/outputs/2026.01.27/11.52.50_HuggingFaceTB/SmolLM2-135M-Instruct
95
+ llm_mode: ete-finetuned
96
+ lora_config:
97
+ bias: none
98
+ lora_alpha: 64
99
+ lora_dropout: 0.05
100
+ r: 32
101
+ task_type: CAUSAL_LM
102
+ max_length: 100
103
+ model_name: SmolLM2-135M-Instruct
104
+ name: HuggingFaceTB/SmolLM2-135M-Instruct
105
+ prompter:
106
+ _target_: llmbc.translator.prompter.smollm2_prompter.SmolLM2Prompter
107
+ use_joint_mlp_projector: true
108
+ use_joint_mlp_projector: true
109
+ use_quantization: false
110
+ llm_do_sample:
111
+ value: false
112
+ llm_orig_expert_feedback:
113
+ value: true
114
+ logging:
115
+ value:
116
+ group: null
117
+ id: null
118
+ mode: online
119
+ name: 2026.01.27-11.52.50_train_llmbc_lowdim_box-close-v2
120
+ project: box-close-v2-training-split-ablation
121
+ resume: true
122
+ tags:
123
+ - train_llmbc_lowdim
124
+ - box-close-v2
125
+ - default
126
+ model_name:
127
+ value: HuggingFaceTB/SmolLM2-135M-Instruct
128
+ multi_run:
129
+ value:
130
+ run_dir: data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2
131
+ wandb_name_base: 2026.01.27-11.52.50_train_llmbc_lowdim_box-close-v2
132
+ n_action_steps:
133
+ value: 1
134
+ n_latency_steps:
135
+ value: 0
136
+ n_obs_steps:
137
+ value: 1
138
+ name:
139
+ value: train_llmbc_lowdim
140
+ obs_dim:
141
+ value: 9
142
+ optimizer:
143
+ value:
144
+ _target_: torch.optim.AdamW
145
+ betas:
146
+ - 0.95
147
+ - 0.999
148
+ eps: 1e-08
149
+ lr: 0.01
150
+ weight_decay: 1e-06
151
+ output_dir:
152
+ value: /work/u1131674/LLM-BC/data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2
153
+ past_action_visible:
154
+ value: false
155
+ policy:
156
+ value:
157
+ _target_: llmbc.policy.llmbc_lowdim_policy.LLMBCLowdimPolicy
158
+ action_dim: 4
159
+ horizon: 1
160
+ llm_discriminator:
161
+ _target_: llmbc.discriminator.llm_ce_discriminator.LLMCEDiscriminator
162
+ llm_translator:
163
+ _target_: llmbc.translator.llm_translator.LLMTranslator
164
+ action_dim: 4
165
+ cfg:
166
+ causal_lm_target: llmbc.model.llm.llama_lowdim_model.LowdimLlamaForCausalLM
167
+ checkpoint: data/outputs/2026.01.26/10.26.52_train_llm_lowdim_box-close-v2/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-box-close-v2/checkpoint-6000
168
+ config_target: llmbc.model.llm.llama_lowdim_model.LowdimLlamaConfig
169
+ finetune_mode: orig
170
+ hydra:
171
+ job:
172
+ override_dirname: HuggingFaceTB/SmolLM2-135M-Instruct
173
+ run:
174
+ dir: data/outputs/2026.01.27/11.52.50_HuggingFaceTB/SmolLM2-135M-Instruct
175
+ llm_mode: ete-finetuned
176
+ lora_config:
177
+ bias: none
178
+ lora_alpha: 64
179
+ lora_dropout: 0.05
180
+ r: 32
181
+ task_type: CAUSAL_LM
182
+ max_length: 100
183
+ model_name: SmolLM2-135M-Instruct
184
+ name: HuggingFaceTB/SmolLM2-135M-Instruct
185
+ prompter:
186
+ _target_: llmbc.translator.prompter.smollm2_prompter.SmolLM2Prompter
187
+ use_joint_mlp_projector: true
188
+ use_joint_mlp_projector: true
189
+ use_quantization: false
190
+ horizon: 1
191
+ n_action_steps: 1
192
+ n_obs_steps: 1
193
+ obs_dim: 9
194
+ task_id: box-close-v2
195
+ loss_bc_weight: 1
196
+ loss_llm_weight: 0.001
197
+ model:
198
+ _target_: llmbc.model.policy.policy_mlp.PolicyMLP
199
+ activation: relu
200
+ hidden_size:
201
+ - 256
202
+ - 256
203
+ input_size: 9
204
+ n_action_steps: 1
205
+ n_obs_steps: 1
206
+ output_size: 4
207
+ n_action_steps: 1
208
+ n_obs_steps: 1
209
+ normalize_llm_loss: true
210
+ obs_dim: 9
211
+ task:
212
+ value:
213
+ action_dim: 4
214
+ dataset:
215
+ _target_: llmbc.dataset.metaworld_lowdim_dataset.MetaworldLowdimDataset
216
+ data_path: datasets/box-close-v2-o4-mini-split0.05.pt
217
+ data_path2: datasets/box-close-v2-o4-mini-split0.05.pt
218
+ dummy_normalizer: true
219
+ horizon: 1
220
+ obs_eef_target: true
221
+ pad_after: 0
222
+ pad_before: 0
223
+ use_manual_normalizer: false
224
+ val_ratio: 0.1
225
+ env_runner:
226
+ _target_: llmbc.env_runner.metaworld_lowdim_runner.MetaworldLowdimRunner
227
+ discount: 0.9
228
+ env_name: llf-metaworld-box-close-v2
229
+ feedback_type:
230
+ - hp
231
+ - hn
232
+ - fp
233
+ instruction_type: b
234
+ max_steps: 30
235
+ n_action_steps: 1
236
+ n_envs: 10
237
+ n_obs_steps: 1
238
+ n_test: 50
239
+ n_train: 10
240
+ visual: false
241
+ instructor:
242
+ _target_: llmbc.translator.instructor.metaworld_instructor.box_close_v2_instructor.BoxCloseV2Instructor
243
+ name: box-close-v2
244
+ obs_dim: 9
245
+ task_name:
246
+ value: box-close-v2
247
+ training:
248
+ value:
249
+ checkpoint_every: 100
250
+ debug: false
251
+ device: cuda:0
252
+ grad_norm_clip: 0.5
253
+ gradient_accumulate_every: 8
254
+ lr_scheduler: cosine
255
+ lr_warmup_steps: 10
256
+ max_train_steps: null
257
+ max_val_steps: null
258
+ num_epochs: 20001
259
+ resume: false
260
+ rollout_every: 100
261
+ sample_every: 100
262
+ sample_max_batch: 128
263
+ seed: 43
264
+ tqdm_interval_sec: 1
265
+ val_every: 1
266
+ val_dataloader:
267
+ value:
268
+ batch_size: 16
269
+ num_workers: 0
270
+ persistent_workers: false
271
+ pin_memory: false
272
+ shuffle: true
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/files/output.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ Eval MetaworldLowdimRunner 1/6: 0%| | 0/30 [00:00<?, ?it/s]/work/u1131674/LLM-BC/llmbc/common/llfbench_util.py:39: UserWarning: Creating a tensor from a list of numpy.ndarrays is extremely slow. Please consider converting the list to a single numpy.ndarray with numpy.array() before converting to a tensor. (Triggered internally at ../torch/csrc/utils/tensor_new.cpp:275.)
2
+ obs = torch.tensor(obs, dtype=torch.float32).unsqueeze(dim=0).to(device)
3
+
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/files/wandb-metadata.json ADDED
@@ -0,0 +1,107 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-4.18.0-513.24.1.el8_9.x86_64-x86_64-with-glibc2.28",
3
+ "python": "3.9.25",
4
+ "startedAt": "2026-01-27T03:53:08.073313Z",
5
+ "args": [
6
+ "--config-path",
7
+ "config/data_efficiency",
8
+ "--config-name",
9
+ "llmbc_box-close-v2_o4-mini_split0.05.yaml",
10
+ "training.seed=43"
11
+ ],
12
+ "program": "/work/u1131674/LLM-BC/./train.py",
13
+ "codePath": "train.py",
14
+ "git": {
15
+ "remote": "https://github.com/CHYang25/LLM-BC.git",
16
+ "commit": "206b9c57fa687707dfa1ee0ebb33b507135a59c6"
17
+ },
18
+ "email": "chris920325@gmail.com",
19
+ "root": "/work/u1131674/LLM-BC/data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2",
20
+ "host": "hgpn06",
21
+ "username": "u1131674",
22
+ "executable": "/home/u1131674/.conda/envs/llm-bc/bin/python3",
23
+ "codePathLocal": "train.py",
24
+ "cpu_count": 112,
25
+ "cpu_count_logical": 112,
26
+ "gpu": "NVIDIA H100 80GB HBM3",
27
+ "gpu_count": 1,
28
+ "disk": {
29
+ "/": {
30
+ "total": "1918024196096",
31
+ "used": "333231804416"
32
+ }
33
+ },
34
+ "memory": {
35
+ "total": "2163623079936"
36
+ },
37
+ "cpu": {
38
+ "count": 112,
39
+ "countLogical": 112
40
+ },
41
+ "gpu_nvidia": [
42
+ {
43
+ "name": "NVIDIA H100 80GB HBM3",
44
+ "memoryTotal": "85520809984",
45
+ "cudaCores": 16896,
46
+ "architecture": "Hopper"
47
+ }
48
+ ],
49
+ "slurm": {
50
+ "cluster_name": "hpc",
51
+ "conf": "/etc/slurm/slurm.conf",
52
+ "cpu_bind": "quiet,mask_cpu:0x0000000000000000000000F0000F",
53
+ "cpu_bind_list": "0x0000000000000000000000F0000F",
54
+ "cpu_bind_type": "mask_cpu:",
55
+ "cpu_bind_verbose": "quiet",
56
+ "cpus_on_node": "8",
57
+ "cpus_per_task": "8",
58
+ "distribution": "cyclic,pack",
59
+ "gpus_on_node": "1",
60
+ "gpus_per_node": "1",
61
+ "gtids": "0",
62
+ "job_account": "mst114558",
63
+ "job_cpus_per_node": "8",
64
+ "job_end_time": "1769658768",
65
+ "job_gid": "106773",
66
+ "job_group": "MST114558",
67
+ "job_id": "104243",
68
+ "job_name": "python3",
69
+ "job_nodelist": "hgpn06",
70
+ "job_num_nodes": "1",
71
+ "job_partition": "normal",
72
+ "job_qos": "normal",
73
+ "job_start_time": "1769485968",
74
+ "job_uid": "41408",
75
+ "job_user": "u1131674",
76
+ "jobid": "104243",
77
+ "launch_node_ipaddr": "172.21.101.1",
78
+ "localid": "0",
79
+ "mem_per_node": "204800",
80
+ "nnodes": "1",
81
+ "nodeid": "0",
82
+ "nodelist": "hgpn06",
83
+ "nprocs": "1",
84
+ "ntasks": "1",
85
+ "prio_process": "0",
86
+ "procid": "0",
87
+ "srun_comm_host": "172.21.101.1",
88
+ "srun_comm_port": "44181",
89
+ "step_gpus": "3",
90
+ "step_id": "0",
91
+ "step_launcher_port": "44181",
92
+ "step_nodelist": "hgpn06",
93
+ "step_num_nodes": "1",
94
+ "step_num_tasks": "1",
95
+ "step_tasks_per_node": "1",
96
+ "stepid": "0",
97
+ "submit_dir": "/work/u1131674/LLM-BC",
98
+ "submit_host": "cbi-lgn01",
99
+ "task_pid": "2419965",
100
+ "tasks_per_node": "1",
101
+ "topology_addr": "ibsw1.hgpn06",
102
+ "topology_addr_pattern": "switch.node",
103
+ "tres_per_task": "cpu:8",
104
+ "umask": "0022"
105
+ },
106
+ "cudaVersion": "12.4"
107
+ }
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_timestamp":1.769524654311829e+09,"global_step":260012,"test/success_rate":0.54,"grad_norm":4.1202642023563385e-05,"lr":2.337295357257574e-11,"_runtime":38668.49954648,"train_loss_llm":9.465217590332031e-05,"test/cumulative_reward":12.559526529061893,"train/success_rate":0.8,"val_loss":2.897437116189394e-05,"train_action_mse_error":1.0403916661516632e-07,"train/cumulative_reward":14.612570521362438,"train/mean_score":6.996024131036938,"test/mean_score":5.711785312482341,"_wandb":{"runtime":38668},"epoch":20000,"train_loss":1.5365032077432681e-07,"_step":260012,"train_loss_bc":9.23023861787442e-08}
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/logs/debug-core.log ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-01-27T11:53:07.47897058+08:00","level":"INFO","msg":"started logging, with flags","port-filename":"/tmp/tmpy8hm3v8r/port-2419965.txt","pid":2419965,"debug":false,"disable-analytics":false}
2
+ {"time":"2026-01-27T11:53:07.478992022+08:00","level":"INFO","msg":"FeatureState","shutdownOnParentExitEnabled":false}
3
+ {"time":"2026-01-27T11:53:07.479516556+08:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":2419965}
4
+ {"time":"2026-01-27T11:53:07.479510234+08:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":40351,"Zone":""}}
5
+ {"time":"2026-01-27T11:53:07.672421694+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:60334"}
6
+ {"time":"2026-01-27T11:53:08.073873525+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"raachjxu","id":"127.0.0.1:60334"}
7
+ {"time":"2026-01-27T11:53:08.190317831+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"raachjxu","id":"127.0.0.1:60334"}
8
+ {"time":"2026-01-27T22:37:36.572632384+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"127.0.0.1:60334"}
9
+ {"time":"2026-01-27T22:37:36.572847244+08:00","level":"INFO","msg":"server is shutting down"}
10
+ {"time":"2026-01-27T22:37:36.572836178+08:00","level":"INFO","msg":"connection: Close: initiating connection closure","id":"127.0.0.1:60334"}
11
+ {"time":"2026-01-27T22:37:36.572940064+08:00","level":"INFO","msg":"connection: Close: connection successfully closed","id":"127.0.0.1:60334"}
12
+ {"time":"2026-01-27T22:37:41.125798856+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"127.0.0.1:60334"}
13
+ {"time":"2026-01-27T22:37:41.125815466+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"127.0.0.1:60334"}
14
+ {"time":"2026-01-27T22:37:41.125826131+08:00","level":"INFO","msg":"server is closed"}
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/logs/debug-internal.log ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-01-27T11:53:08.074868883+08:00","level":"INFO","msg":"using version","core version":"0.18.6"}
2
+ {"time":"2026-01-27T11:53:08.074878267+08:00","level":"INFO","msg":"created symlink","path":"/work/u1131674/LLM-BC/data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/logs/debug-core.log"}
3
+ {"time":"2026-01-27T11:53:08.190280513+08:00","level":"INFO","msg":"created new stream","id":"raachjxu"}
4
+ {"time":"2026-01-27T11:53:08.190312975+08:00","level":"INFO","msg":"stream: started","id":"raachjxu"}
5
+ {"time":"2026-01-27T11:53:08.190338185+08:00","level":"INFO","msg":"sender: started","stream_id":"raachjxu"}
6
+ {"time":"2026-01-27T11:53:08.190329168+08:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"raachjxu"}}
7
+ {"time":"2026-01-27T11:53:08.1903328+08:00","level":"INFO","msg":"handler: started","stream_id":{"value":"raachjxu"}}
8
+ {"time":"2026-01-27T11:53:08.732831596+08:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2026-01-27T22:37:36.57283309+08:00","level":"INFO","msg":"stream: closing","id":"raachjxu"}
10
+ {"time":"2026-01-27T22:37:36.572875878+08:00","level":"INFO","msg":"Stopping system monitor"}
11
+ {"time":"2026-01-27T22:37:36.573489042+08:00","level":"INFO","msg":"Stopped system monitor"}
12
+ {"time":"2026-01-27T22:37:40.850119587+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
13
+ {"time":"2026-01-27T22:37:41.125467393+08:00","level":"INFO","msg":"handler: closed","stream_id":{"value":"raachjxu"}}
14
+ {"time":"2026-01-27T22:37:41.125498907+08:00","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"raachjxu"}}
15
+ {"time":"2026-01-27T22:37:41.125540962+08:00","level":"INFO","msg":"sender: closed","stream_id":"raachjxu"}
16
+ {"time":"2026-01-27T22:37:41.125562117+08:00","level":"INFO","msg":"stream: closed","id":"raachjxu"}
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/logs/debug.log ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-01-27 11:53:08,070 INFO MainThread:2419965 [wandb_setup.py:_flush():79] Current SDK version is 0.18.6
2
+ 2026-01-27 11:53:08,070 INFO MainThread:2419965 [wandb_setup.py:_flush():79] Configure stats pid to 2419965
3
+ 2026-01-27 11:53:08,070 INFO MainThread:2419965 [wandb_setup.py:_flush():79] Loading settings from /home/u1131674/.config/wandb/settings
4
+ 2026-01-27 11:53:08,070 INFO MainThread:2419965 [wandb_setup.py:_flush():79] Loading settings from /work/u1131674/LLM-BC/wandb/settings
5
+ 2026-01-27 11:53:08,070 INFO MainThread:2419965 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2026-01-27 11:53:08,070 INFO MainThread:2419965 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': 'online', '_disable_service': None}
7
+ 2026-01-27 11:53:08,071 INFO MainThread:2419965 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train.py', 'program_abspath': '/work/u1131674/LLM-BC/train.py', 'program': '/work/u1131674/LLM-BC/./train.py'}
8
+ 2026-01-27 11:53:08,071 INFO MainThread:2419965 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2026-01-27 11:53:08,071 INFO MainThread:2419965 [wandb_init.py:_log_setup():533] Logging user logs to /work/u1131674/LLM-BC/data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/logs/debug.log
10
+ 2026-01-27 11:53:08,071 INFO MainThread:2419965 [wandb_init.py:_log_setup():534] Logging internal logs to /work/u1131674/LLM-BC/data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/logs/debug-internal.log
11
+ 2026-01-27 11:53:08,071 INFO MainThread:2419965 [wandb_init.py:init():619] calling init triggers
12
+ 2026-01-27 11:53:08,071 INFO MainThread:2419965 [wandb_init.py:init():626] wandb.init called with sweep_config: {}
13
+ config: {'name': 'train_llmbc_lowdim', '_target_': 'llmbc.workspace.train_llmbc_lowdim_workspace.TrainLLMBCLowdimWorkspace', 'obs_dim': 9, 'action_dim': 4, 'task_name': 'box-close-v2', 'exp_name': 'default', 'model_name': 'HuggingFaceTB/SmolLM2-135M-Instruct', 'horizon': 1, 'n_obs_steps': 1, 'n_action_steps': 1, 'n_latency_steps': 0, 'past_action_visible': False, 'llm_orig_expert_feedback': True, 'llm_do_sample': False, 'policy': {'_target_': 'llmbc.policy.llmbc_lowdim_policy.LLMBCLowdimPolicy', 'model': {'_target_': 'llmbc.model.policy.policy_mlp.PolicyMLP', 'input_size': 9, 'hidden_size': [256, 256], 'output_size': 4, 'activation': 'relu', 'n_obs_steps': 1, 'n_action_steps': 1}, 'obs_dim': 9, 'action_dim': 4, 'llm_discriminator': {'_target_': 'llmbc.discriminator.llm_ce_discriminator.LLMCEDiscriminator', 'task_id': 'box-close-v2', 'llm_translator': {'_target_': 'llmbc.translator.llm_translator.LLMTranslator', 'cfg': {'name': 'HuggingFaceTB/SmolLM2-135M-Instruct', 'model_name': 'SmolLM2-135M-Instruct', 'config_target': 'llmbc.model.llm.llama_lowdim_model.LowdimLlamaConfig', 'causal_lm_target': 'llmbc.model.llm.llama_lowdim_model.LowdimLlamaForCausalLM', 'use_quantization': False, 'use_joint_mlp_projector': True, 'llm_mode': 'ete-finetuned', 'finetune_mode': 'orig', 'checkpoint': 'data/outputs/2026.01.26/10.26.52_train_llm_lowdim_box-close-v2/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-box-close-v2/checkpoint-6000', 'max_length': 100, 'lora_config': {'r': 32, 'lora_alpha': 64, 'lora_dropout': 0.05, 'bias': 'none', 'task_type': 'CAUSAL_LM'}, 'prompter': {'_target_': 'llmbc.translator.prompter.smollm2_prompter.SmolLM2Prompter', 'use_joint_mlp_projector': True}, 'hydra': {'job': {'override_dirname': 'HuggingFaceTB/SmolLM2-135M-Instruct'}, 'run': {'dir': 'data/outputs/2026.01.27/11.52.50_HuggingFaceTB/SmolLM2-135M-Instruct'}}}, 'obs_dim': 9, 'action_dim': 4, 'horizon': 1, 'n_obs_steps': 1, 'n_action_steps': 1}}, 'loss_bc_weight': 1.0, 'loss_llm_weight': 0.001, 'horizon': 1, 'n_obs_steps': 1, 'n_action_steps': 1, 'normalize_llm_loss': True}, 'dataloader': {'batch_size': 16, 'num_workers': 0, 'shuffle': True, 'pin_memory': False, 'persistent_workers': False}, 'val_dataloader': {'batch_size': 16, 'num_workers': 0, 'shuffle': True, 'pin_memory': False, 'persistent_workers': False}, 'optimizer': {'_target_': 'torch.optim.AdamW', 'lr': 0.01, 'betas': [0.95, 0.999], 'eps': 1e-08, 'weight_decay': 1e-06}, 'training': {'device': 'cuda:0', 'seed': 43, 'debug': False, 'resume': False, 'lr_scheduler': 'cosine', 'lr_warmup_steps': 10, 'num_epochs': 20001, 'gradient_accumulate_every': 8, 'grad_norm_clip': 0.5, 'rollout_every': 100, 'checkpoint_every': 100, 'val_every': 1, 'sample_every': 100, 'sample_max_batch': 128, 'max_train_steps': None, 'max_val_steps': None, 'tqdm_interval_sec': 1.0}, 'logging': {'project': 'box-close-v2-training-split-ablation', 'resume': True, 'mode': 'online', 'name': '2026.01.27-11.52.50_train_llmbc_lowdim_box-close-v2', 'tags': ['train_llmbc_lowdim', 'box-close-v2', 'default'], 'id': None, 'group': None}, 'checkpoint': {'topk': {'monitor_key': 'test_success_rate', 'mode': 'max', 'k': 5, 'format_str': 'epoch={epoch:04d}-test_success_rate={test_success_rate:.3f}.ckpt'}, 'save_last_ckpt': True, 'save_last_snapshot': False}, 'multi_run': {'run_dir': 'data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2', 'wandb_name_base': '2026.01.27-11.52.50_train_llmbc_lowdim_box-close-v2'}, 'task': {'name': 'box-close-v2', 'obs_dim': 9, 'action_dim': 4, 'env_runner': {'_target_': 'llmbc.env_runner.metaworld_lowdim_runner.MetaworldLowdimRunner', 'env_name': 'llf-metaworld-box-close-v2', 'n_train': 10, 'n_test': 50, 'n_envs': 10, 'max_steps': 30, 'n_obs_steps': 1, 'n_action_steps': 1, 'instruction_type': 'b', 'feedback_type': ['hp', 'hn', 'fp'], 'visual': False, 'discount': 0.9}, 'dataset': {'_target_': 'llmbc.dataset.metaworld_lowdim_dataset.MetaworldLowdimDataset', 'data_path': 'datasets/box-close-v2-o4-mini-split0.05.pt', 'data_path2': 'datasets/box-close-v2-o4-mini-split0.05.pt', 'horizon': 1, 'pad_before': 0, 'pad_after': 0, 'obs_eef_target': True, 'use_manual_normalizer': False, 'val_ratio': 0.1, 'dummy_normalizer': True}, 'instructor': {'_target_': 'llmbc.translator.instructor.metaworld_instructor.box_close_v2_instructor.BoxCloseV2Instructor'}}, 'llm': {'name': 'HuggingFaceTB/SmolLM2-135M-Instruct', 'model_name': 'SmolLM2-135M-Instruct', 'config_target': 'llmbc.model.llm.llama_lowdim_model.LowdimLlamaConfig', 'causal_lm_target': 'llmbc.model.llm.llama_lowdim_model.LowdimLlamaForCausalLM', 'use_quantization': False, 'use_joint_mlp_projector': True, 'llm_mode': 'ete-finetuned', 'finetune_mode': 'orig', 'checkpoint': 'data/outputs/2026.01.26/10.26.52_train_llm_lowdim_box-close-v2/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-box-close-v2/checkpoint-6000', 'max_length': 100, 'lora_config': {'r': 32, 'lora_alpha': 64, 'lora_dropout': 0.05, 'bias': 'none', 'task_type': 'CAUSAL_LM'}, 'prompter': {'_target_': 'llmbc.translator.prompter.smollm2_prompter.SmolLM2Prompter', 'use_joint_mlp_projector': True}, 'hydra': {'job': {'override_dirname': 'HuggingFaceTB/SmolLM2-135M-Instruct'}, 'run': {'dir': 'data/outputs/2026.01.27/11.52.50_HuggingFaceTB/SmolLM2-135M-Instruct'}}}}
14
+ 2026-01-27 11:53:08,071 INFO MainThread:2419965 [wandb_init.py:init():669] starting backend
15
+ 2026-01-27 11:53:08,071 INFO MainThread:2419965 [wandb_init.py:init():673] sending inform_init request
16
+ 2026-01-27 11:53:08,072 INFO MainThread:2419965 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2026-01-27 11:53:08,073 INFO MainThread:2419965 [wandb_init.py:init():686] backend started and connected
18
+ 2026-01-27 11:53:08,082 INFO MainThread:2419965 [wandb_init.py:init():781] updated telemetry
19
+ 2026-01-27 11:53:08,142 INFO MainThread:2419965 [wandb_init.py:init():814] communicating run to backend with 90.0 second timeout
20
+ 2026-01-27 11:53:08,717 INFO MainThread:2419965 [wandb_init.py:init():859] run resumed
21
+ 2026-01-27 11:53:08,728 INFO MainThread:2419965 [wandb_init.py:init():867] starting run threads in backend
22
+ 2026-01-27 11:53:09,276 INFO MainThread:2419965 [wandb_run.py:_console_start():2451] atexit reg
23
+ 2026-01-27 11:53:09,276 INFO MainThread:2419965 [wandb_run.py:_redirect():2299] redirect: wrap_raw
24
+ 2026-01-27 11:53:09,276 INFO MainThread:2419965 [wandb_run.py:_redirect():2364] Wrapping output streams.
25
+ 2026-01-27 11:53:09,276 INFO MainThread:2419965 [wandb_run.py:_redirect():2389] Redirects installed.
26
+ 2026-01-27 11:53:09,294 INFO MainThread:2419965 [wandb_init.py:init():911] run started, returning control to user process
27
+ 2026-01-27 11:53:09,295 INFO MainThread:2419965 [wandb_run.py:_config_callback():1389] config_cb None None {'output_dir': '/work/u1131674/LLM-BC/data/outputs/2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2'}
28
+ 2026-01-27 22:37:36,572 WARNING MsgRouterThr:2419965 [router.py:message_loop():75] message_loop has been closed
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115308-raachjxu/run-raachjxu.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3781f811f713b9c3f856e19b8a57bcc0d88ac14e4012b98457d6016ab9513a48
3
+ size 218982786
2026.01.27/11.52.50_train_llmbc_lowdim_box-close-v2/wandb/wandb-resume.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"run_id": "raachjxu"}
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/.hydra/config.yaml ADDED
@@ -0,0 +1,163 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: train_llmbc_lowdim
2
+ _target_: llmbc.workspace.train_llmbc_lowdim_workspace.TrainLLMBCLowdimWorkspace
3
+ obs_dim: ${task.obs_dim}
4
+ action_dim: ${task.action_dim}
5
+ task_name: ${task.name}
6
+ exp_name: default
7
+ model_name: ${llm.name}
8
+ horizon: 1
9
+ n_obs_steps: 1
10
+ n_action_steps: 1
11
+ n_latency_steps: 0
12
+ past_action_visible: false
13
+ llm_orig_expert_feedback: true
14
+ llm_do_sample: false
15
+ policy:
16
+ _target_: llmbc.policy.llmbc_lowdim_policy.LLMBCLowdimPolicy
17
+ model:
18
+ _target_: llmbc.model.policy.policy_mlp.PolicyMLP
19
+ input_size: ${eval:'${n_obs_steps}*${obs_dim}'}
20
+ hidden_size:
21
+ - 256
22
+ - 256
23
+ output_size: ${eval:'${n_action_steps}*${action_dim}'}
24
+ activation: relu
25
+ n_obs_steps: ${n_obs_steps}
26
+ n_action_steps: ${n_action_steps}
27
+ obs_dim: ${obs_dim}
28
+ action_dim: ${action_dim}
29
+ llm_discriminator:
30
+ _target_: llmbc.discriminator.llm_ce_discriminator.LLMCEDiscriminator
31
+ task_id: ${task_name}
32
+ llm_translator:
33
+ _target_: llmbc.translator.llm_translator.LLMTranslator
34
+ cfg: ${llm}
35
+ obs_dim: ${task.obs_dim}
36
+ action_dim: ${task.action_dim}
37
+ horizon: ${horizon}
38
+ n_obs_steps: ${n_obs_steps}
39
+ n_action_steps: ${n_action_steps}
40
+ loss_bc_weight: 1.0
41
+ loss_llm_weight: 0.001
42
+ horizon: ${horizon}
43
+ n_obs_steps: ${n_obs_steps}
44
+ n_action_steps: ${n_action_steps}
45
+ normalize_llm_loss: true
46
+ dataloader:
47
+ batch_size: 16
48
+ num_workers: 0
49
+ shuffle: true
50
+ pin_memory: false
51
+ persistent_workers: false
52
+ val_dataloader:
53
+ batch_size: 16
54
+ num_workers: 0
55
+ shuffle: true
56
+ pin_memory: false
57
+ persistent_workers: false
58
+ optimizer:
59
+ _target_: torch.optim.AdamW
60
+ lr: 0.01
61
+ betas:
62
+ - 0.95
63
+ - 0.999
64
+ eps: 1.0e-08
65
+ weight_decay: 1.0e-06
66
+ training:
67
+ device: cuda:0
68
+ seed: 42
69
+ debug: false
70
+ resume: false
71
+ lr_scheduler: cosine
72
+ lr_warmup_steps: 10
73
+ num_epochs: 20001
74
+ gradient_accumulate_every: 8
75
+ grad_norm_clip: 0.5
76
+ rollout_every: 100
77
+ checkpoint_every: 100
78
+ val_every: 1
79
+ sample_every: 100
80
+ sample_max_batch: 128
81
+ max_train_steps: null
82
+ max_val_steps: null
83
+ tqdm_interval_sec: 1.0
84
+ logging:
85
+ project: ${task.name}-training-split-ablation
86
+ resume: true
87
+ mode: online
88
+ name: ${now:%Y.%m.%d-%H.%M.%S}_${name}_${task_name}
89
+ tags:
90
+ - ${name}
91
+ - ${task_name}
92
+ - ${exp_name}
93
+ id: null
94
+ group: null
95
+ checkpoint:
96
+ topk:
97
+ monitor_key: test_success_rate
98
+ mode: max
99
+ k: 5
100
+ format_str: epoch={epoch:04d}-test_success_rate={test_success_rate:.3f}.ckpt
101
+ save_last_ckpt: true
102
+ save_last_snapshot: false
103
+ multi_run:
104
+ run_dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
105
+ wandb_name_base: ${now:%Y.%m.%d-%H.%M.%S}_${name}_${task_name}
106
+ task:
107
+ name: box-close-v2
108
+ obs_dim: 9
109
+ action_dim: 4
110
+ env_runner:
111
+ _target_: llmbc.env_runner.metaworld_lowdim_runner.MetaworldLowdimRunner
112
+ env_name: llf-metaworld-box-close-v2
113
+ n_train: 10
114
+ n_test: 50
115
+ n_envs: 10
116
+ max_steps: 30
117
+ n_obs_steps: ${n_obs_steps}
118
+ n_action_steps: ${n_action_steps}
119
+ instruction_type: b
120
+ feedback_type:
121
+ - hp
122
+ - hn
123
+ - fp
124
+ visual: false
125
+ discount: 0.9
126
+ dataset:
127
+ _target_: llmbc.dataset.metaworld_lowdim_dataset.MetaworldLowdimDataset
128
+ data_path: datasets/box-close-v2-o4-mini-split0.05.pt
129
+ data_path2: datasets/box-close-v2-o4-mini-split0.05.pt
130
+ horizon: ${horizon}
131
+ pad_before: ${eval:'${n_obs_steps}-1'}
132
+ pad_after: ${eval:'${n_action_steps}-1'}
133
+ obs_eef_target: true
134
+ use_manual_normalizer: false
135
+ val_ratio: 0.1
136
+ dummy_normalizer: true
137
+ instructor:
138
+ _target_: llmbc.translator.instructor.metaworld_instructor.box_close_v2_instructor.BoxCloseV2Instructor
139
+ llm:
140
+ name: HuggingFaceTB/SmolLM2-135M-Instruct
141
+ model_name: SmolLM2-135M-Instruct
142
+ config_target: llmbc.model.llm.llama_lowdim_model.LowdimLlamaConfig
143
+ causal_lm_target: llmbc.model.llm.llama_lowdim_model.LowdimLlamaForCausalLM
144
+ use_quantization: false
145
+ use_joint_mlp_projector: true
146
+ llm_mode: ete-finetuned
147
+ finetune_mode: orig
148
+ checkpoint: data/outputs/2026.01.26/10.26.52_train_llm_lowdim_box-close-v2/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-box-close-v2/checkpoint-6000
149
+ max_length: 100
150
+ lora_config:
151
+ r: 32
152
+ lora_alpha: 64
153
+ lora_dropout: 0.05
154
+ bias: none
155
+ task_type: CAUSAL_LM
156
+ prompter:
157
+ _target_: llmbc.translator.prompter.smollm2_prompter.SmolLM2Prompter
158
+ use_joint_mlp_projector: true
159
+ hydra:
160
+ job:
161
+ override_dirname: ${model_name}
162
+ run:
163
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${model_name}
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/.hydra/hydra.yaml ADDED
@@ -0,0 +1,155 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
4
+ sweep:
5
+ dir: data/outputs/${now:%Y.%m.%d}/${now:%H.%M.%S}_${name}_${task_name}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - training.seed=42
116
+ job:
117
+ name: train
118
+ chdir: null
119
+ override_dirname: training.seed=42
120
+ id: ???
121
+ num: ???
122
+ config_name: llmbc_box-close-v2_o4-mini_split0.05.yaml
123
+ env_set: {}
124
+ env_copy: []
125
+ config:
126
+ override_dirname:
127
+ kv_sep: '='
128
+ item_sep: ','
129
+ exclude_keys: []
130
+ runtime:
131
+ version: 1.2.0
132
+ version_base: '1.2'
133
+ cwd: /work/u1131674/LLM-BC
134
+ config_sources:
135
+ - path: hydra.conf
136
+ schema: pkg
137
+ provider: hydra
138
+ - path: /work/u1131674/LLM-BC/config/data_efficiency
139
+ schema: file
140
+ provider: main
141
+ - path: ''
142
+ schema: structured
143
+ provider: schema
144
+ output_dir: /work/u1131674/LLM-BC/data/outputs/2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2
145
+ choices:
146
+ hydra/env: default
147
+ hydra/callbacks: null
148
+ hydra/job_logging: default
149
+ hydra/hydra_logging: default
150
+ hydra/hydra_help: default
151
+ hydra/help: default
152
+ hydra/sweeper: basic
153
+ hydra/launcher: basic
154
+ hydra/output: default
155
+ verbose: false
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/.hydra/overrides.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ - training.seed=42
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=10200-test_success_rate=0.780.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8c36a870df18b492f96b325312fc3238fce7982d9fc259e8970d1179c41132c
3
+ size 864584
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=11400-test_success_rate=0.800.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bd584f31532b75397ab25728c564c692c62f02ca3fa992b538b3ca13a1a0090
3
+ size 864584
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=12200-test_success_rate=0.800.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b7e34cb6bba63eff86f3cf8878f4a0932b7f06f56f4c8d3cde6182a3c92d528
3
+ size 864584
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=12300-test_success_rate=0.780.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3392da6ccd4ac2d41f68d8f6377794f15554518a67fdb6a0d3a8052508e8b4a2
3
+ size 864584
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/checkpoints/epoch=7400-test_success_rate=0.780.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0e75ab803b9efb1e8638d92610ea02fba48a06fd5506a581fa1879d31c77210
3
+ size 864584
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/checkpoints/latest.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ab07a2c125d3a3be246065ef60aaf6cfafa80726d8e1b3fe770559db36dbc0e
3
+ size 864584
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/logs.json.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25aa1e99560ada5a21c6532ce1106f103ee143edcc7b42ef3f34dd1d9f1ea0bb
3
+ size 55577527
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/train.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ [2026-01-27 11:59:33,618][numexpr.utils][INFO] - Note: detected 112 virtual cores but NumExpr set to maximum of 64, check "NUMEXPR_MAX_THREADS" environment variable.
2
+ [2026-01-27 11:59:33,618][numexpr.utils][INFO] - Note: NumExpr detected 112 cores but "NUMEXPR_MAX_THREADS" not set, so enforcing safe limit of 16.
3
+ [2026-01-27 11:59:33,618][numexpr.utils][INFO] - NumExpr defaulting to 16 threads.
4
+ [2026-01-27 11:59:38,326][datasets][INFO] - PyTorch version 2.2.2 available.
5
+ [2026-01-27 11:59:38,327][datasets][INFO] - TensorFlow version 2.15.1 available.
6
+ [2026-01-27 11:59:38,327][datasets][INFO] - JAX version 0.4.30 available.
7
+ [2026-01-27 11:59:45,313][absl][INFO] - MUJOCO_GL=osmesa, attempting to import specified OpenGL backend.
8
+ [2026-01-27 11:59:45,322][absl][INFO] - MuJoCo library version is: 2.3.7
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/debug-internal.log ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-01-27T11:59:46.233224045+08:00","level":"INFO","msg":"using version","core version":"0.18.6"}
2
+ {"time":"2026-01-27T11:59:46.233232608+08:00","level":"INFO","msg":"created symlink","path":"/work/u1131674/LLM-BC/data/outputs/2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/logs/debug-core.log"}
3
+ {"time":"2026-01-27T11:59:46.347396478+08:00","level":"INFO","msg":"created new stream","id":"ccfd9xxl"}
4
+ {"time":"2026-01-27T11:59:46.347428899+08:00","level":"INFO","msg":"stream: started","id":"ccfd9xxl"}
5
+ {"time":"2026-01-27T11:59:46.347452905+08:00","level":"INFO","msg":"sender: started","stream_id":"ccfd9xxl"}
6
+ {"time":"2026-01-27T11:59:46.347448548+08:00","level":"INFO","msg":"handler: started","stream_id":{"value":"ccfd9xxl"}}
7
+ {"time":"2026-01-27T11:59:46.347443454+08:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"ccfd9xxl"}}
8
+ {"time":"2026-01-27T11:59:47.036042163+08:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2026-01-27T22:47:47.651604587+08:00","level":"INFO","msg":"stream: closing","id":"ccfd9xxl"}
10
+ {"time":"2026-01-27T22:47:47.651640552+08:00","level":"INFO","msg":"Stopping system monitor"}
11
+ {"time":"2026-01-27T22:47:47.65220418+08:00","level":"INFO","msg":"Stopped system monitor"}
12
+ {"time":"2026-01-27T22:47:48.903771416+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
13
+ {"time":"2026-01-27T22:47:49.193540342+08:00","level":"INFO","msg":"handler: closed","stream_id":{"value":"ccfd9xxl"}}
14
+ {"time":"2026-01-27T22:47:49.19356658+08:00","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"ccfd9xxl"}}
15
+ {"time":"2026-01-27T22:47:49.193600819+08:00","level":"INFO","msg":"sender: closed","stream_id":"ccfd9xxl"}
16
+ {"time":"2026-01-27T22:47:49.19361559+08:00","level":"INFO","msg":"stream: closed","id":"ccfd9xxl"}
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/debug.log ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-01-27 11:59:46,229 INFO MainThread:888299 [wandb_setup.py:_flush():79] Current SDK version is 0.18.6
2
+ 2026-01-27 11:59:46,229 INFO MainThread:888299 [wandb_setup.py:_flush():79] Configure stats pid to 888299
3
+ 2026-01-27 11:59:46,229 INFO MainThread:888299 [wandb_setup.py:_flush():79] Loading settings from /home/u1131674/.config/wandb/settings
4
+ 2026-01-27 11:59:46,229 INFO MainThread:888299 [wandb_setup.py:_flush():79] Loading settings from /work/u1131674/LLM-BC/wandb/settings
5
+ 2026-01-27 11:59:46,229 INFO MainThread:888299 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2026-01-27 11:59:46,229 INFO MainThread:888299 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': 'online', '_disable_service': None}
7
+ 2026-01-27 11:59:46,229 INFO MainThread:888299 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train.py', 'program_abspath': '/work/u1131674/LLM-BC/train.py', 'program': '/work/u1131674/LLM-BC/./train.py'}
8
+ 2026-01-27 11:59:46,229 INFO MainThread:888299 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2026-01-27 11:59:46,229 INFO MainThread:888299 [wandb_init.py:_log_setup():533] Logging user logs to /work/u1131674/LLM-BC/data/outputs/2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/logs/debug.log
10
+ 2026-01-27 11:59:46,229 INFO MainThread:888299 [wandb_init.py:_log_setup():534] Logging internal logs to /work/u1131674/LLM-BC/data/outputs/2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/logs/debug-internal.log
11
+ 2026-01-27 11:59:46,230 INFO MainThread:888299 [wandb_init.py:init():619] calling init triggers
12
+ 2026-01-27 11:59:46,230 INFO MainThread:888299 [wandb_init.py:init():626] wandb.init called with sweep_config: {}
13
+ config: {'name': 'train_llmbc_lowdim', '_target_': 'llmbc.workspace.train_llmbc_lowdim_workspace.TrainLLMBCLowdimWorkspace', 'obs_dim': 9, 'action_dim': 4, 'task_name': 'box-close-v2', 'exp_name': 'default', 'model_name': 'HuggingFaceTB/SmolLM2-135M-Instruct', 'horizon': 1, 'n_obs_steps': 1, 'n_action_steps': 1, 'n_latency_steps': 0, 'past_action_visible': False, 'llm_orig_expert_feedback': True, 'llm_do_sample': False, 'policy': {'_target_': 'llmbc.policy.llmbc_lowdim_policy.LLMBCLowdimPolicy', 'model': {'_target_': 'llmbc.model.policy.policy_mlp.PolicyMLP', 'input_size': 9, 'hidden_size': [256, 256], 'output_size': 4, 'activation': 'relu', 'n_obs_steps': 1, 'n_action_steps': 1}, 'obs_dim': 9, 'action_dim': 4, 'llm_discriminator': {'_target_': 'llmbc.discriminator.llm_ce_discriminator.LLMCEDiscriminator', 'task_id': 'box-close-v2', 'llm_translator': {'_target_': 'llmbc.translator.llm_translator.LLMTranslator', 'cfg': {'name': 'HuggingFaceTB/SmolLM2-135M-Instruct', 'model_name': 'SmolLM2-135M-Instruct', 'config_target': 'llmbc.model.llm.llama_lowdim_model.LowdimLlamaConfig', 'causal_lm_target': 'llmbc.model.llm.llama_lowdim_model.LowdimLlamaForCausalLM', 'use_quantization': False, 'use_joint_mlp_projector': True, 'llm_mode': 'ete-finetuned', 'finetune_mode': 'orig', 'checkpoint': 'data/outputs/2026.01.26/10.26.52_train_llm_lowdim_box-close-v2/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-box-close-v2/checkpoint-6000', 'max_length': 100, 'lora_config': {'r': 32, 'lora_alpha': 64, 'lora_dropout': 0.05, 'bias': 'none', 'task_type': 'CAUSAL_LM'}, 'prompter': {'_target_': 'llmbc.translator.prompter.smollm2_prompter.SmolLM2Prompter', 'use_joint_mlp_projector': True}, 'hydra': {'job': {'override_dirname': 'HuggingFaceTB/SmolLM2-135M-Instruct'}, 'run': {'dir': 'data/outputs/2026.01.27/11.59.31_HuggingFaceTB/SmolLM2-135M-Instruct'}}}, 'obs_dim': 9, 'action_dim': 4, 'horizon': 1, 'n_obs_steps': 1, 'n_action_steps': 1}}, 'loss_bc_weight': 1.0, 'loss_llm_weight': 0.001, 'horizon': 1, 'n_obs_steps': 1, 'n_action_steps': 1, 'normalize_llm_loss': True}, 'dataloader': {'batch_size': 16, 'num_workers': 0, 'shuffle': True, 'pin_memory': False, 'persistent_workers': False}, 'val_dataloader': {'batch_size': 16, 'num_workers': 0, 'shuffle': True, 'pin_memory': False, 'persistent_workers': False}, 'optimizer': {'_target_': 'torch.optim.AdamW', 'lr': 0.01, 'betas': [0.95, 0.999], 'eps': 1e-08, 'weight_decay': 1e-06}, 'training': {'device': 'cuda:0', 'seed': 42, 'debug': False, 'resume': False, 'lr_scheduler': 'cosine', 'lr_warmup_steps': 10, 'num_epochs': 20001, 'gradient_accumulate_every': 8, 'grad_norm_clip': 0.5, 'rollout_every': 100, 'checkpoint_every': 100, 'val_every': 1, 'sample_every': 100, 'sample_max_batch': 128, 'max_train_steps': None, 'max_val_steps': None, 'tqdm_interval_sec': 1.0}, 'logging': {'project': 'box-close-v2-training-split-ablation', 'resume': True, 'mode': 'online', 'name': '2026.01.27-11.59.31_train_llmbc_lowdim_box-close-v2', 'tags': ['train_llmbc_lowdim', 'box-close-v2', 'default'], 'id': None, 'group': None}, 'checkpoint': {'topk': {'monitor_key': 'test_success_rate', 'mode': 'max', 'k': 5, 'format_str': 'epoch={epoch:04d}-test_success_rate={test_success_rate:.3f}.ckpt'}, 'save_last_ckpt': True, 'save_last_snapshot': False}, 'multi_run': {'run_dir': 'data/outputs/2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2', 'wandb_name_base': '2026.01.27-11.59.31_train_llmbc_lowdim_box-close-v2'}, 'task': {'name': 'box-close-v2', 'obs_dim': 9, 'action_dim': 4, 'env_runner': {'_target_': 'llmbc.env_runner.metaworld_lowdim_runner.MetaworldLowdimRunner', 'env_name': 'llf-metaworld-box-close-v2', 'n_train': 10, 'n_test': 50, 'n_envs': 10, 'max_steps': 30, 'n_obs_steps': 1, 'n_action_steps': 1, 'instruction_type': 'b', 'feedback_type': ['hp', 'hn', 'fp'], 'visual': False, 'discount': 0.9}, 'dataset': {'_target_': 'llmbc.dataset.metaworld_lowdim_dataset.MetaworldLowdimDataset', 'data_path': 'datasets/box-close-v2-o4-mini-split0.05.pt', 'data_path2': 'datasets/box-close-v2-o4-mini-split0.05.pt', 'horizon': 1, 'pad_before': 0, 'pad_after': 0, 'obs_eef_target': True, 'use_manual_normalizer': False, 'val_ratio': 0.1, 'dummy_normalizer': True}, 'instructor': {'_target_': 'llmbc.translator.instructor.metaworld_instructor.box_close_v2_instructor.BoxCloseV2Instructor'}}, 'llm': {'name': 'HuggingFaceTB/SmolLM2-135M-Instruct', 'model_name': 'SmolLM2-135M-Instruct', 'config_target': 'llmbc.model.llm.llama_lowdim_model.LowdimLlamaConfig', 'causal_lm_target': 'llmbc.model.llm.llama_lowdim_model.LowdimLlamaForCausalLM', 'use_quantization': False, 'use_joint_mlp_projector': True, 'llm_mode': 'ete-finetuned', 'finetune_mode': 'orig', 'checkpoint': 'data/outputs/2026.01.26/10.26.52_train_llm_lowdim_box-close-v2/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-box-close-v2/checkpoint-6000', 'max_length': 100, 'lora_config': {'r': 32, 'lora_alpha': 64, 'lora_dropout': 0.05, 'bias': 'none', 'task_type': 'CAUSAL_LM'}, 'prompter': {'_target_': 'llmbc.translator.prompter.smollm2_prompter.SmolLM2Prompter', 'use_joint_mlp_projector': True}, 'hydra': {'job': {'override_dirname': 'HuggingFaceTB/SmolLM2-135M-Instruct'}, 'run': {'dir': 'data/outputs/2026.01.27/11.59.31_HuggingFaceTB/SmolLM2-135M-Instruct'}}}}
14
+ 2026-01-27 11:59:46,230 INFO MainThread:888299 [wandb_init.py:init():669] starting backend
15
+ 2026-01-27 11:59:46,230 INFO MainThread:888299 [wandb_init.py:init():673] sending inform_init request
16
+ 2026-01-27 11:59:46,231 INFO MainThread:888299 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2026-01-27 11:59:46,231 INFO MainThread:888299 [wandb_init.py:init():686] backend started and connected
18
+ 2026-01-27 11:59:46,240 INFO MainThread:888299 [wandb_init.py:init():781] updated telemetry
19
+ 2026-01-27 11:59:46,295 INFO MainThread:888299 [wandb_init.py:init():814] communicating run to backend with 90.0 second timeout
20
+ 2026-01-27 11:59:47,031 INFO MainThread:888299 [wandb_init.py:init():867] starting run threads in backend
21
+ 2026-01-27 11:59:47,512 INFO MainThread:888299 [wandb_run.py:_console_start():2451] atexit reg
22
+ 2026-01-27 11:59:47,512 INFO MainThread:888299 [wandb_run.py:_redirect():2299] redirect: wrap_raw
23
+ 2026-01-27 11:59:47,512 INFO MainThread:888299 [wandb_run.py:_redirect():2364] Wrapping output streams.
24
+ 2026-01-27 11:59:47,512 INFO MainThread:888299 [wandb_run.py:_redirect():2389] Redirects installed.
25
+ 2026-01-27 11:59:47,514 INFO MainThread:888299 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2026-01-27 11:59:47,514 INFO MainThread:888299 [wandb_run.py:_config_callback():1389] config_cb None None {'output_dir': '/work/u1131674/LLM-BC/data/outputs/2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2'}
27
+ 2026-01-27 22:47:47,651 WARNING MsgRouterThr:888299 [router.py:message_loop():75] message_loop has been closed
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/files/config.yaml ADDED
@@ -0,0 +1,271 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _target_:
2
+ value: llmbc.workspace.train_llmbc_lowdim_workspace.TrainLLMBCLowdimWorkspace
3
+ _wandb:
4
+ value:
5
+ cli_version: 0.18.6
6
+ m: []
7
+ python_version: 3.9.25
8
+ t:
9
+ "1":
10
+ - 1
11
+ - 2
12
+ - 3
13
+ - 5
14
+ - 11
15
+ - 12
16
+ - 41
17
+ - 49
18
+ - 50
19
+ - 51
20
+ - 53
21
+ - 55
22
+ - 71
23
+ - 83
24
+ - 95
25
+ - 98
26
+ - 100
27
+ - 105
28
+ "2":
29
+ - 1
30
+ - 2
31
+ - 3
32
+ - 5
33
+ - 11
34
+ - 12
35
+ - 41
36
+ - 49
37
+ - 50
38
+ - 51
39
+ - 53
40
+ - 55
41
+ - 71
42
+ - 83
43
+ - 95
44
+ - 98
45
+ - 100
46
+ - 105
47
+ "3":
48
+ - 13
49
+ - 15
50
+ - 16
51
+ - 23
52
+ - 55
53
+ - 61
54
+ "4": 3.9.25
55
+ "5": 0.18.6
56
+ "6": 4.47.1
57
+ "8":
58
+ - 5
59
+ "12": 0.18.6
60
+ "13": linux-x86_64
61
+ action_dim:
62
+ value: 4
63
+ checkpoint:
64
+ value:
65
+ save_last_ckpt: true
66
+ save_last_snapshot: false
67
+ topk:
68
+ format_str: epoch={epoch:04d}-test_success_rate={test_success_rate:.3f}.ckpt
69
+ k: 5
70
+ mode: max
71
+ monitor_key: test_success_rate
72
+ dataloader:
73
+ value:
74
+ batch_size: 16
75
+ num_workers: 0
76
+ persistent_workers: false
77
+ pin_memory: false
78
+ shuffle: true
79
+ exp_name:
80
+ value: default
81
+ horizon:
82
+ value: 1
83
+ llm:
84
+ value:
85
+ causal_lm_target: llmbc.model.llm.llama_lowdim_model.LowdimLlamaForCausalLM
86
+ checkpoint: data/outputs/2026.01.26/10.26.52_train_llm_lowdim_box-close-v2/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-box-close-v2/checkpoint-6000
87
+ config_target: llmbc.model.llm.llama_lowdim_model.LowdimLlamaConfig
88
+ finetune_mode: orig
89
+ hydra:
90
+ job:
91
+ override_dirname: HuggingFaceTB/SmolLM2-135M-Instruct
92
+ run:
93
+ dir: data/outputs/2026.01.27/11.59.31_HuggingFaceTB/SmolLM2-135M-Instruct
94
+ llm_mode: ete-finetuned
95
+ lora_config:
96
+ bias: none
97
+ lora_alpha: 64
98
+ lora_dropout: 0.05
99
+ r: 32
100
+ task_type: CAUSAL_LM
101
+ max_length: 100
102
+ model_name: SmolLM2-135M-Instruct
103
+ name: HuggingFaceTB/SmolLM2-135M-Instruct
104
+ prompter:
105
+ _target_: llmbc.translator.prompter.smollm2_prompter.SmolLM2Prompter
106
+ use_joint_mlp_projector: true
107
+ use_joint_mlp_projector: true
108
+ use_quantization: false
109
+ llm_do_sample:
110
+ value: false
111
+ llm_orig_expert_feedback:
112
+ value: true
113
+ logging:
114
+ value:
115
+ group: null
116
+ id: null
117
+ mode: online
118
+ name: 2026.01.27-11.59.31_train_llmbc_lowdim_box-close-v2
119
+ project: box-close-v2-training-split-ablation
120
+ resume: true
121
+ tags:
122
+ - train_llmbc_lowdim
123
+ - box-close-v2
124
+ - default
125
+ model_name:
126
+ value: HuggingFaceTB/SmolLM2-135M-Instruct
127
+ multi_run:
128
+ value:
129
+ run_dir: data/outputs/2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2
130
+ wandb_name_base: 2026.01.27-11.59.31_train_llmbc_lowdim_box-close-v2
131
+ n_action_steps:
132
+ value: 1
133
+ n_latency_steps:
134
+ value: 0
135
+ n_obs_steps:
136
+ value: 1
137
+ name:
138
+ value: train_llmbc_lowdim
139
+ obs_dim:
140
+ value: 9
141
+ optimizer:
142
+ value:
143
+ _target_: torch.optim.AdamW
144
+ betas:
145
+ - 0.95
146
+ - 0.999
147
+ eps: 1e-08
148
+ lr: 0.01
149
+ weight_decay: 1e-06
150
+ output_dir:
151
+ value: /work/u1131674/LLM-BC/data/outputs/2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2
152
+ past_action_visible:
153
+ value: false
154
+ policy:
155
+ value:
156
+ _target_: llmbc.policy.llmbc_lowdim_policy.LLMBCLowdimPolicy
157
+ action_dim: 4
158
+ horizon: 1
159
+ llm_discriminator:
160
+ _target_: llmbc.discriminator.llm_ce_discriminator.LLMCEDiscriminator
161
+ llm_translator:
162
+ _target_: llmbc.translator.llm_translator.LLMTranslator
163
+ action_dim: 4
164
+ cfg:
165
+ causal_lm_target: llmbc.model.llm.llama_lowdim_model.LowdimLlamaForCausalLM
166
+ checkpoint: data/outputs/2026.01.26/10.26.52_train_llm_lowdim_box-close-v2/HuggingFaceTB/SmolLM2-135M-Instruct-finetuned-box-close-v2/checkpoint-6000
167
+ config_target: llmbc.model.llm.llama_lowdim_model.LowdimLlamaConfig
168
+ finetune_mode: orig
169
+ hydra:
170
+ job:
171
+ override_dirname: HuggingFaceTB/SmolLM2-135M-Instruct
172
+ run:
173
+ dir: data/outputs/2026.01.27/11.59.31_HuggingFaceTB/SmolLM2-135M-Instruct
174
+ llm_mode: ete-finetuned
175
+ lora_config:
176
+ bias: none
177
+ lora_alpha: 64
178
+ lora_dropout: 0.05
179
+ r: 32
180
+ task_type: CAUSAL_LM
181
+ max_length: 100
182
+ model_name: SmolLM2-135M-Instruct
183
+ name: HuggingFaceTB/SmolLM2-135M-Instruct
184
+ prompter:
185
+ _target_: llmbc.translator.prompter.smollm2_prompter.SmolLM2Prompter
186
+ use_joint_mlp_projector: true
187
+ use_joint_mlp_projector: true
188
+ use_quantization: false
189
+ horizon: 1
190
+ n_action_steps: 1
191
+ n_obs_steps: 1
192
+ obs_dim: 9
193
+ task_id: box-close-v2
194
+ loss_bc_weight: 1
195
+ loss_llm_weight: 0.001
196
+ model:
197
+ _target_: llmbc.model.policy.policy_mlp.PolicyMLP
198
+ activation: relu
199
+ hidden_size:
200
+ - 256
201
+ - 256
202
+ input_size: 9
203
+ n_action_steps: 1
204
+ n_obs_steps: 1
205
+ output_size: 4
206
+ n_action_steps: 1
207
+ n_obs_steps: 1
208
+ normalize_llm_loss: true
209
+ obs_dim: 9
210
+ task:
211
+ value:
212
+ action_dim: 4
213
+ dataset:
214
+ _target_: llmbc.dataset.metaworld_lowdim_dataset.MetaworldLowdimDataset
215
+ data_path: datasets/box-close-v2-o4-mini-split0.05.pt
216
+ data_path2: datasets/box-close-v2-o4-mini-split0.05.pt
217
+ dummy_normalizer: true
218
+ horizon: 1
219
+ obs_eef_target: true
220
+ pad_after: 0
221
+ pad_before: 0
222
+ use_manual_normalizer: false
223
+ val_ratio: 0.1
224
+ env_runner:
225
+ _target_: llmbc.env_runner.metaworld_lowdim_runner.MetaworldLowdimRunner
226
+ discount: 0.9
227
+ env_name: llf-metaworld-box-close-v2
228
+ feedback_type:
229
+ - hp
230
+ - hn
231
+ - fp
232
+ instruction_type: b
233
+ max_steps: 30
234
+ n_action_steps: 1
235
+ n_envs: 10
236
+ n_obs_steps: 1
237
+ n_test: 50
238
+ n_train: 10
239
+ visual: false
240
+ instructor:
241
+ _target_: llmbc.translator.instructor.metaworld_instructor.box_close_v2_instructor.BoxCloseV2Instructor
242
+ name: box-close-v2
243
+ obs_dim: 9
244
+ task_name:
245
+ value: box-close-v2
246
+ training:
247
+ value:
248
+ checkpoint_every: 100
249
+ debug: false
250
+ device: cuda:0
251
+ grad_norm_clip: 0.5
252
+ gradient_accumulate_every: 8
253
+ lr_scheduler: cosine
254
+ lr_warmup_steps: 10
255
+ max_train_steps: null
256
+ max_val_steps: null
257
+ num_epochs: 20001
258
+ resume: false
259
+ rollout_every: 100
260
+ sample_every: 100
261
+ sample_max_batch: 128
262
+ seed: 42
263
+ tqdm_interval_sec: 1
264
+ val_every: 1
265
+ val_dataloader:
266
+ value:
267
+ batch_size: 16
268
+ num_workers: 0
269
+ persistent_workers: false
270
+ pin_memory: false
271
+ shuffle: true
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/files/output.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ Eval MetaworldLowdimRunner 1/6: 0%| | 0/30 [00:00<?, ?it/s]/work/u1131674/LLM-BC/llmbc/common/llfbench_util.py:39: UserWarning: Creating a tensor from a list of numpy.ndarrays is extremely slow. Please consider converting the list to a single numpy.ndarray with numpy.array() before converting to a tensor. (Triggered internally at ../torch/csrc/utils/tensor_new.cpp:275.)
2
+ obs = torch.tensor(obs, dtype=torch.float32).unsqueeze(dim=0).to(device)
3
+
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/files/wandb-metadata.json ADDED
@@ -0,0 +1,107 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-4.18.0-513.24.1.el8_9.x86_64-x86_64-with-glibc2.28",
3
+ "python": "3.9.25",
4
+ "startedAt": "2026-01-27T03:59:46.231929Z",
5
+ "args": [
6
+ "--config-path",
7
+ "config/data_efficiency",
8
+ "--config-name",
9
+ "llmbc_box-close-v2_o4-mini_split0.05.yaml",
10
+ "training.seed=42"
11
+ ],
12
+ "program": "/work/u1131674/LLM-BC/./train.py",
13
+ "codePath": "train.py",
14
+ "git": {
15
+ "remote": "https://github.com/CHYang25/LLM-BC.git",
16
+ "commit": "206b9c57fa687707dfa1ee0ebb33b507135a59c6"
17
+ },
18
+ "email": "chris920325@gmail.com",
19
+ "root": "/work/u1131674/LLM-BC/data/outputs/2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2",
20
+ "host": "hgpn05",
21
+ "username": "u1131674",
22
+ "executable": "/home/u1131674/.conda/envs/llm-bc/bin/python3",
23
+ "codePathLocal": "train.py",
24
+ "cpu_count": 112,
25
+ "cpu_count_logical": 112,
26
+ "gpu": "NVIDIA H100 80GB HBM3",
27
+ "gpu_count": 1,
28
+ "disk": {
29
+ "/": {
30
+ "total": "1918024196096",
31
+ "used": "437314154496"
32
+ }
33
+ },
34
+ "memory": {
35
+ "total": "2163617771520"
36
+ },
37
+ "cpu": {
38
+ "count": 112,
39
+ "countLogical": 112
40
+ },
41
+ "gpu_nvidia": [
42
+ {
43
+ "name": "NVIDIA H100 80GB HBM3",
44
+ "memoryTotal": "85520809984",
45
+ "cudaCores": 16896,
46
+ "architecture": "Hopper"
47
+ }
48
+ ],
49
+ "slurm": {
50
+ "cluster_name": "hpc",
51
+ "conf": "/etc/slurm/slurm.conf",
52
+ "cpu_bind": "quiet,mask_cpu:0x000000000FE04000000000000000",
53
+ "cpu_bind_list": "0x000000000FE04000000000000000",
54
+ "cpu_bind_type": "mask_cpu:",
55
+ "cpu_bind_verbose": "quiet",
56
+ "cpus_on_node": "8",
57
+ "cpus_per_task": "8",
58
+ "distribution": "cyclic,pack",
59
+ "gpus_on_node": "1",
60
+ "gpus_per_node": "1",
61
+ "gtids": "0",
62
+ "job_account": "mst114558",
63
+ "job_cpus_per_node": "8",
64
+ "job_end_time": "1769659170",
65
+ "job_gid": "106773",
66
+ "job_group": "MST114558",
67
+ "job_id": "104250",
68
+ "job_name": "python3",
69
+ "job_nodelist": "hgpn05",
70
+ "job_num_nodes": "1",
71
+ "job_partition": "normal",
72
+ "job_qos": "normal",
73
+ "job_start_time": "1769486370",
74
+ "job_uid": "41408",
75
+ "job_user": "u1131674",
76
+ "jobid": "104250",
77
+ "launch_node_ipaddr": "172.21.101.1",
78
+ "localid": "0",
79
+ "mem_per_node": "204800",
80
+ "nnodes": "1",
81
+ "nodeid": "0",
82
+ "nodelist": "hgpn05",
83
+ "nprocs": "1",
84
+ "ntasks": "1",
85
+ "prio_process": "0",
86
+ "procid": "0",
87
+ "srun_comm_host": "172.21.101.1",
88
+ "srun_comm_port": "43119",
89
+ "step_gpus": "5",
90
+ "step_id": "0",
91
+ "step_launcher_port": "43119",
92
+ "step_nodelist": "hgpn05",
93
+ "step_num_nodes": "1",
94
+ "step_num_tasks": "1",
95
+ "step_tasks_per_node": "1",
96
+ "stepid": "0",
97
+ "submit_dir": "/work/u1131674/LLM-BC",
98
+ "submit_host": "cbi-lgn01",
99
+ "task_pid": "888299",
100
+ "tasks_per_node": "1",
101
+ "topology_addr": "ibsw1.hgpn05",
102
+ "topology_addr_pattern": "switch.node",
103
+ "tres_per_task": "cpu:8",
104
+ "umask": "0022"
105
+ },
106
+ "cudaVersion": "12.4"
107
+ }
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/cumulative_reward":11.42187131867422,"_runtime":38881.419697085,"test/success_rate":0.6,"_wandb":{"runtime":38881},"train_action_mse_error":2.106957985859026e-08,"val_loss":0.00045819077058695257,"grad_norm":1.85015014722012e-05,"global_step":260012,"epoch":20000,"train_loss_llm":0,"test/mean_score":5.574101696679537,"test/cumulative_reward":12.715780319485676,"train_loss_bc":8.472071755249999e-08,"_step":260012,"_timestamp":1.7695252651366467e+09,"lr":2.337295357257574e-11,"train/success_rate":0.5,"train/mean_score":4.583878996955714,"train_loss":8.182333484267718e-08}
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/logs/debug-core.log ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-01-27T11:59:45.620925943+08:00","level":"INFO","msg":"started logging, with flags","port-filename":"/tmp/tmpipd6umd5/port-888299.txt","pid":888299,"debug":false,"disable-analytics":false}
2
+ {"time":"2026-01-27T11:59:45.620947818+08:00","level":"INFO","msg":"FeatureState","shutdownOnParentExitEnabled":false}
3
+ {"time":"2026-01-27T11:59:45.621295865+08:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":888299}
4
+ {"time":"2026-01-27T11:59:45.621289332+08:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":41057,"Zone":""}}
5
+ {"time":"2026-01-27T11:59:45.815580254+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:33124"}
6
+ {"time":"2026-01-27T11:59:46.232446283+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"ccfd9xxl","id":"127.0.0.1:33124"}
7
+ {"time":"2026-01-27T11:59:46.347434312+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"ccfd9xxl","id":"127.0.0.1:33124"}
8
+ {"time":"2026-01-27T22:47:47.651402799+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"127.0.0.1:33124"}
9
+ {"time":"2026-01-27T22:47:47.65160313+08:00","level":"INFO","msg":"connection: Close: initiating connection closure","id":"127.0.0.1:33124"}
10
+ {"time":"2026-01-27T22:47:47.651669651+08:00","level":"INFO","msg":"connection: Close: connection successfully closed","id":"127.0.0.1:33124"}
11
+ {"time":"2026-01-27T22:47:47.651612942+08:00","level":"INFO","msg":"server is shutting down"}
12
+ {"time":"2026-01-27T22:47:49.193798825+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"127.0.0.1:33124"}
13
+ {"time":"2026-01-27T22:47:49.193814452+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"127.0.0.1:33124"}
14
+ {"time":"2026-01-27T22:47:49.193826036+08:00","level":"INFO","msg":"server is closed"}
2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/logs/debug-internal.log ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-01-27T11:59:46.233224045+08:00","level":"INFO","msg":"using version","core version":"0.18.6"}
2
+ {"time":"2026-01-27T11:59:46.233232608+08:00","level":"INFO","msg":"created symlink","path":"/work/u1131674/LLM-BC/data/outputs/2026.01.27/11.59.31_train_llmbc_lowdim_box-close-v2/wandb/run-20260127_115946-ccfd9xxl/logs/debug-core.log"}
3
+ {"time":"2026-01-27T11:59:46.347396478+08:00","level":"INFO","msg":"created new stream","id":"ccfd9xxl"}
4
+ {"time":"2026-01-27T11:59:46.347428899+08:00","level":"INFO","msg":"stream: started","id":"ccfd9xxl"}
5
+ {"time":"2026-01-27T11:59:46.347452905+08:00","level":"INFO","msg":"sender: started","stream_id":"ccfd9xxl"}
6
+ {"time":"2026-01-27T11:59:46.347448548+08:00","level":"INFO","msg":"handler: started","stream_id":{"value":"ccfd9xxl"}}
7
+ {"time":"2026-01-27T11:59:46.347443454+08:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"ccfd9xxl"}}
8
+ {"time":"2026-01-27T11:59:47.036042163+08:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2026-01-27T22:47:47.651604587+08:00","level":"INFO","msg":"stream: closing","id":"ccfd9xxl"}
10
+ {"time":"2026-01-27T22:47:47.651640552+08:00","level":"INFO","msg":"Stopping system monitor"}
11
+ {"time":"2026-01-27T22:47:47.65220418+08:00","level":"INFO","msg":"Stopped system monitor"}
12
+ {"time":"2026-01-27T22:47:48.903771416+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
13
+ {"time":"2026-01-27T22:47:49.193540342+08:00","level":"INFO","msg":"handler: closed","stream_id":{"value":"ccfd9xxl"}}
14
+ {"time":"2026-01-27T22:47:49.19356658+08:00","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"ccfd9xxl"}}
15
+ {"time":"2026-01-27T22:47:49.193600819+08:00","level":"INFO","msg":"sender: closed","stream_id":"ccfd9xxl"}
16
+ {"time":"2026-01-27T22:47:49.19361559+08:00","level":"INFO","msg":"stream: closed","id":"ccfd9xxl"}