hangVLA commited on
Commit
b3d8aa7
·
verified ·
1 Parent(s): cdffc21

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ wandb/run-20251201_171106-oemlqork/run-oemlqork.wandb filter=lfs diff=lfs merge=lfs -text
wandb/debug-internal.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-12-01T17:11:06.83664393+09:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"}
2
+ {"time":"2025-12-01T17:11:07.27891855+09:00","level":"INFO","msg":"stream: created new stream","id":"oemlqork"}
3
+ {"time":"2025-12-01T17:11:07.278973759+09:00","level":"INFO","msg":"stream: started","id":"oemlqork"}
4
+ {"time":"2025-12-01T17:11:07.279001361+09:00","level":"INFO","msg":"handler: started","stream_id":"oemlqork"}
5
+ {"time":"2025-12-01T17:11:07.27901924+09:00","level":"INFO","msg":"sender: started","stream_id":"oemlqork"}
6
+ {"time":"2025-12-01T17:11:07.279151787+09:00","level":"INFO","msg":"writer: started","stream_id":"oemlqork"}
7
+ {"time":"2025-12-02T02:26:38.637293979+09:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"}
8
+ {"time":"2025-12-02T02:44:23.665913541+09:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"}
wandb/debug.log ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4
2
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Configure stats pid to 1280971
3
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Loading settings from /home/vla/.config/wandb/settings
4
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Loading settings from /home/vla/wandb/settings
5
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Loading settings from environment variables
6
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_init.py:setup_run_log_directory():686] Logging user logs to /home/vla/so101_2tasks_continue_pi0train/wandb/run-20251201_171106-oemlqork/logs/debug.log
7
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to /home/vla/so101_2tasks_continue_pi0train/wandb/run-20251201_171106-oemlqork/logs/debug-internal.log
8
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_init.py:init():813] calling init triggers
9
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_init.py:init():818] wandb.init called with sweep_config: {}
10
+ config: {'dataset': {'repo_id': '/home/vla/so101_2tasks_continue', 'root': None, 'episodes': None, 'image_transforms': {'enable': False, 'max_num_transforms': 3, 'random_order': False, 'tfs': {'brightness': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'brightness': [0.8, 1.2]}}, 'contrast': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'contrast': [0.8, 1.2]}}, 'saturation': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'saturation': [0.5, 1.5]}}, 'hue': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'hue': [-0.05, 0.05]}}, 'sharpness': {'weight': 1.0, 'type': 'SharpnessJitter', 'kwargs': {'sharpness': [0.5, 1.5]}}, 'affine': {'weight': 1.0, 'type': 'RandomAffine', 'kwargs': {'degrees': [-5.0, 5.0], 'translate': [0.05, 0.05]}}}}, 'revision': None, 'use_imagenet_stats': True, 'video_backend': 'torchcodec', 'streaming': False}, 'env': None, 'policy': {'type': 'pi0', 'n_obs_steps': 1, 'input_features': {}, 'output_features': {}, 'device': 'cuda', 'use_amp': False, 'push_to_hub': False, 'repo_id': None, 'private': None, 'tags': None, 'license': None, 'pretrained_path': 'lerobot/pi0_base', 'paligemma_variant': 'gemma_2b', 'action_expert_variant': 'gemma_300m', 'dtype': 'bfloat16', 'chunk_size': 50, 'n_action_steps': 50, 'max_state_dim': 32, 'max_action_dim': 32, 'num_inference_steps': 10, 'time_sampling_beta_alpha': 1.5, 'time_sampling_beta_beta': 1.0, 'time_sampling_scale': 0.999, 'time_sampling_offset': 0.001, 'min_period': 0.004, 'max_period': 4.0, 'image_resolution': [224, 224], 'empty_cameras': 0, 'normalization_mapping': {'VISUAL': <NormalizationMode.IDENTITY: 'IDENTITY'>, 'STATE': <NormalizationMode.MEAN_STD: 'MEAN_STD'>, 'ACTION': <NormalizationMode.MEAN_STD: 'MEAN_STD'>}, 'gradient_checkpointing': True, 'compile_model': True, 'compile_mode': 'max-autotune', 'optimizer_lr': 2.5e-05, 'optimizer_betas': [0.9, 0.95], 'optimizer_eps': 1e-08, 'optimizer_weight_decay': 0.01, 'optimizer_grad_clip_norm': 1.0, 'scheduler_warmup_steps': 1000, 'scheduler_decay_steps': 30000, 'scheduler_decay_lr': 2.5e-06, 'tokenizer_max_length': 48}, 'output_dir': '/home/vla/so101_2tasks_continue_pi0train', 'job_name': 'so101_2tasks_continue_pi0train', 'resume': False, 'seed': 1000, 'num_workers': 16, 'batch_size': 24, 'steps': 10000, 'eval_freq': 20000, 'log_freq': 200, 'save_checkpoint': True, 'save_freq': 20000, 'use_policy_training_preset': True, 'optimizer': {'type': 'adamw', 'lr': 2.5e-05, 'weight_decay': 0.01, 'grad_clip_norm': 1.0, 'betas': [0.9, 0.95], 'eps': 1e-08}, 'scheduler': {'type': 'cosine_decay_with_warmup', 'num_warmup_steps': 1000, 'num_decay_steps': 30000, 'peak_lr': 2.5e-05, 'decay_lr': 2.5e-06}, 'eval': {'n_episodes': 50, 'batch_size': 50, 'use_async_envs': False}, 'wandb': {'enable': True, 'disable_artifact': False, 'project': 'lerobot', 'entity': None, 'notes': None, 'run_id': None, 'mode': None}, 'checkpoint_path': None, 'rename_map': {'observation.images.camera1': 'observation.images.wrist', 'observation.images.camera2': 'observation.images.top', 'observation.images.camera3': 'observation.images.front'}, '_wandb': {}}
11
+ 2025-12-01 17:11:06,612 INFO MainThread:1280971 [wandb_init.py:init():854] starting backend
12
+ 2025-12-01 17:11:06,818 INFO MainThread:1280971 [wandb_init.py:init():857] sending inform_init request
13
+ 2025-12-01 17:11:06,826 INFO MainThread:1280971 [wandb_init.py:init():865] backend started and connected
14
+ 2025-12-01 17:11:06,831 INFO MainThread:1280971 [wandb_init.py:init():936] updated telemetry
15
+ 2025-12-01 17:11:06,833 INFO MainThread:1280971 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout
16
+ 2025-12-01 17:11:07,681 INFO MainThread:1280971 [wandb_init.py:init():1011] starting run threads in backend
17
+ 2025-12-01 17:11:07,793 INFO MainThread:1280971 [wandb_run.py:_console_start():2506] atexit reg
18
+ 2025-12-01 17:11:07,793 INFO MainThread:1280971 [wandb_run.py:_redirect():2354] redirect: wrap_raw
19
+ 2025-12-01 17:11:07,794 INFO MainThread:1280971 [wandb_run.py:_redirect():2423] Wrapping output streams.
20
+ 2025-12-01 17:11:07,794 INFO MainThread:1280971 [wandb_run.py:_redirect():2446] Redirects installed.
21
+ 2025-12-01 17:11:07,798 INFO MainThread:1280971 [wandb_init.py:init():1049] run started, returning control to user process
wandb/run-20251201_171106-oemlqork/files/output.log ADDED
@@ -0,0 +1,139 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ INFO 2025-12-01 17:11:07 db_utils.py:102 Logs will be synced with wandb.
2
+ INFO 2025-12-01 17:11:07 db_utils.py:103 Track this run --> https://wandb.ai/wjdfufgusrn1234-konkuk-university/lerobot/runs/oemlqork
3
+ INFO 2025-12-01 17:11:07 ot_train.py:183 Creating dataset
4
+ INFO 2025-12-01 17:11:08 ot_train.py:202 Creating policy
5
+ The PI0 model is a direct port of the OpenPI implementation.
6
+ This implementation follows the original OpenPI structure for compatibility.
7
+ Original implementation: https://github.com/Physical-Intelligence/openpi
8
+ INFO 2025-12-01 17:12:04 ling_pi0.py:553 Enabled gradient checkpointing for PI0Pytorch model
9
+ Loading model from: lerobot/pi0_base
10
+ ✓ Loaded state dict from model.safetensors
11
+ WARNING 2025-12-01 17:12:07 ing_pi0.py:1046 Vision embedding key might need handling: paligemma_with_expert.paligemma.model.vision_tower.vision_model.embeddings.patch_embedding.bias
12
+ WARNING 2025-12-01 17:12:07 ing_pi0.py:1046 Vision embedding key might need handling: paligemma_with_expert.paligemma.model.vision_tower.vision_model.embeddings.patch_embedding.weight
13
+ Remapped 777 state dict keys
14
+ Warning: Could not remap state dict keys: Error(s) in loading state_dict for PI0Policy:
15
+ Missing key(s) in state_dict: "model.paligemma_with_expert.paligemma.model.language_model.embed_tokens.weight".
16
+ INFO 2025-12-01 17:12:14 ot_train.py:247 Creating optimizer and scheduler
17
+ INFO 2025-12-01 17:12:14 hedulers.py:105 Auto-scaling LR scheduler: num_training_steps (10000) < num_decay_steps (30000). Scaling warmup: 1000 → 333, decay: 30000 → 10000 (scale factor: 0.333)
18
+ INFO 2025-12-01 17:12:14 ot_train.py:259 Output dir: /home/vla/so101_2tasks_continue_pi0train
19
+ INFO 2025-12-01 17:12:14 ot_train.py:262 cfg.steps=10000 (10K)
20
+ INFO 2025-12-01 17:12:14 ot_train.py:263 dataset.num_frames=106058 (106K)
21
+ INFO 2025-12-01 17:12:14 ot_train.py:264 dataset.num_episodes=64
22
+ INFO 2025-12-01 17:12:14 ot_train.py:267 Effective batch size: 24 x 2 = 48
23
+ INFO 2025-12-01 17:12:14 ot_train.py:268 num_learnable_params=3501372176 (4B)
24
+ INFO 2025-12-01 17:12:14 ot_train.py:269 num_total_params=3501372176 (4B)
25
+ INFO 2025-12-01 17:12:14 ot_train.py:324 Start offline training on a fixed dataset
26
+ [rank0]:W1201 17:12:22.196000 1280971 site-packages/torch/_logging/_internal.py:1130] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
27
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] Graph break from `Tensor.item()`, consider setting:
28
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] torch._dynamo.config.capture_scalar_outputs = True
29
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] or:
30
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1
31
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] to include these operations in the captured graph.
32
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0]
33
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] Graph break: from user code at:
34
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] File "/home/vla/miniforge3/envs/lerobot/lib/python3.10/site-packages/accelerate/utils/operations.py", line 819, in forward
35
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] return model_forward(*args, **kwargs)
36
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] File "/home/vla/miniforge3/envs/lerobot/lib/python3.10/site-packages/accelerate/utils/operations.py", line 807, in __call__
37
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] return convert_to_fp32(self.model_forward(*args, **kwargs))
38
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] File "/home/vla/miniforge3/envs/lerobot/lib/python3.10/site-packages/torch/amp/autocast_mode.py", line 44, in decorate_autocast
39
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] return func(*args, **kwargs)
40
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] File "/home/vla/lerobot/src/lerobot/policies/pi0/modeling_pi0.py", line 1188, in forward
41
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] "loss": loss.item(),
42
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0]
43
+ [rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0]
44
+ INFO 2025-12-01 17:30:19 ot_train.py:351 step:200 smpl:19K ep:12 epch:0.18 loss:0.404 grdn:5.408 lr:7.6e-06 updt_s:5.368 data_s:0.058
45
+ WARNING 2025-12-01 17:30:19 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
46
+ INFO 2025-12-01 17:44:56 ot_train.py:351 step:400 smpl:38K ep:23 epch:0.36 loss:0.177 grdn:2.557 lr:2.2e-05 updt_s:4.366 data_s:0.019
47
+ WARNING 2025-12-01 17:44:56 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
48
+ INFO 2025-12-01 17:59:35 ot_train.py:351 step:600 smpl:58K ep:35 epch:0.54 loss:0.157 grdn:1.521 lr:2.5e-05 updt_s:4.376 data_s:0.019
49
+ WARNING 2025-12-01 17:59:35 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
50
+ INFO 2025-12-01 18:14:13 ot_train.py:351 step:800 smpl:77K ep:46 epch:0.72 loss:0.141 grdn:1.311 lr:2.5e-05 updt_s:4.368 data_s:0.019
51
+ WARNING 2025-12-01 18:14:13 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
52
+ INFO 2025-12-01 18:28:50 ot_train.py:351 step:1K smpl:96K ep:58 epch:0.91 loss:0.133 grdn:1.178 lr:2.5e-05 updt_s:4.368 data_s:0.019
53
+ WARNING 2025-12-01 18:28:50 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
54
+ INFO 2025-12-01 18:43:28 ot_train.py:351 step:1K smpl:115K ep:70 epch:1.09 loss:0.126 grdn:1.096 lr:2.4e-05 updt_s:4.370 data_s:0.019
55
+ WARNING 2025-12-01 18:43:28 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
56
+ INFO 2025-12-01 18:58:07 ot_train.py:351 step:1K smpl:134K ep:81 epch:1.27 loss:0.118 grdn:1.055 lr:2.4e-05 updt_s:4.373 data_s:0.019
57
+ WARNING 2025-12-01 18:58:07 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
58
+ INFO 2025-12-01 19:12:46 ot_train.py:351 step:2K smpl:154K ep:93 epch:1.45 loss:0.119 grdn:1.016 lr:2.4e-05 updt_s:4.377 data_s:0.019
59
+ WARNING 2025-12-01 19:12:46 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
60
+ INFO 2025-12-01 19:27:23 ot_train.py:351 step:2K smpl:173K ep:104 epch:1.63 loss:0.121 grdn:0.998 lr:2.3e-05 updt_s:4.369 data_s:0.019
61
+ WARNING 2025-12-01 19:27:23 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
62
+ INFO 2025-12-01 19:42:02 ot_train.py:351 step:2K smpl:192K ep:116 epch:1.81 loss:0.111 grdn:0.974 lr:2.3e-05 updt_s:4.373 data_s:0.019
63
+ WARNING 2025-12-01 19:42:02 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
64
+ INFO 2025-12-01 19:56:39 ot_train.py:351 step:2K smpl:211K ep:127 epch:1.99 loss:0.110 grdn:0.907 lr:2.3e-05 updt_s:4.368 data_s:0.019
65
+ WARNING 2025-12-01 19:56:39 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
66
+ INFO 2025-12-01 20:11:22 ot_train.py:351 step:2K smpl:230K ep:139 epch:2.17 loss:0.108 grdn:0.931 lr:2.2e-05 updt_s:4.366 data_s:0.047
67
+ WARNING 2025-12-01 20:11:22 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
68
+ INFO 2025-12-01 20:26:00 ot_train.py:351 step:3K smpl:250K ep:151 epch:2.35 loss:0.112 grdn:0.943 lr:2.2e-05 updt_s:4.370 data_s:0.019
69
+ WARNING 2025-12-01 20:26:00 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
70
+ INFO 2025-12-01 20:40:37 ot_train.py:351 step:3K smpl:269K ep:162 epch:2.53 loss:0.105 grdn:0.912 lr:2.1e-05 updt_s:4.367 data_s:0.019
71
+ WARNING 2025-12-01 20:40:37 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
72
+ INFO 2025-12-01 20:55:14 ot_train.py:351 step:3K smpl:288K ep:174 epch:2.72 loss:0.105 grdn:0.906 lr:2.1e-05 updt_s:4.366 data_s:0.019
73
+ WARNING 2025-12-01 20:55:14 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
74
+ INFO 2025-12-01 21:09:52 ot_train.py:351 step:3K smpl:307K ep:185 epch:2.90 loss:0.103 grdn:0.906 lr:2.0e-05 updt_s:4.370 data_s:0.019
75
+ WARNING 2025-12-01 21:09:52 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
76
+ INFO 2025-12-01 21:24:29 ot_train.py:351 step:3K smpl:326K ep:197 epch:3.08 loss:0.102 grdn:0.910 lr:1.9e-05 updt_s:4.366 data_s:0.019
77
+ WARNING 2025-12-01 21:24:29 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
78
+ INFO 2025-12-01 21:39:07 ot_train.py:351 step:4K smpl:346K ep:209 epch:3.26 loss:0.095 grdn:0.912 lr:1.9e-05 updt_s:4.369 data_s:0.019
79
+ WARNING 2025-12-01 21:39:07 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
80
+ INFO 2025-12-01 21:53:45 ot_train.py:351 step:4K smpl:365K ep:220 epch:3.44 loss:0.097 grdn:0.914 lr:1.8e-05 updt_s:4.372 data_s:0.019
81
+ WARNING 2025-12-01 21:53:45 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
82
+ INFO 2025-12-01 22:08:22 ot_train.py:351 step:4K smpl:384K ep:232 epch:3.62 loss:0.101 grdn:0.901 lr:1.8e-05 updt_s:4.367 data_s:0.019
83
+ WARNING 2025-12-01 22:08:22 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
84
+ INFO 2025-12-01 22:23:02 ot_train.py:351 step:4K smpl:403K ep:243 epch:3.80 loss:0.092 grdn:0.894 lr:1.7e-05 updt_s:4.381 data_s:0.019
85
+ WARNING 2025-12-01 22:23:02 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
86
+ INFO 2025-12-01 22:37:42 ot_train.py:351 step:4K smpl:422K ep:255 epch:3.98 loss:0.094 grdn:0.924 lr:1.6e-05 updt_s:4.378 data_s:0.019
87
+ WARNING 2025-12-01 22:37:42 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
88
+ INFO 2025-12-01 22:52:28 ot_train.py:351 step:5K smpl:442K ep:266 epch:4.16 loss:0.092 grdn:0.889 lr:1.6e-05 updt_s:4.376 data_s:0.053
89
+ WARNING 2025-12-01 22:52:28 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
90
+ INFO 2025-12-01 23:07:05 ot_train.py:351 step:5K smpl:461K ep:278 epch:4.34 loss:0.087 grdn:0.876 lr:1.5e-05 updt_s:4.368 data_s:0.019
91
+ WARNING 2025-12-01 23:07:05 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
92
+ INFO 2025-12-01 23:21:43 ot_train.py:351 step:5K smpl:480K ep:290 epch:4.53 loss:0.097 grdn:0.939 lr:1.4e-05 updt_s:4.368 data_s:0.019
93
+ WARNING 2025-12-01 23:21:43 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
94
+ INFO 2025-12-01 23:36:21 ot_train.py:351 step:5K smpl:499K ep:301 epch:4.71 loss:0.087 grdn:0.929 lr:1.3e-05 updt_s:4.374 data_s:0.019
95
+ WARNING 2025-12-01 23:36:21 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
96
+ INFO 2025-12-01 23:50:59 ot_train.py:351 step:5K smpl:518K ep:313 epch:4.89 loss:0.093 grdn:0.927 lr:1.3e-05 updt_s:4.367 data_s:0.019
97
+ WARNING 2025-12-01 23:50:59 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
98
+ INFO 2025-12-02 00:05:36 ot_train.py:351 step:6K smpl:538K ep:324 epch:5.07 loss:0.094 grdn:0.919 lr:1.2e-05 updt_s:4.370 data_s:0.019
99
+ WARNING 2025-12-02 00:05:36 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
100
+ INFO 2025-12-02 00:20:14 ot_train.py:351 step:6K smpl:557K ep:336 epch:5.25 loss:0.088 grdn:0.919 lr:1.1e-05 updt_s:4.370 data_s:0.019
101
+ WARNING 2025-12-02 00:20:14 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
102
+ INFO 2025-12-02 00:34:56 ot_train.py:351 step:6K smpl:576K ep:348 epch:5.43 loss:0.091 grdn:0.974 lr:1.1e-05 updt_s:4.392 data_s:0.019
103
+ WARNING 2025-12-02 00:34:56 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
104
+ INFO 2025-12-02 00:49:34 ot_train.py:351 step:6K smpl:595K ep:359 epch:5.61 loss:0.086 grdn:0.978 lr:9.9e-06 updt_s:4.368 data_s:0.019
105
+ WARNING 2025-12-02 00:49:34 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
106
+ INFO 2025-12-02 01:04:13 ot_train.py:351 step:6K smpl:614K ep:371 epch:5.79 loss:0.084 grdn:0.954 lr:9.3e-06 updt_s:4.375 data_s:0.019
107
+ WARNING 2025-12-02 01:04:13 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
108
+ INFO 2025-12-02 01:18:51 ot_train.py:351 step:7K smpl:634K ep:382 epch:5.97 loss:0.086 grdn:0.992 lr:8.6e-06 updt_s:4.371 data_s:0.019
109
+ WARNING 2025-12-02 01:18:51 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
110
+ INFO 2025-12-02 01:33:35 ot_train.py:351 step:7K smpl:653K ep:394 epch:6.16 loss:0.084 grdn:0.977 lr:8.0e-06 updt_s:4.379 data_s:0.043
111
+ WARNING 2025-12-02 01:33:35 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
112
+ INFO 2025-12-02 01:48:13 ot_train.py:351 step:7K smpl:672K ep:406 epch:6.34 loss:0.088 grdn:1.017 lr:7.4e-06 updt_s:4.368 data_s:0.019
113
+ WARNING 2025-12-02 01:48:13 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
114
+ INFO 2025-12-02 02:02:51 ot_train.py:351 step:7K smpl:691K ep:417 epch:6.52 loss:0.086 grdn:0.992 lr:6.9e-06 updt_s:4.374 data_s:0.019
115
+ WARNING 2025-12-02 02:02:51 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
116
+ INFO 2025-12-02 02:17:31 ot_train.py:351 step:7K smpl:710K ep:429 epch:6.70 loss:0.087 grdn:0.996 lr:6.3e-06 updt_s:4.377 data_s:0.019
117
+ WARNING 2025-12-02 02:17:31 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
118
+ INFO 2025-12-02 02:32:10 ot_train.py:351 step:8K smpl:730K ep:440 epch:6.88 loss:0.082 grdn:1.009 lr:5.8e-06 updt_s:4.377 data_s:0.019
119
+ WARNING 2025-12-02 02:32:10 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
120
+ INFO 2025-12-02 02:46:48 ot_train.py:351 step:8K smpl:749K ep:452 epch:7.06 loss:0.082 grdn:1.006 lr:5.3e-06 updt_s:4.373 data_s:0.019
121
+ WARNING 2025-12-02 02:46:48 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
122
+ INFO 2025-12-02 03:01:26 ot_train.py:351 step:8K smpl:768K ep:463 epch:7.24 loss:0.082 grdn:1.021 lr:4.9e-06 updt_s:4.369 data_s:0.019
123
+ WARNING 2025-12-02 03:01:26 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
124
+ INFO 2025-12-02 03:16:03 ot_train.py:351 step:8K smpl:787K ep:475 epch:7.42 loss:0.086 grdn:1.056 lr:4.4e-06 updt_s:4.368 data_s:0.019
125
+ WARNING 2025-12-02 03:16:03 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
126
+ INFO 2025-12-02 03:30:42 ot_train.py:351 step:8K smpl:806K ep:487 epch:7.60 loss:0.088 grdn:1.056 lr:4.1e-06 updt_s:4.374 data_s:0.019
127
+ WARNING 2025-12-02 03:30:42 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
128
+ INFO 2025-12-02 03:45:19 ot_train.py:351 step:9K smpl:826K ep:498 epch:7.78 loss:0.079 grdn:1.042 lr:3.7e-06 updt_s:4.367 data_s:0.019
129
+ WARNING 2025-12-02 03:45:19 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
130
+ INFO 2025-12-02 03:59:57 ot_train.py:351 step:9K smpl:845K ep:510 epch:7.97 loss:0.081 grdn:1.035 lr:3.4e-06 updt_s:4.368 data_s:0.019
131
+ WARNING 2025-12-02 03:59:57 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
132
+ INFO 2025-12-02 04:14:40 ot_train.py:351 step:9K smpl:864K ep:521 epch:8.15 loss:0.084 grdn:1.041 lr:3.2e-06 updt_s:4.375 data_s:0.042
133
+ WARNING 2025-12-02 04:14:40 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
134
+ INFO 2025-12-02 04:29:18 ot_train.py:351 step:9K smpl:883K ep:533 epch:8.33 loss:0.083 grdn:1.072 lr:2.9e-06 updt_s:4.368 data_s:0.019
135
+ WARNING 2025-12-02 04:29:18 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
136
+ INFO 2025-12-02 04:43:56 ot_train.py:351 step:9K smpl:902K ep:545 epch:8.51 loss:0.083 grdn:1.084 lr:2.8e-06 updt_s:4.372 data_s:0.019
137
+ WARNING 2025-12-02 04:43:56 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
138
+ INFO 2025-12-02 04:58:33 ot_train.py:351 step:10K smpl:922K ep:556 epch:8.69 loss:0.088 grdn:1.093 lr:2.6e-06 updt_s:4.368 data_s:0.019
139
+ WARNING 2025-12-02 04:58:33 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
wandb/run-20251201_171106-oemlqork/files/requirements.txt ADDED
@@ -0,0 +1,133 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ h11==0.16.0
2
+ wcwidth==0.2.14
3
+ nvidia-cuda-cupti-cu12==12.6.80
4
+ async-timeout==5.0.1
5
+ dill==0.4.0
6
+ typing_extensions==4.15.0
7
+ pytz==2025.2
8
+ cloudpickle==3.1.2
9
+ pydantic==2.12.4
10
+ python-xlib==0.33
11
+ Jinja2==3.1.6
12
+ gmpy2==2.2.1
13
+ aiohappyeyeballs==2.6.1
14
+ einops==0.8.1
15
+ mpmath==1.3.0
16
+ hyperframe==6.1.0
17
+ smmap==5.0.2
18
+ mergedeep==1.3.4
19
+ Farama-Notifications==0.0.4
20
+ shellingham==1.5.4
21
+ yarl==1.22.0
22
+ peft==0.18.0
23
+ regex==2025.11.3
24
+ deepdiff==8.6.1
25
+ nvidia-cufile-cu12==1.11.1.6
26
+ certifi==2025.11.12
27
+ multiprocess==0.70.16
28
+ cmake==4.1.2
29
+ MarkupSafe==3.0.3
30
+ aiohttp==3.13.2
31
+ psutil==7.1.3
32
+ python-dateutil==2.9.0.post0
33
+ h2==4.3.0
34
+ httpcore==1.0.9
35
+ nvidia-cusparse-cu12==12.5.4.2
36
+ click==8.3.0
37
+ frozenlist==1.8.0
38
+ av==15.1.0
39
+ urllib3==2.5.0
40
+ sentry-sdk==2.44.0
41
+ pynput==1.8.1
42
+ pandas==2.3.3
43
+ triton==3.3.1
44
+ aiosignal==1.4.0
45
+ multidict==6.7.0
46
+ mypy_extensions==1.1.0
47
+ transformers==4.53.3
48
+ anyio==4.11.0
49
+ lerobot==0.4.1
50
+ zipp==3.23.0
51
+ platformdirs==4.5.0
52
+ pydantic_core==2.41.5
53
+ PyYAML==6.0.3
54
+ orderly-set==5.5.0
55
+ protobuf==6.33.0
56
+ typing-inspection==0.4.2
57
+ sympy==1.14.0
58
+ flash_attn==2.8.3
59
+ wheel==0.45.1
60
+ attrs==25.4.0
61
+ tqdm==4.67.1
62
+ nvidia-cudnn-cu12==9.5.1.17
63
+ Brotli==1.2.0
64
+ cffi==2.0.0
65
+ fsspec==2025.9.0
66
+ gitdb==4.0.12
67
+ importlib_metadata==8.7.0
68
+ nvidia-cusparselt-cu12==0.6.3
69
+ nvidia-cublas-cu12==12.6.4.1
70
+ opencv-python-headless==4.12.0.88
71
+ exceptiongroup==1.3.0
72
+ filelock==3.20.0
73
+ safetensors==0.6.2
74
+ pyserial==3.5
75
+ annotated-types==0.7.0
76
+ nvidia-cufft-cu12==11.3.0.4
77
+ pyyaml-include==1.4.1
78
+ PySocks==1.7.1
79
+ ImageIO==2.37.2
80
+ nvidia-cusolver-cu12==11.7.1.2
81
+ datasets==4.1.1
82
+ hf_transfer==0.1.9
83
+ networkx==3.4.2
84
+ docopt==0.6.2
85
+ tzdata==2025.2
86
+ num2words==0.5.14
87
+ hf-xet==1.2.0
88
+ nvidia-nccl-cu12==2.26.2
89
+ sniffio==1.3.1
90
+ huggingface-hub==0.35.3
91
+ xxhash==3.6.0
92
+ requests==2.32.5
93
+ idna==3.11
94
+ httpx==0.28.1
95
+ torch==2.7.1
96
+ typing-inspect==0.9.0
97
+ hpack==4.1.0
98
+ tokenizers==0.21.4
99
+ draccus==0.10.0
100
+ charset-normalizer==3.4.4
101
+ evdev==1.9.2
102
+ zstandard==0.25.0
103
+ pfzy==0.3.4
104
+ pillow==12.0.0
105
+ pycparser==2.22
106
+ torchvision==0.22.1
107
+ termcolor==3.2.0
108
+ wandb==0.21.4
109
+ nvidia-cuda-runtime-cu12==12.6.77
110
+ typer-slim==0.20.0
111
+ rerun-sdk==0.26.2
112
+ gymnasium==1.2.2
113
+ pyarrow==22.0.0
114
+ six==1.17.0
115
+ nvidia-nvjitlink-cu12==12.6.85
116
+ GitPython==3.1.45
117
+ prompt_toolkit==3.0.52
118
+ imageio-ffmpeg==0.6.0
119
+ nvidia-nvtx-cu12==12.6.77
120
+ torchcodec==0.5
121
+ nvidia-curand-cu12==10.3.7.77
122
+ diffusers==0.35.2
123
+ propcache==0.4.1
124
+ packaging==25.0
125
+ setuptools==80.9.0
126
+ jsonlines==4.0.0
127
+ numpy==2.2.6
128
+ pip==25.3
129
+ nvidia-cuda-nvrtc-cu12==12.6.77
130
+ accelerate==1.11.0
131
+ inquirerpy==0.3.4
132
+ toml==0.10.2
133
+ lerobot==0.4.1
wandb/run-20251201_171106-oemlqork/files/wandb-metadata.json ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.8.0-88-generic-x86_64-with-glibc2.39",
3
+ "python": "CPython 3.10.19",
4
+ "startedAt": "2025-12-01T08:11:06.610650Z",
5
+ "args": [
6
+ "--dataset.repo_id=/home/vla/so101_2tasks_continue",
7
+ "--policy.type=pi0",
8
+ "--policy.pretrained_path=lerobot/pi0_base",
9
+ "--output_dir=/home/vla/so101_2tasks_continue_pi0train",
10
+ "--job_name=so101_2tasks_continue_pi0train",
11
+ "--steps=10000",
12
+ "--batch_size=24",
13
+ "--num_workers=16",
14
+ "--policy.device=cuda",
15
+ "--wandb.enable=true",
16
+ "--policy.compile_model=true",
17
+ "--policy.gradient_checkpointing=true",
18
+ "--policy.dtype=bfloat16",
19
+ "--policy.push_to_hub=false",
20
+ "--rename_map={\"observation.images.camera1\": \"observation.images.wrist\", \"observation.images.camera2\": \"observation.images.top\", \"observation.images.camera3\": \"observation.images.front\"}"
21
+ ],
22
+ "program": "/home/vla/miniforge3/envs/lerobot/bin/lerobot-train",
23
+ "codePath": "miniforge3/envs/lerobot/bin/lerobot-train",
24
+ "codePathLocal": "miniforge3/envs/lerobot/bin/lerobot-train",
25
+ "root": "/home/vla/so101_2tasks_continue_pi0train",
26
+ "host": "edu02",
27
+ "executable": "/home/vla/miniforge3/envs/lerobot/bin/python3.10",
28
+ "cpu_count": 56,
29
+ "cpu_count_logical": 112,
30
+ "gpu": "NVIDIA RTX A6000",
31
+ "gpu_count": 2,
32
+ "disk": {
33
+ "/": {
34
+ "total": "943412031488",
35
+ "used": "811606069248"
36
+ }
37
+ },
38
+ "memory": {
39
+ "total": "540628525056"
40
+ },
41
+ "gpu_nvidia": [
42
+ {
43
+ "name": "NVIDIA RTX A6000",
44
+ "memoryTotal": "51527024640",
45
+ "cudaCores": 10752,
46
+ "architecture": "Ampere",
47
+ "uuid": "GPU-c895b718-8738-3fff-b42c-92cad284b785"
48
+ },
49
+ {
50
+ "name": "NVIDIA RTX A6000",
51
+ "memoryTotal": "51527024640",
52
+ "cudaCores": 10752,
53
+ "architecture": "Ampere",
54
+ "uuid": "GPU-34849e7e-d621-dccc-ca70-b463192b3472"
55
+ }
56
+ ],
57
+ "cudaVersion": "12.8",
58
+ "writerId": "dhfzrs4kx879100lp67b5x88soub6j73"
59
+ }
wandb/run-20251201_171106-oemlqork/logs/debug-core.log ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {"time":"2025-12-01T17:11:06.649227672+09:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpi7arovmq/port-1280971.txt","pid":1280971,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2025-12-01T17:11:06.650298036+09:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":1280971}
3
+ {"time":"2025-12-01T17:11:06.650164502+09:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-1280971-1280999-1612259489/socket","Net":"unix"}}
4
+ {"time":"2025-12-01T17:11:06.817876718+09:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2025-12-01T17:11:06.836443522+09:00","level":"INFO","msg":"handleInformInit: received","streamId":"oemlqork","id":"1(@)"}
6
+ {"time":"2025-12-01T17:11:07.278985483+09:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"oemlqork","id":"1(@)"}
wandb/run-20251201_171106-oemlqork/logs/debug-internal.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-12-01T17:11:06.83664393+09:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"}
2
+ {"time":"2025-12-01T17:11:07.27891855+09:00","level":"INFO","msg":"stream: created new stream","id":"oemlqork"}
3
+ {"time":"2025-12-01T17:11:07.278973759+09:00","level":"INFO","msg":"stream: started","id":"oemlqork"}
4
+ {"time":"2025-12-01T17:11:07.279001361+09:00","level":"INFO","msg":"handler: started","stream_id":"oemlqork"}
5
+ {"time":"2025-12-01T17:11:07.27901924+09:00","level":"INFO","msg":"sender: started","stream_id":"oemlqork"}
6
+ {"time":"2025-12-01T17:11:07.279151787+09:00","level":"INFO","msg":"writer: started","stream_id":"oemlqork"}
7
+ {"time":"2025-12-02T02:26:38.637293979+09:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"}
8
+ {"time":"2025-12-02T02:44:23.665913541+09:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"}
wandb/run-20251201_171106-oemlqork/logs/debug.log ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4
2
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Configure stats pid to 1280971
3
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Loading settings from /home/vla/.config/wandb/settings
4
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Loading settings from /home/vla/wandb/settings
5
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Loading settings from environment variables
6
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_init.py:setup_run_log_directory():686] Logging user logs to /home/vla/so101_2tasks_continue_pi0train/wandb/run-20251201_171106-oemlqork/logs/debug.log
7
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to /home/vla/so101_2tasks_continue_pi0train/wandb/run-20251201_171106-oemlqork/logs/debug-internal.log
8
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_init.py:init():813] calling init triggers
9
+ 2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_init.py:init():818] wandb.init called with sweep_config: {}
10
+ config: {'dataset': {'repo_id': '/home/vla/so101_2tasks_continue', 'root': None, 'episodes': None, 'image_transforms': {'enable': False, 'max_num_transforms': 3, 'random_order': False, 'tfs': {'brightness': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'brightness': [0.8, 1.2]}}, 'contrast': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'contrast': [0.8, 1.2]}}, 'saturation': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'saturation': [0.5, 1.5]}}, 'hue': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'hue': [-0.05, 0.05]}}, 'sharpness': {'weight': 1.0, 'type': 'SharpnessJitter', 'kwargs': {'sharpness': [0.5, 1.5]}}, 'affine': {'weight': 1.0, 'type': 'RandomAffine', 'kwargs': {'degrees': [-5.0, 5.0], 'translate': [0.05, 0.05]}}}}, 'revision': None, 'use_imagenet_stats': True, 'video_backend': 'torchcodec', 'streaming': False}, 'env': None, 'policy': {'type': 'pi0', 'n_obs_steps': 1, 'input_features': {}, 'output_features': {}, 'device': 'cuda', 'use_amp': False, 'push_to_hub': False, 'repo_id': None, 'private': None, 'tags': None, 'license': None, 'pretrained_path': 'lerobot/pi0_base', 'paligemma_variant': 'gemma_2b', 'action_expert_variant': 'gemma_300m', 'dtype': 'bfloat16', 'chunk_size': 50, 'n_action_steps': 50, 'max_state_dim': 32, 'max_action_dim': 32, 'num_inference_steps': 10, 'time_sampling_beta_alpha': 1.5, 'time_sampling_beta_beta': 1.0, 'time_sampling_scale': 0.999, 'time_sampling_offset': 0.001, 'min_period': 0.004, 'max_period': 4.0, 'image_resolution': [224, 224], 'empty_cameras': 0, 'normalization_mapping': {'VISUAL': <NormalizationMode.IDENTITY: 'IDENTITY'>, 'STATE': <NormalizationMode.MEAN_STD: 'MEAN_STD'>, 'ACTION': <NormalizationMode.MEAN_STD: 'MEAN_STD'>}, 'gradient_checkpointing': True, 'compile_model': True, 'compile_mode': 'max-autotune', 'optimizer_lr': 2.5e-05, 'optimizer_betas': [0.9, 0.95], 'optimizer_eps': 1e-08, 'optimizer_weight_decay': 0.01, 'optimizer_grad_clip_norm': 1.0, 'scheduler_warmup_steps': 1000, 'scheduler_decay_steps': 30000, 'scheduler_decay_lr': 2.5e-06, 'tokenizer_max_length': 48}, 'output_dir': '/home/vla/so101_2tasks_continue_pi0train', 'job_name': 'so101_2tasks_continue_pi0train', 'resume': False, 'seed': 1000, 'num_workers': 16, 'batch_size': 24, 'steps': 10000, 'eval_freq': 20000, 'log_freq': 200, 'save_checkpoint': True, 'save_freq': 20000, 'use_policy_training_preset': True, 'optimizer': {'type': 'adamw', 'lr': 2.5e-05, 'weight_decay': 0.01, 'grad_clip_norm': 1.0, 'betas': [0.9, 0.95], 'eps': 1e-08}, 'scheduler': {'type': 'cosine_decay_with_warmup', 'num_warmup_steps': 1000, 'num_decay_steps': 30000, 'peak_lr': 2.5e-05, 'decay_lr': 2.5e-06}, 'eval': {'n_episodes': 50, 'batch_size': 50, 'use_async_envs': False}, 'wandb': {'enable': True, 'disable_artifact': False, 'project': 'lerobot', 'entity': None, 'notes': None, 'run_id': None, 'mode': None}, 'checkpoint_path': None, 'rename_map': {'observation.images.camera1': 'observation.images.wrist', 'observation.images.camera2': 'observation.images.top', 'observation.images.camera3': 'observation.images.front'}, '_wandb': {}}
11
+ 2025-12-01 17:11:06,612 INFO MainThread:1280971 [wandb_init.py:init():854] starting backend
12
+ 2025-12-01 17:11:06,818 INFO MainThread:1280971 [wandb_init.py:init():857] sending inform_init request
13
+ 2025-12-01 17:11:06,826 INFO MainThread:1280971 [wandb_init.py:init():865] backend started and connected
14
+ 2025-12-01 17:11:06,831 INFO MainThread:1280971 [wandb_init.py:init():936] updated telemetry
15
+ 2025-12-01 17:11:06,833 INFO MainThread:1280971 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout
16
+ 2025-12-01 17:11:07,681 INFO MainThread:1280971 [wandb_init.py:init():1011] starting run threads in backend
17
+ 2025-12-01 17:11:07,793 INFO MainThread:1280971 [wandb_run.py:_console_start():2506] atexit reg
18
+ 2025-12-01 17:11:07,793 INFO MainThread:1280971 [wandb_run.py:_redirect():2354] redirect: wrap_raw
19
+ 2025-12-01 17:11:07,794 INFO MainThread:1280971 [wandb_run.py:_redirect():2423] Wrapping output streams.
20
+ 2025-12-01 17:11:07,794 INFO MainThread:1280971 [wandb_run.py:_redirect():2446] Redirects installed.
21
+ 2025-12-01 17:11:07,798 INFO MainThread:1280971 [wandb_init.py:init():1049] run started, returning control to user process
wandb/run-20251201_171106-oemlqork/run-oemlqork.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bd7a3d420c716066c9534fee57e2c06441ca20b97b2f78fe6dac95c050a8b7e
3
+ size 3375104