Upload folder using huggingface_hub
Browse files- .gitattributes +1 -0
- wandb/debug-internal.log +8 -0
- wandb/debug.log +21 -0
- wandb/run-20251201_171106-oemlqork/files/output.log +139 -0
- wandb/run-20251201_171106-oemlqork/files/requirements.txt +133 -0
- wandb/run-20251201_171106-oemlqork/files/wandb-metadata.json +59 -0
- wandb/run-20251201_171106-oemlqork/logs/debug-core.log +6 -0
- wandb/run-20251201_171106-oemlqork/logs/debug-internal.log +8 -0
- wandb/run-20251201_171106-oemlqork/logs/debug.log +21 -0
- wandb/run-20251201_171106-oemlqork/run-oemlqork.wandb +3 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
wandb/run-20251201_171106-oemlqork/run-oemlqork.wandb filter=lfs diff=lfs merge=lfs -text
|
wandb/debug-internal.log
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2025-12-01T17:11:06.83664393+09:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"}
|
| 2 |
+
{"time":"2025-12-01T17:11:07.27891855+09:00","level":"INFO","msg":"stream: created new stream","id":"oemlqork"}
|
| 3 |
+
{"time":"2025-12-01T17:11:07.278973759+09:00","level":"INFO","msg":"stream: started","id":"oemlqork"}
|
| 4 |
+
{"time":"2025-12-01T17:11:07.279001361+09:00","level":"INFO","msg":"handler: started","stream_id":"oemlqork"}
|
| 5 |
+
{"time":"2025-12-01T17:11:07.27901924+09:00","level":"INFO","msg":"sender: started","stream_id":"oemlqork"}
|
| 6 |
+
{"time":"2025-12-01T17:11:07.279151787+09:00","level":"INFO","msg":"writer: started","stream_id":"oemlqork"}
|
| 7 |
+
{"time":"2025-12-02T02:26:38.637293979+09:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"}
|
| 8 |
+
{"time":"2025-12-02T02:44:23.665913541+09:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"}
|
wandb/debug.log
ADDED
|
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4
|
| 2 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Configure stats pid to 1280971
|
| 3 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Loading settings from /home/vla/.config/wandb/settings
|
| 4 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Loading settings from /home/vla/wandb/settings
|
| 5 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Loading settings from environment variables
|
| 6 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_init.py:setup_run_log_directory():686] Logging user logs to /home/vla/so101_2tasks_continue_pi0train/wandb/run-20251201_171106-oemlqork/logs/debug.log
|
| 7 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to /home/vla/so101_2tasks_continue_pi0train/wandb/run-20251201_171106-oemlqork/logs/debug-internal.log
|
| 8 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_init.py:init():813] calling init triggers
|
| 9 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_init.py:init():818] wandb.init called with sweep_config: {}
|
| 10 |
+
config: {'dataset': {'repo_id': '/home/vla/so101_2tasks_continue', 'root': None, 'episodes': None, 'image_transforms': {'enable': False, 'max_num_transforms': 3, 'random_order': False, 'tfs': {'brightness': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'brightness': [0.8, 1.2]}}, 'contrast': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'contrast': [0.8, 1.2]}}, 'saturation': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'saturation': [0.5, 1.5]}}, 'hue': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'hue': [-0.05, 0.05]}}, 'sharpness': {'weight': 1.0, 'type': 'SharpnessJitter', 'kwargs': {'sharpness': [0.5, 1.5]}}, 'affine': {'weight': 1.0, 'type': 'RandomAffine', 'kwargs': {'degrees': [-5.0, 5.0], 'translate': [0.05, 0.05]}}}}, 'revision': None, 'use_imagenet_stats': True, 'video_backend': 'torchcodec', 'streaming': False}, 'env': None, 'policy': {'type': 'pi0', 'n_obs_steps': 1, 'input_features': {}, 'output_features': {}, 'device': 'cuda', 'use_amp': False, 'push_to_hub': False, 'repo_id': None, 'private': None, 'tags': None, 'license': None, 'pretrained_path': 'lerobot/pi0_base', 'paligemma_variant': 'gemma_2b', 'action_expert_variant': 'gemma_300m', 'dtype': 'bfloat16', 'chunk_size': 50, 'n_action_steps': 50, 'max_state_dim': 32, 'max_action_dim': 32, 'num_inference_steps': 10, 'time_sampling_beta_alpha': 1.5, 'time_sampling_beta_beta': 1.0, 'time_sampling_scale': 0.999, 'time_sampling_offset': 0.001, 'min_period': 0.004, 'max_period': 4.0, 'image_resolution': [224, 224], 'empty_cameras': 0, 'normalization_mapping': {'VISUAL': <NormalizationMode.IDENTITY: 'IDENTITY'>, 'STATE': <NormalizationMode.MEAN_STD: 'MEAN_STD'>, 'ACTION': <NormalizationMode.MEAN_STD: 'MEAN_STD'>}, 'gradient_checkpointing': True, 'compile_model': True, 'compile_mode': 'max-autotune', 'optimizer_lr': 2.5e-05, 'optimizer_betas': [0.9, 0.95], 'optimizer_eps': 1e-08, 'optimizer_weight_decay': 0.01, 'optimizer_grad_clip_norm': 1.0, 'scheduler_warmup_steps': 1000, 'scheduler_decay_steps': 30000, 'scheduler_decay_lr': 2.5e-06, 'tokenizer_max_length': 48}, 'output_dir': '/home/vla/so101_2tasks_continue_pi0train', 'job_name': 'so101_2tasks_continue_pi0train', 'resume': False, 'seed': 1000, 'num_workers': 16, 'batch_size': 24, 'steps': 10000, 'eval_freq': 20000, 'log_freq': 200, 'save_checkpoint': True, 'save_freq': 20000, 'use_policy_training_preset': True, 'optimizer': {'type': 'adamw', 'lr': 2.5e-05, 'weight_decay': 0.01, 'grad_clip_norm': 1.0, 'betas': [0.9, 0.95], 'eps': 1e-08}, 'scheduler': {'type': 'cosine_decay_with_warmup', 'num_warmup_steps': 1000, 'num_decay_steps': 30000, 'peak_lr': 2.5e-05, 'decay_lr': 2.5e-06}, 'eval': {'n_episodes': 50, 'batch_size': 50, 'use_async_envs': False}, 'wandb': {'enable': True, 'disable_artifact': False, 'project': 'lerobot', 'entity': None, 'notes': None, 'run_id': None, 'mode': None}, 'checkpoint_path': None, 'rename_map': {'observation.images.camera1': 'observation.images.wrist', 'observation.images.camera2': 'observation.images.top', 'observation.images.camera3': 'observation.images.front'}, '_wandb': {}}
|
| 11 |
+
2025-12-01 17:11:06,612 INFO MainThread:1280971 [wandb_init.py:init():854] starting backend
|
| 12 |
+
2025-12-01 17:11:06,818 INFO MainThread:1280971 [wandb_init.py:init():857] sending inform_init request
|
| 13 |
+
2025-12-01 17:11:06,826 INFO MainThread:1280971 [wandb_init.py:init():865] backend started and connected
|
| 14 |
+
2025-12-01 17:11:06,831 INFO MainThread:1280971 [wandb_init.py:init():936] updated telemetry
|
| 15 |
+
2025-12-01 17:11:06,833 INFO MainThread:1280971 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout
|
| 16 |
+
2025-12-01 17:11:07,681 INFO MainThread:1280971 [wandb_init.py:init():1011] starting run threads in backend
|
| 17 |
+
2025-12-01 17:11:07,793 INFO MainThread:1280971 [wandb_run.py:_console_start():2506] atexit reg
|
| 18 |
+
2025-12-01 17:11:07,793 INFO MainThread:1280971 [wandb_run.py:_redirect():2354] redirect: wrap_raw
|
| 19 |
+
2025-12-01 17:11:07,794 INFO MainThread:1280971 [wandb_run.py:_redirect():2423] Wrapping output streams.
|
| 20 |
+
2025-12-01 17:11:07,794 INFO MainThread:1280971 [wandb_run.py:_redirect():2446] Redirects installed.
|
| 21 |
+
2025-12-01 17:11:07,798 INFO MainThread:1280971 [wandb_init.py:init():1049] run started, returning control to user process
|
wandb/run-20251201_171106-oemlqork/files/output.log
ADDED
|
@@ -0,0 +1,139 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
INFO 2025-12-01 17:11:07 db_utils.py:102 [1m[34mLogs will be synced with wandb.[0m
|
| 2 |
+
INFO 2025-12-01 17:11:07 db_utils.py:103 Track this run --> [1m[33mhttps://wandb.ai/wjdfufgusrn1234-konkuk-university/lerobot/runs/oemlqork[0m
|
| 3 |
+
INFO 2025-12-01 17:11:07 ot_train.py:183 Creating dataset
|
| 4 |
+
INFO 2025-12-01 17:11:08 ot_train.py:202 Creating policy
|
| 5 |
+
The PI0 model is a direct port of the OpenPI implementation.
|
| 6 |
+
This implementation follows the original OpenPI structure for compatibility.
|
| 7 |
+
Original implementation: https://github.com/Physical-Intelligence/openpi
|
| 8 |
+
INFO 2025-12-01 17:12:04 ling_pi0.py:553 Enabled gradient checkpointing for PI0Pytorch model
|
| 9 |
+
Loading model from: lerobot/pi0_base
|
| 10 |
+
✓ Loaded state dict from model.safetensors
|
| 11 |
+
WARNING 2025-12-01 17:12:07 ing_pi0.py:1046 Vision embedding key might need handling: paligemma_with_expert.paligemma.model.vision_tower.vision_model.embeddings.patch_embedding.bias
|
| 12 |
+
WARNING 2025-12-01 17:12:07 ing_pi0.py:1046 Vision embedding key might need handling: paligemma_with_expert.paligemma.model.vision_tower.vision_model.embeddings.patch_embedding.weight
|
| 13 |
+
Remapped 777 state dict keys
|
| 14 |
+
Warning: Could not remap state dict keys: Error(s) in loading state_dict for PI0Policy:
|
| 15 |
+
Missing key(s) in state_dict: "model.paligemma_with_expert.paligemma.model.language_model.embed_tokens.weight".
|
| 16 |
+
INFO 2025-12-01 17:12:14 ot_train.py:247 Creating optimizer and scheduler
|
| 17 |
+
INFO 2025-12-01 17:12:14 hedulers.py:105 Auto-scaling LR scheduler: num_training_steps (10000) < num_decay_steps (30000). Scaling warmup: 1000 → 333, decay: 30000 → 10000 (scale factor: 0.333)
|
| 18 |
+
INFO 2025-12-01 17:12:14 ot_train.py:259 [1m[33mOutput dir:[0m /home/vla/so101_2tasks_continue_pi0train
|
| 19 |
+
INFO 2025-12-01 17:12:14 ot_train.py:262 cfg.steps=10000 (10K)
|
| 20 |
+
INFO 2025-12-01 17:12:14 ot_train.py:263 dataset.num_frames=106058 (106K)
|
| 21 |
+
INFO 2025-12-01 17:12:14 ot_train.py:264 dataset.num_episodes=64
|
| 22 |
+
INFO 2025-12-01 17:12:14 ot_train.py:267 Effective batch size: 24 x 2 = 48
|
| 23 |
+
INFO 2025-12-01 17:12:14 ot_train.py:268 num_learnable_params=3501372176 (4B)
|
| 24 |
+
INFO 2025-12-01 17:12:14 ot_train.py:269 num_total_params=3501372176 (4B)
|
| 25 |
+
INFO 2025-12-01 17:12:14 ot_train.py:324 Start offline training on a fixed dataset
|
| 26 |
+
[rank0]:W1201 17:12:22.196000 1280971 site-packages/torch/_logging/_internal.py:1130] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
|
| 27 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] Graph break from `Tensor.item()`, consider setting:
|
| 28 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] torch._dynamo.config.capture_scalar_outputs = True
|
| 29 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] or:
|
| 30 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1
|
| 31 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] to include these operations in the captured graph.
|
| 32 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0]
|
| 33 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] Graph break: from user code at:
|
| 34 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] File "/home/vla/miniforge3/envs/lerobot/lib/python3.10/site-packages/accelerate/utils/operations.py", line 819, in forward
|
| 35 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] return model_forward(*args, **kwargs)
|
| 36 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] File "/home/vla/miniforge3/envs/lerobot/lib/python3.10/site-packages/accelerate/utils/operations.py", line 807, in __call__
|
| 37 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] return convert_to_fp32(self.model_forward(*args, **kwargs))
|
| 38 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] File "/home/vla/miniforge3/envs/lerobot/lib/python3.10/site-packages/torch/amp/autocast_mode.py", line 44, in decorate_autocast
|
| 39 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] return func(*args, **kwargs)
|
| 40 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] File "/home/vla/lerobot/src/lerobot/policies/pi0/modeling_pi0.py", line 1188, in forward
|
| 41 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] "loss": loss.item(),
|
| 42 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0]
|
| 43 |
+
[rank0]:W1201 17:12:35.099000 1280971 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0]
|
| 44 |
+
INFO 2025-12-01 17:30:19 ot_train.py:351 step:200 smpl:19K ep:12 epch:0.18 loss:0.404 grdn:5.408 lr:7.6e-06 updt_s:5.368 data_s:0.058
|
| 45 |
+
WARNING 2025-12-01 17:30:19 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 46 |
+
INFO 2025-12-01 17:44:56 ot_train.py:351 step:400 smpl:38K ep:23 epch:0.36 loss:0.177 grdn:2.557 lr:2.2e-05 updt_s:4.366 data_s:0.019
|
| 47 |
+
WARNING 2025-12-01 17:44:56 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 48 |
+
INFO 2025-12-01 17:59:35 ot_train.py:351 step:600 smpl:58K ep:35 epch:0.54 loss:0.157 grdn:1.521 lr:2.5e-05 updt_s:4.376 data_s:0.019
|
| 49 |
+
WARNING 2025-12-01 17:59:35 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 50 |
+
INFO 2025-12-01 18:14:13 ot_train.py:351 step:800 smpl:77K ep:46 epch:0.72 loss:0.141 grdn:1.311 lr:2.5e-05 updt_s:4.368 data_s:0.019
|
| 51 |
+
WARNING 2025-12-01 18:14:13 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 52 |
+
INFO 2025-12-01 18:28:50 ot_train.py:351 step:1K smpl:96K ep:58 epch:0.91 loss:0.133 grdn:1.178 lr:2.5e-05 updt_s:4.368 data_s:0.019
|
| 53 |
+
WARNING 2025-12-01 18:28:50 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 54 |
+
INFO 2025-12-01 18:43:28 ot_train.py:351 step:1K smpl:115K ep:70 epch:1.09 loss:0.126 grdn:1.096 lr:2.4e-05 updt_s:4.370 data_s:0.019
|
| 55 |
+
WARNING 2025-12-01 18:43:28 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 56 |
+
INFO 2025-12-01 18:58:07 ot_train.py:351 step:1K smpl:134K ep:81 epch:1.27 loss:0.118 grdn:1.055 lr:2.4e-05 updt_s:4.373 data_s:0.019
|
| 57 |
+
WARNING 2025-12-01 18:58:07 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 58 |
+
INFO 2025-12-01 19:12:46 ot_train.py:351 step:2K smpl:154K ep:93 epch:1.45 loss:0.119 grdn:1.016 lr:2.4e-05 updt_s:4.377 data_s:0.019
|
| 59 |
+
WARNING 2025-12-01 19:12:46 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 60 |
+
INFO 2025-12-01 19:27:23 ot_train.py:351 step:2K smpl:173K ep:104 epch:1.63 loss:0.121 grdn:0.998 lr:2.3e-05 updt_s:4.369 data_s:0.019
|
| 61 |
+
WARNING 2025-12-01 19:27:23 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 62 |
+
INFO 2025-12-01 19:42:02 ot_train.py:351 step:2K smpl:192K ep:116 epch:1.81 loss:0.111 grdn:0.974 lr:2.3e-05 updt_s:4.373 data_s:0.019
|
| 63 |
+
WARNING 2025-12-01 19:42:02 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 64 |
+
INFO 2025-12-01 19:56:39 ot_train.py:351 step:2K smpl:211K ep:127 epch:1.99 loss:0.110 grdn:0.907 lr:2.3e-05 updt_s:4.368 data_s:0.019
|
| 65 |
+
WARNING 2025-12-01 19:56:39 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 66 |
+
INFO 2025-12-01 20:11:22 ot_train.py:351 step:2K smpl:230K ep:139 epch:2.17 loss:0.108 grdn:0.931 lr:2.2e-05 updt_s:4.366 data_s:0.047
|
| 67 |
+
WARNING 2025-12-01 20:11:22 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 68 |
+
INFO 2025-12-01 20:26:00 ot_train.py:351 step:3K smpl:250K ep:151 epch:2.35 loss:0.112 grdn:0.943 lr:2.2e-05 updt_s:4.370 data_s:0.019
|
| 69 |
+
WARNING 2025-12-01 20:26:00 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 70 |
+
INFO 2025-12-01 20:40:37 ot_train.py:351 step:3K smpl:269K ep:162 epch:2.53 loss:0.105 grdn:0.912 lr:2.1e-05 updt_s:4.367 data_s:0.019
|
| 71 |
+
WARNING 2025-12-01 20:40:37 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 72 |
+
INFO 2025-12-01 20:55:14 ot_train.py:351 step:3K smpl:288K ep:174 epch:2.72 loss:0.105 grdn:0.906 lr:2.1e-05 updt_s:4.366 data_s:0.019
|
| 73 |
+
WARNING 2025-12-01 20:55:14 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 74 |
+
INFO 2025-12-01 21:09:52 ot_train.py:351 step:3K smpl:307K ep:185 epch:2.90 loss:0.103 grdn:0.906 lr:2.0e-05 updt_s:4.370 data_s:0.019
|
| 75 |
+
WARNING 2025-12-01 21:09:52 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 76 |
+
INFO 2025-12-01 21:24:29 ot_train.py:351 step:3K smpl:326K ep:197 epch:3.08 loss:0.102 grdn:0.910 lr:1.9e-05 updt_s:4.366 data_s:0.019
|
| 77 |
+
WARNING 2025-12-01 21:24:29 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 78 |
+
INFO 2025-12-01 21:39:07 ot_train.py:351 step:4K smpl:346K ep:209 epch:3.26 loss:0.095 grdn:0.912 lr:1.9e-05 updt_s:4.369 data_s:0.019
|
| 79 |
+
WARNING 2025-12-01 21:39:07 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 80 |
+
INFO 2025-12-01 21:53:45 ot_train.py:351 step:4K smpl:365K ep:220 epch:3.44 loss:0.097 grdn:0.914 lr:1.8e-05 updt_s:4.372 data_s:0.019
|
| 81 |
+
WARNING 2025-12-01 21:53:45 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 82 |
+
INFO 2025-12-01 22:08:22 ot_train.py:351 step:4K smpl:384K ep:232 epch:3.62 loss:0.101 grdn:0.901 lr:1.8e-05 updt_s:4.367 data_s:0.019
|
| 83 |
+
WARNING 2025-12-01 22:08:22 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 84 |
+
INFO 2025-12-01 22:23:02 ot_train.py:351 step:4K smpl:403K ep:243 epch:3.80 loss:0.092 grdn:0.894 lr:1.7e-05 updt_s:4.381 data_s:0.019
|
| 85 |
+
WARNING 2025-12-01 22:23:02 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 86 |
+
INFO 2025-12-01 22:37:42 ot_train.py:351 step:4K smpl:422K ep:255 epch:3.98 loss:0.094 grdn:0.924 lr:1.6e-05 updt_s:4.378 data_s:0.019
|
| 87 |
+
WARNING 2025-12-01 22:37:42 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 88 |
+
INFO 2025-12-01 22:52:28 ot_train.py:351 step:5K smpl:442K ep:266 epch:4.16 loss:0.092 grdn:0.889 lr:1.6e-05 updt_s:4.376 data_s:0.053
|
| 89 |
+
WARNING 2025-12-01 22:52:28 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 90 |
+
INFO 2025-12-01 23:07:05 ot_train.py:351 step:5K smpl:461K ep:278 epch:4.34 loss:0.087 grdn:0.876 lr:1.5e-05 updt_s:4.368 data_s:0.019
|
| 91 |
+
WARNING 2025-12-01 23:07:05 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 92 |
+
INFO 2025-12-01 23:21:43 ot_train.py:351 step:5K smpl:480K ep:290 epch:4.53 loss:0.097 grdn:0.939 lr:1.4e-05 updt_s:4.368 data_s:0.019
|
| 93 |
+
WARNING 2025-12-01 23:21:43 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 94 |
+
INFO 2025-12-01 23:36:21 ot_train.py:351 step:5K smpl:499K ep:301 epch:4.71 loss:0.087 grdn:0.929 lr:1.3e-05 updt_s:4.374 data_s:0.019
|
| 95 |
+
WARNING 2025-12-01 23:36:21 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 96 |
+
INFO 2025-12-01 23:50:59 ot_train.py:351 step:5K smpl:518K ep:313 epch:4.89 loss:0.093 grdn:0.927 lr:1.3e-05 updt_s:4.367 data_s:0.019
|
| 97 |
+
WARNING 2025-12-01 23:50:59 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 98 |
+
INFO 2025-12-02 00:05:36 ot_train.py:351 step:6K smpl:538K ep:324 epch:5.07 loss:0.094 grdn:0.919 lr:1.2e-05 updt_s:4.370 data_s:0.019
|
| 99 |
+
WARNING 2025-12-02 00:05:36 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 100 |
+
INFO 2025-12-02 00:20:14 ot_train.py:351 step:6K smpl:557K ep:336 epch:5.25 loss:0.088 grdn:0.919 lr:1.1e-05 updt_s:4.370 data_s:0.019
|
| 101 |
+
WARNING 2025-12-02 00:20:14 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 102 |
+
INFO 2025-12-02 00:34:56 ot_train.py:351 step:6K smpl:576K ep:348 epch:5.43 loss:0.091 grdn:0.974 lr:1.1e-05 updt_s:4.392 data_s:0.019
|
| 103 |
+
WARNING 2025-12-02 00:34:56 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 104 |
+
INFO 2025-12-02 00:49:34 ot_train.py:351 step:6K smpl:595K ep:359 epch:5.61 loss:0.086 grdn:0.978 lr:9.9e-06 updt_s:4.368 data_s:0.019
|
| 105 |
+
WARNING 2025-12-02 00:49:34 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 106 |
+
INFO 2025-12-02 01:04:13 ot_train.py:351 step:6K smpl:614K ep:371 epch:5.79 loss:0.084 grdn:0.954 lr:9.3e-06 updt_s:4.375 data_s:0.019
|
| 107 |
+
WARNING 2025-12-02 01:04:13 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 108 |
+
INFO 2025-12-02 01:18:51 ot_train.py:351 step:7K smpl:634K ep:382 epch:5.97 loss:0.086 grdn:0.992 lr:8.6e-06 updt_s:4.371 data_s:0.019
|
| 109 |
+
WARNING 2025-12-02 01:18:51 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 110 |
+
INFO 2025-12-02 01:33:35 ot_train.py:351 step:7K smpl:653K ep:394 epch:6.16 loss:0.084 grdn:0.977 lr:8.0e-06 updt_s:4.379 data_s:0.043
|
| 111 |
+
WARNING 2025-12-02 01:33:35 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 112 |
+
INFO 2025-12-02 01:48:13 ot_train.py:351 step:7K smpl:672K ep:406 epch:6.34 loss:0.088 grdn:1.017 lr:7.4e-06 updt_s:4.368 data_s:0.019
|
| 113 |
+
WARNING 2025-12-02 01:48:13 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 114 |
+
INFO 2025-12-02 02:02:51 ot_train.py:351 step:7K smpl:691K ep:417 epch:6.52 loss:0.086 grdn:0.992 lr:6.9e-06 updt_s:4.374 data_s:0.019
|
| 115 |
+
WARNING 2025-12-02 02:02:51 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 116 |
+
INFO 2025-12-02 02:17:31 ot_train.py:351 step:7K smpl:710K ep:429 epch:6.70 loss:0.087 grdn:0.996 lr:6.3e-06 updt_s:4.377 data_s:0.019
|
| 117 |
+
WARNING 2025-12-02 02:17:31 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 118 |
+
INFO 2025-12-02 02:32:10 ot_train.py:351 step:8K smpl:730K ep:440 epch:6.88 loss:0.082 grdn:1.009 lr:5.8e-06 updt_s:4.377 data_s:0.019
|
| 119 |
+
WARNING 2025-12-02 02:32:10 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 120 |
+
INFO 2025-12-02 02:46:48 ot_train.py:351 step:8K smpl:749K ep:452 epch:7.06 loss:0.082 grdn:1.006 lr:5.3e-06 updt_s:4.373 data_s:0.019
|
| 121 |
+
WARNING 2025-12-02 02:46:48 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 122 |
+
INFO 2025-12-02 03:01:26 ot_train.py:351 step:8K smpl:768K ep:463 epch:7.24 loss:0.082 grdn:1.021 lr:4.9e-06 updt_s:4.369 data_s:0.019
|
| 123 |
+
WARNING 2025-12-02 03:01:26 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 124 |
+
INFO 2025-12-02 03:16:03 ot_train.py:351 step:8K smpl:787K ep:475 epch:7.42 loss:0.086 grdn:1.056 lr:4.4e-06 updt_s:4.368 data_s:0.019
|
| 125 |
+
WARNING 2025-12-02 03:16:03 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 126 |
+
INFO 2025-12-02 03:30:42 ot_train.py:351 step:8K smpl:806K ep:487 epch:7.60 loss:0.088 grdn:1.056 lr:4.1e-06 updt_s:4.374 data_s:0.019
|
| 127 |
+
WARNING 2025-12-02 03:30:42 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 128 |
+
INFO 2025-12-02 03:45:19 ot_train.py:351 step:9K smpl:826K ep:498 epch:7.78 loss:0.079 grdn:1.042 lr:3.7e-06 updt_s:4.367 data_s:0.019
|
| 129 |
+
WARNING 2025-12-02 03:45:19 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 130 |
+
INFO 2025-12-02 03:59:57 ot_train.py:351 step:9K smpl:845K ep:510 epch:7.97 loss:0.081 grdn:1.035 lr:3.4e-06 updt_s:4.368 data_s:0.019
|
| 131 |
+
WARNING 2025-12-02 03:59:57 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 132 |
+
INFO 2025-12-02 04:14:40 ot_train.py:351 step:9K smpl:864K ep:521 epch:8.15 loss:0.084 grdn:1.041 lr:3.2e-06 updt_s:4.375 data_s:0.042
|
| 133 |
+
WARNING 2025-12-02 04:14:40 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 134 |
+
INFO 2025-12-02 04:29:18 ot_train.py:351 step:9K smpl:883K ep:533 epch:8.33 loss:0.083 grdn:1.072 lr:2.9e-06 updt_s:4.368 data_s:0.019
|
| 135 |
+
WARNING 2025-12-02 04:29:18 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 136 |
+
INFO 2025-12-02 04:43:56 ot_train.py:351 step:9K smpl:902K ep:545 epch:8.51 loss:0.083 grdn:1.084 lr:2.8e-06 updt_s:4.372 data_s:0.019
|
| 137 |
+
WARNING 2025-12-02 04:43:56 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
| 138 |
+
INFO 2025-12-02 04:58:33 ot_train.py:351 step:10K smpl:922K ep:556 epch:8.69 loss:0.088 grdn:1.093 lr:2.6e-06 updt_s:4.368 data_s:0.019
|
| 139 |
+
WARNING 2025-12-02 04:58:33 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
|
wandb/run-20251201_171106-oemlqork/files/requirements.txt
ADDED
|
@@ -0,0 +1,133 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
h11==0.16.0
|
| 2 |
+
wcwidth==0.2.14
|
| 3 |
+
nvidia-cuda-cupti-cu12==12.6.80
|
| 4 |
+
async-timeout==5.0.1
|
| 5 |
+
dill==0.4.0
|
| 6 |
+
typing_extensions==4.15.0
|
| 7 |
+
pytz==2025.2
|
| 8 |
+
cloudpickle==3.1.2
|
| 9 |
+
pydantic==2.12.4
|
| 10 |
+
python-xlib==0.33
|
| 11 |
+
Jinja2==3.1.6
|
| 12 |
+
gmpy2==2.2.1
|
| 13 |
+
aiohappyeyeballs==2.6.1
|
| 14 |
+
einops==0.8.1
|
| 15 |
+
mpmath==1.3.0
|
| 16 |
+
hyperframe==6.1.0
|
| 17 |
+
smmap==5.0.2
|
| 18 |
+
mergedeep==1.3.4
|
| 19 |
+
Farama-Notifications==0.0.4
|
| 20 |
+
shellingham==1.5.4
|
| 21 |
+
yarl==1.22.0
|
| 22 |
+
peft==0.18.0
|
| 23 |
+
regex==2025.11.3
|
| 24 |
+
deepdiff==8.6.1
|
| 25 |
+
nvidia-cufile-cu12==1.11.1.6
|
| 26 |
+
certifi==2025.11.12
|
| 27 |
+
multiprocess==0.70.16
|
| 28 |
+
cmake==4.1.2
|
| 29 |
+
MarkupSafe==3.0.3
|
| 30 |
+
aiohttp==3.13.2
|
| 31 |
+
psutil==7.1.3
|
| 32 |
+
python-dateutil==2.9.0.post0
|
| 33 |
+
h2==4.3.0
|
| 34 |
+
httpcore==1.0.9
|
| 35 |
+
nvidia-cusparse-cu12==12.5.4.2
|
| 36 |
+
click==8.3.0
|
| 37 |
+
frozenlist==1.8.0
|
| 38 |
+
av==15.1.0
|
| 39 |
+
urllib3==2.5.0
|
| 40 |
+
sentry-sdk==2.44.0
|
| 41 |
+
pynput==1.8.1
|
| 42 |
+
pandas==2.3.3
|
| 43 |
+
triton==3.3.1
|
| 44 |
+
aiosignal==1.4.0
|
| 45 |
+
multidict==6.7.0
|
| 46 |
+
mypy_extensions==1.1.0
|
| 47 |
+
transformers==4.53.3
|
| 48 |
+
anyio==4.11.0
|
| 49 |
+
lerobot==0.4.1
|
| 50 |
+
zipp==3.23.0
|
| 51 |
+
platformdirs==4.5.0
|
| 52 |
+
pydantic_core==2.41.5
|
| 53 |
+
PyYAML==6.0.3
|
| 54 |
+
orderly-set==5.5.0
|
| 55 |
+
protobuf==6.33.0
|
| 56 |
+
typing-inspection==0.4.2
|
| 57 |
+
sympy==1.14.0
|
| 58 |
+
flash_attn==2.8.3
|
| 59 |
+
wheel==0.45.1
|
| 60 |
+
attrs==25.4.0
|
| 61 |
+
tqdm==4.67.1
|
| 62 |
+
nvidia-cudnn-cu12==9.5.1.17
|
| 63 |
+
Brotli==1.2.0
|
| 64 |
+
cffi==2.0.0
|
| 65 |
+
fsspec==2025.9.0
|
| 66 |
+
gitdb==4.0.12
|
| 67 |
+
importlib_metadata==8.7.0
|
| 68 |
+
nvidia-cusparselt-cu12==0.6.3
|
| 69 |
+
nvidia-cublas-cu12==12.6.4.1
|
| 70 |
+
opencv-python-headless==4.12.0.88
|
| 71 |
+
exceptiongroup==1.3.0
|
| 72 |
+
filelock==3.20.0
|
| 73 |
+
safetensors==0.6.2
|
| 74 |
+
pyserial==3.5
|
| 75 |
+
annotated-types==0.7.0
|
| 76 |
+
nvidia-cufft-cu12==11.3.0.4
|
| 77 |
+
pyyaml-include==1.4.1
|
| 78 |
+
PySocks==1.7.1
|
| 79 |
+
ImageIO==2.37.2
|
| 80 |
+
nvidia-cusolver-cu12==11.7.1.2
|
| 81 |
+
datasets==4.1.1
|
| 82 |
+
hf_transfer==0.1.9
|
| 83 |
+
networkx==3.4.2
|
| 84 |
+
docopt==0.6.2
|
| 85 |
+
tzdata==2025.2
|
| 86 |
+
num2words==0.5.14
|
| 87 |
+
hf-xet==1.2.0
|
| 88 |
+
nvidia-nccl-cu12==2.26.2
|
| 89 |
+
sniffio==1.3.1
|
| 90 |
+
huggingface-hub==0.35.3
|
| 91 |
+
xxhash==3.6.0
|
| 92 |
+
requests==2.32.5
|
| 93 |
+
idna==3.11
|
| 94 |
+
httpx==0.28.1
|
| 95 |
+
torch==2.7.1
|
| 96 |
+
typing-inspect==0.9.0
|
| 97 |
+
hpack==4.1.0
|
| 98 |
+
tokenizers==0.21.4
|
| 99 |
+
draccus==0.10.0
|
| 100 |
+
charset-normalizer==3.4.4
|
| 101 |
+
evdev==1.9.2
|
| 102 |
+
zstandard==0.25.0
|
| 103 |
+
pfzy==0.3.4
|
| 104 |
+
pillow==12.0.0
|
| 105 |
+
pycparser==2.22
|
| 106 |
+
torchvision==0.22.1
|
| 107 |
+
termcolor==3.2.0
|
| 108 |
+
wandb==0.21.4
|
| 109 |
+
nvidia-cuda-runtime-cu12==12.6.77
|
| 110 |
+
typer-slim==0.20.0
|
| 111 |
+
rerun-sdk==0.26.2
|
| 112 |
+
gymnasium==1.2.2
|
| 113 |
+
pyarrow==22.0.0
|
| 114 |
+
six==1.17.0
|
| 115 |
+
nvidia-nvjitlink-cu12==12.6.85
|
| 116 |
+
GitPython==3.1.45
|
| 117 |
+
prompt_toolkit==3.0.52
|
| 118 |
+
imageio-ffmpeg==0.6.0
|
| 119 |
+
nvidia-nvtx-cu12==12.6.77
|
| 120 |
+
torchcodec==0.5
|
| 121 |
+
nvidia-curand-cu12==10.3.7.77
|
| 122 |
+
diffusers==0.35.2
|
| 123 |
+
propcache==0.4.1
|
| 124 |
+
packaging==25.0
|
| 125 |
+
setuptools==80.9.0
|
| 126 |
+
jsonlines==4.0.0
|
| 127 |
+
numpy==2.2.6
|
| 128 |
+
pip==25.3
|
| 129 |
+
nvidia-cuda-nvrtc-cu12==12.6.77
|
| 130 |
+
accelerate==1.11.0
|
| 131 |
+
inquirerpy==0.3.4
|
| 132 |
+
toml==0.10.2
|
| 133 |
+
lerobot==0.4.1
|
wandb/run-20251201_171106-oemlqork/files/wandb-metadata.json
ADDED
|
@@ -0,0 +1,59 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"os": "Linux-6.8.0-88-generic-x86_64-with-glibc2.39",
|
| 3 |
+
"python": "CPython 3.10.19",
|
| 4 |
+
"startedAt": "2025-12-01T08:11:06.610650Z",
|
| 5 |
+
"args": [
|
| 6 |
+
"--dataset.repo_id=/home/vla/so101_2tasks_continue",
|
| 7 |
+
"--policy.type=pi0",
|
| 8 |
+
"--policy.pretrained_path=lerobot/pi0_base",
|
| 9 |
+
"--output_dir=/home/vla/so101_2tasks_continue_pi0train",
|
| 10 |
+
"--job_name=so101_2tasks_continue_pi0train",
|
| 11 |
+
"--steps=10000",
|
| 12 |
+
"--batch_size=24",
|
| 13 |
+
"--num_workers=16",
|
| 14 |
+
"--policy.device=cuda",
|
| 15 |
+
"--wandb.enable=true",
|
| 16 |
+
"--policy.compile_model=true",
|
| 17 |
+
"--policy.gradient_checkpointing=true",
|
| 18 |
+
"--policy.dtype=bfloat16",
|
| 19 |
+
"--policy.push_to_hub=false",
|
| 20 |
+
"--rename_map={\"observation.images.camera1\": \"observation.images.wrist\", \"observation.images.camera2\": \"observation.images.top\", \"observation.images.camera3\": \"observation.images.front\"}"
|
| 21 |
+
],
|
| 22 |
+
"program": "/home/vla/miniforge3/envs/lerobot/bin/lerobot-train",
|
| 23 |
+
"codePath": "miniforge3/envs/lerobot/bin/lerobot-train",
|
| 24 |
+
"codePathLocal": "miniforge3/envs/lerobot/bin/lerobot-train",
|
| 25 |
+
"root": "/home/vla/so101_2tasks_continue_pi0train",
|
| 26 |
+
"host": "edu02",
|
| 27 |
+
"executable": "/home/vla/miniforge3/envs/lerobot/bin/python3.10",
|
| 28 |
+
"cpu_count": 56,
|
| 29 |
+
"cpu_count_logical": 112,
|
| 30 |
+
"gpu": "NVIDIA RTX A6000",
|
| 31 |
+
"gpu_count": 2,
|
| 32 |
+
"disk": {
|
| 33 |
+
"/": {
|
| 34 |
+
"total": "943412031488",
|
| 35 |
+
"used": "811606069248"
|
| 36 |
+
}
|
| 37 |
+
},
|
| 38 |
+
"memory": {
|
| 39 |
+
"total": "540628525056"
|
| 40 |
+
},
|
| 41 |
+
"gpu_nvidia": [
|
| 42 |
+
{
|
| 43 |
+
"name": "NVIDIA RTX A6000",
|
| 44 |
+
"memoryTotal": "51527024640",
|
| 45 |
+
"cudaCores": 10752,
|
| 46 |
+
"architecture": "Ampere",
|
| 47 |
+
"uuid": "GPU-c895b718-8738-3fff-b42c-92cad284b785"
|
| 48 |
+
},
|
| 49 |
+
{
|
| 50 |
+
"name": "NVIDIA RTX A6000",
|
| 51 |
+
"memoryTotal": "51527024640",
|
| 52 |
+
"cudaCores": 10752,
|
| 53 |
+
"architecture": "Ampere",
|
| 54 |
+
"uuid": "GPU-34849e7e-d621-dccc-ca70-b463192b3472"
|
| 55 |
+
}
|
| 56 |
+
],
|
| 57 |
+
"cudaVersion": "12.8",
|
| 58 |
+
"writerId": "dhfzrs4kx879100lp67b5x88soub6j73"
|
| 59 |
+
}
|
wandb/run-20251201_171106-oemlqork/logs/debug-core.log
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2025-12-01T17:11:06.649227672+09:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpi7arovmq/port-1280971.txt","pid":1280971,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
|
| 2 |
+
{"time":"2025-12-01T17:11:06.650298036+09:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":1280971}
|
| 3 |
+
{"time":"2025-12-01T17:11:06.650164502+09:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-1280971-1280999-1612259489/socket","Net":"unix"}}
|
| 4 |
+
{"time":"2025-12-01T17:11:06.817876718+09:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
|
| 5 |
+
{"time":"2025-12-01T17:11:06.836443522+09:00","level":"INFO","msg":"handleInformInit: received","streamId":"oemlqork","id":"1(@)"}
|
| 6 |
+
{"time":"2025-12-01T17:11:07.278985483+09:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"oemlqork","id":"1(@)"}
|
wandb/run-20251201_171106-oemlqork/logs/debug-internal.log
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2025-12-01T17:11:06.83664393+09:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"}
|
| 2 |
+
{"time":"2025-12-01T17:11:07.27891855+09:00","level":"INFO","msg":"stream: created new stream","id":"oemlqork"}
|
| 3 |
+
{"time":"2025-12-01T17:11:07.278973759+09:00","level":"INFO","msg":"stream: started","id":"oemlqork"}
|
| 4 |
+
{"time":"2025-12-01T17:11:07.279001361+09:00","level":"INFO","msg":"handler: started","stream_id":"oemlqork"}
|
| 5 |
+
{"time":"2025-12-01T17:11:07.27901924+09:00","level":"INFO","msg":"sender: started","stream_id":"oemlqork"}
|
| 6 |
+
{"time":"2025-12-01T17:11:07.279151787+09:00","level":"INFO","msg":"writer: started","stream_id":"oemlqork"}
|
| 7 |
+
{"time":"2025-12-02T02:26:38.637293979+09:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"}
|
| 8 |
+
{"time":"2025-12-02T02:44:23.665913541+09:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"}
|
wandb/run-20251201_171106-oemlqork/logs/debug.log
ADDED
|
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4
|
| 2 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Configure stats pid to 1280971
|
| 3 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Loading settings from /home/vla/.config/wandb/settings
|
| 4 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Loading settings from /home/vla/wandb/settings
|
| 5 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_setup.py:_flush():81] Loading settings from environment variables
|
| 6 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_init.py:setup_run_log_directory():686] Logging user logs to /home/vla/so101_2tasks_continue_pi0train/wandb/run-20251201_171106-oemlqork/logs/debug.log
|
| 7 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to /home/vla/so101_2tasks_continue_pi0train/wandb/run-20251201_171106-oemlqork/logs/debug-internal.log
|
| 8 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_init.py:init():813] calling init triggers
|
| 9 |
+
2025-12-01 17:11:06,611 INFO MainThread:1280971 [wandb_init.py:init():818] wandb.init called with sweep_config: {}
|
| 10 |
+
config: {'dataset': {'repo_id': '/home/vla/so101_2tasks_continue', 'root': None, 'episodes': None, 'image_transforms': {'enable': False, 'max_num_transforms': 3, 'random_order': False, 'tfs': {'brightness': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'brightness': [0.8, 1.2]}}, 'contrast': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'contrast': [0.8, 1.2]}}, 'saturation': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'saturation': [0.5, 1.5]}}, 'hue': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'hue': [-0.05, 0.05]}}, 'sharpness': {'weight': 1.0, 'type': 'SharpnessJitter', 'kwargs': {'sharpness': [0.5, 1.5]}}, 'affine': {'weight': 1.0, 'type': 'RandomAffine', 'kwargs': {'degrees': [-5.0, 5.0], 'translate': [0.05, 0.05]}}}}, 'revision': None, 'use_imagenet_stats': True, 'video_backend': 'torchcodec', 'streaming': False}, 'env': None, 'policy': {'type': 'pi0', 'n_obs_steps': 1, 'input_features': {}, 'output_features': {}, 'device': 'cuda', 'use_amp': False, 'push_to_hub': False, 'repo_id': None, 'private': None, 'tags': None, 'license': None, 'pretrained_path': 'lerobot/pi0_base', 'paligemma_variant': 'gemma_2b', 'action_expert_variant': 'gemma_300m', 'dtype': 'bfloat16', 'chunk_size': 50, 'n_action_steps': 50, 'max_state_dim': 32, 'max_action_dim': 32, 'num_inference_steps': 10, 'time_sampling_beta_alpha': 1.5, 'time_sampling_beta_beta': 1.0, 'time_sampling_scale': 0.999, 'time_sampling_offset': 0.001, 'min_period': 0.004, 'max_period': 4.0, 'image_resolution': [224, 224], 'empty_cameras': 0, 'normalization_mapping': {'VISUAL': <NormalizationMode.IDENTITY: 'IDENTITY'>, 'STATE': <NormalizationMode.MEAN_STD: 'MEAN_STD'>, 'ACTION': <NormalizationMode.MEAN_STD: 'MEAN_STD'>}, 'gradient_checkpointing': True, 'compile_model': True, 'compile_mode': 'max-autotune', 'optimizer_lr': 2.5e-05, 'optimizer_betas': [0.9, 0.95], 'optimizer_eps': 1e-08, 'optimizer_weight_decay': 0.01, 'optimizer_grad_clip_norm': 1.0, 'scheduler_warmup_steps': 1000, 'scheduler_decay_steps': 30000, 'scheduler_decay_lr': 2.5e-06, 'tokenizer_max_length': 48}, 'output_dir': '/home/vla/so101_2tasks_continue_pi0train', 'job_name': 'so101_2tasks_continue_pi0train', 'resume': False, 'seed': 1000, 'num_workers': 16, 'batch_size': 24, 'steps': 10000, 'eval_freq': 20000, 'log_freq': 200, 'save_checkpoint': True, 'save_freq': 20000, 'use_policy_training_preset': True, 'optimizer': {'type': 'adamw', 'lr': 2.5e-05, 'weight_decay': 0.01, 'grad_clip_norm': 1.0, 'betas': [0.9, 0.95], 'eps': 1e-08}, 'scheduler': {'type': 'cosine_decay_with_warmup', 'num_warmup_steps': 1000, 'num_decay_steps': 30000, 'peak_lr': 2.5e-05, 'decay_lr': 2.5e-06}, 'eval': {'n_episodes': 50, 'batch_size': 50, 'use_async_envs': False}, 'wandb': {'enable': True, 'disable_artifact': False, 'project': 'lerobot', 'entity': None, 'notes': None, 'run_id': None, 'mode': None}, 'checkpoint_path': None, 'rename_map': {'observation.images.camera1': 'observation.images.wrist', 'observation.images.camera2': 'observation.images.top', 'observation.images.camera3': 'observation.images.front'}, '_wandb': {}}
|
| 11 |
+
2025-12-01 17:11:06,612 INFO MainThread:1280971 [wandb_init.py:init():854] starting backend
|
| 12 |
+
2025-12-01 17:11:06,818 INFO MainThread:1280971 [wandb_init.py:init():857] sending inform_init request
|
| 13 |
+
2025-12-01 17:11:06,826 INFO MainThread:1280971 [wandb_init.py:init():865] backend started and connected
|
| 14 |
+
2025-12-01 17:11:06,831 INFO MainThread:1280971 [wandb_init.py:init():936] updated telemetry
|
| 15 |
+
2025-12-01 17:11:06,833 INFO MainThread:1280971 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout
|
| 16 |
+
2025-12-01 17:11:07,681 INFO MainThread:1280971 [wandb_init.py:init():1011] starting run threads in backend
|
| 17 |
+
2025-12-01 17:11:07,793 INFO MainThread:1280971 [wandb_run.py:_console_start():2506] atexit reg
|
| 18 |
+
2025-12-01 17:11:07,793 INFO MainThread:1280971 [wandb_run.py:_redirect():2354] redirect: wrap_raw
|
| 19 |
+
2025-12-01 17:11:07,794 INFO MainThread:1280971 [wandb_run.py:_redirect():2423] Wrapping output streams.
|
| 20 |
+
2025-12-01 17:11:07,794 INFO MainThread:1280971 [wandb_run.py:_redirect():2446] Redirects installed.
|
| 21 |
+
2025-12-01 17:11:07,798 INFO MainThread:1280971 [wandb_init.py:init():1049] run started, returning control to user process
|
wandb/run-20251201_171106-oemlqork/run-oemlqork.wandb
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6bd7a3d420c716066c9534fee57e2c06441ca20b97b2f78fe6dac95c050a8b7e
|
| 3 |
+
size 3375104
|