Boqian-Li commited on
Commit
4196ae6
·
verified ·
1 Parent(s): f93c489

Upload folder using huggingface_hub

Browse files
Files changed (46) hide show
  1. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/config.json +113 -0
  2. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/experiment_cfg/conf.yaml +3306 -0
  3. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/experiment_cfg/metadata.json +787 -0
  4. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
  5. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
  6. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
  7. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt +3 -0
  8. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/bf16_zero_pp_rank_4_mp_rank_00_optim_states.pt +3 -0
  9. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/bf16_zero_pp_rank_5_mp_rank_00_optim_states.pt +3 -0
  10. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/bf16_zero_pp_rank_6_mp_rank_00_optim_states.pt +3 -0
  11. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/bf16_zero_pp_rank_7_mp_rank_00_optim_states.pt +3 -0
  12. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/mp_rank_00_model_states.pt +3 -0
  13. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/latest +1 -0
  14. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00001-of-00010.safetensors +3 -0
  15. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00002-of-00010.safetensors +3 -0
  16. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00003-of-00010.safetensors +3 -0
  17. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00004-of-00010.safetensors +3 -0
  18. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00005-of-00010.safetensors +3 -0
  19. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00006-of-00010.safetensors +3 -0
  20. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00007-of-00010.safetensors +3 -0
  21. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00008-of-00010.safetensors +3 -0
  22. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00009-of-00010.safetensors +3 -0
  23. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00010-of-00010.safetensors +3 -0
  24. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model.safetensors.index.json +0 -0
  25. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/rng_state_0.pth +3 -0
  26. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/rng_state_1.pth +3 -0
  27. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/rng_state_2.pth +3 -0
  28. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/rng_state_3.pth +3 -0
  29. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/rng_state_4.pth +3 -0
  30. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/rng_state_5.pth +3 -0
  31. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/rng_state_6.pth +3 -0
  32. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/rng_state_7.pth +3 -0
  33. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/scheduler.pt +3 -0
  34. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/trainer_state.json +0 -0
  35. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/wandb_config.json +1 -0
  36. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/zero_to_fp32.py +760 -0
  37. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/loss_log.jsonl +607 -0
  38. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/debug-internal.log +34 -0
  39. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/debug.log +2 -0
  40. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/run-20260330_041509-ao6lob27/files/config.yaml +651 -0
  41. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/run-20260330_041509-ao6lob27/files/output.log +0 -0
  42. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/run-20260330_041509-ao6lob27/files/wandb-summary.json +1 -0
  43. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/run-20260330_041509-ao6lob27/logs/debug-core.log +8 -0
  44. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/run-20260330_041509-ao6lob27/logs/debug-internal.log +34 -0
  45. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/run-20260330_041509-ao6lob27/logs/debug.log +2 -0
  46. dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/run-20260330_041509-ao6lob27/run-ao6lob27.wandb +2 -2
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/config.json ADDED
@@ -0,0 +1,113 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "action_dim": 36,
3
+ "action_head_cfg": {
4
+ "_convert_": "object",
5
+ "_target_": "groot.vla.model.dreamzero.action_head.wan_flow_matching_action_tf.WANPolicyHead",
6
+ "config": {
7
+ "_recursive_": false,
8
+ "_target_": "groot.vla.model.dreamzero.action_head.wan_flow_matching_action_tf.WANPolicyHeadConfig",
9
+ "action_dim": 36,
10
+ "action_horizon": 24,
11
+ "action_loss_embodiment_ids": [
12
+ 26,
13
+ 17,
14
+ 32
15
+ ],
16
+ "add_pos_embed": true,
17
+ "backbone_embedding_dim": 0,
18
+ "backbone_features_projector_cfg": null,
19
+ "decouple_video_action_noise": false,
20
+ "diffusion_model_cfg": {
21
+ "_convert_": "object",
22
+ "_target_": "groot.vla.model.dreamzero.modules.wan_video_dit_action_casual_chunk.CausalWanModel",
23
+ "action_dim": 36,
24
+ "diffusion_model_pretrained_path": "/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P",
25
+ "dim": 5120,
26
+ "eps": 1e-06,
27
+ "ffn_dim": 13824,
28
+ "frame_seqlen": 220,
29
+ "freq_dim": 256,
30
+ "in_dim": 36,
31
+ "max_chunk_size": 4,
32
+ "model_type": "i2v",
33
+ "num_action_per_block": 24,
34
+ "num_frame_per_block": 2,
35
+ "num_heads": 40,
36
+ "num_layers": 40,
37
+ "num_state_per_block": 1,
38
+ "out_dim": 16
39
+ },
40
+ "expand_batch": null,
41
+ "freeze_decode_layer": false,
42
+ "hidden_size": 64,
43
+ "image_encoder_cfg": {
44
+ "_convert_": "object",
45
+ "_target_": "groot.vla.model.dreamzero.modules.wan_video_image_encoder.WanImageEncoder",
46
+ "image_encoder_pretrained_path": "/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_clip_open-clip-xlm-roberta-large-vit-huge-14.pth"
47
+ },
48
+ "init_lora_weights": "kaiming",
49
+ "input_embedding_dim": 1536,
50
+ "load_pretrained_det_decode_layer_path": null,
51
+ "lora_alpha": 4,
52
+ "lora_rank": 4,
53
+ "lora_target_modules": "q,k,v,o,ffn.0,ffn.2",
54
+ "max_action_dim": 36,
55
+ "max_state_dim": 64,
56
+ "model_dtype": "float32",
57
+ "noise_beta_alpha": 1.5,
58
+ "noise_beta_beta": 1.0,
59
+ "noise_s": 0.999,
60
+ "num_frame_per_block": 2,
61
+ "num_frames": 33,
62
+ "num_inference_timesteps": 4,
63
+ "num_timestep_buckets": 1000,
64
+ "repa_coeff": 1.0,
65
+ "repa_layer": 8,
66
+ "skip_component_loading": true,
67
+ "text_encoder_cfg": {
68
+ "_convert_": "object",
69
+ "_target_": "groot.vla.model.dreamzero.modules.wan_video_text_encoder.WanTextEncoder",
70
+ "text_encoder_pretrained_path": "/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_t5_umt5-xxl-enc-bf16.pth"
71
+ },
72
+ "tile_size_height": 34,
73
+ "tile_size_width": 34,
74
+ "tile_stride_height": 18,
75
+ "tile_stride_width": 16,
76
+ "tiled": false,
77
+ "train_architecture": "full",
78
+ "tune_diffusion_model": true,
79
+ "tune_projector": true,
80
+ "use_gradient_checkpointing": true,
81
+ "use_vlln": true,
82
+ "vae_cfg": {
83
+ "_convert_": "object",
84
+ "_target_": "groot.vla.model.dreamzero.modules.wan_video_vae.WanVideoVAE",
85
+ "vae_pretrained_path": "/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/Wan2.1_VAE.pth"
86
+ },
87
+ "video_noise_beta_alpha": 3.0,
88
+ "video_noise_beta_beta": 1.0,
89
+ "vl_self_attention_cfg": {
90
+ "_target_": "groot.vla.model.n1_5.modules.cross_attention_dit.SelfAttentionTransformer",
91
+ "attention_head_dim": 64,
92
+ "dropout": 0.2,
93
+ "final_dropout": true,
94
+ "num_attention_heads": 24,
95
+ "num_layers": 4,
96
+ "positional_embeddings": null
97
+ }
98
+ }
99
+ },
100
+ "action_horizon": 24,
101
+ "architectures": [
102
+ "VLA"
103
+ ],
104
+ "backbone_cfg": {
105
+ "_target_": "groot.vla.model.dreamzero.backbone.identity.IdentityBackbone"
106
+ },
107
+ "hidden_size": 0,
108
+ "model_dtype": "float32",
109
+ "model_type": "vla",
110
+ "resume_path": "./checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute",
111
+ "torch_dtype": "bfloat16",
112
+ "transformers_version": "4.51.3"
113
+ }
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/experiment_cfg/conf.yaml ADDED
@@ -0,0 +1,3306 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ _target_: groot.vla.model.dreamzero.base_vla.VLA
3
+ _convert_: object
4
+ config:
5
+ _target_: groot.vla.model.dreamzero.base_vla.VLAConfig
6
+ _recursive_: false
7
+ model_dtype: float32
8
+ hidden_size: 0
9
+ action_horizon: 24
10
+ action_dim: 36
11
+ backbone_cfg:
12
+ _target_: groot.vla.model.dreamzero.backbone.identity.IdentityBackbone
13
+ action_head_cfg:
14
+ config:
15
+ backbone_features_projector_cfg: null
16
+ _target_: groot.vla.model.dreamzero.action_head.wan_flow_matching_action_tf.WANPolicyHeadConfig
17
+ _recursive_: false
18
+ tiled: false
19
+ tile_size_height: 34
20
+ tile_size_width: 34
21
+ tile_stride_height: 18
22
+ tile_stride_width: 16
23
+ lora_rank: 4
24
+ lora_alpha: 4
25
+ num_frames: 33
26
+ num_frame_per_block: 2
27
+ lora_target_modules: q,k,v,o,ffn.0,ffn.2
28
+ init_lora_weights: kaiming
29
+ train_architecture: full
30
+ use_gradient_checkpointing: true
31
+ add_pos_embed: true
32
+ model_dtype: float32
33
+ max_state_dim: 64
34
+ max_action_dim: 36
35
+ action_loss_embodiment_ids:
36
+ - 26
37
+ - 17
38
+ - 32
39
+ hidden_size: 64
40
+ input_embedding_dim: 1536
41
+ backbone_embedding_dim: 0
42
+ repa_layer: 8
43
+ repa_coeff: 1.0
44
+ load_pretrained_det_decode_layer_path: null
45
+ freeze_decode_layer: false
46
+ expand_batch: null
47
+ use_vlln: true
48
+ vl_self_attention_cfg:
49
+ _target_: groot.vla.model.n1_5.modules.cross_attention_dit.SelfAttentionTransformer
50
+ positional_embeddings: null
51
+ num_layers: 4
52
+ num_attention_heads: 24
53
+ attention_head_dim: 64
54
+ dropout: 0.2
55
+ final_dropout: true
56
+ diffusion_model_cfg:
57
+ _target_: groot.vla.model.dreamzero.modules.wan_video_dit_action_casual_chunk.CausalWanModel
58
+ _convert_: object
59
+ diffusion_model_pretrained_path: /hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P
60
+ model_type: i2v
61
+ frame_seqlen: 220
62
+ dim: 5120
63
+ in_dim: 36
64
+ ffn_dim: 13824
65
+ out_dim: 16
66
+ freq_dim: 256
67
+ eps: 1.0e-06
68
+ num_heads: 40
69
+ num_layers: 40
70
+ max_chunk_size: 4
71
+ num_frame_per_block: 2
72
+ num_action_per_block: 24
73
+ num_state_per_block: 1
74
+ action_dim: 36
75
+ text_encoder_cfg:
76
+ _target_: groot.vla.model.dreamzero.modules.wan_video_text_encoder.WanTextEncoder
77
+ _convert_: object
78
+ text_encoder_pretrained_path: /hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_t5_umt5-xxl-enc-bf16.pth
79
+ image_encoder_cfg:
80
+ _target_: groot.vla.model.dreamzero.modules.wan_video_image_encoder.WanImageEncoder
81
+ _convert_: object
82
+ image_encoder_pretrained_path: /hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_clip_open-clip-xlm-roberta-large-vit-huge-14.pth
83
+ vae_cfg:
84
+ _target_: groot.vla.model.dreamzero.modules.wan_video_vae.WanVideoVAE
85
+ _convert_: object
86
+ vae_pretrained_path: /hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/Wan2.1_VAE.pth
87
+ action_dim: 36
88
+ action_horizon: 24
89
+ num_inference_timesteps: 4
90
+ noise_beta_alpha: 1.5
91
+ noise_beta_beta: 1.0
92
+ noise_s: 0.999
93
+ num_timestep_buckets: 1000
94
+ decouple_video_action_noise: false
95
+ video_noise_beta_alpha: 3.0
96
+ video_noise_beta_beta: 1.0
97
+ tune_projector: true
98
+ tune_diffusion_model: true
99
+ skip_component_loading: true
100
+ _target_: groot.vla.model.dreamzero.action_head.wan_flow_matching_action_tf.WANPolicyHead
101
+ _convert_: object
102
+ train_dataset:
103
+ _target_: groot.vla.data.dataset.lerobot_sharded.ShardedLeRobotMixtureDataset.from_mixture_spec
104
+ _convert_: object
105
+ mixture_spec:
106
+ - dataset_path:
107
+ real_teleop_g1:
108
+ - /hfm/boqian/liboqian_data/data/real_data/gear/g1/Pick_bottle_and_turn_and_pour_into_cup_absolute
109
+ dataset_weight: 1.0
110
+ distribute_weights: true
111
+ dataset_class: groot.vla.data.dataset.lerobot_sharded.ShardedLeRobotSubLangSingleActionChunkDatasetDROID
112
+ all_modality_configs:
113
+ oxe_droid:
114
+ video:
115
+ _target_: groot.vla.data.dataset.ModalityConfig
116
+ delta_indices:
117
+ - 0
118
+ - 1
119
+ - 2
120
+ - 3
121
+ - 4
122
+ - 5
123
+ - 6
124
+ - 7
125
+ - 8
126
+ - 9
127
+ - 10
128
+ - 11
129
+ - 12
130
+ - 13
131
+ - 14
132
+ - 15
133
+ - 16
134
+ - 17
135
+ - 18
136
+ - 19
137
+ - 20
138
+ - 21
139
+ - 22
140
+ - 23
141
+ - 24
142
+ eval_delta_indices:
143
+ - 0
144
+ modality_keys:
145
+ - video.exterior_image_1_left
146
+ - video.exterior_image_2_left
147
+ - video.wrist_image_left
148
+ state:
149
+ _target_: groot.vla.data.dataset.ModalityConfig
150
+ delta_indices:
151
+ - 0
152
+ modality_keys:
153
+ - state.joint_position
154
+ - state.gripper_position
155
+ action:
156
+ _target_: groot.vla.data.dataset.ModalityConfig
157
+ delta_indices:
158
+ - 0
159
+ - 1
160
+ - 2
161
+ - 3
162
+ - 4
163
+ - 5
164
+ - 6
165
+ - 7
166
+ - 8
167
+ - 9
168
+ - 10
169
+ - 11
170
+ - 12
171
+ - 13
172
+ - 14
173
+ - 15
174
+ - 16
175
+ - 17
176
+ - 18
177
+ - 19
178
+ - 20
179
+ - 21
180
+ - 22
181
+ - 23
182
+ modality_keys:
183
+ - action.joint_position
184
+ - action.gripper_position
185
+ language:
186
+ _target_: groot.vla.data.dataset.ModalityConfig
187
+ delta_indices:
188
+ - 0
189
+ modality_keys:
190
+ - annotation.language.language_instruction
191
+ - annotation.language.language_instruction_2
192
+ - annotation.language.language_instruction_3
193
+ lapa_action:
194
+ _target_: groot.vla.data.dataset.ModalityConfig
195
+ delta_indices:
196
+ - 0
197
+ modality_keys:
198
+ - lapa_action
199
+ agibot:
200
+ video:
201
+ _target_: groot.vla.data.dataset.ModalityConfig
202
+ delta_indices:
203
+ - 0
204
+ - 1
205
+ - 2
206
+ - 3
207
+ - 4
208
+ - 5
209
+ - 6
210
+ - 7
211
+ - 8
212
+ - 9
213
+ - 10
214
+ - 11
215
+ - 12
216
+ - 13
217
+ - 14
218
+ - 15
219
+ - 16
220
+ - 17
221
+ - 18
222
+ - 19
223
+ - 20
224
+ - 21
225
+ - 22
226
+ - 23
227
+ - 24
228
+ eval_delta_indices:
229
+ - -3
230
+ - -2
231
+ - -1
232
+ - 0
233
+ modality_keys:
234
+ - video.top_head
235
+ - video.hand_left
236
+ - video.hand_right
237
+ state:
238
+ _target_: groot.vla.data.dataset.ModalityConfig
239
+ delta_indices:
240
+ - 0
241
+ modality_keys:
242
+ - state.left_arm_joint_position
243
+ - state.right_arm_joint_position
244
+ - state.left_effector_position
245
+ - state.right_effector_position
246
+ - state.head_position
247
+ - state.waist_position
248
+ action:
249
+ _target_: groot.vla.data.dataset.ModalityConfig
250
+ delta_indices:
251
+ - 0
252
+ - 1
253
+ - 2
254
+ - 3
255
+ - 4
256
+ - 5
257
+ - 6
258
+ - 7
259
+ - 8
260
+ - 9
261
+ - 10
262
+ - 11
263
+ - 12
264
+ - 13
265
+ - 14
266
+ - 15
267
+ - 16
268
+ - 17
269
+ - 18
270
+ - 19
271
+ - 20
272
+ - 21
273
+ - 22
274
+ - 23
275
+ modality_keys:
276
+ - action.left_arm_joint_position
277
+ - action.right_arm_joint_position
278
+ - action.left_effector_position
279
+ - action.right_effector_position
280
+ - action.head_position
281
+ - action.waist_position
282
+ - action.robot_velocity
283
+ language:
284
+ _target_: groot.vla.data.dataset.ModalityConfig
285
+ delta_indices:
286
+ - 0
287
+ modality_keys:
288
+ - annotation.language.action_text
289
+ yam:
290
+ video:
291
+ _target_: groot.vla.data.dataset.ModalityConfig
292
+ delta_indices:
293
+ - 0
294
+ - 1
295
+ - 2
296
+ - 3
297
+ - 4
298
+ - 5
299
+ - 6
300
+ - 7
301
+ - 8
302
+ - 9
303
+ - 10
304
+ - 11
305
+ - 12
306
+ - 13
307
+ - 14
308
+ - 15
309
+ - 16
310
+ - 17
311
+ - 18
312
+ - 19
313
+ - 20
314
+ - 21
315
+ - 22
316
+ - 23
317
+ - 24
318
+ eval_delta_indices:
319
+ - 0
320
+ modality_keys:
321
+ - video.top_camera-images-rgb
322
+ - video.left_camera-images-rgb
323
+ - video.right_camera-images-rgb
324
+ state:
325
+ _target_: groot.vla.data.dataset.ModalityConfig
326
+ delta_indices:
327
+ - 0
328
+ modality_keys:
329
+ - state.left_joint_pos
330
+ - state.left_gripper_pos
331
+ - state.right_joint_pos
332
+ - state.right_gripper_pos
333
+ action:
334
+ _target_: groot.vla.data.dataset.ModalityConfig
335
+ delta_indices:
336
+ - 0
337
+ - 1
338
+ - 2
339
+ - 3
340
+ - 4
341
+ - 5
342
+ - 6
343
+ - 7
344
+ - 8
345
+ - 9
346
+ - 10
347
+ - 11
348
+ - 12
349
+ - 13
350
+ - 14
351
+ - 15
352
+ - 16
353
+ - 17
354
+ - 18
355
+ - 19
356
+ - 20
357
+ - 21
358
+ - 22
359
+ - 23
360
+ modality_keys:
361
+ - action.left_joint_pos
362
+ - action.left_gripper_pos
363
+ - action.right_joint_pos
364
+ - action.right_gripper_pos
365
+ language:
366
+ _target_: groot.vla.data.dataset.ModalityConfig
367
+ delta_indices:
368
+ - 0
369
+ modality_keys:
370
+ - annotation.task
371
+ real_teleop_g1:
372
+ video:
373
+ _target_: groot.vla.data.dataset.ModalityConfig
374
+ delta_indices:
375
+ - 0
376
+ - 1
377
+ - 2
378
+ - 3
379
+ - 4
380
+ - 5
381
+ - 6
382
+ - 7
383
+ - 8
384
+ - 9
385
+ - 10
386
+ - 11
387
+ - 12
388
+ - 13
389
+ - 14
390
+ - 15
391
+ - 16
392
+ - 17
393
+ - 18
394
+ - 19
395
+ - 20
396
+ - 21
397
+ - 22
398
+ - 23
399
+ - 24
400
+ eval_delta_indices:
401
+ - 0
402
+ modality_keys:
403
+ - video.egocentric
404
+ state:
405
+ _target_: groot.vla.data.dataset.ModalityConfig
406
+ delta_indices:
407
+ - 0
408
+ modality_keys:
409
+ - state.left_hand
410
+ - state.right_hand
411
+ - state.left_arm
412
+ - state.right_arm
413
+ - state.rpy
414
+ - state.height
415
+ action:
416
+ _target_: groot.vla.data.dataset.ModalityConfig
417
+ delta_indices:
418
+ - 0
419
+ - 1
420
+ - 2
421
+ - 3
422
+ - 4
423
+ - 5
424
+ - 6
425
+ - 7
426
+ - 8
427
+ - 9
428
+ - 10
429
+ - 11
430
+ - 12
431
+ - 13
432
+ - 14
433
+ - 15
434
+ - 16
435
+ - 17
436
+ - 18
437
+ - 19
438
+ - 20
439
+ - 21
440
+ - 22
441
+ - 23
442
+ modality_keys:
443
+ - action.left_hand
444
+ - action.right_hand
445
+ - action.left_arm
446
+ - action.right_arm
447
+ - action.rpy
448
+ - action.height
449
+ - action.torso_vx
450
+ - action.torso_vy
451
+ - action.torso_vyaw
452
+ - action.torso_dyaw
453
+ language:
454
+ _target_: groot.vla.data.dataset.ModalityConfig
455
+ delta_indices:
456
+ - 0
457
+ modality_keys:
458
+ - annotation.language.language_instruction
459
+ all_transforms:
460
+ oxe_droid:
461
+ _target_: groot.vla.data.transform.ComposedModalityTransform
462
+ transforms:
463
+ - _target_: groot.vla.data.transform.VideoToTensor
464
+ apply_to:
465
+ - video.exterior_image_1_left
466
+ - video.exterior_image_2_left
467
+ - video.wrist_image_left
468
+ - _target_: groot.vla.data.transform.VideoCrop
469
+ apply_to:
470
+ - video.exterior_image_1_left
471
+ - video.exterior_image_2_left
472
+ - video.wrist_image_left
473
+ scale: 0.95
474
+ mode: random
475
+ - _target_: groot.vla.data.transform.VideoResize
476
+ apply_to:
477
+ - video.exterior_image_1_left
478
+ - video.exterior_image_2_left
479
+ - video.wrist_image_left
480
+ height: 176
481
+ width: 320
482
+ interpolation: linear
483
+ - _target_: groot.vla.data.transform.VideoColorJitter
484
+ apply_to:
485
+ - video.exterior_image_1_left
486
+ - video.exterior_image_2_left
487
+ - video.wrist_image_left
488
+ brightness: 0.3
489
+ contrast: 0.4
490
+ saturation: 0.5
491
+ hue: 0.08
492
+ - _target_: groot.vla.data.transform.VideoToNumpy
493
+ apply_to:
494
+ - video.exterior_image_1_left
495
+ - video.exterior_image_2_left
496
+ - video.wrist_image_left
497
+ - _target_: groot.vla.data.transform.StateActionToTensor
498
+ apply_to:
499
+ - state.joint_position
500
+ - state.gripper_position
501
+ - _target_: groot.vla.data.transform.StateActionTransform
502
+ apply_to:
503
+ - state.joint_position
504
+ - state.gripper_position
505
+ normalization_modes:
506
+ state.joint_position: q99
507
+ state.gripper_position: q99
508
+ - _target_: groot.vla.data.transform.StateActionToTensor
509
+ apply_to:
510
+ - action.joint_position
511
+ - action.gripper_position
512
+ - _target_: groot.vla.data.transform.StateActionTransform
513
+ apply_to:
514
+ - action.joint_position
515
+ - action.gripper_position
516
+ normalization_modes:
517
+ action.joint_position: q99
518
+ action.gripper_position: q99
519
+ - _target_: groot.vla.data.transform.ConcatTransform
520
+ video_concat_order:
521
+ - video.exterior_image_1_left
522
+ - video.exterior_image_2_left
523
+ - video.wrist_image_left
524
+ state_concat_order:
525
+ - state.joint_position
526
+ - state.gripper_position
527
+ action_concat_order:
528
+ - action.joint_position
529
+ - action.gripper_position
530
+ - _target_: groot.vla.model.dreamzero.transform.dreamzero_cotrain.DreamTransform
531
+ default_instruction: Perform the default behavior.
532
+ language_dropout_prob: 0.0
533
+ always_use_default_instruction: false
534
+ max_state_dim: 64
535
+ max_action_dim: 36
536
+ max_length: 512
537
+ state_horizon: 1
538
+ action_horizon: 24
539
+ embodiment_tag_mapping:
540
+ real_gr1_arms_only: 0
541
+ real_gr1_arms_only_annotated: 1
542
+ real_gr1_arms_waist: 2
543
+ real_gr1_arms_waist_annotated: 3
544
+ dexmg_gr1_arms_only_inspire: 4
545
+ dexmg_gr1_arms_only_fourier: 5
546
+ dexmg_gr1_arms_waist_fourier: 6
547
+ robocasa_single_arm: 7
548
+ onex_eve_gripper: 8
549
+ robocasa_gr1_arms_only_inspire_hands: 9
550
+ robocasa_gr1_arms_only_fourier_hands: 10
551
+ robocasa_gr1_fixed_lower_body_inspire_hands: 11
552
+ robocasa_gr1_fixed_lower_body_fourier_hands: 12
553
+ robocasa_panda_omron: 13
554
+ robocasa_bimanual_panda_parallel_gripper: 15
555
+ robocasa_bimanual_panda_inspire_hand: 16
556
+ oxe_droid: 17
557
+ oxe_fractal: 18
558
+ oxe_language_table: 19
559
+ oxe_bridge: 20
560
+ real_panda_single_arm: 21
561
+ hot3d_hands_only: 23
562
+ gr1_unified: 24
563
+ robocasa_gr1_arms_waist_fourier_hands: 25
564
+ agibot: 26
565
+ lapa: 27
566
+ oxe_mutex: 28
567
+ oxe_roboset: 29
568
+ oxe_plex: 30
569
+ dream: 31
570
+ yam: 32
571
+ xdof: 22
572
+ gr1_unified_segmentation: 14
573
+ language_table_sim: 7
574
+ gr1_isaac: 0
575
+ sim_behavior_r1_pro: 31
576
+ mecka_hands: 27
577
+ real_r1_pro_sharpa: 28
578
+ real_teleop_g1: 33
579
+ tokenizer_path: /hfm/boqian/liboqian_data/checkpoints/umt5-xxl
580
+ agibot:
581
+ _target_: groot.vla.data.transform.ComposedModalityTransform
582
+ transforms:
583
+ - _target_: groot.vla.data.transform.VideoToTensor
584
+ apply_to:
585
+ - video.top_head
586
+ - video.hand_left
587
+ - video.hand_right
588
+ - _target_: groot.vla.data.transform.VideoCrop
589
+ apply_to:
590
+ - video.top_head
591
+ - video.hand_left
592
+ - video.hand_right
593
+ scale: 0.95
594
+ mode: random
595
+ - _target_: groot.vla.data.transform.VideoResize
596
+ apply_to:
597
+ - video.top_head
598
+ - video.hand_left
599
+ - video.hand_right
600
+ height: 176
601
+ width: 320
602
+ interpolation: linear
603
+ - _target_: groot.vla.data.transform.VideoColorJitter
604
+ apply_to:
605
+ - video.top_head
606
+ - video.hand_left
607
+ - video.hand_right
608
+ brightness: 0.3
609
+ contrast: 0.4
610
+ saturation: 0.5
611
+ hue: 0.08
612
+ - _target_: groot.vla.data.transform.VideoToNumpy
613
+ apply_to:
614
+ - video.top_head
615
+ - video.hand_left
616
+ - video.hand_right
617
+ - _target_: groot.vla.data.transform.StateActionToTensor
618
+ apply_to:
619
+ - state.left_arm_joint_position
620
+ - state.right_arm_joint_position
621
+ - state.left_effector_position
622
+ - state.right_effector_position
623
+ - state.head_position
624
+ - state.waist_position
625
+ - _target_: groot.vla.data.transform.StateActionTransform
626
+ apply_to:
627
+ - state.left_arm_joint_position
628
+ - state.right_arm_joint_position
629
+ - state.left_effector_position
630
+ - state.right_effector_position
631
+ - state.head_position
632
+ - state.waist_position
633
+ normalization_modes:
634
+ state.left_arm_joint_position: q99
635
+ state.right_arm_joint_position: q99
636
+ state.left_effector_position: q99
637
+ state.right_effector_position: q99
638
+ state.head_position: q99
639
+ state.waist_position: q99
640
+ - _target_: groot.vla.data.transform.StateActionToTensor
641
+ apply_to:
642
+ - action.left_arm_joint_position
643
+ - action.right_arm_joint_position
644
+ - action.left_effector_position
645
+ - action.right_effector_position
646
+ - action.head_position
647
+ - action.waist_position
648
+ - action.robot_velocity
649
+ - _target_: groot.vla.data.transform.StateActionTransform
650
+ apply_to:
651
+ - action.left_arm_joint_position
652
+ - action.right_arm_joint_position
653
+ - action.left_effector_position
654
+ - action.right_effector_position
655
+ - action.head_position
656
+ - action.waist_position
657
+ - action.robot_velocity
658
+ normalization_modes:
659
+ action.left_arm_joint_position: q99
660
+ action.right_arm_joint_position: q99
661
+ action.left_effector_position: q99
662
+ action.right_effector_position: q99
663
+ action.head_position: q99
664
+ action.waist_position: q99
665
+ action.robot_velocity: q99
666
+ - _target_: groot.vla.data.transform.ConcatTransform
667
+ video_concat_order:
668
+ - video.top_head
669
+ - video.hand_left
670
+ - video.hand_right
671
+ state_concat_order:
672
+ - state.left_arm_joint_position
673
+ - state.right_arm_joint_position
674
+ - state.left_effector_position
675
+ - state.right_effector_position
676
+ - state.head_position
677
+ - state.waist_position
678
+ action_concat_order:
679
+ - action.left_arm_joint_position
680
+ - action.right_arm_joint_position
681
+ - action.left_effector_position
682
+ - action.right_effector_position
683
+ - action.head_position
684
+ - action.waist_position
685
+ - action.robot_velocity
686
+ - _target_: groot.vla.model.dreamzero.transform.dreamzero_cotrain.DreamTransform
687
+ default_instruction: Perform the default behavior.
688
+ language_dropout_prob: 0.0
689
+ always_use_default_instruction: false
690
+ max_state_dim: 64
691
+ max_action_dim: 36
692
+ max_length: 512
693
+ state_horizon: 1
694
+ action_horizon: 24
695
+ embodiment_tag_mapping:
696
+ real_gr1_arms_only: 0
697
+ real_gr1_arms_only_annotated: 1
698
+ real_gr1_arms_waist: 2
699
+ real_gr1_arms_waist_annotated: 3
700
+ dexmg_gr1_arms_only_inspire: 4
701
+ dexmg_gr1_arms_only_fourier: 5
702
+ dexmg_gr1_arms_waist_fourier: 6
703
+ robocasa_single_arm: 7
704
+ onex_eve_gripper: 8
705
+ robocasa_gr1_arms_only_inspire_hands: 9
706
+ robocasa_gr1_arms_only_fourier_hands: 10
707
+ robocasa_gr1_fixed_lower_body_inspire_hands: 11
708
+ robocasa_gr1_fixed_lower_body_fourier_hands: 12
709
+ robocasa_panda_omron: 13
710
+ robocasa_bimanual_panda_parallel_gripper: 15
711
+ robocasa_bimanual_panda_inspire_hand: 16
712
+ oxe_droid: 17
713
+ oxe_fractal: 18
714
+ oxe_language_table: 19
715
+ oxe_bridge: 20
716
+ real_panda_single_arm: 21
717
+ hot3d_hands_only: 23
718
+ gr1_unified: 24
719
+ robocasa_gr1_arms_waist_fourier_hands: 25
720
+ agibot: 26
721
+ lapa: 27
722
+ oxe_mutex: 28
723
+ oxe_roboset: 29
724
+ oxe_plex: 30
725
+ dream: 31
726
+ yam: 32
727
+ xdof: 22
728
+ gr1_unified_segmentation: 14
729
+ language_table_sim: 7
730
+ gr1_isaac: 0
731
+ sim_behavior_r1_pro: 31
732
+ mecka_hands: 27
733
+ real_r1_pro_sharpa: 28
734
+ real_teleop_g1: 33
735
+ tokenizer_path: /hfm/boqian/liboqian_data/checkpoints/umt5-xxl
736
+ yam:
737
+ _target_: groot.vla.data.transform.ComposedModalityTransform
738
+ transforms:
739
+ - _target_: groot.vla.data.transform.VideoToTensor
740
+ apply_to:
741
+ - video.top_camera-images-rgb
742
+ - video.left_camera-images-rgb
743
+ - video.right_camera-images-rgb
744
+ - _target_: groot.vla.data.transform.VideoCrop
745
+ apply_to:
746
+ - video.top_camera-images-rgb
747
+ - video.left_camera-images-rgb
748
+ - video.right_camera-images-rgb
749
+ scale: 0.95
750
+ mode: random
751
+ - _target_: groot.vla.data.transform.VideoResize
752
+ apply_to:
753
+ - video.top_camera-images-rgb
754
+ - video.left_camera-images-rgb
755
+ - video.right_camera-images-rgb
756
+ height: 176
757
+ width: 320
758
+ interpolation: linear
759
+ - _target_: groot.vla.data.transform.VideoColorJitter
760
+ apply_to:
761
+ - video.top_camera-images-rgb
762
+ - video.left_camera-images-rgb
763
+ - video.right_camera-images-rgb
764
+ brightness: 0.3
765
+ contrast: 0.4
766
+ saturation: 0.5
767
+ hue: 0.08
768
+ - _target_: groot.vla.data.transform.VideoToNumpy
769
+ apply_to:
770
+ - video.top_camera-images-rgb
771
+ - video.left_camera-images-rgb
772
+ - video.right_camera-images-rgb
773
+ - _target_: groot.vla.data.transform.StateActionToTensor
774
+ apply_to:
775
+ - state.left_joint_pos
776
+ - state.left_gripper_pos
777
+ - state.right_joint_pos
778
+ - state.right_gripper_pos
779
+ - _target_: groot.vla.data.transform.StateActionTransform
780
+ apply_to:
781
+ - state.left_joint_pos
782
+ - state.left_gripper_pos
783
+ - state.right_joint_pos
784
+ - state.right_gripper_pos
785
+ normalization_modes:
786
+ state.left_joint_pos: q99
787
+ state.left_gripper_pos: q99
788
+ state.right_joint_pos: q99
789
+ state.right_gripper_pos: q99
790
+ - _target_: groot.vla.data.transform.StateActionToTensor
791
+ apply_to:
792
+ - action.left_joint_pos
793
+ - action.left_gripper_pos
794
+ - action.right_joint_pos
795
+ - action.right_gripper_pos
796
+ - _target_: groot.vla.data.transform.StateActionTransform
797
+ apply_to:
798
+ - action.left_joint_pos
799
+ - action.left_gripper_pos
800
+ - action.right_joint_pos
801
+ - action.right_gripper_pos
802
+ normalization_modes:
803
+ action.left_joint_pos: q99
804
+ action.left_gripper_pos: q99
805
+ action.right_joint_pos: q99
806
+ action.right_gripper_pos: q99
807
+ - _target_: groot.vla.data.transform.ConcatTransform
808
+ video_concat_order:
809
+ - video.top_camera-images-rgb
810
+ - video.left_camera-images-rgb
811
+ - video.right_camera-images-rgb
812
+ state_concat_order:
813
+ - state.left_joint_pos
814
+ - state.left_gripper_pos
815
+ - state.right_joint_pos
816
+ - state.right_gripper_pos
817
+ action_concat_order:
818
+ - action.left_joint_pos
819
+ - action.left_gripper_pos
820
+ - action.right_joint_pos
821
+ - action.right_gripper_pos
822
+ - _target_: groot.vla.model.dreamzero.transform.dreamzero_cotrain.DreamTransform
823
+ default_instruction: Perform the default behavior.
824
+ language_dropout_prob: 0.0
825
+ always_use_default_instruction: false
826
+ max_state_dim: 64
827
+ max_action_dim: 36
828
+ max_length: 512
829
+ state_horizon: 1
830
+ action_horizon: 24
831
+ embodiment_tag_mapping:
832
+ real_gr1_arms_only: 0
833
+ real_gr1_arms_only_annotated: 1
834
+ real_gr1_arms_waist: 2
835
+ real_gr1_arms_waist_annotated: 3
836
+ dexmg_gr1_arms_only_inspire: 4
837
+ dexmg_gr1_arms_only_fourier: 5
838
+ dexmg_gr1_arms_waist_fourier: 6
839
+ robocasa_single_arm: 7
840
+ onex_eve_gripper: 8
841
+ robocasa_gr1_arms_only_inspire_hands: 9
842
+ robocasa_gr1_arms_only_fourier_hands: 10
843
+ robocasa_gr1_fixed_lower_body_inspire_hands: 11
844
+ robocasa_gr1_fixed_lower_body_fourier_hands: 12
845
+ robocasa_panda_omron: 13
846
+ robocasa_bimanual_panda_parallel_gripper: 15
847
+ robocasa_bimanual_panda_inspire_hand: 16
848
+ oxe_droid: 17
849
+ oxe_fractal: 18
850
+ oxe_language_table: 19
851
+ oxe_bridge: 20
852
+ real_panda_single_arm: 21
853
+ hot3d_hands_only: 23
854
+ gr1_unified: 24
855
+ robocasa_gr1_arms_waist_fourier_hands: 25
856
+ agibot: 26
857
+ lapa: 27
858
+ oxe_mutex: 28
859
+ oxe_roboset: 29
860
+ oxe_plex: 30
861
+ dream: 31
862
+ yam: 32
863
+ xdof: 22
864
+ gr1_unified_segmentation: 14
865
+ language_table_sim: 7
866
+ gr1_isaac: 0
867
+ sim_behavior_r1_pro: 31
868
+ mecka_hands: 27
869
+ real_r1_pro_sharpa: 28
870
+ real_teleop_g1: 33
871
+ tokenizer_path: /hfm/boqian/liboqian_data/checkpoints/umt5-xxl
872
+ real_teleop_g1:
873
+ _target_: groot.vla.data.transform.ComposedModalityTransform
874
+ transforms:
875
+ - _target_: groot.vla.data.transform.VideoToTensor
876
+ apply_to:
877
+ - video.egocentric
878
+ - _target_: groot.vla.data.transform.VideoCrop
879
+ apply_to:
880
+ - video.egocentric
881
+ scale: 0.95
882
+ mode: random
883
+ - _target_: groot.vla.data.transform.VideoResize
884
+ apply_to:
885
+ - video.egocentric
886
+ height: 176
887
+ width: 320
888
+ interpolation: linear
889
+ - _target_: groot.vla.data.transform.VideoColorJitter
890
+ apply_to:
891
+ - video.egocentric
892
+ brightness: 0.3
893
+ contrast: 0.4
894
+ saturation: 0.5
895
+ hue: 0.08
896
+ - _target_: groot.vla.data.transform.VideoToNumpy
897
+ apply_to:
898
+ - video.egocentric
899
+ - _target_: groot.vla.data.transform.StateActionToTensor
900
+ apply_to:
901
+ - state.left_hand
902
+ - state.right_hand
903
+ - state.left_arm
904
+ - state.right_arm
905
+ - state.rpy
906
+ - state.height
907
+ - _target_: groot.vla.data.transform.StateActionTransform
908
+ apply_to:
909
+ - state.left_hand
910
+ - state.right_hand
911
+ - state.left_arm
912
+ - state.right_arm
913
+ - state.rpy
914
+ - state.height
915
+ normalization_modes:
916
+ state.left_hand: q99
917
+ state.right_hand: q99
918
+ state.left_arm: q99
919
+ state.right_arm: q99
920
+ state.rpy: q99
921
+ state.height: q99
922
+ - _target_: groot.vla.data.transform.StateActionToTensor
923
+ apply_to:
924
+ - action.left_hand
925
+ - action.right_hand
926
+ - action.left_arm
927
+ - action.right_arm
928
+ - action.rpy
929
+ - action.height
930
+ - action.torso_vx
931
+ - action.torso_vy
932
+ - action.torso_vyaw
933
+ - action.torso_dyaw
934
+ - _target_: groot.vla.data.transform.StateActionTransform
935
+ apply_to:
936
+ - action.left_hand
937
+ - action.right_hand
938
+ - action.left_arm
939
+ - action.right_arm
940
+ - action.rpy
941
+ - action.height
942
+ - action.torso_vx
943
+ - action.torso_vy
944
+ - action.torso_vyaw
945
+ - action.torso_dyaw
946
+ normalization_modes:
947
+ action.left_hand: q99
948
+ action.right_hand: q99
949
+ action.left_arm: q99
950
+ action.right_arm: q99
951
+ action.rpy: q99
952
+ action.height: q99
953
+ action.torso_vx: q99
954
+ action.torso_vy: q99
955
+ action.torso_vyaw: q99
956
+ action.torso_dyaw: q99
957
+ - _target_: groot.vla.data.transform.ConcatTransform
958
+ video_concat_order:
959
+ - video.egocentric
960
+ state_concat_order:
961
+ - state.left_hand
962
+ - state.right_hand
963
+ - state.left_arm
964
+ - state.right_arm
965
+ - state.rpy
966
+ - state.height
967
+ action_concat_order:
968
+ - action.left_hand
969
+ - action.right_hand
970
+ - action.left_arm
971
+ - action.right_arm
972
+ - action.rpy
973
+ - action.height
974
+ - action.torso_vx
975
+ - action.torso_vy
976
+ - action.torso_vyaw
977
+ - action.torso_dyaw
978
+ - _target_: groot.vla.model.dreamzero.transform.dreamzero_cotrain.DreamTransform
979
+ default_instruction: Perform the default behavior.
980
+ language_dropout_prob: 0.0
981
+ always_use_default_instruction: false
982
+ max_state_dim: 64
983
+ max_action_dim: 36
984
+ max_length: 512
985
+ state_horizon: 1
986
+ action_horizon: 24
987
+ embodiment_tag_mapping:
988
+ real_gr1_arms_only: 0
989
+ real_gr1_arms_only_annotated: 1
990
+ real_gr1_arms_waist: 2
991
+ real_gr1_arms_waist_annotated: 3
992
+ dexmg_gr1_arms_only_inspire: 4
993
+ dexmg_gr1_arms_only_fourier: 5
994
+ dexmg_gr1_arms_waist_fourier: 6
995
+ robocasa_single_arm: 7
996
+ onex_eve_gripper: 8
997
+ robocasa_gr1_arms_only_inspire_hands: 9
998
+ robocasa_gr1_arms_only_fourier_hands: 10
999
+ robocasa_gr1_fixed_lower_body_inspire_hands: 11
1000
+ robocasa_gr1_fixed_lower_body_fourier_hands: 12
1001
+ robocasa_panda_omron: 13
1002
+ robocasa_bimanual_panda_parallel_gripper: 15
1003
+ robocasa_bimanual_panda_inspire_hand: 16
1004
+ oxe_droid: 17
1005
+ oxe_fractal: 18
1006
+ oxe_language_table: 19
1007
+ oxe_bridge: 20
1008
+ real_panda_single_arm: 21
1009
+ hot3d_hands_only: 23
1010
+ gr1_unified: 24
1011
+ robocasa_gr1_arms_waist_fourier_hands: 25
1012
+ agibot: 26
1013
+ lapa: 27
1014
+ oxe_mutex: 28
1015
+ oxe_roboset: 29
1016
+ oxe_plex: 30
1017
+ dream: 31
1018
+ yam: 32
1019
+ xdof: 22
1020
+ gr1_unified_segmentation: 14
1021
+ language_table_sim: 7
1022
+ gr1_isaac: 0
1023
+ sim_behavior_r1_pro: 31
1024
+ mecka_hands: 27
1025
+ real_r1_pro_sharpa: 28
1026
+ real_teleop_g1: 33
1027
+ tokenizer_path: /hfm/boqian/liboqian_data/checkpoints/umt5-xxl
1028
+ metadata_versions:
1029
+ oxe_droid: '0221'
1030
+ agibot: '0221'
1031
+ yam: '0221'
1032
+ real_teleop_g1: '0221'
1033
+ fps:
1034
+ yam: 30
1035
+ real_teleop_g1: 30
1036
+ dataset_kwargs:
1037
+ video_backend: decord
1038
+ use_global_metadata: false
1039
+ max_chunk_size: 4
1040
+ relative_action: false
1041
+ relative_action_keys: []
1042
+ relative_action_per_horizon: false
1043
+ mixture_kwargs:
1044
+ training: true
1045
+ balance_dataset_weights: false
1046
+ seed: 42
1047
+ shard_sampling_rate: 0.1
1048
+ trainer:
1049
+ _target_: groot.vla.experiment.VLATrainer
1050
+ _partial_: true
1051
+ _recursive_: false
1052
+ callbacks: null
1053
+ model: ???
1054
+ train_dataset: ???
1055
+ compute_dtype: ???
1056
+ benchmark_time: false
1057
+ enable_profiling: false
1058
+ profiling_steps: 5
1059
+ enable_prof_callback: false
1060
+ profile_start_step: 50
1061
+ profile_warmup_steps: 1
1062
+ profile_active_steps: 3
1063
+ profile_record_shapes: false
1064
+ profile_with_stack: false
1065
+ profile_memory: false
1066
+ wandb_project: dreamzero
1067
+ output_dir: ./checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute
1068
+ load_from_yaml: null
1069
+ gear_credentials: null
1070
+ upload_checkpoints: false
1071
+ upload_every: 1000
1072
+ upload_last_n_checkpoints: 5
1073
+ remove_unused_columns: false
1074
+ bf16: true
1075
+ tf32: true
1076
+ global_batch_size: null
1077
+ raise_error_if_global_batch_size_not_set: false
1078
+ per_device_train_batch_size: 1
1079
+ per_device_eval_batch_size: 64
1080
+ gradient_accumulation_steps: 1
1081
+ dataloader_num_workers: 1
1082
+ dataloader_pin_memory: false
1083
+ dataloader_persistent_workers: true
1084
+ optim: adamw_torch
1085
+ learning_rate: 0.0001
1086
+ adam_beta1: 0.95
1087
+ adam_beta2: 0.999
1088
+ adam_epsilon: 1.0e-08
1089
+ weight_decay: 1.0e-05
1090
+ lr_scheduler_type: cosine
1091
+ warmup_ratio: 0.05
1092
+ logging_steps: 10.0
1093
+ num_train_epochs: 1000
1094
+ max_steps: 20000
1095
+ save_strategy: steps
1096
+ save_steps: 8000
1097
+ eval_strategy: 'no'
1098
+ save_total_limit: 10
1099
+ report_to: wandb
1100
+ seed: 42
1101
+ do_eval: false
1102
+ gradient_checkpointing: false
1103
+ ddp_find_unused_parameters: false
1104
+ ddp_bucket_cap_mb: 100
1105
+ ray_num_workers: ???
1106
+ eval_bf16: true
1107
+ torch_compile_mode: null
1108
+ pretrained_model_path: /hfm/boqian/liboqian_data/checkpoints/DreamZero-AgiBot
1109
+ only_tune_projectors: false
1110
+ save_llm: false
1111
+ save_lora_only: false
1112
+ save_value_model: false
1113
+ save_q_model: false
1114
+ download_cache: false
1115
+ training_args:
1116
+ _target_: transformers.TrainingArguments
1117
+ output_dir: ./checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute
1118
+ run_name: dreamzero_real_teleop_g1_full_finetune_absolute
1119
+ remove_unused_columns: false
1120
+ deepspeed: groot/vla/configs/deepspeed/zero2_offload.json
1121
+ gradient_checkpointing: false
1122
+ bf16: true
1123
+ tf32: true
1124
+ per_device_train_batch_size: 1
1125
+ per_device_eval_batch_size: 64
1126
+ gradient_accumulation_steps: 1
1127
+ dataloader_num_workers: 1
1128
+ dataloader_pin_memory: false
1129
+ dataloader_persistent_workers: true
1130
+ optim: adamw_torch
1131
+ adam_beta1: 0.95
1132
+ adam_beta2: 0.999
1133
+ adam_epsilon: 1.0e-08
1134
+ learning_rate: 1.0e-05
1135
+ weight_decay: 1.0e-05
1136
+ warmup_ratio: 0.05
1137
+ lr_scheduler_type: cosine
1138
+ logging_steps: 10.0
1139
+ num_train_epochs: 1000
1140
+ max_steps: 20000
1141
+ save_strategy: steps
1142
+ save_steps: 8000
1143
+ save_total_limit: 10
1144
+ report_to: wandb
1145
+ seed: 42
1146
+ do_eval: false
1147
+ ddp_find_unused_parameters: false
1148
+ ddp_bucket_cap_mb: 100
1149
+ torch_compile_mode: null
1150
+ profile_dir: null
1151
+ backbone_hidden_size: 0
1152
+ backbone_cfg:
1153
+ _target_: groot.vla.model.dreamzero.backbone.identity.IdentityBackbone
1154
+ action_head_cfg:
1155
+ config:
1156
+ backbone_features_projector_cfg: null
1157
+ _target_: groot.vla.model.dreamzero.action_head.wan_flow_matching_action_tf.WANPolicyHeadConfig
1158
+ _recursive_: false
1159
+ tiled: false
1160
+ tile_size_height: 34
1161
+ tile_size_width: 34
1162
+ tile_stride_height: 18
1163
+ tile_stride_width: 16
1164
+ lora_rank: 4
1165
+ lora_alpha: 4
1166
+ num_frames: 33
1167
+ num_frame_per_block: 2
1168
+ lora_target_modules: q,k,v,o,ffn.0,ffn.2
1169
+ init_lora_weights: kaiming
1170
+ train_architecture: full
1171
+ use_gradient_checkpointing: true
1172
+ add_pos_embed: true
1173
+ model_dtype: float32
1174
+ max_state_dim: 64
1175
+ max_action_dim: 36
1176
+ action_loss_embodiment_ids:
1177
+ - 26
1178
+ - 17
1179
+ - 32
1180
+ hidden_size: 64
1181
+ input_embedding_dim: 1536
1182
+ backbone_embedding_dim: 0
1183
+ repa_layer: 8
1184
+ repa_coeff: 1.0
1185
+ load_pretrained_det_decode_layer_path: null
1186
+ freeze_decode_layer: false
1187
+ expand_batch: null
1188
+ use_vlln: true
1189
+ vl_self_attention_cfg:
1190
+ _target_: groot.vla.model.n1_5.modules.cross_attention_dit.SelfAttentionTransformer
1191
+ positional_embeddings: null
1192
+ num_layers: 4
1193
+ num_attention_heads: 24
1194
+ attention_head_dim: 64
1195
+ dropout: 0.2
1196
+ final_dropout: true
1197
+ diffusion_model_cfg:
1198
+ _target_: groot.vla.model.dreamzero.modules.wan_video_dit_action_casual_chunk.CausalWanModel
1199
+ _convert_: object
1200
+ diffusion_model_pretrained_path: /hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P
1201
+ model_type: i2v
1202
+ frame_seqlen: 220
1203
+ dim: 5120
1204
+ in_dim: 36
1205
+ ffn_dim: 13824
1206
+ out_dim: 16
1207
+ freq_dim: 256
1208
+ eps: 1.0e-06
1209
+ num_heads: 40
1210
+ num_layers: 40
1211
+ max_chunk_size: 4
1212
+ num_frame_per_block: 2
1213
+ num_action_per_block: 24
1214
+ num_state_per_block: 1
1215
+ action_dim: 36
1216
+ text_encoder_cfg:
1217
+ _target_: groot.vla.model.dreamzero.modules.wan_video_text_encoder.WanTextEncoder
1218
+ _convert_: object
1219
+ text_encoder_pretrained_path: /hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_t5_umt5-xxl-enc-bf16.pth
1220
+ image_encoder_cfg:
1221
+ _target_: groot.vla.model.dreamzero.modules.wan_video_image_encoder.WanImageEncoder
1222
+ _convert_: object
1223
+ image_encoder_pretrained_path: /hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_clip_open-clip-xlm-roberta-large-vit-huge-14.pth
1224
+ vae_cfg:
1225
+ _target_: groot.vla.model.dreamzero.modules.wan_video_vae.WanVideoVAE
1226
+ _convert_: object
1227
+ vae_pretrained_path: /hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/Wan2.1_VAE.pth
1228
+ action_dim: 36
1229
+ action_horizon: 24
1230
+ num_inference_timesteps: 4
1231
+ noise_beta_alpha: 1.5
1232
+ noise_beta_beta: 1.0
1233
+ noise_s: 0.999
1234
+ num_timestep_buckets: 1000
1235
+ decouple_video_action_noise: false
1236
+ video_noise_beta_alpha: 3.0
1237
+ video_noise_beta_beta: 1.0
1238
+ tune_projector: true
1239
+ tune_diffusion_model: true
1240
+ skip_component_loading: true
1241
+ _target_: groot.vla.model.dreamzero.action_head.wan_flow_matching_action_tf.WANPolicyHead
1242
+ _convert_: object
1243
+ add_pos_embed: true
1244
+ hidden_size: 64
1245
+ attn_dropout: 0.2
1246
+ repa_layer: 8
1247
+ repa_coeff: 1.0
1248
+ load_pretrained_det_decode_layer_path: null
1249
+ expand_batch: null
1250
+ dit_version: /hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P
1251
+ text_encoder_pretrained_path: /hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_t5_umt5-xxl-enc-bf16.pth
1252
+ image_encoder_pretrained_path: /hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_clip_open-clip-xlm-roberta-large-vit-huge-14.pth
1253
+ vae_pretrained_path: /hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/Wan2.1_VAE.pth
1254
+ train_architecture: full
1255
+ num_frame_per_block: 2
1256
+ num_action_per_block: 24
1257
+ num_state_per_block: 1
1258
+ frame_seqlen: 220
1259
+ embodiment_tag_to_projector_index:
1260
+ real_gr1_arms_only: 0
1261
+ real_gr1_arms_only_annotated: 1
1262
+ real_gr1_arms_waist: 2
1263
+ real_gr1_arms_waist_annotated: 3
1264
+ dexmg_gr1_arms_only_inspire: 4
1265
+ dexmg_gr1_arms_only_fourier: 5
1266
+ dexmg_gr1_arms_waist_fourier: 6
1267
+ robocasa_single_arm: 7
1268
+ onex_eve_gripper: 8
1269
+ robocasa_gr1_arms_only_inspire_hands: 9
1270
+ robocasa_gr1_arms_only_fourier_hands: 10
1271
+ robocasa_gr1_fixed_lower_body_inspire_hands: 11
1272
+ robocasa_gr1_fixed_lower_body_fourier_hands: 12
1273
+ robocasa_panda_omron: 13
1274
+ robocasa_bimanual_panda_parallel_gripper: 15
1275
+ robocasa_bimanual_panda_inspire_hand: 16
1276
+ oxe_droid: 17
1277
+ oxe_fractal: 18
1278
+ oxe_language_table: 19
1279
+ oxe_bridge: 20
1280
+ real_panda_single_arm: 21
1281
+ hot3d_hands_only: 23
1282
+ gr1_unified: 24
1283
+ robocasa_gr1_arms_waist_fourier_hands: 25
1284
+ agibot: 26
1285
+ lapa: 27
1286
+ oxe_mutex: 28
1287
+ oxe_roboset: 29
1288
+ oxe_plex: 30
1289
+ dream: 31
1290
+ yam: 32
1291
+ xdof: 22
1292
+ gr1_unified_segmentation: 14
1293
+ language_table_sim: 7
1294
+ gr1_isaac: 0
1295
+ sim_behavior_r1_pro: 31
1296
+ mecka_hands: 27
1297
+ real_r1_pro_sharpa: 28
1298
+ real_teleop_g1: 33
1299
+ max_length: 512
1300
+ num_views: 1
1301
+ tokenizer_path: /hfm/boqian/liboqian_data/checkpoints/umt5-xxl
1302
+ data_collator:
1303
+ _target_: groot.vla.model.dreamzero.transform.dreamzero_cotrain.DefaultDataCollator
1304
+ tokenizer_path: /hfm/boqian/liboqian_data/checkpoints/umt5-xxl
1305
+ max_length: 512
1306
+ num_views: 1
1307
+ embodiment_tag_mapping:
1308
+ real_gr1_arms_only: 0
1309
+ real_gr1_arms_only_annotated: 1
1310
+ real_gr1_arms_waist: 2
1311
+ real_gr1_arms_waist_annotated: 3
1312
+ dexmg_gr1_arms_only_inspire: 4
1313
+ dexmg_gr1_arms_only_fourier: 5
1314
+ dexmg_gr1_arms_waist_fourier: 6
1315
+ robocasa_single_arm: 7
1316
+ onex_eve_gripper: 8
1317
+ robocasa_gr1_arms_only_inspire_hands: 9
1318
+ robocasa_gr1_arms_only_fourier_hands: 10
1319
+ robocasa_gr1_fixed_lower_body_inspire_hands: 11
1320
+ robocasa_gr1_fixed_lower_body_fourier_hands: 12
1321
+ robocasa_panda_omron: 13
1322
+ robocasa_bimanual_panda_parallel_gripper: 15
1323
+ robocasa_bimanual_panda_inspire_hand: 16
1324
+ oxe_droid: 17
1325
+ oxe_fractal: 18
1326
+ oxe_language_table: 19
1327
+ oxe_bridge: 20
1328
+ real_panda_single_arm: 21
1329
+ hot3d_hands_only: 23
1330
+ gr1_unified: 24
1331
+ robocasa_gr1_arms_waist_fourier_hands: 25
1332
+ agibot: 26
1333
+ lapa: 27
1334
+ oxe_mutex: 28
1335
+ oxe_roboset: 29
1336
+ oxe_plex: 30
1337
+ dream: 31
1338
+ yam: 32
1339
+ xdof: 22
1340
+ gr1_unified_segmentation: 14
1341
+ language_table_sim: 7
1342
+ gr1_isaac: 0
1343
+ sim_behavior_r1_pro: 31
1344
+ mecka_hands: 27
1345
+ real_r1_pro_sharpa: 28
1346
+ real_teleop_g1: 33
1347
+ num_visual_tokens_per_frame: 16
1348
+ max_state_dim: 64
1349
+ max_action_dim: 36
1350
+ language_dropout_prob: 0.0
1351
+ model_specific_transform:
1352
+ _target_: groot.vla.model.dreamzero.transform.dreamzero_cotrain.DreamTransform
1353
+ default_instruction: Perform the default behavior.
1354
+ language_dropout_prob: 0.0
1355
+ always_use_default_instruction: false
1356
+ max_state_dim: 64
1357
+ max_action_dim: 36
1358
+ max_length: 512
1359
+ state_horizon: 1
1360
+ action_horizon: 24
1361
+ embodiment_tag_mapping:
1362
+ real_gr1_arms_only: 0
1363
+ real_gr1_arms_only_annotated: 1
1364
+ real_gr1_arms_waist: 2
1365
+ real_gr1_arms_waist_annotated: 3
1366
+ dexmg_gr1_arms_only_inspire: 4
1367
+ dexmg_gr1_arms_only_fourier: 5
1368
+ dexmg_gr1_arms_waist_fourier: 6
1369
+ robocasa_single_arm: 7
1370
+ onex_eve_gripper: 8
1371
+ robocasa_gr1_arms_only_inspire_hands: 9
1372
+ robocasa_gr1_arms_only_fourier_hands: 10
1373
+ robocasa_gr1_fixed_lower_body_inspire_hands: 11
1374
+ robocasa_gr1_fixed_lower_body_fourier_hands: 12
1375
+ robocasa_panda_omron: 13
1376
+ robocasa_bimanual_panda_parallel_gripper: 15
1377
+ robocasa_bimanual_panda_inspire_hand: 16
1378
+ oxe_droid: 17
1379
+ oxe_fractal: 18
1380
+ oxe_language_table: 19
1381
+ oxe_bridge: 20
1382
+ real_panda_single_arm: 21
1383
+ hot3d_hands_only: 23
1384
+ gr1_unified: 24
1385
+ robocasa_gr1_arms_waist_fourier_hands: 25
1386
+ agibot: 26
1387
+ lapa: 27
1388
+ oxe_mutex: 28
1389
+ oxe_roboset: 29
1390
+ oxe_plex: 30
1391
+ dream: 31
1392
+ yam: 32
1393
+ xdof: 22
1394
+ gr1_unified_segmentation: 14
1395
+ language_table_sim: 7
1396
+ gr1_isaac: 0
1397
+ sim_behavior_r1_pro: 31
1398
+ mecka_hands: 27
1399
+ real_r1_pro_sharpa: 28
1400
+ real_teleop_g1: 33
1401
+ tokenizer_path: /hfm/boqian/liboqian_data/checkpoints/umt5-xxl
1402
+ use_global_metadata: false
1403
+ num_frames: 33
1404
+ action_horizon: 24
1405
+ state_horizon: 1
1406
+ image_resolution_width: 320
1407
+ image_resolution_height: 176
1408
+ image_resolution_width_single_frame: 256
1409
+ image_resolution_height_single_frame: 256
1410
+ totensor_cfg:
1411
+ _target_: groot.vla.data.transform.VideoToTensor
1412
+ apply_to: ???
1413
+ crop_cfg:
1414
+ _target_: groot.vla.data.transform.VideoCrop
1415
+ apply_to: ???
1416
+ scale: 0.95
1417
+ mode: random
1418
+ resize_cfg:
1419
+ _target_: groot.vla.data.transform.VideoResize
1420
+ apply_to: ???
1421
+ height: 176
1422
+ width: 320
1423
+ interpolation: linear
1424
+ resize_cfg_single_frame:
1425
+ _target_: groot.vla.data.transform.VideoResize
1426
+ apply_to: ???
1427
+ height: 256
1428
+ width: 256
1429
+ interpolation: linear
1430
+ color_jitter_cfg:
1431
+ _target_: groot.vla.data.transform.VideoColorJitter
1432
+ apply_to: ???
1433
+ brightness: 0.3
1434
+ contrast: 0.4
1435
+ saturation: 0.5
1436
+ hue: 0.08
1437
+ random_grayscale_cfg:
1438
+ _target_: groot.vla.data.transform.VideoRandomGrayscale
1439
+ apply_to: ???
1440
+ p: 0.1
1441
+ random_posterize_cfg:
1442
+ _target_: groot.vla.data.transform.VideoRandomPosterize
1443
+ apply_to: ???
1444
+ bits: 4
1445
+ p: 0.1
1446
+ normalize_cfg:
1447
+ _target_: groot.vla.data.transform.VideoNormalize
1448
+ apply_to: ???
1449
+ mean:
1450
+ - 0.5
1451
+ - 0.5
1452
+ - 0.5
1453
+ std:
1454
+ - 0.5
1455
+ - 0.5
1456
+ - 0.5
1457
+ to_numpy_cfg:
1458
+ _target_: groot.vla.data.transform.VideoToNumpy
1459
+ apply_to: ???
1460
+ modality_config_oxe_droid:
1461
+ video:
1462
+ _target_: groot.vla.data.dataset.ModalityConfig
1463
+ delta_indices:
1464
+ - 0
1465
+ - 1
1466
+ - 2
1467
+ - 3
1468
+ - 4
1469
+ - 5
1470
+ - 6
1471
+ - 7
1472
+ - 8
1473
+ - 9
1474
+ - 10
1475
+ - 11
1476
+ - 12
1477
+ - 13
1478
+ - 14
1479
+ - 15
1480
+ - 16
1481
+ - 17
1482
+ - 18
1483
+ - 19
1484
+ - 20
1485
+ - 21
1486
+ - 22
1487
+ - 23
1488
+ - 24
1489
+ eval_delta_indices:
1490
+ - 0
1491
+ modality_keys:
1492
+ - video.exterior_image_1_left
1493
+ - video.exterior_image_2_left
1494
+ - video.wrist_image_left
1495
+ state:
1496
+ _target_: groot.vla.data.dataset.ModalityConfig
1497
+ delta_indices:
1498
+ - 0
1499
+ modality_keys:
1500
+ - state.joint_position
1501
+ - state.gripper_position
1502
+ action:
1503
+ _target_: groot.vla.data.dataset.ModalityConfig
1504
+ delta_indices:
1505
+ - 0
1506
+ - 1
1507
+ - 2
1508
+ - 3
1509
+ - 4
1510
+ - 5
1511
+ - 6
1512
+ - 7
1513
+ - 8
1514
+ - 9
1515
+ - 10
1516
+ - 11
1517
+ - 12
1518
+ - 13
1519
+ - 14
1520
+ - 15
1521
+ - 16
1522
+ - 17
1523
+ - 18
1524
+ - 19
1525
+ - 20
1526
+ - 21
1527
+ - 22
1528
+ - 23
1529
+ modality_keys:
1530
+ - action.joint_position
1531
+ - action.gripper_position
1532
+ language:
1533
+ _target_: groot.vla.data.dataset.ModalityConfig
1534
+ delta_indices:
1535
+ - 0
1536
+ modality_keys:
1537
+ - annotation.language.language_instruction
1538
+ - annotation.language.language_instruction_2
1539
+ - annotation.language.language_instruction_3
1540
+ lapa_action:
1541
+ _target_: groot.vla.data.dataset.ModalityConfig
1542
+ delta_indices:
1543
+ - 0
1544
+ modality_keys:
1545
+ - lapa_action
1546
+ transform_oxe_droid:
1547
+ _target_: groot.vla.data.transform.ComposedModalityTransform
1548
+ transforms:
1549
+ - _target_: groot.vla.data.transform.VideoToTensor
1550
+ apply_to:
1551
+ - video.exterior_image_1_left
1552
+ - video.exterior_image_2_left
1553
+ - video.wrist_image_left
1554
+ - _target_: groot.vla.data.transform.VideoCrop
1555
+ apply_to:
1556
+ - video.exterior_image_1_left
1557
+ - video.exterior_image_2_left
1558
+ - video.wrist_image_left
1559
+ scale: 0.95
1560
+ mode: random
1561
+ - _target_: groot.vla.data.transform.VideoResize
1562
+ apply_to:
1563
+ - video.exterior_image_1_left
1564
+ - video.exterior_image_2_left
1565
+ - video.wrist_image_left
1566
+ height: 176
1567
+ width: 320
1568
+ interpolation: linear
1569
+ - _target_: groot.vla.data.transform.VideoColorJitter
1570
+ apply_to:
1571
+ - video.exterior_image_1_left
1572
+ - video.exterior_image_2_left
1573
+ - video.wrist_image_left
1574
+ brightness: 0.3
1575
+ contrast: 0.4
1576
+ saturation: 0.5
1577
+ hue: 0.08
1578
+ - _target_: groot.vla.data.transform.VideoToNumpy
1579
+ apply_to:
1580
+ - video.exterior_image_1_left
1581
+ - video.exterior_image_2_left
1582
+ - video.wrist_image_left
1583
+ - _target_: groot.vla.data.transform.StateActionToTensor
1584
+ apply_to:
1585
+ - state.joint_position
1586
+ - state.gripper_position
1587
+ - _target_: groot.vla.data.transform.StateActionTransform
1588
+ apply_to:
1589
+ - state.joint_position
1590
+ - state.gripper_position
1591
+ normalization_modes:
1592
+ state.joint_position: q99
1593
+ state.gripper_position: q99
1594
+ - _target_: groot.vla.data.transform.StateActionToTensor
1595
+ apply_to:
1596
+ - action.joint_position
1597
+ - action.gripper_position
1598
+ - _target_: groot.vla.data.transform.StateActionTransform
1599
+ apply_to:
1600
+ - action.joint_position
1601
+ - action.gripper_position
1602
+ normalization_modes:
1603
+ action.joint_position: q99
1604
+ action.gripper_position: q99
1605
+ - _target_: groot.vla.data.transform.ConcatTransform
1606
+ video_concat_order:
1607
+ - video.exterior_image_1_left
1608
+ - video.exterior_image_2_left
1609
+ - video.wrist_image_left
1610
+ state_concat_order:
1611
+ - state.joint_position
1612
+ - state.gripper_position
1613
+ action_concat_order:
1614
+ - action.joint_position
1615
+ - action.gripper_position
1616
+ - _target_: groot.vla.model.dreamzero.transform.dreamzero_cotrain.DreamTransform
1617
+ default_instruction: Perform the default behavior.
1618
+ language_dropout_prob: 0.0
1619
+ always_use_default_instruction: false
1620
+ max_state_dim: 64
1621
+ max_action_dim: 36
1622
+ max_length: 512
1623
+ state_horizon: 1
1624
+ action_horizon: 24
1625
+ embodiment_tag_mapping:
1626
+ real_gr1_arms_only: 0
1627
+ real_gr1_arms_only_annotated: 1
1628
+ real_gr1_arms_waist: 2
1629
+ real_gr1_arms_waist_annotated: 3
1630
+ dexmg_gr1_arms_only_inspire: 4
1631
+ dexmg_gr1_arms_only_fourier: 5
1632
+ dexmg_gr1_arms_waist_fourier: 6
1633
+ robocasa_single_arm: 7
1634
+ onex_eve_gripper: 8
1635
+ robocasa_gr1_arms_only_inspire_hands: 9
1636
+ robocasa_gr1_arms_only_fourier_hands: 10
1637
+ robocasa_gr1_fixed_lower_body_inspire_hands: 11
1638
+ robocasa_gr1_fixed_lower_body_fourier_hands: 12
1639
+ robocasa_panda_omron: 13
1640
+ robocasa_bimanual_panda_parallel_gripper: 15
1641
+ robocasa_bimanual_panda_inspire_hand: 16
1642
+ oxe_droid: 17
1643
+ oxe_fractal: 18
1644
+ oxe_language_table: 19
1645
+ oxe_bridge: 20
1646
+ real_panda_single_arm: 21
1647
+ hot3d_hands_only: 23
1648
+ gr1_unified: 24
1649
+ robocasa_gr1_arms_waist_fourier_hands: 25
1650
+ agibot: 26
1651
+ lapa: 27
1652
+ oxe_mutex: 28
1653
+ oxe_roboset: 29
1654
+ oxe_plex: 30
1655
+ dream: 31
1656
+ yam: 32
1657
+ xdof: 22
1658
+ gr1_unified_segmentation: 14
1659
+ language_table_sim: 7
1660
+ gr1_isaac: 0
1661
+ sim_behavior_r1_pro: 31
1662
+ mecka_hands: 27
1663
+ real_r1_pro_sharpa: 28
1664
+ real_teleop_g1: 33
1665
+ tokenizer_path: /hfm/boqian/liboqian_data/checkpoints/umt5-xxl
1666
+ modality_config_agibot:
1667
+ video:
1668
+ _target_: groot.vla.data.dataset.ModalityConfig
1669
+ delta_indices:
1670
+ - 0
1671
+ - 1
1672
+ - 2
1673
+ - 3
1674
+ - 4
1675
+ - 5
1676
+ - 6
1677
+ - 7
1678
+ - 8
1679
+ - 9
1680
+ - 10
1681
+ - 11
1682
+ - 12
1683
+ - 13
1684
+ - 14
1685
+ - 15
1686
+ - 16
1687
+ - 17
1688
+ - 18
1689
+ - 19
1690
+ - 20
1691
+ - 21
1692
+ - 22
1693
+ - 23
1694
+ - 24
1695
+ eval_delta_indices:
1696
+ - -3
1697
+ - -2
1698
+ - -1
1699
+ - 0
1700
+ modality_keys:
1701
+ - video.top_head
1702
+ - video.hand_left
1703
+ - video.hand_right
1704
+ state:
1705
+ _target_: groot.vla.data.dataset.ModalityConfig
1706
+ delta_indices:
1707
+ - 0
1708
+ modality_keys:
1709
+ - state.left_arm_joint_position
1710
+ - state.right_arm_joint_position
1711
+ - state.left_effector_position
1712
+ - state.right_effector_position
1713
+ - state.head_position
1714
+ - state.waist_position
1715
+ action:
1716
+ _target_: groot.vla.data.dataset.ModalityConfig
1717
+ delta_indices:
1718
+ - 0
1719
+ - 1
1720
+ - 2
1721
+ - 3
1722
+ - 4
1723
+ - 5
1724
+ - 6
1725
+ - 7
1726
+ - 8
1727
+ - 9
1728
+ - 10
1729
+ - 11
1730
+ - 12
1731
+ - 13
1732
+ - 14
1733
+ - 15
1734
+ - 16
1735
+ - 17
1736
+ - 18
1737
+ - 19
1738
+ - 20
1739
+ - 21
1740
+ - 22
1741
+ - 23
1742
+ modality_keys:
1743
+ - action.left_arm_joint_position
1744
+ - action.right_arm_joint_position
1745
+ - action.left_effector_position
1746
+ - action.right_effector_position
1747
+ - action.head_position
1748
+ - action.waist_position
1749
+ - action.robot_velocity
1750
+ language:
1751
+ _target_: groot.vla.data.dataset.ModalityConfig
1752
+ delta_indices:
1753
+ - 0
1754
+ modality_keys:
1755
+ - annotation.language.action_text
1756
+ transform_agibot:
1757
+ _target_: groot.vla.data.transform.ComposedModalityTransform
1758
+ transforms:
1759
+ - _target_: groot.vla.data.transform.VideoToTensor
1760
+ apply_to:
1761
+ - video.top_head
1762
+ - video.hand_left
1763
+ - video.hand_right
1764
+ - _target_: groot.vla.data.transform.VideoCrop
1765
+ apply_to:
1766
+ - video.top_head
1767
+ - video.hand_left
1768
+ - video.hand_right
1769
+ scale: 0.95
1770
+ mode: random
1771
+ - _target_: groot.vla.data.transform.VideoResize
1772
+ apply_to:
1773
+ - video.top_head
1774
+ - video.hand_left
1775
+ - video.hand_right
1776
+ height: 176
1777
+ width: 320
1778
+ interpolation: linear
1779
+ - _target_: groot.vla.data.transform.VideoColorJitter
1780
+ apply_to:
1781
+ - video.top_head
1782
+ - video.hand_left
1783
+ - video.hand_right
1784
+ brightness: 0.3
1785
+ contrast: 0.4
1786
+ saturation: 0.5
1787
+ hue: 0.08
1788
+ - _target_: groot.vla.data.transform.VideoToNumpy
1789
+ apply_to:
1790
+ - video.top_head
1791
+ - video.hand_left
1792
+ - video.hand_right
1793
+ - _target_: groot.vla.data.transform.StateActionToTensor
1794
+ apply_to:
1795
+ - state.left_arm_joint_position
1796
+ - state.right_arm_joint_position
1797
+ - state.left_effector_position
1798
+ - state.right_effector_position
1799
+ - state.head_position
1800
+ - state.waist_position
1801
+ - _target_: groot.vla.data.transform.StateActionTransform
1802
+ apply_to:
1803
+ - state.left_arm_joint_position
1804
+ - state.right_arm_joint_position
1805
+ - state.left_effector_position
1806
+ - state.right_effector_position
1807
+ - state.head_position
1808
+ - state.waist_position
1809
+ normalization_modes:
1810
+ state.left_arm_joint_position: q99
1811
+ state.right_arm_joint_position: q99
1812
+ state.left_effector_position: q99
1813
+ state.right_effector_position: q99
1814
+ state.head_position: q99
1815
+ state.waist_position: q99
1816
+ - _target_: groot.vla.data.transform.StateActionToTensor
1817
+ apply_to:
1818
+ - action.left_arm_joint_position
1819
+ - action.right_arm_joint_position
1820
+ - action.left_effector_position
1821
+ - action.right_effector_position
1822
+ - action.head_position
1823
+ - action.waist_position
1824
+ - action.robot_velocity
1825
+ - _target_: groot.vla.data.transform.StateActionTransform
1826
+ apply_to:
1827
+ - action.left_arm_joint_position
1828
+ - action.right_arm_joint_position
1829
+ - action.left_effector_position
1830
+ - action.right_effector_position
1831
+ - action.head_position
1832
+ - action.waist_position
1833
+ - action.robot_velocity
1834
+ normalization_modes:
1835
+ action.left_arm_joint_position: q99
1836
+ action.right_arm_joint_position: q99
1837
+ action.left_effector_position: q99
1838
+ action.right_effector_position: q99
1839
+ action.head_position: q99
1840
+ action.waist_position: q99
1841
+ action.robot_velocity: q99
1842
+ - _target_: groot.vla.data.transform.ConcatTransform
1843
+ video_concat_order:
1844
+ - video.top_head
1845
+ - video.hand_left
1846
+ - video.hand_right
1847
+ state_concat_order:
1848
+ - state.left_arm_joint_position
1849
+ - state.right_arm_joint_position
1850
+ - state.left_effector_position
1851
+ - state.right_effector_position
1852
+ - state.head_position
1853
+ - state.waist_position
1854
+ action_concat_order:
1855
+ - action.left_arm_joint_position
1856
+ - action.right_arm_joint_position
1857
+ - action.left_effector_position
1858
+ - action.right_effector_position
1859
+ - action.head_position
1860
+ - action.waist_position
1861
+ - action.robot_velocity
1862
+ - _target_: groot.vla.model.dreamzero.transform.dreamzero_cotrain.DreamTransform
1863
+ default_instruction: Perform the default behavior.
1864
+ language_dropout_prob: 0.0
1865
+ always_use_default_instruction: false
1866
+ max_state_dim: 64
1867
+ max_action_dim: 36
1868
+ max_length: 512
1869
+ state_horizon: 1
1870
+ action_horizon: 24
1871
+ embodiment_tag_mapping:
1872
+ real_gr1_arms_only: 0
1873
+ real_gr1_arms_only_annotated: 1
1874
+ real_gr1_arms_waist: 2
1875
+ real_gr1_arms_waist_annotated: 3
1876
+ dexmg_gr1_arms_only_inspire: 4
1877
+ dexmg_gr1_arms_only_fourier: 5
1878
+ dexmg_gr1_arms_waist_fourier: 6
1879
+ robocasa_single_arm: 7
1880
+ onex_eve_gripper: 8
1881
+ robocasa_gr1_arms_only_inspire_hands: 9
1882
+ robocasa_gr1_arms_only_fourier_hands: 10
1883
+ robocasa_gr1_fixed_lower_body_inspire_hands: 11
1884
+ robocasa_gr1_fixed_lower_body_fourier_hands: 12
1885
+ robocasa_panda_omron: 13
1886
+ robocasa_bimanual_panda_parallel_gripper: 15
1887
+ robocasa_bimanual_panda_inspire_hand: 16
1888
+ oxe_droid: 17
1889
+ oxe_fractal: 18
1890
+ oxe_language_table: 19
1891
+ oxe_bridge: 20
1892
+ real_panda_single_arm: 21
1893
+ hot3d_hands_only: 23
1894
+ gr1_unified: 24
1895
+ robocasa_gr1_arms_waist_fourier_hands: 25
1896
+ agibot: 26
1897
+ lapa: 27
1898
+ oxe_mutex: 28
1899
+ oxe_roboset: 29
1900
+ oxe_plex: 30
1901
+ dream: 31
1902
+ yam: 32
1903
+ xdof: 22
1904
+ gr1_unified_segmentation: 14
1905
+ language_table_sim: 7
1906
+ gr1_isaac: 0
1907
+ sim_behavior_r1_pro: 31
1908
+ mecka_hands: 27
1909
+ real_r1_pro_sharpa: 28
1910
+ real_teleop_g1: 33
1911
+ tokenizer_path: /hfm/boqian/liboqian_data/checkpoints/umt5-xxl
1912
+ modality_config_yam:
1913
+ video:
1914
+ _target_: groot.vla.data.dataset.ModalityConfig
1915
+ delta_indices:
1916
+ - 0
1917
+ - 1
1918
+ - 2
1919
+ - 3
1920
+ - 4
1921
+ - 5
1922
+ - 6
1923
+ - 7
1924
+ - 8
1925
+ - 9
1926
+ - 10
1927
+ - 11
1928
+ - 12
1929
+ - 13
1930
+ - 14
1931
+ - 15
1932
+ - 16
1933
+ - 17
1934
+ - 18
1935
+ - 19
1936
+ - 20
1937
+ - 21
1938
+ - 22
1939
+ - 23
1940
+ - 24
1941
+ eval_delta_indices:
1942
+ - 0
1943
+ modality_keys:
1944
+ - video.top_camera-images-rgb
1945
+ - video.left_camera-images-rgb
1946
+ - video.right_camera-images-rgb
1947
+ state:
1948
+ _target_: groot.vla.data.dataset.ModalityConfig
1949
+ delta_indices:
1950
+ - 0
1951
+ modality_keys:
1952
+ - state.left_joint_pos
1953
+ - state.left_gripper_pos
1954
+ - state.right_joint_pos
1955
+ - state.right_gripper_pos
1956
+ action:
1957
+ _target_: groot.vla.data.dataset.ModalityConfig
1958
+ delta_indices:
1959
+ - 0
1960
+ - 1
1961
+ - 2
1962
+ - 3
1963
+ - 4
1964
+ - 5
1965
+ - 6
1966
+ - 7
1967
+ - 8
1968
+ - 9
1969
+ - 10
1970
+ - 11
1971
+ - 12
1972
+ - 13
1973
+ - 14
1974
+ - 15
1975
+ - 16
1976
+ - 17
1977
+ - 18
1978
+ - 19
1979
+ - 20
1980
+ - 21
1981
+ - 22
1982
+ - 23
1983
+ modality_keys:
1984
+ - action.left_joint_pos
1985
+ - action.left_gripper_pos
1986
+ - action.right_joint_pos
1987
+ - action.right_gripper_pos
1988
+ language:
1989
+ _target_: groot.vla.data.dataset.ModalityConfig
1990
+ delta_indices:
1991
+ - 0
1992
+ modality_keys:
1993
+ - annotation.task
1994
+ transform_yam:
1995
+ _target_: groot.vla.data.transform.ComposedModalityTransform
1996
+ transforms:
1997
+ - _target_: groot.vla.data.transform.VideoToTensor
1998
+ apply_to:
1999
+ - video.top_camera-images-rgb
2000
+ - video.left_camera-images-rgb
2001
+ - video.right_camera-images-rgb
2002
+ - _target_: groot.vla.data.transform.VideoCrop
2003
+ apply_to:
2004
+ - video.top_camera-images-rgb
2005
+ - video.left_camera-images-rgb
2006
+ - video.right_camera-images-rgb
2007
+ scale: 0.95
2008
+ mode: random
2009
+ - _target_: groot.vla.data.transform.VideoResize
2010
+ apply_to:
2011
+ - video.top_camera-images-rgb
2012
+ - video.left_camera-images-rgb
2013
+ - video.right_camera-images-rgb
2014
+ height: 176
2015
+ width: 320
2016
+ interpolation: linear
2017
+ - _target_: groot.vla.data.transform.VideoColorJitter
2018
+ apply_to:
2019
+ - video.top_camera-images-rgb
2020
+ - video.left_camera-images-rgb
2021
+ - video.right_camera-images-rgb
2022
+ brightness: 0.3
2023
+ contrast: 0.4
2024
+ saturation: 0.5
2025
+ hue: 0.08
2026
+ - _target_: groot.vla.data.transform.VideoToNumpy
2027
+ apply_to:
2028
+ - video.top_camera-images-rgb
2029
+ - video.left_camera-images-rgb
2030
+ - video.right_camera-images-rgb
2031
+ - _target_: groot.vla.data.transform.StateActionToTensor
2032
+ apply_to:
2033
+ - state.left_joint_pos
2034
+ - state.left_gripper_pos
2035
+ - state.right_joint_pos
2036
+ - state.right_gripper_pos
2037
+ - _target_: groot.vla.data.transform.StateActionTransform
2038
+ apply_to:
2039
+ - state.left_joint_pos
2040
+ - state.left_gripper_pos
2041
+ - state.right_joint_pos
2042
+ - state.right_gripper_pos
2043
+ normalization_modes:
2044
+ state.left_joint_pos: q99
2045
+ state.left_gripper_pos: q99
2046
+ state.right_joint_pos: q99
2047
+ state.right_gripper_pos: q99
2048
+ - _target_: groot.vla.data.transform.StateActionToTensor
2049
+ apply_to:
2050
+ - action.left_joint_pos
2051
+ - action.left_gripper_pos
2052
+ - action.right_joint_pos
2053
+ - action.right_gripper_pos
2054
+ - _target_: groot.vla.data.transform.StateActionTransform
2055
+ apply_to:
2056
+ - action.left_joint_pos
2057
+ - action.left_gripper_pos
2058
+ - action.right_joint_pos
2059
+ - action.right_gripper_pos
2060
+ normalization_modes:
2061
+ action.left_joint_pos: q99
2062
+ action.left_gripper_pos: q99
2063
+ action.right_joint_pos: q99
2064
+ action.right_gripper_pos: q99
2065
+ - _target_: groot.vla.data.transform.ConcatTransform
2066
+ video_concat_order:
2067
+ - video.top_camera-images-rgb
2068
+ - video.left_camera-images-rgb
2069
+ - video.right_camera-images-rgb
2070
+ state_concat_order:
2071
+ - state.left_joint_pos
2072
+ - state.left_gripper_pos
2073
+ - state.right_joint_pos
2074
+ - state.right_gripper_pos
2075
+ action_concat_order:
2076
+ - action.left_joint_pos
2077
+ - action.left_gripper_pos
2078
+ - action.right_joint_pos
2079
+ - action.right_gripper_pos
2080
+ - _target_: groot.vla.model.dreamzero.transform.dreamzero_cotrain.DreamTransform
2081
+ default_instruction: Perform the default behavior.
2082
+ language_dropout_prob: 0.0
2083
+ always_use_default_instruction: false
2084
+ max_state_dim: 64
2085
+ max_action_dim: 36
2086
+ max_length: 512
2087
+ state_horizon: 1
2088
+ action_horizon: 24
2089
+ embodiment_tag_mapping:
2090
+ real_gr1_arms_only: 0
2091
+ real_gr1_arms_only_annotated: 1
2092
+ real_gr1_arms_waist: 2
2093
+ real_gr1_arms_waist_annotated: 3
2094
+ dexmg_gr1_arms_only_inspire: 4
2095
+ dexmg_gr1_arms_only_fourier: 5
2096
+ dexmg_gr1_arms_waist_fourier: 6
2097
+ robocasa_single_arm: 7
2098
+ onex_eve_gripper: 8
2099
+ robocasa_gr1_arms_only_inspire_hands: 9
2100
+ robocasa_gr1_arms_only_fourier_hands: 10
2101
+ robocasa_gr1_fixed_lower_body_inspire_hands: 11
2102
+ robocasa_gr1_fixed_lower_body_fourier_hands: 12
2103
+ robocasa_panda_omron: 13
2104
+ robocasa_bimanual_panda_parallel_gripper: 15
2105
+ robocasa_bimanual_panda_inspire_hand: 16
2106
+ oxe_droid: 17
2107
+ oxe_fractal: 18
2108
+ oxe_language_table: 19
2109
+ oxe_bridge: 20
2110
+ real_panda_single_arm: 21
2111
+ hot3d_hands_only: 23
2112
+ gr1_unified: 24
2113
+ robocasa_gr1_arms_waist_fourier_hands: 25
2114
+ agibot: 26
2115
+ lapa: 27
2116
+ oxe_mutex: 28
2117
+ oxe_roboset: 29
2118
+ oxe_plex: 30
2119
+ dream: 31
2120
+ yam: 32
2121
+ xdof: 22
2122
+ gr1_unified_segmentation: 14
2123
+ language_table_sim: 7
2124
+ gr1_isaac: 0
2125
+ sim_behavior_r1_pro: 31
2126
+ mecka_hands: 27
2127
+ real_r1_pro_sharpa: 28
2128
+ real_teleop_g1: 33
2129
+ tokenizer_path: /hfm/boqian/liboqian_data/checkpoints/umt5-xxl
2130
+ modality_config_real_teleop_g1:
2131
+ video:
2132
+ _target_: groot.vla.data.dataset.ModalityConfig
2133
+ delta_indices:
2134
+ - 0
2135
+ - 1
2136
+ - 2
2137
+ - 3
2138
+ - 4
2139
+ - 5
2140
+ - 6
2141
+ - 7
2142
+ - 8
2143
+ - 9
2144
+ - 10
2145
+ - 11
2146
+ - 12
2147
+ - 13
2148
+ - 14
2149
+ - 15
2150
+ - 16
2151
+ - 17
2152
+ - 18
2153
+ - 19
2154
+ - 20
2155
+ - 21
2156
+ - 22
2157
+ - 23
2158
+ - 24
2159
+ eval_delta_indices:
2160
+ - 0
2161
+ modality_keys:
2162
+ - video.egocentric
2163
+ state:
2164
+ _target_: groot.vla.data.dataset.ModalityConfig
2165
+ delta_indices:
2166
+ - 0
2167
+ modality_keys:
2168
+ - state.left_hand
2169
+ - state.right_hand
2170
+ - state.left_arm
2171
+ - state.right_arm
2172
+ - state.rpy
2173
+ - state.height
2174
+ action:
2175
+ _target_: groot.vla.data.dataset.ModalityConfig
2176
+ delta_indices:
2177
+ - 0
2178
+ - 1
2179
+ - 2
2180
+ - 3
2181
+ - 4
2182
+ - 5
2183
+ - 6
2184
+ - 7
2185
+ - 8
2186
+ - 9
2187
+ - 10
2188
+ - 11
2189
+ - 12
2190
+ - 13
2191
+ - 14
2192
+ - 15
2193
+ - 16
2194
+ - 17
2195
+ - 18
2196
+ - 19
2197
+ - 20
2198
+ - 21
2199
+ - 22
2200
+ - 23
2201
+ modality_keys:
2202
+ - action.left_hand
2203
+ - action.right_hand
2204
+ - action.left_arm
2205
+ - action.right_arm
2206
+ - action.rpy
2207
+ - action.height
2208
+ - action.torso_vx
2209
+ - action.torso_vy
2210
+ - action.torso_vyaw
2211
+ - action.torso_dyaw
2212
+ language:
2213
+ _target_: groot.vla.data.dataset.ModalityConfig
2214
+ delta_indices:
2215
+ - 0
2216
+ modality_keys:
2217
+ - annotation.language.language_instruction
2218
+ transform_real_teleop_g1:
2219
+ _target_: groot.vla.data.transform.ComposedModalityTransform
2220
+ transforms:
2221
+ - _target_: groot.vla.data.transform.VideoToTensor
2222
+ apply_to:
2223
+ - video.egocentric
2224
+ - _target_: groot.vla.data.transform.VideoCrop
2225
+ apply_to:
2226
+ - video.egocentric
2227
+ scale: 0.95
2228
+ mode: random
2229
+ - _target_: groot.vla.data.transform.VideoResize
2230
+ apply_to:
2231
+ - video.egocentric
2232
+ height: 176
2233
+ width: 320
2234
+ interpolation: linear
2235
+ - _target_: groot.vla.data.transform.VideoColorJitter
2236
+ apply_to:
2237
+ - video.egocentric
2238
+ brightness: 0.3
2239
+ contrast: 0.4
2240
+ saturation: 0.5
2241
+ hue: 0.08
2242
+ - _target_: groot.vla.data.transform.VideoToNumpy
2243
+ apply_to:
2244
+ - video.egocentric
2245
+ - _target_: groot.vla.data.transform.StateActionToTensor
2246
+ apply_to:
2247
+ - state.left_hand
2248
+ - state.right_hand
2249
+ - state.left_arm
2250
+ - state.right_arm
2251
+ - state.rpy
2252
+ - state.height
2253
+ - _target_: groot.vla.data.transform.StateActionTransform
2254
+ apply_to:
2255
+ - state.left_hand
2256
+ - state.right_hand
2257
+ - state.left_arm
2258
+ - state.right_arm
2259
+ - state.rpy
2260
+ - state.height
2261
+ normalization_modes:
2262
+ state.left_hand: q99
2263
+ state.right_hand: q99
2264
+ state.left_arm: q99
2265
+ state.right_arm: q99
2266
+ state.rpy: q99
2267
+ state.height: q99
2268
+ - _target_: groot.vla.data.transform.StateActionToTensor
2269
+ apply_to:
2270
+ - action.left_hand
2271
+ - action.right_hand
2272
+ - action.left_arm
2273
+ - action.right_arm
2274
+ - action.rpy
2275
+ - action.height
2276
+ - action.torso_vx
2277
+ - action.torso_vy
2278
+ - action.torso_vyaw
2279
+ - action.torso_dyaw
2280
+ - _target_: groot.vla.data.transform.StateActionTransform
2281
+ apply_to:
2282
+ - action.left_hand
2283
+ - action.right_hand
2284
+ - action.left_arm
2285
+ - action.right_arm
2286
+ - action.rpy
2287
+ - action.height
2288
+ - action.torso_vx
2289
+ - action.torso_vy
2290
+ - action.torso_vyaw
2291
+ - action.torso_dyaw
2292
+ normalization_modes:
2293
+ action.left_hand: q99
2294
+ action.right_hand: q99
2295
+ action.left_arm: q99
2296
+ action.right_arm: q99
2297
+ action.rpy: q99
2298
+ action.height: q99
2299
+ action.torso_vx: q99
2300
+ action.torso_vy: q99
2301
+ action.torso_vyaw: q99
2302
+ action.torso_dyaw: q99
2303
+ - _target_: groot.vla.data.transform.ConcatTransform
2304
+ video_concat_order:
2305
+ - video.egocentric
2306
+ state_concat_order:
2307
+ - state.left_hand
2308
+ - state.right_hand
2309
+ - state.left_arm
2310
+ - state.right_arm
2311
+ - state.rpy
2312
+ - state.height
2313
+ action_concat_order:
2314
+ - action.left_hand
2315
+ - action.right_hand
2316
+ - action.left_arm
2317
+ - action.right_arm
2318
+ - action.rpy
2319
+ - action.height
2320
+ - action.torso_vx
2321
+ - action.torso_vy
2322
+ - action.torso_vyaw
2323
+ - action.torso_dyaw
2324
+ - _target_: groot.vla.model.dreamzero.transform.dreamzero_cotrain.DreamTransform
2325
+ default_instruction: Perform the default behavior.
2326
+ language_dropout_prob: 0.0
2327
+ always_use_default_instruction: false
2328
+ max_state_dim: 64
2329
+ max_action_dim: 36
2330
+ max_length: 512
2331
+ state_horizon: 1
2332
+ action_horizon: 24
2333
+ embodiment_tag_mapping:
2334
+ real_gr1_arms_only: 0
2335
+ real_gr1_arms_only_annotated: 1
2336
+ real_gr1_arms_waist: 2
2337
+ real_gr1_arms_waist_annotated: 3
2338
+ dexmg_gr1_arms_only_inspire: 4
2339
+ dexmg_gr1_arms_only_fourier: 5
2340
+ dexmg_gr1_arms_waist_fourier: 6
2341
+ robocasa_single_arm: 7
2342
+ onex_eve_gripper: 8
2343
+ robocasa_gr1_arms_only_inspire_hands: 9
2344
+ robocasa_gr1_arms_only_fourier_hands: 10
2345
+ robocasa_gr1_fixed_lower_body_inspire_hands: 11
2346
+ robocasa_gr1_fixed_lower_body_fourier_hands: 12
2347
+ robocasa_panda_omron: 13
2348
+ robocasa_bimanual_panda_parallel_gripper: 15
2349
+ robocasa_bimanual_panda_inspire_hand: 16
2350
+ oxe_droid: 17
2351
+ oxe_fractal: 18
2352
+ oxe_language_table: 19
2353
+ oxe_bridge: 20
2354
+ real_panda_single_arm: 21
2355
+ hot3d_hands_only: 23
2356
+ gr1_unified: 24
2357
+ robocasa_gr1_arms_waist_fourier_hands: 25
2358
+ agibot: 26
2359
+ lapa: 27
2360
+ oxe_mutex: 28
2361
+ oxe_roboset: 29
2362
+ oxe_plex: 30
2363
+ dream: 31
2364
+ yam: 32
2365
+ xdof: 22
2366
+ gr1_unified_segmentation: 14
2367
+ language_table_sim: 7
2368
+ gr1_isaac: 0
2369
+ sim_behavior_r1_pro: 31
2370
+ mecka_hands: 27
2371
+ real_r1_pro_sharpa: 28
2372
+ real_teleop_g1: 33
2373
+ tokenizer_path: /hfm/boqian/liboqian_data/checkpoints/umt5-xxl
2374
+ modality_configs:
2375
+ oxe_droid:
2376
+ video:
2377
+ _target_: groot.vla.data.dataset.ModalityConfig
2378
+ delta_indices:
2379
+ - 0
2380
+ - 1
2381
+ - 2
2382
+ - 3
2383
+ - 4
2384
+ - 5
2385
+ - 6
2386
+ - 7
2387
+ - 8
2388
+ - 9
2389
+ - 10
2390
+ - 11
2391
+ - 12
2392
+ - 13
2393
+ - 14
2394
+ - 15
2395
+ - 16
2396
+ - 17
2397
+ - 18
2398
+ - 19
2399
+ - 20
2400
+ - 21
2401
+ - 22
2402
+ - 23
2403
+ - 24
2404
+ eval_delta_indices:
2405
+ - 0
2406
+ modality_keys:
2407
+ - video.exterior_image_1_left
2408
+ - video.exterior_image_2_left
2409
+ - video.wrist_image_left
2410
+ state:
2411
+ _target_: groot.vla.data.dataset.ModalityConfig
2412
+ delta_indices:
2413
+ - 0
2414
+ modality_keys:
2415
+ - state.joint_position
2416
+ - state.gripper_position
2417
+ action:
2418
+ _target_: groot.vla.data.dataset.ModalityConfig
2419
+ delta_indices:
2420
+ - 0
2421
+ - 1
2422
+ - 2
2423
+ - 3
2424
+ - 4
2425
+ - 5
2426
+ - 6
2427
+ - 7
2428
+ - 8
2429
+ - 9
2430
+ - 10
2431
+ - 11
2432
+ - 12
2433
+ - 13
2434
+ - 14
2435
+ - 15
2436
+ - 16
2437
+ - 17
2438
+ - 18
2439
+ - 19
2440
+ - 20
2441
+ - 21
2442
+ - 22
2443
+ - 23
2444
+ modality_keys:
2445
+ - action.joint_position
2446
+ - action.gripper_position
2447
+ language:
2448
+ _target_: groot.vla.data.dataset.ModalityConfig
2449
+ delta_indices:
2450
+ - 0
2451
+ modality_keys:
2452
+ - annotation.language.language_instruction
2453
+ - annotation.language.language_instruction_2
2454
+ - annotation.language.language_instruction_3
2455
+ lapa_action:
2456
+ _target_: groot.vla.data.dataset.ModalityConfig
2457
+ delta_indices:
2458
+ - 0
2459
+ modality_keys:
2460
+ - lapa_action
2461
+ agibot:
2462
+ video:
2463
+ _target_: groot.vla.data.dataset.ModalityConfig
2464
+ delta_indices:
2465
+ - 0
2466
+ - 1
2467
+ - 2
2468
+ - 3
2469
+ - 4
2470
+ - 5
2471
+ - 6
2472
+ - 7
2473
+ - 8
2474
+ - 9
2475
+ - 10
2476
+ - 11
2477
+ - 12
2478
+ - 13
2479
+ - 14
2480
+ - 15
2481
+ - 16
2482
+ - 17
2483
+ - 18
2484
+ - 19
2485
+ - 20
2486
+ - 21
2487
+ - 22
2488
+ - 23
2489
+ - 24
2490
+ eval_delta_indices:
2491
+ - -3
2492
+ - -2
2493
+ - -1
2494
+ - 0
2495
+ modality_keys:
2496
+ - video.top_head
2497
+ - video.hand_left
2498
+ - video.hand_right
2499
+ state:
2500
+ _target_: groot.vla.data.dataset.ModalityConfig
2501
+ delta_indices:
2502
+ - 0
2503
+ modality_keys:
2504
+ - state.left_arm_joint_position
2505
+ - state.right_arm_joint_position
2506
+ - state.left_effector_position
2507
+ - state.right_effector_position
2508
+ - state.head_position
2509
+ - state.waist_position
2510
+ action:
2511
+ _target_: groot.vla.data.dataset.ModalityConfig
2512
+ delta_indices:
2513
+ - 0
2514
+ - 1
2515
+ - 2
2516
+ - 3
2517
+ - 4
2518
+ - 5
2519
+ - 6
2520
+ - 7
2521
+ - 8
2522
+ - 9
2523
+ - 10
2524
+ - 11
2525
+ - 12
2526
+ - 13
2527
+ - 14
2528
+ - 15
2529
+ - 16
2530
+ - 17
2531
+ - 18
2532
+ - 19
2533
+ - 20
2534
+ - 21
2535
+ - 22
2536
+ - 23
2537
+ modality_keys:
2538
+ - action.left_arm_joint_position
2539
+ - action.right_arm_joint_position
2540
+ - action.left_effector_position
2541
+ - action.right_effector_position
2542
+ - action.head_position
2543
+ - action.waist_position
2544
+ - action.robot_velocity
2545
+ language:
2546
+ _target_: groot.vla.data.dataset.ModalityConfig
2547
+ delta_indices:
2548
+ - 0
2549
+ modality_keys:
2550
+ - annotation.language.action_text
2551
+ yam:
2552
+ video:
2553
+ _target_: groot.vla.data.dataset.ModalityConfig
2554
+ delta_indices:
2555
+ - 0
2556
+ - 1
2557
+ - 2
2558
+ - 3
2559
+ - 4
2560
+ - 5
2561
+ - 6
2562
+ - 7
2563
+ - 8
2564
+ - 9
2565
+ - 10
2566
+ - 11
2567
+ - 12
2568
+ - 13
2569
+ - 14
2570
+ - 15
2571
+ - 16
2572
+ - 17
2573
+ - 18
2574
+ - 19
2575
+ - 20
2576
+ - 21
2577
+ - 22
2578
+ - 23
2579
+ - 24
2580
+ eval_delta_indices:
2581
+ - 0
2582
+ modality_keys:
2583
+ - video.top_camera-images-rgb
2584
+ - video.left_camera-images-rgb
2585
+ - video.right_camera-images-rgb
2586
+ state:
2587
+ _target_: groot.vla.data.dataset.ModalityConfig
2588
+ delta_indices:
2589
+ - 0
2590
+ modality_keys:
2591
+ - state.left_joint_pos
2592
+ - state.left_gripper_pos
2593
+ - state.right_joint_pos
2594
+ - state.right_gripper_pos
2595
+ action:
2596
+ _target_: groot.vla.data.dataset.ModalityConfig
2597
+ delta_indices:
2598
+ - 0
2599
+ - 1
2600
+ - 2
2601
+ - 3
2602
+ - 4
2603
+ - 5
2604
+ - 6
2605
+ - 7
2606
+ - 8
2607
+ - 9
2608
+ - 10
2609
+ - 11
2610
+ - 12
2611
+ - 13
2612
+ - 14
2613
+ - 15
2614
+ - 16
2615
+ - 17
2616
+ - 18
2617
+ - 19
2618
+ - 20
2619
+ - 21
2620
+ - 22
2621
+ - 23
2622
+ modality_keys:
2623
+ - action.left_joint_pos
2624
+ - action.left_gripper_pos
2625
+ - action.right_joint_pos
2626
+ - action.right_gripper_pos
2627
+ language:
2628
+ _target_: groot.vla.data.dataset.ModalityConfig
2629
+ delta_indices:
2630
+ - 0
2631
+ modality_keys:
2632
+ - annotation.task
2633
+ real_teleop_g1:
2634
+ video:
2635
+ _target_: groot.vla.data.dataset.ModalityConfig
2636
+ delta_indices:
2637
+ - 0
2638
+ - 1
2639
+ - 2
2640
+ - 3
2641
+ - 4
2642
+ - 5
2643
+ - 6
2644
+ - 7
2645
+ - 8
2646
+ - 9
2647
+ - 10
2648
+ - 11
2649
+ - 12
2650
+ - 13
2651
+ - 14
2652
+ - 15
2653
+ - 16
2654
+ - 17
2655
+ - 18
2656
+ - 19
2657
+ - 20
2658
+ - 21
2659
+ - 22
2660
+ - 23
2661
+ - 24
2662
+ eval_delta_indices:
2663
+ - 0
2664
+ modality_keys:
2665
+ - video.egocentric
2666
+ state:
2667
+ _target_: groot.vla.data.dataset.ModalityConfig
2668
+ delta_indices:
2669
+ - 0
2670
+ modality_keys:
2671
+ - state.left_hand
2672
+ - state.right_hand
2673
+ - state.left_arm
2674
+ - state.right_arm
2675
+ - state.rpy
2676
+ - state.height
2677
+ action:
2678
+ _target_: groot.vla.data.dataset.ModalityConfig
2679
+ delta_indices:
2680
+ - 0
2681
+ - 1
2682
+ - 2
2683
+ - 3
2684
+ - 4
2685
+ - 5
2686
+ - 6
2687
+ - 7
2688
+ - 8
2689
+ - 9
2690
+ - 10
2691
+ - 11
2692
+ - 12
2693
+ - 13
2694
+ - 14
2695
+ - 15
2696
+ - 16
2697
+ - 17
2698
+ - 18
2699
+ - 19
2700
+ - 20
2701
+ - 21
2702
+ - 22
2703
+ - 23
2704
+ modality_keys:
2705
+ - action.left_hand
2706
+ - action.right_hand
2707
+ - action.left_arm
2708
+ - action.right_arm
2709
+ - action.rpy
2710
+ - action.height
2711
+ - action.torso_vx
2712
+ - action.torso_vy
2713
+ - action.torso_vyaw
2714
+ - action.torso_dyaw
2715
+ language:
2716
+ _target_: groot.vla.data.dataset.ModalityConfig
2717
+ delta_indices:
2718
+ - 0
2719
+ modality_keys:
2720
+ - annotation.language.language_instruction
2721
+ transforms:
2722
+ oxe_droid:
2723
+ _target_: groot.vla.data.transform.ComposedModalityTransform
2724
+ transforms:
2725
+ - _target_: groot.vla.data.transform.VideoToTensor
2726
+ apply_to:
2727
+ - video.exterior_image_1_left
2728
+ - video.exterior_image_2_left
2729
+ - video.wrist_image_left
2730
+ - _target_: groot.vla.data.transform.VideoCrop
2731
+ apply_to:
2732
+ - video.exterior_image_1_left
2733
+ - video.exterior_image_2_left
2734
+ - video.wrist_image_left
2735
+ scale: 0.95
2736
+ mode: random
2737
+ - _target_: groot.vla.data.transform.VideoResize
2738
+ apply_to:
2739
+ - video.exterior_image_1_left
2740
+ - video.exterior_image_2_left
2741
+ - video.wrist_image_left
2742
+ height: 176
2743
+ width: 320
2744
+ interpolation: linear
2745
+ - _target_: groot.vla.data.transform.VideoColorJitter
2746
+ apply_to:
2747
+ - video.exterior_image_1_left
2748
+ - video.exterior_image_2_left
2749
+ - video.wrist_image_left
2750
+ brightness: 0.3
2751
+ contrast: 0.4
2752
+ saturation: 0.5
2753
+ hue: 0.08
2754
+ - _target_: groot.vla.data.transform.VideoToNumpy
2755
+ apply_to:
2756
+ - video.exterior_image_1_left
2757
+ - video.exterior_image_2_left
2758
+ - video.wrist_image_left
2759
+ - _target_: groot.vla.data.transform.StateActionToTensor
2760
+ apply_to:
2761
+ - state.joint_position
2762
+ - state.gripper_position
2763
+ - _target_: groot.vla.data.transform.StateActionTransform
2764
+ apply_to:
2765
+ - state.joint_position
2766
+ - state.gripper_position
2767
+ normalization_modes:
2768
+ state.joint_position: q99
2769
+ state.gripper_position: q99
2770
+ - _target_: groot.vla.data.transform.StateActionToTensor
2771
+ apply_to:
2772
+ - action.joint_position
2773
+ - action.gripper_position
2774
+ - _target_: groot.vla.data.transform.StateActionTransform
2775
+ apply_to:
2776
+ - action.joint_position
2777
+ - action.gripper_position
2778
+ normalization_modes:
2779
+ action.joint_position: q99
2780
+ action.gripper_position: q99
2781
+ - _target_: groot.vla.data.transform.ConcatTransform
2782
+ video_concat_order:
2783
+ - video.exterior_image_1_left
2784
+ - video.exterior_image_2_left
2785
+ - video.wrist_image_left
2786
+ state_concat_order:
2787
+ - state.joint_position
2788
+ - state.gripper_position
2789
+ action_concat_order:
2790
+ - action.joint_position
2791
+ - action.gripper_position
2792
+ - _target_: groot.vla.model.dreamzero.transform.dreamzero_cotrain.DreamTransform
2793
+ default_instruction: Perform the default behavior.
2794
+ language_dropout_prob: 0.0
2795
+ always_use_default_instruction: false
2796
+ max_state_dim: 64
2797
+ max_action_dim: 36
2798
+ max_length: 512
2799
+ state_horizon: 1
2800
+ action_horizon: 24
2801
+ embodiment_tag_mapping:
2802
+ real_gr1_arms_only: 0
2803
+ real_gr1_arms_only_annotated: 1
2804
+ real_gr1_arms_waist: 2
2805
+ real_gr1_arms_waist_annotated: 3
2806
+ dexmg_gr1_arms_only_inspire: 4
2807
+ dexmg_gr1_arms_only_fourier: 5
2808
+ dexmg_gr1_arms_waist_fourier: 6
2809
+ robocasa_single_arm: 7
2810
+ onex_eve_gripper: 8
2811
+ robocasa_gr1_arms_only_inspire_hands: 9
2812
+ robocasa_gr1_arms_only_fourier_hands: 10
2813
+ robocasa_gr1_fixed_lower_body_inspire_hands: 11
2814
+ robocasa_gr1_fixed_lower_body_fourier_hands: 12
2815
+ robocasa_panda_omron: 13
2816
+ robocasa_bimanual_panda_parallel_gripper: 15
2817
+ robocasa_bimanual_panda_inspire_hand: 16
2818
+ oxe_droid: 17
2819
+ oxe_fractal: 18
2820
+ oxe_language_table: 19
2821
+ oxe_bridge: 20
2822
+ real_panda_single_arm: 21
2823
+ hot3d_hands_only: 23
2824
+ gr1_unified: 24
2825
+ robocasa_gr1_arms_waist_fourier_hands: 25
2826
+ agibot: 26
2827
+ lapa: 27
2828
+ oxe_mutex: 28
2829
+ oxe_roboset: 29
2830
+ oxe_plex: 30
2831
+ dream: 31
2832
+ yam: 32
2833
+ xdof: 22
2834
+ gr1_unified_segmentation: 14
2835
+ language_table_sim: 7
2836
+ gr1_isaac: 0
2837
+ sim_behavior_r1_pro: 31
2838
+ mecka_hands: 27
2839
+ real_r1_pro_sharpa: 28
2840
+ real_teleop_g1: 33
2841
+ tokenizer_path: /hfm/boqian/liboqian_data/checkpoints/umt5-xxl
2842
+ agibot:
2843
+ _target_: groot.vla.data.transform.ComposedModalityTransform
2844
+ transforms:
2845
+ - _target_: groot.vla.data.transform.VideoToTensor
2846
+ apply_to:
2847
+ - video.top_head
2848
+ - video.hand_left
2849
+ - video.hand_right
2850
+ - _target_: groot.vla.data.transform.VideoCrop
2851
+ apply_to:
2852
+ - video.top_head
2853
+ - video.hand_left
2854
+ - video.hand_right
2855
+ scale: 0.95
2856
+ mode: random
2857
+ - _target_: groot.vla.data.transform.VideoResize
2858
+ apply_to:
2859
+ - video.top_head
2860
+ - video.hand_left
2861
+ - video.hand_right
2862
+ height: 176
2863
+ width: 320
2864
+ interpolation: linear
2865
+ - _target_: groot.vla.data.transform.VideoColorJitter
2866
+ apply_to:
2867
+ - video.top_head
2868
+ - video.hand_left
2869
+ - video.hand_right
2870
+ brightness: 0.3
2871
+ contrast: 0.4
2872
+ saturation: 0.5
2873
+ hue: 0.08
2874
+ - _target_: groot.vla.data.transform.VideoToNumpy
2875
+ apply_to:
2876
+ - video.top_head
2877
+ - video.hand_left
2878
+ - video.hand_right
2879
+ - _target_: groot.vla.data.transform.StateActionToTensor
2880
+ apply_to:
2881
+ - state.left_arm_joint_position
2882
+ - state.right_arm_joint_position
2883
+ - state.left_effector_position
2884
+ - state.right_effector_position
2885
+ - state.head_position
2886
+ - state.waist_position
2887
+ - _target_: groot.vla.data.transform.StateActionTransform
2888
+ apply_to:
2889
+ - state.left_arm_joint_position
2890
+ - state.right_arm_joint_position
2891
+ - state.left_effector_position
2892
+ - state.right_effector_position
2893
+ - state.head_position
2894
+ - state.waist_position
2895
+ normalization_modes:
2896
+ state.left_arm_joint_position: q99
2897
+ state.right_arm_joint_position: q99
2898
+ state.left_effector_position: q99
2899
+ state.right_effector_position: q99
2900
+ state.head_position: q99
2901
+ state.waist_position: q99
2902
+ - _target_: groot.vla.data.transform.StateActionToTensor
2903
+ apply_to:
2904
+ - action.left_arm_joint_position
2905
+ - action.right_arm_joint_position
2906
+ - action.left_effector_position
2907
+ - action.right_effector_position
2908
+ - action.head_position
2909
+ - action.waist_position
2910
+ - action.robot_velocity
2911
+ - _target_: groot.vla.data.transform.StateActionTransform
2912
+ apply_to:
2913
+ - action.left_arm_joint_position
2914
+ - action.right_arm_joint_position
2915
+ - action.left_effector_position
2916
+ - action.right_effector_position
2917
+ - action.head_position
2918
+ - action.waist_position
2919
+ - action.robot_velocity
2920
+ normalization_modes:
2921
+ action.left_arm_joint_position: q99
2922
+ action.right_arm_joint_position: q99
2923
+ action.left_effector_position: q99
2924
+ action.right_effector_position: q99
2925
+ action.head_position: q99
2926
+ action.waist_position: q99
2927
+ action.robot_velocity: q99
2928
+ - _target_: groot.vla.data.transform.ConcatTransform
2929
+ video_concat_order:
2930
+ - video.top_head
2931
+ - video.hand_left
2932
+ - video.hand_right
2933
+ state_concat_order:
2934
+ - state.left_arm_joint_position
2935
+ - state.right_arm_joint_position
2936
+ - state.left_effector_position
2937
+ - state.right_effector_position
2938
+ - state.head_position
2939
+ - state.waist_position
2940
+ action_concat_order:
2941
+ - action.left_arm_joint_position
2942
+ - action.right_arm_joint_position
2943
+ - action.left_effector_position
2944
+ - action.right_effector_position
2945
+ - action.head_position
2946
+ - action.waist_position
2947
+ - action.robot_velocity
2948
+ - _target_: groot.vla.model.dreamzero.transform.dreamzero_cotrain.DreamTransform
2949
+ default_instruction: Perform the default behavior.
2950
+ language_dropout_prob: 0.0
2951
+ always_use_default_instruction: false
2952
+ max_state_dim: 64
2953
+ max_action_dim: 36
2954
+ max_length: 512
2955
+ state_horizon: 1
2956
+ action_horizon: 24
2957
+ embodiment_tag_mapping:
2958
+ real_gr1_arms_only: 0
2959
+ real_gr1_arms_only_annotated: 1
2960
+ real_gr1_arms_waist: 2
2961
+ real_gr1_arms_waist_annotated: 3
2962
+ dexmg_gr1_arms_only_inspire: 4
2963
+ dexmg_gr1_arms_only_fourier: 5
2964
+ dexmg_gr1_arms_waist_fourier: 6
2965
+ robocasa_single_arm: 7
2966
+ onex_eve_gripper: 8
2967
+ robocasa_gr1_arms_only_inspire_hands: 9
2968
+ robocasa_gr1_arms_only_fourier_hands: 10
2969
+ robocasa_gr1_fixed_lower_body_inspire_hands: 11
2970
+ robocasa_gr1_fixed_lower_body_fourier_hands: 12
2971
+ robocasa_panda_omron: 13
2972
+ robocasa_bimanual_panda_parallel_gripper: 15
2973
+ robocasa_bimanual_panda_inspire_hand: 16
2974
+ oxe_droid: 17
2975
+ oxe_fractal: 18
2976
+ oxe_language_table: 19
2977
+ oxe_bridge: 20
2978
+ real_panda_single_arm: 21
2979
+ hot3d_hands_only: 23
2980
+ gr1_unified: 24
2981
+ robocasa_gr1_arms_waist_fourier_hands: 25
2982
+ agibot: 26
2983
+ lapa: 27
2984
+ oxe_mutex: 28
2985
+ oxe_roboset: 29
2986
+ oxe_plex: 30
2987
+ dream: 31
2988
+ yam: 32
2989
+ xdof: 22
2990
+ gr1_unified_segmentation: 14
2991
+ language_table_sim: 7
2992
+ gr1_isaac: 0
2993
+ sim_behavior_r1_pro: 31
2994
+ mecka_hands: 27
2995
+ real_r1_pro_sharpa: 28
2996
+ real_teleop_g1: 33
2997
+ tokenizer_path: /hfm/boqian/liboqian_data/checkpoints/umt5-xxl
2998
+ yam:
2999
+ _target_: groot.vla.data.transform.ComposedModalityTransform
3000
+ transforms:
3001
+ - _target_: groot.vla.data.transform.VideoToTensor
3002
+ apply_to:
3003
+ - video.top_camera-images-rgb
3004
+ - video.left_camera-images-rgb
3005
+ - video.right_camera-images-rgb
3006
+ - _target_: groot.vla.data.transform.VideoCrop
3007
+ apply_to:
3008
+ - video.top_camera-images-rgb
3009
+ - video.left_camera-images-rgb
3010
+ - video.right_camera-images-rgb
3011
+ scale: 0.95
3012
+ mode: random
3013
+ - _target_: groot.vla.data.transform.VideoResize
3014
+ apply_to:
3015
+ - video.top_camera-images-rgb
3016
+ - video.left_camera-images-rgb
3017
+ - video.right_camera-images-rgb
3018
+ height: 176
3019
+ width: 320
3020
+ interpolation: linear
3021
+ - _target_: groot.vla.data.transform.VideoColorJitter
3022
+ apply_to:
3023
+ - video.top_camera-images-rgb
3024
+ - video.left_camera-images-rgb
3025
+ - video.right_camera-images-rgb
3026
+ brightness: 0.3
3027
+ contrast: 0.4
3028
+ saturation: 0.5
3029
+ hue: 0.08
3030
+ - _target_: groot.vla.data.transform.VideoToNumpy
3031
+ apply_to:
3032
+ - video.top_camera-images-rgb
3033
+ - video.left_camera-images-rgb
3034
+ - video.right_camera-images-rgb
3035
+ - _target_: groot.vla.data.transform.StateActionToTensor
3036
+ apply_to:
3037
+ - state.left_joint_pos
3038
+ - state.left_gripper_pos
3039
+ - state.right_joint_pos
3040
+ - state.right_gripper_pos
3041
+ - _target_: groot.vla.data.transform.StateActionTransform
3042
+ apply_to:
3043
+ - state.left_joint_pos
3044
+ - state.left_gripper_pos
3045
+ - state.right_joint_pos
3046
+ - state.right_gripper_pos
3047
+ normalization_modes:
3048
+ state.left_joint_pos: q99
3049
+ state.left_gripper_pos: q99
3050
+ state.right_joint_pos: q99
3051
+ state.right_gripper_pos: q99
3052
+ - _target_: groot.vla.data.transform.StateActionToTensor
3053
+ apply_to:
3054
+ - action.left_joint_pos
3055
+ - action.left_gripper_pos
3056
+ - action.right_joint_pos
3057
+ - action.right_gripper_pos
3058
+ - _target_: groot.vla.data.transform.StateActionTransform
3059
+ apply_to:
3060
+ - action.left_joint_pos
3061
+ - action.left_gripper_pos
3062
+ - action.right_joint_pos
3063
+ - action.right_gripper_pos
3064
+ normalization_modes:
3065
+ action.left_joint_pos: q99
3066
+ action.left_gripper_pos: q99
3067
+ action.right_joint_pos: q99
3068
+ action.right_gripper_pos: q99
3069
+ - _target_: groot.vla.data.transform.ConcatTransform
3070
+ video_concat_order:
3071
+ - video.top_camera-images-rgb
3072
+ - video.left_camera-images-rgb
3073
+ - video.right_camera-images-rgb
3074
+ state_concat_order:
3075
+ - state.left_joint_pos
3076
+ - state.left_gripper_pos
3077
+ - state.right_joint_pos
3078
+ - state.right_gripper_pos
3079
+ action_concat_order:
3080
+ - action.left_joint_pos
3081
+ - action.left_gripper_pos
3082
+ - action.right_joint_pos
3083
+ - action.right_gripper_pos
3084
+ - _target_: groot.vla.model.dreamzero.transform.dreamzero_cotrain.DreamTransform
3085
+ default_instruction: Perform the default behavior.
3086
+ language_dropout_prob: 0.0
3087
+ always_use_default_instruction: false
3088
+ max_state_dim: 64
3089
+ max_action_dim: 36
3090
+ max_length: 512
3091
+ state_horizon: 1
3092
+ action_horizon: 24
3093
+ embodiment_tag_mapping:
3094
+ real_gr1_arms_only: 0
3095
+ real_gr1_arms_only_annotated: 1
3096
+ real_gr1_arms_waist: 2
3097
+ real_gr1_arms_waist_annotated: 3
3098
+ dexmg_gr1_arms_only_inspire: 4
3099
+ dexmg_gr1_arms_only_fourier: 5
3100
+ dexmg_gr1_arms_waist_fourier: 6
3101
+ robocasa_single_arm: 7
3102
+ onex_eve_gripper: 8
3103
+ robocasa_gr1_arms_only_inspire_hands: 9
3104
+ robocasa_gr1_arms_only_fourier_hands: 10
3105
+ robocasa_gr1_fixed_lower_body_inspire_hands: 11
3106
+ robocasa_gr1_fixed_lower_body_fourier_hands: 12
3107
+ robocasa_panda_omron: 13
3108
+ robocasa_bimanual_panda_parallel_gripper: 15
3109
+ robocasa_bimanual_panda_inspire_hand: 16
3110
+ oxe_droid: 17
3111
+ oxe_fractal: 18
3112
+ oxe_language_table: 19
3113
+ oxe_bridge: 20
3114
+ real_panda_single_arm: 21
3115
+ hot3d_hands_only: 23
3116
+ gr1_unified: 24
3117
+ robocasa_gr1_arms_waist_fourier_hands: 25
3118
+ agibot: 26
3119
+ lapa: 27
3120
+ oxe_mutex: 28
3121
+ oxe_roboset: 29
3122
+ oxe_plex: 30
3123
+ dream: 31
3124
+ yam: 32
3125
+ xdof: 22
3126
+ gr1_unified_segmentation: 14
3127
+ language_table_sim: 7
3128
+ gr1_isaac: 0
3129
+ sim_behavior_r1_pro: 31
3130
+ mecka_hands: 27
3131
+ real_r1_pro_sharpa: 28
3132
+ real_teleop_g1: 33
3133
+ tokenizer_path: /hfm/boqian/liboqian_data/checkpoints/umt5-xxl
3134
+ real_teleop_g1:
3135
+ _target_: groot.vla.data.transform.ComposedModalityTransform
3136
+ transforms:
3137
+ - _target_: groot.vla.data.transform.VideoToTensor
3138
+ apply_to:
3139
+ - video.egocentric
3140
+ - _target_: groot.vla.data.transform.VideoCrop
3141
+ apply_to:
3142
+ - video.egocentric
3143
+ scale: 0.95
3144
+ mode: random
3145
+ - _target_: groot.vla.data.transform.VideoResize
3146
+ apply_to:
3147
+ - video.egocentric
3148
+ height: 176
3149
+ width: 320
3150
+ interpolation: linear
3151
+ - _target_: groot.vla.data.transform.VideoColorJitter
3152
+ apply_to:
3153
+ - video.egocentric
3154
+ brightness: 0.3
3155
+ contrast: 0.4
3156
+ saturation: 0.5
3157
+ hue: 0.08
3158
+ - _target_: groot.vla.data.transform.VideoToNumpy
3159
+ apply_to:
3160
+ - video.egocentric
3161
+ - _target_: groot.vla.data.transform.StateActionToTensor
3162
+ apply_to:
3163
+ - state.left_hand
3164
+ - state.right_hand
3165
+ - state.left_arm
3166
+ - state.right_arm
3167
+ - state.rpy
3168
+ - state.height
3169
+ - _target_: groot.vla.data.transform.StateActionTransform
3170
+ apply_to:
3171
+ - state.left_hand
3172
+ - state.right_hand
3173
+ - state.left_arm
3174
+ - state.right_arm
3175
+ - state.rpy
3176
+ - state.height
3177
+ normalization_modes:
3178
+ state.left_hand: q99
3179
+ state.right_hand: q99
3180
+ state.left_arm: q99
3181
+ state.right_arm: q99
3182
+ state.rpy: q99
3183
+ state.height: q99
3184
+ - _target_: groot.vla.data.transform.StateActionToTensor
3185
+ apply_to:
3186
+ - action.left_hand
3187
+ - action.right_hand
3188
+ - action.left_arm
3189
+ - action.right_arm
3190
+ - action.rpy
3191
+ - action.height
3192
+ - action.torso_vx
3193
+ - action.torso_vy
3194
+ - action.torso_vyaw
3195
+ - action.torso_dyaw
3196
+ - _target_: groot.vla.data.transform.StateActionTransform
3197
+ apply_to:
3198
+ - action.left_hand
3199
+ - action.right_hand
3200
+ - action.left_arm
3201
+ - action.right_arm
3202
+ - action.rpy
3203
+ - action.height
3204
+ - action.torso_vx
3205
+ - action.torso_vy
3206
+ - action.torso_vyaw
3207
+ - action.torso_dyaw
3208
+ normalization_modes:
3209
+ action.left_hand: q99
3210
+ action.right_hand: q99
3211
+ action.left_arm: q99
3212
+ action.right_arm: q99
3213
+ action.rpy: q99
3214
+ action.height: q99
3215
+ action.torso_vx: q99
3216
+ action.torso_vy: q99
3217
+ action.torso_vyaw: q99
3218
+ action.torso_dyaw: q99
3219
+ - _target_: groot.vla.data.transform.ConcatTransform
3220
+ video_concat_order:
3221
+ - video.egocentric
3222
+ state_concat_order:
3223
+ - state.left_hand
3224
+ - state.right_hand
3225
+ - state.left_arm
3226
+ - state.right_arm
3227
+ - state.rpy
3228
+ - state.height
3229
+ action_concat_order:
3230
+ - action.left_hand
3231
+ - action.right_hand
3232
+ - action.left_arm
3233
+ - action.right_arm
3234
+ - action.rpy
3235
+ - action.height
3236
+ - action.torso_vx
3237
+ - action.torso_vy
3238
+ - action.torso_vyaw
3239
+ - action.torso_dyaw
3240
+ - _target_: groot.vla.model.dreamzero.transform.dreamzero_cotrain.DreamTransform
3241
+ default_instruction: Perform the default behavior.
3242
+ language_dropout_prob: 0.0
3243
+ always_use_default_instruction: false
3244
+ max_state_dim: 64
3245
+ max_action_dim: 36
3246
+ max_length: 512
3247
+ state_horizon: 1
3248
+ action_horizon: 24
3249
+ embodiment_tag_mapping:
3250
+ real_gr1_arms_only: 0
3251
+ real_gr1_arms_only_annotated: 1
3252
+ real_gr1_arms_waist: 2
3253
+ real_gr1_arms_waist_annotated: 3
3254
+ dexmg_gr1_arms_only_inspire: 4
3255
+ dexmg_gr1_arms_only_fourier: 5
3256
+ dexmg_gr1_arms_waist_fourier: 6
3257
+ robocasa_single_arm: 7
3258
+ onex_eve_gripper: 8
3259
+ robocasa_gr1_arms_only_inspire_hands: 9
3260
+ robocasa_gr1_arms_only_fourier_hands: 10
3261
+ robocasa_gr1_fixed_lower_body_inspire_hands: 11
3262
+ robocasa_gr1_fixed_lower_body_fourier_hands: 12
3263
+ robocasa_panda_omron: 13
3264
+ robocasa_bimanual_panda_parallel_gripper: 15
3265
+ robocasa_bimanual_panda_inspire_hand: 16
3266
+ oxe_droid: 17
3267
+ oxe_fractal: 18
3268
+ oxe_language_table: 19
3269
+ oxe_bridge: 20
3270
+ real_panda_single_arm: 21
3271
+ hot3d_hands_only: 23
3272
+ gr1_unified: 24
3273
+ robocasa_gr1_arms_waist_fourier_hands: 25
3274
+ agibot: 26
3275
+ lapa: 27
3276
+ oxe_mutex: 28
3277
+ oxe_roboset: 29
3278
+ oxe_plex: 30
3279
+ dream: 31
3280
+ yam: 32
3281
+ xdof: 22
3282
+ gr1_unified_segmentation: 14
3283
+ language_table_sim: 7
3284
+ gr1_isaac: 0
3285
+ sim_behavior_r1_pro: 31
3286
+ mecka_hands: 27
3287
+ real_r1_pro_sharpa: 28
3288
+ real_teleop_g1: 33
3289
+ tokenizer_path: /hfm/boqian/liboqian_data/checkpoints/umt5-xxl
3290
+ metadata_versions:
3291
+ oxe_droid: '0221'
3292
+ agibot: '0221'
3293
+ yam: '0221'
3294
+ real_teleop_g1: '0221'
3295
+ fps:
3296
+ yam: 30
3297
+ real_teleop_g1: 30
3298
+ relative_action: false
3299
+ relative_action_per_horizon: false
3300
+ relative_action_keys: []
3301
+ max_chunk_size: 4
3302
+ dataset_shard_sampling_rate: 0.1
3303
+ mixture_dataset_cls: groot.vla.data.dataset.lerobot_sharded.ShardedLeRobotMixtureDataset.from_mixture_spec
3304
+ single_dataset_cls: groot.vla.data.dataset.lerobot_sharded.ShardedLeRobotSubLangSingleActionChunkDatasetDROID
3305
+ real_teleop_g1_data_root: /hfm/boqian/liboqian_data/data/real_data/gear/g1/Pick_bottle_and_turn_and_pour_into_cup_absolute
3306
+ total_training_steps: 1048576000000
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/experiment_cfg/metadata.json ADDED
@@ -0,0 +1,787 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "real_teleop_g1": {
3
+ "statistics": {
4
+ "state": {
5
+ "left_hand": {
6
+ "max": [
7
+ -0.15012885630130768,
8
+ 0.9005824327468872,
9
+ 1.240552306175232,
10
+ -0.001318894443102181,
11
+ -0.029013408347964287,
12
+ -0.0026860591024160385,
13
+ -0.032535798847675323
14
+ ],
15
+ "min": [
16
+ -1.0365108251571655,
17
+ 0.3635019063949585,
18
+ 0.6077943444252014,
19
+ -0.8983967900276184,
20
+ -1.4610702991485596,
21
+ -0.9210805296897888,
22
+ -1.2054297924041748
23
+ ],
24
+ "mean": [
25
+ -0.7106521057783574,
26
+ 0.7397562400548329,
27
+ 0.7932539759105839,
28
+ -0.18307749281402536,
29
+ -0.3996594926481087,
30
+ -0.22440346846797105,
31
+ -0.3648747590553514
32
+ ],
33
+ "std": [
34
+ 0.1800224325521231,
35
+ 0.08892492705684825,
36
+ 0.17347929747061464,
37
+ 0.1250548816433564,
38
+ 0.14322181312600368,
39
+ 0.12746796614686143,
40
+ 0.1004095686025728
41
+ ],
42
+ "q01": [
43
+ -1.0010048174858093,
44
+ 0.5051653385162354,
45
+ 0.6091835498809814,
46
+ -0.5192373794317245,
47
+ -0.7626793467998505,
48
+ -0.5990579128265381,
49
+ -0.6296854817867279
50
+ ],
51
+ "q99": [
52
+ -0.35215023159980774,
53
+ 0.8828813862800599,
54
+ 1.1794202327728271,
55
+ -0.012645758632570512,
56
+ -0.08962718859314919,
57
+ -0.01236837636679411,
58
+ -0.14306417107582092
59
+ ]
60
+ },
61
+ "right_hand": {
62
+ "max": [
63
+ 0.013422729447484016,
64
+ -0.0029037927743047476,
65
+ -0.4307878315448761,
66
+ 1.1691919565200806,
67
+ 1.5508151054382324,
68
+ 1.008196473121643,
69
+ 1.4388744831085205
70
+ ],
71
+ "min": [
72
+ -0.9901090860366821,
73
+ -0.8877645134925842,
74
+ -1.2071785926818848,
75
+ -0.03135598078370094,
76
+ 0.03392980992794037,
77
+ 0.023357456550002098,
78
+ 0.019666638225317
79
+ ],
80
+ "mean": [
81
+ -0.4129581264587618,
82
+ -0.4582520757612821,
83
+ -0.664123836057758,
84
+ 0.4952124991587111,
85
+ 0.9231042625678995,
86
+ 0.5022273107153479,
87
+ 0.669409344487608
88
+ ],
89
+ "std": [
90
+ 0.19006874361243062,
91
+ 0.11976727609652792,
92
+ 0.10119243113959121,
93
+ 0.2266161525462106,
94
+ 0.4655465641238765,
95
+ 0.21432269096327275,
96
+ 0.31467715432459836
97
+ ],
98
+ "q01": [
99
+ -0.9081442606449127,
100
+ -0.8110105991363525,
101
+ -0.9831316405534745,
102
+ 0.023696813061833383,
103
+ 0.037188550531864165,
104
+ 0.024981562048196793,
105
+ 0.02064503453671932
106
+ ],
107
+ "q99": [
108
+ -0.10628944709897117,
109
+ -0.18123969539999962,
110
+ -0.5385138392448425,
111
+ 0.9067972785234452,
112
+ 1.4886025190353394,
113
+ 0.871771514415741,
114
+ 1.3075256443023682
115
+ ]
116
+ },
117
+ "left_arm": {
118
+ "max": [
119
+ 0.6784030199050903,
120
+ 0.5408600568771362,
121
+ 0.261579692363739,
122
+ 1.2853081226348877,
123
+ 0.6677011251449585,
124
+ 0.01986505277454853,
125
+ 0.5779042840003967
126
+ ],
127
+ "min": [
128
+ -0.6697743535041809,
129
+ -0.04027898237109184,
130
+ -0.6574546098709106,
131
+ -0.02732403390109539,
132
+ -0.18146513402462006,
133
+ -1.2264176607131958,
134
+ -0.8434670567512512
135
+ ],
136
+ "mean": [
137
+ 0.06576749867032948,
138
+ 0.25579782742676094,
139
+ -0.06004482964572005,
140
+ 0.808768983211113,
141
+ 0.10850686643277818,
142
+ -0.7254681426808276,
143
+ -0.2878231007376268
144
+ ],
145
+ "std": [
146
+ 0.16998997778227107,
147
+ 0.09886796724754655,
148
+ 0.1532526690961023,
149
+ 0.3788245539398352,
150
+ 0.1304561854786525,
151
+ 0.3305712410884026,
152
+ 0.20458272924414123
153
+ ],
154
+ "q01": [
155
+ -0.472993403673172,
156
+ 0.01923264417797327,
157
+ -0.4175487527251244,
158
+ 0.11523827515542508,
159
+ -0.13969017088413238,
160
+ -1.190878005027771,
161
+ -0.7066819667816162
162
+ ],
163
+ "q99": [
164
+ 0.4539884480834007,
165
+ 0.5193723440170288,
166
+ 0.2179892195761203,
167
+ 1.2710349559783936,
168
+ 0.6194926702976227,
169
+ -0.09952061586081994,
170
+ 0.17473116487264606
171
+ ]
172
+ },
173
+ "right_arm": {
174
+ "max": [
175
+ 0.9002070426940918,
176
+ 0.11829628795385361,
177
+ 0.8515990376472473,
178
+ 1.2844573259353638,
179
+ 0.8022719621658325,
180
+ 0.6399767994880676,
181
+ 1.2358516454696655
182
+ ],
183
+ "min": [
184
+ -1.2026649713516235,
185
+ -0.6969066858291626,
186
+ -0.815610408782959,
187
+ -0.9874761700630188,
188
+ -1.6946533918380737,
189
+ -1.4213329553604126,
190
+ -1.342279314994812
191
+ ],
192
+ "mean": [
193
+ -0.0662775668920557,
194
+ -0.1345353524434497,
195
+ 0.0634982026036378,
196
+ -0.10989526924031702,
197
+ -0.05263142642975812,
198
+ -0.12096835662218453,
199
+ 0.120861711803444
200
+ ],
201
+ "std": [
202
+ 0.4008494462356859,
203
+ 0.1700786431413106,
204
+ 0.21295159275658426,
205
+ 0.43649950708158436,
206
+ 0.4497950790039162,
207
+ 0.36514804111543175,
208
+ 0.1967684071659634
209
+ ],
210
+ "q01": [
211
+ -0.950262793302536,
212
+ -0.5722128254175186,
213
+ -0.4087426000833511,
214
+ -0.8998615843057632,
215
+ -1.200806084871292,
216
+ -1.1520812797546387,
217
+ -0.3325912955403328
218
+ ],
219
+ "q99": [
220
+ 0.7100517773628232,
221
+ 0.10524750150740145,
222
+ 0.6194546324014664,
223
+ 0.9450081789493561,
224
+ 0.559214276075363,
225
+ 0.5243198603391647,
226
+ 0.6973656332492828
227
+ ]
228
+ },
229
+ "rpy": {
230
+ "max": [
231
+ 0.1827869415283203,
232
+ 0.12582087516784668,
233
+ 0.7600905299186707
234
+ ],
235
+ "min": [
236
+ -0.25714462995529175,
237
+ -0.27553248405456543,
238
+ -0.3924732208251953
239
+ ],
240
+ "mean": [
241
+ -0.07450471914543803,
242
+ -0.016358907944713997,
243
+ 0.08461220301037418
244
+ ],
245
+ "std": [
246
+ 0.052686434459038575,
247
+ 0.05521563296798423,
248
+ 0.1871730371455983
249
+ ],
250
+ "q01": [
251
+ -0.21652273923158646,
252
+ -0.15164137467741967,
253
+ -0.3411825066804886
254
+ ],
255
+ "q99": [
256
+ 0.023736179061233972,
257
+ 0.0924126328527927,
258
+ 0.6843682318925858
259
+ ]
260
+ },
261
+ "height": {
262
+ "max": [
263
+ 0.75
264
+ ],
265
+ "min": [
266
+ 0.6516909599304199
267
+ ],
268
+ "mean": [
269
+ 0.717019414027915
270
+ ],
271
+ "std": [
272
+ 0.010611909948945071
273
+ ],
274
+ "q01": [
275
+ 0.681534378528595
276
+ ],
277
+ "q99": [
278
+ 0.7377596861124038
279
+ ]
280
+ }
281
+ },
282
+ "action": {
283
+ "left_hand": {
284
+ "max": [
285
+ -0.1213064193725586,
286
+ 0.9210000038146973,
287
+ 1.2908918857574463,
288
+ 0.0010000000474974513,
289
+ 0.0010000000474974513,
290
+ 0.0010000000474974513,
291
+ 0.0010000000474974513
292
+ ],
293
+ "min": [
294
+ -1.0481975078582764,
295
+ 0.33770662546157837,
296
+ 0.5833444595336914,
297
+ -0.9677509069442749,
298
+ -1.5101021528244019,
299
+ -0.9662452340126038,
300
+ -1.252585768699646
301
+ ],
302
+ "mean": [
303
+ -0.7283108163247403,
304
+ 0.7484512520233001,
305
+ 0.7898890272276098,
306
+ -0.1841689376362202,
307
+ -0.40007694289939594,
308
+ -0.22100483851739403,
309
+ -0.3606009050772844
310
+ ],
311
+ "std": [
312
+ 0.18855574743509407,
313
+ 0.10765775142317323,
314
+ 0.20562452513338061,
315
+ 0.13352257457524555,
316
+ 0.14831147850778914,
317
+ 0.13177602593521492,
318
+ 0.10852811415283538
319
+ ],
320
+ "q01": [
321
+ -1.0481975078582764,
322
+ 0.47301085710525514,
323
+ 0.5974299025535583,
324
+ -0.5396523159742356,
325
+ -0.7698984146118164,
326
+ -0.569393122792244,
327
+ -0.6210935115814209
328
+ ],
329
+ "q99": [
330
+ -0.32839281708002094,
331
+ 0.9210000038146973,
332
+ 1.2100253427028655,
333
+ 0.0010000000474974513,
334
+ -0.07790654666721833,
335
+ 0.0010000000474974513,
336
+ -0.12742940008640297
337
+ ]
338
+ },
339
+ "right_hand": {
340
+ "max": [
341
+ 0.11635252088308334,
342
+ 0.11013627797365189,
343
+ -0.4304533302783966,
344
+ 1.4093444347381592,
345
+ 1.7463293075561523,
346
+ 1.2701730728149414,
347
+ 1.7463293075561523
348
+ ],
349
+ "min": [
350
+ -1.0481975078582764,
351
+ -0.9210000038146973,
352
+ -1.4877983331680298,
353
+ -0.0010000000474974513,
354
+ -0.0010000000474974513,
355
+ -0.0010000000474974513,
356
+ -0.0010000000474974513
357
+ ],
358
+ "mean": [
359
+ -0.3458960654680007,
360
+ -0.6374010453148075,
361
+ -0.7414860319631214,
362
+ 0.6248424045903762,
363
+ 1.0210826742269492,
364
+ 0.624888935664034,
365
+ 0.7309311322729288
366
+ ],
367
+ "std": [
368
+ 0.2477478868975401,
369
+ 0.12149676878810435,
370
+ 0.14591807616135066,
371
+ 0.31889999783341705,
372
+ 0.550414547421302,
373
+ 0.29655947330065524,
374
+ 0.37683330327414893
375
+ ],
376
+ "q01": [
377
+ -0.943056748509407,
378
+ -0.8472843170166016,
379
+ -1.144005218744278,
380
+ -0.0010000000474974513,
381
+ -0.0010000000474974513,
382
+ -0.0010000000474974513,
383
+ -0.0010000000474974513
384
+ ],
385
+ "q99": [
386
+ -0.019174456279724885,
387
+ -0.30395907163619995,
388
+ -0.5490964096784592,
389
+ 1.2382159757614122,
390
+ 1.7386069679260252,
391
+ 1.021970037221905,
392
+ 1.4857414960861206
393
+ ]
394
+ },
395
+ "left_arm": {
396
+ "max": [
397
+ 0.7063568830490112,
398
+ 0.5787747502326965,
399
+ 0.2658364176750183,
400
+ 1.36533784866333,
401
+ 0.7970759272575378,
402
+ 0.050355639308691025,
403
+ 0.6250148415565491
404
+ ],
405
+ "min": [
406
+ -0.6973711848258972,
407
+ -0.04661950096487999,
408
+ -0.7617664933204651,
409
+ -0.0605352558195591,
410
+ -0.21390148997306824,
411
+ -1.2438902854919434,
412
+ -0.8552998304367065
413
+ ],
414
+ "mean": [
415
+ 0.05742595636422137,
416
+ 0.2634212185929878,
417
+ -0.06515526749555424,
418
+ 0.8164667552012993,
419
+ 0.10927570881930437,
420
+ -0.7389155860415086,
421
+ -0.295913822816809
422
+ ],
423
+ "std": [
424
+ 0.16957499887908717,
425
+ 0.0950404677458485,
426
+ 0.15050156924776992,
427
+ 0.38010919103376684,
428
+ 0.13206946446875728,
429
+ 0.32367770103457033,
430
+ 0.20343907571654762
431
+ ],
432
+ "q01": [
433
+ -0.5008356976509094,
434
+ 0.03911703012883663,
435
+ -0.4255762630701065,
436
+ 0.11195705555379391,
437
+ -0.15061241388320923,
438
+ -1.1989330887794494,
439
+ -0.7123385232686996
440
+ ],
441
+ "q99": [
442
+ 0.4477514466643332,
443
+ 0.524748660326004,
444
+ 0.216304791867733,
445
+ 1.2753526496887206,
446
+ 0.6235151898860929,
447
+ -0.11624279774725446,
448
+ 0.17421353697776795
449
+ ]
450
+ },
451
+ "right_arm": {
452
+ "max": [
453
+ 1.0034496784210205,
454
+ 0.6385414600372314,
455
+ 0.8838359117507935,
456
+ 1.3011894226074219,
457
+ 0.9103013873100281,
458
+ 0.6697143316268921,
459
+ 1.300451397895813
460
+ ],
461
+ "min": [
462
+ -1.2567458152770996,
463
+ -0.7329762578010559,
464
+ -0.8994792103767395,
465
+ -1.029470682144165,
466
+ -1.7312105894088745,
467
+ -1.614429235458374,
468
+ -1.554446816444397
469
+ ],
470
+ "mean": [
471
+ -0.06658668015508451,
472
+ -0.1276185496595779,
473
+ 0.05797919617399393,
474
+ -0.11602684028492932,
475
+ -0.03613449704007667,
476
+ -0.1405436867128357,
477
+ 0.1209290054543264
478
+ ],
479
+ "std": [
480
+ 0.4080472941147137,
481
+ 0.19441447416204616,
482
+ 0.22406206897544698,
483
+ 0.46053246124217484,
484
+ 0.4648783663000821,
485
+ 0.3964305914547449,
486
+ 0.21165879464009837
487
+ ],
488
+ "q01": [
489
+ -0.9834537935256957,
490
+ -0.5832101339101792,
491
+ -0.46410051107406614,
492
+ -0.9261943989992142,
493
+ -1.2178711235523223,
494
+ -1.2455947709083557,
495
+ -0.374444442987442
496
+ ],
497
+ "q99": [
498
+ 0.7256432592868804,
499
+ 0.30993463188409803,
500
+ 0.6237642824649811,
501
+ 1.089427736997604,
502
+ 0.6462495481967916,
503
+ 0.5251477748155594,
504
+ 0.7147807055711739
505
+ ]
506
+ },
507
+ "rpy": {
508
+ "max": [
509
+ 0.1827869415283203,
510
+ 0.12582087516784668,
511
+ 0.7600905299186707
512
+ ],
513
+ "min": [
514
+ -0.25714462995529175,
515
+ -0.27553248405456543,
516
+ -0.3924732208251953
517
+ ],
518
+ "mean": [
519
+ -0.07457712200498402,
520
+ -0.016338004632023968,
521
+ 0.08467273931155996
522
+ ],
523
+ "std": [
524
+ 0.05265008883277812,
525
+ 0.05524160676308374,
526
+ 0.18724355327591102
527
+ ],
528
+ "q01": [
529
+ -0.21652273923158646,
530
+ -0.15164137467741967,
531
+ -0.34133789867162706
532
+ ],
533
+ "q99": [
534
+ 0.023736179061233972,
535
+ 0.09255022160708899,
536
+ 0.6843682318925858
537
+ ]
538
+ },
539
+ "height": {
540
+ "max": [
541
+ 0.7456391453742981
542
+ ],
543
+ "min": [
544
+ 0.6516909599304199
545
+ ],
546
+ "mean": [
547
+ 0.7169688045156508
548
+ ],
549
+ "std": [
550
+ 0.01058211087414312
551
+ ],
552
+ "q01": [
553
+ 0.6814439028501511
554
+ ],
555
+ "q99": [
556
+ 0.7374992072582245
557
+ ]
558
+ },
559
+ "torso_vx": {
560
+ "max": [
561
+ 0.3499999940395355
562
+ ],
563
+ "min": [
564
+ 0.0
565
+ ],
566
+ "mean": [
567
+ 0.017269024247052768
568
+ ],
569
+ "std": [
570
+ 0.07580197349073346
571
+ ],
572
+ "q01": [
573
+ 0.0
574
+ ],
575
+ "q99": [
576
+ 0.3499999940395355
577
+ ]
578
+ },
579
+ "torso_vy": {
580
+ "max": [
581
+ 0.5
582
+ ],
583
+ "min": [
584
+ 0.0
585
+ ],
586
+ "mean": [
587
+ 0.0002320065992988245
588
+ ],
589
+ "std": [
590
+ 0.01076798368252486
591
+ ],
592
+ "q01": [
593
+ 0.0
594
+ ],
595
+ "q99": [
596
+ 0.0
597
+ ]
598
+ },
599
+ "torso_vyaw": {
600
+ "max": [
601
+ 0.5
602
+ ],
603
+ "min": [
604
+ -0.5
605
+ ],
606
+ "mean": [
607
+ -0.03733115137363985
608
+ ],
609
+ "std": [
610
+ 0.1199053455233051
611
+ ],
612
+ "q01": [
613
+ -0.5
614
+ ],
615
+ "q99": [
616
+ 0.0
617
+ ]
618
+ },
619
+ "torso_dyaw": {
620
+ "max": [
621
+ 0.3072107136249542
622
+ ],
623
+ "min": [
624
+ -0.1875186562538147
625
+ ],
626
+ "mean": [
627
+ 0.016342195046893805
628
+ ],
629
+ "std": [
630
+ 0.048962870720188865
631
+ ],
632
+ "q01": [
633
+ -0.005849878406152129
634
+ ],
635
+ "q99": [
636
+ 0.22258896410465237
637
+ ]
638
+ }
639
+ }
640
+ },
641
+ "modalities": {
642
+ "video": {
643
+ "egocentric": {
644
+ "resolution": [
645
+ 640,
646
+ 480
647
+ ],
648
+ "channels": 3,
649
+ "fps": 30.0
650
+ }
651
+ },
652
+ "state": {
653
+ "left_hand": {
654
+ "absolute": true,
655
+ "rotation_type": null,
656
+ "shape": [
657
+ 7
658
+ ],
659
+ "continuous": true
660
+ },
661
+ "right_hand": {
662
+ "absolute": true,
663
+ "rotation_type": null,
664
+ "shape": [
665
+ 7
666
+ ],
667
+ "continuous": true
668
+ },
669
+ "left_arm": {
670
+ "absolute": true,
671
+ "rotation_type": null,
672
+ "shape": [
673
+ 7
674
+ ],
675
+ "continuous": true
676
+ },
677
+ "right_arm": {
678
+ "absolute": true,
679
+ "rotation_type": null,
680
+ "shape": [
681
+ 7
682
+ ],
683
+ "continuous": true
684
+ },
685
+ "rpy": {
686
+ "absolute": true,
687
+ "rotation_type": null,
688
+ "shape": [
689
+ 3
690
+ ],
691
+ "continuous": true
692
+ },
693
+ "height": {
694
+ "absolute": true,
695
+ "rotation_type": null,
696
+ "shape": [
697
+ 1
698
+ ],
699
+ "continuous": true
700
+ }
701
+ },
702
+ "action": {
703
+ "left_hand": {
704
+ "absolute": true,
705
+ "rotation_type": null,
706
+ "shape": [
707
+ 7
708
+ ],
709
+ "continuous": true
710
+ },
711
+ "right_hand": {
712
+ "absolute": true,
713
+ "rotation_type": null,
714
+ "shape": [
715
+ 7
716
+ ],
717
+ "continuous": true
718
+ },
719
+ "left_arm": {
720
+ "absolute": true,
721
+ "rotation_type": null,
722
+ "shape": [
723
+ 7
724
+ ],
725
+ "continuous": true
726
+ },
727
+ "right_arm": {
728
+ "absolute": true,
729
+ "rotation_type": null,
730
+ "shape": [
731
+ 7
732
+ ],
733
+ "continuous": true
734
+ },
735
+ "rpy": {
736
+ "absolute": true,
737
+ "rotation_type": null,
738
+ "shape": [
739
+ 3
740
+ ],
741
+ "continuous": true
742
+ },
743
+ "height": {
744
+ "absolute": true,
745
+ "rotation_type": null,
746
+ "shape": [
747
+ 1
748
+ ],
749
+ "continuous": true
750
+ },
751
+ "torso_vx": {
752
+ "absolute": false,
753
+ "rotation_type": null,
754
+ "shape": [
755
+ 1
756
+ ],
757
+ "continuous": true
758
+ },
759
+ "torso_vy": {
760
+ "absolute": false,
761
+ "rotation_type": null,
762
+ "shape": [
763
+ 1
764
+ ],
765
+ "continuous": true
766
+ },
767
+ "torso_vyaw": {
768
+ "absolute": false,
769
+ "rotation_type": null,
770
+ "shape": [
771
+ 1
772
+ ],
773
+ "continuous": true
774
+ },
775
+ "torso_dyaw": {
776
+ "absolute": false,
777
+ "rotation_type": null,
778
+ "shape": [
779
+ 1
780
+ ],
781
+ "continuous": true
782
+ }
783
+ }
784
+ },
785
+ "embodiment_tag": "real_teleop_g1"
786
+ }
787
+ }
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3caff8ec74a8b6a9567475ebe40d3861fe8892f2b8c400d9ca996004310f3761
3
+ size 24726494997
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06d2a1b9ce3e5d4fb81678db9f7b4bf1347e4257d560b6675f2796546b2499eb
3
+ size 24726494869
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f1c8ed0278612e4bb5024be81202eaa2bdc612bb0e5ef8a1cbf8f631aec6f89
3
+ size 24726494933
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:584d0b5d5dfa69327cd95583c1e442fba0c5509b07fe630ed9190c2a0d629983
3
+ size 24726495189
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/bf16_zero_pp_rank_4_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9f62e5478f4b8d0139fe536d4e2ea88c0a8fcb31aaaca1937273ba6add59b11
3
+ size 24726494869
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/bf16_zero_pp_rank_5_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cea7b073d45494887b902099a666ef0d51739f35d8d92def0f1864429cb7fec
3
+ size 24726494805
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/bf16_zero_pp_rank_6_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53a042f0b18e2a32e863d50f6d772b6761f496680288cae301a1edfe2c227d98
3
+ size 24726494805
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/bf16_zero_pp_rank_7_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f383a20036309f2b056fd11380d45b741b9e6637804de3739943ab4baaa9a95
3
+ size 24726495317
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/global_step20000/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:241bbf9b897f8e86fe844221bf2fd2fd3009ee2089b5467a9a859e379de0f754
3
+ size 58729254967
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/latest ADDED
@@ -0,0 +1 @@
 
 
1
+ global_step20000
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00001-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0e0ff763ed8bb6b855782e10f1577d676bc7abce2df0e46cea5d340a05963a5
3
+ size 4935816720
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00002-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3dd6a6ed2735b920badc8b58864ac567f921b5e1ab537f9c755e35e6c65ebd8c
3
+ size 4983107552
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00003-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90a95d54d28bbc3ddbc55719ae709e904548b740f25600c875adb639f6870888
3
+ size 4937616256
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00004-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ca52a1d81944c771ba484b89adef9c8df476c7fca3e4ffac7f2619d0b307746
3
+ size 4987670360
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00005-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5eb8b845be51f2f71959a4b75ffcae684bf87c1832f9affa2c3631be54040b8
3
+ size 4987714608
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00006-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c90d9c99202128fd727ea9f10746b4716184989e5a0edf80f7bc2d3ffbae071
3
+ size 4950963456
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00007-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36a44e22d355d0c600e60523182930d30b5c5139f33758c9fc8f66f153a706fa
3
+ size 4951004896
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00008-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41528a5bcf542bc3e7f76c49d9bd6a14b696313d7183e085e9d331ddd8ce00e1
3
+ size 4950963472
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00009-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5824aff7cb28287d558676dd698be4aa781bb676ac87f2fb12a2b51793a79044
3
+ size 4950984200
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model-00010-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3346e9c1cbcd384e3694dfff4f52b22e1f0b841885faefe941bce96c8ebe40b3
3
+ size 1212823208
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c438dafa01e947a44498e649040021017fe8ac3136f94a6557fd9a73aa6fc87
3
+ size 16389
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67d51182ad9d940b5250d0f4063d4f029f086b6e9cf752e8ce5fef7190bbe2a6
3
+ size 16389
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c2469460a6376abd134137c65e6c5c24654fc30f4aa68654584ad0118a8a799
3
+ size 16389
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66302fb556ed76741228268c5831ac3c0a4d4c66bbc86fdef4bcf6ff391db590
3
+ size 16389
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f366fae6f95875cd115f941788689420c7ea5dde16996514d0da7f7cf0614e0b
3
+ size 16389
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8723611dff729f148415a3396f5beb7a562b2fd9a73d99dd46a6329f42cd678c
3
+ size 16389
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3b337e3ce96bf47ec270f4a05861e7868b21555fb7e8e181e0d106b0a8674e8
3
+ size 16389
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f19001c88c670951f34333ba015a06134a3298870fc8c97ce8c5ec9b787261a1
3
+ size 16389
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28508b0a940eb7f712597070ad628ccf8634a0efd412740f51a02d7843f5d668
3
+ size 1465
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/wandb_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"project": "dreamzero", "run_id": ""}
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/checkpoint-20000/zero_to_fp32.py ADDED
@@ -0,0 +1,760 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/usr/bin/env python
2
+
3
+ # Copyright (c) Microsoft Corporation.
4
+ # SPDX-License-Identifier: Apache-2.0
5
+
6
+ # DeepSpeed Team
7
+
8
+ # This script extracts fp32 consolidated weights from a zero 1, 2 and 3 DeepSpeed checkpoints. It gets
9
+ # copied into the top level checkpoint dir, so the user can easily do the conversion at any point in
10
+ # the future. Once extracted, the weights don't require DeepSpeed and can be used in any
11
+ # application.
12
+ #
13
+ # example:
14
+ # python zero_to_fp32.py . output_dir/
15
+ # or
16
+ # python zero_to_fp32.py . output_dir/ --safe_serialization
17
+
18
+ import argparse
19
+ import torch
20
+ import glob
21
+ import math
22
+ import os
23
+ import re
24
+ import gc
25
+ import json
26
+ import numpy as np
27
+ from tqdm import tqdm
28
+ from collections import OrderedDict
29
+ from dataclasses import dataclass
30
+
31
+ # while this script doesn't use deepspeed to recover data, since the checkpoints are pickled with
32
+ # DeepSpeed data structures it has to be available in the current python environment.
33
+ from deepspeed.utils import logger
34
+ from deepspeed.checkpoint.constants import (DS_VERSION, OPTIMIZER_STATE_DICT, SINGLE_PARTITION_OF_FP32_GROUPS,
35
+ FP32_FLAT_GROUPS, ZERO_STAGE, PARTITION_COUNT, PARAM_SHAPES, BUFFER_NAMES,
36
+ FROZEN_PARAM_SHAPES, FROZEN_PARAM_FRAGMENTS)
37
+
38
+
39
+ @dataclass
40
+ class zero_model_state:
41
+ buffers: dict()
42
+ param_shapes: dict()
43
+ shared_params: list
44
+ ds_version: int
45
+ frozen_param_shapes: dict()
46
+ frozen_param_fragments: dict()
47
+
48
+
49
+ debug = 0
50
+
51
+ # load to cpu
52
+ device = torch.device('cpu')
53
+
54
+
55
+ def atoi(text):
56
+ return int(text) if text.isdigit() else text
57
+
58
+
59
+ def natural_keys(text):
60
+ '''
61
+ alist.sort(key=natural_keys) sorts in human order
62
+ http://nedbatchelder.com/blog/200712/human_sorting.html
63
+ (See Toothy's implementation in the comments)
64
+ '''
65
+ return [atoi(c) for c in re.split(r'(\d+)', text)]
66
+
67
+
68
+ def get_model_state_file(checkpoint_dir, zero_stage):
69
+ if not os.path.isdir(checkpoint_dir):
70
+ raise FileNotFoundError(f"Directory '{checkpoint_dir}' doesn't exist")
71
+
72
+ # there should be only one file
73
+ if zero_stage <= 2:
74
+ file = os.path.join(checkpoint_dir, "mp_rank_00_model_states.pt")
75
+ elif zero_stage == 3:
76
+ file = os.path.join(checkpoint_dir, "zero_pp_rank_0_mp_rank_00_model_states.pt")
77
+
78
+ if not os.path.exists(file):
79
+ raise FileNotFoundError(f"can't find model states file at '{file}'")
80
+
81
+ return file
82
+
83
+
84
+ def get_checkpoint_files(checkpoint_dir, glob_pattern):
85
+ # XXX: need to test that this simple glob rule works for multi-node setup too
86
+ ckpt_files = sorted(glob.glob(os.path.join(checkpoint_dir, glob_pattern)), key=natural_keys)
87
+
88
+ if len(ckpt_files) == 0:
89
+ raise FileNotFoundError(f"can't find {glob_pattern} files in directory '{checkpoint_dir}'")
90
+
91
+ return ckpt_files
92
+
93
+
94
+ def get_optim_files(checkpoint_dir):
95
+ return get_checkpoint_files(checkpoint_dir, "*_optim_states.pt")
96
+
97
+
98
+ def get_model_state_files(checkpoint_dir):
99
+ return get_checkpoint_files(checkpoint_dir, "*_model_states.pt")
100
+
101
+
102
+ def parse_model_states(files):
103
+ zero_model_states = []
104
+ for file in files:
105
+ state_dict = torch.load(file, map_location=device, weights_only=False)
106
+
107
+ if BUFFER_NAMES not in state_dict:
108
+ raise ValueError(f"{file} is not a model state checkpoint")
109
+ buffer_names = state_dict[BUFFER_NAMES]
110
+ if debug:
111
+ print("Found buffers:", buffer_names)
112
+
113
+ # recover just the buffers while restoring them to fp32 if they were saved in fp16
114
+ buffers = {k: v.float() for k, v in state_dict["module"].items() if k in buffer_names}
115
+ param_shapes = state_dict[PARAM_SHAPES]
116
+
117
+ # collect parameters that are included in param_shapes
118
+ param_names = []
119
+ for s in param_shapes:
120
+ for name in s.keys():
121
+ param_names.append(name)
122
+
123
+ # update with frozen parameters
124
+ frozen_param_shapes = state_dict.get(FROZEN_PARAM_SHAPES, None)
125
+ if frozen_param_shapes is not None:
126
+ if debug:
127
+ print(f"Found frozen_param_shapes: {frozen_param_shapes}")
128
+ param_names += list(frozen_param_shapes.keys())
129
+
130
+ # handle shared params
131
+ shared_params = [[k, v] for k, v in state_dict["shared_params"].items()]
132
+
133
+ ds_version = state_dict.get(DS_VERSION, None)
134
+
135
+ frozen_param_fragments = state_dict.get(FROZEN_PARAM_FRAGMENTS, None)
136
+
137
+ z_model_state = zero_model_state(buffers=buffers,
138
+ param_shapes=param_shapes,
139
+ shared_params=shared_params,
140
+ ds_version=ds_version,
141
+ frozen_param_shapes=frozen_param_shapes,
142
+ frozen_param_fragments=frozen_param_fragments)
143
+ zero_model_states.append(z_model_state)
144
+
145
+ return zero_model_states
146
+
147
+
148
+ def parse_optim_states(files, ds_checkpoint_dir):
149
+ total_files = len(files)
150
+ state_dicts = []
151
+ for f in tqdm(files, desc='Loading checkpoint shards'):
152
+ state_dict = torch.load(f, map_location=device, mmap=True, weights_only=False)
153
+ # immediately discard the potentially huge 2 optimizer states as we only care for fp32 master weights
154
+ # and also handle the case where it was already removed by another helper script
155
+ state_dict["optimizer_state_dict"].pop("optimizer_state_dict", None)
156
+ state_dicts.append(state_dict)
157
+
158
+ if ZERO_STAGE not in state_dicts[0][OPTIMIZER_STATE_DICT]:
159
+ raise ValueError(f"{files[0]} is not a zero checkpoint")
160
+ zero_stage = state_dicts[0][OPTIMIZER_STATE_DICT][ZERO_STAGE]
161
+ world_size = state_dicts[0][OPTIMIZER_STATE_DICT][PARTITION_COUNT]
162
+
163
+ # For ZeRO-2 each param group can have different partition_count as data parallelism for expert
164
+ # parameters can be different from data parallelism for non-expert parameters. So we can just
165
+ # use the max of the partition_count to get the dp world_size.
166
+
167
+ if type(world_size) is list:
168
+ world_size = max(world_size)
169
+
170
+ if world_size != total_files:
171
+ raise ValueError(
172
+ f"Expected {world_size} of '*_optim_states.pt' under '{ds_checkpoint_dir}' but found {total_files} files. "
173
+ "Possibly due to an overwrite of an old checkpoint, or a checkpoint didn't get saved by one or more processes."
174
+ )
175
+
176
+ # the groups are named differently in each stage
177
+ if zero_stage <= 2:
178
+ fp32_groups_key = SINGLE_PARTITION_OF_FP32_GROUPS
179
+ elif zero_stage == 3:
180
+ fp32_groups_key = FP32_FLAT_GROUPS
181
+ else:
182
+ raise ValueError(f"unknown zero stage {zero_stage}")
183
+
184
+ fp32_flat_groups = [state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key] for i in range(len(state_dicts))]
185
+ return zero_stage, world_size, fp32_flat_groups
186
+
187
+
188
+ def _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters):
189
+ """
190
+ Returns fp32 state_dict reconstructed from ds checkpoint
191
+
192
+ Args:
193
+ - ``ds_checkpoint_dir``: path to the deepspeed checkpoint folder (where the optimizer files are)
194
+
195
+ """
196
+ print(f"Processing zero checkpoint '{ds_checkpoint_dir}'")
197
+
198
+ optim_files = get_optim_files(ds_checkpoint_dir)
199
+ zero_stage, world_size, fp32_flat_groups = parse_optim_states(optim_files, ds_checkpoint_dir)
200
+ print(f"Detected checkpoint of type zero stage {zero_stage}, world_size: {world_size}")
201
+
202
+ model_files = get_model_state_files(ds_checkpoint_dir)
203
+
204
+ zero_model_states = parse_model_states(model_files)
205
+ print(f'Parsing checkpoint created by deepspeed=={zero_model_states[0].ds_version}')
206
+
207
+ if zero_stage <= 2:
208
+ return _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states,
209
+ exclude_frozen_parameters)
210
+ elif zero_stage == 3:
211
+ return _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states,
212
+ exclude_frozen_parameters)
213
+
214
+
215
+ def _zero2_merge_frozen_params(state_dict, zero_model_states):
216
+ if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0:
217
+ return
218
+
219
+ frozen_param_shapes = zero_model_states[0].frozen_param_shapes
220
+ frozen_param_fragments = zero_model_states[0].frozen_param_fragments
221
+
222
+ if debug:
223
+ num_elem = sum(s.numel() for s in frozen_param_shapes.values())
224
+ print(f'rank 0: {FROZEN_PARAM_SHAPES}.numel = {num_elem}')
225
+
226
+ wanted_params = len(frozen_param_shapes)
227
+ wanted_numel = sum(s.numel() for s in frozen_param_shapes.values())
228
+ avail_numel = sum([p.numel() for p in frozen_param_fragments.values()])
229
+ print(f'Frozen params: Have {avail_numel} numels to process.')
230
+ print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params')
231
+
232
+ total_params = 0
233
+ total_numel = 0
234
+ for name, shape in frozen_param_shapes.items():
235
+ total_params += 1
236
+ unpartitioned_numel = shape.numel()
237
+ total_numel += unpartitioned_numel
238
+
239
+ state_dict[name] = frozen_param_fragments[name]
240
+
241
+ if debug:
242
+ print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ")
243
+
244
+ print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements")
245
+
246
+
247
+ def _has_callable(obj, fn):
248
+ attr = getattr(obj, fn, None)
249
+ return callable(attr)
250
+
251
+
252
+ def _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states):
253
+ param_shapes = zero_model_states[0].param_shapes
254
+
255
+ # Reconstruction protocol:
256
+ #
257
+ # XXX: document this
258
+
259
+ if debug:
260
+ for i in range(world_size):
261
+ for j in range(len(fp32_flat_groups[0])):
262
+ print(f"{FP32_FLAT_GROUPS}[{i}][{j}].shape={fp32_flat_groups[i][j].shape}")
263
+
264
+ # XXX: memory usage doubles here (zero2)
265
+ num_param_groups = len(fp32_flat_groups[0])
266
+ merged_single_partition_of_fp32_groups = []
267
+ for i in range(num_param_groups):
268
+ merged_partitions = [sd[i] for sd in fp32_flat_groups]
269
+ full_single_fp32_vector = torch.cat(merged_partitions, 0)
270
+ merged_single_partition_of_fp32_groups.append(full_single_fp32_vector)
271
+ avail_numel = sum(
272
+ [full_single_fp32_vector.numel() for full_single_fp32_vector in merged_single_partition_of_fp32_groups])
273
+
274
+ if debug:
275
+ wanted_params = sum([len(shapes) for shapes in param_shapes])
276
+ wanted_numel = sum([sum(shape.numel() for shape in shapes.values()) for shapes in param_shapes])
277
+ # not asserting if there is a mismatch due to possible padding
278
+ print(f"Have {avail_numel} numels to process.")
279
+ print(f"Need {wanted_numel} numels in {wanted_params} params.")
280
+
281
+ # params
282
+ # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support
283
+ # out-of-core computing solution
284
+ total_numel = 0
285
+ total_params = 0
286
+ for shapes, full_single_fp32_vector in zip(param_shapes, merged_single_partition_of_fp32_groups):
287
+ offset = 0
288
+ avail_numel = full_single_fp32_vector.numel()
289
+ for name, shape in shapes.items():
290
+
291
+ unpartitioned_numel = shape.numel() if _has_callable(shape, 'numel') else math.prod(shape)
292
+ total_numel += unpartitioned_numel
293
+ total_params += 1
294
+
295
+ if debug:
296
+ print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ")
297
+ state_dict[name] = full_single_fp32_vector.narrow(0, offset, unpartitioned_numel).view(shape)
298
+ offset += unpartitioned_numel
299
+
300
+ # Z2 started to align to 2*world_size to improve nccl performance. Therefore both offset and
301
+ # avail_numel can differ by anywhere between 0..2*world_size. Due to two unrelated complex
302
+ # paddings performed in the code it's almost impossible to predict the exact numbers w/o the
303
+ # live optimizer object, so we are checking that the numbers are within the right range
304
+ align_to = 2 * world_size
305
+
306
+ def zero2_align(x):
307
+ return align_to * math.ceil(x / align_to)
308
+
309
+ if debug:
310
+ print(f"original offset={offset}, avail_numel={avail_numel}")
311
+
312
+ offset = zero2_align(offset)
313
+ avail_numel = zero2_align(avail_numel)
314
+
315
+ if debug:
316
+ print(f"aligned offset={offset}, avail_numel={avail_numel}")
317
+
318
+ # Sanity check
319
+ if offset != avail_numel:
320
+ raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong")
321
+
322
+ print(f"Reconstructed fp32 state dict with {total_params} params {total_numel} elements")
323
+
324
+
325
+ def _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states,
326
+ exclude_frozen_parameters):
327
+ state_dict = OrderedDict()
328
+
329
+ # buffers
330
+ buffers = zero_model_states[0].buffers
331
+ state_dict.update(buffers)
332
+ if debug:
333
+ print(f"added {len(buffers)} buffers")
334
+
335
+ if not exclude_frozen_parameters:
336
+ _zero2_merge_frozen_params(state_dict, zero_model_states)
337
+
338
+ _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states)
339
+
340
+ # recover shared parameters
341
+ for pair in zero_model_states[0].shared_params:
342
+ if pair[1] in state_dict:
343
+ state_dict[pair[0]] = state_dict[pair[1]]
344
+
345
+ return state_dict
346
+
347
+
348
+ def zero3_partitioned_param_info(unpartitioned_numel, world_size):
349
+ remainder = unpartitioned_numel % world_size
350
+ padding_numel = (world_size - remainder) if remainder else 0
351
+ partitioned_numel = math.ceil(unpartitioned_numel / world_size)
352
+ return partitioned_numel, padding_numel
353
+
354
+
355
+ def _zero3_merge_frozen_params(state_dict, world_size, zero_model_states):
356
+ if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0:
357
+ return
358
+
359
+ if debug:
360
+ for i in range(world_size):
361
+ num_elem = sum(s.numel() for s in zero_model_states[i].frozen_param_fragments.values())
362
+ print(f'rank {i}: {FROZEN_PARAM_SHAPES}.numel = {num_elem}')
363
+
364
+ frozen_param_shapes = zero_model_states[0].frozen_param_shapes
365
+ wanted_params = len(frozen_param_shapes)
366
+ wanted_numel = sum(s.numel() for s in frozen_param_shapes.values())
367
+ avail_numel = sum([p.numel() for p in zero_model_states[0].frozen_param_fragments.values()]) * world_size
368
+ print(f'Frozen params: Have {avail_numel} numels to process.')
369
+ print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params')
370
+
371
+ total_params = 0
372
+ total_numel = 0
373
+ for name, shape in zero_model_states[0].frozen_param_shapes.items():
374
+ total_params += 1
375
+ unpartitioned_numel = shape.numel()
376
+ total_numel += unpartitioned_numel
377
+
378
+ param_frags = tuple(model_state.frozen_param_fragments[name] for model_state in zero_model_states)
379
+ state_dict[name] = torch.cat(param_frags, 0).narrow(0, 0, unpartitioned_numel).view(shape)
380
+
381
+ partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size)
382
+
383
+ if debug:
384
+ print(
385
+ f"Frozen params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}"
386
+ )
387
+
388
+ print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements")
389
+
390
+
391
+ class GatheredTensor:
392
+ """
393
+ A pseudo tensor that collects partitioned weights.
394
+ It is more memory efficient when there are multiple groups.
395
+ """
396
+
397
+ def __init__(self, flat_groups, flat_groups_offset, offset, partitioned_numel, shape):
398
+ self.flat_groups = flat_groups
399
+ self.flat_groups_offset = flat_groups_offset
400
+ self.offset = offset
401
+ self.partitioned_numel = partitioned_numel
402
+ self.shape = shape
403
+ self.dtype = self.flat_groups[0][0].dtype
404
+
405
+ def contiguous(self):
406
+ """
407
+ Merge partitioned weights from flat_groups into a single tensor.
408
+ """
409
+ end_idx = self.offset + self.partitioned_numel
410
+ world_size = len(self.flat_groups)
411
+ pad_flat_param_chunks = []
412
+
413
+ for rank_i in range(world_size):
414
+ # for each rank, we need to collect weights from related group/groups
415
+ flat_groups_at_rank_i = self.flat_groups[rank_i]
416
+ start_group_id = None
417
+ end_group_id = None
418
+ for group_id in range(len(self.flat_groups_offset)):
419
+ if self.flat_groups_offset[group_id] <= self.offset < self.flat_groups_offset[group_id + 1]:
420
+ start_group_id = group_id
421
+ if self.flat_groups_offset[group_id] < end_idx <= self.flat_groups_offset[group_id + 1]:
422
+ end_group_id = group_id
423
+ break
424
+ # collect weights from related group/groups
425
+ for group_id in range(start_group_id, end_group_id + 1):
426
+ flat_tensor = flat_groups_at_rank_i[group_id]
427
+ start_offset = self.offset - self.flat_groups_offset[group_id]
428
+ end_offset = min(end_idx, self.flat_groups_offset[group_id + 1]) - self.flat_groups_offset[group_id]
429
+ pad_flat_param_chunks.append(flat_tensor[start_offset:end_offset])
430
+
431
+ # collect weights from all ranks
432
+ pad_flat_param = torch.cat(pad_flat_param_chunks, dim=0)
433
+ param = pad_flat_param[:self.shape.numel()].view(self.shape).contiguous()
434
+ return param
435
+
436
+
437
+ def _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states):
438
+ param_shapes = zero_model_states[0].param_shapes
439
+ avail_numel = sum([flat_group.numel() for flat_group in fp32_flat_groups[0]]) * world_size
440
+
441
+ # Reconstruction protocol: For zero3 we need to zip the partitions together at boundary of each
442
+ # param, re-consolidating each param, while dealing with padding if any
443
+
444
+ # merge list of dicts, preserving order
445
+ param_shapes = {k: v for d in param_shapes for k, v in d.items()}
446
+
447
+ if debug:
448
+ for i in range(world_size):
449
+ print(f"{FP32_FLAT_GROUPS}[{i}].shape={fp32_flat_groups[i].shape}")
450
+
451
+ wanted_params = len(param_shapes)
452
+ wanted_numel = sum(shape.numel() for shape in param_shapes.values())
453
+ # not asserting if there is a mismatch due to possible padding
454
+ avail_numel = fp32_flat_groups[0].numel() * world_size
455
+ print(f"Trainable params: Have {avail_numel} numels to process.")
456
+ print(f"Trainable params: Need {wanted_numel} numels in {wanted_params} params.")
457
+
458
+ # params
459
+ # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support
460
+ # out-of-core computing solution
461
+ offset = 0
462
+ total_numel = 0
463
+ total_params = 0
464
+ flat_groups_offset = [0] + list(np.cumsum([flat_tensor.numel() for flat_tensor in fp32_flat_groups[0]]))
465
+ for name, shape in tqdm(param_shapes.items(), desc='Gathering sharded weights'):
466
+ unpartitioned_numel = shape.numel()
467
+ total_numel += unpartitioned_numel
468
+ total_params += 1
469
+ partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size)
470
+
471
+ if debug:
472
+ print(
473
+ f"Trainable params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}"
474
+ )
475
+
476
+ # memory efficient tensor
477
+ tensor = GatheredTensor(fp32_flat_groups, flat_groups_offset, offset, partitioned_numel, shape)
478
+ state_dict[name] = tensor
479
+ offset += partitioned_numel
480
+
481
+ offset *= world_size
482
+
483
+ # Sanity check
484
+ if offset != avail_numel:
485
+ raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong")
486
+
487
+ print(f"Reconstructed Trainable fp32 state dict with {total_params} params {total_numel} elements")
488
+
489
+
490
+ def _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states,
491
+ exclude_frozen_parameters):
492
+ state_dict = OrderedDict()
493
+
494
+ # buffers
495
+ buffers = zero_model_states[0].buffers
496
+ state_dict.update(buffers)
497
+ if debug:
498
+ print(f"added {len(buffers)} buffers")
499
+
500
+ if not exclude_frozen_parameters:
501
+ _zero3_merge_frozen_params(state_dict, world_size, zero_model_states)
502
+
503
+ _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states)
504
+
505
+ # recover shared parameters
506
+ for pair in zero_model_states[0].shared_params:
507
+ if pair[1] in state_dict:
508
+ state_dict[pair[0]] = state_dict[pair[1]]
509
+
510
+ return state_dict
511
+
512
+
513
+ def to_torch_tensor(state_dict, return_empty_tensor=False):
514
+ """
515
+ Convert state_dict of GatheredTensor to torch tensor
516
+ """
517
+ torch_state_dict = {}
518
+ converted_tensors = {}
519
+ for name, tensor in state_dict.items():
520
+ tensor_id = id(tensor)
521
+ if tensor_id in converted_tensors: # shared tensors
522
+ shared_tensor = torch_state_dict[converted_tensors[tensor_id]]
523
+ torch_state_dict[name] = shared_tensor
524
+ else:
525
+ converted_tensors[tensor_id] = name
526
+ if return_empty_tensor:
527
+ torch_state_dict[name] = torch.empty(tensor.shape, dtype=tensor.dtype)
528
+ else:
529
+ torch_state_dict[name] = tensor.contiguous()
530
+ return torch_state_dict
531
+
532
+
533
+ def get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir,
534
+ tag=None,
535
+ exclude_frozen_parameters=False,
536
+ lazy_mode=False):
537
+ """
538
+ Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated state_dict that can be loaded with
539
+ ``load_state_dict()`` and used for training without DeepSpeed or shared with others, for example
540
+ via a model hub.
541
+
542
+ Args:
543
+ - ``checkpoint_dir``: path to the desired checkpoint folder
544
+ - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in 'latest' file. e.g., ``global_step14``
545
+ - ``exclude_frozen_parameters``: exclude frozen parameters
546
+ - ``lazy_mode``: get state_dict in lazy mode. It returns a dict of pesduo tensor instead of torch tensor, which is more memory efficient.
547
+ Convert the pesduo tensor to torch tensor by ``.contiguous()``
548
+
549
+ Returns:
550
+ - pytorch ``state_dict``
551
+
552
+ A typical usage might be ::
553
+
554
+ from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint
555
+ # do the training and checkpoint saving
556
+ state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir) # already on cpu
557
+ model = model.cpu() # move to cpu
558
+ model.load_state_dict(state_dict)
559
+ # submit to model hub or save the model to share with others
560
+
561
+ In this example the ``model`` will no longer be usable in the deepspeed context of the same
562
+ application. i.e. you will need to re-initialize the deepspeed engine, since
563
+ ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it.
564
+
565
+ If you want it all done for you, use ``load_state_dict_from_zero_checkpoint`` instead.
566
+
567
+ Note: the above usage may not work if your application doesn't have sufficient free CPU memory.
568
+ You may need to use the offline approach using the ``zero_to_fp32.py`` script that is saved with
569
+ the checkpoint. Or you can load state_dict in lazy mode ::
570
+
571
+ from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint
572
+ state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, lazy_mode=True) # not on cpu
573
+ for name, lazy_tensor in state_dict.item():
574
+ tensor = lazy_tensor.contiguous() # to cpu
575
+ print(name, tensor)
576
+ # del tensor to release memory if it no longer in use
577
+ """
578
+ if tag is None:
579
+ latest_path = os.path.join(checkpoint_dir, 'latest')
580
+ if os.path.isfile(latest_path):
581
+ with open(latest_path, 'r') as fd:
582
+ tag = fd.read().strip()
583
+ else:
584
+ raise ValueError(f"Unable to find 'latest' file at {latest_path}")
585
+
586
+ ds_checkpoint_dir = os.path.join(checkpoint_dir, tag)
587
+
588
+ if not os.path.isdir(ds_checkpoint_dir):
589
+ raise FileNotFoundError(f"Directory '{ds_checkpoint_dir}' doesn't exist")
590
+
591
+ state_dict = _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters)
592
+ if lazy_mode:
593
+ return state_dict
594
+ else:
595
+ return to_torch_tensor(state_dict)
596
+
597
+
598
+ def convert_zero_checkpoint_to_fp32_state_dict(checkpoint_dir,
599
+ output_dir,
600
+ max_shard_size="5GB",
601
+ safe_serialization=False,
602
+ tag=None,
603
+ exclude_frozen_parameters=False):
604
+ """
605
+ Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` file that can be
606
+ loaded with ``torch.load(file)`` + ``load_state_dict()`` and used for training without DeepSpeed.
607
+
608
+ Args:
609
+ - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``)
610
+ - ``output_dir``: directory to the pytorch fp32 state_dict output files
611
+ - ``max_shard_size``: the maximum size for a checkpoint before being sharded, default value is 5GB
612
+ - ``safe_serialization``: whether to save the model using `safetensors` or the traditional PyTorch way (that uses `pickle`).
613
+ - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14``
614
+ - ``exclude_frozen_parameters``: exclude frozen parameters
615
+ """
616
+
617
+ # Dependency pre-check
618
+ if safe_serialization:
619
+ try:
620
+ from safetensors.torch import save_file
621
+ except ImportError:
622
+ print('If you want to use `safe_serialization`, please `pip install safetensors`')
623
+ raise
624
+ if max_shard_size is not None:
625
+ try:
626
+ from huggingface_hub import split_torch_state_dict_into_shards
627
+ except ImportError:
628
+ print('If you want to use `max_shard_size`, please `pip install huggingface_hub`')
629
+ raise
630
+
631
+ # Convert zero checkpoint to state_dict
632
+ state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir,
633
+ tag,
634
+ exclude_frozen_parameters,
635
+ lazy_mode=True)
636
+
637
+ # Shard the model if it is too big.
638
+ weights_name = "model.safetensors" if safe_serialization else "pytorch_model.bin"
639
+ if max_shard_size is not None:
640
+ filename_pattern = weights_name.replace(".bin", "{suffix}.bin").replace(".safetensors", "{suffix}.safetensors")
641
+ # an memory-efficient approach for sharding
642
+ empty_state_dict = to_torch_tensor(state_dict, return_empty_tensor=True)
643
+ state_dict_split = split_torch_state_dict_into_shards(empty_state_dict,
644
+ filename_pattern=filename_pattern,
645
+ max_shard_size=max_shard_size)
646
+ else:
647
+ from collections import namedtuple
648
+ StateDictSplit = namedtuple("StateDictSplit", ["is_sharded", "filename_to_tensors"])
649
+ state_dict_split = StateDictSplit(is_sharded=False,
650
+ filename_to_tensors={weights_name: list(state_dict.keys())})
651
+
652
+ # Save the model by shard
653
+ os.makedirs(output_dir, exist_ok=True)
654
+ filename_to_tensors = state_dict_split.filename_to_tensors.items()
655
+ for shard_file, tensors in tqdm(filename_to_tensors, desc="Saving checkpoint shards"):
656
+ shard_state_dict = {tensor_name: state_dict[tensor_name] for tensor_name in tensors}
657
+ shard_state_dict = to_torch_tensor(shard_state_dict)
658
+ output_path = os.path.join(output_dir, shard_file)
659
+ if safe_serialization:
660
+ save_file(shard_state_dict, output_path, metadata={"format": "pt"})
661
+ else:
662
+ torch.save(shard_state_dict, output_path)
663
+ # release the memory of current shard
664
+ for tensor_name in list(shard_state_dict.keys()):
665
+ del state_dict[tensor_name]
666
+ del shard_state_dict[tensor_name]
667
+ del shard_state_dict
668
+ gc.collect()
669
+
670
+ # Save index if sharded
671
+ if state_dict_split.is_sharded:
672
+ index = {
673
+ "metadata": state_dict_split.metadata,
674
+ "weight_map": state_dict_split.tensor_to_filename,
675
+ }
676
+ save_index_file = "model.safetensors.index.json" if safe_serialization else "pytorch_model.bin.index.json"
677
+ save_index_file = os.path.join(output_dir, save_index_file)
678
+ with open(save_index_file, "w", encoding="utf-8") as f:
679
+ content = json.dumps(index, indent=2, sort_keys=True) + "\n"
680
+ f.write(content)
681
+
682
+
683
+ def load_state_dict_from_zero_checkpoint(model, checkpoint_dir, tag=None):
684
+ """
685
+ 1. Put the provided model to cpu
686
+ 2. Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict``
687
+ 3. Load it into the provided model
688
+
689
+ Args:
690
+ - ``model``: the model object to update
691
+ - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``)
692
+ - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14``
693
+
694
+ Returns:
695
+ - ``model`: modified model
696
+
697
+ Make sure you have plenty of CPU memory available before you call this function. If you don't
698
+ have enough use the ``zero_to_fp32.py`` utility to do the conversion. You will find it
699
+ conveniently placed for you in the checkpoint folder.
700
+
701
+ A typical usage might be ::
702
+
703
+ from deepspeed.utils.zero_to_fp32 import load_state_dict_from_zero_checkpoint
704
+ model = load_state_dict_from_zero_checkpoint(trainer.model, checkpoint_dir)
705
+ # submit to model hub or save the model to share with others
706
+
707
+ Note, that once this was run, the ``model`` will no longer be usable in the deepspeed context
708
+ of the same application. i.e. you will need to re-initialize the deepspeed engine, since
709
+ ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it.
710
+
711
+ """
712
+ logger.info("Extracting fp32 weights")
713
+ state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag)
714
+
715
+ logger.info("Overwriting model with fp32 weights")
716
+ model = model.cpu()
717
+ model.load_state_dict(state_dict, strict=False)
718
+
719
+ return model
720
+
721
+
722
+ if __name__ == "__main__":
723
+ parser = argparse.ArgumentParser()
724
+ parser.add_argument("checkpoint_dir",
725
+ type=str,
726
+ help="path to the desired checkpoint folder, e.g., path/checkpoint-12")
727
+ parser.add_argument("output_dir",
728
+ type=str,
729
+ help="directory to the pytorch fp32 state_dict output files"
730
+ "(e.g. path/checkpoint-12-output/)")
731
+ parser.add_argument(
732
+ "--max_shard_size",
733
+ type=str,
734
+ default="5GB",
735
+ help="The maximum size for a checkpoint before being sharded. Checkpoints shard will then be each of size"
736
+ "lower than this size. If expressed as a string, needs to be digits followed by a unit (like `5MB`"
737
+ "We default it to 5GB in order for models to be able to run easily on free-tier google colab instances"
738
+ "without CPU OOM issues.")
739
+ parser.add_argument(
740
+ "--safe_serialization",
741
+ default=False,
742
+ action='store_true',
743
+ help="Whether to save the model using `safetensors` or the traditional PyTorch way (that uses `pickle`).")
744
+ parser.add_argument("-t",
745
+ "--tag",
746
+ type=str,
747
+ default=None,
748
+ help="checkpoint tag used as a unique identifier for checkpoint. e.g., global_step1")
749
+ parser.add_argument("--exclude_frozen_parameters", action='store_true', help="exclude frozen parameters")
750
+ parser.add_argument("-d", "--debug", action='store_true', help="enable debug")
751
+ args = parser.parse_args()
752
+
753
+ debug = args.debug
754
+
755
+ convert_zero_checkpoint_to_fp32_state_dict(args.checkpoint_dir,
756
+ args.output_dir,
757
+ max_shard_size=args.max_shard_size,
758
+ safe_serialization=args.safe_serialization,
759
+ tag=args.tag,
760
+ exclude_frozen_parameters=args.exclude_frozen_parameters)
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/loss_log.jsonl CHANGED
@@ -6290,3 +6290,610 @@
6290
  {"step": 17970, "loss": 0.0586, "learning_rate": 2.7929753142667883e-07}
6291
  {"step": 17970, "dynamics_loss_avg": 0.04188279584050179}
6292
  {"step": 17970, "action_loss_avg": 0.020719322189688683}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6290
  {"step": 17970, "loss": 0.0586, "learning_rate": 2.7929753142667883e-07}
6291
  {"step": 17970, "dynamics_loss_avg": 0.04188279584050179}
6292
  {"step": 17970, "action_loss_avg": 0.020719322189688683}
6293
+ {"step": 17980, "loss": 0.0657, "learning_rate": 2.765795370105401e-07}
6294
+ {"step": 17980, "dynamics_loss_avg": 0.04451025417074561}
6295
+ {"step": 17980, "action_loss_avg": 0.025259029027074574}
6296
+ {"step": 17990, "loss": 0.0769, "learning_rate": 2.738744562455831e-07}
6297
+ {"step": 17990, "dynamics_loss_avg": 0.05019981041550636}
6298
+ {"step": 17990, "action_loss_avg": 0.019757727766409515}
6299
+ {"step": 18000, "loss": 0.0624, "learning_rate": 2.711822965273975e-07}
6300
+ {"step": 18000, "dynamics_loss_avg": 0.03530683349817991}
6301
+ {"step": 18000, "action_loss_avg": 0.01971647096797824}
6302
+ {"step": 18010, "loss": 0.0723, "learning_rate": 2.685030652162424e-07}
6303
+ {"step": 18010, "dynamics_loss_avg": 0.04856907473877072}
6304
+ {"step": 18010, "action_loss_avg": 0.023391773412004113}
6305
+ {"step": 18020, "loss": 0.0713, "learning_rate": 2.658367696370351e-07}
6306
+ {"step": 18020, "dynamics_loss_avg": 0.052192554157227275}
6307
+ {"step": 18020, "action_loss_avg": 0.01916952694300562}
6308
+ {"step": 18030, "loss": 0.0711, "learning_rate": 2.631834170793268e-07}
6309
+ {"step": 18030, "dynamics_loss_avg": 0.05118879219517112}
6310
+ {"step": 18030, "action_loss_avg": 0.018339621974155305}
6311
+ {"step": 18040, "loss": 0.0786, "learning_rate": 2.605430147972804e-07}
6312
+ {"step": 18040, "dynamics_loss_avg": 0.056528258882462976}
6313
+ {"step": 18040, "action_loss_avg": 0.03153302231803536}
6314
+ {"step": 18050, "loss": 0.093, "learning_rate": 2.579155700096575e-07}
6315
+ {"step": 18050, "dynamics_loss_avg": 0.06036507375538349}
6316
+ {"step": 18050, "action_loss_avg": 0.030267918528988958}
6317
+ {"step": 18060, "loss": 0.0572, "learning_rate": 2.5530108989978875e-07}
6318
+ {"step": 18060, "dynamics_loss_avg": 0.033854188490659}
6319
+ {"step": 18060, "action_loss_avg": 0.011101155891083181}
6320
+ {"step": 18070, "loss": 0.0622, "learning_rate": 2.526995816155642e-07}
6321
+ {"step": 18070, "dynamics_loss_avg": 0.03401682879775762}
6322
+ {"step": 18070, "action_loss_avg": 0.016593949124217033}
6323
+ {"step": 18080, "loss": 0.0651, "learning_rate": 2.501110522694089e-07}
6324
+ {"step": 18080, "dynamics_loss_avg": 0.04331094124354422}
6325
+ {"step": 18080, "action_loss_avg": 0.019172144029289483}
6326
+ {"step": 18090, "loss": 0.0722, "learning_rate": 2.475355089382625e-07}
6327
+ {"step": 18090, "dynamics_loss_avg": 0.04457311574369669}
6328
+ {"step": 18090, "action_loss_avg": 0.022115184273570777}
6329
+ {"step": 18100, "loss": 0.0765, "learning_rate": 2.4497295866356295e-07}
6330
+ {"step": 18100, "dynamics_loss_avg": 0.04849362997338176}
6331
+ {"step": 18100, "action_loss_avg": 0.021924686199054123}
6332
+ {"step": 18110, "loss": 0.0662, "learning_rate": 2.424234084512228e-07}
6333
+ {"step": 18110, "dynamics_loss_avg": 0.04284500405192375}
6334
+ {"step": 18110, "action_loss_avg": 0.013206246495246887}
6335
+ {"step": 18120, "loss": 0.0708, "learning_rate": 2.3988686527161686e-07}
6336
+ {"step": 18120, "dynamics_loss_avg": 0.04819774907082319}
6337
+ {"step": 18120, "action_loss_avg": 0.02406782256439328}
6338
+ {"step": 18130, "loss": 0.0681, "learning_rate": 2.373633360595573e-07}
6339
+ {"step": 18130, "dynamics_loss_avg": 0.050570879317820075}
6340
+ {"step": 18130, "action_loss_avg": 0.01779633527621627}
6341
+ {"step": 18140, "loss": 0.0715, "learning_rate": 2.3485282771427586e-07}
6342
+ {"step": 18140, "dynamics_loss_avg": 0.03040293287485838}
6343
+ {"step": 18140, "action_loss_avg": 0.012802653654944152}
6344
+ {"step": 18150, "loss": 0.0647, "learning_rate": 2.3235534709940665e-07}
6345
+ {"step": 18150, "dynamics_loss_avg": 0.038292928226292135}
6346
+ {"step": 18150, "action_loss_avg": 0.01287715877406299}
6347
+ {"step": 18160, "loss": 0.0739, "learning_rate": 2.2987090104296616e-07}
6348
+ {"step": 18160, "dynamics_loss_avg": 0.054934194311499596}
6349
+ {"step": 18160, "action_loss_avg": 0.026963144540786743}
6350
+ {"step": 18170, "loss": 0.0622, "learning_rate": 2.2739949633733549e-07}
6351
+ {"step": 18170, "dynamics_loss_avg": 0.0363878505770117}
6352
+ {"step": 18170, "action_loss_avg": 0.011499037570320069}
6353
+ {"step": 18180, "loss": 0.0633, "learning_rate": 2.2494113973924091e-07}
6354
+ {"step": 18180, "dynamics_loss_avg": 0.04405585685744882}
6355
+ {"step": 18180, "action_loss_avg": 0.01619838015176356}
6356
+ {"step": 18190, "loss": 0.0837, "learning_rate": 2.2249583796973507e-07}
6357
+ {"step": 18190, "dynamics_loss_avg": 0.05213240385055542}
6358
+ {"step": 18190, "action_loss_avg": 0.022103648679330946}
6359
+ {"step": 18200, "loss": 0.0613, "learning_rate": 2.200635977141796e-07}
6360
+ {"step": 18200, "dynamics_loss_avg": 0.053078430145978926}
6361
+ {"step": 18200, "action_loss_avg": 0.01745017133653164}
6362
+ {"step": 18210, "loss": 0.0761, "learning_rate": 2.1764442562222598e-07}
6363
+ {"step": 18210, "dynamics_loss_avg": 0.049139073304831984}
6364
+ {"step": 18210, "action_loss_avg": 0.024428400211036205}
6365
+ {"step": 18220, "loss": 0.0733, "learning_rate": 2.152383283077991e-07}
6366
+ {"step": 18220, "dynamics_loss_avg": 0.050191032327711584}
6367
+ {"step": 18220, "action_loss_avg": 0.02038680133409798}
6368
+ {"step": 18230, "loss": 0.0627, "learning_rate": 2.1284531234907812e-07}
6369
+ {"step": 18230, "dynamics_loss_avg": 0.04079515226185322}
6370
+ {"step": 18230, "action_loss_avg": 0.018291610572487117}
6371
+ {"step": 18240, "loss": 0.0555, "learning_rate": 2.1046538428847464e-07}
6372
+ {"step": 18240, "dynamics_loss_avg": 0.03139160391874611}
6373
+ {"step": 18240, "action_loss_avg": 0.019224531413055957}
6374
+ {"step": 18250, "loss": 0.0781, "learning_rate": 2.0809855063262275e-07}
6375
+ {"step": 18250, "dynamics_loss_avg": 0.05537284519523382}
6376
+ {"step": 18250, "action_loss_avg": 0.01678796047344804}
6377
+ {"step": 18260, "loss": 0.0738, "learning_rate": 2.0574481785235577e-07}
6378
+ {"step": 18260, "dynamics_loss_avg": 0.04461161983199417}
6379
+ {"step": 18260, "action_loss_avg": 0.01744754402898252}
6380
+ {"step": 18270, "loss": 0.0541, "learning_rate": 2.034041923826885e-07}
6381
+ {"step": 18270, "dynamics_loss_avg": 0.03249894678592682}
6382
+ {"step": 18270, "action_loss_avg": 0.013080501649528741}
6383
+ {"step": 18280, "loss": 0.0731, "learning_rate": 2.0107668062280207e-07}
6384
+ {"step": 18280, "dynamics_loss_avg": 0.04757325639948249}
6385
+ {"step": 18280, "action_loss_avg": 0.017159340134821833}
6386
+ {"step": 18290, "loss": 0.0792, "learning_rate": 1.9876228893602355e-07}
6387
+ {"step": 18290, "dynamics_loss_avg": 0.06475771861150861}
6388
+ {"step": 18290, "action_loss_avg": 0.02554315598681569}
6389
+ {"step": 18300, "loss": 0.0815, "learning_rate": 1.9646102364981269e-07}
6390
+ {"step": 18300, "dynamics_loss_avg": 0.05235052909702063}
6391
+ {"step": 18300, "action_loss_avg": 0.02022950756363571}
6392
+ {"step": 18310, "loss": 0.0734, "learning_rate": 1.9417289105574054e-07}
6393
+ {"step": 18310, "dynamics_loss_avg": 0.05568016031756997}
6394
+ {"step": 18310, "action_loss_avg": 0.018663786817342043}
6395
+ {"step": 18320, "loss": 0.0784, "learning_rate": 1.918978974094743e-07}
6396
+ {"step": 18320, "dynamics_loss_avg": 0.051876201014965775}
6397
+ {"step": 18320, "action_loss_avg": 0.02238959036767483}
6398
+ {"step": 18330, "loss": 0.0682, "learning_rate": 1.896360489307597e-07}
6399
+ {"step": 18330, "dynamics_loss_avg": 0.04768909700214863}
6400
+ {"step": 18330, "action_loss_avg": 0.011938971444033086}
6401
+ {"step": 18340, "loss": 0.0702, "learning_rate": 1.8738735180340362e-07}
6402
+ {"step": 18340, "dynamics_loss_avg": 0.04586377870291471}
6403
+ {"step": 18340, "action_loss_avg": 0.017745603155344725}
6404
+ {"step": 18350, "loss": 0.0744, "learning_rate": 1.8515181217525823e-07}
6405
+ {"step": 18350, "dynamics_loss_avg": 0.04679070338606835}
6406
+ {"step": 18350, "action_loss_avg": 0.019874962512403727}
6407
+ {"step": 18360, "loss": 0.0724, "learning_rate": 1.8292943615820458e-07}
6408
+ {"step": 18360, "dynamics_loss_avg": 0.0325282221660018}
6409
+ {"step": 18360, "action_loss_avg": 0.012186819082126021}
6410
+ {"step": 18370, "loss": 0.0692, "learning_rate": 1.8072022982813297e-07}
6411
+ {"step": 18370, "dynamics_loss_avg": 0.03504083342850208}
6412
+ {"step": 18370, "action_loss_avg": 0.017874133703298867}
6413
+ {"step": 18380, "loss": 0.0647, "learning_rate": 1.7852419922492925e-07}
6414
+ {"step": 18380, "dynamics_loss_avg": 0.04656160455197096}
6415
+ {"step": 18380, "action_loss_avg": 0.02956375349313021}
6416
+ {"step": 18390, "loss": 0.0618, "learning_rate": 1.7634135035245692e-07}
6417
+ {"step": 18390, "dynamics_loss_avg": 0.03509335024282336}
6418
+ {"step": 18390, "action_loss_avg": 0.01892667606007308}
6419
+ {"step": 18400, "loss": 0.0679, "learning_rate": 1.7417168917854167e-07}
6420
+ {"step": 18400, "dynamics_loss_avg": 0.04116056151688099}
6421
+ {"step": 18400, "action_loss_avg": 0.01439917974639684}
6422
+ {"step": 18410, "loss": 0.067, "learning_rate": 1.720152216349552e-07}
6423
+ {"step": 18410, "dynamics_loss_avg": 0.04588055443018675}
6424
+ {"step": 18410, "action_loss_avg": 0.015296191233210266}
6425
+ {"step": 18420, "loss": 0.0616, "learning_rate": 1.6987195361739595e-07}
6426
+ {"step": 18420, "dynamics_loss_avg": 0.04652413250878453}
6427
+ {"step": 18420, "action_loss_avg": 0.02111965431831777}
6428
+ {"step": 18430, "loss": 0.0711, "learning_rate": 1.6774189098547833e-07}
6429
+ {"step": 18430, "dynamics_loss_avg": 0.05404613558202982}
6430
+ {"step": 18430, "action_loss_avg": 0.022632275894284248}
6431
+ {"step": 18440, "loss": 0.0579, "learning_rate": 1.656250395627107e-07}
6432
+ {"step": 18440, "dynamics_loss_avg": 0.04214397612959146}
6433
+ {"step": 18440, "action_loss_avg": 0.011458933795802295}
6434
+ {"step": 18450, "loss": 0.0849, "learning_rate": 1.6352140513648418e-07}
6435
+ {"step": 18450, "dynamics_loss_avg": 0.06370990518480539}
6436
+ {"step": 18450, "action_loss_avg": 0.019021764234639704}
6437
+ {"step": 18460, "loss": 0.0576, "learning_rate": 1.6143099345805714e-07}
6438
+ {"step": 18460, "dynamics_loss_avg": 0.03974438169971108}
6439
+ {"step": 18460, "action_loss_avg": 0.011968593345955014}
6440
+ {"step": 18470, "loss": 0.0704, "learning_rate": 1.5935381024253294e-07}
6441
+ {"step": 18470, "dynamics_loss_avg": 0.04946697074919939}
6442
+ {"step": 18470, "action_loss_avg": 0.02251418838277459}
6443
+ {"step": 18480, "loss": 0.0658, "learning_rate": 1.572898611688517e-07}
6444
+ {"step": 18480, "dynamics_loss_avg": 0.049329818598926066}
6445
+ {"step": 18480, "action_loss_avg": 0.0209303927840665}
6446
+ {"step": 18490, "loss": 0.085, "learning_rate": 1.5523915187977135e-07}
6447
+ {"step": 18490, "dynamics_loss_avg": 0.06325942352414131}
6448
+ {"step": 18490, "action_loss_avg": 0.021714000264182687}
6449
+ {"step": 18500, "loss": 0.0679, "learning_rate": 1.532016879818532e-07}
6450
+ {"step": 18500, "dynamics_loss_avg": 0.04309999439865351}
6451
+ {"step": 18500, "action_loss_avg": 0.01855244832113385}
6452
+ {"step": 18510, "loss": 0.0574, "learning_rate": 1.51177475045447e-07}
6453
+ {"step": 18510, "dynamics_loss_avg": 0.03668638826347888}
6454
+ {"step": 18510, "action_loss_avg": 0.014108453458175064}
6455
+ {"step": 18520, "loss": 0.0653, "learning_rate": 1.4916651860467035e-07}
6456
+ {"step": 18520, "dynamics_loss_avg": 0.0457814609631896}
6457
+ {"step": 18520, "action_loss_avg": 0.01547098788432777}
6458
+ {"step": 18530, "loss": 0.0888, "learning_rate": 1.471688241574043e-07}
6459
+ {"step": 18530, "dynamics_loss_avg": 0.06023927703499794}
6460
+ {"step": 18530, "action_loss_avg": 0.01835664624813944}
6461
+ {"step": 18540, "loss": 0.0617, "learning_rate": 1.4518439716526721e-07}
6462
+ {"step": 18540, "dynamics_loss_avg": 0.04002130590379238}
6463
+ {"step": 18540, "action_loss_avg": 0.011030881525948643}
6464
+ {"step": 18550, "loss": 0.0706, "learning_rate": 1.432132430536076e-07}
6465
+ {"step": 18550, "dynamics_loss_avg": 0.05326564330607653}
6466
+ {"step": 18550, "action_loss_avg": 0.028722409531474112}
6467
+ {"step": 18560, "loss": 0.0679, "learning_rate": 1.412553672114869e-07}
6468
+ {"step": 18560, "dynamics_loss_avg": 0.03678571954369545}
6469
+ {"step": 18560, "action_loss_avg": 0.013442619773559272}
6470
+ {"step": 18570, "loss": 0.0762, "learning_rate": 1.3931077499166056e-07}
6471
+ {"step": 18570, "dynamics_loss_avg": 0.051235838793218136}
6472
+ {"step": 18570, "action_loss_avg": 0.014766597165726125}
6473
+ {"step": 18580, "loss": 0.0623, "learning_rate": 1.3737947171057087e-07}
6474
+ {"step": 18580, "dynamics_loss_avg": 0.04611561596393585}
6475
+ {"step": 18580, "action_loss_avg": 0.024290019972249867}
6476
+ {"step": 18590, "loss": 0.0617, "learning_rate": 1.3546146264832583e-07}
6477
+ {"step": 18590, "dynamics_loss_avg": 0.043639526097103955}
6478
+ {"step": 18590, "action_loss_avg": 0.020817835722118615}
6479
+ {"step": 18600, "loss": 0.0542, "learning_rate": 1.3355675304869086e-07}
6480
+ {"step": 18600, "dynamics_loss_avg": 0.041601867880672216}
6481
+ {"step": 18600, "action_loss_avg": 0.018878667848184705}
6482
+ {"step": 18610, "loss": 0.0741, "learning_rate": 1.3166534811906827e-07}
6483
+ {"step": 18610, "dynamics_loss_avg": 0.03837147168815136}
6484
+ {"step": 18610, "action_loss_avg": 0.012793690583202988}
6485
+ {"step": 18620, "loss": 0.0573, "learning_rate": 1.2978725303048666e-07}
6486
+ {"step": 18620, "dynamics_loss_avg": 0.040544414147734645}
6487
+ {"step": 18620, "action_loss_avg": 0.014463109290227294}
6488
+ {"step": 18630, "loss": 0.0763, "learning_rate": 1.2792247291758763e-07}
6489
+ {"step": 18630, "dynamics_loss_avg": 0.047443586960434915}
6490
+ {"step": 18630, "action_loss_avg": 0.01775962356477976}
6491
+ {"step": 18640, "loss": 0.0729, "learning_rate": 1.2607101287860635e-07}
6492
+ {"step": 18640, "dynamics_loss_avg": 0.05059406105428934}
6493
+ {"step": 18640, "action_loss_avg": 0.017567940149456264}
6494
+ {"step": 18650, "loss": 0.0812, "learning_rate": 1.2423287797536654e-07}
6495
+ {"step": 18650, "dynamics_loss_avg": 0.0561508284881711}
6496
+ {"step": 18650, "action_loss_avg": 0.02486546281725168}
6497
+ {"step": 18660, "loss": 0.0649, "learning_rate": 1.2240807323325775e-07}
6498
+ {"step": 18660, "dynamics_loss_avg": 0.03691480532288551}
6499
+ {"step": 18660, "action_loss_avg": 0.013475889107212424}
6500
+ {"step": 18670, "loss": 0.0563, "learning_rate": 1.2059660364122538e-07}
6501
+ {"step": 18670, "dynamics_loss_avg": 0.042761373054236176}
6502
+ {"step": 18670, "action_loss_avg": 0.015165302250534297}
6503
+ {"step": 18680, "loss": 0.0611, "learning_rate": 1.187984741517595e-07}
6504
+ {"step": 18680, "dynamics_loss_avg": 0.038918837811797856}
6505
+ {"step": 18680, "action_loss_avg": 0.014585279813036322}
6506
+ {"step": 18690, "loss": 0.0792, "learning_rate": 1.1701368968087711e-07}
6507
+ {"step": 18690, "dynamics_loss_avg": 0.04762068502604962}
6508
+ {"step": 18690, "action_loss_avg": 0.02017206819728017}
6509
+ {"step": 18700, "loss": 0.0607, "learning_rate": 1.1524225510811116e-07}
6510
+ {"step": 18700, "dynamics_loss_avg": 0.04003286678344011}
6511
+ {"step": 18700, "action_loss_avg": 0.017731298226863146}
6512
+ {"step": 18710, "loss": 0.0674, "learning_rate": 1.1348417527649536e-07}
6513
+ {"step": 18710, "dynamics_loss_avg": 0.04693224634975195}
6514
+ {"step": 18710, "action_loss_avg": 0.027163309697061776}
6515
+ {"step": 18720, "loss": 0.0697, "learning_rate": 1.1173945499255268e-07}
6516
+ {"step": 18720, "dynamics_loss_avg": 0.04715399304404855}
6517
+ {"step": 18720, "action_loss_avg": 0.020895790844224393}
6518
+ {"step": 18730, "loss": 0.0843, "learning_rate": 1.1000809902628307e-07}
6519
+ {"step": 18730, "dynamics_loss_avg": 0.056629661656916144}
6520
+ {"step": 18730, "action_loss_avg": 0.02812079694122076}
6521
+ {"step": 18740, "loss": 0.0694, "learning_rate": 1.0829011211114681e-07}
6522
+ {"step": 18740, "dynamics_loss_avg": 0.04512308090925217}
6523
+ {"step": 18740, "action_loss_avg": 0.012506886711344123}
6524
+ {"step": 18750, "loss": 0.0625, "learning_rate": 1.0658549894405457e-07}
6525
+ {"step": 18750, "dynamics_loss_avg": 0.049353060126304624}
6526
+ {"step": 18750, "action_loss_avg": 0.01917194863781333}
6527
+ {"step": 18760, "loss": 0.0716, "learning_rate": 1.0489426418535342e-07}
6528
+ {"step": 18760, "dynamics_loss_avg": 0.045650312304496767}
6529
+ {"step": 18760, "action_loss_avg": 0.013405046029947697}
6530
+ {"step": 18770, "loss": 0.0768, "learning_rate": 1.0321641245881475e-07}
6531
+ {"step": 18770, "dynamics_loss_avg": 0.046050991769880055}
6532
+ {"step": 18770, "action_loss_avg": 0.02187738287029788}
6533
+ {"step": 18780, "loss": 0.0672, "learning_rate": 1.0155194835162141e-07}
6534
+ {"step": 18780, "dynamics_loss_avg": 0.04976412951946259}
6535
+ {"step": 18780, "action_loss_avg": 0.023234894685447215}
6536
+ {"step": 18790, "loss": 0.0636, "learning_rate": 9.990087641435443e-08}
6537
+ {"step": 18790, "dynamics_loss_avg": 0.05185487000271678}
6538
+ {"step": 18790, "action_loss_avg": 0.016145678516477347}
6539
+ {"step": 18800, "loss": 0.0675, "learning_rate": 9.826320116098131e-08}
6540
+ {"step": 18800, "dynamics_loss_avg": 0.04294224474579096}
6541
+ {"step": 18800, "action_loss_avg": 0.01435599448159337}
6542
+ {"step": 18810, "loss": 0.0695, "learning_rate": 9.663892706884448e-08}
6543
+ {"step": 18810, "dynamics_loss_avg": 0.06062636505812406}
6544
+ {"step": 18810, "action_loss_avg": 0.022002062015235423}
6545
+ {"step": 18820, "loss": 0.0741, "learning_rate": 9.502805857864617e-08}
6546
+ {"step": 18820, "dynamics_loss_avg": 0.045694983564317226}
6547
+ {"step": 18820, "action_loss_avg": 0.014170713955536485}
6548
+ {"step": 18830, "loss": 0.0655, "learning_rate": 9.343060009443961e-08}
6549
+ {"step": 18830, "dynamics_loss_avg": 0.056142124626785514}
6550
+ {"step": 18830, "action_loss_avg": 0.015817865310236812}
6551
+ {"step": 18840, "loss": 0.0752, "learning_rate": 9.184655598361624e-08}
6552
+ {"step": 18840, "dynamics_loss_avg": 0.05955648459494114}
6553
+ {"step": 18840, "action_loss_avg": 0.019305437244474886}
6554
+ {"step": 18850, "loss": 0.0802, "learning_rate": 9.027593057689077e-08}
6555
+ {"step": 18850, "dynamics_loss_avg": 0.05589356829877943}
6556
+ {"step": 18850, "action_loss_avg": 0.02508344443049282}
6557
+ {"step": 18860, "loss": 0.0742, "learning_rate": 8.871872816829442e-08}
6558
+ {"step": 18860, "dynamics_loss_avg": 0.04614152507856488}
6559
+ {"step": 18860, "action_loss_avg": 0.018156480311881752}
6560
+ {"step": 18870, "loss": 0.0594, "learning_rate": 8.717495301515777e-08}
6561
+ {"step": 18870, "dynamics_loss_avg": 0.030260950746014714}
6562
+ {"step": 18870, "action_loss_avg": 0.017760740919038655}
6563
+ {"step": 18880, "loss": 0.0856, "learning_rate": 8.564460933810414e-08}
6564
+ {"step": 18880, "dynamics_loss_avg": 0.0606656925752759}
6565
+ {"step": 18880, "action_loss_avg": 0.022149211214855313}
6566
+ {"step": 18890, "loss": 0.0675, "learning_rate": 8.412770132103454e-08}
6567
+ {"step": 18890, "dynamics_loss_avg": 0.0462961982935667}
6568
+ {"step": 18890, "action_loss_avg": 0.026284475857391954}
6569
+ {"step": 18900, "loss": 0.0674, "learning_rate": 8.262423311111711e-08}
6570
+ {"step": 18900, "dynamics_loss_avg": 0.03955181259661913}
6571
+ {"step": 18900, "action_loss_avg": 0.017146135633811354}
6572
+ {"step": 18910, "loss": 0.0538, "learning_rate": 8.113420881877665e-08}
6573
+ {"step": 18910, "dynamics_loss_avg": 0.027236733864992858}
6574
+ {"step": 18910, "action_loss_avg": 0.007272660755552352}
6575
+ {"step": 18920, "loss": 0.0579, "learning_rate": 7.965763251768289e-08}
6576
+ {"step": 18920, "dynamics_loss_avg": 0.03394540082663298}
6577
+ {"step": 18920, "action_loss_avg": 0.016379949520342052}
6578
+ {"step": 18930, "loss": 0.0757, "learning_rate": 7.819450824473995e-08}
6579
+ {"step": 18930, "dynamics_loss_avg": 0.06098019005730748}
6580
+ {"step": 18930, "action_loss_avg": 0.019424212723970415}
6581
+ {"step": 18940, "loss": 0.0631, "learning_rate": 7.674484000007199e-08}
6582
+ {"step": 18940, "dynamics_loss_avg": 0.05012268982827663}
6583
+ {"step": 18940, "action_loss_avg": 0.012639198126271367}
6584
+ {"step": 18950, "loss": 0.0632, "learning_rate": 7.530863174701752e-08}
6585
+ {"step": 18950, "dynamics_loss_avg": 0.046331493742764}
6586
+ {"step": 18950, "action_loss_avg": 0.016248486703261733}
6587
+ {"step": 18960, "loss": 0.0648, "learning_rate": 7.388588741211511e-08}
6588
+ {"step": 18960, "dynamics_loss_avg": 0.04534003343433142}
6589
+ {"step": 18960, "action_loss_avg": 0.018011249601840973}
6590
+ {"step": 18970, "loss": 0.0579, "learning_rate": 7.247661088509328e-08}
6591
+ {"step": 18970, "dynamics_loss_avg": 0.035985720716416834}
6592
+ {"step": 18970, "action_loss_avg": 0.009873280115425587}
6593
+ {"step": 18980, "loss": 0.0658, "learning_rate": 7.108080601886003e-08}
6594
+ {"step": 18980, "dynamics_loss_avg": 0.04430606840178371}
6595
+ {"step": 18980, "action_loss_avg": 0.015798577875830232}
6596
+ {"step": 18990, "loss": 0.079, "learning_rate": 6.969847662949337e-08}
6597
+ {"step": 18990, "dynamics_loss_avg": 0.047527849674224854}
6598
+ {"step": 18990, "action_loss_avg": 0.022077369946055116}
6599
+ {"step": 19000, "loss": 0.0669, "learning_rate": 6.832962649622799e-08}
6600
+ {"step": 19000, "dynamics_loss_avg": 0.04626642623916268}
6601
+ {"step": 19000, "action_loss_avg": 0.01617791245225817}
6602
+ {"step": 19010, "loss": 0.0754, "learning_rate": 6.697425936144864e-08}
6603
+ {"step": 19010, "dynamics_loss_avg": 0.04557799994945526}
6604
+ {"step": 19010, "action_loss_avg": 0.021255811862647533}
6605
+ {"step": 19020, "loss": 0.0547, "learning_rate": 6.563237893067731e-08}
6606
+ {"step": 19020, "dynamics_loss_avg": 0.03289818889461458}
6607
+ {"step": 19020, "action_loss_avg": 0.023261255864053965}
6608
+ {"step": 19030, "loss": 0.0735, "learning_rate": 6.430398887256329e-08}
6609
+ {"step": 19030, "dynamics_loss_avg": 0.05449823662638664}
6610
+ {"step": 19030, "action_loss_avg": 0.023226954974234104}
6611
+ {"step": 19040, "loss": 0.0756, "learning_rate": 6.298909281887478e-08}
6612
+ {"step": 19040, "dynamics_loss_avg": 0.04155576010234654}
6613
+ {"step": 19040, "action_loss_avg": 0.02150076034013182}
6614
+ {"step": 19050, "loss": 0.0655, "learning_rate": 6.168769436448674e-08}
6615
+ {"step": 19050, "dynamics_loss_avg": 0.035638409107923506}
6616
+ {"step": 19050, "action_loss_avg": 0.018701795022934675}
6617
+ {"step": 19060, "loss": 0.0672, "learning_rate": 6.03997970673742e-08}
6618
+ {"step": 19060, "dynamics_loss_avg": 0.05180228874087334}
6619
+ {"step": 19060, "action_loss_avg": 0.03301493991166353}
6620
+ {"step": 19070, "loss": 0.0723, "learning_rate": 5.9125404448597825e-08}
6621
+ {"step": 19070, "dynamics_loss_avg": 0.05123816449195147}
6622
+ {"step": 19070, "action_loss_avg": 0.03283484075218439}
6623
+ {"step": 19080, "loss": 0.0619, "learning_rate": 5.786451999229836e-08}
6624
+ {"step": 19080, "dynamics_loss_avg": 0.042518378654494884}
6625
+ {"step": 19080, "action_loss_avg": 0.028800006583333017}
6626
+ {"step": 19090, "loss": 0.0646, "learning_rate": 5.6617147145687226e-08}
6627
+ {"step": 19090, "dynamics_loss_avg": 0.052104505337774755}
6628
+ {"step": 19090, "action_loss_avg": 0.03023729813285172}
6629
+ {"step": 19100, "loss": 0.065, "learning_rate": 5.538328931903259e-08}
6630
+ {"step": 19100, "dynamics_loss_avg": 0.04455666309222579}
6631
+ {"step": 19100, "action_loss_avg": 0.016629776498302818}
6632
+ {"step": 19110, "loss": 0.0656, "learning_rate": 5.416294988565551e-08}
6633
+ {"step": 19110, "dynamics_loss_avg": 0.03844494493678212}
6634
+ {"step": 19110, "action_loss_avg": 0.020590703887864947}
6635
+ {"step": 19120, "loss": 0.0788, "learning_rate": 5.29561321819172e-08}
6636
+ {"step": 19120, "dynamics_loss_avg": 0.048755235597491266}
6637
+ {"step": 19120, "action_loss_avg": 0.04096140079200268}
6638
+ {"step": 19130, "loss": 0.0564, "learning_rate": 5.176283950721062e-08}
6639
+ {"step": 19130, "dynamics_loss_avg": 0.03957191361114383}
6640
+ {"step": 19130, "action_loss_avg": 0.021234600781463088}
6641
+ {"step": 19140, "loss": 0.0569, "learning_rate": 5.058307512395333e-08}
6642
+ {"step": 19140, "dynamics_loss_avg": 0.03663741974160075}
6643
+ {"step": 19140, "action_loss_avg": 0.014955365541391074}
6644
+ {"step": 19150, "loss": 0.0559, "learning_rate": 4.941684225757526e-08}
6645
+ {"step": 19150, "dynamics_loss_avg": 0.038839674554765224}
6646
+ {"step": 19150, "action_loss_avg": 0.014020096138119697}
6647
+ {"step": 19160, "loss": 0.0697, "learning_rate": 4.8264144096513145e-08}
6648
+ {"step": 19160, "dynamics_loss_avg": 0.0577778372913599}
6649
+ {"step": 19160, "action_loss_avg": 0.023691162606701253}
6650
+ {"step": 19170, "loss": 0.0836, "learning_rate": 4.712498379219943e-08}
6651
+ {"step": 19170, "dynamics_loss_avg": 0.05882037840783596}
6652
+ {"step": 19170, "action_loss_avg": 0.03158386466093362}
6653
+ {"step": 19180, "loss": 0.0696, "learning_rate": 4.599936445905506e-08}
6654
+ {"step": 19180, "dynamics_loss_avg": 0.04895575139671564}
6655
+ {"step": 19180, "action_loss_avg": 0.018969870824366807}
6656
+ {"step": 19190, "loss": 0.0693, "learning_rate": 4.4887289174480594e-08}
6657
+ {"step": 19190, "dynamics_loss_avg": 0.04373278133571148}
6658
+ {"step": 19190, "action_loss_avg": 0.025109104276634753}
6659
+ {"step": 19200, "loss": 0.0599, "learning_rate": 4.3788760978846215e-08}
6660
+ {"step": 19200, "dynamics_loss_avg": 0.04576329002156854}
6661
+ {"step": 19200, "action_loss_avg": 0.016592811932787298}
6662
+ {"step": 19210, "loss": 0.0666, "learning_rate": 4.2703782875487265e-08}
6663
+ {"step": 19210, "dynamics_loss_avg": 0.055230416171252725}
6664
+ {"step": 19210, "action_loss_avg": 0.02290729838423431}
6665
+ {"step": 19220, "loss": 0.0637, "learning_rate": 4.1632357830692083e-08}
6666
+ {"step": 19220, "dynamics_loss_avg": 0.04439005693420768}
6667
+ {"step": 19220, "action_loss_avg": 0.024878822336904704}
6668
+ {"step": 19230, "loss": 0.0662, "learning_rate": 4.0574488773695855e-08}
6669
+ {"step": 19230, "dynamics_loss_avg": 0.05375898201018572}
6670
+ {"step": 19230, "action_loss_avg": 0.025994909834116696}
6671
+ {"step": 19240, "loss": 0.067, "learning_rate": 3.95301785966723e-08}
6672
+ {"step": 19240, "dynamics_loss_avg": 0.04041470270603895}
6673
+ {"step": 19240, "action_loss_avg": 0.017607183149084448}
6674
+ {"step": 19250, "loss": 0.0556, "learning_rate": 3.849943015472479e-08}
6675
+ {"step": 19250, "dynamics_loss_avg": 0.03672714214771986}
6676
+ {"step": 19250, "action_loss_avg": 0.019279234297573566}
6677
+ {"step": 19260, "loss": 0.0643, "learning_rate": 3.748224626588137e-08}
6678
+ {"step": 19260, "dynamics_loss_avg": 0.04293777644634247}
6679
+ {"step": 19260, "action_loss_avg": 0.016091548628173767}
6680
+ {"step": 19270, "loss": 0.0813, "learning_rate": 3.647862971108307e-08}
6681
+ {"step": 19270, "dynamics_loss_avg": 0.06033924128860235}
6682
+ {"step": 19270, "action_loss_avg": 0.02317564245313406}
6683
+ {"step": 19280, "loss": 0.0676, "learning_rate": 3.5488583234179474e-08}
6684
+ {"step": 19280, "dynamics_loss_avg": 0.04067262168973684}
6685
+ {"step": 19280, "action_loss_avg": 0.01721664038486779}
6686
+ {"step": 19290, "loss": 0.0666, "learning_rate": 3.4512109541920413e-08}
6687
+ {"step": 19290, "dynamics_loss_avg": 0.04518165625631809}
6688
+ {"step": 19290, "action_loss_avg": 0.025825847452506424}
6689
+ {"step": 19300, "loss": 0.0664, "learning_rate": 3.354921130394706e-08}
6690
+ {"step": 19300, "dynamics_loss_avg": 0.048086746409535405}
6691
+ {"step": 19300, "action_loss_avg": 0.02090142834931612}
6692
+ {"step": 19310, "loss": 0.0547, "learning_rate": 3.259989115278639e-08}
6693
+ {"step": 19310, "dynamics_loss_avg": 0.03277552807703614}
6694
+ {"step": 19310, "action_loss_avg": 0.0179377687163651}
6695
+ {"step": 19320, "loss": 0.0756, "learning_rate": 3.1664151683843404e-08}
6696
+ {"step": 19320, "dynamics_loss_avg": 0.05201920215040445}
6697
+ {"step": 19320, "action_loss_avg": 0.025618729740381242}
6698
+ {"step": 19330, "loss": 0.0757, "learning_rate": 3.0741995455394466e-08}
6699
+ {"step": 19330, "dynamics_loss_avg": 0.05810132035985589}
6700
+ {"step": 19330, "action_loss_avg": 0.02606740880291909}
6701
+ {"step": 19340, "loss": 0.0698, "learning_rate": 2.9833424988579554e-08}
6702
+ {"step": 19340, "dynamics_loss_avg": 0.04584471378475428}
6703
+ {"step": 19340, "action_loss_avg": 0.029441210301592947}
6704
+ {"step": 19350, "loss": 0.0621, "learning_rate": 2.8938442767394992e-08}
6705
+ {"step": 19350, "dynamics_loss_avg": 0.035817237198352815}
6706
+ {"step": 19350, "action_loss_avg": 0.013609715178608895}
6707
+ {"step": 19360, "loss": 0.064, "learning_rate": 2.8057051238688515e-08}
6708
+ {"step": 19360, "dynamics_loss_avg": 0.047259863931685686}
6709
+ {"step": 19360, "action_loss_avg": 0.01589830848388374}
6710
+ {"step": 19370, "loss": 0.0646, "learning_rate": 2.71892528121509e-08}
6711
+ {"step": 19370, "dynamics_loss_avg": 0.04148031026124954}
6712
+ {"step": 19370, "action_loss_avg": 0.02225602187681943}
6713
+ {"step": 19380, "loss": 0.0678, "learning_rate": 2.633504986030988e-08}
6714
+ {"step": 19380, "dynamics_loss_avg": 0.04275577375665307}
6715
+ {"step": 19380, "action_loss_avg": 0.017516118101775648}
6716
+ {"step": 19390, "loss": 0.0542, "learning_rate": 2.549444471852347e-08}
6717
+ {"step": 19390, "dynamics_loss_avg": 0.032702566031366585}
6718
+ {"step": 19390, "action_loss_avg": 0.013057024264708162}
6719
+ {"step": 19400, "loss": 0.0557, "learning_rate": 2.4667439684974424e-08}
6720
+ {"step": 19400, "dynamics_loss_avg": 0.045774967316538094}
6721
+ {"step": 19400, "action_loss_avg": 0.014352619275450706}
6722
+ {"step": 19410, "loss": 0.0616, "learning_rate": 2.3854037020662467e-08}
6723
+ {"step": 19410, "dynamics_loss_avg": 0.03391762087121606}
6724
+ {"step": 19410, "action_loss_avg": 0.02073847660794854}
6725
+ {"step": 19420, "loss": 0.0577, "learning_rate": 2.3054238949399287e-08}
6726
+ {"step": 19420, "dynamics_loss_avg": 0.04494827501475811}
6727
+ {"step": 19420, "action_loss_avg": 0.014736767392605543}
6728
+ {"step": 19430, "loss": 0.0617, "learning_rate": 2.2268047657802994e-08}
6729
+ {"step": 19430, "dynamics_loss_avg": 0.04049129285849631}
6730
+ {"step": 19430, "action_loss_avg": 0.016750149230938406}
6731
+ {"step": 19440, "loss": 0.0643, "learning_rate": 2.149546529529034e-08}
6732
+ {"step": 19440, "dynamics_loss_avg": 0.05459021460264921}
6733
+ {"step": 19440, "action_loss_avg": 0.027651762869209052}
6734
+ {"step": 19450, "loss": 0.0685, "learning_rate": 2.0736493974071736e-08}
6735
+ {"step": 19450, "dynamics_loss_avg": 0.0578347940929234}
6736
+ {"step": 19450, "action_loss_avg": 0.022148777451366185}
6737
+ {"step": 19460, "loss": 0.0734, "learning_rate": 1.9991135769145687e-08}
6738
+ {"step": 19460, "dynamics_loss_avg": 0.06578245982527733}
6739
+ {"step": 19460, "action_loss_avg": 0.025048522651195525}
6740
+ {"step": 19470, "loss": 0.0658, "learning_rate": 1.9259392718293247e-08}
6741
+ {"step": 19470, "dynamics_loss_avg": 0.04603032823652029}
6742
+ {"step": 19470, "action_loss_avg": 0.018549220450222493}
6743
+ {"step": 19480, "loss": 0.0842, "learning_rate": 1.8541266822072467e-08}
6744
+ {"step": 19480, "dynamics_loss_avg": 0.05411523971706629}
6745
+ {"step": 19480, "action_loss_avg": 0.024996600742451847}
6746
+ {"step": 19490, "loss": 0.0735, "learning_rate": 1.7836760043811187e-08}
6747
+ {"step": 19490, "dynamics_loss_avg": 0.04837729539722204}
6748
+ {"step": 19490, "action_loss_avg": 0.029992511868476866}
6749
+ {"step": 19500, "loss": 0.0775, "learning_rate": 1.7145874309604793e-08}
6750
+ {"step": 19500, "dynamics_loss_avg": 0.05609335135668516}
6751
+ {"step": 19500, "action_loss_avg": 0.03079576869495213}
6752
+ {"step": 19510, "loss": 0.0737, "learning_rate": 1.6468611508308474e-08}
6753
+ {"step": 19510, "dynamics_loss_avg": 0.038016231171786784}
6754
+ {"step": 19510, "action_loss_avg": 0.022260204795747994}
6755
+ {"step": 19520, "loss": 0.0648, "learning_rate": 1.5804973491532206e-08}
6756
+ {"step": 19520, "dynamics_loss_avg": 0.04410122968256473}
6757
+ {"step": 19520, "action_loss_avg": 0.020528348349034786}
6758
+ {"step": 19530, "loss": 0.0623, "learning_rate": 1.5154962073637425e-08}
6759
+ {"step": 19530, "dynamics_loss_avg": 0.03693989543244243}
6760
+ {"step": 19530, "action_loss_avg": 0.018374732113443315}
6761
+ {"step": 19540, "loss": 0.0586, "learning_rate": 1.4518579031730373e-08}
6762
+ {"step": 19540, "dynamics_loss_avg": 0.03409141227602959}
6763
+ {"step": 19540, "action_loss_avg": 0.020327942539006473}
6764
+ {"step": 19550, "loss": 0.0753, "learning_rate": 1.389582610565876e-08}
6765
+ {"step": 19550, "dynamics_loss_avg": 0.051672781072556975}
6766
+ {"step": 19550, "action_loss_avg": 0.029321196768432856}
6767
+ {"step": 19560, "loss": 0.0669, "learning_rate": 1.3286704998003997e-08}
6768
+ {"step": 19560, "dynamics_loss_avg": 0.049665603414177896}
6769
+ {"step": 19560, "action_loss_avg": 0.023823793698102235}
6770
+ {"step": 19570, "loss": 0.0664, "learning_rate": 1.2691217374080634e-08}
6771
+ {"step": 19570, "dynamics_loss_avg": 0.04450676534324884}
6772
+ {"step": 19570, "action_loss_avg": 0.020256917690858245}
6773
+ {"step": 19580, "loss": 0.0762, "learning_rate": 1.2109364861929707e-08}
6774
+ {"step": 19580, "dynamics_loss_avg": 0.05879650972783566}
6775
+ {"step": 19580, "action_loss_avg": 0.020828585047274827}
6776
+ {"step": 19590, "loss": 0.0646, "learning_rate": 1.1541149052312628e-08}
6777
+ {"step": 19590, "dynamics_loss_avg": 0.045380566362291576}
6778
+ {"step": 19590, "action_loss_avg": 0.019981274381279945}
6779
+ {"step": 19600, "loss": 0.0681, "learning_rate": 1.0986571498710075e-08}
6780
+ {"step": 19600, "dynamics_loss_avg": 0.05374464839696884}
6781
+ {"step": 19600, "action_loss_avg": 0.01966229798272252}
6782
+ {"step": 19610, "loss": 0.0664, "learning_rate": 1.044563371731644e-08}
6783
+ {"step": 19610, "dynamics_loss_avg": 0.04756856011226773}
6784
+ {"step": 19610, "action_loss_avg": 0.03430402267258614}
6785
+ {"step": 19620, "loss": 0.075, "learning_rate": 9.918337187034276e-09}
6786
+ {"step": 19620, "dynamics_loss_avg": 0.054150094650685786}
6787
+ {"step": 19620, "action_loss_avg": 0.02385778082534671}
6788
+ {"step": 19630, "loss": 0.0562, "learning_rate": 9.404683349472642e-09}
6789
+ {"step": 19630, "dynamics_loss_avg": 0.041935459896922114}
6790
+ {"step": 19630, "action_loss_avg": 0.02104444750584662}
6791
+ {"step": 19640, "loss": 0.0703, "learning_rate": 8.904673608940983e-09}
6792
+ {"step": 19640, "dynamics_loss_avg": 0.04481700295582414}
6793
+ {"step": 19640, "action_loss_avg": 0.025399928074330093}
6794
+ {"step": 19650, "loss": 0.0611, "learning_rate": 8.418309332447472e-09}
6795
+ {"step": 19650, "dynamics_loss_avg": 0.04475516434758901}
6796
+ {"step": 19650, "action_loss_avg": 0.02396374901290983}
6797
+ {"step": 19660, "loss": 0.0603, "learning_rate": 7.945591849692901e-09}
6798
+ {"step": 19660, "dynamics_loss_avg": 0.042971633933484556}
6799
+ {"step": 19660, "action_loss_avg": 0.022925299359485507}
6800
+ {"step": 19670, "loss": 0.0587, "learning_rate": 7.486522453069578e-09}
6801
+ {"step": 19670, "dynamics_loss_avg": 0.04090790711343288}
6802
+ {"step": 19670, "action_loss_avg": 0.018234587763436137}
6803
+ {"step": 19680, "loss": 0.0674, "learning_rate": 7.041102397655208e-09}
6804
+ {"step": 19680, "dynamics_loss_avg": 0.04668019823729992}
6805
+ {"step": 19680, "action_loss_avg": 0.01666556135751307}
6806
+ {"step": 19690, "loss": 0.0704, "learning_rate": 6.609332901210686e-09}
6807
+ {"step": 19690, "dynamics_loss_avg": 0.04854318592697382}
6808
+ {"step": 19690, "action_loss_avg": 0.01791277281008661}
6809
+ {"step": 19700, "loss": 0.0628, "learning_rate": 6.191215144178419e-09}
6810
+ {"step": 19700, "dynamics_loss_avg": 0.04074359182268381}
6811
+ {"step": 19700, "action_loss_avg": 0.022232988104224204}
6812
+ {"step": 19710, "loss": 0.0769, "learning_rate": 5.7867502696756785e-09}
6813
+ {"step": 19710, "dynamics_loss_avg": 0.05371684115380049}
6814
+ {"step": 19710, "action_loss_avg": 0.028047747211530805}
6815
+ {"step": 19720, "loss": 0.0738, "learning_rate": 5.395939383494031e-09}
6816
+ {"step": 19720, "dynamics_loss_avg": 0.05972814448177814}
6817
+ {"step": 19720, "action_loss_avg": 0.020521515607833864}
6818
+ {"step": 19730, "loss": 0.0684, "learning_rate": 5.018783554095464e-09}
6819
+ {"step": 19730, "dynamics_loss_avg": 0.04780337158590555}
6820
+ {"step": 19730, "action_loss_avg": 0.01911618495360017}
6821
+ {"step": 19740, "loss": 0.0633, "learning_rate": 4.655283812610156e-09}
6822
+ {"step": 19740, "dynamics_loss_avg": 0.04706993810832501}
6823
+ {"step": 19740, "action_loss_avg": 0.02198672341182828}
6824
+ {"step": 19750, "loss": 0.0565, "learning_rate": 4.305441152831491e-09}
6825
+ {"step": 19750, "dynamics_loss_avg": 0.03720401818864048}
6826
+ {"step": 19750, "action_loss_avg": 0.01753886609803885}
6827
+ {"step": 19760, "loss": 0.0662, "learning_rate": 3.969256531217158e-09}
6828
+ {"step": 19760, "dynamics_loss_avg": 0.045891742780804635}
6829
+ {"step": 19760, "action_loss_avg": 0.017851687874644993}
6830
+ {"step": 19770, "loss": 0.0505, "learning_rate": 3.6467308668824975e-09}
6831
+ {"step": 19770, "dynamics_loss_avg": 0.029992605652660133}
6832
+ {"step": 19770, "action_loss_avg": 0.01609015348367393}
6833
+ {"step": 19780, "loss": 0.0535, "learning_rate": 3.3378650416004965e-09}
6834
+ {"step": 19780, "dynamics_loss_avg": 0.030971918720752}
6835
+ {"step": 19780, "action_loss_avg": 0.012941750744357706}
6836
+ {"step": 19790, "loss": 0.0712, "learning_rate": 3.042659899797906e-09}
6837
+ {"step": 19790, "dynamics_loss_avg": 0.05431404039263725}
6838
+ {"step": 19790, "action_loss_avg": 0.03052880633622408}
6839
+ {"step": 19800, "loss": 0.0776, "learning_rate": 2.7611162485552403e-09}
6840
+ {"step": 19800, "dynamics_loss_avg": 0.045865561440587045}
6841
+ {"step": 19800, "action_loss_avg": 0.026190615445375442}
6842
+ {"step": 19810, "loss": 0.0754, "learning_rate": 2.4932348576017782e-09}
6843
+ {"step": 19810, "dynamics_loss_avg": 0.05829719081521034}
6844
+ {"step": 19810, "action_loss_avg": 0.025177312176674606}
6845
+ {"step": 19820, "loss": 0.0661, "learning_rate": 2.2390164593144583e-09}
6846
+ {"step": 19820, "dynamics_loss_avg": 0.04026942318305373}
6847
+ {"step": 19820, "action_loss_avg": 0.018375073047354817}
6848
+ {"step": 19830, "loss": 0.0754, "learning_rate": 1.9984617487173174e-09}
6849
+ {"step": 19830, "dynamics_loss_avg": 0.05735848154872656}
6850
+ {"step": 19830, "action_loss_avg": 0.029920194949954747}
6851
+ {"step": 19840, "loss": 0.0632, "learning_rate": 1.7715713834776105e-09}
6852
+ {"step": 19840, "dynamics_loss_avg": 0.034946164861321447}
6853
+ {"step": 19840, "action_loss_avg": 0.020855458034202457}
6854
+ {"step": 19850, "loss": 0.063, "learning_rate": 1.5583459839046965e-09}
6855
+ {"step": 19850, "dynamics_loss_avg": 0.04919734569266439}
6856
+ {"step": 19850, "action_loss_avg": 0.02176156654022634}
6857
+ {"step": 19860, "loss": 0.0642, "learning_rate": 1.3587861329489305e-09}
6858
+ {"step": 19860, "dynamics_loss_avg": 0.030154460575431585}
6859
+ {"step": 19860, "action_loss_avg": 0.01854857854777947}
6860
+ {"step": 19870, "loss": 0.064, "learning_rate": 1.1728923761994415e-09}
6861
+ {"step": 19870, "dynamics_loss_avg": 0.045471681468188765}
6862
+ {"step": 19870, "action_loss_avg": 0.02216670736670494}
6863
+ {"step": 19880, "loss": 0.0718, "learning_rate": 1.0006652218819135e-09}
6864
+ {"step": 19880, "dynamics_loss_avg": 0.05042210947722196}
6865
+ {"step": 19880, "action_loss_avg": 0.019688624422997236}
6866
+ {"step": 19890, "loss": 0.0519, "learning_rate": 8.421051408596948e-10}
6867
+ {"step": 19890, "dynamics_loss_avg": 0.023838566755875944}
6868
+ {"step": 19890, "action_loss_avg": 0.010728829819709063}
6869
+ {"step": 19900, "loss": 0.0653, "learning_rate": 6.972125666299123e-10}
6870
+ {"step": 19900, "dynamics_loss_avg": 0.042154907062649724}
6871
+ {"step": 19900, "action_loss_avg": 0.022989201312884688}
6872
+ {"step": 19910, "loss": 0.068, "learning_rate": 5.65987895322917e-10}
6873
+ {"step": 19910, "dynamics_loss_avg": 0.053467546589672565}
6874
+ {"step": 19910, "action_loss_avg": 0.026885805232450367}
6875
+ {"step": 19920, "loss": 0.0676, "learning_rate": 4.48431485701728e-10}
6876
+ {"step": 19920, "dynamics_loss_avg": 0.04784501497633755}
6877
+ {"step": 19920, "action_loss_avg": 0.021005964279174803}
6878
+ {"step": 19930, "loss": 0.0628, "learning_rate": 3.4454365916203326e-10}
6879
+ {"step": 19930, "dynamics_loss_avg": 0.04858131390064955}
6880
+ {"step": 19930, "action_loss_avg": 0.02204345481004566}
6881
+ {"step": 19940, "loss": 0.0743, "learning_rate": 2.5432469972830333e-10}
6882
+ {"step": 19940, "dynamics_loss_avg": 0.0493664912879467}
6883
+ {"step": 19940, "action_loss_avg": 0.016600931715220212}
6884
+ {"step": 19950, "loss": 0.0604, "learning_rate": 1.77774854056012e-10}
6885
+ {"step": 19950, "dynamics_loss_avg": 0.04475926365703344}
6886
+ {"step": 19950, "action_loss_avg": 0.02512040608562529}
6887
+ {"step": 19960, "loss": 0.0679, "learning_rate": 1.1489433142941597e-10}
6888
+ {"step": 19960, "dynamics_loss_avg": 0.04144583009183407}
6889
+ {"step": 19960, "action_loss_avg": 0.015122923068702221}
6890
+ {"step": 19970, "loss": 0.0763, "learning_rate": 6.568330376210964e-11}
6891
+ {"step": 19970, "dynamics_loss_avg": 0.0645992343313992}
6892
+ {"step": 19970, "action_loss_avg": 0.02953599998727441}
6893
+ {"step": 19980, "loss": 0.0735, "learning_rate": 3.014190559424979e-11}
6894
+ {"step": 19980, "dynamics_loss_avg": 0.051318105403333904}
6895
+ {"step": 19980, "action_loss_avg": 0.024603950278833508}
6896
+ {"step": 19990, "loss": 0.0758, "learning_rate": 8.270234094776009e-12}
6897
+ {"step": 19990, "dynamics_loss_avg": 0.047976726107299326}
6898
+ {"step": 19990, "action_loss_avg": 0.02277223765850067}
6899
+ {"step": 20000, "loss": 0.0696, "learning_rate": 6.834906085551041e-14}
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/debug-internal.log CHANGED
@@ -5,3 +5,37 @@
5
  {"time":"2026-03-30T04:15:10.413395221Z","level":"INFO","msg":"writer: started","stream_id":"ao6lob27"}
6
  {"time":"2026-03-30T04:15:10.413394431Z","level":"INFO","msg":"sender: started","stream_id":"ao6lob27"}
7
  {"time":"2026-03-30T22:30:11.007336848Z","level":"INFO","msg":"api: retrying HTTP error","status":502,"url":"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream","body":"\n<html><head>\n<meta http-equiv=\"content-type\" content=\"text/html;charset=utf-8\">\n<title>502 Server Error</title>\n</head>\n<body text=#000000 bgcolor=#ffffff>\n<h1>Error: Server Error</h1>\n<h2>The server encountered a temporary error and could not complete your request.<p>Please try again in 30 seconds.</h2>\n<h2></h2>\n</body></html>\n"}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  {"time":"2026-03-30T04:15:10.413395221Z","level":"INFO","msg":"writer: started","stream_id":"ao6lob27"}
6
  {"time":"2026-03-30T04:15:10.413394431Z","level":"INFO","msg":"sender: started","stream_id":"ao6lob27"}
7
  {"time":"2026-03-30T22:30:11.007336848Z","level":"INFO","msg":"api: retrying HTTP error","status":502,"url":"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream","body":"\n<html><head>\n<meta http-equiv=\"content-type\" content=\"text/html;charset=utf-8\">\n<title>502 Server Error</title>\n</head>\n<body text=#000000 bgcolor=#ffffff>\n<h1>Error: Server Error</h1>\n<h2>The server encountered a temporary error and could not complete your request.<p>Please try again in 30 seconds.</h2>\n<h2></h2>\n</body></html>\n"}
8
+ {"time":"2026-04-01T10:18:44.020890295Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": read tcp 192.168.0.3:41808->35.186.228.49:443: read: connection timed out"}
9
+ {"time":"2026-04-01T10:18:56.087878557Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:38712->127.0.0.53:53: i/o timeout"}
10
+ {"time":"2026-04-01T10:19:10.908212851Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:39469->127.0.0.53:53: i/o timeout"}
11
+ {"time":"2026-04-01T10:19:30.190045118Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:47884->127.0.0.53:53: i/o timeout"}
12
+ {"time":"2026-04-01T10:19:58.216957187Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:34560->127.0.0.53:53: i/o timeout"}
13
+ {"time":"2026-04-01T10:20:40.445735516Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:46537->127.0.0.53:53: i/o timeout"}
14
+ {"time":"2026-04-01T10:21:50.450660421Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:39132->127.0.0.53:53: i/o timeout"}
15
+ {"time":"2026-04-01T10:23:00.454344979Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:60808->127.0.0.53:53: i/o timeout"}
16
+ {"time":"2026-04-01T10:24:10.457186285Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:42899->127.0.0.53:53: i/o timeout"}
17
+ {"time":"2026-04-01T10:25:20.462333271Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:51640->127.0.0.53:53: i/o timeout"}
18
+ {"time":"2026-04-01T10:26:30.466228733Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:54385->127.0.0.53:53: i/o timeout"}
19
+ {"time":"2026-04-01T10:27:40.469992011Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:36994->127.0.0.53:53: i/o timeout"}
20
+ {"time":"2026-04-01T10:28:50.474794045Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:59340->127.0.0.53:53: i/o timeout"}
21
+ {"time":"2026-04-01T10:30:00.479410672Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:53914->127.0.0.53:53: i/o timeout"}
22
+ {"time":"2026-04-01T10:31:10.483284091Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:52683->127.0.0.53:53: i/o timeout"}
23
+ {"time":"2026-04-01T10:32:20.487788622Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:52313->127.0.0.53:53: i/o timeout"}
24
+ {"time":"2026-04-01T10:33:30.492784858Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:38331->127.0.0.53:53: i/o timeout"}
25
+ {"time":"2026-04-01T10:34:40.497056663Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:59696->127.0.0.53:53: i/o timeout"}
26
+ {"time":"2026-04-01T10:35:50.500859105Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:40973->127.0.0.53:53: i/o timeout"}
27
+ {"time":"2026-04-01T10:37:00.5054549Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:45222->127.0.0.53:53: i/o timeout"}
28
+ {"time":"2026-04-01T10:38:10.508535232Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:55362->127.0.0.53:53: i/o timeout"}
29
+ {"time":"2026-04-01T10:39:20.512300509Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:50089->127.0.0.53:53: i/o timeout"}
30
+ {"time":"2026-04-01T10:40:30.517197984Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:34388->127.0.0.53:53: i/o timeout"}
31
+ {"time":"2026-04-01T10:41:40.521767935Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:53489->127.0.0.53:53: i/o timeout"}
32
+ {"time":"2026-04-01T10:42:50.525040035Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:50726->127.0.0.53:53: i/o timeout"}
33
+ {"time":"2026-04-01T10:44:00.527917425Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:53434->127.0.0.53:53: i/o timeout"}
34
+ {"time":"2026-04-01T10:45:10.532222422Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:37341->127.0.0.53:53: i/o timeout"}
35
+ {"time":"2026-04-01T10:46:20.535679123Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:37921->127.0.0.53:53: i/o timeout"}
36
+ {"time":"2026-04-01T10:47:30.539549836Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:38214->127.0.0.53:53: i/o timeout"}
37
+ {"time":"2026-04-01T14:24:28.755750998Z","level":"INFO","msg":"stream: closing","id":"ao6lob27"}
38
+ {"time":"2026-04-01T14:24:29.342449946Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
39
+ {"time":"2026-04-01T14:24:29.828344988Z","level":"INFO","msg":"handler: closed","stream_id":"ao6lob27"}
40
+ {"time":"2026-04-01T14:24:29.82853025Z","level":"INFO","msg":"sender: closed","stream_id":"ao6lob27"}
41
+ {"time":"2026-04-01T14:24:29.8285597Z","level":"INFO","msg":"stream: closed","id":"ao6lob27"}
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/debug.log CHANGED
@@ -20,3 +20,5 @@ config: {'_wandb': {}}
20
  2026-03-30 04:15:10,932 INFO MainThread:977089 [wandb_run.py:_config_callback():1403] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': None, 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': None, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute', '_attn_implementation_autoset': True, 'transformers_version': '4.51.3', 'model_dtype': 'float32', 'hidden_size': 0, 'action_horizon': 24, 'action_dim': 36, 'backbone_cfg': {'_target_': 'groot.vla.model.dreamzero.backbone.identity.IdentityBackbone'}, 'action_head_cfg': {'config': {'backbone_features_projector_cfg': None, '_target_': 'groot.vla.model.dreamzero.action_head.wan_flow_matching_action_tf.WANPolicyHeadConfig', '_recursive_': False, 'tiled': False, 'tile_size_height': 34, 'tile_size_width': 34, 'tile_stride_height': 18, 'tile_stride_width': 16, 'lora_rank': 4, 'lora_alpha': 4, 'num_frames': 33, 'num_frame_per_block': 2, 'lora_target_modules': 'q,k,v,o,ffn.0,ffn.2', 'init_lora_weights': 'kaiming', 'train_architecture': 'full', 'use_gradient_checkpointing': True, 'add_pos_embed': True, 'model_dtype': 'float32', 'max_state_dim': 64, 'max_action_dim': 36, 'action_loss_embodiment_ids': [26, 17, 32], 'hidden_size': 64, 'input_embedding_dim': 1536, 'backbone_embedding_dim': 0, 'repa_layer': 8, 'repa_coeff': 1.0, 'load_pretrained_det_decode_layer_path': None, 'freeze_decode_layer': False, 'expand_batch': None, 'use_vlln': True, 'vl_self_attention_cfg': {'_target_': 'groot.vla.model.n1_5.modules.cross_attention_dit.SelfAttentionTransformer', 'positional_embeddings': None, 'num_layers': 4, 'num_attention_heads': 24, 'attention_head_dim': 64, 'dropout': 0.2, 'final_dropout': True}, 'diffusion_model_cfg': {'_target_': 'groot.vla.model.dreamzero.modules.wan_video_dit_action_casual_chunk.CausalWanModel', '_convert_': 'object', 'diffusion_model_pretrained_path': '/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P', 'model_type': 'i2v', 'frame_seqlen': 220, 'dim': 5120, 'in_dim': 36, 'ffn_dim': 13824, 'out_dim': 16, 'freq_dim': 256, 'eps': 1e-06, 'num_heads': 40, 'num_layers': 40, 'max_chunk_size': 4, 'num_frame_per_block': 2, 'num_action_per_block': 24, 'num_state_per_block': 1, 'action_dim': 36}, 'text_encoder_cfg': {'_target_': 'groot.vla.model.dreamzero.modules.wan_video_text_encoder.WanTextEncoder', '_convert_': 'object', 'text_encoder_pretrained_path': '/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_t5_umt5-xxl-enc-bf16.pth'}, 'image_encoder_cfg': {'_target_': 'groot.vla.model.dreamzero.modules.wan_video_image_encoder.WanImageEncoder', '_convert_': 'object', 'image_encoder_pretrained_path': '/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_clip_open-clip-xlm-roberta-large-vit-huge-14.pth'}, 'vae_cfg': {'_target_': 'groot.vla.model.dreamzero.modules.wan_video_vae.WanVideoVAE', '_convert_': 'object', 'vae_pretrained_path': '/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/Wan2.1_VAE.pth'}, 'action_dim': 36, 'action_horizon': 24, 'num_inference_timesteps': 4, 'noise_beta_alpha': 1.5, 'noise_beta_beta': 1.0, 'noise_s': 0.999, 'num_timestep_buckets': 1000, 'decouple_video_action_noise': False, 'video_noise_beta_alpha': 3.0, 'video_noise_beta_beta': 1.0, 'tune_projector': True, 'tune_diffusion_model': True, 'skip_component_loading': True}, '_target_': 'groot.vla.model.dreamzero.action_head.wan_flow_matching_action_tf.WANPolicyHead', '_convert_': 'object'}, 'resume_path': './checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute', 'model_type': 'vla', 'output_dir': './checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'eval_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 1e-05, 'weight_decay': 1e-05, 'adam_beta1': 0.95, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1000, 'max_steps': 20000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.05, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute/runs/Mar30_04-04-13_nebula100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 8000, 'save_total_limit': 10, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': True, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 1, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'dreamzero_real_teleop_g1_full_finetune_absolute', 'disable_tqdm': False, 'remove_unused_columns': False, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': True, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'tp_size': 0, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'groot/vla/configs/deepspeed/zero2_offload.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': 100, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': False, 'dataloader_persistent_workers': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': None, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'include_for_metrics': [], 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'use_liger_kernel': False, 'eval_use_gather_object': False, 'average_tokens_across_devices': False}
21
  2026-03-30 04:15:10,939 INFO MainThread:977089 [wandb_config.py:__setitem__():155] [no run ID] config set model/num_parameters = 22924196696 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7b9f0d23f190>>
22
  2026-03-30 04:15:10,939 INFO MainThread:977089 [wandb_run.py:_config_callback():1403] config_cb model/num_parameters 22924196696 None
 
 
 
20
  2026-03-30 04:15:10,932 INFO MainThread:977089 [wandb_run.py:_config_callback():1403] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': None, 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': None, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute', '_attn_implementation_autoset': True, 'transformers_version': '4.51.3', 'model_dtype': 'float32', 'hidden_size': 0, 'action_horizon': 24, 'action_dim': 36, 'backbone_cfg': {'_target_': 'groot.vla.model.dreamzero.backbone.identity.IdentityBackbone'}, 'action_head_cfg': {'config': {'backbone_features_projector_cfg': None, '_target_': 'groot.vla.model.dreamzero.action_head.wan_flow_matching_action_tf.WANPolicyHeadConfig', '_recursive_': False, 'tiled': False, 'tile_size_height': 34, 'tile_size_width': 34, 'tile_stride_height': 18, 'tile_stride_width': 16, 'lora_rank': 4, 'lora_alpha': 4, 'num_frames': 33, 'num_frame_per_block': 2, 'lora_target_modules': 'q,k,v,o,ffn.0,ffn.2', 'init_lora_weights': 'kaiming', 'train_architecture': 'full', 'use_gradient_checkpointing': True, 'add_pos_embed': True, 'model_dtype': 'float32', 'max_state_dim': 64, 'max_action_dim': 36, 'action_loss_embodiment_ids': [26, 17, 32], 'hidden_size': 64, 'input_embedding_dim': 1536, 'backbone_embedding_dim': 0, 'repa_layer': 8, 'repa_coeff': 1.0, 'load_pretrained_det_decode_layer_path': None, 'freeze_decode_layer': False, 'expand_batch': None, 'use_vlln': True, 'vl_self_attention_cfg': {'_target_': 'groot.vla.model.n1_5.modules.cross_attention_dit.SelfAttentionTransformer', 'positional_embeddings': None, 'num_layers': 4, 'num_attention_heads': 24, 'attention_head_dim': 64, 'dropout': 0.2, 'final_dropout': True}, 'diffusion_model_cfg': {'_target_': 'groot.vla.model.dreamzero.modules.wan_video_dit_action_casual_chunk.CausalWanModel', '_convert_': 'object', 'diffusion_model_pretrained_path': '/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P', 'model_type': 'i2v', 'frame_seqlen': 220, 'dim': 5120, 'in_dim': 36, 'ffn_dim': 13824, 'out_dim': 16, 'freq_dim': 256, 'eps': 1e-06, 'num_heads': 40, 'num_layers': 40, 'max_chunk_size': 4, 'num_frame_per_block': 2, 'num_action_per_block': 24, 'num_state_per_block': 1, 'action_dim': 36}, 'text_encoder_cfg': {'_target_': 'groot.vla.model.dreamzero.modules.wan_video_text_encoder.WanTextEncoder', '_convert_': 'object', 'text_encoder_pretrained_path': '/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_t5_umt5-xxl-enc-bf16.pth'}, 'image_encoder_cfg': {'_target_': 'groot.vla.model.dreamzero.modules.wan_video_image_encoder.WanImageEncoder', '_convert_': 'object', 'image_encoder_pretrained_path': '/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_clip_open-clip-xlm-roberta-large-vit-huge-14.pth'}, 'vae_cfg': {'_target_': 'groot.vla.model.dreamzero.modules.wan_video_vae.WanVideoVAE', '_convert_': 'object', 'vae_pretrained_path': '/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/Wan2.1_VAE.pth'}, 'action_dim': 36, 'action_horizon': 24, 'num_inference_timesteps': 4, 'noise_beta_alpha': 1.5, 'noise_beta_beta': 1.0, 'noise_s': 0.999, 'num_timestep_buckets': 1000, 'decouple_video_action_noise': False, 'video_noise_beta_alpha': 3.0, 'video_noise_beta_beta': 1.0, 'tune_projector': True, 'tune_diffusion_model': True, 'skip_component_loading': True}, '_target_': 'groot.vla.model.dreamzero.action_head.wan_flow_matching_action_tf.WANPolicyHead', '_convert_': 'object'}, 'resume_path': './checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute', 'model_type': 'vla', 'output_dir': './checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'eval_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 1e-05, 'weight_decay': 1e-05, 'adam_beta1': 0.95, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1000, 'max_steps': 20000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.05, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute/runs/Mar30_04-04-13_nebula100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 8000, 'save_total_limit': 10, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': True, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 1, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'dreamzero_real_teleop_g1_full_finetune_absolute', 'disable_tqdm': False, 'remove_unused_columns': False, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': True, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'tp_size': 0, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'groot/vla/configs/deepspeed/zero2_offload.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': 100, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': False, 'dataloader_persistent_workers': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': None, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'include_for_metrics': [], 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'use_liger_kernel': False, 'eval_use_gather_object': False, 'average_tokens_across_devices': False}
21
  2026-03-30 04:15:10,939 INFO MainThread:977089 [wandb_config.py:__setitem__():155] [no run ID] config set model/num_parameters = 22924196696 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7b9f0d23f190>>
22
  2026-03-30 04:15:10,939 INFO MainThread:977089 [wandb_run.py:_config_callback():1403] config_cb model/num_parameters 22924196696 None
23
+ 2026-04-01 14:24:28,737 INFO wandb-AsyncioManager-main:977089 [service_client.py:_forward_responses():134] Reached EOF.
24
+ 2026-04-01 14:24:28,738 INFO wandb-AsyncioManager-main:977089 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/run-20260330_041509-ao6lob27/files/config.yaml ADDED
@@ -0,0 +1,651 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _attn_implementation_autoset:
2
+ value: true
3
+ _name_or_path:
4
+ value: ./checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute
5
+ _wandb:
6
+ value:
7
+ cli_version: 0.25.0
8
+ e:
9
+ ppsci3bha2i61tuvdhjs2d9rc3ipwu77:
10
+ args:
11
+ - report_to=wandb
12
+ - data=dreamzero/real_teleop_g1_relative
13
+ - wandb_project=dreamzero
14
+ - train_architecture=full
15
+ - num_frames=33
16
+ - action_horizon=24
17
+ - num_views=1
18
+ - model=dreamzero/vla
19
+ - model/dreamzero/action_head=wan_flow_matching_action_tf
20
+ - model/dreamzero/transform=dreamzero_cotrain
21
+ - num_frame_per_block=2
22
+ - num_action_per_block=24
23
+ - num_state_per_block=1
24
+ - seed=42
25
+ - training_args.learning_rate=1e-5
26
+ - training_args.deepspeed=groot/vla/configs/deepspeed/zero2_offload.json
27
+ - save_steps=8000
28
+ - training_args.warmup_ratio=0.05
29
+ - output_dir=./checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute
30
+ - per_device_train_batch_size=1
31
+ - max_steps=20000
32
+ - weight_decay=1e-5
33
+ - save_total_limit=10
34
+ - upload_checkpoints=false
35
+ - bf16=true
36
+ - tf32=true
37
+ - eval_bf16=true
38
+ - dataloader_pin_memory=false
39
+ - dataloader_num_workers=1
40
+ - image_resolution_width=320
41
+ - image_resolution_height=176
42
+ - save_lora_only=false
43
+ - max_chunk_size=4
44
+ - frame_seqlen=220
45
+ - save_strategy=steps
46
+ - real_teleop_g1_data_root=/hfm/boqian/liboqian_data/data/real_data/gear/g1/Pick_bottle_and_turn_and_pour_into_cup_absolute
47
+ - dit_version=/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P
48
+ - text_encoder_pretrained_path=/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_t5_umt5-xxl-enc-bf16.pth
49
+ - image_encoder_pretrained_path=/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_clip_open-clip-xlm-roberta-large-vit-huge-14.pth
50
+ - vae_pretrained_path=/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/Wan2.1_VAE.pth
51
+ - tokenizer_path=/hfm/boqian/liboqian_data/checkpoints/umt5-xxl
52
+ - pretrained_model_path=/hfm/boqian/liboqian_data/checkpoints/DreamZero-AgiBot
53
+ - max_action_dim=36
54
+ - ++action_head_cfg.config.skip_component_loading=true
55
+ codePath: groot/vla/experiment/experiment.py
56
+ codePathLocal: groot/vla/experiment/experiment.py
57
+ cpu_count: 128
58
+ cpu_count_logical: 128
59
+ cudaVersion: "12.9"
60
+ disk:
61
+ /:
62
+ total: "3934171283456"
63
+ used: "3686894223360"
64
+ email: 2074894208@qq.com
65
+ executable: /hfm/boqian/miniconda3/envs/dreamzero/bin/python3.11
66
+ git:
67
+ commit: 21d1fe7f867c819a9feb12ecbfba6d0cd933d126
68
+ remote: git@github.com:boqian-li/DreamZero-private.git
69
+ gpu: NVIDIA A100-SXM4-80GB
70
+ gpu_count: 8
71
+ gpu_nvidia:
72
+ - architecture: Ampere
73
+ cudaCores: 6912
74
+ memoryTotal: "85899345920"
75
+ name: NVIDIA A100-SXM4-80GB
76
+ uuid: GPU-95155859-2785-a5fa-d493-0c9e9eb3332f
77
+ - architecture: Ampere
78
+ cudaCores: 6912
79
+ memoryTotal: "85899345920"
80
+ name: NVIDIA A100-SXM4-80GB
81
+ uuid: GPU-3829db0c-5c30-974b-5c88-9ed4f984eddf
82
+ - architecture: Ampere
83
+ cudaCores: 6912
84
+ memoryTotal: "85899345920"
85
+ name: NVIDIA A100-SXM4-80GB
86
+ uuid: GPU-0e260d0d-d121-6d8c-4c97-6ca0dfce506f
87
+ - architecture: Ampere
88
+ cudaCores: 6912
89
+ memoryTotal: "85899345920"
90
+ name: NVIDIA A100-SXM4-80GB
91
+ uuid: GPU-b99492cb-2347-0f80-e94f-a95e105eaffc
92
+ - architecture: Ampere
93
+ cudaCores: 6912
94
+ memoryTotal: "85899345920"
95
+ name: NVIDIA A100-SXM4-80GB
96
+ uuid: GPU-8a143efc-044d-2012-5820-afb258793ffd
97
+ - architecture: Ampere
98
+ cudaCores: 6912
99
+ memoryTotal: "85899345920"
100
+ name: NVIDIA A100-SXM4-80GB
101
+ uuid: GPU-9fedd571-c6c4-fd5a-1dd7-f82e048386ad
102
+ - architecture: Ampere
103
+ cudaCores: 6912
104
+ memoryTotal: "85899345920"
105
+ name: NVIDIA A100-SXM4-80GB
106
+ uuid: GPU-8cd44ee0-3344-490d-bb51-ba5b38f2b924
107
+ - architecture: Ampere
108
+ cudaCores: 6912
109
+ memoryTotal: "85899345920"
110
+ name: NVIDIA A100-SXM4-80GB
111
+ uuid: GPU-55a7cb99-494a-2d02-b9d8-6f85a0614e8c
112
+ host: nebula100
113
+ memory:
114
+ total: "1082011422720"
115
+ os: Linux-6.8.0-100-generic-x86_64-with-glibc2.39
116
+ program: /hfm/boqian/liboqian_code/DreamZero-private/groot/vla/experiment/experiment.py
117
+ python: CPython 3.11.14
118
+ root: ./checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute
119
+ startedAt: "2026-03-30T04:15:09.855560Z"
120
+ writerId: ppsci3bha2i61tuvdhjs2d9rc3ipwu77
121
+ m:
122
+ - "1": train/global_step
123
+ "6":
124
+ - 3
125
+ "7": []
126
+ - "2": '*'
127
+ "5": 1
128
+ "6":
129
+ - 1
130
+ "7": []
131
+ python_version: 3.11.14
132
+ t:
133
+ "1":
134
+ - 1
135
+ - 11
136
+ - 41
137
+ - 49
138
+ - 50
139
+ - 51
140
+ - 71
141
+ - 80
142
+ - 83
143
+ - 98
144
+ "2":
145
+ - 1
146
+ - 11
147
+ - 41
148
+ - 49
149
+ - 50
150
+ - 51
151
+ - 71
152
+ - 80
153
+ - 83
154
+ - 98
155
+ "3":
156
+ - 7
157
+ - 13
158
+ - 19
159
+ - 62
160
+ - 66
161
+ "4": 3.11.14
162
+ "5": 0.25.0
163
+ "6": 4.51.3
164
+ "9":
165
+ "1": transformers_trainer
166
+ "12": 0.25.0
167
+ "13": linux-x86_64
168
+ accelerator_config:
169
+ value:
170
+ dispatch_batches: null
171
+ even_batches: true
172
+ gradient_accumulation_kwargs: null
173
+ non_blocking: false
174
+ split_batches: false
175
+ use_seedable_sampler: true
176
+ action_dim:
177
+ value: 36
178
+ action_head_cfg:
179
+ value:
180
+ _convert_: object
181
+ _target_: groot.vla.model.dreamzero.action_head.wan_flow_matching_action_tf.WANPolicyHead
182
+ config:
183
+ _recursive_: false
184
+ _target_: groot.vla.model.dreamzero.action_head.wan_flow_matching_action_tf.WANPolicyHeadConfig
185
+ action_dim: 36
186
+ action_horizon: 24
187
+ action_loss_embodiment_ids:
188
+ - 26
189
+ - 17
190
+ - 32
191
+ add_pos_embed: true
192
+ backbone_embedding_dim: 0
193
+ backbone_features_projector_cfg: null
194
+ decouple_video_action_noise: false
195
+ diffusion_model_cfg:
196
+ _convert_: object
197
+ _target_: groot.vla.model.dreamzero.modules.wan_video_dit_action_casual_chunk.CausalWanModel
198
+ action_dim: 36
199
+ diffusion_model_pretrained_path: /hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P
200
+ dim: 5120
201
+ eps: 1e-06
202
+ ffn_dim: 13824
203
+ frame_seqlen: 220
204
+ freq_dim: 256
205
+ in_dim: 36
206
+ max_chunk_size: 4
207
+ model_type: i2v
208
+ num_action_per_block: 24
209
+ num_frame_per_block: 2
210
+ num_heads: 40
211
+ num_layers: 40
212
+ num_state_per_block: 1
213
+ out_dim: 16
214
+ expand_batch: null
215
+ freeze_decode_layer: false
216
+ hidden_size: 64
217
+ image_encoder_cfg:
218
+ _convert_: object
219
+ _target_: groot.vla.model.dreamzero.modules.wan_video_image_encoder.WanImageEncoder
220
+ image_encoder_pretrained_path: /hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_clip_open-clip-xlm-roberta-large-vit-huge-14.pth
221
+ init_lora_weights: kaiming
222
+ input_embedding_dim: 1536
223
+ load_pretrained_det_decode_layer_path: null
224
+ lora_alpha: 4
225
+ lora_rank: 4
226
+ lora_target_modules: q,k,v,o,ffn.0,ffn.2
227
+ max_action_dim: 36
228
+ max_state_dim: 64
229
+ model_dtype: float32
230
+ noise_beta_alpha: 1.5
231
+ noise_beta_beta: 1
232
+ noise_s: 0.999
233
+ num_frame_per_block: 2
234
+ num_frames: 33
235
+ num_inference_timesteps: 4
236
+ num_timestep_buckets: 1000
237
+ repa_coeff: 1
238
+ repa_layer: 8
239
+ skip_component_loading: true
240
+ text_encoder_cfg:
241
+ _convert_: object
242
+ _target_: groot.vla.model.dreamzero.modules.wan_video_text_encoder.WanTextEncoder
243
+ text_encoder_pretrained_path: /hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_t5_umt5-xxl-enc-bf16.pth
244
+ tile_size_height: 34
245
+ tile_size_width: 34
246
+ tile_stride_height: 18
247
+ tile_stride_width: 16
248
+ tiled: false
249
+ train_architecture: full
250
+ tune_diffusion_model: true
251
+ tune_projector: true
252
+ use_gradient_checkpointing: true
253
+ use_vlln: true
254
+ vae_cfg:
255
+ _convert_: object
256
+ _target_: groot.vla.model.dreamzero.modules.wan_video_vae.WanVideoVAE
257
+ vae_pretrained_path: /hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/Wan2.1_VAE.pth
258
+ video_noise_beta_alpha: 3
259
+ video_noise_beta_beta: 1
260
+ vl_self_attention_cfg:
261
+ _target_: groot.vla.model.n1_5.modules.cross_attention_dit.SelfAttentionTransformer
262
+ attention_head_dim: 64
263
+ dropout: 0.2
264
+ final_dropout: true
265
+ num_attention_heads: 24
266
+ num_layers: 4
267
+ positional_embeddings: null
268
+ action_horizon:
269
+ value: 24
270
+ adafactor:
271
+ value: false
272
+ adam_beta1:
273
+ value: 0.95
274
+ adam_beta2:
275
+ value: 0.999
276
+ adam_epsilon:
277
+ value: 1e-08
278
+ add_cross_attention:
279
+ value: false
280
+ architectures:
281
+ value: null
282
+ auto_find_batch_size:
283
+ value: false
284
+ average_tokens_across_devices:
285
+ value: false
286
+ backbone_cfg:
287
+ value:
288
+ _target_: groot.vla.model.dreamzero.backbone.identity.IdentityBackbone
289
+ bad_words_ids:
290
+ value: null
291
+ batch_eval_metrics:
292
+ value: false
293
+ begin_suppress_tokens:
294
+ value: null
295
+ bf16:
296
+ value: true
297
+ bf16_full_eval:
298
+ value: false
299
+ bos_token_id:
300
+ value: null
301
+ chunk_size_feed_forward:
302
+ value: 0
303
+ cross_attention_hidden_size:
304
+ value: null
305
+ data_seed:
306
+ value: null
307
+ dataloader_drop_last:
308
+ value: false
309
+ dataloader_num_workers:
310
+ value: 1
311
+ dataloader_persistent_workers:
312
+ value: true
313
+ dataloader_pin_memory:
314
+ value: false
315
+ dataloader_prefetch_factor:
316
+ value: null
317
+ ddp_backend:
318
+ value: null
319
+ ddp_broadcast_buffers:
320
+ value: null
321
+ ddp_bucket_cap_mb:
322
+ value: 100
323
+ ddp_find_unused_parameters:
324
+ value: false
325
+ ddp_timeout:
326
+ value: 1800
327
+ debug:
328
+ value: []
329
+ decoder_start_token_id:
330
+ value: null
331
+ deepspeed:
332
+ value: groot/vla/configs/deepspeed/zero2_offload.json
333
+ disable_tqdm:
334
+ value: false
335
+ diversity_penalty:
336
+ value: 0
337
+ do_eval:
338
+ value: false
339
+ do_predict:
340
+ value: false
341
+ do_sample:
342
+ value: false
343
+ do_train:
344
+ value: false
345
+ early_stopping:
346
+ value: false
347
+ encoder_no_repeat_ngram_size:
348
+ value: 0
349
+ eos_token_id:
350
+ value: null
351
+ eval_accumulation_steps:
352
+ value: null
353
+ eval_delay:
354
+ value: 0
355
+ eval_do_concat_batches:
356
+ value: true
357
+ eval_on_start:
358
+ value: false
359
+ eval_steps:
360
+ value: null
361
+ eval_strategy:
362
+ value: "no"
363
+ eval_use_gather_object:
364
+ value: false
365
+ exponential_decay_length_penalty:
366
+ value: null
367
+ finetuning_task:
368
+ value: null
369
+ forced_bos_token_id:
370
+ value: null
371
+ forced_eos_token_id:
372
+ value: null
373
+ fp16:
374
+ value: false
375
+ fp16_backend:
376
+ value: auto
377
+ fp16_full_eval:
378
+ value: false
379
+ fp16_opt_level:
380
+ value: O1
381
+ fsdp:
382
+ value: []
383
+ fsdp_config:
384
+ value:
385
+ min_num_params: 0
386
+ xla: false
387
+ xla_fsdp_grad_ckpt: false
388
+ xla_fsdp_v2: false
389
+ fsdp_min_num_params:
390
+ value: 0
391
+ fsdp_transformer_layer_cls_to_wrap:
392
+ value: null
393
+ full_determinism:
394
+ value: false
395
+ gradient_accumulation_steps:
396
+ value: 1
397
+ gradient_checkpointing:
398
+ value: false
399
+ gradient_checkpointing_kwargs:
400
+ value: null
401
+ greater_is_better:
402
+ value: null
403
+ group_by_length:
404
+ value: false
405
+ half_precision_backend:
406
+ value: auto
407
+ hidden_size:
408
+ value: 0
409
+ hub_always_push:
410
+ value: false
411
+ hub_model_id:
412
+ value: null
413
+ hub_private_repo:
414
+ value: null
415
+ hub_strategy:
416
+ value: every_save
417
+ hub_token:
418
+ value: <HUB_TOKEN>
419
+ id2label:
420
+ value:
421
+ "0": LABEL_0
422
+ "1": LABEL_1
423
+ ignore_data_skip:
424
+ value: true
425
+ include_for_metrics:
426
+ value: []
427
+ include_inputs_for_metrics:
428
+ value: false
429
+ include_num_input_tokens_seen:
430
+ value: false
431
+ include_tokens_per_second:
432
+ value: false
433
+ is_decoder:
434
+ value: false
435
+ is_encoder_decoder:
436
+ value: false
437
+ jit_mode_eval:
438
+ value: false
439
+ label_names:
440
+ value: null
441
+ label_smoothing_factor:
442
+ value: 0
443
+ label2id:
444
+ value:
445
+ LABEL_0: 0
446
+ LABEL_1: 1
447
+ learning_rate:
448
+ value: 1e-05
449
+ length_column_name:
450
+ value: length
451
+ length_penalty:
452
+ value: 1
453
+ load_best_model_at_end:
454
+ value: false
455
+ local_rank:
456
+ value: 0
457
+ log_level:
458
+ value: passive
459
+ log_level_replica:
460
+ value: warning
461
+ log_on_each_node:
462
+ value: true
463
+ logging_dir:
464
+ value: ./checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute/runs/Mar30_04-04-13_nebula100
465
+ logging_first_step:
466
+ value: false
467
+ logging_nan_inf_filter:
468
+ value: true
469
+ logging_steps:
470
+ value: 10
471
+ logging_strategy:
472
+ value: steps
473
+ lr_scheduler_type:
474
+ value: cosine
475
+ max_grad_norm:
476
+ value: 1
477
+ max_length:
478
+ value: 20
479
+ max_steps:
480
+ value: 20000
481
+ metric_for_best_model:
482
+ value: null
483
+ min_length:
484
+ value: 0
485
+ model/num_parameters:
486
+ value: 22924196696
487
+ model_dtype:
488
+ value: float32
489
+ model_type:
490
+ value: vla
491
+ mp_parameters:
492
+ value: ""
493
+ neftune_noise_alpha:
494
+ value: null
495
+ no_cuda:
496
+ value: false
497
+ no_repeat_ngram_size:
498
+ value: 0
499
+ num_beam_groups:
500
+ value: 1
501
+ num_beams:
502
+ value: 1
503
+ num_return_sequences:
504
+ value: 1
505
+ num_train_epochs:
506
+ value: 1000
507
+ optim:
508
+ value: adamw_torch
509
+ optim_args:
510
+ value: null
511
+ optim_target_modules:
512
+ value: null
513
+ output_attentions:
514
+ value: false
515
+ output_dir:
516
+ value: ./checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute
517
+ output_hidden_states:
518
+ value: false
519
+ output_scores:
520
+ value: false
521
+ overwrite_output_dir:
522
+ value: false
523
+ pad_token_id:
524
+ value: null
525
+ past_index:
526
+ value: -1
527
+ per_device_eval_batch_size:
528
+ value: 64
529
+ per_device_train_batch_size:
530
+ value: 1
531
+ per_gpu_eval_batch_size:
532
+ value: null
533
+ per_gpu_train_batch_size:
534
+ value: null
535
+ prediction_loss_only:
536
+ value: false
537
+ prefix:
538
+ value: null
539
+ problem_type:
540
+ value: null
541
+ push_to_hub:
542
+ value: false
543
+ push_to_hub_model_id:
544
+ value: null
545
+ push_to_hub_organization:
546
+ value: null
547
+ push_to_hub_token:
548
+ value: <PUSH_TO_HUB_TOKEN>
549
+ ray_scope:
550
+ value: last
551
+ remove_invalid_values:
552
+ value: false
553
+ remove_unused_columns:
554
+ value: false
555
+ repetition_penalty:
556
+ value: 1
557
+ report_to:
558
+ value:
559
+ - wandb
560
+ restore_callback_states_from_checkpoint:
561
+ value: false
562
+ resume_from_checkpoint:
563
+ value: null
564
+ resume_path:
565
+ value: ./checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute
566
+ return_dict:
567
+ value: true
568
+ return_dict_in_generate:
569
+ value: false
570
+ run_name:
571
+ value: dreamzero_real_teleop_g1_full_finetune_absolute
572
+ save_on_each_node:
573
+ value: false
574
+ save_only_model:
575
+ value: false
576
+ save_safetensors:
577
+ value: true
578
+ save_steps:
579
+ value: 8000
580
+ save_strategy:
581
+ value: steps
582
+ save_total_limit:
583
+ value: 10
584
+ seed:
585
+ value: 42
586
+ sep_token_id:
587
+ value: null
588
+ skip_memory_metrics:
589
+ value: true
590
+ suppress_tokens:
591
+ value: null
592
+ task_specific_params:
593
+ value: null
594
+ temperature:
595
+ value: 1
596
+ tf_legacy_loss:
597
+ value: false
598
+ tf32:
599
+ value: true
600
+ tie_encoder_decoder:
601
+ value: false
602
+ tie_word_embeddings:
603
+ value: true
604
+ tokenizer_class:
605
+ value: null
606
+ top_k:
607
+ value: 50
608
+ top_p:
609
+ value: 1
610
+ torch_compile:
611
+ value: false
612
+ torch_compile_backend:
613
+ value: null
614
+ torch_compile_mode:
615
+ value: null
616
+ torch_dtype:
617
+ value: null
618
+ torch_empty_cache_steps:
619
+ value: null
620
+ torchdynamo:
621
+ value: null
622
+ torchscript:
623
+ value: false
624
+ tp_size:
625
+ value: 0
626
+ tpu_metrics_debug:
627
+ value: false
628
+ tpu_num_cores:
629
+ value: null
630
+ transformers_version:
631
+ value: 4.51.3
632
+ typical_p:
633
+ value: 1
634
+ use_bfloat16:
635
+ value: false
636
+ use_cpu:
637
+ value: false
638
+ use_ipex:
639
+ value: false
640
+ use_legacy_prediction_loop:
641
+ value: false
642
+ use_liger_kernel:
643
+ value: false
644
+ use_mps_device:
645
+ value: false
646
+ warmup_ratio:
647
+ value: 0.05
648
+ warmup_steps:
649
+ value: 0
650
+ weight_decay:
651
+ value: 1e-05
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/run-20260330_041509-ao6lob27/files/output.log CHANGED
The diff for this file is too large to render. See raw diff
 
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/run-20260330_041509-ao6lob27/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb":{"runtime":209358},"_step":27600,"train/dynamics_loss_avg":0.047976726107299326,"train/grad_norm":0.745741605758667,"_timestamp":1.7750533435776424e+09,"train_samples_per_second":0.765,"train_loss":0.04518256734460592,"train/training_step_time":16.269297122955322,"train/loss":0.0696,"train_runtime":209234.0757,"train/learning_rate":6.834906085551041e-14,"total_flos":0,"train_steps_per_second":0.096,"train/global_step":20000,"train/epoch":1.1444091796875e-05,"train/model_forward_time":1.6608974933624268,"train/action_loss_avg":0.02277223765850067,"_runtime":209358}
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/run-20260330_041509-ao6lob27/logs/debug-core.log CHANGED
@@ -5,3 +5,11 @@
5
  {"time":"2026-03-30T04:15:10.145622128Z","level":"INFO","msg":"handleInformInit: received","streamId":"ao6lob27","id":"1(@)"}
6
  {"time":"2026-03-30T04:15:10.413378001Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"ao6lob27","id":"1(@)"}
7
  {"time":"2026-03-30T04:15:15.935635806Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"x5vnbp8x770d"}
 
 
 
 
 
 
 
 
 
5
  {"time":"2026-03-30T04:15:10.145622128Z","level":"INFO","msg":"handleInformInit: received","streamId":"ao6lob27","id":"1(@)"}
6
  {"time":"2026-03-30T04:15:10.413378001Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"ao6lob27","id":"1(@)"}
7
  {"time":"2026-03-30T04:15:15.935635806Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"x5vnbp8x770d"}
8
+ {"time":"2026-04-01T14:24:28.744616233Z","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
9
+ {"time":"2026-04-01T14:24:28.746553349Z","level":"INFO","msg":"connection: closing","id":"1(@)"}
10
+ {"time":"2026-04-01T14:24:28.74663126Z","level":"INFO","msg":"server is shutting down"}
11
+ {"time":"2026-04-01T14:24:28.74666882Z","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
12
+ {"time":"2026-04-01T14:24:28.750389742Z","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-977089-981034-2917860159/socket","Net":"unix"}}
13
+ {"time":"2026-04-01T14:24:29.858907981Z","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
14
+ {"time":"2026-04-01T14:24:29.858971311Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
15
+ {"time":"2026-04-01T14:24:29.858997432Z","level":"INFO","msg":"server is closed"}
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/run-20260330_041509-ao6lob27/logs/debug-internal.log CHANGED
@@ -5,3 +5,37 @@
5
  {"time":"2026-03-30T04:15:10.413395221Z","level":"INFO","msg":"writer: started","stream_id":"ao6lob27"}
6
  {"time":"2026-03-30T04:15:10.413394431Z","level":"INFO","msg":"sender: started","stream_id":"ao6lob27"}
7
  {"time":"2026-03-30T22:30:11.007336848Z","level":"INFO","msg":"api: retrying HTTP error","status":502,"url":"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream","body":"\n<html><head>\n<meta http-equiv=\"content-type\" content=\"text/html;charset=utf-8\">\n<title>502 Server Error</title>\n</head>\n<body text=#000000 bgcolor=#ffffff>\n<h1>Error: Server Error</h1>\n<h2>The server encountered a temporary error and could not complete your request.<p>Please try again in 30 seconds.</h2>\n<h2></h2>\n</body></html>\n"}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  {"time":"2026-03-30T04:15:10.413395221Z","level":"INFO","msg":"writer: started","stream_id":"ao6lob27"}
6
  {"time":"2026-03-30T04:15:10.413394431Z","level":"INFO","msg":"sender: started","stream_id":"ao6lob27"}
7
  {"time":"2026-03-30T22:30:11.007336848Z","level":"INFO","msg":"api: retrying HTTP error","status":502,"url":"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream","body":"\n<html><head>\n<meta http-equiv=\"content-type\" content=\"text/html;charset=utf-8\">\n<title>502 Server Error</title>\n</head>\n<body text=#000000 bgcolor=#ffffff>\n<h1>Error: Server Error</h1>\n<h2>The server encountered a temporary error and could not complete your request.<p>Please try again in 30 seconds.</h2>\n<h2></h2>\n</body></html>\n"}
8
+ {"time":"2026-04-01T10:18:44.020890295Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": read tcp 192.168.0.3:41808->35.186.228.49:443: read: connection timed out"}
9
+ {"time":"2026-04-01T10:18:56.087878557Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:38712->127.0.0.53:53: i/o timeout"}
10
+ {"time":"2026-04-01T10:19:10.908212851Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:39469->127.0.0.53:53: i/o timeout"}
11
+ {"time":"2026-04-01T10:19:30.190045118Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:47884->127.0.0.53:53: i/o timeout"}
12
+ {"time":"2026-04-01T10:19:58.216957187Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:34560->127.0.0.53:53: i/o timeout"}
13
+ {"time":"2026-04-01T10:20:40.445735516Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:46537->127.0.0.53:53: i/o timeout"}
14
+ {"time":"2026-04-01T10:21:50.450660421Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:39132->127.0.0.53:53: i/o timeout"}
15
+ {"time":"2026-04-01T10:23:00.454344979Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:60808->127.0.0.53:53: i/o timeout"}
16
+ {"time":"2026-04-01T10:24:10.457186285Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:42899->127.0.0.53:53: i/o timeout"}
17
+ {"time":"2026-04-01T10:25:20.462333271Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:51640->127.0.0.53:53: i/o timeout"}
18
+ {"time":"2026-04-01T10:26:30.466228733Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:54385->127.0.0.53:53: i/o timeout"}
19
+ {"time":"2026-04-01T10:27:40.469992011Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:36994->127.0.0.53:53: i/o timeout"}
20
+ {"time":"2026-04-01T10:28:50.474794045Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:59340->127.0.0.53:53: i/o timeout"}
21
+ {"time":"2026-04-01T10:30:00.479410672Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:53914->127.0.0.53:53: i/o timeout"}
22
+ {"time":"2026-04-01T10:31:10.483284091Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:52683->127.0.0.53:53: i/o timeout"}
23
+ {"time":"2026-04-01T10:32:20.487788622Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:52313->127.0.0.53:53: i/o timeout"}
24
+ {"time":"2026-04-01T10:33:30.492784858Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:38331->127.0.0.53:53: i/o timeout"}
25
+ {"time":"2026-04-01T10:34:40.497056663Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:59696->127.0.0.53:53: i/o timeout"}
26
+ {"time":"2026-04-01T10:35:50.500859105Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:40973->127.0.0.53:53: i/o timeout"}
27
+ {"time":"2026-04-01T10:37:00.5054549Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:45222->127.0.0.53:53: i/o timeout"}
28
+ {"time":"2026-04-01T10:38:10.508535232Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:55362->127.0.0.53:53: i/o timeout"}
29
+ {"time":"2026-04-01T10:39:20.512300509Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:50089->127.0.0.53:53: i/o timeout"}
30
+ {"time":"2026-04-01T10:40:30.517197984Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:34388->127.0.0.53:53: i/o timeout"}
31
+ {"time":"2026-04-01T10:41:40.521767935Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:53489->127.0.0.53:53: i/o timeout"}
32
+ {"time":"2026-04-01T10:42:50.525040035Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:50726->127.0.0.53:53: i/o timeout"}
33
+ {"time":"2026-04-01T10:44:00.527917425Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:53434->127.0.0.53:53: i/o timeout"}
34
+ {"time":"2026-04-01T10:45:10.532222422Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:37341->127.0.0.53:53: i/o timeout"}
35
+ {"time":"2026-04-01T10:46:20.535679123Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:37921->127.0.0.53:53: i/o timeout"}
36
+ {"time":"2026-04-01T10:47:30.539549836Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/boqianli/dreamzero/ao6lob27/file_stream\": dial tcp: lookup api.wandb.ai on 127.0.0.53:53: read udp 127.0.0.1:38214->127.0.0.53:53: i/o timeout"}
37
+ {"time":"2026-04-01T14:24:28.755750998Z","level":"INFO","msg":"stream: closing","id":"ao6lob27"}
38
+ {"time":"2026-04-01T14:24:29.342449946Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
39
+ {"time":"2026-04-01T14:24:29.828344988Z","level":"INFO","msg":"handler: closed","stream_id":"ao6lob27"}
40
+ {"time":"2026-04-01T14:24:29.82853025Z","level":"INFO","msg":"sender: closed","stream_id":"ao6lob27"}
41
+ {"time":"2026-04-01T14:24:29.8285597Z","level":"INFO","msg":"stream: closed","id":"ao6lob27"}
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/run-20260330_041509-ao6lob27/logs/debug.log CHANGED
@@ -20,3 +20,5 @@ config: {'_wandb': {}}
20
  2026-03-30 04:15:10,932 INFO MainThread:977089 [wandb_run.py:_config_callback():1403] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': None, 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': None, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute', '_attn_implementation_autoset': True, 'transformers_version': '4.51.3', 'model_dtype': 'float32', 'hidden_size': 0, 'action_horizon': 24, 'action_dim': 36, 'backbone_cfg': {'_target_': 'groot.vla.model.dreamzero.backbone.identity.IdentityBackbone'}, 'action_head_cfg': {'config': {'backbone_features_projector_cfg': None, '_target_': 'groot.vla.model.dreamzero.action_head.wan_flow_matching_action_tf.WANPolicyHeadConfig', '_recursive_': False, 'tiled': False, 'tile_size_height': 34, 'tile_size_width': 34, 'tile_stride_height': 18, 'tile_stride_width': 16, 'lora_rank': 4, 'lora_alpha': 4, 'num_frames': 33, 'num_frame_per_block': 2, 'lora_target_modules': 'q,k,v,o,ffn.0,ffn.2', 'init_lora_weights': 'kaiming', 'train_architecture': 'full', 'use_gradient_checkpointing': True, 'add_pos_embed': True, 'model_dtype': 'float32', 'max_state_dim': 64, 'max_action_dim': 36, 'action_loss_embodiment_ids': [26, 17, 32], 'hidden_size': 64, 'input_embedding_dim': 1536, 'backbone_embedding_dim': 0, 'repa_layer': 8, 'repa_coeff': 1.0, 'load_pretrained_det_decode_layer_path': None, 'freeze_decode_layer': False, 'expand_batch': None, 'use_vlln': True, 'vl_self_attention_cfg': {'_target_': 'groot.vla.model.n1_5.modules.cross_attention_dit.SelfAttentionTransformer', 'positional_embeddings': None, 'num_layers': 4, 'num_attention_heads': 24, 'attention_head_dim': 64, 'dropout': 0.2, 'final_dropout': True}, 'diffusion_model_cfg': {'_target_': 'groot.vla.model.dreamzero.modules.wan_video_dit_action_casual_chunk.CausalWanModel', '_convert_': 'object', 'diffusion_model_pretrained_path': '/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P', 'model_type': 'i2v', 'frame_seqlen': 220, 'dim': 5120, 'in_dim': 36, 'ffn_dim': 13824, 'out_dim': 16, 'freq_dim': 256, 'eps': 1e-06, 'num_heads': 40, 'num_layers': 40, 'max_chunk_size': 4, 'num_frame_per_block': 2, 'num_action_per_block': 24, 'num_state_per_block': 1, 'action_dim': 36}, 'text_encoder_cfg': {'_target_': 'groot.vla.model.dreamzero.modules.wan_video_text_encoder.WanTextEncoder', '_convert_': 'object', 'text_encoder_pretrained_path': '/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_t5_umt5-xxl-enc-bf16.pth'}, 'image_encoder_cfg': {'_target_': 'groot.vla.model.dreamzero.modules.wan_video_image_encoder.WanImageEncoder', '_convert_': 'object', 'image_encoder_pretrained_path': '/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_clip_open-clip-xlm-roberta-large-vit-huge-14.pth'}, 'vae_cfg': {'_target_': 'groot.vla.model.dreamzero.modules.wan_video_vae.WanVideoVAE', '_convert_': 'object', 'vae_pretrained_path': '/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/Wan2.1_VAE.pth'}, 'action_dim': 36, 'action_horizon': 24, 'num_inference_timesteps': 4, 'noise_beta_alpha': 1.5, 'noise_beta_beta': 1.0, 'noise_s': 0.999, 'num_timestep_buckets': 1000, 'decouple_video_action_noise': False, 'video_noise_beta_alpha': 3.0, 'video_noise_beta_beta': 1.0, 'tune_projector': True, 'tune_diffusion_model': True, 'skip_component_loading': True}, '_target_': 'groot.vla.model.dreamzero.action_head.wan_flow_matching_action_tf.WANPolicyHead', '_convert_': 'object'}, 'resume_path': './checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute', 'model_type': 'vla', 'output_dir': './checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'eval_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 1e-05, 'weight_decay': 1e-05, 'adam_beta1': 0.95, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1000, 'max_steps': 20000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.05, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute/runs/Mar30_04-04-13_nebula100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 8000, 'save_total_limit': 10, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': True, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 1, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'dreamzero_real_teleop_g1_full_finetune_absolute', 'disable_tqdm': False, 'remove_unused_columns': False, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': True, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'tp_size': 0, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'groot/vla/configs/deepspeed/zero2_offload.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': 100, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': False, 'dataloader_persistent_workers': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': None, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'include_for_metrics': [], 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'use_liger_kernel': False, 'eval_use_gather_object': False, 'average_tokens_across_devices': False}
21
  2026-03-30 04:15:10,939 INFO MainThread:977089 [wandb_config.py:__setitem__():155] [no run ID] config set model/num_parameters = 22924196696 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7b9f0d23f190>>
22
  2026-03-30 04:15:10,939 INFO MainThread:977089 [wandb_run.py:_config_callback():1403] config_cb model/num_parameters 22924196696 None
 
 
 
20
  2026-03-30 04:15:10,932 INFO MainThread:977089 [wandb_run.py:_config_callback():1403] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': None, 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': None, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute', '_attn_implementation_autoset': True, 'transformers_version': '4.51.3', 'model_dtype': 'float32', 'hidden_size': 0, 'action_horizon': 24, 'action_dim': 36, 'backbone_cfg': {'_target_': 'groot.vla.model.dreamzero.backbone.identity.IdentityBackbone'}, 'action_head_cfg': {'config': {'backbone_features_projector_cfg': None, '_target_': 'groot.vla.model.dreamzero.action_head.wan_flow_matching_action_tf.WANPolicyHeadConfig', '_recursive_': False, 'tiled': False, 'tile_size_height': 34, 'tile_size_width': 34, 'tile_stride_height': 18, 'tile_stride_width': 16, 'lora_rank': 4, 'lora_alpha': 4, 'num_frames': 33, 'num_frame_per_block': 2, 'lora_target_modules': 'q,k,v,o,ffn.0,ffn.2', 'init_lora_weights': 'kaiming', 'train_architecture': 'full', 'use_gradient_checkpointing': True, 'add_pos_embed': True, 'model_dtype': 'float32', 'max_state_dim': 64, 'max_action_dim': 36, 'action_loss_embodiment_ids': [26, 17, 32], 'hidden_size': 64, 'input_embedding_dim': 1536, 'backbone_embedding_dim': 0, 'repa_layer': 8, 'repa_coeff': 1.0, 'load_pretrained_det_decode_layer_path': None, 'freeze_decode_layer': False, 'expand_batch': None, 'use_vlln': True, 'vl_self_attention_cfg': {'_target_': 'groot.vla.model.n1_5.modules.cross_attention_dit.SelfAttentionTransformer', 'positional_embeddings': None, 'num_layers': 4, 'num_attention_heads': 24, 'attention_head_dim': 64, 'dropout': 0.2, 'final_dropout': True}, 'diffusion_model_cfg': {'_target_': 'groot.vla.model.dreamzero.modules.wan_video_dit_action_casual_chunk.CausalWanModel', '_convert_': 'object', 'diffusion_model_pretrained_path': '/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P', 'model_type': 'i2v', 'frame_seqlen': 220, 'dim': 5120, 'in_dim': 36, 'ffn_dim': 13824, 'out_dim': 16, 'freq_dim': 256, 'eps': 1e-06, 'num_heads': 40, 'num_layers': 40, 'max_chunk_size': 4, 'num_frame_per_block': 2, 'num_action_per_block': 24, 'num_state_per_block': 1, 'action_dim': 36}, 'text_encoder_cfg': {'_target_': 'groot.vla.model.dreamzero.modules.wan_video_text_encoder.WanTextEncoder', '_convert_': 'object', 'text_encoder_pretrained_path': '/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_t5_umt5-xxl-enc-bf16.pth'}, 'image_encoder_cfg': {'_target_': 'groot.vla.model.dreamzero.modules.wan_video_image_encoder.WanImageEncoder', '_convert_': 'object', 'image_encoder_pretrained_path': '/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/models_clip_open-clip-xlm-roberta-large-vit-huge-14.pth'}, 'vae_cfg': {'_target_': 'groot.vla.model.dreamzero.modules.wan_video_vae.WanVideoVAE', '_convert_': 'object', 'vae_pretrained_path': '/hfm/boqian/liboqian_data/checkpoints/Wan2.1-I2V-14B-480P/Wan2.1_VAE.pth'}, 'action_dim': 36, 'action_horizon': 24, 'num_inference_timesteps': 4, 'noise_beta_alpha': 1.5, 'noise_beta_beta': 1.0, 'noise_s': 0.999, 'num_timestep_buckets': 1000, 'decouple_video_action_noise': False, 'video_noise_beta_alpha': 3.0, 'video_noise_beta_beta': 1.0, 'tune_projector': True, 'tune_diffusion_model': True, 'skip_component_loading': True}, '_target_': 'groot.vla.model.dreamzero.action_head.wan_flow_matching_action_tf.WANPolicyHead', '_convert_': 'object'}, 'resume_path': './checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute', 'model_type': 'vla', 'output_dir': './checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'eval_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 1e-05, 'weight_decay': 1e-05, 'adam_beta1': 0.95, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1000, 'max_steps': 20000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.05, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './checkpoints/dreamzero_real_teleop_g1_full_finetune_absolute/runs/Mar30_04-04-13_nebula100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 8000, 'save_total_limit': 10, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': True, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 1, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'dreamzero_real_teleop_g1_full_finetune_absolute', 'disable_tqdm': False, 'remove_unused_columns': False, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': True, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'tp_size': 0, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'groot/vla/configs/deepspeed/zero2_offload.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': 100, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': False, 'dataloader_persistent_workers': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': None, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'include_for_metrics': [], 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'use_liger_kernel': False, 'eval_use_gather_object': False, 'average_tokens_across_devices': False}
21
  2026-03-30 04:15:10,939 INFO MainThread:977089 [wandb_config.py:__setitem__():155] [no run ID] config set model/num_parameters = 22924196696 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7b9f0d23f190>>
22
  2026-03-30 04:15:10,939 INFO MainThread:977089 [wandb_run.py:_config_callback():1403] config_cb model/num_parameters 22924196696 None
23
+ 2026-04-01 14:24:28,737 INFO wandb-AsyncioManager-main:977089 [service_client.py:_forward_responses():134] Reached EOF.
24
+ 2026-04-01 14:24:28,738 INFO wandb-AsyncioManager-main:977089 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.
dreamzero_real_teleop_g1_full_finetune_absolute_Pick_bottle_and_turn_and_pour_into_cup/wandb/run-20260330_041509-ao6lob27/run-ao6lob27.wandb CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d9a7c8981c4b92aea8ed1d6f0294eea264e6dad8984cdf7b6f78537488b5906b
3
- size 94961664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b42c924544506ffebf5d1a80c95d92dd18fa643ba9d19630056da6d28a5df412
3
+ size 116007394