File size: 3,180 Bytes
32f4633
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
{
  "run_id": "0223_QwenLatent_real4tasks_predict_only_2k",
  "run_root_dir": "./runs",
  "seed": 42,
  "trackers": [
    "jsonl",
    "wandb"
  ],
  "wandb_entity": "timsty",
  "wandb_project": "vla_jepa",
  "is_debug": false,
  "framework": {
    "name": "QwenLatent",
    "qwenvl": {
      "base_vlm": "/mnt/data/fangyu/model/Qwen/Qwen3-VL-2B-Instruct",
      "attn_implementation": "flash_attention_2",
      "vl_hidden_dim": 2048,
      "num_data_tokens": 32
    },
    "action_model": {
      "ckpt_path": null,
      "action_size": 37,
      "state_size": 88,
      "hidden_size": 1024,
      "intermediate_size": 3072,
      "dataset_vocab_size": 256,
      "num_data_tokens": 32,
      "min_action_len": 5,
      "num_encoder_layers": 28,
      "num_decoder_layers": 28,
      "num_attention_heads": 16,
      "num_key_value_heads": 8,
      "head_dim": 128,
      "max_position_embeddings": 2048,
      "max_action_chunk_size": 50,
      "rms_norm_eps": 1e-06,
      "attention_dropout": 0.0,
      "use_vae_reparameterization": false,
      "use_ema": false,
      "chunk_size": "${datasets.vla_data.chunk_size}",
      "use_contrastive_loss": false,
      "contrastive_use_proj": false,
      "use_domain_adversarial": false,
      "loss_mode": "predict_only",
      "qwen3_pretrained_name_or_path": "/mnt/data/fangyu/model/Qwen/Qwen3-0.6B"
    }
  },
  "datasets": {
    "vla_data": {
      "dataset_py": "lerobot_datasets",
      "data_root_dir": "/mnt/data/fangyu/dataset/IPEC-COMMUNITY",
      "data_mix": "real_world_4tasks",
      "CoT_prompt": "Task: {instruction}. What are the next 15 actions to take?",
      "default_image_resolution": [
        3,
        224,
        224
      ],
      "per_device_batch_size": 32,
      "load_all_data_for_training": true,
      "obs": [
        "image_0"
      ],
      "image_size": [
        224,
        224
      ],
      "video_backend": "torchcodec",
      "load_video": true,
      "chunk_size": 15,
      "state_use_action_chunk": true,
      "num_history_steps": 0,
      "include_state": true
    }
  },
  "trainer": {
    "epochs": 100,
    "max_train_steps": 2000,
    "num_warmup_steps": 200,
    "num_stable_steps": 0,
    "save_interval": 1000,
    "eval_interval": 50,
    "max_checkpoints_to_keep": 10,
    "learning_rate": {
      "base": 2.5e-05,
      "qwen_vl_interface": 2.5e-05,
      "action_model": 2.5e-05
    },
    "lr_scheduler_type": "warmup_stable_cosine",
    "scheduler_specific_kwargs": {
      "min_lr_ratio": 0.001
    },
    "freeze_modules": "",
    "loss_scale": {
      "align_loss": 1.0,
      "recon_loss": 1.0,
      "predict_loss": 1.0
    },
    "warmup_ratio": 0.1,
    "weight_decay": 0.0,
    "logging_frequency": 10,
    "gradient_clipping": 5.0,
    "gradient_accumulation_steps": 1,
    "optimizer": {
      "name": "AdamW",
      "betas": [
        0.9,
        0.95
      ],
      "eps": 1e-08,
      "weight_decay": 1e-08
    },
    "is_resume": false,
    "resume_epoch": null,
    "resume_step": null,
    "enable_gradient_checkpointing": true,
    "enable_mixed_precision_training": true
  },
  "output_dir": "./runs/0223_QwenLatent_real4tasks_predict_only_2k"
}