| { |
| "run_id": "0223_QwenLatent_real4tasks_predict_only_2k", |
| "run_root_dir": "./runs", |
| "seed": 42, |
| "trackers": [ |
| "jsonl", |
| "wandb" |
| ], |
| "wandb_entity": "timsty", |
| "wandb_project": "vla_jepa", |
| "is_debug": false, |
| "framework": { |
| "name": "QwenLatent", |
| "qwenvl": { |
| "base_vlm": "/mnt/data/fangyu/model/Qwen/Qwen3-VL-2B-Instruct", |
| "attn_implementation": "flash_attention_2", |
| "vl_hidden_dim": 2048, |
| "num_data_tokens": 32 |
| }, |
| "action_model": { |
| "ckpt_path": null, |
| "action_size": 37, |
| "state_size": 88, |
| "hidden_size": 1024, |
| "intermediate_size": 3072, |
| "dataset_vocab_size": 256, |
| "num_data_tokens": 32, |
| "min_action_len": 5, |
| "num_encoder_layers": 28, |
| "num_decoder_layers": 28, |
| "num_attention_heads": 16, |
| "num_key_value_heads": 8, |
| "head_dim": 128, |
| "max_position_embeddings": 2048, |
| "max_action_chunk_size": 50, |
| "rms_norm_eps": 1e-06, |
| "attention_dropout": 0.0, |
| "use_vae_reparameterization": false, |
| "use_ema": false, |
| "chunk_size": "${datasets.vla_data.chunk_size}", |
| "use_contrastive_loss": false, |
| "contrastive_use_proj": false, |
| "use_domain_adversarial": false, |
| "loss_mode": "predict_only", |
| "qwen3_pretrained_name_or_path": "/mnt/data/fangyu/model/Qwen/Qwen3-0.6B" |
| } |
| }, |
| "datasets": { |
| "vla_data": { |
| "dataset_py": "lerobot_datasets", |
| "data_root_dir": "/mnt/data/fangyu/dataset/IPEC-COMMUNITY", |
| "data_mix": "real_world_4tasks", |
| "CoT_prompt": "Task: {instruction}. What are the next 15 actions to take?", |
| "default_image_resolution": [ |
| 3, |
| 224, |
| 224 |
| ], |
| "per_device_batch_size": 32, |
| "load_all_data_for_training": true, |
| "obs": [ |
| "image_0" |
| ], |
| "image_size": [ |
| 224, |
| 224 |
| ], |
| "video_backend": "torchcodec", |
| "load_video": true, |
| "chunk_size": 15, |
| "state_use_action_chunk": true, |
| "num_history_steps": 0, |
| "include_state": true |
| } |
| }, |
| "trainer": { |
| "epochs": 100, |
| "max_train_steps": 2000, |
| "num_warmup_steps": 200, |
| "num_stable_steps": 0, |
| "save_interval": 1000, |
| "eval_interval": 50, |
| "max_checkpoints_to_keep": 10, |
| "learning_rate": { |
| "base": 2.5e-05, |
| "qwen_vl_interface": 2.5e-05, |
| "action_model": 2.5e-05 |
| }, |
| "lr_scheduler_type": "warmup_stable_cosine", |
| "scheduler_specific_kwargs": { |
| "min_lr_ratio": 0.001 |
| }, |
| "freeze_modules": "", |
| "loss_scale": { |
| "align_loss": 1.0, |
| "recon_loss": 1.0, |
| "predict_loss": 1.0 |
| }, |
| "warmup_ratio": 0.1, |
| "weight_decay": 0.0, |
| "logging_frequency": 10, |
| "gradient_clipping": 5.0, |
| "gradient_accumulation_steps": 1, |
| "optimizer": { |
| "name": "AdamW", |
| "betas": [ |
| 0.9, |
| 0.95 |
| ], |
| "eps": 1e-08, |
| "weight_decay": 1e-08 |
| }, |
| "is_resume": false, |
| "resume_epoch": null, |
| "resume_step": null, |
| "enable_gradient_checkpointing": true, |
| "enable_mixed_precision_training": true |
| }, |
| "output_dir": "./runs/0223_QwenLatent_real4tasks_predict_only_2k" |
| } |