File size: 3,180 Bytes
32f4633 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 | {
"run_id": "0223_QwenLatent_real4tasks_predict_only_2k",
"run_root_dir": "./runs",
"seed": 42,
"trackers": [
"jsonl",
"wandb"
],
"wandb_entity": "timsty",
"wandb_project": "vla_jepa",
"is_debug": false,
"framework": {
"name": "QwenLatent",
"qwenvl": {
"base_vlm": "/mnt/data/fangyu/model/Qwen/Qwen3-VL-2B-Instruct",
"attn_implementation": "flash_attention_2",
"vl_hidden_dim": 2048,
"num_data_tokens": 32
},
"action_model": {
"ckpt_path": null,
"action_size": 37,
"state_size": 88,
"hidden_size": 1024,
"intermediate_size": 3072,
"dataset_vocab_size": 256,
"num_data_tokens": 32,
"min_action_len": 5,
"num_encoder_layers": 28,
"num_decoder_layers": 28,
"num_attention_heads": 16,
"num_key_value_heads": 8,
"head_dim": 128,
"max_position_embeddings": 2048,
"max_action_chunk_size": 50,
"rms_norm_eps": 1e-06,
"attention_dropout": 0.0,
"use_vae_reparameterization": false,
"use_ema": false,
"chunk_size": "${datasets.vla_data.chunk_size}",
"use_contrastive_loss": false,
"contrastive_use_proj": false,
"use_domain_adversarial": false,
"loss_mode": "predict_only",
"qwen3_pretrained_name_or_path": "/mnt/data/fangyu/model/Qwen/Qwen3-0.6B"
}
},
"datasets": {
"vla_data": {
"dataset_py": "lerobot_datasets",
"data_root_dir": "/mnt/data/fangyu/dataset/IPEC-COMMUNITY",
"data_mix": "real_world_4tasks",
"CoT_prompt": "Task: {instruction}. What are the next 15 actions to take?",
"default_image_resolution": [
3,
224,
224
],
"per_device_batch_size": 32,
"load_all_data_for_training": true,
"obs": [
"image_0"
],
"image_size": [
224,
224
],
"video_backend": "torchcodec",
"load_video": true,
"chunk_size": 15,
"state_use_action_chunk": true,
"num_history_steps": 0,
"include_state": true
}
},
"trainer": {
"epochs": 100,
"max_train_steps": 2000,
"num_warmup_steps": 200,
"num_stable_steps": 0,
"save_interval": 1000,
"eval_interval": 50,
"max_checkpoints_to_keep": 10,
"learning_rate": {
"base": 2.5e-05,
"qwen_vl_interface": 2.5e-05,
"action_model": 2.5e-05
},
"lr_scheduler_type": "warmup_stable_cosine",
"scheduler_specific_kwargs": {
"min_lr_ratio": 0.001
},
"freeze_modules": "",
"loss_scale": {
"align_loss": 1.0,
"recon_loss": 1.0,
"predict_loss": 1.0
},
"warmup_ratio": 0.1,
"weight_decay": 0.0,
"logging_frequency": 10,
"gradient_clipping": 5.0,
"gradient_accumulation_steps": 1,
"optimizer": {
"name": "AdamW",
"betas": [
0.9,
0.95
],
"eps": 1e-08,
"weight_decay": 1e-08
},
"is_resume": false,
"resume_epoch": null,
"resume_step": null,
"enable_gradient_checkpointing": true,
"enable_mixed_precision_training": true
},
"output_dir": "./runs/0223_QwenLatent_real4tasks_predict_only_2k"
} |