memvla-bridge / config.json
shihao1895's picture
Add files using upload-large-folder tool
86df5b1 verified
{
"action_dim": 7,
"action_model_type": "DiT-L",
"consolidate_type": "tome",
"data_root_dir": "./data/openx_256",
"dataloader_type": "stream",
"fusion_type": "gate",
"future_action_window_size": 15,
"group_size": 16,
"hf_token": "hf_xxx",
"image_aug": true,
"is_resume": false,
"load_all_data_for_training": true,
"mem_length": 16,
"per_token_size": 256,
"pretrained_checkpoint": "./pretrained/CogACT-Large/checkpoints/CogACT-Large.pt",
"repeated_diffusion_steps": 4,
"resume_epoch": 0,
"resume_step": 0,
"retrieval_layers": 2,
"run_id": "memvla-bridge",
"run_id_note": null,
"run_root_dir": "./log/bridge",
"save_interval": 2500,
"seed": 42,
"trackers": [
"jsonl",
"wandb"
],
"update_fused": false,
"use_ema": false,
"use_timestep_pe": true,
"vla": {
"base_vlm": "prism-dinosiglip-224px+7b",
"data_mix": "bridge",
"enable_gradient_checkpointing": true,
"enable_mixed_precision_training": true,
"epochs": 100,
"expected_world_size": 8,
"freeze_llm_backbone": false,
"freeze_vision_backbone": false,
"global_batch_size": 256,
"learning_rate": 2e-05,
"lr_scheduler_type": "constant",
"max_grad_norm": 1.0,
"max_steps": 50000,
"per_device_batch_size": 32,
"reduce_in_full_precision": true,
"shuffle_buffer_size": 128000,
"train_strategy": "fsdp-full-shard",
"type": "prism-dinosiglip-224px+oxe+diffusion",
"unfreeze_last_llm_layer": false,
"vla_id": "prism-dinosiglip-224px+oxe+diffusion",
"warmup_ratio": 0.0,
"weight_decay": 0.0
},
"wandb_entity": "shihao-thu",
"wandb_project": "memvla"
}