zhenyuzhao's picture
Add files using upload-large-folder tool
aceb411 verified
common:
# The number of historical images
img_history_size: 1
# The number of historical point clouds
pcd_history_size: 1
# The number of future actions to predict
action_chunk_size: 16
# The number of cameras to be used in the model
num_cameras: 3
# Dimension for state
state_dim: 14
# Dimension for action
action_dim: 14
# The number of patches in the image
num_patches: 196
dataset:
# We will extract the data from raw dataset
# and store them in the disk buffer by producer
# When training, we will read the data
# randomly from the buffer by consumer
# The producer will replace the data which has been
# read by the consumer with new data
# The path to the buffer (at least 400GB)
buf_path: /ssd/lingxuan/data/buffer
# The number of chunks in the buffer
buf_num_chunks: 512
# The number of samples (step rather than episode) in each chunk
buf_chunk_size: 512
# We will filter the episodes with length less than `epsd_len_thresh_low`
epsd_len_thresh_low: 32
# For those more than `epsd_len_thresh_high`,
# we will randomly sample `epsd_len_thresh_high` steps each time we load the episode
# to better balance the training datasets
epsd_len_thresh_high: 2048
# How to fit the image size
image_aspect_ratio: pad
# Maximum number of language tokens
tokenizer_max_length: 1024
model:
# Config for condition adpators
act_adaptor: mlp3x_silu # Will be reinitialized in finetune mode
st_adaptor: mlp3x_silu # Will be reinitialized in finetune mode
img_adapter: mlp2x_silu # Shared between pretrain and finetune
lang_adapter: mlp2x_silu # Shared between pretrain and finetune
# Config for H-RDT structure (backbone - shared between pretrain and finetune)
hrdt:
hidden_size: 2176
depth: 16
num_heads: 16
norm_eps: 0.00001
# make SwiGLU hidden layer size multiple of large power of 2
multiple_of: 256
ffn_dim_multiplier: null
# Grouped Query Attention
num_kv_heads: 8
# output_size: ${...common.action_dim} # i.e., action dimension (TODO)
output_size: 14
use_flash_attn: true
# For noise scheduler (flow matching)
noise_scheduler:
num_inference_timesteps: 5
timestep_max: 0.999
sampler_type: uniform
time_noise:
a: 5
beta_m: 100
# For EMA (params averaging)
# We do not use EMA currently
ema:
update_after_step: 0
inv_gamma: 1.0
power: 0.75
min_value: 0.0
max_value: 0.9999
# Encoder configurations
vision:
feature_dim: 2176
text:
feature_dim: 4096