# DynaTraj — Training and Data Collection (SB3 + dm_control) This repo provides two entry-point scripts: - `sb3_train.py`: Train an RL policy with Stable-Baselines3 on dm_control tasks and periodically save policy checkpoints - `sb3_collect.py`: Load specific checkpoints and collect fixed-length trajectories into `.npz` datasets with minimal metadata --- ## Environment & Dependencies - Python 3.10+ - Required: - dm_control + MuJoCo: `pip install dm-control mujoco` - Reinforcement learning: `pip install stable-baselines3` - Utilities: `pip install numpy tqdm torch` - Gym API: prefers `gymnasium`, falls back to `gym` (handled automatically) - Optional (only if you need on-screen rendering during collection): - OpenCV UI: `pip install opencv-python` - or Matplotlib: `pip install matplotlib` Note: On headless Linux/remote environments, rendering might require EGL/OSMesa configuration. If you do not render, you do not need any display backend. --- ## Output Layout (Conventions) - Training checkpoints: `weights///ckpt-.pt` - `k` starts from 1; default save interval is every 10,000 timesteps - Examples: `ckpt-1.pt` ≈ 10k steps, `ckpt-2.pt` ≈ 20k steps - Full SB3 model: `weights/sb3__-_seed_.zip` - Collected datasets: `dataset/*.npz` with a companion `*_metadata.pkl` Default absolute paths in this repo: - Checkpoint root: `/home/lau/sim/DynaTraj/weights` - Dataset output: `/home/lau/sim/DynaTraj/dataset` --- ## Training Script: `sb3_train.py` Supported algorithms (argument is case-sensitive and must be lowercase here): - `sac`, `ppo`, `td3` Common args: - `--domain`: dm_control domain, e.g., `cheetah`, `quadruped` (default: `cheetah`) - `--task`: task name, e.g., `run`, `walk` (default: `run`) - `--algo`: `sac|ppo|td3` (default: `sac`) - `--total_timesteps`: total training steps (default: `500000`) - `--n_envs`: number of parallel envs (default: `1`; uses sub-process vectorization if >1) - `--seed`: random seed (default: `0`) - `--device`: `cpu|cuda|auto` (default: `auto`) - `--out_dir`: where to save models & checkpoints (default: `/home/lau/sim/DynaTraj/weights`) Example: ```bash python /home/lau/sim/DynaTraj/train_sb3_dmcontrol.py \ --domain cheetah --task run\ --algo sac \ --total_timesteps 500000 \ --out_dir /home/lau/sim/DynaTraj/weights ``` Checkpointing: - A policy checkpoint is saved every 10,000 global timesteps as `ckpt-.pt` (1-based counter). - A full SB3 model `.zip` is also saved at the end for plain SB3 loading if needed. --- ## Data Collection Script: `sb3_collect.py` Purpose: Run inference with specific training checkpoints and write fixed-length trajectories to `.npz` files plus metadata. Common args: - `--domain` / `--task`: must match training (defaults: `cheetah` / `run`) - `--ckpt_root`: root directory of checkpoints (default: `/home/lau/sim/DynaTraj/weights`) - `--ckpt_indices`: comma-separated checkpoint indices, e.g., `1,5,10` (note: starts from 1) - `--trajectories_per_ckpt`: how many trajectories per checkpoint (default: `5120`) - `--steps_per_trajectory`: steps per trajectory (default: `24`) - `--out_dir`: dataset output directory (default: `/home/lau/sim/DynaTraj/dataset`) - `--device`: inference device (default: `cpu`) - `--render`: optional flag to render frames (requires OpenCV or Matplotlib) - `--algo`: usually unnecessary. The script reads the real algo name from the checkpoint payload. Only used as a fallback (must be UPPERCASE: `SAC|PPO|TD3`). Example: ```bash python sb3_collect.py \ --domain cheetah --task run \ --algo SAC \ --ckpt_root ./weights \ --ckpt_indices 1,20,30,50 \ --trajectories_per_ckpt 1024 \ --steps_per_trajectory 512 \ --out_dir ./dataset \ --device cpu ``` With rendering (optional): ```bash python /home/lau/sim/DynaTraj/sb3_collect.py \ --domain cheetah --task run \ --algo SAC \ --ckpt_root /home/lau/sim/DynaTraj/weights \ --ckpt_indices 50 \ --trajectories_per_ckpt 5120 \ --steps_per_trajectory 24 \ --out_dir /home/lau/sim/DynaTraj/dataset \ --device cpu \ --render ``` For Bouncing ball: python bb_collect.py --trajectories 1024 --steps_per_trajectory 8192 Notes: - The script searches checkpoints under `ckpt_root///` as `ckpt-.pt`. If you see an error for `ckpt-0.pt`, switch to 1-based indices. - If your environment has no display backend, simply omit `--render`. --- ## Output Format For each `k`: - Dataset file: `/sb3___ckpt_.npz` - Metadata file: `/sb3___ckpt__metadata.pkl` Metadata keys include: - `domain`, `task`, `algo`, `seed` - `ckpt_index` (the `k` you collected) - `trajectories_per_ckpt`, `steps_per_trajectory` - `total_trajectories`, `total_steps` - `render` (whether rendering was enabled) The exact arrays inside `.npz` are defined by the internal `TrajectoryBuffer` implementation. Typically they include time-aligned observation/state tensors, actions, rewards, and done flags. --- ## Data Format ### Transition: a_t = pi(s_t) s_t+1 = f(s_t, a_t) ### 1.Cheetah - state - qpos_t(x,z,pitch,joint) = 9 - qvel_t(x,z,pitch,joint) = 9 - action - tau_t(joint) = 6 ## Tips - Prefer absolute paths (all examples above use absolute paths). - Training uses `DummyVecEnv`/`SubprocVecEnv`, automatically flattens dm_control observations and clips actions to env bounds. - Collection reconstructs an SB3 policy and loads the `policy_state_dict` from each checkpoint; it does not require the full `.zip` model. - Algo-arg case: training expects lowercase (`sac|ppo|td3`), collection fallback uses UPPERCASE (`SAC|PPO|TD3`).