SO-101-Project / train_example.py
Jainam03's picture
Uploade robotics project
9c25ee7 verified
#!/usr/bin/env python3
"""
Train LeRobot imitation-learning policies on an SO-101 dataset with W&B logging.
Install:
pip install lerobot huggingface_hub wandb torch pyyaml
pip install 'lerobot[diffusion]' # needed for diffusion policy configs
Login once:
huggingface-cli login # optional for public dataset, useful generally
wandb login
Run:
python train_example.py --config-name act_arch_wide_192
python train_example.py --config-name act_arch_wide_192 \
--dataset-repos user/dataset_a user/dataset_b --aggregate-repo user/a_b_merged
python train_example.py --config-name act_arch_wide_192_hil_finetune
Fine-tuning:
Set `pretrained_policy_path` to a previous `pretrained_model` directory. This
starts a new run initialized from that checkpoint. It is different from
LeRobot `resume`, which continues the exact same interrupted run.
"""
from __future__ import annotations
import argparse
import copy
import hashlib
import os
import shutil
import subprocess
from pathlib import Path
from typing import Any
DEFAULT_DATASET_REPO = "Smencomojica/robotics_class_2"
DEFAULT_CONFIG_FILE = Path(__file__).with_name("train_configs.yaml")
def default_lerobot_home() -> Path:
hf_home = Path(os.environ.get("HF_HOME", Path.home() / ".cache" / "huggingface"))
return Path(os.environ.get("HF_LEROBOT_HOME", hf_home / "lerobot"))
def dataset_local_dir(repo_id: str) -> Path:
return default_lerobot_home() / repo_id
def default_aggregate_repo_id(dataset_repos: list[str]) -> str:
digest = hashlib.sha1("\n".join(dataset_repos).encode("utf-8")).hexdigest()[:8]
return f"local/merged_{digest}"
def looks_downloaded(path: Path) -> bool:
return (
path.exists()
and (path / "meta").exists()
and any(path.rglob("*.parquet"))
)
def normalize_dataset_repos(value: Any) -> list[str]:
if value is None:
return []
if isinstance(value, str):
return [value]
if isinstance(value, list) and all(isinstance(item, str) for item in value):
return value
raise ValueError("`dataset_repos` must be a string or a list of strings.")
def normalize_string_list(value: Any, field_name: str) -> list[str]:
if value is None:
return []
if isinstance(value, str):
return [value]
if isinstance(value, list) and all(isinstance(item, str) for item in value):
return value
raise ValueError(f"`{field_name}` must be a string or a list of strings.")
def pick_device() -> str:
import torch
if torch.cuda.is_available():
return "cuda"
if getattr(torch.backends, "mps", None) and torch.backends.mps.is_available():
return "mps"
return "cpu"
def resolve_config_file(path: str | Path) -> Path:
config_file = Path(path)
if config_file.exists():
return config_file
script_relative = Path(__file__).parent / config_file
if script_relative.exists():
return script_relative
raise FileNotFoundError(f"Could not find config file: {path}")
def load_sweep_configs(config_file: Path) -> dict[str, Any]:
try:
import yaml
except ModuleNotFoundError as exc:
raise RuntimeError(
"PyYAML is required for --config-file support. Install it with `pip install pyyaml`."
) from exc
with config_file.open("r", encoding="utf-8") as f:
data = yaml.safe_load(f) or {}
if not isinstance(data, dict):
raise ValueError(f"{config_file} must contain a YAML mapping.")
if "configs" not in data or not isinstance(data["configs"], dict):
raise ValueError(f"{config_file} must define a `configs` mapping.")
defaults = data.get("defaults", {})
if defaults is None:
defaults = {}
if not isinstance(defaults, dict):
raise ValueError(f"{config_file} `defaults` must be a mapping.")
policy_defaults = data.get("policy_defaults", {})
if policy_defaults is None:
policy_defaults = {}
if not isinstance(policy_defaults, dict):
raise ValueError(f"{config_file} `policy_defaults` must be a mapping.")
return {
"defaults": defaults,
"policy_defaults": policy_defaults,
"configs": data["configs"],
}
def deep_merge(base: dict[str, Any], override: dict[str, Any]) -> dict[str, Any]:
merged = copy.deepcopy(base)
for key, value in override.items():
if isinstance(value, dict) and isinstance(merged.get(key), dict):
merged[key] = deep_merge(merged[key], value)
else:
merged[key] = copy.deepcopy(value)
return merged
def config_by_name(config_file: Path, config_name: str) -> dict[str, Any]:
data = load_sweep_configs(config_file)
configs = data["configs"]
if config_name not in configs:
available = ", ".join(sorted(configs))
raise ValueError(
f"Unknown config name: {config_name}. Available configs: {available}"
)
selected = configs[config_name]
if selected is None:
selected = {}
if not isinstance(selected, dict):
raise ValueError(f"Config `{config_name}` must be a mapping.")
selected_policy = selected.get("policy", {})
if selected_policy is None:
selected_policy = {}
if not isinstance(selected_policy, dict):
raise ValueError(f"Config `{config_name}` field `policy` must be a mapping.")
defaults_policy = data["defaults"].get("policy", {})
if defaults_policy is None:
defaults_policy = {}
if not isinstance(defaults_policy, dict):
raise ValueError(f"{config_file} `defaults.policy` must be a mapping.")
policy_type = selected_policy.get("type") or defaults_policy.get("type") or "act"
policy_defaults = data["policy_defaults"].get(policy_type, {})
if policy_defaults is None:
policy_defaults = {}
if not isinstance(policy_defaults, dict):
raise ValueError(
f"{config_file} `policy_defaults.{policy_type}` must be a mapping."
)
cfg = deep_merge(data["defaults"], policy_defaults)
return deep_merge(cfg, selected)
def str_value(value: Any) -> str:
if isinstance(value, bool):
return str(value).lower()
return str(value)
def add_cli_arg(cmd: list[str], key: str, value: Any) -> None:
if value is None:
return
cmd.append(f"--{key}={str_value(value)}")
def looks_like_local_path(value: str) -> bool:
expanded = Path(value).expanduser()
if expanded.is_absolute():
return True
if value.startswith((".", "~")):
return True
# Hugging Face model IDs are commonly "user/repo". Longer slash-separated
# values are much more likely to be local paths such as outputs/train/...
return len(expanded.parts) > 2
def validate_pretrained_policy_path(value: Any) -> str | None:
if value is None:
return None
if not isinstance(value, str):
raise ValueError("`pretrained_policy_path` must be a string.")
path_value = value.strip()
if not path_value:
raise ValueError("`pretrained_policy_path` must not be empty.")
candidate = Path(path_value).expanduser()
if candidate.exists() or looks_like_local_path(path_value):
required_files = ("config.json", "model.safetensors")
missing = [name for name in required_files if not (candidate / name).is_file()]
if missing:
missing_list = ", ".join(missing)
raise FileNotFoundError(
f"`pretrained_policy_path` must point to a LeRobot `pretrained_model` "
f"directory containing {missing_list}: {candidate}"
)
return path_value
def apply_cli_overrides(cfg: dict[str, Any], args: argparse.Namespace) -> dict[str, Any]:
overridden = copy.deepcopy(cfg)
for key in (
"dataset_repo",
"aggregate_repo",
"output_dir",
"job_name",
"steps",
"batch_size",
"device",
"wandb_project",
"log_freq",
"save_freq",
"policy_repo_id",
"pretrained_policy_path",
):
value = getattr(args, key)
if value is not None:
overridden[key] = value
if args.dataset_repos is not None:
overridden["dataset_repos"] = args.dataset_repos
if args.aggregate_drop_features is not None:
overridden["aggregate_drop_features"] = args.aggregate_drop_features
return overridden
def download_dataset_if_needed(repo_id: str) -> Path:
local_dir = dataset_local_dir(repo_id)
if looks_downloaded(local_dir):
print(f"Dataset already present: {local_dir}")
return local_dir
print(f"Downloading dataset {repo_id} to {local_dir}")
from huggingface_hub import snapshot_download
local_dir.parent.mkdir(parents=True, exist_ok=True)
snapshot_download(
repo_id=repo_id,
repo_type="dataset",
local_dir=str(local_dir),
local_dir_use_symlinks=False,
)
return local_dir
def sanitize_dataset_for_aggregation(
repo_id: str,
root: Path,
drop_features: list[str],
aggregate_repo: str,
source_index: int,
) -> tuple[str, Path]:
if not drop_features:
return repo_id, root
from lerobot.datasets.dataset_tools import remove_feature
from lerobot.datasets.lerobot_dataset import LeRobotDataset
dataset = LeRobotDataset(repo_id, root=root)
features_to_drop = [
feature_name for feature_name in drop_features if feature_name in dataset.meta.features
]
if not features_to_drop:
return repo_id, root
sanitized_repo_id = f"{aggregate_repo}_source_{source_index:02d}_sanitized"
sanitized_root = dataset_local_dir(sanitized_repo_id)
if looks_downloaded(sanitized_root):
print(f"Sanitized dataset already present: {sanitized_root}")
return sanitized_repo_id, sanitized_root
if sanitized_root.exists():
raise FileExistsError(
f"Sanitized dataset path exists but does not look complete: {sanitized_root}. "
"Remove the incomplete directory or choose a different `aggregate_repo`."
)
print(
f"Creating sanitized copy of {repo_id} without features: "
+ ", ".join(features_to_drop)
)
remove_feature(
dataset=dataset,
feature_names=features_to_drop,
output_dir=sanitized_root,
repo_id=sanitized_repo_id,
)
return sanitized_repo_id, sanitized_root
def prepare_training_dataset(cfg: dict[str, Any]) -> tuple[str, Path]:
dataset_repos = normalize_dataset_repos(cfg.get("dataset_repos"))
dataset_repo = cfg.get("dataset_repo", DEFAULT_DATASET_REPO)
if not dataset_repos:
dataset_repos = [dataset_repo]
if len(dataset_repos) == 1:
repo_id = dataset_repos[0]
return repo_id, download_dataset_if_needed(repo_id)
source_roots = [download_dataset_if_needed(repo_id) for repo_id in dataset_repos]
aggregate_repo = cfg.get("aggregate_repo") or default_aggregate_repo_id(dataset_repos)
aggregate_root = dataset_local_dir(aggregate_repo)
drop_features = normalize_string_list(
cfg.get("aggregate_drop_features"), "aggregate_drop_features"
)
if looks_downloaded(aggregate_root):
print(f"Aggregated dataset already present: {aggregate_root}")
return aggregate_repo, aggregate_root
if aggregate_root.exists():
raise FileExistsError(
f"Aggregate dataset path exists but does not look complete: {aggregate_root}. "
"Choose a different `aggregate_repo` or remove the incomplete directory."
)
aggregate_source_repos = []
aggregate_source_roots = []
for source_index, (repo_id, root) in enumerate(zip(dataset_repos, source_roots)):
sanitized_repo_id, sanitized_root = sanitize_dataset_for_aggregation(
repo_id=repo_id,
root=root,
drop_features=drop_features,
aggregate_repo=aggregate_repo,
source_index=source_index,
)
aggregate_source_repos.append(sanitized_repo_id)
aggregate_source_roots.append(sanitized_root)
print("Aggregating datasets:")
for repo_id in aggregate_source_repos:
print(f" {repo_id}")
print(f"Aggregate repo id: {aggregate_repo}")
print(f"Aggregate local dir: {aggregate_root}")
from lerobot.datasets.aggregate import aggregate_datasets
aggregate_datasets(
repo_ids=aggregate_source_repos,
roots=aggregate_source_roots,
aggr_repo_id=aggregate_repo,
aggr_root=aggregate_root,
)
return aggregate_repo, aggregate_root
def main() -> None:
parser = argparse.ArgumentParser()
parser.add_argument("--config-file", default=str(DEFAULT_CONFIG_FILE))
parser.add_argument("--config-name")
parser.add_argument("--list-configs", action="store_true")
parser.add_argument("--dataset-repo", default=None)
parser.add_argument("--dataset-repos", nargs="+", default=None)
parser.add_argument("--aggregate-repo", default=None)
parser.add_argument("--aggregate-drop-features", nargs="+", default=None)
parser.add_argument("--output-dir", default=None)
parser.add_argument("--job-name", default=None)
parser.add_argument("--steps", type=int, default=None)
parser.add_argument("--batch-size", type=int, default=None)
parser.add_argument("--device", default=None)
parser.add_argument("--wandb-project", default=None)
parser.add_argument("--log-freq", type=int, default=None)
parser.add_argument("--save-freq", type=int, default=None)
parser.add_argument("--policy-repo-id", default=None)
parser.add_argument("--pretrained-policy-path", default=None)
args = parser.parse_args()
config_file = resolve_config_file(args.config_file)
sweep_data = load_sweep_configs(config_file)
if args.list_configs:
print("Available configs:")
for name in sorted(sweep_data["configs"]):
print(f" {name}")
return
if not args.config_name:
available = ", ".join(sorted(sweep_data["configs"]))
parser.error(f"--config-name is required. Available configs: {available}")
cfg = apply_cli_overrides(config_by_name(config_file, args.config_name), args)
output_dir = cfg.get("output_dir", "outputs/train/act_so101_lcc")
job_name = cfg.get("job_name", args.config_name)
steps = cfg.get("steps", 20_000)
batch_size = cfg.get("batch_size", 32)
device = cfg.get("device") or pick_device()
wandb_project = cfg.get("wandb_project", "lerobot-so101-act-lcc")
log_freq = cfg.get("log_freq", 100)
save_freq = cfg.get("save_freq", 4_000)
policy_repo_id = cfg.get("policy_repo_id")
pretrained_policy_path = validate_pretrained_policy_path(
cfg.get("pretrained_policy_path")
)
policy_cfg = cfg.get("policy", {})
if not isinstance(policy_cfg, dict):
raise ValueError(f"Config `{args.config_name}` field `policy` must be a mapping.")
print(f"Selected config: {args.config_name}")
print("Selected device:", device)
if shutil.which("lerobot-train") is None:
raise RuntimeError(
"Could not find `lerobot-train`. Install LeRobot first: pip install lerobot"
)
dataset_repo, local_dir = prepare_training_dataset(cfg)
os.environ.setdefault("WANDB_PROJECT", wandb_project)
cmd = [
"lerobot-train",
f"--dataset.repo_id={dataset_repo}",
f"--dataset.root={local_dir}",
f"--output_dir={output_dir}",
f"--job_name={job_name}",
f"--policy.device={device}",
"--wandb.enable=true",
"--policy.push_to_hub=false",
f"--steps={steps}",
f"--batch_size={batch_size}",
f"--log_freq={log_freq}",
f"--save_freq={save_freq}",
]
for key, value in policy_cfg.items():
add_cli_arg(cmd, f"policy.{key}", value)
if pretrained_policy_path:
cmd.append(f"--policy.pretrained_path={pretrained_policy_path}")
if policy_repo_id:
cmd.append(f"--policy.repo_id={policy_repo_id}")
print("\nRunning:\n " + " \\\n ".join(cmd) + "\n")
subprocess.run(cmd, check=True)
if __name__ == "__main__":
main()