|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import logging
|
|
|
import time
|
|
|
from contextlib import nullcontext
|
|
|
from pprint import pformat
|
|
|
from typing import Any
|
|
|
|
|
|
import torch
|
|
|
from accelerate import Accelerator
|
|
|
from termcolor import colored
|
|
|
from torch.optim import Optimizer
|
|
|
|
|
|
from lerobot.configs import parser
|
|
|
from lerobot.configs.train import TrainPipelineConfig
|
|
|
from lerobot.datasets.factory import make_dataset
|
|
|
from lerobot.datasets.sampler import EpisodeAwareSampler
|
|
|
from lerobot.datasets.utils import cycle
|
|
|
from lerobot.envs.factory import make_env
|
|
|
from lerobot.envs.utils import close_envs
|
|
|
from lerobot.optim.factory import make_optimizer_and_scheduler
|
|
|
from lerobot.policies.factory import make_policy, make_pre_post_processors
|
|
|
from lerobot.policies.pretrained import PreTrainedPolicy
|
|
|
from lerobot.rl.wandb_utils import WandBLogger
|
|
|
from lerobot.scripts.lerobot_eval import eval_policy_all
|
|
|
from lerobot.utils.logging_utils import AverageMeter, MetricsTracker
|
|
|
from lerobot.utils.random_utils import set_seed
|
|
|
from lerobot.utils.train_utils import (
|
|
|
get_step_checkpoint_dir,
|
|
|
get_step_identifier,
|
|
|
load_training_state,
|
|
|
save_checkpoint,
|
|
|
update_last_checkpoint,
|
|
|
)
|
|
|
from lerobot.utils.utils import (
|
|
|
format_big_number,
|
|
|
has_method,
|
|
|
init_logging,
|
|
|
)
|
|
|
|
|
|
|
|
|
def update_policy(
|
|
|
train_metrics: MetricsTracker,
|
|
|
policy: PreTrainedPolicy,
|
|
|
batch: Any,
|
|
|
optimizer: Optimizer,
|
|
|
grad_clip_norm: float,
|
|
|
accelerator: Accelerator,
|
|
|
lr_scheduler=None,
|
|
|
lock=None,
|
|
|
) -> tuple[MetricsTracker, dict]:
|
|
|
"""
|
|
|
Performs a single training step to update the policy's weights.
|
|
|
|
|
|
This function executes the forward and backward passes, clips gradients, and steps the optimizer and
|
|
|
learning rate scheduler. Accelerator handles mixed-precision training automatically.
|
|
|
|
|
|
Args:
|
|
|
train_metrics: A MetricsTracker instance to record training statistics.
|
|
|
policy: The policy model to be trained.
|
|
|
batch: A batch of training data.
|
|
|
optimizer: The optimizer used to update the policy's parameters.
|
|
|
grad_clip_norm: The maximum norm for gradient clipping.
|
|
|
accelerator: The Accelerator instance for distributed training and mixed precision.
|
|
|
lr_scheduler: An optional learning rate scheduler.
|
|
|
lock: An optional lock for thread-safe optimizer updates.
|
|
|
|
|
|
Returns:
|
|
|
A tuple containing:
|
|
|
- The updated MetricsTracker with new statistics for this step.
|
|
|
- A dictionary of outputs from the policy's forward pass, for logging purposes.
|
|
|
"""
|
|
|
start_time = time.perf_counter()
|
|
|
policy.train()
|
|
|
|
|
|
|
|
|
with accelerator.autocast():
|
|
|
loss, output_dict = policy.forward(batch)
|
|
|
|
|
|
|
|
|
|
|
|
accelerator.backward(loss)
|
|
|
|
|
|
|
|
|
if grad_clip_norm > 0:
|
|
|
grad_norm = accelerator.clip_grad_norm_(policy.parameters(), grad_clip_norm)
|
|
|
else:
|
|
|
grad_norm = torch.nn.utils.clip_grad_norm_(
|
|
|
policy.parameters(), float("inf"), error_if_nonfinite=False
|
|
|
)
|
|
|
|
|
|
|
|
|
with lock if lock is not None else nullcontext():
|
|
|
optimizer.step()
|
|
|
|
|
|
optimizer.zero_grad()
|
|
|
|
|
|
|
|
|
if lr_scheduler is not None:
|
|
|
lr_scheduler.step()
|
|
|
|
|
|
|
|
|
if has_method(accelerator.unwrap_model(policy, keep_fp32_wrapper=True), "update"):
|
|
|
accelerator.unwrap_model(policy, keep_fp32_wrapper=True).update()
|
|
|
|
|
|
train_metrics.loss = loss.item()
|
|
|
train_metrics.grad_norm = grad_norm.item()
|
|
|
train_metrics.lr = optimizer.param_groups[0]["lr"]
|
|
|
train_metrics.update_s = time.perf_counter() - start_time
|
|
|
return train_metrics, output_dict
|
|
|
|
|
|
|
|
|
@parser.wrap()
|
|
|
def train(cfg: TrainPipelineConfig, accelerator: Accelerator | None = None):
|
|
|
"""
|
|
|
Main function to train a policy.
|
|
|
|
|
|
This function orchestrates the entire training pipeline, including:
|
|
|
- Setting up logging, seeding, and device configuration.
|
|
|
- Creating the dataset, evaluation environment (if applicable), policy, and optimizer.
|
|
|
- Handling resumption from a checkpoint.
|
|
|
- Running the main training loop, which involves fetching data batches and calling `update_policy`.
|
|
|
- Periodically logging metrics, saving model checkpoints, and evaluating the policy.
|
|
|
- Pushing the final trained model to the Hugging Face Hub if configured.
|
|
|
|
|
|
Args:
|
|
|
cfg: A `TrainPipelineConfig` object containing all training configurations.
|
|
|
accelerator: Optional Accelerator instance. If None, one will be created automatically.
|
|
|
"""
|
|
|
cfg.validate()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if accelerator is None:
|
|
|
from accelerate.utils import DistributedDataParallelKwargs
|
|
|
|
|
|
ddp_kwargs = DistributedDataParallelKwargs(find_unused_parameters=True)
|
|
|
accelerator = Accelerator(step_scheduler_with_optimizer=False, kwargs_handlers=[ddp_kwargs])
|
|
|
|
|
|
init_logging(accelerator=accelerator)
|
|
|
|
|
|
|
|
|
|
|
|
is_main_process = accelerator.is_main_process
|
|
|
|
|
|
|
|
|
if is_main_process:
|
|
|
logging.info(pformat(cfg.to_dict()))
|
|
|
|
|
|
|
|
|
if cfg.wandb.enable and cfg.wandb.project and is_main_process:
|
|
|
wandb_logger = WandBLogger(cfg)
|
|
|
else:
|
|
|
wandb_logger = None
|
|
|
if is_main_process:
|
|
|
logging.info(colored("Logs will be saved locally.", "yellow", attrs=["bold"]))
|
|
|
|
|
|
if cfg.seed is not None:
|
|
|
set_seed(cfg.seed, accelerator=accelerator)
|
|
|
|
|
|
|
|
|
device = accelerator.device
|
|
|
torch.backends.cudnn.benchmark = True
|
|
|
torch.backends.cuda.matmul.allow_tf32 = True
|
|
|
|
|
|
|
|
|
if is_main_process:
|
|
|
logging.info("Creating dataset")
|
|
|
dataset = make_dataset(cfg)
|
|
|
|
|
|
accelerator.wait_for_everyone()
|
|
|
|
|
|
|
|
|
if not is_main_process:
|
|
|
dataset = make_dataset(cfg)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
eval_env = None
|
|
|
if cfg.eval_freq > 0 and cfg.env is not None:
|
|
|
if is_main_process:
|
|
|
logging.info("Creating env")
|
|
|
eval_env = make_env(cfg.env, n_envs=cfg.eval.batch_size, use_async_envs=cfg.eval.use_async_envs)
|
|
|
|
|
|
if is_main_process:
|
|
|
logging.info("Creating policy")
|
|
|
policy = make_policy(
|
|
|
cfg=cfg.policy,
|
|
|
ds_meta=dataset.meta,
|
|
|
rename_map=cfg.rename_map,
|
|
|
)
|
|
|
|
|
|
|
|
|
accelerator.wait_for_everyone()
|
|
|
|
|
|
|
|
|
processor_kwargs = {}
|
|
|
postprocessor_kwargs = {}
|
|
|
if (cfg.policy.pretrained_path and not cfg.resume) or not cfg.policy.pretrained_path:
|
|
|
|
|
|
processor_kwargs["dataset_stats"] = dataset.meta.stats
|
|
|
|
|
|
if cfg.policy.pretrained_path is not None:
|
|
|
processor_kwargs["preprocessor_overrides"] = {
|
|
|
"device_processor": {"device": device.type},
|
|
|
"normalizer_processor": {
|
|
|
"stats": dataset.meta.stats,
|
|
|
"features": {**policy.config.input_features, **policy.config.output_features},
|
|
|
"norm_map": policy.config.normalization_mapping,
|
|
|
},
|
|
|
}
|
|
|
processor_kwargs["preprocessor_overrides"]["rename_observations_processor"] = {
|
|
|
"rename_map": cfg.rename_map
|
|
|
}
|
|
|
postprocessor_kwargs["postprocessor_overrides"] = {
|
|
|
"unnormalizer_processor": {
|
|
|
"stats": dataset.meta.stats,
|
|
|
"features": policy.config.output_features,
|
|
|
"norm_map": policy.config.normalization_mapping,
|
|
|
},
|
|
|
}
|
|
|
|
|
|
preprocessor, postprocessor = make_pre_post_processors(
|
|
|
policy_cfg=cfg.policy,
|
|
|
pretrained_path=cfg.policy.pretrained_path,
|
|
|
**processor_kwargs,
|
|
|
**postprocessor_kwargs,
|
|
|
)
|
|
|
|
|
|
if is_main_process:
|
|
|
logging.info("Creating optimizer and scheduler")
|
|
|
optimizer, lr_scheduler = make_optimizer_and_scheduler(cfg, policy)
|
|
|
|
|
|
step = 0
|
|
|
|
|
|
if cfg.resume:
|
|
|
step, optimizer, lr_scheduler = load_training_state(cfg.checkpoint_path, optimizer, lr_scheduler)
|
|
|
|
|
|
num_learnable_params = sum(p.numel() for p in policy.parameters() if p.requires_grad)
|
|
|
num_total_params = sum(p.numel() for p in policy.parameters())
|
|
|
|
|
|
if is_main_process:
|
|
|
logging.info(colored("Output dir:", "yellow", attrs=["bold"]) + f" {cfg.output_dir}")
|
|
|
if cfg.env is not None:
|
|
|
logging.info(f"{cfg.env.task=}")
|
|
|
logging.info(f"{cfg.steps=} ({format_big_number(cfg.steps)})")
|
|
|
logging.info(f"{dataset.num_frames=} ({format_big_number(dataset.num_frames)})")
|
|
|
logging.info(f"{dataset.num_episodes=}")
|
|
|
num_processes = accelerator.num_processes
|
|
|
effective_bs = cfg.batch_size * num_processes
|
|
|
logging.info(f"Effective batch size: {cfg.batch_size} x {num_processes} = {effective_bs}")
|
|
|
logging.info(f"{num_learnable_params=} ({format_big_number(num_learnable_params)})")
|
|
|
logging.info(f"{num_total_params=} ({format_big_number(num_total_params)})")
|
|
|
|
|
|
|
|
|
if hasattr(cfg.policy, "drop_n_last_frames"):
|
|
|
shuffle = False
|
|
|
sampler = EpisodeAwareSampler(
|
|
|
dataset.meta.episodes["dataset_from_index"],
|
|
|
dataset.meta.episodes["dataset_to_index"],
|
|
|
drop_n_last_frames=cfg.policy.drop_n_last_frames,
|
|
|
shuffle=True,
|
|
|
)
|
|
|
else:
|
|
|
shuffle = True
|
|
|
sampler = None
|
|
|
|
|
|
dataloader = torch.utils.data.DataLoader(
|
|
|
dataset,
|
|
|
num_workers=cfg.num_workers,
|
|
|
batch_size=cfg.batch_size,
|
|
|
shuffle=shuffle and not cfg.dataset.streaming,
|
|
|
sampler=sampler,
|
|
|
pin_memory=device.type == "cuda",
|
|
|
drop_last=False,
|
|
|
prefetch_factor=2 if cfg.num_workers > 0 else None,
|
|
|
)
|
|
|
|
|
|
|
|
|
accelerator.wait_for_everyone()
|
|
|
policy, optimizer, dataloader, lr_scheduler = accelerator.prepare(
|
|
|
policy, optimizer, dataloader, lr_scheduler
|
|
|
)
|
|
|
dl_iter = cycle(dataloader)
|
|
|
|
|
|
policy.train()
|
|
|
|
|
|
train_metrics = {
|
|
|
"loss": AverageMeter("loss", ":.3f"),
|
|
|
"grad_norm": AverageMeter("grdn", ":.3f"),
|
|
|
"lr": AverageMeter("lr", ":0.1e"),
|
|
|
"update_s": AverageMeter("updt_s", ":.3f"),
|
|
|
"dataloading_s": AverageMeter("data_s", ":.3f"),
|
|
|
}
|
|
|
|
|
|
|
|
|
effective_batch_size = cfg.batch_size * accelerator.num_processes
|
|
|
train_tracker = MetricsTracker(
|
|
|
effective_batch_size,
|
|
|
dataset.num_frames,
|
|
|
dataset.num_episodes,
|
|
|
train_metrics,
|
|
|
initial_step=step,
|
|
|
accelerator=accelerator,
|
|
|
)
|
|
|
|
|
|
if is_main_process:
|
|
|
logging.info("Start offline training on a fixed dataset")
|
|
|
|
|
|
for _ in range(step, cfg.steps):
|
|
|
start_time = time.perf_counter()
|
|
|
batch = next(dl_iter)
|
|
|
batch = preprocessor(batch)
|
|
|
train_tracker.dataloading_s = time.perf_counter() - start_time
|
|
|
|
|
|
train_tracker, output_dict = update_policy(
|
|
|
train_tracker,
|
|
|
policy,
|
|
|
batch,
|
|
|
optimizer,
|
|
|
cfg.optimizer.grad_clip_norm,
|
|
|
accelerator=accelerator,
|
|
|
lr_scheduler=lr_scheduler,
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
step += 1
|
|
|
train_tracker.step()
|
|
|
is_log_step = cfg.log_freq > 0 and step % cfg.log_freq == 0 and is_main_process
|
|
|
is_saving_step = step % cfg.save_freq == 0 or step == cfg.steps
|
|
|
is_eval_step = cfg.eval_freq > 0 and step % cfg.eval_freq == 0
|
|
|
|
|
|
if is_log_step:
|
|
|
logging.info(train_tracker)
|
|
|
if wandb_logger:
|
|
|
wandb_log_dict = train_tracker.to_dict()
|
|
|
if output_dict:
|
|
|
wandb_log_dict.update(output_dict)
|
|
|
wandb_logger.log_dict(wandb_log_dict, step)
|
|
|
train_tracker.reset_averages()
|
|
|
|
|
|
if cfg.save_checkpoint and is_saving_step:
|
|
|
if is_main_process:
|
|
|
logging.info(f"Checkpoint policy after step {step}")
|
|
|
checkpoint_dir = get_step_checkpoint_dir(cfg.output_dir, cfg.steps, step)
|
|
|
save_checkpoint(
|
|
|
checkpoint_dir=checkpoint_dir,
|
|
|
step=step,
|
|
|
cfg=cfg,
|
|
|
policy=accelerator.unwrap_model(policy),
|
|
|
optimizer=optimizer,
|
|
|
scheduler=lr_scheduler,
|
|
|
preprocessor=preprocessor,
|
|
|
postprocessor=postprocessor,
|
|
|
)
|
|
|
update_last_checkpoint(checkpoint_dir)
|
|
|
if wandb_logger:
|
|
|
wandb_logger.log_policy(checkpoint_dir)
|
|
|
|
|
|
accelerator.wait_for_everyone()
|
|
|
|
|
|
if cfg.env and is_eval_step:
|
|
|
if is_main_process:
|
|
|
step_id = get_step_identifier(step, cfg.steps)
|
|
|
logging.info(f"Eval policy at step {step}")
|
|
|
with torch.no_grad(), accelerator.autocast():
|
|
|
eval_info = eval_policy_all(
|
|
|
envs=eval_env,
|
|
|
policy=accelerator.unwrap_model(policy),
|
|
|
preprocessor=preprocessor,
|
|
|
postprocessor=postprocessor,
|
|
|
n_episodes=cfg.eval.n_episodes,
|
|
|
videos_dir=cfg.output_dir / "eval" / f"videos_step_{step_id}",
|
|
|
max_episodes_rendered=4,
|
|
|
start_seed=cfg.seed,
|
|
|
max_parallel_tasks=cfg.env.max_parallel_tasks,
|
|
|
)
|
|
|
|
|
|
aggregated = eval_info["overall"]
|
|
|
|
|
|
|
|
|
for suite, suite_info in eval_info.items():
|
|
|
logging.info("Suite %s aggregated: %s", suite, suite_info)
|
|
|
|
|
|
|
|
|
eval_metrics = {
|
|
|
"avg_sum_reward": AverageMeter("∑rwrd", ":.3f"),
|
|
|
"pc_success": AverageMeter("success", ":.1f"),
|
|
|
"eval_s": AverageMeter("eval_s", ":.3f"),
|
|
|
}
|
|
|
eval_tracker = MetricsTracker(
|
|
|
cfg.batch_size,
|
|
|
dataset.num_frames,
|
|
|
dataset.num_episodes,
|
|
|
eval_metrics,
|
|
|
initial_step=step,
|
|
|
accelerator=accelerator,
|
|
|
)
|
|
|
eval_tracker.eval_s = aggregated.pop("eval_s")
|
|
|
eval_tracker.avg_sum_reward = aggregated.pop("avg_sum_reward")
|
|
|
eval_tracker.pc_success = aggregated.pop("pc_success")
|
|
|
if wandb_logger:
|
|
|
wandb_log_dict = {**eval_tracker.to_dict(), **eval_info}
|
|
|
wandb_logger.log_dict(wandb_log_dict, step, mode="eval")
|
|
|
wandb_logger.log_video(eval_info["overall"]["video_paths"][0], step, mode="eval")
|
|
|
|
|
|
accelerator.wait_for_everyone()
|
|
|
|
|
|
if eval_env:
|
|
|
close_envs(eval_env)
|
|
|
|
|
|
if is_main_process:
|
|
|
logging.info("End of training")
|
|
|
|
|
|
if cfg.policy.push_to_hub:
|
|
|
unwrapped_policy = accelerator.unwrap_model(policy)
|
|
|
unwrapped_policy.push_model_to_hub(cfg)
|
|
|
preprocessor.push_to_hub(cfg.policy.repo_id)
|
|
|
postprocessor.push_to_hub(cfg.policy.repo_id)
|
|
|
|
|
|
|
|
|
accelerator.wait_for_everyone()
|
|
|
accelerator.end_training()
|
|
|
|
|
|
|
|
|
def main():
|
|
|
train()
|
|
|
|
|
|
|
|
|
if __name__ == "__main__":
|
|
|
main()
|
|
|
|