| | |
| | |
| | |
| | |
| | """Script to replay demonstrations with Isaac Lab environments.""" |
| |
|
| | """Launch Isaac Sim Simulator first.""" |
| |
|
| |
|
| | import argparse |
| |
|
| | from isaaclab.app import AppLauncher |
| |
|
| | |
| | parser = argparse.ArgumentParser(description="Replay demonstrations in Isaac Lab environments.") |
| | parser.add_argument("--num_envs", type=int, default=1, help="Number of environments to replay episodes.") |
| | parser.add_argument("--task", type=str, default=None, help="Force to use the specified task.") |
| | parser.add_argument( |
| | "--select_episodes", |
| | type=int, |
| | nargs="+", |
| | default=[], |
| | help="A list of episode indices to be replayed. Keep empty to replay all in the dataset file.", |
| | ) |
| | parser.add_argument("--dataset_file", type=str, default="datasets/dataset.hdf5", help="Dataset file to be replayed.") |
| | parser.add_argument( |
| | "--validate_states", |
| | action="store_true", |
| | default=False, |
| | help=( |
| | "Validate if the states, if available, match between loaded from datasets and replayed. Only valid if" |
| | " --num_envs is 1." |
| | ), |
| | ) |
| | parser.add_argument( |
| | "--validate_success_rate", |
| | action="store_true", |
| | default=False, |
| | help="Validate the replay success rate using the task environment termination criteria", |
| | ) |
| | parser.add_argument( |
| | "--enable_pinocchio", |
| | action="store_true", |
| | default=False, |
| | help="Enable Pinocchio.", |
| | ) |
| |
|
| | |
| | AppLauncher.add_app_launcher_args(parser) |
| | |
| | args_cli = parser.parse_args() |
| | |
| |
|
| | if args_cli.enable_pinocchio: |
| | |
| | |
| | |
| | import pinocchio |
| |
|
| | |
| | app_launcher = AppLauncher(args_cli) |
| | simulation_app = app_launcher.app |
| |
|
| | """Rest everything follows.""" |
| |
|
| | import contextlib |
| | import os |
| |
|
| | import gymnasium as gym |
| | import torch |
| |
|
| | from isaaclab.devices import Se3Keyboard, Se3KeyboardCfg |
| | from isaaclab.utils.datasets import EpisodeData, HDF5DatasetFileHandler |
| |
|
| | if args_cli.enable_pinocchio: |
| | import isaaclab_tasks.manager_based.locomanipulation.pick_place |
| | import isaaclab_tasks.manager_based.manipulation.pick_place |
| |
|
| | import isaaclab_tasks |
| | from isaaclab_tasks.utils.parse_cfg import parse_env_cfg |
| |
|
| | is_paused = False |
| |
|
| |
|
| | def play_cb(): |
| | global is_paused |
| | is_paused = False |
| |
|
| |
|
| | def pause_cb(): |
| | global is_paused |
| | is_paused = True |
| |
|
| |
|
| | def compare_states(state_from_dataset, runtime_state, runtime_env_index) -> (bool, str): |
| | """Compare states from dataset and runtime. |
| | |
| | Args: |
| | state_from_dataset: State from dataset. |
| | runtime_state: State from runtime. |
| | runtime_env_index: Index of the environment in the runtime states to be compared. |
| | |
| | Returns: |
| | bool: True if states match, False otherwise. |
| | str: Log message if states don't match. |
| | """ |
| | states_matched = True |
| | output_log = "" |
| | for asset_type in ["articulation", "rigid_object"]: |
| | for asset_name in runtime_state[asset_type].keys(): |
| | for state_name in runtime_state[asset_type][asset_name].keys(): |
| | runtime_asset_state = runtime_state[asset_type][asset_name][state_name][runtime_env_index] |
| | dataset_asset_state = state_from_dataset[asset_type][asset_name][state_name] |
| | if len(dataset_asset_state) != len(runtime_asset_state): |
| | raise ValueError(f"State shape of {state_name} for asset {asset_name} don't match") |
| | for i in range(len(dataset_asset_state)): |
| | if abs(dataset_asset_state[i] - runtime_asset_state[i]) > 0.01: |
| | states_matched = False |
| | output_log += f'\tState ["{asset_type}"]["{asset_name}"]["{state_name}"][{i}] don\'t match\r\n' |
| | output_log += f"\t Dataset:\t{dataset_asset_state[i]}\r\n" |
| | output_log += f"\t Runtime: \t{runtime_asset_state[i]}\r\n" |
| | return states_matched, output_log |
| |
|
| |
|
| | def main(): |
| | """Replay episodes loaded from a file.""" |
| | global is_paused |
| |
|
| | |
| | if not os.path.exists(args_cli.dataset_file): |
| | raise FileNotFoundError(f"The dataset file {args_cli.dataset_file} does not exist.") |
| | dataset_file_handler = HDF5DatasetFileHandler() |
| | dataset_file_handler.open(args_cli.dataset_file) |
| | env_name = dataset_file_handler.get_env_name() |
| | episode_count = dataset_file_handler.get_num_episodes() |
| |
|
| | if episode_count == 0: |
| | print("No episodes found in the dataset.") |
| | exit() |
| |
|
| | episode_indices_to_replay = args_cli.select_episodes |
| | if len(episode_indices_to_replay) == 0: |
| | episode_indices_to_replay = list(range(episode_count)) |
| |
|
| | if args_cli.task is not None: |
| | env_name = args_cli.task.split(":")[-1] |
| | if env_name is None: |
| | raise ValueError("Task/env name was not specified nor found in the dataset.") |
| |
|
| | num_envs = args_cli.num_envs |
| |
|
| | env_cfg = parse_env_cfg(env_name, device=args_cli.device, num_envs=num_envs) |
| |
|
| | |
| | success_term = None |
| | if args_cli.validate_success_rate: |
| | if hasattr(env_cfg.terminations, "success"): |
| | success_term = env_cfg.terminations.success |
| | env_cfg.terminations.success = None |
| | else: |
| | print( |
| | "No success termination term was found in the environment." |
| | " Will not be able to mark recorded demos as successful." |
| | ) |
| |
|
| | |
| | env_cfg.recorders = {} |
| | env_cfg.terminations = {} |
| |
|
| | |
| | env = gym.make(args_cli.task, cfg=env_cfg).unwrapped |
| |
|
| | teleop_interface = Se3Keyboard(Se3KeyboardCfg(pos_sensitivity=0.1, rot_sensitivity=0.1)) |
| | teleop_interface.add_callback("N", play_cb) |
| | teleop_interface.add_callback("B", pause_cb) |
| | print('Press "B" to pause and "N" to resume the replayed actions.') |
| |
|
| | |
| | state_validation_enabled = False |
| | if args_cli.validate_states and num_envs == 1: |
| | state_validation_enabled = True |
| | elif args_cli.validate_states and num_envs > 1: |
| | print("Warning: State validation is only supported with a single environment. Skipping state validation.") |
| |
|
| | |
| | if hasattr(env_cfg, "idle_action"): |
| | idle_action = env_cfg.idle_action.repeat(num_envs, 1) |
| | else: |
| | idle_action = torch.zeros(env.action_space.shape) |
| |
|
| | |
| | env.reset() |
| | teleop_interface.reset() |
| |
|
| | |
| | episode_names = list(dataset_file_handler.get_episode_names()) |
| | replayed_episode_count = 0 |
| | recorded_episode_count = 0 |
| |
|
| | |
| | current_episode_indices = [None] * num_envs |
| |
|
| | |
| | failed_demo_ids = [] |
| |
|
| | with contextlib.suppress(KeyboardInterrupt) and torch.inference_mode(): |
| | while simulation_app.is_running() and not simulation_app.is_exiting(): |
| | env_episode_data_map = {index: EpisodeData() for index in range(num_envs)} |
| | first_loop = True |
| | has_next_action = True |
| | episode_ended = [False] * num_envs |
| | while has_next_action: |
| | |
| | actions = idle_action |
| | has_next_action = False |
| | for env_id in range(num_envs): |
| | env_next_action = env_episode_data_map[env_id].get_next_action() |
| | if env_next_action is None: |
| | |
| | if ( |
| | (success_term is not None) |
| | and (current_episode_indices[env_id]) is not None |
| | and (not episode_ended[env_id]) |
| | ): |
| | if bool(success_term.func(env, **success_term.params)[env_id]): |
| | recorded_episode_count += 1 |
| | plural_trailing_s = "s" if recorded_episode_count > 1 else "" |
| |
|
| | print( |
| | f"Successfully replayed {recorded_episode_count} episode{plural_trailing_s} out" |
| | f" of {replayed_episode_count} demos." |
| | ) |
| | else: |
| | |
| | if ( |
| | current_episode_indices[env_id] is not None |
| | and current_episode_indices[env_id] not in failed_demo_ids |
| | ): |
| | failed_demo_ids.append(current_episode_indices[env_id]) |
| |
|
| | episode_ended[env_id] = True |
| |
|
| | next_episode_index = None |
| | while episode_indices_to_replay: |
| | next_episode_index = episode_indices_to_replay.pop(0) |
| |
|
| | if next_episode_index < episode_count: |
| | episode_ended[env_id] = False |
| | break |
| | next_episode_index = None |
| |
|
| | if next_episode_index is not None: |
| | replayed_episode_count += 1 |
| | current_episode_indices[env_id] = next_episode_index |
| | print(f"{replayed_episode_count:4}: Loading #{next_episode_index} episode to env_{env_id}") |
| | episode_data = dataset_file_handler.load_episode( |
| | episode_names[next_episode_index], env.device |
| | ) |
| | env_episode_data_map[env_id] = episode_data |
| | |
| | initial_state = episode_data.get_initial_state() |
| | env.reset_to(initial_state, torch.tensor([env_id], device=env.device), is_relative=True) |
| | |
| | env_next_action = env_episode_data_map[env_id].get_next_action() |
| | has_next_action = True |
| | else: |
| | continue |
| | else: |
| | has_next_action = True |
| | actions[env_id] = env_next_action |
| | if first_loop: |
| | first_loop = False |
| | else: |
| | while is_paused: |
| | env.sim.render() |
| | continue |
| | env.step(actions) |
| |
|
| | if state_validation_enabled: |
| | state_from_dataset = env_episode_data_map[0].get_next_state() |
| | if state_from_dataset is not None: |
| | print( |
| | f"Validating states at action-index: {env_episode_data_map[0].next_state_index - 1:4}", |
| | end="", |
| | ) |
| | current_runtime_state = env.scene.get_state(is_relative=True) |
| | states_matched, comparison_log = compare_states(state_from_dataset, current_runtime_state, 0) |
| | if states_matched: |
| | print("\t- matched.") |
| | else: |
| | print("\t- mismatched.") |
| | print(comparison_log) |
| | break |
| | |
| | plural_trailing_s = "s" if replayed_episode_count > 1 else "" |
| | print(f"Finished replaying {replayed_episode_count} episode{plural_trailing_s}.") |
| |
|
| | |
| | if success_term is not None: |
| | print(f"Successfully replayed: {recorded_episode_count}/{replayed_episode_count}") |
| |
|
| | |
| | if failed_demo_ids: |
| | print(f"\nFailed demo IDs ({len(failed_demo_ids)} total):") |
| | print(f" {sorted(failed_demo_ids)}") |
| |
|
| | env.close() |
| |
|
| |
|
| | if __name__ == "__main__": |
| | |
| | main() |
| | |
| | simulation_app.close() |
| |
|