# Copyright (c) 2022-2026, The Isaac Lab Project Developers (https://github.com/isaac-sim/IsaacLab/blob/main/CONTRIBUTORS.md). # All rights reserved. # # SPDX-License-Identifier: BSD-3-Clause """Launch Isaac Sim Simulator first.""" from isaaclab.app import AppLauncher # launch the simulator app_launcher = AppLauncher(headless=True, enable_cameras=True) simulation_app = app_launcher.app """Rest everything follows.""" import gymnasium as gym import pytest import torch import carb import omni.usd from isaaclab.envs import DirectMARLEnv, multi_agent_to_single_agent from isaaclab_rl.skrl import SkrlVecEnvWrapper import isaaclab_tasks # noqa: F401 from isaaclab_tasks.utils.parse_cfg import parse_env_cfg @pytest.fixture(scope="module") def registered_tasks(): # acquire all Isaac environments names registered_tasks = list() for task_spec in gym.registry.values(): if "Isaac" in task_spec.id: cfg_entry_point = gym.spec(task_spec.id).kwargs.get("skrl_cfg_entry_point") if cfg_entry_point is not None: registered_tasks.append(task_spec.id) # sort environments by name registered_tasks.sort() registered_tasks = registered_tasks[:3] # this flag is necessary to prevent a bug where the simulation gets stuck randomly when running the # test on many environments. carb_settings_iface = carb.settings.get_settings() carb_settings_iface.set_bool("/physics/cooking/ujitsoCollisionCooking", False) # print all existing task names print(">>> All registered environments:", registered_tasks) return registered_tasks def test_random_actions(registered_tasks): """Run random actions and check environments return valid signals.""" # common parameters num_envs = 64 device = "cuda" for task_name in registered_tasks: # Use pytest's subtests print(f">>> Running test for environment: {task_name}") # create a new stage omni.usd.get_context().new_stage() # reset the rtx sensors carb setting to False carb.settings.get_settings().set_bool("/isaaclab/render/rtx_sensors", False) try: # parse configuration env_cfg = parse_env_cfg(task_name, device=device, num_envs=num_envs) # create environment env = gym.make(task_name, cfg=env_cfg) if isinstance(env.unwrapped, DirectMARLEnv): env = multi_agent_to_single_agent(env) # wrap environment env = SkrlVecEnvWrapper(env) except Exception as e: if "env" in locals() and hasattr(env, "_is_closed"): env.close() else: if hasattr(e, "obj") and hasattr(e.obj, "_is_closed"): e.obj.close() pytest.fail(f"Failed to set-up the environment for task {task_name}. Error: {e}") # avoid shutdown of process on simulation stop env.unwrapped.sim._app_control_on_stop_handle = None # reset environment obs, extras = env.reset() # check signal assert _check_valid_tensor(obs) assert _check_valid_tensor(extras) # simulate environment for 100 steps with torch.inference_mode(): for _ in range(100): # sample actions from -1 to 1 actions = 2 * torch.rand(num_envs, *env.action_space.shape, device=env.unwrapped.device) - 1 # apply actions transition = env.step(actions) # check signals for data in transition: assert _check_valid_tensor(data), f"Invalid data: {data}" # close the environment print(f">>> Closing environment: {task_name}") env.close() """ Helper functions. """ @staticmethod def _check_valid_tensor(data: torch.Tensor | dict) -> bool: """Checks if given data does not have corrupted values. Args: data: Data buffer. Returns: True if the data is valid. """ if isinstance(data, torch.Tensor): return not torch.any(torch.isnan(data)) elif isinstance(data, dict): valid_tensor = True for value in data.values(): if isinstance(value, dict): valid_tensor &= _check_valid_tensor(value) elif isinstance(value, torch.Tensor): valid_tensor &= not torch.any(torch.isnan(value)) return valid_tensor else: raise ValueError(f"Input data of invalid type: {type(data)}.")