| | |
| | |
| | |
| | |
| | """Launch Isaac Sim Simulator first.""" |
| |
|
| | from isaaclab.app import AppLauncher |
| |
|
| | |
| | simulation_app = AppLauncher(headless=True).app |
| |
|
| | """Rest everything follows from here.""" |
| |
|
| | import os |
| | import shutil |
| | import tempfile |
| | import uuid |
| |
|
| | import pytest |
| | import torch |
| |
|
| | from isaaclab.utils.datasets import EpisodeData, HDF5DatasetFileHandler |
| |
|
| |
|
| | def create_test_episode(device): |
| | """create a test episode with dummy data.""" |
| | test_episode = EpisodeData() |
| |
|
| | test_episode.seed = 0 |
| | test_episode.success = True |
| |
|
| | test_episode.add("initial_state", torch.tensor([1, 2, 3], device=device)) |
| |
|
| | test_episode.add("actions", torch.tensor([1, 2, 3], device=device)) |
| | test_episode.add("actions", torch.tensor([4, 5, 6], device=device)) |
| | test_episode.add("actions", torch.tensor([7, 8, 9], device=device)) |
| |
|
| | test_episode.add("obs/policy/term1", torch.tensor([1, 2, 3, 4, 5], device=device)) |
| | test_episode.add("obs/policy/term1", torch.tensor([6, 7, 8, 9, 10], device=device)) |
| | test_episode.add("obs/policy/term1", torch.tensor([11, 12, 13, 14, 15], device=device)) |
| |
|
| | return test_episode |
| |
|
| |
|
| | @pytest.fixture |
| | def temp_dir(): |
| | """Create a temporary directory for test datasets.""" |
| | temp_dir = tempfile.mkdtemp() |
| | yield temp_dir |
| | |
| | shutil.rmtree(temp_dir) |
| |
|
| |
|
| | def test_create_dataset_file(temp_dir): |
| | """Test creating a new dataset file.""" |
| | |
| | dataset_file_path = os.path.join(temp_dir, f"{uuid.uuid4()}.hdf5") |
| | dataset_file_handler = HDF5DatasetFileHandler() |
| | dataset_file_handler.create(dataset_file_path, "test_env_name") |
| | dataset_file_handler.close() |
| |
|
| | |
| | assert os.path.exists(dataset_file_path) |
| |
|
| | |
| | dataset_file_path = os.path.join(temp_dir, f"{uuid.uuid4()}") |
| | dataset_file_handler = HDF5DatasetFileHandler() |
| | dataset_file_handler.create(dataset_file_path, "test_env_name") |
| | dataset_file_handler.close() |
| |
|
| | |
| | assert os.path.exists(dataset_file_path + ".hdf5") |
| |
|
| |
|
| | @pytest.mark.parametrize("device", ["cuda:0", "cpu"]) |
| | def test_write_and_load_episode(temp_dir, device): |
| | """Test writing and loading an episode to and from the dataset file.""" |
| | dataset_file_path = os.path.join(temp_dir, f"{uuid.uuid4()}.hdf5") |
| | dataset_file_handler = HDF5DatasetFileHandler() |
| | dataset_file_handler.create(dataset_file_path, "test_env_name") |
| |
|
| | test_episode = create_test_episode(device) |
| |
|
| | |
| | test_episode.pre_export() |
| | dataset_file_handler.write_episode(test_episode) |
| | dataset_file_handler.flush() |
| |
|
| | assert dataset_file_handler.get_num_episodes() == 1 |
| |
|
| | |
| | dataset_file_handler.write_episode(test_episode) |
| | dataset_file_handler.flush() |
| |
|
| | assert dataset_file_handler.get_num_episodes() == 2 |
| |
|
| | |
| | dataset_file_handler.close() |
| |
|
| | |
| | dataset_file_handler = HDF5DatasetFileHandler() |
| | dataset_file_handler.open(dataset_file_path) |
| |
|
| | assert dataset_file_handler.get_env_name() == "test_env_name" |
| |
|
| | loaded_episode_names = dataset_file_handler.get_episode_names() |
| | assert len(list(loaded_episode_names)) == 2 |
| |
|
| | for episode_name in loaded_episode_names: |
| | loaded_episode = dataset_file_handler.load_episode(episode_name, device=device) |
| | assert loaded_episode.env_id == "test_env_name" |
| | assert loaded_episode.seed == test_episode.seed |
| | assert loaded_episode.success == test_episode.success |
| |
|
| | assert torch.equal(loaded_episode.get_initial_state(), test_episode.get_initial_state()) |
| |
|
| | for action in test_episode.data["actions"]: |
| | assert torch.equal(loaded_episode.get_next_action(), action) |
| |
|
| | dataset_file_handler.close() |
| |
|