ConstructTraining / source /isaaclab /test /utils /test_hdf5_dataset_file_handler.py
gerlachje's picture
Upload folder using huggingface_hub
406662d verified
# Copyright (c) 2022-2026, The Isaac Lab Project Developers (https://github.com/isaac-sim/IsaacLab/blob/main/CONTRIBUTORS.md).
# All rights reserved.
#
# SPDX-License-Identifier: BSD-3-Clause
"""Launch Isaac Sim Simulator first."""
from isaaclab.app import AppLauncher
# launch omniverse app in headless mode
simulation_app = AppLauncher(headless=True).app
"""Rest everything follows from here."""
import os
import shutil
import tempfile
import uuid
import pytest
import torch
from isaaclab.utils.datasets import EpisodeData, HDF5DatasetFileHandler
def create_test_episode(device):
"""create a test episode with dummy data."""
test_episode = EpisodeData()
test_episode.seed = 0
test_episode.success = True
test_episode.add("initial_state", torch.tensor([1, 2, 3], device=device))
test_episode.add("actions", torch.tensor([1, 2, 3], device=device))
test_episode.add("actions", torch.tensor([4, 5, 6], device=device))
test_episode.add("actions", torch.tensor([7, 8, 9], device=device))
test_episode.add("obs/policy/term1", torch.tensor([1, 2, 3, 4, 5], device=device))
test_episode.add("obs/policy/term1", torch.tensor([6, 7, 8, 9, 10], device=device))
test_episode.add("obs/policy/term1", torch.tensor([11, 12, 13, 14, 15], device=device))
return test_episode
@pytest.fixture
def temp_dir():
"""Create a temporary directory for test datasets."""
temp_dir = tempfile.mkdtemp()
yield temp_dir
# cleanup after tests
shutil.rmtree(temp_dir)
def test_create_dataset_file(temp_dir):
"""Test creating a new dataset file."""
# create a dataset file given a file name with extension
dataset_file_path = os.path.join(temp_dir, f"{uuid.uuid4()}.hdf5")
dataset_file_handler = HDF5DatasetFileHandler()
dataset_file_handler.create(dataset_file_path, "test_env_name")
dataset_file_handler.close()
# check if the dataset is created
assert os.path.exists(dataset_file_path)
# create a dataset file given a file name without extension
dataset_file_path = os.path.join(temp_dir, f"{uuid.uuid4()}")
dataset_file_handler = HDF5DatasetFileHandler()
dataset_file_handler.create(dataset_file_path, "test_env_name")
dataset_file_handler.close()
# check if the dataset is created
assert os.path.exists(dataset_file_path + ".hdf5")
@pytest.mark.parametrize("device", ["cuda:0", "cpu"])
def test_write_and_load_episode(temp_dir, device):
"""Test writing and loading an episode to and from the dataset file."""
dataset_file_path = os.path.join(temp_dir, f"{uuid.uuid4()}.hdf5")
dataset_file_handler = HDF5DatasetFileHandler()
dataset_file_handler.create(dataset_file_path, "test_env_name")
test_episode = create_test_episode(device)
# write the episode to the dataset
test_episode.pre_export()
dataset_file_handler.write_episode(test_episode)
dataset_file_handler.flush()
assert dataset_file_handler.get_num_episodes() == 1
# write the episode again to test writing 2nd episode
dataset_file_handler.write_episode(test_episode)
dataset_file_handler.flush()
assert dataset_file_handler.get_num_episodes() == 2
# close the dataset file to prepare for testing the load function
dataset_file_handler.close()
# load the episode from the dataset
dataset_file_handler = HDF5DatasetFileHandler()
dataset_file_handler.open(dataset_file_path)
assert dataset_file_handler.get_env_name() == "test_env_name"
loaded_episode_names = dataset_file_handler.get_episode_names()
assert len(list(loaded_episode_names)) == 2
for episode_name in loaded_episode_names:
loaded_episode = dataset_file_handler.load_episode(episode_name, device=device)
assert loaded_episode.env_id == "test_env_name"
assert loaded_episode.seed == test_episode.seed
assert loaded_episode.success == test_episode.success
assert torch.equal(loaded_episode.get_initial_state(), test_episode.get_initial_state())
for action in test_episode.data["actions"]:
assert torch.equal(loaded_episode.get_next_action(), action)
dataset_file_handler.close()