|
|
import argparse |
|
|
import os |
|
|
from pathlib import Path |
|
|
import h5py |
|
|
import numpy as np |
|
|
import json |
|
|
import robosuite |
|
|
import robosuite.utils.transform_utils as T |
|
|
import robosuite.macros as macros |
|
|
|
|
|
import init_path |
|
|
import libero.libero.utils.utils as libero_utils |
|
|
import cv2 |
|
|
from PIL import Image |
|
|
from robosuite.utils import camera_utils |
|
|
|
|
|
from libero.libero.envs import * |
|
|
from libero.libero import get_libero_path |
|
|
|
|
|
def main(): |
|
|
parser = argparse.ArgumentParser() |
|
|
parser.add_argument("--demo-file", default="demo.hdf5") |
|
|
|
|
|
parser.add_argument( |
|
|
"--use-actions", |
|
|
action="store_true", |
|
|
) |
|
|
parser.add_argument("--use-camera-obs", action="store_true") |
|
|
parser.add_argument( |
|
|
"--dataset-path", |
|
|
type=str, |
|
|
default="datasets/", |
|
|
) |
|
|
|
|
|
parser.add_argument( |
|
|
"--dataset-name", |
|
|
type=str, |
|
|
default="training_set", |
|
|
) |
|
|
|
|
|
parser.add_argument("--no-proprio", action="store_true") |
|
|
|
|
|
parser.add_argument( |
|
|
"--use-depth", |
|
|
action="store_true", |
|
|
) |
|
|
|
|
|
args = parser.parse_args() |
|
|
|
|
|
hdf5_path = args.demo_file |
|
|
f = h5py.File(hdf5_path, "r") |
|
|
env_name = f["data"].attrs["env"] |
|
|
|
|
|
env_args = f["data"].attrs["env_info"] |
|
|
env_kwargs = json.loads(f["data"].attrs["env_info"]) |
|
|
|
|
|
problem_info = json.loads(f["data"].attrs["problem_info"]) |
|
|
problem_info["domain_name"] |
|
|
problem_name = problem_info["problem_name"] |
|
|
language_instruction = problem_info["language_instruction"] |
|
|
|
|
|
|
|
|
demos = list(f["data"].keys()) |
|
|
|
|
|
bddl_file_name = f["data"].attrs["bddl_file_name"] |
|
|
|
|
|
bddl_file_dir = os.path.dirname(bddl_file_name) |
|
|
replace_bddl_prefix = "/".join(bddl_file_dir.split("bddl_files/")[:-1] + "bddl_files") |
|
|
|
|
|
hdf5_path = os.path.join(get_libero_path("datasets"), bddl_file_dir.split("bddl_files/")[-1].replace(".bddl", "_demo.hdf5")) |
|
|
|
|
|
output_parent_dir = Path(hdf5_path).parent |
|
|
output_parent_dir.mkdir(parents=True, exist_ok=True) |
|
|
|
|
|
h5py_f = h5py.File(hdf5_path, "w") |
|
|
|
|
|
grp = h5py_f.create_group("data") |
|
|
|
|
|
grp.attrs["env_name"] = env_name |
|
|
grp.attrs["problem_info"] = f["data"].attrs["problem_info"] |
|
|
grp.attrs["macros_image_convention"] = macros.IMAGE_CONVENTION |
|
|
|
|
|
libero_utils.update_env_kwargs( |
|
|
env_kwargs, |
|
|
bddl_file_name=bddl_file_name, |
|
|
has_renderer=not args.use_camera_obs, |
|
|
has_offscreen_renderer=args.use_camera_obs, |
|
|
ignore_done=True, |
|
|
use_camera_obs=args.use_camera_obs, |
|
|
camera_depths=args.use_depth, |
|
|
camera_names=[ |
|
|
"robot0_eye_in_hand", |
|
|
"agentview", |
|
|
], |
|
|
reward_shaping=True, |
|
|
control_freq=20, |
|
|
camera_heights=128, |
|
|
camera_widths=128, |
|
|
camera_segmentations=None, |
|
|
) |
|
|
|
|
|
grp.attrs["bddl_file_name"] = bddl_file_name |
|
|
grp.attrs["bddl_file_content"] = open(bddl_file_name, "r").read() |
|
|
print(grp.attrs["bddl_file_content"]) |
|
|
|
|
|
env = TASK_MAPPING[problem_name]( |
|
|
**env_kwargs, |
|
|
) |
|
|
|
|
|
env_args = { |
|
|
"type": 1, |
|
|
"env_name": env_name, |
|
|
"problem_name": problem_name, |
|
|
"bddl_file": f["data"].attrs["bddl_file_name"], |
|
|
"env_kwargs": env_kwargs, |
|
|
} |
|
|
|
|
|
grp.attrs["env_args"] = json.dumps(env_args) |
|
|
print(grp.attrs["env_args"]) |
|
|
total_len = 0 |
|
|
demos = demos |
|
|
|
|
|
cap_index = 5 |
|
|
|
|
|
for (i, ep) in enumerate(demos): |
|
|
print("Playing back random episode... (press ESC to quit)") |
|
|
|
|
|
|
|
|
|
|
|
model_xml = f["data/{}".format(ep)].attrs["model_file"] |
|
|
reset_success = False |
|
|
while not reset_success: |
|
|
try: |
|
|
env.reset() |
|
|
reset_success = True |
|
|
except: |
|
|
continue |
|
|
|
|
|
model_xml = libero_utils.postprocess_model_xml(model_xml, {}) |
|
|
|
|
|
if not args.use_camera_obs: |
|
|
env.viewer.set_camera(0) |
|
|
|
|
|
|
|
|
states = f["data/{}/states".format(ep)][()] |
|
|
actions = np.array(f["data/{}/actions".format(ep)][()]) |
|
|
|
|
|
num_actions = actions.shape[0] |
|
|
|
|
|
init_idx = 0 |
|
|
env.reset_from_xml_string(model_xml) |
|
|
env.sim.reset() |
|
|
env.sim.set_state_from_flattened(states[init_idx]) |
|
|
env.sim.forward() |
|
|
model_xml = env.sim.model.get_xml() |
|
|
|
|
|
ee_states = [] |
|
|
gripper_states = [] |
|
|
joint_states = [] |
|
|
robot_states = [] |
|
|
|
|
|
agentview_images = [] |
|
|
eye_in_hand_images = [] |
|
|
|
|
|
agentview_depths = [] |
|
|
eye_in_hand_depths = [] |
|
|
|
|
|
agentview_seg = {0: [], 1: [], 2: [], 3: [], 4: []} |
|
|
|
|
|
rewards = [] |
|
|
dones = [] |
|
|
|
|
|
valid_index = [] |
|
|
|
|
|
for j, action in enumerate(actions): |
|
|
|
|
|
obs, reward, done, info = env.step(action) |
|
|
|
|
|
if j < num_actions - 1: |
|
|
|
|
|
state_playback = env.sim.get_state().flatten() |
|
|
|
|
|
err = np.linalg.norm(states[j + 1] - state_playback) |
|
|
|
|
|
if err > 0.01: |
|
|
print( |
|
|
f"[warning] playback diverged by {err:.2f} for ep {ep} at step {j}" |
|
|
) |
|
|
|
|
|
|
|
|
|
|
|
if j < cap_index: |
|
|
continue |
|
|
|
|
|
valid_index.append(j) |
|
|
|
|
|
if not args.no_proprio: |
|
|
if "robot0_gripper_qpos" in obs: |
|
|
gripper_states.append(obs["robot0_gripper_qpos"]) |
|
|
|
|
|
joint_states.append(obs["robot0_joint_pos"]) |
|
|
|
|
|
ee_states.append( |
|
|
np.hstack( |
|
|
( |
|
|
obs["robot0_eef_pos"], |
|
|
T.quat2axisangle(obs["robot0_eef_quat"]), |
|
|
) |
|
|
) |
|
|
) |
|
|
|
|
|
robot_states.append(env.get_robot_state_vector(obs)) |
|
|
|
|
|
if args.use_camera_obs: |
|
|
|
|
|
if args.use_depth: |
|
|
agentview_depths.append(obs["agentview_depth"]) |
|
|
eye_in_hand_depths.append(obs["robot0_eye_in_hand_depth"]) |
|
|
|
|
|
agentview_images.append(obs["agentview_image"]) |
|
|
eye_in_hand_images.append(obs["robot0_eye_in_hand_image"]) |
|
|
else: |
|
|
env.render() |
|
|
|
|
|
|
|
|
states = states[valid_index] |
|
|
actions = actions[valid_index] |
|
|
dones = np.zeros(len(actions)).astype(np.uint8) |
|
|
dones[-1] = 1 |
|
|
rewards = np.zeros(len(actions)).astype(np.uint8) |
|
|
rewards[-1] = 1 |
|
|
print(len(actions), len(agentview_images)) |
|
|
assert len(actions) == len(agentview_images) |
|
|
print(len(actions)) |
|
|
|
|
|
ep_data_grp = grp.create_group(f"demo_{i}") |
|
|
|
|
|
obs_grp = ep_data_grp.create_group("obs") |
|
|
if not args.no_proprio: |
|
|
obs_grp.create_dataset( |
|
|
"gripper_states", data=np.stack(gripper_states, axis=0) |
|
|
) |
|
|
obs_grp.create_dataset("joint_states", data=np.stack(joint_states, axis=0)) |
|
|
obs_grp.create_dataset("ee_states", data=np.stack(ee_states, axis=0)) |
|
|
obs_grp.create_dataset("ee_pos", data=np.stack(ee_states, axis=0)[:, :3]) |
|
|
obs_grp.create_dataset("ee_ori", data=np.stack(ee_states, axis=0)[:, 3:]) |
|
|
|
|
|
obs_grp.create_dataset("agentview_rgb", data=np.stack(agentview_images, axis=0)) |
|
|
obs_grp.create_dataset( |
|
|
"eye_in_hand_rgb", data=np.stack(eye_in_hand_images, axis=0) |
|
|
) |
|
|
if args.use_depth: |
|
|
obs_grp.create_dataset( |
|
|
"agentview_depth", data=np.stack(agentview_depths, axis=0) |
|
|
) |
|
|
obs_grp.create_dataset( |
|
|
"eye_in_hand_depth", data=np.stack(eye_in_hand_depths, axis=0) |
|
|
) |
|
|
|
|
|
ep_data_grp.create_dataset("actions", data=actions) |
|
|
ep_data_grp.create_dataset("states", data=states) |
|
|
ep_data_grp.create_dataset("robot_states", data=np.stack(robot_states, axis=0)) |
|
|
ep_data_grp.create_dataset("rewards", data=rewards) |
|
|
ep_data_grp.create_dataset("dones", data=dones) |
|
|
ep_data_grp.attrs["num_samples"] = len(agentview_images) |
|
|
ep_data_grp.attrs["model_file"] = model_xml |
|
|
ep_data_grp.attrs["init_state"] = states[init_idx] |
|
|
total_len += len(agentview_images) |
|
|
|
|
|
grp.attrs["num_demos"] = len(demos) |
|
|
grp.attrs["total"] = total_len |
|
|
env.close() |
|
|
|
|
|
h5py_f.close() |
|
|
f.close() |
|
|
|
|
|
print("The created dataset is saved in the following path: ") |
|
|
print(hdf5_path) |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
main() |
|
|
|