| | |
| | |
| | |
| | |
| |
|
| | """ |
| | Script to run an environment with a pick and lift state machine. |
| | |
| | The state machine is implemented in the kernel function `infer_state_machine`. |
| | It uses the `warp` library to run the state machine in parallel on the GPU. |
| | |
| | .. code-block:: bash |
| | |
| | ./isaaclab.sh -p scripts/environments/state_machine/lift_cube_sm.py --num_envs 32 |
| | |
| | """ |
| |
|
| | """Launch Omniverse Toolkit first.""" |
| |
|
| | import argparse |
| |
|
| | from isaaclab.app import AppLauncher |
| |
|
| | |
| | parser = argparse.ArgumentParser(description="Pick and lift state machine for lift environments.") |
| | parser.add_argument( |
| | "--disable_fabric", action="store_true", default=False, help="Disable fabric and use USD I/O operations." |
| | ) |
| | parser.add_argument("--num_envs", type=int, default=None, help="Number of environments to simulate.") |
| | |
| | AppLauncher.add_app_launcher_args(parser) |
| | |
| | args_cli = parser.parse_args() |
| |
|
| | |
| | app_launcher = AppLauncher(headless=args_cli.headless) |
| | simulation_app = app_launcher.app |
| |
|
| | """Rest everything else.""" |
| |
|
| | from collections.abc import Sequence |
| |
|
| | import gymnasium as gym |
| | import torch |
| | import warp as wp |
| |
|
| | from isaaclab.assets.rigid_object.rigid_object_data import RigidObjectData |
| |
|
| | import isaaclab_tasks |
| | from isaaclab_tasks.manager_based.manipulation.lift.lift_env_cfg import LiftEnvCfg |
| | from isaaclab_tasks.utils.parse_cfg import parse_env_cfg |
| |
|
| | |
| | wp.init() |
| |
|
| |
|
| | class GripperState: |
| | """States for the gripper.""" |
| |
|
| | OPEN = wp.constant(1.0) |
| | CLOSE = wp.constant(-1.0) |
| |
|
| |
|
| | class PickSmState: |
| | """States for the pick state machine.""" |
| |
|
| | REST = wp.constant(0) |
| | APPROACH_ABOVE_OBJECT = wp.constant(1) |
| | APPROACH_OBJECT = wp.constant(2) |
| | GRASP_OBJECT = wp.constant(3) |
| | LIFT_OBJECT = wp.constant(4) |
| |
|
| |
|
| | class PickSmWaitTime: |
| | """Additional wait times (in s) for states for before switching.""" |
| |
|
| | REST = wp.constant(0.2) |
| | APPROACH_ABOVE_OBJECT = wp.constant(0.5) |
| | APPROACH_OBJECT = wp.constant(0.6) |
| | GRASP_OBJECT = wp.constant(0.3) |
| | LIFT_OBJECT = wp.constant(1.0) |
| |
|
| |
|
| | @wp.func |
| | def distance_below_threshold(current_pos: wp.vec3, desired_pos: wp.vec3, threshold: float) -> bool: |
| | return wp.length(current_pos - desired_pos) < threshold |
| |
|
| |
|
| | @wp.kernel |
| | def infer_state_machine( |
| | dt: wp.array(dtype=float), |
| | sm_state: wp.array(dtype=int), |
| | sm_wait_time: wp.array(dtype=float), |
| | ee_pose: wp.array(dtype=wp.transform), |
| | object_pose: wp.array(dtype=wp.transform), |
| | des_object_pose: wp.array(dtype=wp.transform), |
| | des_ee_pose: wp.array(dtype=wp.transform), |
| | gripper_state: wp.array(dtype=float), |
| | offset: wp.array(dtype=wp.transform), |
| | position_threshold: float, |
| | ): |
| | |
| | tid = wp.tid() |
| | |
| | state = sm_state[tid] |
| | |
| | if state == PickSmState.REST: |
| | des_ee_pose[tid] = ee_pose[tid] |
| | gripper_state[tid] = GripperState.OPEN |
| | |
| | if sm_wait_time[tid] >= PickSmWaitTime.REST: |
| | |
| | sm_state[tid] = PickSmState.APPROACH_ABOVE_OBJECT |
| | sm_wait_time[tid] = 0.0 |
| | elif state == PickSmState.APPROACH_ABOVE_OBJECT: |
| | des_ee_pose[tid] = wp.transform_multiply(offset[tid], object_pose[tid]) |
| | gripper_state[tid] = GripperState.OPEN |
| | if distance_below_threshold( |
| | wp.transform_get_translation(ee_pose[tid]), |
| | wp.transform_get_translation(des_ee_pose[tid]), |
| | position_threshold, |
| | ): |
| | |
| | if sm_wait_time[tid] >= PickSmWaitTime.APPROACH_OBJECT: |
| | |
| | sm_state[tid] = PickSmState.APPROACH_OBJECT |
| | sm_wait_time[tid] = 0.0 |
| | elif state == PickSmState.APPROACH_OBJECT: |
| | des_ee_pose[tid] = object_pose[tid] |
| | gripper_state[tid] = GripperState.OPEN |
| | if distance_below_threshold( |
| | wp.transform_get_translation(ee_pose[tid]), |
| | wp.transform_get_translation(des_ee_pose[tid]), |
| | position_threshold, |
| | ): |
| | if sm_wait_time[tid] >= PickSmWaitTime.APPROACH_OBJECT: |
| | |
| | sm_state[tid] = PickSmState.GRASP_OBJECT |
| | sm_wait_time[tid] = 0.0 |
| | elif state == PickSmState.GRASP_OBJECT: |
| | des_ee_pose[tid] = object_pose[tid] |
| | gripper_state[tid] = GripperState.CLOSE |
| | |
| | if sm_wait_time[tid] >= PickSmWaitTime.GRASP_OBJECT: |
| | |
| | sm_state[tid] = PickSmState.LIFT_OBJECT |
| | sm_wait_time[tid] = 0.0 |
| | elif state == PickSmState.LIFT_OBJECT: |
| | des_ee_pose[tid] = des_object_pose[tid] |
| | gripper_state[tid] = GripperState.CLOSE |
| | if distance_below_threshold( |
| | wp.transform_get_translation(ee_pose[tid]), |
| | wp.transform_get_translation(des_ee_pose[tid]), |
| | position_threshold, |
| | ): |
| | |
| | if sm_wait_time[tid] >= PickSmWaitTime.LIFT_OBJECT: |
| | |
| | sm_state[tid] = PickSmState.LIFT_OBJECT |
| | sm_wait_time[tid] = 0.0 |
| | |
| | sm_wait_time[tid] = sm_wait_time[tid] + dt[tid] |
| |
|
| |
|
| | class PickAndLiftSm: |
| | """A simple state machine in a robot's task space to pick and lift an object. |
| | |
| | The state machine is implemented as a warp kernel. It takes in the current state of |
| | the robot's end-effector and the object, and outputs the desired state of the robot's |
| | end-effector and the gripper. The state machine is implemented as a finite state |
| | machine with the following states: |
| | |
| | 1. REST: The robot is at rest. |
| | 2. APPROACH_ABOVE_OBJECT: The robot moves above the object. |
| | 3. APPROACH_OBJECT: The robot moves to the object. |
| | 4. GRASP_OBJECT: The robot grasps the object. |
| | 5. LIFT_OBJECT: The robot lifts the object to the desired pose. This is the final state. |
| | """ |
| |
|
| | def __init__(self, dt: float, num_envs: int, device: torch.device | str = "cpu", position_threshold=0.01): |
| | """Initialize the state machine. |
| | |
| | Args: |
| | dt: The environment time step. |
| | num_envs: The number of environments to simulate. |
| | device: The device to run the state machine on. |
| | """ |
| | |
| | self.dt = float(dt) |
| | self.num_envs = num_envs |
| | self.device = device |
| | self.position_threshold = position_threshold |
| | |
| | self.sm_dt = torch.full((self.num_envs,), self.dt, device=self.device) |
| | self.sm_state = torch.full((self.num_envs,), 0, dtype=torch.int32, device=self.device) |
| | self.sm_wait_time = torch.zeros((self.num_envs,), device=self.device) |
| |
|
| | |
| | self.des_ee_pose = torch.zeros((self.num_envs, 7), device=self.device) |
| | self.des_gripper_state = torch.full((self.num_envs,), 0.0, device=self.device) |
| |
|
| | |
| | self.offset = torch.zeros((self.num_envs, 7), device=self.device) |
| | self.offset[:, 2] = 0.1 |
| | self.offset[:, -1] = 1.0 |
| |
|
| | |
| | self.sm_dt_wp = wp.from_torch(self.sm_dt, wp.float32) |
| | self.sm_state_wp = wp.from_torch(self.sm_state, wp.int32) |
| | self.sm_wait_time_wp = wp.from_torch(self.sm_wait_time, wp.float32) |
| | self.des_ee_pose_wp = wp.from_torch(self.des_ee_pose, wp.transform) |
| | self.des_gripper_state_wp = wp.from_torch(self.des_gripper_state, wp.float32) |
| | self.offset_wp = wp.from_torch(self.offset, wp.transform) |
| |
|
| | def reset_idx(self, env_ids: Sequence[int] = None): |
| | """Reset the state machine.""" |
| | if env_ids is None: |
| | env_ids = slice(None) |
| | self.sm_state[env_ids] = 0 |
| | self.sm_wait_time[env_ids] = 0.0 |
| |
|
| | def compute(self, ee_pose: torch.Tensor, object_pose: torch.Tensor, des_object_pose: torch.Tensor) -> torch.Tensor: |
| | """Compute the desired state of the robot's end-effector and the gripper.""" |
| | |
| | ee_pose = ee_pose[:, [0, 1, 2, 4, 5, 6, 3]] |
| | object_pose = object_pose[:, [0, 1, 2, 4, 5, 6, 3]] |
| | des_object_pose = des_object_pose[:, [0, 1, 2, 4, 5, 6, 3]] |
| |
|
| | |
| | ee_pose_wp = wp.from_torch(ee_pose.contiguous(), wp.transform) |
| | object_pose_wp = wp.from_torch(object_pose.contiguous(), wp.transform) |
| | des_object_pose_wp = wp.from_torch(des_object_pose.contiguous(), wp.transform) |
| |
|
| | |
| | wp.launch( |
| | kernel=infer_state_machine, |
| | dim=self.num_envs, |
| | inputs=[ |
| | self.sm_dt_wp, |
| | self.sm_state_wp, |
| | self.sm_wait_time_wp, |
| | ee_pose_wp, |
| | object_pose_wp, |
| | des_object_pose_wp, |
| | self.des_ee_pose_wp, |
| | self.des_gripper_state_wp, |
| | self.offset_wp, |
| | self.position_threshold, |
| | ], |
| | device=self.device, |
| | ) |
| |
|
| | |
| | des_ee_pose = self.des_ee_pose[:, [0, 1, 2, 6, 3, 4, 5]] |
| | |
| | return torch.cat([des_ee_pose, self.des_gripper_state.unsqueeze(-1)], dim=-1) |
| |
|
| |
|
| | def main(): |
| | |
| | env_cfg: LiftEnvCfg = parse_env_cfg( |
| | "Isaac-Lift-Cube-Franka-IK-Abs-v0", |
| | device=args_cli.device, |
| | num_envs=args_cli.num_envs, |
| | use_fabric=not args_cli.disable_fabric, |
| | ) |
| | |
| | env = gym.make("Isaac-Lift-Cube-Franka-IK-Abs-v0", cfg=env_cfg) |
| | |
| | env.reset() |
| |
|
| | |
| | actions = torch.zeros(env.unwrapped.action_space.shape, device=env.unwrapped.device) |
| | actions[:, 3] = 1.0 |
| | |
| | desired_orientation = torch.zeros((env.unwrapped.num_envs, 4), device=env.unwrapped.device) |
| | desired_orientation[:, 1] = 1.0 |
| | |
| | pick_sm = PickAndLiftSm( |
| | env_cfg.sim.dt * env_cfg.decimation, env.unwrapped.num_envs, env.unwrapped.device, position_threshold=0.01 |
| | ) |
| |
|
| | while simulation_app.is_running(): |
| | |
| | with torch.inference_mode(): |
| | |
| | dones = env.step(actions)[-2] |
| |
|
| | |
| | |
| | ee_frame_sensor = env.unwrapped.scene["ee_frame"] |
| | tcp_rest_position = ee_frame_sensor.data.target_pos_w[..., 0, :].clone() - env.unwrapped.scene.env_origins |
| | tcp_rest_orientation = ee_frame_sensor.data.target_quat_w[..., 0, :].clone() |
| | |
| | object_data: RigidObjectData = env.unwrapped.scene["object"].data |
| | object_position = object_data.root_pos_w - env.unwrapped.scene.env_origins |
| | |
| | desired_position = env.unwrapped.command_manager.get_command("object_pose")[..., :3] |
| |
|
| | |
| | actions = pick_sm.compute( |
| | torch.cat([tcp_rest_position, tcp_rest_orientation], dim=-1), |
| | torch.cat([object_position, desired_orientation], dim=-1), |
| | torch.cat([desired_position, desired_orientation], dim=-1), |
| | ) |
| |
|
| | |
| | if dones.any(): |
| | pick_sm.reset_idx(dones.nonzero(as_tuple=False).squeeze(-1)) |
| |
|
| | |
| | env.close() |
| |
|
| |
|
| | if __name__ == "__main__": |
| | |
| | main() |
| | |
| | simulation_app.close() |
| |
|