| import numpy as np |
| import pybullet as p |
| from surrol.gym.surrol_env import SurRoLEnv, RENDER_HEIGHT |
| from surrol.robots.ecm import Ecm |
| from surrol.utils.pybullet_utils import ( |
| get_link_pose, |
| reset_camera |
| ) |
| from surrol.utils.robotics import get_pose_2d_from_matrix |
|
|
|
|
| def goal_distance(goal_a, goal_b): |
| assert goal_a.shape == goal_b.shape |
| return np.linalg.norm(goal_a - goal_b, axis=-1) |
|
|
|
|
| class EcmEnv(SurRoLEnv): |
| """ |
| Single arm env using ECM (active_track is not a GoalEnv) |
| Refer to Gym fetch |
| https://github.com/openai/gym/blob/master/gym/envs/robotics/fetch_env.py |
| ravens |
| https://github.com/google-research/ravens/blob/master/ravens/environments/environment.py |
| """ |
| ACTION_SIZE = 3 |
| ACTION_MODE = 'cVc' |
| DISTANCE_THRESHOLD = 0.005 |
| POSE_ECM = ((0.15, 0.0, 0.7524), (0, 30 / 180 * np.pi, 0)) |
| QPOS_ECM = (0, 0.6, 0.04, 0) |
| WORKSPACE_LIMITS = ((0.45, 0.55), (-0.05, 0.05), (0.60, 0.70)) |
| SCALING = 1. |
|
|
| def __init__(self, render_mode: str = None, cid = -1): |
| |
| self.workspace_limits = np.asarray(self.WORKSPACE_LIMITS) |
| self.workspace_limits *= self.SCALING |
|
|
| |
| self.use_camera = False |
|
|
| |
| self.has_object = False |
| self.obj_id = None |
|
|
| super(EcmEnv, self).__init__(render_mode, cid) |
|
|
| |
| self._duration = 0.1 |
|
|
| |
| self.distance_threshold = self.DISTANCE_THRESHOLD * self.SCALING |
|
|
| |
| self._view_matrix = p.computeViewMatrixFromYawPitchRoll( |
| cameraTargetPosition=(0.27 * self.SCALING, -0.20 * self.SCALING, 0.55 * self.SCALING), |
| distance=1.80 * self.SCALING, |
| yaw=150, |
| pitch=-30, |
| roll=0, |
| upAxisIndex=2 |
| ) |
|
|
| def render(self, mode='rgb_array'): |
| |
| if mode != "rgb_array": |
| return np.array([]) |
| rgb_array = super().render(mode) |
| if self.use_camera: |
| rgb_cam, _ = self.ecm.render_image(RENDER_HEIGHT, RENDER_HEIGHT) |
| rgb_array = np.concatenate((rgb_array, rgb_cam), axis=1) |
| return rgb_array |
|
|
| def compute_reward(self, achieved_goal: np.ndarray, desired_goal: np.ndarray, info): |
| """ Sparse reward. """ |
| |
| |
| return self._is_success(achieved_goal, desired_goal).astype(np.float32) - 1. |
|
|
| def _env_setup(self): |
| assert self.ACTION_MODE in ('cVc', 'dmove', 'droll') |
| |
| if self._render_mode == 'human': |
| reset_camera(yaw=150.0, pitch=-30.0, dist=1.50 * self.SCALING, |
| target=(0.27 * self.SCALING, -0.20 * self.SCALING, 0.55 * self.SCALING)) |
|
|
| |
| self.ecm = Ecm(self.POSE_ECM[0], p.getQuaternionFromEuler(self.POSE_ECM[1]), |
| scaling=self.SCALING) |
|
|
| pass |
| |
|
|
| def _get_robot_state(self) -> np.ndarray: |
| |
| |
| pose_rcm = get_pose_2d_from_matrix(self.ecm.get_current_position()) |
| return np.concatenate([ |
| np.array(pose_rcm[0]), np.array(p.getEulerFromQuaternion(pose_rcm[1])) |
| ]) |
|
|
| def _get_obs(self) -> dict: |
| robot_state = self._get_robot_state() |
| |
| if self.has_object: |
| pos, _ = get_link_pose(self.obj_id, -1) |
| object_pos = np.array(pos) |
| else: |
| object_pos = np.zeros(0) |
|
|
| if self.has_object: |
| achieved_goal = object_pos.copy() |
| else: |
| achieved_goal = np.array(get_link_pose(self.ecm.body, self.ecm.EEF_LINK_INDEX)[0]) |
|
|
| observation = np.concatenate([ |
| robot_state, object_pos.ravel(), |
| ]) |
| obs = { |
| 'observation': observation.copy(), |
| 'achieved_goal': achieved_goal.copy(), |
| 'desired_goal': self.goal.copy() |
| } |
| return obs |
|
|
| def _set_action(self, action: np.ndarray): |
| """ |
| delta_position (3) and delta_theta (1); in world coordinate |
| """ |
| |
| assert len(action) == self.ACTION_SIZE |
| action = action.copy() |
| if self.ACTION_MODE == 'cVc': |
| |
| |
| |
| action *= 0.01 * self.SCALING |
| dq = 0.05 * self.ecm.cVc_to_dq(action) |
| result = self.ecm.dmove_joint(dq) |
| if result is False: |
| return False |
| else: |
| return True |
| elif self.ACTION_MODE == 'dmove': |
| |
| action *= 0.01 * self.SCALING |
| pose_rcm = self.ecm.get_current_position() |
| pose_rcm[:3, 3] += action |
| pos, _ = self.ecm.pose_rcm2world(pose_rcm, 'tuple') |
| joint_positions = self.ecm.inverse_kinematics((pos, None), self.ecm.EEF_LINK_INDEX) |
| self.ecm.move_joint(joint_positions[:self.ecm.DoF]) |
| elif self.ACTION_MODE == 'droll': |
| |
| action *= np.deg2rad(3) |
| self.ecm.dmove_joint_one(action[0], 3) |
| else: |
| raise NotImplementedError |
|
|
| def _is_success(self, achieved_goal, desired_goal): |
| """ Indicates whether or not the achieved goal successfully achieved the desired goal. |
| """ |
| d = goal_distance(achieved_goal, desired_goal) |
| return (d < self.distance_threshold).astype(np.float32) |
|
|
| def _sample_goal(self) -> np.ndarray: |
| """ Samples a new goal and returns it. |
| """ |
| raise NotImplementedError |
|
|
| @property |
| def action_size(self): |
| return self.ACTION_SIZE |
|
|
| def get_oracle_action(self, obs) -> np.ndarray: |
| """ |
| Define a scripted oracle strategy |
| """ |
| raise NotImplementedError |
| |
| |
|
|