#!/usr/bin/env python3 """ Visualization script for collected tactile manipulation data. Displays: 1. Camera images (agentview + eye_in_hand) - synchronized with control freq 2. Real-time tactile force distribution heatmaps (4x4 per finger) 3. Force magnitude time series Can visualize from: - Saved HDF5 data files (offline) - Live collection (real-time) Usage: # Visualize saved data python visualize_data.py --data_file ./tactile_data/precision_grasp_data.hdf5 --episode 0 # Live visualization during collection python visualize_data.py --task precision_grasp --live """ import os import sys import argparse import numpy as np import h5py sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) def visualize_offline(data_file, episode_idx=0, playback_speed=1.0): """ Visualize saved episode data from HDF5 file. Shows camera images and tactile heatmaps side by side. """ import matplotlib.pyplot as plt from matplotlib.gridspec import GridSpec from matplotlib.colors import Normalize from matplotlib import cm with h5py.File(data_file, "r") as f: meta = f["metadata"] print(f"Task: {meta.attrs['task']}") print(f"Robot: {meta.attrs['robot']}, Gripper: {meta.attrs['gripper']}") print(f"Tactile sensor: {meta.attrs['tactile_sensor']}") print(f"Control freq: {meta.attrs['control_freq']} Hz, " f"Tactile freq: {meta.attrs['tactile_freq']} Hz") # Per-episode HDF5: data at root level ep = f print(f"\nEpisode: " f"steps={ep.attrs['n_steps']}, " f"success={ep.attrs['success']}") # Load data agentview = ep["agentview_image"][:] if "agentview_image" in ep else None eye_in_hand = ep["eye_in_hand_image"][:] if "eye_in_hand_image" in ep else None tactile_left = ep["tactile_left"][:] if "tactile_left" in ep else None tactile_right = ep["tactile_right"][:] if "tactile_right" in ep else None rewards = ep["rewards"][:] if "rewards" in ep else None eef_pos = ep["eef_pos"][:] if "eef_pos" in ep else None n_frames = len(agentview) if agentview is not None else 0 tactile_ratio = 5 # tactile_freq / control_freq if n_frames == 0: print("No image data to visualize.") return print(f"Frames: {n_frames}, Tactile samples: {len(tactile_left) if tactile_left is not None else 0}") # Setup figure fig = plt.figure(figsize=(16, 10)) gs = GridSpec(3, 4, figure=fig, hspace=0.35, wspace=0.3) # Camera views ax_agent = fig.add_subplot(gs[0:2, 0:2]) ax_agent.set_title("AgentView Camera", fontsize=12, fontweight="bold") ax_agent.axis("off") ax_hand = fig.add_subplot(gs[0:2, 2:4]) ax_hand.set_title("Eye-in-Hand Camera", fontsize=12, fontweight="bold") ax_hand.axis("off") # Tactile heatmaps ax_tleft = fig.add_subplot(gs[2, 0]) ax_tleft.set_title("Left Finger Tactile", fontsize=10, fontweight="bold") ax_tright = fig.add_subplot(gs[2, 1]) ax_tright.set_title("Right Finger Tactile", fontsize=10, fontweight="bold") # Force time series ax_force = fig.add_subplot(gs[2, 2:4]) ax_force.set_title("Tactile Force Magnitude", fontsize=10, fontweight="bold") ax_force.set_xlabel("Step") ax_force.set_ylabel("Force (N)") # Precompute force magnitudes for time series if tactile_left is not None and tactile_right is not None: left_mag_all = np.linalg.norm(tactile_left, axis=-1).mean(axis=(1, 2)) right_mag_all = np.linalg.norm(tactile_right, axis=-1).mean(axis=(1, 2)) else: left_mag_all = np.zeros(1) right_mag_all = np.zeros(1) # Color normalization for tactile vmax = max(left_mag_all.max(), right_mag_all.max(), 0.1) # Initial display img_agent = ax_agent.imshow(agentview[0] if agentview is not None else np.zeros((256, 256, 3), dtype=np.uint8)) img_hand = ax_hand.imshow(eye_in_hand[0] if eye_in_hand is not None else np.zeros((256, 256, 3), dtype=np.uint8)) # Tactile heatmaps if tactile_left is not None: left_mag = np.linalg.norm(tactile_left[0], axis=-1) right_mag = np.linalg.norm(tactile_right[0], axis=-1) else: left_mag = np.zeros((4, 4)) right_mag = np.zeros((4, 4)) hm_left = ax_tleft.imshow(left_mag, cmap="hot", vmin=0, vmax=vmax, interpolation="nearest", aspect="equal") hm_right = ax_tright.imshow(right_mag, cmap="hot", vmin=0, vmax=vmax, interpolation="nearest", aspect="equal") plt.colorbar(hm_left, ax=ax_tleft, fraction=0.046) plt.colorbar(hm_right, ax=ax_tright, fraction=0.046) # Force value text annotations on heatmaps left_texts = [] right_texts = [] for ri in range(4): for ci in range(4): val = left_mag[ri, ci] color = "white" if val > vmax * 0.5 else "black" t = ax_tleft.text(ci, ri, f"{val:.1f}", ha="center", va="center", fontsize=6, color=color, fontweight="bold") left_texts.append(t) val = right_mag[ri, ci] color = "white" if val > vmax * 0.5 else "black" t = ax_tright.text(ci, ri, f"{val:.1f}", ha="center", va="center", fontsize=6, color=color, fontweight="bold") right_texts.append(t) # Add taxel grid labels for ax in [ax_tleft, ax_tright]: ax.set_xticks(range(4)) ax.set_yticks(range(4)) ax.set_xticklabels([f"c{i}" for i in range(4)], fontsize=7) ax.set_yticklabels([f"r{i}" for i in range(4)], fontsize=7) # Force time series line_left, = ax_force.plot([], [], "b-", label="Left finger", linewidth=1) line_right, = ax_force.plot([], [], "r-", label="Right finger", linewidth=1) ax_force.legend(fontsize=8) ax_force.set_xlim(0, len(left_mag_all)) ax_force.set_ylim(0, vmax * 1.1) vline = ax_force.axvline(x=0, color="gray", linestyle="--", alpha=0.5) # Plot full force series line_left.set_data(range(len(left_mag_all)), left_mag_all) line_right.set_data(range(len(right_mag_all)), right_mag_all) fig.suptitle(f"Tactile Manipulation Data Viewer", fontsize=14, fontweight="bold") # Animation delay = (1.0 / 20.0) / playback_speed # 20 Hz control freq plt.ion() plt.show() try: for frame_idx in range(n_frames): # Update camera images if agentview is not None: img_agent.set_data(agentview[frame_idx]) if eye_in_hand is not None: img_hand.set_data(eye_in_hand[frame_idx]) # Update tactile heatmaps (show the last sub-sample for this frame) if tactile_left is not None: t_idx = min(frame_idx * tactile_ratio + tactile_ratio - 1, len(tactile_left) - 1) left_mag = np.linalg.norm(tactile_left[t_idx], axis=-1) right_mag = np.linalg.norm(tactile_right[t_idx], axis=-1) hm_left.set_data(left_mag) hm_right.set_data(right_mag) # Update force value annotations for ri in range(4): for ci in range(4): idx = ri * 4 + ci lv = left_mag[ri, ci] left_texts[idx].set_text(f"{lv:.1f}") left_texts[idx].set_color("white" if lv > vmax * 0.5 else "black") rv = right_mag[ri, ci] right_texts[idx].set_text(f"{rv:.1f}") right_texts[idx].set_color("white" if rv > vmax * 0.5 else "black") # Update time marker vline.set_xdata([frame_idx * tactile_ratio]) # Update title with step info reward_str = f", Reward: {rewards[frame_idx]:.3f}" if rewards is not None else "" fig.suptitle( f"Step {frame_idx}/{n_frames}{reward_str}", fontsize=14, fontweight="bold" ) fig.canvas.draw_idle() fig.canvas.flush_events() plt.pause(delay) except KeyboardInterrupt: pass plt.ioff() print("\nVisualization complete. Close the window to exit.") plt.show() def visualize_live(task_name, n_episodes=1): """ Live visualization during data collection. Renders the MuJoCo scene and shows real-time tactile force distributions. """ import matplotlib matplotlib.use("TkAgg") import matplotlib.pyplot as plt from matplotlib.gridspec import GridSpec from tactile_tasks.uskin_sensor import USkinSensor from tactile_tasks.motion_planner import MotionPlanner from tactile_tasks.collect_data import create_env, TASK_CONFIGS, collect_episode config = TASK_CONFIGS[task_name] # Create env with renderer env = create_env(task_name, has_renderer=True) obs = env.reset() tactile = USkinSensor(env.sim, gripper_prefix="gripper0_right_", noise_std=0.02) planner = MotionPlanner(env, tactile_sensor=tactile) # Setup tactile visualization figure fig, axes = plt.subplots(1, 3, figsize=(14, 4)) ax_left = axes[0] ax_left.set_title("Left Finger Tactile (4x4)", fontweight="bold") hm_left = ax_left.imshow(np.zeros((4, 4)), cmap="hot", vmin=0, vmax=2.0, interpolation="nearest", aspect="equal") plt.colorbar(hm_left, ax=ax_left, label="Force (N)") ax_right = axes[1] ax_right.set_title("Right Finger Tactile (4x4)", fontweight="bold") hm_right = ax_right.imshow(np.zeros((4, 4)), cmap="hot", vmin=0, vmax=2.0, interpolation="nearest", aspect="equal") plt.colorbar(hm_right, ax=ax_right, label="Force (N)") ax_force = axes[2] ax_force.set_title("Force History", fontweight="bold") ax_force.set_xlabel("Step") ax_force.set_ylabel("Avg Force (N)") left_history = [] right_history = [] line_l, = ax_force.plot([], [], "b-", label="Left") line_r, = ax_force.plot([], [], "r-", label="Right") ax_force.legend() ax_force.set_ylim(0, 3) for ax in [ax_left, ax_right]: ax.set_xticks(range(4)) ax.set_yticks(range(4)) plt.ion() plt.tight_layout() plt.show() # Run collection with live visualization plan_fn = config["plan_fn"] phases = plan_fn(planner, env) current_phase_idx = 0 phase_name, phase_init = phases[current_phase_idx] phase_init() print(f"Phase: {phase_name}") step = 0 try: while step < config["horizon"]: action, phase_done = planner.get_action() # Update tactile for _ in range(USkinSensor.FREQ_MULTIPLIER): tactile_data = tactile.update() # Step environment obs, reward, done, info = env.step(action) # Render env.render() # Update tactile visualization mags = tactile.get_force_magnitudes() hm_left.set_data(mags["left_finger"]) hm_right.set_data(mags["right_finger"]) # Update vmax dynamically cur_max = max(mags["left_finger"].max(), mags["right_finger"].max(), 0.5) hm_left.set_clim(0, cur_max) hm_right.set_clim(0, cur_max) # Update force history left_history.append(mags["left_finger"].mean()) right_history.append(mags["right_finger"].mean()) line_l.set_data(range(len(left_history)), left_history) line_r.set_data(range(len(right_history)), right_history) ax_force.set_xlim(0, max(len(left_history), 10)) ax_force.set_ylim(0, max(max(left_history + [0.5]), max(right_history + [0.5])) * 1.2) fig.suptitle(f"Step {step} | Phase: {phase_name} | Reward: {reward:.3f}", fontweight="bold") fig.canvas.draw_idle() fig.canvas.flush_events() step += 1 if phase_done: current_phase_idx += 1 if current_phase_idx < len(phases): phase_name, phase_init = phases[current_phase_idx] phase_init() print(f"Phase: {phase_name}") else: print("All phases complete!") break if done: break except KeyboardInterrupt: pass print(f"Episode done. Steps: {step}, Success: {env._check_success()}") plt.ioff() plt.show() env.close() def generate_video(data_file, episode_idx=None, output_path=None, fps=20, show_tactile=True): """ Generate MP4 video from a per-episode HDF5 file using ffmpeg. Args: data_file (str): path to per-episode HDF5 file episode_idx: deprecated, ignored output_path (str): output video path (default: auto-generated) fps (int): frames per second show_tactile (bool): whether to include tactile heatmaps """ import subprocess import tempfile import shutil try: import matplotlib matplotlib.use("Agg") # non-interactive backend for rendering import matplotlib.pyplot as plt from matplotlib.gridspec import GridSpec except ImportError: print("matplotlib is required for video generation") return # Check ffmpeg availability if shutil.which("ffmpeg") is None: print("ERROR: ffmpeg not found. Install it: sudo apt install ffmpeg") return with h5py.File(data_file, "r") as f: meta = f["metadata"] task_name = meta.attrs["task"] # Per-episode HDF5: data at root level success = f.attrs.get("success", False) n_steps = f.attrs.get("n_steps", 0) agentview = f["agentview_image"][:] if "agentview_image" in f else None eye_in_hand = f["eye_in_hand_image"][:] if "eye_in_hand_image" in f else None tactile_left = f["tactile_left"][:] if "tactile_left" in f else None tactile_right = f["tactile_right"][:] if "tactile_right" in f else None rewards = f["rewards"][:] if "rewards" in f else None if agentview is None or len(agentview) == 0: print("No image data to render.") return n_frames = len(agentview) tactile_ratio = 5 if output_path is None: base = os.path.splitext(data_file)[0] output_path = f"{base}_ep{episode_idx:04d}.mp4" print(f"Generating video: {output_path}") print(f" Task: {task_name}, Episode: {episode_idx}, Steps: {n_steps}, Success: {success}") print(f" Frames: {n_frames}, FPS: {fps}") # Precompute tactile force magnitudes for color scale if show_tactile and tactile_left is not None: left_mag_all = np.linalg.norm(tactile_left, axis=-1) right_mag_all = np.linalg.norm(tactile_right, axis=-1) vmax = max(left_mag_all.max(), right_mag_all.max(), 0.1) left_avg = left_mag_all.mean(axis=(1, 2)) right_avg = right_mag_all.mean(axis=(1, 2)) else: show_tactile = False # Determine figure layout if show_tactile: fig = plt.figure(figsize=(12, 8), dpi=100) gs = GridSpec(3, 4, figure=fig, hspace=0.3, wspace=0.3) ax_agent = fig.add_subplot(gs[0:2, 0:2]) ax_hand = fig.add_subplot(gs[0:2, 2:4]) ax_tleft = fig.add_subplot(gs[2, 0]) ax_tright = fig.add_subplot(gs[2, 1]) ax_force = fig.add_subplot(gs[2, 2:4]) else: fig, axes = plt.subplots(1, 2, figsize=(10, 5), dpi=100) ax_agent, ax_hand = axes # Use a temp directory for frames tmpdir = tempfile.mkdtemp(prefix="tactile_video_") try: for frame_idx in range(n_frames): # Clear and redraw each frame if show_tactile: for ax in [ax_agent, ax_hand, ax_tleft, ax_tright, ax_force]: ax.clear() else: ax_agent.clear() ax_hand.clear() # Camera images ax_agent.imshow(agentview[frame_idx]) ax_agent.set_title("AgentView", fontsize=10, fontweight="bold") ax_agent.axis("off") if eye_in_hand is not None and len(eye_in_hand) > frame_idx: ax_hand.imshow(eye_in_hand[frame_idx]) ax_hand.set_title("Eye-in-Hand", fontsize=10, fontweight="bold") ax_hand.axis("off") if show_tactile: t_idx = min(frame_idx * tactile_ratio + tactile_ratio - 1, len(tactile_left) - 1) left_mag = np.linalg.norm(tactile_left[t_idx], axis=-1) right_mag = np.linalg.norm(tactile_right[t_idx], axis=-1) ax_tleft.imshow(left_mag, cmap="hot", vmin=0, vmax=vmax, interpolation="nearest") ax_tleft.set_title("Left Finger", fontsize=9) ax_tleft.set_xticks(range(4)) ax_tleft.set_yticks(range(4)) # Annotate each taxel with force value for ri in range(4): for ci in range(4): val = left_mag[ri, ci] color = "white" if val > vmax * 0.5 else "black" ax_tleft.text(ci, ri, f"{val:.1f}", ha="center", va="center", fontsize=6, color=color, fontweight="bold") ax_tright.imshow(right_mag, cmap="hot", vmin=0, vmax=vmax, interpolation="nearest") ax_tright.set_title("Right Finger", fontsize=9) ax_tright.set_xticks(range(4)) ax_tright.set_yticks(range(4)) # Annotate each taxel with force value for ri in range(4): for ci in range(4): val = right_mag[ri, ci] color = "white" if val > vmax * 0.5 else "black" ax_tright.text(ci, ri, f"{val:.1f}", ha="center", va="center", fontsize=6, color=color, fontweight="bold") # Force time series up to current frame t_end = min((frame_idx + 1) * tactile_ratio, len(left_avg)) ax_force.plot(range(t_end), left_avg[:t_end], "b-", label="Left", linewidth=1) ax_force.plot(range(t_end), right_avg[:t_end], "r-", label="Right", linewidth=1) ax_force.axvline(x=t_end - 1, color="gray", linestyle="--", alpha=0.5) ax_force.set_xlim(0, len(left_avg)) ax_force.set_ylim(0, vmax * 1.1) ax_force.set_xlabel("Tactile Sample", fontsize=8) ax_force.set_ylabel("Force (N)", fontsize=8) ax_force.legend(fontsize=7, loc="upper right") ax_force.set_title("Tactile Force", fontsize=9) # Title reward_str = f" R={rewards[frame_idx]:.2f}" if rewards is not None else "" status = "SUCCESS" if success else "RUNNING" fig.suptitle( f"{task_name} | Step {frame_idx+1}/{n_frames}{reward_str} | {status}", fontsize=12, fontweight="bold" ) frame_path = os.path.join(tmpdir, f"frame_{frame_idx:06d}.png") fig.savefig(frame_path, bbox_inches="tight", pad_inches=0.1) if frame_idx % 50 == 0: print(f" Rendered frame {frame_idx+1}/{n_frames}") plt.close(fig) # Combine frames with ffmpeg print(f" Encoding video with ffmpeg...") ffmpeg_cmd = [ "ffmpeg", "-y", "-framerate", str(fps), "-i", os.path.join(tmpdir, "frame_%06d.png"), "-vf", "pad=ceil(iw/2)*2:ceil(ih/2)*2", # ensure even dimensions for H.264 "-c:v", "libx264", "-pix_fmt", "yuv420p", "-crf", "23", "-preset", "medium", output_path, ] result = subprocess.run(ffmpeg_cmd, capture_output=True, text=True) if result.returncode != 0: print(f" ffmpeg error: {result.stderr[:500]}") else: file_size = os.path.getsize(output_path) / 1024 print(f" Video saved: {output_path} ({file_size:.0f} KB)") finally: shutil.rmtree(tmpdir, ignore_errors=True) def generate_all_videos(data_dir, output_dir=None, fps=20, show_tactile=True): """Generate videos for all episodes in all HDF5 files in a directory.""" if output_dir is None: output_dir = os.path.join(data_dir, "videos") os.makedirs(output_dir, exist_ok=True) hdf5_files = sorted([f for f in os.listdir(data_dir) if f.endswith(".hdf5")]) if not hdf5_files: print(f"No HDF5 files found in {data_dir}") return for hdf5_file in hdf5_files: filepath = os.path.join(data_dir, hdf5_file) with h5py.File(filepath, "r") as f: episodes = sorted([k for k in f.keys() if k.startswith("episode_")]) task_name = os.path.splitext(hdf5_file)[0].replace("_data", "") for ep_name in episodes: ep_idx = int(ep_name.split("_")[1]) output_path = os.path.join(output_dir, f"{task_name}_{ep_name}.mp4") generate_video(filepath, ep_idx, output_path, fps, show_tactile) print(f"\nAll videos saved to: {output_dir}") def print_data_summary(data_file): """Print summary of saved data file.""" with h5py.File(data_file, "r") as f: print(f"\nData file: {data_file}") print(f"{'='*50}") if "metadata" in f: meta = f["metadata"] for key in meta.attrs: print(f" {key}: {meta.attrs[key]}") episodes = [k for k in f.keys() if k.startswith("episode_")] print(f"\nEpisodes: {len(episodes)}") for ep_name in sorted(episodes): ep = f[ep_name] print(f"\n {ep_name}:") print(f" Success: {ep.attrs.get('success', 'N/A')}") print(f" Steps: {ep.attrs.get('n_steps', 'N/A')}") for key in ep: shape = ep[key].shape dtype = ep[key].dtype print(f" {key}: shape={shape}, dtype={dtype}") def main(): parser = argparse.ArgumentParser(description="Visualize tactile manipulation data") parser.add_argument("--data_file", type=str, default=None, help="HDF5 data file to visualize") parser.add_argument("--episode", type=int, default=0, help="Episode index to visualize") parser.add_argument("--speed", type=float, default=1.0, help="Playback speed multiplier") parser.add_argument("--task", type=str, default="precision_grasp", help="Task name for live visualization") parser.add_argument("--live", action="store_true", help="Live visualization during collection") parser.add_argument("--summary", action="store_true", help="Print data file summary") parser.add_argument("--video", action="store_true", help="Generate MP4 video from saved data") parser.add_argument("--video_all", type=str, default=None, help="Generate videos for all episodes in data directory") parser.add_argument("--output", type=str, default=None, help="Output video file path") parser.add_argument("--fps", type=int, default=20, help="Video frames per second") parser.add_argument("--no_tactile", action="store_true", help="Exclude tactile data from video") args = parser.parse_args() if args.video_all: generate_all_videos(args.video_all, fps=args.fps, show_tactile=not args.no_tactile) elif args.video and args.data_file: generate_video(args.data_file, args.episode, args.output, args.fps, not args.no_tactile) elif args.summary and args.data_file: print_data_summary(args.data_file) elif args.live: visualize_live(args.task) elif args.data_file: visualize_offline(args.data_file, args.episode, args.speed) else: print("Usage:") print(" Offline viz: python visualize_data.py --data_file data.hdf5 --episode 0") print(" Generate video: python visualize_data.py --data_file data.hdf5 --episode 0 --video") print(" All videos: python visualize_data.py --video_all ./tactile_data/") print(" Summary: python visualize_data.py --data_file data.hdf5 --summary") print(" Live viz: python visualize_data.py --task precision_grasp --live") if __name__ == "__main__": main()