| --- |
| license: cc-by-4.0 |
| task_categories: |
| - robotics |
| - image-segmentation |
| - graph-ml |
| language: |
| - en |
| tags: |
| - robotics |
| - manipulation |
| - disassembly |
| - constraint-graph |
| - gnn |
| - world-model |
| - sam2 |
| - segmentation |
| - ur5e |
| size_categories: |
| - 1K<n<10K |
| pretty_name: GNN Disassembly World Model Dataset (v3) |
| --- |
| |
| # GNN Disassembly World Model Dataset (v3) |
|
|
| Real robot disassembly episodes with side-view **per-frame constraint graphs**, SAM2 segmentation masks, 256D feature embeddings, **full 3D depth information (point clouds)**, and synchronized robot states. The robot is labeled as a separate **agent node** with its own mask, embedding, and depth bundle. |
|
|
| **Project:** CoRL 2026 — GNN world model for constraint-aware video generation |
| **Author:** Chang Liu (Texas A&M University) |
| **Hardware:** UR5e + Robotiq 2F-85 gripper, OAK-D Pro (side view) |
| **Format version:** v3 (2026-04-10) |
|
|
| ## Dataset Structure |
|
|
| ``` |
| episode_XX/ |
| ├── metadata.json # Episode metadata, component counts, labeled frame count |
| ├── robot_states.npy # (T, 13) float32 — joint angles + TCP pose + gripper |
| ├── robot_actions.npy # (T-1, 13) float32 — frame-to-frame state deltas |
| ├── timestamps.npy # (T, 3) float64 |
| ├── side/ |
| │ ├── rgb/frame_XXXXXX.png # 1280×720 RGB (side camera) |
| │ └── depth/frame_XXXXXX.npy # 1280×720 uint16 depth (mm) |
| ├── wrist/ # Raw wrist camera data (not used in v3) |
| │ ├── rgb/... |
| │ └── depth/... |
| └── annotations/ |
| ├── side_graph.json # Constraint graph (products only, NO robot) |
| ├── side_masks/ |
| │ └── frame_XXXXXX.npz # {component_id: (H,W) uint8} — products only |
| ├── side_embeddings/ |
| │ └── frame_XXXXXX.npz # {component_id: (256,) float32} — products only |
| ├── side_depth_info/ |
| │ └── frame_XXXXXX.npz # Per-product depth bundle (flat keys) |
| ├── side_robot/ |
| │ └── frame_XXXXXX.npz # Robot bundle — ALWAYS written per labeled frame |
| └── dataset_card.json # Format description |
| ``` |
|
|
| **Alignment guarantee:** every labeled frame has files in all 4 annotation directories. Files are aligned by frame index. |
|
|
| ## Component Types (9 types) |
|
|
| **8 product types** (constraint nodes): |
|
|
| | Index | Type | Color | Notes | |
| |-------|------|-------|-------| |
| | 0 | `cpu_fan` | #FF6B6B | Always visible at start | |
| | 1 | `cpu_bracket` | #4ECDC4 | Hidden at start (under fan) | |
| | 2 | `cpu` | #45B7D1 | Hidden at start | |
| | 3 | `ram_clip` | #96CEB4 | Multiple instances: ram_clip_1, ram_clip_2, ... | |
| | 4 | `ram` | #FFEAA7 | Multiple instances: ram_1, ram_2, ... | |
| | 5 | `connector` | #DDA0DD | Multiple instances: connector_1, connector_2, ... | |
| | 6 | `graphic_card` | #FF8C42 | Always visible | |
| | 7 | `motherboard` | #8B5CF6 | Always visible (base) | |
|
|
| **1 agent type** (NOT in constraint graph): |
|
|
| | Index | Type | Color | Notes | |
| |-------|------|-------|-------| |
| | 8 | `robot` | #F5F5F5 | Labeled but stored separately. Added as agent node at training time. | |
|
|
| ## Graph Semantics |
|
|
| ### Constraint Graph (Sparse, Stored) |
|
|
| `side_graph.json` defines the **physical constraint relationships** between products. Directed edges: `A -> B` means "A must be removed before B can be removed" (A blocks B). |
|
|
| ``` |
| cpu_fan -> cpu_bracket (fan covers bracket) |
| cpu_fan -> motherboard (fan attached to board) |
| cpu_bracket -> cpu (bracket holds CPU) |
| cpu_bracket -> motherboard |
| cpu -> motherboard |
| ram_N -> motherboard |
| ram_clip_N -> motherboard |
| ram_clip_N -> ram_M (user manually pairs) |
| connector_N -> motherboard |
| graphic_card -> motherboard |
| ``` |
|
|
| **Edge states** are delta-encoded in `frame_states`: |
| - `locked: true` (1) — constraint active, component cannot be removed |
| - `locked: false` (0) — constraint released, component is free |
| - Monotonic: once unlocked, stays unlocked |
|
|
| ### Fully Connected Graph (Built at Training Time) |
|
|
| For GNN message passing, the sparse constraint graph is expanded to a **fully connected directed** graph. Every ordered pair `(i, j)` where `i != j` gets an edge. Self-loops are excluded. |
|
|
| **Edge count:** For a graph with N nodes, there are **N × (N - 1)** directed edges (both directions for every pair). |
|
|
| **Edge features (2D):** |
|
|
| | `has_constraint` | `is_locked` | Meaning | |
| |---|---|---| |
| | 1 | 1 | Directed physical constraint `i → j` exists, currently active (locked) | |
| | 1 | 0 | Directed physical constraint `i → j` exists, released (unlocked) | |
| | 0 | 0 | No physical constraint in this direction — message passing only | |
|
|
| **Direction handling is asymmetric.** The physical constraint `A → B` (A blocks B's removal) is a one-way relationship: |
| - Edge `(A, B)` → `has_constraint = 1` |
| - Edge `(B, A)` → `has_constraint = 0` (no reverse constraint; still present for message passing) |
|
|
| For example, if `cpu_fan → cpu_bracket` is a constraint: |
| ``` |
| (cpu_fan, cpu_bracket) → has_constraint=1, is_locked=1 (physical, active) |
| (cpu_bracket, cpu_fan) → has_constraint=0, is_locked=0 (message passing only) |
| ``` |
|
|
| This ensures every node pair communicates during GNN layers while still encoding the directionality of the prerequisite relationship. |
|
|
| **Robot (agent node)** has NO physical constraints. All edges involving the robot (`robot ↔ any_product`) have features `[0, 0]` — context-passing only. |
|
|
| **Node ordering:** Node indices in `edge_index` match the order of `components` in `side_graph.json`. When the robot is added (with `load_pyg_frame_with_robot`), it is appended at index `N_products` (the last position). |
|
|
| ## Data File Schemas |
|
|
| ### `side_graph.json` |
| |
| ```json |
| { |
| "view": "side", |
| "episode_id": "episode_00", |
| "goal_component": "connector_1", |
| "components": [ |
| {"id": "cpu_fan", "type": "cpu_fan", "color": "#FF6B6B"}, |
| {"id": "ram_1", "type": "ram", "color": "#FFEAA7"} |
| ], |
| "edges": [ |
| {"src": "cpu_fan", "dst": "cpu_bracket", "directed": true}, |
| {"src": "ram_clip_1", "dst": "ram_1", "directed": true} |
| ], |
| "frame_states": { |
| "0": { |
| "constraints": {"cpu_fan->cpu_bracket": true}, |
| "visibility": {"cpu_bracket": false, "cpu": false, "robot": true} |
| }, |
| "152": { |
| "constraints": {"cpu_fan->cpu_bracket": false}, |
| "visibility": {"cpu_fan": false, "cpu_bracket": true, "cpu": true} |
| } |
| }, |
| "node_positions": {"cpu_fan": [120, 80]}, |
| "embedding_dim": 256, |
| "feature_extractor": "sam2.1_hiera_base_plus", |
| "type_vocab": ["cpu_fan", "cpu_bracket", "cpu", "ram_clip", "ram", "connector", "graphic_card", "motherboard", "robot"] |
| } |
| ``` |
| |
| **Robot is NOT in components.** Robot is stored in `side_robot/`. |
|
|
| ### `side_depth_info/frame_XXXXXX.npz` |
| |
| **Always contains all 7 keys per component in `graph.components`.** Flat keys prefixed by component_id. |
|
|
| | Key | Shape | Dtype | Description | |
| |-----|-------|-------|-------------| |
| | `{cid}_point_cloud` | (N, 3) | float32 | 3D points in camera frame (meters). Empty (0, 3) if no valid depth. | |
| | `{cid}_pixel_coords` | (N, 2) | int32 | (u, v) pixel coords of valid points | |
| | `{cid}_raw_depths_mm` | (N,) | uint16 | Raw depth values in mm, filtered to [50, 2000] | |
| | `{cid}_centroid` | (3,) | float32 | Mean of point_cloud; [0,0,0] if no valid depth | |
| | `{cid}_bbox_2d` | (4,) | int32 | [x1, y1, x2, y2] from mask | |
| | `{cid}_area` | (1,) | int32 | Mask pixel count | |
| | `{cid}_depth_valid` | (1,) | uint8 | 1 if N > 0 else 0 | |
|
|
| ### `side_robot/frame_XXXXXX.npz` |
|
|
| **Always written per labeled frame** (with `visible=[0]` if robot not in this frame). |
|
|
| | Key | Shape | Dtype | Description | |
| |-----|-------|-------|-------------| |
| | `visible` | (1,) | uint8 | 1 if robot labeled, 0 otherwise | |
| | `mask` | (H, W) | uint8 | Binary mask | |
| | `embedding` | (256,) | float32 | SAM2 256D feature | |
| | `point_cloud` | (N, 3) | float32 | 3D points (meters) | |
| | `pixel_coords` | (N, 2) | int32 | (u, v) pixel coords | |
| | `raw_depths_mm` | (N,) | uint16 | Raw depths in mm | |
| | `centroid` | (3,) | float32 | Mean of point_cloud | |
| | `bbox_2d` | (4,) | int32 | From mask | |
| | `area` | (1,) | int32 | Pixel count | |
| | `depth_valid` | (1,) | uint8 | 1 if N > 0 else 0 | |
|
|
| ### `metadata.json` |
|
|
| ```json |
| { |
| "episode_id": "episode_00", |
| "goal_component": "connector_1", |
| "num_frames": 604, |
| "labeled_frame_count": 246, |
| "annotation_complete": false, |
| "component_counts": { |
| "cpu_fan": 1, "cpu_bracket": 1, "cpu": 1, |
| "ram": 2, "ram_clip": 4, "connector": 4, |
| "graphic_card": 1, "motherboard": 1 |
| }, |
| "format_version": "3.0", |
| "sam2_model": "sam2.1_hiera_b+", |
| "embedding_dim": 256, |
| "fps": 30, |
| "cameras": ["side"], |
| "robot": "UR5e", |
| "gripper": "Robotiq 2F-85" |
| } |
| ``` |
|
|
| ## Test Data Available |
|
|
| Two episodes are fully labeled and validated — you can use them to test the loader: |
|
|
| ### Episode 1: `session_0408_162129/episode_00` |
| |
| | Stat | Value | |
| |------|-------| |
| | Total frames in episode | 604 | |
| | Labeled frames | **346** (range 0–351, 6 gaps) | |
| | Product components | 15 (cpu_fan, cpu_bracket, cpu, graphic_card, motherboard, connector_1..4, ram_1..2, ram_clip_1..4) | |
| | Physical constraints (edges) | 14 | |
| | Robot visibility | Visible in 216 / 346 frames | |
| | Goal component | `cpu_fan` | |
|
|
| ### Episode 2: `session_0410_125013/episode_00` |
| |
| | Stat | Value | |
| |------|-------| |
| | Total frames in episode | 473 | |
| | Labeled frames | **473** (range 0–472, **0 gaps, 100% coverage**) | |
| | Product components | 12 (cpu_fan, cpu_bracket, cpu, graphic_card, motherboard, connector_1..4, ram_1, ram_clip_1..2) | |
| | Physical constraints (edges) | 11 | |
| | Robot visibility | Visible in 448 / 473 frames | |
| | Goal component | `cpu_fan` | |
|
|
| ### Episode 3: `session_0410_125013/episode_01` |
| |
| Continuation episode — cpu_fan was already removed in Episode 2, so this episode starts with all remaining components visible. |
|
|
| | Stat | Value | |
| |------|-------| |
| | Total frames in episode | 196 | |
| | Labeled frames | **196** (range 0–195, **0 gaps, 100% coverage**) | |
| | Product components | 11 (cpu_bracket, cpu, graphic_card, motherboard, connector_1..4, ram_1, ram_clip_1..2) | |
| | Physical constraints (edges) | 11 | |
| | Goal component | `graphic_card` | |
|
|
| **Note on v3.1 data semantics** (applies to all labeled frames): |
| - **Amodal masks** — stored masks represent the full component shape, unchanged by any occluder. |
| - **Full occluder subtraction for depth and embeddings** — at save time, for each target product `cid`: |
|
|
| ``` |
| occluders(cid) = robot_mask ∪ goal_mask if cid != goal_component |
| = robot_mask if cid == goal_component |
| ``` |
|
|
| `compute_depth_info` and embedding re-extraction use `mask & ~occluders`, so `centroid`, `point_cloud`, and 256-D features sample only non-occluded pixels. Rationale: robot holds/moves the goal on top of other components, so both occlude them jointly. The goal itself is only occluded by the robot (it cannot occlude itself). |
| - **Two-flag semantic encoding**: `(visibility, depth_valid)` = `(1,1)` visible & measurable; `(1,0)` fully occluded (robot and/or goal) — component still in scene, unmeasurable this frame; `(0,0)` not in scene (hidden under a blocker, or removed — disambiguated via incoming edge `is_locked`). |
| - **Sparse labeling defensive invariant**: at load time, if `visibility=1` but the component has no entry in that frame's embedding file (user skipped labeling this frame), the PyG loader treats it as effectively `visibility=0`. Distinguishes "unlabeled frame" from "full occlusion" (which has explicit zero embedding stored). |
| - **SAM2 tracking on copy** — goal component and robot get SAM2 bbox+mask-input+centroid prediction across frames; other components simple-copy. |
|
|
| ### Download and Test (3 steps) |
|
|
| **Step 1: Download just one episode (lightweight)** |
|
|
| ```bash |
| pip install huggingface_hub |
| ``` |
|
|
| ```python |
| from huggingface_hub import snapshot_download |
| |
| local_dir = snapshot_download( |
| repo_id="ChangChrisLiu/GNN_Disassembly_WorldModel", |
| repo_type="dataset", |
| allow_patterns=[ |
| "session_0408_162129/episode_00/metadata.json", |
| "session_0408_162129/episode_00/robot_states.npy", |
| "session_0408_162129/episode_00/robot_actions.npy", |
| "session_0408_162129/episode_00/side/rgb/frame_000042.png", |
| "session_0408_162129/episode_00/side/depth/frame_000042.npy", |
| "session_0408_162129/episode_00/annotations/*", |
| ], |
| ) |
| print("Downloaded to:", local_dir) |
| ``` |
|
|
| **Step 2: Save the loader code** (copy the self-contained `gnn_disassembly_loader.py` block below into a file) |
|
|
| **Step 3: Run this test script** — it loads frame 42, prints the full graph anatomy, and verifies everything: |
|
|
| ```python |
| from pathlib import Path |
| from gnn_disassembly_loader import ( |
| load_pyg_frame_products_only, |
| load_pyg_frame_with_robot, |
| list_labeled_frames, |
| load_frame_data, |
| ) |
| |
| # After snapshot_download above: |
| episode = Path(local_dir) / "session_0408_162129" / "episode_00" |
| |
| # 1. Enumerate labeled frames |
| frames = list_labeled_frames(episode) |
| assert len(frames) == 346, f"Expected 346 labeled frames, got {len(frames)}" |
| print(f"✓ Labeled frames: {len(frames)} (range {frames[0]}..{frames[-1]})") |
| |
| # 2. Load frame 42 — products only |
| data1 = load_pyg_frame_products_only(episode, frame_idx=42) |
| assert data1.num_nodes == 15, f"Expected 15 products, got {data1.num_nodes}" |
| assert data1.edge_index.shape[1] == 15 * 14 # fully connected |
| assert data1.edge_attr.shape == (210, 3) # 3D edge features |
| print(f"✓ Products-only: {data1}") |
| |
| # 3. Load frame 42 — with robot agent |
| data2 = load_pyg_frame_with_robot(episode, frame_idx=42) |
| assert data2.num_nodes == 16, f"Expected 15 products + 1 robot = 16, got {data2.num_nodes}" |
| assert data2.edge_index.shape[1] == 16 * 15 |
| assert hasattr(data2, "robot_point_cloud") |
| print(f"✓ With robot: {data2}") |
| print(f" Robot point cloud: {tuple(data2.robot_point_cloud.shape)}") |
| print(f" Robot mask: {tuple(data2.robot_mask.shape)}") |
| |
| # 4. Verify robot edges are all [0, 0, 0] |
| robot_idx = data2.num_nodes - 1 |
| robot_edges = (data2.edge_index[0] == robot_idx) | (data2.edge_index[1] == robot_idx) |
| assert (data2.edge_attr[robot_edges] == 0).all() |
| print(f"✓ Robot edges: {robot_edges.sum().item()} — all [0,0,0]") |
| |
| # 5. Verify edge feature semantics |
| has_c = (data1.edge_attr[:, 0] == 1).sum().item() |
| locked = ((data1.edge_attr[:, 0] == 1) & (data1.edge_attr[:, 1] == 1)).sum().item() |
| src_blocks = ((data1.edge_attr[:, 0] == 1) & (data1.edge_attr[:, 2] == 1)).sum().item() |
| assert has_c == 28 # 14 constraints × 2 directions |
| assert locked == 28 # all locked at frame 42 |
| assert src_blocks == 14 # half the constraint edges have src as blocker |
| print(f"✓ Edge features: {has_c} constraint edges, {locked} locked, {src_blocks} forward-direction") |
| |
| # 6. Verify fully-connected + symmetric structure |
| from collections import Counter |
| pairs = Counter() |
| for i in range(data1.edge_index.shape[1]): |
| src = data1.edge_index[0, i].item() |
| dst = data1.edge_index[1, i].item() |
| pairs[frozenset([src, dst])] += 1 |
| # Every unordered pair should appear exactly twice: (i, j) AND (j, i) |
| assert all(count == 2 for count in pairs.values()) |
| print(f"✓ Structurally symmetric: every pair has both directions") |
| |
| # 7. Raw data access |
| fd = load_frame_data(episode, frame_idx=42) |
| print(f"✓ Raw data: {len(fd.masks)} product masks, robot {'visible' if fd.robot else 'hidden'}") |
| |
| print("\nAll tests passed! The dataset is ready for training.") |
| ``` |
|
|
| Expected output: |
| ``` |
| ✓ Labeled frames: 346 (range 0..351) |
| ✓ Products-only: Data(x=[15, 269], edge_index=[2, 210], edge_attr=[210, 3], y=[1], num_nodes=15) |
| ✓ With robot: Data(x=[16, 269], edge_index=[2, 240], edge_attr=[240, 3], y=[1], num_nodes=16, robot_point_cloud=[5729, 3], robot_pixel_coords=[5729, 2], robot_mask=[720, 1280]) |
| Robot point cloud: (5729, 3) |
| Robot mask: (720, 1280) |
| ✓ Robot edges: 30 — all [0,0,0] |
| ✓ Edge features: 28 constraint edges, 28 locked, 14 forward-direction |
| ✓ Structurally symmetric: every pair has both directions |
| ✓ Raw data: 13 product masks, robot visible |
| |
| All tests passed! The dataset is ready for training. |
| ``` |
|
|
| ## Graph Structure — What You Get Per Frame |
|
|
| Every labeled frame is converted to **one PyTorch Geometric `Data` object**. Here's exactly what it contains: |
|
|
| ### Node Features (269D per node) |
|
|
| ``` |
| ┌───────────────────────────────────────────────────────────────────────┐ |
| │ x[i] = 269D feature vector for node i │ |
| ├───────────────────────────────────────────────────────────────────────┤ |
| │ [0 : 256] SAM2 embedding (256D) │ |
| │ Masked average pool over SAM2 encoder's vision_features. │ |
| │ Captures visual appearance of the component. │ |
| ├───────────────────────────────────────────────────────────────────────┤ |
| │ [256 : 259] 3D position (3D) │ |
| │ Centroid in camera frame, meters. Mean of the valid │ |
| │ depth-backprojected points within the mask. │ |
| │ Zero vector if no valid depth (check depth_valid flag). │ |
| ├───────────────────────────────────────────────────────────────────────┤ |
| │ [259 : 268] Type one-hot (9D) │ |
| │ Index order: cpu_fan, cpu_bracket, cpu, ram_clip, ram, │ |
| │ connector, graphic_card, motherboard, robot. │ |
| │ Multiple instances (e.g. ram_1, ram_2) share the same │ |
| │ one-hot — distinguished by their SAM2 embedding + 3D pos.│ |
| ├───────────────────────────────────────────────────────────────────────┤ |
| │ [268] Visibility (1D) │ |
| │ Binary flag — 1 if visible this frame, 0 if hidden. │ |
| │ Delta-encoded through frame_states in side_graph.json. │ |
| └───────────────────────────────────────────────────────────────────────┘ |
| ``` |
|
|
| ### Graph Topology — Fully Connected, Structurally Symmetric |
|
|
| For N nodes, the PyG graph has: |
| - `edge_index` shape: **(2, N × (N − 1))** |
| - Every ordered pair `(i, j)` with `i ≠ j` has an edge |
| - Both `(i, j)` AND `(j, i)` exist — the graph is **not structurally directed** |
| - Self-loops are excluded |
|
|
| **Why fully connected?** Sparse constraint graphs (just physical prerequisites) would prevent distant nodes from exchanging information through GNN message passing. Making it fully connected ensures every node pair communicates in one layer. |
|
|
| ### Edge Features (3D per edge) |
|
|
| ``` |
| ┌─────────────────┬──────────┬────────────────┬─────────────────────────┐ |
| │ has_constraint │ is_locked│ src_blocks_dst │ Meaning │ |
| ├─────────────────┼──────────┼────────────────┼─────────────────────────┤ |
| │ 0 │ 0 │ 0 │ No physical constraint │ |
| │ │ │ │ (message passing only) │ |
| ├─────────────────┼──────────┼────────────────┼─────────────────────────┤ |
| │ 1 │ 1 │ 1 │ Physical constraint │ |
| │ │ │ │ LOCKED, src is blocker │ |
| ├─────────────────┼──────────┼────────────────┼─────────────────────────┤ |
| │ 1 │ 1 │ 0 │ Physical constraint │ |
| │ │ │ │ LOCKED, src is blocked │ |
| │ │ │ │ (reverse direction) │ |
| ├─────────────────┼──────────┼────────────────┼─────────────────────────┤ |
| │ 1 │ 0 │ 1 │ Physical constraint │ |
| │ │ │ │ RELEASED (unlocked) │ |
| │ │ │ │ src is the blocker │ |
| ├─────────────────┼──────────┼────────────────┼─────────────────────────┤ |
| │ 1 │ 0 │ 0 │ Physical constraint │ |
| │ │ │ │ RELEASED, src is blocked│ |
| └─────────────────┴──────────┴────────────────┴─────────────────────────┘ |
| ``` |
|
|
| **Direction is a feature, not structure.** |
| - `has_constraint` and `is_locked` describe the PAIR — they're the same for both `(i,j)` and `(j,i)`. |
| - `src_blocks_dst` is asymmetric: it flips depending on which direction the edge goes. |
|
|
| **Example:** `cpu_fan` blocks `cpu_bracket` (fan covers bracket). At frame 0 (locked): |
|
|
| ``` |
| edge (cpu_fan, cpu_bracket) → [1, 1, 1] cpu_fan is the blocker |
| edge (cpu_bracket, cpu_fan) → [1, 1, 0] cpu_bracket is the blocked |
| ``` |
|
|
| At frame 152 after the user removes the fan (unlocked): |
|
|
| ``` |
| edge (cpu_fan, cpu_bracket) → [1, 0, 1] |
| edge (cpu_bracket, cpu_fan) → [1, 0, 0] |
| ``` |
|
|
| ### Robot Agent Node (Optional) |
|
|
| When loaded with `load_pyg_frame_with_robot()`, the robot is appended as the **last node** (index `N_products`). All edges involving the robot have features `[0, 0, 0]` — the robot has no physical constraints, it's a context-providing agent node. |
|
|
| The raw robot data (point cloud, pixel coords, full mask) is attached as extra tensors on the `Data` object for optional PointNet-style encoding. |
|
|
| ### Matching a Frame to Its RGB Image |
|
|
| Frame indices in the loader directly map to image files: |
|
|
| ```python |
| frame_idx = 42 |
| rgb_path = episode / "side" / "rgb" / f"frame_{frame_idx:06d}.png" |
| depth_path = episode / "side" / "depth" / f"frame_{frame_idx:06d}.npy" |
| ``` |
|
|
| Example — load PyG frame + matching image + depth: |
|
|
| ```python |
| from pathlib import Path |
| import numpy as np |
| from PIL import Image |
| from gnn_disassembly_loader import load_pyg_frame_with_robot |
| |
| episode = Path("episode_00") |
| frame_idx = 42 |
| |
| # PyG graph for this frame |
| data = load_pyg_frame_with_robot(episode, frame_idx) |
| |
| # Matching RGB image (1280x720 PNG) |
| rgb = np.array(Image.open(episode / "side" / "rgb" / f"frame_{frame_idx:06d}.png")) |
| print("RGB shape:", rgb.shape) # (720, 1280, 3) |
| |
| # Matching depth (1280x720 uint16 mm) |
| depth = np.load(episode / "side" / "depth" / f"frame_{frame_idx:06d}.npy") |
| print("Depth shape:", depth.shape, depth.dtype) # (720, 1280) uint16 |
| |
| # Robot mask is in the PyG data if robot is visible |
| if hasattr(data, "robot_mask"): |
| robot_mask = data.robot_mask.numpy() # (720, 1280) uint8 |
| print("Robot mask area:", robot_mask.sum(), "pixels") |
| ``` |
|
|
| ## Loading the Data — PyTorch Geometric |
|
|
| This section contains **self-contained** code you can copy-paste directly. No need to clone any repo. |
|
|
| ### Prerequisites |
|
|
| ```bash |
| pip install torch numpy torch_geometric pillow |
| ``` |
|
|
| ### Self-contained PyG loader |
|
|
| Copy this into a file called `gnn_disassembly_loader.py`: |
|
|
| ```python |
| """Self-contained PyG loader for the GNN Disassembly dataset. |
| |
| Two loader variants: |
| - load_pyg_frame_products_only(ep, frame) → constraint graph only, no robot |
| - load_pyg_frame_with_robot(ep, frame) → constraint graph + robot agent node |
| |
| Both return torch_geometric.data.Data with: |
| x (N, 268) node features |
| edge_index (2, N*(N-1)) fully connected directed message-passing edges |
| edge_attr (N*(N-1), 3) [has_constraint, is_locked, src_blocks_dst] |
| num_nodes N |
| |
| Notes on the edge feature design: |
| - The graph is FULLY CONNECTED and structurally symmetric. |
| Both (i, j) and (j, i) exist in edge_index for every node pair i != j. |
| - Direction is NOT encoded in the graph structure. It is encoded as |
| a feature: `src_blocks_dst`. |
| - `has_constraint` and `is_locked` are symmetric per pair (same value |
| for both (i, j) and (j, i)). |
| - `src_blocks_dst` is asymmetric: it is 1 if the edge's src node |
| physically blocks its dst node, 0 otherwise. |
| """ |
| |
| import json |
| from dataclasses import dataclass |
| from pathlib import Path |
| from typing import Dict, List, Optional, Tuple |
| |
| import numpy as np |
| import torch |
| from torch_geometric.data import Data |
| |
| |
| # ───────────────────────────────────────────────────────────────────────────── |
| # Helpers |
| # ───────────────────────────────────────────────────────────────────────────── |
| |
| def list_labeled_frames(episode_dir: Path) -> List[int]: |
| """Return sorted list of frame indices that have saved annotations.""" |
| mask_dir = episode_dir / "annotations" / "side_masks" |
| if not mask_dir.exists(): |
| return [] |
| frames = [] |
| for p in mask_dir.glob("frame_*.npz"): |
| try: |
| frames.append(int(p.stem.split("_")[1])) |
| except (ValueError, IndexError): |
| continue |
| return sorted(frames) |
| |
| |
| def resolve_frame_state(graph_json: dict, frame_idx: int) -> Tuple[Dict[str, bool], Dict[str, bool]]: |
| """Resolve delta-encoded constraints + visibility at a frame. |
| |
| Walks frame_states from frame 0 to frame_idx, accumulating deltas. |
| Returns (constraints_dict, visibility_dict). |
| """ |
| constraints: Dict[str, bool] = {} |
| visibility: Dict[str, bool] = {} |
| # Defaults: every component visible, every edge locked |
| for c in graph_json["components"]: |
| visibility[c["id"]] = True |
| for e in graph_json["edges"]: |
| constraints[f"{e['src']}->{e['dst']}"] = True |
| # Walk deltas up to frame_idx |
| fs_dict = graph_json.get("frame_states", {}) |
| for f in sorted([int(k) for k in fs_dict]): |
| if f > frame_idx: |
| break |
| fs = fs_dict[str(f)] |
| for k, v in fs.get("constraints", {}).items(): |
| constraints[k] = v |
| for k, v in fs.get("visibility", {}).items(): |
| visibility[k] = v |
| return constraints, visibility |
| |
| |
| def type_one_hot(comp_type: str, type_vocab: List[str]) -> List[float]: |
| """9-dim one-hot encoding of component type based on type_vocab.""" |
| return [1.0 if t == comp_type else 0.0 for t in type_vocab] |
| |
| |
| # ───────────────────────────────────────────────────────────────────────────── |
| # Raw data loader (NumPy only, no torch) |
| # ───────────────────────────────────────────────────────────────────────────── |
| |
| @dataclass |
| class FrameData: |
| graph: dict |
| masks: Dict[str, np.ndarray] |
| embeddings: Dict[str, np.ndarray] |
| depth_info: dict |
| robot: Optional[dict] |
| constraints: Dict[str, bool] |
| visibility: Dict[str, bool] |
| |
| |
| def load_frame_data(episode_dir: Path, frame_idx: int) -> FrameData: |
| """Load all v3 annotation files for one frame.""" |
| anno = episode_dir / "annotations" |
| |
| with open(anno / "side_graph.json") as f: |
| graph = json.load(f) |
| |
| def _load_npz_dict(path: Path) -> Dict[str, np.ndarray]: |
| if not path.exists(): |
| return {} |
| d = np.load(path) |
| return {k: d[k] for k in d.files} |
| |
| masks = _load_npz_dict(anno / "side_masks" / f"frame_{frame_idx:06d}.npz") |
| embeddings = _load_npz_dict(anno / "side_embeddings" / f"frame_{frame_idx:06d}.npz") |
| depth_info = _load_npz_dict(anno / "side_depth_info" / f"frame_{frame_idx:06d}.npz") |
| |
| robot: Optional[dict] = None |
| robot_path = anno / "side_robot" / f"frame_{frame_idx:06d}.npz" |
| if robot_path.exists(): |
| r = np.load(robot_path) |
| if r["visible"][0] == 1: |
| robot = {k: r[k] for k in r.files} |
| |
| constraints, visibility = resolve_frame_state(graph, frame_idx) |
| return FrameData(graph, masks, embeddings, depth_info, robot, constraints, visibility) |
| |
| |
| # ───────────────────────────────────────────────────────────────────────────── |
| # PyG loader — products only |
| # ───────────────────────────────────────────────────────────────────────────── |
| |
| def load_pyg_frame_products_only(episode_dir: Path, frame_idx: int) -> Data: |
| """Fully connected PyG graph WITHOUT robot. |
| |
| Returns Data( |
| x=[N, 268], |
| edge_index=[2, N*(N-1)], |
| edge_attr=[N*(N-1), 3], # [has_constraint, is_locked, src_blocks_dst] |
| num_nodes=N, |
| ) |
| where N = number of product components (robot excluded). |
| """ |
| fd = load_frame_data(episode_dir, frame_idx) |
| graph = fd.graph |
| type_vocab = graph["type_vocab"] # 9 entries incl. robot |
| nodes = graph["components"] # robot already excluded per spec |
| N = len(nodes) |
| |
| # ── Node features ── |
| # [256D SAM2 embedding, 3D position, 9D type one-hot, 1D visibility] = 269 |
| # NOTE: 256 + 3 + 9 + 1 = 269 (not 268). Adjust if you need a different layout. |
| x_list = [] |
| for node in nodes: |
| cid = node["id"] |
| emb = fd.embeddings.get(cid, np.zeros(256, dtype=np.float32)) |
| |
| depth_valid_key = f"{cid}_depth_valid" |
| centroid_key = f"{cid}_centroid" |
| if (depth_valid_key in fd.depth_info |
| and int(fd.depth_info[depth_valid_key][0]) == 1): |
| pos = fd.depth_info[centroid_key].astype(np.float32) |
| else: |
| pos = np.zeros(3, dtype=np.float32) |
| |
| type_oh = type_one_hot(node["type"], type_vocab) # 9D |
| vis = 1.0 if fd.visibility.get(cid, True) else 0.0 |
| |
| feat = np.concatenate([ |
| emb.astype(np.float32), |
| pos, |
| np.array(type_oh, dtype=np.float32), |
| np.array([vis], dtype=np.float32), |
| ]) |
| x_list.append(feat) |
| x = torch.tensor(np.stack(x_list), dtype=torch.float32) if x_list else torch.empty((0, 269)) |
| |
| # ── Fully connected edges with 3D features ── |
| # Edge feature: [has_constraint, is_locked, src_blocks_dst] |
| # - has_constraint & is_locked are SYMMETRIC for the pair (A, B) |
| # - src_blocks_dst is ASYMMETRIC: 1 if edge's src physically blocks dst |
| constraint_set = {(e["src"], e["dst"]) for e in graph["edges"]} |
| pair_forward = {} # frozenset({a, b}) -> (blocker, blocked) |
| for (s, d) in constraint_set: |
| pair_forward[frozenset([s, d])] = (s, d) |
| |
| src_idx, dst_idx, edge_attr = [], [], [] |
| for i in range(N): |
| for j in range(N): |
| if i == j: |
| continue |
| src_id = nodes[i]["id"] |
| dst_id = nodes[j]["id"] |
| src_idx.append(i) |
| dst_idx.append(j) |
| |
| pair_key = frozenset([src_id, dst_id]) |
| if pair_key in pair_forward: |
| forward = pair_forward[pair_key] |
| constraint_key = f"{forward[0]}->{forward[1]}" |
| is_locked = fd.constraints.get(constraint_key, True) |
| src_blocks_dst = 1.0 if src_id == forward[0] else 0.0 |
| edge_attr.append([ |
| 1.0, |
| 1.0 if is_locked else 0.0, |
| src_blocks_dst, |
| ]) |
| else: |
| edge_attr.append([0.0, 0.0, 0.0]) # message passing only |
| |
| return Data( |
| x=x, |
| edge_index=torch.tensor([src_idx, dst_idx], dtype=torch.long), |
| edge_attr=torch.tensor(edge_attr, dtype=torch.float32), |
| y=torch.tensor([frame_idx], dtype=torch.long), |
| num_nodes=N, |
| ) |
| |
| |
| # ───────────────────────────────────────────────────────────────────────────── |
| # PyG loader — with robot agent node |
| # ───────────────────────────────────────────────────────────────────────────── |
| |
| def load_pyg_frame_with_robot(episode_dir: Path, frame_idx: int) -> Data: |
| """Fully connected PyG graph WITH robot appended as agent node. |
| |
| Robot is node N (the last node). All edges involving the robot have |
| features [0, 0, 0] because the robot has no physical constraints. |
| |
| If the robot is not visible at this frame, returns the products-only graph. |
| Additional attached tensors when robot is visible: |
| data.robot_point_cloud (M, 3) float32 |
| data.robot_pixel_coords (M, 2) int32 |
| data.robot_mask (H, W) uint8 |
| """ |
| data = load_pyg_frame_products_only(episode_dir, frame_idx) |
| fd = load_frame_data(episode_dir, frame_idx) |
| if fd.robot is None: |
| return data |
| |
| graph = fd.graph |
| type_vocab = graph["type_vocab"] |
| products = graph["components"] |
| N_prod = len(products) |
| N = N_prod + 1 |
| |
| # ── Build robot node features ── |
| robot_emb = fd.robot["embedding"].astype(np.float32) |
| robot_pos = (fd.robot["centroid"].astype(np.float32) |
| if int(fd.robot["depth_valid"][0]) == 1 |
| else np.zeros(3, dtype=np.float32)) |
| robot_type_oh = type_one_hot("robot", type_vocab) |
| robot_feat = np.concatenate([ |
| robot_emb, robot_pos, |
| np.array(robot_type_oh, dtype=np.float32), |
| np.array([1.0], dtype=np.float32), |
| ]) |
| x = torch.cat([data.x, torch.tensor(robot_feat, dtype=torch.float32).unsqueeze(0)], dim=0) |
| |
| # ── Rebuild edges with 3D features ── |
| constraint_set = {(e["src"], e["dst"]) for e in graph["edges"]} |
| pair_forward = {} |
| for (s, d) in constraint_set: |
| pair_forward[frozenset([s, d])] = (s, d) |
| |
| src_idx, dst_idx, edge_attr = [], [], [] |
| |
| # Products × Products |
| for i in range(N_prod): |
| for j in range(N_prod): |
| if i == j: |
| continue |
| src_id = products[i]["id"] |
| dst_id = products[j]["id"] |
| src_idx.append(i) |
| dst_idx.append(j) |
| pair_key = frozenset([src_id, dst_id]) |
| if pair_key in pair_forward: |
| forward = pair_forward[pair_key] |
| is_locked = fd.constraints.get(f"{forward[0]}->{forward[1]}", True) |
| src_blocks_dst = 1.0 if src_id == forward[0] else 0.0 |
| edge_attr.append([1.0, 1.0 if is_locked else 0.0, src_blocks_dst]) |
| else: |
| edge_attr.append([0.0, 0.0, 0.0]) |
| |
| # Robot ↔ Products (both directions, message-passing only) |
| robot_idx = N_prod |
| for i in range(N_prod): |
| src_idx.append(robot_idx); dst_idx.append(i); edge_attr.append([0.0, 0.0, 0.0]) |
| src_idx.append(i); dst_idx.append(robot_idx); edge_attr.append([0.0, 0.0, 0.0]) |
| |
| data = Data( |
| x=x, |
| edge_index=torch.tensor([src_idx, dst_idx], dtype=torch.long), |
| edge_attr=torch.tensor(edge_attr, dtype=torch.float32), |
| y=torch.tensor([frame_idx], dtype=torch.long), |
| num_nodes=N, |
| ) |
| data.robot_point_cloud = torch.tensor(fd.robot["point_cloud"], dtype=torch.float32) |
| data.robot_pixel_coords = torch.tensor(fd.robot["pixel_coords"], dtype=torch.int32) |
| data.robot_mask = torch.tensor(fd.robot["mask"], dtype=torch.uint8) |
| return data |
| |
| |
| # ───────────────────────────────────────────────────────────────────────────── |
| # Episode iterator |
| # ───────────────────────────────────────────────────────────────────────────── |
| |
| def iterate_episode(episode_dir: Path, with_robot: bool = True): |
| """Yield (frame_idx, Data) pairs for all labeled frames in an episode.""" |
| loader = load_pyg_frame_with_robot if with_robot else load_pyg_frame_products_only |
| for frame_idx in list_labeled_frames(episode_dir): |
| yield frame_idx, loader(episode_dir, frame_idx) |
| ``` |
|
|
| ### Usage Examples |
|
|
| #### Variant 1: Constraint Graph Only (No Robot) |
|
|
| ```python |
| from pathlib import Path |
| from gnn_disassembly_loader import load_pyg_frame_products_only, list_labeled_frames |
| |
| episode = Path("episode_00") # downloaded from HF |
| |
| # Enumerate labeled frames |
| frames = list_labeled_frames(episode) |
| print(f"Episode has {len(frames)} labeled frames") |
| # → Episode has 346 labeled frames (for session_0408_162129/episode_00) |
| # → Episode has 473 labeled frames (for session_0410_125013/episode_00) |
| |
| # Load one frame as a fully connected PyG graph (products only) |
| data = load_pyg_frame_products_only(episode, frame_idx=42) |
| print(data) |
| # → Data(x=[15, 269], edge_index=[2, 210], edge_attr=[210, 3], y=[1], num_nodes=15) |
| |
| # For N=15 products: edges = 15 * 14 = 210 (fully connected) |
| print("Node features:", data.x.shape) # (15, 269) |
| print("Edges:", data.edge_index.shape) # (2, 210) |
| print("Edge attrs:", data.edge_attr.shape) # (210, 3) = [has_constraint, is_locked, src_blocks_dst] |
| |
| # Count edge feature breakdown |
| has_c = (data.edge_attr[:, 0] == 1).sum().item() |
| locked = ((data.edge_attr[:, 0] == 1) & (data.edge_attr[:, 1] == 1)).sum().item() |
| src_blocks = ((data.edge_attr[:, 0] == 1) & (data.edge_attr[:, 2] == 1)).sum().item() |
| print(f"Edges with physical constraint: {has_c}") |
| print(f" currently locked: {locked}") |
| print(f" where src is the blocker: {src_blocks}") |
| print(f"Message-passing-only edges: {(data.edge_attr[:, 0] == 0).sum().item()}") |
| ``` |
|
|
| #### Variant 2: Constraint Graph + Robot Agent Node |
|
|
| ```python |
| from gnn_disassembly_loader import load_pyg_frame_with_robot |
| |
| data = load_pyg_frame_with_robot(episode, frame_idx=42) |
| print(data) |
| # → Data(x=[16, 269], edge_index=[2, 240], edge_attr=[240, 3], y=[1], num_nodes=16) |
| # Robot is the last node (index 15 for a 15-product graph). |
| # Robot edges: 15 products * 2 directions = 30 extra edges → 210 + 30 = 240 |
| |
| # Verify robot edges are all message-passing (no constraint) |
| robot_idx = data.num_nodes - 1 |
| robot_edges = (data.edge_index[0] == robot_idx) | (data.edge_index[1] == robot_idx) |
| assert (data.edge_attr[robot_edges] == 0).all(), "Robot edges must be [0, 0, 0]" |
| print(f"Robot edges: {robot_edges.sum().item()} — all [0, 0, 0]") |
| |
| # Raw robot data (optional, for PointNet-style encoding) |
| print("Robot point cloud:", data.robot_point_cloud.shape) # (M, 3) — M varies per frame |
| print("Robot mask:", data.robot_mask.shape) # (720, 1280) |
| ``` |
|
|
| #### Edge Feature Semantics |
|
|
| Each row of `data.edge_attr` is 3-dimensional: `[has_constraint, is_locked, src_blocks_dst]`. |
|
|
| ``` |
| ┌──────────────────┬────────────┬────────────────┬─────────────────────────────────┐ |
| │ has_constraint │ is_locked │ src_blocks_dst │ Meaning │ |
| ├──────────────────┼────────────┼────────────────┼─────────────────────────────────┤ |
| │ 0 │ 0 │ 0 │ No physical constraint │ |
| │ │ │ │ Message passing only │ |
| ├──────────────────┼────────────┼────────────────┼─────────────────────────────────┤ |
| │ 1 │ 1 │ 1 │ Edge src physically blocks dst │ |
| │ │ │ │ Constraint currently LOCKED │ |
| ├──────────────────┼────────────┼────────────────┼─────────────────────────────────┤ |
| │ 1 │ 1 │ 0 │ Edge dst physically blocks src │ |
| │ │ │ │ (the reverse direction of the │ |
| │ │ │ │ physical constraint) │ |
| │ │ │ │ Constraint currently LOCKED │ |
| ├──────────────────┼────────────┼────────────────┼─────────────────────────────────┤ |
| │ 1 │ 0 │ 1 │ Edge src physically blocks dst │ |
| │ │ │ │ Constraint RELEASED │ |
| ├──────────────────┼────────────┼────────────────┼─────────────────────────────────┤ |
| │ 1 │ 0 │ 0 │ Edge dst physically blocks src │ |
| │ │ │ │ Constraint RELEASED │ |
| └──────────────────┴────────────┴────────────────┴─────────────────────────────────┘ |
| ``` |
|
|
| **Important:** The graph is **fully connected and structurally symmetric** — both `(A, B)` and `(B, A)` edges exist for every pair. `has_constraint` and `is_locked` are the same for both directions (they describe the unordered pair). `src_blocks_dst` flips between the two directions — it tells you whether the edge's source is the one doing the blocking. |
|
|
| **Example: CPU bracket blocks CPU removal** |
|
|
| If `cpu_bracket → cpu` is an active constraint, the loader produces: |
|
|
| ``` |
| Edge (cpu_bracket, cpu): [1, 1, 1] # cpu_bracket blocks cpu, locked, src=blocker |
| Edge (cpu, cpu_bracket): [1, 1, 0] # same physical pair, src=blocked |
| ``` |
|
|
| When the user unlocks the constraint (e.g., after releasing the bracket): |
| ``` |
| Edge (cpu_bracket, cpu): [1, 0, 1] # constraint released, but bracket still named as blocker |
| Edge (cpu, cpu_bracket): [1, 0, 0] |
| ``` |
|
|
| ### Iterating the Full Episode |
|
|
| ```python |
| from torch_geometric.loader import DataLoader |
| from gnn_disassembly_loader import iterate_episode |
| |
| # Build a dataset list |
| data_list = [data for _, data in iterate_episode(episode, with_robot=True)] |
| print(f"Loaded {len(data_list)} frames") |
| |
| # Batch them for training |
| loader = DataLoader(data_list, batch_size=8, shuffle=True) |
| for batch in loader: |
| print(batch.x.shape, batch.edge_index.shape, batch.edge_attr.shape) |
| break |
| ``` |
|
|
| ### Adding Robot State as Node Features (Graph B) |
|
|
| For the perception + robot state variant, concatenate the 13D robot state to every node: |
|
|
| ```python |
| import numpy as np |
| import torch |
| |
| robot_states = np.load(episode / "robot_states.npy") # (T, 13) |
| |
| def add_robot_state_to_graph(data, frame_idx, robot_states): |
| robot_state_t = torch.tensor(robot_states[frame_idx], dtype=torch.float32) # (13,) |
| broadcast = robot_state_t.unsqueeze(0).expand(data.num_nodes, -1) # (N, 13) |
| data.x = torch.cat([data.x, broadcast], dim=1) # (N, 282) |
| return data |
| |
| data_b = add_robot_state_to_graph(data, frame_idx=42, robot_states=robot_states) |
| print("Graph B node features:", data_b.x.shape) # (16, 282) for with_robot variant |
| ``` |
|
|
| ## Node Feature Layout (269D) |
|
|
| ``` |
| [0 : 256] SAM2 embedding (256D) — masked avg pool over vision_features |
| [256 : 259] 3D position (3D) — centroid in camera frame (meters) |
| [259 : 268] type one-hot (9D) — index by type_vocab (incl. "robot") |
| [268] visibility (1D) — binary flag |
| ``` |
|
|
| Total: **269D per node**. |
|
|
| For Graph B (with robot state broadcast): |
| ``` |
| [0 : 269] Graph A features (269D) |
| [269 : 275] joint positions (6D) — UR5e joint angles (radians) |
| [275 : 281] TCP pose (6D) — [x, y, z, rx, ry, rz] |
| [281] gripper position (1D) — Robotiq 2F-85 (0-255) |
| ``` |
|
|
| Total: **282D per node**. |
|
|
| ## Raw Data Access (No PyG) |
|
|
| If you prefer raw NumPy without PyTorch Geometric: |
|
|
| ```python |
| from scripts.pyg_loader import load_frame_data |
| |
| fd = load_frame_data(episode, frame_idx=42) |
| |
| print("Graph:", fd.graph["components"]) |
| print("Masks:", list(fd.masks.keys())) |
| print("Resolved visibility:", fd.visibility) |
| print("Robot present:", fd.robot is not None) |
| |
| if fd.robot is not None: |
| print("Robot mask shape:", fd.robot["mask"].shape) |
| print("Robot point cloud:", fd.robot["point_cloud"].shape) |
| print("Robot centroid (m):", fd.robot["centroid"]) |
| |
| # Access a specific component's depth info |
| for key in ["point_cloud", "pixel_coords", "centroid", "area", "depth_valid"]: |
| full_key = f"cpu_fan_{key}" |
| if full_key in fd.depth_info: |
| print(f"cpu_fan {key}: {fd.depth_info[full_key]}") |
| ``` |
|
|
| ## Recording Hardware |
|
|
| - **Robot:** UR5e + Robotiq 2F-85 gripper |
| - **Side camera:** Luxonis OAK-D Pro (static viewpoint) |
| - Intrinsics: fx=1033.8, fy=1033.7, cx=632.9, cy=359.9 |
| - **Recording rate:** 30 Hz |
| - **Image size:** 1280 × 720 |
| - **Depth format:** uint16, millimeters |
| - **Teleoperation:** Thrustmaster SOL-R2 HOSAS controllers |
|
|
| ## Annotation Tool |
|
|
| Annotations created with a custom SAM2-based labeling tool: |
|
|
| - **Repository:** https://github.com/ChangChrisLiu/gnn-world-model |
| - **Backend:** FastAPI + SAM2 (`sam2.1_hiera_base_plus`) |
| - **Frontend:** Vanilla HTML/JS, side-only interactive view |
| - **Tools:** BBox, Point, Polygon, Brush, Eraser (all mask-editing operations) |
| - **Features:** Dynamic component instances, AGENT badge for robot, scroll-to-zoom, undo/redo, per-frame delta-encoded visibility |
|
|
| ## License |
|
|
| Released under **CC BY 4.0**. Use, share, and adapt freely with attribution. |
|
|
| ## Acknowledgements |
|
|
| Built using: |
| - [Segment Anything Model 2 (SAM2)](https://github.com/facebookresearch/sam2) by Meta AI |
| - [PyTorch Geometric](https://pytorch-geometric.readthedocs.io/) |
| - [Hugging Face Datasets](https://huggingface.co/docs/datasets) |
|
|