--- dataset_info: features: - name: task_id dtype: int32 - name: annotation_id dtype: int32 - name: image dtype: image - name: instruction_cn dtype: string - name: instruction_en dtype: string - name: sample_id dtype: string - name: gaze_point list: int32 - name: choices struct: - name: activity dtype: string - name: is_same_window dtype: string - name: place dtype: string - name: platform dtype: string - name: scenario dtype: string - name: task type dtype: string - name: ui_type dtype: string - name: target_bbox struct: - name: height dtype: float32 - name: image_rotation dtype: int32 - name: labels list: string - name: original_height dtype: int32 - name: original_width dtype: int32 - name: rotation dtype: float32 - name: width dtype: float32 - name: x dtype: float32 - name: y dtype: float32 - name: is_ok dtype: bool - name: objects struct: - name: bbox list: list: float32 length: 4 - name: category list: class_label: names: '0': target language: - en - zh license: cc-by-4.0 tags: - computer-vision - visual-grounding - xr - egocentric - gui - apple-vision-pro - instruction-following - multimodal task_ids: - visual-grounding - object-detection pretty_name: EgoXR-GUI - Egocentric XR GUI Grounding Dataset --- # EgoXR-GUI: Benchmarking GUI Grounding in Physical–Digital Extended Reality EgoXR-GUI is the first extended reality (XR) specific GUI grounding benchmark. Unlike traditional desktop or mobile GUI benchmarks, EgoXR-GUI evaluates whether multimodal large language models (MLLMs) can effectively reason about virtual interfaces embedded within hybrid digital–physical environments. ## Overview - **Dataset Size:** 1,070 carefully curated examples. (Originally comprising more internal annotations, the final publicly released benchmark validates exactly 1,070 high-quality target grounding instructions across diverse spatial scenarios.) - **Platform:** Apple Vision Pro and other 3D/XR environments. - **Task Types:** 1. **Direct Grounding:** Simple identification. 2. **Spatial Grounding:** Reasoning about UI elements based on 3D spatial properties. 3. **Semantic Grounding:** Reasoning based on the text or icon semantics of the UI elements. - **Language Supported:** English (`instruction_en`) and Chinese (`instruction_cn`). ## Data Fields Each Example contains the following fields: - `task_id` & `annotation_id`: Unique identifiers for tracking the specific visual task. - `sample_id`: External sample identifier linking back to the origin dataset source. - `image`: The egocentric view captured from the XR headset/environment. - `instruction_en`: The grounding prompt in English. - `instruction_cn`: The grounding prompt in Chinese. - `gaze_point`: The tracked eye gaze coordinate `[x, y]` representing the user's attention. - `choices`: Structured dictionary showing context tags: - `is_same_window` - `ui_type` - `platform` - `scenario` - `place` - `activity` - `task type` - `target_bbox`: The exact geometrical target. Contains `x`, `y`, `width`, `height`, spatial `rotation`, and string `labels`. - `objects`: Hugging Face standardized format representing the bounding box for Data Viewer Visualization. - `is_ok`: Quality control boolean indicator.