Datasets:
Tasks:
Visual Question Answering
Modalities:
Image
Formats:
imagefolder
Size:
1K - 10K
Tags:
VPT
License:
| license: cc-by-4.0 | |
| task_categories: | |
| - visual-question-answering | |
| tags: | |
| - VPT | |
| pretty_name: BlenderGaze | |
| size_categories: | |
| - 1K<n<10K | |
| configs: | |
| - config_name: BlenderGaze-Texture-Cone | |
| data_files: | |
| - split: test | |
| path: BlenderGaze-Texture-Cone/* | |
| features: | |
| - name: image | |
| dtype: image | |
| - name: label | |
| dtype: int64 | |
| - config_name: BlenderGaze-Texture-NoCone | |
| data_files: | |
| - split: test | |
| path: BlenderGaze-Texture-NoCone/* | |
| features: | |
| - name: image | |
| dtype: image | |
| - name: label | |
| dtype: int64 | |
| - config_name: BlenderGaze-Isolated-Cone | |
| data_files: | |
| - split: test | |
| path: BlenderGaze-Isolated-Cone/* | |
| features: | |
| - name: image | |
| dtype: image | |
| - name: label | |
| dtype: int64 | |
| - config_name: BlenderGaze-Isolated-NoCone | |
| data_files: | |
| - split: test | |
| path: BlenderGaze-Isolated-NoCone/* | |
| features: | |
| - name: image | |
| dtype: image | |
| - name: label | |
| dtype: int64 | |
| <p align="center"> | |
| <img src="BlenderGaze.png" width="300"> | |
| </p> | |
| # BlenderGaze | |
| The **BlenderGaze** dataset is specifically created to further investigate Visual Perspective Taking (VPT) in Vision Language Models (VLMs). This dataset extends the previously introduced [Isle dataset](https://huggingface.co/datasets/Gracjan/Isle), primarily expanding in terms of size rather than diversity. | |
| BlenderGaze consists of two complementary subsets: | |
| - **BlenderGaze-Isolated:** Contains isolated scenes featuring a humanoid figure and a red box. These scenes are explicitly designed to isolate and test the basic VPT capabilities of models (see Figure 1). | |
| - **BlenderGaze-Texture:** Includes more complex and realistic scenes **with 10 distinct backgrounds**, each featuring 256 unique scenarios involving interactions between humanoids and robots. This subset aims to explore VPT within *human-robot* interaction scenarios. | |
| In both subsets, objects are randomly placed, and camera viewpoints vary. Visibility is determined by a *120-degree field-of-view* cone originating from the humanoid figure's eyes, representing human peripheral vision. Each scene is provided in two versions: | |
| - **With visual cone:** The visibility cone is explicitly shown, serving as a clear indicator of the humanoid figure's perspective (ground truth). | |
| - **Without visual cone:** Simulates a natural viewing experience without explicit indicators of visual perspective. | |
| Due to the rendering process, some minor errors in ground truth annotations might occur, though such **inaccuracies are expected to be under 10%**. | |
| ## Scene and Background Textures | |
| The following environment textures from [Poly Haven](https://polyhaven.com/) were used to generate diverse and realistic backgrounds: | |
| - [Belfast Sunset](https://polyhaven.com/a/belfast_sunset) | |
| - [Christmas Photo Studio 02](https://polyhaven.com/a/christmas_photo_studio_02) | |
| - [Moonlit Golf](https://polyhaven.com/a/moonlit_golf) | |
| - [Park Parking](https://polyhaven.com/a/park_parking) | |
| - [Paul Lobe Haus](https://polyhaven.com/a/paul_lobe_haus) | |
| - [Simon's Town Road](https://polyhaven.com/a/simons_town_road) | |
| - [Studio Country Hall](https://polyhaven.com/a/studio_country_hall) | |
| - [Floral Tent](https://polyhaven.com/a/floral_tent) | |
| - [Orlando Stadium](https://polyhaven.com/a/orlando_stadium) | |
| - [Courtyard Night](https://polyhaven.com/a/courtyard_night) | |
| ### 3D Models | |
| - **Robot Model:** [Free3D - Basic Robot](https://free3d.com/3d-model/robot_basic-166128.html) | |
| - **Humanoid Model:** [Sketchfab - Rigged T-Pose Human Male](https://sketchfab.com/3d-models/rigged-t-pose-human-male-w-50-face-blendshapes-cc7e4596bcd145208a6992c757854c07) | |
| This dataset supports researchers and psychologists studying visual perspective taking, spatial reasoning, and scene understanding in both humans and VLMs. | |
| For experiments conducted on BlenderGaze, see [blog](https://medium.com/@gracjan.highlander/building-trust-with-invisible-robots-3aecd6180fa5). | |
| <p align="center"> | |
| <img src="example.png" width="900"> | |
| <figcaption align="center"> | |
| <h3> | |
| Figure 1. An example of isolated scenes, featuring a humanoid figure and a red box. Left with visible cone, right without. | |
| </h3> | |
| </figcaption> | |
| </p> | |
| ## Citation | |
| ```bibtex | |
| @dataset{blendergaze2025, | |
| title={BlenderGaze}, | |
| author={Gracjan Góral, Chris Thomas and Paweł Budzianowski}, | |
| year={2025}, | |
| note={Extension of Isle dataset: https://github.com/GracjanGoral/ISLE} | |
| } | |
| ``` |