---
license: cc-by-4.0
configs:
- config_name: Isle-Brick-V2
data_files:
- split: test
path: Isle-Brick-V2/*
features:
- name: image
dtype: image
- name: Q1
dtype: int64
- name: Q2
dtype: int64
- name: Q3
dtype: string
- name: Q4
sequence: string
- name: Q5
sequence: string
- name: Q6
dtype: string
- name: Q7
sequence: string
- config_name: Isle-Brick-V2-no_object
data_files:
- split: test
path: Isle-Brick-V2-no_object/*
features:
- name: image
dtype: image
- name: Q5
sequence: string
- config_name: Isle-Brick-V2-visual_hint
data_files:
- split: test
path: Isle-Brick-V2-visual_hint/*
features:
- name: image
dtype: image
- name: Q5
sequence: string
- config_name: Isle-Brick-V2-human
data_files:
- split: test
path: Isle-Brick-V2-human/*
features:
- name: image
dtype: image
- name: Q5
sequence: string
- config_name: Isle-Brick-V2-zoom
data_files:
- split: test
path: Isle-Brick-V2-zoom/*
features:
- name: image
dtype: image
- name: Q5
sequence: string
- name: zoom
dtype: string
- config_name: Isle-Brick-V1
data_files:
- split: test
path: Isle-Brick-V1/*
features:
- name: image
dtype: image
- name: prompt
dtype: string
- name: label
dtype: int64
- config_name: Isle-Dots
data_files:
- split: test
path: Isle-Dots/*
features:
- name: image
dtype: image
- name: level
dtype: int64
- name: prompt
dtype: string
- name: label
dtype: int64
task_categories:
- visual-question-answering
tags:
- VPT
pretty_name: Isle
size_categories:
- n<1K
---
## Dataset Details
The **Isle (I spy with my little eye)** dataset helps researchers study *visual perspective taking* (VPT), scene understanding, and spatial reasoning.
Visual perspective taking is the ability to imagine the world from someone else's viewpoint. This skill is important
for everyday tasks like driving safely, coordinating actions with others, or knowing when it's your _turn to speak_.
This dataset includes high-quality images (over 11 Mpix) and consists of three subsets:
- **Isle-Bricks v1**
- **Isle-Bricks v2**
- **Isle-Dots**
The subsets **Isle-Bricks v1** and **Isle-Dots** come from the study *Seeing Through Their Eyes: Evaluating Visual Perspective Taking in Vision Language Models*,
and were created to test Vision Language Models (VLMs).
**Isle-Bricks v2** comes from the study: *Beyond Recognition: Evaluating Visual Perspective Taking in Vision Language Models* and provides additional images of Lego minifigures from two viewpoints *(See Figure 1)*:
- **surface-level** view
- **bird’s eye** view
Figure 1. Example images from the datasets: bottom left, Isle-Brick v1
bottom right, Isle-Dots; top left, Isle-Dots v2 (surface-level);
and top right, Isle-Dots v2 (bird’s-eye view).
The Isle-Bricks v2 subset includes seven questions (Q1–Q7) to test visual perspective taking and related skills:
- **Q1:** _List and count all objects in the image that are not humanoid minifigures._
- **Q2:** _How many humanoid minifigures are in the image?_
- **Q3:** _Are the humanoid minifigure and the object on the same surface?_
- **Q4:** _In which cardinal direction (north, west, east, or south) is the object located relative to the humanoid minifigure?_
- **Q5:** _Which direction (north, west, east, or south) is the humanoid minifigure facing?_
- **Q6:** _Assuming the humanoid minifigure can see and its eyes are open, does it see the object?_
- **Q7:** _From the perspective of the humanoid minifigure, where is the object located relative to it (front, left, right, or back)?_
**Psychologists can also use this dataset to study human visual perception and understanding.**
# Another related dataset is [BlenderGaze](https://huggingface.co/datasets/Gracjan/BlenderGaze), containing over **2,000** images generated using Blender.
## Dataset Sources
- **Repository:** [GitHub](https://github.com/GracjanGoral/ISLE)
- **Papers:** [Isle-Brick-V1, Isle-Dots](https://arxiv.org/abs/2409.12969), [Isle-Brick-V2](https://arxiv.org/abs/2505.03821)
## Annotation Process
- Three annotators labeled images, and final answers were based on the majority vote.
- Annotators agreed on labels over 99% of the time.
## Citation
```bibtex
@misc{góral2025recognitionevaluatingvisualperspective,
title={Beyond Recognition: Evaluating Visual Perspective Taking in Vision Language Models},
author={Gracjan Góral and Alicja Ziarko and Piotr Miłoś and Michał Nauman and Maciej Wołczyk and Michał Kosiński},
year={2025},
eprint={2505.03821},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2505.03821},
}
@misc{góral2024seeingeyesevaluatingvisual,
title={Seeing Through Their Eyes: Evaluating Visual Perspective Taking in Vision Language Models},
author={Gracjan Góral and Alicja Ziarko and Michal Nauman and Maciej Wołczyk},
year={2024},
eprint={2409.12969},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2409.12969},
}
```
```bibtex
Góral, G., Ziarko, A., Miłoś, P., Nauman, M., Wołczyk, M., & Kosiński, M. (2025). Beyond recognition: Evaluating visual perspective taking in vision language models. arXiv. https://arxiv.org/abs/2505.03821
Góral, G., Ziarko, A., Nauman, M., & Wołczyk, M. (2024). Seeing through their eyes: Evaluating visual perspective taking in vision language models. arXiv. https://arxiv.org/abs/2409.12969
```