patch_tasks_vllm / README.md
yurkes's picture
Update README.md
d89f333 verified
---
dataset_info:
features:
- name: image
dtype: image
- name: question
dtype: string
- name: answer
sequence: int64
splits:
- name: valid
num_bytes: 7094843893.125
num_examples: 14631
- name: train
num_bytes: 140854221157.57
num_examples: 289911
download_size: 51389456693
dataset_size: 147949065050.695
configs:
- config_name: default
data_files:
- split: valid
path: data/valid-*
- split: train
path: data/train-*
license: cc-by-4.0
language:
- en
pretty_name: QA Patches Task Dataset
task_categories:
- image-text-to-text
- visual-question-answering
---
# Dataset Card for Patch-Based Visual Question Answering Dataset
## Dataset Details
### Dataset Description
This dataset contains approximately 305,000 triplets of `question`, `answer`, and `image` designed for patch-based visual reasoning tasks.
A standard question in this dataset is formatted as follows:
> Image Grid: The image is divided into a 4x4 grid of 16 equal-sized patches. Patches are numbered sequentially from the top-left corner and moving right, then down to the next row.
> Task: Identify the patch number(s) that contain a potted plant.
> Response Format: Provide only the relevant patch number(s) as a list (e.g., [3], [5, 12], or [] if none are found).
The dataset is built on top of **COCO-2017**, from which object bounding boxes (bboxes) are used to generate questions and answers.
- **Curated by:** Yurii Potapov
- **Language(s) :** English
- **License:** Annotations and code: CC BY 4.0 (COCO), Images: Flickr Terms of Use
### Dataset Sources
- **Repository:** [Not yet published]
- **Paper:** [Not yet published]
- **Demo:** [More Information Needed]
## Uses
### Direct Use
- Training and evaluating **visual-language models (VLMs)** or other multimodal models.
- Patch-based object detection and reasoning.
- Research in **image question answering**, **visual reasoning**, and **multimodal representation learning**.
### Out-of-Scope Use
- Direct commercial redistribution of original COCO images without following Flickr Terms of Use.
- Use cases where original images are required to be displayed in full, due to copyright restrictions.
## Dataset Structure
- **question**: A textual description of the task referring to a 4x4 patch grid.
- **answer**: List of integers representing the patch indices containing the target object(s).
- **image**: Corresponding COCO-2017 image (PIL Image object or file path).
The dataset contains no explicit splits; users can generate their own train/validation/test splits as needed.
## Dataset Creation
### Curation Rationale
The dataset was created to facilitate **patch-level visual question answering** and to improve the training of visual-language models using real-world images with structured spatial queries.
### Source Data
The dataset is based on COCO-2017 images and annotations. Bounding boxes from COCO are used to determine which patches contain specific objects (e.g., potted plants).
#### Data Collection and Processing
- Images are sourced from COCO-2017 (Flickr) respecting their Terms of Use.
- Bounding boxes from COCO are used to automatically generate 4x4 grid questions.
- Each question asks which patch(es) contain a specific object.
- Answers are stored as lists of patch indices.
#### Who are the source data producers?
Original images were contributed by Flickr users and annotated by the COCO Consortium.
### Annotations
#### Annotation process
Annotations (bounding boxes) are sourced from COCO-2017. Patch assignments and questions were automatically generated programmatically based on bounding box locations.
#### Who are the annotators?
Annotations are from COCO annotators; patch-level questions are generated automatically.
#### Personal and Sensitive Information
The dataset does **not contain personal or sensitive information**.
## Bias, Risks, and Limitations
- Images reflect the distribution and biases present in COCO-2017.
- Models trained on this dataset may inherit biases from the original dataset.
- Limited to the objects annotated in COCO-2017.
### Recommendations
Users should be aware of the **copyright limitations of the original images** and provide attribution for COCO annotations. Use transformed or model-generated outputs rather than raw images for publication if possible.
## Glossary
- **Patch**: One of 16 equally sized blocks in a 4x4 grid over an image.
- **VLM (Visual-Language Model)**: A model that learns joint representations of images and text.
## Dataset Card Authors
Yurii Potapov
## Dataset Card Contact
yurii.a.potapov@gmail.com