|
|
--- |
|
|
pretty_name: "Pokemon BLIP Captions" |
|
|
license: "cc-by-nc-sa-4.0" |
|
|
language: |
|
|
- en |
|
|
tags: |
|
|
- image |
|
|
- text |
|
|
- captioning |
|
|
- text-to-image |
|
|
task_categories: |
|
|
- text-to-image |
|
|
- image-to-text |
|
|
--- |
|
|
|
|
|
|
|
|
|
|
|
This repository hosts a copy of the **Pokemon BLIP Captions** dataset (when available), |
|
|
a small image–text dataset commonly used for experimenting with text-to-image fine-tuning and captioning. |
|
|
|
|
|
It contains Pokemon-style images paired with automatically generated captions (e.g., produced by BLIP-style captioning models). |
|
|
|
|
|
|
|
|
Typical files include: |
|
|
|
|
|
- Image files (or packaged archives/shards) |
|
|
- Captions / metadata (commonly as JSON/CSV/Parquet) |
|
|
|
|
|
A common entry conceptually looks like: |
|
|
```json |
|
|
{ |
|
|
"image": "<image>", |
|
|
"text": "..." |
|
|
} |
|
|
``` |
|
|
|
|
|
|
|
|
```python |
|
|
from datasets import load_dataset |
|
|
ds = load_dataset("lambdalabs/pokemon-blip-captions") |
|
|
``` |
|
|
|
|
|
|
|
|
This dataset is taken from the Hugging Face dataset repository: |
|
|
https://huggingface.co/datasets/lambdalabs/pokemon-blip-captions |
|
|
|
|
|
|
|
|
The Hugging Face dataset card lists the license as **CC BY-NC-SA 4.0**. |
|
|
|
|
|
**Note:** The Hugging Face repository currently shows a **DMCA takedown notice** from The Pokémon Company International, Inc. |
|
|
Please review the dataset page and any applicable legal requirements before using or redistributing any content. |
|
|
|