OpenVTON / README.md
Rivflyyyy's picture
Update README.md
82fdee6 verified
---
dataset_info:
features:
- name: source
dtype: image
- name: mask
dtype: image
- name: target
dtype: image
- name: caption
dtype: string
- name: category
dtype: string
splits:
- name: train
num_examples: 89927
- name: validation
num_examples: 4989
- name: test
num_examples: 5009
license: cc-by-nc-4.0
task_categories:
- image-to-image
tags:
- virtual-try-on
- fashion
- clothing
---
# OpenVTON
A large-scale virtual try-on dataset containing ~100K clothing image pairs with garment masks.
You
## Dataset Structure
Each sample contains:
- **source**: Garment image (clothing item)
- **mask**: Garment segmentation mask
- **target**: Person wearing the garment (ground truth)
- **caption**: Text description of the clothing
- **category**: Clothing category (e.g., pants, jeans, shirt)
## Splits
| Split | Samples |
|-------|---------|
| Train | 89,927 |
| Validation | 4,989 |
| Test | 5,009 |
| **Total** | **99,925** |
## Usage
```python
from datasets import load_dataset
dataset = load_dataset("RenxingIntelligence/OpenVTON")
sample = dataset["train"][0]
sample["source"].show() # garment image
sample["mask"].show() # segmentation mask
sample["target"].show() # person wearing garment
print(sample["caption"])
print(sample["category"])
```
## Benchmark and Paper
This dataset is part of **OpenVTON-Bench**, a large-scale benchmark designed for the systematic evaluation of controllable virtual try-on (VTON) models.
**OpenVTON-Bench** is introduced in our paper:
> **OpenVTON-Bench: A Large-Scale Benchmark for Controllable Virtual Try-On**
> 📄 Paper: [https://arxiv.org/abs/2601.22725](https://arxiv.org/abs/2601.22725)
> 💻 Code: [https://github.com/RenxingIntelligence/OpenVTON-Bench](https://github.com/RenxingIntelligence/OpenVTON-Bench)
OpenVTON-Bench provides a standardized evaluation protocol for modern diffusion-based and transformer-based virtual try-on systems, enabling fair and reproducible comparison across different architectures.
---
## About OpenVTON-Bench
**OpenVTON-Bench** is a **large-scale, high-resolution benchmark** designed for the **systematic evaluation of controllable virtual try-on models**.
Unlike existing datasets and evaluation protocols that struggle with texture details and semantic consistency, OpenVTON-Bench provides:
* 🖼️ **~100K Image Pairs** with resolutions up to **1536×1536**, enabling evaluation of fine-grained texture generation.
* 🏷️ **Fine-Grained Taxonomy** covering **20 garment categories** for balanced semantic evaluation.
* 📐 **Multi-Level Automated Evaluation**, including:
* Pixel fidelity
* Garment consistency
* Semantic realism
This benchmark enables **fair, reproducible, and scalable comparison** across modern virtual try-on systems.
---
## Citation
If you use this dataset or the benchmark in your research, please cite:
```bibtex
@misc{li2026openvtonbenchlargescalehighresolutionbenchmark,
title={OpenVTON-Bench: A Large-Scale High-Resolution Benchmark for Controllable Virtual Try-On Evaluation},
author={Jin Li and Tao Chen and Shuai Jiang and Weijie Wang and Jingwen Luo and Chenhui Wu},
year={2026},
eprint={2601.22725},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2601.22725},
}
```