Add comprehensive model card for OmniTry

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +77 -3
README.md CHANGED
@@ -1,3 +1,77 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ pipeline_tag: image-to-image
4
+ library_name: diffusers
5
+ ---
6
+
7
+ # OmniTry: Virtual Try-On Anything without Masks
8
+
9
+ OmniTry is a unified framework that extends Virtual Try-On (VTON) beyond traditional garments to encompass any wearable objects, such as jewelry and accessories. It operates in a mask-free setting for more practical applications. The framework uses a two-staged pipeline, leveraging large-scale unpaired images for mask-free localization and then fine-tuning with paired images for object appearance consistency.
10
+
11
+ - πŸ“š [**Paper: OmniTry: Virtual Try-On Anything without Masks**](https://huggingface.co/papers/2508.13632)
12
+ - 🌐 [**Project Page**](https://omnitry.github.io/)
13
+ - πŸ’» [**GitHub Repository**](https://github.com/Kunbyte-AI/OmniTry)
14
+ - πŸ€— [**Hugging Face Spaces Demo**](https://huggingface.co/spaces/Kunbyte/OmniTry)
15
+ - πŸ“Š [**Hugging Face Benchmark Dataset**](https://huggingface.co/datasets/Kunbyte/OmniTry-Bench)
16
+
17
+ <p align="center">
18
+ <img src="https://github.com/Kunbyte-AI/OmniTry/raw/main/assets/teaser.png">
19
+ </p>
20
+
21
+ ## Abstract
22
+ Virtual Try-ON (VTON) is a practical and widely-applied task, for which most of existing works focus on clothes. This paper presents OmniTry, a unified framework that extends VTON beyond garment to encompass any wearable objects, e.g., jewelries and accessories, with mask-free setting for more practical application. When extending to various types of objects, data curation is challenging for obtaining paired images, i.e., the object image and the corresponding try-on result. To tackle this problem, we propose a two-staged pipeline: For the first stage, we leverage large-scale unpaired images, i.e., portraits with any wearable items, to train the model for mask-free localization. Specifically, we repurpose the inpainting model to automatically draw objects in suitable positions given an empty mask. For the second stage, the model is further fine-tuned with paired images to transfer the consistency of object appearance. We observed that the model after the first stage shows quick convergence even with few paired samples. OmniTry is evaluated on a comprehensive benchmark consisting of 12 common classes of wearable objects, with both in-shop and in-the-wild images. Experimental results suggest that OmniTry shows better performance on both object localization and ID-preservation compared with existing methods. The code, model weights, and evaluation benchmark of OmniTry will be made publicly available at this https URL .
23
+
24
+ ## News
25
+ * **[2025.08.20]** πŸŽ‰πŸŽ‰πŸŽ‰ We release the model weights, inference demo and evaluation benchmark of OmniTry! To experience our advanced version and other related features, please visit our product website [k-fashionshop](https://marketing.k-fashionshop.com/home) (in Chinese) or [visboom](https://www.visboom.com/) (in English).
26
+
27
+ ## Get Started
28
+
29
+ **Note**: Currently, OmniTry requires at least **28GB** of VRAM for inference under `torch.bfloat16`. We will continue work to decrease memory requirements.
30
+
31
+ ### Download Checkpoints
32
+ 1. Create the checkpoint directory: `mkdir checkpoints`
33
+
34
+ 2. Download the [FLUX.1-Fill-dev](https://huggingface.co/black-forest-labs/FLUX.1-Fill-dev) into `checkpoints/FLUX.1-Fill-dev`
35
+
36
+ 3. Download the [LoRA of OmniTry](https://huggingface.co/Kunbyte/OmniTry/blob/main/omnitry_v1_unified.safetensors) into `checkpoints/omnitry_v1_unified.safetensors`. You can also download the `omnitry_v1_clothes.safetensors` that specifically finetuned on the clothe data only.
37
+
38
+ ### Environment Preparation
39
+ Install the environment with `conda`
40
+ ```bash
41
+ conda env create -f environment.yml
42
+ conda activate omnitry
43
+ ```
44
+ or `pip`:
45
+ ```bash
46
+ pip install -r requirements.txt
47
+ ```
48
+
49
+ (Optional) We recommend to install the [flash-attention](https://github.com/Dao-AILab/flash-attention/tree/main) to accelerate the inference process:
50
+ ```bash
51
+ pip install flash-attn==2.6.3
52
+ ```
53
+
54
+ ### Usage
55
+ For running the gradio demo:
56
+ ```bash
57
+ python gradio_demo.py
58
+ ```
59
+
60
+ To change different versions of checkpoints for OmniTry, replace the `lora_path` in `configs/omnitry_v1_unified.yaml`.
61
+
62
+ ## OmniTry-Bench
63
+ We present a unified evaluation benchmark for OmniTry. Please refer to the [OmniTry-Bench](https://github.com/Kunbyte-AI/OmniTry/blob/main/omnitry_bench/README.MD).
64
+
65
+ ## Acknowledgements
66
+ This project is developed on the [diffusers](https://github.com/huggingface/diffusers) and [FLUX](https://github.com/black-forest-labs/flux). We appreciate the contributors for their awesome works.
67
+
68
+ ## Citation
69
+ If you find this codebase useful for your research, please use the following entry.
70
+ ```BibTeX
71
+ @article{feng2025omnitry,
72
+ title={OmniTry: Virtual Try-On Anything without Masks},
73
+ author={Feng, Yutong and Zhang, Linlin and Cao, Hengyuan and Chen, Yiming and Feng, Xiaoduan and Cao, Jian and Wu, Yuxiong and Wang, Bin},
74
+ journal={arXiv preprint arXiv:2508.13632},
75
+ year={2025}
76
+ }
77
+ ```