feylur commited on
Commit
58e8230
Β·
verified Β·
1 Parent(s): 77bed7f

Upload folder using huggingface_hub

Browse files
Files changed (26) hide show
  1. .gitattributes +7 -0
  2. blobs/1776bb349350ff04eeda7e7b0bcde1be666d0095 +8 -0
  3. blobs/24fa3254ceeb74c8435458994a64b522fb439a3635b7b86ff470457e0413da00 +3 -0
  4. blobs/6d57470c093629c0f55764926f318e04c96a134a +183 -0
  5. blobs/915df7bf19a33bee36a28d5f9ceaef1e2267c47526f98ca9e4c49e90ae5f0fd0 +3 -0
  6. blobs/a1fc093f1b6744623079e6f4e7313411f524e388c4b7467df1e0e7f577cba23a +3 -0
  7. blobs/a6344aac8c09253b3b630fb776ae94478aa0275b +35 -0
  8. blobs/b8a7382001b16e453bad95ca9dbc68ae8f2b839b304cf90eaf5c27fbdb4dae91 +3 -0
  9. blobs/b9cfb1a2a2a88e332c454115b80a3989b4d053443e4b5483083a6a92321b31db +3 -0
  10. blobs/c1134e1786254aff2161d2acc8dd7e9035ee5989 +48 -0
  11. blobs/d65a6c64a9bc48e8009a0006d5272b9332c1077c9c2a74302cbac9f256e84cbf +3 -0
  12. blobs/e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 +0 -0
  13. blobs/e9d7c91ce3b4e7133df56b599fc817b533e3439c5e8d282a59126d2fda339a2a +3 -0
  14. refs/main +1 -0
  15. snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/.gitattributes +35 -0
  16. snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/DensePose/Base-DensePose-RCNN-FPN.yaml +48 -0
  17. snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/DensePose/densepose_rcnn_R_50_FPN_s1x.yaml +8 -0
  18. snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/DensePose/model_final_162be9.pkl +3 -0
  19. snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/README.md +183 -0
  20. snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/SCHP/exp-schp-201908261155-lip.pth +3 -0
  21. snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/SCHP/exp-schp-201908301523-atr.pth +3 -0
  22. snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/config.json +0 -0
  23. snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/dresscode-16k-512/attention/model.safetensors +3 -0
  24. snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/flux-lora/pytorch_lora_weights.safetensors +3 -0
  25. snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/mix-48k-1024/attention/model.safetensors +3 -0
  26. snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/vitonhd-16k-512/attention/model.safetensors +3 -0
.gitattributes CHANGED
@@ -33,3 +33,10 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ blobs/24fa3254ceeb74c8435458994a64b522fb439a3635b7b86ff470457e0413da00 filter=lfs diff=lfs merge=lfs -text
37
+ blobs/915df7bf19a33bee36a28d5f9ceaef1e2267c47526f98ca9e4c49e90ae5f0fd0 filter=lfs diff=lfs merge=lfs -text
38
+ blobs/a1fc093f1b6744623079e6f4e7313411f524e388c4b7467df1e0e7f577cba23a filter=lfs diff=lfs merge=lfs -text
39
+ blobs/b8a7382001b16e453bad95ca9dbc68ae8f2b839b304cf90eaf5c27fbdb4dae91 filter=lfs diff=lfs merge=lfs -text
40
+ blobs/b9cfb1a2a2a88e332c454115b80a3989b4d053443e4b5483083a6a92321b31db filter=lfs diff=lfs merge=lfs -text
41
+ blobs/d65a6c64a9bc48e8009a0006d5272b9332c1077c9c2a74302cbac9f256e84cbf filter=lfs diff=lfs merge=lfs -text
42
+ blobs/e9d7c91ce3b4e7133df56b599fc817b533e3439c5e8d282a59126d2fda339a2a filter=lfs diff=lfs merge=lfs -text
blobs/1776bb349350ff04eeda7e7b0bcde1be666d0095 ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ _BASE_: "Base-DensePose-RCNN-FPN.yaml"
2
+ MODEL:
3
+ WEIGHTS: "detectron2://ImageNetPretrained/MSRA/R-50.pkl"
4
+ RESNETS:
5
+ DEPTH: 50
6
+ SOLVER:
7
+ MAX_ITER: 130000
8
+ STEPS: (100000, 120000)
blobs/24fa3254ceeb74c8435458994a64b522fb439a3635b7b86ff470457e0413da00 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24fa3254ceeb74c8435458994a64b522fb439a3635b7b86ff470457e0413da00
3
+ size 267449349
blobs/6d57470c093629c0f55764926f318e04c96a134a ADDED
@@ -0,0 +1,183 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-sa-4.0
3
+ ---
4
+ # 🐈 CatVTON: Concatenation Is All You Need for Virtual Try-On with Diffusion Models
5
+
6
+ <div style="display: flex; justify-content: center; align-items: center;">
7
+ <a href="http://arxiv.org/abs/2407.15886" style="margin: 0 2px;">
8
+ <img src='https://img.shields.io/badge/arXiv-2407.15886-red?style=flat&logo=arXiv&logoColor=red' alt='arxiv'>
9
+ </a>
10
+ <a href='https://huggingface.co/zhengchong/CatVTON' style="margin: 0 2px;">
11
+ <img src='https://img.shields.io/badge/Hugging Face-ckpts-orange?style=flat&logo=HuggingFace&logoColor=orange' alt='huggingface'>
12
+ </a>
13
+ <a href="https://github.com/Zheng-Chong/CatVTON" style="margin: 0 2px;">
14
+ <img src='https://img.shields.io/badge/GitHub-Repo-blue?style=flat&logo=GitHub' alt='GitHub'>
15
+ </a>
16
+ <a href="http://120.76.142.206:8888" style="margin: 0 2px;">
17
+ <img src='https://img.shields.io/badge/Demo-Gradio-gold?style=flat&logo=Gradio&logoColor=red' alt='Demo'>
18
+ </a>
19
+ <a href="https://huggingface.co/spaces/zhengchong/CatVTON" style="margin: 0 2px;">
20
+ <img src='https://img.shields.io/badge/Space-ZeroGPU-orange?style=flat&logo=Gradio&logoColor=red' alt='Demo'>
21
+ </a>
22
+ <a href='https://zheng-chong.github.io/CatVTON/' style="margin: 0 2px;">
23
+ <img src='https://img.shields.io/badge/Webpage-Project-silver?style=flat&logo=&logoColor=orange' alt='webpage'>
24
+ </a>
25
+ <a href="https://github.com/Zheng-Chong/CatVTON/LICENCE" style="margin: 0 2px;">
26
+ <img src='https://img.shields.io/badge/License-CC BY--NC--SA--4.0-lightgreen?style=flat&logo=Lisence' alt='License'>
27
+ </a>
28
+ </div>
29
+
30
+
31
+ **CatVTON** is a simple and efficient virtual try-on diffusion model with ***1) Lightweight Network (899.06M parameters totally)***, ***2) Parameter-Efficient Training (49.57M parameters trainable)*** and ***3) Simplified Inference (< 8G VRAM for 1024X768 resolution)***.
32
+
33
+
34
+
35
+ ## Updates
36
+ - **`2024/10/17`**:[**Mask-free version**](https://huggingface.co/zhengchong/CatVTON-MaskFree)πŸ€— of CatVTON is release and please try it in our [**Online Demo**](http://120.76.142.206:8888).
37
+ - **`2024/10/13`**: We have built a repo [**Awesome-Try-On-Models**](https://github.com/Zheng-Chong/Awesome-Try-On-Models) that focuses on image, video, and 3D-based try-on models published after 2023, aiming to provide insights into the latest technological trends. If you're interested, feel free to contribute or give it a 🌟 star!
38
+ - **`2024/08/13`**: We localize DensePose & SCHP to avoid certain environment issues.
39
+ - **`2024/08/10`**: Our πŸ€— [**HuggingFace Space**](https://huggingface.co/spaces/zhengchong/CatVTON) is available now! Thanks for the grant from [**ZeroGPU**](https://huggingface.co/zero-gpu-explorers)!
40
+ - **`2024/08/09`**: [**Evaluation code**](https://github.com/Zheng-Chong/CatVTON?tab=readme-ov-file#3-calculate-metrics) is provided to calculate metrics πŸ“š.
41
+ - **`2024/07/27`**: We provide code and workflow for deploying CatVTON on [**ComfyUI**](https://github.com/Zheng-Chong/CatVTON?tab=readme-ov-file#comfyui-workflow) πŸ’₯.
42
+ - **`2024/07/24`**: Our [**Paper on ArXiv**](http://arxiv.org/abs/2407.15886) is available πŸ₯³!
43
+ - **`2024/07/22`**: Our [**App Code**](https://github.com/Zheng-Chong/CatVTON/blob/main/app.py) is released, deploy and enjoy CatVTON on your mechine πŸŽ‰!
44
+ - **`2024/07/21`**: Our [**Inference Code**](https://github.com/Zheng-Chong/CatVTON/blob/main/inference.py) and [**Weights** πŸ€—](https://huggingface.co/zhengchong/CatVTON) are released.
45
+ - **`2024/07/11`**: Our [**Online Demo**](http://120.76.142.206:8888) is released 😁.
46
+
47
+
48
+
49
+
50
+ ## Installation
51
+
52
+ Create a conda environment & Install requirments
53
+ ```shell
54
+ conda create -n catvton python==3.9.0
55
+ conda activate catvton
56
+ cd CatVTON-main # or your path to CatVTON project dir
57
+ pip install -r requirements.txt
58
+ ```
59
+
60
+ ## Deployment
61
+ ### ComfyUI Workflow
62
+ We have modified the main code to enable easy deployment of CatVTON on [ComfyUI](https://github.com/comfyanonymous/ComfyUI). Due to the incompatibility of the code structure, we have released this part in the [Releases](https://github.com/Zheng-Chong/CatVTON/releases/tag/ComfyUI), which includes the code placed under `custom_nodes` of ComfyUI and our workflow JSON files.
63
+
64
+ To deploy CatVTON to your ComfyUI, follow these steps:
65
+ 1. Install all the requirements for both CatVTON and ComfyUI, refer to [Installation Guide for CatVTON](https://github.com/Zheng-Chong/CatVTON/blob/main/INSTALL.md) and [Installation Guide for ComfyUI](https://github.com/comfyanonymous/ComfyUI?tab=readme-ov-file#installing).
66
+ 2. Download [`ComfyUI-CatVTON.zip`](https://github.com/Zheng-Chong/CatVTON/releases/download/ComfyUI/ComfyUI-CatVTON.zip) and unzip it in the `custom_nodes` folder under your ComfyUI project (clone from [ComfyUI](https://github.com/comfyanonymous/ComfyUI)).
67
+ 3. Run the ComfyUI.
68
+ 4. Download [`catvton_workflow.json`](https://github.com/Zheng-Chong/CatVTON/releases/download/ComfyUI/catvton_workflow.json) and drag it into you ComfyUI webpage and enjoy πŸ˜†!
69
+
70
+ > Problems under Windows OS, please refer to [issue#8](https://github.com/Zheng-Chong/CatVTON/issues/8).
71
+ >
72
+ When you run the CatVTON workflow for the first time, the weight files will be automatically downloaded, usually taking dozens of minutes.
73
+
74
+ <div align="center">
75
+ <img src="resource/img/comfyui-1.png" width="100%" height="100%"/>
76
+ </div>
77
+
78
+ <!-- <div align="center">
79
+ <img src="resource/img/comfyui.png" width="100%" height="100%"/>
80
+ </div> -->
81
+
82
+ ### Gradio App
83
+
84
+ To deploy the Gradio App for CatVTON on your machine, run the following command, and checkpoints will be automatically downloaded from HuggingFace.
85
+
86
+ ```PowerShell
87
+ CUDA_VISIBLE_DEVICES=0 python app.py \
88
+ --output_dir="resource/demo/output" \
89
+ --mixed_precision="bf16" \
90
+ --allow_tf32
91
+ ```
92
+ When using `bf16` precision, generating results with a resolution of `1024x768` only requires about `8G` VRAM.
93
+
94
+ ## Inference
95
+ ### 1. Data Preparation
96
+ Before inference, you need to download the [VITON-HD](https://github.com/shadow2496/VITON-HD) or [DressCode](https://github.com/aimagelab/dress-code) dataset.
97
+ Once the datasets are downloaded, the folder structures should look like these:
98
+ ```
99
+ β”œβ”€β”€ VITON-HD
100
+ | β”œβ”€β”€ test_pairs_unpaired.txt
101
+ β”‚ β”œβ”€β”€ test
102
+ | | β”œβ”€β”€ image
103
+ β”‚ β”‚ β”‚ β”œβ”€β”€ [000006_00.jpg | 000008_00.jpg | ...]
104
+ β”‚ β”‚ β”œβ”€β”€ cloth
105
+ β”‚ β”‚ β”‚ β”œβ”€β”€ [000006_00.jpg | 000008_00.jpg | ...]
106
+ β”‚ β”‚ β”œβ”€β”€ agnostic-mask
107
+ β”‚ β”‚ β”‚ β”œβ”€β”€ [000006_00_mask.png | 000008_00.png | ...]
108
+ ...
109
+ ```
110
+
111
+ ```
112
+ β”œβ”€β”€ DressCode
113
+ | β”œβ”€β”€ test_pairs_paired.txt
114
+ | β”œβ”€β”€ test_pairs_unpaired.txt
115
+ β”‚ β”œβ”€β”€ [dresses | lower_body | upper_body]
116
+ | | β”œβ”€β”€ test_pairs_paired.txt
117
+ | | β”œβ”€β”€ test_pairs_unpaired.txt
118
+ β”‚ β”‚ β”œβ”€β”€ images
119
+ β”‚ β”‚ β”‚ β”œβ”€β”€ [013563_0.jpg | 013563_1.jpg | 013564_0.jpg | 013564_1.jpg | ...]
120
+ β”‚ β”‚ β”œβ”€β”€ agnostic_masks
121
+ β”‚ β”‚ β”‚ β”œβ”€β”€ [013563_0.png| 013564_0.png | ...]
122
+ ...
123
+ ```
124
+ For the DressCode dataset, we provide script to preprocessed agnostic masks, run the following command:
125
+ ```PowerShell
126
+ CUDA_VISIBLE_DEVICES=0 python preprocess_agnostic_mask.py \
127
+ --data_root_path <your_path_to_DressCode>
128
+ ```
129
+
130
+ ### 2. Inference on VTIONHD/DressCode
131
+ To run the inference on the DressCode or VITON-HD dataset, run the following command, checkpoints will be automatically downloaded from HuggingFace.
132
+
133
+ ```PowerShell
134
+ CUDA_VISIBLE_DEVICES=0 python inference.py \
135
+ --dataset [dresscode | vitonhd] \
136
+ --data_root_path <path> \
137
+ --output_dir <path>
138
+ --dataloader_num_workers 8 \
139
+ --batch_size 8 \
140
+ --seed 555 \
141
+ --mixed_precision [no | fp16 | bf16] \
142
+ --allow_tf32 \
143
+ --repaint \
144
+ --eval_pair
145
+ ```
146
+ ### 3. Calculate Metrics
147
+
148
+ After obtaining the inference results, calculate the metrics using the following command:
149
+
150
+ ```PowerShell
151
+ CUDA_VISIBLE_DEVICES=0 python eval.py \
152
+ --gt_folder <your_path_to_gt_image_folder> \
153
+ --pred_folder <your_path_to_predicted_image_folder> \
154
+ --paired \
155
+ --batch_size=16 \
156
+ --num_workers=16
157
+ ```
158
+
159
+ - `--gt_folder` and `--pred_folder` should be folders that contain **only images**.
160
+ - To evaluate the results in a paired setting, use `--paired`; for an unpaired setting, simply omit it.
161
+ - `--batch_size` and `--num_workers` should be adjusted based on your machine.
162
+
163
+
164
+ ## Acknowledgement
165
+ Our code is modified based on [Diffusers](https://github.com/huggingface/diffusers). We adopt [Stable Diffusion v1.5 inpainting](https://huggingface.co/runwayml/stable-diffusion-inpainting) as the base model. We use [SCHP](https://github.com/GoGoDuck912/Self-Correction-Human-Parsing/tree/master) and [DensePose](https://github.com/facebookresearch/DensePose) to automatically generate masks in our [Gradio](https://github.com/gradio-app/gradio) App and [ComfyUI](https://github.com/comfyanonymous/ComfyUI) workflow. Thanks to all the contributors!
166
+
167
+ ## License
168
+ All the materials, including code, checkpoints, and demo, are made available under the [Creative Commons BY-NC-SA 4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/) license. You are free to copy, redistribute, remix, transform, and build upon the project for non-commercial purposes, as long as you give appropriate credit and distribute your contributions under the same license.
169
+
170
+
171
+ ## Citation
172
+
173
+ ```bibtex
174
+ @misc{chong2024catvtonconcatenationneedvirtual,
175
+ title={CatVTON: Concatenation Is All You Need for Virtual Try-On with Diffusion Models},
176
+ author={Zheng Chong and Xiao Dong and Haoxiang Li and Shiyue Zhang and Wenqing Zhang and Xujie Zhang and Hanqing Zhao and Xiaodan Liang},
177
+ year={2024},
178
+ eprint={2407.15886},
179
+ archivePrefix={arXiv},
180
+ primaryClass={cs.CV},
181
+ url={https://arxiv.org/abs/2407.15886},
182
+ }
183
+ ```
blobs/915df7bf19a33bee36a28d5f9ceaef1e2267c47526f98ca9e4c49e90ae5f0fd0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:915df7bf19a33bee36a28d5f9ceaef1e2267c47526f98ca9e4c49e90ae5f0fd0
3
+ size 198303368
blobs/a1fc093f1b6744623079e6f4e7313411f524e388c4b7467df1e0e7f577cba23a ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1fc093f1b6744623079e6f4e7313411f524e388c4b7467df1e0e7f577cba23a
3
+ size 198303368
blobs/a6344aac8c09253b3b630fb776ae94478aa0275b ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
blobs/b8a7382001b16e453bad95ca9dbc68ae8f2b839b304cf90eaf5c27fbdb4dae91 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8a7382001b16e453bad95ca9dbc68ae8f2b839b304cf90eaf5c27fbdb4dae91
3
+ size 255757821
blobs/b9cfb1a2a2a88e332c454115b80a3989b4d053443e4b5483083a6a92321b31db ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9cfb1a2a2a88e332c454115b80a3989b4d053443e4b5483083a6a92321b31db
3
+ size 37406344
blobs/c1134e1786254aff2161d2acc8dd7e9035ee5989 ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ VERSION: 2
2
+ MODEL:
3
+ META_ARCHITECTURE: "GeneralizedRCNN"
4
+ BACKBONE:
5
+ NAME: "build_resnet_fpn_backbone"
6
+ RESNETS:
7
+ OUT_FEATURES: ["res2", "res3", "res4", "res5"]
8
+ FPN:
9
+ IN_FEATURES: ["res2", "res3", "res4", "res5"]
10
+ ANCHOR_GENERATOR:
11
+ SIZES: [[32], [64], [128], [256], [512]] # One size for each in feature map
12
+ ASPECT_RATIOS: [[0.5, 1.0, 2.0]] # Three aspect ratios (same for all in feature maps)
13
+ RPN:
14
+ IN_FEATURES: ["p2", "p3", "p4", "p5", "p6"]
15
+ PRE_NMS_TOPK_TRAIN: 2000 # Per FPN level
16
+ PRE_NMS_TOPK_TEST: 1000 # Per FPN level
17
+ # Detectron1 uses 2000 proposals per-batch,
18
+ # (See "modeling/rpn/rpn_outputs.py" for details of this legacy issue)
19
+ # which is approximately 1000 proposals per-image since the default batch size for FPN is 2.
20
+ POST_NMS_TOPK_TRAIN: 1000
21
+ POST_NMS_TOPK_TEST: 1000
22
+
23
+ DENSEPOSE_ON: True
24
+ ROI_HEADS:
25
+ NAME: "DensePoseROIHeads"
26
+ IN_FEATURES: ["p2", "p3", "p4", "p5"]
27
+ NUM_CLASSES: 1
28
+ ROI_BOX_HEAD:
29
+ NAME: "FastRCNNConvFCHead"
30
+ NUM_FC: 2
31
+ POOLER_RESOLUTION: 7
32
+ POOLER_SAMPLING_RATIO: 2
33
+ POOLER_TYPE: "ROIAlign"
34
+ ROI_DENSEPOSE_HEAD:
35
+ NAME: "DensePoseV1ConvXHead"
36
+ POOLER_TYPE: "ROIAlign"
37
+ NUM_COARSE_SEGM_CHANNELS: 2
38
+ DATASETS:
39
+ TRAIN: ("densepose_coco_2014_train", "densepose_coco_2014_valminusminival")
40
+ TEST: ("densepose_coco_2014_minival",)
41
+ SOLVER:
42
+ IMS_PER_BATCH: 16
43
+ BASE_LR: 0.01
44
+ STEPS: (60000, 80000)
45
+ MAX_ITER: 90000
46
+ WARMUP_FACTOR: 0.1
47
+ INPUT:
48
+ MIN_SIZE_TRAIN: (640, 672, 704, 736, 768, 800)
blobs/d65a6c64a9bc48e8009a0006d5272b9332c1077c9c2a74302cbac9f256e84cbf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d65a6c64a9bc48e8009a0006d5272b9332c1077c9c2a74302cbac9f256e84cbf
3
+ size 198303368
blobs/e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 ADDED
File without changes
blobs/e9d7c91ce3b4e7133df56b599fc817b533e3439c5e8d282a59126d2fda339a2a ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9d7c91ce3b4e7133df56b599fc817b533e3439c5e8d282a59126d2fda339a2a
3
+ size 267445237
refs/main ADDED
@@ -0,0 +1 @@
 
 
1
+ 2969fcf85fe62f2036605716f0b56f0b81d01d79
snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/.gitattributes ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/DensePose/Base-DensePose-RCNN-FPN.yaml ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ VERSION: 2
2
+ MODEL:
3
+ META_ARCHITECTURE: "GeneralizedRCNN"
4
+ BACKBONE:
5
+ NAME: "build_resnet_fpn_backbone"
6
+ RESNETS:
7
+ OUT_FEATURES: ["res2", "res3", "res4", "res5"]
8
+ FPN:
9
+ IN_FEATURES: ["res2", "res3", "res4", "res5"]
10
+ ANCHOR_GENERATOR:
11
+ SIZES: [[32], [64], [128], [256], [512]] # One size for each in feature map
12
+ ASPECT_RATIOS: [[0.5, 1.0, 2.0]] # Three aspect ratios (same for all in feature maps)
13
+ RPN:
14
+ IN_FEATURES: ["p2", "p3", "p4", "p5", "p6"]
15
+ PRE_NMS_TOPK_TRAIN: 2000 # Per FPN level
16
+ PRE_NMS_TOPK_TEST: 1000 # Per FPN level
17
+ # Detectron1 uses 2000 proposals per-batch,
18
+ # (See "modeling/rpn/rpn_outputs.py" for details of this legacy issue)
19
+ # which is approximately 1000 proposals per-image since the default batch size for FPN is 2.
20
+ POST_NMS_TOPK_TRAIN: 1000
21
+ POST_NMS_TOPK_TEST: 1000
22
+
23
+ DENSEPOSE_ON: True
24
+ ROI_HEADS:
25
+ NAME: "DensePoseROIHeads"
26
+ IN_FEATURES: ["p2", "p3", "p4", "p5"]
27
+ NUM_CLASSES: 1
28
+ ROI_BOX_HEAD:
29
+ NAME: "FastRCNNConvFCHead"
30
+ NUM_FC: 2
31
+ POOLER_RESOLUTION: 7
32
+ POOLER_SAMPLING_RATIO: 2
33
+ POOLER_TYPE: "ROIAlign"
34
+ ROI_DENSEPOSE_HEAD:
35
+ NAME: "DensePoseV1ConvXHead"
36
+ POOLER_TYPE: "ROIAlign"
37
+ NUM_COARSE_SEGM_CHANNELS: 2
38
+ DATASETS:
39
+ TRAIN: ("densepose_coco_2014_train", "densepose_coco_2014_valminusminival")
40
+ TEST: ("densepose_coco_2014_minival",)
41
+ SOLVER:
42
+ IMS_PER_BATCH: 16
43
+ BASE_LR: 0.01
44
+ STEPS: (60000, 80000)
45
+ MAX_ITER: 90000
46
+ WARMUP_FACTOR: 0.1
47
+ INPUT:
48
+ MIN_SIZE_TRAIN: (640, 672, 704, 736, 768, 800)
snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/DensePose/densepose_rcnn_R_50_FPN_s1x.yaml ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ _BASE_: "Base-DensePose-RCNN-FPN.yaml"
2
+ MODEL:
3
+ WEIGHTS: "detectron2://ImageNetPretrained/MSRA/R-50.pkl"
4
+ RESNETS:
5
+ DEPTH: 50
6
+ SOLVER:
7
+ MAX_ITER: 130000
8
+ STEPS: (100000, 120000)
snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/DensePose/model_final_162be9.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8a7382001b16e453bad95ca9dbc68ae8f2b839b304cf90eaf5c27fbdb4dae91
3
+ size 255757821
snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/README.md ADDED
@@ -0,0 +1,183 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-sa-4.0
3
+ ---
4
+ # 🐈 CatVTON: Concatenation Is All You Need for Virtual Try-On with Diffusion Models
5
+
6
+ <div style="display: flex; justify-content: center; align-items: center;">
7
+ <a href="http://arxiv.org/abs/2407.15886" style="margin: 0 2px;">
8
+ <img src='https://img.shields.io/badge/arXiv-2407.15886-red?style=flat&logo=arXiv&logoColor=red' alt='arxiv'>
9
+ </a>
10
+ <a href='https://huggingface.co/zhengchong/CatVTON' style="margin: 0 2px;">
11
+ <img src='https://img.shields.io/badge/Hugging Face-ckpts-orange?style=flat&logo=HuggingFace&logoColor=orange' alt='huggingface'>
12
+ </a>
13
+ <a href="https://github.com/Zheng-Chong/CatVTON" style="margin: 0 2px;">
14
+ <img src='https://img.shields.io/badge/GitHub-Repo-blue?style=flat&logo=GitHub' alt='GitHub'>
15
+ </a>
16
+ <a href="http://120.76.142.206:8888" style="margin: 0 2px;">
17
+ <img src='https://img.shields.io/badge/Demo-Gradio-gold?style=flat&logo=Gradio&logoColor=red' alt='Demo'>
18
+ </a>
19
+ <a href="https://huggingface.co/spaces/zhengchong/CatVTON" style="margin: 0 2px;">
20
+ <img src='https://img.shields.io/badge/Space-ZeroGPU-orange?style=flat&logo=Gradio&logoColor=red' alt='Demo'>
21
+ </a>
22
+ <a href='https://zheng-chong.github.io/CatVTON/' style="margin: 0 2px;">
23
+ <img src='https://img.shields.io/badge/Webpage-Project-silver?style=flat&logo=&logoColor=orange' alt='webpage'>
24
+ </a>
25
+ <a href="https://github.com/Zheng-Chong/CatVTON/LICENCE" style="margin: 0 2px;">
26
+ <img src='https://img.shields.io/badge/License-CC BY--NC--SA--4.0-lightgreen?style=flat&logo=Lisence' alt='License'>
27
+ </a>
28
+ </div>
29
+
30
+
31
+ **CatVTON** is a simple and efficient virtual try-on diffusion model with ***1) Lightweight Network (899.06M parameters totally)***, ***2) Parameter-Efficient Training (49.57M parameters trainable)*** and ***3) Simplified Inference (< 8G VRAM for 1024X768 resolution)***.
32
+
33
+
34
+
35
+ ## Updates
36
+ - **`2024/10/17`**:[**Mask-free version**](https://huggingface.co/zhengchong/CatVTON-MaskFree)πŸ€— of CatVTON is release and please try it in our [**Online Demo**](http://120.76.142.206:8888).
37
+ - **`2024/10/13`**: We have built a repo [**Awesome-Try-On-Models**](https://github.com/Zheng-Chong/Awesome-Try-On-Models) that focuses on image, video, and 3D-based try-on models published after 2023, aiming to provide insights into the latest technological trends. If you're interested, feel free to contribute or give it a 🌟 star!
38
+ - **`2024/08/13`**: We localize DensePose & SCHP to avoid certain environment issues.
39
+ - **`2024/08/10`**: Our πŸ€— [**HuggingFace Space**](https://huggingface.co/spaces/zhengchong/CatVTON) is available now! Thanks for the grant from [**ZeroGPU**](https://huggingface.co/zero-gpu-explorers)!
40
+ - **`2024/08/09`**: [**Evaluation code**](https://github.com/Zheng-Chong/CatVTON?tab=readme-ov-file#3-calculate-metrics) is provided to calculate metrics πŸ“š.
41
+ - **`2024/07/27`**: We provide code and workflow for deploying CatVTON on [**ComfyUI**](https://github.com/Zheng-Chong/CatVTON?tab=readme-ov-file#comfyui-workflow) πŸ’₯.
42
+ - **`2024/07/24`**: Our [**Paper on ArXiv**](http://arxiv.org/abs/2407.15886) is available πŸ₯³!
43
+ - **`2024/07/22`**: Our [**App Code**](https://github.com/Zheng-Chong/CatVTON/blob/main/app.py) is released, deploy and enjoy CatVTON on your mechine πŸŽ‰!
44
+ - **`2024/07/21`**: Our [**Inference Code**](https://github.com/Zheng-Chong/CatVTON/blob/main/inference.py) and [**Weights** πŸ€—](https://huggingface.co/zhengchong/CatVTON) are released.
45
+ - **`2024/07/11`**: Our [**Online Demo**](http://120.76.142.206:8888) is released 😁.
46
+
47
+
48
+
49
+
50
+ ## Installation
51
+
52
+ Create a conda environment & Install requirments
53
+ ```shell
54
+ conda create -n catvton python==3.9.0
55
+ conda activate catvton
56
+ cd CatVTON-main # or your path to CatVTON project dir
57
+ pip install -r requirements.txt
58
+ ```
59
+
60
+ ## Deployment
61
+ ### ComfyUI Workflow
62
+ We have modified the main code to enable easy deployment of CatVTON on [ComfyUI](https://github.com/comfyanonymous/ComfyUI). Due to the incompatibility of the code structure, we have released this part in the [Releases](https://github.com/Zheng-Chong/CatVTON/releases/tag/ComfyUI), which includes the code placed under `custom_nodes` of ComfyUI and our workflow JSON files.
63
+
64
+ To deploy CatVTON to your ComfyUI, follow these steps:
65
+ 1. Install all the requirements for both CatVTON and ComfyUI, refer to [Installation Guide for CatVTON](https://github.com/Zheng-Chong/CatVTON/blob/main/INSTALL.md) and [Installation Guide for ComfyUI](https://github.com/comfyanonymous/ComfyUI?tab=readme-ov-file#installing).
66
+ 2. Download [`ComfyUI-CatVTON.zip`](https://github.com/Zheng-Chong/CatVTON/releases/download/ComfyUI/ComfyUI-CatVTON.zip) and unzip it in the `custom_nodes` folder under your ComfyUI project (clone from [ComfyUI](https://github.com/comfyanonymous/ComfyUI)).
67
+ 3. Run the ComfyUI.
68
+ 4. Download [`catvton_workflow.json`](https://github.com/Zheng-Chong/CatVTON/releases/download/ComfyUI/catvton_workflow.json) and drag it into you ComfyUI webpage and enjoy πŸ˜†!
69
+
70
+ > Problems under Windows OS, please refer to [issue#8](https://github.com/Zheng-Chong/CatVTON/issues/8).
71
+ >
72
+ When you run the CatVTON workflow for the first time, the weight files will be automatically downloaded, usually taking dozens of minutes.
73
+
74
+ <div align="center">
75
+ <img src="resource/img/comfyui-1.png" width="100%" height="100%"/>
76
+ </div>
77
+
78
+ <!-- <div align="center">
79
+ <img src="resource/img/comfyui.png" width="100%" height="100%"/>
80
+ </div> -->
81
+
82
+ ### Gradio App
83
+
84
+ To deploy the Gradio App for CatVTON on your machine, run the following command, and checkpoints will be automatically downloaded from HuggingFace.
85
+
86
+ ```PowerShell
87
+ CUDA_VISIBLE_DEVICES=0 python app.py \
88
+ --output_dir="resource/demo/output" \
89
+ --mixed_precision="bf16" \
90
+ --allow_tf32
91
+ ```
92
+ When using `bf16` precision, generating results with a resolution of `1024x768` only requires about `8G` VRAM.
93
+
94
+ ## Inference
95
+ ### 1. Data Preparation
96
+ Before inference, you need to download the [VITON-HD](https://github.com/shadow2496/VITON-HD) or [DressCode](https://github.com/aimagelab/dress-code) dataset.
97
+ Once the datasets are downloaded, the folder structures should look like these:
98
+ ```
99
+ β”œβ”€β”€ VITON-HD
100
+ | β”œβ”€β”€ test_pairs_unpaired.txt
101
+ β”‚ β”œβ”€β”€ test
102
+ | | β”œβ”€β”€ image
103
+ β”‚ β”‚ β”‚ β”œβ”€β”€ [000006_00.jpg | 000008_00.jpg | ...]
104
+ β”‚ β”‚ β”œβ”€β”€ cloth
105
+ β”‚ β”‚ β”‚ β”œβ”€β”€ [000006_00.jpg | 000008_00.jpg | ...]
106
+ β”‚ β”‚ β”œβ”€β”€ agnostic-mask
107
+ β”‚ β”‚ β”‚ β”œβ”€β”€ [000006_00_mask.png | 000008_00.png | ...]
108
+ ...
109
+ ```
110
+
111
+ ```
112
+ β”œβ”€β”€ DressCode
113
+ | β”œβ”€β”€ test_pairs_paired.txt
114
+ | β”œβ”€β”€ test_pairs_unpaired.txt
115
+ β”‚ β”œβ”€β”€ [dresses | lower_body | upper_body]
116
+ | | β”œβ”€β”€ test_pairs_paired.txt
117
+ | | β”œβ”€β”€ test_pairs_unpaired.txt
118
+ β”‚ β”‚ β”œβ”€β”€ images
119
+ β”‚ β”‚ β”‚ β”œβ”€β”€ [013563_0.jpg | 013563_1.jpg | 013564_0.jpg | 013564_1.jpg | ...]
120
+ β”‚ β”‚ β”œβ”€β”€ agnostic_masks
121
+ β”‚ β”‚ β”‚ β”œβ”€β”€ [013563_0.png| 013564_0.png | ...]
122
+ ...
123
+ ```
124
+ For the DressCode dataset, we provide script to preprocessed agnostic masks, run the following command:
125
+ ```PowerShell
126
+ CUDA_VISIBLE_DEVICES=0 python preprocess_agnostic_mask.py \
127
+ --data_root_path <your_path_to_DressCode>
128
+ ```
129
+
130
+ ### 2. Inference on VTIONHD/DressCode
131
+ To run the inference on the DressCode or VITON-HD dataset, run the following command, checkpoints will be automatically downloaded from HuggingFace.
132
+
133
+ ```PowerShell
134
+ CUDA_VISIBLE_DEVICES=0 python inference.py \
135
+ --dataset [dresscode | vitonhd] \
136
+ --data_root_path <path> \
137
+ --output_dir <path>
138
+ --dataloader_num_workers 8 \
139
+ --batch_size 8 \
140
+ --seed 555 \
141
+ --mixed_precision [no | fp16 | bf16] \
142
+ --allow_tf32 \
143
+ --repaint \
144
+ --eval_pair
145
+ ```
146
+ ### 3. Calculate Metrics
147
+
148
+ After obtaining the inference results, calculate the metrics using the following command:
149
+
150
+ ```PowerShell
151
+ CUDA_VISIBLE_DEVICES=0 python eval.py \
152
+ --gt_folder <your_path_to_gt_image_folder> \
153
+ --pred_folder <your_path_to_predicted_image_folder> \
154
+ --paired \
155
+ --batch_size=16 \
156
+ --num_workers=16
157
+ ```
158
+
159
+ - `--gt_folder` and `--pred_folder` should be folders that contain **only images**.
160
+ - To evaluate the results in a paired setting, use `--paired`; for an unpaired setting, simply omit it.
161
+ - `--batch_size` and `--num_workers` should be adjusted based on your machine.
162
+
163
+
164
+ ## Acknowledgement
165
+ Our code is modified based on [Diffusers](https://github.com/huggingface/diffusers). We adopt [Stable Diffusion v1.5 inpainting](https://huggingface.co/runwayml/stable-diffusion-inpainting) as the base model. We use [SCHP](https://github.com/GoGoDuck912/Self-Correction-Human-Parsing/tree/master) and [DensePose](https://github.com/facebookresearch/DensePose) to automatically generate masks in our [Gradio](https://github.com/gradio-app/gradio) App and [ComfyUI](https://github.com/comfyanonymous/ComfyUI) workflow. Thanks to all the contributors!
166
+
167
+ ## License
168
+ All the materials, including code, checkpoints, and demo, are made available under the [Creative Commons BY-NC-SA 4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/) license. You are free to copy, redistribute, remix, transform, and build upon the project for non-commercial purposes, as long as you give appropriate credit and distribute your contributions under the same license.
169
+
170
+
171
+ ## Citation
172
+
173
+ ```bibtex
174
+ @misc{chong2024catvtonconcatenationneedvirtual,
175
+ title={CatVTON: Concatenation Is All You Need for Virtual Try-On with Diffusion Models},
176
+ author={Zheng Chong and Xiao Dong and Haoxiang Li and Shiyue Zhang and Wenqing Zhang and Xujie Zhang and Hanqing Zhao and Xiaodan Liang},
177
+ year={2024},
178
+ eprint={2407.15886},
179
+ archivePrefix={arXiv},
180
+ primaryClass={cs.CV},
181
+ url={https://arxiv.org/abs/2407.15886},
182
+ }
183
+ ```
snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/SCHP/exp-schp-201908261155-lip.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24fa3254ceeb74c8435458994a64b522fb439a3635b7b86ff470457e0413da00
3
+ size 267449349
snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/SCHP/exp-schp-201908301523-atr.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9d7c91ce3b4e7133df56b599fc817b533e3439c5e8d282a59126d2fda339a2a
3
+ size 267445237
snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/config.json ADDED
File without changes
snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/dresscode-16k-512/attention/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d65a6c64a9bc48e8009a0006d5272b9332c1077c9c2a74302cbac9f256e84cbf
3
+ size 198303368
snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/flux-lora/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9cfb1a2a2a88e332c454115b80a3989b4d053443e4b5483083a6a92321b31db
3
+ size 37406344
snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/mix-48k-1024/attention/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1fc093f1b6744623079e6f4e7313411f524e388c4b7467df1e0e7f577cba23a
3
+ size 198303368
snapshots/2969fcf85fe62f2036605716f0b56f0b81d01d79/vitonhd-16k-512/attention/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:915df7bf19a33bee36a28d5f9ceaef1e2267c47526f98ca9e4c49e90ae5f0fd0
3
+ size 198303368