lzyhha commited on
Commit
f0ab9f6
·
verified ·
1 Parent(s): da9fc6a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -11
README.md CHANGED
@@ -19,13 +19,6 @@ tags:
19
 
20
  # VisualCloze: A Universal Image Generation Framework via Visual In-Context Learning (Implementation with <strong><span style="color:red">Diffusers</span></strong>)
21
 
22
- **Note**: <strong><span style="color:hotpink">You still need to install our modified version of</span></strong> [<strong><span style="color:hotpink">diffusers</span></strong>](https://github.com/lzyhha/diffusers).
23
-
24
- A model trained with the `resolution` of 384 is released at [Model Card](https://huggingface.co/VisualCloze/VisualClozePipeline-384),
25
- while this model uses the `resolution` of 512. The `resolution` means that each image will be resized to it before being
26
- concatenated to avoid the out-of-memory error. To generate high-resolution images, we use the [SDEdit](https://arxiv.org/abs/2108.01073) technology for upsampling the generated results.
27
-
28
-
29
  <div align="center">
30
 
31
  [[Paper](https://arxiv.org/abs/2504.07960)] &emsp; [[Project Page](https://visualcloze.github.io/)] &emsp; [[Github](https://github.com/lzyhha/VisualCloze)]
@@ -58,7 +51,8 @@ An in-context learning based universal image generation framework.
58
 
59
  ## 🔧 Installation
60
 
61
- Install diffusers from our modified repository.
 
62
  ```bash
63
  git clone https://github.com/lzyhha/diffusers
64
 
@@ -70,7 +64,11 @@ pip install -v -e .
70
 
71
  [![Huggingface VisualCloze](https://img.shields.io/static/v1?label=Demo&message=Huggingface%20Gradio&color=orange)](https://huggingface.co/spaces/VisualCloze/VisualCloze)
72
 
73
- Example with Depth-to-Image:
 
 
 
 
74
 
75
  <img src="./visualcloze_diffusers_example_depthtoimage.jpg" width="60%" height="50%" alt="Example with Depth-to-Image"/>
76
 
@@ -125,8 +123,7 @@ image_result = pipe(
125
  image_result.save("visualcloze.png")
126
  ```
127
 
128
-
129
- Example with Virtual Try-On:
130
 
131
  <img src="./visualcloze_diffusers_example_tryon.jpg" width="60%" height="50%" alt="Example with Virtual Try-On"/>
132
 
 
19
 
20
  # VisualCloze: A Universal Image Generation Framework via Visual In-Context Learning (Implementation with <strong><span style="color:red">Diffusers</span></strong>)
21
 
 
 
 
 
 
 
 
22
  <div align="center">
23
 
24
  [[Paper](https://arxiv.org/abs/2504.07960)] &emsp; [[Project Page](https://visualcloze.github.io/)] &emsp; [[Github](https://github.com/lzyhha/VisualCloze)]
 
51
 
52
  ## 🔧 Installation
53
 
54
+ <strong><span style="color:hotpink">You still need to install our modified version of</span></strong> [diffusers](https://github.com/lzyhha/diffusers).
55
+
56
  ```bash
57
  git clone https://github.com/lzyhha/diffusers
58
 
 
64
 
65
  [![Huggingface VisualCloze](https://img.shields.io/static/v1?label=Demo&message=Huggingface%20Gradio&color=orange)](https://huggingface.co/spaces/VisualCloze/VisualCloze)
66
 
67
+ A model trained with the `resolution` of 384 is released at [Model Card](https://huggingface.co/VisualCloze/VisualClozePipeline-384),
68
+ while this model uses the `resolution` of 512. The `resolution` means that each image will be resized to it before being
69
+ concatenated to avoid the out-of-memory error. To generate high-resolution images, we use the [SDEdit](https://arxiv.org/abs/2108.01073) technology for upsampling the generated results.
70
+
71
+ #### Example with Depth-to-Image:
72
 
73
  <img src="./visualcloze_diffusers_example_depthtoimage.jpg" width="60%" height="50%" alt="Example with Depth-to-Image"/>
74
 
 
123
  image_result.save("visualcloze.png")
124
  ```
125
 
126
+ #### Example with Virtual Try-On:
 
127
 
128
  <img src="./visualcloze_diffusers_example_tryon.jpg" width="60%" height="50%" alt="Example with Virtual Try-On"/>
129