lzyhha commited on
Commit
2021ecd
Β·
verified Β·
1 Parent(s): 54ef1f3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -2
README.md CHANGED
@@ -27,7 +27,8 @@ tags:
27
 
28
  <div align="center">
29
 
30
- [[πŸ€— <strong><span style="color:hotpink">Diffusers</span></strong> Implementation](https://github.com/huggingface/diffusers/tree/main/src/diffusers/pipelines/visualcloze)]
 
31
 
32
  </div>
33
 
@@ -43,6 +44,7 @@ If you find VisualCloze is helpful, please consider to star ⭐ the [<strong><sp
43
 
44
  ## πŸ“° News
45
  - [2025-5-15] πŸ€—πŸ€—πŸ€— VisualCloze has been merged into the [<strong><span style="color:hotpink">official pipelines of diffusers</span></strong>](https://github.com/huggingface/diffusers/tree/main/src/diffusers/pipelines/visualcloze).
 
46
 
47
  ## 🌠 Key Features
48
 
@@ -67,7 +69,11 @@ pip install git+https://github.com/huggingface/diffusers.git
67
 
68
  [![Huggingface VisualCloze](https://img.shields.io/static/v1?label=Demo&message=Huggingface%20Gradio&color=orange)](https://huggingface.co/spaces/VisualCloze/VisualCloze)
69
 
70
- A model trained with the `resolution` of 384 is released at [Model Card](https://huggingface.co/VisualCloze/VisualClozePipeline-384),
 
 
 
 
71
  while this model uses the `resolution` of 512. The `resolution` means that each image will be resized to it before being
72
  concatenated to avoid the out-of-memory error. To generate high-resolution images, we use the [SDEdit](https://arxiv.org/abs/2108.01073) technology for upsampling the generated results.
73
 
@@ -108,6 +114,11 @@ high contrast, photorealistic, intimate, elegant, visually balanced, serene atmo
108
  pipe = VisualClozePipeline.from_pretrained("VisualCloze/VisualClozePipeline-512", resolution=512, torch_dtype=torch.bfloat16)
109
  pipe.to("cuda")
110
 
 
 
 
 
 
111
  # Run the pipeline
112
  image_result = pipe(
113
  task_prompt=task_prompt,
@@ -161,6 +172,11 @@ content_prompt = None
161
  pipe = VisualClozePipeline.from_pretrained("VisualCloze/VisualClozePipeline-512", resolution=512, torch_dtype=torch.bfloat16)
162
  pipe.to("cuda")
163
 
 
 
 
 
 
164
  # Run the pipeline
165
  image_result = pipe(
166
  task_prompt=task_prompt,
 
27
 
28
  <div align="center">
29
 
30
+ [[πŸ€— <strong><span style="color:hotpink">Diffusers</span></strong> Implementation](https://github.com/huggingface/diffusers/tree/main/src/diffusers/pipelines/visualcloze)] &emsp; [[πŸ€— LoRA Model Card for Diffusers]](https://huggingface.co/VisualCloze/VisualClozePipeline-LoRA-512)
31
+
32
 
33
  </div>
34
 
 
44
 
45
  ## πŸ“° News
46
  - [2025-5-15] πŸ€—πŸ€—πŸ€— VisualCloze has been merged into the [<strong><span style="color:hotpink">official pipelines of diffusers</span></strong>](https://github.com/huggingface/diffusers/tree/main/src/diffusers/pipelines/visualcloze).
47
+ - [2025-5-18] πŸ₯³πŸ₯³πŸ₯³ We have released the LoRA weights supporting diffusers at [LoRA Model Card 384](https://huggingface.co/VisualCloze/VisualClozePipeline-LoRA-384) and [LoRA Model Card 512](https://huggingface.co/VisualCloze/VisualClozePipeline-LoRA-512).
48
 
49
  ## 🌠 Key Features
50
 
 
69
 
70
  [![Huggingface VisualCloze](https://img.shields.io/static/v1?label=Demo&message=Huggingface%20Gradio&color=orange)](https://huggingface.co/spaces/VisualCloze/VisualCloze)
71
 
72
+ This model provides the full parameters of our VisualCloze.
73
+ If you find the download size too large, you can use the [LoRA version](https://huggingface.co/VisualCloze/VisualClozePipeline-LoRA-512)
74
+ with the FLUX.1-Fill-dev as the base model.
75
+
76
+ A model trained with the `resolution` of 384 is released at [Full Model Card 384](https://huggingface.co/VisualCloze/VisualClozePipeline-384) and [LoRA Model Card 384](https://huggingface.co/VisualCloze/VisualClozePipeline-LoRA-384),
77
  while this model uses the `resolution` of 512. The `resolution` means that each image will be resized to it before being
78
  concatenated to avoid the out-of-memory error. To generate high-resolution images, we use the [SDEdit](https://arxiv.org/abs/2108.01073) technology for upsampling the generated results.
79
 
 
114
  pipe = VisualClozePipeline.from_pretrained("VisualCloze/VisualClozePipeline-512", resolution=512, torch_dtype=torch.bfloat16)
115
  pipe.to("cuda")
116
 
117
+ # Loading the VisualClozePipeline via LoRA
118
+ # pipe = VisualClozePipeline.from_pretrained("black-forest-labs/FLUX.1-Fill-dev", resolution=512, torch_dtype=torch.bfloat16)
119
+ # pipe.load_lora_weights('VisualCloze/VisualClozePipeline-LoRA-512', weight_name='visualcloze-lora-512.safetensors')
120
+ # pipe.to("cuda")
121
+
122
  # Run the pipeline
123
  image_result = pipe(
124
  task_prompt=task_prompt,
 
172
  pipe = VisualClozePipeline.from_pretrained("VisualCloze/VisualClozePipeline-512", resolution=512, torch_dtype=torch.bfloat16)
173
  pipe.to("cuda")
174
 
175
+ # Loading the VisualClozePipeline via LoRA
176
+ # pipe = VisualClozePipeline.from_pretrained("black-forest-labs/FLUX.1-Fill-dev", resolution=512, torch_dtype=torch.bfloat16)
177
+ # pipe.load_lora_weights('VisualCloze/VisualClozePipeline-LoRA-512', weight_name='visualcloze-lora-512.safetensors')
178
+ # pipe.to("cuda")
179
+
180
  # Run the pipeline
181
  image_result = pipe(
182
  task_prompt=task_prompt,