Update README.md
Browse files
README.md
CHANGED
|
@@ -13,70 +13,43 @@ pipeline_tag: text-to-image
|
|
| 13 |
<table>
|
| 14 |
<tr>
|
| 15 |
<td><img src="assets/teaser_info.png" alt="teaser example 0" width="1200"/></td>
|
|
|
|
|
|
|
| 16 |
<td><img src="assets/teaser_slide.png" alt="teaser example 1" width="1200"/></td>
|
| 17 |
</tr>
|
| 18 |
</table>
|
| 19 |
|
| 20 |
## Abstract
|
| 21 |
<p>
|
| 22 |
-
|
| 23 |
-
|
| 24 |
-
|
| 25 |
-
|
| 26 |
-
|
| 27 |
-
|
| 28 |
</p>
|
| 29 |
<p>
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
|
| 33 |
-
|
| 34 |
-
|
| 35 |
-
|
| 36 |
-
|
| 37 |
-
|
| 38 |
</p>
|
| 39 |
<p>
|
| 40 |
-
|
| 41 |
-
|
| 42 |
-
|
| 43 |
-
|
| 44 |
</p>
|
| 45 |
|
| 46 |
## Model Description
|
| 47 |
|
| 48 |
-
|
| 49 |
-
|
| 50 |
-
|
| 51 |
-
|
| 52 |
-
|
| 53 |
-
## A quick example
|
| 54 |
-
```python
|
| 55 |
-
from diffusers import StableDiffusionXLPipeline, UNet2DConditionModel
|
| 56 |
-
import torch
|
| 57 |
-
|
| 58 |
-
# load pipeline
|
| 59 |
-
inference_dtype = torch.float16
|
| 60 |
-
pipe = StableDiffusionXLPipeline.from_pretrained(
|
| 61 |
-
"SPO-Diffusion-Models/SPO-SDXL_4k-p_10ep",
|
| 62 |
-
torch_dtype=inference_dtype,
|
| 63 |
-
)
|
| 64 |
-
vae = AutoencoderKL.from_pretrained(
|
| 65 |
-
'madebyollin/sdxl-vae-fp16-fix',
|
| 66 |
-
torch_dtype=inference_dtype,
|
| 67 |
-
)
|
| 68 |
-
pipe.vae = vae
|
| 69 |
-
pipe.to('cuda')
|
| 70 |
-
|
| 71 |
-
generator=torch.Generator(device='cuda').manual_seed(42)
|
| 72 |
-
image = pipe(
|
| 73 |
-
prompt='a child and a penguin sitting in front of the moon',
|
| 74 |
-
guidance_scale=5.0,
|
| 75 |
-
generator=generator,
|
| 76 |
-
output_type='pil',
|
| 77 |
-
).images[0]
|
| 78 |
-
image.save('moon.png')
|
| 79 |
-
```
|
| 80 |
|
| 81 |
## Citation
|
| 82 |
If you find our work or codebase useful, please consider giving us a star and citing our work.
|
|
|
|
| 13 |
<table>
|
| 14 |
<tr>
|
| 15 |
<td><img src="assets/teaser_info.png" alt="teaser example 0" width="1200"/></td>
|
| 16 |
+
</tr>
|
| 17 |
+
<tr>
|
| 18 |
<td><img src="assets/teaser_slide.png" alt="teaser example 1" width="1200"/></td>
|
| 19 |
</tr>
|
| 20 |
</table>
|
| 21 |
|
| 22 |
## Abstract
|
| 23 |
<p>
|
| 24 |
+
Recently, state-of-the-art text-to-image generation models, such as Flux and Ideogram 2.0, have made
|
| 25 |
+
significant progress in sentence-level visual text rendering. In this paper, we focus on the more
|
| 26 |
+
challenging scenarios of article-level visual text rendering and address a novel task of generating
|
| 27 |
+
high-quality business content, including infographics and slides, based on user provided article-level
|
| 28 |
+
descriptive prompts and ultra-dense layouts. The fundamental challenges are twofold: significantly
|
| 29 |
+
longer context lengths and the scarcity of high-quality business content data.
|
| 30 |
</p>
|
| 31 |
<p>
|
| 32 |
+
In contrast to most previous works that focus on a limited number of sub-regions and sentence-level
|
| 33 |
+
prompts, ensuring precise adherence to ultra-dense layouts with tens or even hundreds of sub-regions in
|
| 34 |
+
business content is far more challenging. We make two key technical contributions: (i) the construction
|
| 35 |
+
of scalable, high-quality business content dataset, i.e., Infographics-650K, equipped with
|
| 36 |
+
ultra-dense layouts and prompts by implementing a layer-wise retrieval-augmented infographic generation
|
| 37 |
+
scheme; and (ii) a layout-guided cross attention scheme, which injects tens of region-wise prompts into
|
| 38 |
+
a set of cropped region latent space according to the ultra-dense layouts, and refine each sub-regions
|
| 39 |
+
flexibly during inference using a layout conditional CFG.
|
| 40 |
</p>
|
| 41 |
<p>
|
| 42 |
+
We demonstrate the strong results of our system compared to previous SOTA systems such as Flux and SD3
|
| 43 |
+
on our BizEval prompt set. Additionally, we conduct thorough ablation experiments to verify the
|
| 44 |
+
effectiveness of each component. We hope our constructed Infographics-650K and BizEval can encourage
|
| 45 |
+
the broader community to advance the progress of business content generation.
|
| 46 |
</p>
|
| 47 |
|
| 48 |
## Model Description
|
| 49 |
|
| 50 |
+
The ByT5 model is finetuned from [Glyph-ByT5-v2](https://arxiv.org/abs/2406.10208), which supports accerate visual text rendering in ten different languages.
|
| 51 |
+
The [SPO](https://huggingface.co/SPO-Diffusion-Models) model is a substitute for the original sdxl-base-1.0 for aesthetic improvement.
|
| 52 |
+
You can follow our [github]() to organize and run the model.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 53 |
|
| 54 |
## Citation
|
| 55 |
If you find our work or codebase useful, please consider giving us a star and citing our work.
|