Upload README.md
Browse filesReworded the readme for clarity.
README.md
CHANGED
|
@@ -9,7 +9,7 @@ license: apache-2.0
|
|
| 9 |
---
|
| 10 |
|
| 11 |
# Overview
|
| 12 |
-
The Boring Embeddings are negative
|
| 13 |
widely used in the community (10M+ generations across public tools) to suppress low-quality,
|
| 14 |
low-engagement visual patterns and make outputs look more visually appealing.
|
| 15 |
<br>
|
|
@@ -37,10 +37,10 @@ the embedding's name in your NEGATIVE prompt for more interesting outputs.
|
|
| 37 |
|
| 38 |
## Research Summary / Technical Contribution
|
| 39 |
|
| 40 |
-
This project
|
| 41 |
-
|
| 42 |
-
|
| 43 |
-
domain-specific negative conditioning.
|
| 44 |
<br>
|
| 45 |
|
| 46 |
|
|
@@ -54,14 +54,14 @@ depend on manually curated lists of tags describing features people do not want
|
|
| 54 |
* It can only capture named concepts. If there exist unnamed yet visually unappealing concepts that just make an image look wrong, but for reasons that cannot be succinctly explained, they will not be captured by a list of tags.
|
| 55 |
|
| 56 |
To address these problems, we employ textual inversion on a set of images extracted from several large community-tagged art
|
| 57 |
-
datasets with rich metadata. Each of these sites
|
| 58 |
-
|
| 59 |
The Boring embeddings were specifically trained on artworks automatically selected from these sites according to the criteria
|
| 60 |
that no user has ever favorited them, and they have 0 or only a very small number of up or down votes. The Boring embeddings
|
| 61 |
thus learned to produce uninteresting low-quality images, so when they are used in the negative prompt of a stable diffusion image generator,
|
| 62 |
the model avoids making mistakes that would make the generation more boring.
|
| 63 |
Each training sample consisted of an image paired with a comma-separated list of all the tags associated with it on the site it was sourced from, with the embedding’s name prepended to the list.
|
| 64 |
-
This ensures that the model learns to associate the embedding with the characteristics of uninteresting images in a way that
|
| 65 |
<br>
|
| 66 |
|
| 67 |
|
|
@@ -175,6 +175,6 @@ Across nearly all prompt/seed pairs, the versions generated with the embedding e
|
|
| 175 |
|
| 176 |
An independent viewer (not involved in the generation process) remarked that the embedding versions were “clearly better” in the majority of comparisons.
|
| 177 |
|
| 178 |
-
This
|
| 179 |
|
| 180 |
|
|
|
|
| 9 |
---
|
| 10 |
|
| 11 |
# Overview
|
| 12 |
+
The Boring Embeddings are negative textual inversion embeddings for Stable Diffusion models. They are
|
| 13 |
widely used in the community (10M+ generations across public tools) to suppress low-quality,
|
| 14 |
low-engagement visual patterns and make outputs look more visually appealing.
|
| 15 |
<br>
|
|
|
|
| 37 |
|
| 38 |
## Research Summary / Technical Contribution
|
| 39 |
|
| 40 |
+
This project trains negative textual inversion embeddings using automated quality-based sampling from large community-labeled
|
| 41 |
+
image datasets. The method avoids hand-curated defect lists and instead captures visual patterns
|
| 42 |
+
associated with low-engagement images. Typical uses include quality control in generative models, aesthetic
|
| 43 |
+
filtering, and domain-specific negative conditioning.
|
| 44 |
<br>
|
| 45 |
|
| 46 |
|
|
|
|
| 54 |
* It can only capture named concepts. If there exist unnamed yet visually unappealing concepts that just make an image look wrong, but for reasons that cannot be succinctly explained, they will not be captured by a list of tags.
|
| 55 |
|
| 56 |
To address these problems, we employ textual inversion on a set of images extracted from several large community-tagged art
|
| 57 |
+
datasets with rich metadata. Each of these sites contains millions of hand-labeled artworks.
|
| 58 |
+
Users can express their approval of an artwork by either up-voting it or marking it as a favorite.
|
| 59 |
The Boring embeddings were specifically trained on artworks automatically selected from these sites according to the criteria
|
| 60 |
that no user has ever favorited them, and they have 0 or only a very small number of up or down votes. The Boring embeddings
|
| 61 |
thus learned to produce uninteresting low-quality images, so when they are used in the negative prompt of a stable diffusion image generator,
|
| 62 |
the model avoids making mistakes that would make the generation more boring.
|
| 63 |
Each training sample consisted of an image paired with a comma-separated list of all the tags associated with it on the site it was sourced from, with the embedding’s name prepended to the list.
|
| 64 |
+
This ensures that the model learns to associate the embedding with the characteristics of uninteresting images in a way that is consistent with the dataset’s tagging system.
|
| 65 |
<br>
|
| 66 |
|
| 67 |
|
|
|
|
| 175 |
|
| 176 |
An independent viewer (not involved in the generation process) remarked that the embedding versions were “clearly better” in the majority of comparisons.
|
| 177 |
|
| 178 |
+
This dataset makes the embedding’s effect easy to inspect across many prompts and seeds.
|
| 179 |
|
| 180 |
|