FoodDesert commited on
Commit
d964d5f
·
verified ·
1 Parent(s): 0cc5e89

Upload README.md

Browse files

Reworded the readme for clarity.

Files changed (1) hide show
  1. README.md +9 -9
README.md CHANGED
@@ -9,7 +9,7 @@ license: apache-2.0
9
  ---
10
 
11
  # Overview
12
- The Boring Embeddings are negative-control textual inversion embeddings for Stable Diffusion models. They are
13
  widely used in the community (10M+ generations across public tools) to suppress low-quality,
14
  low-engagement visual patterns and make outputs look more visually appealing.
15
  <br>
@@ -37,10 +37,10 @@ the embedding's name in your NEGATIVE prompt for more interesting outputs.
37
 
38
  ## Research Summary / Technical Contribution
39
 
40
- This project demonstrates a novel approach to constructing negative-control embeddings using automated quality-based sampling
41
- from large community-labeled image datasets. The method avoids hand-curated defect lists and instead captures latent features
42
- correlated with low engagement. Applications include quality control in generative models, aesthetic filtering, and
43
- domain-specific negative conditioning.
44
  <br>
45
 
46
 
@@ -54,14 +54,14 @@ depend on manually curated lists of tags describing features people do not want
54
  * It can only capture named concepts. If there exist unnamed yet visually unappealing concepts that just make an image look wrong, but for reasons that cannot be succinctly explained, they will not be captured by a list of tags.
55
 
56
  To address these problems, we employ textual inversion on a set of images extracted from several large community-tagged art
57
- datasets with rich metadata. Each of these sites is a rich resource of millions of
58
- hand-labeled artworks which allow users to express their approval of an artwork by either up-voting it or marking it as a favorite.
59
  The Boring embeddings were specifically trained on artworks automatically selected from these sites according to the criteria
60
  that no user has ever favorited them, and they have 0 or only a very small number of up or down votes. The Boring embeddings
61
  thus learned to produce uninteresting low-quality images, so when they are used in the negative prompt of a stable diffusion image generator,
62
  the model avoids making mistakes that would make the generation more boring.
63
  Each training sample consisted of an image paired with a comma-separated list of all the tags associated with it on the site it was sourced from, with the embedding’s name prepended to the list.
64
- This ensures that the model learns to associate the embedding with the characteristics of uninteresting images in a way that aligns with the dataset’s tagging system.
65
  <br>
66
 
67
 
@@ -175,6 +175,6 @@ Across nearly all prompt/seed pairs, the versions generated with the embedding e
175
 
176
  An independent viewer (not involved in the generation process) remarked that the embedding versions were “clearly better” in the majority of comparisons.
177
 
178
- This artifact provides a transparent, reproducible demonstration of the embedding’s effect across a wide range of seeds and prompts.
179
 
180
 
 
9
  ---
10
 
11
  # Overview
12
+ The Boring Embeddings are negative textual inversion embeddings for Stable Diffusion models. They are
13
  widely used in the community (10M+ generations across public tools) to suppress low-quality,
14
  low-engagement visual patterns and make outputs look more visually appealing.
15
  <br>
 
37
 
38
  ## Research Summary / Technical Contribution
39
 
40
+ This project trains negative textual inversion embeddings using automated quality-based sampling from large community-labeled
41
+ image datasets. The method avoids hand-curated defect lists and instead captures visual patterns
42
+ associated with low-engagement images. Typical uses include quality control in generative models, aesthetic
43
+ filtering, and domain-specific negative conditioning.
44
  <br>
45
 
46
 
 
54
  * It can only capture named concepts. If there exist unnamed yet visually unappealing concepts that just make an image look wrong, but for reasons that cannot be succinctly explained, they will not be captured by a list of tags.
55
 
56
  To address these problems, we employ textual inversion on a set of images extracted from several large community-tagged art
57
+ datasets with rich metadata. Each of these sites contains millions of hand-labeled artworks.
58
+ Users can express their approval of an artwork by either up-voting it or marking it as a favorite.
59
  The Boring embeddings were specifically trained on artworks automatically selected from these sites according to the criteria
60
  that no user has ever favorited them, and they have 0 or only a very small number of up or down votes. The Boring embeddings
61
  thus learned to produce uninteresting low-quality images, so when they are used in the negative prompt of a stable diffusion image generator,
62
  the model avoids making mistakes that would make the generation more boring.
63
  Each training sample consisted of an image paired with a comma-separated list of all the tags associated with it on the site it was sourced from, with the embedding’s name prepended to the list.
64
+ This ensures that the model learns to associate the embedding with the characteristics of uninteresting images in a way that is consistent with the dataset’s tagging system.
65
  <br>
66
 
67
 
 
175
 
176
  An independent viewer (not involved in the generation process) remarked that the embedding versions were “clearly better” in the majority of comparisons.
177
 
178
+ This dataset makes the embedding’s effect easy to inspect across many prompts and seeds.
179
 
180