Add Diffusers weights
#1
by pbuyle - opened
- .gitattributes +0 -1
- README.md +5 -17
- banchan-protogen-v22.ckpt +2 -2
- banchan-protogen-v22.safetensors +0 -3
- model_index.json +1 -1
- safety_checker/config.json +4 -2
- sample_images/00026-3428198346.png +0 -3
- sample_images/00027-3207496684.png +0 -3
- sample_images/00031-290723165.png +0 -3
- sample_images/00147-393566544.png +0 -3
- sample_images/00182-924148568.png +0 -3
- scheduler/scheduler_config.json +1 -1
- text_encoder/config.json +1 -1
- text_encoder/pytorch_model.bin +1 -1
- tokenizer/tokenizer_config.json +1 -0
- unet/config.json +1 -7
- unet/diffusion_pytorch_model.bin +1 -1
- vae/config.json +1 -2
.gitattributes
CHANGED
|
@@ -32,4 +32,3 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 35 |
-
*.png filter=lfs diff=lfs merge=lfs -text
|
|
|
|
| 32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
README.md
CHANGED
|
@@ -7,35 +7,23 @@ language:
|
|
| 7 |
- en
|
| 8 |
library_name: diffusers
|
| 9 |
---
|
| 10 |
-
# Banano Chan - Protogen v2.2 (banchan-protogen-v22)
|
| 11 |
|
| 12 |
|
| 13 |
A potassium rich latent diffusion model. [Protogen v2.2 (Anime)](https://huggingface.co/darkstorm2150/Protogen_v2.2_Official_Release) trained to the likeness of [Banano Chan](https://twitter.com/Banano_Chan/). The digital waifu embodiment of [Banano](https://www.banano.cc), a feeless and super fast meme cryptocurrency.
|
| 14 |
|
| 15 |
This model is intended to produce high-quality, highly detailed images from rich and complex prompts.
|
| 16 |
|
| 17 |
-
``
|
| 18 |
-
Prompt: banchan, 1girl
|
| 19 |
-
Negative prompt: ((disfigured)), ((bad art)), ((deformed)),((extra limbs)), ((bad anatomy)), (((bad proportions)))
|
| 20 |
-
Steps: 20, Sampler: Euler a, CFG scale: 7, Seed: 3207496684, Size:
|
| 21 |
-
768x768, Model hash: 220c1c8ec5, Model: banchanProtogenV2, Clip skip: 2
|
| 22 |
-
```
|
| 23 |
-
|
| 24 |
-
Share your pictures in the [#banano-ai-art Discord channel](https://discord.com/channels/415935345075421194/991823100054355998) or [Community](https://huggingface.co/Banano/banchan-protogen-v22/discussions) tab.
|
| 25 |
|
| 26 |
Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb)
|
| 27 |
-
|
| 28 |
-
|
| 29 |
-
Sample pictures:
|
| 30 |
-

|
| 31 |
-

|
| 32 |
-

|
| 33 |
-

|
| 34 |
-

|
| 35 |
--
|
| 36 |
|
| 37 |
Dreambooth model trained with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook.
|
| 38 |
|
|
|
|
|
|
|
| 39 |
## License
|
| 40 |
|
| 41 |
This model is open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage.
|
|
|
|
| 7 |
- en
|
| 8 |
library_name: diffusers
|
| 9 |
---
|
| 10 |
+
# Banano Chan - Protogen v2.2 (banchan-protogen-v22)
|
| 11 |
|
| 12 |
|
| 13 |
A potassium rich latent diffusion model. [Protogen v2.2 (Anime)](https://huggingface.co/darkstorm2150/Protogen_v2.2_Official_Release) trained to the likeness of [Banano Chan](https://twitter.com/Banano_Chan/). The digital waifu embodiment of [Banano](https://www.banano.cc), a feeless and super fast meme cryptocurrency.
|
| 14 |
|
| 15 |
This model is intended to produce high-quality, highly detailed images from rich and complex prompts.
|
| 16 |
|
| 17 |
+
eg. `(extremely detailed CG unity 8k wallpaper), green eyes, (green hair_0.8), (yellow hair_0.7), (gradient hair_0.4), full shot body photo of banchan, **<insert an action or a place to here, like bowl of sushi, or riding a jetski)>**. nostalgia professional majestic oil painting by Ed Blinkey, Atey Ghailan, Studio Ghibli, by Jeremy Mann, Greg Manchess, Antonio Moro, trending on ArtStation, trending on CGSociety, Intricate, High Detail, Sharp focus, dramatic, photorealistic painting art by midjourney and greg rutkowski, (((masterpiece))), (((best quality)))`
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 18 |
|
| 19 |
Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb)
|
| 20 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 21 |
--
|
| 22 |
|
| 23 |
Dreambooth model trained with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook.
|
| 24 |
|
| 25 |
+
|
| 26 |
+
|
| 27 |
## License
|
| 28 |
|
| 29 |
This model is open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage.
|
banchan-protogen-v22.ckpt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:574bf72be5f84cff7706050a82ebf7cc9e0edb7f28c2a1b229d73a2be2fc35b5
|
| 3 |
+
size 2132866326
|
banchan-protogen-v22.safetensors
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:2677af798c26c4094ca4b6ab6ca5243c933650cdacaf7b3cbfd44a27e04b775c
|
| 3 |
-
size 2132625431
|
|
|
|
|
|
|
|
|
|
|
|
model_index.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
{
|
| 2 |
"_class_name": "StableDiffusionPipeline",
|
| 3 |
-
"_diffusers_version": "0.
|
| 4 |
"feature_extractor": [
|
| 5 |
"transformers",
|
| 6 |
"CLIPFeatureExtractor"
|
|
|
|
| 1 |
{
|
| 2 |
"_class_name": "StableDiffusionPipeline",
|
| 3 |
+
"_diffusers_version": "0.12.0.dev0",
|
| 4 |
"feature_extractor": [
|
| 5 |
"transformers",
|
| 6 |
"CLIPFeatureExtractor"
|
safety_checker/config.json
CHANGED
|
@@ -21,6 +21,7 @@
|
|
| 21 |
"decoder_start_token_id": null,
|
| 22 |
"diversity_penalty": 0.0,
|
| 23 |
"do_sample": false,
|
|
|
|
| 24 |
"early_stopping": false,
|
| 25 |
"encoder_no_repeat_ngram_size": 0,
|
| 26 |
"eos_token_id": 2,
|
|
@@ -79,7 +80,7 @@
|
|
| 79 |
"top_p": 1.0,
|
| 80 |
"torch_dtype": null,
|
| 81 |
"torchscript": false,
|
| 82 |
-
"transformers_version": "4.
|
| 83 |
"typical_p": 1.0,
|
| 84 |
"use_bfloat16": false,
|
| 85 |
"vocab_size": 49408
|
|
@@ -105,6 +106,7 @@
|
|
| 105 |
"decoder_start_token_id": null,
|
| 106 |
"diversity_penalty": 0.0,
|
| 107 |
"do_sample": false,
|
|
|
|
| 108 |
"early_stopping": false,
|
| 109 |
"encoder_no_repeat_ngram_size": 0,
|
| 110 |
"eos_token_id": null,
|
|
@@ -165,7 +167,7 @@
|
|
| 165 |
"top_p": 1.0,
|
| 166 |
"torch_dtype": null,
|
| 167 |
"torchscript": false,
|
| 168 |
-
"transformers_version": "4.
|
| 169 |
"typical_p": 1.0,
|
| 170 |
"use_bfloat16": false
|
| 171 |
},
|
|
|
|
| 21 |
"decoder_start_token_id": null,
|
| 22 |
"diversity_penalty": 0.0,
|
| 23 |
"do_sample": false,
|
| 24 |
+
"dropout": 0.0,
|
| 25 |
"early_stopping": false,
|
| 26 |
"encoder_no_repeat_ngram_size": 0,
|
| 27 |
"eos_token_id": 2,
|
|
|
|
| 80 |
"top_p": 1.0,
|
| 81 |
"torch_dtype": null,
|
| 82 |
"torchscript": false,
|
| 83 |
+
"transformers_version": "4.26.0.dev0",
|
| 84 |
"typical_p": 1.0,
|
| 85 |
"use_bfloat16": false,
|
| 86 |
"vocab_size": 49408
|
|
|
|
| 106 |
"decoder_start_token_id": null,
|
| 107 |
"diversity_penalty": 0.0,
|
| 108 |
"do_sample": false,
|
| 109 |
+
"dropout": 0.0,
|
| 110 |
"early_stopping": false,
|
| 111 |
"encoder_no_repeat_ngram_size": 0,
|
| 112 |
"eos_token_id": null,
|
|
|
|
| 167 |
"top_p": 1.0,
|
| 168 |
"torch_dtype": null,
|
| 169 |
"torchscript": false,
|
| 170 |
+
"transformers_version": "4.26.0.dev0",
|
| 171 |
"typical_p": 1.0,
|
| 172 |
"use_bfloat16": false
|
| 173 |
},
|
sample_images/00026-3428198346.png
DELETED
Git LFS Details
|
sample_images/00027-3207496684.png
DELETED
Git LFS Details
|
sample_images/00031-290723165.png
DELETED
Git LFS Details
|
sample_images/00147-393566544.png
DELETED
Git LFS Details
|
sample_images/00182-924148568.png
DELETED
Git LFS Details
|
scheduler/scheduler_config.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
{
|
| 2 |
"_class_name": "PNDMScheduler",
|
| 3 |
-
"_diffusers_version": "0.
|
| 4 |
"beta_end": 0.012,
|
| 5 |
"beta_schedule": "scaled_linear",
|
| 6 |
"beta_start": 0.00085,
|
|
|
|
| 1 |
{
|
| 2 |
"_class_name": "PNDMScheduler",
|
| 3 |
+
"_diffusers_version": "0.12.0.dev0",
|
| 4 |
"beta_end": 0.012,
|
| 5 |
"beta_schedule": "scaled_linear",
|
| 6 |
"beta_start": 0.00085,
|
text_encoder/config.json
CHANGED
|
@@ -20,6 +20,6 @@
|
|
| 20 |
"pad_token_id": 1,
|
| 21 |
"projection_dim": 768,
|
| 22 |
"torch_dtype": "float32",
|
| 23 |
-
"transformers_version": "4.
|
| 24 |
"vocab_size": 49408
|
| 25 |
}
|
|
|
|
| 20 |
"pad_token_id": 1,
|
| 21 |
"projection_dim": 768,
|
| 22 |
"torch_dtype": "float32",
|
| 23 |
+
"transformers_version": "4.26.0.dev0",
|
| 24 |
"vocab_size": 49408
|
| 25 |
}
|
text_encoder/pytorch_model.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 492307041
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6e6be7b3c28b14ca378c20a9a9e6eddca4951b1088627a030d7dfc0525a549c5
|
| 3 |
size 492307041
|
tokenizer/tokenizer_config.json
CHANGED
|
@@ -19,6 +19,7 @@
|
|
| 19 |
},
|
| 20 |
"errors": "replace",
|
| 21 |
"model_max_length": 77,
|
|
|
|
| 22 |
"pad_token": "<|endoftext|>",
|
| 23 |
"special_tokens_map_file": "./special_tokens_map.json",
|
| 24 |
"tokenizer_class": "CLIPTokenizer",
|
|
|
|
| 19 |
},
|
| 20 |
"errors": "replace",
|
| 21 |
"model_max_length": 77,
|
| 22 |
+
"name_or_path": "openai/clip-vit-large-patch14",
|
| 23 |
"pad_token": "<|endoftext|>",
|
| 24 |
"special_tokens_map_file": "./special_tokens_map.json",
|
| 25 |
"tokenizer_class": "CLIPTokenizer",
|
unet/config.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
{
|
| 2 |
"_class_name": "UNet2DConditionModel",
|
| 3 |
-
"_diffusers_version": "0.
|
| 4 |
"act_fn": "silu",
|
| 5 |
"attention_head_dim": 8,
|
| 6 |
"block_out_channels": [
|
|
@@ -11,8 +11,6 @@
|
|
| 11 |
],
|
| 12 |
"center_input_sample": false,
|
| 13 |
"class_embed_type": null,
|
| 14 |
-
"conv_in_kernel": 3,
|
| 15 |
-
"conv_out_kernel": 3,
|
| 16 |
"cross_attention_dim": 768,
|
| 17 |
"down_block_types": [
|
| 18 |
"CrossAttnDownBlock2D",
|
|
@@ -33,12 +31,8 @@
|
|
| 33 |
"num_class_embeds": null,
|
| 34 |
"only_cross_attention": false,
|
| 35 |
"out_channels": 4,
|
| 36 |
-
"projection_class_embeddings_input_dim": null,
|
| 37 |
"resnet_time_scale_shift": "default",
|
| 38 |
"sample_size": 64,
|
| 39 |
-
"time_cond_proj_dim": null,
|
| 40 |
-
"time_embedding_type": "positional",
|
| 41 |
-
"timestep_post_act": null,
|
| 42 |
"up_block_types": [
|
| 43 |
"UpBlock2D",
|
| 44 |
"CrossAttnUpBlock2D",
|
|
|
|
| 1 |
{
|
| 2 |
"_class_name": "UNet2DConditionModel",
|
| 3 |
+
"_diffusers_version": "0.12.0.dev0",
|
| 4 |
"act_fn": "silu",
|
| 5 |
"attention_head_dim": 8,
|
| 6 |
"block_out_channels": [
|
|
|
|
| 11 |
],
|
| 12 |
"center_input_sample": false,
|
| 13 |
"class_embed_type": null,
|
|
|
|
|
|
|
| 14 |
"cross_attention_dim": 768,
|
| 15 |
"down_block_types": [
|
| 16 |
"CrossAttnDownBlock2D",
|
|
|
|
| 31 |
"num_class_embeds": null,
|
| 32 |
"only_cross_attention": false,
|
| 33 |
"out_channels": 4,
|
|
|
|
| 34 |
"resnet_time_scale_shift": "default",
|
| 35 |
"sample_size": 64,
|
|
|
|
|
|
|
|
|
|
| 36 |
"up_block_types": [
|
| 37 |
"UpBlock2D",
|
| 38 |
"CrossAttnUpBlock2D",
|
unet/diffusion_pytorch_model.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 3438366373
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8e68fa77b32b6b264294fd603c53c2093c73093cdd3113942605c3cdb70a57e3
|
| 3 |
size 3438366373
|
vae/config.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
{
|
| 2 |
"_class_name": "AutoencoderKL",
|
| 3 |
-
"_diffusers_version": "0.
|
| 4 |
"act_fn": "silu",
|
| 5 |
"block_out_channels": [
|
| 6 |
128,
|
|
@@ -20,7 +20,6 @@
|
|
| 20 |
"norm_num_groups": 32,
|
| 21 |
"out_channels": 3,
|
| 22 |
"sample_size": 512,
|
| 23 |
-
"scaling_factor": 0.18215,
|
| 24 |
"up_block_types": [
|
| 25 |
"UpDecoderBlock2D",
|
| 26 |
"UpDecoderBlock2D",
|
|
|
|
| 1 |
{
|
| 2 |
"_class_name": "AutoencoderKL",
|
| 3 |
+
"_diffusers_version": "0.12.0.dev0",
|
| 4 |
"act_fn": "silu",
|
| 5 |
"block_out_channels": [
|
| 6 |
128,
|
|
|
|
| 20 |
"norm_num_groups": 32,
|
| 21 |
"out_channels": 3,
|
| 22 |
"sample_size": 512,
|
|
|
|
| 23 |
"up_block_types": [
|
| 24 |
"UpDecoderBlock2D",
|
| 25 |
"UpDecoderBlock2D",
|