Add Diffusers weights

#1
by pbuyle - opened
.gitattributes CHANGED
@@ -32,4 +32,3 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
- *.png filter=lfs diff=lfs merge=lfs -text
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
README.md CHANGED
@@ -7,35 +7,23 @@ language:
7
  - en
8
  library_name: diffusers
9
  ---
10
- # Banano Chan - Protogen v2.2 (banchan-protogen-v22) V2
11
 
12
 
13
  A potassium rich latent diffusion model. [Protogen v2.2 (Anime)](https://huggingface.co/darkstorm2150/Protogen_v2.2_Official_Release) trained to the likeness of [Banano Chan](https://twitter.com/Banano_Chan/). The digital waifu embodiment of [Banano](https://www.banano.cc), a feeless and super fast meme cryptocurrency.
14
 
15
  This model is intended to produce high-quality, highly detailed images from rich and complex prompts.
16
 
17
- ```
18
- Prompt: banchan, 1girl
19
- Negative prompt: ((disfigured)), ((bad art)), ((deformed)),((extra limbs)), ((bad anatomy)), (((bad proportions)))
20
- Steps: 20, Sampler: Euler a, CFG scale: 7, Seed: 3207496684, Size:
21
- 768x768, Model hash: 220c1c8ec5, Model: banchanProtogenV2, Clip skip: 2
22
- ```
23
-
24
- Share your pictures in the [#banano-ai-art Discord channel](https://discord.com/channels/415935345075421194/991823100054355998) or [Community](https://huggingface.co/Banano/banchan-protogen-v22/discussions) tab.
25
 
26
  Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb)
27
- Or you can run your new concept via `diffusers` [Colab Notebook for Inference](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/sd_dreambooth_inference.ipynb)
28
-
29
- Sample pictures:
30
- ![0](./sample_images/00147-393566544.png)
31
- ![1](./sample_images/00027-3207496684.png)
32
- ![2](./sample_images/00026-3428198346.png)
33
- ![3](./sample_images/00031-290723165.png)
34
- ![4](./sample_images/00182-924148568.png)
35
  --
36
 
37
  Dreambooth model trained with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook.
38
 
 
 
39
  ## License
40
 
41
  This model is open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage.
 
7
  - en
8
  library_name: diffusers
9
  ---
10
+ # Banano Chan - Protogen v2.2 (banchan-protogen-v22)
11
 
12
 
13
  A potassium rich latent diffusion model. [Protogen v2.2 (Anime)](https://huggingface.co/darkstorm2150/Protogen_v2.2_Official_Release) trained to the likeness of [Banano Chan](https://twitter.com/Banano_Chan/). The digital waifu embodiment of [Banano](https://www.banano.cc), a feeless and super fast meme cryptocurrency.
14
 
15
  This model is intended to produce high-quality, highly detailed images from rich and complex prompts.
16
 
17
+ eg. `(extremely detailed CG unity 8k wallpaper), green eyes, (green hair_0.8), (yellow hair_0.7), (gradient hair_0.4), full shot body photo of banchan, **<insert an action or a place to here, like bowl of sushi, or riding a jetski)>**. nostalgia professional majestic oil painting by Ed Blinkey, Atey Ghailan, Studio Ghibli, by Jeremy Mann, Greg Manchess, Antonio Moro, trending on ArtStation, trending on CGSociety, Intricate, High Detail, Sharp focus, dramatic, photorealistic painting art by midjourney and greg rutkowski, (((masterpiece))), (((best quality)))`
 
 
 
 
 
 
 
18
 
19
  Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb)
20
+
 
 
 
 
 
 
 
21
  --
22
 
23
  Dreambooth model trained with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook.
24
 
25
+
26
+
27
  ## License
28
 
29
  This model is open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage.
banchan-protogen-v22.ckpt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:220c1c8ec5d91480f730da7abde187c5b05b4a0250b0ba675e11aafaed4c0790
3
- size 2132868592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:574bf72be5f84cff7706050a82ebf7cc9e0edb7f28c2a1b229d73a2be2fc35b5
3
+ size 2132866326
banchan-protogen-v22.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2677af798c26c4094ca4b6ab6ca5243c933650cdacaf7b3cbfd44a27e04b775c
3
- size 2132625431
 
 
 
 
model_index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
- "_diffusers_version": "0.14.0.dev0",
4
  "feature_extractor": [
5
  "transformers",
6
  "CLIPFeatureExtractor"
 
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.12.0.dev0",
4
  "feature_extractor": [
5
  "transformers",
6
  "CLIPFeatureExtractor"
safety_checker/config.json CHANGED
@@ -21,6 +21,7 @@
21
  "decoder_start_token_id": null,
22
  "diversity_penalty": 0.0,
23
  "do_sample": false,
 
24
  "early_stopping": false,
25
  "encoder_no_repeat_ngram_size": 0,
26
  "eos_token_id": 2,
@@ -79,7 +80,7 @@
79
  "top_p": 1.0,
80
  "torch_dtype": null,
81
  "torchscript": false,
82
- "transformers_version": "4.27.0.dev0",
83
  "typical_p": 1.0,
84
  "use_bfloat16": false,
85
  "vocab_size": 49408
@@ -105,6 +106,7 @@
105
  "decoder_start_token_id": null,
106
  "diversity_penalty": 0.0,
107
  "do_sample": false,
 
108
  "early_stopping": false,
109
  "encoder_no_repeat_ngram_size": 0,
110
  "eos_token_id": null,
@@ -165,7 +167,7 @@
165
  "top_p": 1.0,
166
  "torch_dtype": null,
167
  "torchscript": false,
168
- "transformers_version": "4.27.0.dev0",
169
  "typical_p": 1.0,
170
  "use_bfloat16": false
171
  },
 
21
  "decoder_start_token_id": null,
22
  "diversity_penalty": 0.0,
23
  "do_sample": false,
24
+ "dropout": 0.0,
25
  "early_stopping": false,
26
  "encoder_no_repeat_ngram_size": 0,
27
  "eos_token_id": 2,
 
80
  "top_p": 1.0,
81
  "torch_dtype": null,
82
  "torchscript": false,
83
+ "transformers_version": "4.26.0.dev0",
84
  "typical_p": 1.0,
85
  "use_bfloat16": false,
86
  "vocab_size": 49408
 
106
  "decoder_start_token_id": null,
107
  "diversity_penalty": 0.0,
108
  "do_sample": false,
109
+ "dropout": 0.0,
110
  "early_stopping": false,
111
  "encoder_no_repeat_ngram_size": 0,
112
  "eos_token_id": null,
 
167
  "top_p": 1.0,
168
  "torch_dtype": null,
169
  "torchscript": false,
170
+ "transformers_version": "4.26.0.dev0",
171
  "typical_p": 1.0,
172
  "use_bfloat16": false
173
  },
sample_images/00026-3428198346.png DELETED

Git LFS Details

  • SHA256: abcb304c009c18faa1e90df4f4f8326f5fcf393061f498485ed196828c4982fe
  • Pointer size: 132 Bytes
  • Size of remote file: 1.75 MB
sample_images/00027-3207496684.png DELETED

Git LFS Details

  • SHA256: 65c30e2126efc08ad721257c373557291a4e593b65bfeb69dbf38c81942707a8
  • Pointer size: 131 Bytes
  • Size of remote file: 816 kB
sample_images/00031-290723165.png DELETED

Git LFS Details

  • SHA256: ee879623d15cf16b77b6c4b8d4d67446c096365bb49ab45a9a4392ec3c58ed90
  • Pointer size: 132 Bytes
  • Size of remote file: 2.01 MB
sample_images/00147-393566544.png DELETED

Git LFS Details

  • SHA256: 79d5f3f0715bb978f9af9d98027955e5190a4c57fa618d90ebcb88cec2d7e968
  • Pointer size: 132 Bytes
  • Size of remote file: 1.48 MB
sample_images/00182-924148568.png DELETED

Git LFS Details

  • SHA256: 446fa4e4280a5ca70ae18be6680c795f0512392639c6974214bb7bd96fe7ea6e
  • Pointer size: 132 Bytes
  • Size of remote file: 1.89 MB
scheduler/scheduler_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "PNDMScheduler",
3
- "_diffusers_version": "0.14.0.dev0",
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
 
1
  {
2
  "_class_name": "PNDMScheduler",
3
+ "_diffusers_version": "0.12.0.dev0",
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
text_encoder/config.json CHANGED
@@ -20,6 +20,6 @@
20
  "pad_token_id": 1,
21
  "projection_dim": 768,
22
  "torch_dtype": "float32",
23
- "transformers_version": "4.27.0.dev0",
24
  "vocab_size": 49408
25
  }
 
20
  "pad_token_id": 1,
21
  "projection_dim": 768,
22
  "torch_dtype": "float32",
23
+ "transformers_version": "4.26.0.dev0",
24
  "vocab_size": 49408
25
  }
text_encoder/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c3633a9cf751487bb453c01d75695202a08957044acf274028dc0e914cc7b1f1
3
  size 492307041
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e6be7b3c28b14ca378c20a9a9e6eddca4951b1088627a030d7dfc0525a549c5
3
  size 492307041
tokenizer/tokenizer_config.json CHANGED
@@ -19,6 +19,7 @@
19
  },
20
  "errors": "replace",
21
  "model_max_length": 77,
 
22
  "pad_token": "<|endoftext|>",
23
  "special_tokens_map_file": "./special_tokens_map.json",
24
  "tokenizer_class": "CLIPTokenizer",
 
19
  },
20
  "errors": "replace",
21
  "model_max_length": 77,
22
+ "name_or_path": "openai/clip-vit-large-patch14",
23
  "pad_token": "<|endoftext|>",
24
  "special_tokens_map_file": "./special_tokens_map.json",
25
  "tokenizer_class": "CLIPTokenizer",
unet/config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
- "_diffusers_version": "0.14.0.dev0",
4
  "act_fn": "silu",
5
  "attention_head_dim": 8,
6
  "block_out_channels": [
@@ -11,8 +11,6 @@
11
  ],
12
  "center_input_sample": false,
13
  "class_embed_type": null,
14
- "conv_in_kernel": 3,
15
- "conv_out_kernel": 3,
16
  "cross_attention_dim": 768,
17
  "down_block_types": [
18
  "CrossAttnDownBlock2D",
@@ -33,12 +31,8 @@
33
  "num_class_embeds": null,
34
  "only_cross_attention": false,
35
  "out_channels": 4,
36
- "projection_class_embeddings_input_dim": null,
37
  "resnet_time_scale_shift": "default",
38
  "sample_size": 64,
39
- "time_cond_proj_dim": null,
40
- "time_embedding_type": "positional",
41
- "timestep_post_act": null,
42
  "up_block_types": [
43
  "UpBlock2D",
44
  "CrossAttnUpBlock2D",
 
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.12.0.dev0",
4
  "act_fn": "silu",
5
  "attention_head_dim": 8,
6
  "block_out_channels": [
 
11
  ],
12
  "center_input_sample": false,
13
  "class_embed_type": null,
 
 
14
  "cross_attention_dim": 768,
15
  "down_block_types": [
16
  "CrossAttnDownBlock2D",
 
31
  "num_class_embeds": null,
32
  "only_cross_attention": false,
33
  "out_channels": 4,
 
34
  "resnet_time_scale_shift": "default",
35
  "sample_size": 64,
 
 
 
36
  "up_block_types": [
37
  "UpBlock2D",
38
  "CrossAttnUpBlock2D",
unet/diffusion_pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef36ad5861873c728726be19d4e5073ba2fd20401f7bef48c1c1bd270c48ff6a
3
  size 3438366373
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e68fa77b32b6b264294fd603c53c2093c73093cdd3113942605c3cdb70a57e3
3
  size 3438366373
vae/config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "AutoencoderKL",
3
- "_diffusers_version": "0.14.0.dev0",
4
  "act_fn": "silu",
5
  "block_out_channels": [
6
  128,
@@ -20,7 +20,6 @@
20
  "norm_num_groups": 32,
21
  "out_channels": 3,
22
  "sample_size": 512,
23
- "scaling_factor": 0.18215,
24
  "up_block_types": [
25
  "UpDecoderBlock2D",
26
  "UpDecoderBlock2D",
 
1
  {
2
  "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.12.0.dev0",
4
  "act_fn": "silu",
5
  "block_out_channels": [
6
  128,
 
20
  "norm_num_groups": 32,
21
  "out_channels": 3,
22
  "sample_size": 512,
 
23
  "up_block_types": [
24
  "UpDecoderBlock2D",
25
  "UpDecoderBlock2D",