ttttdiva commited on
Commit
78a2df5
·
1 Parent(s): dcca921

Upload 20 files

Browse files
Elaina-1.3.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:731449dbde2a86bb81f3228880bbcd9e6b14b74255ea506947f89712ea937916
3
+ size 2132856622
Nezuko Kamado.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3afa6161018fd3e09855f35c099c8a64b1ef98cfe82fbcef3f7b13a9c49cf075
3
+ size 3852163249
WD1-3 E8 Clownpiece fairy_girl_blonde_hair_red_eyes 4k.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7dfc6ceea78cb736589b8741961f1bc30d39dbcbed41df0c679e3c42976e1e6f
3
+ size 3852162098
dreambooth_lain_girl.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6110d27b80e83ef5623a2f91685e07574d35209071b72107bc154aeca9186b0f
3
+ size 2132721605
pony-diffusion-v3/README.md ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ tags:
5
+ - stable-diffusion
6
+ - text-to-image
7
+ license: bigscience-bloom-rail-1.0
8
+ inference: false
9
+ thumbnail: "https://imagedelivery.net/_wFNZAzgWNWPmneM1cyjcw/artifact/449d42a8-28c5-44da-afd7-28d7e29a264c/public"
10
+ ---
11
+
12
+ # pony-diffusion-v3 - "kept you waiting huh" edition
13
+
14
+ pony-diffusion is a latent text-to-image diffusion model that has been conditioned on high-quality pony, furry and other non photorealistic SFW and NSFW images through fine-tuning.
15
+
16
+ **WARNING:** This model is capable of producing NSFW content so it's recommended to use 'safe' tag in prompt in combination with negative prompt for image features you may want to suppress (i.e. nudity).
17
+
18
+ **Despite its name, this model is capable of producing wide range of furry and cartoon images as side effect of improving data diversity (with exception of anime stlyes, for which Waifu Diffusion is much stronger choice).**
19
+
20
+ Special thanks to [Waifu-Diffusion](https://huggingface.co/hakurei/waifu-diffusion) for providing finetuning expertise and advising through the process, without their help this project would not exist.
21
+
22
+ [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1rl-R39aKp42RS6sRsM9cVC2elBFugdDd?usp=sharing)
23
+
24
+ **Please join PurpleSmartAI Discord to use this model with our free SD bot and get early access to pony-diffusion-v4.**
25
+
26
+ [![Join Discord Server to try next generation models](https://discordapp.com/api/guilds/670866322619498507/widget.png?style=banner2)](https://discord.gg/pYsdjMfu3q)
27
+
28
+ <img src=https://imagedelivery.net/_wFNZAzgWNWPmneM1cyjcw/artifact/449d42a8-28c5-44da-afd7-28d7e29a264c/public width=25% height=25%>
29
+ <img src=https://imagedelivery.net/_wFNZAzgWNWPmneM1cyjcw/artifact/d72e5089-3857-417e-87a2-0fa59111022d/public width=25% height=25%>
30
+ <img src=https://imagedelivery.net/_wFNZAzgWNWPmneM1cyjcw/artifact/3bf767d5-6770-499b-bac4-353e536b4b7d/public width=25% height=25%>
31
+ <img src=https://imagedelivery.net/_wFNZAzgWNWPmneM1cyjcw/artifact/740ab221-a882-4ad9-bf46-fd3d60c503b9/public width=25% height=25%>
32
+ <img src=https://imagedelivery.net/_wFNZAzgWNWPmneM1cyjcw/artifact/e9700843-24fe-48ef-981b-5171c60f69c4/public width=25% height=25%>
33
+ <img src=https://imagedelivery.net/_wFNZAzgWNWPmneM1cyjcw/artifact/f70fde75-f7ba-4c78-8229-3e39314aaf43/public width=25% height=25%>
34
+ <img src=https://imagedelivery.net/_wFNZAzgWNWPmneM1cyjcw/artifact/422c26b6-38fa-4e0b-9c33-cffe4dc6f879/public width=25% height=25%>
35
+
36
+ You can see more samples at [PurpleSmartAI](https://purplesmart.ai/collection/top?nsfw=0&page=1&gen_type=txt2img&model=5&order=created_desc)
37
+
38
+ [PyTorch Model(Use this with Automatic1111 or other SD UIs)](https://mega.nz/file/9S0iSLpC#NqTvjJIzKoT2ScP9F5tFj7J3p5DhpLbIKiR70hMXqQ4)
39
+
40
+ ## Model Description
41
+
42
+ The model originally used for fine-tuning is [Stable Diffusion V1-5](https://huggingface.co/runwayml/stable-diffusion-v1-5), which is a latent image diffusion model trained on [LAION2B-en](https://huggingface.co/datasets/laion/laion2B-en).
43
+
44
+ This particular checkpoint has been fine-tuned with a learning rate of 5.0e-6 for 20 epochs on approximately 1.7M pony, furry and other cartoon text-image pairs (using metadata from derpibooru, e621 and danbooru).
45
+
46
+ ## Improvements over previous models
47
+
48
+ ### Better disentanglement of tag based prompts
49
+ Aka ["using Hidden States of CLIP’s Penultimate Layer"](https://blog.novelai.net/novelai-improvements-on-stable-diffusion-e10d38db82ac#:~:text=Using%20Hidden%20States%20of%20CLIP%E2%80%99s%20Penultimate%20Layer), a technique adopted by SD2 which should lead to generally higher quality and more tag driven outputs.
50
+ In our experiments using penultimate CLIP was not always the best choice so trying both CLIP skip of 1 and 2 is recommended.
51
+
52
+ ### Better support for non square images and increase of default resolution to 768px
53
+ This should allow you to generate full body image with resolution 512x768px without triggering "double head" glitches.
54
+
55
+ ### Removed SFM/3D bias
56
+ V2 model demostrated bias toward 3d/sfw visual styles so special care has been applied to restrict exposure to 3d/sfm images during training.
57
+
58
+ ### Improver data diversity
59
+ We ran multiple experiments finetuning models on 300k pony only and 600k pony only datasets, the resulting models demostrated worse quality for pony specific data. We concluded that despite more complicated prompting (and lack of pony bias by default), inclusion of large amount of non pony specific highly ranked non photorealistic images generally has strong positive efect.
60
+
61
+ ## License
62
+
63
+ This model is open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage.
64
+ The CreativeML OpenRAIL License specifies:
65
+
66
+ 1. You can't use the model to deliberately produce nor share illegal or harmful outputs or content
67
+ 2. The authors claims no rights on the outputs you generate, you are free to use them and are accountable for their use which must not go against the provisions set in the license
68
+ 3. You may re-distribute the weights and use the model commercially and/or as a service. If you do, please be aware you have to include the same use restrictions as the ones in the license and share a copy of the CreativeML OpenRAIL-M to all your users (please read the license entirely and carefully)
69
+ [Please read the full license here](https://huggingface.co/spaces/CompVis/stable-diffusion-license)
70
+
71
+ ## Downstream Uses
72
+
73
+ This model can be used for entertainment purposes and as a generative art assistant.
74
+
75
+ ## Example Code
76
+
77
+ ```python
78
+ import torch
79
+ from torch import autocast
80
+ from diffusers import StableDiffusionPipeline, DDIMScheduler
81
+ model_id = "AstraliteHeart/pony-diffusion-v3"
82
+ device = "cuda"
83
+ pipe = StableDiffusionPipeline.from_pretrained(
84
+ model_id,
85
+ torch_dtype=torch.float16,
86
+ revision="fp16",
87
+ scheduler=DDIMScheduler(
88
+ beta_start=0.00085,
89
+ beta_end=0.012,
90
+ beta_schedule="scaled_linear",
91
+ clip_sample=False,
92
+ set_alpha_to_one=False,
93
+ ),
94
+ )
95
+ pipe = pipe.to(device)
96
+ prompt = "pinkie pie anthro portrait wedding dress veil intricate highly detailed digital painting artstation concept art smooth sharp focus illustration Unreal Engine 5 8K"
97
+ with autocast("cuda"):
98
+ image = pipe(prompt, guidance_scale=7.5)["sample"][0]
99
+
100
+ image.save("cute_poner.png")
101
+ ```
102
+
103
+ ## Team Members and Acknowledgements
104
+
105
+ This project would not have been possible without the incredible work by the [CompVis Researchers](https://ommer-lab.com/).
106
+
107
+ - [Waifu-Diffusion for helping with finetuning and providing starting checkpoint](https://huggingface.co/hakurei/waifu-diffusion)
108
+
109
+ In order to reach us, you can join our [Discord server](https://discord.gg/WG78ZbSB).
pony-diffusion-v3/feature_extractor/preprocessor_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": {
3
+ "height": 224,
4
+ "width": 224
5
+ },
6
+ "do_center_crop": true,
7
+ "do_convert_rgb": true,
8
+ "do_normalize": true,
9
+ "do_rescale": true,
10
+ "do_resize": true,
11
+ "feature_extractor_type": "CLIPFeatureExtractor",
12
+ "image_mean": [
13
+ 0.48145466,
14
+ 0.4578275,
15
+ 0.40821073
16
+ ],
17
+ "image_processor_type": "CLIPFeatureExtractor",
18
+ "image_std": [
19
+ 0.26862954,
20
+ 0.26130258,
21
+ 0.27577711
22
+ ],
23
+ "resample": 3,
24
+ "rescale_factor": 0.00392156862745098,
25
+ "size": {
26
+ "shortest_edge": 224
27
+ }
28
+ }
pony-diffusion-v3/model_index.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.13.1",
4
+ "feature_extractor": [
5
+ "transformers",
6
+ "CLIPFeatureExtractor"
7
+ ],
8
+ "requires_safety_checker": true,
9
+ "safety_checker": [
10
+ "stable_diffusion",
11
+ "StableDiffusionSafetyChecker"
12
+ ],
13
+ "scheduler": [
14
+ "diffusers",
15
+ "PNDMScheduler"
16
+ ],
17
+ "text_encoder": [
18
+ "transformers",
19
+ "CLIPTextModel"
20
+ ],
21
+ "tokenizer": [
22
+ "transformers",
23
+ "CLIPTokenizer"
24
+ ],
25
+ "unet": [
26
+ "diffusers",
27
+ "UNet2DConditionModel"
28
+ ],
29
+ "vae": [
30
+ "diffusers",
31
+ "AutoencoderKL"
32
+ ]
33
+ }
pony-diffusion-v3/safety_checker/config.json ADDED
@@ -0,0 +1,181 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_commit_hash": "cb41f3a270d63d454d385fc2e4f571c487c253c5",
3
+ "_name_or_path": "CompVis/stable-diffusion-safety-checker",
4
+ "architectures": [
5
+ "StableDiffusionSafetyChecker"
6
+ ],
7
+ "initializer_factor": 1.0,
8
+ "logit_scale_init_value": 2.6592,
9
+ "model_type": "clip",
10
+ "projection_dim": 768,
11
+ "text_config": {
12
+ "_name_or_path": "",
13
+ "add_cross_attention": false,
14
+ "architectures": null,
15
+ "attention_dropout": 0.0,
16
+ "bad_words_ids": null,
17
+ "begin_suppress_tokens": null,
18
+ "bos_token_id": 0,
19
+ "chunk_size_feed_forward": 0,
20
+ "cross_attention_hidden_size": null,
21
+ "decoder_start_token_id": null,
22
+ "diversity_penalty": 0.0,
23
+ "do_sample": false,
24
+ "dropout": 0.0,
25
+ "early_stopping": false,
26
+ "encoder_no_repeat_ngram_size": 0,
27
+ "eos_token_id": 2,
28
+ "exponential_decay_length_penalty": null,
29
+ "finetuning_task": null,
30
+ "forced_bos_token_id": null,
31
+ "forced_eos_token_id": null,
32
+ "hidden_act": "quick_gelu",
33
+ "hidden_size": 768,
34
+ "id2label": {
35
+ "0": "LABEL_0",
36
+ "1": "LABEL_1"
37
+ },
38
+ "initializer_factor": 1.0,
39
+ "initializer_range": 0.02,
40
+ "intermediate_size": 3072,
41
+ "is_decoder": false,
42
+ "is_encoder_decoder": false,
43
+ "label2id": {
44
+ "LABEL_0": 0,
45
+ "LABEL_1": 1
46
+ },
47
+ "layer_norm_eps": 1e-05,
48
+ "length_penalty": 1.0,
49
+ "max_length": 20,
50
+ "max_position_embeddings": 77,
51
+ "min_length": 0,
52
+ "model_type": "clip_text_model",
53
+ "no_repeat_ngram_size": 0,
54
+ "num_attention_heads": 12,
55
+ "num_beam_groups": 1,
56
+ "num_beams": 1,
57
+ "num_hidden_layers": 12,
58
+ "num_return_sequences": 1,
59
+ "output_attentions": false,
60
+ "output_hidden_states": false,
61
+ "output_scores": false,
62
+ "pad_token_id": 1,
63
+ "prefix": null,
64
+ "problem_type": null,
65
+ "projection_dim": 512,
66
+ "pruned_heads": {},
67
+ "remove_invalid_values": false,
68
+ "repetition_penalty": 1.0,
69
+ "return_dict": true,
70
+ "return_dict_in_generate": false,
71
+ "sep_token_id": null,
72
+ "suppress_tokens": null,
73
+ "task_specific_params": null,
74
+ "temperature": 1.0,
75
+ "tf_legacy_loss": false,
76
+ "tie_encoder_decoder": false,
77
+ "tie_word_embeddings": true,
78
+ "tokenizer_class": null,
79
+ "top_k": 50,
80
+ "top_p": 1.0,
81
+ "torch_dtype": null,
82
+ "torchscript": false,
83
+ "transformers_version": "4.26.1",
84
+ "typical_p": 1.0,
85
+ "use_bfloat16": false,
86
+ "vocab_size": 49408
87
+ },
88
+ "text_config_dict": {
89
+ "hidden_size": 768,
90
+ "intermediate_size": 3072,
91
+ "num_attention_heads": 12,
92
+ "num_hidden_layers": 12
93
+ },
94
+ "torch_dtype": "float32",
95
+ "transformers_version": null,
96
+ "vision_config": {
97
+ "_name_or_path": "",
98
+ "add_cross_attention": false,
99
+ "architectures": null,
100
+ "attention_dropout": 0.0,
101
+ "bad_words_ids": null,
102
+ "begin_suppress_tokens": null,
103
+ "bos_token_id": null,
104
+ "chunk_size_feed_forward": 0,
105
+ "cross_attention_hidden_size": null,
106
+ "decoder_start_token_id": null,
107
+ "diversity_penalty": 0.0,
108
+ "do_sample": false,
109
+ "dropout": 0.0,
110
+ "early_stopping": false,
111
+ "encoder_no_repeat_ngram_size": 0,
112
+ "eos_token_id": null,
113
+ "exponential_decay_length_penalty": null,
114
+ "finetuning_task": null,
115
+ "forced_bos_token_id": null,
116
+ "forced_eos_token_id": null,
117
+ "hidden_act": "quick_gelu",
118
+ "hidden_size": 1024,
119
+ "id2label": {
120
+ "0": "LABEL_0",
121
+ "1": "LABEL_1"
122
+ },
123
+ "image_size": 224,
124
+ "initializer_factor": 1.0,
125
+ "initializer_range": 0.02,
126
+ "intermediate_size": 4096,
127
+ "is_decoder": false,
128
+ "is_encoder_decoder": false,
129
+ "label2id": {
130
+ "LABEL_0": 0,
131
+ "LABEL_1": 1
132
+ },
133
+ "layer_norm_eps": 1e-05,
134
+ "length_penalty": 1.0,
135
+ "max_length": 20,
136
+ "min_length": 0,
137
+ "model_type": "clip_vision_model",
138
+ "no_repeat_ngram_size": 0,
139
+ "num_attention_heads": 16,
140
+ "num_beam_groups": 1,
141
+ "num_beams": 1,
142
+ "num_channels": 3,
143
+ "num_hidden_layers": 24,
144
+ "num_return_sequences": 1,
145
+ "output_attentions": false,
146
+ "output_hidden_states": false,
147
+ "output_scores": false,
148
+ "pad_token_id": null,
149
+ "patch_size": 14,
150
+ "prefix": null,
151
+ "problem_type": null,
152
+ "projection_dim": 512,
153
+ "pruned_heads": {},
154
+ "remove_invalid_values": false,
155
+ "repetition_penalty": 1.0,
156
+ "return_dict": true,
157
+ "return_dict_in_generate": false,
158
+ "sep_token_id": null,
159
+ "suppress_tokens": null,
160
+ "task_specific_params": null,
161
+ "temperature": 1.0,
162
+ "tf_legacy_loss": false,
163
+ "tie_encoder_decoder": false,
164
+ "tie_word_embeddings": true,
165
+ "tokenizer_class": null,
166
+ "top_k": 50,
167
+ "top_p": 1.0,
168
+ "torch_dtype": null,
169
+ "torchscript": false,
170
+ "transformers_version": "4.26.1",
171
+ "typical_p": 1.0,
172
+ "use_bfloat16": false
173
+ },
174
+ "vision_config_dict": {
175
+ "hidden_size": 1024,
176
+ "intermediate_size": 4096,
177
+ "num_attention_heads": 16,
178
+ "num_hidden_layers": 24,
179
+ "patch_size": 14
180
+ }
181
+ }
pony-diffusion-v3/safety_checker/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:193490b58ef62739077262e833bf091c66c29488058681ac25cf7df3d8190974
3
+ size 1216061799
pony-diffusion-v3/scheduler/scheduler_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "PNDMScheduler",
3
+ "_diffusers_version": "0.13.1",
4
+ "beta_end": 0.012,
5
+ "beta_schedule": "scaled_linear",
6
+ "beta_start": 0.00085,
7
+ "clip_sample": false,
8
+ "num_train_timesteps": 1000,
9
+ "prediction_type": "epsilon",
10
+ "set_alpha_to_one": false,
11
+ "skip_prk_steps": true,
12
+ "steps_offset": 1,
13
+ "trained_betas": null
14
+ }
pony-diffusion-v3/text_encoder/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "openai/clip-vit-large-patch14",
3
+ "architectures": [
4
+ "CLIPTextModel"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 0,
8
+ "dropout": 0.0,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "quick_gelu",
11
+ "hidden_size": 768,
12
+ "initializer_factor": 1.0,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 77,
17
+ "model_type": "clip_text_model",
18
+ "num_attention_heads": 12,
19
+ "num_hidden_layers": 12,
20
+ "pad_token_id": 1,
21
+ "projection_dim": 768,
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.26.1",
24
+ "vocab_size": 49408
25
+ }
pony-diffusion-v3/text_encoder/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cc329c12fd921cda5679c859ad7025b1f3e9e492aa3cc3e87994de5287d5bd2
3
+ size 492305335
pony-diffusion-v3/tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pony-diffusion-v3/tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
pony-diffusion-v3/tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": {
4
+ "__type": "AddedToken",
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false
10
+ },
11
+ "do_lower_case": true,
12
+ "eos_token": {
13
+ "__type": "AddedToken",
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ },
20
+ "errors": "replace",
21
+ "model_max_length": 77,
22
+ "name_or_path": "openai/clip-vit-large-patch14",
23
+ "pad_token": "<|endoftext|>",
24
+ "special_tokens_map_file": "./special_tokens_map.json",
25
+ "tokenizer_class": "CLIPTokenizer",
26
+ "unk_token": {
27
+ "__type": "AddedToken",
28
+ "content": "<|endoftext|>",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false
33
+ }
34
+ }
pony-diffusion-v3/tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
pony-diffusion-v3/unet/config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.13.1",
4
+ "act_fn": "silu",
5
+ "attention_head_dim": 8,
6
+ "block_out_channels": [
7
+ 320,
8
+ 640,
9
+ 1280,
10
+ 1280
11
+ ],
12
+ "center_input_sample": false,
13
+ "class_embed_type": null,
14
+ "conv_in_kernel": 3,
15
+ "conv_out_kernel": 3,
16
+ "cross_attention_dim": 768,
17
+ "down_block_types": [
18
+ "CrossAttnDownBlock2D",
19
+ "CrossAttnDownBlock2D",
20
+ "CrossAttnDownBlock2D",
21
+ "DownBlock2D"
22
+ ],
23
+ "downsample_padding": 1,
24
+ "dual_cross_attention": false,
25
+ "flip_sin_to_cos": true,
26
+ "freq_shift": 0,
27
+ "in_channels": 4,
28
+ "layers_per_block": 2,
29
+ "mid_block_scale_factor": 1,
30
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
31
+ "norm_eps": 1e-05,
32
+ "norm_num_groups": 32,
33
+ "num_class_embeds": null,
34
+ "only_cross_attention": false,
35
+ "out_channels": 4,
36
+ "projection_class_embeddings_input_dim": null,
37
+ "resnet_time_scale_shift": "default",
38
+ "sample_size": 64,
39
+ "time_cond_proj_dim": null,
40
+ "time_embedding_type": "positional",
41
+ "timestep_post_act": null,
42
+ "up_block_types": [
43
+ "UpBlock2D",
44
+ "CrossAttnUpBlock2D",
45
+ "CrossAttnUpBlock2D",
46
+ "CrossAttnUpBlock2D"
47
+ ],
48
+ "upcast_attention": false,
49
+ "use_linear_projection": false
50
+ }
pony-diffusion-v3/unet/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2580bc61385ca07310a4ca384fb7e2d57366fb8b64ad656ae96924027bc9a1f7
3
+ size 3438354725
pony-diffusion-v3/vae/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.13.1",
4
+ "act_fn": "silu",
5
+ "block_out_channels": [
6
+ 128,
7
+ 256,
8
+ 512,
9
+ 512
10
+ ],
11
+ "down_block_types": [
12
+ "DownEncoderBlock2D",
13
+ "DownEncoderBlock2D",
14
+ "DownEncoderBlock2D",
15
+ "DownEncoderBlock2D"
16
+ ],
17
+ "in_channels": 3,
18
+ "latent_channels": 4,
19
+ "layers_per_block": 2,
20
+ "norm_num_groups": 32,
21
+ "out_channels": 3,
22
+ "sample_size": 512,
23
+ "scaling_factor": 0.18215,
24
+ "up_block_types": [
25
+ "UpDecoderBlock2D",
26
+ "UpDecoderBlock2D",
27
+ "UpDecoderBlock2D",
28
+ "UpDecoderBlock2D"
29
+ ]
30
+ }
pony-diffusion-v3/vae/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8cf5b49d164db18a485d392b2d9a9b4e3636d70613cb756d2e1bc460dd13161
3
+ size 334707217