adhikjoshi commited on
Commit
465c410
·
verified ·
1 Parent(s): 9f013ba

Upload StableDiffusionPipeline

Browse files
README.md CHANGED
@@ -1,11 +1,12 @@
1
  ---
2
  license: creativeml-openrail-m
3
  tags:
4
- - modelslab.com
5
- - stable-diffusion-api
6
- - text-to-image
7
- - ultra-realistic
8
  pinned: true
 
9
  ---
10
 
11
  # Animixed API Inference
 
1
  ---
2
  license: creativeml-openrail-m
3
  tags:
4
+ - modelslab.com
5
+ - stable-diffusion-api
6
+ - text-to-image
7
+ - ultra-realistic
8
  pinned: true
9
+ library_name: diffusers
10
  ---
11
 
12
  # Animixed API Inference
feature_extractor/preprocessor_config.json CHANGED
@@ -8,13 +8,12 @@
8
  "do_normalize": true,
9
  "do_rescale": true,
10
  "do_resize": true,
11
- "feature_extractor_type": "CLIPFeatureExtractor",
12
  "image_mean": [
13
  0.48145466,
14
  0.4578275,
15
  0.40821073
16
  ],
17
- "image_processor_type": "CLIPFeatureExtractor",
18
  "image_std": [
19
  0.26862954,
20
  0.26130258,
@@ -24,6 +23,5 @@
24
  "rescale_factor": 0.00392156862745098,
25
  "size": {
26
  "shortest_edge": 224
27
- },
28
- "use_square_size": false
29
  }
 
8
  "do_normalize": true,
9
  "do_rescale": true,
10
  "do_resize": true,
 
11
  "image_mean": [
12
  0.48145466,
13
  0.4578275,
14
  0.40821073
15
  ],
16
+ "image_processor_type": "CLIPImageProcessor",
17
  "image_std": [
18
  0.26862954,
19
  0.26130258,
 
23
  "rescale_factor": 0.00392156862745098,
24
  "size": {
25
  "shortest_edge": 224
26
+ }
 
27
  }
model_index.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
- "_diffusers_version": "0.28.0.dev0",
4
  "feature_extractor": [
5
  "transformers",
6
- "CLIPFeatureExtractor"
7
  ],
8
  "image_encoder": [
9
  null,
@@ -11,8 +11,8 @@
11
  ],
12
  "requires_safety_checker": true,
13
  "safety_checker": [
14
- "stable_diffusion",
15
- "StableDiffusionSafetyChecker"
16
  ],
17
  "scheduler": [
18
  "diffusers",
 
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.32.2",
4
  "feature_extractor": [
5
  "transformers",
6
+ "CLIPImageProcessor"
7
  ],
8
  "image_encoder": [
9
  null,
 
11
  ],
12
  "requires_safety_checker": true,
13
  "safety_checker": [
14
+ null,
15
+ null
16
  ],
17
  "scheduler": [
18
  "diffusers",
scheduler/scheduler_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "PNDMScheduler",
3
- "_diffusers_version": "0.28.0.dev0",
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
 
1
  {
2
  "_class_name": "PNDMScheduler",
3
+ "_diffusers_version": "0.32.2",
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
text_encoder/config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "architectures": [
3
  "CLIPTextModel"
4
  ],
@@ -18,7 +19,7 @@
18
  "num_hidden_layers": 12,
19
  "pad_token_id": 1,
20
  "projection_dim": 768,
21
- "torch_dtype": "float32",
22
- "transformers_version": "4.35.2",
23
  "vocab_size": 49408
24
  }
 
1
  {
2
+ "_name_or_path": "openai/clip-vit-large-patch14",
3
  "architectures": [
4
  "CLIPTextModel"
5
  ],
 
19
  "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
  "projection_dim": 768,
22
+ "torch_dtype": "float16",
23
+ "transformers_version": "4.43.1",
24
  "vocab_size": 49408
25
  }
text_encoder/model.fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51397c2d800c8c9d89a014d8537b0335c564cfa12ba8ce2aa344c8f23d9385e8
3
+ size 246144152
tokenizer/special_tokens_map.json CHANGED
@@ -13,13 +13,7 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": {
17
- "content": "<|endoftext|>",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
  "unk_token": {
24
  "content": "<|endoftext|>",
25
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": "<|endoftext|>",
 
 
 
 
 
 
17
  "unk_token": {
18
  "content": "<|endoftext|>",
19
  "lstrip": false,
unet/config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
- "_diffusers_version": "0.28.0.dev0",
4
  "act_fn": "silu",
5
  "addition_embed_type": null,
6
  "addition_embed_type_num_heads": 64,
 
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.32.2",
4
  "act_fn": "silu",
5
  "addition_embed_type": null,
6
  "addition_embed_type_num_heads": 64,
unet/diffusion_pytorch_model.fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eec42a428535c34b0a54f333ee467bf1f5fa4af5c909f6b49b451b784bc1a2cd
3
+ size 1719125304
vae/config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "AutoencoderKL",
3
- "_diffusers_version": "0.28.0.dev0",
4
  "act_fn": "silu",
5
  "block_out_channels": [
6
  128,
@@ -20,14 +20,18 @@
20
  "latents_mean": null,
21
  "latents_std": null,
22
  "layers_per_block": 2,
 
23
  "norm_num_groups": 32,
24
  "out_channels": 3,
25
  "sample_size": 512,
26
  "scaling_factor": 0.18215,
 
27
  "up_block_types": [
28
  "UpDecoderBlock2D",
29
  "UpDecoderBlock2D",
30
  "UpDecoderBlock2D",
31
  "UpDecoderBlock2D"
32
- ]
 
 
33
  }
 
1
  {
2
  "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.32.2",
4
  "act_fn": "silu",
5
  "block_out_channels": [
6
  128,
 
20
  "latents_mean": null,
21
  "latents_std": null,
22
  "layers_per_block": 2,
23
+ "mid_block_add_attention": true,
24
  "norm_num_groups": 32,
25
  "out_channels": 3,
26
  "sample_size": 512,
27
  "scaling_factor": 0.18215,
28
+ "shift_factor": null,
29
  "up_block_types": [
30
  "UpDecoderBlock2D",
31
  "UpDecoderBlock2D",
32
  "UpDecoderBlock2D",
33
  "UpDecoderBlock2D"
34
+ ],
35
+ "use_post_quant_conv": true,
36
+ "use_quant_conv": true
37
  }
vae/diffusion_pytorch_model.fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a08607e9d4f473d67a63070490db0c1e95b50d94789dc02b63851ace9a79d44
3
+ size 167335342