johnrachwanpruna commited on
Commit
a8c793b
·
verified ·
1 Parent(s): c71324a

Add files using upload-large-folder tool

Browse files
README.md CHANGED
@@ -1,9 +1,9 @@
1
  ---
2
  library_name: diffusers
3
  tags:
 
4
  - pruna-ai
5
  - safetensors
6
- - pruna_pro-ai
7
  ---
8
 
9
  # Model Card for pruna-test/test-save-tiny-stable-diffusion-pipe-smashed-pro
 
1
  ---
2
  library_name: diffusers
3
  tags:
4
+ - pruna_pro-ai
5
  - pruna-ai
6
  - safetensors
 
7
  ---
8
 
9
  # Model Card for pruna-test/test-save-tiny-stable-diffusion-pipe-smashed-pro
model_index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
- "_diffusers_version": "0.34.0",
4
  "_name_or_path": "hf-internal-testing/tiny-stable-diffusion-pipe",
5
  "feature_extractor": [
6
  "transformers",
 
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.36.0",
4
  "_name_or_path": "hf-internal-testing/tiny-stable-diffusion-pipe",
5
  "feature_extractor": [
6
  "transformers",
safety_checker/config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_name_or_path": "/home/runner/.cache/huggingface/hub/models--hf-internal-testing--tiny-stable-diffusion-pipe/snapshots/3ee6c9f225f088ad5d35b624b6514b091e6a4849/safety_checker",
3
  "architectures": [
4
  "StableDiffusionSafetyChecker"
5
  ],
@@ -13,28 +12,41 @@
13
  "dropout": 0.1,
14
  "eos_token_id": 2,
15
  "gradient_checkpointing": false,
 
16
  "hidden_size": 32,
 
 
17
  "intermediate_size": 37,
 
18
  "max_position_embeddings": 512,
19
  "model_type": "clip_text_model",
20
  "num_attention_heads": 4,
21
  "num_hidden_layers": 5,
 
 
22
  "vocab_size": 99
23
  },
24
  "torch_dtype": "float32",
25
- "transformers_version": "4.48.0",
26
  "vision_config": {
27
  "_attn_implementation_autoset": true,
28
  "attention_dropout": 0.1,
29
  "dropout": 0.1,
30
  "gradient_checkpointing": false,
 
31
  "hidden_size": 32,
32
  "image_size": 30,
 
 
33
  "intermediate_size": 37,
 
34
  "model_type": "clip_vision_model",
35
  "num_attention_heads": 4,
 
36
  "num_hidden_layers": 5,
37
- "patch_size": 2
 
 
38
  },
39
  "vocab_size": 1000
40
  }
 
1
  {
 
2
  "architectures": [
3
  "StableDiffusionSafetyChecker"
4
  ],
 
12
  "dropout": 0.1,
13
  "eos_token_id": 2,
14
  "gradient_checkpointing": false,
15
+ "hidden_act": "quick_gelu",
16
  "hidden_size": 32,
17
+ "initializer_factor": 1.0,
18
+ "initializer_range": 0.02,
19
  "intermediate_size": 37,
20
+ "layer_norm_eps": 1e-05,
21
  "max_position_embeddings": 512,
22
  "model_type": "clip_text_model",
23
  "num_attention_heads": 4,
24
  "num_hidden_layers": 5,
25
+ "projection_dim": 512,
26
+ "torch_dtype": "float32",
27
  "vocab_size": 99
28
  },
29
  "torch_dtype": "float32",
30
+ "transformers_version": "4.51.0",
31
  "vision_config": {
32
  "_attn_implementation_autoset": true,
33
  "attention_dropout": 0.1,
34
  "dropout": 0.1,
35
  "gradient_checkpointing": false,
36
+ "hidden_act": "quick_gelu",
37
  "hidden_size": 32,
38
  "image_size": 30,
39
+ "initializer_factor": 1.0,
40
+ "initializer_range": 0.02,
41
  "intermediate_size": 37,
42
+ "layer_norm_eps": 1e-05,
43
  "model_type": "clip_vision_model",
44
  "num_attention_heads": 4,
45
+ "num_channels": 3,
46
  "num_hidden_layers": 5,
47
+ "patch_size": 2,
48
+ "projection_dim": 512,
49
+ "torch_dtype": "float32"
50
  },
51
  "vocab_size": 1000
52
  }
scheduler/scheduler_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "DDIMScheduler",
3
- "_diffusers_version": "0.34.0",
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
 
1
  {
2
  "_class_name": "DDIMScheduler",
3
+ "_diffusers_version": "0.36.0",
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
text_encoder/config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_name_or_path": "/home/runner/.cache/huggingface/hub/models--hf-internal-testing--tiny-stable-diffusion-pipe/snapshots/3ee6c9f225f088ad5d35b624b6514b091e6a4849/text_encoder",
3
  "architectures": [
4
  "CLIPTextModel"
5
  ],
@@ -20,6 +19,6 @@
20
  "pad_token_id": 1,
21
  "projection_dim": 512,
22
  "torch_dtype": "float32",
23
- "transformers_version": "4.48.0",
24
  "vocab_size": 1000
25
  }
 
1
  {
 
2
  "architectures": [
3
  "CLIPTextModel"
4
  ],
 
19
  "pad_token_id": 1,
20
  "projection_dim": 512,
21
  "torch_dtype": "float32",
22
+ "transformers_version": "4.51.0",
23
  "vocab_size": 1000
24
  }
unet/config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
- "_diffusers_version": "0.34.0",
4
  "_name_or_path": "/home/runner/.cache/huggingface/hub/models--hf-internal-testing--tiny-stable-diffusion-pipe/snapshots/3ee6c9f225f088ad5d35b624b6514b091e6a4849/unet",
5
  "act_fn": "silu",
6
  "addition_embed_type": null,
 
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.36.0",
4
  "_name_or_path": "/home/runner/.cache/huggingface/hub/models--hf-internal-testing--tiny-stable-diffusion-pipe/snapshots/3ee6c9f225f088ad5d35b624b6514b091e6a4849/unet",
5
  "act_fn": "silu",
6
  "addition_embed_type": null,
vae/config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "AutoencoderKL",
3
- "_diffusers_version": "0.34.0",
4
  "_name_or_path": "/home/runner/.cache/huggingface/hub/models--hf-internal-testing--tiny-stable-diffusion-pipe/snapshots/3ee6c9f225f088ad5d35b624b6514b091e6a4849/vae",
5
  "act_fn": "silu",
6
  "block_out_channels": [
 
1
  {
2
  "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.36.0",
4
  "_name_or_path": "/home/runner/.cache/huggingface/hub/models--hf-internal-testing--tiny-stable-diffusion-pipe/snapshots/3ee6c9f225f088ad5d35b624b6514b091e6a4849/vae",
5
  "act_fn": "silu",
6
  "block_out_channels": [