space-jango commited on
Commit
40d4d79
·
verified ·
1 Parent(s): 47eb350

End of training

Browse files
Files changed (37) hide show
  1. .DS_Store +0 -0
  2. 050D7708-D270-46FF-BA9E-D326FF3B07A9_1_201_a.jpeg +0 -0
  3. 2078805C-6398-44A7-ABAF-6E1F710383E9_1_201_a.jpeg +0 -0
  4. 2C80B72C-8E9C-4F0B-B778-B893CC2A136E_1_201_a.jpeg +0 -0
  5. 34F8A143-A77E-4CA5-AD4E-FD6795E999F0_1_201_a.jpeg +0 -0
  6. 51616378-4DF0-40D1-846B-BD2991F165A4_1_201_a.jpeg +0 -0
  7. 5D71B8E9-02E1-4843-A5F7-42DD56AAAF49_1_201_a.jpeg +0 -0
  8. 5FE109B8-8FAB-4274-829B-0ACC6107279A_4_5005_c.jpeg +0 -0
  9. 7782B667-E890-4BBC-8226-6756A30B7E27_1_201_a.jpeg +0 -0
  10. 8083E762-1138-4542-B23B-2B56315B42CA_1_201_a.jpeg +0 -0
  11. 84C29A56-428C-470C-8C09-DA2200888C2B_1_201_a.jpeg +0 -0
  12. 96589AED-7A45-4951-956D-64A95F1562BA_1_201_a.jpeg +0 -0
  13. C404BECE-E500-4098-B0E8-FFF2AFD0F762_1_201_a.jpeg +0 -0
  14. D3637B5F-A77A-440E-93AE-B96F27621DE7_1_201_a.jpeg +0 -0
  15. E723AAAA-56E8-4E2C-84FA-0569B71CB126_1_201_a.jpeg +0 -0
  16. E94E6828-3143-46DC-AD80-E5B5F54BE8CC_1_201_a.jpeg +0 -0
  17. README.md +43 -0
  18. checkpoint-500/optimizer.bin +3 -0
  19. checkpoint-500/random_states_0.pkl +3 -0
  20. checkpoint-500/scheduler.bin +3 -0
  21. checkpoint-500/unet/config.json +68 -0
  22. checkpoint-500/unet/diffusion_pytorch_model.safetensors +3 -0
  23. logs/dreambooth/1725627633.522811/events.out.tfevents.1725627633.trints-mbp-2.marlin-yo.ts.net.27312.1 +3 -0
  24. logs/dreambooth/1725627633.523745/hparams.yml +62 -0
  25. logs/dreambooth/events.out.tfevents.1725627633.trints-mbp-2.marlin-yo.ts.net.27312.0 +3 -0
  26. model_index.json +38 -0
  27. scheduler/scheduler_config.json +20 -0
  28. text_encoder/config.json +25 -0
  29. text_encoder/model.safetensors +3 -0
  30. tokenizer/merges.txt +0 -0
  31. tokenizer/special_tokens_map.json +24 -0
  32. tokenizer/tokenizer_config.json +30 -0
  33. tokenizer/vocab.json +0 -0
  34. unet/config.json +68 -0
  35. unet/diffusion_pytorch_model.safetensors +3 -0
  36. vae/config.json +38 -0
  37. vae/diffusion_pytorch_model.safetensors +3 -0
.DS_Store ADDED
Binary file (6.15 kB). View file
 
050D7708-D270-46FF-BA9E-D326FF3B07A9_1_201_a.jpeg ADDED
2078805C-6398-44A7-ABAF-6E1F710383E9_1_201_a.jpeg ADDED
2C80B72C-8E9C-4F0B-B778-B893CC2A136E_1_201_a.jpeg ADDED
34F8A143-A77E-4CA5-AD4E-FD6795E999F0_1_201_a.jpeg ADDED
51616378-4DF0-40D1-846B-BD2991F165A4_1_201_a.jpeg ADDED
5D71B8E9-02E1-4843-A5F7-42DD56AAAF49_1_201_a.jpeg ADDED
5FE109B8-8FAB-4274-829B-0ACC6107279A_4_5005_c.jpeg ADDED
7782B667-E890-4BBC-8226-6756A30B7E27_1_201_a.jpeg ADDED
8083E762-1138-4542-B23B-2B56315B42CA_1_201_a.jpeg ADDED
84C29A56-428C-470C-8C09-DA2200888C2B_1_201_a.jpeg ADDED
96589AED-7A45-4951-956D-64A95F1562BA_1_201_a.jpeg ADDED
C404BECE-E500-4098-B0E8-FFF2AFD0F762_1_201_a.jpeg ADDED
D3637B5F-A77A-440E-93AE-B96F27621DE7_1_201_a.jpeg ADDED
E723AAAA-56E8-4E2C-84FA-0569B71CB126_1_201_a.jpeg ADDED
E94E6828-3143-46DC-AD80-E5B5F54BE8CC_1_201_a.jpeg ADDED
README.md ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: dreamlike-art/dreamlike-photoreal-2.0
3
+ library_name: diffusers
4
+ license: creativeml-openrail-m
5
+ tags:
6
+ - text-to-image
7
+ - dreambooth
8
+ - diffusers-training
9
+ - stable-diffusion
10
+ - stable-diffusion-diffusers
11
+ inference: true
12
+ instance_prompt: photo of mrclblch
13
+ ---
14
+
15
+ <!-- This model card has been generated automatically according to the information the training script had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
+
18
+
19
+ # DreamBooth - space-jango/mrclblch
20
+
21
+ This is a dreambooth model derived from dreamlike-art/dreamlike-photoreal-2.0. The weights were trained on photo of mrclblch using [DreamBooth](https://dreambooth.github.io/).
22
+ You can find some example images in the following.
23
+
24
+
25
+
26
+ DreamBooth for the text encoder was enabled: False.
27
+
28
+
29
+ ## Intended uses & limitations
30
+
31
+ #### How to use
32
+
33
+ ```python
34
+ # TODO: add an example code snippet for running this diffusion pipeline
35
+ ```
36
+
37
+ #### Limitations and bias
38
+
39
+ [TODO: provide examples of latent issues and potential remediations]
40
+
41
+ ## Training details
42
+
43
+ [TODO: describe the data used to train the model]
checkpoint-500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f42cb91388df7cefa6854a6af1124e5849fd061b19ea89c21ab68dc28f1d4456
3
+ size 6876730964
checkpoint-500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c009b54553980a59225da8bef1f33947de2249eff8406b49b7e90b615a40e94
3
+ size 14084
checkpoint-500/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3644e8028e83d43fa25596dded54dba49fa3868d8762c42218c7ef0b7b627231
3
+ size 1000
checkpoint-500/unet/config.json ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.31.0.dev0",
4
+ "_name_or_path": "dreamlike-art/dreamlike-photoreal-2.0",
5
+ "act_fn": "silu",
6
+ "addition_embed_type": null,
7
+ "addition_embed_type_num_heads": 64,
8
+ "addition_time_embed_dim": null,
9
+ "attention_head_dim": 8,
10
+ "attention_type": "default",
11
+ "block_out_channels": [
12
+ 320,
13
+ 640,
14
+ 1280,
15
+ 1280
16
+ ],
17
+ "center_input_sample": false,
18
+ "class_embed_type": null,
19
+ "class_embeddings_concat": false,
20
+ "conv_in_kernel": 3,
21
+ "conv_out_kernel": 3,
22
+ "cross_attention_dim": 768,
23
+ "cross_attention_norm": null,
24
+ "down_block_types": [
25
+ "CrossAttnDownBlock2D",
26
+ "CrossAttnDownBlock2D",
27
+ "CrossAttnDownBlock2D",
28
+ "DownBlock2D"
29
+ ],
30
+ "downsample_padding": 1,
31
+ "dropout": 0.0,
32
+ "dual_cross_attention": false,
33
+ "encoder_hid_dim": null,
34
+ "encoder_hid_dim_type": null,
35
+ "flip_sin_to_cos": true,
36
+ "freq_shift": 0,
37
+ "in_channels": 4,
38
+ "layers_per_block": 2,
39
+ "mid_block_only_cross_attention": null,
40
+ "mid_block_scale_factor": 1,
41
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
42
+ "norm_eps": 1e-05,
43
+ "norm_num_groups": 32,
44
+ "num_attention_heads": null,
45
+ "num_class_embeds": null,
46
+ "only_cross_attention": false,
47
+ "out_channels": 4,
48
+ "projection_class_embeddings_input_dim": null,
49
+ "resnet_out_scale_factor": 1.0,
50
+ "resnet_skip_time_act": false,
51
+ "resnet_time_scale_shift": "default",
52
+ "reverse_transformer_layers_per_block": null,
53
+ "sample_size": 96,
54
+ "time_cond_proj_dim": null,
55
+ "time_embedding_act_fn": null,
56
+ "time_embedding_dim": null,
57
+ "time_embedding_type": "positional",
58
+ "timestep_post_act": null,
59
+ "transformer_layers_per_block": 1,
60
+ "up_block_types": [
61
+ "UpBlock2D",
62
+ "CrossAttnUpBlock2D",
63
+ "CrossAttnUpBlock2D",
64
+ "CrossAttnUpBlock2D"
65
+ ],
66
+ "upcast_attention": false,
67
+ "use_linear_projection": false
68
+ }
checkpoint-500/unet/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a1cb86556be8ce2942e94fca1979beb7780dbcdc465bc7200f8ab0b833c2152
3
+ size 3438167536
logs/dreambooth/1725627633.522811/events.out.tfevents.1725627633.trints-mbp-2.marlin-yo.ts.net.27312.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d492ba82ac90ad353aae0f8041e3411c2dd251ce18490ddc993930ab85732e2e
3
+ size 3018
logs/dreambooth/1725627633.523745/hparams.yml ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ center_crop: false
7
+ checkpointing_steps: 500
8
+ checkpoints_total_limit: null
9
+ class_data_dir: class
10
+ class_labels_conditioning: null
11
+ class_prompt: A 55-year-old white man, medium build, short hair, small beard, light
12
+ skin, wearing casual clothing, neutral facial expression, standing in various natural
13
+ or urban environments.
14
+ dataloader_num_workers: 0
15
+ enable_xformers_memory_efficient_attention: false
16
+ gradient_accumulation_steps: 1
17
+ gradient_checkpointing: false
18
+ hub_model_id: null
19
+ hub_token: null
20
+ instance_data_dir: resize
21
+ instance_prompt: photo of mrclblch
22
+ learning_rate: 5.0e-06
23
+ local_rank: -1
24
+ logging_dir: logs
25
+ lr_num_cycles: 1
26
+ lr_power: 1.0
27
+ lr_scheduler: constant
28
+ lr_warmup_steps: 0
29
+ max_grad_norm: 1.0
30
+ max_train_steps: 800
31
+ mixed_precision: null
32
+ num_class_images: 101
33
+ num_train_epochs: 8
34
+ num_validation_images: 4
35
+ offset_noise: false
36
+ output_dir: mrclblch
37
+ pre_compute_text_embeddings: false
38
+ pretrained_model_name_or_path: dreamlike-art/dreamlike-photoreal-2.0
39
+ prior_generation_precision: null
40
+ prior_loss_weight: 1.0
41
+ push_to_hub: true
42
+ report_to: tensorboard
43
+ resolution: 512
44
+ resume_from_checkpoint: null
45
+ revision: null
46
+ sample_batch_size: 4
47
+ scale_lr: false
48
+ seed: null
49
+ set_grads_to_none: false
50
+ skip_save_text_encoder: false
51
+ snr_gamma: null
52
+ text_encoder_use_attention_mask: false
53
+ tokenizer_max_length: null
54
+ tokenizer_name: null
55
+ train_batch_size: 1
56
+ train_text_encoder: false
57
+ use_8bit_adam: false
58
+ validation_prompt: null
59
+ validation_scheduler: DPMSolverMultistepScheduler
60
+ validation_steps: 100
61
+ variant: null
62
+ with_prior_preservation: true
logs/dreambooth/events.out.tfevents.1725627633.trints-mbp-2.marlin-yo.ts.net.27312.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb04351979805c0b3a4d6396786be2b5ced63e5fd422c62a1eb00da22bf510ed
3
+ size 67034
model_index.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.31.0.dev0",
4
+ "_name_or_path": "dreamlike-art/dreamlike-photoreal-2.0",
5
+ "feature_extractor": [
6
+ null,
7
+ null
8
+ ],
9
+ "image_encoder": [
10
+ null,
11
+ null
12
+ ],
13
+ "requires_safety_checker": null,
14
+ "safety_checker": [
15
+ null,
16
+ null
17
+ ],
18
+ "scheduler": [
19
+ "diffusers",
20
+ "DDIMScheduler"
21
+ ],
22
+ "text_encoder": [
23
+ "transformers",
24
+ "CLIPTextModel"
25
+ ],
26
+ "tokenizer": [
27
+ "transformers",
28
+ "CLIPTokenizer"
29
+ ],
30
+ "unet": [
31
+ "diffusers",
32
+ "UNet2DConditionModel"
33
+ ],
34
+ "vae": [
35
+ "diffusers",
36
+ "AutoencoderKL"
37
+ ]
38
+ }
scheduler/scheduler_config.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "DDIMScheduler",
3
+ "_diffusers_version": "0.31.0.dev0",
4
+ "beta_end": 0.012,
5
+ "beta_schedule": "scaled_linear",
6
+ "beta_start": 0.00085,
7
+ "clip_sample": false,
8
+ "clip_sample_range": 1.0,
9
+ "dynamic_thresholding_ratio": 0.995,
10
+ "num_train_timesteps": 1000,
11
+ "prediction_type": "epsilon",
12
+ "rescale_betas_zero_snr": false,
13
+ "sample_max_value": 1.0,
14
+ "set_alpha_to_one": false,
15
+ "skip_prk_steps": true,
16
+ "steps_offset": 1,
17
+ "thresholding": false,
18
+ "timestep_spacing": "leading",
19
+ "trained_betas": null
20
+ }
text_encoder/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "dreamlike-art/dreamlike-photoreal-2.0",
3
+ "architectures": [
4
+ "CLIPTextModel"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 0,
8
+ "dropout": 0.0,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "quick_gelu",
11
+ "hidden_size": 768,
12
+ "initializer_factor": 1.0,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 77,
17
+ "model_type": "clip_text_model",
18
+ "num_attention_heads": 12,
19
+ "num_hidden_layers": 12,
20
+ "pad_token_id": 1,
21
+ "projection_dim": 768,
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.44.2",
24
+ "vocab_size": 49408
25
+ }
text_encoder/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb535f506b2581e3cc63e35cda6f7993bd6c53855955d475d91acffbf1ac1227
3
+ size 492265168
tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "49406": {
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "49407": {
13
+ "content": "<|endoftext|>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ }
20
+ },
21
+ "bos_token": "<|startoftext|>",
22
+ "clean_up_tokenization_spaces": true,
23
+ "do_lower_case": true,
24
+ "eos_token": "<|endoftext|>",
25
+ "errors": "replace",
26
+ "model_max_length": 77,
27
+ "pad_token": "<|endoftext|>",
28
+ "tokenizer_class": "CLIPTokenizer",
29
+ "unk_token": "<|endoftext|>"
30
+ }
tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
unet/config.json ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.31.0.dev0",
4
+ "_name_or_path": "dreamlike-art/dreamlike-photoreal-2.0",
5
+ "act_fn": "silu",
6
+ "addition_embed_type": null,
7
+ "addition_embed_type_num_heads": 64,
8
+ "addition_time_embed_dim": null,
9
+ "attention_head_dim": 8,
10
+ "attention_type": "default",
11
+ "block_out_channels": [
12
+ 320,
13
+ 640,
14
+ 1280,
15
+ 1280
16
+ ],
17
+ "center_input_sample": false,
18
+ "class_embed_type": null,
19
+ "class_embeddings_concat": false,
20
+ "conv_in_kernel": 3,
21
+ "conv_out_kernel": 3,
22
+ "cross_attention_dim": 768,
23
+ "cross_attention_norm": null,
24
+ "down_block_types": [
25
+ "CrossAttnDownBlock2D",
26
+ "CrossAttnDownBlock2D",
27
+ "CrossAttnDownBlock2D",
28
+ "DownBlock2D"
29
+ ],
30
+ "downsample_padding": 1,
31
+ "dropout": 0.0,
32
+ "dual_cross_attention": false,
33
+ "encoder_hid_dim": null,
34
+ "encoder_hid_dim_type": null,
35
+ "flip_sin_to_cos": true,
36
+ "freq_shift": 0,
37
+ "in_channels": 4,
38
+ "layers_per_block": 2,
39
+ "mid_block_only_cross_attention": null,
40
+ "mid_block_scale_factor": 1,
41
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
42
+ "norm_eps": 1e-05,
43
+ "norm_num_groups": 32,
44
+ "num_attention_heads": null,
45
+ "num_class_embeds": null,
46
+ "only_cross_attention": false,
47
+ "out_channels": 4,
48
+ "projection_class_embeddings_input_dim": null,
49
+ "resnet_out_scale_factor": 1.0,
50
+ "resnet_skip_time_act": false,
51
+ "resnet_time_scale_shift": "default",
52
+ "reverse_transformer_layers_per_block": null,
53
+ "sample_size": 96,
54
+ "time_cond_proj_dim": null,
55
+ "time_embedding_act_fn": null,
56
+ "time_embedding_dim": null,
57
+ "time_embedding_type": "positional",
58
+ "timestep_post_act": null,
59
+ "transformer_layers_per_block": 1,
60
+ "up_block_types": [
61
+ "UpBlock2D",
62
+ "CrossAttnUpBlock2D",
63
+ "CrossAttnUpBlock2D",
64
+ "CrossAttnUpBlock2D"
65
+ ],
66
+ "upcast_attention": false,
67
+ "use_linear_projection": false
68
+ }
unet/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc3a0a41bd7513abc37babf6bedefe2b61f0ae933a7f17b7e33b927bf15223a6
3
+ size 3438167536
vae/config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.31.0.dev0",
4
+ "_name_or_path": "/Users/jangbelche/.cache/huggingface/hub/models--dreamlike-art--dreamlike-photoreal-2.0/snapshots/d9e27ac81cfa72def39d74ca673219c349f0a0d5/vae",
5
+ "act_fn": "silu",
6
+ "block_out_channels": [
7
+ 128,
8
+ 256,
9
+ 512,
10
+ 512
11
+ ],
12
+ "down_block_types": [
13
+ "DownEncoderBlock2D",
14
+ "DownEncoderBlock2D",
15
+ "DownEncoderBlock2D",
16
+ "DownEncoderBlock2D"
17
+ ],
18
+ "force_upcast": true,
19
+ "in_channels": 3,
20
+ "latent_channels": 4,
21
+ "latents_mean": null,
22
+ "latents_std": null,
23
+ "layers_per_block": 2,
24
+ "mid_block_add_attention": true,
25
+ "norm_num_groups": 32,
26
+ "out_channels": 3,
27
+ "sample_size": 256,
28
+ "scaling_factor": 0.18215,
29
+ "shift_factor": null,
30
+ "up_block_types": [
31
+ "UpDecoderBlock2D",
32
+ "UpDecoderBlock2D",
33
+ "UpDecoderBlock2D",
34
+ "UpDecoderBlock2D"
35
+ ],
36
+ "use_post_quant_conv": true,
37
+ "use_quant_conv": true
38
+ }
vae/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9a6b37f3d69017f413347db4dd1b1a808aed4425c1c8a7767d5a6e349da5bbb
3
+ size 334643268