Upload folder using huggingface_hub
Browse files- checkpoint-epoch-10/config.json +18 -0
- checkpoint-epoch-10/diffusion_pytorch_model.safetensors +3 -0
- checkpoint-epoch-2/config.json +18 -0
- checkpoint-epoch-2/diffusion_pytorch_model.safetensors +3 -0
- checkpoint-epoch-4/config.json +18 -0
- checkpoint-epoch-4/diffusion_pytorch_model.safetensors +3 -0
- checkpoint-epoch-6/config.json +18 -0
- checkpoint-epoch-6/diffusion_pytorch_model.safetensors +3 -0
- checkpoint-epoch-8/config.json +18 -0
- checkpoint-epoch-8/diffusion_pytorch_model.safetensors +3 -0
- model_index.json +49 -0
- scheduler/scheduler_config.json +18 -0
- text_encoder/config.json +24 -0
- text_encoder/model.safetensors +3 -0
- text_encoder_2/config.json +24 -0
- text_encoder_2/model.safetensors +3 -0
- text_encoder_3/config.json +33 -0
- text_encoder_3/model.safetensors +3 -0
- tokenizer/tokenizer.json +0 -0
- tokenizer/tokenizer_config.json +14 -0
- tokenizer_2/tokenizer.json +0 -0
- tokenizer_2/tokenizer_config.json +14 -0
- tokenizer_3/tokenizer.json +0 -0
- tokenizer_3/tokenizer_config.json +116 -0
- transformer/config.json +18 -0
- transformer/diffusion_pytorch_model.safetensors +3 -0
- vae/config.json +38 -0
- vae/diffusion_pytorch_model.safetensors +3 -0
checkpoint-epoch-10/config.json
ADDED
|
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_class_name": "SD3Transformer2DModel",
|
| 3 |
+
"_diffusers_version": "0.37.0",
|
| 4 |
+
"_name_or_path": "stabilityai/stable-diffusion-3-medium-diffusers",
|
| 5 |
+
"attention_head_dim": 64,
|
| 6 |
+
"caption_projection_dim": 1536,
|
| 7 |
+
"dual_attention_layers": [],
|
| 8 |
+
"in_channels": 16,
|
| 9 |
+
"joint_attention_dim": 4096,
|
| 10 |
+
"num_attention_heads": 24,
|
| 11 |
+
"num_layers": 24,
|
| 12 |
+
"out_channels": 16,
|
| 13 |
+
"patch_size": 2,
|
| 14 |
+
"pooled_projection_dim": 2048,
|
| 15 |
+
"pos_embed_max_size": 192,
|
| 16 |
+
"qk_norm": null,
|
| 17 |
+
"sample_size": 128
|
| 18 |
+
}
|
checkpoint-epoch-10/diffusion_pytorch_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:80c25f3c7c73e3fac20822fed0c19983defff965c3805582b2dd4c39cb1d4e48
|
| 3 |
+
size 4169982088
|
checkpoint-epoch-2/config.json
ADDED
|
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_class_name": "SD3Transformer2DModel",
|
| 3 |
+
"_diffusers_version": "0.37.0",
|
| 4 |
+
"_name_or_path": "stabilityai/stable-diffusion-3-medium-diffusers",
|
| 5 |
+
"attention_head_dim": 64,
|
| 6 |
+
"caption_projection_dim": 1536,
|
| 7 |
+
"dual_attention_layers": [],
|
| 8 |
+
"in_channels": 16,
|
| 9 |
+
"joint_attention_dim": 4096,
|
| 10 |
+
"num_attention_heads": 24,
|
| 11 |
+
"num_layers": 24,
|
| 12 |
+
"out_channels": 16,
|
| 13 |
+
"patch_size": 2,
|
| 14 |
+
"pooled_projection_dim": 2048,
|
| 15 |
+
"pos_embed_max_size": 192,
|
| 16 |
+
"qk_norm": null,
|
| 17 |
+
"sample_size": 128
|
| 18 |
+
}
|
checkpoint-epoch-2/diffusion_pytorch_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a6e853b971f8150a48bc42d67f18223da2a748f383014e49b07cf7a209c53882
|
| 3 |
+
size 4169982088
|
checkpoint-epoch-4/config.json
ADDED
|
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_class_name": "SD3Transformer2DModel",
|
| 3 |
+
"_diffusers_version": "0.37.0",
|
| 4 |
+
"_name_or_path": "stabilityai/stable-diffusion-3-medium-diffusers",
|
| 5 |
+
"attention_head_dim": 64,
|
| 6 |
+
"caption_projection_dim": 1536,
|
| 7 |
+
"dual_attention_layers": [],
|
| 8 |
+
"in_channels": 16,
|
| 9 |
+
"joint_attention_dim": 4096,
|
| 10 |
+
"num_attention_heads": 24,
|
| 11 |
+
"num_layers": 24,
|
| 12 |
+
"out_channels": 16,
|
| 13 |
+
"patch_size": 2,
|
| 14 |
+
"pooled_projection_dim": 2048,
|
| 15 |
+
"pos_embed_max_size": 192,
|
| 16 |
+
"qk_norm": null,
|
| 17 |
+
"sample_size": 128
|
| 18 |
+
}
|
checkpoint-epoch-4/diffusion_pytorch_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6843ee54b600caeb5ede3381662a981ca66f2cfc893e11c8daff7bc759b934c4
|
| 3 |
+
size 4169982088
|
checkpoint-epoch-6/config.json
ADDED
|
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_class_name": "SD3Transformer2DModel",
|
| 3 |
+
"_diffusers_version": "0.37.0",
|
| 4 |
+
"_name_or_path": "stabilityai/stable-diffusion-3-medium-diffusers",
|
| 5 |
+
"attention_head_dim": 64,
|
| 6 |
+
"caption_projection_dim": 1536,
|
| 7 |
+
"dual_attention_layers": [],
|
| 8 |
+
"in_channels": 16,
|
| 9 |
+
"joint_attention_dim": 4096,
|
| 10 |
+
"num_attention_heads": 24,
|
| 11 |
+
"num_layers": 24,
|
| 12 |
+
"out_channels": 16,
|
| 13 |
+
"patch_size": 2,
|
| 14 |
+
"pooled_projection_dim": 2048,
|
| 15 |
+
"pos_embed_max_size": 192,
|
| 16 |
+
"qk_norm": null,
|
| 17 |
+
"sample_size": 128
|
| 18 |
+
}
|
checkpoint-epoch-6/diffusion_pytorch_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9dc8727919494f6d7514af06af75f11d158a13577990eacb7417e4f9bcff277e
|
| 3 |
+
size 4169982088
|
checkpoint-epoch-8/config.json
ADDED
|
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_class_name": "SD3Transformer2DModel",
|
| 3 |
+
"_diffusers_version": "0.37.0",
|
| 4 |
+
"_name_or_path": "stabilityai/stable-diffusion-3-medium-diffusers",
|
| 5 |
+
"attention_head_dim": 64,
|
| 6 |
+
"caption_projection_dim": 1536,
|
| 7 |
+
"dual_attention_layers": [],
|
| 8 |
+
"in_channels": 16,
|
| 9 |
+
"joint_attention_dim": 4096,
|
| 10 |
+
"num_attention_heads": 24,
|
| 11 |
+
"num_layers": 24,
|
| 12 |
+
"out_channels": 16,
|
| 13 |
+
"patch_size": 2,
|
| 14 |
+
"pooled_projection_dim": 2048,
|
| 15 |
+
"pos_embed_max_size": 192,
|
| 16 |
+
"qk_norm": null,
|
| 17 |
+
"sample_size": 128
|
| 18 |
+
}
|
checkpoint-epoch-8/diffusion_pytorch_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d06cb435a78131cd5e492bd1dbc8ec51ac4199e93047527c75a2b31ace668d2f
|
| 3 |
+
size 4169982088
|
model_index.json
ADDED
|
@@ -0,0 +1,49 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_class_name": "StableDiffusion3Pipeline",
|
| 3 |
+
"_diffusers_version": "0.37.0",
|
| 4 |
+
"_name_or_path": "stabilityai/stable-diffusion-3-medium-diffusers",
|
| 5 |
+
"feature_extractor": [
|
| 6 |
+
null,
|
| 7 |
+
null
|
| 8 |
+
],
|
| 9 |
+
"image_encoder": [
|
| 10 |
+
null,
|
| 11 |
+
null
|
| 12 |
+
],
|
| 13 |
+
"scheduler": [
|
| 14 |
+
"diffusers",
|
| 15 |
+
"FlowMatchEulerDiscreteScheduler"
|
| 16 |
+
],
|
| 17 |
+
"text_encoder": [
|
| 18 |
+
"transformers",
|
| 19 |
+
"CLIPTextModelWithProjection"
|
| 20 |
+
],
|
| 21 |
+
"text_encoder_2": [
|
| 22 |
+
"transformers",
|
| 23 |
+
"CLIPTextModelWithProjection"
|
| 24 |
+
],
|
| 25 |
+
"text_encoder_3": [
|
| 26 |
+
"transformers",
|
| 27 |
+
"T5EncoderModel"
|
| 28 |
+
],
|
| 29 |
+
"tokenizer": [
|
| 30 |
+
"transformers",
|
| 31 |
+
"CLIPTokenizer"
|
| 32 |
+
],
|
| 33 |
+
"tokenizer_2": [
|
| 34 |
+
"transformers",
|
| 35 |
+
"CLIPTokenizer"
|
| 36 |
+
],
|
| 37 |
+
"tokenizer_3": [
|
| 38 |
+
"transformers",
|
| 39 |
+
"T5Tokenizer"
|
| 40 |
+
],
|
| 41 |
+
"transformer": [
|
| 42 |
+
"diffusers",
|
| 43 |
+
"SD3Transformer2DModel"
|
| 44 |
+
],
|
| 45 |
+
"vae": [
|
| 46 |
+
"diffusers",
|
| 47 |
+
"AutoencoderKL"
|
| 48 |
+
]
|
| 49 |
+
}
|
scheduler/scheduler_config.json
ADDED
|
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_class_name": "FlowMatchEulerDiscreteScheduler",
|
| 3 |
+
"_diffusers_version": "0.37.0",
|
| 4 |
+
"base_image_seq_len": 256,
|
| 5 |
+
"base_shift": 0.5,
|
| 6 |
+
"invert_sigmas": false,
|
| 7 |
+
"max_image_seq_len": 4096,
|
| 8 |
+
"max_shift": 1.15,
|
| 9 |
+
"num_train_timesteps": 1000,
|
| 10 |
+
"shift": 3.0,
|
| 11 |
+
"shift_terminal": null,
|
| 12 |
+
"stochastic_sampling": false,
|
| 13 |
+
"time_shift_type": "exponential",
|
| 14 |
+
"use_beta_sigmas": false,
|
| 15 |
+
"use_dynamic_shifting": false,
|
| 16 |
+
"use_exponential_sigmas": false,
|
| 17 |
+
"use_karras_sigmas": false
|
| 18 |
+
}
|
text_encoder/config.json
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"CLIPTextModelWithProjection"
|
| 4 |
+
],
|
| 5 |
+
"attention_dropout": 0.0,
|
| 6 |
+
"bos_token_id": 0,
|
| 7 |
+
"dropout": 0.0,
|
| 8 |
+
"dtype": "float16",
|
| 9 |
+
"eos_token_id": 2,
|
| 10 |
+
"hidden_act": "quick_gelu",
|
| 11 |
+
"hidden_size": 768,
|
| 12 |
+
"initializer_factor": 1.0,
|
| 13 |
+
"initializer_range": 0.02,
|
| 14 |
+
"intermediate_size": 3072,
|
| 15 |
+
"layer_norm_eps": 1e-05,
|
| 16 |
+
"max_position_embeddings": 77,
|
| 17 |
+
"model_type": "clip_text_model",
|
| 18 |
+
"num_attention_heads": 12,
|
| 19 |
+
"num_hidden_layers": 12,
|
| 20 |
+
"pad_token_id": 1,
|
| 21 |
+
"projection_dim": 768,
|
| 22 |
+
"transformers_version": "5.3.0",
|
| 23 |
+
"vocab_size": 49408
|
| 24 |
+
}
|
text_encoder/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:71e183d11db0c6b6282a4d9e0abb74125edc8692393e89ed8ee5571005f35cb1
|
| 3 |
+
size 247323896
|
text_encoder_2/config.json
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"CLIPTextModelWithProjection"
|
| 4 |
+
],
|
| 5 |
+
"attention_dropout": 0.0,
|
| 6 |
+
"bos_token_id": 0,
|
| 7 |
+
"dropout": 0.0,
|
| 8 |
+
"dtype": "float16",
|
| 9 |
+
"eos_token_id": 2,
|
| 10 |
+
"hidden_act": "gelu",
|
| 11 |
+
"hidden_size": 1280,
|
| 12 |
+
"initializer_factor": 1.0,
|
| 13 |
+
"initializer_range": 0.02,
|
| 14 |
+
"intermediate_size": 5120,
|
| 15 |
+
"layer_norm_eps": 1e-05,
|
| 16 |
+
"max_position_embeddings": 77,
|
| 17 |
+
"model_type": "clip_text_model",
|
| 18 |
+
"num_attention_heads": 20,
|
| 19 |
+
"num_hidden_layers": 32,
|
| 20 |
+
"pad_token_id": 1,
|
| 21 |
+
"projection_dim": 1280,
|
| 22 |
+
"transformers_version": "5.3.0",
|
| 23 |
+
"vocab_size": 49408
|
| 24 |
+
}
|
text_encoder_2/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ec310df2af79c318e24d20511b601a591ca8cd4f1fce1d8dff822a356bcdb1f4
|
| 3 |
+
size 1389382176
|
text_encoder_3/config.json
ADDED
|
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"T5EncoderModel"
|
| 4 |
+
],
|
| 5 |
+
"classifier_dropout": 0.0,
|
| 6 |
+
"d_ff": 10240,
|
| 7 |
+
"d_kv": 64,
|
| 8 |
+
"d_model": 4096,
|
| 9 |
+
"decoder_start_token_id": 0,
|
| 10 |
+
"dense_act_fn": "gelu_new",
|
| 11 |
+
"dropout_rate": 0.1,
|
| 12 |
+
"dtype": "float16",
|
| 13 |
+
"eos_token_id": 1,
|
| 14 |
+
"feed_forward_proj": "gated-gelu",
|
| 15 |
+
"initializer_factor": 1.0,
|
| 16 |
+
"is_decoder": false,
|
| 17 |
+
"is_encoder_decoder": false,
|
| 18 |
+
"is_gated_act": true,
|
| 19 |
+
"layer_norm_epsilon": 1e-06,
|
| 20 |
+
"model_type": "t5",
|
| 21 |
+
"num_decoder_layers": 24,
|
| 22 |
+
"num_heads": 64,
|
| 23 |
+
"num_layers": 24,
|
| 24 |
+
"output_past": true,
|
| 25 |
+
"pad_token_id": 0,
|
| 26 |
+
"relative_attention_max_distance": 128,
|
| 27 |
+
"relative_attention_num_buckets": 32,
|
| 28 |
+
"scale_decoder_outputs": false,
|
| 29 |
+
"tie_word_embeddings": true,
|
| 30 |
+
"transformers_version": "5.3.0",
|
| 31 |
+
"use_cache": false,
|
| 32 |
+
"vocab_size": 32128
|
| 33 |
+
}
|
text_encoder_3/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d0921715ce20abcba4d79360b9547e7ce57368d9182860cace910d812c50666f
|
| 3 |
+
size 11537914408
|
tokenizer/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer/tokenizer_config.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_prefix_space": false,
|
| 3 |
+
"backend": "tokenizers",
|
| 4 |
+
"bos_token": "<|startoftext|>",
|
| 5 |
+
"clean_up_tokenization_spaces": true,
|
| 6 |
+
"do_lower_case": true,
|
| 7 |
+
"eos_token": "<|endoftext|>",
|
| 8 |
+
"errors": "replace",
|
| 9 |
+
"is_local": true,
|
| 10 |
+
"model_max_length": 77,
|
| 11 |
+
"pad_token": "<|endoftext|>",
|
| 12 |
+
"tokenizer_class": "CLIPTokenizer",
|
| 13 |
+
"unk_token": "<|endoftext|>"
|
| 14 |
+
}
|
tokenizer_2/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_2/tokenizer_config.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_prefix_space": false,
|
| 3 |
+
"backend": "tokenizers",
|
| 4 |
+
"bos_token": "<|startoftext|>",
|
| 5 |
+
"clean_up_tokenization_spaces": true,
|
| 6 |
+
"do_lower_case": true,
|
| 7 |
+
"eos_token": "<|endoftext|>",
|
| 8 |
+
"errors": "replace",
|
| 9 |
+
"is_local": true,
|
| 10 |
+
"model_max_length": 77,
|
| 11 |
+
"pad_token": "!",
|
| 12 |
+
"tokenizer_class": "CLIPTokenizer",
|
| 13 |
+
"unk_token": "<|endoftext|>"
|
| 14 |
+
}
|
tokenizer_3/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_3/tokenizer_config.json
ADDED
|
@@ -0,0 +1,116 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_prefix_space": true,
|
| 3 |
+
"backend": "tokenizers",
|
| 4 |
+
"clean_up_tokenization_spaces": true,
|
| 5 |
+
"eos_token": "</s>",
|
| 6 |
+
"extra_ids": 100,
|
| 7 |
+
"extra_special_tokens": [
|
| 8 |
+
"<extra_id_0>",
|
| 9 |
+
"<extra_id_1>",
|
| 10 |
+
"<extra_id_2>",
|
| 11 |
+
"<extra_id_3>",
|
| 12 |
+
"<extra_id_4>",
|
| 13 |
+
"<extra_id_5>",
|
| 14 |
+
"<extra_id_6>",
|
| 15 |
+
"<extra_id_7>",
|
| 16 |
+
"<extra_id_8>",
|
| 17 |
+
"<extra_id_9>",
|
| 18 |
+
"<extra_id_10>",
|
| 19 |
+
"<extra_id_11>",
|
| 20 |
+
"<extra_id_12>",
|
| 21 |
+
"<extra_id_13>",
|
| 22 |
+
"<extra_id_14>",
|
| 23 |
+
"<extra_id_15>",
|
| 24 |
+
"<extra_id_16>",
|
| 25 |
+
"<extra_id_17>",
|
| 26 |
+
"<extra_id_18>",
|
| 27 |
+
"<extra_id_19>",
|
| 28 |
+
"<extra_id_20>",
|
| 29 |
+
"<extra_id_21>",
|
| 30 |
+
"<extra_id_22>",
|
| 31 |
+
"<extra_id_23>",
|
| 32 |
+
"<extra_id_24>",
|
| 33 |
+
"<extra_id_25>",
|
| 34 |
+
"<extra_id_26>",
|
| 35 |
+
"<extra_id_27>",
|
| 36 |
+
"<extra_id_28>",
|
| 37 |
+
"<extra_id_29>",
|
| 38 |
+
"<extra_id_30>",
|
| 39 |
+
"<extra_id_31>",
|
| 40 |
+
"<extra_id_32>",
|
| 41 |
+
"<extra_id_33>",
|
| 42 |
+
"<extra_id_34>",
|
| 43 |
+
"<extra_id_35>",
|
| 44 |
+
"<extra_id_36>",
|
| 45 |
+
"<extra_id_37>",
|
| 46 |
+
"<extra_id_38>",
|
| 47 |
+
"<extra_id_39>",
|
| 48 |
+
"<extra_id_40>",
|
| 49 |
+
"<extra_id_41>",
|
| 50 |
+
"<extra_id_42>",
|
| 51 |
+
"<extra_id_43>",
|
| 52 |
+
"<extra_id_44>",
|
| 53 |
+
"<extra_id_45>",
|
| 54 |
+
"<extra_id_46>",
|
| 55 |
+
"<extra_id_47>",
|
| 56 |
+
"<extra_id_48>",
|
| 57 |
+
"<extra_id_49>",
|
| 58 |
+
"<extra_id_50>",
|
| 59 |
+
"<extra_id_51>",
|
| 60 |
+
"<extra_id_52>",
|
| 61 |
+
"<extra_id_53>",
|
| 62 |
+
"<extra_id_54>",
|
| 63 |
+
"<extra_id_55>",
|
| 64 |
+
"<extra_id_56>",
|
| 65 |
+
"<extra_id_57>",
|
| 66 |
+
"<extra_id_58>",
|
| 67 |
+
"<extra_id_59>",
|
| 68 |
+
"<extra_id_60>",
|
| 69 |
+
"<extra_id_61>",
|
| 70 |
+
"<extra_id_62>",
|
| 71 |
+
"<extra_id_63>",
|
| 72 |
+
"<extra_id_64>",
|
| 73 |
+
"<extra_id_65>",
|
| 74 |
+
"<extra_id_66>",
|
| 75 |
+
"<extra_id_67>",
|
| 76 |
+
"<extra_id_68>",
|
| 77 |
+
"<extra_id_69>",
|
| 78 |
+
"<extra_id_70>",
|
| 79 |
+
"<extra_id_71>",
|
| 80 |
+
"<extra_id_72>",
|
| 81 |
+
"<extra_id_73>",
|
| 82 |
+
"<extra_id_74>",
|
| 83 |
+
"<extra_id_75>",
|
| 84 |
+
"<extra_id_76>",
|
| 85 |
+
"<extra_id_77>",
|
| 86 |
+
"<extra_id_78>",
|
| 87 |
+
"<extra_id_79>",
|
| 88 |
+
"<extra_id_80>",
|
| 89 |
+
"<extra_id_81>",
|
| 90 |
+
"<extra_id_82>",
|
| 91 |
+
"<extra_id_83>",
|
| 92 |
+
"<extra_id_84>",
|
| 93 |
+
"<extra_id_85>",
|
| 94 |
+
"<extra_id_86>",
|
| 95 |
+
"<extra_id_87>",
|
| 96 |
+
"<extra_id_88>",
|
| 97 |
+
"<extra_id_89>",
|
| 98 |
+
"<extra_id_90>",
|
| 99 |
+
"<extra_id_91>",
|
| 100 |
+
"<extra_id_92>",
|
| 101 |
+
"<extra_id_93>",
|
| 102 |
+
"<extra_id_94>",
|
| 103 |
+
"<extra_id_95>",
|
| 104 |
+
"<extra_id_96>",
|
| 105 |
+
"<extra_id_97>",
|
| 106 |
+
"<extra_id_98>",
|
| 107 |
+
"<extra_id_99>"
|
| 108 |
+
],
|
| 109 |
+
"is_local": true,
|
| 110 |
+
"legacy": true,
|
| 111 |
+
"model_max_length": 512,
|
| 112 |
+
"pad_token": "<pad>",
|
| 113 |
+
"sp_model_kwargs": {},
|
| 114 |
+
"tokenizer_class": "T5Tokenizer",
|
| 115 |
+
"unk_token": "<unk>"
|
| 116 |
+
}
|
transformer/config.json
ADDED
|
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_class_name": "SD3Transformer2DModel",
|
| 3 |
+
"_diffusers_version": "0.37.0",
|
| 4 |
+
"_name_or_path": "stabilityai/stable-diffusion-3-medium-diffusers",
|
| 5 |
+
"attention_head_dim": 64,
|
| 6 |
+
"caption_projection_dim": 1536,
|
| 7 |
+
"dual_attention_layers": [],
|
| 8 |
+
"in_channels": 16,
|
| 9 |
+
"joint_attention_dim": 4096,
|
| 10 |
+
"num_attention_heads": 24,
|
| 11 |
+
"num_layers": 24,
|
| 12 |
+
"out_channels": 16,
|
| 13 |
+
"patch_size": 2,
|
| 14 |
+
"pooled_projection_dim": 2048,
|
| 15 |
+
"pos_embed_max_size": 192,
|
| 16 |
+
"qk_norm": null,
|
| 17 |
+
"sample_size": 128
|
| 18 |
+
}
|
transformer/diffusion_pytorch_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:80c25f3c7c73e3fac20822fed0c19983defff965c3805582b2dd4c39cb1d4e48
|
| 3 |
+
size 4169982088
|
vae/config.json
ADDED
|
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_class_name": "AutoencoderKL",
|
| 3 |
+
"_diffusers_version": "0.37.0",
|
| 4 |
+
"_name_or_path": "/root/.cache/huggingface/hub/models--stabilityai--stable-diffusion-3-medium-diffusers/snapshots/ea42f8cef0f178587cf766dc8129abd379c90671/vae",
|
| 5 |
+
"act_fn": "silu",
|
| 6 |
+
"block_out_channels": [
|
| 7 |
+
128,
|
| 8 |
+
256,
|
| 9 |
+
512,
|
| 10 |
+
512
|
| 11 |
+
],
|
| 12 |
+
"down_block_types": [
|
| 13 |
+
"DownEncoderBlock2D",
|
| 14 |
+
"DownEncoderBlock2D",
|
| 15 |
+
"DownEncoderBlock2D",
|
| 16 |
+
"DownEncoderBlock2D"
|
| 17 |
+
],
|
| 18 |
+
"force_upcast": true,
|
| 19 |
+
"in_channels": 3,
|
| 20 |
+
"latent_channels": 16,
|
| 21 |
+
"latents_mean": null,
|
| 22 |
+
"latents_std": null,
|
| 23 |
+
"layers_per_block": 2,
|
| 24 |
+
"mid_block_add_attention": true,
|
| 25 |
+
"norm_num_groups": 32,
|
| 26 |
+
"out_channels": 3,
|
| 27 |
+
"sample_size": 1024,
|
| 28 |
+
"scaling_factor": 1.5305,
|
| 29 |
+
"shift_factor": 0.0609,
|
| 30 |
+
"up_block_types": [
|
| 31 |
+
"UpDecoderBlock2D",
|
| 32 |
+
"UpDecoderBlock2D",
|
| 33 |
+
"UpDecoderBlock2D",
|
| 34 |
+
"UpDecoderBlock2D"
|
| 35 |
+
],
|
| 36 |
+
"use_post_quant_conv": false,
|
| 37 |
+
"use_quant_conv": false
|
| 38 |
+
}
|
vae/diffusion_pytorch_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f9b67a279283625caee39d61eacb5324243848477b4eb535355eaaa8423d4e09
|
| 3 |
+
size 167666654
|