diff --git a/.gitattributes b/.gitattributes index a6344aac8c09253b3b630fb776ae94478aa0275b..f32a7b4800cf1f2d79120a18b0c6805ac9f9feee 100644 --- a/.gitattributes +++ b/.gitattributes @@ -33,3 +33,14 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text *.zip filter=lfs diff=lfs merge=lfs -text *.zst filter=lfs diff=lfs merge=lfs -text *tfevents* filter=lfs diff=lfs merge=lfs -text +models--DeepFloyd--IF-II-L-v1.0/blobs/4c917ec4f9e64a1b45156c7fb81999c0f6e6379ec79cc3a576ad144ffb5bf572 filter=lfs diff=lfs merge=lfs -text +models--DeepFloyd--IF-II-L-v1.0/blobs/c8d32cddbf4701f91fd67b2ba62bb40a0a17a7c427fe1a3ad55a9cc348e9376a filter=lfs diff=lfs merge=lfs -text +models--showlab--show-1-base/blobs/0bff1eff39c04e7d313e5581cf0d8394d25ac277168062b9e43133c399ba058c filter=lfs diff=lfs merge=lfs -text +models--showlab--show-1-base/blobs/73e8b75789dd964d3a51e96a9753359e77cbbc190b1cb83bcf5775ea6374232b filter=lfs diff=lfs merge=lfs -text +models--showlab--show-1-base/blobs/d2bb149576a5164f6d46963535d5bfd8e169b6d22bf2e31a6126dbc25be5aaac filter=lfs diff=lfs merge=lfs -text +models--showlab--show-1-interpolation/blobs/08a84d346b5f7c04036082364c0b4e74a754e62ea546de5bcf9949206b0b9719 filter=lfs diff=lfs merge=lfs -text +models--showlab--show-1-interpolation/blobs/abdda89634be180596b30a7c369ab01ea09a9b18a603f9581f83ac8a4b85fbe2 filter=lfs diff=lfs merge=lfs -text +models--showlab--show-1-interpolation/blobs/b0d673e9c15f0d46df802afc6dd60ac147e5363ed5e17c7b58211e3f793acc2a filter=lfs diff=lfs merge=lfs -text +models--showlab--show-1-sr1/blobs/08a84d346b5f7c04036082364c0b4e74a754e62ea546de5bcf9949206b0b9719 filter=lfs diff=lfs merge=lfs -text +models--showlab--show-1-sr1/blobs/1480c708cae9866e3e6f2255b0b2a4a94dbab09f2b876ef8eb46c1f01d6479b7 filter=lfs diff=lfs merge=lfs -text +models--showlab--show-1-sr1/blobs/abdda89634be180596b30a7c369ab01ea09a9b18a603f9581f83ac8a4b85fbe2 filter=lfs diff=lfs merge=lfs -text diff --git a/models--DeepFloyd--IF-II-L-v1.0/blobs/03be03e1b4ad554235f41609e8810336b29a3d91 b/models--DeepFloyd--IF-II-L-v1.0/blobs/03be03e1b4ad554235f41609e8810336b29a3d91 new file mode 100644 index 0000000000000000000000000000000000000000..03be03e1b4ad554235f41609e8810336b29a3d91 --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/blobs/03be03e1b4ad554235f41609e8810336b29a3d91 @@ -0,0 +1,16 @@ +{ + "_class_name": "DDPMScheduler", + "_diffusers_version": "0.15.0.dev0", + "beta_end": 0.02, + "beta_schedule": "squaredcos_cap_v2", + "beta_start": 0.0001, + "clip_sample": true, + "clip_sample_range": 1.0, + "dynamic_thresholding_ratio": 0.995, + "num_train_timesteps": 1000, + "prediction_type": "epsilon", + "sample_max_value": 1.0, + "thresholding": false, + "trained_betas": null, + "variance_type": "fixed_small" +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/blobs/0d9d33b883843d1b370da781f3943051067e1b2c b/models--DeepFloyd--IF-II-L-v1.0/blobs/0d9d33b883843d1b370da781f3943051067e1b2c new file mode 100644 index 0000000000000000000000000000000000000000..0d9d33b883843d1b370da781f3943051067e1b2c --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/blobs/0d9d33b883843d1b370da781f3943051067e1b2c @@ -0,0 +1,28 @@ +{ + "crop_size": { + "height": 224, + "width": 224 + }, + "do_center_crop": true, + "do_convert_rgb": true, + "do_normalize": true, + "do_rescale": true, + "do_resize": true, + "feature_extractor_type": "CLIPFeatureExtractor", + "image_mean": [ + 0.48145466, + 0.4578275, + 0.40821073 + ], + "image_processor_type": "CLIPImageProcessor", + "image_std": [ + 0.26862954, + 0.26130258, + 0.27577711 + ], + "resample": 3, + "rescale_factor": 0.00392156862745098, + "size": { + "shortest_edge": 224 + } +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/blobs/2c19eb6e3b583f52d34b903b5978d3d30b6b7682 b/models--DeepFloyd--IF-II-L-v1.0/blobs/2c19eb6e3b583f52d34b903b5978d3d30b6b7682 new file mode 100644 index 0000000000000000000000000000000000000000..2c19eb6e3b583f52d34b903b5978d3d30b6b7682 --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/blobs/2c19eb6e3b583f52d34b903b5978d3d30b6b7682 @@ -0,0 +1,107 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": "", + "pad_token": "", + "unk_token": "" +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/blobs/2d626e9ba5506202c79d3b533af4ad4812f4bd9f b/models--DeepFloyd--IF-II-L-v1.0/blobs/2d626e9ba5506202c79d3b533af4ad4812f4bd9f new file mode 100644 index 0000000000000000000000000000000000000000..2d626e9ba5506202c79d3b533af4ad4812f4bd9f --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/blobs/2d626e9ba5506202c79d3b533af4ad4812f4bd9f @@ -0,0 +1,4 @@ +{ + "_class_name": "IFWatermark", + "_diffusers_version": "0.16.0.dev0" +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/blobs/4319cd21d9aa722ea7df1027be09a016711689ca b/models--DeepFloyd--IF-II-L-v1.0/blobs/4319cd21d9aa722ea7df1027be09a016711689ca new file mode 100644 index 0000000000000000000000000000000000000000..4319cd21d9aa722ea7df1027be09a016711689ca --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/blobs/4319cd21d9aa722ea7df1027be09a016711689ca @@ -0,0 +1,183 @@ +{ + "_commit_hash": "8d052a0f05efbaefbc9e8786ba291cfdf93e5bff", + "_name_or_path": "clip-vit-large-patch14/", + "architectures": [ + "IFSafetyChecker" + ], + "initializer_factor": 1.0, + "logit_scale_init_value": 2.6592, + "model_type": "clip", + "projection_dim": 768, + "text_config": { + "_name_or_path": "", + "add_cross_attention": false, + "architectures": null, + "attention_dropout": 0.0, + "bad_words_ids": null, + "begin_suppress_tokens": null, + "bos_token_id": 0, + "chunk_size_feed_forward": 0, + "cross_attention_hidden_size": null, + "decoder_start_token_id": null, + "diversity_penalty": 0.0, + "do_sample": false, + "dropout": 0.0, + "early_stopping": false, + "encoder_no_repeat_ngram_size": 0, + "eos_token_id": 2, + "exponential_decay_length_penalty": null, + "finetuning_task": null, + "forced_bos_token_id": null, + "forced_eos_token_id": null, + "hidden_act": "quick_gelu", + "hidden_size": 768, + "id2label": { + "0": "LABEL_0", + "1": "LABEL_1" + }, + "initializer_factor": 1.0, + "initializer_range": 0.02, + "intermediate_size": 3072, + "is_decoder": false, + "is_encoder_decoder": false, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1 + }, + "layer_norm_eps": 1e-05, + "length_penalty": 1.0, + "max_length": 20, + "max_position_embeddings": 77, + "min_length": 0, + "model_type": "clip_text_model", + "no_repeat_ngram_size": 0, + "num_attention_heads": 12, + "num_beam_groups": 1, + "num_beams": 1, + "num_hidden_layers": 12, + "num_return_sequences": 1, + "output_attentions": false, + "output_hidden_states": false, + "output_scores": false, + "pad_token_id": 1, + "prefix": null, + "problem_type": null, + "projection_dim": 768, + "pruned_heads": {}, + "remove_invalid_values": false, + "repetition_penalty": 1.0, + "return_dict": true, + "return_dict_in_generate": false, + "sep_token_id": null, + "suppress_tokens": null, + "task_specific_params": null, + "temperature": 1.0, + "tf_legacy_loss": false, + "tie_encoder_decoder": false, + "tie_word_embeddings": true, + "tokenizer_class": null, + "top_k": 50, + "top_p": 1.0, + "torch_dtype": null, + "torchscript": false, + "transformers_version": "4.26.0", + "typical_p": 1.0, + "use_bfloat16": false, + "vocab_size": 49408 + }, + "text_config_dict": { + "hidden_size": 768, + "intermediate_size": 3072, + "num_attention_heads": 12, + "num_hidden_layers": 12, + "projection_dim": 768 + }, + "torch_dtype": "float32", + "transformers_version": null, + "vision_config": { + "_name_or_path": "", + "add_cross_attention": false, + "architectures": null, + "attention_dropout": 0.0, + "bad_words_ids": null, + "begin_suppress_tokens": null, + "bos_token_id": null, + "chunk_size_feed_forward": 0, + "cross_attention_hidden_size": null, + "decoder_start_token_id": null, + "diversity_penalty": 0.0, + "do_sample": false, + "dropout": 0.0, + "early_stopping": false, + "encoder_no_repeat_ngram_size": 0, + "eos_token_id": null, + "exponential_decay_length_penalty": null, + "finetuning_task": null, + "forced_bos_token_id": null, + "forced_eos_token_id": null, + "hidden_act": "quick_gelu", + "hidden_size": 1024, + "id2label": { + "0": "LABEL_0", + "1": "LABEL_1" + }, + "image_size": 224, + "initializer_factor": 1.0, + "initializer_range": 0.02, + "intermediate_size": 4096, + "is_decoder": false, + "is_encoder_decoder": false, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1 + }, + "layer_norm_eps": 1e-05, + "length_penalty": 1.0, + "max_length": 20, + "min_length": 0, + "model_type": "clip_vision_model", + "no_repeat_ngram_size": 0, + "num_attention_heads": 16, + "num_beam_groups": 1, + "num_beams": 1, + "num_channels": 3, + "num_hidden_layers": 24, + "num_return_sequences": 1, + "output_attentions": false, + "output_hidden_states": false, + "output_scores": false, + "pad_token_id": null, + "patch_size": 14, + "prefix": null, + "problem_type": null, + "projection_dim": 768, + "pruned_heads": {}, + "remove_invalid_values": false, + "repetition_penalty": 1.0, + "return_dict": true, + "return_dict_in_generate": false, + "sep_token_id": null, + "suppress_tokens": null, + "task_specific_params": null, + "temperature": 1.0, + "tf_legacy_loss": false, + "tie_encoder_decoder": false, + "tie_word_embeddings": true, + "tokenizer_class": null, + "top_k": 50, + "top_p": 1.0, + "torch_dtype": null, + "torchscript": false, + "transformers_version": "4.26.0", + "typical_p": 1.0, + "use_bfloat16": false + }, + "vision_config_dict": { + "hidden_size": 1024, + "intermediate_size": 4096, + "num_attention_heads": 16, + "num_hidden_layers": 24, + "patch_size": 14, + "projection_dim": 768 + } +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/blobs/4c917ec4f9e64a1b45156c7fb81999c0f6e6379ec79cc3a576ad144ffb5bf572 b/models--DeepFloyd--IF-II-L-v1.0/blobs/4c917ec4f9e64a1b45156c7fb81999c0f6e6379ec79cc3a576ad144ffb5bf572 new file mode 100644 index 0000000000000000000000000000000000000000..e3a8219be0678a00ebdba4d21dbaab604bac3d9d --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/blobs/4c917ec4f9e64a1b45156c7fb81999c0f6e6379ec79cc3a576ad144ffb5bf572 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c917ec4f9e64a1b45156c7fb81999c0f6e6379ec79cc3a576ad144ffb5bf572 +size 607990743 diff --git a/models--DeepFloyd--IF-II-L-v1.0/blobs/850651eb7420053e92fba2097ce990f4c03b4e18 b/models--DeepFloyd--IF-II-L-v1.0/blobs/850651eb7420053e92fba2097ce990f4c03b4e18 new file mode 100644 index 0000000000000000000000000000000000000000..850651eb7420053e92fba2097ce990f4c03b4e18 --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/blobs/850651eb7420053e92fba2097ce990f4c03b4e18 @@ -0,0 +1,38 @@ +{ + "_class_name": "IFSuperResolutionPipeline", + "_diffusers_version": "0.15.0.dev0", + "_ignore_files": ["pytorch_model.bin"], + "feature_extractor": [ + "transformers", + "CLIPImageProcessor" + ], + "image_noising_scheduler": [ + "diffusers", + "DDPMScheduler" + ], + "requires_safety_checker": true, + "safety_checker": [ + "deepfloyd_if", + "IFSafetyChecker" + ], + "scheduler": [ + "diffusers", + "DDPMScheduler" + ], + "text_encoder": [ + "transformers", + "T5EncoderModel" + ], + "tokenizer": [ + "transformers", + "T5Tokenizer" + ], + "unet": [ + "diffusers", + "UNet2DConditionModel" + ], + "watermarker": [ + "deepfloyd_if", + "IFWatermarker" + ] +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/blobs/9c5f27589ad1b10a0216c4234265b26975b649e4 b/models--DeepFloyd--IF-II-L-v1.0/blobs/9c5f27589ad1b10a0216c4234265b26975b649e4 new file mode 100644 index 0000000000000000000000000000000000000000..9c5f27589ad1b10a0216c4234265b26975b649e4 --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/blobs/9c5f27589ad1b10a0216c4234265b26975b649e4 @@ -0,0 +1,112 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": "", + "extra_ids": 100, + "model_max_length": 512, + "pad_token": "", + "sp_model_kwargs": {}, + "special_tokens_map_file": "/home/patrick_huggingface_co/.cache/huggingface/hub/models--google--t5-v1_1-xxl/snapshots/3db67ab1af984cf10548a73467f0e5bca2aaaeb2/special_tokens_map.json", + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/blobs/ad38e4ca986f391bbd3641064650aab83122b701bdb143effc8e084425df6110 b/models--DeepFloyd--IF-II-L-v1.0/blobs/ad38e4ca986f391bbd3641064650aab83122b701bdb143effc8e084425df6110 new file mode 100644 index 0000000000000000000000000000000000000000..f19a43cc8a09e61d1563ac45444d1cb8d3e18ed8 Binary files /dev/null and b/models--DeepFloyd--IF-II-L-v1.0/blobs/ad38e4ca986f391bbd3641064650aab83122b701bdb143effc8e084425df6110 differ diff --git a/models--DeepFloyd--IF-II-L-v1.0/blobs/c8d32cddbf4701f91fd67b2ba62bb40a0a17a7c427fe1a3ad55a9cc348e9376a b/models--DeepFloyd--IF-II-L-v1.0/blobs/c8d32cddbf4701f91fd67b2ba62bb40a0a17a7c427fe1a3ad55a9cc348e9376a new file mode 100644 index 0000000000000000000000000000000000000000..53364f8100751cde57be569749af18130af727dd --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/blobs/c8d32cddbf4701f91fd67b2ba62bb40a0a17a7c427fe1a3ad55a9cc348e9376a @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8d32cddbf4701f91fd67b2ba62bb40a0a17a7c427fe1a3ad55a9cc348e9376a +size 2486658568 diff --git a/models--DeepFloyd--IF-II-L-v1.0/blobs/d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86 b/models--DeepFloyd--IF-II-L-v1.0/blobs/d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86 new file mode 100644 index 0000000000000000000000000000000000000000..4e28ff6ebdf584f5372d9de68867399142435d9a Binary files /dev/null and b/models--DeepFloyd--IF-II-L-v1.0/blobs/d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86 differ diff --git a/models--DeepFloyd--IF-II-L-v1.0/blobs/f5b4c08790b715c8e2b64ccaca1a38f46e1f5c3c b/models--DeepFloyd--IF-II-L-v1.0/blobs/f5b4c08790b715c8e2b64ccaca1a38f46e1f5c3c new file mode 100644 index 0000000000000000000000000000000000000000..f5b4c08790b715c8e2b64ccaca1a38f46e1f5c3c --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/blobs/f5b4c08790b715c8e2b64ccaca1a38f46e1f5c3c @@ -0,0 +1,17 @@ +{ + "_class_name": "DDPMScheduler", + "_diffusers_version": "0.15.0.dev0", + "beta_end": 0.02, + "beta_schedule": "squaredcos_cap_v2", + "beta_start": 0.0001, + "clip_sample": true, + "clip_sample_range": 1.0, + "dynamic_thresholding_ratio": 0.95, + "num_train_timesteps": 1000, + "prediction_type": "epsilon", + "sample_max_value": 1.0, + "thresholding": true, + "trained_betas": null, + "variance_type": "learned_range", + "lambda_min_clipped": -5.1 +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/blobs/f7eee79eb024df53fdbbecdb294ebe3877bf0b39 b/models--DeepFloyd--IF-II-L-v1.0/blobs/f7eee79eb024df53fdbbecdb294ebe3877bf0b39 new file mode 100644 index 0000000000000000000000000000000000000000..f7eee79eb024df53fdbbecdb294ebe3877bf0b39 --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/blobs/f7eee79eb024df53fdbbecdb294ebe3877bf0b39 @@ -0,0 +1,68 @@ +{ + "_class_name": "UNet2DConditionModel", + "_diffusers_version": "0.15.0.dev0", + "_name_or_path": "./", + "act_fn": "gelu", + "addition_embed_type": "text", + "attention_head_dim": 64, + "block_out_channels": [ + 160, + 320, + 640, + 960, + 1280 + ], + "center_input_sample": false, + "class_embed_type": "timestep", + "conv_in_kernel": 3, + "conv_out_kernel": 3, + "cross_attention_dim": 1280, + "cross_attention_norm": "group_norm", + "down_block_types": [ + "ResnetDownsampleBlock2D", + "ResnetDownsampleBlock2D", + "ResnetDownsampleBlock2D", + "SimpleCrossAttnDownBlock2D", + "SimpleCrossAttnDownBlock2D" + ], + "downsample_padding": 1, + "dual_cross_attention": false, + "encoder_hid_dim": 4096, + "encoder_hid_dim_type": "text_proj", + "flip_sin_to_cos": true, + "freq_shift": 0, + "in_channels": 6, + "layers_per_block": [ + 2, + 2, + 3, + 5, + 5 + ], + "mid_block_scale_factor": 1.4142271248762552, + "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", + "norm_eps": 1e-05, + "norm_num_groups": 32, + "num_class_embeds": null, + "only_cross_attention": false, + "out_channels": 6, + "projection_class_embeddings_input_dim": null, + "resnet_out_scale_factor": 1.4142271248762552, + "resnet_skip_time_act": true, + "resnet_time_scale_shift": "scale_shift", + "sample_size": 256, + "time_cond_proj_dim": null, + "time_embedding_act_fn": "gelu", + "time_embedding_dim": 1280, + "time_embedding_type": "positional", + "timestep_post_act": null, + "up_block_types": [ + "SimpleCrossAttnUpBlock2D", + "SimpleCrossAttnUpBlock2D", + "ResnetUpsampleBlock2D", + "ResnetUpsampleBlock2D", + "ResnetUpsampleBlock2D" + ], + "upcast_attention": false, + "use_linear_projection": false +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/refs/main b/models--DeepFloyd--IF-II-L-v1.0/refs/main new file mode 100644 index 0000000000000000000000000000000000000000..a185c10ff6199aef1c6d7203a5dca9012294aa50 --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/refs/main @@ -0,0 +1 @@ +609476ce702b2d94aff7d1f944dcc54d4f972901 \ No newline at end of file diff --git a/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/feature_extractor/preprocessor_config.json b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/feature_extractor/preprocessor_config.json new file mode 100644 index 0000000000000000000000000000000000000000..0d9d33b883843d1b370da781f3943051067e1b2c --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/feature_extractor/preprocessor_config.json @@ -0,0 +1,28 @@ +{ + "crop_size": { + "height": 224, + "width": 224 + }, + "do_center_crop": true, + "do_convert_rgb": true, + "do_normalize": true, + "do_rescale": true, + "do_resize": true, + "feature_extractor_type": "CLIPFeatureExtractor", + "image_mean": [ + 0.48145466, + 0.4578275, + 0.40821073 + ], + "image_processor_type": "CLIPImageProcessor", + "image_std": [ + 0.26862954, + 0.26130258, + 0.27577711 + ], + "resample": 3, + "rescale_factor": 0.00392156862745098, + "size": { + "shortest_edge": 224 + } +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/image_noising_scheduler/scheduler_config.json b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/image_noising_scheduler/scheduler_config.json new file mode 100644 index 0000000000000000000000000000000000000000..03be03e1b4ad554235f41609e8810336b29a3d91 --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/image_noising_scheduler/scheduler_config.json @@ -0,0 +1,16 @@ +{ + "_class_name": "DDPMScheduler", + "_diffusers_version": "0.15.0.dev0", + "beta_end": 0.02, + "beta_schedule": "squaredcos_cap_v2", + "beta_start": 0.0001, + "clip_sample": true, + "clip_sample_range": 1.0, + "dynamic_thresholding_ratio": 0.995, + "num_train_timesteps": 1000, + "prediction_type": "epsilon", + "sample_max_value": 1.0, + "thresholding": false, + "trained_betas": null, + "variance_type": "fixed_small" +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/model_index.json b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/model_index.json new file mode 100644 index 0000000000000000000000000000000000000000..850651eb7420053e92fba2097ce990f4c03b4e18 --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/model_index.json @@ -0,0 +1,38 @@ +{ + "_class_name": "IFSuperResolutionPipeline", + "_diffusers_version": "0.15.0.dev0", + "_ignore_files": ["pytorch_model.bin"], + "feature_extractor": [ + "transformers", + "CLIPImageProcessor" + ], + "image_noising_scheduler": [ + "diffusers", + "DDPMScheduler" + ], + "requires_safety_checker": true, + "safety_checker": [ + "deepfloyd_if", + "IFSafetyChecker" + ], + "scheduler": [ + "diffusers", + "DDPMScheduler" + ], + "text_encoder": [ + "transformers", + "T5EncoderModel" + ], + "tokenizer": [ + "transformers", + "T5Tokenizer" + ], + "unet": [ + "diffusers", + "UNet2DConditionModel" + ], + "watermarker": [ + "deepfloyd_if", + "IFWatermarker" + ] +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/safety_checker/config.json b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/safety_checker/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4319cd21d9aa722ea7df1027be09a016711689ca --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/safety_checker/config.json @@ -0,0 +1,183 @@ +{ + "_commit_hash": "8d052a0f05efbaefbc9e8786ba291cfdf93e5bff", + "_name_or_path": "clip-vit-large-patch14/", + "architectures": [ + "IFSafetyChecker" + ], + "initializer_factor": 1.0, + "logit_scale_init_value": 2.6592, + "model_type": "clip", + "projection_dim": 768, + "text_config": { + "_name_or_path": "", + "add_cross_attention": false, + "architectures": null, + "attention_dropout": 0.0, + "bad_words_ids": null, + "begin_suppress_tokens": null, + "bos_token_id": 0, + "chunk_size_feed_forward": 0, + "cross_attention_hidden_size": null, + "decoder_start_token_id": null, + "diversity_penalty": 0.0, + "do_sample": false, + "dropout": 0.0, + "early_stopping": false, + "encoder_no_repeat_ngram_size": 0, + "eos_token_id": 2, + "exponential_decay_length_penalty": null, + "finetuning_task": null, + "forced_bos_token_id": null, + "forced_eos_token_id": null, + "hidden_act": "quick_gelu", + "hidden_size": 768, + "id2label": { + "0": "LABEL_0", + "1": "LABEL_1" + }, + "initializer_factor": 1.0, + "initializer_range": 0.02, + "intermediate_size": 3072, + "is_decoder": false, + "is_encoder_decoder": false, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1 + }, + "layer_norm_eps": 1e-05, + "length_penalty": 1.0, + "max_length": 20, + "max_position_embeddings": 77, + "min_length": 0, + "model_type": "clip_text_model", + "no_repeat_ngram_size": 0, + "num_attention_heads": 12, + "num_beam_groups": 1, + "num_beams": 1, + "num_hidden_layers": 12, + "num_return_sequences": 1, + "output_attentions": false, + "output_hidden_states": false, + "output_scores": false, + "pad_token_id": 1, + "prefix": null, + "problem_type": null, + "projection_dim": 768, + "pruned_heads": {}, + "remove_invalid_values": false, + "repetition_penalty": 1.0, + "return_dict": true, + "return_dict_in_generate": false, + "sep_token_id": null, + "suppress_tokens": null, + "task_specific_params": null, + "temperature": 1.0, + "tf_legacy_loss": false, + "tie_encoder_decoder": false, + "tie_word_embeddings": true, + "tokenizer_class": null, + "top_k": 50, + "top_p": 1.0, + "torch_dtype": null, + "torchscript": false, + "transformers_version": "4.26.0", + "typical_p": 1.0, + "use_bfloat16": false, + "vocab_size": 49408 + }, + "text_config_dict": { + "hidden_size": 768, + "intermediate_size": 3072, + "num_attention_heads": 12, + "num_hidden_layers": 12, + "projection_dim": 768 + }, + "torch_dtype": "float32", + "transformers_version": null, + "vision_config": { + "_name_or_path": "", + "add_cross_attention": false, + "architectures": null, + "attention_dropout": 0.0, + "bad_words_ids": null, + "begin_suppress_tokens": null, + "bos_token_id": null, + "chunk_size_feed_forward": 0, + "cross_attention_hidden_size": null, + "decoder_start_token_id": null, + "diversity_penalty": 0.0, + "do_sample": false, + "dropout": 0.0, + "early_stopping": false, + "encoder_no_repeat_ngram_size": 0, + "eos_token_id": null, + "exponential_decay_length_penalty": null, + "finetuning_task": null, + "forced_bos_token_id": null, + "forced_eos_token_id": null, + "hidden_act": "quick_gelu", + "hidden_size": 1024, + "id2label": { + "0": "LABEL_0", + "1": "LABEL_1" + }, + "image_size": 224, + "initializer_factor": 1.0, + "initializer_range": 0.02, + "intermediate_size": 4096, + "is_decoder": false, + "is_encoder_decoder": false, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1 + }, + "layer_norm_eps": 1e-05, + "length_penalty": 1.0, + "max_length": 20, + "min_length": 0, + "model_type": "clip_vision_model", + "no_repeat_ngram_size": 0, + "num_attention_heads": 16, + "num_beam_groups": 1, + "num_beams": 1, + "num_channels": 3, + "num_hidden_layers": 24, + "num_return_sequences": 1, + "output_attentions": false, + "output_hidden_states": false, + "output_scores": false, + "pad_token_id": null, + "patch_size": 14, + "prefix": null, + "problem_type": null, + "projection_dim": 768, + "pruned_heads": {}, + "remove_invalid_values": false, + "repetition_penalty": 1.0, + "return_dict": true, + "return_dict_in_generate": false, + "sep_token_id": null, + "suppress_tokens": null, + "task_specific_params": null, + "temperature": 1.0, + "tf_legacy_loss": false, + "tie_encoder_decoder": false, + "tie_word_embeddings": true, + "tokenizer_class": null, + "top_k": 50, + "top_p": 1.0, + "torch_dtype": null, + "torchscript": false, + "transformers_version": "4.26.0", + "typical_p": 1.0, + "use_bfloat16": false + }, + "vision_config_dict": { + "hidden_size": 1024, + "intermediate_size": 4096, + "num_attention_heads": 16, + "num_hidden_layers": 24, + "patch_size": 14, + "projection_dim": 768 + } +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/safety_checker/model.fp16.safetensors b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/safety_checker/model.fp16.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e3a8219be0678a00ebdba4d21dbaab604bac3d9d --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/safety_checker/model.fp16.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c917ec4f9e64a1b45156c7fb81999c0f6e6379ec79cc3a576ad144ffb5bf572 +size 607990743 diff --git a/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/scheduler/scheduler_config.json b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/scheduler/scheduler_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f5b4c08790b715c8e2b64ccaca1a38f46e1f5c3c --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/scheduler/scheduler_config.json @@ -0,0 +1,17 @@ +{ + "_class_name": "DDPMScheduler", + "_diffusers_version": "0.15.0.dev0", + "beta_end": 0.02, + "beta_schedule": "squaredcos_cap_v2", + "beta_start": 0.0001, + "clip_sample": true, + "clip_sample_range": 1.0, + "dynamic_thresholding_ratio": 0.95, + "num_train_timesteps": 1000, + "prediction_type": "epsilon", + "sample_max_value": 1.0, + "thresholding": true, + "trained_betas": null, + "variance_type": "learned_range", + "lambda_min_clipped": -5.1 +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/tokenizer/special_tokens_map.json b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/tokenizer/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..2c19eb6e3b583f52d34b903b5978d3d30b6b7682 --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/tokenizer/special_tokens_map.json @@ -0,0 +1,107 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": "", + "pad_token": "", + "unk_token": "" +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/tokenizer/spiece.model b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/tokenizer/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..317a5ccbde45300f5d1d970d4d449af2108b147e --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/tokenizer/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86 +size 791656 diff --git a/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/tokenizer/tokenizer_config.json b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/tokenizer/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..9c5f27589ad1b10a0216c4234265b26975b649e4 --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/tokenizer/tokenizer_config.json @@ -0,0 +1,112 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": "", + "extra_ids": 100, + "model_max_length": 512, + "pad_token": "", + "sp_model_kwargs": {}, + "special_tokens_map_file": "/home/patrick_huggingface_co/.cache/huggingface/hub/models--google--t5-v1_1-xxl/snapshots/3db67ab1af984cf10548a73467f0e5bca2aaaeb2/special_tokens_map.json", + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/unet/config.json b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..f7eee79eb024df53fdbbecdb294ebe3877bf0b39 --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/unet/config.json @@ -0,0 +1,68 @@ +{ + "_class_name": "UNet2DConditionModel", + "_diffusers_version": "0.15.0.dev0", + "_name_or_path": "./", + "act_fn": "gelu", + "addition_embed_type": "text", + "attention_head_dim": 64, + "block_out_channels": [ + 160, + 320, + 640, + 960, + 1280 + ], + "center_input_sample": false, + "class_embed_type": "timestep", + "conv_in_kernel": 3, + "conv_out_kernel": 3, + "cross_attention_dim": 1280, + "cross_attention_norm": "group_norm", + "down_block_types": [ + "ResnetDownsampleBlock2D", + "ResnetDownsampleBlock2D", + "ResnetDownsampleBlock2D", + "SimpleCrossAttnDownBlock2D", + "SimpleCrossAttnDownBlock2D" + ], + "downsample_padding": 1, + "dual_cross_attention": false, + "encoder_hid_dim": 4096, + "encoder_hid_dim_type": "text_proj", + "flip_sin_to_cos": true, + "freq_shift": 0, + "in_channels": 6, + "layers_per_block": [ + 2, + 2, + 3, + 5, + 5 + ], + "mid_block_scale_factor": 1.4142271248762552, + "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", + "norm_eps": 1e-05, + "norm_num_groups": 32, + "num_class_embeds": null, + "only_cross_attention": false, + "out_channels": 6, + "projection_class_embeddings_input_dim": null, + "resnet_out_scale_factor": 1.4142271248762552, + "resnet_skip_time_act": true, + "resnet_time_scale_shift": "scale_shift", + "sample_size": 256, + "time_cond_proj_dim": null, + "time_embedding_act_fn": "gelu", + "time_embedding_dim": 1280, + "time_embedding_type": "positional", + "timestep_post_act": null, + "up_block_types": [ + "SimpleCrossAttnUpBlock2D", + "SimpleCrossAttnUpBlock2D", + "ResnetUpsampleBlock2D", + "ResnetUpsampleBlock2D", + "ResnetUpsampleBlock2D" + ], + "upcast_attention": false, + "use_linear_projection": false +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/unet/diffusion_pytorch_model.fp16.safetensors b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/unet/diffusion_pytorch_model.fp16.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..53364f8100751cde57be569749af18130af727dd --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/unet/diffusion_pytorch_model.fp16.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8d32cddbf4701f91fd67b2ba62bb40a0a17a7c427fe1a3ad55a9cc348e9376a +size 2486658568 diff --git a/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/watermarker/config.json b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/watermarker/config.json new file mode 100644 index 0000000000000000000000000000000000000000..2d626e9ba5506202c79d3b533af4ad4812f4bd9f --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/watermarker/config.json @@ -0,0 +1,4 @@ +{ + "_class_name": "IFWatermark", + "_diffusers_version": "0.16.0.dev0" +} diff --git a/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/watermarker/diffusion_pytorch_model.safetensors b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/watermarker/diffusion_pytorch_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9fe11b35cd7fff0673c9a5e363052a193f3157db --- /dev/null +++ b/models--DeepFloyd--IF-II-L-v1.0/snapshots/609476ce702b2d94aff7d1f944dcc54d4f972901/watermarker/diffusion_pytorch_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad38e4ca986f391bbd3641064650aab83122b701bdb143effc8e084425df6110 +size 15492 diff --git a/models--showlab--show-1-base/blobs/0bff1eff39c04e7d313e5581cf0d8394d25ac277168062b9e43133c399ba058c b/models--showlab--show-1-base/blobs/0bff1eff39c04e7d313e5581cf0d8394d25ac277168062b9e43133c399ba058c new file mode 100644 index 0000000000000000000000000000000000000000..342676fb33609e263399c76472b1afa8bb6a2f1d --- /dev/null +++ b/models--showlab--show-1-base/blobs/0bff1eff39c04e7d313e5581cf0d8394d25ac277168062b9e43133c399ba058c @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0bff1eff39c04e7d313e5581cf0d8394d25ac277168062b9e43133c399ba058c +size 9597033414 diff --git a/models--showlab--show-1-base/blobs/0d9d33b883843d1b370da781f3943051067e1b2c b/models--showlab--show-1-base/blobs/0d9d33b883843d1b370da781f3943051067e1b2c new file mode 100644 index 0000000000000000000000000000000000000000..0d9d33b883843d1b370da781f3943051067e1b2c --- /dev/null +++ b/models--showlab--show-1-base/blobs/0d9d33b883843d1b370da781f3943051067e1b2c @@ -0,0 +1,28 @@ +{ + "crop_size": { + "height": 224, + "width": 224 + }, + "do_center_crop": true, + "do_convert_rgb": true, + "do_normalize": true, + "do_rescale": true, + "do_resize": true, + "feature_extractor_type": "CLIPFeatureExtractor", + "image_mean": [ + 0.48145466, + 0.4578275, + 0.40821073 + ], + "image_processor_type": "CLIPImageProcessor", + "image_std": [ + 0.26862954, + 0.26130258, + 0.27577711 + ], + "resample": 3, + "rescale_factor": 0.00392156862745098, + "size": { + "shortest_edge": 224 + } +} diff --git a/models--showlab--show-1-base/blobs/1a33111ca9725d27f870b6797f3c9726654a0657 b/models--showlab--show-1-base/blobs/1a33111ca9725d27f870b6797f3c9726654a0657 new file mode 100644 index 0000000000000000000000000000000000000000..1a33111ca9725d27f870b6797f3c9726654a0657 --- /dev/null +++ b/models--showlab--show-1-base/blobs/1a33111ca9725d27f870b6797f3c9726654a0657 @@ -0,0 +1,227 @@ +{ + "metadata": { + "total_size": 19575627776 + }, + "weight_map": { + "encoder.block.0.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.11.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.embed_tokens.weight": "pytorch_model-00001-of-00002.bin", + "encoder.final_layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "shared.weight": "pytorch_model-00001-of-00002.bin" + } +} diff --git a/models--showlab--show-1-base/blobs/2c19eb6e3b583f52d34b903b5978d3d30b6b7682 b/models--showlab--show-1-base/blobs/2c19eb6e3b583f52d34b903b5978d3d30b6b7682 new file mode 100644 index 0000000000000000000000000000000000000000..2c19eb6e3b583f52d34b903b5978d3d30b6b7682 --- /dev/null +++ b/models--showlab--show-1-base/blobs/2c19eb6e3b583f52d34b903b5978d3d30b6b7682 @@ -0,0 +1,107 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": "", + "pad_token": "", + "unk_token": "" +} diff --git a/models--showlab--show-1-base/blobs/4f069cd0a20603c3e76d60f34704e65416f12f26 b/models--showlab--show-1-base/blobs/4f069cd0a20603c3e76d60f34704e65416f12f26 new file mode 100644 index 0000000000000000000000000000000000000000..4f069cd0a20603c3e76d60f34704e65416f12f26 --- /dev/null +++ b/models--showlab--show-1-base/blobs/4f069cd0a20603c3e76d60f34704e65416f12f26 @@ -0,0 +1,31 @@ +{ + "_name_or_path": "/mnt/nfs/HDD-30t/models/IF-I-L-v1.0/text_encoder", + "architectures": [ + "T5EncoderModel" + ], + "d_ff": 10240, + "d_kv": 64, + "d_model": 4096, + "decoder_start_token_id": 0, + "dense_act_fn": "gelu_new", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "gated-gelu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": true, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "num_decoder_layers": 24, + "num_heads": 64, + "num_layers": 24, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "tie_word_embeddings": false, + "torch_dtype": "float32", + "transformers_version": "4.29.1", + "use_cache": true, + "vocab_size": 32128 +} diff --git a/models--showlab--show-1-base/blobs/58f6290659e1c3ea9bc80a69742bbfbeac0fbde9 b/models--showlab--show-1-base/blobs/58f6290659e1c3ea9bc80a69742bbfbeac0fbde9 new file mode 100644 index 0000000000000000000000000000000000000000..58f6290659e1c3ea9bc80a69742bbfbeac0fbde9 --- /dev/null +++ b/models--showlab--show-1-base/blobs/58f6290659e1c3ea9bc80a69742bbfbeac0fbde9 @@ -0,0 +1,25 @@ +{ + "_class_name": "TextToVideoIFPipeline", + "_diffusers_version": "0.19.1", + "_name_or_path": "show-1-base", + "feature_extractor": [ + "transformers", + "CLIPImageProcessor" + ], + "scheduler": [ + "diffusers", + "DPMSolverMultistepScheduler" + ], + "text_encoder": [ + "transformers", + "T5EncoderModel" + ], + "tokenizer": [ + "transformers", + "T5Tokenizer" + ], + "unet": [ + "showone.models.unet_3d_condition", + "UNet3DConditionModel" + ] +} diff --git a/models--showlab--show-1-base/blobs/73e8b75789dd964d3a51e96a9753359e77cbbc190b1cb83bcf5775ea6374232b b/models--showlab--show-1-base/blobs/73e8b75789dd964d3a51e96a9753359e77cbbc190b1cb83bcf5775ea6374232b new file mode 100644 index 0000000000000000000000000000000000000000..b9a6ff8995176e8caf41be6a8d548a329ffa467a --- /dev/null +++ b/models--showlab--show-1-base/blobs/73e8b75789dd964d3a51e96a9753359e77cbbc190b1cb83bcf5775ea6374232b @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73e8b75789dd964d3a51e96a9753359e77cbbc190b1cb83bcf5775ea6374232b +size 3454235275 diff --git a/models--showlab--show-1-base/blobs/82964f12bd7ed4b7eae1a084e2b9e0297dac85c3 b/models--showlab--show-1-base/blobs/82964f12bd7ed4b7eae1a084e2b9e0297dac85c3 new file mode 100644 index 0000000000000000000000000000000000000000..82964f12bd7ed4b7eae1a084e2b9e0297dac85c3 --- /dev/null +++ b/models--showlab--show-1-base/blobs/82964f12bd7ed4b7eae1a084e2b9e0297dac85c3 @@ -0,0 +1,24 @@ +{ + "_class_name": "DPMSolverMultistepScheduler", + "_diffusers_version": "0.19.1", + "algorithm_type": "dpmsolver++", + "beta_end": 0.02, + "beta_schedule": "squaredcos_cap_v2", + "beta_start": 0.0001, + "clip_sample": true, + "clip_sample_range": 1.0, + "dynamic_thresholding_ratio": 0.95, + "lambda_min_clipped": -5.1, + "lower_order_final": true, + "num_train_timesteps": 1000, + "prediction_type": "epsilon", + "sample_max_value": 1.5, + "solver_order": 2, + "solver_type": "midpoint", + "steps_offset": 0, + "thresholding": true, + "timestep_spacing": "linspace", + "trained_betas": null, + "use_karras_sigmas": false, + "variance_type": "fixed_small" +} diff --git a/models--showlab--show-1-base/blobs/b8e07730ee2e50b47b667cdaacad799df7f508cf b/models--showlab--show-1-base/blobs/b8e07730ee2e50b47b667cdaacad799df7f508cf new file mode 100644 index 0000000000000000000000000000000000000000..b8e07730ee2e50b47b667cdaacad799df7f508cf --- /dev/null +++ b/models--showlab--show-1-base/blobs/b8e07730ee2e50b47b667cdaacad799df7f508cf @@ -0,0 +1,65 @@ +{ + "_class_name": "UNet3DConditionModel", + "_diffusers_version": "0.19.1", + "act_fn": "gelu", + "addition_embed_type": "text", + "addition_embed_type_num_heads": 64, + "addition_time_embed_dim": null, + "attention_head_dim": 64, + "block_out_channels": [ + 320, + 640, + 960, + 1280 + ], + "center_input_sample": false, + "class_embed_type": null, + "class_embeddings_concat": false, + "conv_in_kernel": 3, + "conv_out_kernel": 3, + "cross_attention_dim": 1280, + "cross_attention_norm": "group_norm", + "down_block_types": [ + "ResnetDownsampleBlock3D", + "SimpleCrossAttnDownBlock3D", + "SimpleCrossAttnDownBlock3D", + "SimpleCrossAttnDownBlock3D" + ], + "downsample_padding": 1, + "dual_cross_attention": false, + "encoder_hid_dim": 4096, + "encoder_hid_dim_type": "text_proj", + "flip_sin_to_cos": true, + "freq_shift": 0, + "in_channels": 3, + "layers_per_block": 3, + "mid_block_only_cross_attention": null, + "mid_block_scale_factor": 1, + "mid_block_type": "UNetMidBlock3DSimpleCrossAttn", + "norm_eps": 1e-05, + "norm_num_groups": 32, + "num_attention_heads": null, + "num_class_embeds": null, + "only_cross_attention": false, + "out_channels": 6, + "projection_class_embeddings_input_dim": null, + "resnet_out_scale_factor": 1.0, + "resnet_skip_time_act": false, + "resnet_time_scale_shift": "scale_shift", + "sample_size": 64, + "time_cond_proj_dim": null, + "time_embedding_act_fn": null, + "time_embedding_dim": null, + "time_embedding_type": "positional", + "timestep_post_act": null, + "transformer_layers_per_block": 1, + "transfromer_in_opt": false, + "up_block_types": [ + "SimpleCrossAttnUpBlock3D", + "SimpleCrossAttnUpBlock3D", + "SimpleCrossAttnUpBlock3D", + "ResnetUpsampleBlock3D" + ], + "upcast_attention": false, + "use_linear_projection": false +} diff --git a/models--showlab--show-1-base/blobs/d2bb149576a5164f6d46963535d5bfd8e169b6d22bf2e31a6126dbc25be5aaac b/models--showlab--show-1-base/blobs/d2bb149576a5164f6d46963535d5bfd8e169b6d22bf2e31a6126dbc25be5aaac new file mode 100644 index 0000000000000000000000000000000000000000..1d926571e27568e3610ce0986aa2aeb5467e7dac --- /dev/null +++ b/models--showlab--show-1-base/blobs/d2bb149576a5164f6d46963535d5bfd8e169b6d22bf2e31a6126dbc25be5aaac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2bb149576a5164f6d46963535d5bfd8e169b6d22bf2e31a6126dbc25be5aaac +size 9452287259 diff --git a/models--showlab--show-1-base/blobs/d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86 b/models--showlab--show-1-base/blobs/d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86 new file mode 100644 index 0000000000000000000000000000000000000000..4e28ff6ebdf584f5372d9de68867399142435d9a Binary files /dev/null and b/models--showlab--show-1-base/blobs/d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86 differ diff --git a/models--showlab--show-1-base/blobs/e99a0d1df1d854da2da3e68e80a83ee379dff248 b/models--showlab--show-1-base/blobs/e99a0d1df1d854da2da3e68e80a83ee379dff248 new file mode 100644 index 0000000000000000000000000000000000000000..e99a0d1df1d854da2da3e68e80a83ee379dff248 --- /dev/null +++ b/models--showlab--show-1-base/blobs/e99a0d1df1d854da2da3e68e80a83ee379dff248 @@ -0,0 +1,112 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": true, + "eos_token": "", + "extra_ids": 100, + "model_max_length": 512, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/models--showlab--show-1-base/refs/main b/models--showlab--show-1-base/refs/main new file mode 100644 index 0000000000000000000000000000000000000000..84ac2a87a64ec582cccc2af597d8690e47858c7d --- /dev/null +++ b/models--showlab--show-1-base/refs/main @@ -0,0 +1 @@ +02c7cbcec48f841e4b4bdbc972077c8420b1ae11 \ No newline at end of file diff --git a/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/feature_extractor/preprocessor_config.json b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/feature_extractor/preprocessor_config.json new file mode 100644 index 0000000000000000000000000000000000000000..0d9d33b883843d1b370da781f3943051067e1b2c --- /dev/null +++ b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/feature_extractor/preprocessor_config.json @@ -0,0 +1,28 @@ +{ + "crop_size": { + "height": 224, + "width": 224 + }, + "do_center_crop": true, + "do_convert_rgb": true, + "do_normalize": true, + "do_rescale": true, + "do_resize": true, + "feature_extractor_type": "CLIPFeatureExtractor", + "image_mean": [ + 0.48145466, + 0.4578275, + 0.40821073 + ], + "image_processor_type": "CLIPImageProcessor", + "image_std": [ + 0.26862954, + 0.26130258, + 0.27577711 + ], + "resample": 3, + "rescale_factor": 0.00392156862745098, + "size": { + "shortest_edge": 224 + } +} diff --git a/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/model_index.json b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/model_index.json new file mode 100644 index 0000000000000000000000000000000000000000..58f6290659e1c3ea9bc80a69742bbfbeac0fbde9 --- /dev/null +++ b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/model_index.json @@ -0,0 +1,25 @@ +{ + "_class_name": "TextToVideoIFPipeline", + "_diffusers_version": "0.19.1", + "_name_or_path": "show-1-base", + "feature_extractor": [ + "transformers", + "CLIPImageProcessor" + ], + "scheduler": [ + "diffusers", + "DPMSolverMultistepScheduler" + ], + "text_encoder": [ + "transformers", + "T5EncoderModel" + ], + "tokenizer": [ + "transformers", + "T5Tokenizer" + ], + "unet": [ + "showone.models.unet_3d_condition", + "UNet3DConditionModel" + ] +} diff --git a/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/scheduler/scheduler_config.json b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/scheduler/scheduler_config.json new file mode 100644 index 0000000000000000000000000000000000000000..82964f12bd7ed4b7eae1a084e2b9e0297dac85c3 --- /dev/null +++ b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/scheduler/scheduler_config.json @@ -0,0 +1,24 @@ +{ + "_class_name": "DPMSolverMultistepScheduler", + "_diffusers_version": "0.19.1", + "algorithm_type": "dpmsolver++", + "beta_end": 0.02, + "beta_schedule": "squaredcos_cap_v2", + "beta_start": 0.0001, + "clip_sample": true, + "clip_sample_range": 1.0, + "dynamic_thresholding_ratio": 0.95, + "lambda_min_clipped": -5.1, + "lower_order_final": true, + "num_train_timesteps": 1000, + "prediction_type": "epsilon", + "sample_max_value": 1.5, + "solver_order": 2, + "solver_type": "midpoint", + "steps_offset": 0, + "thresholding": true, + "timestep_spacing": "linspace", + "trained_betas": null, + "use_karras_sigmas": false, + "variance_type": "fixed_small" +} diff --git a/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/text_encoder/config.json b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/text_encoder/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4f069cd0a20603c3e76d60f34704e65416f12f26 --- /dev/null +++ b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/text_encoder/config.json @@ -0,0 +1,31 @@ +{ + "_name_or_path": "/mnt/nfs/HDD-30t/models/IF-I-L-v1.0/text_encoder", + "architectures": [ + "T5EncoderModel" + ], + "d_ff": 10240, + "d_kv": 64, + "d_model": 4096, + "decoder_start_token_id": 0, + "dense_act_fn": "gelu_new", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "gated-gelu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": true, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "num_decoder_layers": 24, + "num_heads": 64, + "num_layers": 24, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "tie_word_embeddings": false, + "torch_dtype": "float32", + "transformers_version": "4.29.1", + "use_cache": true, + "vocab_size": 32128 +} diff --git a/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/text_encoder/pytorch_model-00001-of-00002.bin b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/text_encoder/pytorch_model-00001-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..1d926571e27568e3610ce0986aa2aeb5467e7dac --- /dev/null +++ b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/text_encoder/pytorch_model-00001-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2bb149576a5164f6d46963535d5bfd8e169b6d22bf2e31a6126dbc25be5aaac +size 9452287259 diff --git a/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/text_encoder/pytorch_model-00002-of-00002.bin b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/text_encoder/pytorch_model-00002-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..342676fb33609e263399c76472b1afa8bb6a2f1d --- /dev/null +++ b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/text_encoder/pytorch_model-00002-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0bff1eff39c04e7d313e5581cf0d8394d25ac277168062b9e43133c399ba058c +size 9597033414 diff --git a/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/text_encoder/pytorch_model.bin.index.json b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/text_encoder/pytorch_model.bin.index.json new file mode 100644 index 0000000000000000000000000000000000000000..1a33111ca9725d27f870b6797f3c9726654a0657 --- /dev/null +++ b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/text_encoder/pytorch_model.bin.index.json @@ -0,0 +1,227 @@ +{ + "metadata": { + "total_size": 19575627776 + }, + "weight_map": { + "encoder.block.0.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.11.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.12.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.13.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.14.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.15.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.16.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.17.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.18.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.19.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.embed_tokens.weight": "pytorch_model-00001-of-00002.bin", + "encoder.final_layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "shared.weight": "pytorch_model-00001-of-00002.bin" + } +} diff --git a/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/tokenizer/special_tokens_map.json b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/tokenizer/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..2c19eb6e3b583f52d34b903b5978d3d30b6b7682 --- /dev/null +++ b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/tokenizer/special_tokens_map.json @@ -0,0 +1,107 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": "", + "pad_token": "", + "unk_token": "" +} diff --git a/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/tokenizer/spiece.model b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/tokenizer/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..317a5ccbde45300f5d1d970d4d449af2108b147e --- /dev/null +++ b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/tokenizer/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86 +size 791656 diff --git a/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/tokenizer/tokenizer_config.json b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/tokenizer/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e99a0d1df1d854da2da3e68e80a83ee379dff248 --- /dev/null +++ b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/tokenizer/tokenizer_config.json @@ -0,0 +1,112 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": true, + "eos_token": "", + "extra_ids": 100, + "model_max_length": 512, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/unet/config.json b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..b8e07730ee2e50b47b667cdaacad799df7f508cf --- /dev/null +++ b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/unet/config.json @@ -0,0 +1,65 @@ +{ + "_class_name": "UNet3DConditionModel", + "_diffusers_version": "0.19.1", + "act_fn": "gelu", + "addition_embed_type": "text", + "addition_embed_type_num_heads": 64, + "addition_time_embed_dim": null, + "attention_head_dim": 64, + "block_out_channels": [ + 320, + 640, + 960, + 1280 + ], + "center_input_sample": false, + "class_embed_type": null, + "class_embeddings_concat": false, + "conv_in_kernel": 3, + "conv_out_kernel": 3, + "cross_attention_dim": 1280, + "cross_attention_norm": "group_norm", + "down_block_types": [ + "ResnetDownsampleBlock3D", + "SimpleCrossAttnDownBlock3D", + "SimpleCrossAttnDownBlock3D", + "SimpleCrossAttnDownBlock3D" + ], + "downsample_padding": 1, + "dual_cross_attention": false, + "encoder_hid_dim": 4096, + "encoder_hid_dim_type": "text_proj", + "flip_sin_to_cos": true, + "freq_shift": 0, + "in_channels": 3, + "layers_per_block": 3, + "mid_block_only_cross_attention": null, + "mid_block_scale_factor": 1, + "mid_block_type": "UNetMidBlock3DSimpleCrossAttn", + "norm_eps": 1e-05, + "norm_num_groups": 32, + "num_attention_heads": null, + "num_class_embeds": null, + "only_cross_attention": false, + "out_channels": 6, + "projection_class_embeddings_input_dim": null, + "resnet_out_scale_factor": 1.0, + "resnet_skip_time_act": false, + "resnet_time_scale_shift": "scale_shift", + "sample_size": 64, + "time_cond_proj_dim": null, + "time_embedding_act_fn": null, + "time_embedding_dim": null, + "time_embedding_type": "positional", + "timestep_post_act": null, + "transformer_layers_per_block": 1, + "transfromer_in_opt": false, + "up_block_types": [ + "SimpleCrossAttnUpBlock3D", + "SimpleCrossAttnUpBlock3D", + "SimpleCrossAttnUpBlock3D", + "ResnetUpsampleBlock3D" + ], + "upcast_attention": false, + "use_linear_projection": false +} diff --git a/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/unet/diffusion_pytorch_model.bin b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..b9a6ff8995176e8caf41be6a8d548a329ffa467a --- /dev/null +++ b/models--showlab--show-1-base/snapshots/02c7cbcec48f841e4b4bdbc972077c8420b1ae11/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73e8b75789dd964d3a51e96a9753359e77cbbc190b1cb83bcf5775ea6374232b +size 3454235275 diff --git a/models--showlab--show-1-interpolation/blobs/08a84d346b5f7c04036082364c0b4e74a754e62ea546de5bcf9949206b0b9719 b/models--showlab--show-1-interpolation/blobs/08a84d346b5f7c04036082364c0b4e74a754e62ea546de5bcf9949206b0b9719 new file mode 100644 index 0000000000000000000000000000000000000000..8e638f104efeb3f994e7b7a1ae33e0a3e3054f1a --- /dev/null +++ b/models--showlab--show-1-interpolation/blobs/08a84d346b5f7c04036082364c0b4e74a754e62ea546de5bcf9949206b0b9719 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08a84d346b5f7c04036082364c0b4e74a754e62ea546de5bcf9949206b0b9719 +size 9725939183 diff --git a/models--showlab--show-1-interpolation/blobs/0d9d33b883843d1b370da781f3943051067e1b2c b/models--showlab--show-1-interpolation/blobs/0d9d33b883843d1b370da781f3943051067e1b2c new file mode 100644 index 0000000000000000000000000000000000000000..0d9d33b883843d1b370da781f3943051067e1b2c --- /dev/null +++ b/models--showlab--show-1-interpolation/blobs/0d9d33b883843d1b370da781f3943051067e1b2c @@ -0,0 +1,28 @@ +{ + "crop_size": { + "height": 224, + "width": 224 + }, + "do_center_crop": true, + "do_convert_rgb": true, + "do_normalize": true, + "do_rescale": true, + "do_resize": true, + "feature_extractor_type": "CLIPFeatureExtractor", + "image_mean": [ + 0.48145466, + 0.4578275, + 0.40821073 + ], + "image_processor_type": "CLIPImageProcessor", + "image_std": [ + 0.26862954, + 0.26130258, + 0.27577711 + ], + "resample": 3, + "rescale_factor": 0.00392156862745098, + "size": { + "shortest_edge": 224 + } +} diff --git a/models--showlab--show-1-interpolation/blobs/1091a86a4c137dcc71b5164aba5c69067bfb9286 b/models--showlab--show-1-interpolation/blobs/1091a86a4c137dcc71b5164aba5c69067bfb9286 new file mode 100644 index 0000000000000000000000000000000000000000..1091a86a4c137dcc71b5164aba5c69067bfb9286 --- /dev/null +++ b/models--showlab--show-1-interpolation/blobs/1091a86a4c137dcc71b5164aba5c69067bfb9286 @@ -0,0 +1,31 @@ +{ + "_name_or_path": "/mnt/efs/home/jarvis/models/IF-I-L-v1.0/text_encoder", + "architectures": [ + "T5EncoderModel" + ], + "d_ff": 10240, + "d_kv": 64, + "d_model": 4096, + "decoder_start_token_id": 0, + "dense_act_fn": "gelu_new", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "gated-gelu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": true, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "num_decoder_layers": 24, + "num_heads": 64, + "num_layers": 24, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "tie_word_embeddings": false, + "torch_dtype": "float16", + "transformers_version": "4.29.1", + "use_cache": true, + "vocab_size": 32128 +} diff --git a/models--showlab--show-1-interpolation/blobs/2c19eb6e3b583f52d34b903b5978d3d30b6b7682 b/models--showlab--show-1-interpolation/blobs/2c19eb6e3b583f52d34b903b5978d3d30b6b7682 new file mode 100644 index 0000000000000000000000000000000000000000..2c19eb6e3b583f52d34b903b5978d3d30b6b7682 --- /dev/null +++ b/models--showlab--show-1-interpolation/blobs/2c19eb6e3b583f52d34b903b5978d3d30b6b7682 @@ -0,0 +1,107 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": "", + "pad_token": "", + "unk_token": "" +} diff --git a/models--showlab--show-1-interpolation/blobs/476aa1bccea9a4c2a726d25d56baca77f75de651 b/models--showlab--show-1-interpolation/blobs/476aa1bccea9a4c2a726d25d56baca77f75de651 new file mode 100644 index 0000000000000000000000000000000000000000..476aa1bccea9a4c2a726d25d56baca77f75de651 --- /dev/null +++ b/models--showlab--show-1-interpolation/blobs/476aa1bccea9a4c2a726d25d56baca77f75de651 @@ -0,0 +1,24 @@ +{ + "_class_name": "DPMSolverMultistepScheduler", + "_diffusers_version": "0.19.3", + "algorithm_type": "dpmsolver++", + "beta_end": 0.02, + "beta_schedule": "squaredcos_cap_v2", + "beta_start": 0.0001, + "clip_sample": true, + "clip_sample_range": 1.0, + "dynamic_thresholding_ratio": 0.95, + "lambda_min_clipped": -5.1, + "lower_order_final": true, + "num_train_timesteps": 1000, + "prediction_type": "epsilon", + "sample_max_value": 1.5, + "solver_order": 2, + "solver_type": "midpoint", + "steps_offset": 0, + "thresholding": true, + "timestep_spacing": "linspace", + "trained_betas": null, + "use_karras_sigmas": false, + "variance_type": "fixed_small" +} diff --git a/models--showlab--show-1-interpolation/blobs/97ed276916890ba4538da9c041cd0af2e86b0873 b/models--showlab--show-1-interpolation/blobs/97ed276916890ba4538da9c041cd0af2e86b0873 new file mode 100644 index 0000000000000000000000000000000000000000..97ed276916890ba4538da9c041cd0af2e86b0873 --- /dev/null +++ b/models--showlab--show-1-interpolation/blobs/97ed276916890ba4538da9c041cd0af2e86b0873 @@ -0,0 +1,227 @@ +{ + "metadata": { + "total_size": 11801079808 + }, + "weight_map": { + "encoder.block.0.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.embed_tokens.weight": "pytorch_model-00001-of-00002.bin", + "encoder.final_layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "shared.weight": "pytorch_model-00001-of-00002.bin" + } +} diff --git a/models--showlab--show-1-interpolation/blobs/abdda89634be180596b30a7c369ab01ea09a9b18a603f9581f83ac8a4b85fbe2 b/models--showlab--show-1-interpolation/blobs/abdda89634be180596b30a7c369ab01ea09a9b18a603f9581f83ac8a4b85fbe2 new file mode 100644 index 0000000000000000000000000000000000000000..a21909bfceeef6beb7d02e56fb34a7a63e02e6b7 --- /dev/null +++ b/models--showlab--show-1-interpolation/blobs/abdda89634be180596b30a7c369ab01ea09a9b18a603f9581f83ac8a4b85fbe2 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abdda89634be180596b30a7c369ab01ea09a9b18a603f9581f83ac8a4b85fbe2 +size 1812025029 diff --git a/models--showlab--show-1-interpolation/blobs/b0d673e9c15f0d46df802afc6dd60ac147e5363ed5e17c7b58211e3f793acc2a b/models--showlab--show-1-interpolation/blobs/b0d673e9c15f0d46df802afc6dd60ac147e5363ed5e17c7b58211e3f793acc2a new file mode 100644 index 0000000000000000000000000000000000000000..2b1d24d092cefc4bb9579aebdb6a311af7f43318 --- /dev/null +++ b/models--showlab--show-1-interpolation/blobs/b0d673e9c15f0d46df802afc6dd60ac147e5363ed5e17c7b58211e3f793acc2a @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0d673e9c15f0d46df802afc6dd60ac147e5363ed5e17c7b58211e3f793acc2a +size 3454460358 diff --git a/models--showlab--show-1-interpolation/blobs/b30c9535bb72995a84dc0e3f30b58e47d7ac3106 b/models--showlab--show-1-interpolation/blobs/b30c9535bb72995a84dc0e3f30b58e47d7ac3106 new file mode 100644 index 0000000000000000000000000000000000000000..b30c9535bb72995a84dc0e3f30b58e47d7ac3106 --- /dev/null +++ b/models--showlab--show-1-interpolation/blobs/b30c9535bb72995a84dc0e3f30b58e47d7ac3106 @@ -0,0 +1,25 @@ +{ + "_class_name": "TextToVideoIFInterpPipeline", + "_diffusers_version": "0.19.3", + "_name_or_path": "show-1-interpolation", + "feature_extractor": [ + "transformers", + "CLIPImageProcessor" + ], + "scheduler": [ + "diffusers", + "DPMSolverMultistepScheduler" + ], + "text_encoder": [ + "transformers", + "T5EncoderModel" + ], + "tokenizer": [ + "transformers", + "T5Tokenizer" + ], + "unet": [ + "showone.models.unet_3d_condition", + "UNet3DConditionModel" + ] +} diff --git a/models--showlab--show-1-interpolation/blobs/d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86 b/models--showlab--show-1-interpolation/blobs/d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86 new file mode 100644 index 0000000000000000000000000000000000000000..4e28ff6ebdf584f5372d9de68867399142435d9a Binary files /dev/null and b/models--showlab--show-1-interpolation/blobs/d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86 differ diff --git a/models--showlab--show-1-interpolation/blobs/e99a0d1df1d854da2da3e68e80a83ee379dff248 b/models--showlab--show-1-interpolation/blobs/e99a0d1df1d854da2da3e68e80a83ee379dff248 new file mode 100644 index 0000000000000000000000000000000000000000..e99a0d1df1d854da2da3e68e80a83ee379dff248 --- /dev/null +++ b/models--showlab--show-1-interpolation/blobs/e99a0d1df1d854da2da3e68e80a83ee379dff248 @@ -0,0 +1,112 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": true, + "eos_token": "", + "extra_ids": 100, + "model_max_length": 512, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/models--showlab--show-1-interpolation/blobs/f08fb177ba057bc33a2f99048011164fce536f94 b/models--showlab--show-1-interpolation/blobs/f08fb177ba057bc33a2f99048011164fce536f94 new file mode 100644 index 0000000000000000000000000000000000000000..f08fb177ba057bc33a2f99048011164fce536f94 --- /dev/null +++ b/models--showlab--show-1-interpolation/blobs/f08fb177ba057bc33a2f99048011164fce536f94 @@ -0,0 +1,65 @@ +{ + "_class_name": "UNet3DConditionModel", + "_diffusers_version": "0.19.3", + "act_fn": "gelu", + "addition_embed_type": "text", + "addition_embed_type_num_heads": 64, + "addition_time_embed_dim": null, + "attention_head_dim": 64, + "block_out_channels": [ + 320, + 640, + 960, + 1280 + ], + "center_input_sample": false, + "class_embed_type": null, + "class_embeddings_concat": false, + "conv_in_kernel": 3, + "conv_out_kernel": 3, + "cross_attention_dim": 1280, + "cross_attention_norm": "group_norm", + "down_block_types": [ + "ResnetDownsampleBlock3D", + "SimpleCrossAttnDownBlock3D", + "SimpleCrossAttnDownBlock3D", + "SimpleCrossAttnDownBlock3D" + ], + "downsample_padding": 1, + "dual_cross_attention": false, + "encoder_hid_dim": 4096, + "encoder_hid_dim_type": "text_proj", + "flip_sin_to_cos": true, + "freq_shift": 0, + "in_channels": 7, + "layers_per_block": 3, + "mid_block_only_cross_attention": null, + "mid_block_scale_factor": 1, + "mid_block_type": "UNetMidBlock3DSimpleCrossAttn", + "norm_eps": 1e-05, + "norm_num_groups": 32, + "num_attention_heads": null, + "num_class_embeds": null, + "only_cross_attention": false, + "out_channels": 6, + "projection_class_embeddings_input_dim": null, + "resnet_out_scale_factor": 1.0, + "resnet_skip_time_act": false, + "resnet_time_scale_shift": "scale_shift", + "sample_size": 64, + "time_cond_proj_dim": null, + "time_embedding_act_fn": null, + "time_embedding_dim": null, + "time_embedding_type": "positional", + "timestep_post_act": null, + "transformer_layers_per_block": 1, + "transfromer_in_opt": false, + "up_block_types": [ + "SimpleCrossAttnUpBlock3D", + "SimpleCrossAttnUpBlock3D", + "SimpleCrossAttnUpBlock3D", + "ResnetUpsampleBlock3D" + ], + "upcast_attention": false, + "use_linear_projection": false +} diff --git a/models--showlab--show-1-interpolation/refs/main b/models--showlab--show-1-interpolation/refs/main new file mode 100644 index 0000000000000000000000000000000000000000..28b57e193775eef116a4e89ec7c4298f30f3a9db --- /dev/null +++ b/models--showlab--show-1-interpolation/refs/main @@ -0,0 +1 @@ +db0295e235766dc9a79dba14b528f0eb7c823dcb \ No newline at end of file diff --git a/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/feature_extractor/preprocessor_config.json b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/feature_extractor/preprocessor_config.json new file mode 100644 index 0000000000000000000000000000000000000000..0d9d33b883843d1b370da781f3943051067e1b2c --- /dev/null +++ b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/feature_extractor/preprocessor_config.json @@ -0,0 +1,28 @@ +{ + "crop_size": { + "height": 224, + "width": 224 + }, + "do_center_crop": true, + "do_convert_rgb": true, + "do_normalize": true, + "do_rescale": true, + "do_resize": true, + "feature_extractor_type": "CLIPFeatureExtractor", + "image_mean": [ + 0.48145466, + 0.4578275, + 0.40821073 + ], + "image_processor_type": "CLIPImageProcessor", + "image_std": [ + 0.26862954, + 0.26130258, + 0.27577711 + ], + "resample": 3, + "rescale_factor": 0.00392156862745098, + "size": { + "shortest_edge": 224 + } +} diff --git a/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/model_index.json b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/model_index.json new file mode 100644 index 0000000000000000000000000000000000000000..b30c9535bb72995a84dc0e3f30b58e47d7ac3106 --- /dev/null +++ b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/model_index.json @@ -0,0 +1,25 @@ +{ + "_class_name": "TextToVideoIFInterpPipeline", + "_diffusers_version": "0.19.3", + "_name_or_path": "show-1-interpolation", + "feature_extractor": [ + "transformers", + "CLIPImageProcessor" + ], + "scheduler": [ + "diffusers", + "DPMSolverMultistepScheduler" + ], + "text_encoder": [ + "transformers", + "T5EncoderModel" + ], + "tokenizer": [ + "transformers", + "T5Tokenizer" + ], + "unet": [ + "showone.models.unet_3d_condition", + "UNet3DConditionModel" + ] +} diff --git a/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/scheduler/scheduler_config.json b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/scheduler/scheduler_config.json new file mode 100644 index 0000000000000000000000000000000000000000..476aa1bccea9a4c2a726d25d56baca77f75de651 --- /dev/null +++ b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/scheduler/scheduler_config.json @@ -0,0 +1,24 @@ +{ + "_class_name": "DPMSolverMultistepScheduler", + "_diffusers_version": "0.19.3", + "algorithm_type": "dpmsolver++", + "beta_end": 0.02, + "beta_schedule": "squaredcos_cap_v2", + "beta_start": 0.0001, + "clip_sample": true, + "clip_sample_range": 1.0, + "dynamic_thresholding_ratio": 0.95, + "lambda_min_clipped": -5.1, + "lower_order_final": true, + "num_train_timesteps": 1000, + "prediction_type": "epsilon", + "sample_max_value": 1.5, + "solver_order": 2, + "solver_type": "midpoint", + "steps_offset": 0, + "thresholding": true, + "timestep_spacing": "linspace", + "trained_betas": null, + "use_karras_sigmas": false, + "variance_type": "fixed_small" +} diff --git a/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/text_encoder/config.json b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/text_encoder/config.json new file mode 100644 index 0000000000000000000000000000000000000000..1091a86a4c137dcc71b5164aba5c69067bfb9286 --- /dev/null +++ b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/text_encoder/config.json @@ -0,0 +1,31 @@ +{ + "_name_or_path": "/mnt/efs/home/jarvis/models/IF-I-L-v1.0/text_encoder", + "architectures": [ + "T5EncoderModel" + ], + "d_ff": 10240, + "d_kv": 64, + "d_model": 4096, + "decoder_start_token_id": 0, + "dense_act_fn": "gelu_new", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "gated-gelu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": true, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "num_decoder_layers": 24, + "num_heads": 64, + "num_layers": 24, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "tie_word_embeddings": false, + "torch_dtype": "float16", + "transformers_version": "4.29.1", + "use_cache": true, + "vocab_size": 32128 +} diff --git a/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/text_encoder/pytorch_model-00001-of-00002.bin b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/text_encoder/pytorch_model-00001-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..8e638f104efeb3f994e7b7a1ae33e0a3e3054f1a --- /dev/null +++ b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/text_encoder/pytorch_model-00001-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08a84d346b5f7c04036082364c0b4e74a754e62ea546de5bcf9949206b0b9719 +size 9725939183 diff --git a/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/text_encoder/pytorch_model-00002-of-00002.bin b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/text_encoder/pytorch_model-00002-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..a21909bfceeef6beb7d02e56fb34a7a63e02e6b7 --- /dev/null +++ b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/text_encoder/pytorch_model-00002-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abdda89634be180596b30a7c369ab01ea09a9b18a603f9581f83ac8a4b85fbe2 +size 1812025029 diff --git a/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/text_encoder/pytorch_model.bin.index.json b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/text_encoder/pytorch_model.bin.index.json new file mode 100644 index 0000000000000000000000000000000000000000..97ed276916890ba4538da9c041cd0af2e86b0873 --- /dev/null +++ b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/text_encoder/pytorch_model.bin.index.json @@ -0,0 +1,227 @@ +{ + "metadata": { + "total_size": 11801079808 + }, + "weight_map": { + "encoder.block.0.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.embed_tokens.weight": "pytorch_model-00001-of-00002.bin", + "encoder.final_layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "shared.weight": "pytorch_model-00001-of-00002.bin" + } +} diff --git a/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/tokenizer/special_tokens_map.json b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/tokenizer/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..2c19eb6e3b583f52d34b903b5978d3d30b6b7682 --- /dev/null +++ b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/tokenizer/special_tokens_map.json @@ -0,0 +1,107 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": "", + "pad_token": "", + "unk_token": "" +} diff --git a/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/tokenizer/spiece.model b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/tokenizer/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..317a5ccbde45300f5d1d970d4d449af2108b147e --- /dev/null +++ b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/tokenizer/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86 +size 791656 diff --git a/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/tokenizer/tokenizer_config.json b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/tokenizer/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e99a0d1df1d854da2da3e68e80a83ee379dff248 --- /dev/null +++ b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/tokenizer/tokenizer_config.json @@ -0,0 +1,112 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": true, + "eos_token": "", + "extra_ids": 100, + "model_max_length": 512, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/unet/config.json b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..f08fb177ba057bc33a2f99048011164fce536f94 --- /dev/null +++ b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/unet/config.json @@ -0,0 +1,65 @@ +{ + "_class_name": "UNet3DConditionModel", + "_diffusers_version": "0.19.3", + "act_fn": "gelu", + "addition_embed_type": "text", + "addition_embed_type_num_heads": 64, + "addition_time_embed_dim": null, + "attention_head_dim": 64, + "block_out_channels": [ + 320, + 640, + 960, + 1280 + ], + "center_input_sample": false, + "class_embed_type": null, + "class_embeddings_concat": false, + "conv_in_kernel": 3, + "conv_out_kernel": 3, + "cross_attention_dim": 1280, + "cross_attention_norm": "group_norm", + "down_block_types": [ + "ResnetDownsampleBlock3D", + "SimpleCrossAttnDownBlock3D", + "SimpleCrossAttnDownBlock3D", + "SimpleCrossAttnDownBlock3D" + ], + "downsample_padding": 1, + "dual_cross_attention": false, + "encoder_hid_dim": 4096, + "encoder_hid_dim_type": "text_proj", + "flip_sin_to_cos": true, + "freq_shift": 0, + "in_channels": 7, + "layers_per_block": 3, + "mid_block_only_cross_attention": null, + "mid_block_scale_factor": 1, + "mid_block_type": "UNetMidBlock3DSimpleCrossAttn", + "norm_eps": 1e-05, + "norm_num_groups": 32, + "num_attention_heads": null, + "num_class_embeds": null, + "only_cross_attention": false, + "out_channels": 6, + "projection_class_embeddings_input_dim": null, + "resnet_out_scale_factor": 1.0, + "resnet_skip_time_act": false, + "resnet_time_scale_shift": "scale_shift", + "sample_size": 64, + "time_cond_proj_dim": null, + "time_embedding_act_fn": null, + "time_embedding_dim": null, + "time_embedding_type": "positional", + "timestep_post_act": null, + "transformer_layers_per_block": 1, + "transfromer_in_opt": false, + "up_block_types": [ + "SimpleCrossAttnUpBlock3D", + "SimpleCrossAttnUpBlock3D", + "SimpleCrossAttnUpBlock3D", + "ResnetUpsampleBlock3D" + ], + "upcast_attention": false, + "use_linear_projection": false +} diff --git a/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/unet/diffusion_pytorch_model.bin b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..2b1d24d092cefc4bb9579aebdb6a311af7f43318 --- /dev/null +++ b/models--showlab--show-1-interpolation/snapshots/db0295e235766dc9a79dba14b528f0eb7c823dcb/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0d673e9c15f0d46df802afc6dd60ac147e5363ed5e17c7b58211e3f793acc2a +size 3454460358 diff --git a/models--showlab--show-1-sr1/blobs/08a84d346b5f7c04036082364c0b4e74a754e62ea546de5bcf9949206b0b9719 b/models--showlab--show-1-sr1/blobs/08a84d346b5f7c04036082364c0b4e74a754e62ea546de5bcf9949206b0b9719 new file mode 100644 index 0000000000000000000000000000000000000000..8e638f104efeb3f994e7b7a1ae33e0a3e3054f1a --- /dev/null +++ b/models--showlab--show-1-sr1/blobs/08a84d346b5f7c04036082364c0b4e74a754e62ea546de5bcf9949206b0b9719 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08a84d346b5f7c04036082364c0b4e74a754e62ea546de5bcf9949206b0b9719 +size 9725939183 diff --git a/models--showlab--show-1-sr1/blobs/0d9d33b883843d1b370da781f3943051067e1b2c b/models--showlab--show-1-sr1/blobs/0d9d33b883843d1b370da781f3943051067e1b2c new file mode 100644 index 0000000000000000000000000000000000000000..0d9d33b883843d1b370da781f3943051067e1b2c --- /dev/null +++ b/models--showlab--show-1-sr1/blobs/0d9d33b883843d1b370da781f3943051067e1b2c @@ -0,0 +1,28 @@ +{ + "crop_size": { + "height": 224, + "width": 224 + }, + "do_center_crop": true, + "do_convert_rgb": true, + "do_normalize": true, + "do_rescale": true, + "do_resize": true, + "feature_extractor_type": "CLIPFeatureExtractor", + "image_mean": [ + 0.48145466, + 0.4578275, + 0.40821073 + ], + "image_processor_type": "CLIPImageProcessor", + "image_std": [ + 0.26862954, + 0.26130258, + 0.27577711 + ], + "resample": 3, + "rescale_factor": 0.00392156862745098, + "size": { + "shortest_edge": 224 + } +} diff --git a/models--showlab--show-1-sr1/blobs/1480c708cae9866e3e6f2255b0b2a4a94dbab09f2b876ef8eb46c1f01d6479b7 b/models--showlab--show-1-sr1/blobs/1480c708cae9866e3e6f2255b0b2a4a94dbab09f2b876ef8eb46c1f01d6479b7 new file mode 100644 index 0000000000000000000000000000000000000000..22fb0f3a2b798ce3b5b8ae1b3defe7cad98346f3 --- /dev/null +++ b/models--showlab--show-1-sr1/blobs/1480c708cae9866e3e6f2255b0b2a4a94dbab09f2b876ef8eb46c1f01d6479b7 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1480c708cae9866e3e6f2255b0b2a4a94dbab09f2b876ef8eb46c1f01d6479b7 +size 1593344706 diff --git a/models--showlab--show-1-sr1/blobs/1ae41d07c327e5793a0cd5e8f8507a34ced27fd0 b/models--showlab--show-1-sr1/blobs/1ae41d07c327e5793a0cd5e8f8507a34ced27fd0 new file mode 100644 index 0000000000000000000000000000000000000000..1ae41d07c327e5793a0cd5e8f8507a34ced27fd0 --- /dev/null +++ b/models--showlab--show-1-sr1/blobs/1ae41d07c327e5793a0cd5e8f8507a34ced27fd0 @@ -0,0 +1,29 @@ +{ + "_class_name": "TextToVideoIFSuperResolutionPipeline_Cond", + "_diffusers_version": "0.19.3", + "_name_or_path": "show-1-sr1", + "feature_extractor": [ + "transformers", + "CLIPImageProcessor" + ], + "image_noising_scheduler": [ + "diffusers", + "DDPMScheduler" + ], + "scheduler": [ + "diffusers", + "DPMSolverMultistepScheduler" + ], + "text_encoder": [ + "transformers", + "T5EncoderModel" + ], + "tokenizer": [ + "transformers", + "T5Tokenizer" + ], + "unet": [ + "showone.models.unet_3d_condition", + "UNet3DConditionModel" + ] +} diff --git a/models--showlab--show-1-sr1/blobs/27b326db7fcacd37be98b5cce8c20a10b6d3f45e b/models--showlab--show-1-sr1/blobs/27b326db7fcacd37be98b5cce8c20a10b6d3f45e new file mode 100644 index 0000000000000000000000000000000000000000..27b326db7fcacd37be98b5cce8c20a10b6d3f45e --- /dev/null +++ b/models--showlab--show-1-sr1/blobs/27b326db7fcacd37be98b5cce8c20a10b6d3f45e @@ -0,0 +1,31 @@ +{ + "_name_or_path": "/mnt/efs/material/jay/models/IF-II-M-v1.0/text_encoder", + "architectures": [ + "T5EncoderModel" + ], + "d_ff": 10240, + "d_kv": 64, + "d_model": 4096, + "decoder_start_token_id": 0, + "dense_act_fn": "gelu_new", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "gated-gelu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": true, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "num_decoder_layers": 24, + "num_heads": 64, + "num_layers": 24, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "tie_word_embeddings": false, + "torch_dtype": "float16", + "transformers_version": "4.29.1", + "use_cache": true, + "vocab_size": 32128 +} diff --git a/models--showlab--show-1-sr1/blobs/2c19eb6e3b583f52d34b903b5978d3d30b6b7682 b/models--showlab--show-1-sr1/blobs/2c19eb6e3b583f52d34b903b5978d3d30b6b7682 new file mode 100644 index 0000000000000000000000000000000000000000..2c19eb6e3b583f52d34b903b5978d3d30b6b7682 --- /dev/null +++ b/models--showlab--show-1-sr1/blobs/2c19eb6e3b583f52d34b903b5978d3d30b6b7682 @@ -0,0 +1,107 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": "", + "pad_token": "", + "unk_token": "" +} diff --git a/models--showlab--show-1-sr1/blobs/3461e7b717210d714c859ed4a34b231a913365ed b/models--showlab--show-1-sr1/blobs/3461e7b717210d714c859ed4a34b231a913365ed new file mode 100644 index 0000000000000000000000000000000000000000..3461e7b717210d714c859ed4a34b231a913365ed --- /dev/null +++ b/models--showlab--show-1-sr1/blobs/3461e7b717210d714c859ed4a34b231a913365ed @@ -0,0 +1,74 @@ +{ + "_class_name": "UNet3DConditionModel", + "_diffusers_version": "0.19.3", + "act_fn": "gelu", + "addition_embed_type": "text", + "addition_embed_type_num_heads": 64, + "addition_time_embed_dim": null, + "attention_head_dim": 64, + "block_out_channels": [ + 128, + 256, + 512, + 768, + 768 + ], + "center_input_sample": false, + "class_embed_type": "timestep", + "class_embeddings_concat": false, + "conv_in_kernel": 3, + "conv_out_kernel": 3, + "cross_attention_dim": 768, + "cross_attention_norm": "group_norm", + "down_block_types": [ + "ResnetDownsampleBlock3D", + "ResnetDownsampleBlock3D", + "ResnetDownsampleBlock3D", + "SimpleCrossAttnDownBlock3D", + "SimpleCrossAttnDownBlock3D" + ], + "downsample_padding": 1, + "dual_cross_attention": false, + "encoder_hid_dim": 4096, + "encoder_hid_dim_type": "text_proj", + "flip_sin_to_cos": true, + "freq_shift": 0, + "in_channels": 6, + "layers_per_block": [ + 2, + 2, + 3, + 4, + 2 + ], + "mid_block_only_cross_attention": null, + "mid_block_scale_factor": 1.4142271248762552, + "mid_block_type": "UNetMidBlock3DSimpleCrossAttn", + "norm_eps": 1e-05, + "norm_num_groups": 32, + "num_attention_heads": null, + "num_class_embeds": null, + "only_cross_attention": false, + "out_channels": 6, + "projection_class_embeddings_input_dim": null, + "resnet_out_scale_factor": 1.4142271248762552, + "resnet_skip_time_act": true, + "resnet_time_scale_shift": "scale_shift", + "sample_size": 256, + "time_cond_proj_dim": null, + "time_embedding_act_fn": "gelu", + "time_embedding_dim": 768, + "time_embedding_type": "positional", + "timestep_post_act": null, + "transformer_layers_per_block": 1, + "transfromer_in_opt": false, + "up_block_types": [ + "SimpleCrossAttnUpBlock3D", + "SimpleCrossAttnUpBlock3D", + "ResnetUpsampleBlock3D", + "ResnetUpsampleBlock3D", + "ResnetUpsampleBlock3D" + ], + "upcast_attention": false, + "use_linear_projection": false +} diff --git a/models--showlab--show-1-sr1/blobs/3a5bf98cae28cc08f87d165f72719b558dfe4bfe b/models--showlab--show-1-sr1/blobs/3a5bf98cae28cc08f87d165f72719b558dfe4bfe new file mode 100644 index 0000000000000000000000000000000000000000..3a5bf98cae28cc08f87d165f72719b558dfe4bfe --- /dev/null +++ b/models--showlab--show-1-sr1/blobs/3a5bf98cae28cc08f87d165f72719b558dfe4bfe @@ -0,0 +1,24 @@ +{ + "_class_name": "DPMSolverMultistepScheduler", + "_diffusers_version": "0.19.3", + "algorithm_type": "dpmsolver++", + "beta_end": 0.02, + "beta_schedule": "squaredcos_cap_v2", + "beta_start": 0.0001, + "clip_sample": true, + "clip_sample_range": 1.0, + "dynamic_thresholding_ratio": 0.95, + "lambda_min_clipped": -5.1, + "lower_order_final": true, + "num_train_timesteps": 1000, + "prediction_type": "v_prediction", + "sample_max_value": 1.0, + "solver_order": 2, + "solver_type": "midpoint", + "steps_offset": 0, + "thresholding": true, + "timestep_spacing": "linspace", + "trained_betas": null, + "use_karras_sigmas": false, + "variance_type": "fixed_small" +} diff --git a/models--showlab--show-1-sr1/blobs/64fca5e4881d23d3033981e2a5aa206119750db6 b/models--showlab--show-1-sr1/blobs/64fca5e4881d23d3033981e2a5aa206119750db6 new file mode 100644 index 0000000000000000000000000000000000000000..64fca5e4881d23d3033981e2a5aa206119750db6 --- /dev/null +++ b/models--showlab--show-1-sr1/blobs/64fca5e4881d23d3033981e2a5aa206119750db6 @@ -0,0 +1,18 @@ +{ + "_class_name": "DDPMScheduler", + "_diffusers_version": "0.19.3", + "beta_end": 0.02, + "beta_schedule": "squaredcos_cap_v2", + "beta_start": 0.0001, + "clip_sample": true, + "clip_sample_range": 1.0, + "dynamic_thresholding_ratio": 0.995, + "num_train_timesteps": 1000, + "prediction_type": "epsilon", + "sample_max_value": 1.0, + "steps_offset": 0, + "thresholding": false, + "timestep_spacing": "leading", + "trained_betas": null, + "variance_type": "fixed_small" +} diff --git a/models--showlab--show-1-sr1/blobs/97ed276916890ba4538da9c041cd0af2e86b0873 b/models--showlab--show-1-sr1/blobs/97ed276916890ba4538da9c041cd0af2e86b0873 new file mode 100644 index 0000000000000000000000000000000000000000..97ed276916890ba4538da9c041cd0af2e86b0873 --- /dev/null +++ b/models--showlab--show-1-sr1/blobs/97ed276916890ba4538da9c041cd0af2e86b0873 @@ -0,0 +1,227 @@ +{ + "metadata": { + "total_size": 11801079808 + }, + "weight_map": { + "encoder.block.0.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.embed_tokens.weight": "pytorch_model-00001-of-00002.bin", + "encoder.final_layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "shared.weight": "pytorch_model-00001-of-00002.bin" + } +} diff --git a/models--showlab--show-1-sr1/blobs/abdda89634be180596b30a7c369ab01ea09a9b18a603f9581f83ac8a4b85fbe2 b/models--showlab--show-1-sr1/blobs/abdda89634be180596b30a7c369ab01ea09a9b18a603f9581f83ac8a4b85fbe2 new file mode 100644 index 0000000000000000000000000000000000000000..a21909bfceeef6beb7d02e56fb34a7a63e02e6b7 --- /dev/null +++ b/models--showlab--show-1-sr1/blobs/abdda89634be180596b30a7c369ab01ea09a9b18a603f9581f83ac8a4b85fbe2 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abdda89634be180596b30a7c369ab01ea09a9b18a603f9581f83ac8a4b85fbe2 +size 1812025029 diff --git a/models--showlab--show-1-sr1/blobs/d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86 b/models--showlab--show-1-sr1/blobs/d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86 new file mode 100644 index 0000000000000000000000000000000000000000..4e28ff6ebdf584f5372d9de68867399142435d9a Binary files /dev/null and b/models--showlab--show-1-sr1/blobs/d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86 differ diff --git a/models--showlab--show-1-sr1/blobs/e99a0d1df1d854da2da3e68e80a83ee379dff248 b/models--showlab--show-1-sr1/blobs/e99a0d1df1d854da2da3e68e80a83ee379dff248 new file mode 100644 index 0000000000000000000000000000000000000000..e99a0d1df1d854da2da3e68e80a83ee379dff248 --- /dev/null +++ b/models--showlab--show-1-sr1/blobs/e99a0d1df1d854da2da3e68e80a83ee379dff248 @@ -0,0 +1,112 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": true, + "eos_token": "", + "extra_ids": 100, + "model_max_length": 512, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/models--showlab--show-1-sr1/refs/main b/models--showlab--show-1-sr1/refs/main new file mode 100644 index 0000000000000000000000000000000000000000..5d40616a321860453ec1eca651e276d589d3e22b --- /dev/null +++ b/models--showlab--show-1-sr1/refs/main @@ -0,0 +1 @@ +51def10fcf940f078b6c6377565937d5458d77ca \ No newline at end of file diff --git a/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/feature_extractor/preprocessor_config.json b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/feature_extractor/preprocessor_config.json new file mode 100644 index 0000000000000000000000000000000000000000..0d9d33b883843d1b370da781f3943051067e1b2c --- /dev/null +++ b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/feature_extractor/preprocessor_config.json @@ -0,0 +1,28 @@ +{ + "crop_size": { + "height": 224, + "width": 224 + }, + "do_center_crop": true, + "do_convert_rgb": true, + "do_normalize": true, + "do_rescale": true, + "do_resize": true, + "feature_extractor_type": "CLIPFeatureExtractor", + "image_mean": [ + 0.48145466, + 0.4578275, + 0.40821073 + ], + "image_processor_type": "CLIPImageProcessor", + "image_std": [ + 0.26862954, + 0.26130258, + 0.27577711 + ], + "resample": 3, + "rescale_factor": 0.00392156862745098, + "size": { + "shortest_edge": 224 + } +} diff --git a/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/image_noising_scheduler/scheduler_config.json b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/image_noising_scheduler/scheduler_config.json new file mode 100644 index 0000000000000000000000000000000000000000..64fca5e4881d23d3033981e2a5aa206119750db6 --- /dev/null +++ b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/image_noising_scheduler/scheduler_config.json @@ -0,0 +1,18 @@ +{ + "_class_name": "DDPMScheduler", + "_diffusers_version": "0.19.3", + "beta_end": 0.02, + "beta_schedule": "squaredcos_cap_v2", + "beta_start": 0.0001, + "clip_sample": true, + "clip_sample_range": 1.0, + "dynamic_thresholding_ratio": 0.995, + "num_train_timesteps": 1000, + "prediction_type": "epsilon", + "sample_max_value": 1.0, + "steps_offset": 0, + "thresholding": false, + "timestep_spacing": "leading", + "trained_betas": null, + "variance_type": "fixed_small" +} diff --git a/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/model_index.json b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/model_index.json new file mode 100644 index 0000000000000000000000000000000000000000..1ae41d07c327e5793a0cd5e8f8507a34ced27fd0 --- /dev/null +++ b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/model_index.json @@ -0,0 +1,29 @@ +{ + "_class_name": "TextToVideoIFSuperResolutionPipeline_Cond", + "_diffusers_version": "0.19.3", + "_name_or_path": "show-1-sr1", + "feature_extractor": [ + "transformers", + "CLIPImageProcessor" + ], + "image_noising_scheduler": [ + "diffusers", + "DDPMScheduler" + ], + "scheduler": [ + "diffusers", + "DPMSolverMultistepScheduler" + ], + "text_encoder": [ + "transformers", + "T5EncoderModel" + ], + "tokenizer": [ + "transformers", + "T5Tokenizer" + ], + "unet": [ + "showone.models.unet_3d_condition", + "UNet3DConditionModel" + ] +} diff --git a/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/scheduler/scheduler_config.json b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/scheduler/scheduler_config.json new file mode 100644 index 0000000000000000000000000000000000000000..3a5bf98cae28cc08f87d165f72719b558dfe4bfe --- /dev/null +++ b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/scheduler/scheduler_config.json @@ -0,0 +1,24 @@ +{ + "_class_name": "DPMSolverMultistepScheduler", + "_diffusers_version": "0.19.3", + "algorithm_type": "dpmsolver++", + "beta_end": 0.02, + "beta_schedule": "squaredcos_cap_v2", + "beta_start": 0.0001, + "clip_sample": true, + "clip_sample_range": 1.0, + "dynamic_thresholding_ratio": 0.95, + "lambda_min_clipped": -5.1, + "lower_order_final": true, + "num_train_timesteps": 1000, + "prediction_type": "v_prediction", + "sample_max_value": 1.0, + "solver_order": 2, + "solver_type": "midpoint", + "steps_offset": 0, + "thresholding": true, + "timestep_spacing": "linspace", + "trained_betas": null, + "use_karras_sigmas": false, + "variance_type": "fixed_small" +} diff --git a/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/text_encoder/config.json b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/text_encoder/config.json new file mode 100644 index 0000000000000000000000000000000000000000..27b326db7fcacd37be98b5cce8c20a10b6d3f45e --- /dev/null +++ b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/text_encoder/config.json @@ -0,0 +1,31 @@ +{ + "_name_or_path": "/mnt/efs/material/jay/models/IF-II-M-v1.0/text_encoder", + "architectures": [ + "T5EncoderModel" + ], + "d_ff": 10240, + "d_kv": 64, + "d_model": 4096, + "decoder_start_token_id": 0, + "dense_act_fn": "gelu_new", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "gated-gelu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": true, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "num_decoder_layers": 24, + "num_heads": 64, + "num_layers": 24, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "tie_word_embeddings": false, + "torch_dtype": "float16", + "transformers_version": "4.29.1", + "use_cache": true, + "vocab_size": 32128 +} diff --git a/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/text_encoder/pytorch_model-00001-of-00002.bin b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/text_encoder/pytorch_model-00001-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..8e638f104efeb3f994e7b7a1ae33e0a3e3054f1a --- /dev/null +++ b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/text_encoder/pytorch_model-00001-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08a84d346b5f7c04036082364c0b4e74a754e62ea546de5bcf9949206b0b9719 +size 9725939183 diff --git a/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/text_encoder/pytorch_model-00002-of-00002.bin b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/text_encoder/pytorch_model-00002-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..a21909bfceeef6beb7d02e56fb34a7a63e02e6b7 --- /dev/null +++ b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/text_encoder/pytorch_model-00002-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abdda89634be180596b30a7c369ab01ea09a9b18a603f9581f83ac8a4b85fbe2 +size 1812025029 diff --git a/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/text_encoder/pytorch_model.bin.index.json b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/text_encoder/pytorch_model.bin.index.json new file mode 100644 index 0000000000000000000000000000000000000000..97ed276916890ba4538da9c041cd0af2e86b0873 --- /dev/null +++ b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/text_encoder/pytorch_model.bin.index.json @@ -0,0 +1,227 @@ +{ + "metadata": { + "total_size": 11801079808 + }, + "weight_map": { + "encoder.block.0.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.0.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.1.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.10.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.11.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.12.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.13.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.14.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.15.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.16.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.17.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.18.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.19.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.2.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.20.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.20.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.21.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.22.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.k.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.o.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.q.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.SelfAttention.v.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.0.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.DenseReluDense.wo.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.23.layer.1.layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.3.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.4.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.5.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.6.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.7.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.8.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.k.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.o.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.q.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.SelfAttention.v.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.0.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wi_0.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wi_1.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.DenseReluDense.wo.weight": "pytorch_model-00001-of-00002.bin", + "encoder.block.9.layer.1.layer_norm.weight": "pytorch_model-00001-of-00002.bin", + "encoder.embed_tokens.weight": "pytorch_model-00001-of-00002.bin", + "encoder.final_layer_norm.weight": "pytorch_model-00002-of-00002.bin", + "shared.weight": "pytorch_model-00001-of-00002.bin" + } +} diff --git a/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/tokenizer/special_tokens_map.json b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/tokenizer/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..2c19eb6e3b583f52d34b903b5978d3d30b6b7682 --- /dev/null +++ b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/tokenizer/special_tokens_map.json @@ -0,0 +1,107 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": "", + "pad_token": "", + "unk_token": "" +} diff --git a/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/tokenizer/spiece.model b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/tokenizer/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..317a5ccbde45300f5d1d970d4d449af2108b147e --- /dev/null +++ b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/tokenizer/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86 +size 791656 diff --git a/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/tokenizer/tokenizer_config.json b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/tokenizer/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e99a0d1df1d854da2da3e68e80a83ee379dff248 --- /dev/null +++ b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/tokenizer/tokenizer_config.json @@ -0,0 +1,112 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": true, + "eos_token": "", + "extra_ids": 100, + "model_max_length": 512, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/unet/config.json b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..3461e7b717210d714c859ed4a34b231a913365ed --- /dev/null +++ b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/unet/config.json @@ -0,0 +1,74 @@ +{ + "_class_name": "UNet3DConditionModel", + "_diffusers_version": "0.19.3", + "act_fn": "gelu", + "addition_embed_type": "text", + "addition_embed_type_num_heads": 64, + "addition_time_embed_dim": null, + "attention_head_dim": 64, + "block_out_channels": [ + 128, + 256, + 512, + 768, + 768 + ], + "center_input_sample": false, + "class_embed_type": "timestep", + "class_embeddings_concat": false, + "conv_in_kernel": 3, + "conv_out_kernel": 3, + "cross_attention_dim": 768, + "cross_attention_norm": "group_norm", + "down_block_types": [ + "ResnetDownsampleBlock3D", + "ResnetDownsampleBlock3D", + "ResnetDownsampleBlock3D", + "SimpleCrossAttnDownBlock3D", + "SimpleCrossAttnDownBlock3D" + ], + "downsample_padding": 1, + "dual_cross_attention": false, + "encoder_hid_dim": 4096, + "encoder_hid_dim_type": "text_proj", + "flip_sin_to_cos": true, + "freq_shift": 0, + "in_channels": 6, + "layers_per_block": [ + 2, + 2, + 3, + 4, + 2 + ], + "mid_block_only_cross_attention": null, + "mid_block_scale_factor": 1.4142271248762552, + "mid_block_type": "UNetMidBlock3DSimpleCrossAttn", + "norm_eps": 1e-05, + "norm_num_groups": 32, + "num_attention_heads": null, + "num_class_embeds": null, + "only_cross_attention": false, + "out_channels": 6, + "projection_class_embeddings_input_dim": null, + "resnet_out_scale_factor": 1.4142271248762552, + "resnet_skip_time_act": true, + "resnet_time_scale_shift": "scale_shift", + "sample_size": 256, + "time_cond_proj_dim": null, + "time_embedding_act_fn": "gelu", + "time_embedding_dim": 768, + "time_embedding_type": "positional", + "timestep_post_act": null, + "transformer_layers_per_block": 1, + "transfromer_in_opt": false, + "up_block_types": [ + "SimpleCrossAttnUpBlock3D", + "SimpleCrossAttnUpBlock3D", + "ResnetUpsampleBlock3D", + "ResnetUpsampleBlock3D", + "ResnetUpsampleBlock3D" + ], + "upcast_attention": false, + "use_linear_projection": false +} diff --git a/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/unet/diffusion_pytorch_model.bin b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..22fb0f3a2b798ce3b5b8ae1b3defe7cad98346f3 --- /dev/null +++ b/models--showlab--show-1-sr1/snapshots/51def10fcf940f078b6c6377565937d5458d77ca/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1480c708cae9866e3e6f2255b0b2a4a94dbab09f2b876ef8eb46c1f01d6479b7 +size 1593344706 diff --git a/version.txt b/version.txt new file mode 100644 index 0000000000000000000000000000000000000000..56a6051ca2b02b04ef92d5150c9ef600403cb1de --- /dev/null +++ b/version.txt @@ -0,0 +1 @@ +1 \ No newline at end of file diff --git a/version_diffusers_cache.txt b/version_diffusers_cache.txt new file mode 100644 index 0000000000000000000000000000000000000000..56a6051ca2b02b04ef92d5150c9ef600403cb1de --- /dev/null +++ b/version_diffusers_cache.txt @@ -0,0 +1 @@ +1 \ No newline at end of file