diff --git a/checkpoint-1000/optimizer.bin b/checkpoint-1000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..f5af6909a58d1a0375c493869677e9b0d2a42aec --- /dev/null +++ b/checkpoint-1000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f0322b5888139ba00a4a41e45908f48b88e2bc4f848e0921ad6e94889ee66c0 +size 2411350917 diff --git a/checkpoint-1000/random_states_0.pkl b/checkpoint-1000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..690597b9e1f318bff17d6a8b1aa6e45f26bb529d --- /dev/null +++ b/checkpoint-1000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c1105e2f308222cc8d3100b64feeb8ef8dd52c41c9d3ff101689419fc1d3aec +size 14663 diff --git a/checkpoint-1000/unet/config.json b/checkpoint-1000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-1000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-1000/unet/diffusion_pytorch_model.bin b/checkpoint-1000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-1000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-1000/unet_ema/config.json b/checkpoint-1000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..79e718fdcaae625155a6c089505e4b9d6d6ef84b --- /dev/null +++ b/checkpoint-1000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 1000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-1000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-1000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..a50dd5fed5bac778eb82d6ba282bdece905a31e8 --- /dev/null +++ b/checkpoint-1000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6c171711b28723be190778ea5147b3509fa144074ac38fbb63c161a487fb4eb +size 1205673685 diff --git a/checkpoint-10000/optimizer.bin b/checkpoint-10000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..f75cd7963681827b68f996b9da26ed845e1e91e8 --- /dev/null +++ b/checkpoint-10000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86f9bbfcb0b5c6cf926251bc8f01c28ebdcc7bedfbeb8cb957f1d9f60133c134 +size 2411350917 diff --git a/checkpoint-10000/random_states_0.pkl b/checkpoint-10000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..d334a4068bc09af505c5651de5f3ca6fc94428fa --- /dev/null +++ b/checkpoint-10000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e220b668168dcdf8d63f73bebd1d2727b7216494653990ee03818930f0deaf70 +size 14663 diff --git a/checkpoint-10000/unet/config.json b/checkpoint-10000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-10000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-10000/unet/diffusion_pytorch_model.bin b/checkpoint-10000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-10000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-10000/unet_ema/config.json b/checkpoint-10000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..3148b00e4f25604312ffb10613406afeb111f193 --- /dev/null +++ b/checkpoint-10000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 10000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-10000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-10000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..e0578966b3b03c197f881ce227aae4ec6aa27abe --- /dev/null +++ b/checkpoint-10000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1727c0c596b9772c999fb8f63f0372cbf11e041ffb1361490ae4595878f73fcc +size 1205673685 diff --git a/checkpoint-10500/optimizer.bin b/checkpoint-10500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..3fa4dcc9a82e14e85ccf26c64b734cc7fbb4cf23 --- /dev/null +++ b/checkpoint-10500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d7bcf85b4ed43a187a644369fba99d63ac40ac8738a1b3cd17cf4ea9aa3a5fe +size 2411350917 diff --git a/checkpoint-10500/random_states_0.pkl b/checkpoint-10500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..8cc261d4a48255cd8e16cfb30ce42b110445fa18 --- /dev/null +++ b/checkpoint-10500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:682adc56d83498cbfb5178a4727df2281a426fa5037cc9dff82dea283b0ba19d +size 14663 diff --git a/checkpoint-10500/unet/config.json b/checkpoint-10500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-10500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-10500/unet/diffusion_pytorch_model.bin b/checkpoint-10500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-10500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-10500/unet_ema/config.json b/checkpoint-10500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..13d5c415c5abbd4ec28903765908a91088c58429 --- /dev/null +++ b/checkpoint-10500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 10500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-10500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-10500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..081bcbfc36fc9e63b6218ab875eec35cbb630586 --- /dev/null +++ b/checkpoint-10500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a2851412c563e03c1000b237729e3a5464b653a0fc8de509590f50e0d71d53a +size 1205673685 diff --git a/checkpoint-11000/optimizer.bin b/checkpoint-11000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..8e05ff38f54abdedabf0096b6ebd9c972576d733 --- /dev/null +++ b/checkpoint-11000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f8901f284acf7e776684a68a0ae5faa7a9c6188af9fa1de8274487fe52697417 +size 2411350917 diff --git a/checkpoint-11000/random_states_0.pkl b/checkpoint-11000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..f71af54d5993478915baca88604384ad845ecdde --- /dev/null +++ b/checkpoint-11000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e4d302b004222cacbbefb8820cc2443e33da75615652cdb67ef109705c6059a7 +size 14663 diff --git a/checkpoint-11000/unet/config.json b/checkpoint-11000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-11000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-11000/unet/diffusion_pytorch_model.bin b/checkpoint-11000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-11000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-11000/unet_ema/config.json b/checkpoint-11000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..b842f089dea4f74278faf54e82c135f98e5a8b6f --- /dev/null +++ b/checkpoint-11000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 11000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-11000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-11000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..e644d802b652df3d5a943ae0791f476050f387e2 --- /dev/null +++ b/checkpoint-11000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:753c1c7f192d084f47a4dca3c521e262793b01bddeb7ac4331c3b612e7ee9c49 +size 1205673685 diff --git a/checkpoint-11500/optimizer.bin b/checkpoint-11500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..223b2e9034dedd191a4ea1e543c770bc927d19d5 --- /dev/null +++ b/checkpoint-11500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68772260b3eeadf3470108661d13b25403fed3698d146515f287f0fab7142ff9 +size 2411350917 diff --git a/checkpoint-11500/random_states_0.pkl b/checkpoint-11500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..c15df74ad9508fb370a908cc9ca06e0e94e8bc5a --- /dev/null +++ b/checkpoint-11500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76ce6ddb7918a19ac947c02441ae4bb08eaf82006be20699e49e0f9bfbb455a6 +size 14663 diff --git a/checkpoint-11500/unet/config.json b/checkpoint-11500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-11500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-11500/unet/diffusion_pytorch_model.bin b/checkpoint-11500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-11500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-11500/unet_ema/config.json b/checkpoint-11500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..1d44e48b59becf10f067ccbd160c70caf0c452af --- /dev/null +++ b/checkpoint-11500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 11500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-11500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-11500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..195aec04e0f759bb29ed55597ce877d3f45d9a29 --- /dev/null +++ b/checkpoint-11500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa3e70a4b7d08e4241cb078ff07768bbdfe1df26f92a8ff093383a0d8d779eb1 +size 1205673685 diff --git a/checkpoint-12000/optimizer.bin b/checkpoint-12000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..d6766004b4b1c8af8ec86aeb3e8a1d3c93cd9b2b --- /dev/null +++ b/checkpoint-12000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6248432f7535de46bb3e5eaf1b88ecdde3a562e2305ca6ff5d342d52d8366299 +size 2411350917 diff --git a/checkpoint-12000/random_states_0.pkl b/checkpoint-12000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..ea2426a61490043f6a9004f1a9b2cdfa8071b36b --- /dev/null +++ b/checkpoint-12000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7a147f4402f1155ffff434f21bfe91d79e1b80b2116f6a338cdb17d5f5dc637 +size 14663 diff --git a/checkpoint-12000/unet/config.json b/checkpoint-12000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-12000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-12000/unet/diffusion_pytorch_model.bin b/checkpoint-12000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-12000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-12000/unet_ema/config.json b/checkpoint-12000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..eba529bfadf1b1522a6dacd7f2fab0ad89042da9 --- /dev/null +++ b/checkpoint-12000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 12000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-12000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-12000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..3fe826329adae5d7b8469bdfc7d598aa60eae515 --- /dev/null +++ b/checkpoint-12000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:487761adf46217a21e6b51fe1db18b7acad256e3885195624589b4d70fd0a0e6 +size 1205673685 diff --git a/checkpoint-12500/optimizer.bin b/checkpoint-12500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..84c215f800047eff3b1b4a6f3b88d624031b3b61 --- /dev/null +++ b/checkpoint-12500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:500e14b20b399b8580ff0caab6078940ab41a15e402d845ea3534a2c2a79aca4 +size 2411350917 diff --git a/checkpoint-12500/random_states_0.pkl b/checkpoint-12500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..b2c9a9dad0fad2cbfdc92c14077ba307a558a168 --- /dev/null +++ b/checkpoint-12500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69c439b2cc2c167b1be19b7ec911955a492aedf9a1a360c7f058c6d1e4d537f1 +size 14663 diff --git a/checkpoint-12500/unet/config.json b/checkpoint-12500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-12500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-12500/unet/diffusion_pytorch_model.bin b/checkpoint-12500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-12500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-12500/unet_ema/config.json b/checkpoint-12500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..9aa9c02eb606ca3f8de6b81d3a5c3c5e6009bdcf --- /dev/null +++ b/checkpoint-12500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 12500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-12500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-12500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..9d94532da1cad2d4c7b3dedf15d2cb9759e2a7dd --- /dev/null +++ b/checkpoint-12500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:00fe1d7950e889a3b18dc37b46e8f40d0ecb034e66106009d2b3373e0fce4fcc +size 1205673685 diff --git a/checkpoint-13000/optimizer.bin b/checkpoint-13000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..ae351d3981d5f99b830c98d4b9a94a1b3a0edc0b --- /dev/null +++ b/checkpoint-13000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e4cd4a6d12f557fcde39293089e1b30acc7c5d94b46b800e5b4e8db84b6d95c +size 2411350917 diff --git a/checkpoint-13000/random_states_0.pkl b/checkpoint-13000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..241cd3b7f32fdc506c133030ac503446ea1ed719 --- /dev/null +++ b/checkpoint-13000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b57a912f09f0a184e1f525e09059a3c46a63c47cd2878d2b122fcb6c08c08ea1 +size 14663 diff --git a/checkpoint-13000/unet/config.json b/checkpoint-13000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-13000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-13000/unet/diffusion_pytorch_model.bin b/checkpoint-13000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-13000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-13000/unet_ema/config.json b/checkpoint-13000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..5705339e420741045924c0c32cdd821fd928f9e3 --- /dev/null +++ b/checkpoint-13000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 13000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-13000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-13000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..4f2cdc532e35632835fec1483ab54da20848a477 --- /dev/null +++ b/checkpoint-13000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b256e206b1d4d8bfeb893ad615df55b6bbdc33bcbda10b455af1e257b71f258b +size 1205673685 diff --git a/checkpoint-13500/optimizer.bin b/checkpoint-13500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..eb99e219bdda7075f07da776465050bb5fc6d05a --- /dev/null +++ b/checkpoint-13500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86eb266f10ccfae17171c9f64bcb0716627a0d0ea2b5fd6abeec9e23d6fb9448 +size 2411350917 diff --git a/checkpoint-13500/random_states_0.pkl b/checkpoint-13500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..8d33b15a2f7702dce9720e3cc000a263b8fc36bd --- /dev/null +++ b/checkpoint-13500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:87b8437bc2aa9f60f41fcc1f43e3779b9003a3ada1265433254221945fe96693 +size 14663 diff --git a/checkpoint-13500/unet/config.json b/checkpoint-13500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-13500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-13500/unet/diffusion_pytorch_model.bin b/checkpoint-13500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-13500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-13500/unet_ema/config.json b/checkpoint-13500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..a700919b1e0d8b92a2d6d6d3bbac5fb40cd5b05f --- /dev/null +++ b/checkpoint-13500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 13500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-13500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-13500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..36efe8de33bc77fd1fcf029ba97240ab7db2ae50 --- /dev/null +++ b/checkpoint-13500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b9013d49162e22574a0fa0d733bc33a8457dacb40b4ff7dcb4b5eceb03125e3a +size 1205673685 diff --git a/checkpoint-14000/optimizer.bin b/checkpoint-14000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..d49377d066d049768758b3774335fe505ee2d926 --- /dev/null +++ b/checkpoint-14000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db1c83ad4e13ac6f2e87f5723b19267c0a1d950f7dfbef515adcbf30ee3bc89a +size 2411350917 diff --git a/checkpoint-14000/random_states_0.pkl b/checkpoint-14000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..53afb6c6a7a948d8f71fdf6757356f60a0ad180d --- /dev/null +++ b/checkpoint-14000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a014fab068d759384de3e177dc1be6aeb11ab936774f52605f2b85627bc96609 +size 14663 diff --git a/checkpoint-14000/unet/config.json b/checkpoint-14000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-14000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-14000/unet/diffusion_pytorch_model.bin b/checkpoint-14000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-14000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-14000/unet_ema/config.json b/checkpoint-14000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..a5966286e6f6e158da70e1b6c026bc92999ee00f --- /dev/null +++ b/checkpoint-14000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 14000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-14000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-14000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..c0768ab8dae5a5847ba0c3ed9106b12a6c21a8dc --- /dev/null +++ b/checkpoint-14000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b62e4226726905782642a8dee740f84ef0894760d9f3632d8f88b0e9bdf1657f +size 1205673685 diff --git a/checkpoint-14500/optimizer.bin b/checkpoint-14500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..352f1e466258b52f212dc96a1e8928e163e24c50 --- /dev/null +++ b/checkpoint-14500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c7f5ceefee9b1ab062bf04d53c1841393c378e3a4e6adbd65254287b4acd86e +size 2411350917 diff --git a/checkpoint-14500/random_states_0.pkl b/checkpoint-14500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..4a9d13434456142609972720e332a6ca01efa2b0 --- /dev/null +++ b/checkpoint-14500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:900f30895cfbc713ff6c10431a0220a1ba9b27db0aa3b074131e46b60c6557ef +size 14663 diff --git a/checkpoint-14500/unet/config.json b/checkpoint-14500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-14500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-14500/unet/diffusion_pytorch_model.bin b/checkpoint-14500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-14500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-14500/unet_ema/config.json b/checkpoint-14500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..96cae266a0a9d0725ed9d41e5cf872f84560a4e4 --- /dev/null +++ b/checkpoint-14500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 14500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-14500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-14500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..e8bb18dadbfd45c47f520dbef94b3ccee690c3cf --- /dev/null +++ b/checkpoint-14500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3dee149431022d024d3a72f4cc53350af6f9812edf673fc15f8f6d869cfbb40a +size 1205673685 diff --git a/checkpoint-1500/optimizer.bin b/checkpoint-1500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..425bea79e962ef77c625533437196ea9b72df205 --- /dev/null +++ b/checkpoint-1500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05b5da5c68548fb4ab11b24202f6881624a28089c0f7d6101e6bca783927fbba +size 2411350917 diff --git a/checkpoint-1500/random_states_0.pkl b/checkpoint-1500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..94c5ca284b9e5d5f36524fc3f3e162699168fd88 --- /dev/null +++ b/checkpoint-1500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c95fd4355842526cf9f62ce87471aaf938fa4f057d16fe660e93f85b26d306e8 +size 14663 diff --git a/checkpoint-1500/unet/config.json b/checkpoint-1500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-1500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-1500/unet/diffusion_pytorch_model.bin b/checkpoint-1500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-1500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-1500/unet_ema/config.json b/checkpoint-1500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..1c05ba02751bde3b37d0f4f45a55f92c184ff661 --- /dev/null +++ b/checkpoint-1500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 1500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-1500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-1500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..8425f7390417c0c1d82f31f1003dfd592648da3b --- /dev/null +++ b/checkpoint-1500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24d1c0166d206592b10edd22fb1defa57713a222feed12c34b260b5fdc866585 +size 1205673685 diff --git a/checkpoint-15000/optimizer.bin b/checkpoint-15000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..7ab188ef15231c2d4b48528b68b07fa537ad388e --- /dev/null +++ b/checkpoint-15000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f5c401cb573e005e3165cade220bb9c45ec9d12a3a3f2e30a5bbfc5938a1eaf +size 2411350917 diff --git a/checkpoint-15000/random_states_0.pkl b/checkpoint-15000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..2b975c6bc39c7c12ba3b3b89895945ccc74977ec --- /dev/null +++ b/checkpoint-15000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de7d4d0cb1f85984b030b392572f4503339b27e4fe002d9ba842f0f43ce356aa +size 14663 diff --git a/checkpoint-15000/unet/config.json b/checkpoint-15000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-15000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-15000/unet/diffusion_pytorch_model.bin b/checkpoint-15000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-15000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-15000/unet_ema/config.json b/checkpoint-15000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..83e576561357eba7b47da4222e756386d664ab22 --- /dev/null +++ b/checkpoint-15000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 15000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-15000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-15000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..e06adcd42fb9b58dbaebe597808ad502d28ef8ed --- /dev/null +++ b/checkpoint-15000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a2b97adb51c119b31acce1ef7af8c32e8b717bb79fc595e7df12863d8468837 +size 1205673685 diff --git a/checkpoint-15500/optimizer.bin b/checkpoint-15500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..23fbdcda8bc4bb4c2a06e4aea6bd175034efd30f --- /dev/null +++ b/checkpoint-15500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e20d84706fc1da7034eab60135a532383befd54beede547ce427e60f51960ef7 +size 2411350917 diff --git a/checkpoint-15500/random_states_0.pkl b/checkpoint-15500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..f8efc1e3b3998da3fa6ddfc2e7469dab12cc6917 --- /dev/null +++ b/checkpoint-15500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c630544873fb41200dba1785da410b2b7d7be17d70f1aff8b26754297d795d47 +size 14663 diff --git a/checkpoint-15500/unet/config.json b/checkpoint-15500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-15500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-15500/unet/diffusion_pytorch_model.bin b/checkpoint-15500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-15500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-15500/unet_ema/config.json b/checkpoint-15500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..5a193751f56ab91f4530aaf8bef4fbb32f1c3000 --- /dev/null +++ b/checkpoint-15500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 15500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-15500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-15500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..3631d907793febc5325681d2d7545812e71fea68 --- /dev/null +++ b/checkpoint-15500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:690fff6217ff8ef90f27b1307832619f78fe802b11fb56d01de6e49e45d30e2e +size 1205673685 diff --git a/checkpoint-16000/optimizer.bin b/checkpoint-16000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..637994c8acdffe897b493a723b41f7a3ccd59e76 --- /dev/null +++ b/checkpoint-16000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb36aa2c4ab459106e6b368ecd69c524cbab5c78a11fcb2d21c50828340c263e +size 2411350917 diff --git a/checkpoint-16000/random_states_0.pkl b/checkpoint-16000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..349bc99c6cd083852a0a46436ba79711d463fd55 --- /dev/null +++ b/checkpoint-16000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39715ccfac486d846772efff3b72caf67645f94b55c582d1c296c83378d46ec9 +size 14663 diff --git a/checkpoint-16000/unet/config.json b/checkpoint-16000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-16000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-16000/unet/diffusion_pytorch_model.bin b/checkpoint-16000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-16000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-16000/unet_ema/config.json b/checkpoint-16000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..0fe218be38e5769bb079a02e957b7dd372874b9d --- /dev/null +++ b/checkpoint-16000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 16000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-16000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-16000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..8cb4441837f656156816efca23bb0977dcaf4341 --- /dev/null +++ b/checkpoint-16000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:564972792166ed3f309d82967ae1ba90dd41528667486f3db34b019c8987e3ac +size 1205673685 diff --git a/checkpoint-16500/optimizer.bin b/checkpoint-16500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..eb33a7fbba622ddfebdcd7320c168e2d7423779c --- /dev/null +++ b/checkpoint-16500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:778e4a4b6739c4906d83e6fcbae137c24af985c538f4b7c87de5d05c36729c3a +size 2411350917 diff --git a/checkpoint-16500/random_states_0.pkl b/checkpoint-16500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..ee5357310acb75df0e2b951c9f3f7837ce68f978 --- /dev/null +++ b/checkpoint-16500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:630c542f16915762028d74f9d297b3616ed307e156fcd6c18305873c02cbdbf3 +size 14663 diff --git a/checkpoint-16500/unet/config.json b/checkpoint-16500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-16500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-16500/unet/diffusion_pytorch_model.bin b/checkpoint-16500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-16500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-16500/unet_ema/config.json b/checkpoint-16500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..19e23848c171ca871682fd390d6150ab76d0b3bf --- /dev/null +++ b/checkpoint-16500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 16500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-16500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-16500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..8243d7b96c14d8c51b8e08a97a6c32a10c48d4f6 --- /dev/null +++ b/checkpoint-16500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ed0a58e63d4b2b393154760b831de65e54584118b96d7ee47c908be78c6f1cdf +size 1205673685 diff --git a/checkpoint-17000/optimizer.bin b/checkpoint-17000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..5ab3a3070f5b90fb02b6d897b9ed4bef45c51ace --- /dev/null +++ b/checkpoint-17000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bcb1e01ddd28f18544e2965d77422f6078965710c0fce6fa61229e9418962aa6 +size 2411350917 diff --git a/checkpoint-17000/random_states_0.pkl b/checkpoint-17000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..b461aa2fc6e66671538c274d0e8c542994b2ef35 --- /dev/null +++ b/checkpoint-17000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f721db0fd9fcc338cf79841d0ddc7f21001e6946c96bff331f7d900222138589 +size 14663 diff --git a/checkpoint-17000/unet/config.json b/checkpoint-17000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-17000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-17000/unet/diffusion_pytorch_model.bin b/checkpoint-17000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-17000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-17000/unet_ema/config.json b/checkpoint-17000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..ccaba7d40dc58503c3a49ea00d62c837556dca07 --- /dev/null +++ b/checkpoint-17000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 17000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-17000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-17000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..cf0fa2c3a72ae6c52057dcfb8ec6f2df98650219 --- /dev/null +++ b/checkpoint-17000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3c3c30143d169ebd2e4af9edd9be117f52b70e84ddbd4c56ba5eac6750382ef +size 1205673685 diff --git a/checkpoint-17500/optimizer.bin b/checkpoint-17500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..72d89a3aacba8b2f8ca1c41791356a6b898fbb51 --- /dev/null +++ b/checkpoint-17500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a90129cb46ed272969770af4d3d8c8f585ab514aff0ac66345280b74bc6f0425 +size 2411350917 diff --git a/checkpoint-17500/random_states_0.pkl b/checkpoint-17500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..b52a8935f40b8f8da1ef1d0dbf09f4bdd24f95ac --- /dev/null +++ b/checkpoint-17500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3cd014d7fe8fce963a19289a28d1b68f260875687783a8a987115b4447fd7b89 +size 14663 diff --git a/checkpoint-17500/unet/config.json b/checkpoint-17500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-17500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-17500/unet/diffusion_pytorch_model.bin b/checkpoint-17500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-17500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-17500/unet_ema/config.json b/checkpoint-17500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..6a1b7fcb0a62f8609643c4e9b053ff53d7090090 --- /dev/null +++ b/checkpoint-17500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 17500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-17500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-17500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..f9fc6bb575a5442ff73b59783ac95ca5b309149e --- /dev/null +++ b/checkpoint-17500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1c5039a88601dee8a3b144c3728e330e9b36ea554b067594ea228fe8916416f +size 1205673685 diff --git a/checkpoint-18000/optimizer.bin b/checkpoint-18000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..905dbef56f23a778bff05fd3aec273a459a10aa1 --- /dev/null +++ b/checkpoint-18000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:163f1e897abcc4d18341ee4bf2fba18789db6dbe8471e3734d4c68ac21d68082 +size 2411350917 diff --git a/checkpoint-18000/random_states_0.pkl b/checkpoint-18000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..0e1c31c3595babf4b760b0261742dbb4725fd059 --- /dev/null +++ b/checkpoint-18000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:264880bee6a5f4137fb4ec67e9802f521140512e2415497d32d7831d6e65c699 +size 14663 diff --git a/checkpoint-18000/unet/config.json b/checkpoint-18000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-18000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-18000/unet/diffusion_pytorch_model.bin b/checkpoint-18000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-18000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-18000/unet_ema/config.json b/checkpoint-18000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..b9be784c9eedbe1d369336cedc2f50a0be1ba2ac --- /dev/null +++ b/checkpoint-18000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 18000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-18000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-18000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..3632c8c2d2640340f794b6b5b5ce8b1f409e3705 --- /dev/null +++ b/checkpoint-18000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d60d677b9031c8bd98eff8a0bc8ffef067ea92f7f0dd9c0a5ccbdbbb06ab89e +size 1205673685 diff --git a/checkpoint-18500/optimizer.bin b/checkpoint-18500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..82147f4f6be222afccfeff0c4d6d4dfbf8652bc8 --- /dev/null +++ b/checkpoint-18500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1cd5ed4b77185dce4e9d4adad380ef81043d11ac15dbfe2c7525bd877a293669 +size 2411350917 diff --git a/checkpoint-18500/random_states_0.pkl b/checkpoint-18500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..8e9028b24dd449d84266189e361fb8f17894c54c --- /dev/null +++ b/checkpoint-18500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c1ee1732c1ed310dc8721f154aa988a7e04370d46422aadc6433eddae3205a3 +size 14663 diff --git a/checkpoint-18500/unet/config.json b/checkpoint-18500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-18500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-18500/unet/diffusion_pytorch_model.bin b/checkpoint-18500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-18500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-18500/unet_ema/config.json b/checkpoint-18500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..5b5d4f65bcc016f7a104b95f88248b9aa3c2023d --- /dev/null +++ b/checkpoint-18500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 18500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-18500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-18500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..630330acdffebd6431dccc3ef9c778fcc5210b65 --- /dev/null +++ b/checkpoint-18500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7fc5789eb705e4eb7a2b35b64181ee987454a2d32bf961ea8b9d975396135a4 +size 1205673685 diff --git a/checkpoint-19000/optimizer.bin b/checkpoint-19000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..ec1fe0ab04fb48434fce6152906953c61b0d46fe --- /dev/null +++ b/checkpoint-19000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3c427b49fe3e831ee188d6e11e7075d9b0bff59db5402b10bd703835bbd3488 +size 2411350917 diff --git a/checkpoint-19000/random_states_0.pkl b/checkpoint-19000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..295c26756ddee3ca4fb43d0c46c21157668e9a4f --- /dev/null +++ b/checkpoint-19000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33cb4cb09649ea30690ec22b301c1f7c689d1e50f26c0f1feab711e283dc712e +size 14663 diff --git a/checkpoint-19000/unet/config.json b/checkpoint-19000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-19000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-19000/unet/diffusion_pytorch_model.bin b/checkpoint-19000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-19000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-19000/unet_ema/config.json b/checkpoint-19000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..981c0cbd8d80296b013c1197c9648a6b5a75b726 --- /dev/null +++ b/checkpoint-19000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 19000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-19000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-19000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..a3a0ffe596e807df5fd33d8b1531c3817e0d5509 --- /dev/null +++ b/checkpoint-19000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f13c6262f0c44a765d7e064ce32ef96621cb49c560ba712eb31abe1d1c04833 +size 1205673685 diff --git a/checkpoint-19500/optimizer.bin b/checkpoint-19500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..cf59f1be673b65f553caa7d1d35b4d08c545aa68 --- /dev/null +++ b/checkpoint-19500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f52ff79484cc019f722c2b9632f83bbca6acf70219708b6ecf90ae091dccb3a +size 2411350917 diff --git a/checkpoint-19500/random_states_0.pkl b/checkpoint-19500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..05f6706a66d586108580f726e3ba5ab9b43ac367 --- /dev/null +++ b/checkpoint-19500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:121f3cebb3b0fcba4bc92d4ced87e9490bb57fcdb082022583de7f3102e15024 +size 14663 diff --git a/checkpoint-19500/unet/config.json b/checkpoint-19500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-19500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-19500/unet/diffusion_pytorch_model.bin b/checkpoint-19500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-19500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-19500/unet_ema/config.json b/checkpoint-19500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..946d8f51ff01d89f618ace318c1a50836265167d --- /dev/null +++ b/checkpoint-19500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 19500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-19500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-19500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..0d3b636a9fbdb5011e40a0e1710caea47b66bd02 --- /dev/null +++ b/checkpoint-19500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c47d7327f9d5b78c27294a1956bffa6a6cf233cba9a4b80030cb4e417b208451 +size 1205673685 diff --git a/checkpoint-2000/optimizer.bin b/checkpoint-2000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..9aa176856dad770b313eadb26aabd3104f20949b --- /dev/null +++ b/checkpoint-2000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df91ebd02adbe1ce1c89f675e1777de263d71238f586cbea15f804b0e16690c5 +size 2411350917 diff --git a/checkpoint-2000/random_states_0.pkl b/checkpoint-2000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..c840a6a8b6c6175dfa33871b660625c57a64e6d4 --- /dev/null +++ b/checkpoint-2000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea095fa0bbd0f63a8e8d0fb88455dc18da19de06c9e853e1a663d3a52b3f1e6a +size 14663 diff --git a/checkpoint-2000/unet/config.json b/checkpoint-2000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-2000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-2000/unet/diffusion_pytorch_model.bin b/checkpoint-2000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-2000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-2000/unet_ema/config.json b/checkpoint-2000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..e737c493f71d4698f784993d84bfa9e1afd1721e --- /dev/null +++ b/checkpoint-2000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 2000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-2000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-2000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..6459ff128fd98dbf0f2a85355c3b16ec7b0df663 --- /dev/null +++ b/checkpoint-2000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a539dd1e7666971cc83d9ee220ebadfe7e4f058583d99892bafb1d7facb9d286 +size 1205673685 diff --git a/checkpoint-20000/optimizer.bin b/checkpoint-20000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..658f767b9c794f2e171c0b25363ea022112fe436 --- /dev/null +++ b/checkpoint-20000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:820fec955edf96577693d9650907ae30a1406413bf400659c6af1d97dd78cbee +size 2411350917 diff --git a/checkpoint-20000/random_states_0.pkl b/checkpoint-20000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..1fbe7529d43a2a5225a0455738c4e4fbad0b8fc3 --- /dev/null +++ b/checkpoint-20000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c85eb12f99f85bd9b49fbcbab0a64d61e83931ef7f02197ccd8cf11a5e39536 +size 14663 diff --git a/checkpoint-20000/unet/config.json b/checkpoint-20000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-20000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-20000/unet/diffusion_pytorch_model.bin b/checkpoint-20000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-20000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-20000/unet_ema/config.json b/checkpoint-20000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..6a9fe9cdba277aa17346ceea23eab9297a13ad50 --- /dev/null +++ b/checkpoint-20000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 20000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-20000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-20000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..4d9bdc96c5f4dd340a4ac63be23404e1cd620a6b --- /dev/null +++ b/checkpoint-20000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33a5b119f0618f262a69b458fed131348f2e5a7378bd68ae24ca87632af6b120 +size 1205673685 diff --git a/checkpoint-20500/optimizer.bin b/checkpoint-20500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..05a6fe0224159ac68657a90a7c7ad6fb31ab874b --- /dev/null +++ b/checkpoint-20500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9bc1d8b84cc5fa410a609aa7d23dabaefc38b028f3bc441776ae639467b2e2c6 +size 2411350917 diff --git a/checkpoint-20500/random_states_0.pkl b/checkpoint-20500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..e27965c33d1ff4e6e85e8c79c3504ad9954f5358 --- /dev/null +++ b/checkpoint-20500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71784a8d68634238da4bc624fb3d5af3cad8b9b914f83f141334d287afc558a1 +size 14663 diff --git a/checkpoint-20500/unet/config.json b/checkpoint-20500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-20500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-20500/unet/diffusion_pytorch_model.bin b/checkpoint-20500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-20500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-20500/unet_ema/config.json b/checkpoint-20500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..3979af86f3357077bc41d05f982902c37bc8df23 --- /dev/null +++ b/checkpoint-20500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 20500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-20500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-20500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..d5f4a9b18f48a808ba7257237aad3a8f1336674f --- /dev/null +++ b/checkpoint-20500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:61e692c0c2c84f8073e2d17ce10676293ffd4eff45c7f92920c3643cd98cd5a6 +size 1205673685 diff --git a/checkpoint-21000/optimizer.bin b/checkpoint-21000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..56108bfc203fe79269cffd609710d4f29d74305e --- /dev/null +++ b/checkpoint-21000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2c9ca6fcfc83c3a04966a46b2a01d2e9a94e61edaa9c58747d5a86cb160e1e6 +size 2411350917 diff --git a/checkpoint-21000/random_states_0.pkl b/checkpoint-21000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..1c548d460e5a83dd532b044dca9be7c43de59f0a --- /dev/null +++ b/checkpoint-21000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4445a18b9461a99c6230ac7ceda1a80d27e5519abef16af606023356117ebd7d +size 14663 diff --git a/checkpoint-21000/unet/config.json b/checkpoint-21000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-21000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-21000/unet/diffusion_pytorch_model.bin b/checkpoint-21000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-21000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-21000/unet_ema/config.json b/checkpoint-21000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..37609f986ca2b165aafb1241a1c1e811d864baff --- /dev/null +++ b/checkpoint-21000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 21000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-21000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-21000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..f980e4090341add102790f5ee4d4118ed43f8eb4 --- /dev/null +++ b/checkpoint-21000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:48686f9ee4e5f8aa57dfefb33d6fc365df5d50869860cd0d4ed5c683e01a9275 +size 1205673685 diff --git a/checkpoint-21500/optimizer.bin b/checkpoint-21500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..3b882c4f13bdaf1d0df939f3d9ed139fc949f0ab --- /dev/null +++ b/checkpoint-21500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02972c44a84cbf13e36950203c2e73c6146d3340b288fd36eb4914117818a374 +size 2411350917 diff --git a/checkpoint-21500/random_states_0.pkl b/checkpoint-21500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..b82a20c3b6c034a11f58a3ced10ea2641e625a5b --- /dev/null +++ b/checkpoint-21500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf8f5a8cf315e3960b90b95f4be3ea4b1804d5d474cb931b8275464548f37365 +size 14663 diff --git a/checkpoint-21500/unet/config.json b/checkpoint-21500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-21500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-21500/unet/diffusion_pytorch_model.bin b/checkpoint-21500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-21500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-21500/unet_ema/config.json b/checkpoint-21500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..df23bc33b47ce0a646a2765594b50afaf827346c --- /dev/null +++ b/checkpoint-21500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 21500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-21500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-21500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..b30900f0dcdb7e1a52acc0de0bb96e6dcab4fcc2 --- /dev/null +++ b/checkpoint-21500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3885dbbc6be801b004f6e4aee526072012217bb5332f14908c1d23c09081a4d8 +size 1205673685 diff --git a/checkpoint-22000/optimizer.bin b/checkpoint-22000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..7376a58eabb0f9be2cc7244c0d4ac5a5e343ab1a --- /dev/null +++ b/checkpoint-22000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31ac68be4d97cc6fec47ef0f14b33f5eefff89904359e789a54812533e91f0a2 +size 2411350917 diff --git a/checkpoint-22000/random_states_0.pkl b/checkpoint-22000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..4975a9c504faa060c53aae9028af990fc9fad082 --- /dev/null +++ b/checkpoint-22000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bdcacad60d951df4a1a07938d7ef596d3d68cae2ea8662a75b2f317d7213323e +size 14663 diff --git a/checkpoint-22000/unet/config.json b/checkpoint-22000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-22000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-22000/unet/diffusion_pytorch_model.bin b/checkpoint-22000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-22000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-22000/unet_ema/config.json b/checkpoint-22000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..75dabe29641897ed6d7af492e714237609eaceae --- /dev/null +++ b/checkpoint-22000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 22000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-22000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-22000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..74179ef00d5445c20ee96f4ad58978d836638733 --- /dev/null +++ b/checkpoint-22000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6349861f2c4c09d622380b0ec95c90691d5b2ffd34eaa8d20e4df83420a9fe6a +size 1205673685 diff --git a/checkpoint-22500/optimizer.bin b/checkpoint-22500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..30153fd968bc6cf475c1d043a2cea6c043098c27 --- /dev/null +++ b/checkpoint-22500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:269b6f5f397e28383c693fff93082da8ddf88389caf211e85a24874479223b27 +size 2411350917 diff --git a/checkpoint-22500/random_states_0.pkl b/checkpoint-22500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..14feb56becb52713c2f69afb10d3109291483821 --- /dev/null +++ b/checkpoint-22500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc6c25d76e13459a529c8e5b57108ee56dbe4807f035694b875795fa25921563 +size 14663 diff --git a/checkpoint-22500/unet/config.json b/checkpoint-22500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-22500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-22500/unet/diffusion_pytorch_model.bin b/checkpoint-22500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-22500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-22500/unet_ema/config.json b/checkpoint-22500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..a977f419b893671d1327594a4062d09030305205 --- /dev/null +++ b/checkpoint-22500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 22500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-22500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-22500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..aa4e3c60acc8d2d956775063f67d0291542b13a1 --- /dev/null +++ b/checkpoint-22500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:637ddba651d87535c35ca4993a46347a23c14a700018a73ad3b0fe6ddd831574 +size 1205673685 diff --git a/checkpoint-23000/optimizer.bin b/checkpoint-23000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..1b6a6438345691ef3a8eaf7dbfae058dcc30f4fe --- /dev/null +++ b/checkpoint-23000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29a463a8b839acb2173ef29128441865e98a8dd8d30fa8358a14abe22286d115 +size 2411350917 diff --git a/checkpoint-23000/random_states_0.pkl b/checkpoint-23000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..1647617cff16b0fc61b899c4c57cdba6c1cce9bb --- /dev/null +++ b/checkpoint-23000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1fb0cc63268c3982d05ec827ebb14e52cd033bf08637a2f3dc490be9a1bac85 +size 14663 diff --git a/checkpoint-23000/unet/config.json b/checkpoint-23000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-23000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-23000/unet/diffusion_pytorch_model.bin b/checkpoint-23000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-23000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-23000/unet_ema/config.json b/checkpoint-23000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..2ad1d625225709107c4493dd0930aea3fac64c5b --- /dev/null +++ b/checkpoint-23000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 23000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-23000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-23000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..e8c35792a0ff3973968909c6e5f7576dac525fe1 --- /dev/null +++ b/checkpoint-23000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:44be7ae7eed54a9592e55a9cd7344a8aa55a2dfb91f0172937733bf03e29f5d9 +size 1205673685 diff --git a/checkpoint-23500/optimizer.bin b/checkpoint-23500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..564304757f7c5f8fda70aa84db8c20de916f05c8 --- /dev/null +++ b/checkpoint-23500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c243ed5c21fa2002d7fac7a85441947702c2becb4422557aadfba3fdcedac89e +size 2411350917 diff --git a/checkpoint-23500/random_states_0.pkl b/checkpoint-23500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..718d5cee3e4f81f54c54309f7f564b4923f9a62d --- /dev/null +++ b/checkpoint-23500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fff24c6598f5bde40755740cceff92de39cb7ca4796ee997cdb3a0727836139e +size 14663 diff --git a/checkpoint-23500/unet/config.json b/checkpoint-23500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-23500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-23500/unet/diffusion_pytorch_model.bin b/checkpoint-23500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-23500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-23500/unet_ema/config.json b/checkpoint-23500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..e7126cdda7a06571ae9ea545edf3e7b935be4191 --- /dev/null +++ b/checkpoint-23500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 23500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-23500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-23500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..d76c9bea6f571d92a05e05316e9cba10a145d848 --- /dev/null +++ b/checkpoint-23500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd0fea7cca16b01c97b5a67c023f0630c7a0ba651c3e3a3402a785364ec44538 +size 1205673685 diff --git a/checkpoint-24000/optimizer.bin b/checkpoint-24000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..d8c31f0b8c9b53310304c7c77a997e66f358e3dd --- /dev/null +++ b/checkpoint-24000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c1312d540e8975b317e23f0199163936178ac589497564514a7fae2d7304f5c +size 2411350917 diff --git a/checkpoint-24000/random_states_0.pkl b/checkpoint-24000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..1aca3ff39a7623debe4b67005c30a1cd9bc3cb94 --- /dev/null +++ b/checkpoint-24000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a9e8c49d9161c866b62c5da251bca7717c2e47ebf12ba72aadb9c643935acf26 +size 14663 diff --git a/checkpoint-24000/unet/config.json b/checkpoint-24000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-24000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-24000/unet/diffusion_pytorch_model.bin b/checkpoint-24000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-24000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-24000/unet_ema/config.json b/checkpoint-24000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..65f6ae8cf133376899a1d6c963fddfea4038932e --- /dev/null +++ b/checkpoint-24000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 24000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-24000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-24000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..59c264e38f04140ed3935d6fd07999682470c284 --- /dev/null +++ b/checkpoint-24000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c14d58ccf634a87a0c312fe018d8ce8efb639fdb88f645da684e5bf8ee45dde2 +size 1205673685 diff --git a/checkpoint-24500/optimizer.bin b/checkpoint-24500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..ebc747468f457c4db273d2626c3bfc49f1b9cf36 --- /dev/null +++ b/checkpoint-24500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:749500d7f564f7c396e2c554f9143b348acba37e4b65e61e8664aca7db6cf6e7 +size 2411350917 diff --git a/checkpoint-24500/random_states_0.pkl b/checkpoint-24500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..1c089ba9c49735760fe53d8277f7b23b32732067 --- /dev/null +++ b/checkpoint-24500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6dd52f631180d9f2e4b260503d4d9091bd58e905204553854696e21785843914 +size 14663 diff --git a/checkpoint-24500/unet/config.json b/checkpoint-24500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-24500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-24500/unet/diffusion_pytorch_model.bin b/checkpoint-24500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-24500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-24500/unet_ema/config.json b/checkpoint-24500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..8e789715c79b83b65224e185bdf7fa44c3d58cbd --- /dev/null +++ b/checkpoint-24500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 24500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-24500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-24500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ab7c1bfbca9aebab0e117bea8e21dcf2ffa16b --- /dev/null +++ b/checkpoint-24500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6ea6fa33d4c15f22b8f5d99bb22bd3271193194f81312fd95715713606582ee +size 1205673685 diff --git a/checkpoint-2500/optimizer.bin b/checkpoint-2500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..5a56fe5af86a75b516ec772ef46d90ed0c31bc1b --- /dev/null +++ b/checkpoint-2500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2c412068e0e4953771a196d5b00d8012a0faac1ef2de6b36221e3194cbeaa7dd +size 2411350917 diff --git a/checkpoint-2500/random_states_0.pkl b/checkpoint-2500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..695352a37b550311d04dbc159a06d2589f769356 --- /dev/null +++ b/checkpoint-2500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a22a812b40c503ea16eb83d55d79964d2c98fa1fc2323be6466929a4fd811180 +size 14663 diff --git a/checkpoint-2500/unet/config.json b/checkpoint-2500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-2500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-2500/unet/diffusion_pytorch_model.bin b/checkpoint-2500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-2500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-2500/unet_ema/config.json b/checkpoint-2500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..5f246997ddd82eb53a93c7946a402a0c0ba33a34 --- /dev/null +++ b/checkpoint-2500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 2500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-2500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-2500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..41a1d2fbcf599d4d3ab0e9fecc16ffb19cc9d805 --- /dev/null +++ b/checkpoint-2500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d644005d4c6990c927151347906efeaaba1187f07b65dbe6face60d17da49bd5 +size 1205673685 diff --git a/checkpoint-25000/optimizer.bin b/checkpoint-25000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..342f075a912df4515d64ee8b60e27d8233d3c49a --- /dev/null +++ b/checkpoint-25000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03619efdc0ed717e442161e4eaec2feaad1aa6644c8b8dfa68db1251bd42ce3a +size 2411350917 diff --git a/checkpoint-25000/random_states_0.pkl b/checkpoint-25000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..05cd3559779592ccafeec6babc5c8d5c34aeab4a --- /dev/null +++ b/checkpoint-25000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a9574a5a796b0f35a7fb64bedd8ef75d5b082fa6937bf1d4996f0bcb96a9d54a +size 14663 diff --git a/checkpoint-25000/unet/config.json b/checkpoint-25000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-25000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-25000/unet/diffusion_pytorch_model.bin b/checkpoint-25000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-25000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-25000/unet_ema/config.json b/checkpoint-25000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..951e423cd01244b4831229c726093b1b7112edaf --- /dev/null +++ b/checkpoint-25000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 25000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-25000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-25000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..4b85fdb30753a36e9a680abdaaf3b6e49575c77a --- /dev/null +++ b/checkpoint-25000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e97a1e4a81b0b188cc20eb69254798a7737dbe06278dd81b868b1aa0e9664fe3 +size 1205673685 diff --git a/checkpoint-25500/optimizer.bin b/checkpoint-25500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..8c7990cc43136b15510f4c5ef73da9a71ebc62f0 --- /dev/null +++ b/checkpoint-25500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02793f98e73a45e3c610560483375f54ad9bd3881c179004721c6313224f523f +size 2411350917 diff --git a/checkpoint-25500/random_states_0.pkl b/checkpoint-25500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..e2083ecf6314b37259fa9f9e5be99b7cf86e8916 --- /dev/null +++ b/checkpoint-25500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7d103d9298208d03b251b703d8fd79e5c32b9d38046624af65cfef1c6afb4d36 +size 14663 diff --git a/checkpoint-25500/unet/config.json b/checkpoint-25500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-25500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-25500/unet/diffusion_pytorch_model.bin b/checkpoint-25500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-25500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-25500/unet_ema/config.json b/checkpoint-25500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..0ab57babd79f778d26b8457161a348eed08e0894 --- /dev/null +++ b/checkpoint-25500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 25500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-25500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-25500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..7b76ee0fe997991b0ad7737f5d15efc2170e8c79 --- /dev/null +++ b/checkpoint-25500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4298ff803889014815e66d3380e77204fe05b1dd4fac3581d174e12efc59296 +size 1205673685 diff --git a/checkpoint-3000/optimizer.bin b/checkpoint-3000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..a55db15bb02e1276091b9e51005d6a7f70586430 --- /dev/null +++ b/checkpoint-3000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2ca208da8b35d06e8f6ac470c03d31d48510f553c37aa56215ea2a1d6c04e9eb +size 2411350917 diff --git a/checkpoint-3000/random_states_0.pkl b/checkpoint-3000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..1269530b21585ca5f4b8a5ae39139dadd901cfba --- /dev/null +++ b/checkpoint-3000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:152e5b0d99e4b965ffdb5d953f563dffcc722e9dfeab9dcc819a55d72f4e1fd4 +size 14663 diff --git a/checkpoint-3000/unet/config.json b/checkpoint-3000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-3000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-3000/unet/diffusion_pytorch_model.bin b/checkpoint-3000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-3000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-3000/unet_ema/config.json b/checkpoint-3000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..0af0b50df6e4eec322b2af3da7a9177414976f40 --- /dev/null +++ b/checkpoint-3000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 3000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-3000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-3000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..2cea7b6f108f2f219e2b0362ce60c55acdfec2a4 --- /dev/null +++ b/checkpoint-3000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:982d28c71ea3d4668684fd334788e3bed7572fe5fcbcd88420ced54ae6593bb2 +size 1205673685 diff --git a/checkpoint-3500/optimizer.bin b/checkpoint-3500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..f4845a536090ba259100612e41ce3b812daf6263 --- /dev/null +++ b/checkpoint-3500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:262494317cbe8ea4005edd96baaf1fef2566550b616d39eaaf93c2b6396bed07 +size 2411350917 diff --git a/checkpoint-3500/random_states_0.pkl b/checkpoint-3500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..79b30d7fea0e41b934bc79cd9d6d9e333ab66fcd --- /dev/null +++ b/checkpoint-3500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3f3e7f51e564aee03b08fd53adcc2d7836d87aa4b005d30ed920a6d1d6134ee4 +size 14663 diff --git a/checkpoint-3500/unet/config.json b/checkpoint-3500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-3500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-3500/unet/diffusion_pytorch_model.bin b/checkpoint-3500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-3500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-3500/unet_ema/config.json b/checkpoint-3500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..c1121c6d81356941a6f19cb3b8ef694ff629496b --- /dev/null +++ b/checkpoint-3500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 3500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-3500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-3500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..c2622aa3580a12ecea7e2a6bb8b371d09542e864 --- /dev/null +++ b/checkpoint-3500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d1d0e43e9b5472d1ff7d407bd50268f1621cf1084d690925ed091d6aeec723d +size 1205673685 diff --git a/checkpoint-4000/optimizer.bin b/checkpoint-4000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..b0a3606ddf3c96c50c9612cef8596a8fea12bed9 --- /dev/null +++ b/checkpoint-4000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb73702f3dde016668d4f727f8807ea1187f92e587686017959a4bce1e4a513d +size 2411350917 diff --git a/checkpoint-4000/random_states_0.pkl b/checkpoint-4000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..8bed8ba259909e5ae30ebb3b6f86c770c76e1d20 --- /dev/null +++ b/checkpoint-4000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f082671efc2fbcc7a2ece7d747b5c56a2aec1cebfed53597e1d199c3a60e5cda +size 14663 diff --git a/checkpoint-4000/unet/config.json b/checkpoint-4000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-4000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-4000/unet/diffusion_pytorch_model.bin b/checkpoint-4000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-4000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-4000/unet_ema/config.json b/checkpoint-4000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..34ef7108fb05952433a72ead3496e437bd126110 --- /dev/null +++ b/checkpoint-4000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 4000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-4000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-4000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..65467df5caedc5d67735373b11c1f4a44aff7cf1 --- /dev/null +++ b/checkpoint-4000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e49d7b5b1863108baf19bbab847735be19ca6470a3637e89e68609d48895a603 +size 1205673685 diff --git a/checkpoint-4500/optimizer.bin b/checkpoint-4500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..4385746c5c21e4c8102cd272548dfa9c62c8bd03 --- /dev/null +++ b/checkpoint-4500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:78a2d7f0efa2a887191677f51b62af1a2044757ca361bb3514d4493e879ef524 +size 2411350917 diff --git a/checkpoint-4500/random_states_0.pkl b/checkpoint-4500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..65a6ae3a146acdbe9bc1ad1f696311da2d2396ff --- /dev/null +++ b/checkpoint-4500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b7950f2db653d422de8a9e9b4b1e1e9fdf1ccf91d7feb7bbcde111a94f63e0f +size 14663 diff --git a/checkpoint-4500/unet/config.json b/checkpoint-4500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-4500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-4500/unet/diffusion_pytorch_model.bin b/checkpoint-4500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-4500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-4500/unet_ema/config.json b/checkpoint-4500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..8d953c764c7e393f9611fd136f9724d9499ca77c --- /dev/null +++ b/checkpoint-4500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 4500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-4500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-4500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..086533b95a1a7efafa2cfa7a09261057b87228b9 --- /dev/null +++ b/checkpoint-4500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ee35feea0e5489ba6d1008134aa3f034a6f1a5bbaabae5e52e0c905ea4408fe +size 1205673685 diff --git a/checkpoint-500/optimizer.bin b/checkpoint-500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..8b50e8fdf6f40cb4972bcf3bf43b86e62ea7d56d --- /dev/null +++ b/checkpoint-500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f322449c65b29826d65b0556e1c061ea424c1a5618d48fc21f2b267c05024e66 +size 2411350917 diff --git a/checkpoint-500/random_states_0.pkl b/checkpoint-500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..44a27227be92ec6fd3186ac5a5abf5a3f71c98ba --- /dev/null +++ b/checkpoint-500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43be2ca024fc7b1154ad40c67b8acfab5c13e0aca8453404a398909410b3261a +size 14663 diff --git a/checkpoint-500/unet/config.json b/checkpoint-500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-500/unet/diffusion_pytorch_model.bin b/checkpoint-500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-500/unet_ema/config.json b/checkpoint-500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..a61a5cfe7ff104403a8eb4561397f9082c81713f --- /dev/null +++ b/checkpoint-500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..d9acf8bc869362b57fe13ad2ca0a5a3641bea61d --- /dev/null +++ b/checkpoint-500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a610c43910f89b904e2695c02a006a26fad7b3e276ca9e011694150d5e6845e7 +size 1205673685 diff --git a/checkpoint-5000/optimizer.bin b/checkpoint-5000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..d897362ad2120bd17a07b87b95963a85e92b7850 --- /dev/null +++ b/checkpoint-5000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:544d42d978b58874253bba11150758fd611f21a4a7141be3f8a3b9a098552b9c +size 2411350917 diff --git a/checkpoint-5000/random_states_0.pkl b/checkpoint-5000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..4f45a5c05298389fa110d5290c59b3052cd830c7 --- /dev/null +++ b/checkpoint-5000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb27537aa6da2f8e02df5193c5ad0e26aa729e9f94b6e67ed46f31a9886f01be +size 14663 diff --git a/checkpoint-5000/unet/config.json b/checkpoint-5000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-5000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-5000/unet/diffusion_pytorch_model.bin b/checkpoint-5000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-5000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-5000/unet_ema/config.json b/checkpoint-5000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..a5b2b197c9845043053b91610eb0b90c2b26051f --- /dev/null +++ b/checkpoint-5000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 5000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-5000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-5000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..1e415e3df35ea5f500c378e8aeac0832c1a324fa --- /dev/null +++ b/checkpoint-5000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2017718ef6674e1339591491ff060b32b7b896842cfcf3a058213abfc5846b47 +size 1205673685 diff --git a/checkpoint-5500/optimizer.bin b/checkpoint-5500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..506e9ded8f8c6ff15a6106ab48b74d4cd41902b5 --- /dev/null +++ b/checkpoint-5500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22b82743b92291968bb0e809ccecb52fc21dc4100e42419cdaf2d2d361355d01 +size 2411350917 diff --git a/checkpoint-5500/random_states_0.pkl b/checkpoint-5500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..2e435d2af410093fca5a09e486eee6ca063e7599 --- /dev/null +++ b/checkpoint-5500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd0ab0909f37730b6e45f8b45917accc139c50e46e73a998da758050f8140262 +size 14663 diff --git a/checkpoint-5500/unet/config.json b/checkpoint-5500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-5500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-5500/unet/diffusion_pytorch_model.bin b/checkpoint-5500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-5500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-5500/unet_ema/config.json b/checkpoint-5500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..9c5554a4dbf04964850d673ff0c04f525a7303df --- /dev/null +++ b/checkpoint-5500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 5500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-5500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-5500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..01a306e8dc97803e0d6e9541dfb4ac3b2cf782e2 --- /dev/null +++ b/checkpoint-5500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e96fdd4f2d0b65f2eb61cf984779764a91a9a1f60f3324b1d6ca054fea098f46 +size 1205673685 diff --git a/checkpoint-6000/optimizer.bin b/checkpoint-6000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..4e5711c274031be12e6d85a7c356ba5b4ff5864a --- /dev/null +++ b/checkpoint-6000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd7ea511065a45ac4fd0e48d35c70ea1c79329f0b94cfa51fa9622257a272185 +size 2411350917 diff --git a/checkpoint-6000/random_states_0.pkl b/checkpoint-6000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..18d57d69d676024f2521d520985d014b2400fc7a --- /dev/null +++ b/checkpoint-6000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b845d6d645012fc073aac313eba75b719bbaf2a82f1e5b97601af07cea264a44 +size 14663 diff --git a/checkpoint-6000/unet/config.json b/checkpoint-6000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-6000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-6000/unet/diffusion_pytorch_model.bin b/checkpoint-6000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-6000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-6000/unet_ema/config.json b/checkpoint-6000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..57a998cce99274015f131a3aaf410c4f0a9b76c0 --- /dev/null +++ b/checkpoint-6000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 6000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-6000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-6000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..b601c256df44caf3af7c0e360fbdcd6d43e13d3d --- /dev/null +++ b/checkpoint-6000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e7d435758180216dd49b0adeee087ef9eebd67055a6f807ab0c302dc9897298 +size 1205673685 diff --git a/checkpoint-6500/optimizer.bin b/checkpoint-6500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..fffc6ecd2dd4893a47e2e89136241cbab77430ec --- /dev/null +++ b/checkpoint-6500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8df826d1bb6dfd235748b226077b591b57e71d7203edd4847f0461295c94269f +size 2411350917 diff --git a/checkpoint-6500/random_states_0.pkl b/checkpoint-6500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..61e32678159511c0a18392b9192ae226db2a7312 --- /dev/null +++ b/checkpoint-6500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2dc7b326ab6650970fbb6871aeba4188541da1d9e50e7c4a6df2cbec3b1d88a4 +size 14663 diff --git a/checkpoint-6500/unet/config.json b/checkpoint-6500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-6500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-6500/unet/diffusion_pytorch_model.bin b/checkpoint-6500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-6500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-6500/unet_ema/config.json b/checkpoint-6500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..318593f11f3c6d62fe8ecc5fec77b01d6afc7b76 --- /dev/null +++ b/checkpoint-6500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 6500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-6500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-6500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..e530c21330f3aca4f4f96f1d54476f221b0124de --- /dev/null +++ b/checkpoint-6500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be5c7c3065bf9ca72041f599104fa3e09c7ca9d5235119a0c174ae2511d3b472 +size 1205673685 diff --git a/checkpoint-7000/optimizer.bin b/checkpoint-7000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..0db0f721274cf62e1c35c3bab6a815d1a8797963 --- /dev/null +++ b/checkpoint-7000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:072510df9c7c97993c37a0980c003ef73f6b55cdaddb5a849fcc6f04f1fdc0a7 +size 2411350917 diff --git a/checkpoint-7000/random_states_0.pkl b/checkpoint-7000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..21d72d7b5bbccf9445fc4b3ece6ab77d76f3c9d2 --- /dev/null +++ b/checkpoint-7000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d58c2d75a1f95fa29d5eae439d031ebaf576d2d15bf5d1b2ad66cc38f4dbcb6 +size 14663 diff --git a/checkpoint-7000/unet/config.json b/checkpoint-7000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-7000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-7000/unet/diffusion_pytorch_model.bin b/checkpoint-7000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-7000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-7000/unet_ema/config.json b/checkpoint-7000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..b4cc126928cc4f833e7b5449ba8dc46ea7e14a77 --- /dev/null +++ b/checkpoint-7000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 7000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-7000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-7000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..ab36abfc477616091637979a609f8877307da515 --- /dev/null +++ b/checkpoint-7000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fd6a3c0fc1edcbdf8d821704bc3cc58b47babef3822e25a6fe1322237991035 +size 1205673685 diff --git a/checkpoint-7500/optimizer.bin b/checkpoint-7500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..b61696b5c7ae542a6c0fc0de658a674812630043 --- /dev/null +++ b/checkpoint-7500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3365a67bae374aa80b1c09aba42e9406971a8695fabb1e467ff824ed1d3a30e +size 2411350917 diff --git a/checkpoint-7500/random_states_0.pkl b/checkpoint-7500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..0686542c85e1d4708f9ac208baee5fe0203071cc --- /dev/null +++ b/checkpoint-7500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e050eb8bfef6b4031079967bfb33b98be894577348700b44c3f08275ea42ddc +size 14663 diff --git a/checkpoint-7500/unet/config.json b/checkpoint-7500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-7500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-7500/unet/diffusion_pytorch_model.bin b/checkpoint-7500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-7500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-7500/unet_ema/config.json b/checkpoint-7500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..59a6a4666af08b7c96f0e40d3b784b6210cf7bf7 --- /dev/null +++ b/checkpoint-7500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 7500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-7500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-7500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..29ae3fe7195aec9051fa6b2d0688301e04f096aa --- /dev/null +++ b/checkpoint-7500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d71dcc2a604b6e027b5f9a848a1f50eae34c02ee66e9b3fae1c4b6c76a1a1964 +size 1205673685 diff --git a/checkpoint-8000/optimizer.bin b/checkpoint-8000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..2d885673138fb39b301723506704d119c0eeb110 --- /dev/null +++ b/checkpoint-8000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a042d8be5414d9d94302757e32f531f4d9321acd79ed25483d933e3828806a6e +size 2411350917 diff --git a/checkpoint-8000/random_states_0.pkl b/checkpoint-8000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..f5ff54ae2432a3e34a3b8ce15cb6dbda1a991adb --- /dev/null +++ b/checkpoint-8000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09b483f914528c0ca2d4185df5ef29af67c84c488870ccd4b323ed7195ba46db +size 14663 diff --git a/checkpoint-8000/unet/config.json b/checkpoint-8000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-8000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-8000/unet/diffusion_pytorch_model.bin b/checkpoint-8000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-8000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-8000/unet_ema/config.json b/checkpoint-8000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..279edd64f27a304a11951a58bf410e678e904610 --- /dev/null +++ b/checkpoint-8000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 8000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-8000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-8000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..a1de5a981a37608e20ad48f777eb32c0902d8eb3 --- /dev/null +++ b/checkpoint-8000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c4e757d8bc8a88b55d807a54db765bde299a3c1da5ba93a5a0d49581fcf93ee7 +size 1205673685 diff --git a/checkpoint-8500/optimizer.bin b/checkpoint-8500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..668ba0877c77cb5e389cc282aa4ef6f8e629b50e --- /dev/null +++ b/checkpoint-8500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0ea452c1b8695bddc963962a4d9b00659767a1652896008ae25c87cf8b6804b +size 2411350917 diff --git a/checkpoint-8500/random_states_0.pkl b/checkpoint-8500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..a4dcfc9d093faaa8f44d4da0f63c6ac355e4e943 --- /dev/null +++ b/checkpoint-8500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:21a3bec345c9d296646f1f26b841940cb70660931d70201820a3a27fc2bb1895 +size 14663 diff --git a/checkpoint-8500/unet/config.json b/checkpoint-8500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-8500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-8500/unet/diffusion_pytorch_model.bin b/checkpoint-8500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-8500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-8500/unet_ema/config.json b/checkpoint-8500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..941896065b6c1aa990198743dfb963321bb4e932 --- /dev/null +++ b/checkpoint-8500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 8500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-8500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-8500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..2a083f81230f193a9370c0c789e69706f67d13cc --- /dev/null +++ b/checkpoint-8500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c78d13435e9b3299fc853a025e1dbab72011203f742bc2c9f5b0e0165fbd447 +size 1205673685 diff --git a/checkpoint-9000/optimizer.bin b/checkpoint-9000/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..055c120d5c5477afb6d7ef6aba8b0cbd8ac6004b --- /dev/null +++ b/checkpoint-9000/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2eb3f2c5671b870f5980072096935ffa3e9f33cc91d4e38e3f79500c7f331f7 +size 2411350917 diff --git a/checkpoint-9000/random_states_0.pkl b/checkpoint-9000/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..9fa9cb3582242144d3a604883b0f8cc1181d2266 --- /dev/null +++ b/checkpoint-9000/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3b830dc992ef3ec00d2bd710557dc53bf8b861212a994a6c1969e3690e5474c +size 14663 diff --git a/checkpoint-9000/unet/config.json b/checkpoint-9000/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-9000/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-9000/unet/diffusion_pytorch_model.bin b/checkpoint-9000/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-9000/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-9000/unet_ema/config.json b/checkpoint-9000/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..b41b5914b0305609ccd705415d80f32c6d715f6f --- /dev/null +++ b/checkpoint-9000/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 9000, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-9000/unet_ema/diffusion_pytorch_model.bin b/checkpoint-9000/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..081b0a4b1175610cbe0022b6f3c7b86a08097207 --- /dev/null +++ b/checkpoint-9000/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1bc5e817459e94b5311055ec12c766b6e2dbfee31f69c1f1f7df10fc43dcc23 +size 1205673685 diff --git a/checkpoint-9500/optimizer.bin b/checkpoint-9500/optimizer.bin new file mode 100644 index 0000000000000000000000000000000000000000..71bffb379f470690d00e65462c5f29f813fee576 --- /dev/null +++ b/checkpoint-9500/optimizer.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5aa06f1c06c48983a13f4adfd7b2be9a57ac9271fa2bf444b9728e42f07493c2 +size 2411350917 diff --git a/checkpoint-9500/random_states_0.pkl b/checkpoint-9500/random_states_0.pkl new file mode 100644 index 0000000000000000000000000000000000000000..22ae96b3bc1b8ec21d809dbcb22d3cd8167ce4d7 --- /dev/null +++ b/checkpoint-9500/random_states_0.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e7beede340d9c3c262a3bc006fdedef9302fcd10d8503d3ca54993e57dd21c4 +size 14663 diff --git a/checkpoint-9500/unet/config.json b/checkpoint-9500/unet/config.json new file mode 100644 index 0000000000000000000000000000000000000000..4410c7e5789aaa7423d4e05ca87e345e6e0ddc47 --- /dev/null +++ b/checkpoint-9500/unet/config.json @@ -0,0 +1,21 @@ +{ + "_class_name": "AutoencoderKL", + "_diffusers_version": "0.13.0.dev0", + "act_fn": "silu", + "block_out_channels": [ + 64 + ], + "down_block_types": [ + "DownEncoderBlock2D" + ], + "in_channels": 3, + "latent_channels": 3, + "layers_per_block": 1, + "norm_num_groups": 32, + "out_channels": 3, + "sample_size": 32, + "scaling_factor": 0.18215, + "up_block_types": [ + "UpDecoderBlock2D" + ] +} diff --git a/checkpoint-9500/unet/diffusion_pytorch_model.bin b/checkpoint-9500/unet/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..73b45f951b41004997b8df1b0fe97b0c677dee17 --- /dev/null +++ b/checkpoint-9500/unet/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d +size 2281457 diff --git a/checkpoint-9500/unet_ema/config.json b/checkpoint-9500/unet_ema/config.json new file mode 100644 index 0000000000000000000000000000000000000000..aaa646031f0943ef87c1faf8e307542001423725 --- /dev/null +++ b/checkpoint-9500/unet_ema/config.json @@ -0,0 +1,30 @@ +{ + "_class_name": "Transformer2DModel", + "_diffusers_version": "0.13.0.dev0", + "activation_fn": "geglu", + "attention_bias": false, + "attention_head_dim": 88, + "cross_attention_dim": null, + "decay": 0.9999, + "dropout": 0.0, + "in_channels": 3, + "inv_gamma": 1.0, + "min_decay": 0.0, + "norm_elementwise_affine": true, + "norm_num_groups": 3, + "norm_type": "ada_norm", + "num_attention_heads": 16, + "num_embeds_ada_norm": 1000, + "num_layers": 8, + "num_vector_embeds": null, + "only_cross_attention": false, + "optimization_step": 9500, + "out_channels": 3, + "patch_size": 8, + "power": 0.75, + "sample_size": 64, + "upcast_attention": false, + "update_after_step": 0, + "use_ema_warmup": true, + "use_linear_projection": false +} diff --git a/checkpoint-9500/unet_ema/diffusion_pytorch_model.bin b/checkpoint-9500/unet_ema/diffusion_pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..9502cd7aa84108f4a8eba8becf92e6f82db84dab --- /dev/null +++ b/checkpoint-9500/unet_ema/diffusion_pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd40e99108f334f7031e8fafda164a392006ef5643baed2a578a588d899af361 +size 1205673685 diff --git a/transformer/diffusion_pytorch_model.bin b/transformer/diffusion_pytorch_model.bin index d3206326d8ff5c34ca26bd3c4c97ee21e89b6536..ef1f3a03cb20b1e1ab055b243a421a3a4dc244f3 100644 --- a/transformer/diffusion_pytorch_model.bin +++ b/transformer/diffusion_pytorch_model.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:03fd5d79cf26c36c86ac384dfe32af201143ea5b6321e7aa73faab90ed476f7c +oid sha256:527d33969f0382c88e15e9cfa79c0e4263e2b8dda99e0a81a87f1071e7c61a4b size 1205675285