Datasets:
Upload folder using huggingface_hub
Browse files- data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/checkpoint-15000/optimizer.bin +3 -0
- data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/checkpoint-15000/random_states_0.pkl +3 -0
- data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/checkpoint-15000/scaler.pt +3 -0
- data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/checkpoint-15000/scheduler.bin +3 -0
- data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/checkpoint-15000/unet/config.json +68 -0
- data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/checkpoint-15000/unet/diffusion_pytorch_model.safetensors +3 -0
- data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/checkpoint-15000/unet_ema/config.json +75 -0
- data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/checkpoint-15000/unet_ema/diffusion_pytorch_model.safetensors +3 -0
- data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/logs/instruct-pix2pix/1746651539.5363052/events.out.tfevents.1746651539.bc051.594383.1 +3 -0
- data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/logs/instruct-pix2pix/1746651539.5406199/hparams.yml +51 -0
- data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/logs/instruct-pix2pix/events.out.tfevents.1746651539.bc051.594383.0 +3 -0
data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/checkpoint-15000/optimizer.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6d0e75a95a2fe5dd13a573b9d1b77014a0160c63c08c3011c2d59e4b85a4d346
|
| 3 |
+
size 6876842324
|
data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/checkpoint-15000/random_states_0.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:58447d8040f926ffd3353ca9c0209a53e11e9a0f8dbe54044fa08a0fc3848697
|
| 3 |
+
size 15124
|
data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/checkpoint-15000/scaler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2d31f9bef150e8dbb2982d04d6ce5e96c20f56651a94795370dfca811943c729
|
| 3 |
+
size 988
|
data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/checkpoint-15000/scheduler.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e4b6fed06972ca6d8f84975e458de6d9d6e977ceb58cb9472f69fec51c80e6ee
|
| 3 |
+
size 1000
|
data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/checkpoint-15000/unet/config.json
ADDED
|
@@ -0,0 +1,68 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_class_name": "UNet2DConditionModel",
|
| 3 |
+
"_diffusers_version": "0.33.0.dev0",
|
| 4 |
+
"_name_or_path": "timbrooks/instruct-pix2pix",
|
| 5 |
+
"act_fn": "silu",
|
| 6 |
+
"addition_embed_type": null,
|
| 7 |
+
"addition_embed_type_num_heads": 64,
|
| 8 |
+
"addition_time_embed_dim": null,
|
| 9 |
+
"attention_head_dim": 8,
|
| 10 |
+
"attention_type": "default",
|
| 11 |
+
"block_out_channels": [
|
| 12 |
+
320,
|
| 13 |
+
640,
|
| 14 |
+
1280,
|
| 15 |
+
1280
|
| 16 |
+
],
|
| 17 |
+
"center_input_sample": false,
|
| 18 |
+
"class_embed_type": null,
|
| 19 |
+
"class_embeddings_concat": false,
|
| 20 |
+
"conv_in_kernel": 3,
|
| 21 |
+
"conv_out_kernel": 3,
|
| 22 |
+
"cross_attention_dim": 768,
|
| 23 |
+
"cross_attention_norm": null,
|
| 24 |
+
"down_block_types": [
|
| 25 |
+
"CrossAttnDownBlock2D",
|
| 26 |
+
"CrossAttnDownBlock2D",
|
| 27 |
+
"CrossAttnDownBlock2D",
|
| 28 |
+
"DownBlock2D"
|
| 29 |
+
],
|
| 30 |
+
"downsample_padding": 1,
|
| 31 |
+
"dropout": 0.0,
|
| 32 |
+
"dual_cross_attention": false,
|
| 33 |
+
"encoder_hid_dim": null,
|
| 34 |
+
"encoder_hid_dim_type": null,
|
| 35 |
+
"flip_sin_to_cos": true,
|
| 36 |
+
"freq_shift": 0,
|
| 37 |
+
"in_channels": 8,
|
| 38 |
+
"layers_per_block": 2,
|
| 39 |
+
"mid_block_only_cross_attention": null,
|
| 40 |
+
"mid_block_scale_factor": 1,
|
| 41 |
+
"mid_block_type": "UNetMidBlock2DCrossAttn",
|
| 42 |
+
"norm_eps": 1e-05,
|
| 43 |
+
"norm_num_groups": 32,
|
| 44 |
+
"num_attention_heads": null,
|
| 45 |
+
"num_class_embeds": null,
|
| 46 |
+
"only_cross_attention": false,
|
| 47 |
+
"out_channels": 4,
|
| 48 |
+
"projection_class_embeddings_input_dim": null,
|
| 49 |
+
"resnet_out_scale_factor": 1.0,
|
| 50 |
+
"resnet_skip_time_act": false,
|
| 51 |
+
"resnet_time_scale_shift": "default",
|
| 52 |
+
"reverse_transformer_layers_per_block": null,
|
| 53 |
+
"sample_size": 64,
|
| 54 |
+
"time_cond_proj_dim": null,
|
| 55 |
+
"time_embedding_act_fn": null,
|
| 56 |
+
"time_embedding_dim": null,
|
| 57 |
+
"time_embedding_type": "positional",
|
| 58 |
+
"timestep_post_act": null,
|
| 59 |
+
"transformer_layers_per_block": 1,
|
| 60 |
+
"up_block_types": [
|
| 61 |
+
"UpBlock2D",
|
| 62 |
+
"CrossAttnUpBlock2D",
|
| 63 |
+
"CrossAttnUpBlock2D",
|
| 64 |
+
"CrossAttnUpBlock2D"
|
| 65 |
+
],
|
| 66 |
+
"upcast_attention": false,
|
| 67 |
+
"use_linear_projection": false
|
| 68 |
+
}
|
data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/checkpoint-15000/unet/diffusion_pytorch_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0374612d3c80aa17f772501ed992b261a1564f647329d62c127e36bd8b6f76ae
|
| 3 |
+
size 3438213624
|
data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/checkpoint-15000/unet_ema/config.json
ADDED
|
@@ -0,0 +1,75 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_class_name": "UNet2DConditionModel",
|
| 3 |
+
"_diffusers_version": "0.33.0.dev0",
|
| 4 |
+
"_name_or_path": "timbrooks/instruct-pix2pix",
|
| 5 |
+
"act_fn": "silu",
|
| 6 |
+
"addition_embed_type": null,
|
| 7 |
+
"addition_embed_type_num_heads": 64,
|
| 8 |
+
"addition_time_embed_dim": null,
|
| 9 |
+
"attention_head_dim": 8,
|
| 10 |
+
"attention_type": "default",
|
| 11 |
+
"block_out_channels": [
|
| 12 |
+
320,
|
| 13 |
+
640,
|
| 14 |
+
1280,
|
| 15 |
+
1280
|
| 16 |
+
],
|
| 17 |
+
"center_input_sample": false,
|
| 18 |
+
"class_embed_type": null,
|
| 19 |
+
"class_embeddings_concat": false,
|
| 20 |
+
"conv_in_kernel": 3,
|
| 21 |
+
"conv_out_kernel": 3,
|
| 22 |
+
"cross_attention_dim": 768,
|
| 23 |
+
"cross_attention_norm": null,
|
| 24 |
+
"decay": 0.9999,
|
| 25 |
+
"down_block_types": [
|
| 26 |
+
"CrossAttnDownBlock2D",
|
| 27 |
+
"CrossAttnDownBlock2D",
|
| 28 |
+
"CrossAttnDownBlock2D",
|
| 29 |
+
"DownBlock2D"
|
| 30 |
+
],
|
| 31 |
+
"downsample_padding": 1,
|
| 32 |
+
"dropout": 0.0,
|
| 33 |
+
"dual_cross_attention": false,
|
| 34 |
+
"encoder_hid_dim": null,
|
| 35 |
+
"encoder_hid_dim_type": null,
|
| 36 |
+
"flip_sin_to_cos": true,
|
| 37 |
+
"freq_shift": 0,
|
| 38 |
+
"in_channels": 8,
|
| 39 |
+
"inv_gamma": 1.0,
|
| 40 |
+
"layers_per_block": 2,
|
| 41 |
+
"mid_block_only_cross_attention": null,
|
| 42 |
+
"mid_block_scale_factor": 1,
|
| 43 |
+
"mid_block_type": "UNetMidBlock2DCrossAttn",
|
| 44 |
+
"min_decay": 0.0,
|
| 45 |
+
"norm_eps": 1e-05,
|
| 46 |
+
"norm_num_groups": 32,
|
| 47 |
+
"num_attention_heads": null,
|
| 48 |
+
"num_class_embeds": null,
|
| 49 |
+
"only_cross_attention": false,
|
| 50 |
+
"optimization_step": 15000,
|
| 51 |
+
"out_channels": 4,
|
| 52 |
+
"power": 0.6666666666666666,
|
| 53 |
+
"projection_class_embeddings_input_dim": null,
|
| 54 |
+
"resnet_out_scale_factor": 1.0,
|
| 55 |
+
"resnet_skip_time_act": false,
|
| 56 |
+
"resnet_time_scale_shift": "default",
|
| 57 |
+
"reverse_transformer_layers_per_block": null,
|
| 58 |
+
"sample_size": 64,
|
| 59 |
+
"time_cond_proj_dim": null,
|
| 60 |
+
"time_embedding_act_fn": null,
|
| 61 |
+
"time_embedding_dim": null,
|
| 62 |
+
"time_embedding_type": "positional",
|
| 63 |
+
"timestep_post_act": null,
|
| 64 |
+
"transformer_layers_per_block": 1,
|
| 65 |
+
"up_block_types": [
|
| 66 |
+
"UpBlock2D",
|
| 67 |
+
"CrossAttnUpBlock2D",
|
| 68 |
+
"CrossAttnUpBlock2D",
|
| 69 |
+
"CrossAttnUpBlock2D"
|
| 70 |
+
],
|
| 71 |
+
"upcast_attention": false,
|
| 72 |
+
"update_after_step": 0,
|
| 73 |
+
"use_ema_warmup": false,
|
| 74 |
+
"use_linear_projection": false
|
| 75 |
+
}
|
data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/checkpoint-15000/unet_ema/diffusion_pytorch_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:56e1877137b0846043103d7b5207d4273165e6f1fc6a95c19bb20b6095b98f0c
|
| 3 |
+
size 3438213624
|
data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/logs/instruct-pix2pix/1746651539.5363052/events.out.tfevents.1746651539.bc051.594383.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4ec707446794554533d9a530f52e7944019a4339b99fcad245bdd053efee461e
|
| 3 |
+
size 2644
|
data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/logs/instruct-pix2pix/1746651539.5406199/hparams.yml
ADDED
|
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
adam_beta1: 0.9
|
| 2 |
+
adam_beta2: 0.999
|
| 3 |
+
adam_epsilon: 1.0e-08
|
| 4 |
+
adam_weight_decay: 0.01
|
| 5 |
+
allow_tf32: false
|
| 6 |
+
cache_dir: /projects01/VICTRE/elena.sizikova/code/skin/stable_diffusion_cache/
|
| 7 |
+
center_crop: false
|
| 8 |
+
checkpointing_steps: 5000
|
| 9 |
+
checkpoints_total_limit: 1
|
| 10 |
+
conditioning_dropout_prob: 0.05
|
| 11 |
+
dataloader_num_workers: 0
|
| 12 |
+
dataset_config_name: null
|
| 13 |
+
dataset_name: /projects01/VICTRE/elena.sizikova/ssynth_artifact/hf_datasets/hair_ssynth_smart_train
|
| 14 |
+
edit_prompt_column: edit_prompt
|
| 15 |
+
edited_image_column: edited_image
|
| 16 |
+
enable_xformers_memory_efficient_attention: true
|
| 17 |
+
gradient_accumulation_steps: 4
|
| 18 |
+
gradient_checkpointing: true
|
| 19 |
+
hub_model_id: null
|
| 20 |
+
hub_token: null
|
| 21 |
+
learning_rate: 5.0e-05
|
| 22 |
+
local_rank: 0
|
| 23 |
+
logging_dir: logs
|
| 24 |
+
lr_scheduler: constant
|
| 25 |
+
lr_warmup_steps: 0
|
| 26 |
+
max_grad_norm: 1.0
|
| 27 |
+
max_train_samples: null
|
| 28 |
+
max_train_steps: 15000
|
| 29 |
+
mixed_precision: fp16
|
| 30 |
+
non_ema_revision: null
|
| 31 |
+
num_train_epochs: 120
|
| 32 |
+
num_validation_images: 4
|
| 33 |
+
original_image_column: input_image
|
| 34 |
+
output_dir: /projects01/VICTRE/elena.sizikova/ssynth_artifact/models/finetune_instructpix2pix_hair_smart
|
| 35 |
+
pretrained_model_name_or_path: timbrooks/instruct-pix2pix
|
| 36 |
+
push_to_hub: false
|
| 37 |
+
random_flip: true
|
| 38 |
+
report_to: tensorboard
|
| 39 |
+
resolution: 512
|
| 40 |
+
resume_from_checkpoint: null
|
| 41 |
+
revision: null
|
| 42 |
+
scale_lr: false
|
| 43 |
+
seed: 42
|
| 44 |
+
train_batch_size: 4
|
| 45 |
+
train_data_dir: null
|
| 46 |
+
use_8bit_adam: false
|
| 47 |
+
use_ema: true
|
| 48 |
+
val_image_url: null
|
| 49 |
+
validation_epochs: 1
|
| 50 |
+
validation_prompt: null
|
| 51 |
+
variant: null
|
data/pretrained_models/diffusion/finetune_instructpix2pix_hair_smart/logs/instruct-pix2pix/events.out.tfevents.1746651539.bc051.594383.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8c741e4be60e9d2d49a9721f3466f8ca466e2164a9fb24ff1220882e26a3a876
|
| 3 |
+
size 734961
|