Upload 5 files
Browse files- feature_extractor/preprocessor_config.json +44 -0
- logs/instruct-pix2pix/1724603119.65034/events.out.tfevents.1724603119.d9-w7525-01.3098137.1 +3 -0
- logs/instruct-pix2pix/1724603119.6559958/hparams.yml +50 -0
- logs/instruct-pix2pix/events.out.tfevents.1724603119.d9-w7525-01.3098137.0 +3 -0
- model_index.json +38 -0
feature_extractor/preprocessor_config.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_valid_processor_keys": [
|
| 3 |
+
"images",
|
| 4 |
+
"do_resize",
|
| 5 |
+
"size",
|
| 6 |
+
"resample",
|
| 7 |
+
"do_center_crop",
|
| 8 |
+
"crop_size",
|
| 9 |
+
"do_rescale",
|
| 10 |
+
"rescale_factor",
|
| 11 |
+
"do_normalize",
|
| 12 |
+
"image_mean",
|
| 13 |
+
"image_std",
|
| 14 |
+
"do_convert_rgb",
|
| 15 |
+
"return_tensors",
|
| 16 |
+
"data_format",
|
| 17 |
+
"input_data_format"
|
| 18 |
+
],
|
| 19 |
+
"crop_size": {
|
| 20 |
+
"height": 224,
|
| 21 |
+
"width": 224
|
| 22 |
+
},
|
| 23 |
+
"do_center_crop": true,
|
| 24 |
+
"do_convert_rgb": true,
|
| 25 |
+
"do_normalize": true,
|
| 26 |
+
"do_rescale": true,
|
| 27 |
+
"do_resize": true,
|
| 28 |
+
"image_mean": [
|
| 29 |
+
0.48145466,
|
| 30 |
+
0.4578275,
|
| 31 |
+
0.40821073
|
| 32 |
+
],
|
| 33 |
+
"image_processor_type": "CLIPImageProcessor",
|
| 34 |
+
"image_std": [
|
| 35 |
+
0.26862954,
|
| 36 |
+
0.26130258,
|
| 37 |
+
0.27577711
|
| 38 |
+
],
|
| 39 |
+
"resample": 3,
|
| 40 |
+
"rescale_factor": 0.00392156862745098,
|
| 41 |
+
"size": {
|
| 42 |
+
"shortest_edge": 224
|
| 43 |
+
}
|
| 44 |
+
}
|
logs/instruct-pix2pix/1724603119.65034/events.out.tfevents.1724603119.d9-w7525-01.3098137.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:89c1b4f4fafcd1c50eaf692f165abdc2d092672da8fbe8e1de0280189fa23862
|
| 3 |
+
size 2444
|
logs/instruct-pix2pix/1724603119.6559958/hparams.yml
ADDED
|
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
adam_beta1: 0.9
|
| 2 |
+
adam_beta2: 0.999
|
| 3 |
+
adam_epsilon: 1.0e-08
|
| 4 |
+
adam_weight_decay: 0.01
|
| 5 |
+
allow_tf32: false
|
| 6 |
+
cache_dir: null
|
| 7 |
+
center_crop: false
|
| 8 |
+
checkpointing_steps: 5000
|
| 9 |
+
checkpoints_total_limit: 1
|
| 10 |
+
conditioning_dropout_prob: null
|
| 11 |
+
dataloader_num_workers: 0
|
| 12 |
+
dataset_config_name: null
|
| 13 |
+
dataset_name: isabelxiao1/ship_pairs7-train
|
| 14 |
+
edit_prompt_column: edit_prompt
|
| 15 |
+
edited_image_column: denoised_image
|
| 16 |
+
enable_xformers_memory_efficient_attention: true
|
| 17 |
+
gradient_accumulation_steps: 4
|
| 18 |
+
gradient_checkpointing: true
|
| 19 |
+
hub_model_id: null
|
| 20 |
+
hub_token: null
|
| 21 |
+
learning_rate: 5.0e-05
|
| 22 |
+
local_rank: -1
|
| 23 |
+
logging_dir: logs
|
| 24 |
+
lr_scheduler: constant
|
| 25 |
+
lr_warmup_steps: 500
|
| 26 |
+
max_grad_norm: 1.0
|
| 27 |
+
max_train_samples: null
|
| 28 |
+
max_train_steps: 20000
|
| 29 |
+
mixed_precision: fp16
|
| 30 |
+
non_ema_revision: null
|
| 31 |
+
num_train_epochs: 1334
|
| 32 |
+
num_validation_images: 4
|
| 33 |
+
original_image_column: original_image
|
| 34 |
+
output_dir: ./bayesian_search_results_ip2p_THREE/trial_11
|
| 35 |
+
pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
|
| 36 |
+
push_to_hub: false
|
| 37 |
+
random_flip: true
|
| 38 |
+
report_to: tensorboard
|
| 39 |
+
resolution: 256
|
| 40 |
+
resume_from_checkpoint: null
|
| 41 |
+
revision: null
|
| 42 |
+
scale_lr: false
|
| 43 |
+
seed: 42
|
| 44 |
+
train_batch_size: 2
|
| 45 |
+
train_data_dir: null
|
| 46 |
+
use_8bit_adam: false
|
| 47 |
+
use_ema: true
|
| 48 |
+
val_image_url: null
|
| 49 |
+
validation_epochs: 1
|
| 50 |
+
validation_prompt: Clean the image.
|
logs/instruct-pix2pix/events.out.tfevents.1724603119.d9-w7525-01.3098137.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a9c099e4f7b3e212da5fd308bb29a583e1458f2e90d44627db39fd8f795313b5
|
| 3 |
+
size 983578
|
model_index.json
ADDED
|
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_class_name": "StableDiffusionInstructPix2PixPipeline",
|
| 3 |
+
"_diffusers_version": "0.29.2",
|
| 4 |
+
"_name_or_path": "runwayml/stable-diffusion-v1-5",
|
| 5 |
+
"feature_extractor": [
|
| 6 |
+
"transformers",
|
| 7 |
+
"CLIPImageProcessor"
|
| 8 |
+
],
|
| 9 |
+
"image_encoder": [
|
| 10 |
+
null,
|
| 11 |
+
null
|
| 12 |
+
],
|
| 13 |
+
"requires_safety_checker": true,
|
| 14 |
+
"safety_checker": [
|
| 15 |
+
"stable_diffusion",
|
| 16 |
+
"StableDiffusionSafetyChecker"
|
| 17 |
+
],
|
| 18 |
+
"scheduler": [
|
| 19 |
+
"diffusers",
|
| 20 |
+
"PNDMScheduler"
|
| 21 |
+
],
|
| 22 |
+
"text_encoder": [
|
| 23 |
+
"transformers",
|
| 24 |
+
"CLIPTextModel"
|
| 25 |
+
],
|
| 26 |
+
"tokenizer": [
|
| 27 |
+
"transformers",
|
| 28 |
+
"CLIPTokenizer"
|
| 29 |
+
],
|
| 30 |
+
"unet": [
|
| 31 |
+
"diffusers",
|
| 32 |
+
"UNet2DConditionModel"
|
| 33 |
+
],
|
| 34 |
+
"vae": [
|
| 35 |
+
"diffusers",
|
| 36 |
+
"AutoencoderKL"
|
| 37 |
+
]
|
| 38 |
+
}
|