End of training
Browse files- image_encoder/config.json +1 -1
- image_encoder/model.safetensors +2 -2
- logs/instruct-pix2pix/1727950434.3417614/events.out.tfevents.1727950434.plato2.1367257.1 +3 -0
- logs/instruct-pix2pix/1727950434.3475065/hparams.yml +51 -0
- logs/instruct-pix2pix/events.out.tfevents.1727950434.plato2.1367257.0 +3 -0
- model_index.json +1 -1
- unet/diffusion_pytorch_model.safetensors +1 -1
image_encoder/config.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
{
|
| 2 |
"_name_or_path": "openai/clip-vit-large-patch14",
|
| 3 |
"architectures": [
|
| 4 |
-
"
|
| 5 |
],
|
| 6 |
"attention_dropout": 0.0,
|
| 7 |
"dropout": 0.0,
|
|
|
|
| 1 |
{
|
| 2 |
"_name_or_path": "openai/clip-vit-large-patch14",
|
| 3 |
"architectures": [
|
| 4 |
+
"CLIPVisionModelWithProjection"
|
| 5 |
],
|
| 6 |
"attention_dropout": 0.0,
|
| 7 |
"dropout": 0.0,
|
image_encoder/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:77b33d2a3a643650857672e880ccf73adbaf114fbbadec36d142ee9d48af7e20
|
| 3 |
+
size 1215912728
|
logs/instruct-pix2pix/1727950434.3417614/events.out.tfevents.1727950434.plato2.1367257.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:739711b73126b26ef861070e7c95bda8738da8b0c8e41aab875be4b07de00f5e
|
| 3 |
+
size 2355
|
logs/instruct-pix2pix/1727950434.3475065/hparams.yml
ADDED
|
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
adam_beta1: 0.9
|
| 2 |
+
adam_beta2: 0.999
|
| 3 |
+
adam_epsilon: 1.0e-08
|
| 4 |
+
adam_weight_decay: 0.01
|
| 5 |
+
allow_tf32: false
|
| 6 |
+
cache_dir: null
|
| 7 |
+
center_crop: false
|
| 8 |
+
checkpointing_steps: 5000
|
| 9 |
+
checkpoints_total_limit: 1
|
| 10 |
+
conditioning_dropout_prob: 0.05
|
| 11 |
+
dataloader_num_workers: 0
|
| 12 |
+
dataset_config_name: null
|
| 13 |
+
dataset_name: osunlp/MagicBrush
|
| 14 |
+
enable_xformers_memory_efficient_attention: false
|
| 15 |
+
gradient_accumulation_steps: 4
|
| 16 |
+
gradient_checkpointing: true
|
| 17 |
+
hub_model_id: null
|
| 18 |
+
hub_token: null
|
| 19 |
+
learning_rate: 5.0e-05
|
| 20 |
+
local_rank: 0
|
| 21 |
+
logging_dir: logs
|
| 22 |
+
lr_scheduler: constant
|
| 23 |
+
lr_warmup_steps: 0
|
| 24 |
+
mask_img_column: mask_img
|
| 25 |
+
max_grad_norm: 1.0
|
| 26 |
+
max_train_samples: null
|
| 27 |
+
max_train_steps: 100
|
| 28 |
+
mixed_precision: 'no'
|
| 29 |
+
non_ema_revision: null
|
| 30 |
+
num_train_epochs: 1
|
| 31 |
+
num_validation_images: 4
|
| 32 |
+
output_dir: instruct-pix2pix-model
|
| 33 |
+
pretrained_model_name_or_path: ''
|
| 34 |
+
push_to_hub: true
|
| 35 |
+
random_flip: true
|
| 36 |
+
report_to: tensorboard
|
| 37 |
+
resolution: 224
|
| 38 |
+
resume_from_checkpoint: null
|
| 39 |
+
revision: null
|
| 40 |
+
scale_lr: false
|
| 41 |
+
seed: 42
|
| 42 |
+
source_img_column: source_img
|
| 43 |
+
target_img_column: target_img
|
| 44 |
+
train_batch_size: 2
|
| 45 |
+
train_data_dir: null
|
| 46 |
+
use_8bit_adam: false
|
| 47 |
+
use_ema: true
|
| 48 |
+
val_image_url: null
|
| 49 |
+
val_image_url2: null
|
| 50 |
+
validation_epochs: 1
|
| 51 |
+
variant: null
|
logs/instruct-pix2pix/events.out.tfevents.1727950434.plato2.1367257.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:67cfabad744d1b04514addb862f5dcf987b0fbf762bb7cb1396a55751ac4f012
|
| 3 |
+
size 4888
|
model_index.json
CHANGED
|
@@ -8,7 +8,7 @@
|
|
| 8 |
],
|
| 9 |
"image_encoder": [
|
| 10 |
"transformers",
|
| 11 |
-
"
|
| 12 |
],
|
| 13 |
"requires_safety_checker": true,
|
| 14 |
"safety_checker": [
|
|
|
|
| 8 |
],
|
| 9 |
"image_encoder": [
|
| 10 |
"transformers",
|
| 11 |
+
"CLIPVisionModelWithProjection"
|
| 12 |
],
|
| 13 |
"requires_safety_checker": true,
|
| 14 |
"safety_checker": [
|
unet/diffusion_pytorch_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 3438213624
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:df424200b5b26974d3160690f6fa98d1af459eb01b51e7530b9f67f89c04ed66
|
| 3 |
size 3438213624
|