End of training
Browse files- README.md +4 -3
- feature_extractor/preprocessor_config.json +0 -17
- learned_embeds.safetensors +3 -0
- logs/textual_inversion/1719884653.746298/events.out.tfevents.1719884653.6a7922459928.1154.1 +3 -0
- logs/textual_inversion/1719884653.7487648/hparams.yml +48 -0
- logs/textual_inversion/events.out.tfevents.1719884653.6a7922459928.1154.0 +3 -0
- model_index.json +1 -1
- safety_checker/config.json +2 -2
- text_encoder/config.json +3 -3
- text_encoder/model.safetensors +2 -2
- tokenizer/added_tokens.json +3 -0
- tokenizer/special_tokens_map.json +1 -7
- tokenizer/tokenizer_config.json +8 -0
- unet/config.json +1 -1
- unet/diffusion_pytorch_model.safetensors +1 -1
- vae/config.json +1 -1
README.md
CHANGED
|
@@ -8,11 +8,12 @@ tags:
|
|
| 8 |
- diffusers-training
|
| 9 |
- stable-diffusion
|
| 10 |
- stable-diffusion-diffusers
|
| 11 |
-
- text-to-image
|
| 12 |
-
- dreambooth
|
| 13 |
-
- diffusers-training
|
| 14 |
- stable-diffusion
|
| 15 |
- stable-diffusion-diffusers
|
|
|
|
|
|
|
|
|
|
|
|
|
| 16 |
inference: true
|
| 17 |
instance_prompt: a photo of t@y
|
| 18 |
---
|
|
|
|
| 8 |
- diffusers-training
|
| 9 |
- stable-diffusion
|
| 10 |
- stable-diffusion-diffusers
|
|
|
|
|
|
|
|
|
|
| 11 |
- stable-diffusion
|
| 12 |
- stable-diffusion-diffusers
|
| 13 |
+
- text-to-image
|
| 14 |
+
- diffusers
|
| 15 |
+
- textual_inversion
|
| 16 |
+
- diffusers-training
|
| 17 |
inference: true
|
| 18 |
instance_prompt: a photo of t@y
|
| 19 |
---
|
feature_extractor/preprocessor_config.json
CHANGED
|
@@ -1,21 +1,4 @@
|
|
| 1 |
{
|
| 2 |
-
"_valid_processor_keys": [
|
| 3 |
-
"images",
|
| 4 |
-
"do_resize",
|
| 5 |
-
"size",
|
| 6 |
-
"resample",
|
| 7 |
-
"do_center_crop",
|
| 8 |
-
"crop_size",
|
| 9 |
-
"do_rescale",
|
| 10 |
-
"rescale_factor",
|
| 11 |
-
"do_normalize",
|
| 12 |
-
"image_mean",
|
| 13 |
-
"image_std",
|
| 14 |
-
"do_convert_rgb",
|
| 15 |
-
"return_tensors",
|
| 16 |
-
"data_format",
|
| 17 |
-
"input_data_format"
|
| 18 |
-
],
|
| 19 |
"crop_size": {
|
| 20 |
"height": 224,
|
| 21 |
"width": 224
|
|
|
|
| 1 |
{
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
"crop_size": {
|
| 3 |
"height": 224,
|
| 4 |
"width": 224
|
learned_embeds.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c782f738753f9cce63c6b3215f2e0a19ee7a9099adb3f2903056e17fd7f17db2
|
| 3 |
+
size 3192
|
logs/textual_inversion/1719884653.746298/events.out.tfevents.1719884653.6a7922459928.1154.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3c3268ccc418025b1dea2affca6d53cea7c43eef52e45ec97642a6c946b32386
|
| 3 |
+
size 2321
|
logs/textual_inversion/1719884653.7487648/hparams.yml
ADDED
|
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
adam_beta1: 0.9
|
| 2 |
+
adam_beta2: 0.999
|
| 3 |
+
adam_epsilon: 1.0e-08
|
| 4 |
+
adam_weight_decay: 0.01
|
| 5 |
+
allow_tf32: false
|
| 6 |
+
center_crop: false
|
| 7 |
+
checkpointing_steps: 100000000
|
| 8 |
+
checkpoints_total_limit: null
|
| 9 |
+
dataloader_num_workers: 0
|
| 10 |
+
enable_xformers_memory_efficient_attention: false
|
| 11 |
+
gradient_accumulation_steps: 4
|
| 12 |
+
gradient_checkpointing: false
|
| 13 |
+
hub_model_id: null
|
| 14 |
+
hub_token: null
|
| 15 |
+
initializer_token: person
|
| 16 |
+
learnable_property: object
|
| 17 |
+
learning_rate: 0.002
|
| 18 |
+
local_rank: -1
|
| 19 |
+
logging_dir: logs
|
| 20 |
+
lr_num_cycles: 1
|
| 21 |
+
lr_scheduler: constant
|
| 22 |
+
lr_warmup_steps: 0
|
| 23 |
+
max_train_steps: 2
|
| 24 |
+
mixed_precision: 'no'
|
| 25 |
+
no_safe_serialization: false
|
| 26 |
+
num_train_epochs: 1
|
| 27 |
+
num_validation_images: 4
|
| 28 |
+
num_vectors: 1
|
| 29 |
+
output_dir: output1
|
| 30 |
+
placeholder_token: <person-glasses>
|
| 31 |
+
pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
|
| 32 |
+
push_to_hub: true
|
| 33 |
+
repeats: 100
|
| 34 |
+
report_to: tensorboard
|
| 35 |
+
resolution: 512
|
| 36 |
+
resume_from_checkpoint: null
|
| 37 |
+
revision: null
|
| 38 |
+
save_as_full_pipeline: false
|
| 39 |
+
save_steps: 500
|
| 40 |
+
scale_lr: true
|
| 41 |
+
seed: null
|
| 42 |
+
tokenizer_name: null
|
| 43 |
+
train_batch_size: 1
|
| 44 |
+
train_data_dir: glasses
|
| 45 |
+
validation_epochs: null
|
| 46 |
+
validation_prompt: null
|
| 47 |
+
validation_steps: 100
|
| 48 |
+
variant: null
|
logs/textual_inversion/events.out.tfevents.1719884653.6a7922459928.1154.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:358c18b433bbeb52953acac212707205e0b883e9d443139a67e7072549d243a1
|
| 3 |
+
size 732
|
model_index.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
{
|
| 2 |
"_class_name": "StableDiffusionPipeline",
|
| 3 |
"_diffusers_version": "0.30.0.dev0",
|
| 4 |
-
"_name_or_path": "
|
| 5 |
"feature_extractor": [
|
| 6 |
"transformers",
|
| 7 |
"CLIPImageProcessor"
|
|
|
|
| 1 |
{
|
| 2 |
"_class_name": "StableDiffusionPipeline",
|
| 3 |
"_diffusers_version": "0.30.0.dev0",
|
| 4 |
+
"_name_or_path": "runwayml/stable-diffusion-v1-5",
|
| 5 |
"feature_extractor": [
|
| 6 |
"transformers",
|
| 7 |
"CLIPImageProcessor"
|
safety_checker/config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "
|
| 3 |
"architectures": [
|
| 4 |
"StableDiffusionSafetyChecker"
|
| 5 |
],
|
|
@@ -15,7 +15,7 @@
|
|
| 15 |
"num_attention_heads": 12
|
| 16 |
},
|
| 17 |
"torch_dtype": "float32",
|
| 18 |
-
"transformers_version": "4.
|
| 19 |
"vision_config": {
|
| 20 |
"dropout": 0.0,
|
| 21 |
"hidden_size": 1024,
|
|
|
|
| 1 |
{
|
| 2 |
+
"_name_or_path": "/root/.cache/huggingface/hub/models--runwayml--stable-diffusion-v1-5/snapshots/1d0c4ebf6ff58a5caecab40fa1406526bca4b5b9/safety_checker",
|
| 3 |
"architectures": [
|
| 4 |
"StableDiffusionSafetyChecker"
|
| 5 |
],
|
|
|
|
| 15 |
"num_attention_heads": 12
|
| 16 |
},
|
| 17 |
"torch_dtype": "float32",
|
| 18 |
+
"transformers_version": "4.42.3",
|
| 19 |
"vision_config": {
|
| 20 |
"dropout": 0.0,
|
| 21 |
"hidden_size": 1024,
|
text_encoder/config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "
|
| 3 |
"architectures": [
|
| 4 |
"CLIPTextModel"
|
| 5 |
],
|
|
@@ -20,6 +20,6 @@
|
|
| 20 |
"pad_token_id": 1,
|
| 21 |
"projection_dim": 768,
|
| 22 |
"torch_dtype": "float32",
|
| 23 |
-
"transformers_version": "4.
|
| 24 |
-
"vocab_size":
|
| 25 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"_name_or_path": "runwayml/stable-diffusion-v1-5",
|
| 3 |
"architectures": [
|
| 4 |
"CLIPTextModel"
|
| 5 |
],
|
|
|
|
| 20 |
"pad_token_id": 1,
|
| 21 |
"projection_dim": 768,
|
| 22 |
"torch_dtype": "float32",
|
| 23 |
+
"transformers_version": "4.42.3",
|
| 24 |
+
"vocab_size": 49409
|
| 25 |
}
|
text_encoder/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8eb8244df7edb6f0cc8db2ebbbbec64444cbbb213b1c3b0fb856b54995340999
|
| 3 |
+
size 492268240
|
tokenizer/added_tokens.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"<person-glasses>": 49408
|
| 3 |
+
}
|
tokenizer/special_tokens_map.json
CHANGED
|
@@ -13,13 +13,7 @@
|
|
| 13 |
"rstrip": false,
|
| 14 |
"single_word": false
|
| 15 |
},
|
| 16 |
-
"pad_token":
|
| 17 |
-
"content": "<|endoftext|>",
|
| 18 |
-
"lstrip": false,
|
| 19 |
-
"normalized": true,
|
| 20 |
-
"rstrip": false,
|
| 21 |
-
"single_word": false
|
| 22 |
-
},
|
| 23 |
"unk_token": {
|
| 24 |
"content": "<|endoftext|>",
|
| 25 |
"lstrip": false,
|
|
|
|
| 13 |
"rstrip": false,
|
| 14 |
"single_word": false
|
| 15 |
},
|
| 16 |
+
"pad_token": "<|endoftext|>",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
"unk_token": {
|
| 18 |
"content": "<|endoftext|>",
|
| 19 |
"lstrip": false,
|
tokenizer/tokenizer_config.json
CHANGED
|
@@ -16,6 +16,14 @@
|
|
| 16 |
"rstrip": false,
|
| 17 |
"single_word": false,
|
| 18 |
"special": true
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 19 |
}
|
| 20 |
},
|
| 21 |
"bos_token": "<|startoftext|>",
|
|
|
|
| 16 |
"rstrip": false,
|
| 17 |
"single_word": false,
|
| 18 |
"special": true
|
| 19 |
+
},
|
| 20 |
+
"49408": {
|
| 21 |
+
"content": "<person-glasses>",
|
| 22 |
+
"lstrip": false,
|
| 23 |
+
"normalized": true,
|
| 24 |
+
"rstrip": false,
|
| 25 |
+
"single_word": false,
|
| 26 |
+
"special": false
|
| 27 |
}
|
| 28 |
},
|
| 29 |
"bos_token": "<|startoftext|>",
|
unet/config.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
{
|
| 2 |
"_class_name": "UNet2DConditionModel",
|
| 3 |
"_diffusers_version": "0.30.0.dev0",
|
| 4 |
-
"_name_or_path": "
|
| 5 |
"act_fn": "silu",
|
| 6 |
"addition_embed_type": null,
|
| 7 |
"addition_embed_type_num_heads": 64,
|
|
|
|
| 1 |
{
|
| 2 |
"_class_name": "UNet2DConditionModel",
|
| 3 |
"_diffusers_version": "0.30.0.dev0",
|
| 4 |
+
"_name_or_path": "runwayml/stable-diffusion-v1-5",
|
| 5 |
"act_fn": "silu",
|
| 6 |
"addition_embed_type": null,
|
| 7 |
"addition_embed_type_num_heads": 64,
|
unet/diffusion_pytorch_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 3438167536
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d27cd69d4a0aa32105087a619f32a51bc087e133be93fe23da92f3c0bcc07d79
|
| 3 |
size 3438167536
|
vae/config.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
{
|
| 2 |
"_class_name": "AutoencoderKL",
|
| 3 |
"_diffusers_version": "0.30.0.dev0",
|
| 4 |
-
"_name_or_path": "
|
| 5 |
"act_fn": "silu",
|
| 6 |
"block_out_channels": [
|
| 7 |
128,
|
|
|
|
| 1 |
{
|
| 2 |
"_class_name": "AutoencoderKL",
|
| 3 |
"_diffusers_version": "0.30.0.dev0",
|
| 4 |
+
"_name_or_path": "runwayml/stable-diffusion-v1-5",
|
| 5 |
"act_fn": "silu",
|
| 6 |
"block_out_channels": [
|
| 7 |
128,
|