ainjarts commited on
Commit
e2e6044
·
1 Parent(s): 6ea2a17

End of training

Browse files
README.md CHANGED
@@ -1,7 +1,7 @@
1
 
2
  ---
3
  license: creativeml-openrail-m
4
- base_model: CompVis/stable-diffusion-v1-4
5
  instance_prompt: a photo of sks cat
6
  tags:
7
  - stable-diffusion
@@ -14,7 +14,7 @@ inference: true
14
 
15
  # DreamBooth - ainjarts/model
16
 
17
- This is a dreambooth model derived from CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks cat using [DreamBooth](https://dreambooth.github.io/).
18
  You can find some example images in the following.
19
 
20
 
 
1
 
2
  ---
3
  license: creativeml-openrail-m
4
+ base_model: SG161222/Realistic_Vision_V2.0
5
  instance_prompt: a photo of sks cat
6
  tags:
7
  - stable-diffusion
 
14
 
15
  # DreamBooth - ainjarts/model
16
 
17
+ This is a dreambooth model derived from SG161222/Realistic_Vision_V2.0. The weights were trained on a photo of sks cat using [DreamBooth](https://dreambooth.github.io/).
18
  You can find some example images in the following.
19
 
20
 
checkpoint-500/optimizer.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc3fe2074421de6a08e1bec7b5ac0d661ba5faf27c01e7a08bcdad42089c7ed4
3
  size 1972598422
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f68592290c9811bc3ed94edc6a27b51eab3cb72bc3caaae3fc6f3af40236440d
3
  size 1972598422
checkpoint-500/random_states_0.pkl CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f5479aa873838b3d3d05ffb30f0868457615ec177d53ef3fff8971bce81b0fa0
3
- size 14344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60192ff37cadf3bf9f9e0e5e67bb1e3c725a0a49ecdc06115ab1e7b3f1b75543
3
+ size 14408
checkpoint-500/text_encoder/config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "CompVis/stable-diffusion-v1-4",
3
  "architectures": [
4
  "CLIPTextModel"
5
  ],
@@ -18,7 +18,7 @@
18
  "num_attention_heads": 12,
19
  "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
- "projection_dim": 512,
22
  "torch_dtype": "float32",
23
  "transformers_version": "4.35.2",
24
  "vocab_size": 49408
 
1
  {
2
+ "_name_or_path": "SG161222/Realistic_Vision_V2.0",
3
  "architectures": [
4
  "CLIPTextModel"
5
  ],
 
18
  "num_attention_heads": 12,
19
  "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
+ "projection_dim": 768,
22
  "torch_dtype": "float32",
23
  "transformers_version": "4.35.2",
24
  "vocab_size": 49408
checkpoint-500/text_encoder/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8d986f55e835d0246193d8715f9f541378b26187bba4c0c829c09ce065cd1a2e
3
  size 492265168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fe3495586c166fb3c9a9832c69a80fc9472987cde2064d2c31b129142fc0bff
3
  size 492265168
checkpoint-500/unet/config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
  "_diffusers_version": "0.24.0.dev0",
4
- "_name_or_path": "CompVis/stable-diffusion-v1-4",
5
  "act_fn": "silu",
6
  "addition_embed_type": null,
7
  "addition_embed_type_num_heads": 64,
 
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
  "_diffusers_version": "0.24.0.dev0",
4
+ "_name_or_path": "SG161222/Realistic_Vision_V2.0",
5
  "act_fn": "silu",
6
  "addition_embed_type": null,
7
  "addition_embed_type_num_heads": 64,
checkpoint-500/unet/diffusion_pytorch_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:408b0f52300b5167b0cce1950277de0710742a32a8e1b48d12f8d995b0cfd092
3
  size 3438167536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b8d35113e58cabb18a74d3ec4bde5bc558fc157823b40f70c220130eb982ac5
3
  size 3438167536
logs/dreambooth/1701169490.5830095/events.out.tfevents.1701169490.53ad7ae7f465.1631.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07f252a117ae8e084c674b5c658e3c5d6851ec00885fae3e90697ef9c1f8e4f7
3
+ size 2979
logs/dreambooth/1701169490.585203/hparams.yml ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ center_crop: false
7
+ checkpointing_steps: 500
8
+ checkpoints_total_limit: null
9
+ class_data_dir: /home/diffusers/examples/dreambooth/dataset/img/class
10
+ class_labels_conditioning: null
11
+ class_prompt: a photo of cat
12
+ dataloader_num_workers: 0
13
+ enable_xformers_memory_efficient_attention: false
14
+ gradient_accumulation_steps: 1
15
+ gradient_checkpointing: true
16
+ hub_model_id: null
17
+ hub_token: null
18
+ instance_data_dir: /home/diffusers/examples/dreambooth/dataset/img/cat
19
+ instance_prompt: a photo of sks cat
20
+ learning_rate: 2.0e-06
21
+ local_rank: -1
22
+ logging_dir: logs
23
+ lr_num_cycles: 1
24
+ lr_power: 1.0
25
+ lr_scheduler: constant
26
+ lr_warmup_steps: 0
27
+ max_grad_norm: 1.0
28
+ max_train_steps: 800
29
+ mixed_precision: null
30
+ num_class_images: 50
31
+ num_train_epochs: 16
32
+ num_validation_images: 4
33
+ offset_noise: false
34
+ output_dir: /home/diffusers/examples/dreambooth/dataset/model
35
+ pre_compute_text_embeddings: false
36
+ pretrained_model_name_or_path: SG161222/Realistic_Vision_V2.0
37
+ prior_generation_precision: null
38
+ prior_loss_weight: 1.0
39
+ push_to_hub: true
40
+ report_to: tensorboard
41
+ resolution: 1024
42
+ resume_from_checkpoint: null
43
+ revision: null
44
+ sample_batch_size: 4
45
+ scale_lr: false
46
+ seed: null
47
+ set_grads_to_none: false
48
+ skip_save_text_encoder: false
49
+ snr_gamma: null
50
+ text_encoder_use_attention_mask: false
51
+ tokenizer_max_length: null
52
+ tokenizer_name: null
53
+ train_batch_size: 1
54
+ train_text_encoder: true
55
+ use_8bit_adam: true
56
+ validation_prompt: null
57
+ validation_scheduler: DPMSolverMultistepScheduler
58
+ validation_steps: 100
59
+ variant: null
60
+ with_prior_preservation: true
logs/dreambooth/events.out.tfevents.1701169490.53ad7ae7f465.1631.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e58d524b916a7619cedc2ff7a570ccf7f3a0a5d70d82ec149ff52000397b4ac4
3
+ size 67034
model_index.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
  "_diffusers_version": "0.24.0.dev0",
4
- "_name_or_path": "CompVis/stable-diffusion-v1-4",
5
  "feature_extractor": [
6
  "transformers",
7
  "CLIPImageProcessor"
 
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
  "_diffusers_version": "0.24.0.dev0",
4
+ "_name_or_path": "SG161222/Realistic_Vision_V2.0",
5
  "feature_extractor": [
6
  "transformers",
7
  "CLIPImageProcessor"
safety_checker/config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/home/.cache/huggingface/hub/models--CompVis--stable-diffusion-v1-4/snapshots/133a221b8aa7292a167afc5127cb63fb5005638b/safety_checker",
3
  "architectures": [
4
  "StableDiffusionSafetyChecker"
5
  ],
 
1
  {
2
+ "_name_or_path": "/home/.cache/huggingface/hub/models--SG161222--Realistic_Vision_V2.0/snapshots/60c57a68e17cc9261e6031cb3babd7ffd6d6dde1/safety_checker",
3
  "architectures": [
4
  "StableDiffusionSafetyChecker"
5
  ],
text_encoder/config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "CompVis/stable-diffusion-v1-4",
3
  "architectures": [
4
  "CLIPTextModel"
5
  ],
@@ -18,7 +18,7 @@
18
  "num_attention_heads": 12,
19
  "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
- "projection_dim": 512,
22
  "torch_dtype": "float32",
23
  "transformers_version": "4.35.2",
24
  "vocab_size": 49408
 
1
  {
2
+ "_name_or_path": "SG161222/Realistic_Vision_V2.0",
3
  "architectures": [
4
  "CLIPTextModel"
5
  ],
 
18
  "num_attention_heads": 12,
19
  "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
+ "projection_dim": 768,
22
  "torch_dtype": "float32",
23
  "transformers_version": "4.35.2",
24
  "vocab_size": 49408
text_encoder/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:074d84b424cfb7c708bba569a40bfde8700fe056f55d5c8a92ad8774c317023a
3
  size 492265168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a169f538e06a896d3e311602e97ddac9c3b144a72bf7bb147bfd1a76bb5cdc3
3
  size 492265168
unet/config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
  "_diffusers_version": "0.24.0.dev0",
4
- "_name_or_path": "CompVis/stable-diffusion-v1-4",
5
  "act_fn": "silu",
6
  "addition_embed_type": null,
7
  "addition_embed_type_num_heads": 64,
 
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
  "_diffusers_version": "0.24.0.dev0",
4
+ "_name_or_path": "SG161222/Realistic_Vision_V2.0",
5
  "act_fn": "silu",
6
  "addition_embed_type": null,
7
  "addition_embed_type_num_heads": 64,
unet/diffusion_pytorch_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:794df22d40c963ee0d3e7315a3a3bcc9bf43263a613359b3dd3809925a4f2ab6
3
  size 3438167536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cabbf3804ebc108c3397ec4e961d1c9365b7391c9f35952bb09924c64fc39b22
3
  size 3438167536
vae/config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "_class_name": "AutoencoderKL",
3
  "_diffusers_version": "0.24.0.dev0",
4
- "_name_or_path": "/home/.cache/huggingface/hub/models--CompVis--stable-diffusion-v1-4/snapshots/133a221b8aa7292a167afc5127cb63fb5005638b/vae",
5
  "act_fn": "silu",
6
  "block_out_channels": [
7
  128,
 
1
  {
2
  "_class_name": "AutoencoderKL",
3
  "_diffusers_version": "0.24.0.dev0",
4
+ "_name_or_path": "/home/.cache/huggingface/hub/models--SG161222--Realistic_Vision_V2.0/snapshots/60c57a68e17cc9261e6031cb3babd7ffd6d6dde1/vae",
5
  "act_fn": "silu",
6
  "block_out_channels": [
7
  128,
vae/diffusion_pytorch_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b4d2b5932bb4151e54e694fd31ccf51fca908223c9485bd56cd0e1d83ad94c49
3
  size 334643268
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86a6170968f284b6c0c111e2611c806a04102e2a48790ebd1d490ca78040d2bb
3
  size 334643268