veddhanth commited on
Commit
1fe53f4
·
verified ·
1 Parent(s): c8ee2be

End of training

Browse files
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ image_0.png filter=lfs diff=lfs merge=lfs -text
37
+ image_1.png filter=lfs diff=lfs merge=lfs -text
<new1>.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f19948796c29590e3639aa8ff9c2929db4b47b505dbdbc08f0638e4d6b283573
3
+ size 3176
README.md ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: CompVis/stable-diffusion-v1-4
3
+ library_name: diffusers
4
+ license: creativeml-openrail-m
5
+ inference: true
6
+ instance_prompt: photo of a <new1> face
7
+ tags:
8
+ - text-to-image
9
+ - diffusers
10
+ - stable-diffusion
11
+ - stable-diffusion-diffusers
12
+ - custom-diffusion
13
+ - diffusers-training
14
+ ---
15
+
16
+ <!-- This model card has been generated automatically according to the information the training script had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+
20
+ # Custom Diffusion - veddhanth/customdiff1989
21
+
22
+ These are Custom Diffusion adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on photo of a <new1> face using [Custom Diffusion](https://www.cs.cmu.edu/~custom-diffusion). You can find some example images in the following.
23
+
24
+ ![img_0](./image_0.png)
25
+ ![img_1](./image_1.png)
26
+
27
+
28
+
29
+ For more details on the training, please follow [this link](https://github.com/huggingface/diffusers/blob/main/examples/custom_diffusion).
30
+
31
+
32
+ ## Intended uses & limitations
33
+
34
+ #### How to use
35
+
36
+ ```python
37
+ # TODO: add an example code snippet for running this diffusion pipeline
38
+ ```
39
+
40
+ #### Limitations and bias
41
+
42
+ [TODO: provide examples of latent issues and potential remediations]
43
+
44
+ ## Training details
45
+
46
+ [TODO: describe the data used to train the model]
image_0.png ADDED

Git LFS Details

  • SHA256: e9b0f598b9f04a2c9cd08018609bde6391c8ebf50deddd1e8cdeefcd941d987d
  • Pointer size: 131 Bytes
  • Size of remote file: 416 kB
image_1.png ADDED

Git LFS Details

  • SHA256: d7161aa8a5f2b3b0eaa94d9908a4d6169a7a6e7581cc1c383cef027b9e0feb6a
  • Pointer size: 131 Bytes
  • Size of remote file: 332 kB
logs/custom-diffusion/1752917826.7060962/events.out.tfevents.1752917826.snuc-HP-Z4-G5-Workstation-Desktop-PC.30001.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:307f52be4fc213161a8f6fbdecd2b34e0a6fb8f0206677fdbb4748976aeda5e6
3
+ size 2936
logs/custom-diffusion/1752917826.7073956/hparams.yml ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ center_crop: false
7
+ checkpointing_steps: 2500
8
+ checkpoints_total_limit: null
9
+ class_data_dir: /home/snuc/Documents/personalizedrep/personalized-rep/synthetic_data/celeba/celeba_class_dir_rv/1989
10
+ class_prompt: face
11
+ concepts_list: null
12
+ dataloader_num_workers: 2
13
+ enable_xformers_memory_efficient_attention: false
14
+ freeze_model: crossattn_kv
15
+ gradient_accumulation_steps: 1
16
+ gradient_checkpointing: false
17
+ hflip: true
18
+ hub_model_id: null
19
+ hub_token: null
20
+ initializer_token: ktn+pll+ucd
21
+ instance_data_dir: /home/snuc/Documents/personalizedrep/dataset/train/1989/
22
+ instance_prompt: photo of a <new1> face
23
+ learning_rate: 1.0e-05
24
+ local_rank: -1
25
+ logging_dir: logs
26
+ lr_scheduler: constant
27
+ lr_warmup_steps: 0
28
+ max_grad_norm: 1.0
29
+ max_train_steps: 1000
30
+ mixed_precision: null
31
+ modifier_token: <new1>
32
+ no_safe_serialization: false
33
+ noaug: false
34
+ num_class_images: 1000
35
+ num_train_epochs: 1
36
+ num_validation_images: 2
37
+ output_dir: customdiff1989
38
+ pretrained_model_name_or_path: CompVis/stable-diffusion-v1-4
39
+ prior_generation_precision: null
40
+ prior_loss_weight: 1.0
41
+ push_to_hub: true
42
+ real_prior: false
43
+ report_to: tensorboard
44
+ resolution: 512
45
+ resume_from_checkpoint: null
46
+ revision: null
47
+ sample_batch_size: 4
48
+ scale_lr: true
49
+ seed: 42
50
+ set_grads_to_none: false
51
+ tokenizer_name: null
52
+ train_batch_size: 1
53
+ use_8bit_adam: false
54
+ validation_prompt: <new1> face with a neutral expression
55
+ validation_steps: 2000
56
+ variant: null
57
+ with_prior_preservation: true
logs/custom-diffusion/1752918358.4833724/events.out.tfevents.1752918358.snuc-HP-Z4-G5-Workstation-Desktop-PC.31561.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:264a674c023ed23382d4305f0c847f6edbe8704cfa2aa70b15c4467238d793ef
3
+ size 2936
logs/custom-diffusion/1752918358.484349/hparams.yml ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ center_crop: false
7
+ checkpointing_steps: 2500
8
+ checkpoints_total_limit: null
9
+ class_data_dir: /home/snuc/Documents/personalizedrep/personalized-rep/synthetic_data/celeba/celeba_class_dir_rv/1989
10
+ class_prompt: face
11
+ concepts_list: null
12
+ dataloader_num_workers: 2
13
+ enable_xformers_memory_efficient_attention: false
14
+ freeze_model: crossattn_kv
15
+ gradient_accumulation_steps: 1
16
+ gradient_checkpointing: false
17
+ hflip: true
18
+ hub_model_id: null
19
+ hub_token: null
20
+ initializer_token: ktn+pll+ucd
21
+ instance_data_dir: /home/snuc/Documents/personalizedrep/dataset/train/1989/
22
+ instance_prompt: photo of a <new1> face
23
+ learning_rate: 1.0e-05
24
+ local_rank: -1
25
+ logging_dir: logs
26
+ lr_scheduler: constant
27
+ lr_warmup_steps: 0
28
+ max_grad_norm: 1.0
29
+ max_train_steps: 1000
30
+ mixed_precision: null
31
+ modifier_token: <new1>
32
+ no_safe_serialization: false
33
+ noaug: false
34
+ num_class_images: 1000
35
+ num_train_epochs: 1
36
+ num_validation_images: 2
37
+ output_dir: customdiff1989
38
+ pretrained_model_name_or_path: CompVis/stable-diffusion-v1-4
39
+ prior_generation_precision: null
40
+ prior_loss_weight: 1.0
41
+ push_to_hub: true
42
+ real_prior: false
43
+ report_to: tensorboard
44
+ resolution: 512
45
+ resume_from_checkpoint: null
46
+ revision: null
47
+ sample_batch_size: 4
48
+ scale_lr: true
49
+ seed: 42
50
+ set_grads_to_none: false
51
+ tokenizer_name: null
52
+ train_batch_size: 1
53
+ use_8bit_adam: false
54
+ validation_prompt: <new1> face with a neutral expression
55
+ validation_steps: 2000
56
+ variant: null
57
+ with_prior_preservation: false
logs/custom-diffusion/events.out.tfevents.1752917826.snuc-HP-Z4-G5-Workstation-Desktop-PC.30001.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fc8227774d9809e26b6c858079d6978c244987e765b6f4b73f9212e69564fd0
3
+ size 35954
logs/custom-diffusion/events.out.tfevents.1752918358.snuc-HP-Z4-G5-Workstation-Desktop-PC.31561.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebb13b3e55ff2812f7bb396a77cdbc20dddfac71fc0feb43e6396c57f4c63796
3
+ size 842279
pytorch_custom_diffusion_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a0aff27aec9a1119df640191ccb9071c18b73b7888cb29b1a10df89613da2b0
3
+ size 76682360