akash-guptag commited on
Commit
9325ccc
·
verified ·
1 Parent(s): 3a6b09f

End of training

Browse files
Qlsv7io4jwrmEPHVUnON9JxqK.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6c25831c049843c2f48472b199a8e0547916414263ff316d29fd9ba76a59ac8
3
+ size 12792952
README.md CHANGED
@@ -1,20 +1,25 @@
1
  ---
2
  tags:
3
  - autotrain
4
- - stable-diffusion
5
- - stable-diffusion-diffusers
6
  - text-to-image
7
  - diffusers
8
  - lora
9
  - template:sd-lora
10
 
11
- base_model: Lykon/DreamShaper
12
- instance_prompt: a crayon-style drawing of any object
13
  license: openrail++
14
  ---
15
  # ModelsLab LoRA DreamBooth Training - stablediffusionapi/my-stablediffusion-lora-4583
16
- These are LoRA adaption weights for Lykon/DreamShaper. The weights were trained on a crayon-style drawing of any object using [ModelsLab](https://modelslab.com).
 
 
 
 
17
  LoRA for the text encoder was enabled: False.
 
18
 
19
  ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers)
20
  ```py
@@ -22,7 +27,7 @@ LoRA for the text encoder was enabled: False.
22
  from diffusers import DiffusionPipeline
23
  import torch
24
 
25
- pipe_id = "Lykon/DreamShaper"
26
  pipe = DiffusionPipeline.from_pretrained(pipe_id, torch_dtype=torch.float16).to("cuda")
27
  pipe.load_lora_weights("stablediffusionapi/my-stablediffusion-lora-4583", weight_name="pytorch_lora_weights.safetensors", adapter_name="abc")
28
  prompt = "abc of a hacker with a hoodie"
@@ -30,8 +35,15 @@ lora_scale = 0.9
30
  image = pipe(
31
  prompt,
32
  num_inference_steps=30,
33
- cross_attention_kwargs={"scale": 0.9},
34
  generator=torch.manual_seed(0)
35
  ).images[0]
36
  image
37
- ```
 
 
 
 
 
 
 
 
1
  ---
2
  tags:
3
  - autotrain
4
+ - stable-diffusion-xl
5
+ - stable-diffusion-xl-diffusers
6
  - text-to-image
7
  - diffusers
8
  - lora
9
  - template:sd-lora
10
 
11
+ base_model: stabilityai/stable-diffusion-xl-base-1.0
12
+ instance_prompt: photo of test men
13
  license: openrail++
14
  ---
15
  # ModelsLab LoRA DreamBooth Training - stablediffusionapi/my-stablediffusion-lora-4583
16
+ <Gallery />
17
+
18
+ ## Model description
19
+ These are stablediffusionapi/my-stablediffusion-lora-4583 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
20
+ The weights were trained using [Modelslab](https://modelslab.com).
21
  LoRA for the text encoder was enabled: False.
22
+ Special VAE used for training: None.
23
 
24
  ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers)
25
  ```py
 
27
  from diffusers import DiffusionPipeline
28
  import torch
29
 
30
+ pipe_id = "stabilityai/stable-diffusion-xl-base-1.0"
31
  pipe = DiffusionPipeline.from_pretrained(pipe_id, torch_dtype=torch.float16).to("cuda")
32
  pipe.load_lora_weights("stablediffusionapi/my-stablediffusion-lora-4583", weight_name="pytorch_lora_weights.safetensors", adapter_name="abc")
33
  prompt = "abc of a hacker with a hoodie"
 
35
  image = pipe(
36
  prompt,
37
  num_inference_steps=30,
38
+ cross_attention_kwargs={"scale": lora_scale},
39
  generator=torch.manual_seed(0)
40
  ).images[0]
41
  image
42
+ ```
43
+
44
+ ## Trigger words
45
+ You should use photo of test men to trigger the image generation.
46
+
47
+ ## Download model
48
+ Weights for this model are available in Safetensors format.
49
+ [Download](stablediffusionapi/my-stablediffusion-lora-4583/tree/main) them in the Files & versions tab.
prompt.txt CHANGED
@@ -1 +1 @@
1
- a crayon-style drawing of any object
 
1
+ photo of test men
pytorch_lora_weights.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6c25831c049843c2f48472b199a8e0547916414263ff316d29fd9ba76a59ac8
3
- size 12792952
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fb204882da40e18ba3a5ea3dbe2a6057252caacddc055b58a71aaa6c9eb3cf8
3
+ size 93065304
pytorch_lora_weights_kohya.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:051eb0114fd6bf9ac401cd1dc6e4c1c7ddfe061dc1dc484f73d887c98846a0a1
3
- size 12811664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f1aae561c7095aeffb4d461ed2067bf8f6c63d2740a9409b456b8ba7b861904
3
+ size 93148104
training_params.json CHANGED
@@ -1 +1 @@
1
- {"model": "Lykon/DreamShaper", "vae_model": null, "revision": null, "tokenizer": null, "image_path": "my-stablediffusion-lora-4583/autotrain-data", "class_image_path": null, "prompt": "a crayon-style drawing of any object", "class_prompt": null, "num_class_images": 100, "class_labels_conditioning": null, "prior_preservation": false, "prior_loss_weight": 1.0, "project_name": "my-stablediffusion-lora-4583", "seed": 42, "resolution": 1024, "center_crop": false, "train_text_encoder": false, "batch_size": 4, "sample_batch_size": 4, "epochs": 1, "num_steps": 16, "checkpointing_steps": 500, "resume_from_checkpoint": null, "gradient_accumulation": 4, "disable_gradient_checkpointing": false, "lr": 0.0005, "scale_lr": false, "scheduler": "constant", "warmup_steps": 0, "num_cycles": 1, "lr_power": 1.0, "dataloader_num_workers": 0, "use_8bit_adam": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_weight_decay": 0.01, "adam_epsilon": 1e-08, "max_grad_norm": 1.0, "allow_tf32": false, "prior_generation_precision": null, "local_rank": -1, "xformers": false, "pre_compute_text_embeddings": false, "tokenizer_max_length": null, "text_encoder_use_attention_mask": false, "rank": 16, "xl": false, "mixed_precision": "fp16", "push_to_hub": true, "username": "stablediffusionapi", "hf_private": false, "validation_prompt": null, "num_validation_images": 4, "validation_epochs": 50, "checkpoints_total_limit": null, "validation_images": null, "logging": false}
 
1
+ {"model": "stabilityai/stable-diffusion-xl-base-1.0", "vae_model": null, "revision": null, "tokenizer": null, "image_path": "my-stablediffusion-lora-4583/autotrain-data", "class_image_path": null, "prompt": "photo of test men", "class_prompt": null, "num_class_images": 100, "class_labels_conditioning": null, "prior_preservation": false, "prior_loss_weight": 1.0, "project_name": "my-stablediffusion-lora-4583", "seed": 42, "resolution": 1024, "center_crop": false, "train_text_encoder": false, "batch_size": 4, "sample_batch_size": 4, "epochs": 1, "num_steps": 10, "checkpointing_steps": 500, "resume_from_checkpoint": null, "gradient_accumulation": 4, "disable_gradient_checkpointing": false, "lr": 0.0005, "scale_lr": false, "scheduler": "constant", "warmup_steps": 0, "num_cycles": 1, "lr_power": 1.0, "dataloader_num_workers": 0, "use_8bit_adam": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_weight_decay": 0.01, "adam_epsilon": 1e-08, "max_grad_norm": 1.0, "allow_tf32": false, "prior_generation_precision": null, "local_rank": -1, "xformers": false, "pre_compute_text_embeddings": false, "tokenizer_max_length": null, "text_encoder_use_attention_mask": false, "rank": 16, "xl": true, "mixed_precision": "fp16", "push_to_hub": true, "username": "stablediffusionapi", "hf_private": false, "validation_prompt": null, "num_validation_images": 4, "validation_epochs": 50, "checkpoints_total_limit": null, "validation_images": null, "logging": false}