End of training
Browse files- Qlsv7io4jwrmEPHVUnON9JxqK.safetensors +3 -0
- README.md +20 -8
- prompt.txt +1 -1
- pytorch_lora_weights.safetensors +2 -2
- pytorch_lora_weights_kohya.safetensors +2 -2
- training_params.json +1 -1
Qlsv7io4jwrmEPHVUnON9JxqK.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e6c25831c049843c2f48472b199a8e0547916414263ff316d29fd9ba76a59ac8
|
| 3 |
+
size 12792952
|
README.md
CHANGED
|
@@ -1,20 +1,25 @@
|
|
| 1 |
---
|
| 2 |
tags:
|
| 3 |
- autotrain
|
| 4 |
-
- stable-diffusion
|
| 5 |
-
- stable-diffusion-diffusers
|
| 6 |
- text-to-image
|
| 7 |
- diffusers
|
| 8 |
- lora
|
| 9 |
- template:sd-lora
|
| 10 |
|
| 11 |
-
base_model:
|
| 12 |
-
instance_prompt:
|
| 13 |
license: openrail++
|
| 14 |
---
|
| 15 |
# ModelsLab LoRA DreamBooth Training - stablediffusionapi/my-stablediffusion-lora-4583
|
| 16 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
LoRA for the text encoder was enabled: False.
|
|
|
|
| 18 |
|
| 19 |
## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers)
|
| 20 |
```py
|
|
@@ -22,7 +27,7 @@ LoRA for the text encoder was enabled: False.
|
|
| 22 |
from diffusers import DiffusionPipeline
|
| 23 |
import torch
|
| 24 |
|
| 25 |
-
pipe_id = "
|
| 26 |
pipe = DiffusionPipeline.from_pretrained(pipe_id, torch_dtype=torch.float16).to("cuda")
|
| 27 |
pipe.load_lora_weights("stablediffusionapi/my-stablediffusion-lora-4583", weight_name="pytorch_lora_weights.safetensors", adapter_name="abc")
|
| 28 |
prompt = "abc of a hacker with a hoodie"
|
|
@@ -30,8 +35,15 @@ lora_scale = 0.9
|
|
| 30 |
image = pipe(
|
| 31 |
prompt,
|
| 32 |
num_inference_steps=30,
|
| 33 |
-
cross_attention_kwargs={"scale":
|
| 34 |
generator=torch.manual_seed(0)
|
| 35 |
).images[0]
|
| 36 |
image
|
| 37 |
-
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
tags:
|
| 3 |
- autotrain
|
| 4 |
+
- stable-diffusion-xl
|
| 5 |
+
- stable-diffusion-xl-diffusers
|
| 6 |
- text-to-image
|
| 7 |
- diffusers
|
| 8 |
- lora
|
| 9 |
- template:sd-lora
|
| 10 |
|
| 11 |
+
base_model: stabilityai/stable-diffusion-xl-base-1.0
|
| 12 |
+
instance_prompt: photo of test men
|
| 13 |
license: openrail++
|
| 14 |
---
|
| 15 |
# ModelsLab LoRA DreamBooth Training - stablediffusionapi/my-stablediffusion-lora-4583
|
| 16 |
+
<Gallery />
|
| 17 |
+
|
| 18 |
+
## Model description
|
| 19 |
+
These are stablediffusionapi/my-stablediffusion-lora-4583 LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
|
| 20 |
+
The weights were trained using [Modelslab](https://modelslab.com).
|
| 21 |
LoRA for the text encoder was enabled: False.
|
| 22 |
+
Special VAE used for training: None.
|
| 23 |
|
| 24 |
## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers)
|
| 25 |
```py
|
|
|
|
| 27 |
from diffusers import DiffusionPipeline
|
| 28 |
import torch
|
| 29 |
|
| 30 |
+
pipe_id = "stabilityai/stable-diffusion-xl-base-1.0"
|
| 31 |
pipe = DiffusionPipeline.from_pretrained(pipe_id, torch_dtype=torch.float16).to("cuda")
|
| 32 |
pipe.load_lora_weights("stablediffusionapi/my-stablediffusion-lora-4583", weight_name="pytorch_lora_weights.safetensors", adapter_name="abc")
|
| 33 |
prompt = "abc of a hacker with a hoodie"
|
|
|
|
| 35 |
image = pipe(
|
| 36 |
prompt,
|
| 37 |
num_inference_steps=30,
|
| 38 |
+
cross_attention_kwargs={"scale": lora_scale},
|
| 39 |
generator=torch.manual_seed(0)
|
| 40 |
).images[0]
|
| 41 |
image
|
| 42 |
+
```
|
| 43 |
+
|
| 44 |
+
## Trigger words
|
| 45 |
+
You should use photo of test men to trigger the image generation.
|
| 46 |
+
|
| 47 |
+
## Download model
|
| 48 |
+
Weights for this model are available in Safetensors format.
|
| 49 |
+
[Download](stablediffusionapi/my-stablediffusion-lora-4583/tree/main) them in the Files & versions tab.
|
prompt.txt
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
|
|
|
|
| 1 |
+
photo of test men
|
pytorch_lora_weights.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4fb204882da40e18ba3a5ea3dbe2a6057252caacddc055b58a71aaa6c9eb3cf8
|
| 3 |
+
size 93065304
|
pytorch_lora_weights_kohya.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8f1aae561c7095aeffb4d461ed2067bf8f6c63d2740a9409b456b8ba7b861904
|
| 3 |
+
size 93148104
|
training_params.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"model": "
|
|
|
|
| 1 |
+
{"model": "stabilityai/stable-diffusion-xl-base-1.0", "vae_model": null, "revision": null, "tokenizer": null, "image_path": "my-stablediffusion-lora-4583/autotrain-data", "class_image_path": null, "prompt": "photo of test men", "class_prompt": null, "num_class_images": 100, "class_labels_conditioning": null, "prior_preservation": false, "prior_loss_weight": 1.0, "project_name": "my-stablediffusion-lora-4583", "seed": 42, "resolution": 1024, "center_crop": false, "train_text_encoder": false, "batch_size": 4, "sample_batch_size": 4, "epochs": 1, "num_steps": 10, "checkpointing_steps": 500, "resume_from_checkpoint": null, "gradient_accumulation": 4, "disable_gradient_checkpointing": false, "lr": 0.0005, "scale_lr": false, "scheduler": "constant", "warmup_steps": 0, "num_cycles": 1, "lr_power": 1.0, "dataloader_num_workers": 0, "use_8bit_adam": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_weight_decay": 0.01, "adam_epsilon": 1e-08, "max_grad_norm": 1.0, "allow_tf32": false, "prior_generation_precision": null, "local_rank": -1, "xformers": false, "pre_compute_text_embeddings": false, "tokenizer_max_length": null, "text_encoder_use_attention_mask": false, "rank": 16, "xl": true, "mixed_precision": "fp16", "push_to_hub": true, "username": "stablediffusionapi", "hf_private": false, "validation_prompt": null, "num_validation_images": 4, "validation_epochs": 50, "checkpoints_total_limit": null, "validation_images": null, "logging": false}
|