| license: other | |
| It is a repository for storing as many LECOs as I can think of, emphasizing quantity over quality. | |
| Files will continue to be added as needed. | |
| Because the guidance_scale parameter is somewhat excessive, these LECOs tend to be very sensitive and too effective; using a weight of -0.1 to -1 is appropriate in most cases. | |
| All LECOs are trained with target eq positive, erase settings. | |
| The target is a one of among danbooru's GENERAL tags what most frequently used in order from the top to the bottom, and sometimes I also add phrases that I have personally come up with. | |
| ``` prompts.yaml | |
| - target: "$query" | |
| positive: "$query" | |
| unconditional: "" | |
| neutral: "" | |
| action: "erase" | |
| guidance_scale: 1.0 | |
| resolution: 512 | |
| batch_size: 4 | |
| ``` | |
| ```config.yaml | |
| prompts_file: prompts.yaml | |
| pretrained_model: | |
| name_or_path: "/storage/model-1892-0000-0000.safetensors" | |
| v2: false | |
| v_pred: false | |
| network: | |
| type: "lierla" | |
| rank: 4 | |
| alpha: 1.0 | |
| training_method: "full" | |
| train: | |
| precision: "bfloat16" | |
| noise_scheduler: "ddim" | |
| iterations: 50 | |
| lr: 1 | |
| optimizer: "Prodigy" | |
| lr_scheduler: "cosine" | |
| max_denoising_steps: 50 | |
| save: | |
| name: "$query" | |
| path: "/stable-diffusion-webui/models/Lora/LECO/" | |
| per_steps: 50 | |
| precision: "float16" | |
| logging: | |
| use_wandb: false | |
| verbose: false | |
| other: | |
| use_xformers: true | |
| ``` | |