Spaces:
Build error
Build error
| # Finetuning Resource Guide | |
| This guide is a resource compilation to facilitate the development of robust LoRA models. | |
| -Need to add resources here | |
| ## Guidelines for SDXL Finetuning | |
| - Set the `Max resolution` to at least 1024x1024, as this is the standard resolution for SDXL. | |
| - The fine-tuning can be done with 24GB GPU memory with the batch size of 1. | |
| - Train U-Net only. | |
| - Use gradient checkpointing. | |
| - Use `--cache_text_encoder_outputs` option and caching latents. | |
| - Use Adafactor optimizer. RMSprop 8bit or Adagrad 8bit may work. AdamW 8bit doesn't seem to work. | |
| - PyTorch 2 seems to use slightly less GPU memory than PyTorch 1. | |
| Example of the optimizer settings for Adafactor with the fixed learning rate: | |
| ``` | |
| optimizer_type = "adafactor" | |
| optimizer_args = [ "scale_parameter=False", "relative_step=False", "warmup_init=False" ] | |
| lr_scheduler = "constant_with_warmup" | |
| lr_warmup_steps = 100 | |
| learning_rate = 4e-7 # SDXL original learning rate | |
| ``` | |
| ## Resource Contributions | |
| If you have valuable resources to add, kindly create a PR on Github. |