| <h1 align="center">Stable Diffusion WebUI Forge - Classic</h1> | |
| <p align="center"><sup> | |
| [ Classic | <a href="https://github.com/Haoming02/sd-webui-forge-classic/tree/neo#stable-diffusion-webui-forge---neo">Neo</a> ] | |
| </sup></p> | |
| <p align="center"><img src="html\ui.webp" width=512 alt="UI"></p> | |
| <blockquote><i> | |
| <b>Stable Diffusion WebUI Forge</b> is a platform on top of the original <a href="https://github.com/AUTOMATIC1111/stable-diffusion-webui">Stable Diffusion WebUI</a> by <ins>AUTOMATIC1111</ins>, to make development easier, optimize resource management, speed up inference, and study experimental features.<br> | |
| The name "Forge" is inspired by "Minecraft Forge". This project aims to become the Forge of Stable Diffusion WebUI.<br> | |
| <p align="right">- <b>lllyasviel</b><br> | |
| <sup>(paraphrased)</sup></p> | |
| </i></blockquote> | |
| <br> | |
| "**Classic**" mainly serves as an archive for the "`previous`" version of Forge, which was built on [Gradio](https://github.com/gradio-app/gradio) `3.41.2` before the major changes *(see the original [announcement](https://github.com/lllyasviel/stable-diffusion-webui-forge/discussions/801))* were introduced. Additionally, this fork is focused exclusively on **SD1** and **SDXL** checkpoints, having various optimizations implemented, with the main goal of being the lightest WebUI without any bloatwares. | |
| > [!Tip] | |
| > [How to Install](#installation) | |
| <br> | |
| ## Features [Oct. 08] | |
| > Most base features of the original [Automatic1111 Webui](https://github.com/AUTOMATIC1111/stable-diffusion-webui) should still function | |
| #### New Features | |
| - [X] Support [uv](https://github.com/astral-sh/uv) package manager | |
| - requires **manually** installing [uv](https://github.com/astral-sh/uv/releases) | |
| - drastically speed up installation | |
| - see [Commandline](#by-classic) | |
| - [X] Support [SageAttention](https://github.com/thu-ml/SageAttention) | |
| - requires **manually** installing [triton](https://github.com/triton-lang/triton) | |
| - [how to install](#install-triton) | |
| - requires RTX **30** + | |
| - ~10% speed up for SDXL | |
| - see [Commandline](#by-classic) | |
| - [X] Support [FlashAttention](https://arxiv.org/abs/2205.14135) | |
| - requires **manually** installing [flash-attn](https://github.com/Dao-AILab/flash-attention) | |
| - [how to install](#install-flash-attn) | |
| - ~10% speed up | |
| - [X] Support fast `fp16_accumulation` | |
| - requires PyTorch **2.7.0** + | |
| - ~25% speed up | |
| - see [Commandline](#by-classic) | |
| - [X] Support fast `cublas` operation *(`CublasLinear`)* | |
| - requires **manually** installing [cublas_ops](https://github.com/aredden/torch-cublas-hgemm) | |
| - [how to install](#install-cublas) | |
| - ~25% speed up | |
| - enable in **Settings/Optimizations** | |
| > [!Important] | |
| > - Both `fp16_accumulation` and `cublas_ops` achieve the same speed up; if you already install/update to PyTorch **2.7.0** +, there is little reason to go for `cublas_ops` | |
| - [X] Support fast `fp8` operation *(`torch._scaled_mm`)* | |
| - requires RTX **40** + | |
| - requires **UNet Weights in fp8** option | |
| - ~10% speed up; reduce quality | |
| - enable in **Settings/Optimizations** | |
| > [!Note] | |
| > - The `fp16_accumulation` and `cublas_ops` require `fp16` precision, thus is not compatible with the `fp8` operation | |
| <br> | |
| - [X] Persistent LoRA Patching | |
| - speed up LoRA loading in subsequent generations | |
| - see [Commandline](#by-classic) | |
| - [X] Patch LoRA in-place | |
| - reduce VRAM usage when loading LoRA | |
| - enable in **Settings/Extra Networks** | |
| - [X] Implement new Samplers | |
| - *(ported from reForge Webui)* | |
| - [X] Implement Scheduler dropdown | |
| - *(backported from Automatic1111 Webui upstream)* | |
| - enable in **Settings/UI Alternatives** | |
| - [X] Add `CFG` slider to the `Hires. fix` section | |
| - [X] Implement RescaleCFG | |
| - reduce burnt colors; mainly for `v-pred` checkpoints | |
| - enable in **Settings/UI Alternatives** | |
| - [X] Implement MaHiRo | |
| - alternative CFG calculation; improve prompt adherence | |
| - enable in **Settings/UI Alternatives** | |
| - [X] Implement [Epsilon Scaling](https://github.com/comfyanonymous/ComfyUI/pull/10132) | |
| - enable in **Settings/Stable Diffusion** | |
| - [X] Implement full precision calculation for `Mask blur` blending | |
| - enable in **Settings/img2img** | |
| - [X] Support loading upscalers in `half` precision | |
| - speed up; reduce quality | |
| - enable in **Settings/Upscaling** | |
| - [X] Support running tile composition on GPU | |
| - enable in **Settings/Upscaling** | |
| - [X] Allow `newline` in LoRA metadata | |
| - *(backported from Automatic1111 Webui upstream)* | |
| - [X] Implement sending parameters from generation result rather than from UI | |
| - **e.g.** send the prompts instead of `Wildcard` syntax | |
| - enable in **Settings/Infotext** | |
| - [X] Implement tiling optimization for VAE | |
| - reduce memory usage; reduce speed | |
| - enable in **Settings/VAE** | |
| - [X] Implement `diskcache` for hashes | |
| - *(backported from Automatic1111 Webui upstream)* | |
| - [X] Implement `skip_early_cond` | |
| - *(backported from Automatic1111 Webui upstream)* | |
| - enable in **Settings/Optimizations** | |
| - [X] Allow inserting the upscaled image to the Gallery instead of overriding the input image | |
| - *(backported from upstream [PR](https://github.com/AUTOMATIC1111/stable-diffusion-webui/pull/16405))* | |
| - [X] Support `v-pred` **SDXL** checkpoints *(**e.g.** [NoobAI](https://civitai.com/models/833294?modelVersionId=1190596))* | |
| - [X] Support new LoRA architectures | |
| - [X] Update `spandrel` | |
| - support new Upscaler architectures | |
| - [X] Add `pillow-heif` package | |
| - support `.avif` and `.heif` images | |
| - [X] Automatically determine the optimal row count for `X/Y/Z Plot` | |
| - [X] Support new LoRA architectures | |
| - [X] `DepthAnything v2` Preprocessor | |
| - [X] Support [NoobAI Inpaint](https://civitai.com/models/1376234/noobai-inpainting-controlnet) ControlNet | |
| - [X] Support [Union](https://huggingface.co/xinsir/controlnet-union-sdxl-1.0) / [ProMax](https://huggingface.co/brad-twinkl/controlnet-union-sdxl-1.0-promax) ControlNet | |
| - they simply always show up in the dropdown | |
| #### Removed Features | |
| - [X] SD2 | |
| - [X] Alt-Diffusion | |
| - [X] Instruct-Pix2Pix | |
| - [X] Hypernetworks | |
| - [X] SVD | |
| - [X] Z123 | |
| - [X] CLIP Interrogator | |
| - [X] Deepbooru Interrogator | |
| - [X] Textual Inversion Training | |
| - [X] Checkpoint Merging | |
| - [X] LDSR | |
| - [X] Most built-in Extensions | |
| - [X] Some built-in Scripts | |
| - [X] Some Samplers | |
| - [X] Sampler in RadioGroup | |
| - [X] `test` scripts | |
| - [X] Some Preprocessors *(ControlNet)* | |
| - [X] `Photopea` and `openpose_editor` *(ControlNet)* | |
| - [X] Unix `.sh` launch scripts | |
| - You can still use this WebUI by simply copying a launch script from other working WebUI | |
| #### Optimizations | |
| - [X] **[Freedom]** Natively integrate the `SD1` and `SDXL` logics | |
| - no longer `git` `clone` any repository on fresh install | |
| - no more random hacks and monkey patches | |
| - [X] Fix `canvas-zoom-and-pan` built-in extension | |
| - no more infinite-resizing bug when using `Send to` buttons | |
| - [X] Fix RAM and VRAM leak when switching checkpoints | |
| - [X] Clean up the `ldm_patched` *(**i.e.** `comfy`)* folder | |
| - [X] Remove unused `cmd_args` | |
| - [X] Remove unused `args_parser` | |
| - [X] Remove unused `shared_options` | |
| - [X] Remove legacy codes | |
| - [X] Fix some typos | |
| - [X] Remove redundant upscaler codes | |
| - put every upscaler inside the `ESRGAN` folder | |
| - [X] Optimize upscaler logics | |
| - [X] Optimize certain operations in `Spandrel` | |
| - [X] Optimize the creation of Extra Networks pages | |
| - *(backported from Automatic1111 Webui upstream)* | |
| - [X] Improve color correction | |
| - [X] Improve hash caching | |
| - [X] Improve error logs | |
| - no longer print `TypeError: 'NoneType' object is not iterable` | |
| - [X] Update the implementation for `uni_pc` sampler | |
| - [X] Revamp settings | |
| - improve formatting | |
| - update descriptions | |
| - [X] Check for Extension updates in parallel | |
| - [X] Move `embeddings` folder into `models` folder | |
| - [X] ControlNet Rewrite | |
| - change Units to `gr.Tab` | |
| - remove multi-inputs, as they are "[misleading](https://github.com/lllyasviel/stable-diffusion-webui-forge/discussions/932)" | |
| - change `visible` toggle to `interactive` toggle; now the UI will no longer jump around | |
| - improve `Presets` application | |
| - fix `Inpaint not masked` mode | |
| - [X] Disable Refiner by default | |
| - enable again in **Settings/UI Alternatives** | |
| - [X] Disable Tree View by default | |
| - enable again in **Settings/Extra Networks** | |
| - [X] Hide Sampler Parameters by default | |
| - enable again by adding **--adv-samplers** flag | |
| - [X] Hide some X/Y/Z Plot options by default | |
| - enable again by adding **--adv-xyz** flag | |
| - [X] Run `text encoder` on CPU by default | |
| - [X] Fix `pydantic` Errors | |
| - [X] Fix `Soft Inpainting` | |
| - [X] Fix `Controllllite` | |
| - [X] Fix `MultiDiffusion` | |
| - [X] Fix `SD Upscale` | |
| - [X] Lint & Format | |
| - [X] Update `Pillow` | |
| - faster image processing | |
| - [X] Update `protobuf` | |
| - faster `insightface` loading | |
| - [X] Update to latest PyTorch | |
| - `torch==2.8.0+cu128` | |
| - `xformers==0.0.32` | |
| > [!Note] | |
| > If your GPU does not support the latest PyTorch, manually [install](#install-older-pytorch) older version of PyTorch | |
| - [X] No longer install `open-clip` twice | |
| - [X] Update some packages to newer versions | |
| - [X] Update recommended Python to `3.11.9` | |
| - [X] many more... :tm: | |
| <br> | |
| ## Commandline | |
| > These flags can be added after the `set COMMANDLINE_ARGS=` line in the `webui-user.bat` *(separate each flag with space)* | |
| #### A1111 built-in | |
| - `--no-download-sd-model`: Do not download a default checkpoint | |
| - can be removed after you download some checkpoints of your choice | |
| - `--xformers`: Install the `xformers` package to speed up generation | |
| - `--port`: Specify a server port to use | |
| - defaults to `7860` | |
| - `--api`: Enable [API](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/API) access | |
| <br> | |
| - Once you have successfully launched the WebUI, you can add the following flags to bypass some validation steps in order to improve the Startup time | |
| - `--skip-prepare-environment` | |
| - `--skip-install` | |
| - `--skip-python-version-check` | |
| - `--skip-torch-cuda-test` | |
| - `--skip-version-check` | |
| > [!Important] | |
| > Remove them if you are installing an Extension, as those also block Extension from installing requirements | |
| #### by. Forge | |
| - For RTX **30** and above, you can add the following flags to slightly increase the performance; but in rare occurrences, they may cause `OutOfMemory` errors or even crash the WebUI; and in certain configurations, they may even lower the speed instead | |
| - `--cuda-malloc` | |
| - `--cuda-stream` | |
| - `--pin-shared-memory` | |
| #### by. Classic | |
| - `--uv`: Replace the `python -m pip` calls with `uv pip` to massively speed up package installation | |
| - requires **uv** to be installed first *(see [Installation](#installation))* | |
| - `--uv-symlink`: Same as above; but additionally pass `--link-mode symlink` to the commands | |
| - significantly reduces installation size (`~7 GB` to `~100 MB`) | |
| > [!Important] | |
| > Using `symlink` means it will directly access the packages from the cache folders; refrain from clearing the cache when setting this option | |
| - `--model-ref`: Points to a central `models` folder that contains all your models | |
| - said folder should contain subfolders like `Stable-diffusion`, `Lora`, `VAE`, `ESRGAN`, etc. | |
| > [!Important] | |
| > This simply **replaces** the `models` folder, rather than adding on top of it | |
| - `--persistent-patches`: Enable the persistent LoRA patching | |
| - no longer apply LoRA every single generation, if the weight is unchanged | |
| - save around 1 second per generation when using LoRA | |
| - `--fast-fp16`: Enable the `allow_fp16_accumulation` option | |
| - requires PyTorch **2.7.0** + | |
| - `--sage`: Install the `sageattention` package to speed up generation | |
| - requires **triton** | |
| - requires RTX **30** + | |
| - only affects **SDXL** | |
| > [!Note] | |
| > For RTX **50** users, you may need to manually [install](#install-sageattention-2) `sageattention 2` instead | |
| <details> | |
| <summary>with SageAttention 2</summary> | |
| - `--sage2-function`: Select the function used by **SageAttention 2** | |
| - **options:** | |
| - `auto` (default) | |
| - `fp16_triton` | |
| - `fp16_cuda` | |
| - `fp8_cuda` | |
| - If you are getting `NaN` errors, try: | |
| ```bash | |
| --sage2-function fp16_cuda --sage-quant-gran per_warp --sage-accum-dtype fp16+fp32 | |
| ``` | |
| </details> | |
| <br> | |
| ## Installation | |
| 0. Install **[git](https://git-scm.com/downloads)** | |
| 1. Clone the Repo | |
| ```bash | |
| git clone https://github.com/Haoming02/sd-webui-forge-classic | |
| ``` | |
| 2. Setup Python | |
| <details> | |
| <summary>Recommended Method</summary> | |
| - Install **[uv](https://github.com/astral-sh/uv#installation)** | |
| - Set up **venv** | |
| ```bash | |
| cd sd-webui-forge-classic | |
| uv venv venv --python 3.11 --seed | |
| ``` | |
| - Add the `--uv` flag to `webui-user.bat` | |
| </details> | |
| <details> | |
| <summary>Standard Method</summary> | |
| - Install **[Python 3.11.9](https://www.python.org/downloads/release/python-3119/)** | |
| - Remember to enable `Add Python to PATH` | |
| </details> | |
| 3. **(Optional)** Configure [Commandline](#commandline) | |
| 4. Launch the WebUI via `webui-user.bat` | |
| 5. During the first launch, it will automatically install all the requirements | |
| 6. Once the installation is finished, the WebUI will start in a browser automatically | |
| <br> | |
| ### Install cublas | |
| <details> | |
| <summary>Expand</summary> | |
| 0. Ensure the WebUI can properly launch already, by following the [installation](#installation) steps first | |
| 1. Open the console in the WebUI directory | |
| ```bash | |
| cd sd-webui-forge-classic | |
| ``` | |
| 2. Start the virtual environment | |
| ```bash | |
| venv\scripts\activate | |
| ``` | |
| 3. Create a new folder | |
| ```bash | |
| mkdir repo | |
| cd repo | |
| ``` | |
| 4. Clone the repo | |
| ```bash | |
| git clone https://github.com/aredden/torch-cublas-hgemm | |
| cd torch-cublas-hgemm | |
| ``` | |
| 5. Install the library | |
| ``` | |
| pip install -e . --no-build-isolation | |
| ``` | |
| - If you installed `uv`, use `uv pip install` instead | |
| - The installation takes a few minutes | |
| </details> | |
| ### Install triton | |
| <details> | |
| <summary>Expand</summary> | |
| 0. Ensure the WebUI can properly launch already, by following the [installation](#installation) steps first | |
| 1. Open the console in the WebUI directory | |
| ```bash | |
| cd sd-webui-forge-classic | |
| ``` | |
| 2. Start the virtual environment | |
| ```bash | |
| venv\scripts\activate | |
| ``` | |
| 3. Install the library | |
| - **Windows** | |
| ```bash | |
| pip install triton-windows | |
| ``` | |
| - **Linux** | |
| ```bash | |
| pip install triton | |
| ``` | |
| - If you installed `uv`, use `uv pip install` instead | |
| </details> | |
| ### Install flash-attn | |
| <details> | |
| <summary>Expand</summary> | |
| 0. Ensure the WebUI can properly launch already, by following the [installation](#installation) steps first | |
| 1. Open the console in the WebUI directory | |
| ```bash | |
| cd sd-webui-forge-classic | |
| ``` | |
| 2. Start the virtual environment | |
| ```bash | |
| venv\scripts\activate | |
| ``` | |
| 3. Install the library | |
| - **Windows** | |
| - Download the pre-built `.whl` package from https://github.com/kingbri1/flash-attention/releases | |
| ```bash | |
| pip install flash_attn...win...whl | |
| ``` | |
| - **Linux** | |
| - Download the pre-built `.whl` package from https://github.com/Dao-AILab/flash-attention/releases | |
| ```bash | |
| pip install flash_attn...linux...whl | |
| ``` | |
| - If you installed `uv`, use `uv pip install` instead | |
| - **Important:** Download the correct `.whl` for your Python and PyTorch version | |
| </details> | |
| ### Install sageattention 2 | |
| <details> | |
| <summary>Expand</summary> | |
| 0. Ensure the WebUI can properly launch already, by following the [installation](#installation) steps first | |
| 1. Open the console in the WebUI directory | |
| ```bash | |
| cd sd-webui-forge-classic | |
| ``` | |
| 2. Start the virtual environment | |
| ```bash | |
| venv\scripts\activate | |
| ``` | |
| 3. Create a new folder | |
| ```bash | |
| mkdir repo | |
| cd repo | |
| ``` | |
| 4. Clone the repo | |
| ```bash | |
| git clone https://github.com/thu-ml/SageAttention | |
| cd SageAttention | |
| ``` | |
| 5. Install the library | |
| ``` | |
| pip install -e . --no-build-isolation | |
| ``` | |
| - If you installed `uv`, use `uv pip install` instead | |
| - The installation takes a few minutes | |
| <br> | |
| ### Alternatively | |
| > for **Windows** | |
| - Download the pre-built `.whl` package from https://github.com/woct0rdho/SageAttention/releases | |
| ```bash | |
| pip install sageattention...win_amd64.whl | |
| ``` | |
| - If you installed `uv`, use `uv pip install` instead | |
| - **Important:** Download the correct `.whl` for your PyTorch version | |
| </details> | |
| ### Install older PyTorch | |
| <details> | |
| <summary>Expand</summary> | |
| 0. Navigate to the WebUI directory | |
| 1. Edit the `webui-user.bat` file | |
| 2. Add a new line to specify an older version: | |
| ```bash | |
| set TORCH_COMMAND=pip install torch==2.1.2 torchvision==0.16.2 --extra-index-url https://download.pytorch.org/whl/cu121 | |
| ``` | |
| </details> | |
| <br> | |
| ## Attention | |
| > [!Important] | |
| > The `--xformers` and `--sage` args are only responsible for installing the packages, **not** whether its respective attention is used *(this also means you can remove them once the packages are successfully installed)* | |
| **Forge Classic** tries to import the packages and automatically choose the first available attention function in the following order: | |
| 1. `SageAttention` | |
| 2. `FlashAttention` | |
| 3. `xformers` | |
| 4. `PyTorch` | |
| 5. `Basic` | |
| > [!Tip] | |
| > To skip a specific attention, add the respective disable arg such as `--disable-sage` | |
| > [!Note] | |
| > The **VAE** only checks for `xformers`, so `--xformers` is still recommended even if you already have `--sage` | |
| In my experience, the speed of each attention function for SDXL is ranked in the following order: | |
| - `SageAttention` ≥ `FlashAttention` > `xformers` > `PyTorch` >> `Basic` | |
| > [!Note] | |
| > `SageAttention` is based on quantization, so its quality might be slightly worse than others | |
| > [!Important] | |
| > When using `SageAttention 2`, both positive prompts and negative prompts are required; omitting negative prompts can cause `NaN` issues | |
| <br> | |
| ## Issues & Requests | |
| - **Issues** about removed features will simply be ignored | |
| - **Issues** regarding installation will be ignored if it's obviously user-error | |
| - **Feature Request** not related to performance or optimization will simply be ignored | |
| - For cutting edge features, check out [reForge](https://github.com/Panchovix/stable-diffusion-webui-reForge) instead | |
| - Non-Windows platforms will not be supported, as I cannot verify nor maintain them | |
| </details> | |
| <hr> | |
| <p align="center"> | |
| Special thanks to <b>AUTOMATIC1111</b>, <b>lllyasviel</b>, and <b>comfyanonymous</b>, <b>kijai</b>, <br> | |
| along with the rest of the contributors, <br> | |
| for their invaluable efforts in the open-source image generation community | |
| </p> | |