|
|
import base64 |
|
|
import datetime |
|
|
import io |
|
|
import ipaddress |
|
|
import itertools |
|
|
import os |
|
|
import time |
|
|
from contextlib import closing |
|
|
from io import BytesIO |
|
|
from secrets import compare_digest |
|
|
from threading import Lock |
|
|
from typing import Any |
|
|
|
|
|
import gradio as gr |
|
|
import modules.shared as shared |
|
|
import numpy as np |
|
|
import piexif |
|
|
import piexif.helper |
|
|
import requests |
|
|
import uvicorn |
|
|
from fastapi import APIRouter, Depends, FastAPI, Request, Response |
|
|
from fastapi.encoders import jsonable_encoder |
|
|
from fastapi.exceptions import HTTPException |
|
|
from fastapi.responses import JSONResponse |
|
|
from fastapi.security import HTTPBasic, HTTPBasicCredentials |
|
|
from modules import ( |
|
|
errors, |
|
|
images, |
|
|
infotext_utils, |
|
|
postprocessing, |
|
|
restart, |
|
|
script_callbacks, |
|
|
scripts, |
|
|
sd_hijack, |
|
|
sd_models, |
|
|
sd_samplers, |
|
|
sd_schedulers, |
|
|
shared_items, |
|
|
ui, |
|
|
) |
|
|
from modules.api import models |
|
|
from modules.processing import ( |
|
|
StableDiffusionProcessingImg2Img, |
|
|
StableDiffusionProcessingTxt2Img, |
|
|
process_images, |
|
|
) |
|
|
from modules.progress import ( |
|
|
add_task_to_queue, |
|
|
create_task_id, |
|
|
current_task, |
|
|
finish_task, |
|
|
start_task, |
|
|
) |
|
|
from modules.sd_models_config import find_checkpoint_config_near_filename |
|
|
from modules.shared import opts |
|
|
from PIL import Image, PngImagePlugin |
|
|
|
|
|
|
|
|
def script_name_to_index(name, scripts): |
|
|
try: |
|
|
return [script.title().lower() for script in scripts].index(name.lower()) |
|
|
except Exception as e: |
|
|
raise HTTPException(status_code=422, detail=f"Script '{name}' not found") from e |
|
|
|
|
|
|
|
|
def validate_sampler_name(name): |
|
|
config = sd_samplers.all_samplers_map.get(name, None) |
|
|
if config is None: |
|
|
raise HTTPException(status_code=404, detail="Sampler not found") |
|
|
|
|
|
return name |
|
|
|
|
|
|
|
|
def setUpscalers(req: dict): |
|
|
reqDict = vars(req) |
|
|
reqDict["extras_upscaler_1"] = reqDict.pop("upscaler_1", None) |
|
|
reqDict["extras_upscaler_2"] = reqDict.pop("upscaler_2", None) |
|
|
return reqDict |
|
|
|
|
|
|
|
|
def verify_url(url): |
|
|
"""Returns True if the url refers to a global resource""" |
|
|
|
|
|
import socket |
|
|
from urllib.parse import urlparse |
|
|
|
|
|
try: |
|
|
parsed_url = urlparse(url) |
|
|
domain_name = parsed_url.netloc |
|
|
host = socket.gethostbyname_ex(domain_name) |
|
|
for ip in host[2]: |
|
|
ip_addr = ipaddress.ip_address(ip) |
|
|
if not ip_addr.is_global: |
|
|
return False |
|
|
except Exception: |
|
|
return False |
|
|
|
|
|
return True |
|
|
|
|
|
|
|
|
def decode_base64_to_image(encoding): |
|
|
if encoding.startswith("http://") or encoding.startswith("https://"): |
|
|
if not opts.api_enable_requests: |
|
|
raise HTTPException(status_code=500, detail="Requests not allowed") |
|
|
|
|
|
if opts.api_forbid_local_requests and not verify_url(encoding): |
|
|
raise HTTPException(status_code=500, detail="Request to local resource not allowed") |
|
|
|
|
|
headers = {"user-agent": opts.api_useragent} if opts.api_useragent else {} |
|
|
response = requests.get(encoding, timeout=30, headers=headers) |
|
|
try: |
|
|
image = Image.open(BytesIO(response.content)) |
|
|
return image |
|
|
except Exception as e: |
|
|
raise HTTPException(status_code=500, detail="Invalid image url") from e |
|
|
|
|
|
if encoding.startswith("data:image/"): |
|
|
encoding = encoding.split(";")[1].split(",")[1] |
|
|
try: |
|
|
image = Image.open(BytesIO(base64.b64decode(encoding))) |
|
|
return image |
|
|
except Exception as e: |
|
|
raise HTTPException(status_code=500, detail="Invalid encoded image") from e |
|
|
|
|
|
|
|
|
def encode_pil_to_base64(image): |
|
|
with io.BytesIO() as output_bytes: |
|
|
if isinstance(image, str): |
|
|
return image |
|
|
if isinstance(image, np.ndarray): |
|
|
image = Image.fromarray(image) |
|
|
if opts.samples_format.lower() == "png": |
|
|
use_metadata = False |
|
|
metadata = PngImagePlugin.PngInfo() |
|
|
for key, value in image.info.items(): |
|
|
if isinstance(key, str) and isinstance(value, str): |
|
|
metadata.add_text(key, value) |
|
|
use_metadata = True |
|
|
image.save(output_bytes, format="PNG", pnginfo=(metadata if use_metadata else None), quality=opts.jpeg_quality) |
|
|
|
|
|
elif opts.samples_format.lower() in ("jpg", "jpeg", "webp"): |
|
|
if image.mode == "RGBA": |
|
|
image = image.convert("RGB") |
|
|
parameters = image.info.get("parameters", None) |
|
|
exif_bytes = piexif.dump({"Exif": {piexif.ExifIFD.UserComment: piexif.helper.UserComment.dump(parameters or "", encoding="unicode")}}) |
|
|
if opts.samples_format.lower() in ("jpg", "jpeg"): |
|
|
image.save(output_bytes, format="JPEG", exif=exif_bytes, quality=opts.jpeg_quality) |
|
|
else: |
|
|
image.save(output_bytes, format="WEBP", exif=exif_bytes, quality=opts.jpeg_quality) |
|
|
|
|
|
else: |
|
|
raise HTTPException(status_code=500, detail="Invalid image format") |
|
|
|
|
|
bytes_data = output_bytes.getvalue() |
|
|
|
|
|
return base64.b64encode(bytes_data) |
|
|
|
|
|
|
|
|
def api_middleware(app: FastAPI): |
|
|
rich_available = False |
|
|
try: |
|
|
if os.environ.get("WEBUI_RICH_EXCEPTIONS", None) is not None: |
|
|
import anyio |
|
|
import starlette |
|
|
from rich.console import Console |
|
|
|
|
|
console = Console() |
|
|
rich_available = True |
|
|
except Exception: |
|
|
pass |
|
|
|
|
|
@app.middleware("http") |
|
|
async def log_and_time(req: Request, call_next): |
|
|
ts = time.time() |
|
|
res: Response = await call_next(req) |
|
|
duration = str(round(time.time() - ts, 4)) |
|
|
res.headers["X-Process-Time"] = duration |
|
|
endpoint = req.scope.get("path", "err") |
|
|
if shared.cmd_opts.api_log and endpoint.startswith("/sdapi"): |
|
|
print( |
|
|
"API {t} {code} {prot}/{ver} {method} {endpoint} {cli} {duration}".format( |
|
|
t=datetime.datetime.now().strftime("%Y-%m-%d %H:%M:%S.%f"), |
|
|
code=res.status_code, |
|
|
ver=req.scope.get("http_version", "0.0"), |
|
|
cli=req.scope.get("client", ("0:0.0.0", 0))[0], |
|
|
prot=req.scope.get("scheme", "err"), |
|
|
method=req.scope.get("method", "err"), |
|
|
endpoint=endpoint, |
|
|
duration=duration, |
|
|
) |
|
|
) |
|
|
return res |
|
|
|
|
|
def handle_exception(request: Request, e: Exception): |
|
|
err = { |
|
|
"error": type(e).__name__, |
|
|
"detail": vars(e).get("detail", ""), |
|
|
"body": vars(e).get("body", ""), |
|
|
"errors": str(e), |
|
|
} |
|
|
if not isinstance(e, HTTPException): |
|
|
message = f"API error: {request.method}: {request.url} {err}" |
|
|
if rich_available: |
|
|
print(message) |
|
|
console.print_exception(show_locals=True, max_frames=2, extra_lines=1, suppress=[anyio, starlette], word_wrap=False, width=min([console.width, 200])) |
|
|
else: |
|
|
errors.report(message, exc_info=True) |
|
|
return JSONResponse(status_code=vars(e).get("status_code", 500), content=jsonable_encoder(err)) |
|
|
|
|
|
@app.middleware("http") |
|
|
async def exception_handling(request: Request, call_next): |
|
|
try: |
|
|
return await call_next(request) |
|
|
except Exception as e: |
|
|
return handle_exception(request, e) |
|
|
|
|
|
@app.exception_handler(Exception) |
|
|
async def fastapi_exception_handler(request: Request, e: Exception): |
|
|
return handle_exception(request, e) |
|
|
|
|
|
@app.exception_handler(HTTPException) |
|
|
async def http_exception_handler(request: Request, e: HTTPException): |
|
|
return handle_exception(request, e) |
|
|
|
|
|
|
|
|
class Api: |
|
|
def __init__(self, app: FastAPI, queue_lock: Lock): |
|
|
if shared.cmd_opts.api_auth: |
|
|
self.credentials = {} |
|
|
for auth in shared.cmd_opts.api_auth.split(","): |
|
|
user, password = auth.split(":") |
|
|
self.credentials[user] = password |
|
|
|
|
|
self.router = APIRouter() |
|
|
self.app = app |
|
|
self.queue_lock = queue_lock |
|
|
api_middleware(self.app) |
|
|
self.add_api_route("/sdapi/v1/txt2img", self.text2imgapi, methods=["POST"], response_model=models.TextToImageResponse) |
|
|
self.add_api_route("/sdapi/v1/img2img", self.img2imgapi, methods=["POST"], response_model=models.ImageToImageResponse) |
|
|
self.add_api_route("/sdapi/v1/extra-single-image", self.extras_single_image_api, methods=["POST"], response_model=models.ExtrasSingleImageResponse) |
|
|
self.add_api_route("/sdapi/v1/extra-batch-images", self.extras_batch_images_api, methods=["POST"], response_model=models.ExtrasBatchImagesResponse) |
|
|
self.add_api_route("/sdapi/v1/png-info", self.pnginfoapi, methods=["POST"], response_model=models.PNGInfoResponse) |
|
|
self.add_api_route("/sdapi/v1/progress", self.progressapi, methods=["GET"], response_model=models.ProgressResponse) |
|
|
self.add_api_route("/sdapi/v1/interrupt", self.interruptapi, methods=["POST"]) |
|
|
self.add_api_route("/sdapi/v1/skip", self.skip, methods=["POST"]) |
|
|
self.add_api_route("/sdapi/v1/options", self.get_config, methods=["GET"], response_model=models.OptionsModel) |
|
|
self.add_api_route("/sdapi/v1/options", self.set_config, methods=["POST"]) |
|
|
self.add_api_route("/sdapi/v1/cmd-flags", self.get_cmd_flags, methods=["GET"], response_model=models.FlagsModel) |
|
|
self.add_api_route("/sdapi/v1/samplers", self.get_samplers, methods=["GET"], response_model=list[models.SamplerItem]) |
|
|
self.add_api_route("/sdapi/v1/schedulers", self.get_schedulers, methods=["GET"], response_model=list[models.SchedulerItem]) |
|
|
self.add_api_route("/sdapi/v1/upscalers", self.get_upscalers, methods=["GET"], response_model=list[models.UpscalerItem]) |
|
|
self.add_api_route("/sdapi/v1/latent-upscale-modes", self.get_latent_upscale_modes, methods=["GET"], response_model=list[models.LatentUpscalerModeItem]) |
|
|
self.add_api_route("/sdapi/v1/sd-models", self.get_sd_models, methods=["GET"], response_model=list[models.SDModelItem]) |
|
|
self.add_api_route("/sdapi/v1/sd-vae", self.get_sd_vaes, methods=["GET"], response_model=list[models.SDVaeItem]) |
|
|
self.add_api_route("/sdapi/v1/face-restorers", self.get_face_restorers, methods=["GET"], response_model=list[models.FaceRestorerItem]) |
|
|
self.add_api_route("/sdapi/v1/prompt-styles", self.get_prompt_styles, methods=["GET"], response_model=list[models.PromptStyleItem]) |
|
|
self.add_api_route("/sdapi/v1/embeddings", self.get_embeddings, methods=["GET"], response_model=models.EmbeddingsResponse) |
|
|
self.add_api_route("/sdapi/v1/refresh-embeddings", self.refresh_embeddings, methods=["POST"]) |
|
|
self.add_api_route("/sdapi/v1/refresh-checkpoints", self.refresh_checkpoints, methods=["POST"]) |
|
|
self.add_api_route("/sdapi/v1/refresh-vae", self.refresh_vae, methods=["POST"]) |
|
|
self.add_api_route("/sdapi/v1/memory", self.get_memory, methods=["GET"], response_model=models.MemoryResponse) |
|
|
self.add_api_route("/sdapi/v1/unload-checkpoint", self.unloadapi, methods=["POST"]) |
|
|
self.add_api_route("/sdapi/v1/reload-checkpoint", self.reloadapi, methods=["POST"]) |
|
|
self.add_api_route("/sdapi/v1/scripts", self.get_scripts_list, methods=["GET"], response_model=models.ScriptsList) |
|
|
self.add_api_route("/sdapi/v1/script-info", self.get_script_info, methods=["GET"], response_model=list[models.ScriptInfo]) |
|
|
self.add_api_route("/sdapi/v1/extensions", self.get_extensions_list, methods=["GET"], response_model=list[models.ExtensionItem]) |
|
|
|
|
|
if shared.cmd_opts.api_server_stop: |
|
|
self.add_api_route("/sdapi/v1/server-kill", self.kill_webui, methods=["POST"]) |
|
|
self.add_api_route("/sdapi/v1/server-restart", self.restart_webui, methods=["POST"]) |
|
|
self.add_api_route("/sdapi/v1/server-stop", self.stop_webui, methods=["POST"]) |
|
|
|
|
|
self.default_script_arg_txt2img = [] |
|
|
self.default_script_arg_img2img = [] |
|
|
|
|
|
txt2img_script_runner = scripts.scripts_txt2img |
|
|
img2img_script_runner = scripts.scripts_img2img |
|
|
|
|
|
if not txt2img_script_runner.scripts or not img2img_script_runner.scripts: |
|
|
ui.create_ui() |
|
|
|
|
|
if not txt2img_script_runner.scripts: |
|
|
txt2img_script_runner.initialize_scripts(False) |
|
|
if not self.default_script_arg_txt2img: |
|
|
self.default_script_arg_txt2img = self.init_default_script_args(txt2img_script_runner) |
|
|
|
|
|
if not img2img_script_runner.scripts: |
|
|
img2img_script_runner.initialize_scripts(True) |
|
|
if not self.default_script_arg_img2img: |
|
|
self.default_script_arg_img2img = self.init_default_script_args(img2img_script_runner) |
|
|
|
|
|
def add_api_route(self, path: str, endpoint, **kwargs): |
|
|
if shared.cmd_opts.api_auth: |
|
|
return self.app.add_api_route(path, endpoint, dependencies=[Depends(self.auth)], **kwargs) |
|
|
return self.app.add_api_route(path, endpoint, **kwargs) |
|
|
|
|
|
def auth(self, credentials: HTTPBasicCredentials = Depends(HTTPBasic())): |
|
|
if credentials.username in self.credentials: |
|
|
if compare_digest(credentials.password, self.credentials[credentials.username]): |
|
|
return True |
|
|
|
|
|
raise HTTPException(status_code=401, detail="Incorrect username or password", headers={"WWW-Authenticate": "Basic"}) |
|
|
|
|
|
def get_selectable_script(self, script_name, script_runner): |
|
|
if script_name is None or script_name == "": |
|
|
return None, None |
|
|
|
|
|
script_idx = script_name_to_index(script_name, script_runner.selectable_scripts) |
|
|
script = script_runner.selectable_scripts[script_idx] |
|
|
return script, script_idx |
|
|
|
|
|
def get_scripts_list(self): |
|
|
t2ilist = [script.name for script in scripts.scripts_txt2img.scripts if script.name is not None] |
|
|
i2ilist = [script.name for script in scripts.scripts_img2img.scripts if script.name is not None] |
|
|
|
|
|
return models.ScriptsList(txt2img=t2ilist, img2img=i2ilist) |
|
|
|
|
|
def get_script_info(self): |
|
|
res = [] |
|
|
|
|
|
for script_list in [scripts.scripts_txt2img.scripts, scripts.scripts_img2img.scripts]: |
|
|
res += [script.api_info for script in script_list if script.api_info is not None] |
|
|
|
|
|
return res |
|
|
|
|
|
def get_script(self, script_name, script_runner): |
|
|
if script_name is None or script_name == "": |
|
|
return None, None |
|
|
|
|
|
script_idx = script_name_to_index(script_name, script_runner.scripts) |
|
|
return script_runner.scripts[script_idx] |
|
|
|
|
|
def init_default_script_args(self, script_runner): |
|
|
|
|
|
last_arg_index = 1 |
|
|
for script in script_runner.scripts: |
|
|
if last_arg_index < script.args_to: |
|
|
last_arg_index = script.args_to |
|
|
|
|
|
script_args = [None] * last_arg_index |
|
|
script_args[0] = 0 |
|
|
|
|
|
|
|
|
with gr.Blocks(): |
|
|
for script in script_runner.scripts: |
|
|
if script.ui(script.is_img2img): |
|
|
ui_default_values = [] |
|
|
for elem in script.ui(script.is_img2img): |
|
|
ui_default_values.append(elem.value) |
|
|
script_args[script.args_from : script.args_to] = ui_default_values |
|
|
return script_args |
|
|
|
|
|
def init_script_args(self, request, default_script_args, selectable_scripts, selectable_idx, script_runner, *, input_script_args=None): |
|
|
script_args = default_script_args.copy() |
|
|
|
|
|
if input_script_args is not None: |
|
|
for index, value in input_script_args.items(): |
|
|
script_args[index] = value |
|
|
|
|
|
|
|
|
if selectable_scripts: |
|
|
script_args[selectable_scripts.args_from : selectable_scripts.args_to] = request.script_args |
|
|
script_args[0] = selectable_idx + 1 |
|
|
|
|
|
|
|
|
if request.alwayson_scripts: |
|
|
for alwayson_script_name in request.alwayson_scripts.keys(): |
|
|
alwayson_script = self.get_script(alwayson_script_name, script_runner) |
|
|
if alwayson_script is None: |
|
|
raise HTTPException(status_code=422, detail=f"always on script {alwayson_script_name} not found") |
|
|
|
|
|
if alwayson_script.alwayson is False: |
|
|
raise HTTPException(status_code=422, detail="Cannot have a selectable script in the always on scripts params") |
|
|
|
|
|
if "args" in request.alwayson_scripts[alwayson_script_name]: |
|
|
|
|
|
for idx in range(0, min((alwayson_script.args_to - alwayson_script.args_from), len(request.alwayson_scripts[alwayson_script_name]["args"]))): |
|
|
script_args[alwayson_script.args_from + idx] = request.alwayson_scripts[alwayson_script_name]["args"][idx] |
|
|
return script_args |
|
|
|
|
|
def apply_infotext(self, request, tabname, *, script_runner=None, mentioned_script_args=None): |
|
|
""" |
|
|
Processes `infotext` field from the `request`, and sets other fields of the `request` according to what's in infotext. |
|
|
If request already has a field set, and that field is encountered in infotext too, the value from infotext is ignored. |
|
|
Additionally, fills `mentioned_script_args` dict with index: value pairs for script arguments read from infotext. |
|
|
""" |
|
|
|
|
|
if not request.infotext: |
|
|
return {} |
|
|
|
|
|
possible_fields = infotext_utils.paste_fields[tabname]["fields"] |
|
|
set_fields = request.model_dump(exclude_unset=True) if hasattr(request, "request") else request.dict(exclude_unset=True) |
|
|
params = infotext_utils.parse_generation_parameters(request.infotext) |
|
|
|
|
|
def get_field_value(field, params): |
|
|
value = field.function(params) if field.function else params.get(field.label) |
|
|
if value is None: |
|
|
return None |
|
|
|
|
|
if field.api in request.__fields__: |
|
|
target_type = request.__fields__[field.api].type_ |
|
|
else: |
|
|
target_type = type(field.component.value) |
|
|
|
|
|
if target_type is type(None): |
|
|
return None |
|
|
|
|
|
if isinstance(value, dict) and value.get("__type__") == "generic_update": |
|
|
value = value.get("value") |
|
|
|
|
|
if value is not None and not isinstance(value, target_type): |
|
|
value = target_type(value) |
|
|
|
|
|
return value |
|
|
|
|
|
for field in possible_fields: |
|
|
if not field.api: |
|
|
continue |
|
|
|
|
|
if field.api in set_fields: |
|
|
continue |
|
|
|
|
|
value = get_field_value(field, params) |
|
|
if value is not None: |
|
|
setattr(request, field.api, value) |
|
|
|
|
|
if request.override_settings is None: |
|
|
request.override_settings = {} |
|
|
|
|
|
overridden_settings = infotext_utils.get_override_settings(params) |
|
|
for _, setting_name, value in overridden_settings: |
|
|
if setting_name not in request.override_settings: |
|
|
request.override_settings[setting_name] = value |
|
|
|
|
|
if script_runner is not None and mentioned_script_args is not None: |
|
|
indexes = {v: i for i, v in enumerate(script_runner.inputs)} |
|
|
script_fields = ((field, indexes[field.component]) for field in possible_fields if field.component in indexes) |
|
|
|
|
|
for field, index in script_fields: |
|
|
value = get_field_value(field, params) |
|
|
|
|
|
if value is None: |
|
|
continue |
|
|
|
|
|
mentioned_script_args[index] = value |
|
|
|
|
|
return params |
|
|
|
|
|
def text2imgapi(self, txt2imgreq: models.StableDiffusionTxt2ImgProcessingAPI): |
|
|
task_id = txt2imgreq.force_task_id or create_task_id("txt2img") |
|
|
|
|
|
script_runner = scripts.scripts_txt2img |
|
|
|
|
|
infotext_script_args = {} |
|
|
self.apply_infotext(txt2imgreq, "txt2img", script_runner=script_runner, mentioned_script_args=infotext_script_args) |
|
|
|
|
|
selectable_scripts, selectable_script_idx = self.get_selectable_script(txt2imgreq.script_name, script_runner) |
|
|
|
|
|
populate = txt2imgreq.copy( |
|
|
update={ |
|
|
"sampler_name": validate_sampler_name(txt2imgreq.sampler_name or txt2imgreq.sampler_index), |
|
|
"do_not_save_samples": not txt2imgreq.save_images, |
|
|
"do_not_save_grid": not txt2imgreq.save_images, |
|
|
} |
|
|
) |
|
|
if populate.sampler_name: |
|
|
populate.sampler_index = None |
|
|
|
|
|
args = vars(populate) |
|
|
args.pop("script_name", None) |
|
|
args.pop("script_args", None) |
|
|
args.pop("alwayson_scripts", None) |
|
|
args.pop("infotext", None) |
|
|
|
|
|
script_args = self.init_script_args(txt2imgreq, self.default_script_arg_txt2img, selectable_scripts, selectable_script_idx, script_runner, input_script_args=infotext_script_args) |
|
|
|
|
|
send_images = args.pop("send_images", True) |
|
|
args.pop("save_images", None) |
|
|
|
|
|
add_task_to_queue(task_id) |
|
|
|
|
|
with self.queue_lock: |
|
|
with closing(StableDiffusionProcessingTxt2Img(sd_model=shared.sd_model, **args)) as p: |
|
|
p.is_api = True |
|
|
p.scripts = script_runner |
|
|
p.outpath_grids = opts.outdir_txt2img_grids |
|
|
p.outpath_samples = opts.outdir_txt2img_samples |
|
|
|
|
|
try: |
|
|
shared.state.begin(job="scripts_txt2img") |
|
|
start_task(task_id) |
|
|
if selectable_scripts is not None: |
|
|
p.script_args = script_args |
|
|
processed = scripts.scripts_txt2img.run(p, *p.script_args) |
|
|
else: |
|
|
p.script_args = tuple(script_args) |
|
|
processed = process_images(p) |
|
|
finish_task(task_id) |
|
|
finally: |
|
|
shared.state.end() |
|
|
shared.total_tqdm.clear() |
|
|
|
|
|
b64images = [encode_pil_to_base64(image) for image in itertools.chain(processed.images, processed.extra_images) if send_images] |
|
|
|
|
|
return models.TextToImageResponse(images=b64images, parameters=vars(txt2imgreq), info=processed.js()) |
|
|
|
|
|
def img2imgapi(self, img2imgreq: models.StableDiffusionImg2ImgProcessingAPI): |
|
|
task_id = img2imgreq.force_task_id or create_task_id("img2img") |
|
|
|
|
|
init_images = img2imgreq.init_images |
|
|
if init_images is None: |
|
|
raise HTTPException(status_code=404, detail="Init image not found") |
|
|
|
|
|
mask = img2imgreq.mask |
|
|
if mask: |
|
|
mask = decode_base64_to_image(mask) |
|
|
|
|
|
script_runner = scripts.scripts_img2img |
|
|
|
|
|
infotext_script_args = {} |
|
|
self.apply_infotext(img2imgreq, "img2img", script_runner=script_runner, mentioned_script_args=infotext_script_args) |
|
|
|
|
|
selectable_scripts, selectable_script_idx = self.get_selectable_script(img2imgreq.script_name, script_runner) |
|
|
|
|
|
populate = img2imgreq.copy( |
|
|
update={ |
|
|
"sampler_name": validate_sampler_name(img2imgreq.sampler_name or img2imgreq.sampler_index), |
|
|
"do_not_save_samples": not img2imgreq.save_images, |
|
|
"do_not_save_grid": not img2imgreq.save_images, |
|
|
"mask": mask, |
|
|
} |
|
|
) |
|
|
if populate.sampler_name: |
|
|
populate.sampler_index = None |
|
|
|
|
|
args = vars(populate) |
|
|
args.pop("include_init_images", None) |
|
|
args.pop("script_name", None) |
|
|
args.pop("script_args", None) |
|
|
args.pop("alwayson_scripts", None) |
|
|
args.pop("infotext", None) |
|
|
|
|
|
script_args = self.init_script_args(img2imgreq, self.default_script_arg_img2img, selectable_scripts, selectable_script_idx, script_runner, input_script_args=infotext_script_args) |
|
|
|
|
|
send_images = args.pop("send_images", True) |
|
|
args.pop("save_images", None) |
|
|
|
|
|
add_task_to_queue(task_id) |
|
|
|
|
|
with self.queue_lock: |
|
|
with closing(StableDiffusionProcessingImg2Img(sd_model=shared.sd_model, **args)) as p: |
|
|
p.init_images = [decode_base64_to_image(x) for x in init_images] |
|
|
p.is_api = True |
|
|
p.scripts = script_runner |
|
|
p.outpath_grids = opts.outdir_img2img_grids |
|
|
p.outpath_samples = opts.outdir_img2img_samples |
|
|
|
|
|
try: |
|
|
shared.state.begin(job="scripts_img2img") |
|
|
start_task(task_id) |
|
|
if selectable_scripts is not None: |
|
|
p.script_args = script_args |
|
|
processed = scripts.scripts_img2img.run(p, *p.script_args) |
|
|
else: |
|
|
p.script_args = tuple(script_args) |
|
|
processed = process_images(p) |
|
|
finish_task(task_id) |
|
|
finally: |
|
|
shared.state.end() |
|
|
shared.total_tqdm.clear() |
|
|
|
|
|
b64images = [encode_pil_to_base64(image) for image in itertools.chain(processed.images, processed.extra_images) if send_images] |
|
|
|
|
|
if not img2imgreq.include_init_images: |
|
|
img2imgreq.init_images = None |
|
|
img2imgreq.mask = None |
|
|
|
|
|
return models.ImageToImageResponse(images=b64images, parameters=vars(img2imgreq), info=processed.js()) |
|
|
|
|
|
def extras_single_image_api(self, req: models.ExtrasSingleImageRequest): |
|
|
reqDict = setUpscalers(req) |
|
|
|
|
|
reqDict["image"] = decode_base64_to_image(reqDict["image"]) |
|
|
|
|
|
with self.queue_lock: |
|
|
result = postprocessing.run_extras(extras_mode=0, image_folder="", input_dir="", output_dir="", save_output=False, **reqDict) |
|
|
|
|
|
return models.ExtrasSingleImageResponse(image=encode_pil_to_base64(result[0][0]), html_info=result[1]) |
|
|
|
|
|
def extras_batch_images_api(self, req: models.ExtrasBatchImagesRequest): |
|
|
reqDict = setUpscalers(req) |
|
|
|
|
|
image_list = reqDict.pop("imageList", []) |
|
|
image_folder = [decode_base64_to_image(x.data) for x in image_list] |
|
|
|
|
|
with self.queue_lock: |
|
|
result = postprocessing.run_extras(extras_mode=1, image_folder=image_folder, image="", input_dir="", output_dir="", save_output=False, **reqDict) |
|
|
|
|
|
return models.ExtrasBatchImagesResponse(images=list(map(encode_pil_to_base64, result[0])), html_info=result[1]) |
|
|
|
|
|
def pnginfoapi(self, req: models.PNGInfoRequest): |
|
|
image = decode_base64_to_image(req.image.strip()) |
|
|
if image is None: |
|
|
return models.PNGInfoResponse(info="") |
|
|
|
|
|
geninfo, items = images.read_info_from_image(image) |
|
|
if geninfo is None: |
|
|
geninfo = "" |
|
|
|
|
|
params = infotext_utils.parse_generation_parameters(geninfo) |
|
|
script_callbacks.infotext_pasted_callback(geninfo, params) |
|
|
|
|
|
return models.PNGInfoResponse(info=geninfo, items=items, parameters=params) |
|
|
|
|
|
def progressapi(self, req: models.ProgressRequest = Depends()): |
|
|
|
|
|
|
|
|
if shared.state.job_count == 0: |
|
|
return models.ProgressResponse(progress=0, eta_relative=0, state=shared.state.dict(), textinfo=shared.state.textinfo) |
|
|
|
|
|
|
|
|
progress = 0.01 |
|
|
|
|
|
if shared.state.job_count > 0: |
|
|
progress += shared.state.job_no / shared.state.job_count |
|
|
if shared.state.sampling_steps > 0: |
|
|
progress += 1 / shared.state.job_count * shared.state.sampling_step / shared.state.sampling_steps |
|
|
|
|
|
time_since_start = time.time() - shared.state.time_start |
|
|
eta = time_since_start / progress |
|
|
eta_relative = eta - time_since_start |
|
|
|
|
|
progress = min(progress, 1) |
|
|
|
|
|
shared.state.set_current_image() |
|
|
|
|
|
current_image = None |
|
|
if shared.state.current_image and not req.skip_current_image: |
|
|
current_image = encode_pil_to_base64(shared.state.current_image) |
|
|
|
|
|
return models.ProgressResponse(progress=progress, eta_relative=eta_relative, state=shared.state.dict(), current_image=current_image, textinfo=shared.state.textinfo, current_task=current_task) |
|
|
|
|
|
def interruptapi(self): |
|
|
shared.state.interrupt() |
|
|
|
|
|
return {} |
|
|
|
|
|
def unloadapi(self): |
|
|
sd_models.unload_model_weights() |
|
|
|
|
|
return {} |
|
|
|
|
|
def reloadapi(self): |
|
|
sd_models.send_model_to_device(shared.sd_model) |
|
|
|
|
|
return {} |
|
|
|
|
|
def skip(self): |
|
|
shared.state.skip() |
|
|
|
|
|
def get_config(self): |
|
|
options = {} |
|
|
for key in shared.opts.data.keys(): |
|
|
metadata = shared.opts.data_labels.get(key) |
|
|
if metadata is not None: |
|
|
options.update({key: shared.opts.data.get(key, shared.opts.data_labels.get(key).default)}) |
|
|
else: |
|
|
options.update({key: shared.opts.data.get(key, None)}) |
|
|
|
|
|
return options |
|
|
|
|
|
def set_config(self, req: dict[str, Any]): |
|
|
checkpoint_name = req.get("sd_model_checkpoint", None) |
|
|
if checkpoint_name is not None and checkpoint_name not in sd_models.checkpoint_aliases: |
|
|
raise RuntimeError(f"model {checkpoint_name!r} not found") |
|
|
|
|
|
for k, v in req.items(): |
|
|
shared.opts.set(k, v, is_api=True) |
|
|
|
|
|
shared.opts.save(shared.config_filename) |
|
|
return |
|
|
|
|
|
def get_cmd_flags(self): |
|
|
return vars(shared.cmd_opts) |
|
|
|
|
|
def get_samplers(self): |
|
|
return [{"name": sampler[0], "aliases": sampler[2], "options": sampler[3]} for sampler in sd_samplers.all_samplers] |
|
|
|
|
|
def get_schedulers(self): |
|
|
return [ |
|
|
{ |
|
|
"name": scheduler.name, |
|
|
"label": scheduler.label, |
|
|
"aliases": scheduler.aliases, |
|
|
"default_rho": scheduler.default_rho, |
|
|
"need_inner_model": scheduler.need_inner_model, |
|
|
} |
|
|
for scheduler in sd_schedulers.schedulers |
|
|
] |
|
|
|
|
|
def get_upscalers(self): |
|
|
return [ |
|
|
{ |
|
|
"name": upscaler.name, |
|
|
"model_name": upscaler.scaler.model_name, |
|
|
"model_path": upscaler.data_path, |
|
|
"model_url": None, |
|
|
"scale": upscaler.scale, |
|
|
} |
|
|
for upscaler in shared.sd_upscalers |
|
|
] |
|
|
|
|
|
def get_latent_upscale_modes(self): |
|
|
return [ |
|
|
{ |
|
|
"name": upscale_mode, |
|
|
} |
|
|
for upscale_mode in [*(shared.latent_upscale_modes or {})] |
|
|
] |
|
|
|
|
|
def get_sd_models(self): |
|
|
import modules.sd_models as sd_models |
|
|
|
|
|
return [{"title": x.title, "model_name": x.model_name, "hash": x.shorthash, "sha256": x.sha256, "filename": x.filename, "config": find_checkpoint_config_near_filename(x)} for x in sd_models.checkpoints_list.values()] |
|
|
|
|
|
def get_sd_vaes(self): |
|
|
import modules.sd_vae as sd_vae |
|
|
|
|
|
return [{"model_name": x, "filename": sd_vae.vae_dict[x]} for x in sd_vae.vae_dict.keys()] |
|
|
|
|
|
def get_face_restorers(self): |
|
|
return [{"name": x.name(), "cmd_dir": getattr(x, "cmd_dir", None)} for x in shared.face_restorers] |
|
|
|
|
|
def get_prompt_styles(self): |
|
|
styleList = [] |
|
|
for k in shared.prompt_styles.styles: |
|
|
style = shared.prompt_styles.styles[k] |
|
|
styleList.append({"name": style[0], "prompt": style[1], "negative_prompt": style[2]}) |
|
|
|
|
|
return styleList |
|
|
|
|
|
def get_embeddings(self): |
|
|
db = sd_hijack.model_hijack.embedding_db |
|
|
|
|
|
def convert_embedding(embedding): |
|
|
return { |
|
|
"step": embedding.step, |
|
|
"sd_checkpoint": embedding.sd_checkpoint, |
|
|
"sd_checkpoint_name": embedding.sd_checkpoint_name, |
|
|
"shape": embedding.shape, |
|
|
"vectors": embedding.vectors, |
|
|
} |
|
|
|
|
|
def convert_embeddings(embeddings): |
|
|
return {embedding.name: convert_embedding(embedding) for embedding in embeddings.values()} |
|
|
|
|
|
return { |
|
|
"loaded": convert_embeddings(db.word_embeddings), |
|
|
"skipped": convert_embeddings(db.skipped_embeddings), |
|
|
} |
|
|
|
|
|
def refresh_embeddings(self): |
|
|
with self.queue_lock: |
|
|
sd_hijack.model_hijack.embedding_db.load_textual_inversion_embeddings(force_reload=True) |
|
|
|
|
|
def refresh_checkpoints(self): |
|
|
with self.queue_lock: |
|
|
shared.refresh_checkpoints() |
|
|
|
|
|
def refresh_vae(self): |
|
|
with self.queue_lock: |
|
|
shared_items.refresh_vae_list() |
|
|
|
|
|
def get_memory(self): |
|
|
try: |
|
|
import os |
|
|
|
|
|
import psutil |
|
|
|
|
|
process = psutil.Process(os.getpid()) |
|
|
res = process.memory_info() |
|
|
ram_total = 100 * res.rss / process.memory_percent() |
|
|
ram = {"free": ram_total - res.rss, "used": res.rss, "total": ram_total} |
|
|
except Exception as err: |
|
|
ram = {"error": f"{err}"} |
|
|
try: |
|
|
import torch |
|
|
|
|
|
if torch.cuda.is_available(): |
|
|
s = torch.cuda.mem_get_info() |
|
|
system = {"free": s[0], "used": s[1] - s[0], "total": s[1]} |
|
|
s = dict(torch.cuda.memory_stats(shared.device)) |
|
|
allocated = {"current": s["allocated_bytes.all.current"], "peak": s["allocated_bytes.all.peak"]} |
|
|
reserved = {"current": s["reserved_bytes.all.current"], "peak": s["reserved_bytes.all.peak"]} |
|
|
active = {"current": s["active_bytes.all.current"], "peak": s["active_bytes.all.peak"]} |
|
|
inactive = {"current": s["inactive_split_bytes.all.current"], "peak": s["inactive_split_bytes.all.peak"]} |
|
|
warnings = {"retries": s["num_alloc_retries"], "oom": s["num_ooms"]} |
|
|
cuda = { |
|
|
"system": system, |
|
|
"active": active, |
|
|
"allocated": allocated, |
|
|
"reserved": reserved, |
|
|
"inactive": inactive, |
|
|
"events": warnings, |
|
|
} |
|
|
else: |
|
|
cuda = {"error": "unavailable"} |
|
|
except Exception as err: |
|
|
cuda = {"error": f"{err}"} |
|
|
return models.MemoryResponse(ram=ram, cuda=cuda) |
|
|
|
|
|
def get_extensions_list(self): |
|
|
from modules import extensions |
|
|
|
|
|
extensions.list_extensions() |
|
|
ext_list = [] |
|
|
for ext in extensions.extensions: |
|
|
ext: extensions.Extension |
|
|
ext.read_info_from_repo() |
|
|
if ext.remote is not None: |
|
|
ext_list.append({"name": ext.name, "remote": ext.remote, "branch": ext.branch, "commit_hash": ext.commit_hash, "commit_date": ext.commit_date, "version": ext.version, "enabled": ext.enabled}) |
|
|
return ext_list |
|
|
|
|
|
def launch(self, server_name, port, root_path): |
|
|
self.app.include_router(self.router) |
|
|
uvicorn.run(self.app, host=server_name, port=port, timeout_keep_alive=shared.cmd_opts.timeout_keep_alive, root_path=root_path, ssl_keyfile=shared.cmd_opts.tls_keyfile, ssl_certfile=shared.cmd_opts.tls_certfile) |
|
|
|
|
|
def kill_webui(self): |
|
|
restart.stop_program() |
|
|
|
|
|
def restart_webui(self): |
|
|
if restart.is_restartable(): |
|
|
restart.restart_program() |
|
|
return Response(status_code=501) |
|
|
|
|
|
def stop_webui(request): |
|
|
shared.state.server_command = "stop" |
|
|
return Response("Stopping.") |
|
|
|