| from datetime import datetime, timedelta |
| import os |
| from pathlib import Path |
| import shutil |
| import sqlite3 |
|
|
|
|
| from scripts.iib.dir_cover_cache import get_top_4_media_info |
| from scripts.iib.tool import ( |
| get_created_date_by_stat, |
| get_video_type, |
| human_readable_size, |
| is_valid_media_path, |
| is_media_file, |
| get_cache_dir, |
| get_formatted_date, |
| is_win, |
| cwd, |
| locale, |
| enable_access_control, |
| get_windows_drives, |
| get_sd_webui_conf, |
| get_valid_img_dirs, |
| open_folder, |
| get_img_geninfo_txt_path, |
| unique_by, |
| create_zip_file, |
| normalize_paths, |
| to_abs_path, |
| is_secret_key_required, |
| open_file_with_default_app, |
| is_exe_ver, |
| backup_db_file, |
| get_current_commit_hash, |
| get_current_tag, |
| get_file_info_by_path |
| ) |
| from fastapi import FastAPI, HTTPException, Header, Response |
| from fastapi.staticfiles import StaticFiles |
| import asyncio |
| from typing import List, Optional |
| from pydantic import BaseModel |
| from fastapi.responses import FileResponse, JSONResponse, StreamingResponse |
| from PIL import Image |
| from fastapi import Depends, FastAPI, HTTPException, Request |
| from fastapi.middleware.cors import CORSMiddleware |
| import hashlib |
| from scripts.iib.db.datamodel import ( |
| DataBase, |
| ExtraPathType, |
| Image as DbImg, |
| Tag, |
| Folder, |
| ImageTag, |
| ExtraPath, |
| FileInfoDict, |
| Cursor, |
| GlobalSetting |
| ) |
| from scripts.iib.db.update_image_data import update_image_data, rebuild_image_index, add_image_data_single |
| from scripts.iib.logger import logger |
| from scripts.iib.seq import seq |
| import urllib.parse |
| from scripts.iib.fastapi_video import range_requests_response, close_video_file_reader |
| from scripts.iib.parsers.index import parse_image_info |
| import scripts.iib.plugin |
|
|
| try: |
| import pillow_avif |
| except Exception as e: |
| logger.error(e) |
|
|
| index_html_path = os.path.join(cwd, "vue/dist/index.html") |
|
|
|
|
| send_img_path = {"value": ""} |
| mem = {"secret_key_hash": None, "extra_paths": [], "all_scanned_paths": []} |
| secret_key = os.getenv("IIB_SECRET_KEY") |
| if secret_key: |
| print("Secret key loaded successfully. ") |
|
|
| WRITEABLE_PERMISSIONS = ["read-write", "write-only"] |
|
|
| is_api_writeable = not (os.getenv("IIB_ACCESS_CONTROL_PERMISSION")) or ( |
| os.getenv("IIB_ACCESS_CONTROL_PERMISSION") in WRITEABLE_PERMISSIONS |
| ) |
| IIB_DEBUG=False |
|
|
|
|
| async def write_permission_required(): |
| if not is_api_writeable: |
| error_msg = ( |
| "User is not authorized to perform this action. Required permission: " |
| + ", ".join(WRITEABLE_PERMISSIONS) |
| ) |
| raise HTTPException(status_code=403, detail=error_msg) |
|
|
|
|
| async def verify_secret(request: Request): |
| if not secret_key: |
| if is_secret_key_required: |
| raise HTTPException(status_code=400, detail={"type": "secret_key_required"}) |
| return |
| token = request.cookies.get("IIB_S") |
| if not token: |
| raise HTTPException(status_code=401, detail="Unauthorized") |
| if not mem["secret_key_hash"]: |
| mem["secret_key_hash"] = hashlib.sha256( |
| (secret_key + "_ciallo").encode("utf-8") |
| ).hexdigest() |
| if mem["secret_key_hash"] != token: |
| raise HTTPException(status_code=401, detail="Unauthorized") |
|
|
| DEFAULT_BASE = "/infinite_image_browsing" |
| def infinite_image_browsing_api(app: FastAPI, **kwargs): |
| backup_db_file(DataBase.get_db_file_path()) |
| api_base = kwargs.get("base") if isinstance(kwargs.get("base"), str) else DEFAULT_BASE |
| fe_public_path = kwargs.get("fe_public_path") if isinstance(kwargs.get("fe_public_path"), str) else api_base |
| cache_base_dir = get_cache_dir() |
|
|
| |
| if IIB_DEBUG or is_exe_ver: |
| @app.exception_handler(Exception) |
| async def exception_handler(request: Request, exc: Exception): |
| error_msg = f"An exception occurred while processing {request.method} {request.url}: {exc}" |
| logger.error(error_msg) |
|
|
| return JSONResponse( |
| status_code=500, content={"message": "Internal Server Error"} |
| ) |
| @app.middleware("http") |
| async def log_requests(request: Request, call_next): |
| path = request.url.path |
| if ( |
| path.find("infinite_image_browsing/image-thumbnail") == -1 |
| and path.find("infinite_image_browsing/file") == -1 |
| and path.find("infinite_image_browsing/fe-static") == -1 |
| ): |
| logger.info(f"Received request: {request.method} {request.url}") |
| if request.query_params: |
| logger.debug(f"Query Params: {request.query_params}") |
| if request.path_params: |
| logger.debug(f"Path Params: {request.path_params}") |
|
|
| try: |
| return await call_next(request) |
| except HTTPException as http_exc: |
| logger.warning( |
| f"HTTPException occurred while processing {request.method} {request.url}: {http_exc}" |
| ) |
| raise http_exc |
| except Exception as exc: |
| logger.error( |
| f"An exception occurred while processing {request.method} {request.url}: {exc}" |
| ) |
|
|
| |
| if kwargs.get("allow_cors"): |
| app.add_middleware( |
| CORSMiddleware, |
| allow_origin_regex="^[\w./:-]+$", |
| allow_methods=["*"], |
| allow_headers=["*"], |
| ) |
|
|
| def get_img_search_dirs(): |
| try: |
| return get_valid_img_dirs(get_sd_webui_conf(**kwargs)) |
| except Exception as e: |
| print(e) |
| return [] |
|
|
| def update_all_scanned_paths(): |
| allowed_paths = os.getenv("IIB_ACCESS_CONTROL_ALLOWED_PATHS") |
| if allowed_paths: |
| sd_webui_conf = get_sd_webui_conf(**kwargs) |
| path_config_key_map = { |
| "save": "outdir_save", |
| "extra": "outdir_extras_samples", |
| "txt2img": "outdir_txt2img_samples", |
| "img2img": "outdir_img2img_samples", |
| } |
|
|
| def path_map(path: str): |
| path = path.strip() |
| if path in path_config_key_map: |
| return sd_webui_conf.get(path_config_key_map.get(path)) |
| return path |
|
|
| paths = normalize_paths( |
| seq(allowed_paths.split(",")) |
| .map(path_map) |
| .filter(lambda x: x) |
| .to_list(), |
| os.getcwd() |
| ) |
| else: |
| paths = ( |
| get_img_search_dirs() |
| + mem["extra_paths"] |
| + kwargs.get("extra_paths_cli", []) |
| ) |
| mem["all_scanned_paths"] = unique_by(paths) |
|
|
| update_all_scanned_paths() |
|
|
| def update_extra_paths(conn: sqlite3.Connection): |
| r = ExtraPath.get_extra_paths(conn) |
| mem["extra_paths"] = [x.path for x in r] |
| update_all_scanned_paths() |
|
|
| def safe_commonpath(seq): |
| try: |
| return os.path.commonpath(seq) |
| except Exception as e: |
| |
| return "" |
|
|
| def is_path_under_parents(path, parent_paths: List[str] = []): |
| """ |
| Check if the given path is under one of the specified parent paths. |
| :param path: The path to check. |
| :param parent_paths: By default, all scanned paths are included in the list of parent paths |
| :return: True if the path is under one of the parent paths, False otherwise. |
| """ |
| try: |
| if not parent_paths: |
| parent_paths = mem["all_scanned_paths"] |
| path = to_abs_path(path) |
| for parent_path in parent_paths: |
| if safe_commonpath([path, parent_path]) == parent_path: |
| return True |
| except Exception as e: |
| logger.error(e) |
| return False |
|
|
| def is_path_trusted(path: str): |
| if not enable_access_control: |
| return True |
| try: |
| parent_paths = mem["all_scanned_paths"] |
| path = to_abs_path(path) |
| for parent_path in parent_paths: |
| if len(path) <= len(parent_path): |
| if parent_path.startswith(path): |
| return True |
| else: |
| if path.startswith(parent_path): |
| return True |
| except: |
| pass |
| return False |
|
|
| def check_path_trust(path: str): |
| if not is_path_trusted(path): |
| raise HTTPException(status_code=403) |
|
|
| def filter_allowed_files(files: List[FileInfoDict]): |
| return [x for x in files if is_path_trusted(x["fullpath"])] |
|
|
|
|
|
|
| class PathsReq(BaseModel): |
| paths: List[str] |
|
|
| @app.get(f"{api_base}/hello") |
| async def greeting(): |
| return "hello" |
|
|
| @app.get(f"{api_base}/global_setting", dependencies=[Depends(verify_secret)]) |
| async def global_setting(): |
| all_custom_tags = [] |
|
|
| extra_paths = [] |
| app_fe_setting = {} |
| try: |
| conn = DataBase.get_conn() |
| all_custom_tags = Tag.get_all_custom_tag(conn) |
| extra_paths = ExtraPath.get_extra_paths(conn) + [ |
| ExtraPath(path, ExtraPathType.cli_only.value) |
| for path in kwargs.get("extra_paths_cli", []) |
| ] |
| update_extra_paths(conn) |
| app_fe_setting = GlobalSetting.get_all_settings(conn) |
| except Exception as e: |
| print(e) |
| return { |
| "global_setting": get_sd_webui_conf(**kwargs), |
| "cwd": cwd, |
| "is_win": is_win, |
| "home": os.environ.get("USERPROFILE") if is_win else os.environ.get("HOME"), |
| "sd_cwd": os.getcwd(), |
| "all_custom_tags": all_custom_tags, |
| "extra_paths": extra_paths, |
| "enable_access_control": enable_access_control, |
| "launch_mode": kwargs.get("launch_mode", "sd"), |
| "export_fe_fn": bool(kwargs.get("export_fe_fn")), |
| "app_fe_setting": app_fe_setting, |
| "is_readonly": not is_api_writeable, |
| } |
| |
| |
| class AppFeSettingReq(BaseModel): |
| name: str |
| value: str |
| |
| @app.post(f"{api_base}/app_fe_setting", dependencies=[Depends(verify_secret), Depends(write_permission_required)]) |
| async def app_fe_setting(req: AppFeSettingReq): |
| conn = DataBase.get_conn() |
| GlobalSetting.save_setting(conn, req.name, req.value) |
|
|
| class AppFeSettingDelReq(BaseModel): |
| name: str |
|
|
| @app.delete(f"{api_base}/app_fe_setting", dependencies=[Depends(verify_secret), Depends(write_permission_required)]) |
| async def remove_app_fe_setting(req: AppFeSettingDelReq): |
| conn = DataBase.get_conn() |
| GlobalSetting.remove_setting(conn, req.name) |
| |
| @app.get(f"{api_base}/version", dependencies=[Depends(verify_secret)]) |
| async def get_version(): |
| return { |
| "hash": get_current_commit_hash(), |
| "tag": get_current_tag(), |
| } |
|
|
| class DeleteFilesReq(BaseModel): |
| file_paths: List[str] |
|
|
| @app.post( |
| api_base + "/delete_files", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| async def delete_files(req: DeleteFilesReq): |
| conn = DataBase.get_conn() |
| for path in req.file_paths: |
| check_path_trust(path) |
| try: |
| if os.path.isdir(path): |
| if len(os.listdir(path)): |
| error_msg = ( |
| "When a folder is not empty, it is not allowed to be deleted." |
| if locale == "en" |
| else "文件夹不为空时不允许删除。" |
| ) |
| raise HTTPException(400, detail=error_msg) |
| shutil.rmtree(path) |
| else: |
| close_video_file_reader(path) |
| os.remove(path) |
| txt_path = get_img_geninfo_txt_path(path) |
| if txt_path: |
| os.remove(txt_path) |
| img = DbImg.get(conn, os.path.normpath(path)) |
| if img: |
| logger.info("delete file: %s", path) |
| ImageTag.remove(conn, img.id) |
| DbImg.remove(conn, img.id) |
| except OSError as e: |
| |
| logger.error("delete failed") |
| error_msg = ( |
| f"Error deleting file {path}: {e}" |
| if locale == "en" |
| else f"删除文件 {path} 时出错:{e}" |
| ) |
| raise HTTPException(400, detail=error_msg) |
|
|
| class CreateFoldersReq(BaseModel): |
| dest_folder: str |
|
|
| @app.post( |
| api_base + "/mkdirs", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| async def create_folders(req: CreateFoldersReq): |
| if enable_access_control: |
| if not is_path_under_parents(req.dest_folder): |
| raise HTTPException(status_code=403) |
| os.makedirs(req.dest_folder, exist_ok=True) |
|
|
| class MoveFilesReq(BaseModel): |
| file_paths: List[str] |
| dest: str |
| create_dest_folder: Optional[bool] = False |
|
|
| @app.post( |
| api_base + "/copy_files", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| async def copy_files(req: MoveFilesReq): |
| for path in req.file_paths: |
| check_path_trust(path) |
| try: |
| shutil.copy(path, req.dest) |
| txt_path = get_img_geninfo_txt_path(path) |
| if txt_path: |
| shutil.copy(txt_path, req.dest) |
| except OSError as e: |
| error_msg = ( |
| f"Error copying file {path} to {req.dest}: {e}" |
| if locale == "en" |
| else f"复制文件 {path} 到 {req.dest} 时出错:{e}" |
| ) |
| raise HTTPException(400, detail=error_msg) |
|
|
| @app.post( |
| api_base + "/move_files", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| async def move_files(req: MoveFilesReq): |
| if req.create_dest_folder: |
| os.makedirs(req.dest, exist_ok=True) |
| elif not os.path.isdir(req.dest): |
| error_msg = ( |
| f"Destination folder {req.dest} does not exist." |
| if locale == "en" |
| else f"目标文件夹 {req.dest} 不存在。" |
| ) |
| raise HTTPException(400, detail=error_msg) |
|
|
| conn = DataBase.get_conn() |
|
|
| def move_file_with_geninfo(path: str, dest: str): |
| path = os.path.normpath(path) |
| txt_path = get_img_geninfo_txt_path(path) |
| if txt_path: |
| shutil.move(txt_path, dest) |
| img = DbImg.get(conn, path) |
| new_path = os.path.normpath(os.path.join(dest, os.path.basename(path))) |
| if img: |
| logger.info(f"update file path: {path} -> {new_path} in db") |
| img.update_path(conn, new_path, force=True) |
|
|
| for path in req.file_paths: |
| check_path_trust(path) |
| path = os.path.normpath(path) |
| base_dir = os.path.dirname(path) |
| try: |
| files = list(os.walk(path)) |
| is_dir = os.path.isdir(path) |
| shutil.move(path, req.dest) |
| if is_dir: |
| for root, _, files in files: |
| relative_path = root[len(base_dir) + 1 :] |
| dest = os.path.join(req.dest, relative_path) |
| for file in files: |
| is_valid = is_media_file(file) |
| if is_valid: |
| move_file_with_geninfo(os.path.join(root, file), dest) |
| else: |
| move_file_with_geninfo(path, req.dest) |
| |
| conn.commit() |
| except OSError as e: |
| |
| conn.rollback() |
| error_msg = ( |
| f"Error moving file {path} to {req.dest}: {e}" |
| if locale == "en" |
| else f"移动文件 {path} 到 {req.dest} 时出错:{e}" |
| ) |
| raise HTTPException(400, detail=error_msg) |
|
|
| @app.get(api_base + "/files", dependencies=[Depends(verify_secret)]) |
| async def get_target_folder_files(folder_path: str): |
| files: List[FileInfoDict] = [] |
| try: |
| if is_win and folder_path == "/": |
| for item in get_windows_drives(): |
| files.append( |
| {"type": "dir", "size": "-", "name": item, "fullpath": item} |
| ) |
| else: |
| if not os.path.exists(folder_path): |
| return {"files": []} |
| folder_path = to_abs_path(folder_path) |
| check_path_trust(folder_path) |
| folder_listing: List[os.DirEntry] = os.scandir(folder_path) |
| is_under_scanned_path = is_path_under_parents(folder_path) |
| for item in folder_listing: |
| if not os.path.exists(item.path): |
| continue |
| fullpath = os.path.normpath(item.path) |
| name = os.path.basename(item.path) |
| stat = item.stat() |
| date = get_formatted_date(stat.st_mtime) |
| created_time = get_created_date_by_stat(stat) |
| if item.is_file(): |
| bytes = stat.st_size |
| size = human_readable_size(bytes) |
| files.append( |
| { |
| "type": "file", |
| "date": date, |
| "size": size, |
| "name": name, |
| "bytes": bytes, |
| "created_time": created_time, |
| "fullpath": fullpath, |
| "is_under_scanned_path": is_under_scanned_path, |
| } |
| ) |
| elif item.is_dir(): |
| files.append( |
| { |
| "type": "dir", |
| "date": date, |
| "created_time": created_time, |
| "size": "-", |
| "name": name, |
| "is_under_scanned_path": is_under_scanned_path, |
| "fullpath": fullpath, |
| } |
| ) |
| except Exception as e: |
| |
| raise HTTPException(status_code=400, detail=str(e)) |
|
|
| return {"files": filter_allowed_files(files)} |
| |
|
|
| @app.post(api_base + "/batch_get_files_info", dependencies=[Depends(verify_secret)]) |
| async def batch_get_files_info(req: PathsReq): |
| res = {} |
| for path in req.paths: |
| check_path_trust(path) |
| res[path] = get_file_info_by_path(path) |
| return res |
|
|
| @app.get(api_base + "/image-thumbnail", dependencies=[Depends(verify_secret)]) |
| async def thumbnail(path: str, t: str, size: str = "256x256"): |
| check_path_trust(path) |
| if not cache_base_dir: |
| return |
| |
| hash_dir = hashlib.md5((path + t).encode("utf-8")).hexdigest() |
| hash = hash_dir + size |
| cache_dir = os.path.join(cache_base_dir, "iib_cache", hash_dir) |
| cache_path = os.path.join(cache_dir, f"{size}.webp") |
|
|
| |
| if os.path.exists(cache_path): |
| return FileResponse( |
| cache_path, |
| media_type="image/webp", |
| headers={"Cache-Control": "max-age=31536000", "ETag": hash}, |
| ) |
|
|
| |
| |
| if os.path.getsize(path) < 64 * 1024: |
| return FileResponse( |
| path, |
| media_type="image/" + path.split(".")[-1], |
| headers={"Cache-Control": "max-age=31536000", "ETag": hash}, |
| ) |
| |
|
|
| |
| with Image.open(path) as img: |
| w, h = size.split("x") |
| img.thumbnail((int(w), int(h))) |
| os.makedirs(cache_dir, exist_ok=True) |
| img.save(cache_path, "webp") |
|
|
| |
| return FileResponse( |
| cache_path, |
| media_type="image/webp", |
| headers={"Cache-Control": "max-age=31536000", "ETag": hash}, |
| ) |
|
|
| @app.get(api_base + "/file", dependencies=[Depends(verify_secret)]) |
| async def get_file(path: str, t: str, disposition: Optional[str] = None): |
| filename = path |
| import mimetypes |
|
|
| check_path_trust(path) |
| if not os.path.exists(filename): |
| raise HTTPException(status_code=404) |
| if not os.path.isfile(filename): |
| raise HTTPException(status_code=400, detail=f"{filename} is not a file") |
| |
| media_type, _ = mimetypes.guess_type(filename) |
| headers = {} |
| if disposition: |
| encoded_filename = urllib.parse.quote(disposition.encode('utf-8')) |
| headers['Content-Disposition'] = f"attachment; filename*=UTF-8''{encoded_filename}" |
|
|
| if is_path_under_parents(filename) and is_valid_media_path( |
| filename |
| ): |
| headers[ |
| "Cache-Control" |
| ] = "public, max-age=31536000" |
| headers["Expires"] = (datetime.now() + timedelta(days=365)).strftime( |
| "%a, %d %b %Y %H:%M:%S GMT" |
| ) |
|
|
| return FileResponse( |
| filename, |
| media_type=media_type, |
| headers=headers, |
| ) |
| |
| @app.get(api_base + "/stream_video", dependencies=[Depends(verify_secret)]) |
| async def stream_video(path: str, request: Request): |
| check_path_trust(path) |
| import mimetypes |
| media_type, _ = mimetypes.guess_type(path) |
| return range_requests_response( |
| request, file_path=path, content_type=media_type |
| ) |
|
|
| @app.get(api_base + "/video_cover", dependencies=[Depends(verify_secret)]) |
| async def video_cover(path: str, t: str): |
| check_path_trust(path) |
| if not cache_base_dir: |
| return |
| |
| if not os.path.exists(path): |
| raise HTTPException(status_code=404) |
| if not os.path.isfile(path) and get_video_type(path): |
| raise HTTPException(status_code=400, detail=f"{path} is not a video file") |
| |
| hash_dir = hashlib.md5((path + t).encode("utf-8")).hexdigest() |
| hash = hash_dir |
| cache_dir = os.path.join(cache_base_dir, "iib_cache", "video_cover", hash_dir) |
| cache_path = os.path.join(cache_dir, "cover.webp") |
|
|
| |
| if os.path.exists(cache_path): |
| return FileResponse( |
| cache_path, |
| media_type="image/webp", |
| headers={"Cache-Control": "max-age=31536000", "ETag": hash}, |
| ) |
| |
| |
| import imageio.v3 as iio |
| frame = iio.imread( |
| path, |
| index=16, |
| plugin="pyav", |
| ) |
| |
| os.makedirs(cache_dir, exist_ok=True) |
| iio.imwrite(cache_path,frame, extension=".webp") |
|
|
| |
| return FileResponse( |
| cache_path, |
| media_type="image/webp", |
| headers={"Cache-Control": "max-age=31536000", "ETag": hash}, |
| ) |
|
|
| @app.post(api_base + "/send_img_path", dependencies=[Depends(verify_secret)]) |
| async def api_set_send_img_path(path: str): |
| send_img_path["value"] = path |
|
|
| |
| @app.get(api_base + "/gen_info_completed", dependencies=[Depends(verify_secret)]) |
| async def api_set_send_img_path(): |
| for _ in range(30): |
| if send_img_path["value"] == "": |
| return True |
| v = send_img_path["value"] |
| |
| await asyncio.sleep(0.1) |
| return send_img_path["value"] == "" |
|
|
| @app.get(api_base + "/image_geninfo", dependencies=[Depends(verify_secret)]) |
| async def image_geninfo(path: str): |
| return parse_image_info(path).raw_info |
|
|
| class GeninfoBatchReq(BaseModel): |
| paths: List[str] |
|
|
| @app.post(api_base + "/image_geninfo_batch", dependencies=[Depends(verify_secret)]) |
| async def image_geninfo_batch(req: GeninfoBatchReq): |
| res = {} |
| conn = DataBase.get_conn() |
| for path in req.paths: |
| try: |
| img = DbImg.get(conn, path) |
| if img: |
| res[path] = img.exif |
| else: |
| res[path] = parse_image_info(path).raw_info |
| except Exception as e: |
| logger.error(e, stack_info=True) |
| return res |
|
|
|
|
| class CheckPathExistsReq(BaseModel): |
| paths: List[str] |
|
|
| @app.post(api_base + "/check_path_exists", dependencies=[Depends(verify_secret)]) |
| async def check_path_exists(req: CheckPathExistsReq): |
| update_all_scanned_paths() |
| res = {} |
| for path in req.paths: |
| res[path] = os.path.exists(path) and is_path_trusted(path) |
| return res |
|
|
| @app.get(api_base) |
| def index_bd(): |
| if fe_public_path: |
| with open(index_html_path, "r", encoding="utf-8") as file: |
| content = file.read().replace(DEFAULT_BASE, fe_public_path) |
| return Response(content=content, media_type="text/html") |
| return FileResponse(index_html_path) |
| |
| static_dir = f"{cwd}/vue/dist" |
| @app.get(api_base + "/fe-static/{file_path:path}") |
| async def serve_static_file(file_path: str): |
| file_full_path = f"{static_dir}/{file_path}" |
| if file_path.endswith(".js"): |
| with open(file_full_path, "r", encoding="utf-8") as file: |
| content = file.read().replace(DEFAULT_BASE, fe_public_path) |
| return Response(content=content, media_type="text/javascript") |
| else: |
| return FileResponse(file_full_path) |
|
|
| class OpenFolderReq(BaseModel): |
| path: str |
|
|
| @app.post( |
| api_base + "/open_folder", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| def open_folder_using_explore(req: OpenFolderReq): |
| if not is_path_trusted(req.path): |
| raise HTTPException(status_code=403) |
| open_folder(*os.path.split(req.path)) |
|
|
| @app.post(api_base + "/shutdown") |
| async def shutdown_app(): |
| |
| if not kwargs.get("enable_shutdown"): |
| raise HTTPException(status_code=403, detail="Shutdown is disabled.") |
| os.kill(os.getpid(), 9) |
| return {"message": "Application is shutting down."} |
|
|
| @app.post( |
| api_base + "/zip", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| def zip_files(req: PathsReq): |
| for path in req.paths: |
| check_path_trust(path) |
| if not os.path.isfile(path): |
| raise HTTPException(400, "The corresponding path must be a file.") |
| now = datetime.now() |
| timestamp = now.strftime("%Y-%m-%d-%H-%M-%S") |
| zip_temp_dir = os.path.join(cwd, "zip_temp") |
| os.makedirs(zip_temp_dir, exist_ok=True) |
| file_path = os.path.join(zip_temp_dir, f"iib_batch_download_{timestamp}.zip") |
| create_zip_file(req.paths, file_path) |
| return FileResponse(file_path, media_type="application/zip") |
| |
| @app.post( |
| api_base + "/open_with_default_app", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| def open_target_file_withDefault_app(req: OpenFolderReq): |
| check_path_trust(req.path) |
| open_file_with_default_app(req.path) |
|
|
|
|
| @app.post( |
| api_base + "/batch_top_4_media_info", |
| dependencies=[Depends(verify_secret)], |
| ) |
| def batch_get_top_4_media_cover_info(req: PathsReq): |
| for path in req.paths: |
| check_path_trust(path) |
| res = {} |
| for path in req.paths: |
| res[path] = get_top_4_media_info(path) |
| return res |
|
|
| db_api_base = api_base + "/db" |
|
|
| @app.get(db_api_base + "/basic_info", dependencies=[Depends(verify_secret)]) |
| async def get_db_basic_info(): |
| conn = DataBase.get_conn() |
| img_count = DbImg.count(conn) |
| tags = Tag.get_all(conn) |
| expired_dirs = Folder.get_expired_dirs(conn) |
| return { |
| "img_count": img_count, |
| "tags": tags, |
| "expired": len(expired_dirs) != 0, |
| "expired_dirs": expired_dirs, |
| } |
|
|
| @app.get(db_api_base + "/expired_dirs", dependencies=[Depends(verify_secret)]) |
| async def get_db_expired(): |
| conn = DataBase.get_conn() |
| expired_dirs = Folder.get_expired_dirs(conn) |
| return { |
| "expired": len(expired_dirs) != 0, |
| "expired_dirs": expired_dirs, |
| } |
|
|
| @app.post( |
| db_api_base + "/update_image_data", |
| dependencies=[Depends(verify_secret)], |
| ) |
| async def update_image_db_data(): |
| try: |
| DataBase._initing = True |
| conn = DataBase.get_conn() |
| img_count = DbImg.count(conn) |
| update_extra_paths(conn) |
| dirs = ( |
| get_img_search_dirs() |
| if img_count == 0 |
| else Folder.get_expired_dirs(conn) |
| ) + mem["extra_paths"] |
|
|
| update_image_data(dirs) |
| finally: |
| DataBase._initing = False |
|
|
| class SearchBySubstrReq(BaseModel): |
| surstr: str |
| cursor: str |
| regexp: str |
| folder_paths: List[str] = None |
| size: Optional[int] = 200 |
|
|
| @app.post(db_api_base + "/search_by_substr", dependencies=[Depends(verify_secret)]) |
| async def search_by_substr(req: SearchBySubstrReq): |
| if IIB_DEBUG: |
| logger.info(req) |
| conn = DataBase.get_conn() |
| folder_paths=normalize_paths(req.folder_paths, os.getcwd()) |
| if(not folder_paths and req.folder_paths): |
| return { "files": [], "cursor": Cursor(has_next=False) } |
| imgs, next_cursor = DbImg.find_by_substring( |
| conn=conn, |
| substring=req.surstr, |
| cursor=req.cursor, |
| limit=req.size, |
| regexp=req.regexp, |
| folder_paths=folder_paths |
| ) |
| return { |
| "files": filter_allowed_files([x.to_file_info() for x in imgs]), |
| "cursor": next_cursor |
| } |
| |
| class MatchImagesByTagsReq(BaseModel): |
| and_tags: List[int] |
| or_tags: List[int] |
| not_tags: List[int] |
| cursor: str |
| folder_paths: List[str] = None |
| size: Optional[int] = 200 |
|
|
| @app.post(db_api_base + "/match_images_by_tags", dependencies=[Depends(verify_secret)]) |
| async def match_image_by_tags(req: MatchImagesByTagsReq): |
| if IIB_DEBUG: |
| logger.info(req) |
| conn = DataBase.get_conn() |
| folder_paths=normalize_paths(req.folder_paths, os.getcwd()) |
| if(not folder_paths and req.folder_paths): |
| return { "files": [], "cursor": Cursor(has_next=False) } |
| imgs, next_cursor = ImageTag.get_images_by_tags( |
| conn=conn, |
| tag_dict={"and": req.and_tags, "or": req.or_tags, "not": req.not_tags}, |
| cursor=req.cursor, |
| folder_paths=folder_paths, |
| limit=req.size |
| ) |
| return { |
| "files": filter_allowed_files([x.to_file_info() for x in imgs]), |
| "cursor": next_cursor |
| } |
|
|
| @app.get(db_api_base + "/img_selected_custom_tag", dependencies=[Depends(verify_secret)]) |
| async def get_img_selected_custom_tag(path: str): |
| path = os.path.normpath(path) |
| if not is_valid_media_path(path): |
| return [] |
| conn = DataBase.get_conn() |
| update_extra_paths(conn) |
| if not is_path_under_parents(path): |
| return [] |
| img = DbImg.get(conn, path) |
| if not img: |
| if DbImg.count(conn) == 0: |
| return [] |
| update_image_data([os.path.dirname(path)]) |
| img = DbImg.get(conn, path) |
| assert img |
| |
| return ImageTag.get_tags_for_image(conn, img.id, type="custom") |
|
|
| @app.post(db_api_base + "/get_image_tags", dependencies=[Depends(verify_secret)]) |
| async def get_img_tags(req: PathsReq): |
| conn = DataBase.get_conn() |
| return ImageTag.batch_get_tags_by_path(conn, req.paths) |
|
|
| class ToggleCustomTagToImgReq(BaseModel): |
| img_path: str |
| tag_id: int |
|
|
| @app.post( |
| db_api_base + "/toggle_custom_tag_to_img", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| async def toggle_custom_tag_to_img(req: ToggleCustomTagToImgReq): |
| conn = DataBase.get_conn() |
| path = os.path.normpath(req.img_path) |
| update_extra_paths(conn) |
| if not is_path_under_parents(path): |
| raise HTTPException( |
| 400, |
| '当前文件不在搜索路径内,你可以将它添加到扫描路径再尝试。在右上角的"更多"里面' |
| if locale == "zh" |
| else 'The current file is not within the scan path. You can add it to the scan path and try again. In the top right corner, click on "More".', |
| ) |
| img = DbImg.get(conn, path) |
| if not img: |
| if DbImg.count(conn): |
| |
| add_image_data_single(path) |
| img = DbImg.get(conn, path) |
| else: |
| raise HTTPException( |
| 400, |
| "你需要先通过图像搜索页生成索引" |
| if locale == "zh" |
| else "You need to generate an index through the image search page first.", |
| ) |
| tags = ImageTag.get_tags_for_image( |
| conn=conn, image_id=img.id, type="custom", tag_id=req.tag_id |
| ) |
| is_remove = len(tags) |
| if is_remove: |
| ImageTag.remove(conn, img.id, tags[0].id) |
| else: |
| ImageTag(img.id, req.tag_id).save(conn) |
| conn.commit() |
| return {"is_remove": is_remove} |
|
|
| class BatchUpdateImageReq(BaseModel): |
| img_paths: List[str] |
| action: str |
| tag_id: int |
|
|
| @app.post( |
| db_api_base + "/batch_update_image_tag", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| async def batch_update_image_tag(req: BatchUpdateImageReq): |
| assert req.action in ["add", "remove"] |
| conn = DataBase.get_conn() |
| paths: List[str] = seq(req.img_paths).map(os.path.normpath).to_list() |
| update_extra_paths(conn) |
| for path in paths: |
| if not is_path_under_parents(path): |
| raise HTTPException( |
| 400, |
| '当前文件不在搜索路径内,你可以将它添加到扫描路径再尝试。在右上角的"更多"里面' |
| if locale == "zh" |
| else 'The current file is not within the scan path. You can add it to the scan path and try again. In the top right corner, click on "More".', |
| ) |
| img = DbImg.get(conn, path) |
| if not img: |
| if DbImg.count(conn): |
| add_image_data_single(path) |
| img = DbImg.get(conn, path) |
| else: |
| raise HTTPException( |
| 400, |
| "你需要先通过图像搜索页生成索引" |
| if locale == "zh" |
| else "You need to generate an index through the image search page first.", |
| ) |
| try: |
| for path in paths: |
| img = DbImg.get(conn, path) |
| if req.action == "add": |
| ImageTag(img.id, req.tag_id).save_or_ignore(conn) |
| else: |
| ImageTag.remove(conn, img.id, req.tag_id) |
| finally: |
| conn.commit() |
|
|
| class AddCustomTagReq(BaseModel): |
| tag_name: str |
|
|
| @app.post( |
| db_api_base + "/add_custom_tag", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| async def add_custom_tag(req: AddCustomTagReq): |
| conn = DataBase.get_conn() |
| tag = Tag.get_or_create(conn, name=req.tag_name, type="custom") |
| conn.commit() |
| return tag |
| |
| class RenameFileReq(BaseModel): |
| path: str |
| name: str |
|
|
| @app.post( |
| db_api_base + "/rename", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| async def rename_file(req: RenameFileReq): |
| conn = DataBase.get_conn() |
| try: |
| |
|
|
| path = os.path.normpath(req.path) |
| new_path = os.path.join(os.path.dirname(path), req.name) |
|
|
| |
| if not os.path.exists(path): |
| raise HTTPException(status_code=404, detail="File not found") |
|
|
| |
| if os.path.exists(new_path): |
| raise HTTPException(status_code=400, detail="A file with the new name already exists") |
| close_video_file_reader(path) |
| img = DbImg.get(conn, path) |
| if img: |
| img.update_path(conn, new_path) |
| conn.commit() |
|
|
| |
| os.rename(path, new_path) |
|
|
|
|
| return {"detail": "File renamed successfully", "new_path": new_path} |
|
|
| except PermissionError: |
| raise HTTPException(status_code=403, detail="Permission denied") |
| except Exception as e: |
| raise HTTPException(status_code=500, detail=str(e)) |
|
|
| class RemoveCustomTagReq(BaseModel): |
| tag_id: int |
|
|
| @app.post( |
| db_api_base + "/remove_custom_tag", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| async def remove_custom_tag(req: RemoveCustomTagReq): |
| conn = DataBase.get_conn() |
| ImageTag.remove(conn, tag_id=req.tag_id) |
| Tag.remove(conn, req.tag_id) |
|
|
| class RemoveCustomTagFromReq(BaseModel): |
| img_id: int |
| tag_id: str |
|
|
| @app.post( |
| db_api_base + "/remove_custom_tag_from_img", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| async def remove_custom_tag_from_img(req: RemoveCustomTagFromReq): |
| conn = DataBase.get_conn() |
| ImageTag.remove(conn, image_id=req.img_id, tag_id=req.tag_id) |
|
|
|
|
|
|
|
|
| class ExtraPathModel(BaseModel): |
| path: str |
| types: List[str] |
|
|
| @app.post( |
| f"{db_api_base}/extra_paths", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| async def create_extra_path(extra_path: ExtraPathModel): |
| if enable_access_control: |
| if not is_path_under_parents(extra_path.path): |
| raise HTTPException(status_code=403) |
| conn = DataBase.get_conn() |
| path = ExtraPath.get_target_path(conn, extra_path.path) |
| if path: |
| for t in extra_path.types: |
| path.types.append(t) |
| path.types = unique_by(path.types) |
| else: |
| path = ExtraPath(extra_path.path, extra_path.types) |
| try: |
| path.save(conn) |
| finally: |
| conn.commit() |
|
|
| class ExtraPathAliasModel(BaseModel): |
| path: str |
| alias: str |
|
|
|
|
| @app.post( |
| f"{db_api_base}/alias_extra_path", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| async def alias_extra_path(req: ExtraPathAliasModel): |
| conn = DataBase.get_conn() |
| path = ExtraPath.get_target_path(conn, req.path) |
| if not path: |
| raise HTTPException(400) |
| path.alias = req.alias |
| try: |
| path.save(conn) |
| finally: |
| conn.commit() |
| return path |
| |
|
|
| @app.get( |
| f"{db_api_base}/extra_paths", |
| dependencies=[Depends(verify_secret)], |
| ) |
| async def read_extra_paths(): |
| conn = DataBase.get_conn() |
| return ExtraPath.get_extra_paths(conn) |
| |
|
|
|
|
| @app.delete( |
| f"{db_api_base}/extra_paths", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| async def delete_extra_path(extra_path: ExtraPathModel): |
| path = to_abs_path(extra_path.path) |
| conn = DataBase.get_conn() |
| ExtraPath.remove(conn, path, extra_path.types, img_search_dirs=get_img_search_dirs()) |
|
|
| |
| @app.post( |
| f"{db_api_base}/rebuild_index", |
| dependencies=[Depends(verify_secret), Depends(write_permission_required)], |
| ) |
| async def rebuild_index(): |
| update_extra_paths(conn = DataBase.get_conn()) |
| rebuild_image_index(search_dirs = get_img_search_dirs() + mem["extra_paths"]) |
|
|
|
|