| | import json, time, os, asyncio, uuid, ssl, re, yaml, shutil
|
| | from datetime import datetime, timezone, timedelta
|
| | from typing import List, Optional, Union, Dict, Any
|
| | from pathlib import Path
|
| | import logging
|
| | from dotenv import load_dotenv
|
| |
|
| | import httpx
|
| | import aiofiles
|
| | from fastapi import FastAPI, HTTPException, Header, Request, Body, Form
|
| | from fastapi.responses import StreamingResponse, HTMLResponse, JSONResponse, RedirectResponse
|
| | from fastapi.staticfiles import StaticFiles
|
| | from pydantic import BaseModel
|
| | from util.streaming_parser import parse_json_array_stream_async
|
| | from collections import deque
|
| | from threading import Lock
|
| |
|
| |
|
| |
|
| | if os.path.exists("/data"):
|
| | DATA_DIR = "/data"
|
| | logger_prefix = "[HF-PRO]"
|
| | else:
|
| | DATA_DIR = "./data"
|
| | logger_prefix = "[LOCAL]"
|
| |
|
| |
|
| | os.makedirs(DATA_DIR, exist_ok=True)
|
| |
|
| |
|
| | ACCOUNTS_FILE = os.path.join(DATA_DIR, "accounts.json")
|
| | SETTINGS_FILE = os.path.join(DATA_DIR, "settings.yaml")
|
| | STATS_FILE = os.path.join(DATA_DIR, "stats.json")
|
| | IMAGE_DIR = os.path.join(DATA_DIR, "images")
|
| |
|
| |
|
| | os.makedirs(IMAGE_DIR, exist_ok=True)
|
| |
|
| |
|
| | from core.auth import verify_api_key
|
| | from core.session_auth import is_logged_in, login_user, logout_user, require_login, generate_session_secret
|
| |
|
| |
|
| | from core.message import (
|
| | get_conversation_key,
|
| | parse_last_message,
|
| | build_full_context_text
|
| | )
|
| | from core.google_api import (
|
| | get_common_headers,
|
| | create_google_session,
|
| | upload_context_file,
|
| | get_session_file_metadata,
|
| | download_image_with_jwt,
|
| | save_image_to_hf
|
| | )
|
| | from core.account import (
|
| | AccountManager,
|
| | MultiAccountManager,
|
| | format_account_expiration,
|
| | load_multi_account_config,
|
| | load_accounts_from_source,
|
| | update_accounts_config as _update_accounts_config,
|
| | delete_account as _delete_account,
|
| | update_account_disabled_status as _update_account_disabled_status
|
| | )
|
| |
|
| |
|
| | from core import uptime as uptime_tracker
|
| |
|
| |
|
| | from fastapi.templating import Jinja2Templates
|
| | from core.config import config_manager, config
|
| | from util.template_helpers import prepare_admin_template_data
|
| |
|
| |
|
| |
|
| |
|
| | log_buffer = deque(maxlen=3000)
|
| | log_lock = Lock()
|
| |
|
| |
|
| | stats_lock = asyncio.Lock()
|
| |
|
| | async def load_stats():
|
| | """加载统计数据(异步)"""
|
| | try:
|
| | if os.path.exists(STATS_FILE):
|
| | async with aiofiles.open(STATS_FILE, 'r', encoding='utf-8') as f:
|
| | content = await f.read()
|
| | return json.loads(content)
|
| | except Exception:
|
| | pass
|
| | return {
|
| | "total_visitors": 0,
|
| | "total_requests": 0,
|
| | "request_timestamps": [],
|
| | "visitor_ips": {},
|
| | "account_conversations": {}
|
| | }
|
| |
|
| | async def save_stats(stats):
|
| | """保存统计数据(异步,避免阻塞事件循环)"""
|
| | try:
|
| | async with aiofiles.open(STATS_FILE, 'w', encoding='utf-8') as f:
|
| | await f.write(json.dumps(stats, ensure_ascii=False, indent=2))
|
| | except Exception as e:
|
| | logger.error(f"[STATS] 保存统计数据失败: {str(e)[:50]}")
|
| |
|
| |
|
| | global_stats = {
|
| | "total_visitors": 0,
|
| | "total_requests": 0,
|
| | "request_timestamps": [],
|
| | "visitor_ips": {},
|
| | "account_conversations": {}
|
| | }
|
| |
|
| | class MemoryLogHandler(logging.Handler):
|
| | """自定义日志处理器,将日志写入内存缓冲区"""
|
| | def emit(self, record):
|
| | log_entry = self.format(record)
|
| |
|
| | beijing_tz = timezone(timedelta(hours=8))
|
| | beijing_time = datetime.fromtimestamp(record.created, tz=beijing_tz)
|
| | with log_lock:
|
| | log_buffer.append({
|
| | "time": beijing_time.strftime("%Y-%m-%d %H:%M:%S"),
|
| | "level": record.levelname,
|
| | "message": record.getMessage()
|
| | })
|
| |
|
| |
|
| | logging.basicConfig(
|
| | level=logging.INFO,
|
| | format="%(asctime)s | %(levelname)s | %(message)s",
|
| | datefmt="%H:%M:%S",
|
| | )
|
| | logger = logging.getLogger("gemini")
|
| |
|
| |
|
| | memory_handler = MemoryLogHandler()
|
| | memory_handler.setFormatter(logging.Formatter("%(asctime)s | %(levelname)s | %(message)s", datefmt="%H:%M:%S"))
|
| | logger.addHandler(memory_handler)
|
| |
|
| |
|
| |
|
| | TIMEOUT_SECONDS = 600
|
| | API_KEY = config.basic.api_key
|
| | PATH_PREFIX = config.security.path_prefix
|
| | ADMIN_KEY = config.security.admin_key
|
| | PROXY = config.basic.proxy
|
| | BASE_URL = config.basic.base_url
|
| | SESSION_SECRET_KEY = config.security.session_secret_key
|
| | SESSION_EXPIRE_HOURS = config.session.expire_hours
|
| |
|
| |
|
| | LOGO_URL = config.public_display.logo_url
|
| | CHAT_URL = config.public_display.chat_url
|
| |
|
| |
|
| | IMAGE_GENERATION_ENABLED = config.image_generation.enabled
|
| | IMAGE_GENERATION_MODELS = config.image_generation.supported_models
|
| |
|
| |
|
| | MAX_NEW_SESSION_TRIES = config.retry.max_new_session_tries
|
| | MAX_REQUEST_RETRIES = config.retry.max_request_retries
|
| | MAX_ACCOUNT_SWITCH_TRIES = config.retry.max_account_switch_tries
|
| | ACCOUNT_FAILURE_THRESHOLD = config.retry.account_failure_threshold
|
| | RATE_LIMIT_COOLDOWN_SECONDS = config.retry.rate_limit_cooldown_seconds
|
| | SESSION_CACHE_TTL_SECONDS = config.retry.session_cache_ttl_seconds
|
| |
|
| |
|
| | MODEL_MAPPING = {
|
| | "gemini-auto": None,
|
| | "gemini-2.5-flash": "gemini-2.5-flash",
|
| | "gemini-2.5-pro": "gemini-2.5-pro",
|
| | "gemini-3-flash-preview": "gemini-3-flash-preview",
|
| | "gemini-3-pro-preview": "gemini-3-pro-preview"
|
| | }
|
| |
|
| |
|
| | http_client = httpx.AsyncClient(
|
| | proxy=PROXY or None,
|
| | verify=False,
|
| | http2=False,
|
| | timeout=httpx.Timeout(TIMEOUT_SECONDS, connect=60.0),
|
| | limits=httpx.Limits(
|
| | max_keepalive_connections=100,
|
| | max_connections=200
|
| | )
|
| | )
|
| |
|
| |
|
| | def get_base_url(request: Request) -> str:
|
| | """获取完整的base URL(优先环境变量,否则从请求自动获取)"""
|
| |
|
| | if BASE_URL:
|
| | return BASE_URL.rstrip("/")
|
| |
|
| |
|
| | forwarded_proto = request.headers.get("x-forwarded-proto", request.url.scheme)
|
| | forwarded_host = request.headers.get("x-forwarded-host", request.headers.get("host"))
|
| |
|
| | return f"{forwarded_proto}://{forwarded_host}"
|
| |
|
| |
|
| | USER_AGENT = "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/140.0.0.0 Safari/537.36"
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| | multi_account_mgr = load_multi_account_config(
|
| | http_client,
|
| | USER_AGENT,
|
| | ACCOUNT_FAILURE_THRESHOLD,
|
| | RATE_LIMIT_COOLDOWN_SECONDS,
|
| | SESSION_CACHE_TTL_SECONDS,
|
| | global_stats
|
| | )
|
| |
|
| |
|
| | if not ADMIN_KEY:
|
| | logger.error("[SYSTEM] 未配置 ADMIN_KEY 环境变量,请设置后重启")
|
| | import sys
|
| | sys.exit(1)
|
| |
|
| |
|
| | if PATH_PREFIX:
|
| | logger.info(f"[SYSTEM] 路径前缀已配置: {PATH_PREFIX[:4]}****")
|
| | logger.info(f"[SYSTEM] API端点: /{PATH_PREFIX}/v1/chat/completions")
|
| | logger.info(f"[SYSTEM] 管理端点: /{PATH_PREFIX}/")
|
| | else:
|
| | logger.info("[SYSTEM] 未配置路径前缀,使用默认路径")
|
| | logger.info("[SYSTEM] API端点: /v1/chat/completions")
|
| | logger.info("[SYSTEM] 管理端点: /admin/")
|
| | logger.info("[SYSTEM] 公开端点: /public/log/html, /public/stats, /public/uptime/html")
|
| | logger.info(f"[SYSTEM] Session过期时间: {SESSION_EXPIRE_HOURS}小时")
|
| | logger.info("[SYSTEM] 系统初始化完成")
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| | app = FastAPI(title="Gemini-Business OpenAI Gateway")
|
| |
|
| |
|
| | templates = Jinja2Templates(directory="templates")
|
| |
|
| |
|
| | if os.getenv("ENV") == "development":
|
| | templates.env.auto_reload = True
|
| | logger.info("[SYSTEM] 模板热更新已启用(开发模式)")
|
| |
|
| |
|
| | app.mount("/static", StaticFiles(directory="static"), name="static")
|
| |
|
| |
|
| | from starlette.middleware.sessions import SessionMiddleware
|
| | app.add_middleware(
|
| | SessionMiddleware,
|
| | secret_key=SESSION_SECRET_KEY,
|
| | max_age=SESSION_EXPIRE_HOURS * 3600,
|
| | same_site="lax",
|
| | https_only=False
|
| | )
|
| |
|
| |
|
| | @app.middleware("http")
|
| | async def track_uptime_middleware(request: Request, call_next):
|
| | """追踪每个请求的成功/失败状态,用于 Uptime 监控"""
|
| |
|
| | path = request.url.path
|
| | if path.startswith("/images/") or path.startswith("/public/") or path.startswith("/favicon"):
|
| | return await call_next(request)
|
| |
|
| | start_time = time.time()
|
| | success = False
|
| | model = None
|
| |
|
| | try:
|
| | response = await call_next(request)
|
| | success = response.status_code < 400
|
| |
|
| |
|
| | if hasattr(request.state, "model"):
|
| | model = request.state.model
|
| |
|
| |
|
| | uptime_tracker.record_request("api_service", success)
|
| |
|
| |
|
| | if model and model in uptime_tracker.SUPPORTED_MODELS:
|
| | uptime_tracker.record_request(model, success)
|
| |
|
| | return response
|
| |
|
| | except Exception as e:
|
| |
|
| | if hasattr(request.state, "model"):
|
| | model = request.state.model
|
| |
|
| | uptime_tracker.record_request("api_service", False)
|
| | if model and model in uptime_tracker.SUPPORTED_MODELS:
|
| | uptime_tracker.record_request(model, False)
|
| | raise
|
| |
|
| |
|
| | os.makedirs(IMAGE_DIR, exist_ok=True)
|
| | app.mount("/images", StaticFiles(directory=IMAGE_DIR), name="images")
|
| | if IMAGE_DIR == "/data/images":
|
| | logger.info(f"[SYSTEM] 图片静态服务已启用: /images/ -> {IMAGE_DIR} (HF Pro持久化)")
|
| | else:
|
| | logger.info(f"[SYSTEM] 图片静态服务已启用: /images/ -> {IMAGE_DIR} (本地持久化)")
|
| |
|
| |
|
| | @app.on_event("startup")
|
| | async def startup_event():
|
| | """应用启动时初始化后台任务"""
|
| | global global_stats
|
| |
|
| |
|
| | old_accounts = "accounts.json"
|
| | if os.path.exists(old_accounts) and not os.path.exists(ACCOUNTS_FILE):
|
| | try:
|
| | shutil.copy(old_accounts, ACCOUNTS_FILE)
|
| | logger.info(f"{logger_prefix} 已迁移 {old_accounts} -> {ACCOUNTS_FILE}")
|
| | except Exception as e:
|
| | logger.warning(f"{logger_prefix} 文件迁移失败: {e}")
|
| |
|
| |
|
| | global_stats = await load_stats()
|
| | logger.info(f"[SYSTEM] 统计数据已加载: {global_stats['total_requests']} 次请求, {global_stats['total_visitors']} 位访客")
|
| |
|
| |
|
| | asyncio.create_task(multi_account_mgr.start_background_cleanup())
|
| | logger.info("[SYSTEM] 后台缓存清理任务已启动(间隔: 5分钟)")
|
| |
|
| |
|
| | asyncio.create_task(uptime_tracker.uptime_aggregation_task())
|
| | logger.info("[SYSTEM] Uptime 数据聚合任务已启动(间隔: 240秒)")
|
| |
|
| |
|
| | def get_sanitized_logs(limit: int = 100) -> list:
|
| | """获取脱敏后的日志列表,按请求ID分组并提取关键事件"""
|
| | with log_lock:
|
| | logs = list(log_buffer)
|
| |
|
| |
|
| | request_logs = {}
|
| | orphan_logs = []
|
| |
|
| | for log in logs:
|
| | message = log["message"]
|
| | req_match = re.search(r'\[req_([a-z0-9]+)\]', message)
|
| |
|
| | if req_match:
|
| | request_id = req_match.group(1)
|
| | if request_id not in request_logs:
|
| | request_logs[request_id] = []
|
| | request_logs[request_id].append(log)
|
| | else:
|
| |
|
| | orphan_logs.append(log)
|
| |
|
| |
|
| |
|
| | for orphan in orphan_logs:
|
| | orphan_time = orphan["time"]
|
| |
|
| | closest_request_id = None
|
| | min_time_diff = None
|
| |
|
| | for request_id, req_logs in request_logs.items():
|
| | if req_logs:
|
| | first_log_time = req_logs[0]["time"]
|
| |
|
| | if first_log_time >= orphan_time:
|
| | if min_time_diff is None or first_log_time < min_time_diff:
|
| | min_time_diff = first_log_time
|
| | closest_request_id = request_id
|
| |
|
| |
|
| | if closest_request_id:
|
| | request_logs[closest_request_id].insert(0, orphan)
|
| |
|
| |
|
| | sanitized = []
|
| | for request_id, req_logs in request_logs.items():
|
| |
|
| | model = None
|
| | message_count = None
|
| | retry_events = []
|
| | final_status = "in_progress"
|
| | duration = None
|
| | start_time = req_logs[0]["time"]
|
| |
|
| |
|
| | for log in req_logs:
|
| | message = log["message"]
|
| |
|
| |
|
| | if '收到请求:' in message and not model:
|
| | model_match = re.search(r'收到请求: ([^ |]+)', message)
|
| | if model_match:
|
| | model = model_match.group(1)
|
| | count_match = re.search(r'(\d+)条消息', message)
|
| | if count_match:
|
| | message_count = int(count_match.group(1))
|
| |
|
| |
|
| |
|
| | if any(keyword in message for keyword in ['切换账户', '选择账户', '失败 (尝试']):
|
| | retry_events.append({
|
| | "time": log["time"],
|
| | "message": message
|
| | })
|
| |
|
| |
|
| | if '响应完成:' in message:
|
| | time_match = re.search(r'响应完成: ([\d.]+)秒', message)
|
| | if time_match:
|
| | duration = time_match.group(1) + 's'
|
| | final_status = "success"
|
| |
|
| |
|
| | if '非流式响应完成' in message:
|
| | final_status = "success"
|
| |
|
| |
|
| | if final_status != "success" and (log['level'] == 'ERROR' or '失败' in message):
|
| | final_status = "error"
|
| |
|
| |
|
| | if final_status != "success" and '超时' in message:
|
| | final_status = "timeout"
|
| |
|
| |
|
| | if not model and final_status == "in_progress":
|
| | continue
|
| |
|
| |
|
| | events = []
|
| |
|
| |
|
| | if model:
|
| | events.append({
|
| | "time": start_time,
|
| | "type": "start",
|
| | "content": f"{model} | {message_count}条消息" if message_count else model
|
| | })
|
| | else:
|
| |
|
| | events.append({
|
| | "time": start_time,
|
| | "type": "start",
|
| | "content": "请求处理中"
|
| | })
|
| |
|
| |
|
| | failure_count = 0
|
| | account_select_count = 0
|
| |
|
| | for i, retry in enumerate(retry_events):
|
| | msg = retry["message"]
|
| |
|
| |
|
| | if '失败 (尝试' in msg:
|
| |
|
| | failure_count += 1
|
| | events.append({
|
| | "time": retry["time"],
|
| | "type": "retry",
|
| | "content": f"服务异常,正在重试({failure_count})"
|
| | })
|
| | elif '选择账户' in msg:
|
| |
|
| | account_select_count += 1
|
| |
|
| |
|
| | next_is_switch = (i + 1 < len(retry_events) and '切换账户' in retry_events[i + 1]["message"])
|
| |
|
| | if not next_is_switch:
|
| | if account_select_count == 1:
|
| |
|
| | events.append({
|
| | "time": retry["time"],
|
| | "type": "select",
|
| | "content": "选择服务节点"
|
| | })
|
| | else:
|
| |
|
| | events.append({
|
| | "time": retry["time"],
|
| | "type": "switch",
|
| | "content": "切换服务节点"
|
| | })
|
| | elif '切换账户' in msg:
|
| |
|
| | events.append({
|
| | "time": retry["time"],
|
| | "type": "switch",
|
| | "content": "切换服务节点"
|
| | })
|
| |
|
| |
|
| | if final_status == "success":
|
| | if duration:
|
| | events.append({
|
| | "time": req_logs[-1]["time"],
|
| | "type": "complete",
|
| | "status": "success",
|
| | "content": f"响应完成 | 耗时{duration}"
|
| | })
|
| | else:
|
| | events.append({
|
| | "time": req_logs[-1]["time"],
|
| | "type": "complete",
|
| | "status": "success",
|
| | "content": "响应完成"
|
| | })
|
| | elif final_status == "error":
|
| | events.append({
|
| | "time": req_logs[-1]["time"],
|
| | "type": "complete",
|
| | "status": "error",
|
| | "content": "请求失败"
|
| | })
|
| | elif final_status == "timeout":
|
| | events.append({
|
| | "time": req_logs[-1]["time"],
|
| | "type": "complete",
|
| | "status": "timeout",
|
| | "content": "请求超时"
|
| | })
|
| |
|
| | sanitized.append({
|
| | "request_id": request_id,
|
| | "start_time": start_time,
|
| | "status": final_status,
|
| | "events": events
|
| | })
|
| |
|
| |
|
| | sanitized.sort(key=lambda x: x["start_time"], reverse=True)
|
| | return sanitized[:limit]
|
| |
|
| | class Message(BaseModel):
|
| | role: str
|
| | content: Union[str, List[Dict[str, Any]]]
|
| |
|
| | class ChatRequest(BaseModel):
|
| | model: str = "gemini-auto"
|
| | messages: List[Message]
|
| | stream: bool = False
|
| | temperature: Optional[float] = 0.7
|
| | top_p: Optional[float] = 1.0
|
| |
|
| | def create_chunk(id: str, created: int, model: str, delta: dict, finish_reason: Union[str, None]) -> str:
|
| | chunk = {
|
| | "id": id,
|
| | "object": "chat.completion.chunk",
|
| | "created": created,
|
| | "model": model,
|
| | "choices": [{
|
| | "index": 0,
|
| | "delta": delta,
|
| | "logprobs": None,
|
| | "finish_reason": finish_reason
|
| | }],
|
| | "system_fingerprint": None
|
| | }
|
| | return json.dumps(chunk)
|
| |
|
| |
|
| |
|
| | def get_admin_template_data(request: Request):
|
| | """获取管理页面模板数据(避免重复代码)"""
|
| | return prepare_admin_template_data(
|
| | request, multi_account_mgr, log_buffer, log_lock,
|
| | api_key=API_KEY, base_url=BASE_URL, proxy=PROXY,
|
| | logo_url=LOGO_URL, chat_url=CHAT_URL, path_prefix=PATH_PREFIX,
|
| | max_new_session_tries=MAX_NEW_SESSION_TRIES,
|
| | max_request_retries=MAX_REQUEST_RETRIES,
|
| | max_account_switch_tries=MAX_ACCOUNT_SWITCH_TRIES,
|
| | account_failure_threshold=ACCOUNT_FAILURE_THRESHOLD,
|
| | rate_limit_cooldown_seconds=RATE_LIMIT_COOLDOWN_SECONDS,
|
| | session_cache_ttl_seconds=SESSION_CACHE_TTL_SECONDS
|
| | )
|
| |
|
| |
|
| |
|
| | @app.get("/")
|
| | async def home(request: Request):
|
| | """首页 - 根据PATH_PREFIX配置决定行为"""
|
| | if PATH_PREFIX:
|
| |
|
| | raise HTTPException(404, "Not Found")
|
| | else:
|
| |
|
| | if is_logged_in(request):
|
| | template_data = get_admin_template_data(request)
|
| | return templates.TemplateResponse("admin/index.html", template_data)
|
| | else:
|
| | return RedirectResponse(url="/login", status_code=302)
|
| |
|
| |
|
| |
|
| |
|
| | @app.get("/login")
|
| | async def admin_login_get(request: Request, error: str = None):
|
| | """登录页面"""
|
| | return templates.TemplateResponse("auth/login.html", {"request": request, "error": error})
|
| |
|
| | @app.post("/login")
|
| | async def admin_login_post(request: Request, admin_key: str = Form(...)):
|
| | """处理登录表单提交"""
|
| | if admin_key == ADMIN_KEY:
|
| | login_user(request)
|
| | logger.info(f"[AUTH] 管理员登录成功")
|
| | return RedirectResponse(url="/", status_code=302)
|
| | else:
|
| | logger.warning(f"[AUTH] 登录失败 - 密钥错误")
|
| | return templates.TemplateResponse("auth/login.html", {"request": request, "error": "密钥错误,请重试"})
|
| |
|
| | @app.post("/logout")
|
| | @require_login(redirect_to_login=False)
|
| | async def admin_logout(request: Request):
|
| | """登出"""
|
| | logout_user(request)
|
| | logger.info(f"[AUTH] 管理员已登出")
|
| | return RedirectResponse(url="/login", status_code=302)
|
| |
|
| |
|
| | if PATH_PREFIX:
|
| | @app.get(f"/{PATH_PREFIX}/login")
|
| | async def admin_login_get_prefixed(request: Request, error: str = None):
|
| | """登录页面(带前缀)"""
|
| | return templates.TemplateResponse("auth/login.html", {"request": request, "error": error})
|
| |
|
| | @app.post(f"/{PATH_PREFIX}/login")
|
| | async def admin_login_post_prefixed(request: Request, admin_key: str = Form(...)):
|
| | """处理登录表单提交(带前缀)"""
|
| | if admin_key == ADMIN_KEY:
|
| | login_user(request)
|
| | logger.info(f"[AUTH] 管理员登录成功")
|
| | return RedirectResponse(url=f"/{PATH_PREFIX}", status_code=302)
|
| | else:
|
| | logger.warning(f"[AUTH] 登录失败 - 密钥错误")
|
| | return templates.TemplateResponse("auth/login.html", {"request": request, "error": "密钥错误,请重试"})
|
| |
|
| | @app.post(f"/{PATH_PREFIX}/logout")
|
| | @require_login(redirect_to_login=False)
|
| | async def admin_logout_prefixed(request: Request):
|
| | """登出(带前缀)"""
|
| | logout_user(request)
|
| | logger.info(f"[AUTH] 管理员已登出")
|
| | return RedirectResponse(url=f"/{PATH_PREFIX}/login", status_code=302)
|
| |
|
| |
|
| |
|
| |
|
| | @app.get("/admin")
|
| | @require_login()
|
| | async def admin_home_no_prefix(request: Request):
|
| | """管理首页"""
|
| | template_data = get_admin_template_data(request)
|
| | return templates.TemplateResponse("admin/index.html", template_data)
|
| |
|
| |
|
| | if PATH_PREFIX:
|
| | @app.get(f"/{PATH_PREFIX}")
|
| | @require_login()
|
| | async def admin_home_prefixed(request: Request):
|
| | """管理首页(带前缀)"""
|
| | return await admin_home_no_prefix(request=request)
|
| |
|
| |
|
| |
|
| | @app.get("/admin/health")
|
| | @require_login()
|
| | async def admin_health(request: Request):
|
| | return {"status": "ok", "time": datetime.utcnow().isoformat()}
|
| |
|
| | @app.get("/admin/accounts")
|
| | @require_login()
|
| | async def admin_get_accounts(request: Request):
|
| | """获取所有账户的状态信息"""
|
| | accounts_info = []
|
| | for account_id, account_manager in multi_account_mgr.accounts.items():
|
| | config = account_manager.config
|
| | remaining_hours = config.get_remaining_hours()
|
| | status, status_color, remaining_display = format_account_expiration(remaining_hours)
|
| | cooldown_seconds, cooldown_reason = account_manager.get_cooldown_info()
|
| |
|
| | accounts_info.append({
|
| | "id": config.account_id,
|
| | "status": status,
|
| | "expires_at": config.expires_at or "未设置",
|
| | "remaining_hours": remaining_hours,
|
| | "remaining_display": remaining_display,
|
| | "is_available": account_manager.is_available,
|
| | "error_count": account_manager.error_count,
|
| | "disabled": config.disabled,
|
| | "cooldown_seconds": cooldown_seconds,
|
| | "cooldown_reason": cooldown_reason,
|
| | "conversation_count": account_manager.conversation_count
|
| | })
|
| |
|
| | return {"total": len(accounts_info), "accounts": accounts_info}
|
| |
|
| | @app.get("/admin/accounts-config")
|
| | @require_login()
|
| | async def admin_get_config(request: Request):
|
| | """获取完整账户配置"""
|
| | try:
|
| | accounts_data = load_accounts_from_source()
|
| | return {"accounts": accounts_data}
|
| | except Exception as e:
|
| | logger.error(f"[CONFIG] 获取配置失败: {str(e)}")
|
| | raise HTTPException(500, f"获取失败: {str(e)}")
|
| |
|
| | @app.put("/admin/accounts-config")
|
| | @require_login()
|
| | async def admin_update_config(request: Request, accounts_data: list = Body(...)):
|
| | """更新整个账户配置"""
|
| | global multi_account_mgr
|
| | try:
|
| | multi_account_mgr = _update_accounts_config(
|
| | accounts_data, multi_account_mgr, http_client, USER_AGENT,
|
| | ACCOUNT_FAILURE_THRESHOLD, RATE_LIMIT_COOLDOWN_SECONDS,
|
| | SESSION_CACHE_TTL_SECONDS, global_stats
|
| | )
|
| | return {"status": "success", "message": "配置已更新", "account_count": len(multi_account_mgr.accounts)}
|
| | except Exception as e:
|
| | logger.error(f"[CONFIG] 更新配置失败: {str(e)}")
|
| | raise HTTPException(500, f"更新失败: {str(e)}")
|
| |
|
| | @app.delete("/admin/accounts/{account_id}")
|
| | @require_login()
|
| | async def admin_delete_account(request: Request, account_id: str):
|
| | """删除单个账户"""
|
| | global multi_account_mgr
|
| | try:
|
| | multi_account_mgr = _delete_account(
|
| | account_id, multi_account_mgr, http_client, USER_AGENT,
|
| | ACCOUNT_FAILURE_THRESHOLD, RATE_LIMIT_COOLDOWN_SECONDS,
|
| | SESSION_CACHE_TTL_SECONDS, global_stats
|
| | )
|
| | return {"status": "success", "message": f"账户 {account_id} 已删除", "account_count": len(multi_account_mgr.accounts)}
|
| | except Exception as e:
|
| | logger.error(f"[CONFIG] 删除账户失败: {str(e)}")
|
| | raise HTTPException(500, f"删除失败: {str(e)}")
|
| |
|
| | @app.put("/admin/accounts/{account_id}/disable")
|
| | @require_login()
|
| | async def admin_disable_account(request: Request, account_id: str):
|
| | """手动禁用账户"""
|
| | global multi_account_mgr
|
| | try:
|
| | multi_account_mgr = _update_account_disabled_status(
|
| | account_id, True, multi_account_mgr, http_client, USER_AGENT,
|
| | ACCOUNT_FAILURE_THRESHOLD, RATE_LIMIT_COOLDOWN_SECONDS,
|
| | SESSION_CACHE_TTL_SECONDS, global_stats
|
| | )
|
| | return {"status": "success", "message": f"账户 {account_id} 已禁用", "account_count": len(multi_account_mgr.accounts)}
|
| | except Exception as e:
|
| | logger.error(f"[CONFIG] 禁用账户失败: {str(e)}")
|
| | raise HTTPException(500, f"禁用失败: {str(e)}")
|
| |
|
| | @app.put("/admin/accounts/{account_id}/enable")
|
| | @require_login()
|
| | async def admin_enable_account(request: Request, account_id: str):
|
| | """启用账户(同时重置错误禁用状态)"""
|
| | global multi_account_mgr
|
| | try:
|
| | multi_account_mgr = _update_account_disabled_status(
|
| | account_id, False, multi_account_mgr, http_client, USER_AGENT,
|
| | ACCOUNT_FAILURE_THRESHOLD, RATE_LIMIT_COOLDOWN_SECONDS,
|
| | SESSION_CACHE_TTL_SECONDS, global_stats
|
| | )
|
| |
|
| |
|
| | if account_id in multi_account_mgr.accounts:
|
| | account_mgr = multi_account_mgr.accounts[account_id]
|
| | account_mgr.is_available = True
|
| | account_mgr.error_count = 0
|
| | account_mgr.last_429_time = 0.0
|
| | logger.info(f"[CONFIG] 账户 {account_id} 错误状态已重置")
|
| |
|
| | return {"status": "success", "message": f"账户 {account_id} 已启用", "account_count": len(multi_account_mgr.accounts)}
|
| | except Exception as e:
|
| | logger.error(f"[CONFIG] 启用账户失败: {str(e)}")
|
| | raise HTTPException(500, f"启用失败: {str(e)}")
|
| |
|
| |
|
| | @app.get("/admin/settings")
|
| | @require_login()
|
| | async def admin_get_settings(request: Request):
|
| | """获取系统设置"""
|
| |
|
| | return {
|
| | "basic": {
|
| | "api_key": config.basic.api_key,
|
| | "base_url": config.basic.base_url,
|
| | "proxy": config.basic.proxy
|
| | },
|
| | "image_generation": {
|
| | "enabled": config.image_generation.enabled,
|
| | "supported_models": config.image_generation.supported_models
|
| | },
|
| | "retry": {
|
| | "max_new_session_tries": config.retry.max_new_session_tries,
|
| | "max_request_retries": config.retry.max_request_retries,
|
| | "max_account_switch_tries": config.retry.max_account_switch_tries,
|
| | "account_failure_threshold": config.retry.account_failure_threshold,
|
| | "rate_limit_cooldown_seconds": config.retry.rate_limit_cooldown_seconds,
|
| | "session_cache_ttl_seconds": config.retry.session_cache_ttl_seconds
|
| | },
|
| | "public_display": {
|
| | "logo_url": config.public_display.logo_url,
|
| | "chat_url": config.public_display.chat_url
|
| | },
|
| | "session": {
|
| | "expire_hours": config.session.expire_hours
|
| | }
|
| | }
|
| |
|
| | @app.put("/admin/settings")
|
| | @require_login()
|
| | async def admin_update_settings(request: Request, new_settings: dict = Body(...)):
|
| | """更新系统设置"""
|
| | global API_KEY, PROXY, BASE_URL, LOGO_URL, CHAT_URL
|
| | global IMAGE_GENERATION_ENABLED, IMAGE_GENERATION_MODELS
|
| | global MAX_NEW_SESSION_TRIES, MAX_REQUEST_RETRIES, MAX_ACCOUNT_SWITCH_TRIES
|
| | global ACCOUNT_FAILURE_THRESHOLD, RATE_LIMIT_COOLDOWN_SECONDS, SESSION_CACHE_TTL_SECONDS
|
| | global SESSION_EXPIRE_HOURS, multi_account_mgr, http_client
|
| |
|
| | try:
|
| |
|
| | old_proxy = PROXY
|
| | old_retry_config = {
|
| | "account_failure_threshold": ACCOUNT_FAILURE_THRESHOLD,
|
| | "rate_limit_cooldown_seconds": RATE_LIMIT_COOLDOWN_SECONDS,
|
| | "session_cache_ttl_seconds": SESSION_CACHE_TTL_SECONDS
|
| | }
|
| |
|
| |
|
| | config_manager.save_yaml(new_settings)
|
| |
|
| |
|
| | config_manager.reload()
|
| |
|
| |
|
| | API_KEY = config.basic.api_key
|
| | PROXY = config.basic.proxy
|
| | BASE_URL = config.basic.base_url
|
| | LOGO_URL = config.public_display.logo_url
|
| | CHAT_URL = config.public_display.chat_url
|
| | IMAGE_GENERATION_ENABLED = config.image_generation.enabled
|
| | IMAGE_GENERATION_MODELS = config.image_generation.supported_models
|
| | MAX_NEW_SESSION_TRIES = config.retry.max_new_session_tries
|
| | MAX_REQUEST_RETRIES = config.retry.max_request_retries
|
| | MAX_ACCOUNT_SWITCH_TRIES = config.retry.max_account_switch_tries
|
| | ACCOUNT_FAILURE_THRESHOLD = config.retry.account_failure_threshold
|
| | RATE_LIMIT_COOLDOWN_SECONDS = config.retry.rate_limit_cooldown_seconds
|
| | SESSION_CACHE_TTL_SECONDS = config.retry.session_cache_ttl_seconds
|
| | SESSION_EXPIRE_HOURS = config.session.expire_hours
|
| |
|
| |
|
| | if old_proxy != PROXY:
|
| | logger.info(f"[CONFIG] 代理配置已变化,重建 HTTP 客户端")
|
| | await http_client.aclose()
|
| | http_client = httpx.AsyncClient(
|
| | proxy=PROXY or None,
|
| | verify=False,
|
| | http2=False,
|
| | timeout=httpx.Timeout(TIMEOUT_SECONDS, connect=60.0),
|
| | limits=httpx.Limits(
|
| | max_keepalive_connections=100,
|
| | max_connections=200
|
| | )
|
| | )
|
| |
|
| | multi_account_mgr.update_http_client(http_client)
|
| |
|
| |
|
| | retry_changed = (
|
| | old_retry_config["account_failure_threshold"] != ACCOUNT_FAILURE_THRESHOLD or
|
| | old_retry_config["rate_limit_cooldown_seconds"] != RATE_LIMIT_COOLDOWN_SECONDS or
|
| | old_retry_config["session_cache_ttl_seconds"] != SESSION_CACHE_TTL_SECONDS
|
| | )
|
| |
|
| | if retry_changed:
|
| | logger.info(f"[CONFIG] 重试策略已变化,更新账户管理器配置")
|
| |
|
| | multi_account_mgr.cache_ttl = SESSION_CACHE_TTL_SECONDS
|
| | for account_id, account_mgr in multi_account_mgr.accounts.items():
|
| | account_mgr.account_failure_threshold = ACCOUNT_FAILURE_THRESHOLD
|
| | account_mgr.rate_limit_cooldown_seconds = RATE_LIMIT_COOLDOWN_SECONDS
|
| |
|
| | logger.info(f"[CONFIG] 系统设置已更新并实时生效")
|
| | return {"status": "success", "message": "设置已保存并实时生效!"}
|
| | except Exception as e:
|
| | logger.error(f"[CONFIG] 更新设置失败: {str(e)}")
|
| | raise HTTPException(500, f"更新失败: {str(e)}")
|
| |
|
| | @app.get("/admin/log")
|
| | @require_login()
|
| | async def admin_get_logs(
|
| | request: Request,
|
| | limit: int = 1500,
|
| | level: str = None,
|
| | search: str = None,
|
| | start_time: str = None,
|
| | end_time: str = None
|
| | ):
|
| | with log_lock:
|
| | logs = list(log_buffer)
|
| |
|
| | stats_by_level = {}
|
| | error_logs = []
|
| | chat_count = 0
|
| | for log in logs:
|
| | level_name = log.get("level", "INFO")
|
| | stats_by_level[level_name] = stats_by_level.get(level_name, 0) + 1
|
| | if level_name in ["ERROR", "CRITICAL"]:
|
| | error_logs.append(log)
|
| | if "收到请求" in log.get("message", ""):
|
| | chat_count += 1
|
| |
|
| | if level:
|
| | level = level.upper()
|
| | logs = [log for log in logs if log["level"] == level]
|
| | if search:
|
| | logs = [log for log in logs if search.lower() in log["message"].lower()]
|
| | if start_time:
|
| | logs = [log for log in logs if log["time"] >= start_time]
|
| | if end_time:
|
| | logs = [log for log in logs if log["time"] <= end_time]
|
| |
|
| | limit = min(limit, 3000)
|
| | filtered_logs = logs[-limit:]
|
| |
|
| | return {
|
| | "total": len(filtered_logs),
|
| | "limit": limit,
|
| | "filters": {"level": level, "search": search, "start_time": start_time, "end_time": end_time},
|
| | "logs": filtered_logs,
|
| | "stats": {
|
| | "memory": {"total": len(log_buffer), "by_level": stats_by_level, "capacity": log_buffer.maxlen},
|
| | "errors": {"count": len(error_logs), "recent": error_logs[-10:]},
|
| | "chat_count": chat_count
|
| | }
|
| | }
|
| |
|
| | @app.delete("/admin/log")
|
| | @require_login()
|
| | async def admin_clear_logs(request: Request, confirm: str = None):
|
| | if confirm != "yes":
|
| | raise HTTPException(400, "需要 confirm=yes 参数确认清空操作")
|
| | with log_lock:
|
| | cleared_count = len(log_buffer)
|
| | log_buffer.clear()
|
| | logger.info("[LOG] 日志已清空")
|
| | return {"status": "success", "message": "已清空内存日志", "cleared_count": cleared_count}
|
| |
|
| | @app.get("/admin/log/html")
|
| | @require_login()
|
| | async def admin_logs_html_route(request: Request):
|
| | """返回美化的 HTML 日志查看界面"""
|
| | return templates.TemplateResponse("admin/logs.html", {"request": request})
|
| |
|
| |
|
| | if PATH_PREFIX:
|
| | @app.get(f"/{PATH_PREFIX}/health")
|
| | @require_login()
|
| | async def admin_health_prefixed(request: Request):
|
| | return await admin_health(request=request)
|
| |
|
| | @app.get(f"/{PATH_PREFIX}/accounts")
|
| | @require_login()
|
| | async def admin_get_accounts_prefixed(request: Request):
|
| | return await admin_get_accounts(request=request)
|
| |
|
| | @app.get(f"/{PATH_PREFIX}/accounts-config")
|
| | @require_login()
|
| | async def admin_get_config_prefixed(request: Request):
|
| | return await admin_get_config(request=request)
|
| |
|
| | @app.put(f"/{PATH_PREFIX}/accounts-config")
|
| | @require_login()
|
| | async def admin_update_config_prefixed(request: Request, accounts_data: list = Body(...)):
|
| | return await admin_update_config(request=request, accounts_data=accounts_data)
|
| |
|
| | @app.delete(f"/{PATH_PREFIX}/accounts/{{account_id}}")
|
| | @require_login()
|
| | async def admin_delete_account_prefixed(request: Request, account_id: str):
|
| | return await admin_delete_account(request=request, account_id=account_id)
|
| |
|
| | @app.put(f"/{PATH_PREFIX}/accounts/{{account_id}}/disable")
|
| | @require_login()
|
| | async def admin_disable_account_prefixed(request: Request, account_id: str):
|
| | return await admin_disable_account(request=request, account_id=account_id)
|
| |
|
| | @app.put(f"/{PATH_PREFIX}/accounts/{{account_id}}/enable")
|
| | @require_login()
|
| | async def admin_enable_account_prefixed(request: Request, account_id: str):
|
| | return await admin_enable_account(request=request, account_id=account_id)
|
| |
|
| | @app.get(f"/{PATH_PREFIX}/log")
|
| | @require_login()
|
| | async def admin_get_logs_prefixed(
|
| | request: Request,
|
| | limit: int = 1500,
|
| | level: str = None,
|
| | search: str = None,
|
| | start_time: str = None,
|
| | end_time: str = None
|
| | ):
|
| | return await admin_get_logs(request=request, limit=limit, level=level, search=search, start_time=start_time, end_time=end_time)
|
| |
|
| | @app.delete(f"/{PATH_PREFIX}/log")
|
| | @require_login()
|
| | async def admin_clear_logs_prefixed(request: Request, confirm: str = None):
|
| | return await admin_clear_logs(request=request, confirm=confirm)
|
| |
|
| | @app.get(f"/{PATH_PREFIX}/log/html")
|
| | @require_login()
|
| | async def admin_logs_html_route_prefixed(request: Request):
|
| | return await admin_logs_html_route(request=request)
|
| |
|
| | @app.get(f"/{PATH_PREFIX}/settings")
|
| | @require_login()
|
| | async def admin_get_settings_prefixed(request: Request):
|
| | return await admin_get_settings(request=request)
|
| |
|
| | @app.put(f"/{PATH_PREFIX}/settings")
|
| | @require_login()
|
| | async def admin_update_settings_prefixed(request: Request, new_settings: dict = Body(...)):
|
| | return await admin_update_settings(request=request, new_settings=new_settings)
|
| |
|
| |
|
| |
|
| | @app.get("/v1/models")
|
| | async def list_models(authorization: str = Header(None)):
|
| | verify_api_key(API_KEY, authorization)
|
| | data = []
|
| | now = int(time.time())
|
| | for m in MODEL_MAPPING.keys():
|
| | data.append({"id": m, "object": "model", "created": now, "owned_by": "google", "permission": []})
|
| | return {"object": "list", "data": data}
|
| |
|
| | @app.get("/v1/models/{model_id}")
|
| | async def get_model(model_id: str, authorization: str = Header(None)):
|
| | verify_api_key(API_KEY, authorization)
|
| | return {"id": model_id, "object": "model"}
|
| |
|
| |
|
| | if PATH_PREFIX:
|
| | @app.get(f"/{PATH_PREFIX}/v1/models")
|
| | async def list_models_prefixed(authorization: str = Header(None)):
|
| | return await list_models(authorization)
|
| |
|
| | @app.get(f"/{PATH_PREFIX}/v1/models/{{model_id}}")
|
| | async def get_model_prefixed(model_id: str, authorization: str = Header(None)):
|
| | return await get_model(model_id, authorization)
|
| |
|
| |
|
| |
|
| | @app.post("/v1/chat/completions")
|
| | async def chat(
|
| | req: ChatRequest,
|
| | request: Request,
|
| | authorization: Optional[str] = Header(None)
|
| | ):
|
| |
|
| | verify_api_key(API_KEY, authorization)
|
| |
|
| | return await chat_impl(req, request, authorization)
|
| |
|
| | if PATH_PREFIX:
|
| | @app.post(f"/{PATH_PREFIX}/v1/chat/completions")
|
| | async def chat_prefixed(
|
| | req: ChatRequest,
|
| | request: Request,
|
| | authorization: Optional[str] = Header(None)
|
| | ):
|
| | return await chat(req, request, authorization)
|
| |
|
| |
|
| | async def chat_impl(
|
| | req: ChatRequest,
|
| | request: Request,
|
| | authorization: Optional[str]
|
| | ):
|
| |
|
| | request_id = str(uuid.uuid4())[:6]
|
| |
|
| |
|
| | client_ip = request.headers.get("x-forwarded-for")
|
| | if client_ip:
|
| | client_ip = client_ip.split(",")[0].strip()
|
| | else:
|
| | client_ip = request.client.host if request.client else "unknown"
|
| |
|
| |
|
| | async with stats_lock:
|
| | global_stats["total_requests"] += 1
|
| | global_stats["request_timestamps"].append(time.time())
|
| | await save_stats(global_stats)
|
| |
|
| |
|
| | if req.model not in MODEL_MAPPING:
|
| | logger.error(f"[CHAT] [req_{request_id}] 不支持的模型: {req.model}")
|
| | raise HTTPException(
|
| | status_code=404,
|
| | detail=f"Model '{req.model}' not found. Available models: {list(MODEL_MAPPING.keys())}"
|
| | )
|
| |
|
| |
|
| | request.state.model = req.model
|
| |
|
| |
|
| | conv_key = get_conversation_key([m.model_dump() for m in req.messages], client_ip)
|
| | session_lock = await multi_account_mgr.acquire_session_lock(conv_key)
|
| |
|
| |
|
| | async with session_lock:
|
| | cached_session = multi_account_mgr.global_session_cache.get(conv_key)
|
| |
|
| | if cached_session:
|
| |
|
| | account_id = cached_session["account_id"]
|
| | account_manager = await multi_account_mgr.get_account(account_id, request_id)
|
| | google_session = cached_session["session_id"]
|
| | is_new_conversation = False
|
| | logger.info(f"[CHAT] [{account_id}] [req_{request_id}] 继续会话: {google_session[-12:]}")
|
| | else:
|
| |
|
| | max_account_tries = min(MAX_NEW_SESSION_TRIES, len(multi_account_mgr.accounts))
|
| | last_error = None
|
| |
|
| | for attempt in range(max_account_tries):
|
| | try:
|
| | account_manager = await multi_account_mgr.get_account(None, request_id)
|
| | google_session = await create_google_session(account_manager, http_client, USER_AGENT, request_id)
|
| |
|
| | await multi_account_mgr.set_session_cache(
|
| | conv_key,
|
| | account_manager.config.account_id,
|
| | google_session
|
| | )
|
| | is_new_conversation = True
|
| | logger.info(f"[CHAT] [{account_manager.config.account_id}] [req_{request_id}] 新会话创建并绑定账户")
|
| |
|
| | uptime_tracker.record_request("account_pool", True)
|
| | break
|
| | except Exception as e:
|
| | last_error = e
|
| | error_type = type(e).__name__
|
| |
|
| | account_id = account_manager.config.account_id if 'account_manager' in locals() and account_manager else 'unknown'
|
| | logger.error(f"[CHAT] [req_{request_id}] 账户 {account_id} 创建会话失败 (尝试 {attempt + 1}/{max_account_tries}) - {error_type}: {str(e)}")
|
| |
|
| | uptime_tracker.record_request("account_pool", False)
|
| | if attempt == max_account_tries - 1:
|
| | logger.error(f"[CHAT] [req_{request_id}] 所有账户均不可用")
|
| | raise HTTPException(503, f"All accounts unavailable: {str(last_error)[:100]}")
|
| |
|
| |
|
| |
|
| | if req.messages:
|
| | last_content = req.messages[-1].content
|
| | if isinstance(last_content, str):
|
| |
|
| | if len(last_content) > 500:
|
| | preview = last_content[:500] + "...(已截断)"
|
| | else:
|
| | preview = last_content
|
| | else:
|
| | preview = f"[多模态: {len(last_content)}部分]"
|
| | else:
|
| | preview = "[空消息]"
|
| |
|
| |
|
| | logger.info(f"[CHAT] [{account_manager.config.account_id}] [req_{request_id}] 收到请求: {req.model} | {len(req.messages)}条消息 | stream={req.stream}")
|
| |
|
| |
|
| | logger.info(f"[CHAT] [{account_manager.config.account_id}] [req_{request_id}] 用户消息: {preview}")
|
| |
|
| |
|
| | last_text, current_images = await parse_last_message(req.messages, http_client, request_id)
|
| |
|
| |
|
| | if is_new_conversation:
|
| |
|
| | text_to_send = last_text
|
| | is_retry_mode = True
|
| | else:
|
| |
|
| | text_to_send = last_text
|
| | is_retry_mode = False
|
| |
|
| | await multi_account_mgr.update_session_time(conv_key)
|
| |
|
| | chat_id = f"chatcmpl-{uuid.uuid4()}"
|
| | created_time = int(time.time())
|
| |
|
| |
|
| | async def response_wrapper():
|
| | nonlocal account_manager
|
| |
|
| | retry_count = 0
|
| | max_retries = MAX_REQUEST_RETRIES
|
| |
|
| | current_text = text_to_send
|
| | current_retry_mode = is_retry_mode
|
| |
|
| |
|
| | current_file_ids = []
|
| |
|
| |
|
| | failed_accounts = set()
|
| |
|
| |
|
| | while retry_count <= max_retries:
|
| | try:
|
| |
|
| | cached = multi_account_mgr.global_session_cache.get(conv_key)
|
| | if not cached:
|
| | logger.warning(f"[CHAT] [{account_manager.config.account_id}] [req_{request_id}] 缓存已清理,重建Session")
|
| | new_sess = await create_google_session(account_manager, http_client, USER_AGENT, request_id)
|
| | await multi_account_mgr.set_session_cache(
|
| | conv_key,
|
| | account_manager.config.account_id,
|
| | new_sess
|
| | )
|
| | current_session = new_sess
|
| | current_retry_mode = True
|
| | current_file_ids = []
|
| | else:
|
| | current_session = cached["session_id"]
|
| |
|
| |
|
| |
|
| | if current_images and not current_file_ids:
|
| | for img in current_images:
|
| | fid = await upload_context_file(current_session, img["mime"], img["data"], account_manager, http_client, USER_AGENT, request_id)
|
| | current_file_ids.append(fid)
|
| |
|
| |
|
| | if current_retry_mode:
|
| | current_text = build_full_context_text(req.messages)
|
| |
|
| |
|
| | async for chunk in stream_chat_generator(
|
| | current_session,
|
| | current_text,
|
| | current_file_ids,
|
| | req.model,
|
| | chat_id,
|
| | created_time,
|
| | account_manager,
|
| | req.stream,
|
| | request_id,
|
| | request
|
| | ):
|
| | yield chunk
|
| |
|
| |
|
| | account_manager.is_available = True
|
| | account_manager.error_count = 0
|
| | account_manager.conversation_count += 1
|
| |
|
| |
|
| | uptime_tracker.record_request("account_pool", True)
|
| |
|
| |
|
| | async with stats_lock:
|
| | if "account_conversations" not in global_stats:
|
| | global_stats["account_conversations"] = {}
|
| | global_stats["account_conversations"][account_manager.config.account_id] = account_manager.conversation_count
|
| | await save_stats(global_stats)
|
| |
|
| | break
|
| |
|
| | except (httpx.HTTPError, ssl.SSLError, HTTPException) as e:
|
| |
|
| | failed_accounts.add(account_manager.config.account_id)
|
| |
|
| |
|
| | uptime_tracker.record_request("account_pool", False)
|
| |
|
| |
|
| | is_rate_limit = isinstance(e, HTTPException) and e.status_code == 429
|
| |
|
| |
|
| | account_manager.last_error_time = time.time()
|
| | if is_rate_limit:
|
| | account_manager.last_429_time = time.time()
|
| |
|
| | account_manager.error_count += 1
|
| | if account_manager.error_count >= ACCOUNT_FAILURE_THRESHOLD:
|
| | account_manager.is_available = False
|
| | if is_rate_limit:
|
| | logger.error(f"[ACCOUNT] [{account_manager.config.account_id}] [req_{request_id}] 遇到429错误{account_manager.error_count}次,账户已禁用(需休息{RATE_LIMIT_COOLDOWN_SECONDS}秒)")
|
| | else:
|
| | logger.error(f"[ACCOUNT] [{account_manager.config.account_id}] [req_{request_id}] 请求连续失败{account_manager.error_count}次,账户已永久禁用")
|
| |
|
| | retry_count += 1
|
| |
|
| |
|
| | error_type = type(e).__name__
|
| | error_detail = str(e)
|
| |
|
| |
|
| | if isinstance(e, HTTPException):
|
| | if is_rate_limit:
|
| | logger.error(f"[CHAT] [{account_manager.config.account_id}] [req_{request_id}] 遇到429限流错误,账户将休息{RATE_LIMIT_COOLDOWN_SECONDS}秒")
|
| | else:
|
| | logger.error(f"[CHAT] [{account_manager.config.account_id}] [req_{request_id}] HTTP错误 {e.status_code}: {e.detail}")
|
| | else:
|
| | logger.error(f"[CHAT] [{account_manager.config.account_id}] [req_{request_id}] {error_type}: {error_detail}")
|
| |
|
| |
|
| | if retry_count <= max_retries:
|
| | logger.warning(f"[CHAT] [{account_manager.config.account_id}] [req_{request_id}] 正在重试 ({retry_count}/{max_retries})")
|
| |
|
| | try:
|
| |
|
| | max_account_tries = MAX_ACCOUNT_SWITCH_TRIES
|
| | new_account = None
|
| |
|
| | for _ in range(max_account_tries):
|
| | candidate = await multi_account_mgr.get_account(None, request_id)
|
| | if candidate.config.account_id not in failed_accounts:
|
| | new_account = candidate
|
| | break
|
| |
|
| | if not new_account:
|
| | logger.error(f"[CHAT] [req_{request_id}] 所有账户均已失败,无可用账户")
|
| | if req.stream: yield f"data: {json.dumps({'error': {'message': 'All Accounts Failed'}})}\n\n"
|
| | return
|
| |
|
| | logger.info(f"[CHAT] [req_{request_id}] 切换账户: {account_manager.config.account_id} -> {new_account.config.account_id}")
|
| |
|
| |
|
| | new_sess = await create_google_session(new_account, http_client, USER_AGENT, request_id)
|
| |
|
| |
|
| | await multi_account_mgr.set_session_cache(
|
| | conv_key,
|
| | new_account.config.account_id,
|
| | new_sess
|
| | )
|
| |
|
| |
|
| | account_manager = new_account
|
| |
|
| |
|
| | current_retry_mode = True
|
| | current_file_ids = []
|
| |
|
| | except Exception as create_err:
|
| | error_type = type(create_err).__name__
|
| | logger.error(f"[CHAT] [req_{request_id}] 账户切换失败 ({error_type}): {str(create_err)}")
|
| |
|
| | uptime_tracker.record_request("account_pool", False)
|
| | if req.stream: yield f"data: {json.dumps({'error': {'message': 'Account Failover Failed'}})}\n\n"
|
| | return
|
| | else:
|
| |
|
| | logger.error(f"[CHAT] [req_{request_id}] 已达到最大重试次数 ({max_retries}),请求失败")
|
| | if req.stream: yield f"data: {json.dumps({'error': {'message': f'Max retries ({max_retries}) exceeded: {e}'}})}\n\n"
|
| | return
|
| |
|
| | if req.stream:
|
| | return StreamingResponse(response_wrapper(), media_type="text/event-stream")
|
| |
|
| | full_content = ""
|
| | full_reasoning = ""
|
| | async for chunk_str in response_wrapper():
|
| | if chunk_str.startswith("data: [DONE]"): break
|
| | if chunk_str.startswith("data: "):
|
| | try:
|
| | data = json.loads(chunk_str[6:])
|
| | delta = data["choices"][0]["delta"]
|
| | if "content" in delta:
|
| | full_content += delta["content"]
|
| | if "reasoning_content" in delta:
|
| | full_reasoning += delta["reasoning_content"]
|
| | except json.JSONDecodeError as e:
|
| | logger.error(f"[CHAT] [{account_manager.config.account_id}] [req_{request_id}] JSON解析失败: {str(e)}")
|
| | except (KeyError, IndexError) as e:
|
| | logger.error(f"[CHAT] [{account_manager.config.account_id}] [req_{request_id}] 响应格式错误 ({type(e).__name__}): {str(e)}")
|
| |
|
| |
|
| | message = {"role": "assistant", "content": full_content}
|
| | if full_reasoning:
|
| | message["reasoning_content"] = full_reasoning
|
| |
|
| |
|
| | logger.info(f"[CHAT] [{account_manager.config.account_id}] [req_{request_id}] 非流式响应完成")
|
| |
|
| |
|
| | response_preview = full_content[:500] + "...(已截断)" if len(full_content) > 500 else full_content
|
| | logger.info(f"[CHAT] [{account_manager.config.account_id}] [req_{request_id}] AI响应: {response_preview}")
|
| |
|
| | return {
|
| | "id": chat_id,
|
| | "object": "chat.completion",
|
| | "created": created_time,
|
| | "model": req.model,
|
| | "choices": [{"index": 0, "message": message, "finish_reason": "stop"}],
|
| | "usage": {"prompt_tokens": 0, "completion_tokens": 0, "total_tokens": 0}
|
| | }
|
| |
|
| |
|
| | def parse_images_from_response(data_list: list) -> tuple[list, str]:
|
| | """从API响应中解析图片文件引用
|
| | 返回: (file_ids_list, session_name)
|
| | file_ids_list: [{"fileId": str, "mimeType": str}, ...]
|
| | """
|
| | file_ids = []
|
| | session_name = ""
|
| |
|
| | for data in data_list:
|
| | sar = data.get("streamAssistResponse")
|
| | if not sar:
|
| | continue
|
| |
|
| |
|
| | session_info = sar.get("sessionInfo", {})
|
| | if session_info.get("session"):
|
| | session_name = session_info["session"]
|
| |
|
| | answer = sar.get("answer") or {}
|
| | replies = answer.get("replies") or []
|
| |
|
| | for reply in replies:
|
| | gc = reply.get("groundedContent", {})
|
| | content = gc.get("content", {})
|
| |
|
| |
|
| | file_info = content.get("file")
|
| | if file_info and file_info.get("fileId"):
|
| | file_ids.append({
|
| | "fileId": file_info["fileId"],
|
| | "mimeType": file_info.get("mimeType", "image/png")
|
| | })
|
| |
|
| | return file_ids, session_name
|
| |
|
| |
|
| | async def stream_chat_generator(session: str, text_content: str, file_ids: List[str], model_name: str, chat_id: str, created_time: int, account_manager: AccountManager, is_stream: bool = True, request_id: str = "", request: Request = None):
|
| | start_time = time.time()
|
| |
|
| |
|
| | text_preview = text_content[:500] + "...(已截断)" if len(text_content) > 500 else text_content
|
| | logger.info(f"[API] [{account_manager.config.account_id}] [req_{request_id}] 发送内容: {text_preview}")
|
| | if file_ids:
|
| | logger.info(f"[API] [{account_manager.config.account_id}] [req_{request_id}] 附带文件: {len(file_ids)}个")
|
| |
|
| | jwt = await account_manager.get_jwt(request_id)
|
| | headers = get_common_headers(jwt, USER_AGENT)
|
| |
|
| |
|
| | tools_spec = {
|
| | "webGroundingSpec": {},
|
| | "toolRegistry": "default_tool_registry",
|
| | }
|
| |
|
| | if IMAGE_GENERATION_ENABLED and model_name in IMAGE_GENERATION_MODELS:
|
| | tools_spec["imageGenerationSpec"] = {}
|
| | tools_spec["videoGenerationSpec"] = {}
|
| |
|
| | body = {
|
| | "configId": account_manager.config.config_id,
|
| | "additionalParams": {"token": "-"},
|
| | "streamAssistRequest": {
|
| | "session": session,
|
| | "query": {"parts": [{"text": text_content}]},
|
| | "filter": "",
|
| | "fileIds": file_ids,
|
| | "answerGenerationMode": "NORMAL",
|
| | "toolsSpec": tools_spec,
|
| | "languageCode": "zh-CN",
|
| | "userMetadata": {"timeZone": "Asia/Shanghai"},
|
| | "assistSkippingMode": "REQUEST_ASSIST"
|
| | }
|
| | }
|
| |
|
| | target_model_id = MODEL_MAPPING.get(model_name)
|
| | if target_model_id:
|
| | body["streamAssistRequest"]["assistGenerationConfig"] = {
|
| | "modelId": target_model_id
|
| | }
|
| |
|
| | if is_stream:
|
| | chunk = create_chunk(chat_id, created_time, model_name, {"role": "assistant"}, None)
|
| | yield f"data: {chunk}\n\n"
|
| |
|
| |
|
| | json_objects = []
|
| | file_ids_info = None
|
| |
|
| | async with http_client.stream(
|
| | "POST",
|
| | "https://biz-discoveryengine.googleapis.com/v1alpha/locations/global/widgetStreamAssist",
|
| | headers=headers,
|
| | json=body,
|
| | ) as r:
|
| | if r.status_code != 200:
|
| | error_text = await r.aread()
|
| | raise HTTPException(status_code=r.status_code, detail=f"Upstream Error {error_text.decode()}")
|
| |
|
| |
|
| | try:
|
| | async for json_obj in parse_json_array_stream_async(r.aiter_lines()):
|
| | json_objects.append(json_obj)
|
| |
|
| |
|
| | for reply in json_obj.get("streamAssistResponse", {}).get("answer", {}).get("replies", []):
|
| | content_obj = reply.get("groundedContent", {}).get("content", {})
|
| | text = content_obj.get("text", "")
|
| |
|
| | if not text:
|
| | continue
|
| |
|
| |
|
| | if content_obj.get("thought"):
|
| |
|
| | chunk = create_chunk(chat_id, created_time, model_name, {"reasoning_content": text}, None)
|
| | yield f"data: {chunk}\n\n"
|
| | else:
|
| |
|
| | chunk = create_chunk(chat_id, created_time, model_name, {"content": text}, None)
|
| | yield f"data: {chunk}\n\n"
|
| |
|
| |
|
| | if json_objects:
|
| | file_ids, session_name = parse_images_from_response(json_objects)
|
| | if file_ids and session_name:
|
| | file_ids_info = (file_ids, session_name)
|
| | logger.info(f"[IMAGE] [{account_manager.config.account_id}] [req_{request_id}] 检测到{len(file_ids)}张生成图片")
|
| |
|
| | except ValueError as e:
|
| | logger.error(f"[API] [{account_manager.config.account_id}] [req_{request_id}] JSON解析失败: {str(e)}")
|
| | except Exception as e:
|
| | error_type = type(e).__name__
|
| | logger.error(f"[API] [{account_manager.config.account_id}] [req_{request_id}] 流处理错误 ({error_type}): {str(e)}")
|
| | raise
|
| |
|
| |
|
| | if file_ids_info:
|
| | file_ids, session_name = file_ids_info
|
| | try:
|
| | base_url = get_base_url(request) if request else ""
|
| | file_metadata = await get_session_file_metadata(account_manager, session_name, http_client, USER_AGENT, request_id)
|
| |
|
| |
|
| | download_tasks = []
|
| | for file_info in file_ids:
|
| | fid = file_info["fileId"]
|
| | mime = file_info["mimeType"]
|
| | meta = file_metadata.get(fid, {})
|
| | correct_session = meta.get("session") or session_name
|
| | task = download_image_with_jwt(account_manager, correct_session, fid, http_client, USER_AGENT, request_id)
|
| | download_tasks.append((fid, mime, task))
|
| |
|
| | results = await asyncio.gather(*[task for _, _, task in download_tasks], return_exceptions=True)
|
| |
|
| |
|
| | success_count = 0
|
| | for idx, ((fid, mime, _), result) in enumerate(zip(download_tasks, results), 1):
|
| | if isinstance(result, Exception):
|
| | logger.error(f"[IMAGE] [{account_manager.config.account_id}] [req_{request_id}] 图片{idx}下载失败: {type(result).__name__}: {str(result)[:100]}")
|
| |
|
| | error_msg = f"\n\n⚠️ 图片 {idx} 下载失败\n\n"
|
| | chunk = create_chunk(chat_id, created_time, model_name, {"content": error_msg}, None)
|
| | yield f"data: {chunk}\n\n"
|
| | continue
|
| |
|
| | try:
|
| | image_url = save_image_to_hf(result, chat_id, fid, mime, base_url, IMAGE_DIR)
|
| | logger.info(f"[IMAGE] [{account_manager.config.account_id}] [req_{request_id}] 图片{idx}已保存: {image_url}")
|
| | success_count += 1
|
| |
|
| | markdown = f"\n\n\n\n"
|
| | chunk = create_chunk(chat_id, created_time, model_name, {"content": markdown}, None)
|
| | yield f"data: {chunk}\n\n"
|
| | except Exception as save_error:
|
| | logger.error(f"[IMAGE] [{account_manager.config.account_id}] [req_{request_id}] 图片{idx}保存失败: {str(save_error)[:100]}")
|
| | error_msg = f"\n\n⚠️ 图片 {idx} 保存失败\n\n"
|
| | chunk = create_chunk(chat_id, created_time, model_name, {"content": error_msg}, None)
|
| | yield f"data: {chunk}\n\n"
|
| |
|
| | logger.info(f"[IMAGE] [{account_manager.config.account_id}] [req_{request_id}] 图片处理完成: {success_count}/{len(file_ids)} 成功")
|
| |
|
| | except Exception as e:
|
| | logger.error(f"[IMAGE] [{account_manager.config.account_id}] [req_{request_id}] 图片处理失败: {type(e).__name__}: {str(e)[:100]}")
|
| |
|
| | error_msg = f"\n\n⚠️ 图片处理失败: {type(e).__name__}\n\n"
|
| | chunk = create_chunk(chat_id, created_time, model_name, {"content": error_msg}, None)
|
| | yield f"data: {chunk}\n\n"
|
| |
|
| | total_time = time.time() - start_time
|
| | logger.info(f"[API] [{account_manager.config.account_id}] [req_{request_id}] 响应完成: {total_time:.2f}秒")
|
| |
|
| | if is_stream:
|
| | final_chunk = create_chunk(chat_id, created_time, model_name, {}, "stop")
|
| | yield f"data: {final_chunk}\n\n"
|
| | yield "data: [DONE]\n\n"
|
| |
|
| |
|
| | @app.get("/public/uptime")
|
| | async def get_public_uptime(days: int = 90):
|
| | """获取 Uptime 监控数据(JSON格式)"""
|
| | if days < 1 or days > 90:
|
| | days = 90
|
| | return await uptime_tracker.get_uptime_summary(days)
|
| |
|
| | @app.get("/public/uptime/html")
|
| | async def get_public_uptime_html(request: Request):
|
| | """Uptime 监控页面(类似 status.openai.com)"""
|
| | return templates.TemplateResponse("public/uptime.html", {"request": request})
|
| |
|
| | @app.get("/public/stats")
|
| | async def get_public_stats():
|
| | """获取公开统计信息"""
|
| | async with stats_lock:
|
| |
|
| | current_time = time.time()
|
| | global_stats["request_timestamps"] = [
|
| | ts for ts in global_stats["request_timestamps"]
|
| | if current_time - ts < 3600
|
| | ]
|
| |
|
| |
|
| | recent_minute = [
|
| | ts for ts in global_stats["request_timestamps"]
|
| | if current_time - ts < 60
|
| | ]
|
| | requests_per_minute = len(recent_minute)
|
| |
|
| |
|
| | if requests_per_minute < 10:
|
| | load_status = "low"
|
| | load_color = "#10b981"
|
| | elif requests_per_minute < 30:
|
| | load_status = "medium"
|
| | load_color = "#f59e0b"
|
| | else:
|
| | load_status = "high"
|
| | load_color = "#ef4444"
|
| |
|
| | return {
|
| | "total_visitors": global_stats["total_visitors"],
|
| | "total_requests": global_stats["total_requests"],
|
| | "requests_per_minute": requests_per_minute,
|
| | "load_status": load_status,
|
| | "load_color": load_color
|
| | }
|
| |
|
| | @app.get("/public/log")
|
| | async def get_public_logs(request: Request, limit: int = 100):
|
| | """获取脱敏后的日志(JSON格式)"""
|
| | try:
|
| |
|
| |
|
| | client_ip = request.headers.get("x-forwarded-for")
|
| | if client_ip:
|
| |
|
| | client_ip = client_ip.split(",")[0].strip()
|
| | else:
|
| |
|
| | client_ip = request.client.host if request.client else "unknown"
|
| |
|
| | current_time = time.time()
|
| |
|
| | async with stats_lock:
|
| |
|
| | if "visitor_ips" not in global_stats:
|
| | global_stats["visitor_ips"] = {}
|
| |
|
| | expired_ips = [
|
| | ip for ip, timestamp in global_stats["visitor_ips"].items()
|
| | if current_time - timestamp > 86400
|
| | ]
|
| | for ip in expired_ips:
|
| | del global_stats["visitor_ips"][ip]
|
| |
|
| |
|
| | if client_ip not in global_stats["visitor_ips"]:
|
| | global_stats["visitor_ips"][client_ip] = current_time
|
| |
|
| |
|
| | global_stats["total_visitors"] = len(global_stats["visitor_ips"])
|
| | await save_stats(global_stats)
|
| |
|
| | sanitized_logs = get_sanitized_logs(limit=min(limit, 1000))
|
| | return {
|
| | "total": len(sanitized_logs),
|
| | "logs": sanitized_logs
|
| | }
|
| | except Exception as e:
|
| | logger.error(f"[LOG] 获取公开日志失败: {e}")
|
| | return {"total": 0, "logs": [], "error": str(e)}
|
| |
|
| | @app.get("/public/log/html")
|
| | async def get_public_logs_html(request: Request):
|
| | """公开的脱敏日志查看器"""
|
| | return templates.TemplateResponse("public/logs.html", {
|
| | "request": request,
|
| | "logo_url": LOGO_URL,
|
| | "chat_url": CHAT_URL
|
| | })
|
| |
|
| |
|
| |
|
| | @app.exception_handler(404)
|
| | async def not_found_handler(request: Request, exc: HTTPException):
|
| | """全局 404 处理器"""
|
| | return JSONResponse(
|
| | status_code=404,
|
| | content={"detail": "Not Found"}
|
| | )
|
| |
|
| | if __name__ == "__main__":
|
| | import uvicorn
|
| | uvicorn.run(app, host="0.0.0.0", port=7860) |