Upload 48 files
Browse files- Dockerfile +19 -19
- Dockerfile_git +14 -0
- app/api/auth.py +14 -0
- app/api/client_disconnect.py +120 -0
- app/api/dashboard.py +1 -1
- app/api/gemini_handlers.py +57 -0
- app/api/logging_utils.py +11 -0
- app/api/nonstream_handlers.py +157 -0
- app/api/request_handlers.py +165 -0
- app/api/routes.py +47 -741
- app/api/stream_handlers.py +284 -0
- app/config/settings.py +11 -4
- app/main.py +13 -8
- app/services/gemini.py +16 -1
- app/templates/index.html +44 -21
- app/utils/maintenance.py +46 -5
- app/utils/version.py +8 -9
- readme.md +11 -2
- requirements.txt +6 -6
- version.txt +1 -1
- wiki/Termux.md +84 -0
- wiki/docker.md +91 -0
- wiki/wiki.md +19 -10
Dockerfile
CHANGED
|
@@ -1,20 +1,20 @@
|
|
| 1 |
-
FROM python:3.11-slim
|
| 2 |
-
|
| 3 |
-
WORKDIR /app
|
| 4 |
-
|
| 5 |
-
# 安装 unzip 工具
|
| 6 |
-
RUN apt-get update && apt-get install -y unzip && rm -rf /var/lib/apt/lists/*
|
| 7 |
-
|
| 8 |
-
COPY app.zip .
|
| 9 |
-
COPY requirements.txt .
|
| 10 |
-
COPY version.txt .
|
| 11 |
-
RUN mkdir -p app
|
| 12 |
-
# 解压 app.zip 文件
|
| 13 |
-
RUN unzip app.zip -d app && rm app.zip
|
| 14 |
-
|
| 15 |
-
RUN pip install --no-cache-dir -r requirements.txt
|
| 16 |
-
|
| 17 |
-
# 环境变量 (在 Hugging Face Spaces 中设置)
|
| 18 |
-
# ENV GEMINI_API_KEYS=your_key_1,your_key_2,your_key_3
|
| 19 |
-
|
| 20 |
CMD ["uvicorn", "app.main:app", "--host", "0.0.0.0", "--port", "7860"]
|
|
|
|
| 1 |
+
FROM python:3.11-slim
|
| 2 |
+
|
| 3 |
+
WORKDIR /app
|
| 4 |
+
|
| 5 |
+
# 安装 unzip 工具
|
| 6 |
+
RUN apt-get update && apt-get install -y unzip && rm -rf /var/lib/apt/lists/*
|
| 7 |
+
|
| 8 |
+
COPY app.zip .
|
| 9 |
+
COPY requirements.txt .
|
| 10 |
+
COPY version.txt .
|
| 11 |
+
RUN mkdir -p app
|
| 12 |
+
# 解压 app.zip 文件
|
| 13 |
+
RUN unzip app.zip -d app && rm app.zip
|
| 14 |
+
|
| 15 |
+
RUN pip install --no-cache-dir -r requirements.txt
|
| 16 |
+
|
| 17 |
+
# 环境变量 (在 Hugging Face Spaces 中设置)
|
| 18 |
+
# ENV GEMINI_API_KEYS=your_key_1,your_key_2,your_key_3
|
| 19 |
+
|
| 20 |
CMD ["uvicorn", "app.main:app", "--host", "0.0.0.0", "--port", "7860"]
|
Dockerfile_git
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
FROM python:3.11-slim
|
| 2 |
+
|
| 3 |
+
WORKDIR /app
|
| 4 |
+
|
| 5 |
+
COPY ./app /app/app
|
| 6 |
+
COPY requirements.txt .
|
| 7 |
+
COPY version.txt .
|
| 8 |
+
|
| 9 |
+
RUN pip install --no-cache-dir -r requirements.txt
|
| 10 |
+
|
| 11 |
+
# 环境变量 (在 Hugging Face Spaces 中设置)
|
| 12 |
+
# ENV GEMINI_API_KEYS=your_key_1,your_key_2,your_key_3
|
| 13 |
+
|
| 14 |
+
CMD ["uvicorn", "app.main:app", "--host", "0.0.0.0", "--port", "7860"]
|
app/api/auth.py
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from fastapi import HTTPException, Request
|
| 2 |
+
|
| 3 |
+
# 密码验证依赖
|
| 4 |
+
async def verify_password(request: Request, PASSWORD: str = None):
|
| 5 |
+
"""验证请求中的Bearer令牌是否与配置的密码匹配"""
|
| 6 |
+
if PASSWORD:
|
| 7 |
+
auth_header = request.headers.get("Authorization")
|
| 8 |
+
if not auth_header or not auth_header.startswith("Bearer "):
|
| 9 |
+
raise HTTPException(
|
| 10 |
+
status_code=401, detail="Unauthorized: Missing or invalid token")
|
| 11 |
+
token = auth_header.split(" ")[1]
|
| 12 |
+
if token != PASSWORD:
|
| 13 |
+
raise HTTPException(
|
| 14 |
+
status_code=401, detail="Unauthorized: Invalid token")
|
app/api/client_disconnect.py
ADDED
|
@@ -0,0 +1,120 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import asyncio
|
| 2 |
+
import time
|
| 3 |
+
from fastapi import Request
|
| 4 |
+
from app.models import ChatCompletionRequest
|
| 5 |
+
from app.utils import create_error_response
|
| 6 |
+
from .logging_utils import log
|
| 7 |
+
|
| 8 |
+
# 客户端断开检测函数
|
| 9 |
+
async def check_client_disconnect(http_request: Request, current_api_key: str, request_type: str, model: str):
|
| 10 |
+
"""检查客户端是否断开连接"""
|
| 11 |
+
while True:
|
| 12 |
+
if await http_request.is_disconnected():
|
| 13 |
+
extra_log = {'key': current_api_key[:8], 'request_type': request_type, 'model': model, 'error_message': '检测到客户端断开连接'}
|
| 14 |
+
log('info', "客户端连接已中断,等待API请求完成", extra=extra_log)
|
| 15 |
+
return True
|
| 16 |
+
await asyncio.sleep(0.5)
|
| 17 |
+
|
| 18 |
+
# 客户端断开处理函数
|
| 19 |
+
async def handle_client_disconnect(
|
| 20 |
+
gemini_task: asyncio.Task,
|
| 21 |
+
chat_request: ChatCompletionRequest,
|
| 22 |
+
request_type: str,
|
| 23 |
+
current_api_key: str,
|
| 24 |
+
response_cache_manager,
|
| 25 |
+
cache_key: str = None,
|
| 26 |
+
client_ip: str = None
|
| 27 |
+
):
|
| 28 |
+
try:
|
| 29 |
+
# 等待API任务完成,使用shield防止它被取消
|
| 30 |
+
response_content = await asyncio.shield(gemini_task)
|
| 31 |
+
|
| 32 |
+
# 检查响应文本是否为空
|
| 33 |
+
if response_content is None or response_content.text == "":
|
| 34 |
+
if response_content is None:
|
| 35 |
+
log('info', "客户端断开后API任务返回None",
|
| 36 |
+
extra={'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model})
|
| 37 |
+
else:
|
| 38 |
+
extra_log = {'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model, 'status_code': 204}
|
| 39 |
+
log('info', "客户端断开后Gemini API 返回空响应", extra=extra_log)
|
| 40 |
+
|
| 41 |
+
# 删除任何现有缓存,因为响应为空
|
| 42 |
+
if cache_key and cache_key in response_cache_manager.cache:
|
| 43 |
+
log('info', f"因空响应,删除缓存: {cache_key[:8]}...",
|
| 44 |
+
extra={'cache_operation': 'remove-on-empty', 'request_type': request_type})
|
| 45 |
+
del response_cache_manager.cache[cache_key]
|
| 46 |
+
|
| 47 |
+
# 返回错误响应而不是None
|
| 48 |
+
return create_error_response(chat_request.model, "AI未返回任何内容,请重试")
|
| 49 |
+
|
| 50 |
+
# 首先检查是否有现有缓存
|
| 51 |
+
cached_response, cache_hit = response_cache_manager.get(cache_key)
|
| 52 |
+
if cache_hit:
|
| 53 |
+
log('info', f"客户端断开但找到已存在缓存,将删除: {cache_key[:8]}...",
|
| 54 |
+
extra={'cache_operation': 'disconnect-found-cache', 'request_type': request_type})
|
| 55 |
+
|
| 56 |
+
# 安全删除缓存
|
| 57 |
+
if cache_key in response_cache_manager.cache:
|
| 58 |
+
del response_cache_manager.cache[cache_key]
|
| 59 |
+
|
| 60 |
+
# 不返回缓存,而是创建新响应并缓存
|
| 61 |
+
|
| 62 |
+
# 创建新响应
|
| 63 |
+
from app.utils.response import create_response
|
| 64 |
+
response = create_response(chat_request, response_content)
|
| 65 |
+
|
| 66 |
+
# 客户端已断开,此响应不会实际发送,可以考虑将其缓存以供后续使用
|
| 67 |
+
# 如果确实需要缓存,则可以取消下面的注释
|
| 68 |
+
# cache_response(response, cache_key, client_ip)
|
| 69 |
+
|
| 70 |
+
return response
|
| 71 |
+
except asyncio.CancelledError:
|
| 72 |
+
# 对于取消异常,仍然尝试继续完成任务
|
| 73 |
+
log('info', "客户端断开后任务被取消,但我们仍会尝试完成",
|
| 74 |
+
extra={'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model})
|
| 75 |
+
|
| 76 |
+
# 检查任务是否已经完成
|
| 77 |
+
if gemini_task.done() and not gemini_task.cancelled():
|
| 78 |
+
try:
|
| 79 |
+
response_content = gemini_task.result()
|
| 80 |
+
|
| 81 |
+
# 首先检查是否有现有缓存
|
| 82 |
+
cached_response, cache_hit = response_cache_manager.get(cache_key)
|
| 83 |
+
if cache_hit:
|
| 84 |
+
log('info', f"任务被取消但找到已存在缓存,将删除: {cache_key[:8]}...",
|
| 85 |
+
extra={'cache_operation': 'cancel-found-cache', 'request_type': request_type})
|
| 86 |
+
|
| 87 |
+
# 安全删除缓存
|
| 88 |
+
if cache_key in response_cache_manager.cache:
|
| 89 |
+
del response_cache_manager.cache[cache_key]
|
| 90 |
+
|
| 91 |
+
# 创建但不缓存响应
|
| 92 |
+
from app.utils.response import create_response
|
| 93 |
+
response = create_response(chat_request, response_content)
|
| 94 |
+
return response
|
| 95 |
+
except Exception as inner_e:
|
| 96 |
+
log('error', f"客户端断开后从已完成任务获取结果失败: {str(inner_e)}",
|
| 97 |
+
extra={'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model})
|
| 98 |
+
|
| 99 |
+
# 删除缓存,因为出现错误
|
| 100 |
+
if cache_key and cache_key in response_cache_manager.cache:
|
| 101 |
+
log('info', f"因任务获取结果失败,删除缓存: {cache_key[:8]}...",
|
| 102 |
+
extra={'cache_operation': 'remove-on-error', 'request_type': request_type})
|
| 103 |
+
del response_cache_manager.cache[cache_key]
|
| 104 |
+
|
| 105 |
+
# 创建错误响应而不是返回None
|
| 106 |
+
return create_error_response(chat_request.model, "请求处理过程中发生错误,请重试")
|
| 107 |
+
except Exception as e:
|
| 108 |
+
# 处理API任务异常
|
| 109 |
+
error_msg = str(e)
|
| 110 |
+
extra_log = {'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model, 'error_message': error_msg}
|
| 111 |
+
log('error', f"客户端断开后处理API响应时出错: {error_msg}", extra=extra_log)
|
| 112 |
+
|
| 113 |
+
# 删除缓存,因为出现错误
|
| 114 |
+
if cache_key and cache_key in response_cache_manager.cache:
|
| 115 |
+
log('info', f"因API响应错误,删除缓存: {cache_key[:8]}...",
|
| 116 |
+
extra={'cache_operation': 'remove-on-error', 'request_type': request_type})
|
| 117 |
+
del response_cache_manager.cache[cache_key]
|
| 118 |
+
|
| 119 |
+
# 创建错误响应而不是返回None
|
| 120 |
+
return create_error_response(chat_request.model, f"请求处理错误: {error_msg}")
|
app/api/dashboard.py
CHANGED
|
@@ -95,7 +95,7 @@ async def get_dashboard_data():
|
|
| 95 |
api_key_stats.sort(key=lambda x: x['usage_percent'], reverse=True)
|
| 96 |
|
| 97 |
# 获取最近的日志
|
| 98 |
-
recent_logs = log_manager.get_recent_logs(
|
| 99 |
|
| 100 |
# 返回JSON格式的数据
|
| 101 |
return {
|
|
|
|
| 95 |
api_key_stats.sort(key=lambda x: x['usage_percent'], reverse=True)
|
| 96 |
|
| 97 |
# 获取最近的日志
|
| 98 |
+
recent_logs = log_manager.get_recent_logs(500) # 获取最近50条日志
|
| 99 |
|
| 100 |
# 返回JSON格式的数据
|
| 101 |
return {
|
app/api/gemini_handlers.py
ADDED
|
@@ -0,0 +1,57 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import asyncio
|
| 2 |
+
from app.models import ChatCompletionRequest
|
| 3 |
+
from app.services import GeminiClient
|
| 4 |
+
from .logging_utils import log
|
| 5 |
+
|
| 6 |
+
# Gemini完成请求函数
|
| 7 |
+
async def run_gemini_completion(
|
| 8 |
+
gemini_client,
|
| 9 |
+
chat_request: ChatCompletionRequest,
|
| 10 |
+
contents,
|
| 11 |
+
system_instruction,
|
| 12 |
+
request_type: str,
|
| 13 |
+
current_api_key: str,
|
| 14 |
+
safety_settings,
|
| 15 |
+
safety_settings_g2
|
| 16 |
+
):
|
| 17 |
+
"""运行Gemini非流式请求"""
|
| 18 |
+
# 记录函数调用状态
|
| 19 |
+
run_fn = run_gemini_completion
|
| 20 |
+
|
| 21 |
+
try:
|
| 22 |
+
# 创建一个不会被客户端断开影响的任务
|
| 23 |
+
response_future = asyncio.create_task(
|
| 24 |
+
asyncio.to_thread(
|
| 25 |
+
gemini_client.complete_chat,
|
| 26 |
+
chat_request,
|
| 27 |
+
contents,
|
| 28 |
+
safety_settings_g2 if 'gemini-2.0-flash-exp' in chat_request.model else safety_settings,
|
| 29 |
+
system_instruction
|
| 30 |
+
)
|
| 31 |
+
)
|
| 32 |
+
|
| 33 |
+
# 使用shield防止任务被外部取消
|
| 34 |
+
response_content = await asyncio.shield(response_future)
|
| 35 |
+
|
| 36 |
+
# 只在第一次调用时记录完成日志
|
| 37 |
+
if not hasattr(run_fn, 'logged_complete'):
|
| 38 |
+
log('info', "非流式请求成功完成", extra={'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model})
|
| 39 |
+
run_fn.logged_complete = True
|
| 40 |
+
return response_content
|
| 41 |
+
except asyncio.CancelledError:
|
| 42 |
+
# 即使任务被取消,我们也确保正在进行的API请求能够完成
|
| 43 |
+
if 'response_future' in locals() and not response_future.done():
|
| 44 |
+
try:
|
| 45 |
+
# 使用shield确保任务不被取消,并等待它完成
|
| 46 |
+
response_content = await asyncio.shield(response_future)
|
| 47 |
+
log('info', "API请求在客户端断开后完成", extra={'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model})
|
| 48 |
+
return response_content
|
| 49 |
+
except Exception as e:
|
| 50 |
+
extra_log_gemini_cancel = {'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model, 'error_message': f'API请求在客户端断开后失败: {str(e)}'}
|
| 51 |
+
log('info', "API调用因客户端断开而失败", extra=extra_log_gemini_cancel)
|
| 52 |
+
raise
|
| 53 |
+
|
| 54 |
+
# 如果任务尚未开始或已经失败,记录日志
|
| 55 |
+
extra_log_gemini_cancel = {'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model, 'error_message': '客户端断开导致API调用取消'}
|
| 56 |
+
log('info', "API调用因客户端断开而取消", extra=extra_log_gemini_cancel)
|
| 57 |
+
raise
|
app/api/logging_utils.py
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import logging
|
| 2 |
+
from app.utils import format_log_message
|
| 3 |
+
|
| 4 |
+
# 获取logger
|
| 5 |
+
logger = logging.getLogger("my_logger")
|
| 6 |
+
|
| 7 |
+
# 日志记录函数
|
| 8 |
+
def log(level: str, message: str, **extra):
|
| 9 |
+
"""简化日志记录的统一函数"""
|
| 10 |
+
msg = format_log_message(level.upper(), message, extra=extra)
|
| 11 |
+
getattr(logger, level.lower())(msg)
|
app/api/nonstream_handlers.py
ADDED
|
@@ -0,0 +1,157 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import asyncio
|
| 2 |
+
from fastapi import HTTPException, status, Request
|
| 3 |
+
from app.models import ChatCompletionRequest
|
| 4 |
+
from app.services import GeminiClient
|
| 5 |
+
from app.utils import cache_response, update_api_call_stats
|
| 6 |
+
from .logging_utils import log
|
| 7 |
+
from .client_disconnect import check_client_disconnect, handle_client_disconnect
|
| 8 |
+
from .gemini_handlers import run_gemini_completion
|
| 9 |
+
|
| 10 |
+
# 非流式请求处理函数
|
| 11 |
+
async def process_nonstream_request(
|
| 12 |
+
chat_request: ChatCompletionRequest,
|
| 13 |
+
http_request: Request,
|
| 14 |
+
request_type: str,
|
| 15 |
+
contents,
|
| 16 |
+
system_instruction,
|
| 17 |
+
current_api_key: str,
|
| 18 |
+
response_cache_manager,
|
| 19 |
+
active_requests_manager,
|
| 20 |
+
safety_settings,
|
| 21 |
+
safety_settings_g2,
|
| 22 |
+
api_call_stats,
|
| 23 |
+
cache_key: str = None,
|
| 24 |
+
client_ip: str = None
|
| 25 |
+
):
|
| 26 |
+
"""处理非流式API请求"""
|
| 27 |
+
gemini_client = GeminiClient(current_api_key)
|
| 28 |
+
|
| 29 |
+
# 创建任务
|
| 30 |
+
gemini_task = asyncio.create_task(
|
| 31 |
+
run_gemini_completion(
|
| 32 |
+
gemini_client,
|
| 33 |
+
chat_request,
|
| 34 |
+
contents,
|
| 35 |
+
system_instruction,
|
| 36 |
+
request_type,
|
| 37 |
+
current_api_key,
|
| 38 |
+
safety_settings,
|
| 39 |
+
safety_settings_g2
|
| 40 |
+
)
|
| 41 |
+
)
|
| 42 |
+
|
| 43 |
+
disconnect_task = asyncio.create_task(
|
| 44 |
+
check_client_disconnect(
|
| 45 |
+
http_request,
|
| 46 |
+
current_api_key,
|
| 47 |
+
request_type,
|
| 48 |
+
chat_request.model
|
| 49 |
+
)
|
| 50 |
+
)
|
| 51 |
+
|
| 52 |
+
try:
|
| 53 |
+
# 先等待看是否API任务先完成,或者客户端先断开连接
|
| 54 |
+
done, pending = await asyncio.wait(
|
| 55 |
+
[gemini_task, disconnect_task],
|
| 56 |
+
return_when=asyncio.FIRST_COMPLETED
|
| 57 |
+
)
|
| 58 |
+
|
| 59 |
+
if disconnect_task in done:
|
| 60 |
+
# 客户端已断开连接,但我们仍继续完成API请求以便缓存结果
|
| 61 |
+
return await handle_client_disconnect(
|
| 62 |
+
gemini_task,
|
| 63 |
+
chat_request,
|
| 64 |
+
request_type,
|
| 65 |
+
current_api_key,
|
| 66 |
+
response_cache_manager,
|
| 67 |
+
cache_key,
|
| 68 |
+
client_ip
|
| 69 |
+
)
|
| 70 |
+
else:
|
| 71 |
+
# API任务先完成,取消断开检测任务
|
| 72 |
+
disconnect_task.cancel()
|
| 73 |
+
|
| 74 |
+
# 获取响应内容
|
| 75 |
+
response_content = await gemini_task
|
| 76 |
+
|
| 77 |
+
# 检查缓存是否已经存在,如果存在则不再创建新缓存
|
| 78 |
+
cached_response, cache_hit = response_cache_manager.get(cache_key)
|
| 79 |
+
if cache_hit:
|
| 80 |
+
log('info', f"缓存已存在,直接返回: {cache_key[:8]}...",
|
| 81 |
+
extra={'cache_operation': 'use-existing', 'request_type': request_type})
|
| 82 |
+
|
| 83 |
+
# 安全删除缓存
|
| 84 |
+
if cache_key in response_cache_manager.cache:
|
| 85 |
+
del response_cache_manager.cache[cache_key]
|
| 86 |
+
log('info', f"缓存使用后已删除: {cache_key[:8]}...",
|
| 87 |
+
extra={'cache_operation': 'used-and-removed', 'request_type': request_type})
|
| 88 |
+
|
| 89 |
+
return cached_response
|
| 90 |
+
|
| 91 |
+
# 创建响应
|
| 92 |
+
from app.utils.response import create_response
|
| 93 |
+
response = create_response(chat_request, response_content)
|
| 94 |
+
|
| 95 |
+
# 缓存响应
|
| 96 |
+
cache_response(response, cache_key, client_ip, response_cache_manager, update_api_call_stats, api_key=current_api_key)
|
| 97 |
+
|
| 98 |
+
# 立即删除缓存,确保只能使用一次
|
| 99 |
+
if cache_key and cache_key in response_cache_manager.cache:
|
| 100 |
+
del response_cache_manager.cache[cache_key]
|
| 101 |
+
log('info', f"缓存创建后立即删除: {cache_key[:8]}...",
|
| 102 |
+
extra={'cache_operation': 'store-and-remove', 'request_type': request_type})
|
| 103 |
+
|
| 104 |
+
# 返回响应
|
| 105 |
+
return response
|
| 106 |
+
|
| 107 |
+
except asyncio.CancelledError:
|
| 108 |
+
extra_log = {'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model, 'error_message':"请求被取消"}
|
| 109 |
+
log('info', "请求取消", extra=extra_log)
|
| 110 |
+
|
| 111 |
+
# 在请求被取消时先检查缓存中是否已有结果
|
| 112 |
+
cached_response, cache_hit = response_cache_manager.get(cache_key)
|
| 113 |
+
if cache_hit:
|
| 114 |
+
log('info', f"请求取消但找到有效缓存,使用缓存响应: {cache_key[:8]}...",
|
| 115 |
+
extra={'cache_operation': 'use-cache-on-cancel', 'request_type': request_type})
|
| 116 |
+
|
| 117 |
+
# 安全删除缓存
|
| 118 |
+
if cache_key in response_cache_manager.cache:
|
| 119 |
+
del response_cache_manager.cache[cache_key]
|
| 120 |
+
log('info', f"缓存使用后已删除: {cache_key[:8]}...",
|
| 121 |
+
extra={'cache_operation': 'used-and-removed', 'request_type': request_type})
|
| 122 |
+
|
| 123 |
+
return cached_response
|
| 124 |
+
|
| 125 |
+
# 尝试完成正在进行的API请求
|
| 126 |
+
if not gemini_task.done():
|
| 127 |
+
log('info', "请求取消但API���求尚未完成,继续等待...",
|
| 128 |
+
extra={'key': current_api_key[:8], 'request_type': request_type})
|
| 129 |
+
|
| 130 |
+
# 使用shield确保任务不会被取消
|
| 131 |
+
response_content = await asyncio.shield(gemini_task)
|
| 132 |
+
|
| 133 |
+
# 创建响应
|
| 134 |
+
from app.utils.response import create_response
|
| 135 |
+
response = create_response(chat_request, response_content)
|
| 136 |
+
|
| 137 |
+
# 不缓存这个响应,直接返回
|
| 138 |
+
return response
|
| 139 |
+
else:
|
| 140 |
+
# 任务已完成,获取结果
|
| 141 |
+
response_content = gemini_task.result()
|
| 142 |
+
|
| 143 |
+
# 创建响应
|
| 144 |
+
from app.utils.response import create_response
|
| 145 |
+
response = create_response(chat_request, response_content)
|
| 146 |
+
|
| 147 |
+
# 不缓存这个响应,直接返回
|
| 148 |
+
return response
|
| 149 |
+
|
| 150 |
+
except HTTPException as e:
|
| 151 |
+
if e.status_code == status.HTTP_408_REQUEST_TIMEOUT:
|
| 152 |
+
extra_log = {'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model,
|
| 153 |
+
'status_code': 408, 'error_message': '客户端连接中断'}
|
| 154 |
+
log('error', "客户端连接中断,终止后续重试", extra=extra_log)
|
| 155 |
+
raise
|
| 156 |
+
else:
|
| 157 |
+
raise
|
app/api/request_handlers.py
ADDED
|
@@ -0,0 +1,165 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import asyncio
|
| 2 |
+
import json
|
| 3 |
+
from typing import Literal
|
| 4 |
+
from fastapi import HTTPException, Request, status
|
| 5 |
+
from fastapi.responses import StreamingResponse
|
| 6 |
+
from app.models import ChatCompletionRequest
|
| 7 |
+
from app.services import GeminiClient
|
| 8 |
+
from app.utils import protect_from_abuse, handle_gemini_error, handle_api_error
|
| 9 |
+
from .logging_utils import log
|
| 10 |
+
from .stream_handlers import process_stream_request
|
| 11 |
+
from .nonstream_handlers import process_nonstream_request
|
| 12 |
+
|
| 13 |
+
# 请求处理函数
|
| 14 |
+
async def process_request(
|
| 15 |
+
chat_request: ChatCompletionRequest,
|
| 16 |
+
http_request: Request,
|
| 17 |
+
request_type: Literal['stream', 'non-stream'],
|
| 18 |
+
key_manager,
|
| 19 |
+
response_cache_manager,
|
| 20 |
+
active_requests_manager,
|
| 21 |
+
safety_settings,
|
| 22 |
+
safety_settings_g2,
|
| 23 |
+
api_call_stats,
|
| 24 |
+
FAKE_STREAMING,
|
| 25 |
+
FAKE_STREAMING_INTERVAL,
|
| 26 |
+
MAX_REQUESTS_PER_MINUTE,
|
| 27 |
+
MAX_REQUESTS_PER_DAY_PER_IP,
|
| 28 |
+
cache_key: str = None,
|
| 29 |
+
client_ip: str = None
|
| 30 |
+
):
|
| 31 |
+
"""处理API请求的主函数,根据需要处理流式或非流式请求"""
|
| 32 |
+
global current_api_key
|
| 33 |
+
|
| 34 |
+
# 请求前基本检查
|
| 35 |
+
protect_from_abuse(
|
| 36 |
+
http_request, MAX_REQUESTS_PER_MINUTE, MAX_REQUESTS_PER_DAY_PER_IP)
|
| 37 |
+
if chat_request.model not in GeminiClient.AVAILABLE_MODELS:
|
| 38 |
+
error_msg = "无效的模型"
|
| 39 |
+
extra_log = {'request_type': request_type, 'model': chat_request.model, 'status_code': 400, 'error_message': error_msg}
|
| 40 |
+
log('error', error_msg, extra=extra_log)
|
| 41 |
+
raise HTTPException(
|
| 42 |
+
status_code=status.HTTP_400_BAD_REQUEST, detail=error_msg)
|
| 43 |
+
|
| 44 |
+
# 重置已尝试的密钥
|
| 45 |
+
key_manager.reset_tried_keys_for_request()
|
| 46 |
+
|
| 47 |
+
# 转换消息格式
|
| 48 |
+
contents, system_instruction = GeminiClient.convert_messages(
|
| 49 |
+
GeminiClient, chat_request.messages)
|
| 50 |
+
|
| 51 |
+
# 设置重试次数(使用可用API密钥数量作为最大重试次数)
|
| 52 |
+
retry_attempts = len(key_manager.api_keys) if key_manager.api_keys else 1
|
| 53 |
+
|
| 54 |
+
# 尝试使用不同API密钥
|
| 55 |
+
for attempt in range(1, retry_attempts + 1):
|
| 56 |
+
# 获取下一个密钥
|
| 57 |
+
current_api_key = key_manager.get_available_key()
|
| 58 |
+
|
| 59 |
+
# 检查API密钥是否可用
|
| 60 |
+
if current_api_key is None:
|
| 61 |
+
log('warning', "没有可用的 API 密钥,跳过本次尝试",
|
| 62 |
+
extra={'request_type': request_type, 'model': chat_request.model, 'status_code': 'N/A'})
|
| 63 |
+
break
|
| 64 |
+
|
| 65 |
+
# 记录当前尝试的密钥信息
|
| 66 |
+
log('info', f"第 {attempt}/{retry_attempts} 次尝试 ... 使用密钥: {current_api_key[:8]}...",
|
| 67 |
+
extra={'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model})
|
| 68 |
+
|
| 69 |
+
# 服务器错误重试逻辑
|
| 70 |
+
server_error_retries = 3
|
| 71 |
+
for server_retry in range(1, server_error_retries + 1):
|
| 72 |
+
try:
|
| 73 |
+
# 根据请求类型分别处理
|
| 74 |
+
if chat_request.stream:
|
| 75 |
+
try:
|
| 76 |
+
return await process_stream_request(
|
| 77 |
+
chat_request,
|
| 78 |
+
http_request,
|
| 79 |
+
contents,
|
| 80 |
+
system_instruction,
|
| 81 |
+
current_api_key,
|
| 82 |
+
key_manager,
|
| 83 |
+
safety_settings,
|
| 84 |
+
safety_settings_g2,
|
| 85 |
+
api_call_stats,
|
| 86 |
+
FAKE_STREAMING,
|
| 87 |
+
FAKE_STREAMING_INTERVAL
|
| 88 |
+
)
|
| 89 |
+
except Exception as e:
|
| 90 |
+
# 捕获流式请求的异常,但不立即返回错误
|
| 91 |
+
# 记录错误并继续尝试下一个API密钥
|
| 92 |
+
error_detail = handle_gemini_error(e, current_api_key, key_manager)
|
| 93 |
+
log('error', f"流式请求失败: {error_detail}",
|
| 94 |
+
extra={'key': current_api_key[:8], 'request_type': 'stream', 'model': chat_request.model})
|
| 95 |
+
# 不返回错误,而是抛出异常让外层循环处理
|
| 96 |
+
raise
|
| 97 |
+
else:
|
| 98 |
+
return await process_nonstream_request(
|
| 99 |
+
chat_request,
|
| 100 |
+
http_request,
|
| 101 |
+
request_type,
|
| 102 |
+
contents,
|
| 103 |
+
system_instruction,
|
| 104 |
+
current_api_key,
|
| 105 |
+
response_cache_manager,
|
| 106 |
+
active_requests_manager,
|
| 107 |
+
safety_settings,
|
| 108 |
+
safety_settings_g2,
|
| 109 |
+
api_call_stats,
|
| 110 |
+
cache_key,
|
| 111 |
+
client_ip
|
| 112 |
+
)
|
| 113 |
+
except HTTPException as e:
|
| 114 |
+
if e.status_code == status.HTTP_408_REQUEST_TIMEOUT:
|
| 115 |
+
log('error', "客户端连接中断",
|
| 116 |
+
extra={'key': current_api_key[:8], 'request_type': request_type,
|
| 117 |
+
'model': chat_request.model, 'status_code': 408})
|
| 118 |
+
raise
|
| 119 |
+
else:
|
| 120 |
+
raise
|
| 121 |
+
except Exception as e:
|
| 122 |
+
# 使用统一的API错误处理函数
|
| 123 |
+
error_result = await handle_api_error(
|
| 124 |
+
e,
|
| 125 |
+
current_api_key,
|
| 126 |
+
key_manager,
|
| 127 |
+
request_type,
|
| 128 |
+
chat_request.model,
|
| 129 |
+
server_retry - 1
|
| 130 |
+
)
|
| 131 |
+
|
| 132 |
+
# 如果需要删除缓存,清除缓存
|
| 133 |
+
if error_result.get('remove_cache', False) and cache_key and cache_key in response_cache_manager.cache:
|
| 134 |
+
log('info', f"因API错误,删除缓存: {cache_key[:8]}...",
|
| 135 |
+
extra={'cache_operation': 'remove-on-error', 'request_type': request_type})
|
| 136 |
+
del response_cache_manager.cache[cache_key]
|
| 137 |
+
|
| 138 |
+
if error_result.get('should_retry', False):
|
| 139 |
+
# 服务器错误需要重试(等待已在handle_api_error中完成)
|
| 140 |
+
continue
|
| 141 |
+
elif error_result.get('should_switch_key', False) and attempt < retry_attempts:
|
| 142 |
+
# 跳出服务器错误重试循环,获取下一个可用密钥
|
| 143 |
+
log('info', f"API密钥 {current_api_key[:8]}... 失败,准备尝试下一个密钥",
|
| 144 |
+
extra={'key': current_api_key[:8], 'request_type': request_type})
|
| 145 |
+
break
|
| 146 |
+
else:
|
| 147 |
+
# 无法处理的错误或已达到重试上限
|
| 148 |
+
break
|
| 149 |
+
|
| 150 |
+
# 如果所有尝试都失败
|
| 151 |
+
msg = "所有API密钥均请求失败,请稍后重试"
|
| 152 |
+
log('error', "API key 替换失败,所有API key都已尝试,请重新配置或稍后重试", extra={'key': 'N/A', 'request_type': 'switch_key', 'status_code': 'N/A'})
|
| 153 |
+
|
| 154 |
+
# 对于流式请求,创建一个特殊的StreamingResponse返回错误
|
| 155 |
+
if chat_request.stream:
|
| 156 |
+
async def error_generator():
|
| 157 |
+
error_json = json.dumps({'error': {'message': msg, 'type': 'api_error'}})
|
| 158 |
+
yield f"data: {error_json}\n\n"
|
| 159 |
+
yield "data: [DONE]\n\n"
|
| 160 |
+
|
| 161 |
+
return StreamingResponse(error_generator(), media_type="text/event-stream")
|
| 162 |
+
else:
|
| 163 |
+
# 非流式请求使用标准HTTP异常
|
| 164 |
+
raise HTTPException(
|
| 165 |
+
status_code=status.HTTP_500_INTERNAL_SERVER_ERROR, detail=msg)
|
app/api/routes.py
CHANGED
|
@@ -1,33 +1,25 @@
|
|
| 1 |
from fastapi import APIRouter, HTTPException, Request, Depends, status
|
| 2 |
from fastapi.responses import JSONResponse, StreamingResponse
|
| 3 |
from app.models import ChatCompletionRequest, ChatCompletionResponse, ErrorResponse, ModelList
|
| 4 |
-
from app.services import GeminiClient
|
| 5 |
from app.utils import (
|
| 6 |
-
handle_gemini_error,
|
| 7 |
-
protect_from_abuse,
|
| 8 |
-
APIKeyManager,
|
| 9 |
-
test_api_key,
|
| 10 |
-
format_log_message,
|
| 11 |
-
log_manager,
|
| 12 |
generate_cache_key,
|
| 13 |
cache_response,
|
| 14 |
create_chat_response,
|
| 15 |
-
create_error_response
|
| 16 |
-
handle_api_error,
|
| 17 |
-
update_api_call_stats
|
| 18 |
)
|
| 19 |
-
import json
|
| 20 |
-
import asyncio
|
| 21 |
-
import time
|
| 22 |
-
import logging
|
| 23 |
-
import random
|
| 24 |
-
from typing import Literal
|
| 25 |
from app.config.settings import (
|
| 26 |
api_call_stats,
|
| 27 |
BLOCKED_MODELS
|
| 28 |
)
|
| 29 |
-
|
| 30 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 31 |
|
| 32 |
# 创建路由器
|
| 33 |
router = APIRouter()
|
|
@@ -76,23 +68,9 @@ def init_router(
|
|
| 76 |
MAX_REQUESTS_PER_MINUTE = _max_requests_per_minute
|
| 77 |
MAX_REQUESTS_PER_DAY_PER_IP = _max_requests_per_day_per_ip
|
| 78 |
|
| 79 |
-
#
|
| 80 |
-
def
|
| 81 |
-
|
| 82 |
-
msg = format_log_message(level.upper(), message, extra=extra)
|
| 83 |
-
getattr(logger, level.lower())(msg)
|
| 84 |
-
|
| 85 |
-
# 密码验证依赖
|
| 86 |
-
async def verify_password(request: Request):
|
| 87 |
-
if PASSWORD:
|
| 88 |
-
auth_header = request.headers.get("Authorization")
|
| 89 |
-
if not auth_header or not auth_header.startswith("Bearer "):
|
| 90 |
-
raise HTTPException(
|
| 91 |
-
status_code=401, detail="Unauthorized: Missing or invalid token")
|
| 92 |
-
token = auth_header.split(" ")[1]
|
| 93 |
-
if token != PASSWORD:
|
| 94 |
-
raise HTTPException(
|
| 95 |
-
status_code=401, detail="Unauthorized: Invalid token")
|
| 96 |
|
| 97 |
# API路由
|
| 98 |
@router.get("/v1/models", response_model=ModelList)
|
|
@@ -102,13 +80,27 @@ def list_models():
|
|
| 102 |
return ModelList(data=[{"id": model, "object": "model", "created": 1678888888, "owned_by": "organization-owner"} for model in filtered_models])
|
| 103 |
|
| 104 |
@router.post("/v1/chat/completions", response_model=ChatCompletionResponse)
|
| 105 |
-
async def chat_completions(request: ChatCompletionRequest, http_request: Request, _: None = Depends(
|
| 106 |
# 获取客户端IP
|
| 107 |
client_ip = http_request.client.host if http_request.client else "unknown"
|
| 108 |
|
| 109 |
# 流式请求直接处理,不使用缓存
|
| 110 |
if request.stream:
|
| 111 |
-
return await process_request(
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 112 |
|
| 113 |
# 生成完整缓存键 - 用于精确匹配
|
| 114 |
cache_key = generate_cache_key(request)
|
|
@@ -191,7 +183,23 @@ async def chat_completions(request: ChatCompletionRequest, http_request: Request
|
|
| 191 |
|
| 192 |
# 创建请求处理任务
|
| 193 |
process_task = asyncio.create_task(
|
| 194 |
-
process_request(
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 195 |
)
|
| 196 |
|
| 197 |
# 将任务添加到活跃请求池
|
|
@@ -213,706 +221,4 @@ async def chat_completions(request: ChatCompletionRequest, http_request: Request
|
|
| 213 |
return cached_response
|
| 214 |
|
| 215 |
# 重新抛出异常
|
| 216 |
-
raise
|
| 217 |
-
|
| 218 |
-
# 请求处理函数
|
| 219 |
-
async def process_request(chat_request: ChatCompletionRequest, http_request: Request, request_type: Literal['stream', 'non-stream'], cache_key: str = None, client_ip: str = None):
|
| 220 |
-
"""处理API请求的主函数,根据需要处理流式或非流式请求"""
|
| 221 |
-
global current_api_key
|
| 222 |
-
|
| 223 |
-
# 请求前基本检查
|
| 224 |
-
protect_from_abuse(
|
| 225 |
-
http_request, MAX_REQUESTS_PER_MINUTE, MAX_REQUESTS_PER_DAY_PER_IP)
|
| 226 |
-
if chat_request.model not in GeminiClient.AVAILABLE_MODELS:
|
| 227 |
-
error_msg = "无效的模型"
|
| 228 |
-
extra_log = {'request_type': request_type, 'model': chat_request.model, 'status_code': 400, 'error_message': error_msg}
|
| 229 |
-
log('error', error_msg, extra=extra_log)
|
| 230 |
-
raise HTTPException(
|
| 231 |
-
status_code=status.HTTP_400_BAD_REQUEST, detail=error_msg)
|
| 232 |
-
|
| 233 |
-
# 重置已尝试的密钥
|
| 234 |
-
key_manager.reset_tried_keys_for_request()
|
| 235 |
-
|
| 236 |
-
# 转换消息格式
|
| 237 |
-
contents, system_instruction = GeminiClient.convert_messages(
|
| 238 |
-
GeminiClient, chat_request.messages)
|
| 239 |
-
|
| 240 |
-
# 设置重试次数(使用可用API密钥数量作为最大重试次数)
|
| 241 |
-
retry_attempts = len(key_manager.api_keys) if key_manager.api_keys else 1
|
| 242 |
-
|
| 243 |
-
# 尝试使用不同API密钥
|
| 244 |
-
for attempt in range(1, retry_attempts + 1):
|
| 245 |
-
# 获取下一个密钥
|
| 246 |
-
current_api_key = key_manager.get_available_key()
|
| 247 |
-
|
| 248 |
-
# 检查API密钥是否可用
|
| 249 |
-
if current_api_key is None:
|
| 250 |
-
log('warning', "没有可用的 API 密钥,跳过本次尝试",
|
| 251 |
-
extra={'request_type': request_type, 'model': chat_request.model, 'status_code': 'N/A'})
|
| 252 |
-
break
|
| 253 |
-
|
| 254 |
-
# 记录当前尝试的密钥信息
|
| 255 |
-
log('info', f"第 {attempt}/{retry_attempts} 次尝试 ... 使用密钥: {current_api_key[:8]}...",
|
| 256 |
-
extra={'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model})
|
| 257 |
-
|
| 258 |
-
# 服务器错误重试逻辑
|
| 259 |
-
server_error_retries = 3
|
| 260 |
-
for server_retry in range(1, server_error_retries + 1):
|
| 261 |
-
try:
|
| 262 |
-
# 根据请求类型分别处理
|
| 263 |
-
if chat_request.stream:
|
| 264 |
-
try:
|
| 265 |
-
return await process_stream_request(
|
| 266 |
-
chat_request,
|
| 267 |
-
http_request,
|
| 268 |
-
contents,
|
| 269 |
-
system_instruction,
|
| 270 |
-
current_api_key
|
| 271 |
-
)
|
| 272 |
-
except Exception as e:
|
| 273 |
-
# 捕获流式请求的异常,但不立即返回错误
|
| 274 |
-
# 记录错误并继续尝试下一个API密钥
|
| 275 |
-
error_detail = handle_gemini_error(e, current_api_key, key_manager)
|
| 276 |
-
log('error', f"流式请求失败: {error_detail}",
|
| 277 |
-
extra={'key': current_api_key[:8], 'request_type': 'stream', 'model': chat_request.model})
|
| 278 |
-
# 不返回错误,而是抛出异常让外层循环处理
|
| 279 |
-
raise
|
| 280 |
-
else:
|
| 281 |
-
return await process_nonstream_request(
|
| 282 |
-
chat_request,
|
| 283 |
-
http_request,
|
| 284 |
-
request_type,
|
| 285 |
-
contents,
|
| 286 |
-
system_instruction,
|
| 287 |
-
current_api_key,
|
| 288 |
-
cache_key,
|
| 289 |
-
client_ip
|
| 290 |
-
)
|
| 291 |
-
except HTTPException as e:
|
| 292 |
-
if e.status_code == status.HTTP_408_REQUEST_TIMEOUT:
|
| 293 |
-
log('error', "客户端连接中断",
|
| 294 |
-
extra={'key': current_api_key[:8], 'request_type': request_type,
|
| 295 |
-
'model': chat_request.model, 'status_code': 408})
|
| 296 |
-
raise
|
| 297 |
-
else:
|
| 298 |
-
raise
|
| 299 |
-
except Exception as e:
|
| 300 |
-
# 使用统一的API错误处理函数
|
| 301 |
-
error_result = await handle_api_error(
|
| 302 |
-
e,
|
| 303 |
-
current_api_key,
|
| 304 |
-
key_manager,
|
| 305 |
-
request_type,
|
| 306 |
-
chat_request.model,
|
| 307 |
-
server_retry - 1
|
| 308 |
-
)
|
| 309 |
-
|
| 310 |
-
# 如果需要删除缓存,清除缓存
|
| 311 |
-
if error_result.get('remove_cache', False) and cache_key and cache_key in response_cache_manager.cache:
|
| 312 |
-
log('info', f"因API错误,删除缓存: {cache_key[:8]}...",
|
| 313 |
-
extra={'cache_operation': 'remove-on-error', 'request_type': request_type})
|
| 314 |
-
del response_cache_manager.cache[cache_key]
|
| 315 |
-
|
| 316 |
-
if error_result.get('should_retry', False):
|
| 317 |
-
# 服务器错误需要重试(等待已在handle_api_error中完成)
|
| 318 |
-
continue
|
| 319 |
-
elif error_result.get('should_switch_key', False) and attempt < retry_attempts:
|
| 320 |
-
# 跳出服务器错误重试循环,获取下一个可用密钥
|
| 321 |
-
log('info', f"API密钥 {current_api_key[:8]}... 失败,准备尝试下一个密钥",
|
| 322 |
-
extra={'key': current_api_key[:8], 'request_type': request_type})
|
| 323 |
-
break
|
| 324 |
-
else:
|
| 325 |
-
# 无法处理的错误或已达到重试上限
|
| 326 |
-
break
|
| 327 |
-
|
| 328 |
-
# 如果所有尝试都失败
|
| 329 |
-
msg = "所有API密钥均请求失败,请稍后重试"
|
| 330 |
-
log('error', "API key 替换失败,所有API key都已尝试,请重新配置或稍后重试", extra={'key': 'N/A', 'request_type': 'switch_key', 'status_code': 'N/A'})
|
| 331 |
-
|
| 332 |
-
# 对于流式请求,创建一个特殊的StreamingResponse返回错误
|
| 333 |
-
if chat_request.stream:
|
| 334 |
-
async def error_generator():
|
| 335 |
-
error_json = json.dumps({'error': {'message': msg, 'type': 'api_error'}})
|
| 336 |
-
yield f"data: {error_json}\n\n"
|
| 337 |
-
yield "data: [DONE]\n\n"
|
| 338 |
-
|
| 339 |
-
return StreamingResponse(error_generator(), media_type="text/event-stream")
|
| 340 |
-
else:
|
| 341 |
-
# 非流式请求使用标准HTTP异常
|
| 342 |
-
raise HTTPException(
|
| 343 |
-
status_code=status.HTTP_500_INTERNAL_SERVER_ERROR, detail=msg)
|
| 344 |
-
|
| 345 |
-
# 流式请求处理函数
|
| 346 |
-
async def process_stream_request(
|
| 347 |
-
chat_request: ChatCompletionRequest,
|
| 348 |
-
http_request: Request,
|
| 349 |
-
contents,
|
| 350 |
-
system_instruction,
|
| 351 |
-
current_api_key: str
|
| 352 |
-
) -> StreamingResponse:
|
| 353 |
-
"""处理流式API请求"""
|
| 354 |
-
|
| 355 |
-
# 创建一个直接流式响应的生成器函数
|
| 356 |
-
async def stream_response_generator():
|
| 357 |
-
# 如果启用了假流式模式,使用随机遍历API密钥的方式
|
| 358 |
-
if FAKE_STREAMING:
|
| 359 |
-
# 创建一个队列用于在任务之间传递数据
|
| 360 |
-
queue = asyncio.Queue()
|
| 361 |
-
keep_alive_task = None
|
| 362 |
-
api_request_task = None
|
| 363 |
-
|
| 364 |
-
try:
|
| 365 |
-
# 创建一个保持连接的任务,持续发送换行符
|
| 366 |
-
async def keep_alive_sender():
|
| 367 |
-
try:
|
| 368 |
-
# 创建一个Gemini客户端用于发送保持连接的换行符
|
| 369 |
-
keep_alive_client = GeminiClient(current_api_key)
|
| 370 |
-
|
| 371 |
-
# 启动保持连接的生成器
|
| 372 |
-
keep_alive_generator = keep_alive_client.stream_chat(
|
| 373 |
-
chat_request,
|
| 374 |
-
contents,
|
| 375 |
-
safety_settings_g2 if 'gemini-2.0-flash-exp' in chat_request.model else safety_settings,
|
| 376 |
-
system_instruction
|
| 377 |
-
)
|
| 378 |
-
|
| 379 |
-
# 持续发送换行符直到被取消
|
| 380 |
-
async for line in keep_alive_generator:
|
| 381 |
-
if line == "\n":
|
| 382 |
-
# 将换行符格式化为SSE格式
|
| 383 |
-
formatted_chunk = {
|
| 384 |
-
"id": "chatcmpl-keepalive",
|
| 385 |
-
"object": "chat.completion.chunk",
|
| 386 |
-
"created": int(time.time()),
|
| 387 |
-
"model": chat_request.model,
|
| 388 |
-
"choices": [{"delta": {"content": ""}, "index": 0, "finish_reason": None}]
|
| 389 |
-
}
|
| 390 |
-
# 将格式化的换行符放入队列
|
| 391 |
-
await queue.put(f"data: {json.dumps(formatted_chunk)}\n\n")
|
| 392 |
-
except asyncio.CancelledError:
|
| 393 |
-
log('info', "保持连接任务被取消",
|
| 394 |
-
extra={'key': current_api_key[:8], 'request_type': 'fake-stream'})
|
| 395 |
-
raise
|
| 396 |
-
except Exception as e:
|
| 397 |
-
log('error', f"保持连接任务出错: {str(e)}",
|
| 398 |
-
extra={'key': current_api_key[:8], 'request_type': 'fake-stream'})
|
| 399 |
-
# 将错误放入队列
|
| 400 |
-
await queue.put(None)
|
| 401 |
-
raise
|
| 402 |
-
|
| 403 |
-
# 创建一个任务来随机遍历API密钥并请求内容
|
| 404 |
-
async def api_request_handler():
|
| 405 |
-
success = False
|
| 406 |
-
try:
|
| 407 |
-
# 重置已尝试的密钥
|
| 408 |
-
key_manager.reset_tried_keys_for_request()
|
| 409 |
-
|
| 410 |
-
# 获取可用的API密钥
|
| 411 |
-
available_keys = key_manager.api_keys.copy()
|
| 412 |
-
random.shuffle(available_keys) # 随机打乱密钥顺序
|
| 413 |
-
|
| 414 |
-
# 遍历所有API密钥尝试获取响应
|
| 415 |
-
for attempt, api_key in enumerate(available_keys, 1):
|
| 416 |
-
try:
|
| 417 |
-
log('info', f"假流式模式: 尝试API密钥 {api_key[:8]}... ({attempt}/{len(available_keys)})",
|
| 418 |
-
extra={'key': api_key[:8], 'request_type': 'fake-stream', 'model': chat_request.model})
|
| 419 |
-
|
| 420 |
-
# 创建一个新的客户端使用当前API密钥
|
| 421 |
-
non_stream_client = GeminiClient(api_key)
|
| 422 |
-
|
| 423 |
-
# 使用非流式方式请求内容
|
| 424 |
-
response_content = await asyncio.to_thread(
|
| 425 |
-
non_stream_client.complete_chat,
|
| 426 |
-
chat_request,
|
| 427 |
-
contents,
|
| 428 |
-
safety_settings_g2 if 'gemini-2.0-flash-exp' in chat_request.model else safety_settings,
|
| 429 |
-
system_instruction
|
| 430 |
-
)
|
| 431 |
-
|
| 432 |
-
# 检查响应是否有效
|
| 433 |
-
if response_content and response_content.text:
|
| 434 |
-
log('info', f"假流式模式: API密钥 {api_key[:8]}... 成功获取响应",
|
| 435 |
-
extra={'key': api_key[:8], 'request_type': 'fake-stream', 'model': chat_request.model})
|
| 436 |
-
|
| 437 |
-
# 将完整响应分割成小块,模拟流式返回
|
| 438 |
-
full_text = response_content.text
|
| 439 |
-
chunk_size = max(len(full_text) // 10, 1) # 至少分成10块,每块至少1个字符
|
| 440 |
-
|
| 441 |
-
for i in range(0, len(full_text), chunk_size):
|
| 442 |
-
chunk = full_text[i:i+chunk_size]
|
| 443 |
-
formatted_chunk = {
|
| 444 |
-
"id": "chatcmpl-someid",
|
| 445 |
-
"object": "chat.completion.chunk",
|
| 446 |
-
"created": int(time.time()),
|
| 447 |
-
"model": chat_request.model,
|
| 448 |
-
"choices": [{"delta": {"role": "assistant", "content": chunk}, "index": 0, "finish_reason": None}]
|
| 449 |
-
}
|
| 450 |
-
# 将格式化的内容块放入队列
|
| 451 |
-
await queue.put(f"data: {json.dumps(formatted_chunk)}\n\n")
|
| 452 |
-
|
| 453 |
-
success = True
|
| 454 |
-
# 更新API调用统计
|
| 455 |
-
from app.utils.stats import update_api_call_stats
|
| 456 |
-
update_api_call_stats(api_call_stats,api_key)
|
| 457 |
-
break # 成功获取响应,退出循环
|
| 458 |
-
else:
|
| 459 |
-
log('warning', f"假流式模式: API密钥 {api_key[:8]}... 返回空响应",
|
| 460 |
-
extra={'key': api_key[:8], 'request_type': 'fake-stream', 'model': chat_request.model})
|
| 461 |
-
except Exception as e:
|
| 462 |
-
error_detail = handle_gemini_error(e, api_key, key_manager)
|
| 463 |
-
log('error', f"假流式模式: API密钥 {api_key[:8]}... 请求失败: {error_detail}",
|
| 464 |
-
extra={'key': api_key[:8], 'request_type': 'fake-stream', 'model': chat_request.model})
|
| 465 |
-
# 继续尝试下一个API密钥
|
| 466 |
-
|
| 467 |
-
# 如果所有API密钥都尝试失败
|
| 468 |
-
if not success:
|
| 469 |
-
error_msg = "所有API密钥均请求失败,请稍后重试"
|
| 470 |
-
log('error', error_msg,
|
| 471 |
-
extra={'key': 'ALL', 'request_type': 'fake-stream', 'model': chat_request.model})
|
| 472 |
-
|
| 473 |
-
# 添加错误信息到队列
|
| 474 |
-
error_json = {
|
| 475 |
-
"id": "chatcmpl-error",
|
| 476 |
-
"object": "chat.completion.chunk",
|
| 477 |
-
"created": int(time.time()),
|
| 478 |
-
"model": chat_request.model,
|
| 479 |
-
"choices": [{"delta": {"content": f"\n\n[错误: {error_msg}]"}, "index": 0, "finish_reason": "error"}]
|
| 480 |
-
}
|
| 481 |
-
await queue.put(f"data: {json.dumps(error_json)}\n\n")
|
| 482 |
-
|
| 483 |
-
# 添加完成标记到队列
|
| 484 |
-
await queue.put("data: [DONE]\n\n")
|
| 485 |
-
# 添加None表示队列结束
|
| 486 |
-
await queue.put(None)
|
| 487 |
-
|
| 488 |
-
except asyncio.CancelledError:
|
| 489 |
-
log('info', "API请求任务被取消",
|
| 490 |
-
extra={'key': current_api_key[:8], 'request_type': 'fake-stream'})
|
| 491 |
-
# 添加None表示队列结束
|
| 492 |
-
await queue.put(None)
|
| 493 |
-
raise
|
| 494 |
-
except Exception as e:
|
| 495 |
-
log('error', f"API请求任务出错: {str(e)}",
|
| 496 |
-
extra={'key': current_api_key[:8], 'request_type': 'fake-stream'})
|
| 497 |
-
# 添加错误信息到队列
|
| 498 |
-
error_json = {
|
| 499 |
-
"id": "chatcmpl-error",
|
| 500 |
-
"object": "chat.completion.chunk",
|
| 501 |
-
"created": int(time.time()),
|
| 502 |
-
"model": chat_request.model,
|
| 503 |
-
"choices": [{"delta": {"content": f"\n\n[错误: {str(e)}]"}, "index": 0, "finish_reason": "error"}]
|
| 504 |
-
}
|
| 505 |
-
await queue.put(f"data: {json.dumps(error_json)}\n\n")
|
| 506 |
-
await queue.put("data: [DONE]\n\n")
|
| 507 |
-
# 添加None表示队列结束
|
| 508 |
-
await queue.put(None)
|
| 509 |
-
raise
|
| 510 |
-
|
| 511 |
-
# 启动保持连接的任务
|
| 512 |
-
keep_alive_task = asyncio.create_task(keep_alive_sender())
|
| 513 |
-
# 启动API请求任务
|
| 514 |
-
api_request_task = asyncio.create_task(api_request_handler())
|
| 515 |
-
|
| 516 |
-
# 从队列中获取数据并发送给客户端
|
| 517 |
-
while True:
|
| 518 |
-
chunk = await queue.get()
|
| 519 |
-
if chunk is None: # None表示队列结束
|
| 520 |
-
break
|
| 521 |
-
yield chunk
|
| 522 |
-
|
| 523 |
-
# 如果API请求任务已完成,取消保持连接任务
|
| 524 |
-
if api_request_task.done() and not keep_alive_task.done():
|
| 525 |
-
keep_alive_task.cancel()
|
| 526 |
-
|
| 527 |
-
except asyncio.CancelledError:
|
| 528 |
-
log('info', "流式响应生成器被取消",
|
| 529 |
-
extra={'key': current_api_key[:8], 'request_type': 'fake-stream'})
|
| 530 |
-
# 取消所有任务
|
| 531 |
-
if keep_alive_task and not keep_alive_task.done():
|
| 532 |
-
keep_alive_task.cancel()
|
| 533 |
-
if api_request_task and not api_request_task.done():
|
| 534 |
-
api_request_task.cancel()
|
| 535 |
-
except Exception as e:
|
| 536 |
-
log('error', f"流式响应生成器出错: {str(e)}",
|
| 537 |
-
extra={'key': current_api_key[:8], 'request_type': 'fake-stream'})
|
| 538 |
-
# 取消所有任务
|
| 539 |
-
if keep_alive_task and not keep_alive_task.done():
|
| 540 |
-
keep_alive_task.cancel()
|
| 541 |
-
if api_request_task and not api_request_task.done():
|
| 542 |
-
api_request_task.cancel()
|
| 543 |
-
# 发送错误信息给客户端
|
| 544 |
-
error_json = {
|
| 545 |
-
"id": "chatcmpl-error",
|
| 546 |
-
"object": "chat.completion.chunk",
|
| 547 |
-
"created": int(time.time()),
|
| 548 |
-
"model": chat_request.model,
|
| 549 |
-
"choices": [{"delta": {"content": f"\n\n[错误: {str(e)}]"}, "index": 0, "finish_reason": "error"}]
|
| 550 |
-
}
|
| 551 |
-
yield f"data: {json.dumps(error_json)}\n\n"
|
| 552 |
-
yield "data: [DONE]\n\n"
|
| 553 |
-
finally:
|
| 554 |
-
# 确保所有任务都被取消
|
| 555 |
-
if keep_alive_task and not keep_alive_task.done():
|
| 556 |
-
keep_alive_task.cancel()
|
| 557 |
-
if api_request_task and not api_request_task.done():
|
| 558 |
-
api_request_task.cancel()
|
| 559 |
-
else:
|
| 560 |
-
# 原始流式请求处理逻辑
|
| 561 |
-
gemini_client = GeminiClient(current_api_key)
|
| 562 |
-
success = False
|
| 563 |
-
|
| 564 |
-
try:
|
| 565 |
-
# 直接迭代生成器并发送响应块
|
| 566 |
-
async for chunk in gemini_client.stream_chat(
|
| 567 |
-
chat_request,
|
| 568 |
-
contents,
|
| 569 |
-
safety_settings_g2 if 'gemini-2.0-flash-exp' in chat_request.model else safety_settings,
|
| 570 |
-
system_instruction
|
| 571 |
-
):
|
| 572 |
-
# 空字符串跳过
|
| 573 |
-
if not chunk:
|
| 574 |
-
continue
|
| 575 |
-
|
| 576 |
-
formatted_chunk = {
|
| 577 |
-
"id": "chatcmpl-someid",
|
| 578 |
-
"object": "chat.completion.chunk",
|
| 579 |
-
"created": int(time.time()),
|
| 580 |
-
"model": chat_request.model,
|
| 581 |
-
"choices": [{"delta": {"role": "assistant", "content": chunk}, "index": 0, "finish_reason": None}]
|
| 582 |
-
}
|
| 583 |
-
success = True # 只要有一个chunk成功,就标记为成功
|
| 584 |
-
yield f"data: {json.dumps(formatted_chunk)}\n\n"
|
| 585 |
-
|
| 586 |
-
# 如果成功获取到响应,更新API调用统计
|
| 587 |
-
if success:
|
| 588 |
-
from app.utils.stats import update_api_call_stats
|
| 589 |
-
update_api_call_stats(api_call_stats, current_api_key)
|
| 590 |
-
|
| 591 |
-
yield "data: [DONE]\n\n"
|
| 592 |
-
|
| 593 |
-
except asyncio.CancelledError:
|
| 594 |
-
extra_log_cancel = {'key': current_api_key[:8], 'request_type': 'stream', 'model': chat_request.model, 'error_message': '客户端已断开连接'}
|
| 595 |
-
log('info', "客户端连接已中断", extra=extra_log_cancel)
|
| 596 |
-
except Exception as e:
|
| 597 |
-
error_detail = handle_gemini_error(e, current_api_key, key_manager)
|
| 598 |
-
log('error', f"流式请求失败: {error_detail}",
|
| 599 |
-
extra={'key': current_api_key[:8], 'request_type': 'stream', 'model': chat_request.model})
|
| 600 |
-
# 发送错误信息给客户端
|
| 601 |
-
error_json = {
|
| 602 |
-
"id": "chatcmpl-error",
|
| 603 |
-
"object": "chat.completion.chunk",
|
| 604 |
-
"created": int(time.time()),
|
| 605 |
-
"model": chat_request.model,
|
| 606 |
-
"choices": [{"delta": {"content": f"\n\n[错误: {error_detail}]"}, "index": 0, "finish_reason": "error"}]
|
| 607 |
-
}
|
| 608 |
-
yield f"data: {json.dumps(error_json)}\n\n"
|
| 609 |
-
yield "data: [DONE]\n\n"
|
| 610 |
-
# 重新抛出异常,这样process_request可以捕获它
|
| 611 |
-
raise e
|
| 612 |
-
|
| 613 |
-
return StreamingResponse(stream_response_generator(), media_type="text/event-stream")
|
| 614 |
-
|
| 615 |
-
# Gemini完成请求函数
|
| 616 |
-
async def run_gemini_completion(
|
| 617 |
-
gemini_client,
|
| 618 |
-
chat_request: ChatCompletionRequest,
|
| 619 |
-
contents,
|
| 620 |
-
system_instruction,
|
| 621 |
-
request_type: str,
|
| 622 |
-
current_api_key: str
|
| 623 |
-
):
|
| 624 |
-
"""运行Gemini非流式请求"""
|
| 625 |
-
# 记录函数调用状态
|
| 626 |
-
run_fn = run_gemini_completion
|
| 627 |
-
|
| 628 |
-
try:
|
| 629 |
-
# 创建一个不会被客户端断开影响的任务
|
| 630 |
-
response_future = asyncio.create_task(
|
| 631 |
-
asyncio.to_thread(
|
| 632 |
-
gemini_client.complete_chat,
|
| 633 |
-
chat_request,
|
| 634 |
-
contents,
|
| 635 |
-
safety_settings_g2 if 'gemini-2.0-flash-exp' in chat_request.model else safety_settings,
|
| 636 |
-
system_instruction
|
| 637 |
-
)
|
| 638 |
-
)
|
| 639 |
-
|
| 640 |
-
# 使用shield防止任务被外部取消
|
| 641 |
-
response_content = await asyncio.shield(response_future)
|
| 642 |
-
|
| 643 |
-
# 只在第一次调用时记录完成日志
|
| 644 |
-
if not hasattr(run_fn, 'logged_complete'):
|
| 645 |
-
log('info', "非流式请求成功完成", extra={'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model})
|
| 646 |
-
run_fn.logged_complete = True
|
| 647 |
-
return response_content
|
| 648 |
-
except asyncio.CancelledError:
|
| 649 |
-
# 即使任务被取消,我们也确保正在进行的API请求能够完成
|
| 650 |
-
if 'response_future' in locals() and not response_future.done():
|
| 651 |
-
try:
|
| 652 |
-
# 使用shield确保任务不被取消,并等待它完成
|
| 653 |
-
response_content = await asyncio.shield(response_future)
|
| 654 |
-
log('info', "API请求在客户端断开后完成", extra={'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model})
|
| 655 |
-
return response_content
|
| 656 |
-
except Exception as e:
|
| 657 |
-
extra_log_gemini_cancel = {'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model, 'error_message': f'API请求在客户端断开后失败: {str(e)}'}
|
| 658 |
-
log('info', "API调用因客户端断开而失败", extra=extra_log_gemini_cancel)
|
| 659 |
-
raise
|
| 660 |
-
|
| 661 |
-
# 如果任务尚未开始或已经失败,记录日志
|
| 662 |
-
extra_log_gemini_cancel = {'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model, 'error_message': '客户端断开导致API调用取消'}
|
| 663 |
-
log('info', "API调用因客户端断开而取消", extra=extra_log_gemini_cancel)
|
| 664 |
-
raise
|
| 665 |
-
|
| 666 |
-
# 客户端断开检测函数
|
| 667 |
-
async def check_client_disconnect(http_request: Request, current_api_key: str, request_type: str, model: str):
|
| 668 |
-
"""检查客户端是否断开连接"""
|
| 669 |
-
while True:
|
| 670 |
-
if await http_request.is_disconnected():
|
| 671 |
-
extra_log = {'key': current_api_key[:8], 'request_type': request_type, 'model': model, 'error_message': '检测到客户端断开连接'}
|
| 672 |
-
log('info', "客户端连接已中断,等待API请求完成", extra=extra_log)
|
| 673 |
-
return True
|
| 674 |
-
await asyncio.sleep(0.5)
|
| 675 |
-
|
| 676 |
-
# 客户端断开处理函数
|
| 677 |
-
async def handle_client_disconnect(
|
| 678 |
-
gemini_task: asyncio.Task,
|
| 679 |
-
chat_request: ChatCompletionRequest,
|
| 680 |
-
request_type: str,
|
| 681 |
-
current_api_key: str,
|
| 682 |
-
cache_key: str = None,
|
| 683 |
-
client_ip: str = None
|
| 684 |
-
):
|
| 685 |
-
try:
|
| 686 |
-
# 等待API任务完成,使用shield防止它被取消
|
| 687 |
-
response_content = await asyncio.shield(gemini_task)
|
| 688 |
-
|
| 689 |
-
# 检查响应文本是否为空
|
| 690 |
-
if response_content is None or response_content.text == "":
|
| 691 |
-
if response_content is None:
|
| 692 |
-
log('info', "客户端断开后API任务返回None",
|
| 693 |
-
extra={'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model})
|
| 694 |
-
else:
|
| 695 |
-
extra_log = {'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model, 'status_code': 204}
|
| 696 |
-
log('info', "客户端断开后Gemini API 返回空响应", extra=extra_log)
|
| 697 |
-
|
| 698 |
-
# 删除任何现有缓存,因为响应为空
|
| 699 |
-
if cache_key and cache_key in response_cache_manager.cache:
|
| 700 |
-
log('info', f"因空响应,删除缓存: {cache_key[:8]}...",
|
| 701 |
-
extra={'cache_operation': 'remove-on-empty', 'request_type': request_type})
|
| 702 |
-
del response_cache_manager.cache[cache_key]
|
| 703 |
-
|
| 704 |
-
# 返回错误响应而不是None
|
| 705 |
-
return create_error_response(chat_request.model, "AI未返回任何内容,请重试")
|
| 706 |
-
|
| 707 |
-
# 首先检查是否有现有缓存
|
| 708 |
-
cached_response, cache_hit = response_cache_manager.get(cache_key)
|
| 709 |
-
if cache_hit:
|
| 710 |
-
log('info', f"客户端断开但找到已存在缓存,将删除: {cache_key[:8]}...",
|
| 711 |
-
extra={'cache_operation': 'disconnect-found-cache', 'request_type': request_type})
|
| 712 |
-
|
| 713 |
-
# 安全删除缓存
|
| 714 |
-
if cache_key in response_cache_manager.cache:
|
| 715 |
-
del response_cache_manager.cache[cache_key]
|
| 716 |
-
|
| 717 |
-
# 不返回缓存,而是创建新响应并缓存
|
| 718 |
-
|
| 719 |
-
# 创建新响应
|
| 720 |
-
from app.utils.response import create_response
|
| 721 |
-
response = create_response(chat_request, response_content)
|
| 722 |
-
|
| 723 |
-
# 客户端已断开,此响应不会实际发送,可以考虑将其缓存以供后续使用
|
| 724 |
-
# 如果确实需要缓存,则可以取消下面的注释
|
| 725 |
-
# cache_response(response, cache_key, client_ip)
|
| 726 |
-
|
| 727 |
-
return response
|
| 728 |
-
except asyncio.CancelledError:
|
| 729 |
-
# 对于取消异常,仍然尝试继续完成任务
|
| 730 |
-
log('info', "客户端断开后任务被取消,但我们仍会尝试完成",
|
| 731 |
-
extra={'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model})
|
| 732 |
-
|
| 733 |
-
# 检查任务是否已经完成
|
| 734 |
-
if gemini_task.done() and not gemini_task.cancelled():
|
| 735 |
-
try:
|
| 736 |
-
response_content = gemini_task.result()
|
| 737 |
-
|
| 738 |
-
# 首先检查是否有现有缓存
|
| 739 |
-
cached_response, cache_hit = response_cache_manager.get(cache_key)
|
| 740 |
-
if cache_hit:
|
| 741 |
-
log('info', f"任务被取消但找到已存在缓存,将删除: {cache_key[:8]}...",
|
| 742 |
-
extra={'cache_operation': 'cancel-found-cache', 'request_type': request_type})
|
| 743 |
-
|
| 744 |
-
# 安全删除缓存
|
| 745 |
-
if cache_key in response_cache_manager.cache:
|
| 746 |
-
del response_cache_manager.cache[cache_key]
|
| 747 |
-
|
| 748 |
-
# 创建但不缓存响应
|
| 749 |
-
from app.utils.response import create_response
|
| 750 |
-
response = create_response(chat_request, response_content)
|
| 751 |
-
return response
|
| 752 |
-
except Exception as inner_e:
|
| 753 |
-
log('error', f"客户端断开后从已完成任务获取结果失败: {str(inner_e)}",
|
| 754 |
-
extra={'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model})
|
| 755 |
-
|
| 756 |
-
# 删除缓存,因为出现错误
|
| 757 |
-
if cache_key and cache_key in response_cache_manager.cache:
|
| 758 |
-
log('info', f"因任务获取结果失败,删除缓存: {cache_key[:8]}...",
|
| 759 |
-
extra={'cache_operation': 'remove-on-error', 'request_type': request_type})
|
| 760 |
-
del response_cache_manager.cache[cache_key]
|
| 761 |
-
|
| 762 |
-
# 创建错误响应而不是返回None
|
| 763 |
-
return create_error_response(chat_request.model, "请求处理过程中发生错误,请重试")
|
| 764 |
-
except Exception as e:
|
| 765 |
-
# 处理API任务异常
|
| 766 |
-
error_msg = str(e)
|
| 767 |
-
extra_log = {'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model, 'error_message': error_msg}
|
| 768 |
-
log('error', f"客户端断开后处理API响应时出错: {error_msg}", extra=extra_log)
|
| 769 |
-
|
| 770 |
-
# 删除缓存,因为出现错误
|
| 771 |
-
if cache_key and cache_key in response_cache_manager.cache:
|
| 772 |
-
log('info', f"因API响应错误,删除缓存: {cache_key[:8]}...",
|
| 773 |
-
extra={'cache_operation': 'remove-on-error', 'request_type': request_type})
|
| 774 |
-
del response_cache_manager.cache[cache_key]
|
| 775 |
-
|
| 776 |
-
# 创建错误响应而不是返回None
|
| 777 |
-
return create_error_response(chat_request.model, f"请求处理错误: {error_msg}")
|
| 778 |
-
|
| 779 |
-
# 非流式请求处理函数
|
| 780 |
-
async def process_nonstream_request(
|
| 781 |
-
chat_request: ChatCompletionRequest,
|
| 782 |
-
http_request: Request,
|
| 783 |
-
request_type: str,
|
| 784 |
-
contents,
|
| 785 |
-
system_instruction,
|
| 786 |
-
current_api_key: str,
|
| 787 |
-
cache_key: str = None,
|
| 788 |
-
client_ip: str = None
|
| 789 |
-
):
|
| 790 |
-
"""处理非流式API请求"""
|
| 791 |
-
gemini_client = GeminiClient(current_api_key)
|
| 792 |
-
|
| 793 |
-
# 创建任务
|
| 794 |
-
gemini_task = asyncio.create_task(
|
| 795 |
-
run_gemini_completion(
|
| 796 |
-
gemini_client,
|
| 797 |
-
chat_request,
|
| 798 |
-
contents,
|
| 799 |
-
system_instruction,
|
| 800 |
-
request_type,
|
| 801 |
-
current_api_key
|
| 802 |
-
)
|
| 803 |
-
)
|
| 804 |
-
|
| 805 |
-
disconnect_task = asyncio.create_task(
|
| 806 |
-
check_client_disconnect(
|
| 807 |
-
http_request,
|
| 808 |
-
current_api_key,
|
| 809 |
-
request_type,
|
| 810 |
-
chat_request.model
|
| 811 |
-
)
|
| 812 |
-
)
|
| 813 |
-
|
| 814 |
-
try:
|
| 815 |
-
# 先等待看是否API任务先完成,或者客户端先断开连接
|
| 816 |
-
done, pending = await asyncio.wait(
|
| 817 |
-
[gemini_task, disconnect_task],
|
| 818 |
-
return_when=asyncio.FIRST_COMPLETED
|
| 819 |
-
)
|
| 820 |
-
|
| 821 |
-
if disconnect_task in done:
|
| 822 |
-
# 客户端已断开连接,但我们仍继续完成API请求以便缓存结果
|
| 823 |
-
return await handle_client_disconnect(
|
| 824 |
-
gemini_task,
|
| 825 |
-
chat_request,
|
| 826 |
-
request_type,
|
| 827 |
-
current_api_key,
|
| 828 |
-
cache_key,
|
| 829 |
-
client_ip
|
| 830 |
-
)
|
| 831 |
-
else:
|
| 832 |
-
# API任务先完成,取消断开检测任务
|
| 833 |
-
disconnect_task.cancel()
|
| 834 |
-
|
| 835 |
-
# 获取响应内容
|
| 836 |
-
response_content = await gemini_task
|
| 837 |
-
|
| 838 |
-
# 检查缓存是否已经存在,如果存在则不再创建新缓存
|
| 839 |
-
cached_response, cache_hit = response_cache_manager.get(cache_key)
|
| 840 |
-
if cache_hit:
|
| 841 |
-
log('info', f"缓存已存在,直接返回: {cache_key[:8]}...",
|
| 842 |
-
extra={'cache_operation': 'use-existing', 'request_type': request_type})
|
| 843 |
-
|
| 844 |
-
# 安全删除缓存
|
| 845 |
-
if cache_key in response_cache_manager.cache:
|
| 846 |
-
del response_cache_manager.cache[cache_key]
|
| 847 |
-
log('info', f"缓存使用后已删除: {cache_key[:8]}...",
|
| 848 |
-
extra={'cache_operation': 'used-and-removed', 'request_type': request_type})
|
| 849 |
-
|
| 850 |
-
return cached_response
|
| 851 |
-
|
| 852 |
-
# 创建响应
|
| 853 |
-
from app.utils.response import create_response
|
| 854 |
-
response = create_response(chat_request, response_content)
|
| 855 |
-
|
| 856 |
-
# 缓存响应
|
| 857 |
-
cache_response(response, cache_key, client_ip, response_cache_manager, update_api_call_stats, api_key=current_api_key)
|
| 858 |
-
|
| 859 |
-
# 立即删除缓存,确保只能使用一次
|
| 860 |
-
if cache_key and cache_key in response_cache_manager.cache:
|
| 861 |
-
del response_cache_manager.cache[cache_key]
|
| 862 |
-
log('info', f"缓存创建后立即删除: {cache_key[:8]}...",
|
| 863 |
-
extra={'cache_operation': 'store-and-remove', 'request_type': request_type})
|
| 864 |
-
|
| 865 |
-
# 返回响应
|
| 866 |
-
return response
|
| 867 |
-
|
| 868 |
-
except asyncio.CancelledError:
|
| 869 |
-
extra_log = {'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model, 'error_message':"请求被取消"}
|
| 870 |
-
log('info', "请求取消", extra=extra_log)
|
| 871 |
-
|
| 872 |
-
# 在请求被取消时先检查缓存中是否已有结果
|
| 873 |
-
cached_response, cache_hit = response_cache_manager.get(cache_key)
|
| 874 |
-
if cache_hit:
|
| 875 |
-
log('info', f"请求取消但找到有效缓存,使用缓存响应: {cache_key[:8]}...",
|
| 876 |
-
extra={'cache_operation': 'use-cache-on-cancel', 'request_type': request_type})
|
| 877 |
-
|
| 878 |
-
# 安全删除缓存
|
| 879 |
-
if cache_key in response_cache_manager.cache:
|
| 880 |
-
del response_cache_manager.cache[cache_key]
|
| 881 |
-
log('info', f"缓存使用后已删除: {cache_key[:8]}...",
|
| 882 |
-
extra={'cache_operation': 'used-and-removed', 'request_type': request_type})
|
| 883 |
-
|
| 884 |
-
return cached_response
|
| 885 |
-
|
| 886 |
-
# 尝试完成正在进行的API请求
|
| 887 |
-
if not gemini_task.done():
|
| 888 |
-
log('info', "请求取消但API请求尚未完成,继续等待...",
|
| 889 |
-
extra={'key': current_api_key[:8], 'request_type': request_type})
|
| 890 |
-
|
| 891 |
-
# 使用shield确保任务不会被取消
|
| 892 |
-
response_content = await asyncio.shield(gemini_task)
|
| 893 |
-
|
| 894 |
-
# 创建响应
|
| 895 |
-
from app.utils.response import create_response
|
| 896 |
-
response = create_response(chat_request, response_content)
|
| 897 |
-
|
| 898 |
-
# 不缓存这个响应,直接返回
|
| 899 |
-
return response
|
| 900 |
-
else:
|
| 901 |
-
# 任务已完成,获取结果
|
| 902 |
-
response_content = gemini_task.result()
|
| 903 |
-
|
| 904 |
-
# 创建响应
|
| 905 |
-
from app.utils.response import create_response
|
| 906 |
-
response = create_response(chat_request, response_content)
|
| 907 |
-
|
| 908 |
-
# 不缓存这个响应,直接返回
|
| 909 |
-
return response
|
| 910 |
-
|
| 911 |
-
except HTTPException as e:
|
| 912 |
-
if e.status_code == status.HTTP_408_REQUEST_TIMEOUT:
|
| 913 |
-
extra_log = {'key': current_api_key[:8], 'request_type': request_type, 'model': chat_request.model,
|
| 914 |
-
'status_code': 408, 'error_message': '客户端连接中断'}
|
| 915 |
-
log('error', "客户端连接中断,终止后续重试", extra=extra_log)
|
| 916 |
-
raise
|
| 917 |
-
else:
|
| 918 |
-
raise
|
|
|
|
| 1 |
from fastapi import APIRouter, HTTPException, Request, Depends, status
|
| 2 |
from fastapi.responses import JSONResponse, StreamingResponse
|
| 3 |
from app.models import ChatCompletionRequest, ChatCompletionResponse, ErrorResponse, ModelList
|
| 4 |
+
from app.services import GeminiClient
|
| 5 |
from app.utils import (
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
generate_cache_key,
|
| 7 |
cache_response,
|
| 8 |
create_chat_response,
|
| 9 |
+
create_error_response
|
|
|
|
|
|
|
| 10 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
from app.config.settings import (
|
| 12 |
api_call_stats,
|
| 13 |
BLOCKED_MODELS
|
| 14 |
)
|
| 15 |
+
import asyncio
|
| 16 |
+
import time
|
| 17 |
+
import logging
|
| 18 |
+
|
| 19 |
+
# 导入拆分后的模块
|
| 20 |
+
from .auth import verify_password
|
| 21 |
+
from .logging_utils import log
|
| 22 |
+
from .request_handlers import process_request
|
| 23 |
|
| 24 |
# 创建路由器
|
| 25 |
router = APIRouter()
|
|
|
|
| 68 |
MAX_REQUESTS_PER_MINUTE = _max_requests_per_minute
|
| 69 |
MAX_REQUESTS_PER_DAY_PER_IP = _max_requests_per_day_per_ip
|
| 70 |
|
| 71 |
+
# 自定义密码验证依赖
|
| 72 |
+
async def custom_verify_password(request: Request):
|
| 73 |
+
await verify_password(request, PASSWORD)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 74 |
|
| 75 |
# API路由
|
| 76 |
@router.get("/v1/models", response_model=ModelList)
|
|
|
|
| 80 |
return ModelList(data=[{"id": model, "object": "model", "created": 1678888888, "owned_by": "organization-owner"} for model in filtered_models])
|
| 81 |
|
| 82 |
@router.post("/v1/chat/completions", response_model=ChatCompletionResponse)
|
| 83 |
+
async def chat_completions(request: ChatCompletionRequest, http_request: Request, _: None = Depends(custom_verify_password)):
|
| 84 |
# 获取客户端IP
|
| 85 |
client_ip = http_request.client.host if http_request.client else "unknown"
|
| 86 |
|
| 87 |
# 流式请求直接处理,不使用缓存
|
| 88 |
if request.stream:
|
| 89 |
+
return await process_request(
|
| 90 |
+
request,
|
| 91 |
+
http_request,
|
| 92 |
+
"stream",
|
| 93 |
+
key_manager,
|
| 94 |
+
response_cache_manager,
|
| 95 |
+
active_requests_manager,
|
| 96 |
+
safety_settings,
|
| 97 |
+
safety_settings_g2,
|
| 98 |
+
api_call_stats,
|
| 99 |
+
FAKE_STREAMING,
|
| 100 |
+
FAKE_STREAMING_INTERVAL,
|
| 101 |
+
MAX_REQUESTS_PER_MINUTE,
|
| 102 |
+
MAX_REQUESTS_PER_DAY_PER_IP
|
| 103 |
+
)
|
| 104 |
|
| 105 |
# 生成完整缓存键 - 用于精确匹配
|
| 106 |
cache_key = generate_cache_key(request)
|
|
|
|
| 183 |
|
| 184 |
# 创建请求处理任务
|
| 185 |
process_task = asyncio.create_task(
|
| 186 |
+
process_request(
|
| 187 |
+
request,
|
| 188 |
+
http_request,
|
| 189 |
+
"non-stream",
|
| 190 |
+
key_manager,
|
| 191 |
+
response_cache_manager,
|
| 192 |
+
active_requests_manager,
|
| 193 |
+
safety_settings,
|
| 194 |
+
safety_settings_g2,
|
| 195 |
+
api_call_stats,
|
| 196 |
+
FAKE_STREAMING,
|
| 197 |
+
FAKE_STREAMING_INTERVAL,
|
| 198 |
+
MAX_REQUESTS_PER_MINUTE,
|
| 199 |
+
MAX_REQUESTS_PER_DAY_PER_IP,
|
| 200 |
+
cache_key,
|
| 201 |
+
client_ip
|
| 202 |
+
)
|
| 203 |
)
|
| 204 |
|
| 205 |
# 将任务添加到活跃请求池
|
|
|
|
| 221 |
return cached_response
|
| 222 |
|
| 223 |
# 重新抛出异常
|
| 224 |
+
raise
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
app/api/stream_handlers.py
ADDED
|
@@ -0,0 +1,284 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import asyncio
|
| 2 |
+
import json
|
| 3 |
+
import time
|
| 4 |
+
import random
|
| 5 |
+
from fastapi import Request
|
| 6 |
+
from fastapi.responses import StreamingResponse
|
| 7 |
+
from app.models import ChatCompletionRequest
|
| 8 |
+
from app.services import GeminiClient
|
| 9 |
+
from app.utils import handle_gemini_error, update_api_call_stats
|
| 10 |
+
from .logging_utils import log
|
| 11 |
+
|
| 12 |
+
# 流式请求处理函数
|
| 13 |
+
async def process_stream_request(
|
| 14 |
+
chat_request: ChatCompletionRequest,
|
| 15 |
+
http_request: Request,
|
| 16 |
+
contents,
|
| 17 |
+
system_instruction,
|
| 18 |
+
current_api_key: str,
|
| 19 |
+
key_manager,
|
| 20 |
+
safety_settings,
|
| 21 |
+
safety_settings_g2,
|
| 22 |
+
api_call_stats,
|
| 23 |
+
FAKE_STREAMING,
|
| 24 |
+
FAKE_STREAMING_INTERVAL
|
| 25 |
+
) -> StreamingResponse:
|
| 26 |
+
"""处理流式API请求"""
|
| 27 |
+
|
| 28 |
+
# 创建一个直接流式响应的生成器函数
|
| 29 |
+
async def stream_response_generator():
|
| 30 |
+
# 如果启用了假流式模式,使用随机遍历API密钥的方式
|
| 31 |
+
if FAKE_STREAMING:
|
| 32 |
+
# 创建一个队列用于在任务之间传递数据
|
| 33 |
+
queue = asyncio.Queue()
|
| 34 |
+
keep_alive_task = None
|
| 35 |
+
api_request_task = None
|
| 36 |
+
|
| 37 |
+
try:
|
| 38 |
+
# 创建一个保持连接的任务,持续发送换行符
|
| 39 |
+
async def keep_alive_sender():
|
| 40 |
+
try:
|
| 41 |
+
# 创建一个Gemini客户端用于发送保持连接的换行符
|
| 42 |
+
keep_alive_client = GeminiClient(current_api_key)
|
| 43 |
+
|
| 44 |
+
# 启动保持连接的生成器
|
| 45 |
+
keep_alive_generator = keep_alive_client.stream_chat(
|
| 46 |
+
chat_request,
|
| 47 |
+
contents,
|
| 48 |
+
safety_settings_g2 if 'gemini-2.0-flash-exp' in chat_request.model else safety_settings,
|
| 49 |
+
system_instruction
|
| 50 |
+
)
|
| 51 |
+
|
| 52 |
+
# 持续发送换行符直到被取消
|
| 53 |
+
async for line in keep_alive_generator:
|
| 54 |
+
if line == "\n":
|
| 55 |
+
# 将换行符格式化为SSE格式
|
| 56 |
+
formatted_chunk = {
|
| 57 |
+
"id": "chatcmpl-keepalive",
|
| 58 |
+
"object": "chat.completion.chunk",
|
| 59 |
+
"created": int(time.time()),
|
| 60 |
+
"model": chat_request.model,
|
| 61 |
+
"choices": [{"delta": {"content": ""}, "index": 0, "finish_reason": None}]
|
| 62 |
+
}
|
| 63 |
+
# 将格式化的换行符放入队列
|
| 64 |
+
await queue.put(f"data: {json.dumps(formatted_chunk)}\n\n")
|
| 65 |
+
except asyncio.CancelledError:
|
| 66 |
+
# log('info', "保持连接任务被取消",
|
| 67 |
+
# extra={'key': current_api_key[:8], 'request_type': 'fake-stream'})
|
| 68 |
+
raise
|
| 69 |
+
except Exception as e:
|
| 70 |
+
log('error', f"保持连接任务出错: {str(e)}",
|
| 71 |
+
extra={'key': current_api_key[:8], 'request_type': 'fake-stream'})
|
| 72 |
+
# 将错误放入队列
|
| 73 |
+
await queue.put(None)
|
| 74 |
+
raise
|
| 75 |
+
|
| 76 |
+
# 创建一个任务来随机遍历API密钥并请求内容
|
| 77 |
+
async def api_request_handler():
|
| 78 |
+
success = False
|
| 79 |
+
try:
|
| 80 |
+
# 重置已尝试的密钥
|
| 81 |
+
key_manager.reset_tried_keys_for_request()
|
| 82 |
+
|
| 83 |
+
# 获取可用的API密钥
|
| 84 |
+
available_keys = key_manager.api_keys.copy()
|
| 85 |
+
random.shuffle(available_keys) # 随机打乱密钥顺序
|
| 86 |
+
|
| 87 |
+
# 遍历所有API密钥尝试获取响应
|
| 88 |
+
for attempt, api_key in enumerate(available_keys, 1):
|
| 89 |
+
try:
|
| 90 |
+
log('info', f"假流式模式: 尝试API密钥 {api_key[:8]}... ({attempt}/{len(available_keys)})",
|
| 91 |
+
extra={'key': api_key[:8], 'request_type': 'fake-stream', 'model': chat_request.model})
|
| 92 |
+
|
| 93 |
+
# 创建一个新的客户端使用当前API密钥
|
| 94 |
+
non_stream_client = GeminiClient(api_key)
|
| 95 |
+
|
| 96 |
+
# 使用非流式方式请求内容
|
| 97 |
+
response_content = await asyncio.to_thread(
|
| 98 |
+
non_stream_client.complete_chat,
|
| 99 |
+
chat_request,
|
| 100 |
+
contents,
|
| 101 |
+
safety_settings_g2 if 'gemini-2.0-flash-exp' in chat_request.model else safety_settings,
|
| 102 |
+
system_instruction
|
| 103 |
+
)
|
| 104 |
+
|
| 105 |
+
# 检查响应是否有效
|
| 106 |
+
if response_content and response_content.text:
|
| 107 |
+
log('info', f"假流式模式: API密钥 {api_key[:8]}... 成功获取响应",
|
| 108 |
+
extra={'key': api_key[:8], 'request_type': 'fake-stream', 'model': chat_request.model})
|
| 109 |
+
|
| 110 |
+
# 将完整响应分割成小块,模拟流式返回
|
| 111 |
+
full_text = response_content.text
|
| 112 |
+
chunk_size = max(len(full_text) // 10, 1) # 至少分成10块,每块至少1个字符
|
| 113 |
+
|
| 114 |
+
for i in range(0, len(full_text), chunk_size):
|
| 115 |
+
chunk = full_text[i:i+chunk_size]
|
| 116 |
+
formatted_chunk = {
|
| 117 |
+
"id": "chatcmpl-someid",
|
| 118 |
+
"object": "chat.completion.chunk",
|
| 119 |
+
"created": int(time.time()),
|
| 120 |
+
"model": chat_request.model,
|
| 121 |
+
"choices": [{"delta": {"role": "assistant", "content": chunk}, "index": 0, "finish_reason": None}]
|
| 122 |
+
}
|
| 123 |
+
# 将格式化的内容块放入队列
|
| 124 |
+
await queue.put(f"data: {json.dumps(formatted_chunk)}\n\n")
|
| 125 |
+
|
| 126 |
+
success = True
|
| 127 |
+
# 更新API调用统计
|
| 128 |
+
update_api_call_stats(api_call_stats, api_key)
|
| 129 |
+
break # 成功获取响应,退出循环
|
| 130 |
+
else:
|
| 131 |
+
log('warning', f"假流式模式: API密钥 {api_key[:8]}... 返回空响应",
|
| 132 |
+
extra={'key': api_key[:8], 'request_type': 'fake-stream', 'model': chat_request.model})
|
| 133 |
+
except Exception as e:
|
| 134 |
+
error_detail = handle_gemini_error(e, api_key, key_manager)
|
| 135 |
+
log('error', f"假流式模式: API密钥 {api_key[:8]}... 请求失败: {error_detail}",
|
| 136 |
+
extra={'key': api_key[:8], 'request_type': 'fake-stream', 'model': chat_request.model})
|
| 137 |
+
# 继续尝试下一个API密钥
|
| 138 |
+
|
| 139 |
+
# 如果所有API密钥都尝试失败
|
| 140 |
+
if not success:
|
| 141 |
+
error_msg = "所有API密钥均请求失败,请稍后重试"
|
| 142 |
+
log('error', error_msg,
|
| 143 |
+
extra={'key': 'ALL', 'request_type': 'fake-stream', 'model': chat_request.model})
|
| 144 |
+
|
| 145 |
+
# 添加错误信息到队列
|
| 146 |
+
error_json = {
|
| 147 |
+
"id": "chatcmpl-error",
|
| 148 |
+
"object": "chat.completion.chunk",
|
| 149 |
+
"created": int(time.time()),
|
| 150 |
+
"model": chat_request.model,
|
| 151 |
+
"choices": [{"delta": {"content": f"\n\n[错误: {error_msg}]"}, "index": 0, "finish_reason": "error"}]
|
| 152 |
+
}
|
| 153 |
+
await queue.put(f"data: {json.dumps(error_json)}\n\n")
|
| 154 |
+
|
| 155 |
+
# 添加完成标记到队列
|
| 156 |
+
await queue.put("data: [DONE]\n\n")
|
| 157 |
+
# 添加None表示队列结束
|
| 158 |
+
await queue.put(None)
|
| 159 |
+
|
| 160 |
+
except asyncio.CancelledError:
|
| 161 |
+
log('info', "API请求任务被取消",
|
| 162 |
+
extra={'key': current_api_key[:8], 'request_type': 'fake-stream'})
|
| 163 |
+
# 添加None表示队列结束
|
| 164 |
+
await queue.put(None)
|
| 165 |
+
raise
|
| 166 |
+
except Exception as e:
|
| 167 |
+
log('error', f"API请求任务出错: {str(e)}",
|
| 168 |
+
extra={'key': current_api_key[:8], 'request_type': 'fake-stream'})
|
| 169 |
+
# 添加错误信息到队列
|
| 170 |
+
error_json = {
|
| 171 |
+
"id": "chatcmpl-error",
|
| 172 |
+
"object": "chat.completion.chunk",
|
| 173 |
+
"created": int(time.time()),
|
| 174 |
+
"model": chat_request.model,
|
| 175 |
+
"choices": [{"delta": {"content": f"\n\n[错误: {str(e)}]"}, "index": 0, "finish_reason": "error"}]
|
| 176 |
+
}
|
| 177 |
+
await queue.put(f"data: {json.dumps(error_json)}\n\n")
|
| 178 |
+
await queue.put("data: [DONE]\n\n")
|
| 179 |
+
# 添加None表示队列结束
|
| 180 |
+
await queue.put(None)
|
| 181 |
+
raise
|
| 182 |
+
|
| 183 |
+
# 启动保持连接的任务
|
| 184 |
+
keep_alive_task = asyncio.create_task(keep_alive_sender())
|
| 185 |
+
# 启动API请求任务
|
| 186 |
+
api_request_task = asyncio.create_task(api_request_handler())
|
| 187 |
+
|
| 188 |
+
# 从队列中获取数据并发送给客户端
|
| 189 |
+
while True:
|
| 190 |
+
chunk = await queue.get()
|
| 191 |
+
if chunk is None: # None表示队列结束
|
| 192 |
+
break
|
| 193 |
+
yield chunk
|
| 194 |
+
|
| 195 |
+
# 如果API请求任务已完成,取消保持连接任务
|
| 196 |
+
if api_request_task.done() and not keep_alive_task.done():
|
| 197 |
+
keep_alive_task.cancel()
|
| 198 |
+
|
| 199 |
+
except asyncio.CancelledError:
|
| 200 |
+
log('info', "流式响应生成器被取消",
|
| 201 |
+
extra={'key': current_api_key[:8], 'request_type': 'fake-stream'})
|
| 202 |
+
# 取消所有任务
|
| 203 |
+
if keep_alive_task and not keep_alive_task.done():
|
| 204 |
+
keep_alive_task.cancel()
|
| 205 |
+
if api_request_task and not api_request_task.done():
|
| 206 |
+
api_request_task.cancel()
|
| 207 |
+
except Exception as e:
|
| 208 |
+
log('error', f"流式响应生成器出错: {str(e)}",
|
| 209 |
+
extra={'key': current_api_key[:8], 'request_type': 'fake-stream'})
|
| 210 |
+
# 取消所有任务
|
| 211 |
+
if keep_alive_task and not keep_alive_task.done():
|
| 212 |
+
keep_alive_task.cancel()
|
| 213 |
+
if api_request_task and not api_request_task.done():
|
| 214 |
+
api_request_task.cancel()
|
| 215 |
+
# 发送错误信息给客户端
|
| 216 |
+
error_json = {
|
| 217 |
+
"id": "chatcmpl-error",
|
| 218 |
+
"object": "chat.completion.chunk",
|
| 219 |
+
"created": int(time.time()),
|
| 220 |
+
"model": chat_request.model,
|
| 221 |
+
"choices": [{"delta": {"content": f"\n\n[错误: {str(e)}]"}, "index": 0, "finish_reason": "error"}]
|
| 222 |
+
}
|
| 223 |
+
yield f"data: {json.dumps(error_json)}\n\n"
|
| 224 |
+
yield "data: [DONE]\n\n"
|
| 225 |
+
finally:
|
| 226 |
+
# 确保所有任务都被取消
|
| 227 |
+
if keep_alive_task and not keep_alive_task.done():
|
| 228 |
+
keep_alive_task.cancel()
|
| 229 |
+
if api_request_task and not api_request_task.done():
|
| 230 |
+
api_request_task.cancel()
|
| 231 |
+
else:
|
| 232 |
+
# 原始流式请求处理逻辑
|
| 233 |
+
gemini_client = GeminiClient(current_api_key)
|
| 234 |
+
success = False
|
| 235 |
+
|
| 236 |
+
try:
|
| 237 |
+
# 直接迭代生成器并发送响应块
|
| 238 |
+
async for chunk in gemini_client.stream_chat(
|
| 239 |
+
chat_request,
|
| 240 |
+
contents,
|
| 241 |
+
safety_settings_g2 if 'gemini-2.0-flash-exp' in chat_request.model else safety_settings,
|
| 242 |
+
system_instruction
|
| 243 |
+
):
|
| 244 |
+
# 空字符串跳过
|
| 245 |
+
if not chunk:
|
| 246 |
+
continue
|
| 247 |
+
|
| 248 |
+
formatted_chunk = {
|
| 249 |
+
"id": "chatcmpl-someid",
|
| 250 |
+
"object": "chat.completion.chunk",
|
| 251 |
+
"created": int(time.time()),
|
| 252 |
+
"model": chat_request.model,
|
| 253 |
+
"choices": [{"delta": {"role": "assistant", "content": chunk}, "index": 0, "finish_reason": None}]
|
| 254 |
+
}
|
| 255 |
+
success = True # 只要有一个chunk成功,就标记为成功
|
| 256 |
+
yield f"data: {json.dumps(formatted_chunk)}\n\n"
|
| 257 |
+
|
| 258 |
+
# 如果成功获取到响应,更新API调用统计
|
| 259 |
+
if success:
|
| 260 |
+
update_api_call_stats(api_call_stats, current_api_key)
|
| 261 |
+
|
| 262 |
+
yield "data: [DONE]\n\n"
|
| 263 |
+
|
| 264 |
+
except asyncio.CancelledError:
|
| 265 |
+
extra_log_cancel = {'key': current_api_key[:8], 'request_type': 'stream', 'model': chat_request.model, 'error_message': '客户端已断开连接'}
|
| 266 |
+
log('info', "客户端连接已中断", extra=extra_log_cancel)
|
| 267 |
+
except Exception as e:
|
| 268 |
+
error_detail = handle_gemini_error(e, current_api_key, key_manager)
|
| 269 |
+
log('error', f"流式请求失败: {error_detail}",
|
| 270 |
+
extra={'key': current_api_key[:8], 'request_type': 'stream', 'model': chat_request.model})
|
| 271 |
+
# 发送错误信息给客户端
|
| 272 |
+
error_json = {
|
| 273 |
+
"id": "chatcmpl-error",
|
| 274 |
+
"object": "chat.completion.chunk",
|
| 275 |
+
"created": int(time.time()),
|
| 276 |
+
"model": chat_request.model,
|
| 277 |
+
"choices": [{"delta": {"content": f"\n\n[错误: {error_detail}]"}, "index": 0, "finish_reason": "error"}]
|
| 278 |
+
}
|
| 279 |
+
yield f"data: {json.dumps(error_json)}\n\n"
|
| 280 |
+
yield "data: [DONE]\n\n"
|
| 281 |
+
# 重新抛出异常,这样process_request可以捕获它
|
| 282 |
+
raise e
|
| 283 |
+
|
| 284 |
+
return StreamingResponse(stream_response_generator(), media_type="text/event-stream")
|
app/config/settings.py
CHANGED
|
@@ -11,6 +11,11 @@ FAKE_STREAMING = os.environ.get("FAKE_STREAMING", "true").lower() in ["true", "1
|
|
| 11 |
# 假流式请求的空内容返回间隔(秒)
|
| 12 |
FAKE_STREAMING_INTERVAL = float(os.environ.get("FAKE_STREAMING_INTERVAL", "1"))
|
| 13 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 14 |
# 日志配置
|
| 15 |
logging.getLogger("uvicorn").disabled = True
|
| 16 |
logging.getLogger("uvicorn.access").disabled = True
|
|
@@ -35,10 +40,12 @@ REMOVE_CACHE_AFTER_USE = os.environ.get("REMOVE_CACHE_AFTER_USE", "true").lower(
|
|
| 35 |
REQUEST_HISTORY_EXPIRY_TIME = int(os.environ.get("REQUEST_HISTORY_EXPIRY_TIME", "600")) # 默认10分钟
|
| 36 |
ENABLE_RECONNECT_DETECTION = os.environ.get("ENABLE_RECONNECT_DETECTION", "true").lower() in ["true", "1", "yes"]
|
| 37 |
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
|
| 41 |
-
|
|
|
|
|
|
|
| 42 |
|
| 43 |
# API调用统计
|
| 44 |
api_call_stats = {
|
|
|
|
| 11 |
# 假流式请求的空内容返回间隔(秒)
|
| 12 |
FAKE_STREAMING_INTERVAL = float(os.environ.get("FAKE_STREAMING_INTERVAL", "1"))
|
| 13 |
|
| 14 |
+
#随机字符串
|
| 15 |
+
RANDOM_STRING = os.environ.get("RANDOM_STRING", "true").lower() in ["true", "1", "yes"]
|
| 16 |
+
RANDOM_STRING_LENGTH = int(os.environ.get("RANDOM_STRING_LENGTH", "20"))
|
| 17 |
+
|
| 18 |
+
|
| 19 |
# 日志配置
|
| 20 |
logging.getLogger("uvicorn").disabled = True
|
| 21 |
logging.getLogger("uvicorn.access").disabled = True
|
|
|
|
| 40 |
REQUEST_HISTORY_EXPIRY_TIME = int(os.environ.get("REQUEST_HISTORY_EXPIRY_TIME", "600")) # 默认10分钟
|
| 41 |
ENABLE_RECONNECT_DETECTION = os.environ.get("ENABLE_RECONNECT_DETECTION", "true").lower() in ["true", "1", "yes"]
|
| 42 |
|
| 43 |
+
|
| 44 |
+
version={
|
| 45 |
+
"local_version":"0.0.0",
|
| 46 |
+
"remote_version":"0.0.0",
|
| 47 |
+
"has_update":False
|
| 48 |
+
}
|
| 49 |
|
| 50 |
# API调用统计
|
| 51 |
api_call_stats = {
|
app/main.py
CHANGED
|
@@ -22,6 +22,8 @@ from app.api import router, init_router, dashboard_router, init_dashboard_router
|
|
| 22 |
from app.config.settings import (
|
| 23 |
FAKE_STREAMING,
|
| 24 |
FAKE_STREAMING_INTERVAL,
|
|
|
|
|
|
|
| 25 |
PASSWORD,
|
| 26 |
MAX_REQUESTS_PER_MINUTE,
|
| 27 |
MAX_REQUESTS_PER_DAY_PER_IP,
|
|
@@ -34,9 +36,7 @@ from app.config.settings import (
|
|
| 34 |
ENABLE_RECONNECT_DETECTION,
|
| 35 |
api_call_stats,
|
| 36 |
client_request_history,
|
| 37 |
-
|
| 38 |
-
remote_version,
|
| 39 |
-
has_update,
|
| 40 |
API_KEY_DAILY_LIMIT
|
| 41 |
)
|
| 42 |
from app.config.safety import SAFETY_SETTINGS, SAFETY_SETTINGS_G2
|
|
@@ -115,7 +115,6 @@ async def startup_event():
|
|
| 115 |
|
| 116 |
# 检查版本
|
| 117 |
await check_version()
|
| 118 |
-
|
| 119 |
available_keys = await check_keys()
|
| 120 |
if available_keys:
|
| 121 |
key_manager.api_keys = available_keys
|
|
@@ -202,7 +201,7 @@ async def root(request: Request):
|
|
| 202 |
continue
|
| 203 |
|
| 204 |
# 获取最近的日志
|
| 205 |
-
recent_logs = log_manager.get_recent_logs(
|
| 206 |
|
| 207 |
# 获取缓存统计
|
| 208 |
total_cache = len(response_cache_manager.cache)
|
|
@@ -267,9 +266,15 @@ async def root(request: Request):
|
|
| 267 |
"current_time": datetime.now().strftime('%H:%M:%S'),
|
| 268 |
"logs": recent_logs,
|
| 269 |
# 添加版本信息
|
| 270 |
-
"local_version": local_version,
|
| 271 |
-
"remote_version": remote_version,
|
| 272 |
-
"has_update": has_update,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 273 |
# 添加缓存信息
|
| 274 |
"cache_entries": total_cache,
|
| 275 |
"valid_cache": valid_cache,
|
|
|
|
| 22 |
from app.config.settings import (
|
| 23 |
FAKE_STREAMING,
|
| 24 |
FAKE_STREAMING_INTERVAL,
|
| 25 |
+
RANDOM_STRING,
|
| 26 |
+
RANDOM_STRING_LENGTH,
|
| 27 |
PASSWORD,
|
| 28 |
MAX_REQUESTS_PER_MINUTE,
|
| 29 |
MAX_REQUESTS_PER_DAY_PER_IP,
|
|
|
|
| 36 |
ENABLE_RECONNECT_DETECTION,
|
| 37 |
api_call_stats,
|
| 38 |
client_request_history,
|
| 39 |
+
version,
|
|
|
|
|
|
|
| 40 |
API_KEY_DAILY_LIMIT
|
| 41 |
)
|
| 42 |
from app.config.safety import SAFETY_SETTINGS, SAFETY_SETTINGS_G2
|
|
|
|
| 115 |
|
| 116 |
# 检查版本
|
| 117 |
await check_version()
|
|
|
|
| 118 |
available_keys = await check_keys()
|
| 119 |
if available_keys:
|
| 120 |
key_manager.api_keys = available_keys
|
|
|
|
| 201 |
continue
|
| 202 |
|
| 203 |
# 获取最近的日志
|
| 204 |
+
recent_logs = log_manager.get_recent_logs(500) # 获取最近50条日志
|
| 205 |
|
| 206 |
# 获取缓存统计
|
| 207 |
total_cache = len(response_cache_manager.cache)
|
|
|
|
| 266 |
"current_time": datetime.now().strftime('%H:%M:%S'),
|
| 267 |
"logs": recent_logs,
|
| 268 |
# 添加版本信息
|
| 269 |
+
"local_version": version["local_version"],
|
| 270 |
+
"remote_version": version["remote_version"],
|
| 271 |
+
"has_update": version["has_update"],
|
| 272 |
+
# 添加流式响应配置
|
| 273 |
+
"fake_streaming": FAKE_STREAMING,
|
| 274 |
+
"fake_streaming_interval": FAKE_STREAMING_INTERVAL,
|
| 275 |
+
# 添加随机字符串配置
|
| 276 |
+
"random_string": RANDOM_STRING,
|
| 277 |
+
"random_string_length": RANDOM_STRING_LENGTH,
|
| 278 |
# 添加缓存信息
|
| 279 |
"cache_entries": total_cache,
|
| 280 |
"valid_cache": valid_cache,
|
app/services/gemini.py
CHANGED
|
@@ -8,7 +8,18 @@ from dataclasses import dataclass
|
|
| 8 |
from typing import Optional, Dict, Any, List
|
| 9 |
import httpx
|
| 10 |
import logging
|
|
|
|
|
|
|
| 11 |
from app.utils import format_log_message
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 12 |
|
| 13 |
logger = logging.getLogger('my_logger')
|
| 14 |
|
|
@@ -269,7 +280,6 @@ class GeminiClient:
|
|
| 269 |
for i, message in enumerate(messages):
|
| 270 |
role = message.role
|
| 271 |
content = message.content
|
| 272 |
-
|
| 273 |
if isinstance(content, str):
|
| 274 |
if is_system_phase and role == 'system':
|
| 275 |
if system_instruction_text:
|
|
@@ -331,6 +341,11 @@ class GeminiClient:
|
|
| 331 |
if errors:
|
| 332 |
return errors
|
| 333 |
else:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 334 |
return gemini_history, {"parts": [{"text": system_instruction_text}]}
|
| 335 |
|
| 336 |
@staticmethod
|
|
|
|
| 8 |
from typing import Optional, Dict, Any, List
|
| 9 |
import httpx
|
| 10 |
import logging
|
| 11 |
+
import secrets
|
| 12 |
+
import string
|
| 13 |
from app.utils import format_log_message
|
| 14 |
+
from app.config.settings import (
|
| 15 |
+
RANDOM_STRING,
|
| 16 |
+
RANDOM_STRING_LENGTH
|
| 17 |
+
)
|
| 18 |
+
|
| 19 |
+
def generate_secure_random_string(length):
|
| 20 |
+
all_characters = string.ascii_letters + string.digits
|
| 21 |
+
secure_random_string = ''.join(secrets.choice(all_characters) for _ in range(length))
|
| 22 |
+
return secure_random_string
|
| 23 |
|
| 24 |
logger = logging.getLogger('my_logger')
|
| 25 |
|
|
|
|
| 280 |
for i, message in enumerate(messages):
|
| 281 |
role = message.role
|
| 282 |
content = message.content
|
|
|
|
| 283 |
if isinstance(content, str):
|
| 284 |
if is_system_phase and role == 'system':
|
| 285 |
if system_instruction_text:
|
|
|
|
| 341 |
if errors:
|
| 342 |
return errors
|
| 343 |
else:
|
| 344 |
+
if RANDOM_STRING:
|
| 345 |
+
gemini_history.insert(1,{'role': 'user', 'parts': [{'text': generate_secure_random_string(RANDOM_STRING_LENGTH)}]})
|
| 346 |
+
gemini_history.insert(len(gemini_history)-1,{'role': 'user', 'parts': [{'text': generate_secure_random_string(RANDOM_STRING_LENGTH)}]})
|
| 347 |
+
log_msg = format_log_message('INFO', "伪装消息成功")
|
| 348 |
+
logger.info(log_msg)
|
| 349 |
return gemini_history, {"parts": [{"text": system_instruction_text}]}
|
| 350 |
|
| 351 |
@staticmethod
|
app/templates/index.html
CHANGED
|
@@ -240,6 +240,15 @@
|
|
| 240 |
.progress-bar.high {
|
| 241 |
background-color: #dc3545; /* 红色 - 高使用率 */
|
| 242 |
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 243 |
</style>
|
| 244 |
</head>
|
| 245 |
<body>
|
|
@@ -294,6 +303,9 @@
|
|
| 294 |
|
| 295 |
<div class="info-box">
|
| 296 |
<h2 class="section-title">⚙️ 环境配置</h2>
|
|
|
|
|
|
|
|
|
|
| 297 |
<div class="stats-grid">
|
| 298 |
<div class="stat-card">
|
| 299 |
<div class="stat-value">{{ max_requests_per_minute }}</div>
|
|
@@ -308,30 +320,41 @@
|
|
| 308 |
<div class="stat-label">当前服务器时间</div>
|
| 309 |
</div>
|
| 310 |
</div>
|
| 311 |
-
|
| 312 |
-
|
| 313 |
-
|
| 314 |
-
<
|
| 315 |
-
|
| 316 |
-
|
| 317 |
-
|
| 318 |
</div>
|
| 319 |
-
|
| 320 |
-
|
| 321 |
-
<div
|
| 322 |
-
|
| 323 |
-
|
| 324 |
-
|
| 325 |
-
|
| 326 |
-
|
| 327 |
-
|
| 328 |
-
|
| 329 |
-
|
| 330 |
-
|
| 331 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 332 |
</div>
|
|
|
|
| 333 |
</div>
|
| 334 |
-
{% endif %}
|
| 335 |
</div>
|
| 336 |
</div>
|
| 337 |
|
|
|
|
| 240 |
.progress-bar.high {
|
| 241 |
background-color: #dc3545; /* 红色 - 高使用率 */
|
| 242 |
}
|
| 243 |
+
|
| 244 |
+
/* 版本更新状态样式 */
|
| 245 |
+
.update-needed {
|
| 246 |
+
color: #dc3545; /* 红色 - 需要更新 */
|
| 247 |
+
}
|
| 248 |
+
|
| 249 |
+
.up-to-date {
|
| 250 |
+
color: #28a745; /* 绿色 - 已是最新 */
|
| 251 |
+
}
|
| 252 |
</style>
|
| 253 |
</head>
|
| 254 |
<body>
|
|
|
|
| 303 |
|
| 304 |
<div class="info-box">
|
| 305 |
<h2 class="section-title">⚙️ 环境配置</h2>
|
| 306 |
+
|
| 307 |
+
<!-- 基本配置 -->
|
| 308 |
+
<h3 class="section-title">基本配置</h3>
|
| 309 |
<div class="stats-grid">
|
| 310 |
<div class="stat-card">
|
| 311 |
<div class="stat-value">{{ max_requests_per_minute }}</div>
|
|
|
|
| 320 |
<div class="stat-label">当前服务器时间</div>
|
| 321 |
</div>
|
| 322 |
</div>
|
| 323 |
+
|
| 324 |
+
<!-- 功能配置 -->
|
| 325 |
+
<h3 class="section-title">功能配置</h3>
|
| 326 |
+
<div class="stats-grid">
|
| 327 |
+
<div class="stat-card">
|
| 328 |
+
<div class="stat-value">{{ "启用" if fake_streaming else "禁用" }}</div>
|
| 329 |
+
<div class="stat-label">假流式响应</div>
|
| 330 |
</div>
|
| 331 |
+
<div class="stat-card">
|
| 332 |
+
<div class="stat-value">{{ fake_streaming_interval }}秒</div>
|
| 333 |
+
<div class="stat-label">假流式间隔</div>
|
| 334 |
+
</div>
|
| 335 |
+
<div class="stat-card">
|
| 336 |
+
<div class="stat-value">{{ "启用" if random_string else "禁用" }}</div>
|
| 337 |
+
<div class="stat-label">伪装信息</div>
|
| 338 |
+
</div>
|
| 339 |
+
</div>
|
| 340 |
+
|
| 341 |
+
<!-- 版本信息 -->
|
| 342 |
+
<h3 class="section-title">版本信息</h3>
|
| 343 |
+
<div class="stats-grid">
|
| 344 |
+
<div class="stat-card">
|
| 345 |
+
<div class="stat-value">{{ local_version }}</div>
|
| 346 |
+
<div class="stat-label">当前版本</div>
|
| 347 |
+
</div>
|
| 348 |
+
<div class="stat-card">
|
| 349 |
+
<div class="stat-value">{{ remote_version }}</div>
|
| 350 |
+
<div class="stat-label">最新版本</div>
|
| 351 |
+
</div>
|
| 352 |
+
<div class="stat-card">
|
| 353 |
+
<div class="stat-value {% if has_update %}update-needed{% else %}up-to-date{% endif %}">
|
| 354 |
+
{{ "需要更新" if has_update else "已是最新" }}
|
| 355 |
</div>
|
| 356 |
+
<div class="stat-label">更新状态</div>
|
| 357 |
</div>
|
|
|
|
| 358 |
</div>
|
| 359 |
</div>
|
| 360 |
|
app/utils/maintenance.py
CHANGED
|
@@ -1,9 +1,12 @@
|
|
| 1 |
-
import sys
|
| 2 |
-
from apscheduler.schedulers.background import BackgroundScheduler
|
|
|
|
| 3 |
from app.utils.logging import log
|
| 4 |
from app.utils.stats import clean_expired_stats
|
| 5 |
from app.config import api_call_stats
|
| 6 |
from app.utils import check_version
|
|
|
|
|
|
|
| 7 |
def handle_exception(exc_type, exc_value, exc_traceback):
|
| 8 |
"""
|
| 9 |
全局异常处理函数
|
|
@@ -16,7 +19,7 @@ def handle_exception(exc_type, exc_value, exc_traceback):
|
|
| 16 |
from app.utils.error_handling import translate_error
|
| 17 |
error_message = translate_error(str(exc_value))
|
| 18 |
log('error', f"未捕获的异常: {error_message}", status_code=500, error_message=error_message)
|
| 19 |
-
|
| 20 |
def schedule_cache_cleanup(response_cache_manager, active_requests_manager):
|
| 21 |
"""
|
| 22 |
设置定期清理缓存和活跃请求的定时任务
|
|
@@ -30,7 +33,45 @@ def schedule_cache_cleanup(response_cache_manager, active_requests_manager):
|
|
| 30 |
scheduler.add_job(active_requests_manager.clean_completed, 'interval', seconds=30) # 每30秒清理已完成的活跃请求
|
| 31 |
scheduler.add_job(active_requests_manager.clean_long_running, 'interval', minutes=5, args=[300]) # 每5分钟清理运行超过5分钟的任务
|
| 32 |
scheduler.add_job(clean_expired_stats, 'interval', minutes=5,args=[api_call_stats]) # 每5分钟清理过期的统计数据
|
| 33 |
-
scheduler.add_job(check_version, 'interval', minutes=
|
| 34 |
scheduler.start()
|
| 35 |
|
| 36 |
-
return scheduler
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import sys,asyncio
|
| 2 |
+
#from apscheduler.schedulers.background import BackgroundScheduler
|
| 3 |
+
from apscheduler.schedulers.asyncio import AsyncIOScheduler # 替换为异步调度器
|
| 4 |
from app.utils.logging import log
|
| 5 |
from app.utils.stats import clean_expired_stats
|
| 6 |
from app.config import api_call_stats
|
| 7 |
from app.utils import check_version
|
| 8 |
+
from zoneinfo import ZoneInfo
|
| 9 |
+
from app.config import settings
|
| 10 |
def handle_exception(exc_type, exc_value, exc_traceback):
|
| 11 |
"""
|
| 12 |
全局异常处理函数
|
|
|
|
| 19 |
from app.utils.error_handling import translate_error
|
| 20 |
error_message = translate_error(str(exc_value))
|
| 21 |
log('error', f"未捕获的异常: {error_message}", status_code=500, error_message=error_message)
|
| 22 |
+
'''
|
| 23 |
def schedule_cache_cleanup(response_cache_manager, active_requests_manager):
|
| 24 |
"""
|
| 25 |
设置定期清理缓存和活跃请求的定时任务
|
|
|
|
| 33 |
scheduler.add_job(active_requests_manager.clean_completed, 'interval', seconds=30) # 每30秒清理已完成的活跃请求
|
| 34 |
scheduler.add_job(active_requests_manager.clean_long_running, 'interval', minutes=5, args=[300]) # 每5分钟清理运行超过5分钟的任务
|
| 35 |
scheduler.add_job(clean_expired_stats, 'interval', minutes=5,args=[api_call_stats]) # 每5分钟清理过期的统计数据
|
| 36 |
+
scheduler.add_job(check_version, 'interval', minutes=1) # 每4小时检查更新
|
| 37 |
scheduler.start()
|
| 38 |
|
| 39 |
+
return scheduler
|
| 40 |
+
|
| 41 |
+
'''
|
| 42 |
+
def schedule_cache_cleanup(response_cache_manager, active_requests_manager):
|
| 43 |
+
"""
|
| 44 |
+
设置定期清理缓存和活跃请求的定时任务
|
| 45 |
+
顺便定时检查更新
|
| 46 |
+
Args:
|
| 47 |
+
response_cache_manager: 响应缓存管理器实例
|
| 48 |
+
active_requests_manager: 活跃请求管理器实例
|
| 49 |
+
"""
|
| 50 |
+
beijing_tz = ZoneInfo("Asia/Shanghai")
|
| 51 |
+
scheduler = AsyncIOScheduler(timezone=beijing_tz) # 使用 AsyncIOScheduler 替代 BackgroundScheduler
|
| 52 |
+
|
| 53 |
+
# 添加任务时直接传递异步函数(无需额外包装)
|
| 54 |
+
scheduler.add_job(response_cache_manager.clean_expired, 'interval', minutes=1)
|
| 55 |
+
scheduler.add_job(active_requests_manager.clean_completed, 'interval', seconds=30)
|
| 56 |
+
scheduler.add_job(active_requests_manager.clean_long_running, 'interval', minutes=5, args=[300])
|
| 57 |
+
scheduler.add_job(clean_expired_stats, 'interval', minutes=5, args=[api_call_stats])
|
| 58 |
+
scheduler.add_job(check_version, 'interval', hours=4)
|
| 59 |
+
scheduler.add_job(api_call_stats_clean, 'cron', hour=16,minute=0)
|
| 60 |
+
scheduler.start()
|
| 61 |
+
return scheduler
|
| 62 |
+
|
| 63 |
+
def api_call_stats_clean():
|
| 64 |
+
settings.api_call_stats = {
|
| 65 |
+
'last_24h': {
|
| 66 |
+
'total': {},
|
| 67 |
+
'by_endpoint': {}
|
| 68 |
+
},
|
| 69 |
+
'hourly': {
|
| 70 |
+
'total': {},
|
| 71 |
+
'by_endpoint': {}
|
| 72 |
+
},
|
| 73 |
+
'minute': {
|
| 74 |
+
'total': {},
|
| 75 |
+
'by_endpoint': {}
|
| 76 |
+
}
|
| 77 |
+
}
|
app/utils/version.py
CHANGED
|
@@ -13,18 +13,17 @@ async def check_version():
|
|
| 13 |
# 读取本地版本
|
| 14 |
with open("./version.txt", "r") as f:
|
| 15 |
version_line = f.read().strip()
|
| 16 |
-
settings.local_version = version_line.split("=")[1] if "=" in version_line else "0.0.0"
|
| 17 |
|
| 18 |
# 获取远程版本
|
| 19 |
github_url = "https://raw.githubusercontent.com/wyeeeee/hajimi/refs/heads/main/version.txt"
|
| 20 |
response = requests.get(github_url, timeout=5)
|
| 21 |
if response.status_code == 200:
|
| 22 |
version_line = response.text.strip()
|
| 23 |
-
settings.remote_version
|
| 24 |
-
|
| 25 |
# 比较版本号
|
| 26 |
-
local_parts = [int(x) for x in settings.local_version.split(".")]
|
| 27 |
-
remote_parts = [int(x) for x in settings.remote_version.split(".")]
|
| 28 |
|
| 29 |
# 确保两个列表长度相同
|
| 30 |
while len(local_parts) < len(remote_parts):
|
|
@@ -33,18 +32,18 @@ async def check_version():
|
|
| 33 |
remote_parts.append(0)
|
| 34 |
|
| 35 |
# 比较版本号
|
| 36 |
-
settings.has_update = False
|
| 37 |
for i in range(len(local_parts)):
|
| 38 |
if remote_parts[i] > local_parts[i]:
|
| 39 |
-
settings.has_update = True
|
| 40 |
break
|
| 41 |
elif remote_parts[i] < local_parts[i]:
|
| 42 |
break
|
| 43 |
|
| 44 |
-
log('info', f"版本检查: 本地版本 {settings.local_version}, 远程版本 {settings.remote_version}, 有更新: {settings.has_update}")
|
| 45 |
else:
|
| 46 |
log('warning', f"无法获取远程版本信息,HTTP状态码: {response.status_code}")
|
| 47 |
except Exception as e:
|
| 48 |
log('error', f"版本检查失败: {str(e)}")
|
| 49 |
|
| 50 |
-
return settings.
|
|
|
|
| 13 |
# 读取本地版本
|
| 14 |
with open("./version.txt", "r") as f:
|
| 15 |
version_line = f.read().strip()
|
| 16 |
+
settings.version['local_version'] = version_line.split("=")[1] if "=" in version_line else "0.0.0"
|
| 17 |
|
| 18 |
# 获取远程版本
|
| 19 |
github_url = "https://raw.githubusercontent.com/wyeeeee/hajimi/refs/heads/main/version.txt"
|
| 20 |
response = requests.get(github_url, timeout=5)
|
| 21 |
if response.status_code == 200:
|
| 22 |
version_line = response.text.strip()
|
| 23 |
+
settings.version['remote_version']= version_line.split("=")[1] if "=" in version_line else "0.0.0"
|
|
|
|
| 24 |
# 比较版本号
|
| 25 |
+
local_parts = [int(x) for x in settings.version['local_version'].split(".")]
|
| 26 |
+
remote_parts = [int(x) for x in settings.version['remote_version'].split(".")]
|
| 27 |
|
| 28 |
# 确保两个列表长度相同
|
| 29 |
while len(local_parts) < len(remote_parts):
|
|
|
|
| 32 |
remote_parts.append(0)
|
| 33 |
|
| 34 |
# 比较版本号
|
| 35 |
+
settings.version['has_update'] = False
|
| 36 |
for i in range(len(local_parts)):
|
| 37 |
if remote_parts[i] > local_parts[i]:
|
| 38 |
+
settings.version['has_update'] = True
|
| 39 |
break
|
| 40 |
elif remote_parts[i] < local_parts[i]:
|
| 41 |
break
|
| 42 |
|
| 43 |
+
log('info', f"版本检查: 本地版本 {settings.version['local_version']}, 远程版本 {settings.version['remote_version']}, 有更新: {settings.version['has_update']}")
|
| 44 |
else:
|
| 45 |
log('warning', f"无法获取远程版本信息,HTTP状态码: {response.status_code}")
|
| 46 |
except Exception as e:
|
| 47 |
log('error', f"版本检查失败: {str(e)}")
|
| 48 |
|
| 49 |
+
return settings.version['has_update']
|
readme.md
CHANGED
|
@@ -4,13 +4,22 @@
|
|
| 4 |
|
| 5 |
这是一个基于 FastAPI 构建的 Gemini API 代理,旨在提供一个简单、安全且可配置的方式来访问 Google 的 Gemini 模型。适用于在 Hugging Face Spaces 上部署,并支持openai api格式的工具集成。
|
| 6 |
### 使用文档
|
| 7 |
-
- [使用文档](
|
| 8 |
### 更新日志
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
* v0.0.4
|
| 10 |
* 修改版本更新逻辑,现在为每四小时检查一次版本更新
|
| 11 |
* 前端界面所有数据数据实现动态更新
|
| 12 |
-
* 新增
|
| 13 |
* 在环境变量中新增`API_KEY_DAILY_LIMIT`,为单api 24小时最大使用次数,默认值为25
|
|
|
|
| 14 |
|
| 15 |
* v0.0.3beta
|
| 16 |
* 完善了客户端断开连接的处理逻辑(感谢[@warming-afternoon](https://github.com/warming-afternoon))
|
|
|
|
| 4 |
|
| 5 |
这是一个基于 FastAPI 构建的 Gemini API 代理,旨在提供一个简单、安全且可配置的方式来访问 Google 的 Gemini 模型。适用于在 Hugging Face Spaces 上部署,并支持openai api格式的工具集成。
|
| 6 |
### 使用文档
|
| 7 |
+
- [使用文档](./wiki/wiki.md)
|
| 8 |
### 更新日志
|
| 9 |
+
* v0.0.5beta
|
| 10 |
+
* 新增“**伪装信息**功能,默认开启,可在转发消息中添加随机字符串伪装消息,防止被检测
|
| 11 |
+
* 修复若干bug
|
| 12 |
+
* 为前端界面新增**功能配置**栏目,可检查功能是否开启
|
| 13 |
+
* 北京时间16点自动重置统计数据
|
| 14 |
+
* 在环境变量中新增`RANDOM_STRING`,是否启用伪装信息,默认值为true
|
| 15 |
+
* 在环境变量中新增`RANDOM_STRING_LENGTH`,伪装信息长度,默认为20
|
| 16 |
+
* 为git用户提供单独的`Dockerfile_git`
|
| 17 |
* v0.0.4
|
| 18 |
* 修改版本更新逻辑,现在为每四小时检查一次版本更新
|
| 19 |
* 前端界面所有数据数据实现动态更新
|
| 20 |
+
* 新增**单api使用次数统计**,在原API调用统计下方新增可折叠的单api使用次数统计,同时提供进度条查看剩余使用次数
|
| 21 |
* 在环境变量中新增`API_KEY_DAILY_LIMIT`,为单api 24小时最大使用次数,默认值为25
|
| 22 |
+
* 在环境变量中新增`BLOCKED_MODELS`,为需要屏蔽的模型名称,多个模型用英文逗号分隔
|
| 23 |
|
| 24 |
* v0.0.3beta
|
| 25 |
* 完善了客户端断开连接的处理逻辑(感谢[@warming-afternoon](https://github.com/warming-afternoon))
|
requirements.txt
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
-
fastapi
|
| 2 |
-
uvicorn
|
| 3 |
-
httpx
|
| 4 |
-
python-dotenv
|
| 5 |
-
requests
|
| 6 |
-
apscheduler
|
| 7 |
jinja2
|
|
|
|
| 1 |
+
fastapi
|
| 2 |
+
uvicorn
|
| 3 |
+
httpx
|
| 4 |
+
python-dotenv
|
| 5 |
+
requests
|
| 6 |
+
apscheduler
|
| 7 |
jinja2
|
version.txt
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
version=0.0.5
|
|
|
|
| 1 |
+
version=0.0.5
|
wiki/Termux.md
ADDED
|
@@ -0,0 +1,84 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Termux 安装与配置 Hajimi 项目教程 由[@天命不又](https://github.com/tmby)编写
|
| 2 |
+
|
| 3 |
+
## 一、切换 Termux 清华源(有“魔法”的可跳过)
|
| 4 |
+
|
| 5 |
+
```bash
|
| 6 |
+
sed -i 's@^\(deb.*stable main\)$@#\1\ndeb https://mirrors.tuna.tsinghua.edu.cn/termux/apt/termux-main stable main@' $PREFIX/etc/apt/sources.list
|
| 7 |
+
```
|
| 8 |
+
|
| 9 |
+
## 二、安装必要依赖
|
| 10 |
+
|
| 11 |
+
```bash
|
| 12 |
+
apt update && apt --yes upgrade && apt --yes install git python rust
|
| 13 |
+
```
|
| 14 |
+
|
| 15 |
+
## 三、配置 pip 使用清华源(有“魔法”的可跳过)
|
| 16 |
+
|
| 17 |
+
```bash
|
| 18 |
+
pip config set global.index-url https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple
|
| 19 |
+
```
|
| 20 |
+
|
| 21 |
+
## 四、配置 Rust 使用清华源(有“魔法”的可跳过)
|
| 22 |
+
|
| 23 |
+
```bash
|
| 24 |
+
export CARGO_REGISTRIES_CRATES_IO_INDEX="https://mirrors.tuna.tsinghua.edu.cn/git/crates.io-index.git"
|
| 25 |
+
```
|
| 26 |
+
|
| 27 |
+
## 五、克隆项目源码(有“魔法”的可去除加速链接)
|
| 28 |
+
|
| 29 |
+
```bash
|
| 30 |
+
git clone https://github.boki.moe/https://github.com/wyeeeee/hajimi
|
| 31 |
+
```
|
| 32 |
+
|
| 33 |
+
## 六、进入项目目录
|
| 34 |
+
|
| 35 |
+
```bash
|
| 36 |
+
cd hajimi
|
| 37 |
+
```
|
| 38 |
+
|
| 39 |
+
## 七、安装 Python 依赖
|
| 40 |
+
|
| 41 |
+
```bash
|
| 42 |
+
pip install tzdata -r requirements.txt
|
| 43 |
+
```
|
| 44 |
+
|
| 45 |
+
> **提示**:安装和构建过程会比较慢,请耐心等待。
|
| 46 |
+
|
| 47 |
+
---
|
| 48 |
+
|
| 49 |
+
## 八、配置 Termux Widget 启动脚本
|
| 50 |
+
|
| 51 |
+
### 1. 安装 Termux Widget 插件
|
| 52 |
+
项目地址:[termux-widget](https://github.com/termux/termux-widget)
|
| 53 |
+
|
| 54 |
+
### 2. 使用 MT 管理器:
|
| 55 |
+
- 打开 Termux 的根目录
|
| 56 |
+
- 新建 `.shortcuts` 文件夹
|
| 57 |
+
- 在其中创建一个空白脚本文件(例如 `哈基米启动`),内容如下:
|
| 58 |
+
|
| 59 |
+
```bash
|
| 60 |
+
#!/data/data/com.termux/files/usr/bin/bash
|
| 61 |
+
|
| 62 |
+
HAJIMI_PATH=/data/data/com.termux/files/home/hajimi
|
| 63 |
+
export GEMINI_API_KEYS="key1,key2,key3"
|
| 64 |
+
|
| 65 |
+
termux-wake-lock
|
| 66 |
+
cd $HAJIMI_PATH
|
| 67 |
+
uvicorn app.main:app --reload --host 127.0.0.1 --port 7860
|
| 68 |
+
```
|
| 69 |
+
|
| 70 |
+
### 3. 添加桌面小部件
|
| 71 |
+
- 在桌面添加 Termux Widget 快捷方式
|
| 72 |
+
- 点击刚才创建的脚本,即可一键启动 Hajimi 服务
|
| 73 |
+
|
| 74 |
+
---
|
| 75 |
+
|
| 76 |
+
## 九、访问前端页面
|
| 77 |
+
|
| 78 |
+
启动服务后,在浏览器中访问以下地址查看前端界面:
|
| 79 |
+
|
| 80 |
+
```
|
| 81 |
+
http://127.0.0.1:7860
|
| 82 |
+
```
|
| 83 |
+
|
| 84 |
+
> **注意**:此地址只能在当前设备本地访问,如需远程访问请进行端口转发或内网穿透设置。
|
wiki/docker.md
ADDED
|
@@ -0,0 +1,91 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# 使用 Docker 部署 Gemini 轮询魔改版教程 由 **北极星星** 编写
|
| 2 |
+
|
| 3 |
+
## 一、准备工作
|
| 4 |
+
|
| 5 |
+
### 1.1 下载项目
|
| 6 |
+
|
| 7 |
+
从 releases 下载最新版本,并解压到任意目录。
|
| 8 |
+
|
| 9 |
+
### 1.2 配置环境变量
|
| 10 |
+
|
| 11 |
+
在项目根目录下创建 `.env` 文件,配置必要环境变量,示例如下:
|
| 12 |
+
|
| 13 |
+
```
|
| 14 |
+
GEMINI_API_KEYS=key1,key2,key3
|
| 15 |
+
PASSWORD=your_password
|
| 16 |
+
TZ=Asia/Shanghai
|
| 17 |
+
```
|
| 18 |
+
|
| 19 |
+
按需修改上述值,注意⚠️key必须使用英文逗号间隔。
|
| 20 |
+
|
| 21 |
+
## 二、构建并运行 Docker 容器
|
| 22 |
+
|
| 23 |
+
### 2.1 构建镜像
|
| 24 |
+
|
| 25 |
+
在项目根目录打开终端,执行命令构建 Docker 镜像:
|
| 26 |
+
|
| 27 |
+
```bash
|
| 28 |
+
cd 项目文件夹完整路径(例如:cd /volume3/docker/hagemi)
|
| 29 |
+
docker build -t hajimi-app .
|
| 30 |
+
```
|
| 31 |
+
|
| 32 |
+
此过程可能需一些时间,取决于网络和系统性能。
|
| 33 |
+
|
| 34 |
+
### 2.2 运行容器
|
| 35 |
+
|
| 36 |
+
镜像构建完成后,执行命令启动 Docker 容器,如端口被占用需修改左侧端口号:
|
| 37 |
+
|
| 38 |
+
```bash
|
| 39 |
+
docker run -d -p 7860:7860 --env-file .env hajimi-app
|
| 40 |
+
```
|
| 41 |
+
|
| 42 |
+
## 三、验证部署
|
| 43 |
+
|
| 44 |
+
### 3.1 检查容器状态
|
| 45 |
+
|
| 46 |
+
打开 docker 查看 hajimi-app 容器运行状态,确认正常启动。
|
| 47 |
+
|
| 48 |
+
### 3.2 访问应用
|
| 49 |
+
|
| 50 |
+
打开浏览器,访问 http://localhost:7860,若看到应用界面,则部署成功。
|
| 51 |
+
|
| 52 |
+
API 地址:http://localhost:7860/v1
|
| 53 |
+
key:PASSWORD=your_password
|
| 54 |
+
|
| 55 |
+
## 四、容器更新
|
| 56 |
+
|
| 57 |
+
### 4.1 更新脚本
|
| 58 |
+
|
| 59 |
+
将下面 gemini_docker_update.sh 脚本按需修改保存为一个 .sh 文件:
|
| 60 |
+
|
| 61 |
+
```bash
|
| 62 |
+
# 停止容器
|
| 63 |
+
docker stop hajimi-app
|
| 64 |
+
# 删除容器
|
| 65 |
+
docker rm hajimi-app
|
| 66 |
+
# 进入项目所在目录
|
| 67 |
+
cd /volume3/docker/hagemi
|
| 68 |
+
# 使用以下命令拉取最新代码
|
| 69 |
+
git pull origin main
|
| 70 |
+
# 构建新的 Docker 镜像
|
| 71 |
+
docker build -t hajimi-app .
|
| 72 |
+
# 运行新容器
|
| 73 |
+
docker run -d -p 7860:7860 --env-file .env hajimi-app
|
| 74 |
+
# 查看容器状态
|
| 75 |
+
docker ps -a | grep hajimi-app
|
| 76 |
+
```
|
| 77 |
+
|
| 78 |
+
### 4.2 脚本存放位置
|
| 79 |
+
|
| 80 |
+
把 gemini_docker_update.sh 脚本存放在项目根目录,例如项目文件路径是 /volume3/docker/hagemi,便将脚本存放在 /volume3/docker/hagemi。
|
| 81 |
+
|
| 82 |
+
### 4.3 执行更新
|
| 83 |
+
|
| 84 |
+
进入终端输入命令,执行更新脚本:
|
| 85 |
+
|
| 86 |
+
```bash
|
| 87 |
+
cd /volume3/docker/hagemi
|
| 88 |
+
./gemini_docker_update.sh
|
| 89 |
+
```
|
| 90 |
+
|
| 91 |
+
通过以上步骤,即可使用 Docker 成功部署 Gemini 轮询魔改版应用。
|
wiki/wiki.md
CHANGED
|
@@ -7,16 +7,16 @@
|
|
| 7 |
|
| 8 |
### 1.2 使用huggingface space部署项目
|
| 9 |
- 在[huggingface](https://huggingface.co)注册账号。
|
| 10 |
-
- 注册完成后,进入[spaces](https://huggingface.co/spaces)如图所示,点击new spaces。<br>是否启用假流式传输,默认为true。
|
| 29 |
- `API_KEY_DAILY_LIMIT`: 单api 24小时最大使用次数,默认值为25
|
| 30 |
- `BLOCKED_MODELS`,(可选)需要屏蔽的模型名称,多个模型用英文逗号分隔
|
| 31 |
-
|
|
|
|
| 32 |
### 1.4 在酒馆中使用
|
| 33 |
在酒馆api连接配置中,选择兼容openai格式,URL格式为`https://(huggingface用户名)-(huggingface项目名).hf.space/v1`(注意为https)。自定义 API 密钥为1.3中配置的`PASSWORD`。
|
| 34 |
|
|
@@ -47,7 +48,15 @@
|
|
| 47 |
2. 设置环境变量(如上所述)。
|
| 48 |
3. 运行:`uvicorn app.main:app --reload --host 0.0.0.0 --port 7860`
|
| 49 |
|
| 50 |
-
###
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 51 |
|
| 52 |
1. 在连接中选择OpenAI
|
| 53 |
2. 在API Base URL中填入`https://(huggingface用户名)-(huggingface项目名).hf.space/v1`
|
|
|
|
| 7 |
|
| 8 |
### 1.2 使用huggingface space部署项目
|
| 9 |
- 在[huggingface](https://huggingface.co)注册账号。
|
| 10 |
+
- 注册完成后,进入[spaces](https://huggingface.co/spaces)如图所示,点击new spaces。<br>
|
| 11 |
+
- 如图所示,填入创建选项,注意space name不要使用符号,填写完成后,点击create space<br>
|
| 12 |
+
- 如图所示,选择upload files<br>
|
| 13 |
+
- 如图所示,将1.1下载并解压的文件全部拖入,上传完成后点击Commit changes to main<br>
|
| 14 |
|
| 15 |
### 1.2 配置环境变量
|
| 16 |
+
- 如图所示,进入settings界面<br>
|
| 17 |
+
- 如图所示,在settings界面中找到Variables and secrets,点击new secrets<br>
|
| 18 |
+
- 添加环境变量,如图所示为添加GEMINI_API_KEYS环境变量,在value中填入具体apikey<br>
|
| 19 |
+
- 等待项目部署完成,app界面显示如图界面,即完成<br>
|
| 20 |
|
| 21 |
### 1.3 环境变量说明
|
| 22 |
#### 重要环境变量
|
|
|
|
| 28 |
- `FAKE_STREAMING`:(可选)是否启用假流式传输,默认为true。
|
| 29 |
- `API_KEY_DAILY_LIMIT`: 单api 24小时最大使用次数,默认值为25
|
| 30 |
- `BLOCKED_MODELS`,(可选)需要屏蔽的模型名称,多个模型用英文逗号分隔
|
| 31 |
+
- `RANDOM_STRING`,是否启用伪装信息,默认值为true
|
| 32 |
+
- `RANDOM_STRING_LENGTH`,伪装信息长度,默认为20
|
| 33 |
### 1.4 在酒馆中使用
|
| 34 |
在酒馆api连接配置中,选择兼容openai格式,URL格式为`https://(huggingface用户名)-(huggingface项目名).hf.space/v1`(注意为https)。自定义 API 密钥为1.3中配置的`PASSWORD`。
|
| 35 |
|
|
|
|
| 48 |
2. 设置环境变量(如上所述)。
|
| 49 |
3. 运行:`uvicorn app.main:app --reload --host 0.0.0.0 --port 7860`
|
| 50 |
|
| 51 |
+
### 1.8 💻 手机本地运行(可选,安装耗时长):
|
| 52 |
+
|
| 53 |
+
[Termux 安装与配置 Hajimi 项目教程 由**天命不又**编写](./Termux.md)
|
| 54 |
+
|
| 55 |
+
### 1.9 💻 docker部署运行:
|
| 56 |
+
|
| 57 |
+
[docker 安装与配置 Hajimi 项目教程 由 **北极星星** 编写](./docker.md)
|
| 58 |
+
|
| 59 |
+
### 1.10 🔌 接入其他服务
|
| 60 |
|
| 61 |
1. 在连接中选择OpenAI
|
| 62 |
2. 在API Base URL中填入`https://(huggingface用户名)-(huggingface项目名).hf.space/v1`
|