File size: 17,702 Bytes
c72f0db
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
import base64
import json
import os
import shutil
import subprocess
import tempfile
import time
from typing import Optional

from curl_cffi import AsyncSession, Response
from fastapi import FastAPI, Depends, HTTPException
from fastapi.security import HTTPBearer, HTTPAuthorizationCredentials
from loguru import logger
from starlette.middleware.cors import CORSMiddleware

from app.config import SCRIPT_URL, FP, API_KEY, MODELS, SYSTEM_PROMPT_INJECT, TIMEOUT, PROXY, USER_PROMPT_INJECT, \
    X_IS_HUMAN_SERVER_URL, ENABLE_FUNCTION_CALLING, TRUNCATION_CONTINUE, TRUNCATION_MAX_RETRIES, EMPTY_RETRY_MAX_RETRIES
from app.errors import CursorWebError
from app.models import ChatCompletionRequest, Message, ModelsResponse, Model, Usage, OpenAIMessageContent, ToolCall
from app.utils import error_wrapper, to_async, generate_random_string, non_stream_chat_completion, \
    stream_chat_completion, safe_stream_wrapper, match_tool_name, truncation_continue_wrapper, empty_retry_wrapper

main_code = open('./jscode/main.js', 'r', encoding='utf-8').read()
env_code = open('./jscode/env.js', 'r', encoding='utf-8').read()
app = FastAPI()

security = HTTPBearer()

app.add_middleware(
    CORSMiddleware,
    allow_origins=["*"],
    allow_credentials=True,
    allow_methods=["*"],
    allow_headers=["*"],
)


@app.post("/v1/chat/completions")
async def chat_completions(
        request: ChatCompletionRequest,
        credentials: HTTPAuthorizationCredentials = Depends(security),
):
    """处理聊天完成请求"""

    if credentials.credentials != API_KEY:
        raise HTTPException(401, 'api key 错误')

    # 空回复重试包装器(始终启用)
    chat_func = lambda req: empty_retry_wrapper(cursor_chat, req, max_retries=EMPTY_RETRY_MAX_RETRIES)

    if TRUNCATION_CONTINUE:
        chat_generator = truncation_continue_wrapper(chat_func, request, max_retries=TRUNCATION_MAX_RETRIES)
    else:
        chat_generator = chat_func(request)

    # async for c in chat_generator:
    #     logger.debug(c)

    if request.stream:
        return await error_wrapper(safe_stream_wrapper, stream_chat_completion, request, chat_generator)
    else:
        return await error_wrapper(non_stream_chat_completion, request, chat_generator)


@app.get("/v1/models")
async def list_models(credentials: HTTPAuthorizationCredentials = Depends(security)):
    models = MODELS.split(',')
    model_list = []

    for model_id in models:
        model_list.append(
            Model(
                id=model_id,  # 使用model name作为对外的id
                object="model",
                created=int(time.time()),
                owned_by='',
            )
        )

    return ModelsResponse(object="list", data=model_list)


def inject_system_prompt(list_openai_message: list[Message], inject_prompt: str):
    # 查找是否存在system角色的消息
    system_message_found = False

    for message in list_openai_message:
        if message.role == "system":
            system_message_found = True
            # 处理content字段,需要考虑不同的数据类型
            if message.content is None:
                message.content = inject_prompt
            elif isinstance(message.content, str):
                message.content += f'\n{inject_prompt}'
            elif isinstance(message.content, list):
                # 如果content是列表,需要找到text类型的内容进行追加
                # 或者添加一个新的text内容项
                text_content_found = False
                for content_item in message.content:
                    if content_item.type == "text" and content_item.text:
                        content_item.text += f'\n{inject_prompt}'
                        text_content_found = True
                        break

                # 如果没有找到text内容,添加一个新的text内容项
                if not text_content_found:
                    new_text_content = OpenAIMessageContent(
                        type="text",
                        text=inject_prompt
                        , image_url=None)
                    message.content.append(new_text_content)
            break  # 找到第一个system消息后就退出循环

    # 如果没有找到system消息,在列表开头插入一个新的system消息
    if not system_message_found:
        system_message = Message(
            role="system",
            content=inject_prompt
            , tool_call_id=None, tool_calls=None)
        list_openai_message.insert(0, system_message)


def collect_developer_messages(list_openai_message: list[Message]) -> str:
    collected_contents = []

    # 从后往前遍历,避免删除元素时索引变化的问题
    for i in range(len(list_openai_message) - 1, -1, -1):
        message = list_openai_message[i]

        if message.role == "developer":
            # 提取消息内容
            content_text = ""

            if message.content is None:
                content_text = ""
            elif isinstance(message.content, str):
                content_text = message.content
            elif isinstance(message.content, list):
                # 如果content是列表,提取所有text类型的内容
                text_parts = []
                for content_item in message.content:
                    if content_item.type == "text" and content_item.text:
                        text_parts.append(content_item.text)
                content_text = " ".join(text_parts)  # 多个text内容用空格连接

            # 将内容添加到收集列表的开头,保持原始顺序
            collected_contents.insert(0, content_text)

            # 删除该消息
            list_openai_message.pop(i)

    # 将收集到的内容按\n拼接并返回
    return "\n".join(collected_contents)


def to_cursor_messages(request: ChatCompletionRequest):
    list_openai_message: list[Message] = request.messages
    if list_openai_message is None:
        list_openai_message = []

    developer_messages = collect_developer_messages(list_openai_message)
    inject_system_prompt(list_openai_message, developer_messages)

    if ENABLE_FUNCTION_CALLING:
        if request.tools:
            tools = [tool.model_dump_json() for tool in request.tools]
            inject_system_prompt(list_openai_message, "你可用的工具: " + json.dumps(tools))
            inject_system_prompt(list_openai_message, "不允许使用tool_calls: xxxx调用工具,请使用原生的工具调用方法")

    if SYSTEM_PROMPT_INJECT:
        inject_system_prompt(list_openai_message, SYSTEM_PROMPT_INJECT)
    if USER_PROMPT_INJECT:
        list_openai_message.append(Message(role='user', content=USER_PROMPT_INJECT, tool_calls=None, tool_call_id=None))

    result: list[dict[str, str]] = []

    for m in list_openai_message:
        if not m:
            continue

        if ENABLE_FUNCTION_CALLING:
            if m.tool_calls:
                message = {
                    'role': m.role,
                    'parts': [{
                        'type': 'text',
                        'text': f"tool_calls: {json.dumps(m.tool_calls, ensure_ascii=False)}"
                    }]
                }
                result.append(message)
                continue

            if m.tool_call_id:
                message = {
                    'role': 'user',
                    'parts': [{
                        'type': 'text',
                        'text': f"{m.role}: tool_call_id: {m.tool_call_id} {m.content}"
                    }]
                }
                result.append(message)
                continue

        text = ''
        if isinstance(m.content, str):
            text = m.content
        else:
            for content in m.content:
                if not content.text:
                    continue
                text = text + content.text
        message = {
            'role': m.role,
            'parts': [{
                'type': 'text',
                'text': text
            }]
        }
        result.append(message)

    if result[0]['role'] == 'system' and not result[0]['parts'][0]['text']:
        result.pop(0)

    return result


def parse_sse_line(line: str) -> Optional[str]:
    """解析SSE数据行"""
    line = line.strip()
    if line.startswith("data: "):
        return line[6:]  # 去掉 'data: ' 前缀
    return None


async def cursor_chat(request: ChatCompletionRequest):
    # 提取可用工具名列表,用于后续修正
    available_tool_names = []
    if ENABLE_FUNCTION_CALLING and request.tools:
        available_tool_names = [tool.function.name for tool in request.tools]

    json_data = {
        "context": [

        ],
        "model": request.model,
        "id": generate_random_string(16),
        "messages": to_cursor_messages(request),
        "trigger": "submit-message"
    }
    async with AsyncSession(impersonate='chrome', timeout=TIMEOUT, proxy=PROXY) as session:
        if X_IS_HUMAN_SERVER_URL:
            x_is_human = await get_x_is_human_server(session)
        else:
            x_is_human = await get_x_is_human(session)
        logger.debug(x_is_human)
        headers = {
            'User-Agent': FP.get("userAgent"),
            # 'Accept-Encoding': 'gzip, deflate, br, zstd',
            'Content-Type': 'application/json',
            'sec-ch-ua-platform': '"Windows"',
            'x-path': '/api/chat',
            'sec-ch-ua': '"Chromium";v="140", "Not=A?Brand";v="24", "Google Chrome";v="140"',
            'x-method': 'POST',
            'sec-ch-ua-bitness': '"64"',
            'sec-ch-ua-mobile': '?0',
            'sec-ch-ua-arch': '"x86"',
            'x-is-human': x_is_human,
            'sec-ch-ua-platform-version': '"19.0.0"',
            'origin': 'https://cursor.com',
            'sec-fetch-site': 'same-origin',
            'sec-fetch-mode': 'cors',
            'sec-fetch-dest': 'empty',
            'referer': 'https://cursor.com/en-US/learn/how-ai-models-work',
            'accept-language': 'zh-CN,zh;q=0.9,en;q=0.8',
            'priority': 'u=1, i',
        }
        logger.debug(json_data)
        async with session.stream("POST", 'https://cursor.com/api/chat', headers=headers, json=json_data,
                                  impersonate='chrome') as response:
            response: Response
            # logger.debug(await response.atext())

            if response.status_code != 200:
                text = await response.atext()
                if 'Attention Required! | Cloudflare' in text:
                    text = 'Cloudflare 403'
                raise CursorWebError(response.status_code, text)
            content_type = response.headers['content-type']
            if 'text/event-stream' not in content_type:
                text = await response.atext()
                raise CursorWebError(response.status_code, "响应非事件流: " + text)
            async for line in response.aiter_lines():
                line = line.decode("utf-8")
                logger.debug(line)
                data = parse_sse_line(line)
                if not data:
                    continue
                if data and data.strip():
                    try:
                        event_data = json.loads(data)
                        if event_data.get('type') == 'error':
                            err_msg = event_data.get('errorText', 'errorText为空')
                            if 'The content field in the Message object at' in err_msg:
                                err_msg = "消息为空,很可能你的消息只包含图片,本接口不支持图片\n" + err_msg
                            raise CursorWebError(response.status_code, err_msg)
                        if event_data.get('type') == 'finish':
                            usage = event_data.get('messageMetadata', {}).get('usage')
                            if not usage:
                                continue
                            yield Usage(prompt_tokens=usage.get('inputTokens'),
                                        completion_tokens=usage.get('outputTokens'),
                                        total_tokens=usage.get('totalTokens'))
                            return
                        if ENABLE_FUNCTION_CALLING:
                            if event_data.get('type') == 'tool-input-error':
                                tool_call_id = event_data.get('toolCallId')
                                tool_name = event_data.get('toolName')
                                tool_input = event_data.get('input')
                                if isinstance(tool_input, str):
                                    tool_input_str = tool_input
                                else:
                                    tool_input_str = json.dumps(tool_input)

                                # 修正工具名称
                                if available_tool_names:
                                    tool_name = match_tool_name(tool_name, available_tool_names)

                                response.close()  # 工具返回了直接掐断
                                yield ToolCall(toolId=tool_call_id, toolInput=tool_input_str, toolName=tool_name)
                                return

                        delta = event_data.get('delta')
                        # logger.debug(delta)
                        if not delta:
                            continue
                        yield delta
                    except json.JSONDecodeError:
                        continue


async def get_x_is_human_server(session: AsyncSession):
    headers = {
        'User-Agent': FP.get("userAgent"),
        # 'Accept-Encoding': 'gzip, deflate, br, zstd',
        'sec-ch-ua-arch': '"x86"',
        'sec-ch-ua-platform': '"Windows"',
        'sec-ch-ua': '"Chromium";v="140", "Not=A?Brand";v="24", "Google Chrome";v="140"',
        'sec-ch-ua-bitness': '"64"',
        'sec-ch-ua-mobile': '?0',
        'sec-ch-ua-platform-version': '"19.0.0"',
        'sec-fetch-site': 'same-origin',
        'sec-fetch-mode': 'no-cors',
        'sec-fetch-dest': 'script',
        'referer': 'https://cursor.com/en-US/learn/how-ai-models-work',
        'accept-language': 'zh-CN,zh;q=0.9,en;q=0.8',
    }

    response = await session.get(SCRIPT_URL,
                                 headers=headers,
                                 impersonate='chrome')
    cursor_js = response.text
    js_b64 = base64.b64encode(cursor_js.encode('utf-8')).decode("utf-8")

    response = await session.post(X_IS_HUMAN_SERVER_URL, json={
        "jscode": js_b64,
        "fp": FP
    })
    try:
        s = response.json().get('s')
    except json.decoder.JSONDecodeError:
        raise CursorWebError(response.status_code, '纯算服务器返回结果错误: ' + response.text)
    if not s:
        raise CursorWebError(response.status_code, '纯算服务器返回结果错误: ' + response.text)

    return response.text


async def get_x_is_human(session: AsyncSession):
    headers = {
        'User-Agent': FP.get("userAgent"),
        # 'Accept-Encoding': 'gzip, deflate, br, zstd',
        'sec-ch-ua-arch': '"x86"',
        'sec-ch-ua-platform': '"Windows"',
        'sec-ch-ua': '"Chromium";v="140", "Not=A?Brand";v="24", "Google Chrome";v="140"',
        'sec-ch-ua-bitness': '"64"',
        'sec-ch-ua-mobile': '?0',
        'sec-ch-ua-platform-version': '"19.0.0"',
        'sec-fetch-site': 'same-origin',
        'sec-fetch-mode': 'no-cors',
        'sec-fetch-dest': 'script',
        'referer': 'https://cursor.com/en-US/learn/how-ai-models-work',
        'accept-language': 'zh-CN,zh;q=0.9,en;q=0.8',
    }

    response = await session.get(SCRIPT_URL,
                                 headers=headers,
                                 impersonate='chrome')
    cursor_js = response.text

    # 替换指纹
    main = (main_code.replace("$$currentScriptSrc$$", SCRIPT_URL)
            .replace("$$UNMASKED_VENDOR_WEBGL$$", FP.get("UNMASKED_VENDOR_WEBGL"))
            .replace("$$UNMASKED_RENDERER_WEBGL$$", FP.get("UNMASKED_RENDERER_WEBGL"))
            .replace("$$userAgent$$", FP.get("userAgent")))

    # 替换代码
    main = main.replace('$$env_jscode$$', env_code)
    main = main.replace("$$cursor_jscode$$", cursor_js)
    return await runjs(main)


@to_async
def runjs(jscode: str) -> str:
    """
    执行 JavaScript 代码并返回标准输出内容。

    Args:
        jscode: 要执行的 JavaScript 代码字符串

    Returns:
        Node.js 程序的标准输出内容

    Raises:
        FileNotFoundError: Node.js 未安装或不在系统 PATH 中
        subprocess.CalledProcessError: Node.js 程序执行失败,异常信息包含 stdout 和 stderr
    """
    temp_dir = tempfile.mkdtemp()
    try:
        js_file_path = os.path.join(temp_dir, "script.js")
        with open(js_file_path, "w", encoding="utf-8") as f:
            f.write(jscode)

        result = subprocess.run(
            ['node', js_file_path],
            capture_output=True,
            text=True,
            encoding="utf-8"
        )

        if result.returncode != 0:
            error_msg = f"Node.js 执行失败 (退出码: {result.returncode})\nSTDOUT:\n{result.stdout}\nSTDERR:\n{result.stderr}"
            logger.error(error_msg)
            raise subprocess.CalledProcessError(result.returncode, ['node', js_file_path], result.stdout, result.stderr)

        return result.stdout.strip()
    finally:
        shutil.rmtree(temp_dir)


if __name__ == "__main__":
    import uvicorn

    uvicorn.run(
        "main:app",
        host="0.0.0.0",
        port=8000,
        reload=False,
        log_level="info",
    )