File size: 2,168 Bytes
4ceec83
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
from fastapi import FastAPI
from fastapi.responses import StreamingResponse
from pydantic import BaseModel
from transformers import (
    AutoTokenizer,
    AutoModelForCausalLM,
    BitsAndBytesConfig,
    TextIteratorStreamer
)
import torch
import threading

app = FastAPI()

MODEL_NAME = "Qwen/Qwen2.5-Coder-7B"

# ---- Quantization config (CPU safe) ----
bnb_config = BitsAndBytesConfig(
    load_in_4bit=True,
    bnb_4bit_compute_dtype=torch.float32,
    bnb_4bit_use_double_quant=True,
    bnb_4bit_quant_type="nf4"
)

tokenizer = AutoTokenizer.from_pretrained(
    MODEL_NAME,
    trust_remote_code=True
)

model = AutoModelForCausalLM.from_pretrained(
    MODEL_NAME,
    device_map="cpu",
    quantization_config=bnb_config,
    trust_remote_code=True
)

class Prompt(BaseModel):
    message: str

# -------------------------------------------------
# ✅ NORMAL CHAT (UNCHANGED)
# -------------------------------------------------
@app.post("/chat")
def chat(prompt: Prompt):
    inputs = tokenizer(prompt.message, return_tensors="pt")

    outputs = model.generate(
        **inputs,
        max_new_tokens=200,
        temperature=0.7,
        do_sample=True
    )

    response = tokenizer.decode(outputs[0], skip_special_tokens=True)
    return {"response": response}


# -------------------------------------------------
# 🚀 STREAMING CHAT (CHATGPT-LIKE)
# -------------------------------------------------
@app.post("/chat-stream")
def chat_stream(prompt: Prompt):

    inputs = tokenizer(prompt.message, return_tensors="pt")

    streamer = TextIteratorStreamer(
        tokenizer,
        skip_special_tokens=True,
        skip_prompt=True
    )

    generation_kwargs = dict(
        **inputs,
        streamer=streamer,
        max_new_tokens=200,
        temperature=0.7,
        do_sample=True
    )

    # Run generation in background thread
    thread = threading.Thread(
        target=model.generate,
        kwargs=generation_kwargs
    )
    thread.start()

    def token_generator():
        for token in streamer:
            yield token

    return StreamingResponse(
        token_generator(),
        media_type="text/plain"
    )