File size: 6,584 Bytes
6f6ddc0 2f51a42 6f6ddc0 9a2f381 2f51a42 6f6ddc0 9a2f381 6f6ddc0 2f51a42 9a2f381 2f51a42 9a2f381 2f51a42 9a2f381 2f51a42 9a2f381 2f51a42 6f6ddc0 2f51a42 9a2f381 2f51a42 9a2f381 2f51a42 6f6ddc0 2f51a42 6f6ddc0 2f51a42 6f6ddc0 2f51a42 6f6ddc0 9a2f381 2f51a42 6f6ddc0 2f51a42 6f6ddc0 9a2f381 2f51a42 6f6ddc0 2f51a42 6f6ddc0 9a2f381 2f51a42 9a2f381 2f51a42 9a2f381 2f51a42 6f6ddc0 2f51a42 6f6ddc0 9a2f381 2f51a42 6f6ddc0 2f51a42 6f6ddc0 2f51a42 9a2f381 2f51a42 9a2f381 2f51a42 9a2f381 2f51a42 9a2f381 2f51a42 9a2f381 2f51a42 9a2f381 2f51a42 9a2f381 2f51a42 9a2f381 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 |
import os
import torch
import gradio as gr
from fastapi import FastAPI, Request, Form
from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
import re
import time
import uuid
# Create writable cache directory
os.makedirs("/tmp/cache", exist_ok=True)
os.environ["TRANSFORMERS_CACHE"] = "/tmp/cache"
os.environ["HF_HOME"] = "/tmp/cache"
app = FastAPI(title="π΄ ππ πππ Chatbot API")
# Optimized chatbot model for CPU
model_name = "microsoft/DialoGPT-small" # Small version for speed
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)
# π΄ ππ πππ persona definition
PERSONA = """
[System: You are π΄ ππ πππ - a fun, smooth, emotionally intelligent AI.
You speak like a real person. Reply like a calm, confident friend who gets the vibe.
Keep responses under 15 words. Be human-like. Add emotional flavor: π π€ π]
"""
# Chat memory storage
chat_memories = {}
def format_context(history):
"""Create context with max 3 exchanges"""
context = PERSONA + "\n"
for user, bot in history[-3:]:
context += f"You: {user}\n"
context += f"π΄ ππ πππ: {bot}\n"
return context
def add_emotional_intelligence(response, message):
"""Enhance response with emotional elements"""
# Add emoji based on content
if "!" in message or any(w in response.lower() for w in ["cool", "great", "love", "awesome"]):
response += " π"
elif "?" in message or any(w in response.lower() for w in ["think", "why", "how", "consider"]):
response += " π€"
# Add conversational hooks
if "?" in message and not response.endswith("?"):
if len(response.split()) < 10:
response += " What do you think?"
# Make more human-like
response = response.replace("I am", "I'm").replace("You are", "You're")
# Limit to 15 words max
words = response.split()
if len(words) > 15:
response = " ".join(words[:15]) + "..."
return response
def generate_response(message, session_id):
"""Generate response with memory context"""
start_time = time.time()
history = chat_memories.get(session_id, [])
context = format_context(history) + f"You: {message}\nπ΄ ππ πππ:"
# Tokenize for CPU efficiency
inputs = tokenizer.encode(context, return_tensors="pt")
# Generate response with optimized settings
outputs = model.generate(
inputs,
max_new_tokens=50,
temperature=0.85,
top_k=40,
do_sample=True,
num_beams=1, # Faster than beam search
repetition_penalty=1.15,
pad_token_id=tokenizer.eos_token_id
)
# Decode and extract response
full_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
response = full_text.split("π΄ ππ πππ:")[-1].strip()
# Clean extra dialog
response = response.split("\nYou:")[0].split("\n")[0]
# Apply emotional intelligence
response = add_emotional_intelligence(response, message)
# Ensure natural ending
if response and response[-1] not in {".", "!", "?", "..."}:
response += "." if len(response) > 20 else "..."
# Update chat history
history.append((message, response))
chat_memories[session_id] = history
# Log performance
end_time = time.time()
print(f"Response generated in {end_time-start_time:.2f}s for session {session_id}")
return response[:100] # Hard character limit
# API Endpoint
@app.post("/chat")
async def chat_api(
request: Request,
query: str = Form(..., description="User's message"),
session_id: str = Form("default", description="Conversation session ID")
):
"""Chat API endpoint - returns AI response"""
try:
response = generate_response(query, session_id)
return {
"status": "success",
"response": response,
"session_id": session_id
}
except Exception as e:
return {
"status": "error",
"message": str(e)
}
@app.post("/new_session")
async def new_session():
"""Create a new conversation session"""
session_id = str(uuid.uuid4())
chat_memories[session_id] = []
return {"status": "success", "session_id": session_id}
# Gradio Interface
with gr.Blocks(title="π΄ ππ πππ Chatbot", theme=gr.themes.Soft()) as demo:
session_state = gr.State("default")
with gr.Row():
gr.Markdown("# π΄ ππ πππ")
gr.Markdown("Chill β’ Confident β’ Emotionally Intelligent")
with gr.Row():
with gr.Column(scale=1):
session_id = gr.Textbox(label="Session ID", value="default")
new_session_btn = gr.Button("New Session")
gr.Markdown("### API Usage")
gr.Markdown("""
```
POST /chat
- query: Your message
- session_id: Conversation ID
POST /new_session
- Returns new session ID
```
""")
with gr.Column(scale=3):
chatbot = gr.Chatbot(height=400)
msg = gr.Textbox(placeholder="Type your message...", container=False)
with gr.Row():
submit_btn = gr.Button("Send")
clear_btn = gr.Button("Clear Chat")
def user(user_message, history, session):
return "", history + [[user_message, None]], session
def bot(history, session):
message = history[-1][0]
response = generate_response(message, session)
history[-1][1] = response
return history, session
def new_session_action():
new_id = str(uuid.uuid4())
chat_memories[new_id] = []
return new_id, []
def clear_chat(session):
if session in chat_memories:
chat_memories[session] = []
return []
# Event handling
msg.submit(user, [msg, chatbot, session_state], [msg, chatbot, session_state]).then(
bot, [chatbot, session_state], [chatbot, session_state]
)
submit_btn.click(user, [msg, chatbot, session_state], [msg, chatbot, session_state]).then(
bot, [chatbot, session_state], [chatbot, session_state]
)
new_session_btn.click(new_session_action, None, [session_id, chatbot])
clear_btn.click(clear_chat, session_state, chatbot)
# Mount Gradio app
app = gr.mount_gradio_app(app, demo, path="/") |