|
|
import os |
|
|
import streamlit as st |
|
|
import torch |
|
|
from groq import Groq |
|
|
from transformers import pipeline, AutoModelForCausalLM, AutoTokenizer |
|
|
|
|
|
|
|
|
st.set_page_config(page_title="AI Study Assistant", page_icon="π€", layout="wide") |
|
|
|
|
|
|
|
|
GROQ_API_KEY = "your_groq_api_key_here" |
|
|
os.environ["GROQ_API_KEY"] = GROQ_API_KEY |
|
|
|
|
|
|
|
|
client = Groq(api_key=GROQ_API_KEY) |
|
|
|
|
|
|
|
|
try: |
|
|
import accelerate |
|
|
except ImportError: |
|
|
st.error("β οΈ `accelerate` library is required. Install it with: `pip install accelerate`") |
|
|
|
|
|
|
|
|
MODEL_NAME = "deepseek-ai/DeepSeek-R1" |
|
|
|
|
|
try: |
|
|
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, trust_remote_code=True) |
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
MODEL_NAME, |
|
|
trust_remote_code=True, |
|
|
torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32, |
|
|
device_map="auto" if torch.cuda.is_available() else None |
|
|
) |
|
|
|
|
|
def generate_response_hf(user_message): |
|
|
inputs = tokenizer(user_message, return_tensors="pt").to(model.device) |
|
|
outputs = model.generate(**inputs, max_length=200) |
|
|
return tokenizer.decode(outputs[0], skip_special_tokens=True) |
|
|
|
|
|
except Exception as e: |
|
|
st.error(f"β Error loading DeepSeek-R1: {str(e)}") |
|
|
generate_response_hf = lambda x: "β οΈ Error: Model not loaded." |
|
|
|
|
|
|
|
|
st.title("π Subject-specific AI Chatbot") |
|
|
st.write("Hello! I'm your AI Study Assistant. You can ask me any questions related to your subjects, and I'll try to help.") |
|
|
|
|
|
|
|
|
st.sidebar.header("βοΈ Settings") |
|
|
chat_model = st.sidebar.radio("Choose AI Model:", ["Groq API", "DeepSeek R1 (Hugging Face)"]) |
|
|
|
|
|
|
|
|
if 'conversation_history' not in st.session_state: |
|
|
st.session_state.conversation_history = [] |
|
|
|
|
|
|
|
|
subjects = ["Chemistry", "Computer", "English", "Islamiat", "Mathematics", "Physics", "Urdu"] |
|
|
|
|
|
def generate_chatbot_response(user_message): |
|
|
related_subject = next((subject for subject in subjects if subject.lower() in user_message.lower()), None) |
|
|
|
|
|
if "kisne banaya" in user_message.lower() or "who created you" in user_message.lower(): |
|
|
return "I was created by Abdul Basit π" |
|
|
|
|
|
prompt = f"You are a helpful AI chatbot for studying {related_subject if related_subject else 'general knowledge'}. The user is asking: {user_message}. Provide a detailed, helpful response." |
|
|
|
|
|
if chat_model == "Groq API": |
|
|
chat_completion = client.chat.completions.create( |
|
|
messages=[{"role": "user", "content": prompt}], |
|
|
model="deepseek-chat" |
|
|
) |
|
|
return chat_completion.choices[0].message.content |
|
|
else: |
|
|
return generate_response_hf(prompt) |
|
|
|
|
|
|
|
|
st.markdown("### π¬ Chat with me") |
|
|
user_input = st.chat_input("Ask me a subject-related question:") |
|
|
|
|
|
if user_input: |
|
|
chatbot_response = generate_chatbot_response(user_input) |
|
|
st.session_state.conversation_history.append(("User: " + user_input, "Chatbot: " + chatbot_response)) |
|
|
|
|
|
|
|
|
st.markdown("---") |
|
|
st.markdown("### π¨οΈ Chat History") |
|
|
for question, answer in st.session_state.conversation_history: |
|
|
st.write(f"<div class='chat-bubble'><b>{question}</b></div>", unsafe_allow_html=True) |
|
|
st.write(f"<div class='chat-bubble'>{answer}</div>", unsafe_allow_html=True) |
|
|
|