subramaniyam's picture
Update app.py
0c26562 verified
# Import necessary modules
from langchain.prompts import ChatPromptTemplate # type: ignore
from langchain.llms import Ollama # type: ignore
import streamlit as st # type: ignore
# Streamlit setup
st.title("Subbu Chat Bot")
input_txt = st.text_input("Enter your queries here...")
# Add a dropdown for model selection
model_choice = st.selectbox("Select the model:", ["Llama 3.2", "Llama 3.1", "Code Llama"])
# Define the prompt template
prompt = ChatPromptTemplate.from_messages(
[("system", "You are a helpful AI assistant. Your name is Subbu Assistant."),
("user", "user query: {query}")]
)
# Initialize each model (adjust the model names based on available models)
llm_3_2 = Ollama(model="llama3.2")
llm_3_1 = Ollama(model="llama3.1")
code_llama = Ollama(model="codellama")
# Process input and display the response
if input_txt:
# Select model based on user choice
if model_choice == "Llama 3.2":
response = llm_3_2(prompt.format(query=input_txt))
elif model_choice == "Llama 3.1":
response = llm_3_1(prompt.format(query=input_txt))
elif model_choice == "Code Llama":
response = code_llama(prompt.format(query=input_txt))
# Display the response
st.write(response)