Spaces:
Runtime error
Runtime error
| import streamlit as st | |
| from langchain import PromptTemplate, LLMChain | |
| from langchain.memory import StreamlitChatMessageHistory | |
| from streamlit_chat import message | |
| import numpy as np | |
| from langchain.chains import LLMChain | |
| from langchain.prompts import PromptTemplate | |
| from langchain.memory import ConversationBufferMemory | |
| from langchain.memory.chat_message_histories import StreamlitChatMessageHistory | |
| from streamlit.components.v1 import html | |
| from langchain import HuggingFaceHub | |
| import os | |
| from dotenv import load_dotenv | |
| load_dotenv() | |
| st.set_page_config(page_title="Cheers! Open AI Chat Assistant", layout="wide") | |
| st.subheader("Cheers! Open AI Chat Assistant: Life Enhancing with AI!") | |
| css_file = "main.css" | |
| with open(css_file) as f: | |
| st.markdown("<style>{}</style>".format(f.read()), unsafe_allow_html=True) | |
| HUGGINGFACEHUB_API_TOKEN = os.getenv('HUGGINGFACEHUB_API_TOKEN') | |
| repo_id = os.environ.get('repo_id') | |
| llm = HuggingFaceHub(repo_id=repo_id, | |
| model_kwargs={"min_length":100, | |
| "max_new_tokens":1024, "do_sample":True, | |
| "temperature":0.1, | |
| "top_k":50, | |
| "top_p":0.95, "eos_token_id":49155}) | |
| prompt_template = """You are a very helpful AI assistant. Please response to the user's input question with as many details as possible. | |
| Question: {user_question} | |
| Helpufl AI AI Repsonse: | |
| """ | |
| llm_chain = LLMChain(llm=llm, prompt=PromptTemplate.from_template(prompt_template)) | |
| user_query = st.text_input("Enter your query here:") | |
| with st.spinner("AI Thinking...Please wait a while to Cheers!"): | |
| if user_query != "": | |
| initial_response=llm_chain.run(user_query) | |
| temp_ai_response_1=initial_response.partition('<|end|>\n<|user|>\n')[0] | |
| temp_ai_response_2=temp_ai_response_1.replace('<|end|>\n<|assistant|>\n', '') | |
| final_ai_response=temp_ai_response_2.replace('<|end|>\n<|system|>\n', '') | |
| st.write("AI Response:") | |
| st.write(final_ai_response) |