Update app.py
Browse files
app.py
CHANGED
|
@@ -166,40 +166,40 @@ else:
|
|
| 166 |
|
| 167 |
#############
|
| 168 |
|
| 169 |
-
# LLaMA 7B model from Hugging Face
|
| 170 |
-
# MODEL_NAME = "huggyllama/llama-7b" # Example of a LLaMA model
|
| 171 |
|
| 172 |
-
# Try this OpenAssistant model available on Hugging Face
|
| 173 |
-
MODEL_NAME = "OpenAssistant/oasst-sft-1-pythia-12b" # Example of an OpenAssistant model
|
| 174 |
|
| 175 |
-
import streamlit as st
|
| 176 |
-
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 177 |
-
import torch
|
| 178 |
|
| 179 |
-
# Load the model and tokenizer (OpenAssistant or LLaMA)
|
| 180 |
-
MODEL_NAME = "OpenAssistant/oasst-sft-1-pythia-12b" # Replace with "huggyllama/llama-7b" for LLaMA
|
| 181 |
-
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
|
| 182 |
-
model = AutoModelForCausalLM.from_pretrained(MODEL_NAME)
|
| 183 |
|
| 184 |
-
# Streamlit UI for input
|
| 185 |
-
st.markdown("<h3 style='text-align: center; font-size: 20px;'>Chat with OpenAssistant/LLaMA</h3>", unsafe_allow_html=True)
|
| 186 |
|
| 187 |
-
# Input text area
|
| 188 |
-
user_input = st.text_area("You:", "", height=150)
|
| 189 |
|
| 190 |
-
if st.button('Generate Response'):
|
| 191 |
-
|
| 192 |
-
|
| 193 |
-
|
| 194 |
-
|
| 195 |
|
| 196 |
-
|
| 197 |
-
|
| 198 |
|
| 199 |
-
|
| 200 |
-
|
| 201 |
-
|
| 202 |
-
|
| 203 |
|
| 204 |
|
| 205 |
|
|
|
|
| 166 |
|
| 167 |
#############
|
| 168 |
|
| 169 |
+
# # LLaMA 7B model from Hugging Face
|
| 170 |
+
# # MODEL_NAME = "huggyllama/llama-7b" # Example of a LLaMA model
|
| 171 |
|
| 172 |
+
# # Try this OpenAssistant model available on Hugging Face
|
| 173 |
+
# MODEL_NAME = "OpenAssistant/oasst-sft-1-pythia-12b" # Example of an OpenAssistant model
|
| 174 |
|
| 175 |
+
# import streamlit as st
|
| 176 |
+
# from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 177 |
+
# import torch
|
| 178 |
|
| 179 |
+
# # Load the model and tokenizer (OpenAssistant or LLaMA)
|
| 180 |
+
# MODEL_NAME = "OpenAssistant/oasst-sft-1-pythia-12b" # Replace with "huggyllama/llama-7b" for LLaMA
|
| 181 |
+
# tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
|
| 182 |
+
# model = AutoModelForCausalLM.from_pretrained(MODEL_NAME)
|
| 183 |
|
| 184 |
+
# # Streamlit UI for input
|
| 185 |
+
# st.markdown("<h3 style='text-align: center; font-size: 20px;'>Chat with OpenAssistant/LLaMA</h3>", unsafe_allow_html=True)
|
| 186 |
|
| 187 |
+
# # Input text area
|
| 188 |
+
# user_input = st.text_area("You:", "", height=150)
|
| 189 |
|
| 190 |
+
# if st.button('Generate Response'):
|
| 191 |
+
# if user_input:
|
| 192 |
+
# # Tokenize the input and generate response
|
| 193 |
+
# inputs = tokenizer(user_input, return_tensors="pt")
|
| 194 |
+
# outputs = model.generate(**inputs, max_length=150)
|
| 195 |
|
| 196 |
+
# # Decode the generated response
|
| 197 |
+
# response = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
| 198 |
|
| 199 |
+
# # Display the model's response
|
| 200 |
+
# st.write("Assistant: ", response)
|
| 201 |
+
# else:
|
| 202 |
+
# st.warning('Please enter some text to get a response!')
|
| 203 |
|
| 204 |
|
| 205 |
|