Update app.py
Browse files
app.py
CHANGED
|
@@ -12,8 +12,8 @@ def load_resources():
|
|
| 12 |
load_dotenv()
|
| 13 |
huggingface_token = os.getenv("HUGGINGFACE_TOKEN")
|
| 14 |
subprocess.run(["huggingface-cli", "login", "--token", huggingface_token], capture_output=True)
|
| 15 |
-
tokenizer = AutoTokenizer.from_pretrained("istiak101/TinyLlama-1.1B-
|
| 16 |
-
model = AutoModelForCausalLM.from_pretrained("istiak101/TinyLlama-1.1B-
|
| 17 |
return model, tokenizer
|
| 18 |
|
| 19 |
def create_test_prompt(question, context, tokenizer):
|
|
@@ -79,6 +79,14 @@ def generate_pdf(convo, topic):
|
|
| 79 |
return None
|
| 80 |
return result
|
| 81 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 82 |
# --- Session Init ---
|
| 83 |
if "chat_sessions" not in st.session_state:
|
| 84 |
st.session_state.chat_sessions = {}
|
|
@@ -255,17 +263,18 @@ if st.session_state.current_conversation:
|
|
| 255 |
st.error("❌ Failed to generate PDF.")
|
| 256 |
|
| 257 |
# --- User Prompt ---
|
| 258 |
-
question_input = st.text_input("Enter your question:")
|
| 259 |
-
context_input = st.text_area("Enter the context:")
|
|
|
|
|
|
|
|
|
|
| 260 |
|
| 261 |
# Button to submit
|
| 262 |
if st.button("Submit"):
|
| 263 |
if question_input and context_input:
|
| 264 |
-
combined_input = f"Question: {question_input}<br>Context: {context_input}"
|
| 265 |
convo.append({"role": "user", "text": combined_input})
|
| 266 |
|
| 267 |
-
st.session_state.question_input = ""
|
| 268 |
-
st.session_state.context_input = ""
|
| 269 |
# Avoid rerunning unnecessarily
|
| 270 |
st.session_state.chat_sessions[st.session_state.current_conversation] = convo
|
| 271 |
st.rerun()
|
|
|
|
| 12 |
load_dotenv()
|
| 13 |
huggingface_token = os.getenv("HUGGINGFACE_TOKEN")
|
| 14 |
subprocess.run(["huggingface-cli", "login", "--token", huggingface_token], capture_output=True)
|
| 15 |
+
tokenizer = AutoTokenizer.from_pretrained("istiak101/TinyLlama-1.1B-Finetuned")
|
| 16 |
+
model = AutoModelForCausalLM.from_pretrained("istiak101/TinyLlama-1.1B-Finetuned")
|
| 17 |
return model, tokenizer
|
| 18 |
|
| 19 |
def create_test_prompt(question, context, tokenizer):
|
|
|
|
| 79 |
return None
|
| 80 |
return result
|
| 81 |
|
| 82 |
+
def submit1():
|
| 83 |
+
st.session_state.question_input = st.session_state.qi
|
| 84 |
+
st.session_state.qi = ""
|
| 85 |
+
|
| 86 |
+
def submit2():
|
| 87 |
+
st.session_state.context_input = st.session_state.ci
|
| 88 |
+
st.session_state.ci = ""
|
| 89 |
+
|
| 90 |
# --- Session Init ---
|
| 91 |
if "chat_sessions" not in st.session_state:
|
| 92 |
st.session_state.chat_sessions = {}
|
|
|
|
| 263 |
st.error("❌ Failed to generate PDF.")
|
| 264 |
|
| 265 |
# --- User Prompt ---
|
| 266 |
+
# question_input = st.text_input("Enter your question:")
|
| 267 |
+
# context_input = st.text_area("Enter the context:")
|
| 268 |
+
|
| 269 |
+
st.text_input("Enter your question:", key="qi", on_change=submit1)
|
| 270 |
+
st.text_area("Enter the context:", key="ci", on_change=submit2)
|
| 271 |
|
| 272 |
# Button to submit
|
| 273 |
if st.button("Submit"):
|
| 274 |
if question_input and context_input:
|
| 275 |
+
combined_input = f"Question: {question_input}<br><br>Context: {context_input}"
|
| 276 |
convo.append({"role": "user", "text": combined_input})
|
| 277 |
|
|
|
|
|
|
|
| 278 |
# Avoid rerunning unnecessarily
|
| 279 |
st.session_state.chat_sessions[st.session_state.current_conversation] = convo
|
| 280 |
st.rerun()
|