Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -1381,11 +1381,11 @@ if filename is not None:
|
|
| 1381 |
session_state = {}
|
| 1382 |
if "search_queries" not in session_state:
|
| 1383 |
session_state["search_queries"] = []
|
|
|
|
| 1384 |
example_input = st.text_input("AI Search ArXiV Scholarly Articles", value=session_state["search_queries"][-1] if session_state["search_queries"] else "")
|
|
|
|
| 1385 |
if example_input:
|
| 1386 |
session_state["search_queries"].append(example_input)
|
| 1387 |
-
|
| 1388 |
-
# Search AI
|
| 1389 |
query=example_input
|
| 1390 |
if query:
|
| 1391 |
result = search_arxiv(query)
|
|
@@ -1462,27 +1462,14 @@ client = OpenAI(api_key= os.getenv('OPENAI_API_KEY'), organization=os.getenv('OP
|
|
| 1462 |
#MODEL = "gpt-4o"
|
| 1463 |
MODEL = "gpt-4o-2024-05-13"
|
| 1464 |
|
| 1465 |
-
def
|
| 1466 |
-
text_input = st.text_input("Enter your text:")
|
| 1467 |
-
if text_input:
|
| 1468 |
-
completion = client.chat.completions.create(
|
| 1469 |
-
model=MODEL,
|
| 1470 |
-
messages=[
|
| 1471 |
-
{"role": "system", "content": "You are a helpful assistant. Help me with my math homework!"},
|
| 1472 |
-
{"role": "user", "content": f"Hello! Could you solve {text_input}?"}
|
| 1473 |
-
]
|
| 1474 |
-
)
|
| 1475 |
-
st.write("Assistant: " + completion.choices[0].message.content)
|
| 1476 |
-
|
| 1477 |
-
|
| 1478 |
-
|
| 1479 |
-
def process_text():
|
| 1480 |
-
text_input = st.text_input("Enter your text:")
|
| 1481 |
if text_input:
|
| 1482 |
|
| 1483 |
st.session_state.messages.append({"role": "user", "content": text_input})
|
|
|
|
| 1484 |
with st.chat_message("user"):
|
| 1485 |
st.markdown(text_input)
|
|
|
|
| 1486 |
with st.chat_message("assistant"):
|
| 1487 |
completion = client.chat.completions.create(
|
| 1488 |
model=MODEL,
|
|
@@ -1490,14 +1477,13 @@ def process_text():
|
|
| 1490 |
{"role": m["role"], "content": m["content"]}
|
| 1491 |
for m in st.session_state.messages
|
| 1492 |
],
|
| 1493 |
-
stream=
|
| 1494 |
)
|
| 1495 |
return_text = completion.choices[0].message.content
|
| 1496 |
st.write("Assistant: " + return_text)
|
| 1497 |
filename = generate_filename(text_input, "md")
|
| 1498 |
create_file(filename, text_input, return_text, should_save)
|
| 1499 |
-
|
| 1500 |
-
st.session_state.messages.append({"role": "assistant", "content": response})
|
| 1501 |
|
| 1502 |
#st.write("Assistant: " + completion.choices[0].message.content)
|
| 1503 |
|
|
@@ -1673,8 +1659,12 @@ def process_audio_and_video(video_input):
|
|
| 1673 |
],
|
| 1674 |
temperature=0,
|
| 1675 |
)
|
| 1676 |
-
|
| 1677 |
-
st.markdown(
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1678 |
|
| 1679 |
|
| 1680 |
def main():
|
|
@@ -1682,7 +1672,10 @@ def main():
|
|
| 1682 |
st.markdown("##### GPT-4o Omni Model: Text, Audio, Image, & Video")
|
| 1683 |
option = st.selectbox("Select an option", ("Text", "Image", "Audio", "Video"))
|
| 1684 |
if option == "Text":
|
| 1685 |
-
|
|
|
|
|
|
|
|
|
|
| 1686 |
elif option == "Image":
|
| 1687 |
image_input = st.file_uploader("Upload an image", type=["jpg", "jpeg", "png"])
|
| 1688 |
image_response = process_image(image_input)
|
|
|
|
| 1381 |
session_state = {}
|
| 1382 |
if "search_queries" not in session_state:
|
| 1383 |
session_state["search_queries"] = []
|
| 1384 |
+
|
| 1385 |
example_input = st.text_input("AI Search ArXiV Scholarly Articles", value=session_state["search_queries"][-1] if session_state["search_queries"] else "")
|
| 1386 |
+
|
| 1387 |
if example_input:
|
| 1388 |
session_state["search_queries"].append(example_input)
|
|
|
|
|
|
|
| 1389 |
query=example_input
|
| 1390 |
if query:
|
| 1391 |
result = search_arxiv(query)
|
|
|
|
| 1462 |
#MODEL = "gpt-4o"
|
| 1463 |
MODEL = "gpt-4o-2024-05-13"
|
| 1464 |
|
| 1465 |
+
def process_text(text_input):
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1466 |
if text_input:
|
| 1467 |
|
| 1468 |
st.session_state.messages.append({"role": "user", "content": text_input})
|
| 1469 |
+
|
| 1470 |
with st.chat_message("user"):
|
| 1471 |
st.markdown(text_input)
|
| 1472 |
+
|
| 1473 |
with st.chat_message("assistant"):
|
| 1474 |
completion = client.chat.completions.create(
|
| 1475 |
model=MODEL,
|
|
|
|
| 1477 |
{"role": m["role"], "content": m["content"]}
|
| 1478 |
for m in st.session_state.messages
|
| 1479 |
],
|
| 1480 |
+
stream=False
|
| 1481 |
)
|
| 1482 |
return_text = completion.choices[0].message.content
|
| 1483 |
st.write("Assistant: " + return_text)
|
| 1484 |
filename = generate_filename(text_input, "md")
|
| 1485 |
create_file(filename, text_input, return_text, should_save)
|
| 1486 |
+
st.session_state.messages.append({"role": "assistant", "content": return_text})
|
|
|
|
| 1487 |
|
| 1488 |
#st.write("Assistant: " + completion.choices[0].message.content)
|
| 1489 |
|
|
|
|
| 1659 |
],
|
| 1660 |
temperature=0,
|
| 1661 |
)
|
| 1662 |
+
results = response.choices[0].message.content
|
| 1663 |
+
st.markdown(results)
|
| 1664 |
+
|
| 1665 |
+
filename = generate_filename(transcript, "md")
|
| 1666 |
+
create_file(filename, transcript, results, should_save)
|
| 1667 |
+
|
| 1668 |
|
| 1669 |
|
| 1670 |
def main():
|
|
|
|
| 1672 |
st.markdown("##### GPT-4o Omni Model: Text, Audio, Image, & Video")
|
| 1673 |
option = st.selectbox("Select an option", ("Text", "Image", "Audio", "Video"))
|
| 1674 |
if option == "Text":
|
| 1675 |
+
text_input = st.text_input("Enter your text:")
|
| 1676 |
+
if (text_input > ''):
|
| 1677 |
+
textResponse = process_text(text_input)
|
| 1678 |
+
|
| 1679 |
elif option == "Image":
|
| 1680 |
image_input = st.file_uploader("Upload an image", type=["jpg", "jpeg", "png"])
|
| 1681 |
image_response = process_image(image_input)
|