Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -153,8 +153,11 @@ if prompt := st.chat_input(f"Hi I'm {selected_model}, ask me a question"):
|
|
| 153 |
|
| 154 |
# Display assistant response in chat message container
|
| 155 |
with st.chat_message("assistant"):
|
| 156 |
-
|
| 157 |
model=model_links[selected_model],
|
|
|
|
|
|
|
|
|
|
| 158 |
messages=[
|
| 159 |
{"role": m["role"], "content": m["content"]}
|
| 160 |
for m in st.session_state.messages
|
|
|
|
| 153 |
|
| 154 |
# Display assistant response in chat message container
|
| 155 |
with st.chat_message("assistant"):
|
| 156 |
+
client = InferenceClient(
|
| 157 |
model=model_links[selected_model],
|
| 158 |
+
headers)
|
| 159 |
+
|
| 160 |
+
output = client.text_generation(
|
| 161 |
messages=[
|
| 162 |
{"role": m["role"], "content": m["content"]}
|
| 163 |
for m in st.session_state.messages
|