Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -85,7 +85,13 @@ async def respond(message, history, model_name="command-a-03-2025", temperature=
|
|
| 85 |
yield f"Error: {str(e)}"
|
| 86 |
|
| 87 |
with gr.Blocks() as demo:
|
| 88 |
-
gr.Markdown("## Modular Chatbot
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 89 |
|
| 90 |
# State components to track current values
|
| 91 |
temperature_state = gr.State(value=0.7)
|
|
|
|
| 85 |
yield f"Error: {str(e)}"
|
| 86 |
|
| 87 |
with gr.Blocks() as demo:
|
| 88 |
+
gr.Markdown("""## Modular TTS-Chatbot
|
| 89 |
+
Status: In Development
|
| 90 |
+
|
| 91 |
+
The goal of this project is to enable voice-chat with any supported LLM which currently do not have speech ability similar to Gemini or GPT-4o.
|
| 92 |
+
|
| 93 |
+
|
| 94 |
+
""")
|
| 95 |
|
| 96 |
# State components to track current values
|
| 97 |
temperature_state = gr.State(value=0.7)
|