Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -41,7 +41,7 @@ DEFAULT_REASONER_CONFIG = {
|
|
| 41 |
}
|
| 42 |
|
| 43 |
DEFAULT_LLM_CONFIG = {
|
| 44 |
-
"
|
| 45 |
provider="groq",
|
| 46 |
api_key=PROVIDERS_API_KEYS.get("groq"),
|
| 47 |
model="meta-llama/llama-4-scout-17b-16e-instruct",
|
|
@@ -63,14 +63,12 @@ DEFAULT_LLM_CONFIG = {
|
|
| 63 |
async def chat_profile():
|
| 64 |
return [
|
| 65 |
cl.ChatProfile(
|
| 66 |
-
name="
|
| 67 |
markdown_description="Talk with the lastest Llm models! Powered by AiCore, check it on GitHub, link in Readme",
|
| 68 |
-
icon="https://picsum.photos/200",
|
| 69 |
),
|
| 70 |
cl.ChatProfile(
|
| 71 |
name="OpenAi",
|
| 72 |
markdown_description="Talk with the lastest Llm models! Powered by AiCore, check it on GitHub, link in Readme",
|
| 73 |
-
icon="https://picsum.photos/200",
|
| 74 |
)
|
| 75 |
]
|
| 76 |
|
|
|
|
| 41 |
}
|
| 42 |
|
| 43 |
DEFAULT_LLM_CONFIG = {
|
| 44 |
+
"ZeppFusion": LlmConfig(
|
| 45 |
provider="groq",
|
| 46 |
api_key=PROVIDERS_API_KEYS.get("groq"),
|
| 47 |
model="meta-llama/llama-4-scout-17b-16e-instruct",
|
|
|
|
| 63 |
async def chat_profile():
|
| 64 |
return [
|
| 65 |
cl.ChatProfile(
|
| 66 |
+
name="ZeppFusion",
|
| 67 |
markdown_description="Talk with the lastest Llm models! Powered by AiCore, check it on GitHub, link in Readme",
|
|
|
|
| 68 |
),
|
| 69 |
cl.ChatProfile(
|
| 70 |
name="OpenAi",
|
| 71 |
markdown_description="Talk with the lastest Llm models! Powered by AiCore, check it on GitHub, link in Readme",
|
|
|
|
| 72 |
)
|
| 73 |
]
|
| 74 |
|