Update app.py
Browse files
app.py
CHANGED
|
@@ -358,7 +358,7 @@ async def search_api(query: str = Form(...), job_id: str = Form(...)):
|
|
| 358 |
try:
|
| 359 |
chat_completion = groq_client.chat.completions.create(
|
| 360 |
messages=[{"role": "user", "content": rag_prompt}],
|
| 361 |
-
model="moonshotai/kimi-k2-instruct-0905",
|
| 362 |
temperature=0.1,
|
| 363 |
max_tokens=1024
|
| 364 |
)
|
|
@@ -415,7 +415,7 @@ async def describe_clusters_api(job_id: str = Form(...)):
|
|
| 415 |
messages=[
|
| 416 |
{"role": "system", "content": "JSON Output Only."},
|
| 417 |
{"role": "user", "content": master_prompt},
|
| 418 |
-
], model="meta-llama/llama-
|
| 419 |
)
|
| 420 |
|
| 421 |
response_content = chat_completion.choices[0].message.content
|
|
|
|
| 358 |
try:
|
| 359 |
chat_completion = groq_client.chat.completions.create(
|
| 360 |
messages=[{"role": "user", "content": rag_prompt}],
|
| 361 |
+
model="moonshotai/kimi-k2-instruct-0905",
|
| 362 |
temperature=0.1,
|
| 363 |
max_tokens=1024
|
| 364 |
)
|
|
|
|
| 415 |
messages=[
|
| 416 |
{"role": "system", "content": "JSON Output Only."},
|
| 417 |
{"role": "user", "content": master_prompt},
|
| 418 |
+
], model="meta-llama/llama-4-maverick-17b-128e-instruct", temperature=0.2,
|
| 419 |
)
|
| 420 |
|
| 421 |
response_content = chat_completion.choices[0].message.content
|