Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -100,7 +100,7 @@ Reference json:
|
|
| 100 |
}
|
| 101 |
"""
|
| 102 |
|
| 103 |
-
def ask_llm(user_message, model='
|
| 104 |
client = Groq(api_key=os.environ.get("GROQ_API_KEY"))
|
| 105 |
|
| 106 |
response = client.chat.completions.create(
|
|
@@ -120,7 +120,7 @@ def ask_llm(user_message, model='deepseek-r1-distill-llama-70b', system_prompt="
|
|
| 120 |
|
| 121 |
return response.choices[0].message.content
|
| 122 |
|
| 123 |
-
def ask_ollama(user_message, model='
|
| 124 |
client = Groq(api_key=os.environ.get("GROQ_API_KEY"))
|
| 125 |
|
| 126 |
response = client.chat.completions.create(
|
|
|
|
| 100 |
}
|
| 101 |
"""
|
| 102 |
|
| 103 |
+
def ask_llm(user_message, model='llama-3.3-70b-versatile', system_prompt="You are a helpful assistant."):
|
| 104 |
client = Groq(api_key=os.environ.get("GROQ_API_KEY"))
|
| 105 |
|
| 106 |
response = client.chat.completions.create(
|
|
|
|
| 120 |
|
| 121 |
return response.choices[0].message.content
|
| 122 |
|
| 123 |
+
def ask_ollama(user_message, model='llama-3.3-70b-versatile', system_prompt=search_prompt):
|
| 124 |
client = Groq(api_key=os.environ.get("GROQ_API_KEY"))
|
| 125 |
|
| 126 |
response = client.chat.completions.create(
|