Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -20,7 +20,7 @@ pipe = pipe.to("cuda" if torch.cuda.is_available() else "cpu")
|
|
| 20 |
# ===== PROMPT ENHANCER USING GROQ =====
|
| 21 |
def enhance_prompt(user_prompt):
|
| 22 |
response = client.chat.completions.create(
|
| 23 |
-
model="
|
| 24 |
messages=[
|
| 25 |
{
|
| 26 |
"role": "system",
|
|
|
|
| 20 |
# ===== PROMPT ENHANCER USING GROQ =====
|
| 21 |
def enhance_prompt(user_prompt):
|
| 22 |
response = client.chat.completions.create(
|
| 23 |
+
model="llama-3.3-70b-versatile",
|
| 24 |
messages=[
|
| 25 |
{
|
| 26 |
"role": "system",
|