Spaces:
Runtime error
Runtime error
Parimal Kalpande
commited on
Commit
Β·
dbcc933
1
Parent(s):
f09eba4
final update
Browse files- DEPLOYMENT_GUIDE.md +2 -1
- GROQ_MODEL_UPDATE.md +36 -0
- modules/llm_handler.py +1 -1
DEPLOYMENT_GUIDE.md
CHANGED
|
@@ -4,7 +4,8 @@
|
|
| 4 |
1. β
Missing dependencies - FIXED
|
| 5 |
2. β
Missing GROQ_API_KEY - NEEDS CONFIGURATION
|
| 6 |
3. β
File paths - FIXED
|
| 7 |
-
4.
|
|
|
|
| 8 |
|
| 9 |
## Solution 1: Update Requirements.txt with Compatible Gradio Version
|
| 10 |
|
|
|
|
| 4 |
1. β
Missing dependencies - FIXED
|
| 5 |
2. β
Missing GROQ_API_KEY - NEEDS CONFIGURATION
|
| 6 |
3. β
File paths - FIXED
|
| 7 |
+
4. β
Gradio version bug - FIXED (downgraded to 4.40.0)
|
| 8 |
+
5. β
Decommissioned model - FIXED (updated to llama-3.1-70b-versatile)
|
| 9 |
|
| 10 |
## Solution 1: Update Requirements.txt with Compatible Gradio Version
|
| 11 |
|
GROQ_MODEL_UPDATE.md
ADDED
|
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# GROQ MODEL UPDATE GUIDE
|
| 2 |
+
|
| 3 |
+
## Issue Fixed: Model Decommissioned Error
|
| 4 |
+
|
| 5 |
+
The error you encountered:
|
| 6 |
+
```
|
| 7 |
+
The model llama3-70b-8192 has been decommissioned and is no longer supported
|
| 8 |
+
```
|
| 9 |
+
|
| 10 |
+
## β
SOLUTION APPLIED:
|
| 11 |
+
Updated the model in `modules/llm_handler.py` from:
|
| 12 |
+
- β `llama3-70b-8192` (decommissioned)
|
| 13 |
+
- β
`llama-3.1-70b-versatile` (active)
|
| 14 |
+
|
| 15 |
+
## Currently Supported Groq Models (as of Oct 2024):
|
| 16 |
+
|
| 17 |
+
### **Recommended for Your Use Case:**
|
| 18 |
+
- `llama-3.1-70b-versatile` - Best for detailed interview evaluation
|
| 19 |
+
- `llama-3.1-8b-instant` - Faster, good for quick responses
|
| 20 |
+
- `mixtral-8x7b-32768` - Good alternative with large context
|
| 21 |
+
|
| 22 |
+
### **All Available Models:**
|
| 23 |
+
- `llama-3.1-405b-reasoning` - Most powerful (if available)
|
| 24 |
+
- `llama-3.1-70b-versatile` - High quality, good balance
|
| 25 |
+
- `llama-3.1-8b-instant` - Fast responses
|
| 26 |
+
- `llama3-groq-70b-8192-tool-use-preview` - With tool support
|
| 27 |
+
- `llama3-groq-8b-8192-tool-use-preview` - Faster with tools
|
| 28 |
+
- `mixtral-8x7b-32768` - Mixtral model
|
| 29 |
+
- `gemma2-9b-it` - Google's Gemma
|
| 30 |
+
- `gemma-7b-it` - Smaller Gemma
|
| 31 |
+
|
| 32 |
+
## For Hugging Face Spaces Deployment:
|
| 33 |
+
Your app should now work properly with the updated model. The interview functionality will work normally without the API error.
|
| 34 |
+
|
| 35 |
+
## Testing:
|
| 36 |
+
Run your app and try creating an interview - the question generation should now work without the decommissioned model error.
|
modules/llm_handler.py
CHANGED
|
@@ -15,7 +15,7 @@ if not groq_api_key:
|
|
| 15 |
MODEL = None
|
| 16 |
else:
|
| 17 |
client = Groq(api_key=groq_api_key)
|
| 18 |
-
MODEL = "
|
| 19 |
|
| 20 |
def generate_question(interview_type, document_text):
|
| 21 |
if not client:
|
|
|
|
| 15 |
MODEL = None
|
| 16 |
else:
|
| 17 |
client = Groq(api_key=groq_api_key)
|
| 18 |
+
MODEL = "llama-3.1-70b-versatile" # Updated to supported model
|
| 19 |
|
| 20 |
def generate_question(interview_type, document_text):
|
| 21 |
if not client:
|