Spaces:
Sleeping
Sleeping
| # models_config.py | |
| MODELS = [ | |
| { | |
| "id": "meta-llama/Meta-Llama-3.1-8B-Instruct", | |
| "name": "Llama-3.1-8B-Instruct", | |
| "quantized": False, | |
| "provider": "groq" # Groq usually reliable for Llama-3.1 | |
| }, | |
| { | |
| "id": "google/gemma-2-9b-it", | |
| "name": "Gemma-2-9B-it", | |
| "quantized": False, | |
| "provider": "nebius" # Nebius AI – currently active for Gemma-2-9b-it | |
| # Alternative: "featherless-ai" if Nebius has issues | |
| }, | |
| # Add more if needed (avoid deprecated combinations) | |
| ] | |
| JUDGE_MODEL = "meta-llama/Meta-Llama-3.1-70B-Instruct" # or use "groq" provider below if needed | |
| # JUDGE_PROVIDER = "groq" # uncomment if judge fails on default |