| # π GitHub Models Setup (Recommended) | |
| ## Why Use GitHub Models? | |
| - **Free tier**: No credit card required | |
| - **Excellent tool calling**: More reliable than small local models | |
| - **Stable cloud endpoint**: No disconnects | |
| - **Fast responses**: 2-5 seconds per query | |
| - **Easy setup**: Just need a GitHub personal access token | |
| ## Quick Setup | |
| ### 1. Get a GitHub Personal Access Token | |
| - Go to [GitHub tokens](https://github.com/settings/tokens) | |
| - Click "Generate new token (classic)" | |
| - Name it (e.g., `Multi-Agent Backend Testing`) | |
| - Select scopes: `repo` (if needed), `read:org` (optional) | |
| - Click "Generate token" and copy it | |
| ### 2. Configure Environment | |
| ```powershell | |
| notepad .env | |
| # Add your token: | |
| GITHUB_TOKEN=ghp_xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx | |
| ``` | |
| ### 3. Test Your Setup | |
| ```powershell | |
| uv run test_agents.py | |
| # Should see: Using GitHub Models: openai/gpt-5-mini via https://models.github.ai | |
| ``` | |
| ## LLM Priority Order | |
| 1. GitHub Models (if `GITHUB_TOKEN` set) | |
| 2. OpenAI (if `OPENAI_API_KEY` set) | |
| 3. Google GenAI (if `GOOGLE_API_KEY` set) | |
| 4. Ollama (local fallback) | |
| ## Troubleshooting | |
| - **Initialization failed**: Check token validity and format (`ghp_` or `github_pat_`, 40+ chars) | |
| - **Rate limit exceeded**: Wait 1 minute or use another provider | |
| - **Model not available**: List available models: | |
| ```powershell | |
| curl -H "Authorization: Bearer YOUR_TOKEN" -H "Accept: application/vnd.github+json" https://models.github.ai/models | |
| ``` | |
| ## Alternative Models | |
| If `gpt-5-mini` has issues, try: | |
| - Claude: `anthropic/claude-3-5-sonnet` | |
| - Llama: `meta-llama/Meta-Llama-3.1-8B-Instruct` | |
| - GPT-4: `openai/gpt-4` | |
| Edit `.env` or [agents.py](agents.py) to change the model. | |
| ## Comparison: GitHub Models vs Ollama | |
| | Feature | GitHub Models | Ollama (qwen3:0.6b) | | |
| |--------------- |--------------|---------------------| | |
| | Setup | 2 min | 10+ min | | |
| | Cost | Free | Free (local) | | |
| | Speed | 2-5 sec | 5-15 sec | | |
| | Reliability | 98% | 50% (disconnects) | | |
| | Tool Calling | Excellent | Poor | | |
| | RAM Usage | 0 MB | 1-2 GB | | |
| | GPU Needed | No | Optional | | |
| | Quality | High | Low | | |
| ## Production Deployment | |
| - Use paid GitHub Models tier for higher limits | |
| - OpenAI API for maximum reliability | |
| - Azure OpenAI for enterprise features | |
| Automatic fallback supported in codebase | |
| ## Reverting to Ollama | |
| Comment out `GITHUB_TOKEN` in `.env` and set: | |
| ```powershell | |
| OLLAMA_BASE_URL=http://localhost:11434 | |
| OLLAMA_MODEL=llama3.2 | |
| ``` | |
| ## Summary | |
| GitHub Models is the **recommended default** for this project: | |
| - Free, easy, production-quality responses | |
| - No local resource requirements | |
| - Excellent tool calling for agentic workflows | |
| [Get started in 2 minutes](https://github.com/settings/tokens) | |
| π Happy testing! | |