cr-agent / docs /GITHUB_MODELS_SETUP.md
Sibi Krishnamoorthy
fix workflow
48a5851
# πŸš€ GitHub Models Setup (Recommended)
## Why Use GitHub Models?
- **Free tier**: No credit card required
- **Excellent tool calling**: More reliable than small local models
- **Stable cloud endpoint**: No disconnects
- **Fast responses**: 2-5 seconds per query
- **Easy setup**: Just need a GitHub personal access token
## Quick Setup
### 1. Get a GitHub Personal Access Token
- Go to [GitHub tokens](https://github.com/settings/tokens)
- Click "Generate new token (classic)"
- Name it (e.g., `Multi-Agent Backend Testing`)
- Select scopes: `repo` (if needed), `read:org` (optional)
- Click "Generate token" and copy it
### 2. Configure Environment
```powershell
notepad .env
# Add your token:
GITHUB_TOKEN=ghp_xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx
```
### 3. Test Your Setup
```powershell
uv run test_agents.py
# Should see: Using GitHub Models: openai/gpt-5-mini via https://models.github.ai
```
## LLM Priority Order
1. GitHub Models (if `GITHUB_TOKEN` set)
2. OpenAI (if `OPENAI_API_KEY` set)
3. Google GenAI (if `GOOGLE_API_KEY` set)
4. Ollama (local fallback)
## Troubleshooting
- **Initialization failed**: Check token validity and format (`ghp_` or `github_pat_`, 40+ chars)
- **Rate limit exceeded**: Wait 1 minute or use another provider
- **Model not available**: List available models:
```powershell
curl -H "Authorization: Bearer YOUR_TOKEN" -H "Accept: application/vnd.github+json" https://models.github.ai/models
```
## Alternative Models
If `gpt-5-mini` has issues, try:
- Claude: `anthropic/claude-3-5-sonnet`
- Llama: `meta-llama/Meta-Llama-3.1-8B-Instruct`
- GPT-4: `openai/gpt-4`
Edit `.env` or [agents.py](agents.py) to change the model.
## Comparison: GitHub Models vs Ollama
| Feature | GitHub Models | Ollama (qwen3:0.6b) |
|--------------- |--------------|---------------------|
| Setup | 2 min | 10+ min |
| Cost | Free | Free (local) |
| Speed | 2-5 sec | 5-15 sec |
| Reliability | 98% | 50% (disconnects) |
| Tool Calling | Excellent | Poor |
| RAM Usage | 0 MB | 1-2 GB |
| GPU Needed | No | Optional |
| Quality | High | Low |
## Production Deployment
- Use paid GitHub Models tier for higher limits
- OpenAI API for maximum reliability
- Azure OpenAI for enterprise features
Automatic fallback supported in codebase
## Reverting to Ollama
Comment out `GITHUB_TOKEN` in `.env` and set:
```powershell
OLLAMA_BASE_URL=http://localhost:11434
OLLAMA_MODEL=llama3.2
```
## Summary
GitHub Models is the **recommended default** for this project:
- Free, easy, production-quality responses
- No local resource requirements
- Excellent tool calling for agentic workflows
[Get started in 2 minutes](https://github.com/settings/tokens)
πŸŽ‰ Happy testing!