Spaces:
Runtime error
A newer version of the Gradio SDK is available:
6.2.0
π How to Run the AI Development Agent
This guide provides sequential instructions to set up and run all components of the AI Development Agent: the MCP Server (Backend/Integration Hub) and the Web Dashboard (Frontend).
π Prerequisites
- Python 3.10+ (Recommended: 3.11 or 3.12)
- Note: Python 3.13 requires a specific fix for Gradio (included in instructions).
- JIRA Account (for real integration)
- Git
π οΈ Step 1: Setup & Run MCP Server
The MCP Server is the core "brain" that handles RAG, Fine-tuning queries, and JIRA integration.
1. Navigate to directory
cd mcp
2. Create Virtual Environment
python3 -m venv venv
source venv/bin/activate
3. Install Dependencies
pip install -r requirements.txt
# β οΈ Python 3.13 Fix: If you are using Python 3.13, run this extra command:
pip install audioop-lts
4. Configure Environment
Create a .env file in the mcp/ directory:
touch .env
Add your credentials to .env:
# JIRA Configuration
JIRA_URL="https://your-domain.atlassian.net"
JIRA_EMAIL="your-email@example.com"
JIRA_API_TOKEN="your-api-token"
JIRA_PROJECT_KEY="PROJ"
# RAG Configuration
RAG_ENABLED="true"
# URL from Step 1.5 below
RAG_API_URL="https://your-modal-url.modal.run"
# Fine-tuned Model Configuration
# URL from Step 1.6 below
FINETUNED_MODEL_API_URL="https://your-finetuned-model-url.modal.run"
5. Start the Server
python mcp_server.py
β
Success: You should see Running on local URL: http://0.0.0.0:7860
π Step 1.5: Deploy RAG System (Optional)
To enable real RAG capabilities instead of mock data, deploy the RAG system on Modal.
1. Deploy the RAG App
cd .. # Go back to root if in mcp/
./venv/bin/modal deploy src/rag/modal-rag-product-design.py
2. Get the URL
After deployment, you will see a URL ending in ...-api-query.modal.run.
Copy this URL and add it to your mcp/.env file as RAG_API_URL.
To retrieve the URL later:
./venv/bin/modal app list
Look for insurance-rag-product-design and note the endpoint URL.
π Step 1.6: Deploy Fine-Tuned Model (Optional)
To enable the real fine-tuned model for domain insights.
1. Deploy the Model Endpoint
cd .. # Go back to root if in mcp/
./venv/bin/modal deploy src/finetune/api_endpoint_vllm.py
2. Get the URL
After deployment, you will see a URL ending in ...-model-ask.modal.run.
Copy this URL and add it to your mcp/.env file as FINETUNED_MODEL_API_URL.
To retrieve the URL later:
./venv/bin/modal app list
Look for phi3-inference-vllm and note the endpoint URL.
π₯οΈ Step 2: Setup & Run Dashboard
The Dashboard is the user interface where you interact with the agent.
1. Open a new terminal and navigate
cd dashboard
2. Create Virtual Environment
python3 -m venv venv
source venv/bin/activate
3. Install Dependencies
pip install -r requirements.txt
4. Start the Dashboard
python server.py
β
Success: You should see Uvicorn running on http://0.0.0.0:8000
π Step 3: Access the Application
- Open your browser to http://localhost:8000
- Enter a requirement (e.g., "Create a login page with 2FA")
- Watch the agent analyze, query RAG, and create JIRA epics/stories!
π§ Advanced: Fine-Tuning Pipeline
If you want to train your own domain-specific model, follow these steps.
Dataset Generation Results (Reference)
- Training Samples: 201,651
- Validation Samples: 22,407
- Total Dataset: 224,058 high-quality QA pairs
Step 1: Fine-Tune the Model
Run the fine-tuning job on Modal with H200 GPU:
cd /Users/veeru/agents/mcp-hack
./venv/bin/modal run --detach src/finetune/finetune_modal.py
Step 2: Evaluate the Model
After training completes, test the model:
./venv/bin/modal run src/finetune/eval_finetuned.py
Step 3: Deploy Inference API
Option B: High-Performance vLLM Endpoint (Recommended)
- Merge Model:
./venv/bin/modal run src/finetune/merge_model.py - Deploy vLLM Endpoint:
./venv/bin/modal deploy src/finetune/api_endpoint_vllm.py
Step 4: Test the API
curl -X POST https://YOUR-MODAL-URL/ask \
-H "Content-Type: application/json" \
-d '{
"question": "What is the population of Tokyo?",
"context": "Japan Census data"
}'
Troubleshooting Fine-Tuning
- Logs:
modal app logs mcp-hack::finetune-phi3-modal - Volumes:
modal volume list
π Deploying to Hugging Face Spaces
Deploy your MCP Server to Hugging Face Spaces for public access.
Prerequisites
- Hugging Face account (sign up here)
- Git configured locally
Step 1: Create a Space
- Go to huggingface.co/new-space
- Configure the Space:
- Owner: Select your username or organization (e.g.,
MCP-1st-Birthday) - Space name: Choose a name (e.g.,
sdlc-agent) - License: MIT
- SDK: Gradio
- Space hardware: CPU basic (free)
- Visibility: Public
- Owner: Select your username or organization (e.g.,
- Click Create Space
Step 2: Push Your Code
# Navigate to project root
cd /Users/veeru/agents/mcp-hack
# Add Hugging Face as a remote (replace ORG and SPACE_NAME)
git remote add hf https://huggingface.co/spaces/ORG/SPACE_NAME
# Ensure .env is ignored
grep -q "mcp/.env" .gitignore || echo "mcp/.env" >> .gitignore
# Pull initial Space files
git pull hf main --allow-unrelated-histories
# Resolve any conflicts (usually just README)
git checkout --ours README.md
git add README.md
git commit -m "Merge Hugging Face Space initial files"
# Push to Hugging Face
git push hf main
Step 3: Configure Secrets
- Go to your Space's Settings tab
- Scroll to Repository secrets
- Add these secrets:
JIRA_URLJIRA_EMAILJIRA_API_TOKENJIRA_PROJECT_KEYRAG_ENABLED=trueRAG_API_URLFINETUNED_MODEL_API_URL
Step 4: Monitor Deployment
- Check the Logs tab to monitor the build
- Once complete, your app will be live at:
https://huggingface.co/spaces/ORG/SPACE_NAME
Important Notes
β οΈ Limitations:
- Only the MCP Server (
mcp/mcp_server.py) will be deployed - The Dashboard requires a separate deployment (use Render, Railway, or Fly.io)
β Your Space is now live and accessible to anyone!