Omraghu07's picture
Upload 11 files
b71ce4b verified
# src/server.py
"""
Free-Plan Friendly SEO Keyword Research API
Optimized to minimize SerpAPI calls while maximizing keyword discovery
Key Features:
- Configurable keyword count (5, 10, 20, 50, etc.)
- Only 1 SerpAPI call per seed for candidate collection
- Mock scoring for initial ranking
- Optional SerpAPI verification for top N results
- Strict mode for free plan protection (max 5 API calls per request)
"""
import os
import logging
import time
import math
import re
import io
from typing import List, Dict, Any, Optional, Tuple
from datetime import datetime
from collections import Counter
from fastapi import FastAPI, HTTPException, Query, Request
from fastapi.middleware.cors import CORSMiddleware
from fastapi.responses import JSONResponse, StreamingResponse
from pydantic import BaseModel, Field
from dotenv import load_dotenv
try:
import pandas as pd
HAS_PANDAS = True
except ImportError:
HAS_PANDAS = False
try:
from serpapi import GoogleSearch
HAS_SERPAPI = True
except ImportError:
try:
from google_search_results import GoogleSearch
HAS_SERPAPI = True
except ImportError:
HAS_SERPAPI = False
# Load environment
load_dotenv()
# Configure logging
logging.basicConfig(
level=logging.INFO,
format='%(asctime)s - %(levelname)s - %(message)s'
)
logger = logging.getLogger(__name__)
# Initialize FastAPI
app = FastAPI(
title="Free-Plan Friendly SEO Keyword API",
description="Efficient keyword research optimized for SerpAPI free plan",
version="4.0.0",
docs_url="/docs"
)
# CORS
app.add_middleware(
CORSMiddleware,
allow_origins=["*"],
allow_credentials=True,
allow_methods=["GET", "POST", "OPTIONS"],
allow_headers=["*"],
)
# Configuration
SERPAPI_KEY = os.getenv("SERPAPI_KEY")
API_AUTH_KEY = os.getenv("API_AUTH_KEY")
USE_SERPAPI_STRICT_MODE = os.getenv("USE_SERPAPI_STRICT_MODE", "true").lower() == "true"
MAX_SERPAPI_CALLS_STRICT = 5 # Maximum API calls in strict mode
MAX_SERPAPI_CALLS_NORMAL = 20 # Maximum API calls in normal mode
# Rate limiting
REQUEST_TIMES = {}
RATE_LIMIT_WINDOW = 60
RATE_LIMIT_MAX_REQUESTS = 30
# Request counter for monitoring
API_CALL_COUNTER = {"total": 0, "session_start": time.time()}
class KeywordResponse(BaseModel):
"""API response model."""
success: bool = True
seed: str
requested: int
returned: int
results: List[Dict[str, Any]]
processing_time: float
api_calls_used: int
api_budget_remaining: int
data_source: str
timestamp: str
def count_api_call():
"""Track API usage."""
API_CALL_COUNTER["total"] += 1
logger.info(f"API call #{API_CALL_COUNTER['total']} - Session time: {time.time() - API_CALL_COUNTER['session_start']:.1f}s")
def get_api_budget() -> int:
"""Calculate remaining API budget for this request."""
max_calls = MAX_SERPAPI_CALLS_STRICT if USE_SERPAPI_STRICT_MODE else MAX_SERPAPI_CALLS_NORMAL
used = API_CALL_COUNTER["total"]
return max(0, max_calls - used)
def heuristic_competition_score(keyword: str) -> float:
"""
Calculate mock competition score based on keyword characteristics.
Does NOT use any API calls.
"""
words = keyword.lower().split()
word_count = len(words)
# Base competition by word count
base_scores = {1: 0.8, 2: 0.6, 3: 0.4, 4: 0.25, 5: 0.2}
base_score = base_scores.get(word_count, max(0.15, 0.3 - (word_count * 0.02)))
# Adjust for question keywords (lower competition)
question_words = ["how", "what", "why", "when", "where", "who", "which", "can", "should", "is", "are", "does"]
if any(word in words for word in question_words):
base_score *= 0.7
# Adjust for commercial intent (higher competition)
commercial_words = ["buy", "best", "top", "review", "price", "cheap", "discount"]
if any(word in words for word in commercial_words):
base_score *= 1.3
# Adjust for specific/niche keywords (lower competition)
specific_words = ["beginner", "tutorial", "guide", "explained", "step", "diy", "simple"]
if any(word in words for word in specific_words):
base_score *= 0.8
# Add some deterministic variation based on keyword hash
variation = (hash(keyword) % 20) / 100 # -0.1 to +0.1
base_score += variation
return max(0.05, min(0.95, base_score))
def heuristic_search_volume(keyword: str) -> int:
"""
Estimate search volume based on keyword characteristics.
Does NOT use any API calls.
"""
words = keyword.lower().split()
word_count = len(words)
# Base volumes
base_volumes = {1: 10000, 2: 5000, 3: 2000, 4: 800, 5: 400}
base_volume = base_volumes.get(word_count, max(100, 500 - (word_count * 50)))
# Adjust for popular terms
popular_terms = ["free", "online", "best", "how", "tutorial", "guide"]
if any(term in words for term in popular_terms):
base_volume = int(base_volume * 1.5)
# Adjust for very specific/niche terms
niche_terms = ["advanced", "professional", "enterprise", "custom"]
if any(term in words for term in niche_terms):
base_volume = int(base_volume * 0.6)
# Add deterministic variation
variation_factor = 1 + ((hash(keyword) % 40) - 20) / 100 # 0.8 to 1.2
volume = int(base_volume * variation_factor)
return max(10, min(100000, volume))
def calculate_opportunity_score(volume: int, competition: float) -> float:
"""Calculate opportunity score."""
volume_score = math.log10(volume + 1)
return volume_score / (competition + 0.1)
def score_keyword_heuristic(keyword: str) -> Dict[str, Any]:
"""
Score a keyword using only heuristics (NO API calls).
Fast and free method for initial ranking.
"""
competition = heuristic_competition_score(keyword)
volume = heuristic_search_volume(keyword)
opportunity = calculate_opportunity_score(volume, competition)
# Determine difficulty
if competition < 0.3:
difficulty = "Easy"
elif competition < 0.5:
difficulty = "Medium"
elif competition < 0.7:
difficulty = "Hard"
else:
difficulty = "Very Hard"
# Estimate ranking potential
if competition < 0.4 and volume >= 300:
ranking_chance = "High"
elif competition < 0.6 and volume >= 100:
ranking_chance = "Medium"
else:
ranking_chance = "Low"
return {
"keyword": keyword,
"monthly_searches": volume,
"competition_score": round(competition, 4),
"opportunity_score": round(opportunity, 2),
"difficulty": difficulty,
"ranking_chance": ranking_chance,
"data_source": "heuristic"
}
def enrich_with_serpapi(keyword: str) -> Optional[Dict[str, Any]]:
"""
Enrich a keyword with real SerpAPI data.
Uses 1 API call per keyword.
"""
if not HAS_SERPAPI or not SERPAPI_KEY:
logger.warning("SerpAPI not available for enrichment")
return None
try:
count_api_call()
params = {
"engine": "google",
"q": keyword,
"api_key": SERPAPI_KEY,
"hl": "en",
"gl": "us",
"num": 10
}
search = GoogleSearch(params)
results = search.get_dict()
if "error" in results:
logger.error(f"SerpAPI error: {results['error']}")
return None
# Extract metrics
search_info = results.get("search_information", {})
total_results_raw = search_info.get("total_results") or search_info.get("total_results_raw") or ""
total_results = 0
if isinstance(total_results_raw, int):
total_results = total_results_raw
elif isinstance(total_results_raw, str):
nums = re.sub(r"[^\d]", "", total_results_raw)
total_results = int(nums) if nums else 0
ads_count = len(results.get("ads_results", []))
has_featured_snippet = bool(results.get("featured_snippet") or results.get("answer_box"))
has_paa = bool(results.get("related_questions") or results.get("people_also_ask"))
has_kg = bool(results.get("knowledge_graph"))
# Calculate real competition
normalized_results = min(math.log10(total_results + 1) / 7, 1.0) if total_results > 0 else 0
ads_score = min(ads_count / 3, 1.0)
competition = (
0.40 * normalized_results +
0.25 * ads_score +
0.15 * (1 if has_featured_snippet else 0) +
0.10 * (1 if has_paa else 0) +
0.10 * (1 if has_kg else 0)
)
competition = max(0.0, min(1.0, competition))
# Estimate volume from signals
word_count = len(keyword.split())
base_volume = max(100, 8000 // (word_count + 1))
if ads_count > 2:
base_volume = int(base_volume * 1.5)
if has_featured_snippet:
base_volume = int(base_volume * 1.2)
volume = min(base_volume, 50000)
opportunity = calculate_opportunity_score(volume, competition)
# Determine difficulty
if competition < 0.3:
difficulty = "Easy"
elif competition < 0.5:
difficulty = "Medium"
elif competition < 0.7:
difficulty = "Hard"
else:
difficulty = "Very Hard"
# Ranking chance
if competition < 0.35:
ranking_chance = "High"
elif competition < 0.55:
ranking_chance = "Medium"
else:
ranking_chance = "Low"
return {
"keyword": keyword,
"monthly_searches": volume,
"competition_score": round(competition, 4),
"opportunity_score": round(opportunity, 2),
"difficulty": difficulty,
"ranking_chance": ranking_chance,
"total_results": total_results,
"ads_count": ads_count,
"featured_snippet": "Yes" if has_featured_snippet else "No",
"people_also_ask": "Yes" if has_paa else "No",
"knowledge_graph": "Yes" if has_kg else "No",
"data_source": "serpapi"
}
except Exception as e:
logger.error(f"SerpAPI enrichment failed for '{keyword}': {e}")
return None
def collect_candidates_from_seed(seed: str) -> Tuple[List[str], int]:
"""
Collect keyword candidates using ONLY 1 SerpAPI call.
Returns (candidates, api_calls_used)
"""
candidates = set()
candidates.add(seed) # Always include seed
api_calls = 0
# Generate synthetic candidates (NO API calls)
question_words = ["how to", "what is", "why", "when", "where", "can i", "should i"]
modifiers = ["best", "free", "online", "guide", "tutorial", "tips", "examples",
"for beginners", "explained", "2024", "2025", "cheap", "review"]
for q in question_words[:5]:
candidates.add(f"{q} {seed}")
for mod in modifiers[:15]:
candidates.add(f"{seed} {mod}")
candidates.add(f"{mod} {seed}")
# Make ONE SerpAPI call to get real related keywords
if HAS_SERPAPI and SERPAPI_KEY:
try:
count_api_call()
api_calls = 1
params = {
"engine": "google",
"q": seed,
"api_key": SERPAPI_KEY,
"hl": "en",
"gl": "us"
}
search = GoogleSearch(params)
results = search.get_dict()
if "error" not in results:
# Extract related searches
for item in results.get("related_searches", [])[:20]:
query = item.get("query", "")
if query and len(query.split()) <= 6:
candidates.add(query.lower().strip())
# Extract PAA questions
for item in results.get("related_questions", [])[:15]:
question = item.get("question", "")
if question:
candidates.add(question.lower().strip())
logger.info(f"SerpAPI call successful: collected real suggestions")
else:
logger.warning(f"SerpAPI error: {results.get('error')}")
except Exception as e:
logger.error(f"SerpAPI collection failed: {e}")
final_candidates = list(candidates)
logger.info(f"Collected {len(final_candidates)} candidates ({api_calls} API call)")
return final_candidates, api_calls
def check_rate_limit(client_ip: str) -> bool:
"""Rate limiting."""
current_time = time.time()
if client_ip not in REQUEST_TIMES:
REQUEST_TIMES[client_ip] = []
REQUEST_TIMES[client_ip] = [
t for t in REQUEST_TIMES[client_ip]
if current_time - t < RATE_LIMIT_WINDOW
]
if len(REQUEST_TIMES[client_ip]) >= RATE_LIMIT_MAX_REQUESTS:
return False
REQUEST_TIMES[client_ip].append(current_time)
return True
@app.on_event("startup")
async def startup():
"""Startup logging."""
logger.info("=" * 60)
logger.info("SEO Keyword API - Free Plan Optimized")
logger.info(f"Strict Mode: {USE_SERPAPI_STRICT_MODE}")
logger.info(f"Max API calls per request: {MAX_SERPAPI_CALLS_STRICT if USE_SERPAPI_STRICT_MODE else MAX_SERPAPI_CALLS_NORMAL}")
logger.info(f"SerpAPI Available: {HAS_SERPAPI and bool(SERPAPI_KEY)}")
logger.info("=" * 60)
@app.get("/")
async def root():
"""Root endpoint."""
return {
"service": "Free-Plan Friendly SEO Keyword API",
"version": "4.0.0",
"strict_mode": USE_SERPAPI_STRICT_MODE,
"max_api_calls": MAX_SERPAPI_CALLS_STRICT if USE_SERPAPI_STRICT_MODE else MAX_SERPAPI_CALLS_NORMAL,
"strategy": "1 API call for candidate collection + optional enrichment for top N",
"endpoints": {
"/keywords": "Main keyword research (configurable count)",
"/health": "Health check",
"/stats": "API usage statistics"
}
}
@app.get("/health")
async def health():
"""Health check."""
return {
"status": "healthy",
"timestamp": datetime.utcnow().isoformat(),
"serpapi_available": HAS_SERPAPI and bool(SERPAPI_KEY),
"strict_mode": USE_SERPAPI_STRICT_MODE,
"session_api_calls": API_CALL_COUNTER["total"]
}
@app.get("/stats")
async def stats():
"""API usage statistics."""
uptime = time.time() - API_CALL_COUNTER["session_start"]
return {
"session_start": datetime.fromtimestamp(API_CALL_COUNTER["session_start"]).isoformat(),
"uptime_seconds": round(uptime, 1),
"total_api_calls": API_CALL_COUNTER["total"],
"strict_mode": USE_SERPAPI_STRICT_MODE,
"max_calls_per_request": MAX_SERPAPI_CALLS_STRICT if USE_SERPAPI_STRICT_MODE else MAX_SERPAPI_CALLS_NORMAL
}
@app.get("/keywords", response_model=KeywordResponse)
async def get_keywords(
request: Request,
seed: str = Query(..., description="Seed keyword", min_length=1, max_length=100),
top: int = Query(50, description="Number of keywords to return", ge=1, le=100),
enrich_top: int = Query(4, description="Number of top results to enrich with SerpAPI", ge=0, le=20)
):
"""
Main keyword research endpoint.
Strategy:
1. Make 1 SerpAPI call to collect candidates from seed
2. Score all candidates with heuristics (free)
3. Optionally enrich top N with real SerpAPI data
Parameters:
- seed: Your main keyword
- top: How many keywords you want (e.g., 5, 10, 20, 50)
- enrich_top: How many of the top results to verify with SerpAPI (0 = none, saves API calls)
Example: top=10, enrich_top=3 means:
- 1 API call to collect candidates
- Return 10 keywords scored with heuristics
- Enrich the top 3 with real SerpAPI data (3 more API calls)
- Total: 4 API calls
"""
start_time = time.time()
client_ip = request.client.host or "unknown"
# Authentication
if API_AUTH_KEY:
auth = request.headers.get("Authorization", "").replace("Bearer ", "")
if auth != API_AUTH_KEY:
raise HTTPException(401, "Invalid or missing API key")
# Rate limiting
if not check_rate_limit(client_ip):
raise HTTPException(429, "Rate limit exceeded")
# Validate
seed = seed.strip().lower()
if not seed:
raise HTTPException(400, "Invalid seed keyword")
# Check API budget
max_calls = MAX_SERPAPI_CALLS_STRICT if USE_SERPAPI_STRICT_MODE else MAX_SERPAPI_CALLS_NORMAL
if enrich_top > 0:
required_calls = 1 + enrich_top # 1 for collection + N for enrichment
if required_calls > max_calls:
raise HTTPException(
400,
f"Request would use {required_calls} API calls, but budget is {max_calls}. "
f"Reduce enrich_top to {max_calls - 1} or less."
)
try:
logger.info(f"Request: seed='{seed}', top={top}, enrich_top={enrich_top}")
# Step 1: Collect candidates (1 API call)
candidates, api_calls_used = collect_candidates_from_seed(seed)
if not candidates:
raise HTTPException(404, "No candidates found")
# Step 2: Score all candidates with heuristics (FREE - no API calls)
logger.info(f"Scoring {len(candidates)} candidates with heuristics...")
scored_candidates = []
for candidate in candidates:
try:
result = score_keyword_heuristic(candidate)
scored_candidates.append(result)
except Exception as e:
logger.warning(f"Heuristic scoring failed for '{candidate}': {e}")
continue
# Sort by opportunity score (highest first)
scored_candidates.sort(key=lambda x: x["opportunity_score"], reverse=True)
# Get top N requested
top_results = scored_candidates[:top]
# Step 3: Optionally enrich top results with real SerpAPI data
data_source = "heuristic"
if enrich_top > 0 and HAS_SERPAPI and SERPAPI_KEY:
logger.info(f"Enriching top {enrich_top} results with SerpAPI...")
for i in range(min(enrich_top, len(top_results))):
keyword = top_results[i]["keyword"]
# Check budget before each call
if api_calls_used >= max_calls:
logger.warning(f"API budget exhausted at {api_calls_used} calls")
break
enriched = enrich_with_serpapi(keyword)
if enriched:
top_results[i] = enriched
api_calls_used += 1
data_source = "mixed"
# Small delay between calls
time.sleep(0.2)
logger.info(f"Enrichment complete: {api_calls_used} total API calls used")
# Add ranking
for rank, result in enumerate(top_results, 1):
result["rank"] = rank
processing_time = time.time() - start_time
budget_remaining = max_calls - api_calls_used
logger.info(
f"SUCCESS: Returned {len(top_results)} keywords, "
f"API calls: {api_calls_used}/{max_calls}, "
f"Time: {processing_time:.2f}s"
)
return KeywordResponse(
success=True,
seed=seed,
requested=top,
returned=len(top_results),
results=top_results,
processing_time=round(processing_time, 2),
api_calls_used=api_calls_used,
api_budget_remaining=budget_remaining,
data_source=data_source,
timestamp=datetime.utcnow().isoformat()
)
except HTTPException:
raise
except Exception as e:
logger.error(f"Request failed: {e}")
raise HTTPException(500, f"Processing error: {str(e)}")
@app.get("/export/csv")
async def export_csv(
seed: str = Query(...),
top: int = Query(50),
enrich_top: int = Query(0)
):
"""Export results as CSV."""
if not HAS_PANDAS:
raise HTTPException(500, "CSV export unavailable (pandas not installed)")
# Get keyword data
response = await get_keywords(Request(scope={"type": "http", "client": ("127.0.0.1", 0), "headers": []}), seed, top, enrich_top)
# Convert to DataFrame
df = pd.DataFrame(response.results)
# Create CSV
output = io.StringIO()
df.to_csv(output, index=False)
output.seek(0)
return StreamingResponse(
iter([output.getvalue()]),
media_type="text/csv",
headers={"Content-Disposition": f"attachment; filename=keywords_{seed.replace(' ', '_')}.csv"}
)
if __name__ == "__main__":
import uvicorn
port = int(os.getenv("PORT", 8000))
logger.info(f"Starting server on port {port}")
uvicorn.run(
app,
host="0.0.0.0",
port=port,
log_level="info"
)