|
|
from functools import lru_cache |
|
|
from typing import List, Tuple, Optional |
|
|
import aiohttp |
|
|
import elevenlabs |
|
|
from contextlib import asynccontextmanager |
|
|
from logger import setup_logger, log_execution_time, log_async_execution_time |
|
|
|
|
|
logger = setup_logger("api_clients") |
|
|
|
|
|
class OpenRouterClient: |
|
|
"""Handles OpenRouter API interactions with comprehensive logging and error tracking""" |
|
|
|
|
|
def __init__(self, api_key: str): |
|
|
logger.info("Initializing OpenRouter client") |
|
|
if not api_key or len(api_key) < 32: |
|
|
logger.error("Invalid API key format") |
|
|
raise ValueError("Invalid OpenRouter API key") |
|
|
|
|
|
self.api_key = api_key |
|
|
self.base_url = "https://openrouter.ai/api/v1" |
|
|
self.headers = { |
|
|
"Authorization": f"Bearer {api_key}", |
|
|
"Content-Type": "application/json", |
|
|
} |
|
|
logger.debug("OpenRouter client initialized successfully") |
|
|
|
|
|
@asynccontextmanager |
|
|
async def get_session(self): |
|
|
logger.debug("Creating new aiohttp session") |
|
|
async with aiohttp.ClientSession(headers=self.headers) as session: |
|
|
yield session |
|
|
|
|
|
@lru_cache(maxsize=1) |
|
|
async def get_models(self) -> List[Tuple[str, str]]: |
|
|
""" |
|
|
Fetch available models from OpenRouter API |
|
|
|
|
|
Returns: |
|
|
List of tuples containing (model_id, model_description) |
|
|
|
|
|
Raises: |
|
|
ValueError: If API request fails |
|
|
""" |
|
|
logger.info("Fetching available models from OpenRouter") |
|
|
async with self.get_session() as session: |
|
|
async with session.get(f"{self.base_url}/models") as response: |
|
|
if response.status != 200: |
|
|
error_msg = await response.text() |
|
|
logger.error(f"Failed to fetch models: {error_msg}") |
|
|
raise ValueError(f"Failed to fetch models: {error_msg}") |
|
|
|
|
|
models = await response.json() |
|
|
logger.info(f"Successfully fetched {len(models)} models") |
|
|
logger.debug(f"Available models: {[model['name'] for model in models]}") |
|
|
return [(model['id'], f"{model['name']} ({model['context_length']} tokens)") |
|
|
for model in models] |
|
|
|
|
|
@log_async_execution_time(logger) |
|
|
async def generate_script(self, content: str, prompt: str, model_id: str) -> str: |
|
|
""" |
|
|
Generate a podcast script with detailed progress tracking and validation |
|
|
|
|
|
Performance metrics and content analysis are logged at each step. |
|
|
""" |
|
|
logger.info(f"Starting script generation with model: {model_id}") |
|
|
logger.debug(f"Input metrics - Content: {len(content)} chars, Prompt: {len(prompt)} chars") |
|
|
|
|
|
|
|
|
if not content or len(content) < 100: |
|
|
logger.error("Content too short for meaningful script generation") |
|
|
raise ValueError("Insufficient content for script generation") |
|
|
|
|
|
if not prompt or len(prompt) < 10: |
|
|
logger.error("Prompt too short or missing") |
|
|
raise ValueError("Please provide a more detailed prompt") |
|
|
|
|
|
try: |
|
|
async with self.get_session() as session: |
|
|
logger.debug("Preparing script generation request") |
|
|
response = await self._make_script_request(session, content, prompt, model_id) |
|
|
|
|
|
script = response['choices'][0]['message']['content'] |
|
|
logger.info(f"Script generated successfully: {len(script)} chars") |
|
|
logger.debug(f"Script preview: {script[:200]}...") |
|
|
|
|
|
return script |
|
|
except Exception as e: |
|
|
logger.error(f"Script generation failed", exc_info=True) |
|
|
raise |
|
|
|
|
|
async def _make_script_request(self, session, content, prompt, model_id): |
|
|
async with session.post( |
|
|
f"{self.base_url}/chat/completions", |
|
|
json={ |
|
|
"model": model_id, |
|
|
"messages": [ |
|
|
{ |
|
|
"role": "system", |
|
|
"content": "You are an expert podcast script writer. Create engaging, conversational content." |
|
|
}, |
|
|
{ |
|
|
"role": "user", |
|
|
"content": f"""Based on this content: {content} |
|
|
Create a 3-minute podcast script focusing on: {prompt} |
|
|
Format as a natural conversation with clear speaker parts. |
|
|
Include [HOST] and [GUEST] markers for different voices.""" |
|
|
} |
|
|
] |
|
|
} |
|
|
) as response: |
|
|
logger.debug("Sending script generation request") |
|
|
|
|
|
if response.status != 200: |
|
|
error_msg = await response.text() |
|
|
logger.error(f"Script generation failed: {error_msg}") |
|
|
raise ValueError(f"Script generation failed: {error_msg}") |
|
|
|
|
|
return await response.json() |
|
|
|
|
|
class ElevenLabsClient: |
|
|
"""Handles ElevenLabs API interactions with detailed performance tracking""" |
|
|
|
|
|
def __init__(self, api_key: str): |
|
|
logger.info("Initializing ElevenLabs client") |
|
|
self.api_key = api_key |
|
|
elevenlabs.set_api_key(api_key) |
|
|
|
|
|
@lru_cache(maxsize=1) |
|
|
def get_voices(self) -> List[Tuple[str, str]]: |
|
|
""" |
|
|
Fetch available voices from ElevenLabs |
|
|
|
|
|
Returns: |
|
|
List of tuples containing (voice_id, voice_name) |
|
|
""" |
|
|
logger.info("Fetching available voices from ElevenLabs") |
|
|
voices = elevenlabs.voices() |
|
|
logger.info(f"Successfully fetched {len(voices)} voices") |
|
|
logger.debug(f"Available voices: {[voice.name for voice in voices]}") |
|
|
return [(voice.voice_id, voice.name) for voice in voices] |
|
|
|
|
|
@log_execution_time(logger) |
|
|
def generate_audio(self, text: str, voice_id: str) -> bytes: |
|
|
""" |
|
|
Generate audio with comprehensive error handling and quality checks |
|
|
|
|
|
Logs detailed metrics about the input text and resulting audio. |
|
|
""" |
|
|
logger.info(f"Starting audio generation with voice: {voice_id}") |
|
|
logger.debug(f"Input text length: {len(text)} chars") |
|
|
|
|
|
if len(text) > 5000: |
|
|
logger.warning(f"Long text detected ({len(text)} chars), may impact performance") |
|
|
|
|
|
try: |
|
|
start_time = time.time() |
|
|
audio = elevenlabs.generate( |
|
|
text=text, |
|
|
voice=voice_id, |
|
|
model="eleven_monolingual_v1" |
|
|
) |
|
|
|
|
|
duration = time.time() - start_time |
|
|
audio_size = len(audio) |
|
|
logger.info(f"Audio generated: {audio_size} bytes in {duration:.2f} seconds") |
|
|
logger.debug(f"Audio generation rate: {len(text)/duration:.2f} chars/second") |
|
|
|
|
|
return audio |
|
|
except Exception as e: |
|
|
logger.error("Audio generation failed", exc_info=True) |
|
|
raise |
|
|
|