NotebookLM-Clone / backend /llm_client.py
rahulrb99
Use openai/gpt-oss-20b:groq (Groq-supported model)
e46cb4d
raw
history blame contribute delete
628 Bytes
"""Helper for shared LLM-based features."""
import os
from openai import OpenAI
DEFAULT_MODEL = os.getenv("LLM_MODEL", "openai/gpt-oss-20b:groq")
_client: OpenAI | None = None
def get_llm_client() -> OpenAI:
"""Return a cached OpenAI client configured with the Hugging Face router."""
global _client
if _client is None:
token = os.getenv("HF_TOKEN")
if not token:
raise RuntimeError("HF_TOKEN environment variable is required for LLM calls.")
_client = OpenAI(
base_url="https://router.huggingface.co/v1",
api_key=token,
)
return _client