Frontier-CS / research /scripts /llm_interface.py
andylizf's picture
Upload folder using huggingface_hub
5fed0fc verified
from abc import ABC, abstractmethod
from typing import Any, Tuple, Optional
from openai import OpenAI, APITimeoutError
import google.generativeai as genai
import os
from dotenv import load_dotenv
from anthropic import Anthropic, APITimeoutError as AnthropicAPITimeoutError
load_dotenv()
class LLMInterface(ABC):
"""
Abstract base class for integrating Large Language Models (LLMs) into a competitive programming context.
"""
def __init__(self):
"""
Initialize the LLMInterface with a predefined prompt for generating competitive programming solutions.
"""
self.prompt = """
You are a competitive programmer. You will be given a problem statement, please implement a solution in C++. The execution time and memory limit are also stated in the statement so be aware of the complexity of the program. Please wrap the code in ```cpp and ``` so that it is properly formatted. Your response should ONLY contain the C++ code, with no additional explanation or text.
"""
@abstractmethod
def call_llm(self, user_prompt: str) -> Tuple[str, Any]:
"""
Abstract method to interact with the LLM.
"""
pass
def generate_solution(self, problem_statement: str) -> Tuple[str, Any]:
"""
Generates a solution to a given competitive programming problem using the LLM.
"""
user_prompt = self.prompt + problem_statement
response, meta = self.call_llm(user_prompt)
return response, meta
class GPT(LLMInterface):
"""Concrete implementation of LLMInterface using OpenAI chat models."""
def __init__(
self,
model: str = "gpt-5",
reasoning_effort: Optional[str] = "high",
timeout: float = 600.0,
base_url: Optional[str] = None,
api_key: Optional[str] = None,
):
super().__init__()
resolved_key = api_key or os.getenv("OPENAI_API_KEY")
client_kwargs = {"api_key": resolved_key, "timeout": timeout}
if base_url:
client_kwargs["base_url"] = base_url
self.client = OpenAI(**client_kwargs)
self.name = 'gpt'
self.model = model
self.reasoning_effort = reasoning_effort
def call_llm(self, user_prompt: str) -> Tuple[str, Any]:
"""Sends the user prompt to the configured OpenAI model."""
try:
request_kwargs = {
"model": self.model,
"messages": [{"role": "user", "content": user_prompt}],
}
if self.reasoning_effort:
request_kwargs["reasoning_effort"] = self.reasoning_effort
completion = self.client.chat.completions.create(**request_kwargs)
return completion.choices[0].message.content, str(completion)
except APITimeoutError as e:
print(f"OpenAI API request timed out: {e}")
return "", str(e)
except Exception as e:
print(f"An unexpected error occurred while calling the OpenAI API: {e}")
return "", str(e)
class Gemini(LLMInterface):
"""
Concrete implementation of LLMInterface using Google's Gemini 2.5 Pro model.
Attributes:
model (genai.GenerativeModel): Instance for interacting with the Gemini API.
"""
def __init__(self, model: str = 'gemini-2.5-pro', timeout: float = 600.0, api_key: Optional[str] = None):
"""
Initializes the GeminiLLM class by configuring the API key and creating an
instance of the Gemini model.
"""
super().__init__()
try:
key_candidates = [
api_key,
os.getenv("GOOGLE_API_KEY"),
os.getenv("GEMINI_API_KEY"),
]
resolved_key = next((k for k in key_candidates if k), None)
if not resolved_key:
raise ValueError("GOOGLE_API_KEY not found in environment variables.")
self.api_key = resolved_key
genai.configure(api_key=self.api_key)
# Using a powerful and recent model. You can change this to other available models.
self.model_name = model
self.timeout = timeout
self.model = genai.GenerativeModel(self.model_name)
except Exception as e:
print(f"Error during Gemini initialization: {e}")
self.model = None
self.name = 'gemini'
def call_llm(self, user_prompt: str) -> Tuple[str, Any]:
"""
Sends the user prompt to the Gemini model and retrieves the solution.
"""
if not self.model:
return "Error: Model not initialized.", None
try:
if hasattr(self, "api_key") and self.api_key:
genai.configure(api_key=self.api_key)
response = self.model.generate_content(
user_prompt,
request_options={"timeout": self.timeout}
)
solution_text = response.text
return solution_text, response
except Exception as e:
print(f"An error occurred while calling the Gemini API: {e}")
return f"Error: {e}", None
class ClaudeBase(LLMInterface):
"""Shared Anthropic client wrapper."""
def __init__(
self,
model: str,
name: str = 'claude',
max_tokens: int = 32000,
thinking_budget: Optional[int] = 20000,
timeout: float = 600.0,
api_key: Optional[str] = None,
):
super().__init__()
resolved_key = api_key or os.getenv("ANTHROPIC_API_KEY")
self.client = Anthropic(api_key=resolved_key, timeout=timeout)
self.name = name
self.model = model
self.max_tokens = max_tokens
self.thinking_budget = thinking_budget
def call_llm(self, user_prompt: str) -> Tuple[str, Any]:
"""Sends the combined user prompt to Anthropic's model."""
try:
request_kwargs = {
"model": self.model,
"max_tokens": self.max_tokens,
"messages": [{"role": "user", "content": user_prompt}],
}
if self.thinking_budget:
request_kwargs["thinking"] = {
"type": "enabled",
"budget_tokens": self.thinking_budget,
}
completion = self.client.messages.create(**request_kwargs)
final_text = ""
if hasattr(completion, 'content') and completion.content:
for block in completion.content:
if getattr(block, 'type', None) == 'text' and hasattr(block, 'text'):
final_text += block.text
return final_text, str(completion)
except AnthropicAPITimeoutError as e:
print(f"Anthropic API request timed out: {e}")
return "", str(e)
except Exception as e:
print(f"An unexpected error occurred while calling the Anthropic API: {e}")
return "", str(e)
class Claude(ClaudeBase):
def __init__(self, model: str = "claude-sonnet-4-20250514", **kwargs):
super().__init__(model=model, **kwargs)
class Claude_Opus(ClaudeBase):
def __init__(self, model: str = "claude-opus-4-1-20250805", **kwargs):
super().__init__(model=model, **kwargs)
class Claude_Sonnet_4_5(ClaudeBase):
def __init__(self, model: str = "claude-sonnet-4-5-20250929", **kwargs):
super().__init__(model=model, **kwargs)
class DeepSeek(LLMInterface):
"""
Concrete implementation of LLMInterface using DeepSeek's models.
For deepseek-reasoner (R1/V3.2), max_tokens controls total output
including Chain-of-Thought reasoning. Default 32K, max 64K.
"""
def __init__(
self,
model: str = "deepseek-reasoner",
max_tokens: int = 32000,
timeout: float = 600.0,
base_url: str = "https://api.deepseek.com",
api_key: Optional[str] = None,
):
super().__init__()
resolved_key = api_key or os.getenv("DEEPSEEK_API_KEY")
self.client = OpenAI(
api_key=resolved_key,
base_url=base_url,
timeout=timeout
)
self.name = 'deepseek'
self.model = model
self.max_tokens = max_tokens
def call_llm(self, user_prompt: str) -> Tuple[str, Any]:
"""Sends the user prompt to DeepSeek's model."""
try:
request_kwargs = {
"model": self.model,
"messages": [{"role": "user", "content": user_prompt}],
"max_tokens": self.max_tokens,
}
completion = self.client.chat.completions.create(**request_kwargs)
return completion.choices[0].message.content, str(completion)
except APITimeoutError as e:
print(f"DeepSeek API request timed out: {e}")
return "", str(e)
except Exception as e:
print(f"An unexpected error occurred while calling the DeepSeek API: {e}")
return "", str(e)
class Grok(LLMInterface):
"""
Concrete implementation of LLMInterface using xAI's Grok models.
"""
def __init__(
self,
model: str = "grok-4",
reasoning_effort: Optional[str] = "high",
timeout: float = 600.0,
base_url: str = "https://api.x.ai/v1",
api_key: Optional[str] = None,
):
"""
Initializes the Grok class by creating an instance of the OpenAI client
pointed at the Grok API endpoint.
"""
super().__init__()
resolved_key = api_key or os.getenv("XAI_API_KEY") or os.getenv("GROK_API_KEY")
self.client = OpenAI(
api_key=resolved_key,
base_url=base_url,
timeout=timeout
)
self.name = 'grok'
self.model = model
self.reasoning_effort = reasoning_effort
def call_llm(self, user_prompt: str) -> Tuple[str, Any]:
"""
Sends the combined user prompt to Grok's model.
Args:
user_prompt (str): The complete prompt (system + problem).
Returns:
Tuple[str, Any]: The LLM's response and metadata.
"""
try:
# Reverted to the simpler, single-message format
request_kwargs = {
"model": self.model,
"messages": [
{"role": "user", "content": user_prompt}
]
}
if self.reasoning_effort:
request_kwargs["reasoning_effort"] = self.reasoning_effort
completion = self.client.chat.completions.create(**request_kwargs)
return completion.choices[0].message.content, str(completion)
except APITimeoutError as e:
print(f"Grok (xAI) API request timed out: {e}")
return "", str(e)
except Exception as e:
print(f"An unexpected error occurred while calling the Grok (xAI) API: {e}")
return "", str(e)