| from abc import ABC, abstractmethod |
| from typing import Any, Tuple, Optional |
| from openai import OpenAI, APITimeoutError |
| import google.generativeai as genai |
| import os |
| from dotenv import load_dotenv |
| from anthropic import Anthropic, APITimeoutError as AnthropicAPITimeoutError |
|
|
| load_dotenv() |
|
|
| class LLMInterface(ABC): |
| """ |
| Abstract base class for integrating Large Language Models (LLMs) into a competitive programming context. |
| """ |
|
|
| def __init__(self): |
| """ |
| Initialize the LLMInterface with a predefined prompt for generating competitive programming solutions. |
| """ |
| self.prompt = """ |
| You are a competitive programmer. You will be given a problem statement, please implement a solution in C++. The execution time and memory limit are also stated in the statement so be aware of the complexity of the program. Please wrap the code in ```cpp and ``` so that it is properly formatted. Your response should ONLY contain the C++ code, with no additional explanation or text. |
| """ |
|
|
| @abstractmethod |
| def call_llm(self, user_prompt: str) -> Tuple[str, Any]: |
| """ |
| Abstract method to interact with the LLM. |
| """ |
| pass |
|
|
| def generate_solution(self, problem_statement: str) -> Tuple[str, Any]: |
| """ |
| Generates a solution to a given competitive programming problem using the LLM. |
| """ |
| user_prompt = self.prompt + problem_statement |
| response, meta = self.call_llm(user_prompt) |
| return response, meta |
|
|
|
|
| class GPT(LLMInterface): |
| """Concrete implementation of LLMInterface using OpenAI chat models.""" |
|
|
| def __init__( |
| self, |
| model: str = "gpt-5", |
| reasoning_effort: Optional[str] = "high", |
| timeout: float = 600.0, |
| base_url: Optional[str] = None, |
| api_key: Optional[str] = None, |
| ): |
| super().__init__() |
| resolved_key = api_key or os.getenv("OPENAI_API_KEY") |
| client_kwargs = {"api_key": resolved_key, "timeout": timeout} |
| if base_url: |
| client_kwargs["base_url"] = base_url |
| self.client = OpenAI(**client_kwargs) |
| self.name = 'gpt' |
| self.model = model |
| self.reasoning_effort = reasoning_effort |
|
|
| def call_llm(self, user_prompt: str) -> Tuple[str, Any]: |
| """Sends the user prompt to the configured OpenAI model.""" |
| try: |
| request_kwargs = { |
| "model": self.model, |
| "messages": [{"role": "user", "content": user_prompt}], |
| } |
| if self.reasoning_effort: |
| request_kwargs["reasoning_effort"] = self.reasoning_effort |
| completion = self.client.chat.completions.create(**request_kwargs) |
| return completion.choices[0].message.content, str(completion) |
| except APITimeoutError as e: |
| print(f"OpenAI API request timed out: {e}") |
| return "", str(e) |
| except Exception as e: |
| print(f"An unexpected error occurred while calling the OpenAI API: {e}") |
| return "", str(e) |
|
|
| class Gemini(LLMInterface): |
| """ |
| Concrete implementation of LLMInterface using Google's Gemini 2.5 Pro model. |
| |
| Attributes: |
| model (genai.GenerativeModel): Instance for interacting with the Gemini API. |
| """ |
|
|
| def __init__(self, model: str = 'gemini-2.5-pro', timeout: float = 600.0, api_key: Optional[str] = None): |
| """ |
| Initializes the GeminiLLM class by configuring the API key and creating an |
| instance of the Gemini model. |
| """ |
| super().__init__() |
| try: |
| key_candidates = [ |
| api_key, |
| os.getenv("GOOGLE_API_KEY"), |
| os.getenv("GEMINI_API_KEY"), |
| ] |
| resolved_key = next((k for k in key_candidates if k), None) |
| if not resolved_key: |
| raise ValueError("GOOGLE_API_KEY not found in environment variables.") |
| self.api_key = resolved_key |
| genai.configure(api_key=self.api_key) |
| |
| self.model_name = model |
| self.timeout = timeout |
| self.model = genai.GenerativeModel(self.model_name) |
| except Exception as e: |
| print(f"Error during Gemini initialization: {e}") |
| self.model = None |
| self.name = 'gemini' |
|
|
| def call_llm(self, user_prompt: str) -> Tuple[str, Any]: |
| """ |
| Sends the user prompt to the Gemini model and retrieves the solution. |
| """ |
| if not self.model: |
| return "Error: Model not initialized.", None |
|
|
| try: |
| if hasattr(self, "api_key") and self.api_key: |
| genai.configure(api_key=self.api_key) |
| response = self.model.generate_content( |
| user_prompt, |
| request_options={"timeout": self.timeout} |
| ) |
| solution_text = response.text |
| return solution_text, response |
| except Exception as e: |
| print(f"An error occurred while calling the Gemini API: {e}") |
| return f"Error: {e}", None |
|
|
|
|
| class ClaudeBase(LLMInterface): |
| """Shared Anthropic client wrapper.""" |
|
|
| def __init__( |
| self, |
| model: str, |
| name: str = 'claude', |
| max_tokens: int = 32000, |
| thinking_budget: Optional[int] = 20000, |
| timeout: float = 600.0, |
| api_key: Optional[str] = None, |
| ): |
| super().__init__() |
| resolved_key = api_key or os.getenv("ANTHROPIC_API_KEY") |
| self.client = Anthropic(api_key=resolved_key, timeout=timeout) |
| self.name = name |
| self.model = model |
| self.max_tokens = max_tokens |
| self.thinking_budget = thinking_budget |
|
|
| def call_llm(self, user_prompt: str) -> Tuple[str, Any]: |
| """Sends the combined user prompt to Anthropic's model.""" |
| try: |
| request_kwargs = { |
| "model": self.model, |
| "max_tokens": self.max_tokens, |
| "messages": [{"role": "user", "content": user_prompt}], |
| } |
| if self.thinking_budget: |
| request_kwargs["thinking"] = { |
| "type": "enabled", |
| "budget_tokens": self.thinking_budget, |
| } |
|
|
| completion = self.client.messages.create(**request_kwargs) |
|
|
| final_text = "" |
| if hasattr(completion, 'content') and completion.content: |
| for block in completion.content: |
| if getattr(block, 'type', None) == 'text' and hasattr(block, 'text'): |
| final_text += block.text |
|
|
| return final_text, str(completion) |
|
|
| except AnthropicAPITimeoutError as e: |
| print(f"Anthropic API request timed out: {e}") |
| return "", str(e) |
| except Exception as e: |
| print(f"An unexpected error occurred while calling the Anthropic API: {e}") |
| return "", str(e) |
|
|
|
|
| class Claude(ClaudeBase): |
| def __init__(self, model: str = "claude-sonnet-4-20250514", **kwargs): |
| super().__init__(model=model, **kwargs) |
|
|
|
|
| class Claude_Opus(ClaudeBase): |
| def __init__(self, model: str = "claude-opus-4-1-20250805", **kwargs): |
| super().__init__(model=model, **kwargs) |
|
|
|
|
| class Claude_Sonnet_4_5(ClaudeBase): |
| def __init__(self, model: str = "claude-sonnet-4-5-20250929", **kwargs): |
| super().__init__(model=model, **kwargs) |
|
|
| class DeepSeek(LLMInterface): |
| """ |
| Concrete implementation of LLMInterface using DeepSeek's models. |
| |
| For deepseek-reasoner (R1/V3.2), max_tokens controls total output |
| including Chain-of-Thought reasoning. Default 32K, max 64K. |
| """ |
|
|
| def __init__( |
| self, |
| model: str = "deepseek-reasoner", |
| max_tokens: int = 32000, |
| timeout: float = 600.0, |
| base_url: str = "https://api.deepseek.com", |
| api_key: Optional[str] = None, |
| ): |
| super().__init__() |
| resolved_key = api_key or os.getenv("DEEPSEEK_API_KEY") |
| self.client = OpenAI( |
| api_key=resolved_key, |
| base_url=base_url, |
| timeout=timeout |
| ) |
| self.name = 'deepseek' |
| self.model = model |
| self.max_tokens = max_tokens |
|
|
| def call_llm(self, user_prompt: str) -> Tuple[str, Any]: |
| """Sends the user prompt to DeepSeek's model.""" |
| try: |
| request_kwargs = { |
| "model": self.model, |
| "messages": [{"role": "user", "content": user_prompt}], |
| "max_tokens": self.max_tokens, |
| } |
| completion = self.client.chat.completions.create(**request_kwargs) |
| return completion.choices[0].message.content, str(completion) |
| except APITimeoutError as e: |
| print(f"DeepSeek API request timed out: {e}") |
| return "", str(e) |
| except Exception as e: |
| print(f"An unexpected error occurred while calling the DeepSeek API: {e}") |
| return "", str(e) |
|
|
|
|
| class Grok(LLMInterface): |
| """ |
| Concrete implementation of LLMInterface using xAI's Grok models. |
| """ |
|
|
| def __init__( |
| self, |
| model: str = "grok-4", |
| reasoning_effort: Optional[str] = "high", |
| timeout: float = 600.0, |
| base_url: str = "https://api.x.ai/v1", |
| api_key: Optional[str] = None, |
| ): |
| """ |
| Initializes the Grok class by creating an instance of the OpenAI client |
| pointed at the Grok API endpoint. |
| """ |
| super().__init__() |
| resolved_key = api_key or os.getenv("XAI_API_KEY") or os.getenv("GROK_API_KEY") |
| self.client = OpenAI( |
| api_key=resolved_key, |
| base_url=base_url, |
| timeout=timeout |
| ) |
| self.name = 'grok' |
| self.model = model |
| self.reasoning_effort = reasoning_effort |
|
|
| def call_llm(self, user_prompt: str) -> Tuple[str, Any]: |
| """ |
| Sends the combined user prompt to Grok's model. |
| |
| Args: |
| user_prompt (str): The complete prompt (system + problem). |
| |
| Returns: |
| Tuple[str, Any]: The LLM's response and metadata. |
| """ |
| try: |
| |
| request_kwargs = { |
| "model": self.model, |
| "messages": [ |
| {"role": "user", "content": user_prompt} |
| ] |
| } |
| if self.reasoning_effort: |
| request_kwargs["reasoning_effort"] = self.reasoning_effort |
| completion = self.client.chat.completions.create(**request_kwargs) |
| return completion.choices[0].message.content, str(completion) |
| except APITimeoutError as e: |
| print(f"Grok (xAI) API request timed out: {e}") |
| return "", str(e) |
| except Exception as e: |
| print(f"An unexpected error occurred while calling the Grok (xAI) API: {e}") |
| return "", str(e) |
|
|