| """
|
| Model name utilities for Frontier-CS.
|
|
|
| Provides consistent model prefix conversion used across:
|
| - Solution generation (generate_solutions.py)
|
| - Solution matrix checking (frontier-eval check)
|
| - Batch evaluation (frontier-eval batch)
|
| """
|
|
|
| import re
|
| from typing import Dict, List, Optional, Tuple
|
|
|
|
|
| MODEL_PREFIX_ALIASES: Dict[str, str] = {
|
| "grokcodefast1_": "grok4fastreasoning_",
|
| }
|
|
|
|
|
| def get_model_prefix(model: str) -> str:
|
| """
|
| Convert model name to the prefix format used in solution folder names.
|
|
|
| This is the canonical function for model prefix conversion. All tools
|
| should use this to ensure consistent naming.
|
|
|
| Examples:
|
| >>> get_model_prefix("gpt-5")
|
| 'gpt5'
|
| >>> get_model_prefix("gpt-5.1-preview")
|
| 'gpt5.1'
|
| >>> get_model_prefix("gemini/gemini-2.5-pro")
|
| 'gemini2.5pro'
|
| >>> get_model_prefix("claude-sonnet-4-5-20250929")
|
| 'claude4.5sonnet'
|
| >>> get_model_prefix("grok-3-fast-reasoning")
|
| 'grok3fastreasoning'
|
|
|
| Args:
|
| model: Model name (e.g., "gpt-5", "gemini/gemini-2.5-pro")
|
|
|
| Returns:
|
| Normalized prefix for solution directory names
|
| """
|
| original = model
|
|
|
|
|
| if "/" in model:
|
| model = model.split("/", 1)[1]
|
|
|
| model_lower = model.lower().strip()
|
|
|
|
|
|
|
| if model_lower.startswith("gpt-5.1") or model_lower.startswith("gpt5.1"):
|
| return "gpt5.1"
|
| if model_lower.startswith("gpt-5") or model_lower.startswith("gpt5"):
|
| return "gpt5"
|
|
|
|
|
| if "gemini-2.5-pro" in model_lower or "gemini2.5pro" in model_lower:
|
| return "gemini2.5pro"
|
|
|
|
|
| gemini_match = re.match(r"gemini-?(\d+\.?\d*)-?pro", model_lower)
|
| if gemini_match:
|
| version = gemini_match.group(1)
|
| return f"gemini{version}pro"
|
|
|
|
|
| claude_match = re.match(r"claude-([a-z]+)-(\d+)-(\d+)", model_lower)
|
| if claude_match:
|
| family = claude_match.group(1)
|
| major = claude_match.group(2)
|
| minor = claude_match.group(3)
|
| return f"claude{major}.{minor}{family}"
|
|
|
|
|
| if "grok" in model_lower:
|
| sanitized = re.sub(r"[^a-zA-Z0-9]+", "", model_lower)
|
| if sanitized:
|
| return sanitized
|
|
|
|
|
| sanitized = re.sub(r"[^a-zA-Z0-9]+", "", model_lower)
|
| if not sanitized:
|
| raise ValueError(f"Unable to derive model prefix from '{original}'")
|
| return sanitized
|
|
|
|
|
| def normalize_solution_name(name: str) -> str:
|
| """
|
| Normalize a solution directory name by applying prefix aliases.
|
|
|
| Args:
|
| name: Solution directory name
|
|
|
| Returns:
|
| Normalized name with aliases applied
|
| """
|
| for old_prefix, new_prefix in MODEL_PREFIX_ALIASES.items():
|
| if name.startswith(old_prefix):
|
| return new_prefix + name[len(old_prefix):]
|
| return name
|
|
|
|
|
| def sanitize_problem_name(problem: str) -> str:
|
| """
|
| Convert problem path to solution name suffix.
|
|
|
| Examples:
|
| >>> sanitize_problem_name("flash_attn")
|
| 'flash_attn'
|
| >>> sanitize_problem_name("gemm_optimization/squares")
|
| 'gemm_optimization_squares'
|
|
|
| Args:
|
| problem: Problem ID (may contain slashes)
|
|
|
| Returns:
|
| Sanitized problem name for use in solution directory names
|
| """
|
| return problem.replace("/", "_")
|
|
|
|
|
| def parse_solution_name(solution_name: str) -> Tuple[str, str, int]:
|
| """
|
| Parse a solution directory name into components.
|
|
|
| Examples:
|
| >>> parse_solution_name("gpt5_flash_attn")
|
| ('gpt5', 'flash_attn', 0)
|
| >>> parse_solution_name("gpt5_flash_attn_1")
|
| ('gpt5', 'flash_attn', 1)
|
| >>> parse_solution_name("claude4.5sonnet_gemm_optimization_squares_2")
|
| ('claude4.5sonnet', 'gemm_optimization_squares', 2)
|
|
|
| Args:
|
| solution_name: Solution directory name
|
|
|
| Returns:
|
| Tuple of (model_prefix, problem_slug, variant_index)
|
| """
|
|
|
| parts = solution_name.rsplit("_", 1)
|
| variant_index = 0
|
| base_name = solution_name
|
|
|
| if len(parts) == 2 and parts[1].isdigit():
|
| variant_index = int(parts[1])
|
| base_name = parts[0]
|
|
|
|
|
|
|
| first_underscore = base_name.find("_")
|
| if first_underscore == -1:
|
| return (base_name, "", variant_index)
|
|
|
| model_prefix = base_name[:first_underscore]
|
| problem_slug = base_name[first_underscore + 1:]
|
|
|
| return (model_prefix, problem_slug, variant_index)
|
|
|
|
|
| def build_solution_name(model: str, problem: str, variant_index: int = 0) -> str:
|
| """
|
| Build a solution directory name from components.
|
|
|
| Args:
|
| model: Model name (will be converted to prefix)
|
| problem: Problem ID
|
| variant_index: Variant index (0 = no suffix)
|
|
|
| Returns:
|
| Solution directory name
|
| """
|
| prefix = get_model_prefix(model)
|
| slug = sanitize_problem_name(problem)
|
| suffix = "" if variant_index == 0 else f"_{variant_index}"
|
| return f"{prefix}_{slug}{suffix}"
|
|
|
|
|
| def detect_provider(model: str) -> str:
|
| """
|
| Detect the LLM provider from model name.
|
|
|
| Args:
|
| model: Model name
|
|
|
| Returns:
|
| Provider name: 'openai', 'google', 'anthropic', 'xai', 'deepseek', 'openrouter'
|
| """
|
| normalized = model.strip()
|
| if "/" in normalized:
|
| provider_hint, actual_model = normalized.split("/", 1)
|
| else:
|
| provider_hint, actual_model = "", normalized
|
|
|
| provider_hint = provider_hint.lower()
|
| actual_lower = actual_model.lower()
|
|
|
| if (provider_hint in {"", "openai", "azure", "azure_openai"}) and actual_lower.startswith("gpt"):
|
| return "openai"
|
| if provider_hint in {"gemini", "google"} or "gemini" in actual_lower:
|
| return "google"
|
| if provider_hint == "anthropic" or "claude" in actual_lower:
|
| return "anthropic"
|
| if provider_hint == "xai" or "grok" in actual_lower:
|
| return "xai"
|
| if provider_hint == "deepseek" or "deepseek" in actual_lower:
|
| return "deepseek"
|
|
|
| return provider_hint or "openai"
|
|
|
|
|
| def is_reasoning_model(model: str, override: Optional[bool] = None) -> bool:
|
| """
|
| Determine if a model is a reasoning model.
|
|
|
| Args:
|
| model: Model name
|
| override: If set, use this value instead of auto-detection
|
|
|
| Returns:
|
| True if the model is a reasoning model
|
| """
|
| if override is not None:
|
| return override
|
|
|
| prefixes = ("gpt-5", "o1", "o3", "deepseek-reasoner")
|
| if any(model.startswith(p) for p in prefixes):
|
| return True
|
|
|
| normalized = model.lower()
|
| if "reasoning" in normalized and normalized.startswith("grok-"):
|
| return True
|
|
|
| return False
|
|
|