|
|
""" |
|
|
LLM utility functions for DeepCode project. |
|
|
|
|
|
This module provides common LLM-related utilities to avoid circular imports |
|
|
and reduce code duplication across the project. |
|
|
""" |
|
|
|
|
|
import os |
|
|
import yaml |
|
|
from typing import Any, Type, Dict, Tuple |
|
|
|
|
|
|
|
|
from mcp_agent.workflows.llm.augmented_llm_anthropic import AnthropicAugmentedLLM |
|
|
from mcp_agent.workflows.llm.augmented_llm_openai import OpenAIAugmentedLLM |
|
|
|
|
|
|
|
|
def get_preferred_llm_class(config_path: str = "mcp_agent.secrets.yaml") -> Type[Any]: |
|
|
""" |
|
|
Automatically select the LLM class based on API key availability in configuration. |
|
|
|
|
|
Reads from YAML config file and returns AnthropicAugmentedLLM if anthropic.api_key |
|
|
is available, otherwise returns OpenAIAugmentedLLM. |
|
|
|
|
|
Args: |
|
|
config_path: Path to the YAML configuration file |
|
|
|
|
|
Returns: |
|
|
class: The preferred LLM class |
|
|
""" |
|
|
try: |
|
|
|
|
|
if os.path.exists(config_path): |
|
|
with open(config_path, "r", encoding="utf-8") as f: |
|
|
config = yaml.safe_load(f) |
|
|
|
|
|
|
|
|
anthropic_config = config.get("anthropic", {}) |
|
|
anthropic_key = anthropic_config.get("api_key", "") |
|
|
|
|
|
if anthropic_key and anthropic_key.strip() and not anthropic_key == "": |
|
|
|
|
|
return AnthropicAugmentedLLM |
|
|
else: |
|
|
|
|
|
return OpenAIAugmentedLLM |
|
|
else: |
|
|
print(f"π€ Config file {config_path} not found, using OpenAIAugmentedLLM") |
|
|
return OpenAIAugmentedLLM |
|
|
|
|
|
except Exception as e: |
|
|
print(f"π€ Error reading config file {config_path}: {e}") |
|
|
print("π€ Falling back to OpenAIAugmentedLLM") |
|
|
return OpenAIAugmentedLLM |
|
|
|
|
|
|
|
|
def get_default_models(config_path: str = "mcp_agent.config.yaml"): |
|
|
""" |
|
|
Get default models from configuration file. |
|
|
|
|
|
Args: |
|
|
config_path: Path to the configuration file |
|
|
|
|
|
Returns: |
|
|
dict: Dictionary with 'anthropic' and 'openai' default models |
|
|
""" |
|
|
try: |
|
|
if os.path.exists(config_path): |
|
|
with open(config_path, "r", encoding="utf-8") as f: |
|
|
config = yaml.safe_load(f) |
|
|
|
|
|
|
|
|
anthropic_config = config.get("anthropic") or {} |
|
|
openai_config = config.get("openai") or {} |
|
|
|
|
|
anthropic_model = anthropic_config.get( |
|
|
"default_model", "claude-sonnet-4-20250514" |
|
|
) |
|
|
openai_model = openai_config.get("default_model", "o3-mini") |
|
|
|
|
|
return {"anthropic": anthropic_model, "openai": openai_model} |
|
|
else: |
|
|
print(f"Config file {config_path} not found, using default models") |
|
|
return {"anthropic": "claude-sonnet-4-20250514", "openai": "o3-mini"} |
|
|
|
|
|
except Exception as e: |
|
|
print(f"βError reading config file {config_path}: {e}") |
|
|
return {"anthropic": "claude-sonnet-4-20250514", "openai": "o3-mini"} |
|
|
|
|
|
|
|
|
def get_document_segmentation_config( |
|
|
config_path: str = "mcp_agent.config.yaml", |
|
|
) -> Dict[str, Any]: |
|
|
""" |
|
|
Get document segmentation configuration from config file. |
|
|
|
|
|
Args: |
|
|
config_path: Path to the main configuration file |
|
|
|
|
|
Returns: |
|
|
Dict containing segmentation configuration with default values |
|
|
""" |
|
|
try: |
|
|
if os.path.exists(config_path): |
|
|
with open(config_path, "r", encoding="utf-8") as f: |
|
|
config = yaml.safe_load(f) |
|
|
|
|
|
|
|
|
seg_config = config.get("document_segmentation", {}) |
|
|
return { |
|
|
"enabled": seg_config.get("enabled", True), |
|
|
"size_threshold_chars": seg_config.get("size_threshold_chars", 50000), |
|
|
} |
|
|
else: |
|
|
print( |
|
|
f"π Config file {config_path} not found, using default segmentation settings" |
|
|
) |
|
|
return {"enabled": True, "size_threshold_chars": 50000} |
|
|
|
|
|
except Exception as e: |
|
|
print(f"π Error reading segmentation config from {config_path}: {e}") |
|
|
print("π Using default segmentation settings") |
|
|
return {"enabled": True, "size_threshold_chars": 50000} |
|
|
|
|
|
|
|
|
def should_use_document_segmentation( |
|
|
document_content: str, config_path: str = "mcp_agent.config.yaml" |
|
|
) -> Tuple[bool, str]: |
|
|
""" |
|
|
Determine whether to use document segmentation based on configuration and document size. |
|
|
|
|
|
Args: |
|
|
document_content: The content of the document to analyze |
|
|
config_path: Path to the configuration file |
|
|
|
|
|
Returns: |
|
|
Tuple of (should_segment, reason) where: |
|
|
- should_segment: Boolean indicating whether to use segmentation |
|
|
- reason: String explaining the decision |
|
|
""" |
|
|
seg_config = get_document_segmentation_config(config_path) |
|
|
|
|
|
if not seg_config["enabled"]: |
|
|
return False, "Document segmentation disabled in configuration" |
|
|
|
|
|
doc_size = len(document_content) |
|
|
threshold = seg_config["size_threshold_chars"] |
|
|
|
|
|
if doc_size > threshold: |
|
|
return ( |
|
|
True, |
|
|
f"Document size ({doc_size:,} chars) exceeds threshold ({threshold:,} chars)", |
|
|
) |
|
|
else: |
|
|
return ( |
|
|
False, |
|
|
f"Document size ({doc_size:,} chars) below threshold ({threshold:,} chars)", |
|
|
) |
|
|
|
|
|
|
|
|
def get_adaptive_agent_config( |
|
|
use_segmentation: bool, search_server_names: list = None |
|
|
) -> Dict[str, list]: |
|
|
""" |
|
|
Get adaptive agent configuration based on whether to use document segmentation. |
|
|
|
|
|
Args: |
|
|
use_segmentation: Whether to include document-segmentation server |
|
|
search_server_names: Base search server names (from get_search_server_names) |
|
|
|
|
|
Returns: |
|
|
Dict containing server configurations for different agents |
|
|
""" |
|
|
if search_server_names is None: |
|
|
search_server_names = [] |
|
|
|
|
|
|
|
|
config = { |
|
|
"concept_analysis": [], |
|
|
"algorithm_analysis": search_server_names.copy(), |
|
|
"code_planner": search_server_names.copy(), |
|
|
} |
|
|
|
|
|
|
|
|
if use_segmentation: |
|
|
config["concept_analysis"] = ["document-segmentation"] |
|
|
if "document-segmentation" not in config["algorithm_analysis"]: |
|
|
config["algorithm_analysis"].append("document-segmentation") |
|
|
if "document-segmentation" not in config["code_planner"]: |
|
|
config["code_planner"].append("document-segmentation") |
|
|
else: |
|
|
config["concept_analysis"] = ["filesystem"] |
|
|
if "filesystem" not in config["algorithm_analysis"]: |
|
|
config["algorithm_analysis"].append("filesystem") |
|
|
if "filesystem" not in config["code_planner"]: |
|
|
config["code_planner"].append("filesystem") |
|
|
|
|
|
return config |
|
|
|
|
|
|
|
|
def get_adaptive_prompts(use_segmentation: bool) -> Dict[str, str]: |
|
|
""" |
|
|
Get appropriate prompt versions based on segmentation usage. |
|
|
|
|
|
Args: |
|
|
use_segmentation: Whether to use segmented reading prompts |
|
|
|
|
|
Returns: |
|
|
Dict containing prompt configurations |
|
|
""" |
|
|
|
|
|
from prompts.code_prompts import ( |
|
|
PAPER_CONCEPT_ANALYSIS_PROMPT, |
|
|
PAPER_ALGORITHM_ANALYSIS_PROMPT, |
|
|
CODE_PLANNING_PROMPT, |
|
|
PAPER_CONCEPT_ANALYSIS_PROMPT_TRADITIONAL, |
|
|
PAPER_ALGORITHM_ANALYSIS_PROMPT_TRADITIONAL, |
|
|
CODE_PLANNING_PROMPT_TRADITIONAL, |
|
|
) |
|
|
|
|
|
if use_segmentation: |
|
|
return { |
|
|
"concept_analysis": PAPER_CONCEPT_ANALYSIS_PROMPT, |
|
|
"algorithm_analysis": PAPER_ALGORITHM_ANALYSIS_PROMPT, |
|
|
"code_planning": CODE_PLANNING_PROMPT, |
|
|
} |
|
|
else: |
|
|
return { |
|
|
"concept_analysis": PAPER_CONCEPT_ANALYSIS_PROMPT_TRADITIONAL, |
|
|
"algorithm_analysis": PAPER_ALGORITHM_ANALYSIS_PROMPT_TRADITIONAL, |
|
|
"code_planning": CODE_PLANNING_PROMPT_TRADITIONAL, |
|
|
} |
|
|
|