| | """ |
| | LLM utility functions for DeepCode project. |
| | |
| | This module provides common LLM-related utilities to avoid circular imports |
| | and reduce code duplication across the project. |
| | """ |
| |
|
| | import os |
| | import yaml |
| | from typing import Any, Type, Dict, Tuple |
| |
|
| | |
| | from mcp_agent.workflows.llm.augmented_llm_anthropic import AnthropicAugmentedLLM |
| | from mcp_agent.workflows.llm.augmented_llm_openai import OpenAIAugmentedLLM |
| |
|
| |
|
| | def get_preferred_llm_class(config_path: str = "mcp_agent.secrets.yaml") -> Type[Any]: |
| | """ |
| | Automatically select the LLM class based on API key availability in configuration. |
| | |
| | Reads from YAML config file and returns AnthropicAugmentedLLM if anthropic.api_key |
| | is available, otherwise returns OpenAIAugmentedLLM. |
| | |
| | Args: |
| | config_path: Path to the YAML configuration file |
| | |
| | Returns: |
| | class: The preferred LLM class |
| | """ |
| | try: |
| | |
| | if os.path.exists(config_path): |
| | with open(config_path, "r", encoding="utf-8") as f: |
| | config = yaml.safe_load(f) |
| |
|
| | |
| | anthropic_config = config.get("anthropic", {}) |
| | anthropic_key = anthropic_config.get("api_key", "") |
| |
|
| | if anthropic_key and anthropic_key.strip() and not anthropic_key == "": |
| | |
| | return AnthropicAugmentedLLM |
| | else: |
| | |
| | return OpenAIAugmentedLLM |
| | else: |
| | print(f"π€ Config file {config_path} not found, using OpenAIAugmentedLLM") |
| | return OpenAIAugmentedLLM |
| |
|
| | except Exception as e: |
| | print(f"π€ Error reading config file {config_path}: {e}") |
| | print("π€ Falling back to OpenAIAugmentedLLM") |
| | return OpenAIAugmentedLLM |
| |
|
| |
|
| | def get_default_models(config_path: str = "mcp_agent.config.yaml"): |
| | """ |
| | Get default models from configuration file. |
| | |
| | Args: |
| | config_path: Path to the configuration file |
| | |
| | Returns: |
| | dict: Dictionary with 'anthropic' and 'openai' default models |
| | """ |
| | try: |
| | if os.path.exists(config_path): |
| | with open(config_path, "r", encoding="utf-8") as f: |
| | config = yaml.safe_load(f) |
| |
|
| | |
| | anthropic_config = config.get("anthropic") or {} |
| | openai_config = config.get("openai") or {} |
| |
|
| | anthropic_model = anthropic_config.get( |
| | "default_model", "claude-sonnet-4-20250514" |
| | ) |
| | openai_model = openai_config.get("default_model", "o3-mini") |
| |
|
| | return {"anthropic": anthropic_model, "openai": openai_model} |
| | else: |
| | print(f"Config file {config_path} not found, using default models") |
| | return {"anthropic": "claude-sonnet-4-20250514", "openai": "o3-mini"} |
| |
|
| | except Exception as e: |
| | print(f"βError reading config file {config_path}: {e}") |
| | return {"anthropic": "claude-sonnet-4-20250514", "openai": "o3-mini"} |
| |
|
| |
|
| | def get_document_segmentation_config( |
| | config_path: str = "mcp_agent.config.yaml", |
| | ) -> Dict[str, Any]: |
| | """ |
| | Get document segmentation configuration from config file. |
| | |
| | Args: |
| | config_path: Path to the main configuration file |
| | |
| | Returns: |
| | Dict containing segmentation configuration with default values |
| | """ |
| | try: |
| | if os.path.exists(config_path): |
| | with open(config_path, "r", encoding="utf-8") as f: |
| | config = yaml.safe_load(f) |
| |
|
| | |
| | seg_config = config.get("document_segmentation", {}) |
| | return { |
| | "enabled": seg_config.get("enabled", True), |
| | "size_threshold_chars": seg_config.get("size_threshold_chars", 50000), |
| | } |
| | else: |
| | print( |
| | f"π Config file {config_path} not found, using default segmentation settings" |
| | ) |
| | return {"enabled": True, "size_threshold_chars": 50000} |
| |
|
| | except Exception as e: |
| | print(f"π Error reading segmentation config from {config_path}: {e}") |
| | print("π Using default segmentation settings") |
| | return {"enabled": True, "size_threshold_chars": 50000} |
| |
|
| |
|
| | def should_use_document_segmentation( |
| | document_content: str, config_path: str = "mcp_agent.config.yaml" |
| | ) -> Tuple[bool, str]: |
| | """ |
| | Determine whether to use document segmentation based on configuration and document size. |
| | |
| | Args: |
| | document_content: The content of the document to analyze |
| | config_path: Path to the configuration file |
| | |
| | Returns: |
| | Tuple of (should_segment, reason) where: |
| | - should_segment: Boolean indicating whether to use segmentation |
| | - reason: String explaining the decision |
| | """ |
| | seg_config = get_document_segmentation_config(config_path) |
| |
|
| | if not seg_config["enabled"]: |
| | return False, "Document segmentation disabled in configuration" |
| |
|
| | doc_size = len(document_content) |
| | threshold = seg_config["size_threshold_chars"] |
| |
|
| | if doc_size > threshold: |
| | return ( |
| | True, |
| | f"Document size ({doc_size:,} chars) exceeds threshold ({threshold:,} chars)", |
| | ) |
| | else: |
| | return ( |
| | False, |
| | f"Document size ({doc_size:,} chars) below threshold ({threshold:,} chars)", |
| | ) |
| |
|
| |
|
| | def get_adaptive_agent_config( |
| | use_segmentation: bool, search_server_names: list = None |
| | ) -> Dict[str, list]: |
| | """ |
| | Get adaptive agent configuration based on whether to use document segmentation. |
| | |
| | Args: |
| | use_segmentation: Whether to include document-segmentation server |
| | search_server_names: Base search server names (from get_search_server_names) |
| | |
| | Returns: |
| | Dict containing server configurations for different agents |
| | """ |
| | if search_server_names is None: |
| | search_server_names = [] |
| |
|
| | |
| | config = { |
| | "concept_analysis": [], |
| | "algorithm_analysis": search_server_names.copy(), |
| | "code_planner": search_server_names.copy(), |
| | } |
| |
|
| | |
| | if use_segmentation: |
| | config["concept_analysis"] = ["document-segmentation"] |
| | if "document-segmentation" not in config["algorithm_analysis"]: |
| | config["algorithm_analysis"].append("document-segmentation") |
| | if "document-segmentation" not in config["code_planner"]: |
| | config["code_planner"].append("document-segmentation") |
| | else: |
| | config["concept_analysis"] = ["filesystem"] |
| | if "filesystem" not in config["algorithm_analysis"]: |
| | config["algorithm_analysis"].append("filesystem") |
| | if "filesystem" not in config["code_planner"]: |
| | config["code_planner"].append("filesystem") |
| |
|
| | return config |
| |
|
| |
|
| | def get_adaptive_prompts(use_segmentation: bool) -> Dict[str, str]: |
| | """ |
| | Get appropriate prompt versions based on segmentation usage. |
| | |
| | Args: |
| | use_segmentation: Whether to use segmented reading prompts |
| | |
| | Returns: |
| | Dict containing prompt configurations |
| | """ |
| | |
| | from prompts.code_prompts import ( |
| | PAPER_CONCEPT_ANALYSIS_PROMPT, |
| | PAPER_ALGORITHM_ANALYSIS_PROMPT, |
| | CODE_PLANNING_PROMPT, |
| | PAPER_CONCEPT_ANALYSIS_PROMPT_TRADITIONAL, |
| | PAPER_ALGORITHM_ANALYSIS_PROMPT_TRADITIONAL, |
| | CODE_PLANNING_PROMPT_TRADITIONAL, |
| | ) |
| |
|
| | if use_segmentation: |
| | return { |
| | "concept_analysis": PAPER_CONCEPT_ANALYSIS_PROMPT, |
| | "algorithm_analysis": PAPER_ALGORITHM_ANALYSIS_PROMPT, |
| | "code_planning": CODE_PLANNING_PROMPT, |
| | } |
| | else: |
| | return { |
| | "concept_analysis": PAPER_CONCEPT_ANALYSIS_PROMPT_TRADITIONAL, |
| | "algorithm_analysis": PAPER_ALGORITHM_ANALYSIS_PROMPT_TRADITIONAL, |
| | "code_planning": CODE_PLANNING_PROMPT_TRADITIONAL, |
| | } |
| |
|