adaptai / projects /ui /DeepCode /workflows /codebase_index_workflow.py
ADAPT-Chase's picture
Add files using upload-large-folder tool
af1432a verified
"""
Codebase Index Workflow
This workflow integrates the functionality of run_indexer.py and code_indexer.py
to build intelligent relationships between existing codebase and target structure.
Features:
- Extract target file structure from initial_plan.txt
- Analyze codebase and build indexes
- Generate relationship mappings and statistical reports
- Provide reference basis for code reproduction
"""
import asyncio
import json
import logging
import os
import re
import sys
from pathlib import Path
from typing import Dict, Any, Optional
import yaml
# Add tools directory to path
sys.path.append(str(Path(__file__).parent.parent / "tools"))
from tools.code_indexer import CodeIndexer
class CodebaseIndexWorkflow:
"""Codebase Index Workflow Class"""
def __init__(self, logger=None):
"""
Initialize workflow
Args:
logger: Logger instance
"""
self.logger = logger or self._setup_default_logger()
self.indexer = None
def _setup_default_logger(self) -> logging.Logger:
"""Setup default logger"""
logger = logging.getLogger("CodebaseIndexWorkflow")
logger.setLevel(logging.INFO)
if not logger.handlers:
handler = logging.StreamHandler()
formatter = logging.Formatter(
"%(asctime)s - %(name)s - %(levelname)s - %(message)s"
)
handler.setFormatter(formatter)
logger.addHandler(handler)
return logger
def extract_file_tree_from_plan(self, plan_content: str) -> Optional[str]:
"""
Extract file tree structure from initial_plan.txt content
Args:
plan_content: Content of the initial_plan.txt file
Returns:
Extracted file tree structure as string
"""
# Look for file structure section, specifically "## File Structure" format
file_structure_pattern = r"## File Structure[^\n]*\n```[^\n]*\n(.*?)\n```"
match = re.search(file_structure_pattern, plan_content, re.DOTALL)
if match:
file_tree = match.group(1).strip()
lines = file_tree.split("\n")
# Clean tree structure - remove empty lines and comments not part of structure
cleaned_lines = []
for line in lines:
# Keep tree structure lines
if line.strip() and (
any(char in line for char in ["β”œβ”€β”€", "└──", "β”‚"])
or line.strip().endswith("/")
or "." in line.split("/")[-1] # has file extension
or line.strip().endswith(".py")
or line.strip().endswith(".txt")
or line.strip().endswith(".md")
or line.strip().endswith(".yaml")
):
cleaned_lines.append(line)
if len(cleaned_lines) >= 5:
file_tree = "\n".join(cleaned_lines)
self.logger.info(
f"πŸ“Š Extracted file tree structure from ## File Structure section ({len(cleaned_lines)} lines)"
)
return file_tree
# Fallback: look for any code block containing project structure
code_block_patterns = [
r"```[^\n]*\n(project/.*?(?:β”œβ”€β”€|└──).*?)\n```",
r"```[^\n]*\n(src/.*?(?:β”œβ”€β”€|└──).*?)\n```",
r"```[^\n]*\n(core/.*?(?:β”œβ”€β”€|└──).*?)\n```",
r"```[^\n]*\n(.*?(?:β”œβ”€β”€|└──).*?(?:\.py|\.txt|\.md|\.yaml).*?)\n```",
]
for pattern in code_block_patterns:
match = re.search(pattern, plan_content, re.DOTALL)
if match:
file_tree = match.group(1).strip()
lines = [line for line in file_tree.split("\n") if line.strip()]
if len(lines) >= 5:
self.logger.info(
f"πŸ“Š Extracted file tree structure from code block ({len(lines)} lines)"
)
return file_tree
# Final fallback: extract file paths from file mentions and create basic structure
self.logger.warning(
"⚠️ No standard file tree found, trying to extract from file mentions..."
)
# Search for file paths in backticks throughout the document
file_mentions = re.findall(
r"`([^`]*(?:\.py|\.txt|\.md|\.yaml|\.yml)[^`]*)`", plan_content
)
if file_mentions:
# Organize files into directory structure
dirs = set()
files_by_dir = {}
for file_path in file_mentions:
file_path = file_path.strip()
if "/" in file_path:
dir_path = "/".join(file_path.split("/")[:-1])
filename = file_path.split("/")[-1]
dirs.add(dir_path)
if dir_path not in files_by_dir:
files_by_dir[dir_path] = []
files_by_dir[dir_path].append(filename)
else:
if "root" not in files_by_dir:
files_by_dir["root"] = []
files_by_dir["root"].append(file_path)
# Create tree structure
structure_lines = []
# Determine root directory name from common patterns
if any("src/" in f for f in file_mentions):
root_name = "src"
elif any("core/" in f for f in file_mentions):
root_name = "core"
elif any("lib/" in f for f in file_mentions):
root_name = "lib"
else:
root_name = "project"
structure_lines.append(f"{root_name}/")
# Add directories and files
sorted_dirs = sorted(dirs) if dirs else []
for i, dir_path in enumerate(sorted_dirs):
is_last_dir = i == len(sorted_dirs) - 1
prefix = "└──" if is_last_dir else "β”œβ”€β”€"
structure_lines.append(f"{prefix} {dir_path}/")
if dir_path in files_by_dir:
files = sorted(files_by_dir[dir_path])
for j, filename in enumerate(files):
is_last_file = j == len(files) - 1
if is_last_dir:
file_prefix = " └──" if is_last_file else " β”œβ”€β”€"
else:
file_prefix = "β”‚ └──" if is_last_file else "β”‚ β”œβ”€β”€"
structure_lines.append(f"{file_prefix} {filename}")
# Add root files (if any)
if "root" in files_by_dir:
root_files = sorted(files_by_dir["root"])
for i, filename in enumerate(root_files):
is_last = (i == len(root_files) - 1) and not sorted_dirs
prefix = "└──" if is_last else "β”œβ”€β”€"
structure_lines.append(f"{prefix} {filename}")
if len(structure_lines) >= 3:
file_tree = "\n".join(structure_lines)
self.logger.info(
f"πŸ“Š Generated file tree from file mentions ({len(structure_lines)} lines)"
)
return file_tree
# If no file tree found, return None
self.logger.warning("⚠️ No file tree structure found in initial plan")
return None
def load_target_structure_from_plan(self, plan_path: str) -> str:
"""
Load target structure from initial_plan.txt and extract file tree
Args:
plan_path: Path to initial_plan.txt file
Returns:
Extracted file tree structure
"""
try:
# Load complete plan content
with open(plan_path, "r", encoding="utf-8") as f:
plan_content = f.read()
self.logger.info(f"πŸ“„ Loaded initial plan ({len(plan_content)} characters)")
# Extract file tree structure
file_tree = self.extract_file_tree_from_plan(plan_content)
if file_tree:
self.logger.info(
"βœ… Successfully extracted file tree from initial plan"
)
self.logger.info("πŸ“‹ Extracted structure preview:")
# Show first few lines of extracted tree
preview_lines = file_tree.split("\n")[:8]
for line in preview_lines:
self.logger.info(f" {line}")
if len(file_tree.split("\n")) > 8:
self.logger.info(
f" ... {len(file_tree.split('\n')) - 8} more lines"
)
return file_tree
else:
self.logger.warning("⚠️ Unable to extract file tree from initial plan")
self.logger.info("πŸ”„ Falling back to default target structure")
return self.get_default_target_structure()
except Exception as e:
self.logger.error(f"❌ Failed to load initial plan file {plan_path}: {e}")
self.logger.info("πŸ”„ Falling back to default target structure")
return self.get_default_target_structure()
def get_default_target_structure(self) -> str:
"""Get default target structure"""
return """
project/
β”œβ”€β”€ src/
β”‚ β”œβ”€β”€ core/
β”‚ β”‚ β”œβ”€β”€ gcn.py # GCN encoder
β”‚ β”‚ β”œβ”€β”€ diffusion.py # forward/reverse processes
β”‚ β”‚ β”œβ”€β”€ denoiser.py # denoising MLP
β”‚ β”‚ └── fusion.py # fusion combiner
β”‚ β”œβ”€β”€ models/ # model wrapper classes
β”‚ β”‚ └── recdiff.py
β”‚ β”œβ”€β”€ utils/
β”‚ β”‚ β”œβ”€β”€ data.py # loading & preprocessing
β”‚ β”‚ β”œβ”€β”€ predictor.py # scoring functions
β”‚ β”‚ β”œβ”€β”€ loss.py # loss functions
β”‚ β”‚ β”œβ”€β”€ metrics.py # NDCG, Recall etc.
β”‚ β”‚ └── sched.py # beta/alpha schedule utils
β”‚ └── configs/
β”‚ └── default.yaml # hyperparameters, paths
β”œβ”€β”€ tests/
β”‚ β”œβ”€β”€ test_gcn.py
β”‚ β”œβ”€β”€ test_diffusion.py
β”‚ β”œβ”€β”€ test_denoiser.py
β”‚ β”œβ”€β”€ test_loss.py
β”‚ └── test_pipeline.py
β”œβ”€β”€ docs/
β”‚ β”œβ”€β”€ architecture.md
β”‚ β”œβ”€β”€ api_reference.md
β”‚ └── README.md
β”œβ”€β”€ experiments/
β”‚ β”œβ”€β”€ run_experiment.py
β”‚ └── notebooks/
β”‚ └── analysis.ipynb
β”œβ”€β”€ requirements.txt
└── setup.py
"""
def load_or_create_indexer_config(self, paper_dir: str) -> Dict[str, Any]:
"""
Load or create indexer configuration
Args:
paper_dir: Paper directory path
Returns:
Configuration dictionary
"""
# Try to load existing configuration file
config_path = Path(__file__).parent.parent / "tools" / "indexer_config.yaml"
try:
if config_path.exists():
with open(config_path, "r", encoding="utf-8") as f:
config = yaml.safe_load(f)
# Update path configuration to current paper directory
if "paths" not in config:
config["paths"] = {}
config["paths"]["code_base_path"] = os.path.join(paper_dir, "code_base")
config["paths"]["output_dir"] = os.path.join(paper_dir, "indexes")
# Adjust performance settings for workflow
if "performance" in config:
config["performance"]["enable_concurrent_analysis"] = (
False # Disable concurrency to avoid API limits
)
if "debug" in config:
config["debug"]["verbose_output"] = True # Enable verbose output
if "llm" in config:
config["llm"]["request_delay"] = 0.5 # Increase request delay
self.logger.info(f"Loaded configuration file: {config_path}")
return config
except Exception as e:
self.logger.warning(f"Failed to load configuration file: {e}")
# If loading fails, use default configuration
self.logger.info("Using default configuration")
default_config = {
"paths": {
"code_base_path": os.path.join(paper_dir, "code_base"),
"output_dir": os.path.join(paper_dir, "indexes"),
},
"llm": {
"model_provider": "anthropic",
"max_tokens": 4000,
"temperature": 0.3,
"request_delay": 0.5, # Increase request delay
"max_retries": 3,
"retry_delay": 1.0,
},
"file_analysis": {
"max_file_size": 1048576, # 1MB
"max_content_length": 3000,
"supported_extensions": [
".py",
".js",
".ts",
".java",
".cpp",
".c",
".h",
".hpp",
".cs",
".php",
".rb",
".go",
".rs",
".scala",
".kt",
".yaml",
".yml",
".json",
".xml",
".toml",
".md",
".txt",
],
"skip_directories": [
"__pycache__",
"node_modules",
"target",
"build",
"dist",
"venv",
"env",
".git",
".svn",
"data",
"datasets",
],
},
"relationships": {
"min_confidence_score": 0.3,
"high_confidence_threshold": 0.7,
"relationship_types": {
"direct_match": 1.0,
"partial_match": 0.8,
"reference": 0.6,
"utility": 0.4,
},
},
"performance": {
"enable_concurrent_analysis": False, # Disable concurrency to avoid API limits
"max_concurrent_files": 3,
"enable_content_caching": True,
"max_cache_size": 100,
},
"debug": {
"verbose_output": True,
"save_raw_responses": False,
"mock_llm_responses": False,
},
"output": {
"generate_summary": True,
"generate_statistics": True,
"include_metadata": True,
"json_indent": 2,
},
"logging": {"level": "INFO", "log_to_file": False},
}
return default_config
async def run_indexing_workflow(
self,
paper_dir: str,
initial_plan_path: Optional[str] = None,
config_path: str = "mcp_agent.secrets.yaml",
) -> Dict[str, Any]:
"""
Run the complete code indexing workflow
Args:
paper_dir: Paper directory path
initial_plan_path: Initial plan file path (optional)
config_path: API configuration file path
Returns:
Index result dictionary
"""
try:
self.logger.info("πŸš€ Starting codebase index workflow...")
# Step 1: Determine initial plan file path
if not initial_plan_path:
initial_plan_path = os.path.join(paper_dir, "initial_plan.txt")
# Step 2: Load target structure
if os.path.exists(initial_plan_path):
self.logger.info(
f"πŸ“ Loading target structure from {initial_plan_path}"
)
target_structure = self.load_target_structure_from_plan(
initial_plan_path
)
else:
self.logger.warning(
f"⚠️ Initial plan file does not exist: {initial_plan_path}"
)
self.logger.info("πŸ“ Using default target structure")
target_structure = self.get_default_target_structure()
# Step 3: Check codebase path
code_base_path = os.path.join(paper_dir, "code_base")
if not os.path.exists(code_base_path):
self.logger.error(f"❌ Codebase path does not exist: {code_base_path}")
return {
"status": "error",
"message": f"Code base path does not exist: {code_base_path}",
"output_files": {},
}
# Step 4: Create output directory
output_dir = os.path.join(paper_dir, "indexes")
os.makedirs(output_dir, exist_ok=True)
# Step 5: Load configuration
indexer_config = self.load_or_create_indexer_config(paper_dir)
self.logger.info(f"πŸ“ Codebase path: {code_base_path}")
self.logger.info(f"πŸ“€ Output directory: {output_dir}")
# Step 6: Create code indexer
self.indexer = CodeIndexer(
code_base_path=code_base_path,
target_structure=target_structure,
output_dir=output_dir,
config_path=config_path,
enable_pre_filtering=True,
)
# Apply configuration settings
self.indexer.indexer_config = indexer_config
# Directly set configuration attributes to indexer
if "file_analysis" in indexer_config:
file_config = indexer_config["file_analysis"]
self.indexer.supported_extensions = set(
file_config.get(
"supported_extensions", self.indexer.supported_extensions
)
)
self.indexer.skip_directories = set(
file_config.get("skip_directories", self.indexer.skip_directories)
)
self.indexer.max_file_size = file_config.get(
"max_file_size", self.indexer.max_file_size
)
self.indexer.max_content_length = file_config.get(
"max_content_length", self.indexer.max_content_length
)
if "llm" in indexer_config:
llm_config = indexer_config["llm"]
self.indexer.model_provider = llm_config.get(
"model_provider", self.indexer.model_provider
)
self.indexer.llm_max_tokens = llm_config.get(
"max_tokens", self.indexer.llm_max_tokens
)
self.indexer.llm_temperature = llm_config.get(
"temperature", self.indexer.llm_temperature
)
self.indexer.request_delay = llm_config.get(
"request_delay", self.indexer.request_delay
)
self.indexer.max_retries = llm_config.get(
"max_retries", self.indexer.max_retries
)
self.indexer.retry_delay = llm_config.get(
"retry_delay", self.indexer.retry_delay
)
if "relationships" in indexer_config:
rel_config = indexer_config["relationships"]
self.indexer.min_confidence_score = rel_config.get(
"min_confidence_score", self.indexer.min_confidence_score
)
self.indexer.high_confidence_threshold = rel_config.get(
"high_confidence_threshold", self.indexer.high_confidence_threshold
)
self.indexer.relationship_types = rel_config.get(
"relationship_types", self.indexer.relationship_types
)
if "performance" in indexer_config:
perf_config = indexer_config["performance"]
self.indexer.enable_concurrent_analysis = perf_config.get(
"enable_concurrent_analysis",
self.indexer.enable_concurrent_analysis,
)
self.indexer.max_concurrent_files = perf_config.get(
"max_concurrent_files", self.indexer.max_concurrent_files
)
self.indexer.enable_content_caching = perf_config.get(
"enable_content_caching", self.indexer.enable_content_caching
)
self.indexer.max_cache_size = perf_config.get(
"max_cache_size", self.indexer.max_cache_size
)
if "debug" in indexer_config:
debug_config = indexer_config["debug"]
self.indexer.verbose_output = debug_config.get(
"verbose_output", self.indexer.verbose_output
)
self.indexer.save_raw_responses = debug_config.get(
"save_raw_responses", self.indexer.save_raw_responses
)
self.indexer.mock_llm_responses = debug_config.get(
"mock_llm_responses", self.indexer.mock_llm_responses
)
if "output" in indexer_config:
output_config = indexer_config["output"]
self.indexer.generate_summary = output_config.get(
"generate_summary", self.indexer.generate_summary
)
self.indexer.generate_statistics = output_config.get(
"generate_statistics", self.indexer.generate_statistics
)
self.indexer.include_metadata = output_config.get(
"include_metadata", self.indexer.include_metadata
)
self.logger.info("πŸ”§ Indexer configuration completed")
self.logger.info(f"πŸ€– Model provider: {self.indexer.model_provider}")
self.logger.info(
f"⚑ Concurrent analysis: {'Enabled' if self.indexer.enable_concurrent_analysis else 'Disabled'}"
)
self.logger.info(
f"πŸ—„οΈ Content caching: {'Enabled' if self.indexer.enable_content_caching else 'Disabled'}"
)
self.logger.info(
f"πŸ” Pre-filtering: {'Enabled' if self.indexer.enable_pre_filtering else 'Disabled'}"
)
self.logger.info("=" * 60)
self.logger.info("πŸš€ Starting code indexing process...")
# Step 7: Build all indexes
output_files = await self.indexer.build_all_indexes()
# Step 8: Generate summary report
if output_files:
summary_report = self.indexer.generate_summary_report(output_files)
self.logger.info("=" * 60)
self.logger.info("βœ… Indexing completed successfully!")
self.logger.info(f"πŸ“Š Processed {len(output_files)} repositories")
self.logger.info("πŸ“ Generated index files:")
for repo_name, file_path in output_files.items():
self.logger.info(f" πŸ“„ {repo_name}: {file_path}")
self.logger.info(f"πŸ“‹ Summary report: {summary_report}")
# Statistics (if enabled)
if self.indexer.generate_statistics:
self.logger.info("\nπŸ“ˆ Processing statistics:")
total_relationships = 0
high_confidence_relationships = 0
for file_path in output_files.values():
try:
with open(file_path, "r", encoding="utf-8") as f:
index_data = json.load(f)
relationships = index_data.get("relationships", [])
total_relationships += len(relationships)
high_confidence_relationships += len(
[
r
for r in relationships
if r.get("confidence_score", 0)
> self.indexer.high_confidence_threshold
]
)
except Exception as e:
self.logger.warning(
f" ⚠️ Unable to load statistics from {file_path}: {e}"
)
self.logger.info(
f" πŸ”— Total relationships found: {total_relationships}"
)
self.logger.info(
f" ⭐ High confidence relationships: {high_confidence_relationships}"
)
self.logger.info(
f" πŸ“Š Average relationships per repository: {total_relationships / len(output_files) if output_files else 0:.1f}"
)
self.logger.info("\nπŸŽ‰ Code indexing process completed successfully!")
return {
"status": "success",
"message": f"Successfully indexed {len(output_files)} repositories",
"output_files": output_files,
"summary_report": summary_report,
"statistics": {
"total_repositories": len(output_files),
"total_relationships": total_relationships,
"high_confidence_relationships": high_confidence_relationships,
}
if self.indexer.generate_statistics
else None,
}
else:
self.logger.warning("⚠️ No index files generated")
return {
"status": "warning",
"message": "No index files were generated",
"output_files": {},
}
except Exception as e:
self.logger.error(f"❌ Index workflow failed: {e}")
# If there are detailed error messages, log them
import traceback
self.logger.error(f"Detailed error information: {traceback.format_exc()}")
return {"status": "error", "message": str(e), "output_files": {}}
def print_banner(self):
"""Print application banner"""
banner = """
╔═══════════════════════════════════════════════════════════════════════╗
β•‘ πŸ” Codebase Index Workflow v1.0 β•‘
β•‘ Intelligent Code Relationship Analysis Tool β•‘
╠═══════════════════════════════════════════════════════════════════════╣
β•‘ πŸ“ Analyzes existing codebases β•‘
β•‘ πŸ”— Builds intelligent relationships with target structure β•‘
β•‘ πŸ€– Powered by LLM analysis β•‘
β•‘ πŸ“Š Generates detailed JSON indexes β•‘
β•‘ 🎯 Provides reference for code reproduction β•‘
β•šβ•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•
"""
print(banner)
# Convenience function for direct workflow invocation
async def run_codebase_indexing(
paper_dir: str,
initial_plan_path: Optional[str] = None,
config_path: str = "mcp_agent.secrets.yaml",
logger=None,
) -> Dict[str, Any]:
"""
Convenience function to run codebase indexing
Args:
paper_dir: Paper directory path
initial_plan_path: Initial plan file path (optional)
config_path: API configuration file path
logger: Logger instance (optional)
Returns:
Index result dictionary
"""
workflow = CodebaseIndexWorkflow(logger=logger)
workflow.print_banner()
return await workflow.run_indexing_workflow(
paper_dir=paper_dir,
initial_plan_path=initial_plan_path,
config_path=config_path,
)
# Main function for testing
async def main():
"""Main function for testing workflow"""
import logging
# Setup logging
logging.basicConfig(level=logging.INFO)
logger = logging.getLogger(__name__)
# Test parameters
paper_dir = "./deepcode_lab/papers/1"
initial_plan_path = os.path.join(paper_dir, "initial_plan.txt")
# Run workflow
result = await run_codebase_indexing(
paper_dir=paper_dir, initial_plan_path=initial_plan_path, logger=logger
)
logger.info(f"Index result: {result}")
if __name__ == "__main__":
asyncio.run(main())