|
|
""" |
|
|
Codebase Index Workflow |
|
|
|
|
|
This workflow integrates the functionality of run_indexer.py and code_indexer.py |
|
|
to build intelligent relationships between existing codebase and target structure. |
|
|
|
|
|
Features: |
|
|
- Extract target file structure from initial_plan.txt |
|
|
- Analyze codebase and build indexes |
|
|
- Generate relationship mappings and statistical reports |
|
|
- Provide reference basis for code reproduction |
|
|
""" |
|
|
|
|
|
import asyncio |
|
|
import json |
|
|
import logging |
|
|
import os |
|
|
import re |
|
|
import sys |
|
|
from pathlib import Path |
|
|
from typing import Dict, Any, Optional |
|
|
import yaml |
|
|
|
|
|
|
|
|
sys.path.append(str(Path(__file__).parent.parent / "tools")) |
|
|
|
|
|
from tools.code_indexer import CodeIndexer |
|
|
|
|
|
|
|
|
class CodebaseIndexWorkflow: |
|
|
"""Codebase Index Workflow Class""" |
|
|
|
|
|
def __init__(self, logger=None): |
|
|
""" |
|
|
Initialize workflow |
|
|
|
|
|
Args: |
|
|
logger: Logger instance |
|
|
""" |
|
|
self.logger = logger or self._setup_default_logger() |
|
|
self.indexer = None |
|
|
|
|
|
def _setup_default_logger(self) -> logging.Logger: |
|
|
"""Setup default logger""" |
|
|
logger = logging.getLogger("CodebaseIndexWorkflow") |
|
|
logger.setLevel(logging.INFO) |
|
|
|
|
|
if not logger.handlers: |
|
|
handler = logging.StreamHandler() |
|
|
formatter = logging.Formatter( |
|
|
"%(asctime)s - %(name)s - %(levelname)s - %(message)s" |
|
|
) |
|
|
handler.setFormatter(formatter) |
|
|
logger.addHandler(handler) |
|
|
|
|
|
return logger |
|
|
|
|
|
def extract_file_tree_from_plan(self, plan_content: str) -> Optional[str]: |
|
|
""" |
|
|
Extract file tree structure from initial_plan.txt content |
|
|
|
|
|
Args: |
|
|
plan_content: Content of the initial_plan.txt file |
|
|
|
|
|
Returns: |
|
|
Extracted file tree structure as string |
|
|
""" |
|
|
|
|
|
file_structure_pattern = r"## File Structure[^\n]*\n```[^\n]*\n(.*?)\n```" |
|
|
|
|
|
match = re.search(file_structure_pattern, plan_content, re.DOTALL) |
|
|
if match: |
|
|
file_tree = match.group(1).strip() |
|
|
lines = file_tree.split("\n") |
|
|
|
|
|
|
|
|
cleaned_lines = [] |
|
|
for line in lines: |
|
|
|
|
|
if line.strip() and ( |
|
|
any(char in line for char in ["βββ", "βββ", "β"]) |
|
|
or line.strip().endswith("/") |
|
|
or "." in line.split("/")[-1] |
|
|
or line.strip().endswith(".py") |
|
|
or line.strip().endswith(".txt") |
|
|
or line.strip().endswith(".md") |
|
|
or line.strip().endswith(".yaml") |
|
|
): |
|
|
cleaned_lines.append(line) |
|
|
|
|
|
if len(cleaned_lines) >= 5: |
|
|
file_tree = "\n".join(cleaned_lines) |
|
|
self.logger.info( |
|
|
f"π Extracted file tree structure from ## File Structure section ({len(cleaned_lines)} lines)" |
|
|
) |
|
|
return file_tree |
|
|
|
|
|
|
|
|
code_block_patterns = [ |
|
|
r"```[^\n]*\n(project/.*?(?:βββ|βββ).*?)\n```", |
|
|
r"```[^\n]*\n(src/.*?(?:βββ|βββ).*?)\n```", |
|
|
r"```[^\n]*\n(core/.*?(?:βββ|βββ).*?)\n```", |
|
|
r"```[^\n]*\n(.*?(?:βββ|βββ).*?(?:\.py|\.txt|\.md|\.yaml).*?)\n```", |
|
|
] |
|
|
|
|
|
for pattern in code_block_patterns: |
|
|
match = re.search(pattern, plan_content, re.DOTALL) |
|
|
if match: |
|
|
file_tree = match.group(1).strip() |
|
|
lines = [line for line in file_tree.split("\n") if line.strip()] |
|
|
if len(lines) >= 5: |
|
|
self.logger.info( |
|
|
f"π Extracted file tree structure from code block ({len(lines)} lines)" |
|
|
) |
|
|
return file_tree |
|
|
|
|
|
|
|
|
self.logger.warning( |
|
|
"β οΈ No standard file tree found, trying to extract from file mentions..." |
|
|
) |
|
|
|
|
|
|
|
|
file_mentions = re.findall( |
|
|
r"`([^`]*(?:\.py|\.txt|\.md|\.yaml|\.yml)[^`]*)`", plan_content |
|
|
) |
|
|
|
|
|
if file_mentions: |
|
|
|
|
|
dirs = set() |
|
|
files_by_dir = {} |
|
|
|
|
|
for file_path in file_mentions: |
|
|
file_path = file_path.strip() |
|
|
if "/" in file_path: |
|
|
dir_path = "/".join(file_path.split("/")[:-1]) |
|
|
filename = file_path.split("/")[-1] |
|
|
dirs.add(dir_path) |
|
|
if dir_path not in files_by_dir: |
|
|
files_by_dir[dir_path] = [] |
|
|
files_by_dir[dir_path].append(filename) |
|
|
else: |
|
|
if "root" not in files_by_dir: |
|
|
files_by_dir["root"] = [] |
|
|
files_by_dir["root"].append(file_path) |
|
|
|
|
|
|
|
|
structure_lines = [] |
|
|
|
|
|
|
|
|
if any("src/" in f for f in file_mentions): |
|
|
root_name = "src" |
|
|
elif any("core/" in f for f in file_mentions): |
|
|
root_name = "core" |
|
|
elif any("lib/" in f for f in file_mentions): |
|
|
root_name = "lib" |
|
|
else: |
|
|
root_name = "project" |
|
|
structure_lines.append(f"{root_name}/") |
|
|
|
|
|
|
|
|
sorted_dirs = sorted(dirs) if dirs else [] |
|
|
for i, dir_path in enumerate(sorted_dirs): |
|
|
is_last_dir = i == len(sorted_dirs) - 1 |
|
|
prefix = "βββ" if is_last_dir else "βββ" |
|
|
structure_lines.append(f"{prefix} {dir_path}/") |
|
|
|
|
|
if dir_path in files_by_dir: |
|
|
files = sorted(files_by_dir[dir_path]) |
|
|
for j, filename in enumerate(files): |
|
|
is_last_file = j == len(files) - 1 |
|
|
if is_last_dir: |
|
|
file_prefix = " βββ" if is_last_file else " βββ" |
|
|
else: |
|
|
file_prefix = "β βββ" if is_last_file else "β βββ" |
|
|
structure_lines.append(f"{file_prefix} {filename}") |
|
|
|
|
|
|
|
|
if "root" in files_by_dir: |
|
|
root_files = sorted(files_by_dir["root"]) |
|
|
for i, filename in enumerate(root_files): |
|
|
is_last = (i == len(root_files) - 1) and not sorted_dirs |
|
|
prefix = "βββ" if is_last else "βββ" |
|
|
structure_lines.append(f"{prefix} {filename}") |
|
|
|
|
|
if len(structure_lines) >= 3: |
|
|
file_tree = "\n".join(structure_lines) |
|
|
self.logger.info( |
|
|
f"π Generated file tree from file mentions ({len(structure_lines)} lines)" |
|
|
) |
|
|
return file_tree |
|
|
|
|
|
|
|
|
self.logger.warning("β οΈ No file tree structure found in initial plan") |
|
|
return None |
|
|
|
|
|
def load_target_structure_from_plan(self, plan_path: str) -> str: |
|
|
""" |
|
|
Load target structure from initial_plan.txt and extract file tree |
|
|
|
|
|
Args: |
|
|
plan_path: Path to initial_plan.txt file |
|
|
|
|
|
Returns: |
|
|
Extracted file tree structure |
|
|
""" |
|
|
try: |
|
|
|
|
|
with open(plan_path, "r", encoding="utf-8") as f: |
|
|
plan_content = f.read() |
|
|
|
|
|
self.logger.info(f"π Loaded initial plan ({len(plan_content)} characters)") |
|
|
|
|
|
|
|
|
file_tree = self.extract_file_tree_from_plan(plan_content) |
|
|
|
|
|
if file_tree: |
|
|
self.logger.info( |
|
|
"β
Successfully extracted file tree from initial plan" |
|
|
) |
|
|
self.logger.info("π Extracted structure preview:") |
|
|
|
|
|
preview_lines = file_tree.split("\n")[:8] |
|
|
for line in preview_lines: |
|
|
self.logger.info(f" {line}") |
|
|
if len(file_tree.split("\n")) > 8: |
|
|
self.logger.info( |
|
|
f" ... {len(file_tree.split('\n')) - 8} more lines" |
|
|
) |
|
|
return file_tree |
|
|
else: |
|
|
self.logger.warning("β οΈ Unable to extract file tree from initial plan") |
|
|
self.logger.info("π Falling back to default target structure") |
|
|
return self.get_default_target_structure() |
|
|
|
|
|
except Exception as e: |
|
|
self.logger.error(f"β Failed to load initial plan file {plan_path}: {e}") |
|
|
self.logger.info("π Falling back to default target structure") |
|
|
return self.get_default_target_structure() |
|
|
|
|
|
def get_default_target_structure(self) -> str: |
|
|
"""Get default target structure""" |
|
|
return """ |
|
|
project/ |
|
|
βββ src/ |
|
|
β βββ core/ |
|
|
β β βββ gcn.py # GCN encoder |
|
|
β β βββ diffusion.py # forward/reverse processes |
|
|
β β βββ denoiser.py # denoising MLP |
|
|
β β βββ fusion.py # fusion combiner |
|
|
β βββ models/ # model wrapper classes |
|
|
β β βββ recdiff.py |
|
|
β βββ utils/ |
|
|
β β βββ data.py # loading & preprocessing |
|
|
β β βββ predictor.py # scoring functions |
|
|
β β βββ loss.py # loss functions |
|
|
β β βββ metrics.py # NDCG, Recall etc. |
|
|
β β βββ sched.py # beta/alpha schedule utils |
|
|
β βββ configs/ |
|
|
β βββ default.yaml # hyperparameters, paths |
|
|
βββ tests/ |
|
|
β βββ test_gcn.py |
|
|
β βββ test_diffusion.py |
|
|
β βββ test_denoiser.py |
|
|
β βββ test_loss.py |
|
|
β βββ test_pipeline.py |
|
|
βββ docs/ |
|
|
β βββ architecture.md |
|
|
β βββ api_reference.md |
|
|
β βββ README.md |
|
|
βββ experiments/ |
|
|
β βββ run_experiment.py |
|
|
β βββ notebooks/ |
|
|
β βββ analysis.ipynb |
|
|
βββ requirements.txt |
|
|
βββ setup.py |
|
|
""" |
|
|
|
|
|
def load_or_create_indexer_config(self, paper_dir: str) -> Dict[str, Any]: |
|
|
""" |
|
|
Load or create indexer configuration |
|
|
|
|
|
Args: |
|
|
paper_dir: Paper directory path |
|
|
|
|
|
Returns: |
|
|
Configuration dictionary |
|
|
""" |
|
|
|
|
|
config_path = Path(__file__).parent.parent / "tools" / "indexer_config.yaml" |
|
|
|
|
|
try: |
|
|
if config_path.exists(): |
|
|
with open(config_path, "r", encoding="utf-8") as f: |
|
|
config = yaml.safe_load(f) |
|
|
|
|
|
|
|
|
if "paths" not in config: |
|
|
config["paths"] = {} |
|
|
config["paths"]["code_base_path"] = os.path.join(paper_dir, "code_base") |
|
|
config["paths"]["output_dir"] = os.path.join(paper_dir, "indexes") |
|
|
|
|
|
|
|
|
if "performance" in config: |
|
|
config["performance"]["enable_concurrent_analysis"] = ( |
|
|
False |
|
|
) |
|
|
if "debug" in config: |
|
|
config["debug"]["verbose_output"] = True |
|
|
if "llm" in config: |
|
|
config["llm"]["request_delay"] = 0.5 |
|
|
|
|
|
self.logger.info(f"Loaded configuration file: {config_path}") |
|
|
return config |
|
|
|
|
|
except Exception as e: |
|
|
self.logger.warning(f"Failed to load configuration file: {e}") |
|
|
|
|
|
|
|
|
self.logger.info("Using default configuration") |
|
|
default_config = { |
|
|
"paths": { |
|
|
"code_base_path": os.path.join(paper_dir, "code_base"), |
|
|
"output_dir": os.path.join(paper_dir, "indexes"), |
|
|
}, |
|
|
"llm": { |
|
|
"model_provider": "anthropic", |
|
|
"max_tokens": 4000, |
|
|
"temperature": 0.3, |
|
|
"request_delay": 0.5, |
|
|
"max_retries": 3, |
|
|
"retry_delay": 1.0, |
|
|
}, |
|
|
"file_analysis": { |
|
|
"max_file_size": 1048576, |
|
|
"max_content_length": 3000, |
|
|
"supported_extensions": [ |
|
|
".py", |
|
|
".js", |
|
|
".ts", |
|
|
".java", |
|
|
".cpp", |
|
|
".c", |
|
|
".h", |
|
|
".hpp", |
|
|
".cs", |
|
|
".php", |
|
|
".rb", |
|
|
".go", |
|
|
".rs", |
|
|
".scala", |
|
|
".kt", |
|
|
".yaml", |
|
|
".yml", |
|
|
".json", |
|
|
".xml", |
|
|
".toml", |
|
|
".md", |
|
|
".txt", |
|
|
], |
|
|
"skip_directories": [ |
|
|
"__pycache__", |
|
|
"node_modules", |
|
|
"target", |
|
|
"build", |
|
|
"dist", |
|
|
"venv", |
|
|
"env", |
|
|
".git", |
|
|
".svn", |
|
|
"data", |
|
|
"datasets", |
|
|
], |
|
|
}, |
|
|
"relationships": { |
|
|
"min_confidence_score": 0.3, |
|
|
"high_confidence_threshold": 0.7, |
|
|
"relationship_types": { |
|
|
"direct_match": 1.0, |
|
|
"partial_match": 0.8, |
|
|
"reference": 0.6, |
|
|
"utility": 0.4, |
|
|
}, |
|
|
}, |
|
|
"performance": { |
|
|
"enable_concurrent_analysis": False, |
|
|
"max_concurrent_files": 3, |
|
|
"enable_content_caching": True, |
|
|
"max_cache_size": 100, |
|
|
}, |
|
|
"debug": { |
|
|
"verbose_output": True, |
|
|
"save_raw_responses": False, |
|
|
"mock_llm_responses": False, |
|
|
}, |
|
|
"output": { |
|
|
"generate_summary": True, |
|
|
"generate_statistics": True, |
|
|
"include_metadata": True, |
|
|
"json_indent": 2, |
|
|
}, |
|
|
"logging": {"level": "INFO", "log_to_file": False}, |
|
|
} |
|
|
|
|
|
return default_config |
|
|
|
|
|
async def run_indexing_workflow( |
|
|
self, |
|
|
paper_dir: str, |
|
|
initial_plan_path: Optional[str] = None, |
|
|
config_path: str = "mcp_agent.secrets.yaml", |
|
|
) -> Dict[str, Any]: |
|
|
""" |
|
|
Run the complete code indexing workflow |
|
|
|
|
|
Args: |
|
|
paper_dir: Paper directory path |
|
|
initial_plan_path: Initial plan file path (optional) |
|
|
config_path: API configuration file path |
|
|
|
|
|
Returns: |
|
|
Index result dictionary |
|
|
""" |
|
|
try: |
|
|
self.logger.info("π Starting codebase index workflow...") |
|
|
|
|
|
|
|
|
if not initial_plan_path: |
|
|
initial_plan_path = os.path.join(paper_dir, "initial_plan.txt") |
|
|
|
|
|
|
|
|
if os.path.exists(initial_plan_path): |
|
|
self.logger.info( |
|
|
f"π Loading target structure from {initial_plan_path}" |
|
|
) |
|
|
target_structure = self.load_target_structure_from_plan( |
|
|
initial_plan_path |
|
|
) |
|
|
else: |
|
|
self.logger.warning( |
|
|
f"β οΈ Initial plan file does not exist: {initial_plan_path}" |
|
|
) |
|
|
self.logger.info("π Using default target structure") |
|
|
target_structure = self.get_default_target_structure() |
|
|
|
|
|
|
|
|
code_base_path = os.path.join(paper_dir, "code_base") |
|
|
if not os.path.exists(code_base_path): |
|
|
self.logger.error(f"β Codebase path does not exist: {code_base_path}") |
|
|
return { |
|
|
"status": "error", |
|
|
"message": f"Code base path does not exist: {code_base_path}", |
|
|
"output_files": {}, |
|
|
} |
|
|
|
|
|
|
|
|
output_dir = os.path.join(paper_dir, "indexes") |
|
|
os.makedirs(output_dir, exist_ok=True) |
|
|
|
|
|
|
|
|
indexer_config = self.load_or_create_indexer_config(paper_dir) |
|
|
|
|
|
self.logger.info(f"π Codebase path: {code_base_path}") |
|
|
self.logger.info(f"π€ Output directory: {output_dir}") |
|
|
|
|
|
|
|
|
self.indexer = CodeIndexer( |
|
|
code_base_path=code_base_path, |
|
|
target_structure=target_structure, |
|
|
output_dir=output_dir, |
|
|
config_path=config_path, |
|
|
enable_pre_filtering=True, |
|
|
) |
|
|
|
|
|
|
|
|
self.indexer.indexer_config = indexer_config |
|
|
|
|
|
|
|
|
if "file_analysis" in indexer_config: |
|
|
file_config = indexer_config["file_analysis"] |
|
|
self.indexer.supported_extensions = set( |
|
|
file_config.get( |
|
|
"supported_extensions", self.indexer.supported_extensions |
|
|
) |
|
|
) |
|
|
self.indexer.skip_directories = set( |
|
|
file_config.get("skip_directories", self.indexer.skip_directories) |
|
|
) |
|
|
self.indexer.max_file_size = file_config.get( |
|
|
"max_file_size", self.indexer.max_file_size |
|
|
) |
|
|
self.indexer.max_content_length = file_config.get( |
|
|
"max_content_length", self.indexer.max_content_length |
|
|
) |
|
|
|
|
|
if "llm" in indexer_config: |
|
|
llm_config = indexer_config["llm"] |
|
|
self.indexer.model_provider = llm_config.get( |
|
|
"model_provider", self.indexer.model_provider |
|
|
) |
|
|
self.indexer.llm_max_tokens = llm_config.get( |
|
|
"max_tokens", self.indexer.llm_max_tokens |
|
|
) |
|
|
self.indexer.llm_temperature = llm_config.get( |
|
|
"temperature", self.indexer.llm_temperature |
|
|
) |
|
|
self.indexer.request_delay = llm_config.get( |
|
|
"request_delay", self.indexer.request_delay |
|
|
) |
|
|
self.indexer.max_retries = llm_config.get( |
|
|
"max_retries", self.indexer.max_retries |
|
|
) |
|
|
self.indexer.retry_delay = llm_config.get( |
|
|
"retry_delay", self.indexer.retry_delay |
|
|
) |
|
|
|
|
|
if "relationships" in indexer_config: |
|
|
rel_config = indexer_config["relationships"] |
|
|
self.indexer.min_confidence_score = rel_config.get( |
|
|
"min_confidence_score", self.indexer.min_confidence_score |
|
|
) |
|
|
self.indexer.high_confidence_threshold = rel_config.get( |
|
|
"high_confidence_threshold", self.indexer.high_confidence_threshold |
|
|
) |
|
|
self.indexer.relationship_types = rel_config.get( |
|
|
"relationship_types", self.indexer.relationship_types |
|
|
) |
|
|
|
|
|
if "performance" in indexer_config: |
|
|
perf_config = indexer_config["performance"] |
|
|
self.indexer.enable_concurrent_analysis = perf_config.get( |
|
|
"enable_concurrent_analysis", |
|
|
self.indexer.enable_concurrent_analysis, |
|
|
) |
|
|
self.indexer.max_concurrent_files = perf_config.get( |
|
|
"max_concurrent_files", self.indexer.max_concurrent_files |
|
|
) |
|
|
self.indexer.enable_content_caching = perf_config.get( |
|
|
"enable_content_caching", self.indexer.enable_content_caching |
|
|
) |
|
|
self.indexer.max_cache_size = perf_config.get( |
|
|
"max_cache_size", self.indexer.max_cache_size |
|
|
) |
|
|
|
|
|
if "debug" in indexer_config: |
|
|
debug_config = indexer_config["debug"] |
|
|
self.indexer.verbose_output = debug_config.get( |
|
|
"verbose_output", self.indexer.verbose_output |
|
|
) |
|
|
self.indexer.save_raw_responses = debug_config.get( |
|
|
"save_raw_responses", self.indexer.save_raw_responses |
|
|
) |
|
|
self.indexer.mock_llm_responses = debug_config.get( |
|
|
"mock_llm_responses", self.indexer.mock_llm_responses |
|
|
) |
|
|
|
|
|
if "output" in indexer_config: |
|
|
output_config = indexer_config["output"] |
|
|
self.indexer.generate_summary = output_config.get( |
|
|
"generate_summary", self.indexer.generate_summary |
|
|
) |
|
|
self.indexer.generate_statistics = output_config.get( |
|
|
"generate_statistics", self.indexer.generate_statistics |
|
|
) |
|
|
self.indexer.include_metadata = output_config.get( |
|
|
"include_metadata", self.indexer.include_metadata |
|
|
) |
|
|
|
|
|
self.logger.info("π§ Indexer configuration completed") |
|
|
self.logger.info(f"π€ Model provider: {self.indexer.model_provider}") |
|
|
self.logger.info( |
|
|
f"β‘ Concurrent analysis: {'Enabled' if self.indexer.enable_concurrent_analysis else 'Disabled'}" |
|
|
) |
|
|
self.logger.info( |
|
|
f"ποΈ Content caching: {'Enabled' if self.indexer.enable_content_caching else 'Disabled'}" |
|
|
) |
|
|
self.logger.info( |
|
|
f"π Pre-filtering: {'Enabled' if self.indexer.enable_pre_filtering else 'Disabled'}" |
|
|
) |
|
|
|
|
|
self.logger.info("=" * 60) |
|
|
self.logger.info("π Starting code indexing process...") |
|
|
|
|
|
|
|
|
output_files = await self.indexer.build_all_indexes() |
|
|
|
|
|
|
|
|
if output_files: |
|
|
summary_report = self.indexer.generate_summary_report(output_files) |
|
|
|
|
|
self.logger.info("=" * 60) |
|
|
self.logger.info("β
Indexing completed successfully!") |
|
|
self.logger.info(f"π Processed {len(output_files)} repositories") |
|
|
self.logger.info("π Generated index files:") |
|
|
for repo_name, file_path in output_files.items(): |
|
|
self.logger.info(f" π {repo_name}: {file_path}") |
|
|
self.logger.info(f"π Summary report: {summary_report}") |
|
|
|
|
|
|
|
|
if self.indexer.generate_statistics: |
|
|
self.logger.info("\nπ Processing statistics:") |
|
|
total_relationships = 0 |
|
|
high_confidence_relationships = 0 |
|
|
|
|
|
for file_path in output_files.values(): |
|
|
try: |
|
|
with open(file_path, "r", encoding="utf-8") as f: |
|
|
index_data = json.load(f) |
|
|
relationships = index_data.get("relationships", []) |
|
|
total_relationships += len(relationships) |
|
|
high_confidence_relationships += len( |
|
|
[ |
|
|
r |
|
|
for r in relationships |
|
|
if r.get("confidence_score", 0) |
|
|
> self.indexer.high_confidence_threshold |
|
|
] |
|
|
) |
|
|
except Exception as e: |
|
|
self.logger.warning( |
|
|
f" β οΈ Unable to load statistics from {file_path}: {e}" |
|
|
) |
|
|
|
|
|
self.logger.info( |
|
|
f" π Total relationships found: {total_relationships}" |
|
|
) |
|
|
self.logger.info( |
|
|
f" β High confidence relationships: {high_confidence_relationships}" |
|
|
) |
|
|
self.logger.info( |
|
|
f" π Average relationships per repository: {total_relationships / len(output_files) if output_files else 0:.1f}" |
|
|
) |
|
|
|
|
|
self.logger.info("\nπ Code indexing process completed successfully!") |
|
|
|
|
|
return { |
|
|
"status": "success", |
|
|
"message": f"Successfully indexed {len(output_files)} repositories", |
|
|
"output_files": output_files, |
|
|
"summary_report": summary_report, |
|
|
"statistics": { |
|
|
"total_repositories": len(output_files), |
|
|
"total_relationships": total_relationships, |
|
|
"high_confidence_relationships": high_confidence_relationships, |
|
|
} |
|
|
if self.indexer.generate_statistics |
|
|
else None, |
|
|
} |
|
|
else: |
|
|
self.logger.warning("β οΈ No index files generated") |
|
|
return { |
|
|
"status": "warning", |
|
|
"message": "No index files were generated", |
|
|
"output_files": {}, |
|
|
} |
|
|
|
|
|
except Exception as e: |
|
|
self.logger.error(f"β Index workflow failed: {e}") |
|
|
|
|
|
import traceback |
|
|
|
|
|
self.logger.error(f"Detailed error information: {traceback.format_exc()}") |
|
|
return {"status": "error", "message": str(e), "output_files": {}} |
|
|
|
|
|
def print_banner(self): |
|
|
"""Print application banner""" |
|
|
banner = """ |
|
|
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
|
|
β π Codebase Index Workflow v1.0 β |
|
|
β Intelligent Code Relationship Analysis Tool β |
|
|
β ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ£ |
|
|
β π Analyzes existing codebases β |
|
|
β π Builds intelligent relationships with target structure β |
|
|
β π€ Powered by LLM analysis β |
|
|
β π Generates detailed JSON indexes β |
|
|
β π― Provides reference for code reproduction β |
|
|
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
|
|
""" |
|
|
print(banner) |
|
|
|
|
|
|
|
|
|
|
|
async def run_codebase_indexing( |
|
|
paper_dir: str, |
|
|
initial_plan_path: Optional[str] = None, |
|
|
config_path: str = "mcp_agent.secrets.yaml", |
|
|
logger=None, |
|
|
) -> Dict[str, Any]: |
|
|
""" |
|
|
Convenience function to run codebase indexing |
|
|
|
|
|
Args: |
|
|
paper_dir: Paper directory path |
|
|
initial_plan_path: Initial plan file path (optional) |
|
|
config_path: API configuration file path |
|
|
logger: Logger instance (optional) |
|
|
|
|
|
Returns: |
|
|
Index result dictionary |
|
|
""" |
|
|
workflow = CodebaseIndexWorkflow(logger=logger) |
|
|
workflow.print_banner() |
|
|
|
|
|
return await workflow.run_indexing_workflow( |
|
|
paper_dir=paper_dir, |
|
|
initial_plan_path=initial_plan_path, |
|
|
config_path=config_path, |
|
|
) |
|
|
|
|
|
|
|
|
|
|
|
async def main(): |
|
|
"""Main function for testing workflow""" |
|
|
import logging |
|
|
|
|
|
|
|
|
logging.basicConfig(level=logging.INFO) |
|
|
logger = logging.getLogger(__name__) |
|
|
|
|
|
|
|
|
paper_dir = "./deepcode_lab/papers/1" |
|
|
initial_plan_path = os.path.join(paper_dir, "initial_plan.txt") |
|
|
|
|
|
|
|
|
result = await run_codebase_indexing( |
|
|
paper_dir=paper_dir, initial_plan_path=initial_plan_path, logger=logger |
|
|
) |
|
|
|
|
|
logger.info(f"Index result: {result}") |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
asyncio.run(main()) |
|
|
|