adaptai / projects /ui /DeepCode /tools /indexer_config.yaml
ADAPT-Chase's picture
Add files using upload-large-folder tool
74d8e8f verified
# Code Indexer Configuration File
# Configure various aspects of the code indexing process
# Paths Configuration
paths:
code_base_path: "D:/Documents/GitHub/Code-Agent/examples/input/paper1/code_base"
output_dir: "D:/Documents/GitHub/Code-Agent/examples/input/paper1/indexes"
# File Analysis Settings
file_analysis:
# Supported file extensions for analysis
supported_extensions:
- ".py" # Python
- ".js" # JavaScript
- ".ts" # TypeScript
- ".java" # Java
- ".cpp" # C++
- ".c" # C
- ".h" # C Header
- ".hpp" # C++ Header
- ".cs" # C#
- ".php" # PHP
- ".rb" # Ruby
- ".go" # Go
- ".rs" # Rust
- ".scala" # Scala
- ".kt" # Kotlin
- ".swift" # Swift
- ".r" # R
- ".sql" # SQL
- ".sh" # Shell Script
- ".bat" # Batch File
- ".ps1" # PowerShell
- ".yaml" # YAML
- ".yml" # YAML
- ".json" # JSON
- ".xml" # XML
- ".toml" # TOML
# Directories to skip during traversal
skip_directories:
- "__pycache__"
- "node_modules"
- "target"
- "build"
- "dist"
- "venv"
- "env"
- ".git"
- ".svn"
- ".hg"
- "coverage"
- ".pytest_cache"
- ".mypy_cache"
# Maximum file size to analyze (in bytes)
max_file_size: 1048576 # 1MB
# Maximum content length to send to LLM (in characters)
max_content_length: 3000
# LLM Configuration
llm:
# Model selection: "anthropic" or "openai"
model_provider: "openai"
# Request parameters
max_tokens: 4000
temperature: 0.3
# System prompt for analysis
system_prompt: "You are a code analysis expert. Provide precise, structured analysis of code relationships and similarities."
# Rate limiting (seconds between requests)
request_delay: 0.1
# Retry configuration
max_retries: 3
retry_delay: 1.0
# Relationship Analysis Settings
relationships:
# Minimum confidence score to include a relationship
min_confidence_score: 0.3
# High confidence threshold for reporting
high_confidence_threshold: 0.7
# Relationship types and their priorities
relationship_types:
direct_match: 1.0 # Direct implementation match
partial_match: 0.8 # Partial functionality match
reference: 0.6 # Reference or utility function
utility: 0.4 # General utility or helper
# Output Configuration
output:
# JSON formatting options
json_indent: 2
ensure_ascii: false
# Generate additional report files
generate_summary: true
generate_statistics: true
# Include metadata in output
include_metadata: true
# File naming pattern (use {repo_name} placeholder)
index_filename_pattern: "{repo_name}_index.json"
summary_filename: "indexing_summary.json"
stats_filename: "indexing_statistics.json"
# Logging Configuration
logging:
level: "INFO" # DEBUG, INFO, WARNING, ERROR
log_to_file: true
log_file: "indexer.log"
log_format: "%(asctime)s - %(name)s - %(levelname)s - %(message)s"
# Performance Settings
performance:
# Enable concurrent processing of files within a repository
enable_concurrent_analysis: true
max_concurrent_files: 5
# Memory optimization
enable_content_caching: false
max_cache_size: 100
# Debug and Development Settings
debug:
# Save raw LLM responses for debugging
save_raw_responses: false
raw_responses_dir: "debug_responses"
# Verbose output during processing
verbose_output: false
# Skip LLM calls for testing (uses mock responses)
mock_llm_responses: false