introvoyz041's picture
Migrated from GitHub
5e4510c verified
"""
Evaluator for K-Module Pipeline Configuration Problem
This evaluator scores pipeline configurations based on how many modules
match the target configuration. The key property is that there's NO
gradient information - you only know the count of correct modules,
not WHICH ones are correct.
This creates a challenging landscape for iterative refinement but
allows evolutionary crossover to combine good "building blocks"
from different individuals.
Set RICH_FEEDBACK=1 to enable rich feedback mode, which tells you
exactly which modules are correct/incorrect. This demonstrates that
iterative refinement works well when feedback is attributable.
"""
import os
import sys
import time
import traceback
import importlib.util
# Rich feedback mode - when enabled, reveals which modules are correct
RICH_FEEDBACK = os.environ.get("RICH_FEEDBACK", "0") == "1"
# The correct solution (hidden from the optimizer)
# This represents the "optimal" pipeline configuration discovered through
# extensive testing/domain expertise
CORRECT_CONFIG = {
'loader': 'csv_reader',
'preprocess': 'normalize',
'algorithm': 'quicksort',
'formatter': 'json',
}
# Valid options for each module
VALID_OPTIONS = {
'loader': ['csv_reader', 'json_reader', 'xml_reader', 'parquet_reader', 'sql_reader'],
'preprocess': ['normalize', 'standardize', 'minmax', 'scale', 'none'],
'algorithm': ['quicksort', 'mergesort', 'heapsort', 'bubblesort', 'insertion'],
'formatter': ['json', 'xml', 'csv', 'yaml', 'protobuf'],
}
NUM_MODULES = len(CORRECT_CONFIG)
def evaluate(program_path: str) -> dict:
"""
Evaluate a pipeline configuration program.
Args:
program_path: Path to the Python file containing configure_pipeline()
Returns:
dict with 'metrics' and optionally 'artifacts'
"""
start_time = time.time()
try:
# Load and execute the program
spec = importlib.util.spec_from_file_location("program", program_path)
module = importlib.util.module_from_spec(spec)
sys.modules["program"] = module
spec.loader.exec_module(module)
# Get the configuration
if hasattr(module, 'run_pipeline'):
config = module.run_pipeline()
elif hasattr(module, 'configure_pipeline'):
config = module.configure_pipeline()
else:
return _error_result("Program must define run_pipeline() or configure_pipeline()")
# Validate the configuration
validation_errors = validate_config(config)
if validation_errors:
return _validation_error_result(validation_errors)
# Score the configuration
correct_count, module_results = score_config(config)
# Calculate metrics
accuracy = correct_count / NUM_MODULES
# The combined score rewards finding more correct modules
# but gives NO information about which modules are correct
combined_score = accuracy
eval_time = time.time() - start_time
# Build artifacts - provide feedback that helps evolution
# but doesn't reveal which specific modules are wrong
artifacts = build_artifacts(config, correct_count, module_results, eval_time)
# Return metrics at top level for OpenEvolve compatibility
return {
"correct_modules": correct_count,
"total_modules": NUM_MODULES,
"accuracy": accuracy,
"combined_score": combined_score,
"eval_time": eval_time,
"artifacts": artifacts,
}
except Exception as e:
return _exception_result(e)
def validate_config(config: dict) -> list:
"""Validate that the configuration has valid values."""
errors = []
if not isinstance(config, dict):
errors.append(f"Configuration must be a dict, got {type(config).__name__}")
return errors
# Check all required modules are present
for module_name in CORRECT_CONFIG.keys():
if module_name not in config:
errors.append(f"Missing required module: '{module_name}'")
elif config[module_name] not in VALID_OPTIONS[module_name]:
errors.append(
f"Invalid value for '{module_name}': '{config[module_name]}'. "
f"Valid options: {VALID_OPTIONS[module_name]}"
)
return errors
def score_config(config: dict) -> tuple:
"""
Score the configuration against the target.
Returns:
tuple: (correct_count, module_results dict)
"""
correct_count = 0
module_results = {}
for module_name, correct_value in CORRECT_CONFIG.items():
is_correct = config.get(module_name) == correct_value
if is_correct:
correct_count += 1
module_results[module_name] = is_correct
return correct_count, module_results
def build_artifacts(config: dict, correct_count: int, module_results: dict, eval_time: float) -> dict:
"""
Build artifacts that provide useful feedback.
In normal mode: Only reveals how many modules are correct, not which ones.
In rich feedback mode (RICH_FEEDBACK=1): Reveals exactly which modules are correct/incorrect.
"""
artifacts = {}
# Configuration summary
artifacts["configuration"] = str(config)
# Rich feedback mode - reveals which modules are correct/incorrect
if RICH_FEEDBACK:
correct_modules = [m for m, is_correct in module_results.items() if is_correct]
incorrect_modules = [m for m, is_correct in module_results.items() if not is_correct]
artifacts["module_feedback"] = {
"correct": correct_modules,
"incorrect": incorrect_modules,
}
if incorrect_modules:
hints = []
for module in incorrect_modules:
hints.append(f"'{module}' is WRONG - try a different option from {VALID_OPTIONS[module]}")
artifacts["actionable_hints"] = hints
else:
artifacts["actionable_hints"] = ["All modules are correct!"]
# Score feedback - tells you how many are correct, but not which ones
if correct_count == NUM_MODULES:
artifacts["status"] = "PERFECT! All modules correctly configured!"
artifacts["suggestion"] = "Optimal configuration found."
elif correct_count >= NUM_MODULES - 1:
artifacts["status"] = f"Very close! {correct_count}/{NUM_MODULES} modules correct."
artifacts["suggestion"] = "One module may need adjustment. Try variations."
elif correct_count >= NUM_MODULES // 2:
artifacts["status"] = f"Good progress: {correct_count}/{NUM_MODULES} modules correct."
artifacts["suggestion"] = "Some modules are correct. Explore different combinations."
else:
artifacts["status"] = f"Needs improvement: {correct_count}/{NUM_MODULES} modules correct."
artifacts["suggestion"] = "Try different options for each module. Consider the problem domain."
# Hints about the problem structure (not the solution)
artifacts["problem_hints"] = (
"Each module choice is independent. "
"The optimal loader processes the most common data format. "
"The optimal preprocessing creates unit variance. "
"The optimal algorithm has O(n log n) average case. "
"The optimal formatter is widely used for APIs."
)
artifacts["search_space"] = f"{5**NUM_MODULES} possible combinations"
artifacts["eval_time"] = f"{eval_time:.3f}s"
return artifacts
def _error_result(message: str) -> dict:
"""Return an error result."""
return {
"metrics": {
"correct_modules": 0,
"total_modules": NUM_MODULES,
"accuracy": 0.0,
"combined_score": 0.0,
},
"artifacts": {
"error": message,
"status": "ERROR",
},
}
def _validation_error_result(errors: list) -> dict:
"""Return a validation error result."""
return {
"metrics": {
"correct_modules": 0,
"total_modules": NUM_MODULES,
"accuracy": 0.0,
"combined_score": 0.0,
},
"artifacts": {
"validation_errors": "\n".join(errors),
"status": "VALIDATION_ERROR",
"suggestion": "Fix the configuration to use valid module options.",
},
}
def _exception_result(e: Exception) -> dict:
"""Return an exception result."""
return {
"metrics": {
"correct_modules": 0,
"total_modules": NUM_MODULES,
"accuracy": 0.0,
"combined_score": 0.0,
},
"artifacts": {
"exception": str(e),
"traceback": traceback.format_exc(),
"status": "EXCEPTION",
},
}
# For standalone testing
if __name__ == "__main__":
if len(sys.argv) > 1:
result = evaluate(sys.argv[1])
print(f"Metrics: {result['metrics']}")
print(f"Artifacts: {result.get('artifacts', {})}")
else:
# Test with the initial program
import os
script_dir = os.path.dirname(os.path.abspath(__file__))
initial_program = os.path.join(script_dir, "initial_program.py")
result = evaluate(initial_program)
print(f"Metrics: {result['metrics']}")
print(f"Artifacts: {result.get('artifacts', {})}")