|
|
""" |
|
|
Main controller for OpenEvolve |
|
|
""" |
|
|
|
|
|
import asyncio |
|
|
import logging |
|
|
import os |
|
|
import signal |
|
|
import time |
|
|
import uuid |
|
|
from pathlib import Path |
|
|
from typing import Any, Dict, List, Optional, Union |
|
|
|
|
|
from openevolve.config import Config, load_config |
|
|
from openevolve.database import Program, ProgramDatabase |
|
|
from openevolve.evaluator import Evaluator |
|
|
from openevolve.evolution_trace import EvolutionTracer |
|
|
from openevolve.llm.ensemble import LLMEnsemble |
|
|
from openevolve.process_parallel import ProcessParallelController |
|
|
from openevolve.prompt.sampler import PromptSampler |
|
|
from openevolve.utils.code_utils import extract_code_language |
|
|
from openevolve.utils.format_utils import format_improvement_safe, format_metrics_safe |
|
|
|
|
|
logger = logging.getLogger(__name__) |
|
|
|
|
|
|
|
|
def _format_metrics(metrics: Dict[str, Any]) -> str: |
|
|
"""Safely format metrics, handling both numeric and string values""" |
|
|
formatted_parts = [] |
|
|
for name, value in metrics.items(): |
|
|
if isinstance(value, (int, float)) and not isinstance(value, bool): |
|
|
try: |
|
|
formatted_parts.append(f"{name}={value:.4f}") |
|
|
except (ValueError, TypeError): |
|
|
formatted_parts.append(f"{name}={value}") |
|
|
else: |
|
|
formatted_parts.append(f"{name}={value}") |
|
|
return ", ".join(formatted_parts) |
|
|
|
|
|
|
|
|
def _format_improvement(improvement: Dict[str, Any]) -> str: |
|
|
"""Safely format improvement metrics""" |
|
|
formatted_parts = [] |
|
|
for name, diff in improvement.items(): |
|
|
if isinstance(diff, (int, float)) and not isinstance(diff, bool): |
|
|
try: |
|
|
formatted_parts.append(f"{name}={diff:+.4f}") |
|
|
except (ValueError, TypeError): |
|
|
formatted_parts.append(f"{name}={diff}") |
|
|
else: |
|
|
formatted_parts.append(f"{name}={diff}") |
|
|
return ", ".join(formatted_parts) |
|
|
|
|
|
|
|
|
class OpenEvolve: |
|
|
""" |
|
|
Main controller for OpenEvolve |
|
|
|
|
|
Orchestrates the evolution process, coordinating between the prompt sampler, |
|
|
LLM ensemble, evaluator, and program database. |
|
|
|
|
|
Features: |
|
|
- Tracks the absolute best program across evolution steps |
|
|
- Ensures the best solution is not lost during the MAP-Elites process |
|
|
- Always includes the best program in the selection process for inspiration |
|
|
- Maintains detailed logs and metadata about improvements |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
initial_program_path: str, |
|
|
evaluation_file: str, |
|
|
config: Config, |
|
|
output_dir: Optional[str] = None, |
|
|
): |
|
|
|
|
|
self.config = config |
|
|
|
|
|
|
|
|
self.output_dir = output_dir or os.path.join( |
|
|
os.path.dirname(initial_program_path), "openevolve_output" |
|
|
) |
|
|
os.makedirs(self.output_dir, exist_ok=True) |
|
|
|
|
|
|
|
|
self._setup_logging() |
|
|
|
|
|
|
|
|
if self.config.random_seed is not None: |
|
|
import hashlib |
|
|
import random |
|
|
|
|
|
import numpy as np |
|
|
|
|
|
|
|
|
random.seed(self.config.random_seed) |
|
|
np.random.seed(self.config.random_seed) |
|
|
|
|
|
|
|
|
base_seed = str(self.config.random_seed).encode("utf-8") |
|
|
llm_seed = int(hashlib.md5(base_seed + b"llm").hexdigest()[:8], 16) % (2**31) |
|
|
|
|
|
|
|
|
self.config.llm.random_seed = llm_seed |
|
|
for model_cfg in self.config.llm.models: |
|
|
if not hasattr(model_cfg, "random_seed") or model_cfg.random_seed is None: |
|
|
model_cfg.random_seed = llm_seed |
|
|
for model_cfg in self.config.llm.evaluator_models: |
|
|
if not hasattr(model_cfg, "random_seed") or model_cfg.random_seed is None: |
|
|
model_cfg.random_seed = llm_seed |
|
|
|
|
|
logger.info(f"Set random seed to {self.config.random_seed} for reproducibility") |
|
|
logger.debug(f"Generated LLM seed: {llm_seed}") |
|
|
|
|
|
|
|
|
self.initial_program_path = initial_program_path |
|
|
self.initial_program_code = self._load_initial_program() |
|
|
if not self.config.language: |
|
|
self.config.language = extract_code_language(self.initial_program_code) |
|
|
|
|
|
|
|
|
self.file_extension = os.path.splitext(initial_program_path)[1] |
|
|
if not self.file_extension: |
|
|
|
|
|
self.file_extension = ".py" |
|
|
else: |
|
|
|
|
|
if not self.file_extension.startswith("."): |
|
|
self.file_extension = f".{self.file_extension}" |
|
|
|
|
|
|
|
|
if not hasattr(self.config, "file_suffix") or self.config.file_suffix == ".py": |
|
|
self.config.file_suffix = self.file_extension |
|
|
|
|
|
|
|
|
self.llm_ensemble = LLMEnsemble(self.config.llm.models) |
|
|
self.llm_evaluator_ensemble = LLMEnsemble(self.config.llm.evaluator_models) |
|
|
|
|
|
self.prompt_sampler = PromptSampler(self.config.prompt) |
|
|
self.evaluator_prompt_sampler = PromptSampler(self.config.prompt) |
|
|
self.evaluator_prompt_sampler.set_templates("evaluator_system_message") |
|
|
|
|
|
|
|
|
if self.config.random_seed is not None: |
|
|
self.config.database.random_seed = self.config.random_seed |
|
|
|
|
|
self.config.database.novelty_llm = self.llm_ensemble |
|
|
self.database = ProgramDatabase(self.config.database) |
|
|
|
|
|
self.evaluator = Evaluator( |
|
|
self.config.evaluator, |
|
|
evaluation_file, |
|
|
self.llm_evaluator_ensemble, |
|
|
self.evaluator_prompt_sampler, |
|
|
database=self.database, |
|
|
suffix=Path(self.initial_program_path).suffix, |
|
|
) |
|
|
self.evaluation_file = evaluation_file |
|
|
|
|
|
logger.info(f"Initialized OpenEvolve with {initial_program_path}") |
|
|
|
|
|
|
|
|
if self.config.evolution_trace.enabled: |
|
|
trace_output_path = self.config.evolution_trace.output_path |
|
|
if not trace_output_path: |
|
|
|
|
|
trace_output_path = os.path.join( |
|
|
self.output_dir, f"evolution_trace.{self.config.evolution_trace.format}" |
|
|
) |
|
|
|
|
|
self.evolution_tracer = EvolutionTracer( |
|
|
output_path=trace_output_path, |
|
|
format=self.config.evolution_trace.format, |
|
|
include_code=self.config.evolution_trace.include_code, |
|
|
include_prompts=self.config.evolution_trace.include_prompts, |
|
|
enabled=True, |
|
|
buffer_size=self.config.evolution_trace.buffer_size, |
|
|
compress=self.config.evolution_trace.compress, |
|
|
) |
|
|
logger.info(f"Evolution tracing enabled: {trace_output_path}") |
|
|
else: |
|
|
self.evolution_tracer = None |
|
|
|
|
|
|
|
|
self.parallel_controller = None |
|
|
|
|
|
def _setup_logging(self) -> None: |
|
|
"""Set up logging""" |
|
|
log_dir = self.config.log_dir or os.path.join(self.output_dir, "logs") |
|
|
os.makedirs(log_dir, exist_ok=True) |
|
|
|
|
|
|
|
|
root_logger = logging.getLogger() |
|
|
root_logger.setLevel(getattr(logging, self.config.log_level)) |
|
|
|
|
|
|
|
|
log_file = os.path.join(log_dir, f"openevolve_{time.strftime('%Y%m%d_%H%M%S')}.log") |
|
|
file_handler = logging.FileHandler(log_file) |
|
|
file_handler.setFormatter( |
|
|
logging.Formatter("%(asctime)s - %(name)s - %(levelname)s - %(message)s") |
|
|
) |
|
|
root_logger.addHandler(file_handler) |
|
|
|
|
|
|
|
|
console_handler = logging.StreamHandler() |
|
|
console_handler.setFormatter(logging.Formatter("%(asctime)s - %(levelname)s - %(message)s")) |
|
|
root_logger.addHandler(console_handler) |
|
|
|
|
|
logger.info(f"Logging to {log_file}") |
|
|
|
|
|
def _load_initial_program(self) -> str: |
|
|
"""Load the initial program from file""" |
|
|
with open(self.initial_program_path, "r") as f: |
|
|
return f.read() |
|
|
|
|
|
async def run( |
|
|
self, |
|
|
iterations: Optional[int] = None, |
|
|
target_score: Optional[float] = None, |
|
|
checkpoint_path: Optional[str] = None, |
|
|
) -> Optional[Program]: |
|
|
""" |
|
|
Run the evolution process with improved parallel processing |
|
|
|
|
|
Args: |
|
|
iterations: Maximum number of iterations (uses config if None) |
|
|
target_score: Target score to reach (continues until reached if specified) |
|
|
checkpoint_path: Path to resume from checkpoint |
|
|
|
|
|
Returns: |
|
|
Best program found |
|
|
""" |
|
|
max_iterations = iterations or self.config.max_iterations |
|
|
|
|
|
|
|
|
start_iteration = 0 |
|
|
if checkpoint_path and os.path.exists(checkpoint_path): |
|
|
self._load_checkpoint(checkpoint_path) |
|
|
start_iteration = self.database.last_iteration + 1 |
|
|
logger.info(f"Resuming from checkpoint at iteration {start_iteration}") |
|
|
else: |
|
|
start_iteration = self.database.last_iteration |
|
|
|
|
|
|
|
|
should_add_initial = ( |
|
|
start_iteration == 0 |
|
|
and len(self.database.programs) == 0 |
|
|
and not any( |
|
|
p.code == self.initial_program_code for p in self.database.programs.values() |
|
|
) |
|
|
) |
|
|
|
|
|
if should_add_initial: |
|
|
logger.info("Adding initial program to database") |
|
|
initial_program_id = str(uuid.uuid4()) |
|
|
|
|
|
|
|
|
initial_metrics = await self.evaluator.evaluate_program( |
|
|
self.initial_program_code, initial_program_id |
|
|
) |
|
|
|
|
|
initial_program = Program( |
|
|
id=initial_program_id, |
|
|
code=self.initial_program_code, |
|
|
language=self.config.language, |
|
|
metrics=initial_metrics, |
|
|
iteration_found=start_iteration, |
|
|
) |
|
|
|
|
|
self.database.add(initial_program) |
|
|
|
|
|
|
|
|
if "combined_score" not in initial_metrics: |
|
|
|
|
|
numeric_metrics = [ |
|
|
v |
|
|
for v in initial_metrics.values() |
|
|
if isinstance(v, (int, float)) and not isinstance(v, bool) |
|
|
] |
|
|
if numeric_metrics: |
|
|
avg_score = sum(numeric_metrics) / len(numeric_metrics) |
|
|
logger.warning( |
|
|
f"⚠️ No 'combined_score' metric found in evaluation results. " |
|
|
f"Using average of all numeric metrics ({avg_score:.4f}) for evolution guidance. " |
|
|
f"For better evolution results, please modify your evaluator to return a 'combined_score' " |
|
|
f"metric that properly weights different aspects of program performance." |
|
|
) |
|
|
else: |
|
|
logger.info( |
|
|
f"Skipping initial program addition (resuming from iteration {start_iteration} " |
|
|
f"with {len(self.database.programs)} existing programs)" |
|
|
) |
|
|
|
|
|
|
|
|
try: |
|
|
self.parallel_controller = ProcessParallelController( |
|
|
self.config, |
|
|
self.evaluation_file, |
|
|
self.database, |
|
|
self.evolution_tracer, |
|
|
file_suffix=self.config.file_suffix, |
|
|
) |
|
|
|
|
|
|
|
|
def signal_handler(signum, frame): |
|
|
logger.info(f"Received signal {signum}, initiating graceful shutdown...") |
|
|
self.parallel_controller.request_shutdown() |
|
|
|
|
|
|
|
|
def force_exit_handler(signum, frame): |
|
|
logger.info("Force exit requested - terminating immediately") |
|
|
import sys |
|
|
|
|
|
sys.exit(0) |
|
|
|
|
|
signal.signal(signal.SIGINT, force_exit_handler) |
|
|
|
|
|
signal.signal(signal.SIGINT, signal_handler) |
|
|
signal.signal(signal.SIGTERM, signal_handler) |
|
|
|
|
|
self.parallel_controller.start() |
|
|
|
|
|
|
|
|
|
|
|
evolution_start = start_iteration |
|
|
evolution_iterations = max_iterations |
|
|
|
|
|
|
|
|
if should_add_initial and start_iteration == 0: |
|
|
evolution_start = 1 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
await self._run_evolution_with_checkpoints( |
|
|
evolution_start, evolution_iterations, target_score |
|
|
) |
|
|
|
|
|
finally: |
|
|
|
|
|
if self.parallel_controller: |
|
|
self.parallel_controller.stop() |
|
|
self.parallel_controller = None |
|
|
|
|
|
|
|
|
if self.evolution_tracer: |
|
|
self.evolution_tracer.close() |
|
|
logger.info("Evolution tracer closed") |
|
|
|
|
|
|
|
|
best_program = None |
|
|
if self.database.best_program_id: |
|
|
best_program = self.database.get(self.database.best_program_id) |
|
|
logger.info(f"Using tracked best program: {self.database.best_program_id}") |
|
|
|
|
|
if best_program is None: |
|
|
best_program = self.database.get_best_program() |
|
|
logger.info("Using calculated best program (tracked program not found)") |
|
|
|
|
|
if best_program: |
|
|
if ( |
|
|
hasattr(self, "parallel_controller") |
|
|
and self.parallel_controller |
|
|
and self.parallel_controller.early_stopping_triggered |
|
|
): |
|
|
logger.info( |
|
|
f"🛑 Evolution complete via early stopping. Best program has metrics: " |
|
|
f"{format_metrics_safe(best_program.metrics)}" |
|
|
) |
|
|
else: |
|
|
logger.info( |
|
|
f"Evolution complete. Best program has metrics: " |
|
|
f"{format_metrics_safe(best_program.metrics)}" |
|
|
) |
|
|
self._save_best_program(best_program) |
|
|
return best_program |
|
|
else: |
|
|
logger.warning("No valid programs found during evolution") |
|
|
return None |
|
|
|
|
|
def _log_iteration( |
|
|
self, |
|
|
iteration: int, |
|
|
parent: Program, |
|
|
child: Program, |
|
|
elapsed_time: float, |
|
|
) -> None: |
|
|
""" |
|
|
Log iteration progress |
|
|
|
|
|
Args: |
|
|
iteration: Iteration number |
|
|
parent: Parent program |
|
|
child: Child program |
|
|
elapsed_time: Elapsed time in seconds |
|
|
""" |
|
|
|
|
|
improvement_str = format_improvement_safe(parent.metrics, child.metrics) |
|
|
|
|
|
logger.info( |
|
|
f"Iteration {iteration+1}: Child {child.id} from parent {parent.id} " |
|
|
f"in {elapsed_time:.2f}s. Metrics: " |
|
|
f"{format_metrics_safe(child.metrics)} " |
|
|
f"(Δ: {improvement_str})" |
|
|
) |
|
|
|
|
|
def _save_checkpoint(self, iteration: int) -> None: |
|
|
""" |
|
|
Save a checkpoint |
|
|
|
|
|
Args: |
|
|
iteration: Current iteration number |
|
|
""" |
|
|
checkpoint_dir = os.path.join(self.output_dir, "checkpoints") |
|
|
os.makedirs(checkpoint_dir, exist_ok=True) |
|
|
|
|
|
|
|
|
checkpoint_path = os.path.join(checkpoint_dir, f"checkpoint_{iteration}") |
|
|
os.makedirs(checkpoint_path, exist_ok=True) |
|
|
|
|
|
|
|
|
self.database.save(checkpoint_path, iteration) |
|
|
|
|
|
|
|
|
best_program = None |
|
|
if self.database.best_program_id: |
|
|
best_program = self.database.get(self.database.best_program_id) |
|
|
else: |
|
|
best_program = self.database.get_best_program() |
|
|
|
|
|
if best_program: |
|
|
|
|
|
best_program_path = os.path.join(checkpoint_path, f"best_program{self.file_extension}") |
|
|
with open(best_program_path, "w") as f: |
|
|
f.write(best_program.code) |
|
|
|
|
|
|
|
|
best_program_info_path = os.path.join(checkpoint_path, "best_program_info.json") |
|
|
with open(best_program_info_path, "w") as f: |
|
|
import json |
|
|
|
|
|
json.dump( |
|
|
{ |
|
|
"id": best_program.id, |
|
|
"generation": best_program.generation, |
|
|
"iteration": best_program.iteration_found, |
|
|
"current_iteration": iteration, |
|
|
"metrics": best_program.metrics, |
|
|
"language": best_program.language, |
|
|
"timestamp": best_program.timestamp, |
|
|
"saved_at": time.time(), |
|
|
}, |
|
|
f, |
|
|
indent=2, |
|
|
) |
|
|
|
|
|
logger.info( |
|
|
f"Saved best program at checkpoint {iteration} with metrics: " |
|
|
f"{format_metrics_safe(best_program.metrics)}" |
|
|
) |
|
|
|
|
|
logger.info(f"Saved checkpoint at iteration {iteration} to {checkpoint_path}") |
|
|
|
|
|
def _load_checkpoint(self, checkpoint_path: str) -> None: |
|
|
"""Load state from a checkpoint directory""" |
|
|
if not os.path.exists(checkpoint_path): |
|
|
raise FileNotFoundError(f"Checkpoint directory {checkpoint_path} not found") |
|
|
|
|
|
logger.info(f"Loading checkpoint from {checkpoint_path}") |
|
|
self.database.load(checkpoint_path) |
|
|
logger.info(f"Checkpoint loaded successfully (iteration {self.database.last_iteration})") |
|
|
|
|
|
async def _run_evolution_with_checkpoints( |
|
|
self, start_iteration: int, max_iterations: int, target_score: Optional[float] |
|
|
) -> None: |
|
|
"""Run evolution with checkpoint saving support""" |
|
|
logger.info(f"Using island-based evolution with {self.config.database.num_islands} islands") |
|
|
self.database.log_island_status() |
|
|
|
|
|
|
|
|
await self.parallel_controller.run_evolution( |
|
|
start_iteration, max_iterations, target_score, checkpoint_callback=self._save_checkpoint |
|
|
) |
|
|
|
|
|
|
|
|
if self.parallel_controller.shutdown_event.is_set(): |
|
|
logger.info("Evolution stopped due to shutdown request") |
|
|
return |
|
|
elif self.parallel_controller.early_stopping_triggered: |
|
|
logger.info("Evolution stopped due to early stopping - saving final checkpoint") |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
final_iteration = start_iteration + max_iterations - 1 |
|
|
if final_iteration > 0 and final_iteration % self.config.checkpoint_interval == 0: |
|
|
self._save_checkpoint(final_iteration) |
|
|
|
|
|
def _save_best_program(self, program: Optional[Program] = None) -> None: |
|
|
""" |
|
|
Save the best program |
|
|
|
|
|
Args: |
|
|
program: Best program (if None, uses the tracked best program) |
|
|
""" |
|
|
|
|
|
if program is None: |
|
|
if self.database.best_program_id: |
|
|
program = self.database.get(self.database.best_program_id) |
|
|
else: |
|
|
|
|
|
program = self.database.get_best_program() |
|
|
|
|
|
if not program: |
|
|
logger.warning("No best program found to save") |
|
|
return |
|
|
|
|
|
best_dir = os.path.join(self.output_dir, "best") |
|
|
os.makedirs(best_dir, exist_ok=True) |
|
|
|
|
|
|
|
|
filename = f"best_program{self.file_extension}" |
|
|
code_path = os.path.join(best_dir, filename) |
|
|
|
|
|
with open(code_path, "w") as f: |
|
|
f.write(program.code) |
|
|
|
|
|
|
|
|
info_path = os.path.join(best_dir, "best_program_info.json") |
|
|
with open(info_path, "w") as f: |
|
|
import json |
|
|
|
|
|
json.dump( |
|
|
{ |
|
|
"id": program.id, |
|
|
"generation": program.generation, |
|
|
"iteration": program.iteration_found, |
|
|
"timestamp": program.timestamp, |
|
|
"parent_id": program.parent_id, |
|
|
"metrics": program.metrics, |
|
|
"language": program.language, |
|
|
"saved_at": time.time(), |
|
|
}, |
|
|
f, |
|
|
indent=2, |
|
|
) |
|
|
|
|
|
logger.info(f"Saved best program to {code_path} with program info to {info_path}") |
|
|
|