|
|
""" |
|
|
High-level API for using OpenEvolve as a library |
|
|
""" |
|
|
|
|
|
import asyncio |
|
|
import tempfile |
|
|
import os |
|
|
import uuid |
|
|
import inspect |
|
|
from typing import Union, Callable, Optional, List, Dict, Any, Tuple |
|
|
from dataclasses import dataclass |
|
|
from pathlib import Path |
|
|
|
|
|
from openevolve.controller import OpenEvolve |
|
|
from openevolve.config import Config, load_config, LLMModelConfig |
|
|
from openevolve.database import Program |
|
|
|
|
|
|
|
|
@dataclass |
|
|
class EvolutionResult: |
|
|
"""Result of an evolution run""" |
|
|
|
|
|
best_program: Optional[Program] |
|
|
best_score: float |
|
|
best_code: str |
|
|
metrics: Dict[str, Any] |
|
|
output_dir: Optional[str] |
|
|
|
|
|
def __repr__(self): |
|
|
return f"EvolutionResult(best_score={self.best_score:.4f})" |
|
|
|
|
|
|
|
|
def run_evolution( |
|
|
initial_program: Union[str, Path, List[str]], |
|
|
evaluator: Union[str, Path, Callable], |
|
|
config: Union[str, Path, Config, None] = None, |
|
|
iterations: Optional[int] = None, |
|
|
output_dir: Optional[str] = None, |
|
|
cleanup: bool = True, |
|
|
) -> EvolutionResult: |
|
|
""" |
|
|
Run evolution with flexible inputs - the main library API |
|
|
|
|
|
Args: |
|
|
initial_program: Can be: |
|
|
- Path to a program file (str or Path) |
|
|
- Program code as a string |
|
|
- List of code lines |
|
|
evaluator: Can be: |
|
|
- Path to an evaluator file (str or Path) |
|
|
- Callable function that takes (program_path) and returns metrics dict |
|
|
config: Can be: |
|
|
- Path to config YAML file (str or Path) |
|
|
- Config object |
|
|
- None for defaults |
|
|
iterations: Number of iterations (overrides config) |
|
|
output_dir: Output directory (None for temp directory) |
|
|
cleanup: If True, clean up temp files after evolution |
|
|
|
|
|
Returns: |
|
|
EvolutionResult with best program and metrics |
|
|
|
|
|
Examples: |
|
|
# Using file paths (original way) |
|
|
result = run_evolution( |
|
|
'program.py', |
|
|
'evaluator.py' |
|
|
) |
|
|
|
|
|
# Using code strings |
|
|
result = run_evolution( |
|
|
initial_program=''' |
|
|
# EVOLVE-BLOCK-START |
|
|
def solve(x): |
|
|
return x * 2 |
|
|
# EVOLVE-BLOCK-END |
|
|
''', |
|
|
evaluator=lambda path: {"score": evaluate_program(path)}, |
|
|
iterations=100 |
|
|
) |
|
|
|
|
|
# Using a custom evaluator function |
|
|
def my_evaluator(program_path): |
|
|
# Run tests, benchmarks, etc. |
|
|
return {"score": 0.95, "runtime": 1.2} |
|
|
|
|
|
result = run_evolution( |
|
|
initial_program=generate_initial_code(), |
|
|
evaluator=my_evaluator |
|
|
) |
|
|
""" |
|
|
return asyncio.run( |
|
|
_run_evolution_async(initial_program, evaluator, config, iterations, output_dir, cleanup) |
|
|
) |
|
|
|
|
|
|
|
|
async def _run_evolution_async( |
|
|
initial_program: Union[str, Path, List[str]], |
|
|
evaluator: Union[str, Path, Callable], |
|
|
config: Union[str, Path, Config, None], |
|
|
iterations: Optional[int], |
|
|
output_dir: Optional[str], |
|
|
cleanup: bool, |
|
|
) -> EvolutionResult: |
|
|
"""Async implementation of run_evolution""" |
|
|
|
|
|
temp_dir = None |
|
|
temp_files = [] |
|
|
|
|
|
try: |
|
|
|
|
|
if config is None: |
|
|
config_obj = Config() |
|
|
elif isinstance(config, Config): |
|
|
config_obj = config |
|
|
else: |
|
|
config_obj = load_config(str(config)) |
|
|
|
|
|
|
|
|
if not config_obj.llm.models: |
|
|
raise ValueError( |
|
|
"No LLM models configured. Please provide a config with LLM models, or set up " |
|
|
"your configuration with models. For example:\n\n" |
|
|
"from openevolve.config import Config, LLMModelConfig\n" |
|
|
"config = Config()\n" |
|
|
"config.llm.models = [LLMModelConfig(name='gpt-4', api_key='your-key')]\n" |
|
|
"result = run_evolution(program, evaluator, config=config)" |
|
|
) |
|
|
|
|
|
|
|
|
if output_dir is None and cleanup: |
|
|
temp_dir = tempfile.mkdtemp(prefix="openevolve_") |
|
|
actual_output_dir = temp_dir |
|
|
else: |
|
|
actual_output_dir = output_dir or "openevolve_output" |
|
|
os.makedirs(actual_output_dir, exist_ok=True) |
|
|
|
|
|
|
|
|
program_path = _prepare_program(initial_program, temp_dir, temp_files) |
|
|
|
|
|
|
|
|
evaluator_path = _prepare_evaluator(evaluator, temp_dir, temp_files) |
|
|
|
|
|
|
|
|
controller = OpenEvolve( |
|
|
initial_program_path=program_path, |
|
|
evaluation_file=evaluator_path, |
|
|
config=config_obj, |
|
|
output_dir=actual_output_dir, |
|
|
) |
|
|
|
|
|
best_program = await controller.run(iterations=iterations) |
|
|
|
|
|
|
|
|
best_score = 0.0 |
|
|
metrics = {} |
|
|
best_code = "" |
|
|
|
|
|
if best_program: |
|
|
best_code = best_program.code |
|
|
metrics = best_program.metrics or {} |
|
|
|
|
|
if "combined_score" in metrics: |
|
|
best_score = metrics["combined_score"] |
|
|
elif metrics: |
|
|
numeric_metrics = [v for v in metrics.values() if isinstance(v, (int, float))] |
|
|
if numeric_metrics: |
|
|
best_score = sum(numeric_metrics) / len(numeric_metrics) |
|
|
|
|
|
return EvolutionResult( |
|
|
best_program=best_program, |
|
|
best_score=best_score, |
|
|
best_code=best_code, |
|
|
metrics=metrics, |
|
|
output_dir=actual_output_dir if not cleanup else None, |
|
|
) |
|
|
|
|
|
finally: |
|
|
|
|
|
if cleanup: |
|
|
for temp_file in temp_files: |
|
|
try: |
|
|
os.unlink(temp_file) |
|
|
except: |
|
|
pass |
|
|
if temp_dir and os.path.exists(temp_dir): |
|
|
import shutil |
|
|
|
|
|
try: |
|
|
shutil.rmtree(temp_dir) |
|
|
except: |
|
|
pass |
|
|
|
|
|
|
|
|
def _prepare_program( |
|
|
initial_program: Union[str, Path, List[str]], temp_dir: Optional[str], temp_files: List[str] |
|
|
) -> str: |
|
|
"""Convert program input to a file path""" |
|
|
|
|
|
|
|
|
if isinstance(initial_program, (str, Path)): |
|
|
if os.path.exists(str(initial_program)): |
|
|
return str(initial_program) |
|
|
|
|
|
|
|
|
if isinstance(initial_program, list): |
|
|
code = "\n".join(initial_program) |
|
|
else: |
|
|
code = str(initial_program) |
|
|
|
|
|
|
|
|
if "EVOLVE-BLOCK-START" not in code: |
|
|
|
|
|
code = f"""# EVOLVE-BLOCK-START |
|
|
{code} |
|
|
# EVOLVE-BLOCK-END""" |
|
|
|
|
|
|
|
|
if temp_dir is None: |
|
|
temp_dir = tempfile.gettempdir() |
|
|
|
|
|
program_file = os.path.join(temp_dir, f"program_{uuid.uuid4().hex[:8]}.py") |
|
|
with open(program_file, "w") as f: |
|
|
f.write(code) |
|
|
temp_files.append(program_file) |
|
|
|
|
|
return program_file |
|
|
|
|
|
|
|
|
def _prepare_evaluator( |
|
|
evaluator: Union[str, Path, Callable], temp_dir: Optional[str], temp_files: List[str] |
|
|
) -> str: |
|
|
"""Convert evaluator input to a file path""" |
|
|
|
|
|
|
|
|
if isinstance(evaluator, (str, Path)): |
|
|
if os.path.exists(str(evaluator)): |
|
|
return str(evaluator) |
|
|
|
|
|
|
|
|
if callable(evaluator): |
|
|
|
|
|
evaluator_id = f"_openevolve_evaluator_{uuid.uuid4().hex[:8]}" |
|
|
|
|
|
|
|
|
globals()[evaluator_id] = evaluator |
|
|
|
|
|
evaluator_code = f""" |
|
|
# Wrapper for user-provided evaluator function |
|
|
import {__name__} as api_module |
|
|
|
|
|
def evaluate(program_path): |
|
|
'''Wrapper for user-provided evaluator function''' |
|
|
user_evaluator = getattr(api_module, '{evaluator_id}') |
|
|
return user_evaluator(program_path) |
|
|
""" |
|
|
else: |
|
|
|
|
|
evaluator_code = str(evaluator) |
|
|
|
|
|
|
|
|
if "def evaluate" not in evaluator_code: |
|
|
raise ValueError("Evaluator code must contain an 'evaluate(program_path)' function") |
|
|
|
|
|
|
|
|
if temp_dir is None: |
|
|
temp_dir = tempfile.gettempdir() |
|
|
|
|
|
eval_file = os.path.join(temp_dir, f"evaluator_{uuid.uuid4().hex[:8]}.py") |
|
|
with open(eval_file, "w") as f: |
|
|
f.write(evaluator_code) |
|
|
temp_files.append(eval_file) |
|
|
|
|
|
return eval_file |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def evolve_function( |
|
|
func: Callable, test_cases: List[Tuple[Any, Any]], iterations: int = 100, **kwargs |
|
|
) -> EvolutionResult: |
|
|
""" |
|
|
Evolve a Python function based on test cases |
|
|
|
|
|
Args: |
|
|
func: Initial function to evolve |
|
|
test_cases: List of (input, expected_output) tuples |
|
|
iterations: Number of evolution iterations |
|
|
**kwargs: Additional arguments for run_evolution |
|
|
|
|
|
Returns: |
|
|
EvolutionResult with optimized function |
|
|
|
|
|
Example: |
|
|
def initial_sort(arr): |
|
|
# Slow bubble sort |
|
|
for i in range(len(arr)): |
|
|
for j in range(len(arr)-1): |
|
|
if arr[j] > arr[j+1]: |
|
|
arr[j], arr[j+1] = arr[j+1], arr[j] |
|
|
return arr |
|
|
|
|
|
result = evolve_function( |
|
|
initial_sort, |
|
|
test_cases=[ |
|
|
([3, 1, 2], [1, 2, 3]), |
|
|
([5, 2, 8, 1], [1, 2, 5, 8]), |
|
|
], |
|
|
iterations=50 |
|
|
) |
|
|
print(f"Optimized function score: {result.best_score}") |
|
|
""" |
|
|
|
|
|
|
|
|
func_source = inspect.getsource(func) |
|
|
func_name = func.__name__ |
|
|
|
|
|
|
|
|
if "EVOLVE-BLOCK-START" not in func_source: |
|
|
|
|
|
lines = func_source.split("\n") |
|
|
func_def_line = next(i for i, line in enumerate(lines) if line.strip().startswith("def ")) |
|
|
|
|
|
|
|
|
indent = len(lines[func_def_line]) - len(lines[func_def_line].lstrip()) |
|
|
func_end = len(lines) |
|
|
for i in range(func_def_line + 1, len(lines)): |
|
|
if lines[i].strip() and (len(lines[i]) - len(lines[i].lstrip())) <= indent: |
|
|
func_end = i |
|
|
break |
|
|
|
|
|
|
|
|
lines.insert(func_def_line + 1, " " * (indent + 4) + "# EVOLVE-BLOCK-START") |
|
|
lines.insert(func_end + 1, " " * (indent + 4) + "# EVOLVE-BLOCK-END") |
|
|
func_source = "\n".join(lines) |
|
|
|
|
|
|
|
|
def evaluator(program_path): |
|
|
import importlib.util |
|
|
import sys |
|
|
|
|
|
|
|
|
spec = importlib.util.spec_from_file_location("evolved", program_path) |
|
|
if spec is None or spec.loader is None: |
|
|
return {"score": 0.0, "error": "Failed to load program"} |
|
|
|
|
|
module = importlib.util.module_from_spec(spec) |
|
|
|
|
|
try: |
|
|
spec.loader.exec_module(module) |
|
|
except Exception as e: |
|
|
return {"score": 0.0, "error": f"Failed to execute program: {str(e)}"} |
|
|
|
|
|
if not hasattr(module, func_name): |
|
|
return {"score": 0.0, "error": f"Function '{func_name}' not found"} |
|
|
|
|
|
evolved_func = getattr(module, func_name) |
|
|
correct = 0 |
|
|
total = len(test_cases) |
|
|
errors = [] |
|
|
|
|
|
for input_val, expected in test_cases: |
|
|
try: |
|
|
|
|
|
if isinstance(input_val, list): |
|
|
test_input = input_val.copy() |
|
|
else: |
|
|
test_input = input_val |
|
|
|
|
|
result = evolved_func(test_input) |
|
|
if result == expected: |
|
|
correct += 1 |
|
|
else: |
|
|
errors.append(f"Input {input_val}: expected {expected}, got {result}") |
|
|
except Exception as e: |
|
|
errors.append(f"Input {input_val}: {str(e)}") |
|
|
|
|
|
return { |
|
|
"score": correct / total, |
|
|
"test_pass_rate": correct / total, |
|
|
"tests_passed": correct, |
|
|
"total_tests": total, |
|
|
"errors": errors[:3], |
|
|
} |
|
|
|
|
|
return run_evolution( |
|
|
initial_program=func_source, evaluator=evaluator, iterations=iterations, **kwargs |
|
|
) |
|
|
|
|
|
|
|
|
def evolve_algorithm( |
|
|
algorithm_class: type, benchmark: Callable, iterations: int = 100, **kwargs |
|
|
) -> EvolutionResult: |
|
|
""" |
|
|
Evolve an algorithm class based on a benchmark |
|
|
|
|
|
Args: |
|
|
algorithm_class: Initial algorithm class to evolve |
|
|
benchmark: Function that takes an instance and returns metrics |
|
|
iterations: Number of evolution iterations |
|
|
**kwargs: Additional arguments for run_evolution |
|
|
|
|
|
Returns: |
|
|
EvolutionResult with optimized algorithm |
|
|
|
|
|
Example: |
|
|
class SortAlgorithm: |
|
|
def sort(self, arr): |
|
|
# Simple bubble sort |
|
|
return sorted(arr) # placeholder |
|
|
|
|
|
def benchmark_sort(instance): |
|
|
import time |
|
|
test_data = [list(range(100, 0, -1))] # Reverse sorted |
|
|
|
|
|
start = time.time() |
|
|
for data in test_data: |
|
|
result = instance.sort(data.copy()) |
|
|
if result != sorted(data): |
|
|
return {"score": 0.0} |
|
|
|
|
|
duration = time.time() - start |
|
|
return { |
|
|
"score": 1.0, |
|
|
"runtime": duration, |
|
|
"performance": 1.0 / (duration + 0.001) |
|
|
} |
|
|
|
|
|
result = evolve_algorithm(SortAlgorithm, benchmark_sort, iterations=50) |
|
|
""" |
|
|
|
|
|
|
|
|
class_source = inspect.getsource(algorithm_class) |
|
|
|
|
|
|
|
|
if "EVOLVE-BLOCK-START" not in class_source: |
|
|
lines = class_source.split("\n") |
|
|
|
|
|
class_def_line = next( |
|
|
i for i, line in enumerate(lines) if line.strip().startswith("class ") |
|
|
) |
|
|
|
|
|
|
|
|
indent = len(lines[class_def_line]) - len(lines[class_def_line].lstrip()) |
|
|
lines.insert(class_def_line + 1, " " * (indent + 4) + "# EVOLVE-BLOCK-START") |
|
|
lines.append(" " * (indent + 4) + "# EVOLVE-BLOCK-END") |
|
|
class_source = "\n".join(lines) |
|
|
|
|
|
|
|
|
def evaluator(program_path): |
|
|
import importlib.util |
|
|
|
|
|
|
|
|
spec = importlib.util.spec_from_file_location("evolved", program_path) |
|
|
if spec is None or spec.loader is None: |
|
|
return {"score": 0.0, "error": "Failed to load program"} |
|
|
|
|
|
module = importlib.util.module_from_spec(spec) |
|
|
|
|
|
try: |
|
|
spec.loader.exec_module(module) |
|
|
except Exception as e: |
|
|
return {"score": 0.0, "error": f"Failed to execute program: {str(e)}"} |
|
|
|
|
|
if not hasattr(module, algorithm_class.__name__): |
|
|
return {"score": 0.0, "error": f"Class '{algorithm_class.__name__}' not found"} |
|
|
|
|
|
AlgorithmClass = getattr(module, algorithm_class.__name__) |
|
|
|
|
|
try: |
|
|
instance = AlgorithmClass() |
|
|
metrics = benchmark(instance) |
|
|
return metrics if isinstance(metrics, dict) else {"score": metrics} |
|
|
except Exception as e: |
|
|
return {"score": 0.0, "error": str(e)} |
|
|
|
|
|
return run_evolution( |
|
|
initial_program=class_source, evaluator=evaluator, iterations=iterations, **kwargs |
|
|
) |
|
|
|
|
|
|
|
|
def evolve_code( |
|
|
initial_code: str, evaluator: Callable[[str], Dict[str, Any]], iterations: int = 100, **kwargs |
|
|
) -> EvolutionResult: |
|
|
""" |
|
|
Evolve arbitrary code with a custom evaluator |
|
|
|
|
|
Args: |
|
|
initial_code: Initial code to evolve |
|
|
evaluator: Function that takes a program path and returns metrics |
|
|
iterations: Number of evolution iterations |
|
|
**kwargs: Additional arguments for run_evolution |
|
|
|
|
|
Returns: |
|
|
EvolutionResult with optimized code |
|
|
|
|
|
Example: |
|
|
initial_code = ''' |
|
|
def fibonacci(n): |
|
|
if n <= 1: |
|
|
return n |
|
|
return fibonacci(n-1) + fibonacci(n-2) |
|
|
''' |
|
|
|
|
|
def eval_fib(program_path): |
|
|
# Evaluate fibonacci implementation |
|
|
import importlib.util |
|
|
import time |
|
|
|
|
|
spec = importlib.util.spec_from_file_location("fib", program_path) |
|
|
module = importlib.util.module_from_spec(spec) |
|
|
spec.loader.exec_module(module) |
|
|
|
|
|
try: |
|
|
start = time.time() |
|
|
result = module.fibonacci(20) |
|
|
duration = time.time() - start |
|
|
|
|
|
correct = result == 6765 |
|
|
return { |
|
|
"score": 1.0 if correct else 0.0, |
|
|
"runtime": duration, |
|
|
"correctness": correct |
|
|
} |
|
|
except: |
|
|
return {"score": 0.0} |
|
|
|
|
|
result = evolve_code(initial_code, eval_fib, iterations=50) |
|
|
""" |
|
|
return run_evolution( |
|
|
initial_program=initial_code, evaluator=evaluator, iterations=iterations, **kwargs |
|
|
) |
|
|
|