OpenEvolve / data /tests /integration /test_evolution_pipeline.py
introvoyz041's picture
Migrated from GitHub
5e4510c verified
"""
Integration tests for the full evolution pipeline with real LLM inference
"""
import pytest
import asyncio
from openevolve.controller import OpenEvolve
class TestEvolutionPipeline:
"""Test complete evolution with real LLM generation"""
@pytest.mark.slow
@pytest.mark.asyncio
async def test_full_evolution_loop(
self,
optillm_server,
evolution_config,
test_program_file,
test_evaluator_file,
evolution_output_dir
):
"""Test complete evolution with real LLM"""
# Configure smaller iteration count for testing
evolution_config.max_iterations = 8
evolution_config.checkpoint_interval = 4
# Run evolution
controller = OpenEvolve(
initial_program_path=str(test_program_file),
evaluation_file=str(test_evaluator_file),
config=evolution_config,
output_dir=str(evolution_output_dir)
)
best_program = await controller.run(iterations=3)
# Verify basic evolution functionality
assert len(controller.database.programs) >= 1, "Should have at least the initial program"
assert best_program is not None, "Should have a best program"
# Check no duplicate chains (validates our per-island MAP-Elites fix)
program_ids = list(controller.database.programs.keys())
migrant_programs = [pid for pid in program_ids if "_migrant_" in pid]
assert len(migrant_programs) == 0, f"Found programs with _migrant_ suffix: {migrant_programs}"
# Print stats for debugging
total_programs = len(controller.database.programs)
evolved_programs = [p for p in controller.database.programs.values() if p.iteration_found > 0]
print(f"Evolution results: {total_programs} total programs, {len(evolved_programs)} evolved programs")
# Verify evolution completed successfully
assert len(controller.database.programs) >= 1, "Should have at least the initial program"
# Check that programs are distributed across islands
island_counts = {i: 0 for i in range(evolution_config.database.num_islands)}
for program in controller.database.programs.values():
island = program.metadata.get("island", 0)
island_counts[island] += 1
# At least one island should have programs
populated_islands = [i for i, count in island_counts.items() if count > 0]
assert len(populated_islands) >= 1, "At least one island should have programs"
@pytest.mark.slow
@pytest.mark.asyncio
async def test_island_feature_maps_populated(
self,
optillm_server,
evolution_config,
test_program_file,
test_evaluator_file,
evolution_output_dir
):
"""Test that island feature maps are properly populated during evolution"""
evolution_config.max_iterations = 6
evolution_config.database.num_islands = 3
controller = OpenEvolve(
initial_program_path=str(test_program_file),
evaluation_file=str(test_evaluator_file),
config=evolution_config,
output_dir=str(evolution_output_dir)
)
await controller.run(iterations=6)
# Check that island feature maps have been populated
total_mapped_programs = 0
for island_idx, island_map in enumerate(controller.database.island_feature_maps):
program_count = len(island_map)
total_mapped_programs += program_count
print(f"Island {island_idx}: {program_count} programs in feature map")
assert total_mapped_programs > 0, "Island feature maps should be populated"
# Verify that all programs in feature maps exist in database
for island_idx, island_map in enumerate(controller.database.island_feature_maps):
for coord, program_id in island_map.items():
assert program_id in controller.database.programs, \
f"Program {program_id} in island {island_idx} feature map not found in database"
@pytest.mark.slow
@pytest.mark.asyncio
async def test_evolution_with_small_model_succeeds(
self,
optillm_server,
evolution_config,
test_program_file,
test_evaluator_file,
evolution_output_dir
):
"""Test that evolution works with small local model (may not be perfect but should not crash)"""
evolution_config.max_iterations = 4
evolution_config.evaluator.timeout = 30 # Longer timeout for small model
controller = OpenEvolve(
initial_program_path=str(test_program_file),
evaluation_file=str(test_evaluator_file),
config=evolution_config,
output_dir=str(evolution_output_dir)
)
# This should not crash, even if some LLM generations fail
best_program = await controller.run(iterations=4)
# Basic sanity checks
assert controller.database.programs, "Should have at least the initial program"
assert best_program is not None or len(controller.database.programs) >= 1, \
"Should have a best program or at least the initial program"
# Check that output directory was created and has some structure
assert evolution_output_dir.exists(), "Output directory should exist"
logs_dir = evolution_output_dir / "logs"
if logs_dir.exists():
log_files = list(logs_dir.glob("*.log"))
# It's okay if no log files - depends on config
print(f"Found {len(log_files)} log files")
@pytest.mark.slow
@pytest.mark.asyncio
async def test_best_program_tracking(
self,
optillm_server,
evolution_config,
test_program_file,
test_evaluator_file,
evolution_output_dir
):
"""Test that best program tracking works correctly"""
evolution_config.max_iterations = 5
controller = OpenEvolve(
initial_program_path=str(test_program_file),
evaluation_file=str(test_evaluator_file),
config=evolution_config,
output_dir=str(evolution_output_dir)
)
best_program = await controller.run(iterations=5)
# Check best program tracking
if controller.database.best_program_id:
best_from_db = controller.database.get(controller.database.best_program_id)
assert best_from_db is not None, "Best program should exist in database"
if best_program:
assert best_program.id == controller.database.best_program_id, \
"Returned best program should match tracked best program"
# Alternative check: get best program from database
best_from_query = controller.database.get_best_program()
assert best_from_query is not None, "Should be able to get best program from database"