OpenEvolve / data /tests /test_process_parallel.py
introvoyz041's picture
Migrated from GitHub
5e4510c verified
"""
Tests for process-based parallel controller
"""
import asyncio
import os
import tempfile
import unittest
from unittest.mock import Mock, patch, MagicMock
import time
from concurrent.futures import Future
# Set dummy API key for testing
os.environ["OPENAI_API_KEY"] = "test"
from openevolve.config import Config, DatabaseConfig, EvaluatorConfig, LLMConfig, PromptConfig
from openevolve.database import Program, ProgramDatabase
from openevolve.process_parallel import ProcessParallelController, SerializableResult
class TestProcessParallel(unittest.TestCase):
"""Tests for process-based parallel controller"""
def setUp(self):
"""Set up test environment"""
self.test_dir = tempfile.mkdtemp()
# Create test config
self.config = Config()
self.config.max_iterations = 10
self.config.evaluator.parallel_evaluations = 2
self.config.evaluator.timeout = 10
self.config.database.num_islands = 2
self.config.database.in_memory = True
self.config.checkpoint_interval = 5
# Create test evaluation file
self.eval_content = """
def evaluate(program_path):
return {"score": 0.5, "performance": 0.6}
"""
self.eval_file = os.path.join(self.test_dir, "evaluator.py")
with open(self.eval_file, "w") as f:
f.write(self.eval_content)
# Create test database
self.database = ProgramDatabase(self.config.database)
# Add some test programs
for i in range(3):
program = Program(
id=f"test_{i}",
code=f"def func_{i}(): return {i}",
language="python",
metrics={"score": 0.5 + i * 0.1, "performance": 0.4 + i * 0.1},
iteration_found=0,
)
self.database.add(program)
def tearDown(self):
"""Clean up test environment"""
import shutil
shutil.rmtree(self.test_dir, ignore_errors=True)
def test_controller_initialization(self):
"""Test that controller initializes correctly"""
controller = ProcessParallelController(self.config, self.eval_file, self.database)
self.assertEqual(controller.num_workers, 2)
self.assertIsNone(controller.executor)
self.assertIsNotNone(controller.shutdown_event)
def test_controller_start_stop(self):
"""Test starting and stopping the controller"""
controller = ProcessParallelController(self.config, self.eval_file, self.database)
# Start controller
controller.start()
self.assertIsNotNone(controller.executor)
# Stop controller
controller.stop()
self.assertIsNone(controller.executor)
self.assertTrue(controller.shutdown_event.is_set())
def test_database_snapshot_creation(self):
"""Test creating database snapshot for workers"""
controller = ProcessParallelController(self.config, self.eval_file, self.database)
snapshot = controller._create_database_snapshot()
# Verify snapshot structure
self.assertIn("programs", snapshot)
self.assertIn("islands", snapshot)
self.assertIn("current_island", snapshot)
self.assertIn("artifacts", snapshot)
# Verify programs are serialized
self.assertEqual(len(snapshot["programs"]), 3)
for pid, prog_dict in snapshot["programs"].items():
self.assertIsInstance(prog_dict, dict)
self.assertIn("id", prog_dict)
self.assertIn("code", prog_dict)
def test_run_evolution_basic(self):
"""Test basic evolution run"""
async def run_test():
controller = ProcessParallelController(self.config, self.eval_file, self.database)
# Mock the executor to avoid actually spawning processes
with patch.object(controller, "_submit_iteration") as mock_submit:
# Create mock futures that complete immediately
mock_future1 = MagicMock()
mock_result1 = SerializableResult(
child_program_dict={
"id": "child_1",
"code": "def evolved(): return 1",
"language": "python",
"parent_id": "test_0",
"generation": 1,
"metrics": {"score": 0.7, "performance": 0.8},
"iteration_found": 1,
"metadata": {"changes": "test", "island": 0},
},
parent_id="test_0",
iteration_time=0.1,
iteration=1,
)
mock_future1.done.return_value = True
mock_future1.result.return_value = mock_result1
mock_future1.cancel.return_value = True
mock_submit.return_value = mock_future1
# Start controller
controller.start()
# Run evolution for 1 iteration
result = await controller.run_evolution(
start_iteration=1, max_iterations=1, target_score=None
)
# Verify iteration was submitted with island_id
mock_submit.assert_called_once_with(1, 0)
# Verify program was added to database
self.assertIn("child_1", self.database.programs)
child = self.database.get("child_1")
self.assertEqual(child.metrics["score"], 0.7)
# Run the async test
asyncio.run(run_test())
def test_request_shutdown(self):
"""Test graceful shutdown request"""
controller = ProcessParallelController(self.config, self.eval_file, self.database)
# Request shutdown
controller.request_shutdown()
# Verify shutdown event is set
self.assertTrue(controller.shutdown_event.is_set())
def test_serializable_result(self):
"""Test SerializableResult dataclass"""
result = SerializableResult(
child_program_dict={"id": "test", "code": "pass"},
parent_id="parent",
iteration_time=1.5,
iteration=10,
error=None,
)
# Verify attributes
self.assertEqual(result.child_program_dict["id"], "test")
self.assertEqual(result.parent_id, "parent")
self.assertEqual(result.iteration_time, 1.5)
self.assertEqual(result.iteration, 10)
self.assertIsNone(result.error)
# Test with error
error_result = SerializableResult(error="Test error", iteration=5)
self.assertEqual(error_result.error, "Test error")
self.assertIsNone(error_result.child_program_dict)
if __name__ == "__main__":
unittest.main()