|
|
""" |
|
|
Tests for process-based parallel controller |
|
|
""" |
|
|
|
|
|
import asyncio |
|
|
import os |
|
|
import tempfile |
|
|
import unittest |
|
|
from unittest.mock import Mock, patch, MagicMock |
|
|
import time |
|
|
from concurrent.futures import Future |
|
|
|
|
|
|
|
|
os.environ["OPENAI_API_KEY"] = "test" |
|
|
|
|
|
from openevolve.config import Config, DatabaseConfig, EvaluatorConfig, LLMConfig, PromptConfig |
|
|
from openevolve.database import Program, ProgramDatabase |
|
|
from openevolve.process_parallel import ProcessParallelController, SerializableResult |
|
|
|
|
|
|
|
|
class TestProcessParallel(unittest.TestCase): |
|
|
"""Tests for process-based parallel controller""" |
|
|
|
|
|
def setUp(self): |
|
|
"""Set up test environment""" |
|
|
self.test_dir = tempfile.mkdtemp() |
|
|
|
|
|
|
|
|
self.config = Config() |
|
|
self.config.max_iterations = 10 |
|
|
self.config.evaluator.parallel_evaluations = 2 |
|
|
self.config.evaluator.timeout = 10 |
|
|
self.config.database.num_islands = 2 |
|
|
self.config.database.in_memory = True |
|
|
self.config.checkpoint_interval = 5 |
|
|
|
|
|
|
|
|
self.eval_content = """ |
|
|
def evaluate(program_path): |
|
|
return {"score": 0.5, "performance": 0.6} |
|
|
""" |
|
|
self.eval_file = os.path.join(self.test_dir, "evaluator.py") |
|
|
with open(self.eval_file, "w") as f: |
|
|
f.write(self.eval_content) |
|
|
|
|
|
|
|
|
self.database = ProgramDatabase(self.config.database) |
|
|
|
|
|
|
|
|
for i in range(3): |
|
|
program = Program( |
|
|
id=f"test_{i}", |
|
|
code=f"def func_{i}(): return {i}", |
|
|
language="python", |
|
|
metrics={"score": 0.5 + i * 0.1, "performance": 0.4 + i * 0.1}, |
|
|
iteration_found=0, |
|
|
) |
|
|
self.database.add(program) |
|
|
|
|
|
def tearDown(self): |
|
|
"""Clean up test environment""" |
|
|
import shutil |
|
|
|
|
|
shutil.rmtree(self.test_dir, ignore_errors=True) |
|
|
|
|
|
def test_controller_initialization(self): |
|
|
"""Test that controller initializes correctly""" |
|
|
controller = ProcessParallelController(self.config, self.eval_file, self.database) |
|
|
|
|
|
self.assertEqual(controller.num_workers, 2) |
|
|
self.assertIsNone(controller.executor) |
|
|
self.assertIsNotNone(controller.shutdown_event) |
|
|
|
|
|
def test_controller_start_stop(self): |
|
|
"""Test starting and stopping the controller""" |
|
|
controller = ProcessParallelController(self.config, self.eval_file, self.database) |
|
|
|
|
|
|
|
|
controller.start() |
|
|
self.assertIsNotNone(controller.executor) |
|
|
|
|
|
|
|
|
controller.stop() |
|
|
self.assertIsNone(controller.executor) |
|
|
self.assertTrue(controller.shutdown_event.is_set()) |
|
|
|
|
|
def test_database_snapshot_creation(self): |
|
|
"""Test creating database snapshot for workers""" |
|
|
controller = ProcessParallelController(self.config, self.eval_file, self.database) |
|
|
|
|
|
snapshot = controller._create_database_snapshot() |
|
|
|
|
|
|
|
|
self.assertIn("programs", snapshot) |
|
|
self.assertIn("islands", snapshot) |
|
|
self.assertIn("current_island", snapshot) |
|
|
self.assertIn("artifacts", snapshot) |
|
|
|
|
|
|
|
|
self.assertEqual(len(snapshot["programs"]), 3) |
|
|
for pid, prog_dict in snapshot["programs"].items(): |
|
|
self.assertIsInstance(prog_dict, dict) |
|
|
self.assertIn("id", prog_dict) |
|
|
self.assertIn("code", prog_dict) |
|
|
|
|
|
def test_run_evolution_basic(self): |
|
|
"""Test basic evolution run""" |
|
|
|
|
|
async def run_test(): |
|
|
controller = ProcessParallelController(self.config, self.eval_file, self.database) |
|
|
|
|
|
|
|
|
with patch.object(controller, "_submit_iteration") as mock_submit: |
|
|
|
|
|
mock_future1 = MagicMock() |
|
|
mock_result1 = SerializableResult( |
|
|
child_program_dict={ |
|
|
"id": "child_1", |
|
|
"code": "def evolved(): return 1", |
|
|
"language": "python", |
|
|
"parent_id": "test_0", |
|
|
"generation": 1, |
|
|
"metrics": {"score": 0.7, "performance": 0.8}, |
|
|
"iteration_found": 1, |
|
|
"metadata": {"changes": "test", "island": 0}, |
|
|
}, |
|
|
parent_id="test_0", |
|
|
iteration_time=0.1, |
|
|
iteration=1, |
|
|
) |
|
|
mock_future1.done.return_value = True |
|
|
mock_future1.result.return_value = mock_result1 |
|
|
mock_future1.cancel.return_value = True |
|
|
|
|
|
mock_submit.return_value = mock_future1 |
|
|
|
|
|
|
|
|
controller.start() |
|
|
|
|
|
|
|
|
result = await controller.run_evolution( |
|
|
start_iteration=1, max_iterations=1, target_score=None |
|
|
) |
|
|
|
|
|
|
|
|
mock_submit.assert_called_once_with(1, 0) |
|
|
|
|
|
|
|
|
self.assertIn("child_1", self.database.programs) |
|
|
child = self.database.get("child_1") |
|
|
self.assertEqual(child.metrics["score"], 0.7) |
|
|
|
|
|
|
|
|
asyncio.run(run_test()) |
|
|
|
|
|
def test_request_shutdown(self): |
|
|
"""Test graceful shutdown request""" |
|
|
controller = ProcessParallelController(self.config, self.eval_file, self.database) |
|
|
|
|
|
|
|
|
controller.request_shutdown() |
|
|
|
|
|
|
|
|
self.assertTrue(controller.shutdown_event.is_set()) |
|
|
|
|
|
def test_serializable_result(self): |
|
|
"""Test SerializableResult dataclass""" |
|
|
result = SerializableResult( |
|
|
child_program_dict={"id": "test", "code": "pass"}, |
|
|
parent_id="parent", |
|
|
iteration_time=1.5, |
|
|
iteration=10, |
|
|
error=None, |
|
|
) |
|
|
|
|
|
|
|
|
self.assertEqual(result.child_program_dict["id"], "test") |
|
|
self.assertEqual(result.parent_id, "parent") |
|
|
self.assertEqual(result.iteration_time, 1.5) |
|
|
self.assertEqual(result.iteration, 10) |
|
|
self.assertIsNone(result.error) |
|
|
|
|
|
|
|
|
error_result = SerializableResult(error="Test error", iteration=5) |
|
|
self.assertEqual(error_result.error, "Test error") |
|
|
self.assertIsNone(error_result.child_program_dict) |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
unittest.main() |
|
|
|