|
|
""" |
|
|
Tests for cascade evaluation validation functionality in openevolve.evaluator |
|
|
""" |
|
|
|
|
|
import unittest |
|
|
import tempfile |
|
|
import os |
|
|
from unittest.mock import patch, MagicMock |
|
|
from openevolve.config import Config |
|
|
from openevolve.evaluator import Evaluator |
|
|
from openevolve.evaluation_result import EvaluationResult |
|
|
|
|
|
|
|
|
class TestCascadeValidation(unittest.IsolatedAsyncioTestCase): |
|
|
"""Tests for cascade evaluation configuration validation""" |
|
|
|
|
|
def setUp(self): |
|
|
"""Set up test evaluator with cascade validation""" |
|
|
self.config = Config() |
|
|
|
|
|
|
|
|
self.temp_dir = tempfile.mkdtemp() |
|
|
|
|
|
def tearDown(self): |
|
|
"""Clean up temporary files""" |
|
|
|
|
|
import shutil |
|
|
|
|
|
shutil.rmtree(self.temp_dir, ignore_errors=True) |
|
|
|
|
|
def _create_evaluator_file(self, filename: str, content: str) -> str: |
|
|
"""Helper to create temporary evaluator file""" |
|
|
file_path = os.path.join(self.temp_dir, filename) |
|
|
with open(file_path, "w") as f: |
|
|
f.write(content) |
|
|
return file_path |
|
|
|
|
|
def test_cascade_validation_with_valid_evaluator(self): |
|
|
"""Test cascade validation with evaluator that has cascade functions""" |
|
|
|
|
|
evaluator_content = """ |
|
|
def evaluate_stage1(program_path): |
|
|
return {"stage1_score": 0.5} |
|
|
|
|
|
def evaluate_stage2(program_path): |
|
|
return {"stage2_score": 0.7} |
|
|
|
|
|
def evaluate_stage3(program_path): |
|
|
return {"stage3_score": 0.9} |
|
|
|
|
|
def evaluate(program_path): |
|
|
return {"final_score": 1.0} |
|
|
""" |
|
|
evaluator_path = self._create_evaluator_file("valid_cascade.py", evaluator_content) |
|
|
|
|
|
|
|
|
self.config.evaluator.cascade_evaluation = True |
|
|
self.config.evaluator.evaluation_file = evaluator_path |
|
|
|
|
|
|
|
|
with patch("openevolve.evaluator.logger") as mock_logger: |
|
|
evaluator = Evaluator(self.config.evaluator, evaluator_path) |
|
|
|
|
|
|
|
|
mock_logger.warning.assert_not_called() |
|
|
|
|
|
def test_cascade_validation_warning_for_missing_functions(self): |
|
|
"""Test cascade validation warns when cascade functions are missing""" |
|
|
|
|
|
evaluator_content = """ |
|
|
def evaluate(program_path): |
|
|
return {"score": 0.5} |
|
|
""" |
|
|
evaluator_path = self._create_evaluator_file("no_cascade.py", evaluator_content) |
|
|
|
|
|
|
|
|
self.config.evaluator.cascade_evaluation = True |
|
|
self.config.evaluator.evaluation_file = evaluator_path |
|
|
|
|
|
|
|
|
with patch("openevolve.evaluator.logger") as mock_logger: |
|
|
evaluator = Evaluator(self.config.evaluator, evaluator_path) |
|
|
|
|
|
|
|
|
mock_logger.warning.assert_called() |
|
|
warning_call = mock_logger.warning.call_args[0][0] |
|
|
self.assertIn("cascade_evaluation: true", warning_call) |
|
|
self.assertIn("evaluate_stage1", warning_call) |
|
|
|
|
|
def test_cascade_validation_partial_functions(self): |
|
|
"""Test cascade validation with only some cascade functions""" |
|
|
|
|
|
evaluator_content = """ |
|
|
def evaluate_stage1(program_path): |
|
|
return {"stage1_score": 0.5} |
|
|
|
|
|
def evaluate(program_path): |
|
|
return {"score": 0.5} |
|
|
""" |
|
|
evaluator_path = self._create_evaluator_file("partial_cascade.py", evaluator_content) |
|
|
|
|
|
|
|
|
self.config.evaluator.cascade_evaluation = True |
|
|
self.config.evaluator.evaluation_file = evaluator_path |
|
|
|
|
|
|
|
|
with patch("openevolve.evaluator.logger") as mock_logger: |
|
|
evaluator = Evaluator(self.config.evaluator, evaluator_path) |
|
|
|
|
|
|
|
|
mock_logger.warning.assert_called_once() |
|
|
warning_call = mock_logger.warning.call_args[0][0] |
|
|
self.assertIn( |
|
|
"defines 'evaluate_stage1' but no additional cascade stages", warning_call |
|
|
) |
|
|
|
|
|
def test_no_cascade_validation_when_disabled(self): |
|
|
"""Test no validation when cascade evaluation is disabled""" |
|
|
|
|
|
evaluator_content = """ |
|
|
def evaluate(program_path): |
|
|
return {"score": 0.5} |
|
|
""" |
|
|
evaluator_path = self._create_evaluator_file("no_cascade.py", evaluator_content) |
|
|
|
|
|
|
|
|
self.config.evaluator.cascade_evaluation = False |
|
|
self.config.evaluator.evaluation_file = evaluator_path |
|
|
|
|
|
|
|
|
with patch("openevolve.evaluator.logger") as mock_logger: |
|
|
evaluator = Evaluator(self.config.evaluator, evaluator_path) |
|
|
|
|
|
|
|
|
mock_logger.warning.assert_not_called() |
|
|
|
|
|
async def test_direct_evaluate_supports_evaluation_result(self): |
|
|
"""Test that _direct_evaluate supports EvaluationResult returns""" |
|
|
|
|
|
evaluator_content = """ |
|
|
from openevolve.evaluation_result import EvaluationResult |
|
|
|
|
|
def evaluate(program_path): |
|
|
return EvaluationResult( |
|
|
metrics={"score": 0.8, "accuracy": 0.9}, |
|
|
artifacts={"debug_info": "test data"} |
|
|
) |
|
|
""" |
|
|
evaluator_path = self._create_evaluator_file("result_evaluator.py", evaluator_content) |
|
|
|
|
|
self.config.evaluator.cascade_evaluation = False |
|
|
self.config.evaluator.evaluation_file = evaluator_path |
|
|
self.config.evaluator.timeout = 10 |
|
|
|
|
|
evaluator = Evaluator(self.config.evaluator, evaluator_path) |
|
|
|
|
|
|
|
|
program_path = self._create_evaluator_file("test_program.py", "def test(): pass") |
|
|
|
|
|
|
|
|
def mock_evaluate(path): |
|
|
return EvaluationResult( |
|
|
metrics={"score": 0.8, "accuracy": 0.9}, artifacts={"debug_info": "test data"} |
|
|
) |
|
|
|
|
|
evaluator.evaluate_function = mock_evaluate |
|
|
|
|
|
|
|
|
result = await evaluator._direct_evaluate(program_path) |
|
|
|
|
|
|
|
|
self.assertIsInstance(result, EvaluationResult) |
|
|
self.assertEqual(result.metrics["score"], 0.8) |
|
|
self.assertEqual(result.artifacts["debug_info"], "test data") |
|
|
|
|
|
async def test_direct_evaluate_supports_dict_result(self): |
|
|
"""Test that _direct_evaluate still supports dict returns""" |
|
|
|
|
|
evaluator_content = """ |
|
|
def evaluate(program_path): |
|
|
return {"score": 0.7, "performance": 0.85} |
|
|
""" |
|
|
evaluator_path = self._create_evaluator_file("dict_evaluator.py", evaluator_content) |
|
|
|
|
|
self.config.evaluator.cascade_evaluation = False |
|
|
self.config.evaluator.evaluation_file = evaluator_path |
|
|
self.config.evaluator.timeout = 10 |
|
|
|
|
|
evaluator = Evaluator(self.config.evaluator, evaluator_path) |
|
|
|
|
|
|
|
|
program_path = self._create_evaluator_file("test_program.py", "def test(): pass") |
|
|
|
|
|
|
|
|
def mock_evaluate(path): |
|
|
return {"score": 0.7, "performance": 0.85} |
|
|
|
|
|
evaluator.evaluate_function = mock_evaluate |
|
|
|
|
|
|
|
|
result = await evaluator._direct_evaluate(program_path) |
|
|
|
|
|
|
|
|
self.assertIsInstance(result, dict) |
|
|
self.assertEqual(result["score"], 0.7) |
|
|
self.assertEqual(result["performance"], 0.85) |
|
|
|
|
|
def test_cascade_validation_with_class_based_evaluator(self): |
|
|
"""Test cascade validation with class-based evaluator""" |
|
|
|
|
|
evaluator_content = """ |
|
|
class Evaluator: |
|
|
def evaluate_stage1(self, program_path): |
|
|
return {"stage1_score": 0.5} |
|
|
|
|
|
def evaluate_stage2(self, program_path): |
|
|
return {"stage2_score": 0.7} |
|
|
|
|
|
def evaluate(self, program_path): |
|
|
return {"score": 0.5} |
|
|
|
|
|
# Module-level functions (what validation looks for) |
|
|
def evaluate_stage1(program_path): |
|
|
evaluator = Evaluator() |
|
|
return evaluator.evaluate_stage1(program_path) |
|
|
|
|
|
def evaluate_stage2(program_path): |
|
|
evaluator = Evaluator() |
|
|
return evaluator.evaluate_stage2(program_path) |
|
|
|
|
|
def evaluate(program_path): |
|
|
evaluator = Evaluator() |
|
|
return evaluator.evaluate(program_path) |
|
|
""" |
|
|
evaluator_path = self._create_evaluator_file("class_cascade.py", evaluator_content) |
|
|
|
|
|
|
|
|
self.config.evaluator.cascade_evaluation = True |
|
|
self.config.evaluator.evaluation_file = evaluator_path |
|
|
|
|
|
|
|
|
with patch("openevolve.evaluator.logger") as mock_logger: |
|
|
evaluator = Evaluator(self.config.evaluator, evaluator_path) |
|
|
|
|
|
mock_logger.warning.assert_not_called() |
|
|
|
|
|
def test_cascade_validation_with_syntax_error(self): |
|
|
"""Test cascade validation handles syntax errors gracefully""" |
|
|
|
|
|
evaluator_content = """ |
|
|
def evaluate_stage1(program_path) # Missing colon |
|
|
return {"stage1_score": 0.5} |
|
|
""" |
|
|
evaluator_path = self._create_evaluator_file("syntax_error.py", evaluator_content) |
|
|
|
|
|
|
|
|
self.config.evaluator.cascade_evaluation = True |
|
|
self.config.evaluator.evaluation_file = evaluator_path |
|
|
|
|
|
|
|
|
with self.assertRaises(Exception): |
|
|
evaluator = Evaluator(self.config.evaluator, evaluator_path) |
|
|
|
|
|
def test_cascade_validation_nonexistent_file(self): |
|
|
"""Test cascade validation with nonexistent evaluator file""" |
|
|
|
|
|
nonexistent_path = "/nonexistent/path.py" |
|
|
self.config.evaluator.cascade_evaluation = True |
|
|
self.config.evaluator.evaluation_file = nonexistent_path |
|
|
|
|
|
|
|
|
with self.assertRaises(ValueError) as context: |
|
|
evaluator = Evaluator(self.config.evaluator, nonexistent_path) |
|
|
|
|
|
self.assertIn("not found", str(context.exception)) |
|
|
|
|
|
def test_process_evaluation_result_with_artifacts(self): |
|
|
"""Test that _process_evaluation_result handles artifacts correctly""" |
|
|
evaluator_content = """ |
|
|
def evaluate(program_path): |
|
|
return {"score": 0.5} |
|
|
""" |
|
|
evaluator_path = self._create_evaluator_file("dummy.py", evaluator_content) |
|
|
|
|
|
self.config.evaluator.cascade_evaluation = False |
|
|
self.config.evaluator.evaluation_file = evaluator_path |
|
|
evaluator = Evaluator(self.config.evaluator, evaluator_path) |
|
|
|
|
|
|
|
|
eval_result = EvaluationResult( |
|
|
metrics={"score": 0.9}, artifacts={"log": "test log", "data": [1, 2, 3]} |
|
|
) |
|
|
|
|
|
result = evaluator._process_evaluation_result(eval_result) |
|
|
|
|
|
self.assertEqual(result.metrics, {"score": 0.9}) |
|
|
self.assertEqual(result.artifacts, {"log": "test log", "data": [1, 2, 3]}) |
|
|
|
|
|
def test_process_evaluation_result_with_dict(self): |
|
|
"""Test that _process_evaluation_result handles dict results correctly""" |
|
|
evaluator_content = """ |
|
|
def evaluate(program_path): |
|
|
return {"score": 0.5} |
|
|
""" |
|
|
evaluator_path = self._create_evaluator_file("dummy.py", evaluator_content) |
|
|
|
|
|
self.config.evaluator.cascade_evaluation = False |
|
|
self.config.evaluator.evaluation_file = evaluator_path |
|
|
evaluator = Evaluator(self.config.evaluator, evaluator_path) |
|
|
|
|
|
|
|
|
dict_result = {"score": 0.7, "accuracy": 0.8} |
|
|
|
|
|
result = evaluator._process_evaluation_result(dict_result) |
|
|
|
|
|
self.assertEqual(result.metrics, {"score": 0.7, "accuracy": 0.8}) |
|
|
self.assertEqual(result.artifacts, {}) |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
unittest.main() |
|
|
|