|
|
""" |
|
|
Test utilities for OpenEvolve tests |
|
|
Provides common functions and constants for consistent testing |
|
|
""" |
|
|
|
|
|
import os |
|
|
import sys |
|
|
import time |
|
|
import subprocess |
|
|
import requests |
|
|
import socket |
|
|
from typing import Optional, Tuple |
|
|
from openai import OpenAI |
|
|
from openevolve.config import Config, LLMModelConfig |
|
|
|
|
|
|
|
|
TEST_MODEL = "google/gemma-3-270m-it" |
|
|
DEFAULT_PORT = 8000 |
|
|
DEFAULT_BASE_URL = f"http://localhost:{DEFAULT_PORT}/v1" |
|
|
|
|
|
def find_free_port(start_port: int = 8000, max_tries: int = 100) -> int: |
|
|
"""Find a free port starting from start_port""" |
|
|
for port in range(start_port, start_port + max_tries): |
|
|
sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) |
|
|
try: |
|
|
sock.bind(('localhost', port)) |
|
|
sock.close() |
|
|
return port |
|
|
except OSError: |
|
|
continue |
|
|
finally: |
|
|
sock.close() |
|
|
raise RuntimeError(f"Could not find free port in range {start_port}-{start_port + max_tries}") |
|
|
|
|
|
def setup_test_env(): |
|
|
"""Set up test environment with local inference""" |
|
|
os.environ["OPTILLM_API_KEY"] = "optillm" |
|
|
return TEST_MODEL |
|
|
|
|
|
def get_test_client(base_url: str = DEFAULT_BASE_URL) -> OpenAI: |
|
|
"""Get OpenAI client configured for local optillm""" |
|
|
return OpenAI(api_key="optillm", base_url=base_url) |
|
|
|
|
|
def start_test_server(model: str = TEST_MODEL, port: Optional[int] = None) -> Tuple[subprocess.Popen, int]: |
|
|
""" |
|
|
Start optillm server for testing |
|
|
Returns tuple of (process_handle, actual_port_used) |
|
|
""" |
|
|
if port is None: |
|
|
port = find_free_port() |
|
|
|
|
|
|
|
|
env = os.environ.copy() |
|
|
env["OPTILLM_API_KEY"] = "optillm" |
|
|
|
|
|
|
|
|
if "HF_TOKEN" in os.environ: |
|
|
env["HF_TOKEN"] = os.environ["HF_TOKEN"] |
|
|
|
|
|
print(f"Starting optillm server on port {port}...") |
|
|
|
|
|
|
|
|
proc = subprocess.Popen([ |
|
|
"optillm", |
|
|
"--model", model, |
|
|
"--port", str(port) |
|
|
], env=env) |
|
|
|
|
|
|
|
|
for i in range(30): |
|
|
try: |
|
|
response = requests.get(f"http://localhost:{port}/health", timeout=2) |
|
|
if response.status_code == 200: |
|
|
print(f"✅ optillm server started successfully on port {port}") |
|
|
return proc, port |
|
|
except Exception as e: |
|
|
if i < 5: |
|
|
print(f"Attempt {i+1}: Waiting for server... ({e})") |
|
|
pass |
|
|
time.sleep(1) |
|
|
|
|
|
|
|
|
error_msg = f"optillm server failed to start on port {port}" |
|
|
print(f"❌ {error_msg} - check that optillm is installed and model is available") |
|
|
|
|
|
|
|
|
try: |
|
|
proc.terminate() |
|
|
proc.wait(timeout=5) |
|
|
except subprocess.TimeoutExpired: |
|
|
proc.kill() |
|
|
proc.wait() |
|
|
|
|
|
raise RuntimeError(error_msg) |
|
|
|
|
|
def stop_test_server(proc: subprocess.Popen): |
|
|
"""Stop the test server""" |
|
|
try: |
|
|
proc.terminate() |
|
|
proc.wait(timeout=5) |
|
|
except subprocess.TimeoutExpired: |
|
|
proc.kill() |
|
|
proc.wait() |
|
|
|
|
|
def is_server_running(port: int = DEFAULT_PORT) -> bool: |
|
|
"""Check if optillm server is running on the given port""" |
|
|
try: |
|
|
response = requests.get(f"http://localhost:{port}/health", timeout=2) |
|
|
return response.status_code == 200 |
|
|
except: |
|
|
return False |
|
|
|
|
|
def get_integration_config(port: int = DEFAULT_PORT) -> Config: |
|
|
"""Get config for integration tests with optillm""" |
|
|
config = Config() |
|
|
config.max_iterations = 5 |
|
|
config.checkpoint_interval = 2 |
|
|
config.database.in_memory = True |
|
|
config.evaluator.parallel_evaluations = 2 |
|
|
config.evaluator.timeout = 10 |
|
|
|
|
|
|
|
|
config.evaluator.cascade_evaluation = False |
|
|
|
|
|
|
|
|
config.llm.retries = 0 |
|
|
config.llm.timeout = 120 |
|
|
|
|
|
|
|
|
base_url = f"http://localhost:{port}/v1" |
|
|
config.llm.api_base = base_url |
|
|
config.llm.models = [ |
|
|
LLMModelConfig( |
|
|
name=TEST_MODEL, |
|
|
api_key="optillm", |
|
|
api_base=base_url, |
|
|
weight=1.0, |
|
|
timeout=120, |
|
|
retries=0 |
|
|
) |
|
|
] |
|
|
|
|
|
return config |
|
|
|
|
|
def get_simple_test_messages(): |
|
|
"""Get simple test messages for basic validation""" |
|
|
return [ |
|
|
{"role": "system", "content": "You are a helpful coding assistant."}, |
|
|
{"role": "user", "content": "Write a simple Python function that returns 'hello'."} |
|
|
] |
|
|
|
|
|
def get_evolution_test_program(): |
|
|
"""Get a simple program for evolution testing""" |
|
|
return """# EVOLVE-BLOCK-START |
|
|
def solve(x): |
|
|
return x * 2 |
|
|
# EVOLVE-BLOCK-END |
|
|
""" |
|
|
|
|
|
def get_evolution_test_evaluator(): |
|
|
"""Get a simple evaluator for evolution testing""" |
|
|
return """def evaluate(program_path): |
|
|
return {"score": 0.5, "complexity": 10, "combined_score": 0.5} |
|
|
""" |