#!/bin/bash # Full test: AlphaEvolve AC with eval service and WandB # # Prerequisites: Eval service should be running # bash scripts/dev/start_eval_server.sh # uv run eval_agent/ev2_service_standalone.py --host "0.0.0.0" --port 8711 set -e # ============================================================================ # Configuration # ============================================================================ EXPERIMENT_NAME_PREFIX="alphaevolve_ac_mmv1_3" NUM_GENERATIONS=200 MAX_PARALLEL_JOBS=5 META_INTERVAL=10 LLM_MODELS="native-gemini-3-flash-preview" LLM_SELECTION="ucb1" LLM_TEMPERATURES="0.0 0.5 1.0" LLM_MAX_TOKENS=65536 USE_EVAL_SERVICE="--use-eval-service" EVAL_SERVICE_URL="http://localhost:8711" EVAL_TRIGGER_MODE="periodic" EVAL_TRIGGER_INTERVAL=10 EXPERIMENT_NAME=${EXPERIMENT_NAME_PREFIX}_gen${NUM_GENERATIONS}_${EVAL_TRIGGER_MODE}${EVAL_TRIGGER_INTERVAL} USE_WANDB="--use-wandb" WANDB_PROJECT="ev2" WANDB_TAGS="alphaevolve_ac eval-service full-experiment ${EVAL_TRIGGER_MODE}" # ============================================================================ # Run Experiment # ============================================================================ .venv/bin/python tasks/alphaevolve_ac/run_experiment.py \ --experiment-name "$EXPERIMENT_NAME" \ --num-generations "$NUM_GENERATIONS" \ --max-parallel-jobs "$MAX_PARALLEL_JOBS" \ --meta-interval "$META_INTERVAL" \ --llm-models $LLM_MODELS \ --llm-selection "$LLM_SELECTION" \ --llm-temperatures $LLM_TEMPERATURES \ --llm-max-tokens "$LLM_MAX_TOKENS" \ $USE_EVAL_SERVICE \ --eval-service-url "$EVAL_SERVICE_URL" \ --eval-trigger-mode "$EVAL_TRIGGER_MODE" \ --eval-trigger-interval "$EVAL_TRIGGER_INTERVAL" \ $USE_WANDB \ --wandb-project "$WANDB_PROJECT" \ --wandb-tags $WANDB_TAGS \ --verbose