|
|
"""
|
|
|
Batch evaluator for running multiple evaluations with incremental progress.
|
|
|
|
|
|
Supports:
|
|
|
- Batch evaluation of multiple solution-problem pairs
|
|
|
- Incremental evaluation (resume from where you left off)
|
|
|
- Parallel execution (configurable concurrency)
|
|
|
- Both Docker and SkyPilot backends
|
|
|
- Bucket storage for results (S3/GCS)
|
|
|
- Progress bar with tqdm
|
|
|
- Export failed/pending/aggregated results
|
|
|
"""
|
|
|
|
|
|
import logging
|
|
|
import sys
|
|
|
import time
|
|
|
from concurrent.futures import ThreadPoolExecutor, as_completed
|
|
|
from pathlib import Path
|
|
|
from typing import Callable, Iterator, List, Optional
|
|
|
|
|
|
try:
|
|
|
from tqdm import tqdm
|
|
|
HAS_TQDM = True
|
|
|
except ImportError:
|
|
|
HAS_TQDM = False
|
|
|
|
|
|
from ..runner.base import EvaluationResult, EvaluationStatus
|
|
|
from ..runner.docker import DockerRunner
|
|
|
from .pair import Pair, expand_pairs, read_pairs_file, read_problems_file, read_models_file, read_variants_file
|
|
|
from .state import EvaluationState, PairResult
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
class BatchEvaluator:
|
|
|
"""
|
|
|
Batch evaluator with incremental progress tracking.
|
|
|
|
|
|
Example usage:
|
|
|
# Evaluate all problems for a model
|
|
|
evaluator = BatchEvaluator(results_dir="results/gpt5")
|
|
|
evaluator.evaluate_model("gpt-5", problems=["flash_attn", "cross_entropy"])
|
|
|
|
|
|
# Resume an interrupted evaluation
|
|
|
evaluator = BatchEvaluator(results_dir="results/gpt5")
|
|
|
evaluator.resume()
|
|
|
|
|
|
# Evaluate from a pairs file
|
|
|
evaluator = BatchEvaluator(results_dir="results/batch1")
|
|
|
evaluator.# Use batch.scan_solutions_dir() or evaluate_pairs()
|
|
|
"""
|
|
|
|
|
|
STATE_FILE = ".state.json"
|
|
|
|
|
|
def __init__(
|
|
|
self,
|
|
|
results_dir: Path,
|
|
|
*,
|
|
|
base_dir: Optional[Path] = None,
|
|
|
backend: str = "docker",
|
|
|
max_concurrent: int = 1,
|
|
|
timeout: Optional[int] = None,
|
|
|
bucket_url: Optional[str] = None,
|
|
|
):
|
|
|
"""
|
|
|
Initialize batch evaluator.
|
|
|
|
|
|
Args:
|
|
|
results_dir: Directory for results and state
|
|
|
base_dir: Frontier-CS base directory (auto-detected if None)
|
|
|
backend: Evaluation backend ("docker" or "skypilot")
|
|
|
max_concurrent: Maximum concurrent evaluations
|
|
|
timeout: Default timeout for evaluations (seconds)
|
|
|
bucket_url: Optional bucket URL for result storage (s3://... or gs://...)
|
|
|
Only used with skypilot backend. Results are written directly
|
|
|
to the bucket and synced incrementally.
|
|
|
"""
|
|
|
self.results_dir = Path(results_dir)
|
|
|
self.base_dir = base_dir or self._find_base_dir()
|
|
|
self.backend = backend
|
|
|
self.max_concurrent = max_concurrent
|
|
|
self.timeout = timeout
|
|
|
self.bucket_url = bucket_url
|
|
|
self._bucket_storage = None
|
|
|
|
|
|
|
|
|
if bucket_url:
|
|
|
from ..storage.bucket import BucketStorage
|
|
|
self._bucket_storage = BucketStorage(
|
|
|
bucket_url,
|
|
|
local_cache=self.results_dir / ".bucket_cache",
|
|
|
)
|
|
|
|
|
|
self.state_path = self.results_dir / self.STATE_FILE
|
|
|
self.state = EvaluationState.load(self.state_path)
|
|
|
|
|
|
|
|
|
if backend == "docker":
|
|
|
self._runner = DockerRunner(base_dir=self.base_dir)
|
|
|
else:
|
|
|
from ..runner.skypilot import SkyPilotRunner
|
|
|
self._runner = SkyPilotRunner(
|
|
|
base_dir=self.base_dir,
|
|
|
bucket_url=bucket_url,
|
|
|
)
|
|
|
|
|
|
def _find_base_dir(self) -> Path:
|
|
|
"""Find the Frontier-CS base directory."""
|
|
|
candidates = [
|
|
|
Path(__file__).parents[4],
|
|
|
Path.cwd(),
|
|
|
Path.cwd().parent,
|
|
|
]
|
|
|
for candidate in candidates:
|
|
|
if (candidate / "research").is_dir() and (candidate / "pyproject.toml").exists():
|
|
|
return candidate
|
|
|
raise RuntimeError("Could not find Frontier-CS base directory")
|
|
|
|
|
|
def _save_state(self) -> None:
|
|
|
"""Save current state to disk."""
|
|
|
self.state.save(self.state_path)
|
|
|
|
|
|
def sync_from_bucket(self) -> int:
|
|
|
"""
|
|
|
Sync results from bucket to local state.
|
|
|
|
|
|
Downloads new results from the bucket and merges them into the local state.
|
|
|
Uses incremental sync (--size-only) to only download changed files.
|
|
|
|
|
|
Returns:
|
|
|
Number of results synced from bucket
|
|
|
"""
|
|
|
if not self._bucket_storage:
|
|
|
logger.warning("No bucket storage configured")
|
|
|
return 0
|
|
|
|
|
|
|
|
|
count = self._bucket_storage.sync_from_bucket()
|
|
|
|
|
|
if count == 0:
|
|
|
return 0
|
|
|
|
|
|
|
|
|
bucket_results = self._bucket_storage.read_all_results()
|
|
|
merged = 0
|
|
|
|
|
|
for pair_id, result_data in bucket_results.items():
|
|
|
|
|
|
existing = self.state.results.get(pair_id)
|
|
|
|
|
|
|
|
|
should_update = (
|
|
|
existing is None
|
|
|
or not existing.is_complete
|
|
|
or (result_data.status == "success" and existing.status != "success")
|
|
|
)
|
|
|
|
|
|
if should_update:
|
|
|
self.state.results[pair_id] = PairResult(
|
|
|
pair_id=pair_id,
|
|
|
score=result_data.score,
|
|
|
status=result_data.status,
|
|
|
message=result_data.message,
|
|
|
duration_seconds=result_data.duration_seconds,
|
|
|
timestamp=result_data.timestamp,
|
|
|
)
|
|
|
merged += 1
|
|
|
|
|
|
if merged > 0:
|
|
|
self._save_state()
|
|
|
logger.info(f"Merged {merged} results from bucket into local state")
|
|
|
|
|
|
return merged
|
|
|
|
|
|
def evaluate_pairs(
|
|
|
self,
|
|
|
pairs: List[Pair],
|
|
|
*,
|
|
|
resume: bool = True,
|
|
|
on_progress: Optional[Callable[[Pair, EvaluationResult], None]] = None,
|
|
|
show_progress: bool = True,
|
|
|
) -> EvaluationState:
|
|
|
"""
|
|
|
Evaluate a list of pairs.
|
|
|
|
|
|
Args:
|
|
|
pairs: List of pairs to evaluate
|
|
|
resume: Skip already-completed pairs
|
|
|
on_progress: Callback after each evaluation
|
|
|
show_progress: Show tqdm progress bar
|
|
|
|
|
|
Returns:
|
|
|
Final evaluation state
|
|
|
"""
|
|
|
|
|
|
if self._bucket_storage:
|
|
|
self.sync_from_bucket()
|
|
|
|
|
|
|
|
|
if not self.state.started_at:
|
|
|
self.state.started_at = time.strftime("%Y-%m-%dT%H:%M:%S")
|
|
|
self.state.backend = self.backend
|
|
|
self.state.total_pairs = len(pairs)
|
|
|
self._save_state()
|
|
|
|
|
|
|
|
|
if resume:
|
|
|
pending = self.state.get_pending_pairs(pairs)
|
|
|
if len(pending) < len(pairs):
|
|
|
logger.info(f"Resuming: {len(pairs) - len(pending)} pairs already complete")
|
|
|
else:
|
|
|
pending = pairs
|
|
|
|
|
|
if not pending:
|
|
|
logger.info("All pairs already evaluated")
|
|
|
self._export_all_results(pairs)
|
|
|
return self.state
|
|
|
|
|
|
logger.info(f"Evaluating {len(pending)} pairs (max_concurrent={self.max_concurrent})")
|
|
|
|
|
|
|
|
|
if self.max_concurrent == 1:
|
|
|
self._evaluate_sequential(pending, on_progress, show_progress)
|
|
|
else:
|
|
|
self._evaluate_parallel(pending, on_progress, show_progress)
|
|
|
|
|
|
|
|
|
self._export_all_results(pairs)
|
|
|
|
|
|
return self.state
|
|
|
|
|
|
def _export_all_results(self, all_pairs: Optional[List[Pair]] = None) -> None:
|
|
|
"""Export all result files."""
|
|
|
|
|
|
if self._bucket_storage:
|
|
|
self.sync_from_bucket()
|
|
|
|
|
|
|
|
|
self.state.export_csv(self.results_dir / "results.csv")
|
|
|
self.state.export_summary(self.results_dir / "summary.txt")
|
|
|
|
|
|
|
|
|
failed_count = self.state.export_failed(self.results_dir / "failed.txt")
|
|
|
pending_count = self.state.export_pending(self.results_dir / "pending.txt", all_pairs)
|
|
|
skipped_count = self.state.export_skipped(self.results_dir / "skipped.txt")
|
|
|
|
|
|
|
|
|
self.state.export_aggregated_csv(self.results_dir / "by_model.csv", by="model")
|
|
|
self.state.export_aggregated_csv(self.results_dir / "by_problem.csv", by="problem")
|
|
|
|
|
|
logger.info(f"Results exported to {self.results_dir}")
|
|
|
if failed_count > 0:
|
|
|
logger.info(f" - {failed_count} failed pairs in failed.txt")
|
|
|
if pending_count > 0:
|
|
|
logger.info(f" - {pending_count} pending pairs in pending.txt")
|
|
|
|
|
|
def _evaluate_sequential(
|
|
|
self,
|
|
|
pairs: List[Pair],
|
|
|
on_progress: Optional[Callable[[Pair, EvaluationResult], None]],
|
|
|
show_progress: bool = True,
|
|
|
) -> None:
|
|
|
"""Evaluate pairs sequentially with optional progress bar."""
|
|
|
iterator = pairs
|
|
|
pbar = None
|
|
|
|
|
|
if show_progress and HAS_TQDM:
|
|
|
pbar = tqdm(total=len(pairs), desc="Evaluating", unit="pair", dynamic_ncols=True)
|
|
|
iterator = pairs
|
|
|
|
|
|
try:
|
|
|
for i, pair in enumerate(pairs, 1):
|
|
|
if pbar:
|
|
|
pbar.set_postfix_str(pair.id[:40])
|
|
|
else:
|
|
|
logger.info(f"[{i}/{len(pairs)}] Evaluating {pair.id}")
|
|
|
|
|
|
self.state.mark_running(pair)
|
|
|
self._save_state()
|
|
|
|
|
|
try:
|
|
|
result = self._evaluate_pair(pair)
|
|
|
self._record_result(pair, result)
|
|
|
|
|
|
if on_progress:
|
|
|
on_progress(pair, result)
|
|
|
|
|
|
|
|
|
status_str = "OK" if result.success else "FAIL"
|
|
|
score_str = str(result.score) if result.success else (result.message or "error")
|
|
|
if pbar:
|
|
|
pbar.write(f" [{status_str}] {pair.id}: {score_str}")
|
|
|
|
|
|
except Exception as e:
|
|
|
logger.error(f"Error evaluating {pair.id}: {e}")
|
|
|
self.state.record_result(
|
|
|
pair,
|
|
|
score=None,
|
|
|
status="error",
|
|
|
message=str(e),
|
|
|
)
|
|
|
if pbar:
|
|
|
pbar.write(f" [ERROR] {pair.id}: {e}")
|
|
|
finally:
|
|
|
self._save_state()
|
|
|
if pbar:
|
|
|
pbar.update(1)
|
|
|
finally:
|
|
|
if pbar:
|
|
|
pbar.close()
|
|
|
|
|
|
def _evaluate_parallel(
|
|
|
self,
|
|
|
pairs: List[Pair],
|
|
|
on_progress: Optional[Callable[[Pair, EvaluationResult], None]],
|
|
|
show_progress: bool = True,
|
|
|
) -> None:
|
|
|
"""Evaluate pairs in parallel with optional progress bar."""
|
|
|
pbar = None
|
|
|
if show_progress and HAS_TQDM:
|
|
|
pbar = tqdm(total=len(pairs), desc="Evaluating", unit="pair", dynamic_ncols=True)
|
|
|
|
|
|
try:
|
|
|
with ThreadPoolExecutor(max_workers=self.max_concurrent) as executor:
|
|
|
futures = {}
|
|
|
for pair in pairs:
|
|
|
self.state.mark_running(pair)
|
|
|
future = executor.submit(self._evaluate_pair, pair)
|
|
|
futures[future] = pair
|
|
|
self._save_state()
|
|
|
|
|
|
for future in as_completed(futures):
|
|
|
pair = futures[future]
|
|
|
try:
|
|
|
result = future.result()
|
|
|
self._record_result(pair, result)
|
|
|
|
|
|
if on_progress:
|
|
|
on_progress(pair, result)
|
|
|
|
|
|
|
|
|
status_str = "OK" if result.success else "FAIL"
|
|
|
score_str = str(result.score) if result.success else (result.message or "error")
|
|
|
if pbar:
|
|
|
pbar.write(f" [{status_str}] {pair.id}: {score_str}")
|
|
|
|
|
|
except Exception as e:
|
|
|
logger.error(f"Error evaluating {pair.id}: {e}")
|
|
|
self.state.record_result(
|
|
|
pair,
|
|
|
score=None,
|
|
|
status="error",
|
|
|
message=str(e),
|
|
|
)
|
|
|
if pbar:
|
|
|
pbar.write(f" [ERROR] {pair.id}: {e}")
|
|
|
finally:
|
|
|
self._save_state()
|
|
|
if pbar:
|
|
|
pbar.update(1)
|
|
|
finally:
|
|
|
if pbar:
|
|
|
pbar.close()
|
|
|
|
|
|
def _evaluate_pair(self, pair: Pair) -> EvaluationResult:
|
|
|
"""Evaluate a single pair using the configured runner."""
|
|
|
|
|
|
solution_dir = self.base_dir / "solutions" / pair.solution / "resources"
|
|
|
solution_file = solution_dir / "solution.py"
|
|
|
|
|
|
if not solution_file.exists():
|
|
|
|
|
|
solution_file = self.base_dir / "solutions" / pair.solution / "solution.py"
|
|
|
|
|
|
if not solution_file.exists():
|
|
|
return EvaluationResult(
|
|
|
problem_id=pair.problem,
|
|
|
status=EvaluationStatus.ERROR,
|
|
|
message=f"Solution file not found: {pair.solution}",
|
|
|
)
|
|
|
|
|
|
return self._runner.evaluate_file(
|
|
|
pair.problem,
|
|
|
solution_file,
|
|
|
timeout=self.timeout,
|
|
|
solution_id=pair.solution,
|
|
|
)
|
|
|
|
|
|
def _record_result(self, pair: Pair, result: EvaluationResult) -> None:
|
|
|
"""Record an evaluation result to state."""
|
|
|
status_map = {
|
|
|
EvaluationStatus.SUCCESS: "success",
|
|
|
EvaluationStatus.ERROR: "error",
|
|
|
EvaluationStatus.TIMEOUT: "timeout",
|
|
|
EvaluationStatus.SKIPPED: "skipped",
|
|
|
}
|
|
|
self.state.record_result(
|
|
|
pair,
|
|
|
score=result.score,
|
|
|
status=status_map.get(result.status, "error"),
|
|
|
message=result.message,
|
|
|
duration_seconds=result.duration_seconds,
|
|
|
)
|
|
|
|
|
|
def evaluate_model(
|
|
|
self,
|
|
|
model: str,
|
|
|
problems: List[str],
|
|
|
*,
|
|
|
variants: Optional[List[int]] = None,
|
|
|
resume: bool = True,
|
|
|
on_progress: Optional[Callable[[Pair, EvaluationResult], None]] = None,
|
|
|
) -> EvaluationState:
|
|
|
"""
|
|
|
Evaluate all problems for a given model.
|
|
|
|
|
|
Args:
|
|
|
model: Model name (e.g., "gpt-5", "claude-sonnet-4-5")
|
|
|
problems: List of problem IDs
|
|
|
variants: List of variant indices (default: [0])
|
|
|
resume: Skip already-completed pairs
|
|
|
on_progress: Callback after each evaluation
|
|
|
|
|
|
Returns:
|
|
|
Final evaluation state
|
|
|
"""
|
|
|
solutions_dir = self.base_dir / "solutions"
|
|
|
pairs = expand_pairs(
|
|
|
problems,
|
|
|
[model],
|
|
|
variants,
|
|
|
solutions_dir=solutions_dir,
|
|
|
validate_paths=True,
|
|
|
)
|
|
|
|
|
|
if not pairs:
|
|
|
logger.warning(f"No valid pairs found for model {model}")
|
|
|
return self.state
|
|
|
|
|
|
return self.evaluate_pairs(pairs, resume=resume, on_progress=on_progress)
|
|
|
|
|
|
def evaluate_problem(
|
|
|
self,
|
|
|
problem: str,
|
|
|
models: List[str],
|
|
|
*,
|
|
|
variants: Optional[List[int]] = None,
|
|
|
resume: bool = True,
|
|
|
on_progress: Optional[Callable[[Pair, EvaluationResult], None]] = None,
|
|
|
) -> EvaluationState:
|
|
|
"""
|
|
|
Evaluate a problem across all given models.
|
|
|
|
|
|
Args:
|
|
|
problem: Problem ID (e.g., "flash_attn")
|
|
|
models: List of model names
|
|
|
variants: List of variant indices (default: [0])
|
|
|
resume: Skip already-completed pairs
|
|
|
on_progress: Callback after each evaluation
|
|
|
|
|
|
Returns:
|
|
|
Final evaluation state
|
|
|
"""
|
|
|
solutions_dir = self.base_dir / "solutions"
|
|
|
pairs = expand_pairs(
|
|
|
[problem],
|
|
|
models,
|
|
|
variants,
|
|
|
solutions_dir=solutions_dir,
|
|
|
validate_paths=True,
|
|
|
)
|
|
|
|
|
|
if not pairs:
|
|
|
logger.warning(f"No valid pairs found for problem {problem}")
|
|
|
return self.state
|
|
|
|
|
|
return self.evaluate_pairs(pairs, resume=resume, on_progress=on_progress)
|
|
|
|
|
|
def evaluate_pairs_file(
|
|
|
self,
|
|
|
pairs_file: Path,
|
|
|
*,
|
|
|
resume: bool = True,
|
|
|
on_progress: Optional[Callable[[Pair, EvaluationResult], None]] = None,
|
|
|
) -> EvaluationState:
|
|
|
"""
|
|
|
Evaluate pairs from a pairs file.
|
|
|
|
|
|
Args:
|
|
|
pairs_file: Path to pairs file (solution:problem per line)
|
|
|
resume: Skip already-completed pairs
|
|
|
on_progress: Callback after each evaluation
|
|
|
|
|
|
Returns:
|
|
|
Final evaluation state
|
|
|
"""
|
|
|
pairs = read_pairs_file(pairs_file)
|
|
|
return self.evaluate_pairs(pairs, resume=resume, on_progress=on_progress)
|
|
|
|
|
|
def evaluate_from_files(
|
|
|
self,
|
|
|
problems_file: Path,
|
|
|
models_file: Path,
|
|
|
*,
|
|
|
variants_file: Optional[Path] = None,
|
|
|
resume: bool = True,
|
|
|
on_progress: Optional[Callable[[Pair, EvaluationResult], None]] = None,
|
|
|
) -> EvaluationState:
|
|
|
"""
|
|
|
Evaluate pairs by expanding problems × models × variants.
|
|
|
|
|
|
Args:
|
|
|
problems_file: Path to problems file (one per line)
|
|
|
models_file: Path to models file (one per line)
|
|
|
variants_file: Path to variants file (optional, defaults to [0])
|
|
|
resume: Skip already-completed pairs
|
|
|
on_progress: Callback after each evaluation
|
|
|
|
|
|
Returns:
|
|
|
Final evaluation state
|
|
|
"""
|
|
|
problems = read_problems_file(problems_file)
|
|
|
models = read_models_file(models_file)
|
|
|
variants = read_variants_file(variants_file) if variants_file else [0]
|
|
|
|
|
|
solutions_dir = self.base_dir / "solutions"
|
|
|
pairs = expand_pairs(
|
|
|
problems,
|
|
|
models,
|
|
|
variants,
|
|
|
solutions_dir=solutions_dir,
|
|
|
validate_paths=True,
|
|
|
)
|
|
|
|
|
|
logger.info(f"Expanded {len(problems)} problems × {len(models)} models × {len(variants)} variants = {len(pairs)} pairs")
|
|
|
|
|
|
return self.evaluate_pairs(pairs, resume=resume, on_progress=on_progress)
|
|
|
|
|
|
def resume(
|
|
|
self,
|
|
|
on_progress: Optional[Callable[[Pair, EvaluationResult], None]] = None,
|
|
|
) -> EvaluationState:
|
|
|
"""
|
|
|
Resume an interrupted evaluation.
|
|
|
|
|
|
Requires that the state file exists from a previous run.
|
|
|
|
|
|
Returns:
|
|
|
Final evaluation state
|
|
|
"""
|
|
|
if not self.state.results:
|
|
|
raise ValueError("No previous evaluation state found")
|
|
|
|
|
|
|
|
|
pairs = [
|
|
|
Pair(solution=pair_id.split(":")[0], problem=pair_id.split(":")[1])
|
|
|
for pair_id in self.state.results.keys()
|
|
|
]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
return self.evaluate_pairs(pairs, resume=True, on_progress=on_progress)
|
|
|
|
|
|
def get_status(self) -> dict:
|
|
|
"""Get current evaluation status."""
|
|
|
return {
|
|
|
"total_pairs": self.state.total_pairs,
|
|
|
"completed": self.state.completed_count,
|
|
|
"successful": self.state.success_count,
|
|
|
"errors": self.state.error_count,
|
|
|
"pending": self.state.total_pairs - self.state.completed_count,
|
|
|
"started_at": self.state.started_at,
|
|
|
"updated_at": self.state.updated_at,
|
|
|
}
|
|
|
|
|
|
def retry_failed(
|
|
|
self,
|
|
|
on_progress: Optional[Callable[[Pair, EvaluationResult], None]] = None,
|
|
|
show_progress: bool = True,
|
|
|
) -> EvaluationState:
|
|
|
"""
|
|
|
Retry all failed pairs from the current state.
|
|
|
|
|
|
Returns:
|
|
|
Final evaluation state
|
|
|
"""
|
|
|
failed_pairs = self.state.get_failed_pairs()
|
|
|
if not failed_pairs:
|
|
|
logger.info("No failed pairs to retry")
|
|
|
return self.state
|
|
|
|
|
|
logger.info(f"Retrying {len(failed_pairs)} failed pairs")
|
|
|
|
|
|
|
|
|
for pair in failed_pairs:
|
|
|
del self.state.results[pair.id]
|
|
|
self._save_state()
|
|
|
|
|
|
return self.evaluate_pairs(failed_pairs, resume=False, on_progress=on_progress, show_progress=show_progress)
|
|
|
|
|
|
def evaluate_missing(
|
|
|
self,
|
|
|
problems: List[str],
|
|
|
models: List[str],
|
|
|
*,
|
|
|
variants: Optional[List[int]] = None,
|
|
|
on_progress: Optional[Callable[[Pair, EvaluationResult], None]] = None,
|
|
|
show_progress: bool = True,
|
|
|
) -> EvaluationState:
|
|
|
"""
|
|
|
Evaluate only missing pairs (those not yet in results).
|
|
|
|
|
|
Useful for completing an evaluation run after adding new models/problems.
|
|
|
|
|
|
Args:
|
|
|
problems: List of all problem IDs
|
|
|
models: List of all model names
|
|
|
variants: List of variant indices (default: [0])
|
|
|
on_progress: Callback after each evaluation
|
|
|
show_progress: Show progress bar
|
|
|
|
|
|
Returns:
|
|
|
Final evaluation state
|
|
|
"""
|
|
|
solutions_dir = self.base_dir / "solutions"
|
|
|
all_pairs = expand_pairs(
|
|
|
problems,
|
|
|
models,
|
|
|
variants,
|
|
|
solutions_dir=solutions_dir,
|
|
|
validate_paths=True,
|
|
|
)
|
|
|
|
|
|
|
|
|
missing = [p for p in all_pairs if p.id not in self.state.results]
|
|
|
|
|
|
if not missing:
|
|
|
logger.info("No missing pairs to evaluate")
|
|
|
return self.state
|
|
|
|
|
|
logger.info(f"Found {len(missing)} missing pairs out of {len(all_pairs)} total")
|
|
|
|
|
|
|
|
|
self.state.total_pairs = len(all_pairs)
|
|
|
self._save_state()
|
|
|
|
|
|
return self.evaluate_pairs(missing, resume=False, on_progress=on_progress, show_progress=show_progress)
|
|
|
|