| import requests |
| import time |
| import os |
| import re |
| import json |
| import concurrent.futures |
| import argparse |
| import sys |
| import threading |
| from llm_interface import * |
|
|
| |
| API_WORKERS = 16 |
| JUDGE_WORKERS = 8 |
| |
| script_dir = os.path.dirname(os.path.abspath(__file__)) |
|
|
| |
| solution_dir = os.path.join(script_dir, 'solutions') |
|
|
| |
| AVAILABLE_MODELS = { |
| 'gemini', 'gpt', 'claude', 'claude-opus', |
| 'claude-sonnet-4-5', 'gemini3', 'Grok', 'claude-opus-4-5' |
| } |
|
|
| |
| judge_throttler = threading.BoundedSemaphore(JUDGE_WORKERS) |
|
|
| class LocalJudge: |
| def __init__(self, judge_url="http://localhost:8081"): |
| self.judge_url = judge_url |
| self.session = requests.Session() |
|
|
| def get_all_problems(self): |
| try: |
| response = self.session.get(f"{self.judge_url}/problems") |
| response.raise_for_status() |
| return [p['id'] for p in response.json().get('problems', [])] |
| except requests.RequestException: |
| print(f"Error connecting to the judge at {self.judge_url}. Is it running?") |
| return None |
|
|
| def get_problem_statement(self, pid): |
| try: |
| response = self.session.get(f"{self.judge_url}/problem/{pid}/statement") |
| response.raise_for_status() |
| return response.text |
| except requests.RequestException: |
| return None |
|
|
| def submit_solution(self, pid, code): |
| files = {'code': ('solution.cpp', code)} |
| data = {'pid': pid, 'lang': 'cpp'} |
| try: |
| response = self.session.post(f"{self.judge_url}/submit", files=files, data=data) |
| response.raise_for_status() |
| return response.json().get('sid') |
| except requests.RequestException: |
| return None |
|
|
| def get_result(self, sid, poll_interval=60): |
| |
| while True: |
| try: |
| response = self.session.get(f"{self.judge_url}/result/{sid}") |
| if response.status_code == 404: |
| time.sleep(poll_interval) |
| continue |
| |
| response.raise_for_status() |
| result = response.json() |
| |
| if result.get('status') in ['done', 'error']: |
| return result |
| |
| time.sleep(poll_interval) |
| |
| except requests.RequestException: |
| time.sleep(poll_interval) |
|
|
| def extract_cpp_code(response_text): |
| match = re.search(r'```cpp\n(.*?)```', response_text, re.DOTALL) |
| if match: |
| return match.group(1).strip() |
| return response_text |
|
|
| def get_llm_instance(model_name): |
| """Factory function to instantiate the correct LLM class.""" |
| if model_name == 'gemini': |
| return Gemini() |
| elif model_name == 'gpt': |
| return GPT() |
| elif model_name == 'claude': |
| return Claude() |
| elif model_name == 'claude-opus': |
| return Claude_Opus() |
| elif model_name == 'claude-sonnet-4-5': |
| return Claude_Sonnet_4_5() |
| elif model_name == 'gemini3': |
| return Gemini3() |
| elif model_name == 'Grok': |
| return Grok() |
| elif model_name == 'claude-opus-4-5': |
| return Claude_Opus_4_5() |
| else: |
| raise ValueError(f"Unknown model: {model_name}") |
|
|
| def process_single_attempt(pid, model_name): |
| """ |
| Generates one solution, judges it, and saves the results. |
| Returns the score (float/int) for the attempt. |
| Returns 0 if a critical error occurs before grading. |
| """ |
| run_id = f"{pid}_{model_name}" |
| result_filename = f"{solution_dir}/{run_id}_result.json" |
| solution_filename = f"{solution_dir}/{run_id}_solution.cpp" |
|
|
| print(f"[Processing {pid}] Starting attempt using {model_name}") |
| |
| judge = LocalJudge() |
| |
| try: |
| llm = get_llm_instance(model_name) |
| except ValueError as e: |
| print(e) |
| return 0 |
|
|
| try: |
| |
| statement = judge.get_problem_statement(pid) |
| if not statement: |
| print(f"[Processing {pid}] Failed to get statement.") |
| return 0 |
|
|
| |
| llm_response, _ = llm.generate_solution(statement) |
| if not llm_response: |
| print(f"[Processing {pid}] LLM failed.") |
| final_result = {"status": "error", "error": "LLM_TIMEOUT_OR_FAILURE", "score": 0} |
| with open(result_filename, 'w', encoding='utf-8') as f: |
| json.dump(final_result, f, indent=4) |
| return 0 |
|
|
| llm_text_content = str(llm_response) |
| solution_code = extract_cpp_code(llm_text_content) |
| |
| with open(solution_filename, 'w', encoding='utf-8') as f: |
| f.write(solution_code) |
|
|
| |
| with judge_throttler: |
| submission_id = judge.submit_solution(pid, solution_code) |
| if not submission_id: |
| final_result = {"status": "error", "error": "SUBMISSION_FAILED", "score": 0} |
| with open(result_filename, 'w', encoding='utf-8') as f: |
| json.dump(final_result, f, indent=4) |
| return 0 |
|
|
| final_result = judge.get_result(submission_id) |
| |
| |
| new_score = final_result.get('score', -1) |
| with open(result_filename, 'w', encoding='utf-8') as f: |
| json.dump(final_result, f, indent=4) |
| |
| print(f"[Processing {pid}] Finished. Score: {new_score}") |
| |
| return new_score |
|
|
| except Exception as e: |
| print(f"[Processing {pid}] A critical error occurred: {e}") |
| error_result = {"status": "error", "error": str(e), "score": 0} |
| try: |
| with open(result_filename, 'w', encoding='utf-8') as f: |
| json.dump(error_result, f, indent=4) |
| except Exception as e_file: |
| print(f" ...Additionally, FAILED to write error log: {e_file}") |
| return 0 |
|
|
|
|
| if __name__ == "__main__": |
| |
| parser = argparse.ArgumentParser(description="Run LLM Judge Benchmarks") |
| parser.add_argument( |
| "model", |
| type=str, |
| help=f"The model to run. Options: {', '.join(AVAILABLE_MODELS)}" |
| ) |
| args = parser.parse_args() |
| |
| selected_model = args.model |
|
|
| if selected_model not in AVAILABLE_MODELS: |
| print(f"Error: '{selected_model}' is not a valid model.") |
| print(f"Available options: {', '.join(AVAILABLE_MODELS)}") |
| sys.exit(1) |
|
|
| if not os.path.exists(solution_dir): |
| os.makedirs(solution_dir) |
|
|
| judge = LocalJudge() |
| print("Fetching list of problems...") |
| problem_ids = judge.get_all_problems() |
| if not problem_ids: |
| print("No problems found on the judge.") |
| sys.exit(0) |
| |
| print(f"Found {len(problem_ids)} problems. Starting single run for each using model: {selected_model}") |
|
|
| collected_scores = [] |
|
|
| with concurrent.futures.ThreadPoolExecutor(max_workers=API_WORKERS) as executor: |
| futures = [executor.submit(process_single_attempt, pid, selected_model) for pid in problem_ids] |
| |
| for future in concurrent.futures.as_completed(futures): |
| try: |
| |
| raw_score = future.result() |
| |
| |
| if raw_score == -1: |
| effective_score = 0 |
| else: |
| effective_score = raw_score |
| |
| collected_scores.append(effective_score) |
| |
| except Exception as e: |
| print(f"A job failed in the thread pool: {e}") |
| collected_scores.append(0) |
|
|
| print("\nAll problems have been processed.") |
| |
| |
| if collected_scores: |
| avg_score = sum(collected_scores) / len(collected_scores) |
| print("-" * 30) |
| print(f"Final Average Score: {avg_score:.2f}") |
| print("-" * 30) |
| else: |
| print("No scores collected.") |