| import argparse |
| import subprocess |
| import yaml |
| import os |
| import json |
| import logging |
| import datetime |
| from typing import Dict, Any, List, Set, Tuple |
| from datasets import Dataset, Features, Value, Image as HFImage |
| from tqdm import tqdm |
| from PIL import Image as PILImage |
|
|
| |
| GREEN = '\033[92m' |
| RED = '\033[91m' |
| RESET = '\033[0m' |
| YELLOW = '\033[93m' |
| CYAN = '\033[96m' |
| MAGENTA = '\033[95m' |
|
|
| |
| from utils import load_api_key |
| from llm_interface import get_openrouter_prediction |
| |
| from evaluation import calculate_exam_scores, calculate_single_question_score_details |
|
|
| |
| logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s') |
|
|
|
|
| def get_git_sha() -> str | None: |
| """Returns the current git commit SHA, or None if not in a git repo.""" |
| try: |
| result = subprocess.run( |
| ["git", "rev-parse", "HEAD"], |
| capture_output=True, text=True, timeout=5 |
| ) |
| if result.returncode == 0: |
| return result.stdout.strip() |
| except Exception: |
| pass |
| return None |
|
|
|
|
| def get_available_models(config_path: str) -> List[str]: |
| """Loads models from the benchmark configuration YAML file.""" |
| try: |
| with open(config_path, 'r') as f: |
| config = yaml.safe_load(f) |
| models = config.get("openrouter_models", []) |
| if not models: |
| logging.warning(f"No models found in {config_path} under 'openrouter_models'.") |
| return models |
| except FileNotFoundError: |
| logging.error(f"Configuration file not found at {config_path} for model retrieval.") |
| return [] |
| except yaml.YAMLError as e: |
| logging.error(f"Error parsing configuration file {config_path} for model retrieval: {e}") |
| return [] |
| except Exception as e: |
| logging.error(f"Unexpected error retrieving models from {config_path}: {e}") |
| return [] |
|
|
| def get_available_exam_details(metadata_path: str) -> Tuple[List[str], List[str]]: |
| """Reads metadata.jsonl to get unique exam names and years.""" |
| exam_names: Set[str] = set() |
| exam_years: Set[str] = set() |
| try: |
| with open(metadata_path, 'r') as f: |
| for line in f: |
| try: |
| data = json.loads(line) |
| if 'exam_name' in data: |
| exam_names.add(data['exam_name']) |
| if 'exam_year' in data: |
| exam_years.add(str(data['exam_year'])) |
| except json.JSONDecodeError: |
| logging.warning(f"Skipping malformed JSON line in {metadata_path}: {line.strip()}") |
| |
| sorted_exam_names = sorted(list(exam_names)) |
| sorted_exam_years = sorted(list(exam_years)) |
| |
| if not sorted_exam_names: |
| logging.warning(f"No exam names found in {metadata_path}.") |
| if not sorted_exam_years: |
| logging.warning(f"No exam years found in {metadata_path}.") |
| |
| return sorted_exam_names, sorted_exam_years |
| except FileNotFoundError: |
| logging.error(f"Metadata file not found at {metadata_path}.") |
| return [], [] |
| except Exception as e: |
| logging.error(f"Unexpected error reading or parsing {metadata_path}: {e}") |
| return [], [] |
|
|
| def load_config(config_path: str) -> dict: |
| """Loads the benchmark configuration from a YAML file.""" |
| try: |
| with open(config_path, 'r') as f: |
| config = yaml.safe_load(f) |
| logging.info(f"Configuration loaded from {config_path}") |
| return config |
| except FileNotFoundError: |
| logging.error(f"Configuration file not found at {config_path}") |
| raise |
| except yaml.YAMLError as e: |
| logging.error(f"Error parsing configuration file {config_path}: {e}") |
| raise |
|
|
| def append_prediction(result: Dict[str, Any], filepath: str): |
| """Appends a single prediction result to a JSONL file.""" |
| |
| |
| prediction_data = result.copy() |
| prediction_data.pop('marks_awarded', None) |
| prediction_data.pop('evaluation_status', None) |
| prediction_data.pop('predicted_answer', None) |
| prediction_data.pop('ground_truth', None) |
| try: |
| with open(filepath, 'a') as f: |
| json.dump(prediction_data, f) |
| f.write('\n') |
| except IOError as e: |
| logging.error(f"Failed to append prediction to {filepath}: {e}") |
| except Exception as e: |
| logging.error(f"Unexpected error appending prediction to {filepath}: {e}") |
|
|
| def append_summary_detail(result_detail: Dict[str, Any], filepath: str): |
| """Appends a single question's summary details (evaluation status, marks, predicted, truth) to a JSONL file.""" |
| try: |
| with open(filepath, 'a') as f: |
| json.dump(result_detail, f) |
| f.write('\n') |
| except IOError as e: |
| logging.error(f"Failed to append summary detail to {filepath}: {e}") |
| except Exception as e: |
| logging.error(f"Unexpected error appending summary detail to {filepath}: {e}") |
|
|
|
|
| |
|
|
| def generate_markdown_summary(summary: Dict[str, Any], filepath: str): |
| """Generates a human-readable Markdown summary from the results dictionary.""" |
| try: |
| md_content = [] |
| model_name = summary.get("model_name", "N/A") |
| exam_name = summary.get("exam_name", "N/A") |
| exam_year = summary.get("exam_year", "N/A") |
| timestamp = summary.get("timestamp", "N/A") |
| total_questions_in_dataset = summary.get("total_questions_in_dataset", 0) |
| total_questions_processed_in_run = summary.get("total_questions_processed_in_run", 0) |
|
|
| filtered_questions_count = 0 |
| if total_questions_in_dataset > 0 and total_questions_processed_in_run > 0: |
| filtered_questions_count = total_questions_in_dataset - total_questions_processed_in_run |
|
|
|
|
| md_content.append(f"# Benchmark Results: {model_name}") |
| if exam_name and exam_name not in ["N/A", "All_Exams"]: |
| md_content.append(f"**Exam Name:** {exam_name}") |
| if exam_year and exam_year not in ["N/A", "All_Years"]: |
| md_content.append(f"**Exam Year:** {exam_year}") |
| md_content.append(f"**Timestamp:** {timestamp}") |
| md_content.append(f"**Total Questions in Dataset:** {total_questions_in_dataset if total_questions_in_dataset > 0 else 'N/A'}") |
| if filtered_questions_count > 0: |
| md_content.append(f"**Questions Filtered Out:** {filtered_questions_count}") |
| md_content.append(f"**Total Questions Processed in this Run:** {total_questions_processed_in_run}") |
|
|
| |
| total_cost_usd = summary.get("total_cost_usd") |
| total_tokens = summary.get("total_tokens") |
| avg_latency = summary.get("avg_response_latency_ms") |
| median_latency = summary.get("median_response_latency_ms") |
| if total_tokens or total_cost_usd or avg_latency: |
| md_content.append("") |
| md_content.append("### API Usage") |
| if summary.get("total_prompt_tokens"): |
| md_content.append(f"- **Prompt Tokens:** {summary['total_prompt_tokens']:,}") |
| if summary.get("total_completion_tokens"): |
| md_content.append(f"- **Completion Tokens:** {summary['total_completion_tokens']:,}") |
| if total_tokens: |
| md_content.append(f"- **Total Tokens:** {total_tokens:,}") |
| if total_cost_usd: |
| md_content.append(f"- **Total Cost:** ${total_cost_usd:.4f}") |
| if avg_latency: |
| md_content.append(f"- **Avg Response Latency:** {avg_latency:,} ms") |
| if median_latency: |
| md_content.append(f"- **Median Response Latency:** {median_latency:,} ms") |
|
|
| md_content.append("\n---\n") |
|
|
| |
| if "overall_score" in summary and "section_breakdown" in summary: |
| total_processed = summary.get("total_questions_processed", 0) |
| |
| overall_score = summary.get('overall_score', 'N/A') |
| total_possible_score = summary.get('total_possible_score_for_processed_questions', 'N/A') |
| correct_full_count = summary.get('overall_correct_full', 'N/A') |
| partial_correct_count = summary.get('overall_partial_correct', 'N/A') |
| incorrect_choice_count = summary.get('overall_incorrect_choice', 'N/A') |
| skipped_count = summary.get('overall_skipped', 'N/A') |
| failures_count = summary.get('overall_api_parse_failures', 'N/A') |
| unmapped_count = summary.get('unmapped_section_questions', 'N/A') |
|
|
| md_content.append("## Exam Scoring Results") |
| md_content.append(f"**Overall Score:** **{overall_score}** / **{total_possible_score}**") |
| md_content.append(f"- **Fully Correct Answers:** {correct_full_count}") |
| if partial_correct_count != 'N/A' and partial_correct_count > 0 : |
| md_content.append(f"- **Partially Correct Answers:** {partial_correct_count}") |
| md_content.append(f"- **Incorrectly Answered (Choice Made):** {incorrect_choice_count}") |
| md_content.append(f"- **Skipped Questions:** {skipped_count}") |
| md_content.append(f"- **API/Parse Failures:** {failures_count}") |
| md_content.append(f"- **Total Questions Processed:** {total_processed}") |
| if unmapped_count > 0: |
| md_content.append(f"- **Unmapped Section Questions:** {unmapped_count} *(Not included in section breakdown)*") |
|
|
| md_content.append("\n### Detailed Score Calculation by Question Type") |
| question_type_breakdown = summary.get("question_type_breakdown", {}) |
| if question_type_breakdown: |
| sorted_q_types = sorted(question_type_breakdown.keys()) |
| for q_type in sorted_q_types: |
| stats = question_type_breakdown[q_type] |
| q_type_display = q_type.replace('_', ' ').title() |
|
|
| correct_count_q = stats.get('correct_full', 0) |
| partial_count_q = stats.get('partial_correct', 0) |
| incorrect_count_q = stats.get('incorrect_choice', 0) |
| skipped_count_q = stats.get('skipped', 0) |
| api_fail_count_q = stats.get('api_parse_failures', 0) |
| score_q = stats.get('score', 0) |
|
|
| breakdown_parts = [] |
| if correct_count_q > 0: |
| breakdown_parts.append(f"{correct_count_q} Correct") |
| if partial_count_q > 0: |
| breakdown_parts.append(f"{partial_count_q} Partial") |
| if incorrect_count_q > 0: |
| breakdown_parts.append(f"{incorrect_count_q} Incorrect") |
| if skipped_count_q > 0: |
| breakdown_parts.append(f"{skipped_count_q} Skipped") |
| if api_fail_count_q > 0: |
| breakdown_parts.append(f"{api_fail_count_q} API/Parse Fail") |
|
|
| breakdown_str = ", ".join(breakdown_parts) if breakdown_parts else "No questions processed" |
|
|
| md_content.append(f"**{q_type_display} ({stats.get('count', 0)} questions):** {score_q} marks") |
| md_content.append(f" *Breakdown:* {breakdown_str}") |
| else: |
| md_content.append("No question type breakdown available.") |
|
|
| md_content.append("\n### Section Breakdown") |
| md_content.append("| Section | Score | Fully Correct | Partially Correct | Incorrect Choice | Skipped | API/Parse Failures |") |
| md_content.append("|---------------|-------|---------------|-------------------|------------------|---------|--------------------|") |
| section_breakdown = summary.get("section_breakdown", {}) |
| |
| sorted_section_names = sorted(section_breakdown.keys()) |
| if not sorted_section_names and section_breakdown: |
| logging.warning("Could not sort section names for Markdown summary; using unsorted.") |
| sorted_section_names = list(section_breakdown.keys()) |
|
|
| for section_name in sorted_section_names: |
| stats = section_breakdown.get(section_name, {}) |
| score = stats.get('score', 'N/A') |
| s_correct = stats.get('correct', 'N/A') |
| s_partial = stats.get('partial_correct', 'N/A') |
| s_incorrect = stats.get('incorrect', 'N/A') |
| s_skipped = stats.get('skipped', 'N/A') |
| s_failures = stats.get('api_parse_failures', 'N/A') |
| display_section_name = section_name.replace('_', ' ') |
| md_content.append(f"| {display_section_name:<13} | {score:<5} | {s_correct:<13} | {s_partial:<17} | {s_incorrect:<16} | {s_skipped:<7} | {s_failures:<18} |") |
| if not sorted_section_names: |
| md_content.append("| No section data available | N/A | N/A | N/A | N/A | N/A | N/A |") |
| |
| |
| elif "accuracy_on_parsed" in summary: |
| md_content.append("## Simple Accuracy Results (Fallback)") |
| md_content.append(f"- **Accuracy (on successfully parsed non-skipped):** {summary.get('accuracy_on_parsed', 'N/A'):.4f}") |
| md_content.append(f"- **Total Processed Attempts:** {summary.get('total_processed_attempts', 'N/A')}") |
| |
| else: |
| md_content.append("## Summary") |
| md_content.append("*(No specific Exam Scoring or Accuracy metrics found in summary)*") |
|
|
|
|
| with open(filepath, 'w') as f: |
| f.write("\n".join(md_content)) |
| logging.info(f"Markdown summary saved to {filepath}") |
|
|
| except IOError as e: |
| logging.error(f"Failed to save markdown summary to {filepath}: {e}") |
| except Exception as e: |
| logging.error(f"Unexpected error generating or saving markdown summary to {filepath}: {e}") |
|
|
|
|
| def process_question( |
| model_id: str, |
| api_key: str, |
| config: dict, |
| example: dict, |
| image: PILImage.Image, |
| predictions_path: str, |
| summary_details_path: str, |
| attempt: int = 1, |
| ) -> Dict[str, Any]: |
| """Handles a single question: API call, re-prompt on parse failure, scoring, writing to disk. |
| |
| Returns the result_data dict. On API exception, raises so the caller can handle retry queueing. |
| """ |
| question_id = example["question_id"] |
| subject = example["subject"] |
| exam_name_from_data = example.get("exam_name", "UNKNOWN_EXAM") |
| question_type_from_data = example.get("question_type", "MCQ_SINGLE_CORRECT") |
| truth = json.loads(example["correct_answer"]) |
|
|
| result_data = { |
| "question_id": question_id, |
| "subject": subject, |
| "exam_name": exam_name_from_data, |
| "question_type": question_type_from_data, |
| "ground_truth": truth, |
| "predicted_answer": None, |
| "raw_response": None, |
| "parse_successful": False, |
| "api_call_successful": False, |
| "error": None, |
| "attempt": attempt, |
| "previous_raw_response_on_reprompt": None, |
| "response_metadata": None, |
| } |
|
|
| |
| logging.info(f"Attempting API call for question: {question_id} with model: {model_id}") |
| parsed_answer, raw_response, response_metadata = get_openrouter_prediction( |
| model_identifier=model_id, |
| api_key=api_key, |
| image=image, |
| exam_name=exam_name_from_data, |
| exam_year=str(example.get("exam_year", "UNKNOWN_YEAR")), |
| question_type=question_type_from_data, |
| max_tokens=config.get("max_tokens", 100), |
| request_timeout=config.get("request_timeout", 60), |
| temperature=config.get("temperature", 0), |
| ) |
|
|
| api_success = raw_response is not None |
| parse_success = parsed_answer is not None |
|
|
| |
| if api_success and not parse_success and raw_response is not None: |
| logging.warning(f"Question {question_id}: Initial parse failed. Attempting re-prompt.") |
| result_data["previous_raw_response_on_reprompt"] = raw_response |
| try: |
| parsed_answer_rp, raw_response_rp, rp_metadata = get_openrouter_prediction( |
| model_identifier=model_id, |
| api_key=api_key, |
| previous_raw_response=raw_response, |
| question_type=question_type_from_data, |
| max_tokens=config.get("max_tokens", 100), |
| request_timeout=config.get("request_timeout", 60), |
| temperature=config.get("temperature", 0), |
| ) |
| if isinstance(parsed_answer_rp, list): |
| parsed_answer_rp = [str(item) for item in parsed_answer_rp] |
|
|
| result_data.update({ |
| "predicted_answer": parsed_answer_rp, |
| "raw_response": raw_response_rp, |
| "parse_successful": parsed_answer_rp is not None, |
| "api_call_successful": True, |
| "attempt": attempt + 1, |
| "response_metadata": rp_metadata, |
| }) |
| logging.info(f"Question {question_id}: Re-prompt {'succeeded' if result_data['parse_successful'] else 'failed to parse'}.") |
| except Exception as e_rp: |
| logging.error(f"Re-prompt API call failed for question {question_id}: {e_rp}") |
| result_data.update({ |
| "predicted_answer": None, |
| "raw_response": raw_response, |
| "parse_successful": False, |
| "api_call_successful": True, |
| "error": f"Initial parse failed. Re-prompt API call failed: {str(e_rp)}", |
| "attempt": attempt, |
| "response_metadata": response_metadata, |
| }) |
| else: |
| current_error = result_data.get("error") |
| if not api_success: |
| current_error = "API call returned empty content." |
|
|
| if isinstance(parsed_answer, list): |
| parsed_answer = [str(item) for item in parsed_answer] |
|
|
| result_data.update({ |
| "predicted_answer": parsed_answer, |
| "raw_response": raw_response, |
| "parse_successful": parse_success, |
| "api_call_successful": api_success, |
| "error": current_error, |
| "attempt": attempt, |
| "response_metadata": response_metadata, |
| }) |
|
|
| |
| score_details = calculate_single_question_score_details(result_data) |
| result_data["marks_awarded"] = score_details.get("marks_awarded") |
| result_data["evaluation_status"] = score_details.get("evaluation_status") |
|
|
| metadata = result_data.get("response_metadata") or {} |
| append_summary_detail( |
| { |
| "question_id": question_id, |
| "marks_awarded": result_data["marks_awarded"], |
| "evaluation_status": result_data["evaluation_status"], |
| "predicted_answer": result_data["predicted_answer"], |
| "ground_truth": result_data["ground_truth"], |
| "attempt": result_data["attempt"], |
| "prompt_tokens": metadata.get("prompt_tokens"), |
| "completion_tokens": metadata.get("completion_tokens"), |
| "cost": metadata.get("cost"), |
| "response_latency_ms": metadata.get("response_latency_ms"), |
| }, |
| summary_details_path, |
| ) |
| append_prediction(result_data, predictions_path) |
|
|
| return result_data |
|
|
|
|
| def log_question_result(result_data: Dict[str, Any], prefix: str = "") -> str: |
| """Logs a color-coded result for a single question. |
| |
| Returns a category string: 'correct', 'incorrect', 'skipped', 'parse_fail', or 'api_fail'. |
| """ |
| question_id = result_data["question_id"] |
| log_message_prefix = f"{prefix}Question {question_id}:" |
| log_message_suffix = f"(Attempt {result_data['attempt']})" |
|
|
| if not result_data["api_call_successful"]: |
| print(f"{MAGENTA}{log_message_prefix} API Call Failed {log_message_suffix}{RESET}") |
| return "api_fail" |
|
|
| if not result_data["parse_successful"]: |
| print(f"{CYAN}{log_message_prefix} Failed to parse answer {log_message_suffix}{RESET}") |
| return "parse_fail" |
|
|
| if result_data["predicted_answer"] == "SKIP": |
| print(f"{YELLOW}{log_message_prefix} Skipped {log_message_suffix}{RESET}") |
| return "skipped" |
|
|
| marks_awarded = result_data.get("marks_awarded", 0) |
| evaluation_status_value = result_data.get("evaluation_status") |
|
|
| is_considered_correct = False |
| log_display_status = "N/A" |
| status_check_string = "" |
|
|
| if evaluation_status_value is True: |
| is_considered_correct = True |
| log_display_status = "True (Boolean)" |
| status_check_string = "CORRECT_TRUE_BOOLEAN" |
| elif isinstance(evaluation_status_value, str): |
| log_display_status = evaluation_status_value |
| status_check_string = evaluation_status_value.strip().upper() |
| if "CORRECT" in status_check_string: |
| is_considered_correct = True |
| elif evaluation_status_value is None: |
| log_display_status = "None" |
| status_check_string = "NONE_STATUS" |
| else: |
| log_display_status = str(evaluation_status_value) |
| status_check_string = str(evaluation_status_value).strip().upper() |
|
|
| known_eval_skip_statuses = ["SKIPPED_BY_EVAL", "SKIPPED"] |
|
|
| if is_considered_correct: |
| print(f"{GREEN}{log_message_prefix} Correct - Marks: {marks_awarded}, Status: {log_display_status} {log_message_suffix}{RESET}") |
| return "correct" |
| elif status_check_string in known_eval_skip_statuses: |
| print(f"{YELLOW}{log_message_prefix} Skipped by Eval - Marks: {marks_awarded}, Status: {log_display_status} {log_message_suffix}{RESET}") |
| return "skipped" |
| else: |
| print(f"{RED}{log_message_prefix} Incorrect - Marks: {marks_awarded}, Status: {log_display_status} {log_message_suffix}{RESET}") |
| return "incorrect" |
|
|
|
|
| def load_completed_question_ids(summary_details_path: str) -> set: |
| """Reads summary.jsonl and returns a set of question_ids that have already been processed.""" |
| completed_ids = set() |
| if not os.path.exists(summary_details_path): |
| return completed_ids |
| try: |
| with open(summary_details_path, 'r') as f: |
| for line in f: |
| try: |
| data = json.loads(line) |
| qid = data.get("question_id") |
| if qid: |
| completed_ids.add(qid) |
| except json.JSONDecodeError: |
| continue |
| except IOError as e: |
| logging.warning(f"Could not read {summary_details_path} for resume: {e}") |
| return completed_ids |
|
|
|
|
| def run_benchmark( |
| config: dict, |
| api_key: str, |
| model_to_run: str, |
| output_dir_override: str | None = None, |
| exam_name_choice: str | None = None, |
| exam_year_choice: str | None = None, |
| question_ids_str: str | None = None, |
| resume_dir: str | None = None |
| ): |
| """Runs the benchmark evaluation loop with incremental saving and retries.""" |
|
|
| |
| models_to_run = [model_to_run] |
| logging.info(f"Target model for this run: {model_to_run}") |
|
|
| |
| base_output_dir = output_dir_override if output_dir_override else config.get("results_base_dir", "results") |
| os.makedirs(base_output_dir, exist_ok=True) |
|
|
| |
| metadata_path = config.get("metadata_path", "images/metadata.jsonl") |
| images_base_dir = config.get("images_base_dir", "images") |
| try: |
| records = [] |
| with open(metadata_path, 'r', encoding='utf-8') as f: |
| for line in f: |
| row = json.loads(line) |
| row["image"] = os.path.join(images_base_dir, row.pop("file_name")) |
| row.setdefault("paper_id", None) |
| records.append(row) |
|
|
| features = Features({ |
| "image": HFImage(decode=True), |
| "question_id": Value("string"), |
| "exam_name": Value("string"), |
| "exam_year": Value("int32"), |
| "subject": Value("string"), |
| "question_type": Value("string"), |
| "correct_answer": Value("string"), |
| "paper_id": Value("int64"), |
| }) |
| dataset = Dataset.from_list(records, features=features) |
| logging.info(f"Dataset loaded successfully from {metadata_path}. Original number of questions: {len(dataset)}") |
| except Exception as e: |
| logging.error(f"Failed to load dataset from '{metadata_path}': {e}") |
| logging.error("Ensure 'metadata.jsonl' exists and image paths are valid.") |
| return |
|
|
| |
| original_dataset_size = len(dataset) |
| |
| |
| if exam_name_choice and exam_name_choice.lower() != "all": |
| logging.info(f"Filtering dataset for exam_name: '{exam_name_choice}'") |
| dataset = dataset.filter(lambda example: example.get('exam_name') == exam_name_choice) |
| logging.info(f"Dataset size after exam_name filter: {len(dataset)} questions.") |
|
|
| |
| if exam_year_choice and exam_year_choice.lower() != "all": |
| try: |
| filter_year_int = int(exam_year_choice) |
| logging.info(f"Filtering dataset for exam_year: {filter_year_int}") |
| dataset = dataset.filter(lambda example: example.get('exam_year') == filter_year_int) |
| logging.info(f"Dataset size after exam_year filter: {len(dataset)} questions.") |
| except ValueError: |
| logging.error(f"Invalid exam_year provided: '{exam_year_choice}'. Must be an integer or 'all'. Year filtering skipped.") |
|
|
| |
| if question_ids_str: |
| try: |
| target_question_ids = {q_id.strip() for q_id in question_ids_str.split(',') if q_id.strip()} |
| if target_question_ids: |
| logging.info(f"Filtering dataset for specific question IDs: {target_question_ids}") |
| dataset = dataset.filter(lambda example: example.get('question_id') in target_question_ids) |
| logging.info(f"Dataset size after question_id filter: {len(dataset)} questions.") |
| else: |
| logging.warning("Empty or invalid question_ids string provided. No question ID filtering applied.") |
| except Exception as e: |
| logging.error(f"Error processing question_ids_str '{question_ids_str}': {e}. No question ID filtering applied.") |
| |
| if len(dataset) < original_dataset_size: |
| logging.info(f"Final dataset size after all filters: {len(dataset)} (originally {original_dataset_size}).") |
| |
| if len(dataset) == 0: |
| logging.warning("No questions to process after filtering. Skipping model benchmark.") |
| return |
|
|
| |
| for model_id in models_to_run: |
| logging.info(f"--- Starting benchmark for model: {model_id} ---") |
|
|
| |
| if resume_dir: |
| model_output_dir = resume_dir |
| timestamp = os.path.basename(resume_dir).rsplit('_', 2)[-2] + '_' + os.path.basename(resume_dir).rsplit('_', 1)[-1] if '_' in os.path.basename(resume_dir) else datetime.datetime.now().strftime("%Y%m%d_%H%M%S") |
| else: |
| timestamp = datetime.datetime.now().strftime("%Y%m%d_%H%M%S") |
| safe_model_name = model_id.replace('/', '_') |
| dir_name_parts = [safe_model_name] |
|
|
| current_exam_name_for_dir = exam_name_choice if exam_name_choice and exam_name_choice.lower() != "all" else "AllExams" |
| current_exam_year_for_dir = exam_year_choice if exam_year_choice and exam_year_choice.lower() != "all" else "AllYears" |
|
|
| if current_exam_name_for_dir != "AllExams": |
| dir_name_parts.append(current_exam_name_for_dir.replace('/', '_')) |
| if current_exam_year_for_dir != "AllYears": |
| dir_name_parts.append(str(current_exam_year_for_dir)) |
|
|
| dir_name_parts.append(timestamp) |
|
|
| model_output_dir_name = "_".join(filter(None, dir_name_parts)) |
| model_output_dir = os.path.join(base_output_dir, model_output_dir_name) |
|
|
| os.makedirs(model_output_dir, exist_ok=True) |
| predictions_path = os.path.join(model_output_dir, "predictions.jsonl") |
| summary_details_path = os.path.join(model_output_dir, "summary.jsonl") |
| markdown_summary_path = os.path.join(model_output_dir, "summary.md") |
| logging.info(f"Results for {model_id} will be saved to: {model_output_dir}") |
|
|
| |
| if resume_dir: |
| completed_ids = load_completed_question_ids(summary_details_path) |
| if completed_ids: |
| logging.info(f"Resuming: found {len(completed_ids)} already-completed questions. Skipping them.") |
| dataset = dataset.filter(lambda example: example.get('question_id') not in completed_ids) |
| logging.info(f"Remaining questions to process: {len(dataset)}") |
| if len(dataset) == 0: |
| logging.info("All questions already completed. Nothing to resume.") |
| return |
|
|
| current_total_questions = len(dataset) |
| logging.info(f"Processing {current_total_questions} questions for model: {model_id}") |
|
|
| model_results = [] |
| failed_questions_data = [] |
|
|
| |
| initial_correct_count = 0 |
| initial_incorrect_count = 0 |
| initial_skipped_count = 0 |
| initial_parse_fail_count = 0 |
| initial_api_fail_count = 0 |
|
|
| |
| pbar_initial = tqdm(dataset, desc=f"Processing {model_id} (Initial Pass)", total=current_total_questions) |
| for example in pbar_initial: |
| try: |
| result_data = process_question( |
| model_id=model_id, api_key=api_key, config=config, |
| example=example, image=example["image"], |
| predictions_path=predictions_path, summary_details_path=summary_details_path, |
| attempt=1, |
| ) |
| model_results.append(result_data) |
| category = log_question_result(result_data) |
| except Exception as e: |
| logging.error(f"Initial API call failed for question {example['question_id']} (Attempt 1): {e}") |
| failed_questions_data.append(example) |
| category = "api_fail" |
|
|
| if category == "correct": |
| initial_correct_count += 1 |
| elif category == "incorrect": |
| initial_incorrect_count += 1 |
| elif category == "skipped": |
| initial_skipped_count += 1 |
| elif category == "parse_fail": |
| initial_parse_fail_count += 1 |
| elif category == "api_fail": |
| initial_api_fail_count += 1 |
| pbar_initial.set_postfix_str(f"✓:{initial_correct_count} ✗:{initial_incorrect_count} S:{initial_skipped_count} P!:{initial_parse_fail_count} A!:{initial_api_fail_count}") |
|
|
| pbar_initial.close() |
|
|
| |
| if failed_questions_data: |
| logging.info(f"--- Retrying {len(failed_questions_data)} questions with initial API failures for model: {model_id} ---") |
|
|
| retry_correct_count = 0 |
| retry_incorrect_count = 0 |
| retry_skipped_count = 0 |
| retry_parse_fail_count = 0 |
| retry_api_fail_count = 0 |
|
|
| pbar_retry = tqdm(failed_questions_data, desc=f"Processing {model_id} (API Retry Pass)", total=len(failed_questions_data)) |
| for example_retry in pbar_retry: |
| try: |
| result_data_retry = process_question( |
| model_id=model_id, api_key=api_key, config=config, |
| example=example_retry, image=example_retry["image"], |
| predictions_path=predictions_path, summary_details_path=summary_details_path, |
| attempt=2, |
| ) |
| model_results.append(result_data_retry) |
| category = log_question_result(result_data_retry, prefix="(Retry) ") |
| except Exception as e_retry_api: |
| logging.error(f"API call failed permanently for question {example_retry['question_id']} (Attempt 2 API Retry): {e_retry_api}") |
| |
| fail_data = { |
| "question_id": example_retry["question_id"], |
| "subject": example_retry["subject"], |
| "exam_name": example_retry.get("exam_name", "UNKNOWN_EXAM"), |
| "question_type": example_retry.get("question_type", "MCQ_SINGLE_CORRECT"), |
| "ground_truth": json.loads(example_retry["correct_answer"]), |
| "predicted_answer": None, "raw_response": None, |
| "parse_successful": False, "api_call_successful": False, |
| "error": f"Initial API fail. Retry API call also failed: {str(e_retry_api)}", |
| "attempt": 2, "previous_raw_response_on_reprompt": None, |
| } |
| score_details = calculate_single_question_score_details(fail_data) |
| fail_data["marks_awarded"] = score_details.get("marks_awarded") |
| fail_data["evaluation_status"] = score_details.get("evaluation_status") |
| append_summary_detail( |
| {"question_id": fail_data["question_id"], "marks_awarded": fail_data["marks_awarded"], |
| "evaluation_status": fail_data["evaluation_status"], "predicted_answer": None, |
| "ground_truth": fail_data["ground_truth"], "attempt": 2}, |
| summary_details_path, |
| ) |
| append_prediction(fail_data, predictions_path) |
| model_results.append(fail_data) |
| category = "api_fail" |
|
|
| if category == "correct": |
| retry_correct_count += 1 |
| elif category == "incorrect": |
| retry_incorrect_count += 1 |
| elif category == "skipped": |
| retry_skipped_count += 1 |
| elif category == "parse_fail": |
| retry_parse_fail_count += 1 |
| elif category == "api_fail": |
| retry_api_fail_count += 1 |
| pbar_retry.set_postfix_str(f"✓:{retry_correct_count} ✗:{retry_incorrect_count} S:{retry_skipped_count} P!:{retry_parse_fail_count} A!:{retry_api_fail_count}") |
| pbar_retry.close() |
|
|
| |
| logging.info(f"--- Calculating final results for model: {model_id} ---") |
| |
| |
| evaluation_summary = calculate_exam_scores(model_results) |
| |
| |
| summary_exam_name_display = exam_name_choice if exam_name_choice and exam_name_choice.lower() != "all" else "All_Exams" |
| summary_exam_year_display = exam_year_choice if exam_year_choice and exam_year_choice.lower() != "all" else "All_Years" |
|
|
| |
| total_prompt_tokens = 0 |
| total_completion_tokens = 0 |
| total_cost = 0.0 |
| latencies = [] |
| for r in model_results: |
| meta = r.get("response_metadata") or {} |
| if meta.get("prompt_tokens"): |
| total_prompt_tokens += meta["prompt_tokens"] |
| if meta.get("completion_tokens"): |
| total_completion_tokens += meta["completion_tokens"] |
| if meta.get("cost"): |
| total_cost += meta["cost"] |
| if meta.get("response_latency_ms"): |
| latencies.append(meta["response_latency_ms"]) |
|
|
| summary = { |
| "model_name": model_id, |
| "exam_name": summary_exam_name_display, |
| "exam_year": summary_exam_year_display, |
| "question_ids_filter": question_ids_str if question_ids_str else "None", |
| "timestamp": timestamp, |
| "git_sha": get_git_sha(), |
| "temperature": config.get("temperature", 0), |
| "max_tokens": config.get("max_tokens", 100), |
| "total_questions_in_dataset": original_dataset_size, |
| "total_questions_processed_in_run": len(dataset), |
| "total_prompt_tokens": total_prompt_tokens, |
| "total_completion_tokens": total_completion_tokens, |
| "total_tokens": total_prompt_tokens + total_completion_tokens, |
| "total_cost_usd": round(total_cost, 6), |
| "avg_response_latency_ms": round(sum(latencies) / len(latencies)) if latencies else None, |
| "median_response_latency_ms": round(sorted(latencies)[len(latencies) // 2]) if latencies else None, |
| **evaluation_summary |
| } |
| logging.info(f"Overall Score: {summary.get('overall_score')}") |
| logging.info(f"Full Correct: {summary.get('overall_correct_full')}, Partial Correct: {summary.get('overall_partial_correct')}, Incorrect Choice: {summary.get('overall_incorrect_choice')}, Skipped: {summary.get('overall_skipped')}, API/Parse Failures: {summary.get('overall_api_parse_failures')}") |
| |
| logging.info(f"--- Results Summary for model: {model_id} ---") |
| logging.info(json.dumps(summary, indent=2, sort_keys=True)) |
| logging.info("-------------------------------------") |
|
|
| |
| |
| |
| |
|
|
|
|
| |
| |
| generate_markdown_summary(summary, markdown_summary_path) |
|
|
| logging.info("Benchmark run completed.") |
|
|
|
|
| if __name__ == "__main__": |
| |
| |
| default_config_path = "configs/benchmark_config.yaml" |
| default_metadata_path = "images/metadata.jsonl" |
|
|
| available_models = get_available_models(default_config_path) |
| available_exam_names, available_exam_years = get_available_exam_details(default_metadata_path) |
|
|
| |
| exam_name_choices = ["all"] + available_exam_names |
| exam_year_choices = ["all"] + available_exam_years |
|
|
| parser = argparse.ArgumentParser(description="Run JEE/NEET LLM Benchmark.") |
| parser.add_argument( |
| "--config", |
| type=str, |
| default=default_config_path, |
| help=f"Path to the benchmark configuration YAML file (default: {default_config_path})." |
| ) |
| parser.add_argument( |
| "--model", |
| type=str, |
| required=True if available_models else False, |
| choices=available_models if available_models else None, |
| help="Select the model to run." + (f" Available: {', '.join(available_models)}." if available_models else " (No models found in config)") |
| ) |
| parser.add_argument( |
| "--output_dir", |
| type=str, |
| help="Override the base output directory specified in the config file." |
| ) |
| parser.add_argument( |
| "--exam_name", |
| type=str, |
| default="all", |
| choices=exam_name_choices if exam_name_choices else ["all"], |
| help="Select the exam name to run, or 'all' for all exams." + (f" Available: {', '.join(available_exam_names)}." if available_exam_names else "") |
| ) |
| parser.add_argument( |
| "--exam_year", |
| type=str, |
| default="all", |
| choices=exam_year_choices if exam_year_choices else ["all"], |
| help="Select the exam year to run, or 'all' for all years." + (f" Available: {', '.join(available_exam_years)}." if available_exam_years else "") |
| ) |
| parser.add_argument( |
| "--question_ids", |
| type=str, |
| default=None, |
| help="Optional: Comma-separated list of specific question IDs to run (e.g., ID1,ID2,ID3)." |
| ) |
| parser.add_argument( |
| "--resume", |
| type=str, |
| default=None, |
| help="Optional: Path to an existing results directory to resume an interrupted run." |
| ) |
| parser.add_argument( |
| "--temperature", |
| type=float, |
| default=None, |
| help="Override sampling temperature (default: from config, typically 0 for deterministic output)." |
| ) |
| parser.add_argument( |
| "--num_runs", |
| type=int, |
| default=1, |
| help="Number of independent runs per model (default: 1). Use 3+ for publication-grade variance analysis." |
| ) |
| args = parser.parse_args() |
|
|
| |
| if args.config != default_config_path: |
| logging.info(f"User provided config path: {args.config}. Re-fetching models if necessary.") |
| |
| if not available_models or args.model not in available_models: |
| user_config_models = get_available_models(args.config) |
| if args.model not in user_config_models: |
| logging.error(f"Selected model '{args.model}' not found in the specified config '{args.config}'. Exiting.") |
| exit(1) |
| |
| |
|
|
| try: |
| |
| api_key = load_api_key() |
| |
| config = load_config(args.config) |
|
|
| |
| if args.temperature is not None: |
| config["temperature"] = args.temperature |
|
|
| |
| if args.model not in config.get("openrouter_models", []): |
| logging.error(f"The model '{args.model}' is not listed in '{args.config}'. Please check the model name or the config file.") |
| exit(1) |
|
|
| |
| for run_num in range(1, args.num_runs + 1): |
| if args.num_runs > 1: |
| logging.info(f"=== Starting run {run_num}/{args.num_runs} ===") |
| run_benchmark( |
| config=config, |
| api_key=api_key, |
| model_to_run=args.model, |
| output_dir_override=args.output_dir, |
| exam_name_choice=args.exam_name, |
| exam_year_choice=args.exam_year, |
| question_ids_str=args.question_ids, |
| resume_dir=args.resume if run_num == 1 else None, |
| ) |
| except (ValueError, FileNotFoundError, yaml.YAMLError) as e: |
| logging.error(f"Setup failed: {e}") |
| except Exception as e: |
| logging.error(f"An unexpected error occurred during benchmark execution: {e}", exc_info=True) |
|
|