| import json |
| import asyncio |
| import aiofiles |
| from tqdm import tqdm |
| import os |
| import argparse |
| from openai import AsyncOpenAI |
| from math_verify import parse |
| from evaluate import load |
|
|
| math = load("competition_math") |
|
|
|
|
| async def generate_single_answer(client, question: str, model_name: str) -> str: |
| """Generate a single answer for a question using the language model.""" |
| |
| problem = question |
| |
| prompt = f""" |
| The following is a math problem: |
| |
| [Math Problem] |
| |
| {problem} |
| |
| Your task is to solve it step by step. |
| |
| In your thinking, when you think you just get the result of one complete solution, |
| just stop (do not neec to self-check or rethink). |
| """ |
| |
| |
| try: |
| response = await client.chat.completions.create( |
| model=model_name, |
| messages=[ |
| {"role": "user", "content": prompt} |
| ], |
| max_tokens=8192, |
| temperature=0.6, |
| top_p=0.95, |
| n=1 |
| ) |
| return response.choices[0].message.content.strip() |
| except Exception as e: |
| print(f"Error in generate_single_answer: {str(e)}") |
| return None |
|
|
|
|
| async def evaluate_single_problem( |
| prob: dict, |
| client: AsyncOpenAI, |
| model_name: str, |
| sem: asyncio.Semaphore |
| ) -> dict: |
| async with sem: |
| try: |
| print("Evaluating problem: {}".format(prob["question"])) |
| |
| |
| answer = await generate_single_answer(client, prob["question"], model_name) |
| if answer is None: |
| return None |
| |
| |
| extracted_ans = parse(answer) |
| pass_at_1 = 1 if math.compute(references=[prob["expected_answer"]], predictions=[extracted_ans])["accuracy"] > 0.99 else 0 |
| |
| print("------------------------------------------------------------") |
| print("Question:", prob["question"]) |
| print("Expected answer:", prob["expected_answer"]) |
| print("Generated answer:", answer) |
| print("Pass@1:", pass_at_1) |
| |
| result = { |
| "question": prob["question"], |
| "expected_answer": prob["expected_answer"], |
| "generated_answer": answer, |
| "pass@1": pass_at_1 |
| } |
| return result |
| except Exception as e: |
| print(f"Error in evaluate_single_problem: {str(e)}") |
| return None |
|
|
|
|
| async def save_results_async(output_file: str, data: dict): |
| async with aiofiles.open(output_file, 'a') as f: |
| await f.write(json.dumps(data) + '\n') |
|
|
|
|
| async def main(debug: bool = False, resume: bool = False): |
| |
| client = AsyncOpenAI( |
| base_url="http://localhost:8014/v1", |
| api_key="token-abc123" |
| ) |
| |
| model_name = "DeepSeek-R1-Distill-Qwen-14B" |
| |
| |
| problems = [] |
| with open('./test500.jsonl', 'r') as f: |
| for line in f: |
| problem = json.loads(line) |
| problems.append({ |
| 'question': problem['problem'], |
| 'expected_answer': problem['answer'] |
| }) |
| |
| |
| if debug: |
| problems = problems[:50] |
| print("DEBUG MODE: processing only the first 50 problems.") |
|
|
| |
| output_file = "pass_at_1_simple_results.jsonl" |
| if resume: |
| if os.path.exists(output_file): |
| |
| dedup = {} |
| with open(output_file, 'r') as res_file: |
| for line in res_file: |
| if line.strip(): |
| try: |
| rec = json.loads(line) |
| question = rec.get("question") |
| if question is not None: |
| dedup[question] = rec |
| except Exception as e: |
| continue |
|
|
| |
| with open(output_file, 'w') as res_file: |
| for rec in dedup.values(): |
| res_file.write(json.dumps(rec) + "\n") |
|
|
| evaluated_questions = set(dedup.keys()) |
| original_count = len(problems) |
| problems = [p for p in problems if p["question"] not in evaluated_questions] |
| skipped = original_count - len(problems) |
| print(f"Resuming evaluation: Skipping {skipped} already evaluated problems.") |
| else: |
| print("No previous evaluation results found. Starting from scratch.") |
|
|
| |
| sem = asyncio.Semaphore(30) |
| |
| |
| tasks = [ |
| asyncio.create_task(evaluate_single_problem(prob, client, model_name, sem)) |
| for prob in problems |
| ] |
| |
| results = [] |
| |
| for future in tqdm(asyncio.as_completed(tasks), total=len(tasks), desc='Processing problems'): |
| result = await future |
| if result is not None: |
| results.append(result) |
| |
| await save_results_async(output_file, result) |
|
|
| if results: |
| total_pass_at_1 = sum(result["pass@1"] for result in results) |
| pass_at_1_rate = total_pass_at_1 / len(results) * 100 |
| print(f"\nFinal Pass@1 Rate: {pass_at_1_rate:.2f}%") |
|
|
| print(f"Evaluation complete. Processed {len(results)} problems successfully.") |
| print(f"Results saved to {output_file}") |
|
|
|
|
| if __name__ == "__main__": |
| parser = argparse.ArgumentParser() |
| parser.add_argument("--debug", action="store_true", help="Run in debug mode (only evaluate the first 50 problems)") |
| parser.add_argument("--resume", action="store_true", help="Resume evaluation by skipping already evaluated problems") |
| args = parser.parse_args() |
| asyncio.run(main(debug=args.debug, resume=args.resume)) |