#!/usr/bin/env python3 """ Evaluation script for the fine-tuned Qwen2.5-3B evaluation agent model. This script evaluates the trained model on various tasks including: - VBench evaluation (text-to-video generation quality assessment) - T2I-CompBench evaluation (text-to-image generation quality assessment) - Open-ended evaluation queries The model uses CoT-like reasoning format for quality assessment. """ import json import argparse import requests import time from typing import Dict, List, Any, Optional from pathlib import Path import pandas as pd from tqdm import tqdm class EvalAgentTester: """Tester for the fine-tuned evaluation agent model.""" def __init__(self, model_url: str = "http://0.0.0.0:12333/v1/chat/completions", model_name: str = "eval-agent"): """ Initialize the evaluation tester. Args: model_url: URL of the model server (launched via vLLM) model_name: Name of the served model """ self.model_url = model_url self.model_name = model_name self.test_data = {} def load_test_data(self, data_path: str = "data/postprocess_20250819/ea_cot_dataset_10k.json"): """Load test dataset.""" try: with open(data_path, 'r', encoding='utf-8') as f: self.test_data = json.load(f) print(f"Loaded {len(self.test_data)} test samples from {data_path}") except FileNotFoundError: print(f"Test data file not found: {data_path}") print("Please run the data preprocessing script first.") return False except json.JSONDecodeError as e: print(f"Error parsing JSON file: {e}") return False return True def call_model(self, instruction: str, input_text: str = "", system: str = "", history: List = None, max_tokens: int = 2048, temperature: float = 0.7) -> Optional[str]: """ Call the fine-tuned model via API. Args: instruction: Main instruction/question input_text: Additional input context system: System prompt history: Conversation history max_tokens: Maximum response tokens temperature: Sampling temperature Returns: Model response or None if error """ # Construct message based on alpaca format messages = [] if system: messages.append({"role": "system", "content": system}) # Add history if provided if history: for human, assistant in history: messages.append({"role": "user", "content": human}) messages.append({"role": "assistant", "content": assistant}) # Construct user message user_content = instruction if input_text: user_content = f"{instruction}\n\n{input_text}" messages.append({"role": "user", "content": user_content}) payload = { "model": self.model_name, "messages": messages, "max_tokens": max_tokens, "temperature": temperature, "stream": False } try: response = requests.post(self.model_url, json=payload, timeout=60) response.raise_for_status() result = response.json() return result["choices"][0]["message"]["content"] except requests.exceptions.RequestException as e: print(f"API request failed: {e}") return None except (KeyError, IndexError) as e: print(f"Unexpected response format: {e}") return None def evaluate_sample(self, sample: Dict[str, Any]) -> Dict[str, Any]: """ Evaluate a single sample. Args: sample: Test sample with instruction, input, output, system, history Returns: Evaluation result with prediction and metadata """ instruction = sample.get("instruction", "") input_text = sample.get("input", "") expected_output = sample.get("output", "") system = sample.get("system", "") history = sample.get("history", []) # Get model prediction prediction = self.call_model( instruction=instruction, input_text=input_text, system=system, history=history ) result = { "instruction": instruction, "input": input_text, "expected_output": expected_output, "prediction": prediction, "system": system, "history": history, "success": prediction is not None } return result def run_evaluation(self, num_samples: int = 100, save_results: bool = True, output_path: str = "evaluation_results.json") -> Dict[str, Any]: """ Run evaluation on a subset of test data. Args: num_samples: Number of samples to evaluate save_results: Whether to save results to file output_path: Path to save results Returns: Evaluation summary and results """ if not self.test_data: print("No test data loaded. Please call load_test_data() first.") return {} # Select samples to evaluate test_samples = list(self.test_data)[:num_samples] if isinstance(self.test_data, list) else list(self.test_data.values())[:num_samples] print(f"Evaluating {len(test_samples)} samples...") results = [] successful_calls = 0 failed_calls = 0 for i, sample in enumerate(tqdm(test_samples, desc="Evaluating")): result = self.evaluate_sample(sample) results.append(result) if result["success"]: successful_calls += 1 else: failed_calls += 1 # Add delay to avoid overwhelming the server time.sleep(0.1) # Compile summary summary = { "total_samples": len(test_samples), "successful_calls": successful_calls, "failed_calls": failed_calls, "success_rate": successful_calls / len(test_samples) if test_samples else 0, "results": results } # Save results if requested if save_results: with open(output_path, 'w', encoding='utf-8') as f: json.dump(summary, f, indent=2, ensure_ascii=False) print(f"Results saved to {output_path}") return summary def analyze_results(self, results: Dict[str, Any]) -> None: """ Analyze and print evaluation results. Args: results: Results from run_evaluation() """ print("\n" + "="*50) print("EVALUATION SUMMARY") print("="*50) print(f"Total samples evaluated: {results['total_samples']}") print(f"Successful API calls: {results['successful_calls']}") print(f"Failed API calls: {results['failed_calls']}") print(f"Success rate: {results['success_rate']:.2%}") if results['results']: print("\n" + "-"*50) print("SAMPLE RESULTS") print("-"*50) # Show first few successful predictions successful_results = [r for r in results['results'] if r['success']] for i, result in enumerate(successful_results[:3]): print(f"\nSample {i+1}:") print(f"Instruction: {result['instruction'][:100]}...") print(f"Input: {result['input'][:50]}..." if result['input'] else "Input: (empty)") print(f"Expected: {result['expected_output'][:100]}...") print(f"Predicted: {result['prediction'][:100]}..." if result['prediction'] else "Predicted: (failed)") print("\n" + "="*50) def test_specific_tasks(self) -> None: """Test the model on specific evaluation tasks.""" print("\n" + "="*50) print("TESTING SPECIFIC EVALUATION TASKS") print("="*50) # Test 1: VBench-style video evaluation print("\n1. Testing VBench-style video evaluation:") vbench_instruction = "How accurately does the model generate specific object classes as described in the text prompt?" vbench_system = """ You are an expert in evaluating video generation models. Your task is to dynamically explore the model's capabilities step by step, simulating the process of human exploration. Dynamic evaluation refers to initially providing a preliminary focus based on the user's question, and then continuously adjusting what aspects to focus on according to the intermediate evaluation results. Please provide your analysis using the following format: Sub-aspect: The specific aspect you want to focus on. Tool: The evaluation tool you choose to use. Thought: Detailed explanation of your reasoning. """ response = self.call_model( instruction=vbench_instruction, system=vbench_system ) print(f"Response: {response[:500]}..." if response else "Failed to get response") # Test 2: T2I-CompBench-style image evaluation print("\n2. Testing T2I-CompBench-style image evaluation:") t2i_instruction = "How well does the model handle color accuracy in generated images?" t2i_system = """ You are an expert evaluator for text-to-image generation models. Evaluate the model's performance on color accuracy. Provide your assessment with reasoning and specific examples. """ response = self.call_model( instruction=t2i_instruction, system=t2i_system ) print(f"Response: {response[:500]}..." if response else "Failed to get response") # Test 3: Open-ended evaluation print("\n3. Testing open-ended evaluation:") open_instruction = "What are the key strengths and weaknesses of this image generation model?" response = self.call_model( instruction=open_instruction ) print(f"Response: {response[:500]}..." if response else "Failed to get response") print("\n" + "="*50) def main(): """Main evaluation function.""" parser = argparse.ArgumentParser(description="Evaluate the fine-tuned Qwen2.5-3B evaluation agent") parser.add_argument("--model_url", default="http://0.0.0.0:12333/v1/chat/completions", help="URL of the model server") parser.add_argument("--model_name", default="eval-agent", help="Name of the served model") parser.add_argument("--data_path", default="data/postprocess_20250819/ea_cot_dataset_10k.json", help="Path to test dataset") parser.add_argument("--num_samples", type=int, default=50, help="Number of samples to evaluate") parser.add_argument("--output_path", default="evaluation_results.json", help="Path to save evaluation results") parser.add_argument("--test_tasks", action="store_true", help="Run specific task tests instead of full evaluation") args = parser.parse_args() # Initialize tester print("Initializing Evaluation Agent Tester...") tester = EvalAgentTester(model_url=args.model_url, model_name=args.model_name) if args.test_tasks: # Run specific task tests tester.test_specific_tasks() else: # Load test data print(f"Loading test data from {args.data_path}...") if not tester.load_test_data(args.data_path): print("Failed to load test data. Exiting.") return # Run evaluation print("Starting evaluation...") results = tester.run_evaluation( num_samples=args.num_samples, save_results=True, output_path=args.output_path ) # Analyze results tester.analyze_results(results) print(f"\nEvaluation complete! Results saved to {args.output_path}") if __name__ == "__main__": main()