| from memories.datasets.base import QuestionAnswerPair, MemoryDataset |
| from inference_utils.operators import ( |
| QuestionAnsweringOperator, |
| LLMExactMatch, |
| ) |
| from memories import DATASET_MAPPING |
| import numpy as np |
| import argparse |
| import json |
| import os |
| from typing import ( |
| List, |
| Dict, |
| Any, |
| Optional, |
| Tuple, |
| ) |
|
|
|
|
| def _build_context_text(retrieved_memories: List[Dict[str, Any]]) -> str: |
| contents = [] |
| for i, mem in enumerate(retrieved_memories): |
| content = mem.get("used_content", '') |
| if not isinstance(content, str): |
| raise AssertionError("The used_content is not a string for the current memory unit.") |
| if not content: |
| raise AssertionError("The used_content is empty for the current memory unit.") |
| contents.append(f"### Memory {i + 1}:\n{content}") |
| return "\n\n".join(contents) |
|
|
|
|
| def answer_questions( |
| retrievals: List[Dict[str, Any]], |
| qa_model: str, |
| qa_batch_size: int = 4, |
| add_question_timestamp: bool = False, |
| dataset_type: Optional[str] = None, |
| interface_kwargs: Optional[Dict[str, Any]] = None, |
| ) -> List[Dict[str, Any]]: |
| interface_kwargs = interface_kwargs or {} |
| questions: List[str] = [] |
| contexts: List[str] = [] |
|
|
| for item in retrievals: |
| qa_pair: QuestionAnswerPair = item["qa_pair"] |
| question_text = qa_pair.question |
| if add_question_timestamp: |
| question_text = f"{question_text}\nQuestion Timestamp: {qa_pair.get_string_timestamp()}" |
| questions.append(question_text) |
| base_ctx = _build_context_text(item["retrieved_memories"]) |
| contexts.append(base_ctx) |
| |
| |
| has_graph = any( |
| mem.get("metadata", {}).get("has_graph_relations", False) |
| for item in retrievals |
| for mem in item["retrieved_memories"] |
| ) |
| if dataset_type is not None: |
| dataset_cls = DATASET_MAPPING[dataset_type] |
| prompt_name = dataset_cls.get_qa_prompt_name(has_graph) |
| else: |
| prompt_name = "question-answering" |
|
|
| print(f"Using prompt: {prompt_name} for question answering.") |
| |
| qa_operator = QuestionAnsweringOperator( |
| prompt_name=prompt_name, |
| model_name=qa_model, |
| **interface_kwargs, |
| ) |
|
|
| responses = qa_operator( |
| questions, |
| contexts, |
| batch_size=qa_batch_size, |
| aggregate=False, |
| temperature=0.0, |
| ) |
| return responses |
|
|
| def evaluate_answers( |
| retrievals: List[Dict[str, Any]], |
| predictions: List[Dict[str, Any]], |
| judge_model: str, |
| judge_batch_size: int = 4, |
| dataset_type: Optional[str] = None, |
| interface_kwargs: Optional[Dict[str, Any]] = None, |
| ) -> List[Dict[str, Any]]: |
| interface_kwargs = interface_kwargs or {} |
| question_list: List[str] = [] |
| golden_answers_list: List[List[str]] = [] |
| prediction_list: List[str] = [] |
| prompt_name_per_index: List[Tuple[str, str]] = [] |
| category_per_index: List[Any] = [] |
| for i, item in enumerate(retrievals): |
| qa_pair: QuestionAnswerPair = item["qa_pair"] |
| question_list.append(qa_pair.question) |
| golden_answers_list.append([ans for ans in qa_pair.answer_list]) |
| pred = predictions[i].get("processed_content") |
| |
| if pred is None: |
| raise ValueError(f"The prediction is None for the question {qa_pair.question}.") |
| prediction_list.append(pred) |
|
|
| category = qa_pair.metadata.get("category", "unknown") |
| category_per_index.append(category) |
|
|
| if dataset_type is not None: |
| dataset_cls = DATASET_MAPPING[dataset_type] |
| prompt_name, qtype = dataset_cls.get_judge_prompt_info(qa_pair) |
| else: |
| qtype = qa_pair.metadata.get("question_type", "normal") |
| prompt_name = "exact-match" |
| |
| prompt_name_per_index.append((prompt_name, qtype)) |
|
|
| print(f"Using prompt: {prompt_name} for judgment.") |
| judge_operator = LLMExactMatch( |
| prompt_name=prompt_name, |
| model_name=judge_model, |
| **interface_kwargs, |
| ) |
|
|
| groups: Dict[Tuple[str, str], List[int]] = {} |
| for idx, p in enumerate(prompt_name_per_index): |
| if p not in groups: |
| groups[p] = [] |
| groups[p].append(idx) |
|
|
| judge_outputs: List[Optional[Dict[str, Any]]] = [None] * len(retrievals) |
| correctness_flags: List[Optional[bool]] = [None] * len(retrievals) |
|
|
| for (prompt_name, qtype), idx_list in groups.items(): |
| judge_operator.set_prompt(prompt_name) |
| batched_questions = [question_list[i] for i in idx_list] |
| batched_golden = [golden_answers_list[i] for i in idx_list] |
| batched_predictions = [prediction_list[i] for i in idx_list] |
| results = judge_operator( |
| batched_questions, |
| batched_golden, |
| batched_predictions, |
| batch_size=judge_batch_size, |
| aggregate=False, |
| temperature=0.0, |
| ) |
| for local_pos, global_idx in enumerate(idx_list): |
| out = results[local_pos] |
| judge_outputs[global_idx] = out |
| content = out.get("processed_content") |
| if content is None: |
| raise ValueError(f"The content is None for the question {batched_questions[local_pos]}.") |
| is_correct = "yes" in content.lower() or "correct" in content.lower() |
| correctness_flags[global_idx] = is_correct |
| |
| accuracy = np.mean( |
| [correctness_flags[global_idx] for global_idx in idx_list] |
| ).item() |
| print(f"The accuracy for {qtype} (prompt name: {prompt_name}) is {accuracy:.4f}.") |
| |
| |
| print("\n" + "="*60) |
| accuracy = np.mean(correctness_flags).item() |
| print(f"Overall Accuracy: {accuracy:.4f}") |
| print("="*60 + "\n") |
|
|
| finalized = [] |
| for i in range(len(retrievals)): |
| finalized.append( |
| { |
| "judge_response": judge_outputs[i], |
| "is_correct": correctness_flags[i], |
| "category": category_per_index[i], |
| } |
| ) |
| return finalized |
|
|
| if __name__ == "__main__": |
| parser = argparse.ArgumentParser( |
| description="A script to evaluate the answers of the search results." |
| ) |
| parser.add_argument( |
| "--search-results-path", |
| type=str, |
| required=True, |
| help="Path to the search results." |
| ) |
| parser.add_argument( |
| "--qa-model", |
| type=str, |
| default="gpt-4o-mini", |
| help="Model name/path for question answering." |
| ) |
| parser.add_argument( |
| "--judge-model", |
| type=str, |
| default="gpt-4o-mini", |
| help="Model name/path for judgment (exact match)." |
| ) |
| parser.add_argument( |
| "--qa-batch-size", |
| type=int, |
| default=4, |
| help="Batch size for QA generation." |
| ) |
| parser.add_argument( |
| "--judge-batch-size", |
| type=int, |
| default=4, |
| help="Batch size for judge model." |
| ) |
| parser.add_argument( |
| "--api-config-path", |
| type=str, |
| default=None, |
| help="Path to the API config file." |
| ) |
| parser.add_argument( |
| "--dataset-type", |
| type=str, |
| required=True, |
| help="The type of dataset (e.g., LoCoMo, LongMemEval)." |
| ) |
| args = parser.parse_args() |
|
|
| |
| interface_kwargs: Dict[str, Any] = {} |
| if args.api_config_path is not None: |
| with open(args.api_config_path, 'r') as f: |
| api_config = json.load(f) |
| interface_kwargs["api_keys"] = api_config["api_keys"] |
| interface_kwargs["base_urls"] = api_config["base_urls"] |
| elif os.environ.get("OPENAI_API_KEY") is not None: |
| interface_kwargs["api_keys"] = [os.environ.get("OPENAI_API_KEY")] |
| interface_kwargs["base_urls"] = [os.environ.get("OPENAI_API_BASE")] |
| |
| with open(args.search_results_path, 'r') as f: |
| retrievals = json.load(f) |
| for item in retrievals: |
| item["qa_pair"] = QuestionAnswerPair(**item["qa_pair"]) |
| print(f"Loaded {len(retrievals)} search results from {args.search_results_path}.") |
|
|
| |
| print("Generating answers with QA model...") |
| qa_responses = answer_questions( |
| retrievals, |
| qa_model=args.qa_model, |
| qa_batch_size=args.qa_batch_size, |
| dataset_type=args.dataset_type, |
| interface_kwargs=interface_kwargs, |
| ) |
|
|
| |
| print("Evaluating answers with judge model...") |
| judge_results = evaluate_answers( |
| retrievals, |
| qa_responses, |
| judge_model=args.judge_model, |
| judge_batch_size=args.judge_batch_size, |
| dataset_type=args.dataset_type, |
| interface_kwargs=interface_kwargs, |
| ) |
|
|
| |
| final_results: List[Dict[str, Any]] = [] |
| for i, item in enumerate(retrievals): |
| qa_pair: QuestionAnswerPair = item["qa_pair"] |
| ans_dict = qa_responses[i] |
| judge_dict = judge_results[i] |
| final_results.append( |
| { |
| "qa_pair": qa_pair.model_dump(), |
| "prediction": ans_dict["processed_content"], |
| "judge_response": judge_dict["judge_response"], |
| "is_correct": judge_dict["is_correct"], |
| "retrieved_memories": item["retrieved_memories"], |
| "user_id": item["user_id"], |
| } |
| ) |
|
|
| output_path = args.search_results_path.rsplit('.', 1)[0] + "_evaluation.json" |
| with open( |
| output_path, |
| 'w', |
| encoding="utf-8" |
| ) as f: |
| json.dump( |
| final_results, |
| f, |
| ensure_ascii=False, |
| indent=4, |
| ) |
| print(f"Saved {len(final_results)} results to {output_path}.") |
|
|