| import argparse |
| import json |
| from concurrent.futures import ThreadPoolExecutor, as_completed |
| from datetime import datetime |
| from tqdm import tqdm |
| import time |
| import threading |
| from importlib.util import find_spec |
| from copy import deepcopy |
| import os |
| from contextlib import nullcontext |
|
|
| from token_monitor import ( |
| CostStateManager, |
| token_monitor, |
| CostState, |
| get_tokenizer_for_model, |
| ) |
| from monkey_patch import ( |
| PatchSpec, |
| MonkeyPatcher, |
| make_attr_patch, |
| ) |
|
|
| from memories import ( |
| CONFIG_MAPPING, |
| MEMORY_LAYERS_MAPPING, |
| DATASET_MAPPING, |
| ) |
| from memories.datasets.base import Trajectory |
| from typing import ( |
| Dict, |
| Any, |
| Optional, |
| Tuple, |
| List, |
| Callable, |
| ) |
| from memories.datasets.base import Trajectory, Message, QuestionAnswerPair, Session |
| from importlib import import_module |
|
|
| _LOCK = threading.Lock() |
|
|
| def _check_langchain_core_imports() -> None: |
| """Check if `langchain_core` is installed.""" |
| if find_spec("langchain_core") is None: |
| raise ImportError("`langchain_core` is not installed. Please install it to use this function.") |
|
|
| def _normalize_langmem_messages(*args, **kwargs) -> Dict[str, List[Dict[str, str]] | str | float | int]: |
| """A helper function to process the messages of LangMem.""" |
| _check_langchain_core_imports() |
| from langchain_core.messages import ( |
| HumanMessage, |
| SystemMessage, |
| AIMessage, |
| ToolMessage, |
| ) |
|
|
| messages = kwargs.get("messages", args[0]) |
| assert len(messages) == 1, "Unconsidered Case." |
| messages = messages[0] |
| normalized_messages = [] |
| for message in messages: |
| if isinstance(message, SystemMessage): |
| normalized_messages.append( |
| { |
| "role": "system", |
| "content": message.content |
| } |
| ) |
| elif isinstance(message, HumanMessage): |
| normalized_messages.append( |
| { |
| "role": "user", |
| "content": message.content |
| } |
| ) |
| elif isinstance(message, AIMessage): |
| if message.content is not None and not isinstance(message.content, str): |
| raise ValueError( |
| f"The content of the message is not a string: {type(message.content)}." |
| ) |
| normalized_messages.append( |
| { |
| "role": "assistant", |
| "content": message.content, |
| "tool_calls": [ |
| { |
| "id": tool_call["id"], |
| "type": "function", |
| "function": { |
| "name": tool_call["name"], |
| "arguments": str(tool_call["args"]), |
| } |
| } |
| for tool_call in message.tool_calls |
| ], |
| } |
| ) |
| elif isinstance(message, ToolMessage): |
| |
| normalized_messages.append( |
| { |
| "role": "tool", |
| "tool_call_id": message.tool_call_id, |
| "content": message.content, |
| } |
| ) |
| else: |
| raise ValueError(f"Unsupported message type: {type(message)}.") |
| |
| return normalized_messages |
|
|
| def _extract_langmem_model( |
| llm_model: str, |
| query_model: Optional[str], |
| *args, |
| **kwargs |
| ) -> Tuple[str, Dict[str, Any]]: |
| """A helper function to extract the model name and metadata for LangMem.""" |
| _check_langchain_core_imports() |
| from langchain_core.messages import HumanMessage |
|
|
| llm_model = llm_model.split(':', 1)[1] |
| query_model = query_model.split(':', 1)[1] if query_model is not None else None |
|
|
| messages = kwargs.get("messages", args[0]) |
| assert len(messages) == 1, "Unconsidered Case." |
| messages = messages[0] |
| |
| metadata = { |
| "tools": kwargs.get("tools"), |
| "tool_choice": kwargs.get("tool_choice"), |
| } |
| if isinstance(messages[0], HumanMessage) and messages[0].content.startswith( |
| "Use parallel tool calling to search for distinct memories relevant to this conversation." |
| ): |
| if query_model is None: |
| raise ValueError("Query model is not provided.") |
| return query_model, metadata |
| return llm_model, metadata |
|
|
| def _extract_langmem_output(response) -> Dict[str, List[Dict[str, str]] | str | float | int]: |
| """A helper function to extract the output for LangMem.""" |
| assert len(response.generations) == 1, "Unconsidered Case." |
| assert len(response.generations[0]) == 1, "Unconsidered Case." |
| return { |
| "messages": _normalize_langmem_messages( |
| [[response.generations[0][0].message]] |
| ) |
| } |
| |
|
|
| def memory_construction( |
| layer_type: str, |
| user_id: str, |
| trajectory: Trajectory, |
| config: Optional[Dict[str, Any]] = None, |
| rerun: bool = False, |
| message_preprocessor: Optional[ |
| Callable[[Message | QuestionAnswerPair, Session], Dict[str, Any]] |
| ] = None, |
| **kwargs |
| ) -> Dict[str, float]: |
| """Given a specific interaction trajectory, this function builds a memory.""" |
| config = config or {} |
| llm_model = config["llm_model"] |
| |
| config["user_id"] = user_id |
| |
| config["save_dir"] = f"{layer_type}_{llm_model}/{user_id}" |
| |
| config_cls = CONFIG_MAPPING[layer_type] |
| config = config_cls(**config) |
| dataset_type = kwargs["dataset_type"] |
| with _LOCK: |
| layer_cls = MEMORY_LAYERS_MAPPING[layer_type] |
| layer = layer_cls(config) |
|
|
| output = { |
| "total_add_time": 0.0, |
| "avg_add_time": 0.0, |
| } |
|
|
| with _LOCK: |
| |
| if not rerun and layer.load_memory(user_id): |
| print(f"The memory for user {user_id} is loaded successfully.") |
| return output |
| |
| if layer_type == "A-MEM": |
| |
| |
| |
| getter, setter = make_attr_patch(layer.memory_layer.llm_controller.llm, "get_completion") |
| spec = PatchSpec( |
| name=f"{layer.memory_layer.llm_controller.llm.__class__.__name__}.get_completion", |
| getter=getter, |
| setter=setter, |
| wrapper=token_monitor( |
| extract_model_name=lambda *args, **kwargs: (config.llm_model, {}), |
| extract_input_dict=lambda *args, **kwargs: { |
| "messages": [ |
| { |
| "role": "system", |
| "content": "You must respond with a JSON object." |
| }, |
| { |
| "role": "user", |
| "content": kwargs.get("prompt", args[0]) |
| } |
| ], |
| "metadata": { |
| "op_type": ( |
| "generation" |
| if kwargs.get("prompt", args[0]).startswith("Generate a structured analysis") |
| else "update" |
| ) |
| } |
| }, |
| extract_output_dict=lambda result: { |
| "messages": result |
| }, |
| ), |
| ) |
| specs = [spec] |
| patch_ctx = MonkeyPatcher(specs) |
| elif layer_type == "LangMem": |
| getter, setter = make_attr_patch(layer.llm_model, "generate") |
| spec = PatchSpec( |
| name=f"{layer.llm_model.__class__.__name__}.generate", |
| getter=getter, |
| setter=setter, |
| wrapper=token_monitor( |
| extract_model_name=lambda *args, **kwargs: _extract_langmem_model( |
| config.llm_model, |
| config.query_model, |
| *args, |
| **kwargs |
| ), |
| extract_input_dict=lambda *args, **kwargs: { |
| |
| |
| "messages": _normalize_langmem_messages(*args, **kwargs), |
| "metadata": { |
| "op_type": "generation, update" |
| } |
| }, |
| extract_output_dict=lambda response: _extract_langmem_output(response) |
| ) |
| ) |
| specs = [spec] |
| patch_ctx = MonkeyPatcher(specs) |
| elif layer_type in ("MemZero", "MemZeroGraph"): |
| getter, setter = make_attr_patch(layer.memory_layer.llm, "generate_response") |
| spec = PatchSpec( |
| name=f"{layer.memory_layer.llm.__class__.__name__}.generate_response", |
| getter=getter, |
| setter=setter, |
| wrapper=token_monitor( |
| extract_model_name = lambda *args, **kwargs: (config.llm_model, {}), |
| extract_input_dict = lambda *args, **kwargs: { |
| |
| "messages": kwargs["messages"], |
| "metadata": { |
| "op_type": ( |
| "generation" if kwargs["messages"][0]["content"].startswith("You are a Personal Information Organizer") else "update" |
| ) |
| } |
| }, |
| |
| extract_output_dict = lambda response: {"messages": response.get("content", "")} if isinstance(response, dict) else {"messages": response} |
| ) |
| ) |
| specs = [spec] |
| patch_ctx = MonkeyPatcher(specs) |
| elif layer_type == "FullContext": |
| |
| patch_ctx = nullcontext() |
| elif layer_type == "NaiveRAG": |
| |
| patch_ctx = nullcontext() |
| else: |
| raise ValueError(f"Unsupported memory type: {layer_type}.") |
|
|
| |
| with patch_ctx: |
| total_msgs = sum(len(session) for session in trajectory) |
|
|
| pbar_desc = f" {layer_type} | {user_id}" |
| pbar = tqdm( |
| total=total_msgs, |
| desc=pbar_desc, |
| leave=False, |
| ) |
| |
| for session in trajectory: |
| |
| for message in session: |
| start_time = datetime.now() |
| msg_dict = {"role": message.role, "content": message.content} |
| if message_preprocessor is not None: |
| msg_dict = message_preprocessor(message, session) |
| |
| timestamp = session.get_string_timestamp() |
| layer.add_message( |
| msg_dict, |
| timestamp=timestamp |
| ) |
|
|
| end_time = datetime.now() |
| output["total_add_time"] += (end_time - start_time).total_seconds() |
|
|
| pbar.update(1) |
|
|
| time.sleep(0.2) |
| pbar.close() |
|
|
| |
| if layer_type == "A-MEM": |
| |
| with _LOCK: |
| layer.consolidate_memories() |
| |
| with _LOCK: |
| layer.save_memory() |
|
|
| output["avg_add_time"] = output["total_add_time"] / len(trajectory) |
|
|
| return output |
|
|
| if __name__ == "__main__": |
| parser = argparse.ArgumentParser( |
| description="A script used to evaluate various memory layers on various datasets." |
| ) |
| parser.add_argument( |
| "--memory-type", |
| choices=list(MEMORY_LAYERS_MAPPING.keys()), |
| type=str, |
| required=True, |
| help="The type of the memory layer to be evaluated." |
| ) |
| parser.add_argument( |
| "--dataset-type", |
| choices=list(DATASET_MAPPING.keys()), |
| type=str, |
| required=True, |
| help="The type of the dataset used to evaluate the memory layer." |
| ) |
| parser.add_argument( |
| "--dataset-path", |
| type=str, |
| required=True, |
| help="The path to the dataset." |
| ) |
| parser.add_argument( |
| "--num-workers", |
| type=int, |
| default=4, |
| help="The number of threads to use for the evaluation." |
| ) |
| parser.add_argument( |
| "--seed", |
| type=int, |
| default=42, |
| help="Random seed used to sample the dataset if the user provides the sample size." |
| ) |
| parser.add_argument( |
| "--sample-size", |
| type=int, |
| default=None, |
| help="Subset size from dataset." |
| ) |
| parser.add_argument( |
| "--rerun", |
| action="store_true", |
| help="Ignore saved memory; rebuild from scratch." |
| ) |
| parser.add_argument( |
| "--config-path", |
| type=str, |
| default=None, |
| help="Path to JSON config for memory method." |
| ) |
| parser.add_argument( |
| "--token-cost-save-filename", |
| type=str, |
| default="token_cost", |
| help="Path to save the statistics related to the token consumption." |
| ) |
| parser.add_argument( |
| "--start-idx", |
| type=int, |
| default=None, |
| help="The starting index of the trajectories to be processed." |
| ) |
| parser.add_argument( |
| "--end-idx", |
| type=int, |
| default=None, |
| help="The ending index of the trajectories to be processed." |
| ) |
| parser.add_argument( |
| "--tokenizer-path", |
| type=str, |
| default=None, |
| help="The path to the tokenizer (only for backbone model)." |
| ) |
| parser.add_argument( |
| "--message-preprocessor", |
| type=str, |
| default=None, |
| help=( |
| "Dotted path to a message preprocessor function, in the form " |
| "'some_module:some_function'. The function should accept " |
| "(message, session) and return a dict with at least 'role' " |
| "and 'content' keys." |
| ) |
| ) |
| parser.add_argument( |
| "--use-gpt4o-caption", |
| action="store_true", |
| default=False, |
| help=( |
| "If set, use GPT-4o to generate image captions in datasets that support it " |
| "(e.g., LoCoMo)." |
| ) |
| ) |
| args = parser.parse_args() |
|
|
| message_preprocessor = None |
| if args.message_preprocessor is not None: |
| module_path, func_name = args.message_preprocessor.split(":", 1) |
| module = import_module(module_path) |
| message_preprocessor = getattr(module, func_name) |
| |
| if args.use_gpt4o_caption: |
| os.environ["USE_GPT4O_CAPTION"] = "1" |
| else: |
| os.environ["USE_GPT4O_CAPTION"] = "0" |
|
|
| |
| ds_cls = DATASET_MAPPING[args.dataset_type] |
| dataset = ds_cls.read_raw_data(args.dataset_path) |
| if args.sample_size is not None: |
| dataset = dataset.sample(size=args.sample_size, seed=args.seed) |
| print("The dataset is loaded successfully.") |
| |
| print(repr(dataset)) |
| print() |
|
|
| config = None |
| if args.config_path is not None: |
| with open(args.config_path, 'r', encoding="utf-8") as f: |
| config = json.load(f) |
|
|
| |
| |
| config_cls = CONFIG_MAPPING[args.memory_type] |
| if config is None: |
| dummy_user_id = "guest" |
| dummy_config = config_cls(user_id=dummy_user_id) |
| else: |
| dummy_config = config_cls(**config) |
| |
| |
| if os.path.exists(args.token_cost_save_filename + ".json"): |
| with open(args.token_cost_save_filename + ".json", 'r') as f: |
| token_cost = json.load(f) |
| for model, state in token_cost.items(): |
| is_dict = all(isinstance(value, dict) for value in state.values()) |
| if not is_dict: |
| token_cost[model] = CostState.from_dict(state) |
| else: |
| token_cost[model] = { |
| op: CostState.from_dict(cs) for op, cs in state.items() |
| } |
| else: |
| token_cost = {} |
| |
| |
| |
| |
| llm_model = dummy_config.llm_model |
| if args.memory_type == "LangMem": |
| llm_model = llm_model.split(':', 1)[1] |
| if dummy_config.query_model is not None: |
| query_model = dummy_config.query_model.split(':', 1)[1] |
| if query_model != llm_model: |
| state = token_cost.get(query_model) |
| if state is not None: |
| print( |
| f"There is a saved checkpoint for monitoring the token consumption of {query_model}. " |
| "It will be loaded into `CostStateManager`." |
| ) |
| CostStateManager.register(query_model, state=state) |
| state = token_cost.get(llm_model) |
| if state is not None: |
| print( |
| f"There is a saved checkpoint for monitoring the token consumption of {llm_model}. " |
| "It will be loaded into `CostStateManager`." |
| ) |
| if args.tokenizer_path is not None: |
| tokenizer = get_tokenizer_for_model(args.tokenizer_path) |
| else: |
| tokenizer = None |
| CostStateManager.register(llm_model, state=state, tokenizer=tokenizer) |
| del dummy_config |
| print(f"The LLM model being used is {llm_model}. It has been registered in `CostStateManager`.") |
| print() |
|
|
| if args.start_idx is None: |
| args.start_idx = 0 |
| if args.end_idx is None: |
| args.end_idx = len(dataset) |
| args.start_idx, args.end_idx = max(0, args.start_idx), min(args.end_idx, len(dataset)) |
| if args.start_idx >= args.end_idx: |
| raise ValueError("The starting index must be less than the ending index.") |
|
|
| results = [] |
| with ThreadPoolExecutor(max_workers=args.num_workers) as executor: |
| futures = [] |
| for trajectory, _ in zip(*dataset[args.start_idx: args.end_idx]): |
| |
| user_id = f"user_{dataset.__class__.__name__}_{trajectory.metadata['id']}" |
| future = executor.submit( |
| memory_construction, |
| args.memory_type, |
| user_id, |
| trajectory, |
| config=deepcopy(config), |
| rerun=args.rerun, |
| message_preprocessor=message_preprocessor, |
| dataset_type=args.dataset_type |
| ) |
| futures.append(future) |
| for future in tqdm( |
| as_completed(futures), total=len(futures), desc="Processing trajectories" |
| ): |
| try: |
| result = future.result() |
| results.append(result) |
| except Exception as e: |
| print(f"Error processing trajectory: {e}") |
|
|
| if len(results) == args.end_idx - args.start_idx: |
| print("The memory construction process is completed successfully.") |
|
|
| total_time = 0.0 |
| avg_time_per_add_session = 0.0 |
| num_vaild_trajectories = 0 |
| for result in results: |
| |
| if result["total_add_time"] > 0: |
| total_time += result["total_add_time"] |
| avg_time_per_add_session += result["avg_add_time"] |
| num_vaild_trajectories += 1 |
| avg_time = total_time / max(num_vaild_trajectories, 1) |
| avg_time_per_add_session = avg_time_per_add_session / max(num_vaild_trajectories, 1) |
| print( |
| f"For {args.memory_type}, the average time per trajectory " |
| f"({num_vaild_trajectories} in {len(results)}) is {avg_time:.2f} seconds." |
| ) |
| print( |
| f"For {args.memory_type}, the average time per operation of adding new session " |
| f"is {avg_time_per_add_session:.2f} seconds." |
| ) |
|
|
| |
| CostStateManager.save_to_json_file(args.token_cost_save_filename) |
| CostStateManager.reset() |
|
|