|
|
from __future__ import annotations |
|
|
|
|
|
import logging |
|
|
import threading |
|
|
from abc import ABC, abstractmethod |
|
|
from collections import defaultdict |
|
|
from copy import copy |
|
|
from dataclasses import asdict, dataclass |
|
|
from enum import Enum |
|
|
|
|
|
from anthropic.types import MessageParam, MessageTokensCount |
|
|
from dotenv import load_dotenv |
|
|
|
|
|
log = logging.getLogger(__name__) |
|
|
|
|
|
|
|
|
class TokenCountEstimator(ABC): |
|
|
@abstractmethod |
|
|
def estimate_token_count(self, text: str) -> int: |
|
|
""" |
|
|
Estimate the number of tokens in the given text. |
|
|
This is an abstract method that should be implemented by subclasses. |
|
|
""" |
|
|
|
|
|
|
|
|
class TiktokenCountEstimator(TokenCountEstimator): |
|
|
""" |
|
|
Approximate token count using tiktoken. |
|
|
""" |
|
|
|
|
|
def __init__(self, model_name: str = "gpt-4o"): |
|
|
""" |
|
|
The tokenizer will be downloaded on the first initialization, which may take some time. |
|
|
|
|
|
:param model_name: see `tiktoken.model` to see available models. |
|
|
""" |
|
|
import tiktoken |
|
|
|
|
|
log.info(f"Loading tiktoken encoding for model {model_name}, this may take a while on the first run.") |
|
|
self._encoding = tiktoken.encoding_for_model(model_name) |
|
|
|
|
|
def estimate_token_count(self, text: str) -> int: |
|
|
return len(self._encoding.encode(text)) |
|
|
|
|
|
|
|
|
class AnthropicTokenCount(TokenCountEstimator): |
|
|
""" |
|
|
The exact count using the Anthropic API. |
|
|
Counting is free, but has a rate limit and will require an API key, |
|
|
(typically, set through an env variable). |
|
|
See https://docs.anthropic.com/en/docs/build-with-claude/token-counting |
|
|
""" |
|
|
|
|
|
def __init__(self, model_name: str = "claude-sonnet-4-20250514", api_key: str | None = None): |
|
|
import anthropic |
|
|
|
|
|
self._model_name = model_name |
|
|
if api_key is None: |
|
|
load_dotenv() |
|
|
self._anthropic_client = anthropic.Anthropic(api_key=api_key) |
|
|
|
|
|
def _send_count_tokens_request(self, text: str) -> MessageTokensCount: |
|
|
return self._anthropic_client.messages.count_tokens( |
|
|
model=self._model_name, |
|
|
messages=[MessageParam(role="user", content=text)], |
|
|
) |
|
|
|
|
|
def estimate_token_count(self, text: str) -> int: |
|
|
return self._send_count_tokens_request(text).input_tokens |
|
|
|
|
|
|
|
|
_registered_token_estimator_instances_cache: dict[RegisteredTokenCountEstimator, TokenCountEstimator] = {} |
|
|
|
|
|
|
|
|
class RegisteredTokenCountEstimator(Enum): |
|
|
TIKTOKEN_GPT4O = "TIKTOKEN_GPT4O" |
|
|
ANTHROPIC_CLAUDE_SONNET_4 = "ANTHROPIC_CLAUDE_SONNET_4" |
|
|
|
|
|
@classmethod |
|
|
def get_valid_names(cls) -> list[str]: |
|
|
""" |
|
|
Get a list of all registered token count estimator names. |
|
|
""" |
|
|
return [estimator.name for estimator in cls] |
|
|
|
|
|
def _create_estimator(self) -> TokenCountEstimator: |
|
|
match self: |
|
|
case RegisteredTokenCountEstimator.TIKTOKEN_GPT4O: |
|
|
return TiktokenCountEstimator(model_name="gpt-4o") |
|
|
case RegisteredTokenCountEstimator.ANTHROPIC_CLAUDE_SONNET_4: |
|
|
return AnthropicTokenCount(model_name="claude-sonnet-4-20250514") |
|
|
case _: |
|
|
raise ValueError(f"Unknown token count estimator: {self.value}") |
|
|
|
|
|
def load_estimator(self) -> TokenCountEstimator: |
|
|
estimator_instance = _registered_token_estimator_instances_cache.get(self) |
|
|
if estimator_instance is None: |
|
|
estimator_instance = self._create_estimator() |
|
|
_registered_token_estimator_instances_cache[self] = estimator_instance |
|
|
return estimator_instance |
|
|
|
|
|
|
|
|
class ToolUsageStats: |
|
|
""" |
|
|
A class to record and manage tool usage statistics. |
|
|
""" |
|
|
|
|
|
def __init__(self, token_count_estimator: RegisteredTokenCountEstimator = RegisteredTokenCountEstimator.TIKTOKEN_GPT4O): |
|
|
self._token_count_estimator = token_count_estimator.load_estimator() |
|
|
self._token_estimator_name = token_count_estimator.value |
|
|
self._tool_stats: dict[str, ToolUsageStats.Entry] = defaultdict(ToolUsageStats.Entry) |
|
|
self._tool_stats_lock = threading.Lock() |
|
|
|
|
|
@property |
|
|
def token_estimator_name(self) -> str: |
|
|
""" |
|
|
Get the name of the registered token count estimator used. |
|
|
""" |
|
|
return self._token_estimator_name |
|
|
|
|
|
@dataclass(kw_only=True) |
|
|
class Entry: |
|
|
num_times_called: int = 0 |
|
|
input_tokens: int = 0 |
|
|
output_tokens: int = 0 |
|
|
|
|
|
def update_on_call(self, input_tokens: int, output_tokens: int) -> None: |
|
|
""" |
|
|
Update the entry with the number of tokens used for a single call. |
|
|
""" |
|
|
self.num_times_called += 1 |
|
|
self.input_tokens += input_tokens |
|
|
self.output_tokens += output_tokens |
|
|
|
|
|
def _estimate_token_count(self, text: str) -> int: |
|
|
return self._token_count_estimator.estimate_token_count(text) |
|
|
|
|
|
def get_stats(self, tool_name: str) -> ToolUsageStats.Entry: |
|
|
""" |
|
|
Get (a copy of) the current usage statistics for a specific tool. |
|
|
""" |
|
|
with self._tool_stats_lock: |
|
|
return copy(self._tool_stats[tool_name]) |
|
|
|
|
|
def record_tool_usage(self, tool_name: str, input_str: str, output_str: str) -> None: |
|
|
input_tokens = self._estimate_token_count(input_str) |
|
|
output_tokens = self._estimate_token_count(output_str) |
|
|
with self._tool_stats_lock: |
|
|
entry = self._tool_stats[tool_name] |
|
|
entry.update_on_call(input_tokens, output_tokens) |
|
|
|
|
|
def get_tool_stats_dict(self) -> dict[str, dict[str, int]]: |
|
|
with self._tool_stats_lock: |
|
|
return {name: asdict(entry) for name, entry in self._tool_stats.items()} |
|
|
|
|
|
def clear(self) -> None: |
|
|
with self._tool_stats_lock: |
|
|
self._tool_stats.clear() |
|
|
|