#pragma once #include #include #include "llm_manager_types.h" std::optional estimate_chat_tokens( const json &payload, const LimitsConfig &limits, std::string &error); bool request_stream_enabled(const json &payload);