Strategic Navigation or Stochastic Search? How Agents and Humans Reason Over Document Collections
Paper
• 2603.12180 • Published
• 47
None defined yet.
hf-mem v0.4.1 now also estimates KV cache memory requirements for any context length and batch size with the --experimental flag!uvx hf-mem --model-id ... --experimental will automatically pull the required information from the Hugging Face Hub to include the KV cache estimation, when applicable.--max-model-len, --batch-size and --kv-cache-dtype arguments (à la vLLM) manually if preferred. kernel-builder 0.7.0: https://github.com/huggingface/kernel-builder/releases/tag/v0.7.0kernelize function:torch.compile kernel, it will use that kernel since it is compatible with inference as well. kernels!kernelize will pick the kernel depending on whether you are going to do training or inference.