Running Featured 65 QED-Nano: Teaching a Tiny Model to Prove Hard Theorems π 65 Who needs 1T parameters? Olympiad proofs with a 4B model
Running on CPU Upgrade Featured 3.04k The Smol Training Playbook π 3.04k The secrets to building world-class LLMs
SmolVLM Collection State-of-the-art compact VLMs for on-device applications: Base, Synthetic, and Instruct. Check our blog: https://huggingface.co/blog/smolvlm β’ 5 items β’ Updated May 5, 2025 β’ 42
Running on CPU Upgrade Featured 1.25k Open ASR Leaderboard π 1.25k Explore ASR model performance across languages and datasets
view article Article Prefill and Decode for Concurrent Requests - Optimizing LLM Performance Apr 16, 2025 β’ 66
view article Article The Transformers Library: standardizing model definitions +2 May 15, 2025 β’ 121