Cerebras REAP Collection Sparse MoE models compressed using REAP (Router-weighted Expert Activation Pruning) method • 24 items • Updated 3 days ago • 92
VTP Collection Towards Scalable Pre-training of Visual Tokenizers for Generation • 4 items • Updated Dec 16, 2025 • 40
Teacher Logits Collection Logits captured from large models to act as the teacher for distillation • 3 items • Updated Dec 15, 2025 • 7
Ministral 3 Collection Mistral Ministral 3: new multimodal models in Base, Instruct, and Reasoning variants, available in 3B, 8B, and 14B sizes. • 36 items • Updated 4 days ago • 28
Ministral 3 Collection A collection of edge models, with Base, Instruct and Reasoning variants, in 3 different sizes: 3B, 8B and 14B. All with vision capabilities. • 9 items • Updated Dec 2, 2025 • 149
Trinity Collection Collection of Arcee AI models in the Trinity family • 8 items • Updated Dec 11, 2025 • 21
Olmo 3 Pre-training Collection All artifacts related to Olmo 3 pre-training • 10 items • Updated Dec 23, 2025 • 32
BERT Hash Nano Models Collection Set of BERT models with a modified embeddings layer • 4 items • Updated Dec 22, 2025 • 9
TOUCAN: Synthesizing 1.5M Tool-Agentic Data from Real-World MCP Environments Paper • 2510.01179 • Published Oct 1, 2025 • 26
💧 LFM2 Collection LFM2 is a new generation of hybrid models, designed for on-device deployment. • 27 items • Updated 14 days ago • 136
view article Article Welcome EmbeddingGemma, Google's new efficient embedding model +4 Sep 4, 2025 • 271