Planning in 8 Tokens: A Compact Discrete Tokenizer for Latent World Model Paper • 2603.05438 • Published 12 days ago • 38
Evaluation of a Robust Control System in Real-World Cable-Driven Parallel Robots Paper • 2510.08270 • Published Oct 9, 2025 • 2
view post Post 3343 MiniMax M2.5 is now available on the hub 🚀 MiniMaxAI/MiniMax-M2.5✨ 229B - Modified MIT license✨37% faster than M2.1✨ ~$1/hour at 100 TPS See translation 2 replies · 🚀 8 8 👀 1 1 + Reply
view post Post 664 RynnBrain 🤖 a physics aware embodied brain for robots from Alibaba DAMO https://huggingface.co/collections/Alibaba-DAMO-Academy/rynnbrain✨ 2B/8B/30B (3B active) ✨ Apache 2.0✨ Understands egocentric scenes with strong spatial awareness✨ Tracks objects and motion over time See translation 2 replies · 🔥 2 2 + Reply
view post Post 3759 Game on 🎮🚀While Seedance 2.0’s videos are all over the timeline, DeepSeek quietly pushed a new model update in its app.GLM-5 from Z.ai adds more momentum.Ming-flash-omni from Ant Group , MiniCPM-SALA from OpenBMB , and the upcoming MiniMax M2.5 keep the heat on 🔥 Spring Festival is around the corner, no one’s sleeping! ✨ More releases coming, stay tunedhttps://huggingface.co/collections/zh-ai-community/2026-february-china-open-source-highlights See translation 🔥 5 5 🤯 1 1 + Reply
view post Post 3902 Ming-flash-omni 2.0 🚀 New open omni-MLLM released by Ant Group inclusionAI/Ming-flash-omni-2.0✨ MIT license✨ MoE - 100B/6B active✨ Zero-shot voice cloning + controllable audio✨ Fine-grained visual knowledge grounding See translation 2 replies · 🔥 9 9 + Reply
view post Post 758 LLaDA 2.1 is out 🔥 A new series of MoE diffusion language model released by AntGroup inclusionAI/LLaDA2.1-mini inclusionAI/LLaDA2.1-flash✨LLaDA2.1-mini: 16B - Apache2.0✨LLaDA2.1-flash: 100B - Apache2.0 ✨Both delivers editable generation, RL-trained diffusion reasoning and fast inference See translation 2 replies · 🔥 2 2 + Reply
view post Post 2589 AI for science is moving fast🚀Intern-S1-Pro 🔬 a MoE multimodal scientific reasoning model from Shanghai AI Lab internlm/Intern-S1-Pro✨ 1T total / 22B active ✨ Apache 2.0✨ SoTA scientific reasoning performance ✨ FoPE enables scalable modeling of long physical time series (10⁰–10⁶) See translation 2 replies · 🔥 6 6 😎 2 2 👍 2 2 + Reply
Organize the Web: Constructing Domains Enhances Pre-Training Data Curation Paper • 2502.10341 • Published Feb 14, 2025 • 3
olmOCR: Unlocking Trillions of Tokens in PDFs with Vision Language Models Paper • 2502.18443 • Published Feb 25, 2025 • 9
DataDecide: How to Predict Best Pretraining Data with Small Experiments Paper • 2504.11393 • Published Apr 15, 2025 • 18
Teaching Models to Understand (but not Generate) High-risk Data Paper • 2505.03052 • Published May 5, 2025 • 6
The Common Pile v0.1: An 8TB Dataset of Public Domain and Openly Licensed Text Paper • 2506.05209 • Published Jun 5, 2025 • 60
DR Tulu: Reinforcement Learning with Evolving Rubrics for Deep Research Paper • 2511.19399 • Published Nov 24, 2025 • 63
Bolmo: Byteifying the Next Generation of Language Models Paper • 2512.15586 • Published Dec 17, 2025 • 17