view post Post 5121 We collaborated with Hugging Face to enable you to train MoE models 12ร faster with 35% less VRAM via our new Triton kernels (no accuracy loss). ๐คTrain gpt-oss locally on 12.8GB VRAM with our free notebooks: https://unsloth.ai/docs/new/faster-moe See translation 1 reply ยท ๐ฅ 29 29 ๐ค 5 5 + Reply
PaddleOCR-VL: Boosting Multilingual Document Parsing via a 0.9B Ultra-Compact Vision-Language Model Paper โข 2510.14528 โข Published Oct 16, 2025 โข 118
Conditional Memory via Scalable Lookup: A New Axis of Sparsity for Large Language Models Paper โข 2601.07372 โข Published Jan 12 โข 41
Scaling Text-to-Image Diffusion Transformers with Representation Autoencoders Paper โข 2601.16208 โข Published about 1 month ago โข 52
Efficient Memory Management for Large Language Model Serving with PagedAttention Paper โข 2309.06180 โข Published Sep 12, 2023 โข 37
Running on Zero MCP Featured 538 FLUX.2 [Klein] 9B ๐ป 538 Generate or edit images from text prompts with optional input images
MobileLLM Collection Optimizing Sub-billion Parameter Language Models for On-Device Use Cases (ICML 2024) https://arxiv.org/abs/2402.14905 โข 50 items โข Updated Dec 11, 2025 โข 137
view post Post 3881 Mistral's new Ministral 3 models can now be Run & Fine-tuned locally! (16GB RAM)Ministral 3 have vision support and the best-in-class performance for their sizes.14B Instruct GGUF: unsloth/Ministral-3-14B-Instruct-2512-GGUF14B Reasoning GGUF: unsloth/Ministral-3-14B-Reasoning-2512-GGUF๐ฑ Step-by-step Guide: https://docs.unsloth.ai/new/ministral-3All GGUFs, BnB, FP8 etc. variants uploads: https://huggingface.co/collections/unsloth/ministral-3 See translation 3 replies ยท ๐ฅ 17 17 ๐ค 7 7 โค๏ธ 5 5 ๐ 3 3 + Reply
view article Article Releasing Common Corpus: the largest public domain dataset for training LLMs Mar 20, 2024 โข 32
FilmAgent: A Multi-Agent Framework for End-to-End Film Automation in Virtual 3D Spaces Paper โข 2501.12909 โข Published Jan 22, 2025 โข 74