byteshape/Devstral-Small-2-24B-Instruct-2512-GGUF Text Generation • 24B • Updated 26 days ago • 11.6k • 19
view post Post 5192 We collaborated with Hugging Face to enable you to train MoE models 12× faster with 35% less VRAM via our new Triton kernels (no accuracy loss). 🤗Train gpt-oss locally on 12.8GB VRAM with our free notebooks: https://unsloth.ai/docs/new/faster-moe See translation 1 reply · 🔥 29 29 🤗 5 5 + Reply
nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16 Text Generation • 32B • Updated about 17 hours ago • 925k • 669
noctrex/Nemotron-3-Nano-30B-A3B-MXFP4_MOE-GGUF Text Generation • 32B • Updated Dec 21, 2025 • 3.32k • 17