mlx-community/DeepSeek-R1-Distill-Qwen-32B-abliterated-4bit Text Generation • 5B • Updated Feb 20, 2025 • 586 • 5
mlx-community/DeepSeek-R1-Distill-Qwen-32B-abliterated Text Generation • 33B • Updated Feb 20, 2025 • 248 • • 3
mlx-community/DeepSeek-R1-Distill-Qwen-32B-4bit Text Generation • 5B • Updated Feb 26, 2025 • 3.12k • 46
Running 3.74k The Ultra-Scale Playbook 🌌 3.74k The ultimate guide to training LLM on large GPU Clusters