Running 3.6k The Ultra-Scale Playbook 🌌 3.6k The ultimate guide to training LLM on large GPU Clusters
view article Article Fine-tuning LLMs to 1.58bit: extreme quantization made easy +4 Sep 18, 2024 • 272
rhdeng/transport-qa-taiwan-70b-instruct-ds-dgxh100 Text Generation • 71B • Updated Jul 15, 2024 • 8
rhdeng/transport-qa-taiwan-70b-instruct-dgxh100 Text Generation • 71B • Updated Jul 12, 2024 • 7