Sparse Finetuning for Inference Acceleration of Large Language Models
Paper • 2310.06927 • Published • 15
Explore our breakthrough in sparse fine-tuning LLMs! Our novel method maintains downstream accuracy even with >70% sparsity.
Solve math problems with chat-based guidance