Sparse Finetuning for Inference Acceleration of Large Language Models
Paper
•
2310.06927
•
Published
•
14
Explore our breakthrough in sparse fine-tuning LLMs! Our novel method maintains downstream accuracy even with >70% sparsity.
Solve math problems with chat-based guidance