Abhinav Agarwalla, Abhay Gupta, Alexandre Marques, Shubhra Pandit, Michael Goin...
TL;DR通过稀疏性,我们能够以较小的模型实现更快的训练和推理加速,并且不牺牲准确性。
Abstract
large language models (LLMs) have revolutionized Natural Language Processing (NLP), but their size creates computational bottlenecks. We introduce a novel approach to create accurate, sparse foundational versions