BriefGPT.xyz
Nov, 2023
稀疏低秩的预训练语言模型适应
Sparse Low-rank Adaptation of Pre-trained Language Models
HTML
PDF
Ning Ding, Xingtai Lv, Qiaosen Wang, Yulin Chen, Bowen Zhou...
TL;DR
在对大规模预训练语言模型进行提升调优的过程中,我们通过引入稀疏低秩适应性的创新方法(SoRA),使得适应过程中能够动态地调整内在秩,从而提高LoRA的表现能力,同时通过更新稀疏方式高效地控制参数数量。实验结果表明,SoRA在保留70%参数和训练时间的情况下,能够胜过其他基准模型。
Abstract
fine-tuning
pre-trained large language models
in a
parameter-efficient
manner is widely studied for its effectiveness and efficiency. The
→