BriefGPT.xyz
Feb, 2024
PeriodicLoRA: 打破 LoRA 优化中的低秩瓶颈
PeriodicLoRA: Breaking the Low-Rank Bottleneck in LoRA Optimization
HTML
PDF
Xiangdi Meng, Damai Dai, Weiyao Luo, Zhe Yang, Shaoxiang Wu...
TL;DR
本研究探讨了一种改进的LoRA优化方法,称为PeriodicLoRA(PLoRA),通过多次积累低秩更新矩阵来提高更新秩,并引入一种基于动量的卸载策略以减轻训练不稳定性。实验结果表明,PLoRA具有更强的学习能力,最高可达到LoRA学习能力的1.8倍,但不增加内存使用。
Abstract
supervised fine-tuning
is the most common method to adapt large language models (LLMs) to downstream tasks, but full fine-tuning LLMs requires massive computational resources. Recently,
parameter-efficient fine-tuning
→