BriefGPT.xyz
Oct, 2024
ALLoRA:自适应学习率缓解LoRA致命缺陷
ALLoRA: Adaptive Learning Rate Mitigates LoRA Fatal Flaws
HTML
PDF
Hai Huang, Randall Balestriero
TL;DR
本研究针对LoRA在有限数据和训练步骤下的微调局限性,提出了一种新的方法:自适应学习率的ALLoRA。通过消除Dropout和比例因子的影响,ALLoRA显著提高了训练的动态性和收敛性,实验证明在多种设置中其精度优于原有的LoRA及其变体。
Abstract
Low-Rank Adaptation
(LoRA) is the bread and butter of
Large Language Model
(LLM) finetuning. LoRA learns an additive low-rank perturbation, $AB$, of a pretrained matrix parameter $W$ to align the model to a new t
→