Hwichan Kim, Shota Sasaki, Sho Hoshino, Ukyo Honda
TL;DR一个单线性层产生了任务适应的低秩矩阵,此方法在效果上与LoRA相当,但可训练参数更少。
Abstract
low-rank adaptation (LoRA) is a widely used parameter-efficient fine-tuning (PEFT) method that updates an initial weight matrix $W_0$ with a delta matrix $\Delta W$ consisted by two low-rank matrices $A$ and $B$.