BriefGPT.xyz
Jul, 2024
通过循环卷积实现参数高效的微调
Parameter-Efficient Fine-Tuning via Circular Convolution
HTML
PDF
Aochuan Chen, Ziqi Gao, Zijing Liu, Yu Li, Jia Li
TL;DR
本研究解决了低秩适应方法(LoRA)在高性能微调中的局限性,尤其是在计算和内存效率方面。提出的循环卷积适应方法(C$^3$A)不仅实现了更高的适应性,还在资源利用上表现优越,实验结果表明其在各种微调任务中持续超越LoRA及其变种。
Abstract
Low-Rank Adaptation
(LoRA) has gained popularity for
Fine-tuning
large foundation models, leveraging low-rank matrices $\mathbf{A}$ and $\mathbf{B}$ to represent weight changes (\textit{i.e.,} $\Delta \mathbf{W}
→