BriefGPT.xyz
Sep, 2024
平坦的LoRA:在平坦损失景观上的低秩适应
Flat-LoRA: Low-Rank Adaption over a Flat Loss Landscape
HTML
PDF
Tao Li, Zhengbao He, Yujun Li, Yasheng Wang, Lifeng Shang...
TL;DR
本研究解决了大规模预训练模型微调的高昂计算和内存成本问题,提出了Flat-LoRA方法,旨在寻找位于全参数空间平坦区域的低秩适应。通过随机权重扰动与贝叶斯期望损失目标相结合,Flat-LoRA在自然语言处理和图像分类任务中表现出了卓越的性能,提高了微调模型的效率和泛化能力。
Abstract
Fine-Tuning
large-scale pre-trained models is prohibitively expensive in terms of computational and memory costs.
Low-Rank Adaptation
(LoRA), a popular
→