BriefGPT.xyz
Oct, 2024
LoRA与全面微调:等效性的幻觉
LoRA vs Full Fine-tuning: An Illusion of Equivalence
HTML
PDF
Reece Shuttleworth, Jacob Andreas, Antonio Torralba, Pratyusha Sharma
TL;DR
本研究探讨了LoRA与全面微调在训练后模型中权重矩阵的差异,揭示了这两种微调方法的结果并不等效。研究发现,LoRA模型中出现的高排名奇异向量称为“入侵维度”,这些维度在全面微调中并不存在,并且尽管在目标任务上表现相似,LoRA模型在适应多个任务时的稳健性较差。这一发现对理解不同微调方法影响的模型能力具有重要意义。
Abstract
Fine-tuning
is a crucial paradigm for adapting pre-trained large language models to downstream tasks. Recently, methods like Low-Rank Adaptation (
LoRA
) have been shown to match the performance of fully fine-tuned
→