BriefGPT.xyz
Nov, 2023
用于细调的Rank稳定化缩放因子及其LoRA方法
A Rank Stabilization Scaling Factor for Fine-Tuning with LoRA
HTML
PDF
Damjan Kalajdzievski
TL;DR
通过修改Low-Rank Adapters方法中的缩放因子,我们提出了一种称为rank-stabilized LoRA (rsLoRA)的方法,可以在训练期间用更多的计算资源来换取更好的fine-tuning性能,并且在推理计算成本不变的情况下实现了fine-tuning计算性能的折中。
Abstract
As
large language models
(LLMs) have become increasingly compute and memory intensive,
parameter-efficient fine-tuning
(PEFT) methods are now a common strategy to fine-tune LLMs. A popular PEFT method is
→