BriefGPT.xyz
Jun, 2021
LoRA: 大型语言模型的低秩适应
LoRA: Low-Rank Adaptation of Large Language Models
HTML
PDF
Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li...
TL;DR
本文提出了一种低秩适应方法(Low-Rank Adaptation,简称LoRA),通过将可训练秩分解矩阵注入变压器结构的每个层中,极大地减少了下游任务中的可训练参数,并且性能与微调相当或更好,同时具有更高的训练吞吐量和没有额外推理延迟,这解决了大规模预训练模型对于微调参数和GPU内存占用过高的问题。
Abstract
The dominant paradigm of
natural language processing
consists of large-scale
pre-training
on general domain data and adaptation to particular tasks or domains. As we pre-train larger models, conventional fine-tun
→