BriefGPT.xyz
Oct, 2024
LoRTA:大语言模型的低秩张量适配
LoRTA: Low Rank Tensor Adaptation of Large Language Models
HTML
PDF
Ignacio Hounie, Charilaos Kanatsoulis, Arnuv Tandon, Alejandro Ribeiro
TL;DR
本研究针对低秩适配方法在适应大型预训练模型时参数数量过高的限制,提出了一种新颖的低秩张量参数化方法。这一方法显著降低了可训练参数的数量,并在多项基准测试中展示了在保持性能的同时实现高效微调的潜力。
Abstract
Low Rank Adaptation
(LoRA) is a popular Parameter Efficient
Fine Tuning
(PEFT) method that effectively adapts large pre-trained models for downstream tasks. LoRA parameterizes model updates using low-rank matrice
→