BriefGPT.xyz
Oct, 2023
LoftQ: 针对大型语言模型的LoRA微调感知量化
LoftQ: LoRA-Fine-Tuning-Aware Quantization for Large Language Models
HTML
PDF
Yixiao Li, Yifan Yu, Chen Liang, Pengcheng He, Nikos Karampatziakis...
TL;DR
我们提出了LoftQ,这是一个新的量化框架,旨在同时量化LLM并为LoRA fine-tuning找到适当的低秩初始化,以改善量化和全精度模型之间的差异,并显著提高下游任务的泛化性能。
Abstract
quantization
is an indispensable technique for serving Large Language Models (LLMs) and has recently found its way into
lora fine-tuning
. In this work we focus on the scenario where
→