BriefGPT.xyz
Feb, 2024
L4Q:大型语言模型的参数高效量化感知训练:基于LoRA-wise LSQ
L4Q: Parameter Efficient Quantization-Aware Training on Large Language Models via LoRA-wise LSQ
HTML
PDF
Hyesung Jeon, Yulhwa Kim, Jae-joon Kim
TL;DR
L4Q是一种参数高效的量化感知训练算法,利用LLMs中学到的低秩适应性量化步长,实现对高精度模型的同时量化和微调,达到亚4位精度并保持与应用PEFT在量化模型上相当的训练时间。
Abstract
post-training quantization
(PTQ) and
quantization-aware training
(QAT) methods are gaining popularity in mitigating the high memory and computational costs associated with Large Language Models (LLMs). In resourc
→