BriefGPT.xyz
Jan, 2024
量化侧调优:快速和内存高效优化量化大型语言模型
Quantized Side Tuning: Fast and Memory-Efficient Tuning of Quantized Large Language Models
HTML
PDF
Zhengxin Zhang, Dan Zhao, Xupeng Miao, Gabriele Oliaro, Qing Li...
TL;DR
通过使用量化、分离网络和低秩适配器等方法,Quantized Side Tuning (QST)能够实现大型语言模型(LLMs)的内存高效、快速的微调,并在减少内存占用的同时达到与最先进方法相媲美的性能,可将总内存占用减少最多7倍。
Abstract
finetuning
large language models
(LLMs) has been empirically effective on a variety of downstream tasks. Existing approaches to
finetuning
→