BriefGPT.xyz
Mar, 2023
大型语言模型的后训练量化综合研究
A Comprehensive Study on Post-Training Quantization for Large Language Models
HTML
PDF
Zhewei Yao, Cheng Li, Xiaoxia Wu, Stephen Youn, Yuxiong He
TL;DR
通过数万次的零-shot实验,我们对后期训练量化(PTQ)的各种组成成分和效应进行了全面的研究,发现细粒度量化和PTQ方法是获得良好准确性所必需的,并且粗粒度量化的高比特位(例如,5比特)比非常细粒度量化的低比特位(例如,4比特)更强大。
Abstract
post-training quantization
(\ptq) had been recently shown as a compromising method to reduce the
memory consumption
and/or
compute cost
fo
→