BriefGPT.xyz
Apr, 2020
自适应取整的后训练量化:向上或向下?
Up or Down? Adaptive Rounding for Post-Training Quantization
HTML
PDF
Markus Nagel, Rana Ali Amjad, Mart van Baalen, Christos Louizos, Tijmen Blankevoort
TL;DR
本文提出了AdaRound,它是一种更好的后训练量化权重舍入机制,能够适应数据和任务损失,不需要对网络进行微调,并且只使用少量无标签数据。
Abstract
When
quantizing neural networks
, assigning each floating-point weight to its nearest
fixed-point value
is the predominant approach. We find that, perhaps surprisingly, this is not the best we can do. In this pape
→