BriefGPT.xyz
Nov, 2019
有损后训练量化
Loss Aware Post-training Quantization
HTML
PDF
Yury Nahshan, Brian Chmiel, Chaim Baskin, Evgenii Zheltonozhskii, Ron Banner...
TL;DR
研究神经网络量化对损失函数的结构的影响,证明在轻量量化时,损失函数的结构是平坦且可分离的,从而使得一些简单的后量化方法能够获得良好的结果。同时,提出了一种通过同时量化层参数来提高精度的方法。
Abstract
neural network quantization
enables the deployment of large models on resource-constrained devices. Current
post-training quantization
methods fall short in terms of accuracy for INT4 (or lower) but provide reaso
→