BriefGPT.xyz
Mar, 2021
可学习的压缩量化技术用于精确低比特神经网络
Learnable Companding Quantization for Accurate Low-bit Neural Networks
HTML
PDF
Kohei Yamamoto
TL;DR
本文提出了一种可学习的压缩量化方法(LCQ), 该方法能够灵活地通过优化模型权重和可学习的压缩函数来控制权重和激活的压缩级别,从而优于传统的最先进方法,并缩小量化模型与全精度模型之间的差距。
Abstract
quantizing
deep neural networks
is an effective method for reducing memory consumption and improving inference speed, and is thus useful for implementation in
→