BriefGPT.xyz
Nov, 2021
FQ-ViT: 完全量化视觉Transformer的后训练量化
FQ-ViT: Fully Quantized Vision Transformer without Retraining
HTML
PDF
Yang Lin, Tianyu Zhang, Peiqin Sun, Zheng Li, Shuchang Zhou
TL;DR
本研究开发了基于Power-of-Two Factor(PTF)和Log-Int-Softmax(LIS)的技术来简化全量化视觉Transformer的推理复杂度,并应用于各种基于Transformer的体系结构和基准测试中,以达到更高的性能。
Abstract
network quantization
significantly reduces model inference complexity and has been widely used in real-world deployments. However, most existing quantization methods have been developed and tested mainly on
convolutiona
→