BriefGPT.xyz
Jan, 2024
一次性权重耦合学习的无需重新训练的模型量化
Retraining-free Model Quantization via One-Shot Weight-Coupling Learning
HTML
PDF
Chen Tang, Yuan Meng, Jiacheng Jiang, Shuzhao Xie, Rongwei Lu...
TL;DR
通过一种一次性训练-搜索范式,本文研究了混合精度模型压缩的问题,并提出了一种动态冻结和信息失真缓解技术来优化位宽配置和性能下降。
Abstract
quantization
is of significance for compressing the over-parameterized deep neural models and deploying them on resource-limited devices. Fixed-precision
quantization
suffers from performance drop due to the limi
→