BriefGPT.xyz
May, 2019
使用8位浮点数进行混合精度训练
Mixed Precision Training With 8-bit Floating Point
HTML
PDF
Naveen Mellempudi, Sudarshan Srinivasan, Dipankar Das, Bharat Kaul
TL;DR
本文介绍了一个使用8位浮点表示法训练深度神经网络的方法,减少计算精度和主权重复制的精度要求,并且通过强化误差传播和降低量化噪声的方法来提高模型性能。实验表明,所提出方法在多个数据集和不同工作负载下与精度基线相比不降反升。
Abstract
reduced precision computation
for
deep neural networks
is one of the key areas addressing the widening compute gap driven by an exponential growth in model size. In recent years, deep learning training has largel
→