BriefGPT.xyz
Dec, 2018
使用8位浮点数训练深度神经网络
Training Deep Neural Networks with 8-bit Floating Point Numbers
HTML
PDF
Naigang Wang, Jungwook Choi, Daniel Brand, Chia-Yu Chen, Kailash Gopalakrishnan
TL;DR
本文介绍了使用较低的精度来训练深度神经网络的成功实践,通过引入基于块的操作和浮点随机取整等技术,成功地实现了在 8 位浮点数下对多种深度学习模型和数据集进行了精确的训练。这些新技术为新一代硬件训练平台奠定了基础,并具有提高 2-4 倍吞吐量的潜力。
Abstract
The state-of-the-art
hardware platforms
for
training
deep neural networks
(DNNs) are moving from traditional single
→