BriefGPT.xyz
Jul, 2019
从零开始的稀疏神经网络:在不降低性能的情况下更快地训练
Sparse Networks from Scratch: Faster Training without Losing Performance
HTML
PDF
Tim Dettmers, Luke Zettlemoyer
TL;DR
本文研究了通过发展一种称为稀疏动量的算法,实现在深度神经网络训练过程中,保持稀疏权重的同时实现稠密表现水平的加速训练方法,实验证明稀疏动量可靠地重现稠密表现水平并提供最多5.61倍的训练加速度。
Abstract
We demonstrate the possibility of what we call
sparse learning
: accelerated
training
of
deep neural networks
that maintain sparse weights
→