BriefGPT.xyz
Mar, 2018
异步共享内存中随机梯度下降算法的收敛性
The Convergence of Stochastic Gradient Descent in Asynchronous Shared Memory
HTML
PDF
Dan Alistarh, Christopher De Sa, Nikola Konstantinov
TL;DR
本文研究并发训练中Stochastic Gradient Descent算法的收敛性,提出了在异步共享内存模型下的新收敛界限,同时指出了最大延迟和算法收敛速度之间的基本权衡关系。
Abstract
stochastic gradient descent
(SGD) is a fundamental algorithm in
machine learning
, representing the optimization backbone for training several classic models, from regression to neural networks. Given the recent p
→