BriefGPT.xyz
Feb, 2018
非凸优化中具有概率保障的随机梯度下降泛化误差界
Generalization Error Bounds with Probabilistic Guarantee for SGD in Nonconvex Optimization
HTML
PDF
Yi Zhou, Yingbin Liang, Huishuai Zhang
TL;DR
本文探讨了深度学习模型的一种优化方法——随机梯度下降在泛化能力上的稳定性,提出了一种基于梯度方差的稳定性指标,并在此基础上分别分析了常规非凸损失函数、梯度主导性损失函数和带强凸规则化器的问题,得到了一系列改进的泛化误差界。
Abstract
The success of
deep learning
has led to a rising interest in the generalization property of the
stochastic gradient descent
(SGD) method, and stability is one popular approach to study it. Existing works based on
→