BriefGPT.xyz
Feb, 2019
非凸学习中带噪声梯度方法的泛化误差界
On Generalization Error Bounds of Noisy Gradient Methods for Non-Convex Learning
HTML
PDF
Jian Li, Xuanyuan Luo, Mingda Qiao
TL;DR
本文应用Bayes-Stability框架证明算法相关的广义误差界,得到了随机梯度 Langevin 动力学以及其他一些带噪声梯度的方法(例如加动量,小批量和加速,熵-SGD)的数据相关的新广义误差界,论文结果较之前相关研究更紧凑。
Abstract
generalization error
(also known as the out-of-sample error) measures how well the hypothesis obtained from the training data can generalize to previously unseen data. Obtaining tight
generalization error
bounds
→