BriefGPT.xyz
Jul, 2019
有序SGD: 一种新的经验风险最小化随机优化框架
A Stochastic First-Order Method for Ordered Empirical Risk Minimization
HTML
PDF
Kenji Kawaguchi, Haihao Lu
TL;DR
论文提出了一种新的随机优化方法,它有针对性地偏向于高损失值的观测结果,并证明该算法对于凸损失具有亚线性收敛率,对于弱凸损失(非凸)具有关键点,同时在 SVM、逻辑回归和深度学习等模型中获得了更好的测试误差。
Abstract
We propose a new stochastic first-order method for
empirical risk minimization
problems such as those that arise in machine learning. The traditional approaches, such as (mini-batch) stochastic gradient descent (SGD), utilize an unbiased
→