TL;DR本文提出了一种改进reSGLD的方法来加速处理非凸学习,采用方差约减的策略并在 Markov jump process 下进行非渐进的分析,结果表明我们在优化和不确定性预测方面实现了最先进的效果。
Abstract
Replica exchange stochastic gradient Langevin dynamics (resgld) has shown promise in accelerating the convergence in non-convex learning; however, an excessively large correction for avoiding biases from noisy en