Gianluca Detommaso, Tiangang Cui, Youssef Marzouk, Robert Scheichl, Alessio Spantini
TL;DR本文提出了一种改进的 Stein 变分梯度下降(SVGD)算法,采用函数空间的 Newton 迭代近似二阶信息以加速算法,同时还介绍了二阶信息在选择核函数时的更有效作用。在多个测试案例中,我们观察到与原始 SVGD 算法相比,有了显著的计算优势。
Abstract
stein variational gradient descent (SVGD) was recently proposed as a general purpose nonparametric variational inference algorithm [Liu & Wang, NIPS 2016]: it minimizes the Kullback-Leibler divergence between the