BriefGPT.xyz
Jul, 2018
变分贝叶斯的dropout: 陷阱与修正
Variational Bayesian dropout: pitfalls and fixes
HTML
PDF
Jiri Hron, Alexander G. de G. Matthews, Zoubin Ghahramani
TL;DR
将Dropout重新解释为贝叶斯神经网络的近似推理算法,提出了一个有用的理论框架,但对于使用不当的先验概率,存在未定义或病态行为的真后验分布问题;对于近似分布相对于真后验分布的奇异性而言,近似难以定义。为了解决这些问题,提出了Quasi-KL(QKL)差异作为新的近似推理目标。
Abstract
dropout
, a stochastic regularisation technique for training of neural networks, has recently been reinterpreted as a specific type of
approximate inference
algorithm for
→