BriefGPT.xyz
Feb, 2018
双层神经网络优化景观中的虚假峰谷
Neural Networks with Finite Intrinsic Dimension have no Spurious Valleys
HTML
PDF
Luca Venturi, Afonso Bandeira, Joan Bruna
TL;DR
本文主要研究神经网络中存在的局部极小值问题。针对两层神经网络,定义了其固有维度,并证明了有限的固有维度保证了超参数化的模型不存在局部极小值,而无限的固有维度意味着在某些数据分布下必然存在局部极小值。此外,尽管在一般情况下可能存在局部极小值,但其出现在低风险水平,并高概率地避免在超参数化的模型上。
Abstract
neural networks
provide a rich class of high-dimensional,
non-convex optimization
problems. Despite their non-convexity, gradient-descent methods often successfully optimize these models. This has motivated a rec
→