BriefGPT.xyz
Nov, 2018
超参数神经网络中的学习和泛化:超越两层
Learning and Generalization in Overparameterized Neural Networks, Going Beyond Two Layers
HTML
PDF
Zeyuan Allen-Zhu, Yuanzhi Li, Yingyu Liang
TL;DR
本文通过分析神经网络在超参数化情况下的学习理论,证明了神经网络能够通过SGD算法简单地学习某些重要的概念并且样本复杂度几乎独立于网络参数的数量。此外,本文还建立了一个神经网络的二次近似概念,并将其与如何逃离鞍点的SGD理论联系起来。
Abstract
neural networks
have great success in many machine learning applications, but the fundamental
learning theory
behind them remains largely unsolved. Learning
→