BriefGPT.xyz
Dec, 2017
梯度下降学习一层卷积神经网络:不必担心虚假局部极小值
Gradient Descent Learns One-hidden-layer CNN: Don't be Afraid of Spurious Local Minima
HTML
PDF
Simon S. Du, Jason D. Lee, Yuandong Tian, Barnabas Poczos, Aarti Singh
TL;DR
研究了在卷积层和ReLU激活下的一层神经网络的学习问题,证明了随机初始化并使用归一化权重的梯度下降可以恢复真实参数,但存在虚假局部最小值,且该局部最小值在梯度下降的动力学中起到了重要作用。
Abstract
We consider the problem of learning a one-hidden-layer
neural network
with non-overlapping
convolutional layer
and
relu activation
functio
→