BriefGPT.xyz
Nov, 2019
深度ReLU网络学习所需的过度参数化程度是多少?
How Much Over-parameterization Is Sufficient to Learn Deep ReLU Networks?
HTML
PDF
Zixiang Chen, Yuan Cao, Difan Zou, Quanquan Gu
TL;DR
本研究探讨了在过度参数化的深度神经网络中,当网络宽度大于训练样本大小和目标误差的高次多项式的倒数时,通过(随机)梯度下降学习的深度神经网络可以获得良好的优化和泛化性能。此外,我们还构建了深层ReLU网络的学习保证,使得网络宽度对n和ϵ的对数具有良好保证。
Abstract
A recent line of research on
deep learning
focuses on the extremely
over-parameterized setting
, and shows that when the
network width
is l
→