BriefGPT.xyz
Mar, 2019
早停梯度下降在过度参数化的神经网络上被证明对标签噪声具有鲁棒性
Gradient Descent with Early Stopping is Provably Robust to Label Noise for Overparameterized Neural Networks
HTML
PDF
Mingchen Li, Mahdi Soltanolkotabi, Samet Oymak
TL;DR
本文研究神经网络的训练,证明使用梯度下降法可以在一定的噪声或污染下保证稳健性,避免过拟合。
Abstract
Modern
neural networks
are typically trained in an
over-parameterized
regime where the parameters of the model far exceed the size of the training data. Due to over-parameterization these
→