BriefGPT.xyz
Jun, 2021
神经网络中良性过拟合现象的理解探讨
Towards an Understanding of Benign Overfitting in Neural Networks
HTML
PDF
Zhu Li, Zhi-Hua Zhou, Arthur Gretton
TL;DR
本研究探讨了现代机器学习模型中广泛存在的过度拟合现象及理论预测,表明超学习风险会在满足一定条件的情况下逐渐减小,并且在两层神经网络中使用ReLU激活函数的情况下具有近最小化学习率的能力。同时,还发现当网络参数数量超过O(n^2)时,超学习风险开始增加,这与最近的实证结果相符。
Abstract
Modern
machine learning
models often employ a huge number of parameters and are typically optimized to have zero training loss; yet surprisingly, they possess near-optimal prediction performance, contradicting classical
→