BriefGPT.xyz
Mar, 2020
双峰下的双重麻烦:懒惰模式中的偏差与方差
Double Trouble in Double Descent : Bias and Variance(s) in the Lazy Regime
HTML
PDF
Stéphane d'Ascoli, Maria Refinetti, Giulio Biroli, Florent Krzakala
TL;DR
研究发现,通过过度参数化,深度神经网络能够在插值训练数据的同时实现卓越的泛化性能,并且在测试误差上具有双下降现象,该现象可以通过集成平均估计器进行抑制。
Abstract
deep neural networks
can achieve remarkable generalization performances while interpolating the training data perfectly. Rather than the U-curve emblematic of the bias-variance trade-off, their test error often follows a
→