BriefGPT.xyz
Jun, 2020
梯度下降遵循普通损失的正则化路径
Gradient descent follows the regularization path for general losses
HTML
PDF
Ziwei Ji, Miroslav Dudík, Robert E. Schapire, Matus Telgarsky
TL;DR
本论文研究了机器学习中隐含的偏差及其对应的正则化解,并且根据理论证明我们使用的指数型损失函数的正则化效果,可达到最大保边缘的方向,相应的其他损失函数可能会导致收敛于边缘较差的方向。
Abstract
Recent work across many
machine learning
disciplines has highlighted that standard descent methods, even without explicit
regularization
, do not merely minimize the training error, but also exhibit an
→