BriefGPT.xyz
Mar, 2024
使用预处理改进最小二乘问题的隐式正则化 SGD
Improving Implicit Regularization of SGD with Preconditioning for Least Square Problems
HTML
PDF
Junwei Su, Difan Zou, Chuan Wu
TL;DR
通过对预处理的随机梯度下降(SGD)和岭回归的综合比较研究,我们建立了预处理的SGD和岭回归的过度风险界限,并证明了存在一个简单的预处理矩阵能够优于标准的和预处理的岭回归,从而展示了预处理的SGD的增强正则化效果。
Abstract
stochastic gradient descent
(SGD) exhibits strong algorithmic regularization effects in practice and plays an important role in the generalization of modern machine learning. However, prior research has revealed instances where the
→