BriefGPT.xyz
Aug, 2021
隐式稀疏正则化:深度和提前停止的影响
Implicit Sparse Regularization: The Impact of Depth and Early Stopping
HTML
PDF
Jiangyuan Li, Thanh V. Nguyen, Chinmay Hegde, Raymond K. W. Wong
TL;DR
本文研究了梯度下降的隐式偏差对于稀疏回归的影响,并将关于二次参数化的回归结果扩展到更一般的深度为N的网络,结果表明通过提前停止来实现隐式稀疏规则化至关重要,并且对于一般深度参数N,足够小的初始化和步长可以实现最小化最优的稀疏恢复。
Abstract
In this paper, we study the
implicit bias
of
gradient descent
for
sparse regression
. We extend results on regression with quadratic parame
→