BriefGPT.xyz
Jun, 2023
在过度参数化模型中放松随机线搜索
Don't be so Monotone: Relaxing Stochastic Line Search in Over-Parameterized Models
HTML
PDF
Leonardo Galli, Holger Rauhut, Mark Schmidt
TL;DR
本文提出了一种名为PoNoS的算法,采用非单调线搜索方法和Polyak初始步进大小,可优化SGD/Adam的收敛速度和一般性能,初步运行对比表明此基于线搜索的算法优于传统算法。
Abstract
Recent works have shown that
line search methods
can speed up
stochastic gradient descent
(SGD) and
adam
in modern over-parameterized sett
→