BriefGPT.xyz
Oct, 2019
利用无限宽深度神经网络在小数据任务中的能力
Harnessing the Power of Infinitely Wide Deep Nets on Small-data Tasks
HTML
PDF
Sanjeev Arora, Simon S. Du, Zhiyuan Li, Ruslan Salakhutdinov, Ruosong Wang...
TL;DR
该研究表明:(a) 在无穷宽度神经网络(NNs)上应用l2 损失(通过梯度下降法)训练,并将学习率设置为无穷小,与 (b) 基于所谓的神经切向核(NTK)的核回归是相等的。在此基础上,对NTK进行高效计算的算法已被提出,表明NTK在低数据任务上表现良好。
Abstract
Recent research shows that the following two models are equivalent: (a) infinitely wide
neural networks
(NNs) trained under l2 loss by gradient descent with infinitesimally small learning rate (b)
kernel regression
→