BriefGPT.xyz
Feb, 2021
轻度超参数化双层神经网络的局部收敛理论
A Local Convergence Theory for Mildly Over-Parameterized Two-Layer Neural Network
HTML
PDF
Mo Zhou, Rong Ge, Chi Jin
TL;DR
本文研究了过度参数化对于神经网络构建优化的成功至关重要的原因,并发现过度参数化下的神经网络具有非常稳定的收敛性质,其学生节点不断向教师节点收敛,并有一个漂亮的收敛速率定理,该速率与学生节点的数量无关。
Abstract
While
over-parameterization
is widely believed to be crucial for the success of optimization for the
neural networks
, most existing theories on
o
→