BriefGPT.xyz
May, 2023
卷积神经网络初步凝结的理解
Understanding the Initial Condensation of Convolutional Neural Networks
HTML
PDF
Zhangchen Zhou, Hanxu Zhou, Yuqing Li, Zhi-Qin John Xu
TL;DR
本研究探讨了卷积神经网络在小初始化和梯度训练方法下内核权重的凝聚现象,实验证明该现象在卷积神经网络中同样存在且显著。理论上,本研究证明在有限的训练期间,具有小初始化的两层卷积神经网络内核将收敛至一个或几个方向,为对具有专业结构的神经网络表现出的非线性训练行为的更好理解迈出了一步。
Abstract
Previous research has shown that fully-connected networks with small initialization and gradient-based
training methods
exhibit a phenomenon known as
condensation
during training. This phenomenon refers to the in
→