BriefGPT.xyz
May, 2024
一种用于替代梯度学习的广义神经切向核
A generalized neural tangent kernel for surrogate gradient learning
HTML
PDF
Luke Eilers, Raoul-Martin Memmesheimer, Sven Goedeke
TL;DR
研究了神经网络训练方法中激活函数导数不可用时的问题,提出了代理梯度学习(SGL)的理论基础,并利用神经切向核(NTK)的推广——代理梯度NTK分析了SGL,通过数值实验验证了SGL在具有有限宽度和符号激活函数的网络中的有效性。
Abstract
State-of-the-art
neural network training methods
depend on the gradient of the network function. Therefore, they cannot be applied to networks whose
activation functions
do not have useful derivatives, such as bi
→