BriefGPT.xyz
Oct, 2019
正交梯度下降用于连续学习
Orthogonal Gradient Descent for Continual Learning
HTML
PDF
Mehrdad Farajtabar, Navid Azizan, Alex Mott, Ang Li
TL;DR
该论文通过从参数空间的角度解决神经网络的连续学习问题,提出了正交梯度下降(OGD)方法,并通过限制梯度更新方向来避免遗忘之前学习的数据。该方法可以更有效地利用神经网络的高容量,并不需要存储先前学习的数据。实验证明,所提出的OGD方法是有效的。
Abstract
neural networks
are achieving state of the art and sometimes super-human performance on learning tasks across a variety of domains. Whenever these problems require learning in a continual or sequential manner, however,
→