BriefGPT.xyz
May, 2021
分层优化:通过梯度分解进行持续学习
Layerwise Optimization by Gradient Decomposition for Continual Learning
HTML
PDF
Shixiang Tang, Dapeng Chen, Jinguo Zhu, Shijie Yu, Wanli Ouyang
TL;DR
通过梯度分解的方式,实现一个新的连续学习方法,在保留之前任务特定知识的同时,促进共同知识整合,避免了梯度大小变化的影响并实现了当前领域最佳结果。
Abstract
deep neural networks
achieve
state-of-the-art
and sometimes super-human performance across various domains. However, when learning tasks sequentially, the networks easily forget the knowledge of previous tasks, k
→