BriefGPT.xyz
May, 2024
稠密提取累积知识用于连续学习
Densely Distilling Cumulative Knowledge for Continual Learning
HTML
PDF
Zenglin Shi, Pei Liu, Tong Su, Yunpeng Wu, Kuien Liu...
TL;DR
连续学习面临的致命遗忘问题可以通过稠密知识蒸馏方法得到改善,该方法可以在所有任务间蒸馏累积知识,提高模型的稳定性。
Abstract
continual learning
, involving sequential training on diverse tasks, often faces
catastrophic forgetting
. While
knowledge distillation
-base
→