BriefGPT.xyz
Jan, 2024
连续学习的步长优化
Step-size Optimization for Continual Learning
HTML
PDF
Thomas Degris, Khurram Javed, Arsalan Sharifnassab, Yuxin Liu, Richard Sutton
TL;DR
在这篇论文中,我们展示了常用的算法(如RMSProp和Adam)在调整步长向量时忽略了其适应过程对整体目标函数的影响,并通过简单问题的实验显示,与RMSProp和Adam相比,IDBD算法可以持续改进步长向量。我们讨论了两种方法的差异和各自的限制,并建议将两种方法结合起来以提高神经网络在持续学习中的性能。
Abstract
In
continual learning
, a learner has to keep learning from the data over its whole life time. A key issue is to decide what knowledge to keep and what knowledge to let go. In a
neural network
, this can be impleme
→