BriefGPT.xyz
Jun, 2024
深入剖析语言模型微调中的遗忘现象:基于示例关联的统计分析
Demystifying Forgetting in Language Model Fine-Tuning with Statistical Analysis of Example Associations
HTML
PDF
Xisen Jin, Xiang Ren
TL;DR
本文通过对语言模型进行经验分析,发现忘记常常可以通过上游示例和新学习任务的简单乘法关系来近似,并揭示了特定子集示例的复杂忘记模式,在基于经验关联的矩阵补全方法中预测了在学习新任务时发生在上游示例上的遗忘,优于依赖可训练语言模型的先前方法。
Abstract
language models
(LMs) are known to suffer from
forgetting
of previously learned examples when fine-tuned, breaking stability of deployed LM systems. Despite efforts on mitigating
→