BriefGPT.xyz
Feb, 2025
揭示并缓解知识编辑中的过度关注问题
Revealing and Mitigating Over-Attention in Knowledge Editing
HTML
PDF
Pinzheng Wang, Zecheng Tang, Keyan Zhou, Juntao Li, Qiaoming Zhu...
TL;DR
本研究解决了大型语言模型在知识编辑过程中出现的特定性失败问题,即编辑知识时可能损害现有知识的稳定性。通过引入选择性注意力漂移限制(SADR)方法,限制注意力权重分布的变化,显著减轻了特定性失败现象,从而提高知识编辑的效果。实验结果表明,该方法在多个主流大型语言模型中具有显著的有效性。
Abstract
Large Language Models
have demonstrated superior performance across a wide range of tasks, but they still exhibit undesirable errors due to incorrect knowledge learned from the training data. To avoid this,
Knowledge Ed
→