BriefGPT.xyz
May, 2024
对于每个(文本序列)的独立性:改进大型语言模型中的记忆数据遗忘
To Each (Textual Sequence) Its Own: Improving Memorized-Data Unlearning in Large Language Models
HTML
PDF
George-Octavian Barbulescu, Peter Triantafillou
TL;DR
通过新的度量衡、对抗攻击以及基于梯度上升和任务算术的两种新的遗忘方法,本研究提供了关于LLMs隐私保护和遗忘的新视角,并在大量NLP任务上进行了全面的性能评估。
Abstract
llms
have been found to memorize training
textual sequences
and regurgitate verbatim said sequences during text generation time. This fact is known to be the cause of
→