BriefGPT.xyz
Jun, 2024
通过对语言模型中的序列遗忘进行近似优化参数保护隐私
Protecting Privacy Through Approximating Optimal Parameters for Sequence Unlearning in Language Models
HTML
PDF
Dohyun Lee, Daniel Rim, Minseok Choi, Jaegul Choo
TL;DR
我们提出了一种名为“POP”的新型遗忘方法,通过对参数应用最佳梯度更新,从预训练的语言模型中有效地遗忘目标令牌序列,实现隐私保护,具有出色的遗忘后保留性能,优于现有技术水平。
Abstract
Although
language models
(LMs) demonstrate exceptional capabilities on various tasks, they are potentially vulnerable to extraction attacks, which represent a significant privacy risk. To mitigate the
privacy concerns
→