BriefGPT.xyz
Jun, 2021
使用不可靠的教师进行可信对抗蒸馏
Reliable Adversarial Distillation with Unreliable Teachers
HTML
PDF
Jianing Zhu, Jiangchao Yao, Bo Han, Jingfeng Zhang, Tongliang Liu...
TL;DR
提出一种新的神经网络训练方法,叫做可靠的内省式敌对蒸馏(IAD),用于提高神经网络对抗攻击的能力。通过在不同情况下,对不同来源的标签进行部分可信任处理,以提高神经网络的稳定性。实验结果表明,IAD在提高抗对抗性方面的效果显著。
Abstract
In ordinary distillation, student networks are trained with
soft labels
(SLs) given by pretrained teacher networks, and students are expected to improve upon teachers since SLs are stronger supervision than the original
→