BriefGPT.xyz
May, 2023
模拟退火自蒸馏纠正改进了对抗训练
Annealing Self-Distillation Rectification Improves Adversarial Training
HTML
PDF
Yu-Yu Wu, Hung-Jui Wang, Shang-Tse Chen
TL;DR
提出一种通过生成软标签辅助来增强对抗训练模型的方法 Annealing Self-Distillation Rectification (ADR),该方法显著改善了模型的鲁棒性并且易于与其他对抗训练技术进行无缝集成。
Abstract
In standard
adversarial training
, models are optimized to fit one-hot labels within allowable
adversarial perturbation
budgets. However, the ignorance of underlying distribution shifts brought by perturbations ca
→