BriefGPT.xyz
May, 2021
通过分解表示进行自监督的对抗样本检测
Self-Supervised Adversarial Example Detection by Disentangled Representation
HTML
PDF
Zhaoxi Zhang, Leo Yu Zhang, Xufei Zheng, Shengshan Hu, Jinyu Tian...
TL;DR
该论文提出了一种名为DRR的新方法,它使用自编码器结构进行对抗性样本检测,并采用解缠表示来减少不必要的自编码器泛化能力,能够在视觉任务和自然语言任务中提供抗适应性对抗性攻击,并在各种度量标准下表现出更好的性能。
Abstract
Deep learning models are known to be vulnerable to
adversarial examples
that are elaborately designed for malicious purposes and are imperceptible to the human perceptual system.
autoencoder
, when trained solely
→