论文标题
开放式对抗防御和清洁对抗的相互学习
Open-set Adversarial Defense with Clean-Adversarial Mutual Learning
论文作者
论文摘要
开放式认可和对抗性国防研究深度学习的两个关键方面对于现实世界的部署至关重要。开放式识别的目的是在测试过程中识别开放式类别的样本,而对抗防御的目的是稳健地对网络来抗网络,以抵抗被不可察觉的对抗性噪声扰动的图像。本文表明,开放式识别系统容易受到对抗样本的影响。此外,本文表明,在已知类别中训练的对抗性防御机制无法很好地推广到开放式样本。在这些观察结果的推动下,我们强调了开放设定的对抗防御(OSAD)机制的必要性。本文提出了一个开放式防御网络,该防御网络具有清洁的相互学习(OSDN-CAML),以解决OSAD问题。拟议的网络设计了一个编码器,具有双练习功能刺激层,并与分类器一起学习,以学习无噪声的潜在特征表示,该特征表示可以自适应地消除以通道和空间良好的细心过滤器引导的对抗性噪声。利用了几种技术来学习无噪声和信息丰富的潜在特征空间,以改善对抗性防御和开放式识别的性能。首先,我们合并了一个解码器,以确保可以从获得的潜在特征中重建干净的图像。然后,使用自学来确保潜在功能足以完成辅助任务。最后,为了利用从干净的图像分类来利用更多的互补知识,以促进特征降级并寻找更概括的局部最低限度以进行开放集识别,我们进一步提出了清洁对流的相互学习,在此进一步引入同伴网络(分类干净的图像)以与分类器(分类的对手图像分类)相互学习。
Open-set recognition and adversarial defense study two key aspects of deep learning that are vital for real-world deployment. The objective of open-set recognition is to identify samples from open-set classes during testing, while adversarial defense aims to robustify the network against images perturbed by imperceptible adversarial noise. This paper demonstrates that open-set recognition systems are vulnerable to adversarial samples. Furthermore, this paper shows that adversarial defense mechanisms trained on known classes are unable to generalize well to open-set samples. Motivated by these observations, we emphasize the necessity of an Open-Set Adversarial Defense (OSAD) mechanism. This paper proposes an Open-Set Defense Network with Clean-Adversarial Mutual Learning (OSDN-CAML) as a solution to the OSAD problem. The proposed network designs an encoder with dual-attentive feature-denoising layers coupled with a classifier to learn a noise-free latent feature representation, which adaptively removes adversarial noise guided by channel and spatial-wise attentive filters. Several techniques are exploited to learn a noise-free and informative latent feature space with the aim of improving the performance of adversarial defense and open-set recognition. First, we incorporate a decoder to ensure that clean images can be well reconstructed from the obtained latent features. Then, self-supervision is used to ensure that the latent features are informative enough to carry out an auxiliary task. Finally, to exploit more complementary knowledge from clean image classification to facilitate feature denoising and search for a more generalized local minimum for open-set recognition, we further propose clean-adversarial mutual learning, where a peer network (classifying clean images) is further introduced to mutually learn with the classifier (classifying adversarial images).