论文标题

Tensorshield:基于张量的防御图像对对抗性攻击

TensorShield: Tensor-based Defense Against Adversarial Attacks on Images

论文作者

Entezari, Negin, Papalexakis, Evangelos E.

论文摘要

最近的研究表明,诸如深神经网络(DNN)之类的机器学习方法很容易被对抗攻击所欺骗。数据的微妙和不可察觉的扰动能够改变深度神经网络的结果。利用脆弱的机器学习方法引起了许多问题,尤其是在安全性是重要因素的领域。因此,对于针对对抗性攻击设计防御机制至关重要。对于图像分类的任务,不明显的扰动主要发生在图像的高频频谱中。在本文中,我们利用张量分解技术作为预处理步骤,以找到图像的低级别近似值,这些图像可以显着丢弃高频扰动。最近,一个称为SHIELD的防御框架可以通过在Imagenet数据集上的局部图像贴片上执行随机质量JPEG压缩,以“接种”卷积神经网络(CNN)对抗示例。我们的基于张量的防御机制优于SHIELD的SLQ方法14%,而在保持可比的速度的同时,对快速降级下降(FGSM)对抗攻击。

Recent studies have demonstrated that machine learning approaches like deep neural networks (DNNs) are easily fooled by adversarial attacks. Subtle and imperceptible perturbations of the data are able to change the result of deep neural networks. Leveraging vulnerable machine learning methods raises many concerns especially in domains where security is an important factor. Therefore, it is crucial to design defense mechanisms against adversarial attacks. For the task of image classification, unnoticeable perturbations mostly occur in the high-frequency spectrum of the image. In this paper, we utilize tensor decomposition techniques as a preprocessing step to find a low-rank approximation of images which can significantly discard high-frequency perturbations. Recently a defense framework called Shield could "vaccinate" Convolutional Neural Networks (CNN) against adversarial examples by performing random-quality JPEG compressions on local patches of images on the ImageNet dataset. Our tensor-based defense mechanism outperforms the SLQ method from Shield by 14% against FastGradient Descent (FGSM) adversarial attacks, while maintaining comparable speed.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源