论文标题
对抗排名攻击和防御
Adversarial Ranking Attack and Defense
论文作者
论文摘要
深神经网络(DNN)分类器容易受到对抗性攻击的影响,在这种攻击中,不可察觉的扰动可能导致错误分类。但是,基于DNN的图像排名系统的脆弱性仍然不足。在本文中,我们提出了针对深度排名系统的两次攻击,即候选攻击和查询攻击,可以通过对抗性扰动提高或降低所选候选人的排名。具体而言,预期的排名顺序首先被表示为一组不等式,然后设计出单一的目标函数以获得最佳的扰动。相反,还提出了一种防御方法来改善排名系统的鲁棒性,该系统可以同时减轻所有提议的攻击。我们的对抗性排名攻击和防御在包括MNIST,时尚狂热者和斯坦福 - 在线产品在内的数据集上进行了评估。实验结果表明,我们的攻击可以有效地损害典型的深度排名系统。同时,通过我们的防御能力可以适度改善系统的鲁棒性。此外,我们对手的可转移和通用性能说明了逼真的黑盒攻击的可能性。
Deep Neural Network (DNN) classifiers are vulnerable to adversarial attack, where an imperceptible perturbation could result in misclassification. However, the vulnerability of DNN-based image ranking systems remains under-explored. In this paper, we propose two attacks against deep ranking systems, i.e., Candidate Attack and Query Attack, that can raise or lower the rank of chosen candidates by adversarial perturbations. Specifically, the expected ranking order is first represented as a set of inequalities, and then a triplet-like objective function is designed to obtain the optimal perturbation. Conversely, a defense method is also proposed to improve the ranking system robustness, which can mitigate all the proposed attacks simultaneously. Our adversarial ranking attacks and defense are evaluated on datasets including MNIST, Fashion-MNIST, and Stanford-Online-Products. Experimental results demonstrate that a typical deep ranking system can be effectively compromised by our attacks. Meanwhile, the system robustness can be moderately improved with our defense. Furthermore, the transferable and universal properties of our adversary illustrate the possibility of realistic black-box attack.