论文标题

捍卫语音身份验证的数据中毒攻击

Defend Data Poisoning Attacks on Voice Authentication

论文作者

Li, Ke, Baird, Cameron, Lin, Dan

论文摘要

随着深度学习的进步,演讲者的验证取得了很高的准确性,并且在我们日常生活的许多场景中,尤其是Web服务的不断增长的市场中,作为一种生物识别验证选项的流行。与传统密码相比,“声音密码”更加方便,因为它们可以减轻人们记住不同密码的记忆。但是,新的机器学习攻击使这些语音身份验证系统处于危险之中。如果没有强大的安全保证,攻击者可以通过欺骗基于深层神经网络(DNN)的语音识别模型来访问合法用户的Web帐户。在本文中,我们证明了对语音身份验证系统的易于实现的数据中毒攻击,这几乎无法通过现有的防御机制来捕获。因此,我们提出了一种更强大的防御方法,称为Guardian,该方法是基于卷积神经网络的歧视者。 《卫报》歧视者整合了一系列新型技术,包括减少偏见,输入增强和集成学习。我们的方法能够将约95%的攻击帐户与普通帐户区分开,这比仅准确性60%的现有方法更有效。

With the advances in deep learning, speaker verification has achieved very high accuracy and is gaining popularity as a type of biometric authentication option in many scenes of our daily life, especially the growing market of web services. Compared to traditional passwords, "vocal passwords" are much more convenient as they relieve people from memorizing different passwords. However, new machine learning attacks are putting these voice authentication systems at risk. Without a strong security guarantee, attackers could access legitimate users' web accounts by fooling the deep neural network (DNN) based voice recognition models. In this paper, we demonstrate an easy-to-implement data poisoning attack to the voice authentication system, which can hardly be captured by existing defense mechanisms. Thus, we propose a more robust defense method, called Guardian, which is a convolutional neural network-based discriminator. The Guardian discriminator integrates a series of novel techniques including bias reduction, input augmentation, and ensemble learning. Our approach is able to distinguish about 95% of attacked accounts from normal accounts, which is much more effective than existing approaches with only 60% accuracy.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源