首页 | 本学科首页   官方微博 | 高级检索  
     

面向深度学习的对抗样本差异性检测方法
引用本文:王曙燕,侯则昱,孙家泽.面向深度学习的对抗样本差异性检测方法[J].计算机应用,2021,41(7):1849-1856.
作者姓名:王曙燕  侯则昱  孙家泽
作者单位:西安邮电大学 可信软件实验室, 西安 710121
基金项目:2020年陕西省重点研发计划项目(2020GY-010);2019年西安市科技计划项目(2019218114GXRC017CG018-GXYD17.10)。
摘    要:深度神经网络(DNN)在许多深度学习关键系统如人脸识别、智能驾驶中被证明容易受到对抗样本攻击,而对多种类对抗样本的检测还存在着检测不充分以及检测效率低的问题,为此,提出一种面向深度学习模型的对抗样本差异性检测方法。首先,构建工业化生产中常用的残差神经网络模型作为对抗样本生成与检测系统的模型;然后,利用多种对抗攻击攻击深度学习模型以产生对抗样本组;最终,构建样本差异性检测系统,包含置信度检测、感知度检测及抗干扰度检测三个子检测系统共7项检测方法。在MNIST与Cifar-10数据集上的实验结果表明,属于不同对抗攻击的对抗样本在置信度、感知度、抗干扰度等各项性能检测上存在明显差异,如感知度各项指标优异的对抗样本在置信度以及抗干扰度的检测中,相较于其他类的对抗样本表现出明显不足;同时,证明了在两个数据集上呈现出差异的一致性。通过运用该检测方法,能有效提升模型对对抗样本检测的全面性与多样性。

关 键 词:深度神经网络  对抗攻击  对抗样本  残差神经网络  差异性检测  
收稿时间:2020-08-27
修稿时间:2020-11-29

Difference detection method of adversarial samples oriented to deep learning
WANG Shuyan,HOU Zeyu,SUN Jiaze.Difference detection method of adversarial samples oriented to deep learning[J].journal of Computer Applications,2021,41(7):1849-1856.
Authors:WANG Shuyan  HOU Zeyu  SUN Jiaze
Affiliation:Trusted Software Laboratory, Xi'an University of Posts and Telecommunications, Xi'an Shaanxi 710121, China
Abstract:Deep Neural Network (DNN) is proved to be vulnerable to adversarial sample attacks in many key deep learning systems such as face recognition and intelligent driving. And the detection of various types of adversarial samples has problems of insufficient detection and low detection efficiency. Therefore, a deep learning model oriented adversarial sample difference detection method was proposed. Firstly, the residual neural network model commonly used in industrial production was constructed as the model of the adversarial sample generation and detection system. Then, multiple kinds of adversarial attacks were used to attack the deep learning model to generate adversarial sample groups. Finally, a sample difference detection system was constructed, containing total 7 adversarial sample difference detection methods in sample confidence detection, perception detection and anti-interference degree detection. Empirical research was carried out by the constructed method on the MNIST and Cifar-10 datasets. The results show that the adversarial samples belonging to different adversarial attacks have obvious differences in the performance detection on confidence, perception and anti-interference degrees, for example, in the detection of confidence and anti-interference, the adversarial samples with excellent performance indicators in perception show significant insufficiencies compared to other types of adversarial samples. At the same time, it is proved that there is consistency of the differences in the two datasets. By using this detection method, the comprehensiveness and diversity of the model's detection of adversarial samples can be effectively improved.
Keywords:Deep Neural Network (DNN)  adversarial attack  adversarial sample  residual neural network  difference detection  
本文献已被 万方数据 等数据库收录!
点击此处可从《计算机应用》浏览原始摘要信息
点击此处可从《计算机应用》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号