首页 | 本学科首页   官方微博 | 高级检索  
     


Adversarial image detection in deep neural networks
Authors:Carrara  Fabio  Falchi  Fabrizio  Caldelli  Roberto  Amato  Giuseppe  Becarelli  Rudy
Affiliation:1.ISTI - CNR, Via G. Moruzzi, 1, Pisa, Italy
;2.CNIT Research Unit at MICC - University of Florence, Viale Morgagni, 65, Firenze, Italy
;
Abstract:

Deep neural networks are more and more pervading many computer vision applications and in particular image classification. Notwithstanding that, recent works have demonstrated that it is quite easy to create adversarial examples, i.e., images malevolently modified to cause deep neural networks to fail. Such images contain changes unnoticeable to the human eye but sufficient to mislead the network. This represents a serious threat for machine learning methods. In this paper, we investigate the robustness of the representations learned by the fooled neural network, analyzing the activations of its hidden layers. Specifically, we tested scoring approaches used for kNN classification, in order to distinguish between correctly classified authentic images and adversarial examples. These scores are obtained searching only between the very same images used for training the network. The results show that hidden layers activations can be used to reveal incorrect classifications caused by adversarial attacks.

Keywords:
本文献已被 SpringerLink 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号