Technical Program

Paper Detail

Paper IDB-2-3.4
Paper Title Detection of Adversarial Examples Based on Sensitivities to Noise Removal Filter
Authors Akinori Higashi, Minoru Kuribayashi, Nobuo Funabiki, Okayama University, Japan; Huy Nguyen, Isao Echizen, National Institute of Informatics, Japan
Session B-2-3: Deep Generative Models for Media Clones and Its Detection
TimeWednesday, 09 December, 17:15 - 19:15
Presentation Time:Wednesday, 09 December, 18:00 - 18:15 Check your Time Zone
All times are in New Zealand Time (UTC +13)
Topic Multimedia Security and Forensics (MSF): Special Session: Deep Generative Models for Media Clones and Its Detection
Abstract An injection of malicious noise causes a serious problem in machine learning system. Due to the uncertainty of the system, the noise may misleads the system to the wrong output determined by a malicious party. The created images, videos, speeches are called adversarial examples. The study of fooling an image classifier have been reported as a potential threat for the CNN-based systems. The noise is well-designed so that the existence in an image is kept hidden from human eyes as well as computer-based classifiers. In this paper, we propose a novel method for detecting adversarial images by using the sensitivities of image classifiers. As adversarial images are created by adding noise, we focus on the behavior of outputs of image classifier for differently filtered images. Our idea is to observe the outputs by changing the strength of a noise removal filtering operation, which is called operation-oriented characteristics. With the increase of the strength, the output from a softmax function in an image classifier is drastically changed in case of adversarial images, while it is rather stable in case of normal images. We investigate the operation-oriented characteristics for some noise removal operations and the propose a simple detector of adversarial images. The performance is quantitatively evaluated by experiments for some typical attacks.