Abstract

Neural network classifiers (NNCs) are known to be vulnerable to malicious adversarial perturbations of inputs including those modifying a small fraction of the input features named sparse or L0 attacks. Effective and fast L0 attacks, such as the widely used Jacobian-based Saliency Map Attack (JSMA) are practical to fool NNCs but also to improve their robustness. In this paper, we show that penalising saliency maps of JSMA by the output probabilities and the input features of the NNC leads to more powerful attack algorithms that better take into account each input’s characteristics. This leads us to introduce improved versions of JSMA, named Weighted JSMA (WJSMA) and Taylor JSMA (TJSMA), and demonstrate through a variety of white-box and black-box experiments on three different datasets (MNIST, CIFAR-10 and GTSRB), that they are both significantly faster and more efficient than the original targeted and non-targeted versions of JSMA. Experiments also demonstrate, in some cases, very competitive results of our attacks in comparison with the Carlini-Wagner (CW) L0 attack, while remaining, like JSMA, significantly faster (WJSMA and TJSMA are more than 50 times faster than CW L0 on CIFAR-10). Therefore, our new attacks provide good trade-offs between JSMA and CW for L0 real-time adversarial testing on datasets such as the ones previously cited.

Highlights

  • Deep learning classifiers are used in a wide variety of situations, such as vision, speech recognition, financial fraud detection, malware detection, autonomous driving, defense, and more.The ubiquity of deep learning algorithms in many applications, especially those that are critical such as autonomous driving [1,2] or that pertain to security and privacy [3,4] makes their attack useful

  • We provide a deep comparison between Weighted JSMA (WJSMA) and Taylor JSMA (TJSMA) which is of independent interest

  • The first part focuses on targeted attacks and provides intensive comparisons between Jacobian-based Saliency Map Attack (JSMA), WJSMA and TJSMA on deep Neural network classifiers (NNCs) on MNIST and CIFAR-10 as well as comparisons with CW L0

Read more

Summary

Introduction

Deep learning classifiers are used in a wide variety of situations, such as vision, speech recognition, financial fraud detection, malware detection, autonomous driving, defense, and more. Since the L0 distance is not convenient for gradient descent, the authors of [11] solve this problem by making use of their L2 attack and an algorithm that iteratively eliminates the components without much effect on the output classification They obtain an effective L0 attack that has a net advantage over JSMA. WJSMA applies a simple weighting to saliency maps by the output probabilities, and TJSMA does the same, while penalising extremal input features. Both attacks are more efficient than JSMA according to several metrics Appendices A and B are appendices dedicated to supplementary results and materials

Targeted Attacks
Non-Targeted Attacks
Experiments
Experiments on Targeted Attacks
Avoid Confusion
Run-Time Comparison
Comparison with CW L0 Attack
Experiments on Non-Targeted Attacks
White-Box Experiments
Black-Box Experiments
Comparison with Non-L0 Attacks
Conclusions
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.