Abstract

Convolutional neural networks (CNNs) have achieved tremendous success in solving complex classification problems. Motivated by this success, there have been proposed various compression methods for downsizing the CNNs to deploy them on resource-constrained embedded systems. However, a new type of vulnerability of compressed CNNs known as the adversarial examples has been discovered recently, which is critical for security-sensitive systems because the adversarial examples can cause malfunction of CNNs and can be crafted easily in many cases. In this paper, we proposed a compression framework to produce compressed CNNs robust against such adversarial examples. To achieve the goal, our framework uses both pruning and knowledge distillation with adversarial training. We formulate our framework as an optimization problem and provide a solution algorithm based on the proximal gradient method, which is more memory-efficient than the popular ADMM-based compression approaches. In experiments, we show that our framework can improve the trade-off between adversarial robustness and compression rate compared to the existing state-of-the-art adversarial pruning approach.

Highlights

  • In the past few years, convolutional neural networks (CNNs) have achieved great success in many applications including image classification and object detection

  • We focused on the original accuracy of the early stage of the optimization to show how well Adversarial Pruning with Distillation (APD) preserved the original accuracy of the baseline model during the adversarial pruning

  • The adversarial robustness of the compressed CNNs is essential for deploying them to the real-world embedded systems

Read more

Summary

Introduction

In the past few years, convolutional neural networks (CNNs) have achieved great success in many applications including image classification and object detection. The excessively large amount of learning parameters and the vulnerability for the adversarial examples [1,2,3,4,5,6,7,8] are making it difficult to deploy CNNs especially on resource-constrained environments such as smartphones, automobiles, and wearable devices To overcome this drawback, various model compression methods have been proposed, where many are based on weight pruning [9,10,11,12,13,14,15,16,17]. We found that consistently providing information about the pretrained original network during adversarial training can improve the robustness of the resulting compressed network With this intuition, we propose a novel robust pruning framework that jointly uses pruning and knowledge distillation [21] within the adversarial training procedure. Our method showed a better trade-off between adversarial robustness and compression rate compared to the state-of-the-art methods [15,19,22]

Adversarial Attacks
Adversarial Training
Weight Pruning
Knowledge Distillation
Adversarially Robust Model Compression
Methods
The Attack Model
Adversarial Pruning with Distillation
Optimization
The Effect of Knowledge Distillation
Element-Wise Pruning
Filter Pruning
The Convergence Behavior
Comparison with the State-of-the-Art Methods
Computational and Space Complexity
Effectiveness of Knowledge Distillation on Other Attack Methods
Findings
Conclusions
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.