Abstract

Deep neural networks have achieved significant development and wide applications for their amazing performance. However, their complex structure, high computation and storage resource limit their applications in mobile or embedding devices such as sensor platforms. Neural network pruning is an efficient way to design a lightweight model from a well-trained complex deep neural network. In this paper, we propose an evolutionary multi-objective one-shot filter pruning method for designing a lightweight convolutional neural network. Firstly, unlike some famous iterative pruning methods, a one-shot pruning framework only needs to perform filter pruning and model fine-tuning once. Moreover, we built a constraint multi-objective filter pruning problem in which two objectives represent the filter pruning ratio and the accuracy of the pruned convolutional neural network, respectively. A non-dominated sorting-based evolutionary multi-objective algorithm was used to solve the filter pruning problem, and it provides a set of Pareto solutions which consists of a series of different trade-off pruned models. Finally, some models are uniformly selected from the set of Pareto solutions to be fine-tuned as the output of our method. The effectiveness of our method was demonstrated in experimental studies on four designed models, LeNet and AlexNet. Our method can prune over 85%, 82%, 75%, 65%, 91% and 68% filters with little accuracy loss on four designed models, LeNet and AlexNet, respectively.

Highlights

  • For evolutionary multi-objective one-shot filter pruning method (EMOFP), the fine-tuned models are better than the original model, except in the case where Compression ratio (CR) is 7.11

  • We will show the experimental results on LeNet, which is one of the most familiar convolutional neural networks

  • AlexNet was the deepest convolutional neural network used to examine the performance of EMOFP in the experimental studies

Read more

Summary

Introduction

Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. In this paper, we will study how to design a lightweight convolutional neural network based on pruning methods that can be deployed on resource-limited devices. The unstructured pruning methods can obtain a sparse model, but it is hard to be deployed on resource-limited devices without specific sparse computing algorithms. The time cost of a human is very expensive for iterative pruning methods, especially for designing a series of similar models. Whether it is iterative pruning or one-shot pruning, we need to design each model independently for each device. We propose an evolutionary multi-objective one-shot filter pruning method (EMOFP) for designing a lightweight convolutional neural network. We will give the concluding remarks of this paper

Neural Network Pruning
Evolutionary Multi-Objective Optimization
Methodology
Framework of EMOFP
Multi-Objective Filter Pruning Model
Evolutionary Multi-Objective Filter Pruning Algorithm
Fine-Tuning Strategy
Computational Complexity of EMOFP
Experimental Studies
Description of Model Variants and Datasets
Experimental Setting
Results on Designed Models
Method
Results on LeNet
Results on AlexNet
Fine-Tuning with Shared Weights or Randomly Initial Weights
Practical Example of Cat and Dog Classification
Conclusions and Future Works
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call