Abstract

In recent years, thanks to the development of GPU (Graphic Processing Unit) computing power, deep neural network (DNN) has drawn extensive attention from academia and industry and has been widely used in various fields of artificial intelligence. However, the mainstream deep learning models have some problems such as large storage space consumption and high computational complexity, which makes it difficult for the deep neural network model to be deployed to mobile devices with limited computing resources or applications with strict delay requirements. Therefore, on the premise of minimizing the loss of model accuracy, the algorithm researches of deep neural network compression and acceleration which can lighten the model has been carried out. In recent years, model compression methods such as pruning, quantification and knowledge distillation have emerged. This paper focuses on neural network pruning algorithm. Firstly, this paper expounds the basic methods and classification of neural network pruning algorithms. Then according to the classification of pruning granularity, a more detailed classification, analysis and comparison of specific models are carried out from the algorithm level. Then based on cifar-10 data set, validity tests of some neural network pruning algorithms and contrastive analyses of the experimental data from different models such as the parameters are carried out. Finally, in combination of the recent results of neural network pruning algorithm, we put forward research dilemmas and problems to be solved. Then the future research trends and methods of neural network pruning algorithm are prospected.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call