Abstract

Convolutional neural networks (CNNs) suffer from issues of large parameter size and high computational complexity. To address this, we propose an adaptive pruning algorithm based on self-distillation. The algorithm introduces a trainable parameter for each channel to control channel pruning and integrates the pruning process into network training, enabling pruning and fine-tuning in a single training iteration to derive the final pruned model. Moreover, this framework requires only a single overall pruning rate to achieve adaptive pruning for each layer, avoiding tedious hyperparameter tuning for a less iterative, simple and efficient pruning process. Additionally, self-distillation is utilized in the pruning algorithm, leveraging the knowledge from the pretrained CNN to guide its own pruning process, facilitating the recovery from performance degradation caused by pruning and achieving higher accuracy. Extensive pruning experiments on various CNN models over different datasets demonstrate that at least 75% of redundant parameters can be reduced without sacrificing model accuracy.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.