Abstract

To address the limitations of existing pruning methods in practical applications, such as the necessity of training from scratch with sparsity regularization or complex data-driven optimization, we set out from a novel perspective to explore parameter redundancy and accelerate deep CNNs. Precisely, we argue that channels revealing similar feature information have functional overlap and that each such similarity group can be reduced to a few representatives with little impact on the representational power of the model. After deriving an effective metric for evaluating channel similarity via probabilistic modeling, we introduce a similarity-based pruning framework based on hierarchical clustering.In particular, the proposed algorithm can be directly applied to all kinds of pre-trained CNN models for better trade-offs between latency and accuracy. Moreover, rather than relying on a pre-defined target structure, it automatically discovers resource-efficient ones out of the original model under given budgets, which is in the same flavor as NAS. Extensive experiments on benchmark datasets well demonstrate the superior performance of our approach over prior arts. On ImageNet, our pruned ResNet-50 with 30% FLOPs reduced outperforms the original model. We further extend our algorithm to a GAN-based generative model and achieve 2× acceleration, showing its remarkable generalization capability and flexibility.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call