Abstract
Deep neural network pruning is an effective model compression and acceleration method. In the initial pruning stage, maintaining the integrity of the input channel of the convolution layer is very important to improve the performance of the pruning model. This paper proposes a two-stage multi-strategy progressive kernel pruning method with adjustable input channels. First, a Hybrid Norm Sparse Index (HNSI) is defined as the basis for selecting the number of retained kernels, and then a two-stage progressive pruning technique is adopted. In the first stage, HNSI is used in the groups for moderate kernel pruning. HNSI in the group can reserve at least one kernel in each group, allowing the input channel information of each layer to be mapped to the next layer and ensuring better network optimization through moderate pruning. In the second stage, HNSI is used in the layers for adjustable full kernel pruning. At this stage, the HNSI of each layer is the basis of preserving the kernel number, and the pruning process is divided into two strategies. The first strategy is kernel pruning in the low-level layer. On the basis of the overall kernel pruning in the layer, each group is forced to retain at least one kernel, thus ensuring that the primary feature of each input channel can be transmitted to the next layer. The second strategy is kernel pruning in the high-level layer. Because of the stronger information abstraction ability in the high-level layer, only the valid input channel information can be passed to the next layer, no longer forcing each group to retain the kernel, which can greatly improve the efficiency of network pruning. Model analysis and experiments show that the two-stage kernel pruning can not only obtain better network optimization direction under moderate pruning but also obtain better network performance under a higher pruning rate.
Similar content being viewed by others
References
Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778
Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu C-Y, Berg AC (2016) SSD: Single shot multibox detector. In: European conference on computer vision, pp 21–37. Springer
Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: Unified, real-time object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 779–788
Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3431–3440
Li Y, Qi H, Dai J, Ji X, Wei Y (2017) Fully convolutional instance-aware semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 2359–2367
Yan S, Teng Y, Smith JS, Zhang B (2016) Driver behavior recognition based on deep convolutional neural networks. In: 2016 12th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery (ICNC-FSKD), IEEE, pp 636–641
Reiß S, Roitberg A, Haurilet M, Stiefelhagen R (2020) Deep classification-driven domain adaptation for cross-modal driver behavior recognition. In: 2020 IEEE Intelligent Vehicles Symposium (IV), IEEE, pp 1042–1047
Takahashi N, Gygli M, Van Gool L (2017) Aenet: Learning deep audio features for video analysis. IEEE Transactions on Multimedia 20(3):513–524
Chen J, Li K, Deng Q, Li K, Philip SY (2019) Distributed deep learning model for intelligent video surveillance systems with edge computing. IEEE Transactions on Industrial Informatics
Mozaffari S, Al-Jarrah OY, Dianati M, Jennings P, Mouzakitis A (2019) Deep learning-based vehicle behaviour prediction for autonomous driving applications: A review. arXiv preprint arXiv:1912.11676
Fayjie AR, Hossain S, Oualid D, Lee D-J (2018) Driverless car: Autonomous driving using deep reinforcement learning in urban environment. In: 2018 15th International Conference on Ubiquitous Robots (UR), IEEE, pp 896–901
LeCun Y, Denker JS, Solla A (1990) Optimal brain damage. In: Advances in neural information processing systems, pp 598–605
Hassibi B, Stork DG (1993) Second order derivatives for network pruning: Optimal brain surgeon. In: Advances in neural information processing systems, pp 164–171
Han S, Pool J, Tran J, Dally W (2015) Learning both weights and connections for efficient neural network. In: Advances in neural information processing systems, pp 1135–1143
Han S, Mao H, Dally WJ (2015) Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149
Ullrich K, Meeds E, Welling M (2017) Soft weight-sharing for neural network compression. arXiv preprint arXiv:1702.04008
Xiao X, Wang Z, Rajasekaran S (2019) Autoprune: Automatic network pruning by regularizing auxiliary parameters. In: Advances in Neural Information Processing Systems, pp 13681–13691
Wang Y, Xu C, You S, Tao D, Xu C (2016) Cnnpack: Packing convolutional neural networks in the frequency domain. In: Advances in neural information processing systems, pp 253–261
Liu Z, Xu J, Peng X, Xiong R (2018) Frequency-domain dynamic pruning for convolutional neural networks. In: Advances in Neural Information Processing Systems, pp 1043–1053
Li H, Kadav A, Durdanovic I, Samet H, Graf HP (2016) Pruning filters for efficient convnets. arXiv preprint arXiv:1608.08710
He Y, Kang G, Dong X, Fu Y, Yang Y (2018) Soft filter pruning for accelerating deep convolutional neural networks. arXiv preprint arXiv:1808.06866
He Y, Liu P, Wang Z, Hu Z, Yang Y (2019) Filter pruning via geometric median for deep convolutional neural networks acceleration. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 4340–4349
Luo J-H, Wu J, Lin W (2017) ThiNet: A filter level pruning method for deep neural network compression. In: Proceedings of the IEEE international conference on computer vision, pp 5058–5066
Yu R, Li A, Chen C-F, Lai J-H, Morariu VI, Han X, Gao M, Lin C-Y, Davis LS (2018) Nisp: Pruning networks using neuron importance score propagation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 9194–9203
Zhuang Z, Tan M, Zhuang B, Liu J, Guo Y, Wu Q, Huang J, Zhu J (2018) Discrimination-aware channel pruning for deep neural networks. In: Advances in Neural Information Processing Systems, pp 875–886
He Y, Zhang X, Sun J (2017) Channel pruning for accelerating very deep neural networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp 1389–1397
Liu Z, Li J, Shen Z, Huang G, Yan S, Zhang C (2017) Learning efficient convolutional networks through network slimming. In: Proceedings of the IEEE International Conference on Computer Vision, pp 2736–2744
Lin S, Ji R, Li Y, Wu Y, Huang F, Zhang B (2018) Accelerating convolutional networks via global & dynamic filter pruning.. In: IJCAI, pp 2425–2432
Lin S, Ji R, Li Y, Deng C, Li X (2019) Toward compact convnets via structure-sparsity regularized filter pruning. IEEE transactions on neural networks and learning systems 31(2):574–588
Zhu X, Zhou W, Li H (2018) Improving deep neural network sparsity through decorrelation regularization.. In: IJCAI, pp 3264–3270
Liu C, Wang Y, Han K, Xu C, Xu C (2019) Learning instance-wise sparsity for accelerating deep models. arXiv preprint arXiv:1907.11840
Huang Z, Wang N (2018) Data-driven sparse structure selection for deep neural networks. In: Proceedings of the European conference on computer vision (ECCV), pp 304–320
Louizos C, Welling M, Kingma DP (2017) Learning sparse neural networks through l_0 regularization. arXiv preprint arXiv:1712.01312
Molchanov D, Ashukha A, Vetrov D (2017) Variational dropout sparsifies deep neural networks. In: Proceedings of the 34th International Conference on Machine Learning-Volume 70, JMLR. org, pp 2498–2507
Wang Z, Lin S, Xie J, Lin Y (2019) Pruning blocks for CNN compression and acceleration via online ensemble distillation. IEEE Access 7:175703–175716
Lin S, Ji R, Yan C, Zhang B, Cao L, Ye Q, Huang F, Doermann D (2019) Towards optimal structured CNN pruning via generative adversarial learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 2790–2799
Liu N, Ma X, Xu Z, Wang Y, Tang J, Ye J (2020) Autocompress: An automatic DNN structured pruning framework for ultra-high compression rates.. In: AAAI, pp 4876–4883
He Y, Lin J, Liu Z, Wang H, Li L-J, Han S (2018) Amc: Automl for model compression and acceleration on mobile devices. In: Proceedings of the European Conference on Computer Vision (ECCV), pp 784–800
Lin C, Zhong Z, Wei W, Yan J (2018) Synaptic strength for convolutional neural network. In: Advances in Neural Information Processing Systems, pp 10149–10158
Li Y, Lin S, Zhang B, Liu J, Doermann D, Wu Y, Huang F, Ji R (2019) Exploiting kernel sparsity and entropy for interpretable CNN compression. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 2800–2809
Mao H, Han S, Pool J, Li W, Liu X, Wang Y, Dally WJ (2017) Exploring the regularity of sparse structure in convolutional neural networks. arXiv preprint arXiv:1705.08922
Wen W, Wu C, Wang Y, Chen Y, Li H (2016) Learning structured sparsity in deep neural networks. In: Advances in neural information processing systems, pp 2074–2082
Wang H, Zhang Q, Wang Y, Hu R (2018) Structured deep neural network pruning by varying regularization parameters. ArXiv preprint:1804.09461 3
Aubry M, Russell BC (2015) Understanding deep features with computer-generated imagery. In: Proceedings of the IEEE International Conference on Computer Vision, pp 2875–2883
Zhang Q, Nian Wu Y, Zhu S-C (2018) Interpretable convolutional neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 8827–8836
Wagner J, Kohler JM, Gindele T, Hetzel L, Wiedemer JT, Behnke S (2019) Interpretable and fine-grained visual explanations for convolutional neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 9097–9107
Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L-C (2018) Mobilenetv2: Inverted residuals and linear bottlenecks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4510–4520
Krizhevsky A (2009) Learning multiple layers of features from tiny images. Master’s thesis, University of Tront
Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L (2009) Imagenet: A large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition, Ieee, pp 248–255
Adam P, Sam G, Soumith C, Gregory C, Edward Y, Zachary D, Zeming L, Alban D, Luca A, Adam L (2017) Automatic differentiation in pytorch. In: Proceedings of Neural Information Processing Systems
Li Z, Gong Y, Ma X, Liu S, Sun M, Zhan Z, Kong Z, Yuan G, Wang Y (2020) SS-Auto: A single-shot, automatic structured weight pruning framework of DNNs with ultra-high efficiency. arXiv preprint arXiv:2001.08839
Zuo Y, Chen B, Shi T, Sun M (2020) Filter pruning without damaging networks capacity. IEEE Access 8:90924–90930
Ding G, Zhang S, Jia Z, Zhong J, Han J (2020) Where to prune: Using lstm to guide data-dependent soft pruning. IEEE Transactions on Image Processing 30:293–304
Ding X, Ding G, Guo Y, Han J (2019) Centripetal sgd for pruning very deep convolutional networks with complicated structure. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 4943–4953
Acknowledgements
This work was supported in part by the National Natural Science Foundation of China under Grant(62071303, 61871269), Guangdong Basic and Applied Basic Research Foundation(2019A1515011861), Shenzhen Science and Technology Projection (JCYJ20190808151615540).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of Interests
The authors declare that they have no conflict of interest.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Zhu, J., Pei, J. Progressive kernel pruning CNN compression method with an adjustable input channel. Appl Intell 52, 10519–10540 (2022). https://doi.org/10.1007/s10489-021-02932-z
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-021-02932-z