Abstract
With the rapid development of hardware GPU and the advent of the era of big data, neural networks have developed rapidly and greatly improved the recognition performance in various fields. The application of nerual networks to intelligent mobile embedded military equipment will become the wave of the development of a new generation of deep learning. However, neural networks are both computationally intensive and memory intensive, making them difficult to deploy on embedded systems with limited hardware resources. To address this limitation, we propose a new method called “deep compression”, which including three stages: pruning, quantification and Huffmann encoding, to reduce the storage requirement of neural networks without impacting original accuracy. Our method first prunes the network by learning only the important connections. Next, we quantize the weights to enforce weight sharing, finally, weapply Huffmann encoding. We evaluated our method on both MNIST and ImageNet. On the ImageNet dataset, our method reduced the storage of AlexNet by 35× without loss of accuracy and compressed VGG-16 model by 49×, also with no loss of accuracy. Our method is an efficient solution for real-time multi-objective recognition based on lightweight deep neural networks.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. Adv. Neural Inform. Process. Syst. 25, 1097–1105 (2012)
Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)
Szegedy, C., Liu, W., Jia, Y., et al.: Going deeper with convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–9 (2015)
He, K., Zhang, X., Ren, S., et al.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016)
Karpathy, A., Fei-Fei, L.: Deep visual-semantic alignments for generating image descriptions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3128–3137 (2015)
Zhang, C., Li, P., Sun, G., Guan, Y., Xiao, B., Cong, J.: Optimizing FPGA-based accelerator design for deep convolutional neural networks. In: Proceedings of the 2015 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays, pp. 161–170 (2015)
Motamedi, M., Gysel, P., Akella, V., et al.: Design space exploration of FPGA-based deep convolutional neural networks. In: Design Automation Conference (ASP-DAC), 2016 21st Asia and South Pacific, pp. 575–580. IEEE (2016)
Qiu, J., Wang, J., Yao, S., et al.: Going deeper with embedded fpga platform for convolutional neural network. In: Proceedings of the 2016 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays, pp. 26–35 (2016)
Dundar, A., Jin, J., Gokhale, V., et al.: Accelerating deep neural networks on mobile processor with embedded programmable logic. In: Neural Information Processing Systems Conference (2013)
LeCun, Y., Bottou, L., Bengio, Y., et al.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278–2324 (1998)
Acknowledgments
This work is supported by Xiamen Major Science and Technology Projects (No. 3502Z20201017).
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Wei, C., Lu, Z., Lin, Z., Zhong, C. (2021). Research on Deep Neural Network Model Compression Based on Quantification Pruning and Huffmann Encoding. In: Huang, DS., Jo, KH., Li, J., Gribova, V., Hussain, A. (eds) Intelligent Computing Theories and Application. ICIC 2021. Lecture Notes in Computer Science(), vol 12837. Springer, Cham. https://doi.org/10.1007/978-3-030-84529-2_44
Download citation
DOI: https://doi.org/10.1007/978-3-030-84529-2_44
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-84528-5
Online ISBN: 978-3-030-84529-2
eBook Packages: Computer ScienceComputer Science (R0)