Abstract
In this work, we study network binarization (i.e., binary neural networks, BNNs), which is one of the most promising techniques in network compression for convolutional neural networks (CNNs). Although prior work has introduced many binarization methods that improve the accuracy of BNNs by minimizing the quantization error, there remains a non-negligible performance gap between the binarized model and the full-precision model. Given that feature representation is critical for deep neural networks and that in BNNs, the features only differ in signs, we argue that the impact on the accuracy of BNNs may be strongly related to the sign distribution of the network parameters in addition to the quantization error. To this end, Self-Distribution Binary Neural Network (SD-BNN) is proposed. First, we utilize Activation Self Distribution (ASD) to adaptively adjust the sign distribution of activations, thereby improving the sign differences of the outputs of the convolution. Second, we adjust the sign distribution of weights through Weight Self Distribution (WSD) and then fine-tune the sign distribution of the outputs of the convolution. Extensive experiments on the CIFAR-10 and ImageNet datasets with various network structures show that the proposed SD-BNN consistently outperforms state-of-the-art (SOTA) BNNs (e.g., 92.5% on CIFAR-10 and 66.5% on ImageNet with ResNet-18) with lower computational cost. Our code is available at https://github.com/pingxue-hfut/SD-BNN.
Similar content being viewed by others
References
Ding Y, Ma Z, Wen S, Xie J, Chang D, Si Z, Wu M, Ling H (2021) AP-CNN: Weakly supervised attention pyramid convolutional neural network for fine-grained visual classification. IEEE Trans Image Process 30:2826–2836. https://doi.org/10.1109/TIP.2021.3055617
Hu J, Shen L, Albanie S, Sun G, Wu E (2020) Squeeze-and-excitation networks. IEEE Trans Pattern Anal Mach Intell 42(8):2011–2023. https://doi.org/10.1109/TPAMI.2019.2913372
Wang Z, Lu J, Wu Z, Zhou J (2021) Learning efficient binarized object detectors with information compression. IEEE Trans Pattern Anal Mach Intell. https://doi.org/10.1109/TPAMI.2021.3050464
Chen C, Wei J, Peng C, Qin H (2021) Depth-quality-aware salient object detection. IEEE Trans Image Process 30:2350–2363. https://doi.org/10.1109/TIP.2021.3052069
Kalayeh MM, Shah M (2021) On symbiosis of attribute prediction and semantic segmentation. IEEE Trans Pattern Anal Mach Intell 43(5):1620–1635. https://doi.org/10.1109/TPAMI.2019.2956039
Tong Z, Xu P, Denoeux T (2021) Evidential fully convolutional network for semantic segmentation. Appl Intell 51(9):6376–6399. https://doi.org/10.1007/s10489-021-02327-0
Ding G, Zhang S, Jia Z, Zhong J, Han J (2021) Where to prune: Using LSTM to guide data-dependent soft pruning. IEEE Trans Image Process 30:293–304. https://doi.org/10.1109/TIP.2020.3035028
Singh P, Verma VK, Rai P, Namboodiri VP (2020) Acceleration of deep convolutional neural networks using adaptive filter pruning. IEEE J Sel Top Signal Process 14 (4):838–847. https://doi.org/10.1109/JSTSP.2020.2992390
Gao H, Wang Z, Cai L, Ji S (2021) Channelnets: Compact and efficient convolutional neural networks via channel-wise convolutions. IEEE Trans Pattern Anal Mach Intell 43(8):2570–2581. https://doi.org/10.1109/TPAMI.2020.2975796
Han K, Wang Y, Xu C, Xu C, Wu E, Tao D (2021) Learning versatile convolution filters for efficient visual recognition. IEEE Trans Pattern Anal Mach Intell. https://doi.org/10.1109/TPAMI.2021.3114368
Li X, Li S, Omar B, Wu F, Li X (2021) Reskd: Residual-guided knowledge distillation. IEEE Trans Image Process 30:4735–4746. https://doi.org/10.1109/TIP.2021.3066051
Gou J, Yu B, Maybank SJ, Tao D (2021) Knowledge distillation: A survey. Int J Comput Vis 129(6):1789–1819. https://doi.org/10.1007/s11263-021-01453-z
Hubara I, Courbariaux M, Soudry D, El-yaniv R, Bengio Y (2017) Quantized neural networks: training neural networks with low precision weights and activations. J Mach Learn Res 18:187–118730
Gong R, Liu X, Jiang S, Li T, Hu P, Lin J, Yu F, Yan J (2019) Differentiable soft quantization: Bridging full-precision and low-bit neural networks. In: 2019 IEEE/CVF international conference on computer vision, ICCV 2019, Seoul, Korea (South), October 27 - November 2, 2019, pp 4851–4860. https://doi.org/10.1109/ICCV.2019.00495
Tung F, Mori G (2020) Deep neural network compression by in-parallel pruning-quantization. IEEE Trans Pattern Anal Mach Intell 42(3):568–579. https://doi.org/10.1109/TPAMI.2018.2886192
Huang C, Liu P, Fang L (2021) MXQN: Mixed quantization for reducing bit-width of weights and activations in deep convolutional neural networks. Appl Intell 51 (7):4561–4574. https://doi.org/10.1007/s10489-020-02109-0
Fan Y, Pang W, Lu S (2021) HFPQ: Deep neural network compression by hardware-friendly pruning-quantization. Appl Intell 51(10):7016–7028. https://doi.org/10.1007/s10489-020-01968-x
Hubara I, Courbariaux M, Soudry D, El-yaniv R, Bengio Y (2016) Binarized neural networks. In: Annual conference on neural information processing systems 2016, December 5-10, 2016, Barcelona, Spain, pp 4107–4115
Qiao GC, Hu S, Chen TP, Rong LM, Ning N, Yu Q, Liu Y (2020) STBNN: Hardware-friendly spatio-temporal binary neural network with high pattern recognition accuracy. Neurocomputing 409:351–360. https://doi.org/10.1016/j.neucom.2020.06.084
Rastegari M, Ordonez V, Redmon J, Farhadi A (2016) Xnor-net: Imagenet classification using binary convolutional neural networks. In: Computer Vision - ECCV 2016 - 14th European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part IV, pp 525–542. https://doi.org/10.1007/978-3-319-46493-0_32
Bulat A, Tzimiropoulos G (2019) Xnor-net++: Improved binary neural networks. In: 30Th british machine vision conference 2019, BMVC 2019, Cardiff, UK, September 9-12, 2019, pp 62
Martínez B, Yang J, Bulat A, Tzimiropoulos G (2020) Training binary neural networks with real-to-binary convolutions. In: 8Th international conference on learning representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020
Qin H, Gong R, Liu X, Shen M, Wei Z, Yu F, Song J (2020) Forward and backward information retention for accurate binary neural networks. In: 2020 IEEE/CVF Conference on computer vision and pattern recognition, CVPR 2020, Seattle, WA, USA, June 13-19, 2020, pp 2247–2256. https://doi.org/10.1109/CVPR42600.2020.00232
Shen M, Liu X, Gong R, Han K (2020) Balanced binary neural networks with gated residual. In: 2020 IEEE International conference on acoustics, speech and signal processing, ICASSP 2020, Barcelona, Spain, May 4-8, 2020, pp 4197–4201. https://doi.org/10.1109/ICASSP40776.2020.9054599
Ding R, Chin T, Liu Z, Marculescu D (2019) Regularizing activation distribution for training binarized deep networks. In: IEEE Conference on computer vision and pattern recognition, CVPR 2019, Long Beach, CA, USA, June 16-20, 2019, pp 11408–11417. https://doi.org/10.1109/CVPR.2019.01167
Wang Z, Lu J, Zhou J (2021) Learning channel-wise interactions for binary convolutional neural networks. IEEE Trans Pattern Anal Mach Intell 43(10):3432–3445. https://doi.org/10.1109/TPAMI.2020.2988262
Torralba A, Fergus R, Freeman WT (2008) 80 million tiny images: a large data set for nonparametric object and scene recognition. IEEE Trans Pattern Anal Mach Intell 30(11):1958–1970. https://doi.org/10.1109/TPAMI.2008.128
Deng J, Dong W, Socher R, Li L, Li K, Fei-fei L (2009) Imagenet: A large-scale hierarchical image database. In: 2009 IEEE Computer society conference on computer vision and pattern recognition (CVPR 2009), 20-25 June 2009, Miami, Florida, USA, pp 248–255
Liu Z, Luo W, Wu B, Yang X, Liu W, Cheng K (2020) Bi-real net: Binarizing deep network towards real-network performance. Int J Comput Vis 128 (1):202–219. https://doi.org/10.1007/s11263-019-01227-8
Wu L, Lin X, Chen Z, Huang J, Liu H, Yang Y (2021) An efficient binary convolutional neural network with numerous skip connections for fog computing. IEEE Internet Things J 8(14):11357–11367. https://doi.org/10.1109/JIOT.2021.3052105
Kim H, Kim K, Kim J, Kim J (2020) Binaryduo: Reducing gradient mismatch in binary activation network by coupling binary activations. In: 8Th international conference on learning representations, ICLR 2020, Addis Ababa, ethiopia, April 26-30, 2020
Lin M, Ji R, Xu Z, Zhang B, Wang Y, Wu Y, Huang F, Lin C (2020) Rotated binary neural network. In: Advances in neural information processing systems 33: annual conference on neural information processing systems 2020, neurIPS 2020, December 6-12, 2020, virtual
Lan W, Lan L (2021) Compressing deep convolutional neural networks by stacking low-dimensional binary convolution filters. In: Thirty-fifth AAAI conference on artificial intelligence, AAAI 2021, virtual event, February 2-9, 2021, pp 8235–8242
Lin X, Zhao C, Pan W (2017) Towards accurate binary convolutional neural network. In: Advances in neural information processing systems 30: annual conference on neural information processing systems 2017, December 4-9, 2017, Long Beach, CA, USA, pp 345–353
Pouransari H, Tu Z, Tuzel O (2020) Least squares binary quantization of neural networks. In: 2020 IEEE/CVF Conference on computer vision and pattern recognition, CVPR workshops 2020, Seattle, WA, USA, June 14-19, 2020, pp 2986–2996
Liu C, Ding W, Hu Y, Xia X, Zhang B, Liu J, Doermann D (2020) Circulant binary convolutional networks for object recognition. IEEE J Sel Top Signal Process 14(4):884–893. https://doi.org/10.1109/JSTSP.2020.2969516
Bethge J, Bartz C, Yang H, Chen Y, Meinel C (2021) Meliusnet: An improved network architecture for binary neural networks. In: IEEE Winter conference on applications of computer vision, WACV 2021, Waikoloa, HI, USA, January 3-8, 2021, pp 1438–1447. https://doi.org/10.1109/WACV48630.2021.00148
Zhuang B, Shen C, Tan M, Liu L, Reid ID (2019) Structured binary neural networks for accurate image classification and semantic segmentation. In: IEEE Conference on computer vision and pattern recognition, CVPR 2019, Long Beach, CA, USA, June 16-20, 2019, pp 413–422. https://doi.org/10.1109/CVPR.2019.00050
Zhu S, Dong X, Su H (2019) Binary ensemble neural network: More bits per network or more networks per bit?. In: IEEE Conference on computer vision and pattern recognition, CVPR 2019, Long Beach, CA, USA, June 16-20, 2019, pp 4923–4932
Zagoruyko S, Komodakis N (2017) Paying more attention to attention: Improving the performance of convolutional neural networks via attention transfer. In: 5Th international conference on learning representations, ICLR 2017, Toulon, France, April 24-26, 2017, conference track proceedings
Bengio Y, Lėonard N, Courville AC (2013) Estimating or propagating gradients through stochastic neurons for conditional computation. CoRR arXiv:abs/1308.3432
Huang K, Ni B, Yang X (2019) Efficient quantization for neural networks with binary weights and low bitwidth activations. In: The thirty-third AAAI conference on artificial intelligence, AAAI 2019, Honolulu, Hawaii, USA, January 27 - February 1, 2019, pp 3854–3861. https://doi.org/10.1609/aaai.v33i01.33013854
Zhang D, Yang J, Ye D, Hua G (2018) Lq-nets: Learned quantization for highly accurate and compact deep neural networks. In: Computer Vision - ECCV 2018 - 15th European conference, Munich, Germany, September 8-14, 2018, Proceedings, Part VIII, pp 373–390. https://doi.org/10.1007/978-3-030-01237-3_23
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: 2016 IEEE Conference on computer vision and pattern recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, pp 770–778. https://doi.org/10.1109/CVPR.2016.90
Han S, Pool J, Tran J, Dally WJ (2015) Learning both weights and connections for efficient neural network. In: Advances in neural information processing systems 28: annual conference on neural information processing systems 2015, December 7-12, 2015, Montreal, Quebec, Canada, pp 1135–1143
Bulat A, Tzimiropoulos G, Kossaifi J, Pantic M (2019) Improved training of binary networks for human pose estimation and image recognition. CoRR arXiv:abs/1904.05868
He K, Zhang X, Ren S, Sun J (2015) Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In: 2015 IEEE International conference on computer vision, ICCV 2015, Santiago, Chile, December 7-13, 2015, pp 1026–1034. https://doi.org/10.1109/ICCV.2015.123
Acknowledgments
This work was supported in part by the Anhui Provincial Key Research and Development Program under Grant 202004a05020040, in part by the National Key Research and Development Program under Grant 2018YFC0604404, in part by Intelligent Network and New Energy Vehicle Special Project of Intelligent Manufacturing Institute of HFUT under Grant IMIWL2019003, and in part by Fundamental Research Funds for the Central Universities under Grant PA2021GDGP0061.
Author information
Authors and Affiliations
Corresponding authors
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Xue, P., Lu, Y., Chang, J. et al. Self-distribution binary neural networks. Appl Intell 52, 13870–13882 (2022). https://doi.org/10.1007/s10489-022-03348-z
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-022-03348-z