Abstract
There has been rising interest in reduced precision in the training of deep neural networks (DNNs) that is used from the single-precision (FP32) to different precision format (FP16, FP8, bfloat16) due to the rapid increase in model sizes, which require less representational space when stored in lower precision. However, training a DNN in reduced precision format (FP16, FP8, and bfloat16) is challenging, because the data format may be inadequate for representing the gradients during backpropagation. In this research paper, we compare the various novel approaches to train a DNN using the different reduced precision formats and explore the challenges that we get during the training of DNN in reduced precision. Besides, we also examine various layers in the neural network which make a significant reduction in the backpropagation and also observe that when should get sufficient precision.
Equal contribution
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Micikevicius, P., Narang, S., Alben, J., Diamos, G., Elsen, E., Garcia, D., et al. (2017). Mixed precision training. arXiv preprint arXiv:1710.03740.
Wang, N., Choi, J., Brand, D., Chen, C.-Y., & Gopalakrishnan, K. (2018). Training deep neural networks with 8-bit floating point numbers. In Advances in neural information processing systems (pp. 7675–7684).
Mellempudi, N., Srinivasan, S., Das, D., & Kaul, B. (2019). Mixed precision training with 8-bit floating point. arXiv preprint arXiv:1905.12334.
He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 770–778).
Wu, Y., Schuster, M., Chen, Z., Le, Q. V., Norouzi, M., Macherey, W., et al. (2016). Google’s neural machine translation system: Bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144.
Graves, A., Mohamed, A.-R., & Hinton, G. (2013). Speech recognition with deep recurrent neural networks. In 2013 IEEE international conference on acoustics, speech and signal processing (pp. 6645–6649). IEEE.
Hannun, A., Case, C., Casper, J., Catanzaro, B., Diamos, G., Elsen, E., et al. (2014). Deep speech: Scaling up end-to-end speech recognition. arXiv preprint arXiv:1412.5567.
Matt Post. (2018). A call for clarity in reporting bleu scores. arXiv preprint arXiv:1804.08771.
Ioffe, S., & Szegedy, C. (2015). Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv e-prints.
Training With Mixed Precision. https://docs.nvidia.com/deeplearning/sdk/mixed-precision-training/index.html.
Amodei, D., & Hernandez, D. AI and compute. https://openai.com/blog/ai-and-compute.
Neural Network Terminology. https://www.cs.toronto.edu/~lczhang/360/lec/w02/terms.html.
Krizhevsky, Sutskever, A., Hinton, I., & Geoffrey. (2012). ImageNet classification with deep convolutional neural networks. Neural Information Processing Systems,25. https://doi.org/10.1145/3065386.
He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 770–778).
Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., & Wojna, Z. (2016). Rethinking the inception architecture for computer vision. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 2818–2826).
Wu, Y., Schuster, M., Chen, Z., Le, Q.V., Norouzi, M., Macherey, W., et al. (2016). Google’s neural machine translation system: Bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144.
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, Ł., et al. (2017). Attention is all you need. In Advances in neural information processing systems (pp. 5998–6008).
Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., et al. (2015). Imagenet large scale visual recognition challenge. International Journal of Computer Vision, 115(3), 211–252.
Sennrich, R., Haddow, B., & Birch, A. (2016). Edinburgh neural machine translation systems for WMT 16. arXiv preprint arXiv:1606.02891.
Kalamkar, D., Mudigere, D., Mellempudi, N., Das, D., Banerjee, K., Avancha, S., et al. (2019). A study of bfloat16 for deep learning training. arXiv preprint arXiv:1905.12322.
Gupta, S., Agrawal, A., Gopalakrishnan, K., & Narayanan, P. (2015). Deep learning with limited numerical precision. In International conference on machine learning (pp. 1737–1746).
Micikevicius, P., Narang, S., Alben, J., Diamos, G., Elsen, E., Garcia, D., et al. (2017). Mixed precision training. arXiv preprint arXiv:1710.03740.
UrsKöster, Webb, T., Wang, X., Nassar, M., Bansal, A. K., Constable, W., et al. (2017). Flexpoint: An adaptive numerical format for efficient training of deep neural networks. In Advances in neural information processing systems (pp. 1742–1752).
Das, D., Mellempudi, N., Mudigere, D., Kalamkar, D., Avancha, S., Banerjee, K., et al. (2018). Mixed precision training of convolutional neural networks using integer operations. arXiv preprint arXiv:1802.00930.
Banner, R., Hubara, I., Hoffer, E., & Soudry, D. (2018). Scalable methods for 8-bit training of neural networks. In Advances in neural information processing systems (pp. 5145–5153).
Das, D., Mellempudi, N., Mudigere, D., Kalamkar, D., Avancha, S., Banerjee, K., et al. (2018). Mixed precision training of convolutional neural networks using integer operations. arXiv preprint arXiv:1802.00930.
Cambier, L., Bhiwandiwalla, A., Gong, T., Nekuii, M., Elibol, O. H., & Tang, H. (2020). Shifted and squeezed 8-bit floating point format for low-precision training of deep neural networks. arXiv preprint arXiv:2001.05674.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Gupta, R.R., Ranga, V. (2021). Comparative Study of Different Reduced Precision Techniques in Deep Neural Network. In: Tiwari, S., Suryani, E., Ng, A.K., Mishra, K.K., Singh, N. (eds) Proceedings of International Conference on Big Data, Machine Learning and their Applications. Lecture Notes in Networks and Systems, vol 150. Springer, Singapore. https://doi.org/10.1007/978-981-15-8377-3_11
Download citation
DOI: https://doi.org/10.1007/978-981-15-8377-3_11
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-15-8376-6
Online ISBN: 978-981-15-8377-3
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)