Skip to main content

Deep Learning Frameworks for Convolutional Neural Networks—A Benchmark Test

  • Conference paper
  • First Online:
Cyber-physical Systems and Digital Twins (REV2019 2019)

Part of the book series: Lecture Notes in Networks and Systems ((LNNS,volume 80))

Abstract

Deep neural networks are state of the art for many machine-learning problems. The architecture of deep neural networks is inspired by the hierarchical structure of the brain. Deep neural networks feature a hierarchical, layer-wise arrangement of nonlinear activation functions (neurons) fed by inputs scaled by linear weights (synapses). Deep Learning frameworks simplify model development and training by providing high-level primitives for complex and error-prone mathematical transformations, like gradient descent, back-propagation, and inference. The main goal of this study is to compare the performance of two trending frameworks, Caffe and TensorFlow for Deep Machine Learning. As benchmark for the comparison with other approaches in deep learning was a well-analyzed database for handwritten cipher classification chosen. These two frameworks (Caffe and TensorFlow) were chosen out of nine different frameworks, after checking which ones better fits the proposed selection criteria. Caffe and TensorFlow frameworks were selected after nine different frameworks were analyzed using the following selection criteria: programming language, functionality level, algorithms and network architecture, pretrained models and community activity. As the performance is heavily affected by the number of parameters in the networks, four well-known convolutional neural network (CNN) architectures (LeNet, AlexNet, VGG19 and GoogLeNet) were trained and tested using both frameworks. Using architectures with different deepness will allow investigating how the number of hidden layers improves the accuracy of the CNN. As a CNN requires high computational effort, two computers equipped with different NVIDIA GPU were used to ease the effort and they were used to investigate how the hardware improves the performance of the CNN and if it is worthy to invest on it. As the CNN are widely used for image classification, it was defined that the used architecture was going to be used for classification of handwritten numbers, because this example of classification is very well analyzed and can serve as a benchmark for comparison. Due to this and considering that the training of these networks requires a huge amount of data, MNIST database was set for training and testing them. All the architectures were adapted to the3 MNIST database and they were developed for Caffe and TensorFlow frameworks and analyzed on the named architectures. The biggest differences on our hardware we have got in training the VVG19 and the GoogLeNet architectures in TensorFlow and Caffe.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 229.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 299.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Chetlur, S., et al.: cuDNN: Efficient Primitives for Deep Learning. CoRR abs/1410.0759 (2014)

    Google Scholar 

  2. Abadi, M., et al.: TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems. https://www.tensorflow.org/about/bib, 9 Nov 2015 (14 June 2018)

  3. Collobert, R., Kavukcuoglu, K., Farabet, C.: Torch7: A Matlab-Like Environment for Machine Learning (2011)

    Google Scholar 

  4. Agarwal, A., Akchurin, E., Basoglu, C.: An Introduction to Computational Networks and the Computational Network Toolkit. https://www.microsoft.com/en-us/research/publication/an-introduction-to-computational-networks-and-the-computational-network-toolkit/, 1 Oct 2014 (20 June 2018)

  5. Jia, Y., et al.: Caffe: Convolutional Architecture for Fast Feature Embedding (2014). CoRR. [On-line] abs/1408.5093. Available: https://arxiv.org/abs/1408.5093, 20 June 2018

  6. Chen, T., et al.: MXNet: A Flexible and Efficient Machine Learning Library for Heterogeneous Distributed Systems (2015). CoRR. [On-line] abs/1512.01274. Available: https://arxiv.org/abs/1512.01274, 21 June 2018

  7. Tokui, S., Oono, K., Hido, S.: Chainer: A Next-Generation Open Source Framework for Deep Learning (2015)

    Google Scholar 

  8. Chollet, F.: Keras (2015). https://github.com/fchollet/keras, 21 June 2018

  9. Gibson, A., Nicholson, C., Patterson J.: Deeplearning4j (2017). https://deeplearning4j.org/, 22 June 2018

  10. Al-Rfou, R., Alain, G., Almahairi, A.: Theano: A Python Framework For Fast Computation Of Mathematical Expressions (2016). https://arxiv.org/abs/1605.02688, 22 June 2018

  11. Lopez de Guereña, X.: Deep Learning Frameworks for Convolutional Neural Networks. Master Thesis at CUAS. Unpublished (2018)

    Google Scholar 

  12. van Rossum, G.: Python tutorial (1995). https://www.python.org/, 23 June 2018

  13. Lecun, Y., et al.: The MNIST Dataset of Handwritten Digits (Images) (1999). http://yann.lecun.com/exdb/mnist/, 23 June 2018

  14. Lecun, Y., et al.: Gradient-based learning applied to document recognition. In: Proceedings of the IEEE, vol. 86, no. 11, pp. 2278–2324, Nov 1998. https://doi.org/10.1109/5.726791

    Article  Google Scholar 

  15. Krizhevsky, A., Sutskever, I., Hinton, G.: ImageNet classification with deep convolutional neural networks. In: NIPS’12 Proceedings of the 25th International Conference on Neural Information Processing Systems, vol. 1, pp 1097–1105, Lake Tahoe, Nevada, 3 Dec 2012

    Google Scholar 

  16. Russakovsky, O., et al.: ImageNet large scale visual recognition challenge. IJCV (2015)

    Google Scholar 

  17. Simonyan, K., Zisserman, A.: Very Deep Convolutional Networks for Large-scale Image Recognition (2014). CoRR. [On-line] abs/1409.1556. https://arxiv.org/abs/1409.1556, 24 June 2018

  18. Szegedy, C., et al.: Going deeper with convolutions. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, pp. 1–9 (2015). https://doi.org/10.1109/cvpr.2015.7298594

  19. Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. Frechen, MITP (2018)

    MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Andreas Pester .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Pester, A., Madritsch, C., Klinger, T., de Guereña, X.L. (2020). Deep Learning Frameworks for Convolutional Neural Networks—A Benchmark Test. In: Auer, M., Ram B., K. (eds) Cyber-physical Systems and Digital Twins. REV2019 2019. Lecture Notes in Networks and Systems, vol 80. Springer, Cham. https://doi.org/10.1007/978-3-030-23162-0_74

Download citation

Publish with us

Policies and ethics