Skip to main content

End to End Deep Neural Network Classifier Design for Universal Sign Recognition

  • Conference paper
  • First Online:
Pattern Recognition Applications and Methods (ICPRAM 2020)

Part of the book series: Lecture Notes in Computer Science ((LNIP,volume 12594))

  • 258 Accesses

Abstract

Self-driving cars and Advanced Driver Assistance Systems rely heavily on Traffic Sign Recognition for safe maneuvering on the roads. But traffic signs can vary from one country to another, thereby necessitating multiple classifiers or a single universal classifier which can handle variations across countries. This paper reports our attempt at building a universal classifier. This classifier has to deal with large intra-class variations in the classes and also similarities among various difficult to distinguish traffic sign classes. This paper is an extension of our previous work in which we proposed a hierarchical classifier for traffic signs of a specific country. In hierarchical classification, dedicated classifiers are trained for classes which are more difficult to distinguish. Such similar classes are grouped together automatically by learning category hierarchy from the confusion matrix of a flat classifier (building block). In this paper, we use attention network for country independent classification. Here, CNN itself pays attention to regions in an image which are more discriminative and thus results in better classification for such problems. The aim here is to design a traffic sign recognition framework which can be used for multiple countries and be able to classify even the hard to distinguish classes by exploiting category hierarchy of traffic signs. The model is evaluated on traffic signs of seven countries namely Belgium, China, Croatia, Russia, Spain, Germany and Italy. The new building block architecture shows significant improvement of classification accuracy that is 97.7% as compared to building block architecture (VGG) used in our previous paper that is 95.1%.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Sengar, V., Rameshan, R.M., Ponkumar, S.: Hierarchical traffic sign recognition for autonomous driving. In: International Conference on Pattern Recognition Applications and Methods (ICPRAM), February 2020

    Google Scholar 

  2. Wang, Y., Morariu, V.I., Davis, L.S.: Learning a discriminative filter bank within a cnn for fine-grained recognition. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2018

    Google Scholar 

  3. Saha, S., Amit Kamran, S., Shihab Sabbir, A.: Total recall: understanding traffic signs using deep convolutional neural network. In: 2018 21st International Conference of Computer and Information Technology (ICCIT), pp. 1–6, December 2018

    Google Scholar 

  4. JuriÅi, F., Filkovi, I., Kalafati, Z.: Multiple-dataset traffic sign classification with one CNN. In: 2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR), pp. 614–618, November 2015

    Google Scholar 

  5. Rosario, G., Sonderman, T., Zhu, X.: Deep transfer learning for traffic sign recognition. In: 2018 IEEE International Conference on Information Reuse and Integration (IRI), pp. 178–185, July 2018

    Google Scholar 

  6. Yang, Y., Liu, S., Ma, W., Wang, Q., Liu, Z.: Efficient traffic-sign recognition with scale-aware CNN. CoRR, vol. abs/1805.12289 (2018)

    Google Scholar 

  7. Schroff, F., Kalenichenko, D., Philbin, J.: Facenet: a unified embedding for face recognition and clustering. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 815–823, June 2015

    Google Scholar 

  8. Liu, W., Wen, Y., Yu, Z., Yang, M.: Large-margin softmax loss for convolutional neural networks. In: Proceedings of the 33rd International Conference on International72 Conference on Machine Learning, ICML 2016, vol. 48, pp. 507–516. JMLR.org (2016)

    Google Scholar 

  9. Liu, W., Wen, Y., Yu, Z., Li, M., Raj, B., Song, L.: Sphereface: deep hypersphere embedding for face recognition. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6738–6746, July 2017

    Google Scholar 

  10. Qi, C., Su, F.: Contrastive-center loss for deep neural networks. In: in 2017 IEEE International Conference on Image Processing (ICIP), pp. 2851–2855, September 2017

    Google Scholar 

  11. Wang, H., et al.: Cosface: large margin cosine loss for deep face recognition. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5265–5274, June 2018

    Google Scholar 

  12. Wang, F., Cheng, J., Liu, W., Liu, H.: Additive margin softmax for face verification. IEEE Signal Process. Lett. 25, 926–930 (2018)

    Article  Google Scholar 

  13. Yang, H., Zhang, X., Yin, F., Liu, C.: Robust classification with convolutional prototype learning. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3474–3482, June 2018

    Google Scholar 

  14. Valev, K., Schumann, A., Sommer, L.W., Beyerer, J.: A systematic evaluation of recent deep learning architectures for fine-grained vehicle classification. CoRR, vol. abs/1806.02987 (2018)

    Google Scholar 

  15. Yang, Z., Luo, T., Wang, D., Hu, Z., Gao, J., Wang, L.: Learning to navigate for fine-grained classification. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018, Part XIV. LNCS, vol. 11218, pp. 438–454. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01264-9_26

    Chapter  Google Scholar 

  16. Wang, F., et al.: Residual attention network for image classification. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6450–6458, July 2017

    Google Scholar 

  17. Fu, J., Zheng, H., Mei, T.: Look closer to see better: recurrent attention convolutional neural network for fine-grained image recognition. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4476–4484, July 2017

    Google Scholar 

  18. BelgiumTS Dataset. https://btsd.ethz.ch/shareddata/

  19. Chinese Traffic Sign Database. http://www.nlpr.ia.ac.cn/pal/trafficdata/detection.html

  20. The MASTIF datasets. http://www.zemris.fer.hr/~ssegvic/mastif/datasets.shtml

  21. Russian Traffic sign recognition Dataset. http://graphics.cs.msu.ru/en/research/projects/imagerecognition/trafficsign

  22. Spanish Traffic Sign Dataset. https://daus-lab.github.io/spanish-traffic-sign-dataset/

  23. GTSRB Dataset. http://benchmark.ini.rub.de/?section=gtsrb&subsection=dataset

  24. DITS - Data set of Italian Traffic Signs. http://users.diag.uniroma1.it/bloisi/ds/dits.html

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Vartika Sengar .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Sengar, V., Rameshan, R.M., Ponkumar, S. (2020). End to End Deep Neural Network Classifier Design for Universal Sign Recognition. In: De Marsico, M., Sanniti di Baja, G., Fred, A. (eds) Pattern Recognition Applications and Methods. ICPRAM 2020. Lecture Notes in Computer Science(), vol 12594. Springer, Cham. https://doi.org/10.1007/978-3-030-66125-0_1

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-66125-0_1

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-66124-3

  • Online ISBN: 978-3-030-66125-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics