Skip to main content
Log in

Automatic identification of commodity label images using lightweight attention network

  • Original Article
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

Recent research has raised interest in applying image classification techniques to automatically identify the commodity label images for the business automation of retail enterprises. These techniques can help enterprises improve their service efficiency and realize digital transformation. In this work, we developed a lightweight attention network with a small size and comparable precision, namely MS-DenseNet, to identify the commodity label images. MS-DenseNet is based on the recent well-known DenseNet architecture, where we replaced the regular planner convolution in dense blocks with depthwise separable convolution to compress the model size. Further, the SE modules were incorporated in the proposed network to highlight the useful feature channels while suppressing the useless feature channels, which made good use of interdependencies between channels and realized the maximum reuse of inter-channel relations. Besides, the two-stage progressive strategy was adopted in model training. The proposed procedure achieved significant performance gain with an average accuracy of 97.60% on the identification of commodity label images task. Also, it realized a 94.90% average accuracy on public datasets. The experimental findings present a substantial performance compared with existing methods and also demonstrate the effectiveness and extensibility of the proposed procedure. Our code is available at https://github.com/xtu502/Automatic-identification-of-commodity-label-images.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12

Similar content being viewed by others

References

  1. Wang Y, Wang Z (2019) A survey of recent work on fine-grained image classification techniques. J Vis Commun Image Represent 59:210–214

    Article  Google Scholar 

  2. Gomes SL et al (2017) Embedded real-time speed limit sign recognition using image processing and machine learning techniques. Neural Comput Appl 28(1):573–584

    Article  Google Scholar 

  3. Li CH et al (2013) Algorithm research of two-dimensional size measurement on parts based on machine vision. Adv Mater Res 694:1945–1948

    Article  Google Scholar 

  4. Gökmen V, Sügüt I (2007) A non-contact computer vision based analysis of color in foods. Int J Food Eng 3:1–13

    Article  Google Scholar 

  5. Ciocca G, Napoletano P, Schettini R (2018) CNN-based features for retrieval and classification of food images. Comput Vis Image Underst 176:70–77

    Article  Google Scholar 

  6. Geetharamani G, Pandian A (2019) Identification of plant leaf diseases using a nine-layer deep convolutional neural network. Comput Electric Eng 76:323–338

    Article  Google Scholar 

  7. Priyadharshini RA et al (2019) Maize leaf disease classification using deep convolutional neural networks. Neural Comput Appl 31(12):8887–8895

    Article  Google Scholar 

  8. Miki Y et al (2017) Classification of teeth in cone-beam CT using deep convolutional neural network. Comput Biol Med 80:24–29

    Article  Google Scholar 

  9. Mahbod A et al (2020) (2020) Transfer learning using a multi-scale and multi-network ensemble for skin lesion classification. Comput Methods Programs Biomed 193:105475

    Article  Google Scholar 

  10. Mondal S, Bours P (2017) A study on continuous authentication using a combination of keystroke and mouse biometrics. Neurocomputing 230:1–22

    Article  Google Scholar 

  11. Duan Y et al (2017) SAR Image segmentation based on convolutional-wavelet neural network and markov random field. Pattern Recogn 64:255–267

    Article  Google Scholar 

  12. Precup R-E et al (2020) Evolving fuzzy models for prosthetic hand myoelectric-based control. IEEE Trans Instrum Meas 69:4625–4636

    Article  Google Scholar 

  13. Li X et al (2020) Fault diagnostics between different type of components: a transfer learning approach. Appl Soft Comput 86:105950

    Article  Google Scholar 

  14. Ma J et al. (2019) Machine learning based cross-border E-commerce commodity customs product name recognition algorithm. In: Pacific Rim International Conference on Artificial Intelligence. Springer, Cham, pp 247–256

  15. Ahmed MU et al (2019) A machine learning approach to classify pedestrians’ events based on IMU and GPS. Int J Artif Intell 17(2):154–167

    Google Scholar 

  16. Zhang T, Chen E (2019) Product recognition algorithm based on HOG and bag of words model. In: 2019 8th International Symposium on Next Generation Electronics (ISNE), pp 1–3. IEEE

  17. Kussul N et al (2017) Deep learning classification of land cover and crop types using remote sensing data. IEEE Geosci Remote Sens Lett 14(5):778–782

    Article  Google Scholar 

  18. Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. Adv Neural Inf Process Syst 25:1097–1105

    Google Scholar 

  19. Ji Y et al (2019) Graph model-based salient object detection using objectness and multiple saliency cues. Neurocomputing 323:188–202

    Article  Google Scholar 

  20. Zou X et al (2020) Multi-task cascade deep convolutional neural networks for large-scale commodity recognition. Neural Comput Appl 32(10):5633–5647

    Article  Google Scholar 

  21. Chen C, Yang R, Wang C (2017) Research and realization of commodity image retrieval system based on deep learning. In: International Symposium on Parallel Architecture, Algorithm and Programming, vol 729, pp. 376–385. Springer, Singapore

  22. Cao Z, Shaomin Mu, Dong M (2020) Two-attribute e-commerce image classification based on a convolutional neural network. Vis Comput 36:1619–1634

    Article  Google Scholar 

  23. Huang G et al. (2017) Densely connected convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp 2261–2269

  24. Howard AG et al. (2017) Mobilenets: efficient convolutional neural networks for mobile vision applications, arXiv preprint arXiv:1704.04861. pp 1–9

  25. Sifre L, Mallat S (2014) Rigid-motion scattering for image classification. Ph. D. thesis

  26. Kaiser L, Gomez AN, Chollet F (2017) Depthwise separable convolutions for neural machine translation, arXiv preprint arXiv:1706.03059. pp 1–10.

  27. Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7132–7141

  28. Glorot X, Bordes A, Bengio Y (2011) Deep sparse rectifier neural networks. Proceedings of the fourteenth international conference on artificial intelligence and statistics, pp 315–323

  29. Zoph B et al. (2018) Learning transferable architectures for scalable image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8697–8710

  30. Zhang X et al. (2018) Shufflenet: an extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6848–6856

  31. Tan M, Le QV (2019) Efficientnet: rethinking model scaling for convolutional neural networks, arXiv preprint arXiv:1905.11946, pp 6105–6114

  32. Lin T-Y et al. (2017) Focal loss for dense object detection. In: Proceedings of the IEEE international conference on computer vision, pp 2980–2988

  33. Kingma DP, Ba J (2014) Adam: a method for stochastic optimization, arXiv preprint arXiv:1412.6980, pp 1–15

  34. Ghazi MM, Yanikoglu B, Aptoula E (2017) Plant identification using deep neural networks via optimization of transfer learning parameters. Neurocomputing 235:228–235

    Article  Google Scholar 

  35. Sandler M et al. (2018) Mobilenetv2: inverted residuals and linear bottlenecks. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 4510–4520

  36. Wang RJ, Li X, Ling CX (2018) Pelee: a real-time object detection system on mobile devices. Advances in Neural Information Processing Systems, pp 1–10

  37. Pereira DG, Afonso A, Medeiros FM (2015) Overview of Friedman’s test and post-hoc analysis. Commun Statis Simul Comput 44(10):2636–2653

    Article  MathSciNet  Google Scholar 

  38. Irigaray D et al. (2019) Accelerating the calculation of Friedman test tables on many-core processors. In: Latin American High Performance Computing Conference. Springer, Cham, pp 122–135

  39. Lazebnik S, Schmid C, Ponce J (2006) Beyond bags of features: spatial pyramid matching for recognizing natural scene categories. In: 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06), vol 2. IEEE, pp 2169–2178

  40. Li F-F, Fergus R, Perona P (2004) Learning generative visual models from few training examples: an incremental bayesian approach tested on 101 object categories. In: 2004 conference on computer vision and pattern recognition workshop. IEEE, pp 178–178

  41. Chua T-S et al. (2009) NUS-WIDE: a real-world web image database from National University of Singapore. In: Proceedings of the ACM international conference on image and video retrieval, pp 1–9

  42. Yang J et al. (2009) Linear spatial pyramid matching using sparse coding for image classification. In: 2009 IEEE Conference on computer vision and pattern recognition. IEEE, pp 1794–1801

  43. Rasiwasia N, Vasconcelos N (2008) Scene classification with low-dimensional semantic spaces and weak supervision. In: 2008 IEEE Conference on Computer Vision and Pattern Recognition. IEEE, pp 1–6

  44. Li L et al (2010) Object bank: a high-level image representation for scene classification & semantic feature sparsification. Adv Neural Inf Process Syst 23:1378–1386

    Google Scholar 

  45. Sun Y et al (2019) Image classification base on PCA of multi-view deep representation. J Vis Commun Image Repres 62(2019):253–258

    Article  Google Scholar 

Download references

Acknowledgments

The research is funded by the National Natural Science Foundation of China (61672439) and the Fundamental Research Funds for the Central Universities (20720181004). The authors also thank editors and all unknown reviewers for constructive suggestions.

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Shuangyuan Yang or Defu Zhang.

Ethics declarations

Conflicts of interest

The authors declare no conflicts of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Chen, J., Zeb, A., Yang, S. et al. Automatic identification of commodity label images using lightweight attention network. Neural Comput & Applic 33, 14413–14428 (2021). https://doi.org/10.1007/s00521-021-06081-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-021-06081-9

Keywords

Navigation