Skip to main content
Log in

Model-based recognition in robot vision for monitoring built environments

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

In contrast to the machine, people have strong perception to determine the multiple objects easily in various environments under certain conditions. It is evident that the objects may exist in different orientations, scales and/or shapes, however it does not affect our ability to correct object recognition accuracy. The computer vision based object recognition is one of the most tedious tasks. So, objective is to design and development of the robust object recognition system to recognize the multiple objects effectively since different variations in the natural images are existed. In this paper, authors have presented comparative analysis and studies of some state of art object-recognition models for robot vision based problems, where the main goal is to recognize position, orientations and the identity of objects/parts, where parts have been taken from the industry. In the dynamic environment problem, the parts are recognized in a complex scenario. This paper presented based on the object representations using the recognition algorithms. Three steps are discussed and examined in details which are common to each and every category, namely, feature extraction, modeling, and matching. Test results confirmed that proposed model-based recognition in robot vision is fast and provides 93.00% accuracy. The proposed algorithm for object-recognition model has been compared with existing industrial part-recognition models and has, provided insights for progress toward future robot vision systems.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Algorithm 1
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13

Similar content being viewed by others

Code availability

Authors confirm that the data supporting the findings of this study are available within the article.

References

  1. Chen SY (2011) Kalman filter for robot vision: a survey. IEEE Trans Ind Electron 59(11):4409–4420

    Article  Google Scholar 

  2. Zhang Zhaoxiang, Tan Tieniu, Huang Kaiqi, Wang Yunhong (2011) Three-dimensional deformable-model-based localization and recognition of road vehicles. IEEE Trans Image Process 21(1):1–13

    Article  MathSciNet  Google Scholar 

  3. Aghaie S, Khanmohammadi S, Moghadam-Fard H, Samadi F (2014) Adaptive vision-based control of robot manipulators using the interpolating polynomial. Trans Inst Meas Control 36(6):837–844

    Article  Google Scholar 

  4. Bijalwan V, Semwal VB, Mandal TK (2021) Fusion of multi-sensor-based biomechanical gait analysis using vision and wearable sensor. IEEE Sens J 21(13):14213–14220

    Article  Google Scholar 

  5. Abed SH, Al-Waisy AS, Mohammed HJ, Al-Fahdawi S (2021) A modern deep learning framework in robot vision for automated bean leaves diseases detection. Int J Intell Robotics Appl 5(2):235–251

    Article  Google Scholar 

  6. Chang CK, Siagian C, Itti L (2010) Mobile robot vision navigation & localization using gist and saliency. In: 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 4147–4154. IEEE

  7. Shaheed K, Mao A, Qureshi I, Kumar M, Hussain S, Ullah I, Zhang X (2022) DS-CNN: A pre-trained Xception model based on depth-wise separable convolutional neural network for finger vein recognition. Expert Syst Appl 191:116288

    Article  Google Scholar 

  8. Khan A, Mineo C, Dobie G, Macleod C, Pierce G (2021) Vision guided robotic inspection for parts in manufacturing and remanufacturing industry. J Remanufacturing 11(1):49–70

    Article  Google Scholar 

  9. Khan A, Li JP, Husain MA (2023) Power grid stability analysis using pipeline machine. Multimed Tools Appl, 1–25

  10. Li Z, Yang C, Su CY, Deng J, Zhang W (2015) Vision-based model predictive control for steering of a nonholonomic mobile robot. IEEE Trans Control Syst Technol 24(2):553–564

    Google Scholar 

  11. Liao R, Yu S, An W, Huang Y (2020) A model-based gait recognition method with body pose and human prior knowledge. Pattern Recognit 98:107069

    Article  Google Scholar 

  12. Sünderhauf N, Brock O, Scheirer W, Hadsell R, Fox D, Leitner J, Upcroft B, Abbeel P, Burgard W, Milford M et al (2018) The limits and potentials of deep learning for robotics. Int J Robot Res 37(4–5):405–420

    Article  Google Scholar 

  13. Li X, Makihara Y, Xu C, Yagi Y, Yu S, Ren M (2020) End-to-end model-based gait recognition. In: Proceedings of the Asian conference on computer vision

  14. Xu D, Han L, Tan M, Li YF (2009) Ceiling-based visual positioning for an indoor mobile robot with monocular vision. IEEE Trans Ind Electron 56(5):1617–1628

    Article  Google Scholar 

  15. Han G, Xu Z, Zhu H, Ge Y, Peng J (2023) A two-stage model based on a complex-valued separate residual network for cross-domain IIoT devices identification. IEEE Trans Ind Inform

  16. Shao L, Han J, Xu D, Shotton J (2013) Computer vision for RGB-D sensors: Kinect and its applications [special issue intro.]. IEEE Trans Cybern 43(5):1314–1317

  17. Sheng B, Xiao F, Sha L, Sun L (2020) Deep spatial-temporal model based cross-scene action recognition using commodity WiFi. IEEE Internet Things J 7(4):3592–3601

    Article  Google Scholar 

  18. Huang C, Huang X, Fang Y, Xu J, Qu Y, Zhai P, Fan L, Yin H, Xu Y, Li J (2020) Sample imbalance disease classification model based on association rule feature selection. Pattern Recognit Lett 133:280–286

    Article  Google Scholar 

  19. Zou Y, Zhang Y, Yan J, Jiang X, Huang T, Fan H, Cui Z (2020) A robust license plate recognition model based on bi-LSTM. IEEE Access 8:211630–211641

    Article  Google Scholar 

  20. Wan S, Goudos S (2020) Faster R-CNN for multi-class fruit detection using a robotic vision system. Comput Netw 168:107036

    Article  Google Scholar 

  21. Poma Y, Melin P, González CI, Martinez GE (2020) Optimal recognition model based on convolutional neural networks and fuzzy gravitational search algorithm method. In: Hybrid intelligent systems in control, pattern recognition and medicine, pp. 71–81. Springer, ???

  22. Kakani V, Nguyen VH, Kumar BP, Kim H, Pasupuleti VR (2020) A critical review on computer vision and artificial intelligence in food industry. J Agric Food Res 2:100033

    Google Scholar 

  23. Martinez P, Al-Hussein M, Ahmad R (2019) A scientometric analysis and critical review of computer vision applications for construction. Autom Constr 107:102947

    Article  Google Scholar 

  24. Ashry S, Ogawa T, Gomaa W (2020) CHARM-deep: Continuous human activity recognition model based on deep neural network using IMU sensors of smartwatch. IEEE Sens J 20(15):8757–8770

    Article  Google Scholar 

  25. Yang W, Tan RT, Wang S, Fang Y, Liu J (2020) Single image deraining: From model-based to data-driven and beyond. IEEE Trans Pattern Anal Mach Intell 43(11):4059–4077

    Article  Google Scholar 

  26. Abbood WT, Abdullah OI, Khalid EA (2020) A real-time automated sorting of robotic vision system based on the interactive design approach. Int J Interact Des Manuf 14(1):201–209

    Article  Google Scholar 

  27. Zeng R, Wen Y, Zhao W, Liu YJ (2020) View planning in robot active vision: A survey of systems, algorithms, and applications. Comput Vis Media 6(3):225–245

    Article  Google Scholar 

  28. Shlezinger N, Whang J, Eldar YC, Dimakis AG (2020) Model-based deep learning. arXiv preprint arXiv:2012.08405

  29. Li J, Yin J (2021) Deng L (2021) A robot vision navigation method using deep learning in edge computing environment. EURASIP J Adv Signal Process 1:1–20

    Google Scholar 

  30. Recht B, Roelofs R, Schmidt L, Shankar V (2018) Do cifar-10 classifiers generalize to cifar-10?. arXiv preprint arXiv:1806.00451

  31. Xu B, Wang N, Chen T, Li M (2015) Empirical evaluation of rectified activations in convolutional network. arXiv preprint arXiv:1505.00853

  32. Cohen G, Afshar S, Tapson J, Van Schaik A (2017) EMNIST: Extending MNIST to handwritten letters. In: 2017 international joint conference on neural networks (IJCNN), pp. 2921–2926. IEEE

  33. Singh H, Swagatika S, Venkat RS, Saxena S (2019) Justification of STL-10 dataset using a competent CNN model trained on CIFAR-10. In: 2019 3rd International conference on Electronics, Communication and Aerospace Technology (ICECA), pp. 1254–1257. IEEE

  34. Sermanet P, Chintala S, LeCun Y (2012) Convolutional neural networks applied to house numbers digit classification. In: Proceedings of the 21st international conference on pattern recognition (ICPR2012), pp. 3288–3291. IEEE

  35. Greff K, Kaufman RL, Kabra R, Watters N, Burgess C, Zoran D, Matthey L, Botvinick M, Lerchner A (2019) Multi-object representation learning with iterative variational inference. In: International Conference on Machine Learning, 2424–2433. PMLR

  36. Chen Y, Tu Z, Kang D, Bao L, Zhang Y, Zhe X, Chen R, Yuan J (2021) Model-based 3d hand reconstruction via self-supervised learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10451–10460

  37. Khan A, Li JP, Khan J, Jasim KM, Alam R, Ahamed VN (2018) Complex environment fuzzy vision computing. In: 2018 15th International Computer Conference on Wavelet Active Media Technology and Information Processing (ICCWAMTIP), pp. 232–236. IEEE

  38. Khan A, Li JP, Malik A, Yusuf Khan, M (2019) Vision-based inceptive integration for robotic control. In: Soft Computing and Signal Processing, pp. 95–105. Springer, ???

  39. Theagarajan R, Bhanu B, Erpek T, Hue YK, Schwieterman R, Davaslioglu K, Shi Y, Sagduyu YE (2020) Integrating deep learning-based data driven and model-based approaches for inverse synthetic aperture radar target recognition. Opt Eng 59(5):051407

    Article  Google Scholar 

Download references

Acknowledgements

This work supported by the Department of Computer Application, Integral University, Lucknow, UP, India.

Funding

We certify that there is no conflict of interest with any financial organization regarding the material discussed in the manuscript.

Author information

Authors and Affiliations

Authors

Contributions

Asif Khan: Conceptualization, Methodology, Software, Validation, Investigation, Data Creation, Writing - Original Draft. Naushad Varish: Writing - Review and Editing

Corresponding author

Correspondence to Naushad Varish.

Ethics declarations

Conflict of Interest

The authors have no relevant financial or non-financial interests to disclose.

Ethical standard

This work does not require ethics approval.

Consent to participate

This work does not require consent to participate, because it does not involve human subjects.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Khan, A., Varish, N., Pandey, D. et al. Model-based recognition in robot vision for monitoring built environments. Multimed Tools Appl (2024). https://doi.org/10.1007/s11042-024-19323-4

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s11042-024-19323-4

Keywords

Navigation