Abstract
In contrast to the machine, people have strong perception to determine the multiple objects easily in various environments under certain conditions. It is evident that the objects may exist in different orientations, scales and/or shapes, however it does not affect our ability to correct object recognition accuracy. The computer vision based object recognition is one of the most tedious tasks. So, objective is to design and development of the robust object recognition system to recognize the multiple objects effectively since different variations in the natural images are existed. In this paper, authors have presented comparative analysis and studies of some state of art object-recognition models for robot vision based problems, where the main goal is to recognize position, orientations and the identity of objects/parts, where parts have been taken from the industry. In the dynamic environment problem, the parts are recognized in a complex scenario. This paper presented based on the object representations using the recognition algorithms. Three steps are discussed and examined in details which are common to each and every category, namely, feature extraction, modeling, and matching. Test results confirmed that proposed model-based recognition in robot vision is fast and provides 93.00% accuracy. The proposed algorithm for object-recognition model has been compared with existing industrial part-recognition models and has, provided insights for progress toward future robot vision systems.
Similar content being viewed by others
Code availability
Authors confirm that the data supporting the findings of this study are available within the article.
References
Chen SY (2011) Kalman filter for robot vision: a survey. IEEE Trans Ind Electron 59(11):4409–4420
Zhang Zhaoxiang, Tan Tieniu, Huang Kaiqi, Wang Yunhong (2011) Three-dimensional deformable-model-based localization and recognition of road vehicles. IEEE Trans Image Process 21(1):1–13
Aghaie S, Khanmohammadi S, Moghadam-Fard H, Samadi F (2014) Adaptive vision-based control of robot manipulators using the interpolating polynomial. Trans Inst Meas Control 36(6):837–844
Bijalwan V, Semwal VB, Mandal TK (2021) Fusion of multi-sensor-based biomechanical gait analysis using vision and wearable sensor. IEEE Sens J 21(13):14213–14220
Abed SH, Al-Waisy AS, Mohammed HJ, Al-Fahdawi S (2021) A modern deep learning framework in robot vision for automated bean leaves diseases detection. Int J Intell Robotics Appl 5(2):235–251
Chang CK, Siagian C, Itti L (2010) Mobile robot vision navigation & localization using gist and saliency. In: 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 4147–4154. IEEE
Shaheed K, Mao A, Qureshi I, Kumar M, Hussain S, Ullah I, Zhang X (2022) DS-CNN: A pre-trained Xception model based on depth-wise separable convolutional neural network for finger vein recognition. Expert Syst Appl 191:116288
Khan A, Mineo C, Dobie G, Macleod C, Pierce G (2021) Vision guided robotic inspection for parts in manufacturing and remanufacturing industry. J Remanufacturing 11(1):49–70
Khan A, Li JP, Husain MA (2023) Power grid stability analysis using pipeline machine. Multimed Tools Appl, 1–25
Li Z, Yang C, Su CY, Deng J, Zhang W (2015) Vision-based model predictive control for steering of a nonholonomic mobile robot. IEEE Trans Control Syst Technol 24(2):553–564
Liao R, Yu S, An W, Huang Y (2020) A model-based gait recognition method with body pose and human prior knowledge. Pattern Recognit 98:107069
Sünderhauf N, Brock O, Scheirer W, Hadsell R, Fox D, Leitner J, Upcroft B, Abbeel P, Burgard W, Milford M et al (2018) The limits and potentials of deep learning for robotics. Int J Robot Res 37(4–5):405–420
Li X, Makihara Y, Xu C, Yagi Y, Yu S, Ren M (2020) End-to-end model-based gait recognition. In: Proceedings of the Asian conference on computer vision
Xu D, Han L, Tan M, Li YF (2009) Ceiling-based visual positioning for an indoor mobile robot with monocular vision. IEEE Trans Ind Electron 56(5):1617–1628
Han G, Xu Z, Zhu H, Ge Y, Peng J (2023) A two-stage model based on a complex-valued separate residual network for cross-domain IIoT devices identification. IEEE Trans Ind Inform
Shao L, Han J, Xu D, Shotton J (2013) Computer vision for RGB-D sensors: Kinect and its applications [special issue intro.]. IEEE Trans Cybern 43(5):1314–1317
Sheng B, Xiao F, Sha L, Sun L (2020) Deep spatial-temporal model based cross-scene action recognition using commodity WiFi. IEEE Internet Things J 7(4):3592–3601
Huang C, Huang X, Fang Y, Xu J, Qu Y, Zhai P, Fan L, Yin H, Xu Y, Li J (2020) Sample imbalance disease classification model based on association rule feature selection. Pattern Recognit Lett 133:280–286
Zou Y, Zhang Y, Yan J, Jiang X, Huang T, Fan H, Cui Z (2020) A robust license plate recognition model based on bi-LSTM. IEEE Access 8:211630–211641
Wan S, Goudos S (2020) Faster R-CNN for multi-class fruit detection using a robotic vision system. Comput Netw 168:107036
Poma Y, Melin P, González CI, Martinez GE (2020) Optimal recognition model based on convolutional neural networks and fuzzy gravitational search algorithm method. In: Hybrid intelligent systems in control, pattern recognition and medicine, pp. 71–81. Springer, ???
Kakani V, Nguyen VH, Kumar BP, Kim H, Pasupuleti VR (2020) A critical review on computer vision and artificial intelligence in food industry. J Agric Food Res 2:100033
Martinez P, Al-Hussein M, Ahmad R (2019) A scientometric analysis and critical review of computer vision applications for construction. Autom Constr 107:102947
Ashry S, Ogawa T, Gomaa W (2020) CHARM-deep: Continuous human activity recognition model based on deep neural network using IMU sensors of smartwatch. IEEE Sens J 20(15):8757–8770
Yang W, Tan RT, Wang S, Fang Y, Liu J (2020) Single image deraining: From model-based to data-driven and beyond. IEEE Trans Pattern Anal Mach Intell 43(11):4059–4077
Abbood WT, Abdullah OI, Khalid EA (2020) A real-time automated sorting of robotic vision system based on the interactive design approach. Int J Interact Des Manuf 14(1):201–209
Zeng R, Wen Y, Zhao W, Liu YJ (2020) View planning in robot active vision: A survey of systems, algorithms, and applications. Comput Vis Media 6(3):225–245
Shlezinger N, Whang J, Eldar YC, Dimakis AG (2020) Model-based deep learning. arXiv preprint arXiv:2012.08405
Li J, Yin J (2021) Deng L (2021) A robot vision navigation method using deep learning in edge computing environment. EURASIP J Adv Signal Process 1:1–20
Recht B, Roelofs R, Schmidt L, Shankar V (2018) Do cifar-10 classifiers generalize to cifar-10?. arXiv preprint arXiv:1806.00451
Xu B, Wang N, Chen T, Li M (2015) Empirical evaluation of rectified activations in convolutional network. arXiv preprint arXiv:1505.00853
Cohen G, Afshar S, Tapson J, Van Schaik A (2017) EMNIST: Extending MNIST to handwritten letters. In: 2017 international joint conference on neural networks (IJCNN), pp. 2921–2926. IEEE
Singh H, Swagatika S, Venkat RS, Saxena S (2019) Justification of STL-10 dataset using a competent CNN model trained on CIFAR-10. In: 2019 3rd International conference on Electronics, Communication and Aerospace Technology (ICECA), pp. 1254–1257. IEEE
Sermanet P, Chintala S, LeCun Y (2012) Convolutional neural networks applied to house numbers digit classification. In: Proceedings of the 21st international conference on pattern recognition (ICPR2012), pp. 3288–3291. IEEE
Greff K, Kaufman RL, Kabra R, Watters N, Burgess C, Zoran D, Matthey L, Botvinick M, Lerchner A (2019) Multi-object representation learning with iterative variational inference. In: International Conference on Machine Learning, 2424–2433. PMLR
Chen Y, Tu Z, Kang D, Bao L, Zhang Y, Zhe X, Chen R, Yuan J (2021) Model-based 3d hand reconstruction via self-supervised learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10451–10460
Khan A, Li JP, Khan J, Jasim KM, Alam R, Ahamed VN (2018) Complex environment fuzzy vision computing. In: 2018 15th International Computer Conference on Wavelet Active Media Technology and Information Processing (ICCWAMTIP), pp. 232–236. IEEE
Khan A, Li JP, Malik A, Yusuf Khan, M (2019) Vision-based inceptive integration for robotic control. In: Soft Computing and Signal Processing, pp. 95–105. Springer, ???
Theagarajan R, Bhanu B, Erpek T, Hue YK, Schwieterman R, Davaslioglu K, Shi Y, Sagduyu YE (2020) Integrating deep learning-based data driven and model-based approaches for inverse synthetic aperture radar target recognition. Opt Eng 59(5):051407
Acknowledgements
This work supported by the Department of Computer Application, Integral University, Lucknow, UP, India.
Funding
We certify that there is no conflict of interest with any financial organization regarding the material discussed in the manuscript.
Author information
Authors and Affiliations
Contributions
Asif Khan: Conceptualization, Methodology, Software, Validation, Investigation, Data Creation, Writing - Original Draft. Naushad Varish: Writing - Review and Editing
Corresponding author
Ethics declarations
Conflict of Interest
The authors have no relevant financial or non-financial interests to disclose.
Ethical standard
This work does not require ethics approval.
Consent to participate
This work does not require consent to participate, because it does not involve human subjects.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Khan, A., Varish, N., Pandey, D. et al. Model-based recognition in robot vision for monitoring built environments. Multimed Tools Appl (2024). https://doi.org/10.1007/s11042-024-19323-4
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11042-024-19323-4