Abstract
You Only Look Once (YOLO) is a popular problem-solving time visual perception framework that utilizes an individual autoencoder network to detect entity captured in an image. The key idea behind YOLO is to perform object detection in one forward pass of the network, rather than using a two-stage pipeline as in many other object detection frameworks. The framework functions by segmenting an illustration into a matrix of sections and allocating each unit the responsibility of detecting objects. The network then predicts the envelope and category probabilities for objects within each cell. YOLO uses ConvNet architecture for visual perception. The network takes an image as input and outputs a collection of envelope and category probabilities for objects within the visual representation. YOLO has proven to be effective in real-time object detection and has found extensive usage in various domains. However; it has some limitations, such as a lower accuracy compared to other frameworks and difficulty detecting smaller objects. Despite these limitations, YOLO remains a popular choice for real-time object detection due to its efficiency and speed.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 580–587
Wang Z, Qi Y, Wan J (2020) Dense object detection using point R-CNN. arXiv preprint arXiv:1912.07155
Girshick R (2015) Fast R-CNN. In: Proceedings of the IEEE international conference on computer vision, pp 1440–1448
Redmon J, Farhadi A (2016) You only look once: unified, real-time object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 779–788
Bochkovskiy A, Ros G (2020) YOLOv4: optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934
He K, Gkioxari G, Dollar P, Girshick R (2017) Mask R-CNN. In: Proceedings of the IEEE international conference on computer vision, pp 2961–2969
Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu CY, Berg AC (2016) SSD: single shot multibox detector. In: European conference on computer vision, pp 21–37
Redmon J, Farhadi A (2017) YOLO9000: better, faster, stronger. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7263–7271
Huang J, Rathod V, Sun C, Zhu M, Korattikara A, Fathi A, Fischer I, Wojna Z, Song Y, Guadarrama S, Murphy K (2017) Speed/accuracy trade-offs for modern convolutional object detectors. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7310–7317
Ren S, He K, Girshick R, Sun J (2015) Faster R-CNN: towards real-time object detection with region proposal networks. In: Advances in neural information processing systems, pp 91–99
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Ajith Babu, R.R., Dhushyanth, H.M., Hemanth, R., Naveen Kumar, M., Sushma, B.A., Loganayagi, B. (2023). Fast and Accurate YOLO Framework for Live Object Detection. In: Ranganathan, G., Papakostas, G.A., Rocha, Á. (eds) Inventive Communication and Computational Technologies. ICICCT 2023. Lecture Notes in Networks and Systems, vol 757. Springer, Singapore. https://doi.org/10.1007/978-981-99-5166-6_38
Download citation
DOI: https://doi.org/10.1007/978-981-99-5166-6_38
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-5165-9
Online ISBN: 978-981-99-5166-6
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)