Skip to main content
Log in

An object detection method for the work of an unmanned sweeper in a noisy environment on an improved YOLO algorithm

  • Original Paper
  • Published:
Signal, Image and Video Processing Aims and scope Submit manuscript

Abstract

Efficient and accurate object detection is crucial for the widespread use of low-cost unmanned sweepers. This paper focuses on the low-cost sweeper in practical working scenarios and proposes a traffic participant detection method based on an enhanced YOLO-v5 model. To train the model on noise knowledge, three types of noise are added to the data set in the offline phase, according to the vibration response of the mathematical model, and the impact of the low-cost camera. The loss function was optimized to balance detection accuracy and real-time performance while focusing on traffic participant detection using YOLO-v5. CTDS and BFSA modules were proposed based on the attention mechanism to enhance the YOLO-v5 model. Comparative experiments demonstrated the effectiveness of the proposed method, with the enhanced YOLO-v5 model achieving a 4.5% higher mean average precision than the traditional YOLO-v5 network. Moreover, the proposed method can process images at a frame per second of 89 while ensuring real-time performance, meeting the object detection requirements of actual sweeper.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8

Similar content being viewed by others

Data availability

Due to the nature of this study and in order to protect the privacy of study participants, participants of this study did not agree for their data to be shared publicly, so supporting data are not available.

References

  1. Yu, F., Chen, H., Wang, X., Xian, W., Chen, Y., Liu, F., Madhavan, V., Darrell, T.: BDD100K: A diverse driving dataset for heterogeneous multitask learning. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2633–2642 (2020). https://doi.org/10.1109/CVPR42600.2020.00271

  2. Papageorgiou, C.P., Oren, M., Poggio, T.: A general framework for object detection. In:1998 6th International Conference on Computer Vision(ICCV), pp. 555–562 (1998). https://doi.org/10.1109/ICCV.1998.710772

  3. Matsumoto, M.: SVM-based parameter setting of self-quotient -filter and its application to noise robust human detection. In: 2011 3rd International Conference on Agents and Artificial Intelligence(ICAART), pp. 290–295 (2011)

  4. Moubtahij, R.E., Merad, D., Damoisaux, J.L., Drap, P.: Mine detection based on adaboost and polynomial image decomposition. In: 2017 19th International Conference on Image Analysis and Processing (ICIAP), pp. 660–670 (2017). https://doi.org/10.1007/978-3-319-68560-1_59

  5. Ali, A., Olaleye, O.G., Bayoumi, M.: Fast region-based DPM object detection for autonomous vehicles. In: 2016 59th IEEE International Midwest Symposium on Circuits and Systems (MWSCAS), pp. 691–694 (2016)

  6. Zou, Z., Shi, Z., Guo, Y., Ye, J.: Object detection in 20 years: A survey. P IEEE, pp. 257–276 (2019). https://doi.org/10.48550/arXiv.1905.05055

  7. Ruxin, W., Dacheng, T.: Training very deep CNNs for general non-blind deconvolution. IEEE T Image Process. 27(6), 2897–2910 (2018). https://doi.org/10.1109/TIP.2018.2815084

    Article  MathSciNet  MATH  Google Scholar 

  8. Wang, K., Zhou, W.: Pedestrian and cyclist detection based on deep neural network fast R-CNN. Int. J. Adv. Robot Syst. 16(78), 96 (2019)

    Google Scholar 

  9. Lai, K.C., Zhao, J., Liu, D.J., Huang, X.N., Wang, L.: Research on pedestrian detection using optimized mask R-CNN algorithm in low-light road environment. J. Phys Conf. Ser. 1777(1), 12057 (2021). https://doi.org/10.1088/1742-6596/1777/1/012057

    Article  Google Scholar 

  10. Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C., Berg, A.C.: SSD: Single Shot MultiBox Detector. In: 2016 14th European Conference Computer Vision(ECCV), pp. 21–37,(2016).https://doi.org/10.1007/978-3-319-46448-0_2

  11. Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 779–788,(2016).https://doi.org/10.1109/CVPR.2016.91

  12. Nakahara, H., Yonekawa, H., Fujii, T., Sato, S.: A Lightweight YOLOv2: A binarized CNN with a parallel support vector regression for an FPGA. In: 2018 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays (FPGA), pp. 31–40,(2018).https://doi.org/10.1145/3174243.3174266

  13. Redmon, J., Farhadi, A.: YOLO9000: Better, faster, stronger. In: 2017 30th IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6517–6525,(2017).https://doi.org/10.1109/CVPR.2017.690

  14. Qu, H.Q., Yuan, T.Y., Sheng, Z.Y., Zhang, Y.: A pedestrian detection method based on YOLOv3 model and image enhanced by Retinex. In: 2018 11th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), (2018).

  15. Bochkovskiy, A., Wang, C.Y., Liao, H.: YOLOv4: Optimal speed and accuracy of object detection, (2020). https://doi.org/10.48550/arXiv.2004.10934

  16. Mekhalfi, ML, Nicolo, C, Bazi, Y, Rahhal, M, Alsharif, NA, Maghayreh, EA (2022): Contrasting YOLOv5, transformer, and efficient Det detectors for crop circle detection in desert. Geosci. Remotes. 19: 78 96

  17. Ge, Z., Liu, S., Wang, F., Li, Z., Sun, J.: YOLOX: Exceeding YOLO series in 2021,(2021). https://doi.org/10.48550/arXiv.2107.08430

  18. Wu, D., Liao, M.W., Zhang, W.T., Wang, X.G., Bai, X., Cheng, W.Q., Liu, W.Y.: YOLOP: You only look once for panoptic driving perception. 19(6), 13 (2022). arXiv:2108.11250

  19. Wang, C.Y., Bochkovskiy, A., Liao, H.: YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors, (2022). https://doi.org/10.48550/arXiv.2207.02696

  20. Liu, S., Qi, L., Qin, H., Shi, J., Jia, J.: Path aggregation network for instance segmentation. In: 2018 31st IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8759–8768,(2018). https://doi.org/10.1109/CVPR.2018.00913

  21. Woo, S.H., Park, J., Lee, J.Y., Kweon, I.S.: CBAM: Convolutional block attention module. In:2018 15th European Conference on Computer Vision (ECCV), pp. 3–19,(2018).https://doi.org/10.1007/978-3-030-01234-2_1

  22. Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Houlsby, N.: An image is worth 16x16 words: Transformers for image recognition at scale, (2020). https://doi.org/10.48550/arXiv.2010.11929

  23. Huang, G.J., Zhang, Y.L., Ou, J.Y.: Transfer remaining useful life estimation of bearing using depth-wise separable convolution recurrent network. MEASUREMENT. 176, (2021). https://doi.org/10.1016/j.measurement.2021.109090

  24. Yung, N., Wong, W.K., Juwono, F.H., Sim, Z.A., IEEE: Safety helmet detection using deep learning: implementation and comparative study using YOLOv5, YOLOv6, and YOLOv7. In: 2022 International Conference on Green Energy, Computing and Sustainable Technology (GECOST), pp. 164–170,(2022). https://doi.org/10.1109/GECOST55694.2022.10010490

  25. Ramachandran, P., Parmar, N., Vaswani, A., Bello, I., Levskaya, A., Shlens, J.: Stand-alone self-attention in vision models. In: 2019 33rd Conference on Neural Information Processing Systems (NeurIPS), (2019)

  26. Dumoulin, V., Visin, F.: A guide to convolution arithmetic for deep learning, (2016). https://doi.org/10.48550/arXiv.1603.07285

  27. Cho, Y.J.: Weighted intersection over union (wIoU): A new evaluation metric for image segmentation, (2021). https://doi.org/10.48550/arXiv.2107.09858

  28. Mafi, M., Izquierdo, W., Cabrerizo, M., Barreto, A., Andrian, J., Rishe, N.D., Adjouadi, M.: Survey on mixed impulse and Gaussian denoising filters. IET Image Process. 14(16), 4027–4038 (2020). https://doi.org/10.1049/iet-ipr.2018.6335

    Article  Google Scholar 

  29. Wenhao, W., Shangbing, G., Jingbo, Z., Vunyang, Y.: Research on denoising algorithm for salt and pepper noise. J. Data Acquis. Process. 30(5), 1091–1098 (2015). https://doi.org/10.16337/j.1004-9037.2015.05.021

    Article  Google Scholar 

Download references

Funding

The authors did not receive support from any organization for the submitted work.

Author information

Authors and Affiliations

Authors

Contributions

Material preparation, data collection, analysis and modification, experiment were performed by JH, the first draft of the manuscript was written by JH, All authors read and approved the final manuscript.

Corresponding author

Correspondence to Baijun Shi.

Ethics declarations

Conflict of interest

The authors declare that there is no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Huo, J., Shi, B. & Zhang, Y. An object detection method for the work of an unmanned sweeper in a noisy environment on an improved YOLO algorithm. SIViP 17, 4219–4227 (2023). https://doi.org/10.1007/s11760-023-02654-4

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11760-023-02654-4

Keywords

Navigation