Skip to main content
Log in

DBH-YOLO: a surgical instrument detection method based on feature separation in laparoscopic surgery

  • Original Article
  • Published:
International Journal of Computer Assisted Radiology and Surgery Aims and scope Submit manuscript

Abstract

Purpose

Accurately locating and analysing surgical instruments in laparoscopic surgical videos can assist doctors in postoperative quality assessment. This can provide patients with more scientific and rational solutions for healing surgical complications. Therefore, we propose an end-to-end algorithm for the detection of surgical instruments.

Methods

Dual-Branched Head (DBH) and Overall Intersection over Union Loss (OIoU Loss) are introduced to solve the problem of inaccurate surgical instrument detection, both in terms of localization and classification. An effective method (DBHYOLO) for the detection for laparoscopic surgery in complex scenarios is proposed. This study manually annotates a new laparoscopic gastric cancer resection surgical instrument location dataset LGIL, which provides a better validation platform for surgical instrument detection methods.

Results

The proposed method's performance was tested using the m2cai16-tool-locations, LGIL, and Onyeogulu datasets. The mean Average Precision (mAP) values obtained were 96.8%, 95.6%, and 98.4%, respectively, which were higher than the other classical models compared. The improved model is more effective than the benchmark network in distinguishing between surgical instrument classes with high similarity and avoiding too many missed detection cases.

Conclusions

In this paper, the problem of inaccurate detection of surgical instruments is addressed from two different perspectives: classification and localization. And the experimental results on three representative datasets verify the performance of DBH-YOLO. It is shown that this method has a good generalization capability.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

Notes

  1. https://github.com/onyeogulu/Object-detection-of-surgical-instrument-based-on-YOLOv5/tree/main.

References

  1. Zhao Z, Voros S, Weng Y, Chang F, Li R (2017) Tracking-by-detection of surgical instruments in minimally invasive surgery via the convolutional neural network deep learning-based method. Comput Assist Surg 22(sup1):26–35. https://doi.org/10.1080/24699322.2017.1378777

    Article  Google Scholar 

  2. Liu K, Zhao Z, Shi P, Li F, Song H (2022) Real-time surgical tool detection in computer-aided surgery based on enhanced feature-fusion convolutional neural network. J Comput Des Eng 9(3):1123–1134. https://doi.org/10.1093/jcde/qwac049

    Article  Google Scholar 

  3. Peng J, Chen Q, Kang L, Jie H, Han Y (2022) Autonomous recognition of multiple surgical instruments tips based on arrow OBB-YOLO network. IEEE Trans Instrum Meas 71:1–13. https://doi.org/10.1109/TIM.2022.3162596

    Article  Google Scholar 

  4. Orentlicher D (2000) Medical malpractice: treating the causes instead of the symptoms. Med Care 38(3):247–249. https://doi.org/10.1097/00005650-200003000-00001

    Article  CAS  PubMed  Google Scholar 

  5. Ramesh A, Beniwal M, Uppar A M, Vikas V, Rao M (2021) Microsurgical tool detection and characterization in intra-operative neurosurgical videos. In: 2021 43rd Annual international conference of the IEEE engineering in medicine & biology society (EMBC), pp 2676–2681. https://ieeexplore.ieee.org/document/9630274

  6. Yang C, Zhao Z, Hu S (2020) Image-based laparoscopic tool detection and tracking using convolutional neural networks: a review of the literature. Comput Assist Surg 25(1):15–28. https://doi.org/10.1080/24699322.2020.1801842

    Article  Google Scholar 

  7. Sarikaya D, Corso J, Guru K (2017) Detection and localization of robotic tools in robot-assisted surgery videos using deep neural networks for region proposal and detection. IEEE Trans Med Imaging 36(7):1542–1549. https://doi.org/10.1109/TMI.2017.2665671

    Article  PubMed  Google Scholar 

  8. Zhang B, Wang S, Dong L, Chen P (2020) Surgical tools detection based on modulated anchoring network in laparoscopic videos. IEEE Access 8:23748–23758. https://doi.org/10.1109/ACCESS.2020.2969885

    Article  Google Scholar 

  9. Lee JD, Chien JC, Hsu YT, Wu CT (2021) Automatic surgical instrument recognition—a case of comparison study between the faster R-CNN, mask R-CNN, and single-shot multi-box detectors. Appl Sci 11(17):8097. https://doi.org/10.3390/app11178097

    Article  CAS  Google Scholar 

  10. Jo K, Choi Y, Choi J, Chung JW (2019) Robust real-time detection of laparoscopic instruments in robot surgery using convolutional neural networks with motion vector prediction. Appl Sci 9(14):2865. https://doi.org/10.3390/app9142865

    Article  Google Scholar 

  11. Li L, Li X, Ouyang B, Ding S, Yang S, Qu Y (2021) Autonomous multiple instruments tracking for robot-assisted laparoscopic surgery with visual tracking space vector method. IEEE/ASME Trans Mechatron 27(2):733–743. https://doi.org/10.1109/TMECH.2021.3070553

    Article  Google Scholar 

  12. Zheng Z, Wang P, Liu W, Li J, Ye R, Ren D (2020) Distance-IoU loss: faster and better learning for bounding box regression. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 12993–13000. https://doi.org/10.1609/aaai.v34i07.6999

  13. Wu Y, Chen Y, Yuan L, Liu Z, Wang L, Li H, Fu Y (2020) Rethinking classification and localization for object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 10186–10195. https://doi.org/10.1109/CVPR42600.2020.01020

  14. Song G, Liu Y, Wang X (2020) Revisiting the sibling head in object detector. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 11563–11572. https://doi.org/10.48550/arXiv.2003.07540

  15. Bodla N, Singh B, Chellappa R, Davis LS (2017) Soft-NMS--improving object detection with one line of code. In: Proceedings of the IEEE international conference on computer vision, pp 5561–5569. https://doi.org/10.48550/arXiv.1704.04503

  16. Jin A, Yeung S, Jopling J, Krause J, Azagury D, Milstein A, Fei-Fei L (2018) Tool detection and operative skill assessment in surgical videos using region-based convolutional neural networks. In: 2018 IEEE winter conference on applications of computer vision (WACV), IEEE, pp 691–699. https://doi.org/10.1109/WACV.2018.00081

  17. Onyeogulu T, Khan S, Teeti I, Islam A, Jin K, Rubio-Solis A, Naik R, Mylonas G, Cuzzolin F, (2022) Situation awareness for automated surgical check-listing in AI-assisted operating room. arXiv preprint arXiv:2209.05056. https://doi.org/10.48550/arXiv.2209.05056

  18. Jha D, Ali S, Emanuelsen K, Hicks S A, Halvorsen P (2021) Kvasir-instrument: diagnostic and therapeutic tool segmentation dataset in gastrointestinal endoscopy. https://doi.org/10.1007/978-3-030-67835-7_19

  19. Bochkovskiy A, Wang CY, Liao HYM (2020) Yolov4: optimal speed and accuracy of object detection. https://doi.org/10.48550/arXiv.2004.10934

  20. Wang CY, Bochkovskiy A, Liao HYM (2023) YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 7464–7475. https://doi.org/10.48550/arXiv.2207.02696

  21. Tan M, Pang R, Le QV (2020) Efficientdet: scalable and efficient object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 10781–10790. https://doi.org/10.48550/arXiv.1911.09070

  22. Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu CY, Berg AC (2016) SSD: single shot multibox detector. In: Computer vision–ECCV 2016: 14th European conference, Amsterdam, The Netherlands, October 11–14, 2016, proceedings, Part I, vol 14, pp 21–37. https://arxiv.org/abs/1512.02325

  23. Lin TY, Goyal P, Girshick R, He K, Dollár P (2017) Focal loss for dense object detection. In: Proceedings of the IEEE international conference on computer vision, pp 2980–2988. https://doi.org/10.48550/arXiv.1708.02002

  24. Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S (2020) End-to-end object detection with transformers. In: European conference on computer vision, pp 213–229. https://doi.org/10.1007/978-3-030-58452-8_13

  25. Ren S, He K, Girshick R, Sun J (2015) Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans Pattern Anal Mach Intell 39(6):1137–1149. https://doi.org/10.1109/TPAMI.2016.2577031

    Article  Google Scholar 

  26. Abdulbaki Alshirbaji T, Jalal NA, Docherty PD, Neumuth T, Möller K (2022) Robustness of convolutional neural networks for surgical tool classification in laparoscopic videos from multiple sources and of multiple types: a systematic evaluation. Electronics 11(18):2849. https://doi.org/10.3390/electronics11182849

    Article  Google Scholar 

  27. Wang X, Girdhar R, Yu SX, Misra I (2023) Cut and learn for unsupervised object detection and instance segmentation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 3124–3134. https://doi.org/10.48550/arXiv.2301.11320

Download references

Funding

The Key Industry Innovation Chain of Shaanxi (2022ZDLSF04-05).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xiaoying Pan.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Ethical approval

For this type of research, formal consent is not required.

Informed consent

This paper contains patient data from two publicly available datasets and one self-constructed dataset. Informed consent was obtained from patients for non-public records.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Pan, X., Bi, M., Wang, H. et al. DBH-YOLO: a surgical instrument detection method based on feature separation in laparoscopic surgery. Int J CARS (2024). https://doi.org/10.1007/s11548-024-03115-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s11548-024-03115-0

Keywords

Navigation