Skip to main content
Log in

Automatic fetal ultrasound image segmentation of first trimester for measuring biometric parameters based on deep learning

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Transvaginal ultrasonography (TVS) is a common method used by doctors to monitor the embryonic development. In the early stage of pregnancy, doctors assess the growth and development of the embryo by measuring biological indicators such as gestational sac area (GSA), yolk sac diameter (YSD), and crown-rump length (CRL) in TVS images. Even though these indicators can be manually obtained by experienced physicians, the manual measurement process is time-consuming, inefficient, and heavily dependent on the sonographer's expertise. To improve this situation, we, here, aimed to establish a modified Unet model, namely AFG-net, which is capable of automatically obtaining the related clinical values required for measuring embryonic development. Using this method, the essential values, including gestational sac (GS), yolk sac (YS) and embryo region in the TVS image, were easily and accurately identified and located, which were further completely separated by image segmentation to obtain the corresponding measurement values. Notably, this model is able to achieve superior segmentation effect even when the input image with poor quality, low contrast, fuzzy region boundary and complex anatomical shape by applying some advanced methods such as attention fusion and guide filter. Consequently, our results showed our model demonstrated a higher average precision, Intersection Over Union (IOU), and Dice coefficient (Dice) of GS, YS and embryo compared to a normal Unet, with 94.75%, 86.15% and 92.11% versus 92.01%, 83.00%, and 90.00%, respectively. The absolute error between the biological indicators (GSA, YSD and CRL) automatically extracted from the segmentation results and the manual measurement results is 0.66mm. The automatic segmentation and measurement process significantly reduces the subjectivity of manual measurement and reduces the clinician workload. It also helps to improve diagnostic accuracy, enables repeatability and standardization in clinical practice, and provides a valuable tool for prenatal care and monitoring.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

Data availability

The data that support the findings of this study are available from Reproductive Center and the Imaging Department of the Reproductive and Genetic Hospital of CITIC-Xiangya but restrictions apply to the availability of these data, which were used under licence for the current study, and so are not publicly available. Data are however available from the authors upon reasonable request and with permission of Reproductive Center and the Imaging Department of the Reproductive and Genetic Hospital of CITIC-Xiangya.

References

  1. Yin C, Wang Y, Zhang Q, Han F, Yuan Z, Yao Y (2022) An Accurate Segmentation framework for static ultrasound images of the gestational sac. J Med Biol Eng 42(1):49–62. https://doi.org/10.1007/s40846-021-00674-4

    Article  Google Scholar 

  2. Hu X, Miao M, Bai Y, Cheng N, Ren X (2018) Reproductive factors and risk of spontaneous abortion in the jinchang cohort. Int J Environ Res Public Health 15(11):2444. https://doi.org/10.3390/ijerph15112444

    Article  PubMed  PubMed Central  Google Scholar 

  3. Bilagi A, Burke DL, Riley RD, Mills I, Kilby MD, Katie Morris R (2017) Association of maternal serum PAPP-A levels, nuchal translucency and crown-rump length in first trimester with adverse pregnancy outcomes: retrospective cohort study: first trimester PAPP-A and adverse pregnancy outcome. Prenat Diagn 37(7):705–711. https://doi.org/10.1002/pd.5069

    Article  CAS  PubMed  Google Scholar 

  4. Papaioannou GI, Syngelaki A, Maiz N, Ross JA, Nicolaides KH (2011) Ultrasonographic prediction of early miscarriage. Hum Reprod (Oxford, England) 26(7):1685–1692. https://doi.org/10.1093/humrep/der130

    Article  Google Scholar 

  5. Błaszczyk K, Wojcieszyn M, Biernat M, Lukasik A, Wilk M, Poreba R (2000) Predicting the risk of poor pregnancy outcome by ultrasound examination of yolk sac diameter. Ginekologia Polska 71(8):699–703. http://refhub.elsevier.com/S0169-2607(20)31457-7/sbref0006. Accessed 23 Jul 2023

  6. Jeve Y, Rana R, Bhide A, Thangaratinam S (2011) Accuracy of first-trimester ultrasound in the diagnosis of early embryonic demise: a systematic review. Ultrasound Obstet Gynecol 38(5):489–496. https://doi.org/10.1002/uog.10108

    Article  CAS  PubMed  Google Scholar 

  7. El-Mekkawi SF, El-Shahawy HF, Alyamni OM (2015) Prediction of spontaneous miscarriage risk by the use of first trimester ultrasound measurements and maternal serum progesterone level at the 7th week of pregnancy. Middle East Fertil Soc J 20(1):16–20. https://doi.org/10.1016/j.mefs.2014.04.006

    Article  Google Scholar 

  8. Yu Z, Tan E-L, Ni D, Qin J, Chen S, Li S, Lei B, Wang T (2018) A Deep convolutional neural network-based framework for automatic fetal facial standard plane recognition. IEEE J Biomed Health Inform 22(3):874–885. https://doi.org/10.1109/JBHI.2017.2705031

    Article  PubMed  Google Scholar 

  9. Doubilet PM, Benson CB, Bourne T, Blaivas M (2013) Diagnostic criteria for nonviable pregnancy early in the first trimester. N Engl J Med 369(15):1443–1451. https://doi.org/10.1056/NEJMra1302417

    Article  CAS  PubMed  Google Scholar 

  10. Ibrahim DA, Al-Assam H, Du H, Farren J, Al-karawi D, Bourne T, Jassim S (2016) In:  Agaian SS, Jassim SA (eds) Automatic segmentation and measurements of gestational sac using static B-mode ultrasound images, pp 98690B. https://doi.org/10.1117/12.2224514

  11. Chakkarwar VA, Joshi MS, Revankar PS (2010) Automated analysis of gestational sac in medical image processing. 2010 IEEE 2nd International Advance Computing Conference (IACC), pp 304–309. https://doi.org/10.1109/IADCC.2010.5422938

  12. Khazendar S, Farren J, Al-Assam H, Du H, Sayasneh A, Bourne T, Jassim S (2014) Automatic identification of early miscarriage based on multiple features extracted from ultrasound images. Annual Conference on Medical Image Understanding and Analysis. https://api.semanticscholar.org/CorpusID:6568420. Accessed 5 Mar 2022

  13. Krizhevsky A, Sutskever I, Hinton GE (2017) ImageNet classification with deep convolutional neural networks. Commun ACM 60(6):84–90. https://doi.org/10.1145/3065386

    Article  Google Scholar 

  14. Ejbali R, Zaied M (2018) A dyadic multi-resolution deep convolutional neural wavelet network for image classification. Multimed Tools Appl 77(5):6149–6163. https://doi.org/10.1007/s11042-017-4523-2

    Article  Google Scholar 

  15. He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 770–778.https://doi.org/10.1109/CVPR.2016.90

  16. Ren S, He K, Girshick R, Sun J (2017) Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans Pattern Anal Mach Intell 39(6):1137–1149. https://doi.org/10.1109/TPAMI.2016.2577031

    Article  PubMed  Google Scholar 

  17. Lu X, Wang W, Shen J, Crandall D, Luo J (2020) Zero-shot video object segmentation with co-attention siamese networks. IEEE Trans Pattern Anal Mach Intell:1–1. https://doi.org/10.1109/TPAMI.2020.3040258

  18. Zhao H, Shi J, Qi X, Wang X, Jia J (2017) Pyramid scene parsing network (arXiv:1612.01105). arXiv. http://arxiv.org/abs/1612.01105. Accessed 9 Jul 2023

  19. Lu X, Wang W, Shen J, Crandall DJ, Van Gool L (2022) Segmenting objects from relational visual data. IEEE Trans Pattern Anal Mach Intell 44(11):7885–7897. https://doi.org/10.1109/TPAMI.2021.3115815

    Article  PubMed  Google Scholar 

  20. Shelhamer E, Long J, Darrell T (2017) Fully convolutional networks for semantic segmentation. IEEE Trans Pattern Anal Mach Intell 39(4):640–651. https://doi.org/10.1109/TPAMI.2016.2572683

    Article  PubMed  Google Scholar 

  21. Ronneberger O, Fischer P, Brox T (2015) U-Net: convolutional networks for biomedical image segmentation (arXiv:1505.04597). arXiv. http://arxiv.org/abs/1505.04597. Accessed 5 Mar 2022

  22. Wu L, Xin Y, Li S, Wang T, Heng P-A, Ni D (2017) Cascaded Fully Convolutional Networks for automatic prenatal ultrasound image segmentation. 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017), pp 663–666. https://doi.org/10.1109/ISBI.2017.7950607

  23. Qiao D, Zulkernine F (2020) Dilated squeeze-and-excitation U-Net for fetal ultrasound image segmentation. 2020 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB), pp 1–7. https://doi.org/10.1109/CIBCB48159.2020.9277667

  24. Fu H, Xu Y, Lin S, Kee Wong DW, Liu J (2016) DeepVessel: retinal vessel segmentation via deep learning and conditional random field. In: Ourselin S, Joskowicz L, Sabuncu M, Unal G, Wells W (eds) Medical Image Computing and Computer-Assisted Intervention – MICCAI 2016. MICCAI 2016. Lecture Notes in Computer Science, vol 9901. Springer, Cham. https://doi.org/10.1007/978-3-319-46723-8_16

  25. Oktay O, Schlemper J, Folgoc LL, Lee M, Heinrich M, Misawa K, Mori K, McDonagh S, Hammerla NY, Kainz B, Glocker B, Rueckert D (2018) Attention U-Net: learning where to look for the pancreas (arXiv:1804.03999). arXiv. http://arxiv.org/abs/1804.03999. Accessed 5 Mar 2022

  26. Lu X, Wang W, Ma C, Shen J, Shao L, Porikli F (2019) See more, know more: unsupervised video object segmentation with co-attention Siamese networks. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 3618–3627. https://doi.org/10.1109/CVPR.2019.00374

  27. Wang W, Lu X, Shen J, Crandall D, Shao L (2019) Zero-shot video object segmentation via attentive graph neural networks. 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pp 9235–9244https://doi.org/10.1109/ICCV.2019.00933

  28. He K, Sun J, Tang X (2013) Guided image filtering. IEEE Trans Pattern Anal Mach Intell 35(6):1397–1409. https://doi.org/10.1109/TPAMI.2012.213

    Article  PubMed  Google Scholar 

  29. Westin C-F, Knutsson H, Kikinis R (2009) Chapter 2—Adaptive image filtering. In: Bankman IN (ed) Handbook of medical image processing and analysis, 2nd edn. Academic Press, pp 19–33. https://doi.org/10.1016/B978-012373904-9.50009-X

  30. Zhang S, Fu H, Yan Y, Zhang Y, Wu Q, Yang M, Tan M, Xu Y (2019) Attention guided network for retinal image segmentation. Medical Image Computing and Computer Assisted Intervention – MICCAI 2019: 22nd International Conference, Shenzhen, China, October 13–17, 2019, Proceedings, Part I, pp 797–805. https://doi.org/10.1007/978-3-030-32239-7_88

  31. Rueda S, Fathima S, Knight CL, Yaqub M, Papageorghiou AT, Rahmatullah B, Foi A, Maggioni M, Pepe A, Tohka J, Stebbing RV, McManigle JE, Ciurte A, Bresson X, Cuadra MB, Sun C, Ponomarev GV, Gelfand MS, Kazanov MD … Noble JA (2014) Evaluation and comparison of current fetal ultrasound image segmentation methods for biometric measurements: a grand challenge. IEEE Trans Med Imaging 33(4):797–813. https://doi.org/10.1109/TMI.2013.2276943

  32. Suguna B, Sukanya K (2019) Yolk sac size & shape as predictors of first trimester pregnancy outcome: a prospective observational study. J Gynecol Obstet Hum Reprod 48(3):159–164. https://doi.org/10.1016/j.jogoh.2018.10.016

    Article  CAS  PubMed  Google Scholar 

  33. Islam MN, Mustafina SN, Mahmud T, Khan NI (2022) Machine learning to predict pregnancy outcomes: a systematic review, synthesizing framework and future research agenda. BMC Pregnancy Childbirth 22(1):348. https://doi.org/10.1186/s12884-022-04594-2

    Article  PubMed  PubMed Central  Google Scholar 

  34. Yi Y, Li X, Ouyang Y, Lin G, Lu G, Gong F (2016) Discriminant analysis forecasting model of first trimester pregnancy outcomes developed by following 9,963 infertile patients after in vitro fertilization. Fertil Steril 105(5):1261–1265. https://doi.org/10.1016/j.fertnstert.2016.01.033

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

The authors gratefully acknowledge the financial support provided by Hunan Provincial Natural Science Foundation of China (2023JJ60491) and the Open Project of Xiangjiang Laboratory (22XJ02005).

Author information

Authors and Affiliations

Authors

Contributions

Conceptualization: L.L., D.T. and X.L.; methodology: L.L., D.T., X.L. and Y.O.; software: D.T.; resources: L.L., D.T. and X.L.; data curation: X.L. and Y.O.; data analysis and interpretation: L.L., D.T., X.L., and Y.O.; writing—original draft preparation: L.L. and D.T.; writing—review and editing: L.L., D.T. and X.L.; project administration: L.L. and Y.O. The first four authors contributed equally to the paper. All authors have read and agreed to the published version of the manuscript.

Corresponding author

Correspondence to Xihong Li.

Ethics declarations

There is nothing to declare by the authors.

Competing interests

We declare that we do not have any commercial or associative interest that represents a conflict of interest in connection with the work submitted.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Liu, L., Tang, D., Li, X. et al. Automatic fetal ultrasound image segmentation of first trimester for measuring biometric parameters based on deep learning. Multimed Tools Appl 83, 27283–27304 (2024). https://doi.org/10.1007/s11042-023-16565-6

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-023-16565-6

Keywords

Navigation