Skip to main content
Log in

The Segmentation of Multiple Types of Uterine Lesions in Magnetic Resonance Images Using a Sequential Deep Learning Method with Image-Level Annotations

  • Published:
Journal of Imaging Informatics in Medicine Aims and scope Submit manuscript

Abstract

Fully supervised medical image segmentation methods use pixel-level labels to achieve good results, but obtaining such large-scale, high-quality labels is cumbersome and time consuming. This study aimed to develop a weakly supervised model that only used image-level labels to achieve automatic segmentation of four types of uterine lesions and three types of normal tissues on magnetic resonance images. The MRI data of the patients were retrospectively collected from the database of our institution, and the T2-weighted sequence images were selected and only image-level annotations were made. The proposed two-stage model can be divided into four sequential parts: the pixel correlation module, the class re-activation map module, the inter-pixel relation network module, and the Deeplab v3 + module. The dice similarity coefficient (DSC), the Hausdorff distance (HD), and the average symmetric surface distance (ASSD) were employed to evaluate the performance of the model. The original dataset consisted of 85,730 images from 316 patients with four different types of lesions (i.e., endometrial cancer, uterine leiomyoma, endometrial polyps, and atypical hyperplasia of endometrium). A total number of 196, 57, and 63 patients were randomly selected for model training, validation, and testing. After being trained from scratch, the proposed model showed a good segmentation performance with an average DSC of 83.5%, HD of 29.3 mm, and ASSD of 8.83 mm, respectively. As far as the weakly supervised methods using only image-level labels are concerned, the performance of the proposed model is equivalent to the state-of-the-art weakly supervised methods.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

Data Availability

Data that support the findings of this study are available on request from the corresponding author. Data are not publicly available due to privacy or ethical restrictions.

Abbreviations

WSSS:

Weakly supervised semantic segmentation

FSSS:

Fully supervised semantic segmentation

CAM:

Class activation map

GAP:

Global average pooling

PCM:

Pixel correlation module

ReCAM:

Class re-activation map

IRNet:

Inter-pixel relation network

GT:

Ground truth

SCE:

Softmax cross entropy

BCE:

Binary cross entropy

DSC:

Dice similarity coefficient

HD:

Hausdorff distance

ASSD:

Average symmetric surface distance

EC:

Endometrial cancer

UL:

Uterine leiomyoma

EP:

Endometrial polyps

AHE:

Atypical hyperplasia of endometrium

ROI:

Region of interest

References

  1. Berek JS, Matias-Guiu X, Creutzberg C, Fotopoulou C, Gaffney D, Kehoe S, Lindemann K, Mutch D, Concin N; Endometrial Cancer Staging Subcommittee, FIGO Women's Cancer Committee. FIGO staging of endometrial cancer: 2023. Int J Gynaecol Obstet. 2023; 162 (2): 383–394. https://doi.org/10.1002/ijgo.14923.

  2. Rahimpour M, Saint Martin MJ, Frouin F, Akl P, Orlhac F, Koole M, Malhaire C. Visual ensemble selection of deep convolutional neural networks for 3D segmentation of breast tumors on dynamic contrast enhanced MRI. Eur Radiol. 2023; 33 (2): 959-969. https://doi.org/10.1007/s00330-022-09113-7.

    Article  PubMed  Google Scholar 

  3. Opfer R, Krüger J, Spies L, Ostwaldt AC, Kitzler HH, Schippling S, Buchert R. Automatic segmentation of the thalamus using a massively trained 3D convolutional neural network: higher sensitivity for the detection of reduced thalamus volume by improved inter-scanner stability. Eur Radiol. 2023; 33 (3): 1852-1861. https://doi.org/10.1007/s00330-022-09170-y.

    Article  CAS  PubMed  Google Scholar 

  4. Chen C, Zhang T, Teng Y, Yu Y, Shu X, Zhang L, Zhao F, Xu J. Automated segmentation of craniopharyngioma on MR images using U-Net-based deep convolutional neural network. Eur Radiol. 2023; 33 (4): 2665-2675. https://doi.org/10.1007/s00330-022-09216-1.

    Article  PubMed  Google Scholar 

  5. Jávorszky N, Homonnay B, Gerstenblith G, Bluemke D, Kiss P, Török M, Celentano D, Lai H, Lai S, Kolossváry M. Deep learning-based atherosclerotic coronary plaque segmentation on coronary CT angiography. Eur Radiol. 2022; 32 (10): 7217-7226. https://doi.org/10.1007/s00330-022-08801-8.

    Article  PubMed  Google Scholar 

  6. Rouvière O, Moldovan PC, Vlachomitrou A, Gouttard S, Riche B, Groth A, Rabotnikov M, Ruffion A, Colombel M, Crouzet S, Weese J, Rabilloud M. Combined model-based and deep learning-based automated 3D zonal segmentation of the prostate on T2-weighted MR images: clinical evaluation. Eur Radiol. 2022; 32 (5): 3248-3259. https://doi.org/10.1007/s00330-021-08408-5.

    Article  CAS  PubMed  Google Scholar 

  7. Corrado PA, Wentland AL, Starekova J, Dhyani A, Goss KN, Wieben O. Fully automated intracardiac 4D flow MRI post-processing using deep learning for biventricular segmentation. Eur Radiol. 2022; 32 (8): 5669-5678. https://doi.org/10.1007/s00330-022-08616-7.

    Article  PubMed  Google Scholar 

  8. Li Y, Wu Y, He J, Jiang W, Wang J, Peng Y, Jia Y, Xiong T, Jia K, Yi Z, Chen M. Automatic coronary artery segmentation and diagnosis of stenosis by deep learning based on computed tomographic coronary angiography. Eur Radiol. 2022; 32 (9): 6037-6045. https://doi.org/10.1007/s00330-022-08761-z.

    Article  PubMed  Google Scholar 

  9. Zheng Q, Zhang Y, Li H, Tong X, Ouyang M. How segmentation methods affect hippocampal radiomic feature accuracy in Alzheimer's disease analysis? Eur Radiol. 2022; 32 (10): 6965-6976. https://doi.org/10.1007/s00330-022-09081-y.

    Article  PubMed  Google Scholar 

  10. Cayot B, Milot L, Nempont O, Vlachomitrou AS, Langlois-Jacques C, Dumortier J, Boillot O, Arnaud K, Barten TRM, Drenth JPH, Valette PJ. Polycystic liver: automatic segmentation using deep learning on CT is faster and as accurate compared to manual segmentation. Eur Radiol. 2022; 32 (7): 4780-4790. https://doi.org/10.1007/s00330-022-08549-1.

    Article  PubMed  Google Scholar 

  11. Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A. Learning Deep Features for Discriminative Localization. In: IEEE Conference on Computer Vision and Pattern Recognition. 2016; pp. 2921–2929. https://doi.org/10.1109/CVPR.2016.319.

  12. Chen Z, Tian ZQ, Zhu JH, Li C, Du SY. C-CAM: Causal CAM for Weakly Supervised Semantic Segmentation on Medical Image. In: IEEE Conference on Computer Vision and Pattern Recognition. 2022; pp. 11676–11685.

  13. Wang Y, Zhang J, Kan M, Shan S, Chen X. Self-supervised Equivariant Attention Mechanism for Weakly Supervised Semantic Segmentation. In: IEEE Conference on Computer Vision and Pattern Recognition. 2020; pp. 12272–12281. https://doi.org/10.48550/arXiv.2004.04581.

  14. Jo SH, Yu IJ. Puzzle-cam: Improved local- ization via matching partial and full features. In: IEEE International Conference on Image Processing. 2021; pp. 639–643. https://doi.org/10.1109/ICIP42928.2021.9506058.

  15. Jiang PT, Yang YQ, Hou QB, Wei YC. L2G: A Simple Local-to-Global Knowledge Transfer Framework for Weakly Supervised Semantic Segmentation. In: IEEE Conference on Computer Vision and Pattern Recognition. 2022; pp. 16886–16896. https://doi.org/10.1109/CVPR52688.2022.01638.

  16. Chen ZZ, Wang T, Wu XW, Hua XS, Zhang HW, Sun QR. Class Re-Activation Maps for Weakly-Supervised Semantic Segmentation. In: IEEE Conference on Computer Vision and Pattern Recognition. 2022; pp. 959–968. https://doi.org/10.1109/cvpr52688.2022.00104.

  17. Ahn J, Kwak S. Learning pixel-level semantic affinitywith image-level supervision for weakly supervised semantic segmentation. In: IEEE Conference on Computer Vision and Pattern Recognition. 2018; pp. 4981–4990. http://https://doi.org/10.1109/CVPR.2018.00523.

  18. Ahn J, Cho S, Kwak S. Weakly Supervised Learning of Instance Segmentation with Inter-pixel Relations. In: IEEE Conference on Computer Vision and Pattern Recognition. 2019; pp. 2204–2213. https://doi.org/10.1109/CVPR.2019.00231.

  19. Ou Y, Huang SX, Wong KK, Cummock J, Volpi J, Wang JZ, Wong STC. BBox-Guided Segmentor: Leveraging expert knowledge for accurate stroke lesion segmentation using weakly supervised bounding box prior. Comput Med Imaging Graph. 2023; 107, 102236. https://doi.org/10.1016/j.compmedimag.2023.102236.

    Article  PubMed  Google Scholar 

  20. Lin D, Dai J, Jia J, He K, Sun J. Scribblesup: scribble-supervised convolutional networks for semantic segmentation. In: IEEE Conference on Computer Vision and Pattern Recognition. pp. 3159–3167. https://doi.org/10.1109/CVPR.2016.344.

  21. Gao F, Hu M, Zhong ME, Feng S, Tian X, Meng X, Ni-Jia-Ti MY, Huang Z, Lv M, Song T, Zhang X, Zou X, Wu X. Segmentation only uses sparse annotations: Unified weakly and semi-supervised learning in medical images. Med Image Anal. 2022; 80: 102515. https://doi.org/10.1016/j.media.2022.102515.

    Article  PubMed  Google Scholar 

  22. Luo X, Wang G, Liao W, Chen J, Song T, Chen Y, Zhang S, Metaxas DN, Zhang S. Semi-supervised medical image segmentation via uncertainty rectified pyramid consistency. Med Image Anal. 2022; 80: 102517. https://doi.org/10.1016/j.media.2022.102517.

    Article  PubMed  Google Scholar 

  23. Wang K, Zhan B, Zu C, Wu X, Zhou J, Zhou L, Wang Y. Semi-supervised medical image segmentation via a tripled-uncertainty guided mean teacher model with contrastive learning. Med Image Anal. 2022; 79: 102447. https://doi.org/10.1016/j.media.2022.102447.

    Article  PubMed  Google Scholar 

  24. Huang M, Zhou S, Chen X, Lai H, Feng, Q. Semi-supervised hybrid spine network for segmentation of spine MR images. Comput Med Imaging Graph. 2023; 107, 102245. https://doi.org/10.1016/j.compmedimag.2023.102245.

    Article  PubMed  Google Scholar 

  25. Chen LC, Zhu Y, Papandreou G, Schroff F, Adam H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In: Proceedings of the European conference on computer vision. pp. 801–818.

Download references

Funding

This study has received funding by the National Natural Science Foundation of China (No. 82003843), Natural Science Foundation of Liaoning Province (No. 2020-BS-221), and Dalian Medical Science Research Program (No.2211023).

Author information

Authors and Affiliations

Authors

Contributions

Yu-meng Cui: conceptualization; methodology; writing; funding acquisition; Hua-li Wang: conceptualization; resources; supervision; Rui Cao: resources; supervision; Hong Bai: methodology; validation; statistical analysis; Dan Sun: data curation; project administration; Jiu-Xiang Feng: data curation; project administration; Xue-feng Lu: conceptualization; review and editing; coding; funding acquisition.

Corresponding author

Correspondence to Xue-feng Lu.

Ethics declarations

Ethics Approval

This study was approved by the Institutional Review Board of Dalian Women and Children’s Medical Group before patient information was accessed.

Consent to Participate

This study was performed under a waiver for informed consent due to the retrospective nature of the analysis, the anonymity of the data.

Competing Interests

The authors declare no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Below is the link to the electronic supplementary material.

Supplementary file1 (DOCX 173 KB)

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Cui, Ym., Wang, Hl., Cao, R. et al. The Segmentation of Multiple Types of Uterine Lesions in Magnetic Resonance Images Using a Sequential Deep Learning Method with Image-Level Annotations. J Digit Imaging. Inform. med. 37, 374–385 (2024). https://doi.org/10.1007/s10278-023-00931-9

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10278-023-00931-9

Keywords

Navigation