Skip to main content
Log in

Meningioma segmentation with GV-UNet: a hybrid model using a ghost module and vision transformer

  • Original Paper
  • Published:
Signal, Image and Video Processing Aims and scope Submit manuscript

Abstract

Meningiomas are the most common intracranial tumors in adults. The size and shape of a tumor mostly rely on manual measurement by a neurosurgeon. In recent years, deep learning has rapidly developed and has great potential for medical image segmentation. However, most segmentation models still cannot balance the number of parameters and accuracy. In this study, we proposed a novel segmentation network (named GV-UNet) based on a CNN and a transformer for T1-enhanced images of meningiomas to improve the accuracy and efficiency of tumor segmentation. GV-UNet uses an encoder–decoder as the main structure. In the downsampling process, features are extracted through a standard convolutional layer, and a ConvMixer Layer is used to optimize feature extraction with different sizes of meningiomas. Then, a lightweight transformer block is built to model long-range dependencies. In the final layer of the encoder, we propose an innovative Ghost-CA block, which extracts deep features via feature mapping rather than by elevating dimensionality to reduce the number of parameters. In the upsampling process, we add a SimAM that can incorporate a three-dimensional attention mechanism without increasing network parameters, effectively capturing the relationships between features and the spatial structure of the target. GV-UNet was trained and validated using numerous pathologically confirmed T1-enhanced images of meningiomas from Tianjin Huanhu Hospital. We also utilized meningioma images from the Kaggle dataset to test the robustness of the model.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

Data availability

The original contributions presented in the study are included in the article/supplementary material; further inquiries can be directed to the corresponding author.

References

  1. Wiemels, J., Wrensch, M., Claus, E.B.: Epidemiology and etiology of meningioma. J. Neuro-Oncol. Numer. Math. 99(3), 307–314 (2010)

    Article  Google Scholar 

  2. Holleczek, B., et al.: Incidence, mortality and outcome of meningiomas: a population-based study from Germany. Cancer Epidemiol. 62(101562) (2019)

  3. Hwang, K.L., Hwang, W.L., Bussire, M.R., Shih, H.A.: The role of radiotherapy in the management of high-grade meningiomas. Chin. Clin. Oncol. 6(Suppl 1), S5–S5 (2017)

    Article  PubMed  Google Scholar 

  4. Preusser, M., Brastianos, P.K., Mawrin, C.: Advances in meningioma genetics: novel therapeutic opportunities. Nat. Rev. Neurol. 14(2), 106–115 (2018)

    Article  CAS  PubMed  Google Scholar 

  5. Kim, D., et al.: Histopathological prognostic factors of recurrence following definitive therapy for atypical and malignant meningiomas. J. Neurosurg. 128(4), 1123–1132 (2018)

    Article  PubMed  Google Scholar 

  6. Louis, D.N., et al.: The 2016 World Health Organization classification of tumors of the central nervous system: a summary. Acta Neuropathol. 131(6), 803–820 (2016)

    Article  PubMed  Google Scholar 

  7. Goldbrunner, R., et al.: EANO guidelines for the diagnosis and treatment of meningiomas. Lancet Oncol. 17(9), E383–E391 (2016)

    Article  PubMed  Google Scholar 

  8. Mawrin, C., Perry, A.: Pathological classification and molecular genetics of meningiomas. J. Neurooncol. 99(3), 379–391 (2010)

    Article  CAS  PubMed  Google Scholar 

  9. Huang, W., et al.: Feature pyramid network with level-aware attention for meningioma segmentation. IEEE Trans. Emerg. Top. Comput. Intell. 6(5), 1201–1210 (2022)

    Article  Google Scholar 

  10. Ke, C., et al.: Differentiation between benign and nonbenign meningiomas by using texture analysis from multiparametric MRI. J. Magn. Reson. Imaging 51(6), 1810–1820 (2020)

    Article  PubMed  Google Scholar 

  11. Huang, R.Y., et al.: Imaging and diagnostic advances for intracranial meningiomas. Neuro Oncol. 21, I44–I61 (2019)

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  12. Spille, D.C., Sporns, P.B., Hess, K., Stummer, W., Brokinkel, B.: Prediction of high-grade histology and recurrence in meningiomas using routine preoperative magnetic resonance imaging: a systematic review. World Neurosurg. 128, 174–181 (2019)

    Article  PubMed  Google Scholar 

  13. Duan, C.F., et al.: Comparison of different radiomic models based on enhanced T1-weighted images to predict the meningioma grade. Clin. Radiol. 77(4), E302–E307 (2022)

    Article  CAS  PubMed  Google Scholar 

  14. Chen, C.Y., Guo, X.Y., Wang, J., Guo, W., Ma, X.L., Xu J.G.: The diagnostic value of radiomics-based machine learning in predicting the grade of meningiomas using conventional magnetic resonance imaging: a preliminary study. Front. Oncol. 9(1338) (2019)

  15. Li, X.X., et al.: Meningioma grading using conventional MRI histogram analysis based on 3D tumor measurement. Eur. J. Radiol. 110, 45–53 (2019)

    Article  PubMed  Google Scholar 

  16. Vassantachart, A. et al.: Automatic differentiation of Grade I and II meningiomas on magnetic resonance image using an asymmetric convolutional neural network. Sci. Rep. 12(1), Art no. 3806 (2022)

  17. Chen, C.Y. et.al.: Automatic meningioma segmentation and grading prediction: a hybrid deep-learning method. J. Personal. Med. 11(8), 786 (2021)

  18. Zhu, H., Fang, Q.H., He, H.Z., Hu, J.F.D., Jiang, H., Xu, K.: Automatic prediction of meningioma grade image based on data amplification and improved convolutional neural network. Comput. Math. Methods Med. 2019, 7289273 (2019)

  19. Zhang, H., et al.: Deep learning model for the automated detection and histopathological prediction of meningioma. Neuroinformatics 19(3), 393–402 (2021)

    Article  PubMed  Google Scholar 

  20. Zhu, Y.B., et al.: A deep learning radiomics model for preoperative grading in meningioma. Eur. J. Radiol. 116, 128–134 (2019)

    Article  PubMed  Google Scholar 

  21. Yang, L.P., et al.: A deep learning radiomics model may help to improve the prediction performance of preoperative grading in meningioma. Neuroradiology 64(7), 1373–1382 (2022)

    Article  PubMed  Google Scholar 

  22. Shirokikh, B., Dalechina, A., Shevtsov, A., Krivov, E., Belyaev, M.: Deep Learning for Brain Tumor Segmentation in Radiosurgery: Prospective Clinical Evaluation. Glioma, Multiple Sclerosis. Stroke and Traumatic Brain Injuries, Brainlesion (2020)

  23. Otsu, N.: A threshold selection method from gray-level histograms. IEEE Trans. Syst. Man Cybern. 9(1), 62–66 (2007)

    Article  Google Scholar 

  24. Prastawa, M., Bullitt, E., Gerig, G.: Simulation of brain tumors in MR images for evaluation of segmentation efficacy. Med. Image Anal. 13(2), 297–311 (2009)

    Article  PubMed  Google Scholar 

  25. Corso, J.J., Sharon, E., Dube, S., El-Saden, S., Sinha, U., Yuille, A.: Efficient multilevel brain tumor segmentation with integrated Bayesian model classification. IEEE Trans. Med. Imaging 27(5), 629–640 (2008)

    Article  CAS  PubMed  Google Scholar 

  26. Chen, L.C., Papandreou, G., Kokkinos, I., Murphy, K., Yuille, A.L.: DeepLab: semantic image segmentation with deep convolutional nets, Atrous convolution, and fully connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 40(4), 834–848 (2018)

    Article  PubMed  Google Scholar 

  27. Chen, L.C., Zhu, G. Papandreou, F., Schroff, Adam, H.: Encoder–decoder with atrous separable convolution for semantic image segmentation. In: European Conference on Computer Vision (2018)

  28. Chen, L.C., Papandreou, G., Schroff, F., Adam, H.: Rethinking atrous convolution for semantic image segmentation (2017)

  29. Zhao, H., et al.: SC2Net: a novel segmentation-based classification network for detection of COVID-19 in chest X-ray images. IEEE J. Biomed. Health Inform. 26(8), 4032–4043 (2022)

    Article  PubMed  Google Scholar 

  30. Wang, M., Jiang, H., Shi, T., Yao, Y.D.: SCL-Net: Structured collaborative learning for PET/CT based tumor segmentation. IEEE J. Biomed. Health Inform. 27(2), 1048–1059 (2023)

  31. Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: International Conference on Medical Image Computing and Computer Assisted Intervention, Springer, Cham, pp. 234–241 (2015)

  32. Zhou, Z., Siddiquee, M., Tajbakhsh, N., Liang, J.: UNet++: a nested U-net architecture for medical image segmentation (2018)

  33. Zhang, Y., Wu, J., Liu, Y., Chen, Y., Wu, E.X., Tang, X.: MI-UNet: multi-inputs UNet incorporating brain parcellation for stroke lesion segmentation from T1-weighted magnetic resonance images. IEEE J. Biomed. Health Inform. 25(2), 526–535 (2021)

    Article  PubMed  Google Scholar 

  34. Li, H., et al.: CR-Unet: a composite network for ovary and follicle segmentation in ultrasound images. IEEE J. Biomed. Health Inform. 24(4), 974–983 (2020)

    Article  MathSciNet  PubMed  Google Scholar 

  35. Guerrero, R., et al.: White matter hyperintensity and stroke lesion segmentation and differentiation using convolutional neural networks. NeuroImage: Clin. 17, 918–934 (2018)

  36. Oktay, O. et al.: Attention U-net: learning where to look for the pancreas (2018)

  37. Guo, C., Szemenyei, M., Yi, Y., Wang, W., Chen, B., Fan, C. SA-UNet: spatial attention U-Net for retinal vessel segmentation (2020)

  38. Zheng, S., Lu, J., Zhao, H., Zhu, X., Zhang, L.: Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective with Transformers (2020)

  39. Cao, H., et al.: Swin-Unet: Unet-like pure transformer for medical image segmentation (2021)

  40. Chen, J., Lu, Y., Yu, Q., Luo, X., Zhou, Y.: TransUNet: transformers make strong encoders for medical image segmentation (2021)

  41. Valanarasu, J.M.J., Oza, P., Hacihaliloglu, I., Patel, V.M.: Medical transformer: gated axial-attention for medical image segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention (2021)

  42. Wang, H., Cao, P., Wang, J., Zaiane, O.R.: UCTransNet: rethinking the skip connections in U-Net from a channel-wise perspective with transformer (2021)

  43. Huang, W., Shu, X., Wang, Z., Zhang, L., Chen, C., Xu, J., Yi, Z.: Feature pyramid network with level-aware attention for meningioma segmentation. IEEE Trans. Emerg. Top. Comput. Intell. 6(5), 1201–1210 (2022)

    Article  Google Scholar 

  44. Ma, X., Zhao, Y., Lu, Y., Li, P., Li, X., Mei, N., Yin, B.: A dual-branch hybrid dilated CNN model for the AI-assisted segmentation of meningiomas in MR images. Comput. Biol. Med. 151, 106279 (2022)

    Article  PubMed  Google Scholar 

  45. Wang, W., Chen, C., Ding, M., Yu, H., Zha, S., Li, J.: Transbts: multimodal brain tumor segmentation using transformer. In: Medical Image Computing and Computer Assisted Intervention-MICCAI 2021: 24th International Conference, Strasbourg, France, September 27–October 1, 2021, Proceedings, Part I 24 (pp. 109–119). Springer (2021)

  46. Liu, Y., Wang, H., Chen, Z., Huangliang, K., Zhang, H.: TransUNet+: redesigning the skip connection to enhance features in medical image segmentation. Knowl.-Based Syst. 256, 109859 (2022)

    Article  Google Scholar 

  47. Zou, Y., Ge, Y., Zhao, L., Li, W.: MR-Trans: multiresolution transformer for medical image segmentation. Comput. Biol. Med. 165, 107456 (2023)

    Article  PubMed  Google Scholar 

  48. Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. JMLR (2015)

  49. He, K., Zhang, X., Ren, S., Sun, J.: Delving deep into rectifiers: surpassing human-level performance on ImageNet Classification. IEEE Computer Society (2012)

  50. Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Rabinovich, A.: Going deeper with convolutions. IEEE Computer Society (2014)

  51. Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. Computer Science (2014)

  52. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. IEEE (2016)

  53. Trockman, A., Kolter, J.Z.: Patches are all you need? arXiv e-prints (2022)

  54. Mehta, S., Rastegari, M.: MobileViT: Light-Weight, General-purpose, and Mobile-friendly Vision Transformer (2021)

  55. Tang, Y., Han, K., Guo, J., Xu, C., Xu, C., Wang, Y.: GhostNetV2: enhance cheap operation with long-range attention. arXiv preprint arXiv:2211.12905 (2022)

  56. Hou, Q., Zhou, D., Feng, J.: Coordinate attention for efficient mobile network design (2021)

  57. Yang, L., Zhang, R.Y., Li, L., Xie, X.: SimAM: a simple, parameter-free attention module for convolutional neural networks. In: International Conference on Machine Learning (2021)

  58. Zhang, H., Wang, Y., Dayoub, F., Snderhauf, N.: VarifocalNet: An IoU-Aware Dense Object Detector (2020)

  59. Kingma, D., Ba, J.: Adam: a method for stochastic optimization. Comput. Sci. (2014)

  60. Mehta, R.: Introducing Dice, Jaccard, and Other Label Overlap Measures To ITK (2015)

  61. Beauchemin, M., Thomson, K.P., Edwards, G.: On the Hausdorff distance used for the evaluation of segmentation results. Can. J. Remote. Sens. 24(1), 3–8 (1998)

    Article  ADS  Google Scholar 

  62. Huang, Z., Wang, X., Wei, Y., Huang, L., Huang, T.S.: CCNet: criss-cross attention for semantic segmentation. IEEE Trans. Pattern Anal. Mach. Intell., vol. PP, 99, 1–1 (2020)

  63. Badrinarayanan, V., Kendall, A., Cipolla, R.: SegNet: a deep convolutional encoder–decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 39(12), 2481–2495 (2017)

    Article  PubMed  Google Scholar 

Download references

Funding

This work was supported in part by National Natural Science Foundation of China (61201106, 82102014), the Natural Science Basic Research Program of Shaanxi Province (2022JQ-792) and Tianjin Research Innovation Project for Postgraduate Students (2022SKY126).

Author information

Authors and Affiliations

Authors

Contributions

H.B. and Z.Z. contributed to conceptualization; H.B. provided methodology; Z.Z. provided software; C.N., Q.M. and J.S. performed validation; C.N. carried out formal analysis; Q.G. and Y.Y. performed investigation; Y.Y. provided resources; Z.Z. provided data; Z.Z. performed writing original draft preparation; H.B., Q.M. and J.S. performed writing review and editing; Z.Z. contributed to visualization; H.B. performed supervision; Q.M. performed project administration; H.B. contributed to funding acquisition. All authors have read and agreed to the published version of the manuscript.

Corresponding authors

Correspondence to Quanfeng Ma or Jian Song.

Ethics declarations

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Ethics approval

Ethical review and approval was not required for the study on human participants in accordance with the local legislation and institutional requirements. The patients/participants provided their written informed consent to participate in this study.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix A

Appendix A

1.1 A.1 Explanation of evaluation indicators

Fig. 5
figure 5

Trade-off between parameters and segmentation accuracy of different models in Dataset A

Fig. 6
figure 6

Trade-off between parameters and segmentation accuracy of different models in Dataset B

The DSC is an evaluation metric used to quantify the overlapping ratio between two binary sets, as defined in Eq. (15).

$$\begin{aligned} Dice=\frac{2TP}{FP+2TP+FN} \end{aligned}$$
(15)

where TP represents the true positive; FP represents the false positive; TN represents the true negative; and FN represents the false negative. The range of all the values is [0, 1], and the higher the Dice value is, the closer the tumor segmentation result is to the annotated result, and the better the segmentation effect.

IoU is another evaluation index in segmentation models to measure the overlap between the ground truth and the predicted segmentation, as defined in Eq. (16).

$$\begin{aligned} IoU=\frac{P\bigcap T}{P\bigcup T} \end{aligned}$$
(16)

where P and T represent the predicted result and manually annotated result, respectively.

The Hausdorff distance is a measure of the similarity between two sets of points, describing the maximum shortest distance between the segmentation result and the ground truth, as defined in Eq. (17).

$$\begin{aligned} {{d}_{H}}(A,B)=\max \{{{d}_{AB}},{{d}_{BA}}\} \end{aligned}$$
$$\begin{aligned} max \{\underset{a\in A}{\mathop {\max }}\,\underset{b\in B}{\mathop {\min d}}\,(a,b),\underset{b\in B}{\mathop {\max }} \,\underset{a\in A}{\mathop {\min }}\,d(b,a)\} \end{aligned}$$
(17)

where A and B represent the point set of the model prediction results and the point set of the ground truth results, respectively; d represents the Euclidean distance; and \({{d}_{AB}}\) and \({{d}_{BA}}\) represent the one-way Hausdorff distance from A to B and the one-way Hausdorff distance from B to A, respectively. The smaller the value is, the better the segmentation results.

ASD represents the average symmetrical surface distance, which indicates the average surface distance of all points in the segmented point set and is sensitive to the boundary, as defined in Eq. (18)-(20). The lower the value is, the better the segmentation results.

$$\begin{aligned} {{B}_{seg}}= & {} \left\{ \forall {{p}_{1}}\in {{A}_{seg}},closet\_distance({{p}_{1}},{{p}_{2}})|\exists {{p}_{2}}\in {{A}_{gt}} \right\} \end{aligned}$$
(18)
$$\begin{aligned} {{B}_{gt}}= & {} \left\{ \forall {{p}_{2}}\in {{A}_{gt}},closet\_distance({{p}_{1}},{{p}_{2}})|\exists {{p}_{1}}\in {{A}_{seg}} \right\} \end{aligned}$$
(19)
$$\begin{aligned} ASD= & {} mean(\{{{B}_{seg}},{{B}_{gt}}\}) \end{aligned}$$
(20)

where \({{A}_{seg}}\) represents the pixels of the boundary of the model’s predicted result and \({{A}_{gt}}\) represents the pixels of the boundary of the ground truth.

1.2 A.2 Parameters

The proposed GV-UNet presents a solution to the trade-off between computational power and segmentation accuracy in clinical scenarios involving meningiomas. In this section, we focus on the relationship between model parameters and segmentation accuracy.

As shown in Figs. 5 and 6, the segmentation performance of SegNet, SwinUNet and TransUNet in the two datasets show large gaps, especially TransUNet, which contains many parameters (421.47 M) but exhibits the worst results for meningioma segmentation, indicating that the complex ViT structure cannot achieve the desired performance in a small sample dataset. In addition, UCTransNet is able to maintain a relatively high segmentation accuracy in both datasets, but its parameter quantity is still too large (266.16 M). In contrast, the proposed GV-UNet achieved the highest segmentation accuracy in both datasets with the smallest parameters (50.20 M).

General medical image segmentation networks have potential applicability. However, their complex network structure may lead to the degradation of model performance due to the distinct boundary of meningioma in MRI. In addition, the smaller sample size can easily lead to overfitting of the general networks during training. To address the above problems, we redesigned the encoder structure to limit the downsampling size range to 512. The results demonstrate that the proposed GV-UNet can enhance feature extraction while maintaining the appropriate parameters.

1.3 A.3 Performance analysis of GV-UNet

As shown in Tables 1 and 2, the Dice score, IoU, HD95 and ASD of the proposed GV-UNet outperform the other SOTA models. As shown in Tables 3 and 4, the generalization ability of GV-UNet also outperforms their competitors for several reasons. First, GV-UNet expands the input image into 64-channel feature maps by a simple convolution layer, preserving numerous original features. Second, the large kernel convolution in the ConvMixer Layer is used to blend more distant spatial locations. Third, the MobileViT block is utilized to learn the global representation from different perspectives, which not only has the advantage of inductive bias inherent to CNNs but also enables adaptive weighting and global dependency of ViT. Last, the Ghost-CA layer generates many redundant maps to obtain the final feature information. Table 5 demonstrates the effectiveness of each module. Compared to the baseline network UNet, the Dice scores of GV-UNet improved by 4.4704% (Dataset A) and 2.1865% (Dataset B). As shown in Table 6, the proposed Equilibrium Loss outperforms all competing schemes because Equilibrium Loss not only has the advantage of Dice Loss global evaluation but also utilizes Varifocal Loss to mitigate the extreme imbalance between the foreground and the background.

1.4 A.4 Analysis on redundant maps

As shown in Fig. 2, we discovered that many redundant maps are generated during the feature extraction process. To verify whether these redundant maps contain feature information of the tumor, we performed ablation experiments. As shown in Table 5, the Dice scores of the model with only Ghost-CA added are 90.1068% and 89.0342% in the two datasets, with increases of 1.4409% and 0.1634%, respectively, over those of the baseline network, proving that the redundant maps can enhance the feature extraction capability of the model. Furthermore, Ghost-CA combines simple linear operations with coordinated attention to generate more redundant maps and thus orchestrate both remote dependencies and location capability to enhance the focus on the region of interest.

1.5 A.5 Limitations and future directions

The proposed GV-UNet can be developed into a powerful computer-assisted segmentation system for automating the segmentation process and easing the burden on clinical staff.

Presently, Ghost and ViT are utilized to balance parameters and accuracy. By adding the attention mechanism, the tumor region can be better localized. Although the proposed GV-UNet outperforms other models in terms of the number of parameters, the inherent structure of the transformer leads to the model’s floating points of operations (FLOPs) being unsatisfactory. In addition, our work only focuses on the segmentation performance of the model in T1-enhanced images without trying other modalities. To address these limitations, future work will focus on reducing the model complexity to achieve faster meningioma segmentation, as well as building novel segmentation models for T2, DWI and other modalities. To further address the clinical issues of meningioma, we will also work on the prediction of benign and malignant grades of meningioma by computer technology.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Bai, H., Zhang, Z., Yang, Y. et al. Meningioma segmentation with GV-UNet: a hybrid model using a ghost module and vision transformer. SIViP 18, 2377–2390 (2024). https://doi.org/10.1007/s11760-023-02914-3

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11760-023-02914-3

Keywords

Navigation