Skip to main content
Log in

A cycle generative adversarial network for generating synthetic contrast-enhanced computed tomographic images from non-contrast images in the internal jugular lymph node-bearing area

  • Original Article
  • Published:
Odontology Aims and scope Submit manuscript

Abstract

The objectives of this study were to create a mutual conversion system between contrast-enhanced computed tomography (CECT) and non-CECT images using a cycle generative adversarial network (cycleGAN) for the internal jugular region. Image patches were cropped from CT images in 25 patients who underwent both CECT and non-CECT imaging. Using a cycleGAN, synthetic CECT and non-CECT images were generated from original non-CECT and CECT images, respectively. The peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) were calculated. Visual Turing tests were used to determine whether oral and maxillofacial radiologists could tell the difference between synthetic versus original images, and receiver operating characteristic (ROC) analyses were used to assess the radiologists’ performances in discriminating lymph nodes from blood vessels. The PSNR of non-CECT images was higher than that of CECT images, while the SSIM was higher in CECT images. The Visual Turing test showed a higher perceptual quality in CECT images. The area under the ROC curve showed almost perfect performances in synthetic as well as original CECT images. In conclusion, synthetic CECT images created by cycleGAN appeared to have the potential to provide effective information in patients who could not receive contrast enhancement.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

Data availability

The dataset used in current study is available reasonable request to the corresponding author.

References

  1. Groell R, Willfurth P, Schaffler GJ, Mayer R, Schmidt F, Uggowitzer MM, et al. Contrast-enhanced spiral CT of the head and neck: comparison of contrast material injection rates. AJNR Am J Neuroradiol. 1999;20(9):1732–6.

    CAS  PubMed  PubMed Central  Google Scholar 

  2. Lakhal K, Ehrmann S, Robert-Edan V. Iodinated contrast medium: is there a re(n)al problem? a clinical vignette-based review. Crit Care. 2020;24(1):641. https://doi.org/10.1186/s13054-020-03365-9.

    Article  PubMed  PubMed Central  Google Scholar 

  3. Nielsen Y, Thomsen H. Optimal management of acute nonrenal adverse reactions to iodine-based contrast media Reports Med. Imaging. 2013;6:49e55.

    Google Scholar 

  4. Morzycki A, Bhatia A, Murphy KJ. Adverse reactions to contrast material: a canadian update. Can Assoc Radiol J. 2017;68(2):187–93. https://doi.org/10.1016/j.carj.2016.05.006.

    Article  PubMed  Google Scholar 

  5. Goodfellow J, Pougetbadie M, Mirza B, Xu D, Warde Farley S, Ozair A. Generative adversarial nets. In: Ghahramani Z, Welling M, Cortes C, Lawrence ND, Weinberger KQ, editors. Advances in Neural Information Processing Systems. New York, NY: ACM; 2014.

    Google Scholar 

  6. Zhu JY, Park T, Isola P, Efros AA. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy. 2017;22–29:2223–32.

    Google Scholar 

  7. Wang J, Zhao Y, Noble JH, Dawant BM. Conditional generative adversarial networks for metal artifact reduction in CT images of the ear. Med Image Comput Comput Assist Interv. 2018;11070:3–11. https://doi.org/10.1007/978-3-030-00928-1_1.

    Article  PubMed  PubMed Central  Google Scholar 

  8. Tang C, Li J, Wang L, Li Z, Jiang L, Cai A, et al. Unpaired low-dose CT denoising network based on cycle-consistent generative adversarial network with prior image information. Comput Math Methods Med. 2019;7(2019):8639825. https://doi.org/10.1155/2019/8639825.

    Article  Google Scholar 

  9. Wang J, Noble JH, Dawant BM. Metal artifact reduction for the segmentation of the intra cochlear anatomy in CT images of the ear with 3D-conditional GANs. Med Image Anal. 2019;58: 101553. https://doi.org/10.1016/j.media.2019.101553.

    Article  PubMed  PubMed Central  Google Scholar 

  10. Kida S, Kaji S, Nawa K, Imae T, Nakamoto T, Ozaki S, et al. Visual enhancement of cone-beam CT by use of CycleGAN. Med Phys. 2020;47(3):998–1010. https://doi.org/10.1002/mp.13963.

    Article  PubMed  Google Scholar 

  11. Kearney V, Ziemer BP, Perry A, Wang T, Chan JW, Ma L, et al. Attention-aware discrimination for MR-to-CT image translation using cycle-consistent generative adversarial networks. Radiol Artif Intell. 2020;2(2): e190027. https://doi.org/10.1148/ryai.2020190027.

    Article  PubMed  PubMed Central  Google Scholar 

  12. Li Z, Shi W, Xing Q, Miao Y, He W, Yang H, et al. Low-dose CT image denoising with improving WGAN and hybrid loss function. Comput Math Methods Med. 2021;26(2021):2973108. https://doi.org/10.1155/2021/2973108.

    Article  Google Scholar 

  13. Haubold J, Hosch R, Umutlu L, Wetter A, Haubold P, Radbruch A, et al. Contrast agent dose reduction in computed tomography with deep learning using a conditional generative adversarial network. Eur Radiol. 2021;31(8):6087–95. https://doi.org/10.1007/s00330-021-07714-2.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  14. Choi JW, Cho YJ, Ha JY, Lee SB, Lee S, Choi YH, et al. Generating synthetic contrast enhancement from non-contrast chest computed tomography using a generative adversarial network. Sci Rep. 2021;11(1):20403. https://doi.org/10.1038/s41598-021-00058-3.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  15. Denck J, Guehring J, Maier A, Rothgang E. MR-contrast-aware image-to-image translations with generative adversarial networks. Int J Comput Assist Radiol Surg. 2021;16(12):2069–78. https://doi.org/10.1007/s11548-021-02433-x.

    Article  PubMed  PubMed Central  Google Scholar 

  16. Gomi T, Kijima Y, Kobayashi T, Koibuchi Y. Evaluation of a generative adversarial network to improve image quality and reduce radiation-dose during digital breast tomosynthesis. Diagnostics (Basel). 2022;12(2):495. https://doi.org/10.3390/diagnostics12020495.

    Article  PubMed  Google Scholar 

  17. Huang Z, Zhang G, Lin J, Pang Y, Wang H, Bai T, et al. Multi-modal feature-fusion for CT metal artifact reduction using edge-enhanced generative adversarial networks. Comput Methods Programs Biomed. 2022;217: 106700. https://doi.org/10.1016/j.cmpb.2022.106700.

    Article  PubMed  Google Scholar 

  18. Wang X, Yu Z, Wang L, Zheng P. An enhanced priori knowledge GAN for CT images generation of early lung nodules with small-size labelled samples. Oxid Med Cell Longev. 2022;14(2022):2129303. https://doi.org/10.1155/2022/2129303.

    Article  Google Scholar 

  19. Li J, Qu Z, Yang Y, Zhang F, Li M, Hu S. TCGAN: a transformer-enhanced GAN for PET synthetic CT. Biomed Opt Express. 2022;13(11):6003–18. https://doi.org/10.1364/BOE.467683.

    Article  PubMed  PubMed Central  Google Scholar 

  20. Wang Z, Bovik AC, Sheikh HR, Simoncelli EP. Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process. 2004;13:600–12.

    Article  PubMed  Google Scholar 

  21. Fleiss JL. Measuring nominal scale agreement among many raters. Psychol Bull. 1971;76:378–82.

    Article  Google Scholar 

  22. Landis JR, Koch GG. The measurement of observer agreement for categorical data. Biometrics. 1977;33:159–74.

    Article  CAS  PubMed  Google Scholar 

  23. Hosmer DW Jr, Lemeshow S, Sturdivant RX. Applied logistic regression. 3rd ed. New York: Wiley; 2013. p. 177.

    Book  Google Scholar 

  24. Power M, Fell G, Wright M. Principles for high-quality, high-value testing. Evidence Based Medicine. 2013;18(1):5–10.

    Article  PubMed  PubMed Central  Google Scholar 

  25. Kim HS, Ha EG, Lee A, Choi YJ, Jeon KJ, Han SS, et al. Refinement of image quality in panoramic radiography using a generative adversarial network. Dentomaxillofac Radiol. 2023;52:20230007.

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

We thank Helen Jeays, BDSc AE, from Edanz (https://jp.edanz.com/ac) for editing a draft of this manuscript.

Author information

Authors and Affiliations

Authors

Contributions

Motoki Fukuda conducted the full experiment involving the deep learning method, compiled the results, and wrote the full manuscript. Yoshiko Ariji, Shinya Kotaki, and Michihito Nozawa evaluated the CT images as the observers. Chiaki Kuwada, Yoshitaka Kise, and Eiichiro Ariji supervised the whole experiment and provided important instructions and advice. All authors reviewed the manuscript.

Corresponding author

Correspondence to Motoki Fukuda.

Ethics declarations

Conflict of interest

All authors declare that they have no conflicts of interest. These include grants, patent licensing arrangements, consultancies, stock or other equity ownership, advisory board memberships, or payments for conducting or publicizing the study.

Human rights

This study was approved by Aichi Gakuin University’s ethics committee (approval no. 586). All procedures were in accordance with the ethical standards of the responsible committee on human experimentation (institutional and national) and with the 1975 Declaration of Helsinki, as amended in 2008. As this study was a retrospective study, informed consent could not be obtained from all patients. The purposes and methods of this study were made public, and patients were given the right to opt out of participation.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Fukuda, M., Kotaki, S., Nozawa, M. et al. A cycle generative adversarial network for generating synthetic contrast-enhanced computed tomographic images from non-contrast images in the internal jugular lymph node-bearing area. Odontology (2024). https://doi.org/10.1007/s10266-024-00933-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10266-024-00933-1

Keywords

Navigation