Abstract
This chapter presents five deep learning architectures for identification of Human papillomavirus (HPV) through generation of super-resolution (SR) images by fourfolds. Specifically, generative adversarial deep learning networks (GAN) and a texture-based vision transformer (TTSR) architecture are applied and evaluated. As such, the generated SR images are able to display the same way a high-resolution image offers in identification of HPV-like structures. In comparison, TTSR appears to perform the best with PSNR and SSIM being 28.70 and 0.8778, respectively, whereas 25.80/0.7910, 18.35/0.5059. 30.31/0.8013 and 28.07/0.6074 are observed for the methods of RCAN, Pix2Pix, CycleGAN and ESRGAN, respectively. With regard to sensitivity and specificity when detecting HPV clusters, TTSR also leads with 83.6% and 83.33%, respectively. It appears the computational SR images are capable to differentiate distinguishing features of HPV-like particles and to determine the effectiveness of anti-HPV agents, holding promise providing insights into the formation stage of a cancer from HPV in the near future. Latest version of May 2022.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Gao, X.W., Wen, X., Li, D., Liu, W., Xiong, J., Xu, B., Liu, J., Zhang, H., Liu, X.: Evaluation of GAN architectures for visualisation of HPV viruses from microscopic images. In: ICMLA 2021, Virtual, Dec. 13–16, 2021 (2021)
IARC Working Group on the Evaluation of Carcinogenic Risks to Humans, Human Papillomaviruses. Lyon (FR): International Agency for Research on Cancer. https://www.ncbi.nlm.nih.gov/books/NBK321770/. Retrieved in August 2020 (2007)
Braaten, K.P., Laufer, M.R.: Human papillomavirus (HPV), HPV-related disease, and the HPV vaccine. Rev. Obstet. Gynecol. 1(1), 2–10 (2008)
Chelimo, C., Wouldes, T.A., Cameron, L.D., Elwood, J.M.: Risk factors for and prevention of human papillomaviruses (HPV), genital warts and cervical cancer. J. Infect. 66, 207–217 (2013)
Song, B., Ding, C., Chen, W., Sun, H., Zhang, M., Chen, W.: Incidence and mortality of cervical cancer in China, 2013. Chin. J. Cancer Res. 29(6), 471–476 (2017)
Crewe, A.V., Isaacson, M., Johnson, D.: A simple scanning electron microscope. Rev. Sci. Instrum. 40(2), 241–246 (1969)
Born, M., Wolf, E.: Principles of Optics, 7th edn. Cambridge University Press, Cambridge, UK (1997)
Hell, S.W.: Microscopy and its focal switch. Nat. Methods 6, 24–32 (2009)
Rust, M.J., Bates, M., Zhuang, X.: Sub-diffraction-limit imaging by stochastic optical reconstruction microscopy (STORM). Nat. Methods 3, 793–795 (2006)
Hell, S.W.: Far-field optical nanoscopy. Science 316, 1153–1158 (2007)
Schermelleh, L., Ferrand, A., Huser, T., et al.: Super-resolution microscopy demystified. Nat. Cell Biol. 21, 72–84 (2019)
Wang, H., Rivenson, Y., Jin, Y., Wei, Z., Gao, R., Bentolila, L., Kural, C., Ozcan, A.: Deep learning enables cross-modality super-resolution in fluorescence microscopy. Nat. Methods 16, 103–110 (2019)
Dong, C., Loy, C.C., He, K., Tang, X.: Learning a deep convolutional network for image super-resolution. In: ECCV 2014 (2014)
Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative Adversarial Networks, Proceedings of the International Conference on Neural Information Processing Systems (NIPS 2014). pp. 2672–2680 (2014)
LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521, 436–444 (2015)
Kim, J., Lee, J.K., Lee, K.M.: Deeply-recursive convolutional network for image super-resolution. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016)
Haris, M., Shakhnarovich, G., Ukita, N.: Deep networks for super resolution. In: CVPR (2018)
Zhang, Y., Li, K., Li, K., Wang, L., Zhong, B., Fu, Y.: Image super-resolution using very deep residual channel attention networks. In: ECCV18 (2018)
Ledig, C., Theis, L., Huszár, F., et al.: Photo-realistic single image super-resolution using a generative adversarial network (2016). arXiv:1609.04802
Johnson, J., Alahi, A., Li, F.F.: Perceptual losses for real-time style transfer and super-resolution. In: ECCV’16 (2016)
Bruna, J., Sprechmann, P., LeCun, Y., Super-resolution with deep convolutional sufficient statistics. In: ICLR’15 (2015)
Wang, X., Yu, K., Dong, C., Loy, C.C.: Recovering realistic texture in image super-resolution by deep spatial feature transform. In: CVPR’18 (2018)
Wang, X., Yu, K., Wu, S., Gu, J., Liu, Y., Dong, C., Qiao, Y., Loy, C.C.: ESR-GAN: enhanced super-resolution generative adversarial networks. In: Proceeding of the ECCV Workshops (2018)
Jolicoeur-Martineau, A.: The relativistic discriminator: a key element missing from standard GAN (2018). arXiv:1807.00734
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR’16 (2016)
Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. In: ICLR 2021 (2021)
Gao, X.W., Wen, X., Li, D., Liu, W., Xiong, J., Xu, B., Liu, J., Zhang, H., Liu, X.: Detection of human papillomavirus (HPV) from super resolution microscopic images applying a texture transformer network. In: SPIE Medical Imaging, 20–24 Feb. 2022, San Diego, USA (2022)
Vaswani, A., Shazeer, N., Parmar, N., et al.: Attention is all you need. In NIPS (2017)
Brown, T.B., Mann, B., Ryder, N., Subbiah, M., et al.: Language models are few-shot learners (2020)
Yang, F., Yang, H., Fu, J., et al.: Learning texture transformer network for image super-resolution. In: CVPR 2020 (2020)
Zhu JY, Park T, Isola P, Efros AA, Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks, Computer Vision (ICCV), 2017 IEEE International Conference on, 2017.
Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: ICCV 2017 (2017)
Wang, T.C., Liu, M.Y., Zhu, J.Y., Tao, A., Kautz, J., Catanzaro, B.: High-resolution image synthesis and semantic manipulation with conditional GANs. In: CVPR 2018 (2018)
Pix2pixHD (2021). https://github.com/NVIDIA/pix2pixHD. Accessed in Jan. 2021
ESRGAN-Pytorch (2020). https://github.com/wonbeomjang/ESRGAN-pytorch. Retrieved in Aug. 2020
Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: ICLR’15 (2015)
Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., Courville, A.C.: Improved training of Wasserstein GANs. In: NeurIPS, pp. 5767–5777 (2017)
Zhang, Y., Li, K., Wang, L., et al.: Image super-resolution using very deep residual channel attention networks. In: ECCV 2018 (2018)
Netherton, C., Moffat, K., Brooks, E., Wileman, T.: A guide to viral inclusions, membrane rearrangements, factories, and viroplasm produced during virus replication. Adv. Virus Res. 70, 101–182 (2007)
Acknowledgements
This work is financially supported by The Royal Society in the UK (Ref: IEC\NSFC\181557). Their support is gratefully acknowledged. The authors would also like to thank Dr. Natalie Allcock from University of Leicester, UK for providing the service of acquisition of TEM data.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Appendices
Appendix A. Sample Preparation for Confocal Fluorescent Microscopy
The cervical cancer cell lines, CaSki containing HPV16 sequences and C33a without HPV, were used in this study. Both cell lines were grown on cover slips in a six-well culture plate containing RPMI culture media with 10% Fetal calf serum and 1% penicillin/streptomycin. They were kept in a humidified incubator with 95% air and 5% carbon dioxide for 2 days until the cells reached 70–80% confluence. The cells were washed by PBS three times with 1 min each time before they were fixed by 4% paraformaldehyde for 10 min. They were then exposed to 0.1% Triton-100 for 5 min following PBS washes. 50% house serum was then added for 8 min, then it was removed. Next, 200ul 1 in 100 dilution of anti-mouse HPV E6/E7 antibody (Abcam, UK) was added in and the cells were left at the room temperature for 2 h before they were washed again and 100ul biotinylated second antibody (ABC Universal Kit, Vector lab, UK) was added. After 30 min, the cells were washed again and then tertiary antibody was added and left for 20 min. Finally, 100 ul TSA/FITC amplification reagent exposed to the cells for 6 min in the darkness before the cells were washed. DAPI containing mounting media was added on the labelled slides and cells attached on the cover slips were sealed inside for microscopic viewing.
Appendix B. Sample Preparation for TEM Scanning
Cells were grown and prepared in a similar way as described above in Section S1 until the procedure reached at fixation step. Instead of fixing cells by paraformaldehyde, 2.5% glutaraldehyde in PBS buffer at pH 7 was used and the cells were fixed for 3Â h at RT before further steps taking place.
In addition, before the scanning, these samples undertake a series of standard preparation processes, including (1) flat embedding into EM capsules and polymerize for 6 h at 16 °C, (2) sectioned to 70 nm thick using a Leica UC7 ultramicrotome, (3) collected onto copper mesh grids and (4) stained in lead citrate for 5 min.
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this chapter
Cite this chapter
Gao, X.W. et al. (2023). Identification of Human Papillomavirus from Super-Resolution Microscopic Images Generated Using Deep Learning Architectures. In: Wani, M.A., Palade , V. (eds) Deep Learning Applications, Volume 4. Advances in Intelligent Systems and Computing, vol 1434. Springer, Singapore. https://doi.org/10.1007/978-981-19-6153-3_1
Download citation
DOI: https://doi.org/10.1007/978-981-19-6153-3_1
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-19-6152-6
Online ISBN: 978-981-19-6153-3
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)