Abstract
Medical imaging classification tasks require models that can provide high accuracy results. Training these models requires large annotated datasets. Such datasets are not openly available, are very costly, and annotations require professional knowledge in the medical domain. In the medical field specifically, datasets can also be inherently small. Self-supervised methods allow the construction of models that learn image representations on large unlabeled image sets; these models can then be fine-tuned on smaller datasets for various tasks. With breast cancer being a leading cause of death among women worldwide, precise lesion classification is crucial for detecting malignant cases. Through a set of experiments on 30K unlabeled mammography (MG) and ultrasound (US) breast images, we demonstrate a practical way to use self-supervised contrastive learning to improve breast cancer classification. Contrastive learning is a machine learning technique that teaches the model which data points are similar or different by using representations that force similar elements to be equal and dissimilar elements to be different. Our goal is to show the advantages of using self-supervised pre-training on a large unlabeled set, compared to training small sets from scratch. We compare training from scratch on small labeled MG and US datasets to using self-supervised contrastive methods and supervised pre-training. Our results demonstrate that the improvement in biopsy classification using self-supervision is consistent on both modalities. We show how to use self-supervised methods on medical data and propose a novel method of training contrastive learning on MG, which results in higher specificity classification.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Azizi, S., et al.: Big self-supervised models advance medical image classification. arXiv preprint arXiv:2101.05224 (2021)
Bray, F., Ferlay, J., Soerjomataram, I., Siegel, R.L., Torre, L.A., Jemal, A.: Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA: Cancer J. Clin. 68(6), 394–424 (2018)
Calli, E., Sogancioglu, E., Scholten, E.T., Murphy, K., van Ginneken, B.: Handling label noise through model confidence and uncertainty: application to chest radiograph classification. In: Mori, K., Hahn, H.K. (eds.) Medical Imaging 2019: Computer-Aided Diagnosis, vol. 10950, pp. 289–296. International Society for Optics and Photonics, SPIE (2019)
Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations. In: International Conference on Machine Learning, pp. 1597–1607. PMLR (2020)
Chen, X., Fan, H., Girshick, R., He, K.: Improved baselines with momentum contrastive learning. arXiv preprint arXiv:2003.04297 (2020)
DeLong, E.R., DeLong, D.M., Clarke-Pearson, D.L.: Comparing the areas under two or more correlated receiver operating characteristic curves: a nonparametric approach. Biometrics 837–845 (1988)
Hadad, O., Bakalo, R., Ben-Ari, R., Hashoul, S., Amit, G.: Classification of breast lesions using cross-modal deep learning. In: 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017), pp. 109–112. IEEE (2017)
He, K., Fan, H., Wu, Y., Xie, S., Girshick, R.: Momentum contrast for unsupervised visual representation learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9729–9738 (2020)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016)
Jamaludin, A., Kadir, T., Zisserman, A.: Self-supervised learning for spinal MRIs. In: Cardoso, M.J., et al. (eds.) DLMIA/ML-CDS -2017. LNCS, vol. 10553, pp. 294–302. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-67558-9_34
Khan, S., Islam, N., Jan, Z., Din, I.U., Rodrigues, J.J.P.C.: A novel deep learning based framework for the detection and classification of breast cancer using transfer learning. Pattern Recogn. Lett. 125, 1–6 (2019)
Lehman, C.D., Arao, R.F., et al.: National performance benchmarks for modern screening digital mammography: update from the breast cancer surveillance consortium. Radiology 283(1), 49–58 (2017)
Li, Y., Zhang, Y., Zhu, Z.: Learning deep networks under noisy labels for remote sensing image scene classification. In: IGARSS 2019–2019 IEEE International Geoscience and Remote Sensing Symposium, pp. 3025–3028 (2019)
McKinney, S.M., et al.: International evaluation of an AI system for breast cancer screening. Nature 577(7788), 89–94 (2020)
Oord, A.V.D., Li, Y., Vinyals, O.: Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748 (2018)
Schaffter, T., et al.: Evaluation of combined artificial intelligence and radiologist assessment to interpret screening mammograms. JAMA Netw. Open 3(3), e200265–e200265 (2020)
Shi, J., Zhou, S., Liu, X., Zhang, Q., Lu, M., Wang, T.: Stacked deep polynomial network based representation learning for tumor classification with small ultrasound image dataset. Neurocomputing 194, 87–94 (2016)
Szegedy, C., Ioffe, S., et al.: Inception-v4, Inception-ResNet and the impact of residual connections on learning (2016)
Tabár, L., et al.: The incidence of fatal breast cancer measures the increased effectiveness of therapy in women participating in mammography screening. Cancer 125(4), 515–523 (2019)
Zeimarani, B., Costa, M.G.F., Nurani, N.Z., Bianco, S.R., De Albuquerque Pereira, W.C., Filho, C.F.F.C.: Breast lesion classification in ultrasound images using deep convolutional neural network. IEEE Access 8, 133349–133359 (2020). https://doi.org/10.1109/ACCESS.2020.3010863
Zhang, X., et al.: Whole mammogram image classification with convolutional neural networks. In: 2017 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pp. 700–704 (2017)
Zhou, H.-Y., Yu, S., Bian, C., Hu, Y., Ma, K., Zheng, Y.: Comparing to learn: surpassing ImageNet pretraining on radiographs by comparing image representations. In: Martel, A.L., et al. (eds.) MICCAI 2020. LNCS, vol. 12261, pp. 398–407. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-59710-8_39
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Perek, S., Amit, M., Hexter, E. (2021). Self Supervised Contrastive Learning on Multiple Breast Modalities Boosts Classification Performance. In: Rekik, I., Adeli, E., Park, S.H., Schnabel, J. (eds) Predictive Intelligence in Medicine. PRIME 2021. Lecture Notes in Computer Science(), vol 12928. Springer, Cham. https://doi.org/10.1007/978-3-030-87602-9_11
Download citation
DOI: https://doi.org/10.1007/978-3-030-87602-9_11
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-87601-2
Online ISBN: 978-3-030-87602-9
eBook Packages: Computer ScienceComputer Science (R0)