Skip to main content

Advertisement

Log in

Effect of Training Data Volume on Performance of Convolutional Neural Network Pneumothorax Classifiers

  • Published:
Journal of Digital Imaging Aims and scope Submit manuscript

Abstract

Large datasets with high-quality labels required to train deep neural networks are challenging to obtain in the radiology domain. This work investigates the effect of training dataset size on the performance of deep learning classifiers, focusing on chest radiograph pneumothorax detection as a proxy visual task in the radiology domain. Two open-source datasets (ChestX-ray14 and CheXpert) comprising 291,454 images were merged and convolutional neural networks trained with stepwise increase in training dataset sizes. Model iterations at each dataset volume were evaluated on an external test set of 525 emergency department chest radiographs. Learning curve analysis was performed to fit the observed AUCs for all models generated. For all three network architectures tested, model AUCs and accuracy increased rapidly from 2 × 103 to 20 × 103 training samples, with more gradual increase until the maximum training dataset size of 291 × 103 images. AUCs for models trained with the maximum tested dataset size of 291 × 103 images were significantly higher than models trained with 20 × 103 images: ResNet-50: AUC20k = 0.86, AUC291k = 0.95, p < 0.001; DenseNet-121 AUC20k = 0.85, AUC291k = 0.93, p < 0.001; EfficientNet AUC20k = 0.92, AUC 291 k = 0.98, p < 0.001. Our study established learning curves describing the relationship between dataset training size and model performance of deep learning convolutional neural networks applied to a typical radiology binary classification task. These curves suggest a point of diminishing performance returns for increasing training data volumes, which algorithm developers should consider given the high costs of obtaining and labelling radiology data.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1

source chest radiograph datasets. NLP = natural language processing

Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

Availability of Data and Material

Open-source training data as described in the “Materials and Methods” section.

Code Availability

Available on Github.

References

  1. Sun, C., Shrivastava, A., Singh, S., & Gupta, A. Revisiting unreasonable effectiveness of data in deep learning era. In Proceedings of the IEEE international conference on computer vision pp. 843–852, 2017.

  2. Alwosheel A, van Cranenburgh S, Chorus CG. Is your dataset big enough? Sample size requirements when using artificial neural networks for discrete choice analysis. Journal of choice modelling. 28:167-82, 2018.

    Article  Google Scholar 

  3. Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., & Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition pp. 248–255, 2009.

  4. Parkhi, O.M., Vedaldi, A., & Zisserman, A. Deep face recognition. In bmvc, vol. 1, p.6, 2015.

    Google Scholar 

  5. Kelly CJ, Karthikesalingam A, Suleyman M, Corrado G, King D. Key challenges for delivering clinical impact with artificial intelligence. BMC medicine. 17:1-9 2019.

    Article  CAS  Google Scholar 

  6. Price WN, Cohen IG. Privacy in the age of medical big data. Nat Med. 25:37-43, 2019.

    Article  CAS  Google Scholar 

  7. Prevedello LM, Halabi SS, Shih G, Wu CC, Kohli MD, Chokshi FH, Erickson BJ, Kalpathy-Cramer J, Andriole KP, Flanders AE. Challenges related to artificial intelligence research in medical imaging and the importance of image analysis competitions. Radiol Artif Intell. 1:e180031, 2019.

  8. Willemink MJ, Koszek WA, Hardell C, Wu J, Fleischmann D, Harvey H, Folio LR, Summers RM, Rubin DL, Lungren MP. Preparing medical imaging data for machine learning. Radiology. 295:4-15, 2020.

    Article  Google Scholar 

  9. Kim DH, MacKinnon T. Artificial intelligence in fracture detection: transfer learning from deep convolutional neural networks. Clin Radiol. 73:439-45, 2018.

    Article  CAS  Google Scholar 

  10. Majkowska A, Mittal S, Steiner DF, Reicher JJ, McKinney SM, Duggan GE, Eswaran K, Cameron Chen PH, Liu Y, Kalidindi SR, Ding A. Chest radiograph interpretation with deep learning models: assessment with radiologist-adjudicated reference standards and population-adjusted evaluation. Radiology. 294:421-31, 2020.

    Article  Google Scholar 

  11. Cho J, Lee K, Shin E, Choy G, Do S. How much data is needed to train a medical image deep learning system to achieve necessary high accuracy? arXiv Prepr arXiv151106348, 2015.

  12. Narayana, P. A., Coronado, I., Sujit, S. J., Wolinsky, J. S., Lublin, F. D., & Gabr, R. E. Deep-Learning-Based Neural Tissue Segmentation of MRI in Multiple Sclerosis: Effect of Training Set Size. J Magn Reson Imaging. 51:1487–1496, 2020.

    Article  Google Scholar 

  13. Wang X, Peng Y, Lu L, Lu Z, Bagheri M, Summers R. Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. InIEEE CVPR pp. 3462–3471, 2017.

  14. Irvin J, Rajpurkar P, Ko M, Yu Y, Ciurea-Ilcus S, Chute C, Marklund H, Haghgoo B, Ball R, Shpanskaya K, Seekins J. Chexpert: A large chest radiograph dataset with uncertainty labels and expert comparison. In Proceedings of the AAAI Conference on Artificial Intelligenc. 33: 590-597, 2019.

    Article  Google Scholar 

  15. Oakden-Rayner L. Exploring large-scale public medical image datasets. Acad Radiol. 27:106-12, 2020.

    Article  Google Scholar 

  16. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition 770–778, 2016.

  17. Huang G, Liu Z, Van Der Maaten L, Weinberger KQ. Densely connected convolutional networks. In Proceedings of the IEEE conference on computer vision and pattern recognition 4700–4708, 2017.

  18. Tan M, Le Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In International Conference on Machine Learning 6105–6114, 2019.

  19. Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference On Computer Vision 618–626, 2017.

  20. Krawczyk B. Learning from imbalanced data: open challenges and future directions. Progress in Artificial Intelligence. 5:221-32, 2016.

    Article  Google Scholar 

  21. Kohli MD, Summers RM, Geis JR. Medical image data and datasets in the era of machine learning—whitepaper from the 2016 C-MIMI meeting dataset session. J Digit Imaging 30:392-9, 2017

    Article  Google Scholar 

  22. Figueroa RL, Zeng-Treitler Q, Kandula S, Ngo LH. Predicting sample size required for classification performance. BMC Med Inform Decis Mak. 12:8, 2012.

    Article  Google Scholar 

  23. Hestness J, Narang S, Ardalani N, Diamos G, Jun H, Kianinejad H, Patwary M, Ali M, Yang Y, Zhou Y. Deep learning scaling is predictable, empirically. arXiv preprint arXiv:1712.00409, 2017 .

  24. Balki I, Amirabadi A, Levman J, Martel AL, Emersic Z, Meden B, Garcia-Pedrero A, Ramirez SC, Kong D, Moody AR, Tyrrell PN. Sample-size determination methodologies for machine learning in medical imaging research: a systematic review. Can Assoc of Radiol J.;70:344-53, 2019.

    Article  Google Scholar 

  25. Dunnmon JA, Yi D, Langlotz CP, Ré C, Rubin DL, Lungren MP. Assessment of convolutional neural networks for automated classification of chest radiographs. Radiology. 290:537-44, 2019

    Article  Google Scholar 

  26. Krause J, Gebru T, Deng J, Li LJ, Fei-Fei L. Learning features and parts for fine-grained recognition. In 2014 22nd International Conference on Pattern Recognition. 26–33, 2014. IEEE.

Download references

Funding

This research was supported by the NUHS Internal Grant Funding under NUHS Seed Fund (NUHSRO/2018/097/R05 + 5/Seed-Nov/07), NUHS-NHIC Joint MedTech Grant (NUHS-NHIC MT2020-02), NUHSRO/2018/019/RO5 + 5/NUHS), and NMRC Health Service Research Grant (HSRG-OC17nov004).

Author information

Authors and Affiliations

Authors

Contributions

Study design, YLT, DWN, MF; data acquisition, YLT, JTPDH, PJ, SYS, JSAM, QST; data analysis, YLT, DWN, JTPDH, MF; literature search, YLT, DWN, MF; clinical studies, YLT, DWN, JTPDH, PJ, SYS, JSAM, QST, MF; statistical analysis, YLT, DWN, MF; manuscript editing, YLT, DWN, JTPDH, PJ, SYS, JSAM, QST, MF.

Corresponding author

Correspondence to Yee Liang Thian.

Ethics declarations

Conflict of Interest

The authors declare no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Thian, Y.L., Ng, D.W., Hallinan, J.T.P.D. et al. Effect of Training Data Volume on Performance of Convolutional Neural Network Pneumothorax Classifiers. J Digit Imaging 35, 881–892 (2022). https://doi.org/10.1007/s10278-022-00594-y

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10278-022-00594-y

Keywords

Navigation