Abstract
Artificial intelligence is a chaotic field with conflicting terminology, philosophies, and architectures. This occurs in part due to the rapid merging of practitioners in computer graphics, image processing, and computer science. The entry of large corporations, attempting to simplify and streamline machine learning, creates an environment providing for reduced technical understanding by users. Competition for noteworthy breakthroughs drive rapid prototyping and early reporting. This collision yields a vivid spectrum and hidden risk. This chapter provides an orientation to important components and terminology in machine learning, plus painless examples of success and failure.
Do not include reference citations or undefined abbreviations in abstracts since these are often read independently of the actual chapter and without access to the reference list.
Similar content being viewed by others
References
Bzdok D, Altman N, Krzywinski M. Statistics versus machine learning. Nat Methods. 2018;15(4):233–4. https://doi.org/10.1038/nmeth.4642.
Ebigbo A, Palm C, Probst A, Mendel R, Manzeneder J, Prinz F, de Souza LA, Papa JP, Siersema P, Messmann H. A technical review of artificial intelligence as applied to gastrointestinal endoscopy: clarifying the terminology. Endosc Int Open. 2019;07(12):E1616–23. https://doi.org/10.1055/a-1010-5705.
Fournier A, Fussell D. Stochastic modeling in computer graphics. 1980. https://doi.org/10.1145/800250.807477.
Gu J, Oelke D. Understanding bias in machine learning. 1st Workshop on Visualization for AI Explainability in 2018 IEEE Vis, 2019..
Gunning D. DARPA’s explainable artificial intelligence (XAI) program. 2019. https://doi.org/10.1145/3301275.3308446.
Guo W. Explainable Artificial Intelligence for 6G: improving trust between human and machine. IEEE Commun Mag. 2020;58(6):39–45. https://doi.org/10.1109/mcom.001.2000050.
Kerasidou A. Artificial intelligence and the ongoing need for empathy, compassion and trust in healthcare. Bull World Health Organ. 2020;98(4):245–50. https://doi.org/10.2471/blt.19.237198.
Kumar RSS, O’Brien D, Albert K, Viljoen S, Snover J. Failure modes in machine learning systems. 2019.
Lapuschkin S, Wäldchen S, Binder A, Montavon G, Samek W, Müller K-R. Unmasking Clever Hans predictors and assessing what machines really learn. Nat Commun. 2019;10(1):1096. https://doi.org/10.1038/s41467-019-08987-4.
Norkin VI. Method of generalized gradient descent. Cybern Syst Anal. 1986;21(4):495–505. https://doi.org/10.1007/bf01070609.
Reynolds CW. Flocks, herds and schools: a distributed behavioral model. 1987. https://doi.org/10.1145/37401.37406.
Savage N. How AI is improving cancer diagnostics – Artificial intelligence can spot subtle patterns that can easily be missed by humans. Nature Outlook, 25 March 2020. https://doi.org/10.1038/d41586-020-00847-2.
Schönberger D. Artificial intelligence in healthcare: a critical analysis of the legal and ethical implications. Int J Law Inf Technol. 2019;27(2):171–203. https://doi.org/10.1093/ijlit/eaz004.
Yala A, et al. A deep learning mammography-based model for improved breast cancer risk prediction. Radiology. 2019;292:60–6. https://doi.org/10.1148/radiol.2019182716.
Yu AC, Eng J. One algorithm may not fit all: how selection bias affects machine learning performance. RadioGraphics. 2020;40(7):1932–7. https://doi.org/10.1148/rg.2020200040.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this entry
Cite this entry
Gospic, K.A.M., Passmore, G. (2021). Importance of AI in Medicine. In: Lidströmer, N., Ashrafian, H. (eds) Artificial Intelligence in Medicine. Springer, Cham. https://doi.org/10.1007/978-3-030-58080-3_277-1
Download citation
DOI: https://doi.org/10.1007/978-3-030-58080-3_277-1
Received:
Accepted:
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-58080-3
Online ISBN: 978-3-030-58080-3
eBook Packages: Springer Reference MedicineReference Module Medicine