Abstract
The large language model (LLM) ChatGPT-3.5, a member of generative pre-training transformer (GPT) models created by artificial intelligence (AI), is an updated and finely tuned version of previously launched AI chatbots. It is trained on a large volume of text data available on the internet, can produce human-like responses to a range of prompts and inquiries, and interprets and conversationally creates text, making it suitable for participating in interactive human conversations on a range of topics. Since its release in November 2022, ChatGPT has gained quick popularity not only among the general population but also among healthcare workers and researchers in all fields owing to its versatile, potent, and reliable benefits regarding scientific writing and medical education. Consequently, many studies have been done on ChatGPT regarding scientific research and medical education that have greatly enlightened medical professionals about its efficient use and probable risks. However, its effectiveness in healthcare and clinical medicine is still being determined, given its criticism regarding the authenticity of its diagnostic decisions. This review aims to highlight its viable usage in clinical practice, such as virtual assistance and patient communication, and the potential associated drawbacks, including a lack of human judgment and interactivity, data quality, accountability, and transparency, the risk of overreliance, and medico-legal and ethical considerations.
Similar content being viewed by others
References
Eysenbach, G.: The role of ChatGPT, generative language models, and artificial intelligence in medical education: a conversation with ChatGPT and a call for papers. JMIR Med. Educ. 6(9), e46885 (2023). https://doi.org/10.2196/46885. (PMID: 36863937; PMCID: PMC10028514)
Temsah, M.H., Jamal, A., Al-Tawfiq, J.A.: Reflection with ChatGPT about the excess death after the COVID-19 pandemic. New Microbes New Infect. 52, 101103 (2023). https://doi.org/10.1016/j.nmni.2023.101103. (Epub 2023 Feb 18. PMID: 36844147; PMCID: PMC9937725)
Xu, L., Sanders, L., Li, K., Chow, J.C.L.: Chatbot for health care and oncology applications using artificial intelligence and machine learning: systematic review. JMIR Cancer. 7(4), e27850 (2021). https://doi.org/10.2196/27850. (PMID: 34847056; PMCID: PMC8669585)
Limna, P., Kraiwanit, T., Jangjarat, K., Klayklung, P., Chocksathaporn, P.: The use of ChatGPT in the digital era: perspectives on chatbot implementation. J. Appl. Learn. Teach. (2023). https://doi.org/10.37074/jalt.2023.6.1.32
Gilson, A., Safranek, C.W., Huang, T., Socrates, V., Chi, L., Taylor, R.A., Chartash, D.: How does ChatGPT perform on the United States medical licensing examination? The implications of large language models for medical education and knowledge assessment. JMIR Med. Educ. 8(9), e45312 (2023). https://doi.org/10.2196/45312. (PMID: 36753318; PMCID: PMC9947764)
Strong, E., DiGiammarino, A., Weng, Y., Basaviah, P., Hosamani, P., Kumar, A., Nevins, A., Kugler, J., Hom, J., Chen, J.H.: Performance of ChatGPT on free-response, clinical reasoning exams. medRxiv (2023). https://doi.org/10.1101/2023.03.24.23287731. (PMID: 37034742; PMCID: PMC10081420)
Kitamura, F.C.: ChatGPT is shaping the future of medical writing but still requires human judgment. Radiology 307(2), e230171 (2023). https://doi.org/10.1148/radiol.230171. (Epub 2023 Feb 2. PMID: 36728749)
Khader, F., Han, T., Müller-Franzes, G., Huck, L., Schad, P., Keil, S., Barzakova, E., Schulze-Hagen, M., Pedersoli, F., Schulz, V., Zimmermann, M., Nebelung, L., Kather, J., Hamesch, K., Haarburger, C., Marx, G., Stegmaier, J., Kuhl, C., Bruners, P., Nebelung, S., Truhn, D.: Artificial intelligence for clinical interpretation of bedside chest radiographs. Radiology 307(1), e220510 (2023). https://doi.org/10.1148/radiol.220510. (Epub 2022 Dec 6. PMID: 36472534)
Shen, Y., Heacock, L., Elias, J., Hentel, K.D., Reig, B., Shih, G., Moy, L.: ChatGPT and other large language models are double-edged swords. Radiology 307(2), e230163 (2023). https://doi.org/10.1148/radiol.230163. (Epub 2023 Jan 26. PMID: 36700838)
Fayyad, U., Piatetsky-Shapiro, G., Smyth, P.: From data mining to knowledge discovery in databases. AI Mag. 17(3), 37–37 (1996)
Rezapour, M., KhavaninZadeh, M., Sepehri, M.M.: Implementation of predictive data mining techniques for identifying risk factors of early AVF failure in hemodialysis patients. Comput. Math. Methods Med.. Math. Methods Med. 2013, 830745 (2013). https://doi.org/10.1155/2013/830745. (Epub 2013 Jun 4. PMID: 23861725; PMCID: PMC3686075)
Sepehri, M. M., Khavaninzadeh, M., Rezapour, M., Teimourpour, B.: A data mining approach to fistula surgery failure analysis in hemodialysis patients. In: 2011 18th Iranian Conference of Biomedical Engineering (ICBME), Tehran, Iran, pp. 15–20, (2011) https://doi.org/10.1109/ICBME.2011.6168546
Rezapour, M., KhavaninZadeh, M., Sepehri, M.M., et al.: Less primary fistula failure in hypertensive patients. J. Hum. Hypertens.Hypertens. 32, 311–318 (2018). https://doi.org/10.1038/s41371-018-0052-3
Barash, Y., Klang, E., Konen, E., Sorin, V.: ChatGPT-4 assistance in optimizing emergency department radiology referrals and imaging selection. J. Am. Coll. Radiol.Radiol. 20(10), 998–1003 (2023)
Egli, A.: ChatGPT, GPT-4, and other large language models: the next revolution for clinical microbiology? Clin. Infect. Dis.. Infect. Dis. 77(9), 1322–1328 (2023)
Sallam, M.: ChatGPT utility in healthcare education, research, and practice: systematic review on the promising perspectives and valid concerns. Healthcare (Basel). 11(6), 887 (2023). https://doi.org/10.3390/healthcare11060887. (PMID: 36981544; PMCID: PMC10048148)
Fui-Hoon Nah, F., Zheng, R., Cai, J., Siau, K., Chen, L.: Generative AI and ChatGPT: applications, challenges, and AI-human collaboration. J. Inform. Technol Case Appl. Res. 25(3), 277–304 (2023)
Xue, V.W., Lei, P., Cho, W.C.: The potential impact of ChatGPT in clinical and translational medicine. Clin. Transl. Med.. Transl. Med. 13(3), e1216 (2023). https://doi.org/10.1002/ctm2.1216. (PMID: 36856370; PMCID: PMC9976604)
Palanica, A., Flaschner, P., Thommandram, A., Li, M., Fossat, Y.: Physicians’ perceptions of Chatbots in health care: cross-sectional web-based survey. J. Med. Internet Res. 21(4), e12887 (2019). https://doi.org/10.2196/12887. (PMID: 30950796; PMCID: PMC6473203)
Parviainen, J., Rantala, J.: Chatbot breakthrough in the 2020s? An ethical reflection on the trend of automated consultations in health care. Med. Health Care Philos. 25(1), 61–71 (2022). https://doi.org/10.1007/s11019-021-10049-w. (Epub 2021 Sep 4. PMID: 34480711; PMCID: PMC8416570)
Cascella, M., Montomoli, J., Bellini, V., Bignami, E.: Evaluating the feasibility of ChatGPT in healthcare: an analysis of multiple clinical and research scenarios. J. Med. Syst. 47(1), 33 (2023). https://doi.org/10.1007/s10916-023-01925-4. (PMID: 36869927; PMCID: PMC9985086)
Tools such as ChatGPT threaten transparent science; here are our ground rules for their use. Nature. 613(7945):612. (2023) https://doi.org/10.1038/d41586-023-00191-1(PMID: 36694020)
Stokel-Walker, C., Van Noorden, R.: What ChatGPT and generative AI mean for science. Nature 614(7947), 214–216 (2023)
Cahan, P., Treutlein, B.: A conversation with ChatGPT on the role of computational systems biology in stem cell research. Stem Cell Reports. 18(1), 1–2 (2023). https://doi.org/10.1016/j.stemcr.2022.12.009. (PMID: 36630899; PMCID: PMC9860153)
Borji, A.: A categorical archive of ChatGPT failures. arXiv preprint https://arXiv.org/2302.03494 (2023)
OpenAI OpenAI: Models GPT-3. https://beta.openai.com/docs/models. Accessed on 14 Jan 2023
Shaw, G.D.: ChatGPT will see you now: the promise and peril of chatbots in patient communication. Neurol. Today 23(14), 14–15 (2023)
Geis, J.R., Brady, A.P., Wu, C.C., Spencer, J., Ranschaert, E., Jaremko, J.L., Langer, S.G., Kitts, A.B., Birch, J., Shields, W.F., van den Hoven-van-Genderen, R., Kotter, E., Gichoya, J.W., Cook, T.S., Morgan, M.B., Tang, A., Safdar, N.M., Kohli, M.: Ethics of artificial intelligence in radiology: summary of the joint European and North American Multisociety statement. J. Am. Coll. Radiol.Radiol. 16(11), 1516–1521 (2019). https://doi.org/10.1016/j.jacr.2019.07.028
Jiang, L., Wu, Z., Xu, X., Zhan, Y., Jin, X., Wang, L., Qiu, Y.: Opportunities and challenges of artificial intelligence in the medical field: current application, emerging problems, and problem-solving strategies. J. Int. Med. Res. 49(3), 3000605211000157 (2021). https://doi.org/10.1177/03000605211000157. (PMID: 33771068; PMCID: PMC8165857)
Goisauf, M., Cano, A.M.: Ethics of AI in radiology: a review of ethical and societal implications. Front Big Data. 14(5), 850383 (2022). https://doi.org/10.3389/fdata.2022.850383. (PMID: 35910490; PMCID: PMC9329694)
Deng, J., Lin, Y.: The benefits and challenges of ChatGPT: an overview. Front. Comput. Intell. Syst. 2(2), 81–83 (2022)
Caliskan, A., Bryson, J.J., Narayanan, A.: Semantics derived automatically from language corpora contain human-like biases. Science 356(6334), 183–186 (2017)
Ma, X., Huo, Y.: Are users willing to embrace ChatGPT? Exploring the factors on the acceptance of chatbots from the perspective of AIDUA framework. Technol. Soc. 1(75), 102362 (2023)
Funding
None.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
None.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Alam, K., Kumar, A. & Samiullah, F.N.U. Prospectives and drawbacks of ChatGPT in healthcare and clinical medicine. AI Ethics (2024). https://doi.org/10.1007/s43681-024-00434-5
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s43681-024-00434-5