Abstract
The use of technology has been increasing in health care settings. These health care technologies (e.g., avatar-assisted therapy) offer promising solutions to address accessibility among individuals who are facing mental health challenges. To increase awareness and encourage the use of these avatar-assisted therapies, two studies were conducted to determine the factors that could pose a barrier or an opportunity toward adopting such technology for mental health care interventions. The samples for these studies are individuals who are medically diagnosed patients. The sample for Study 1 was recruited from a crowdsourcing platform with screening questions that filter out suitable respondents. Study 1 tests a model that determines the role of mediators (trust and attitude) between the perceived usability of the avatar and an individual’s intention to use the avatar-assisted therapy. Study 2 is a proof of concept that serves as a program evaluation that validates the model and informs the use of avatar-assisted therapy. The sample for Study 2 was recruited from a local northwestern hospital. The results from the two studies show that individuals are open to using avatar-assisted therapy, but the content and the usability of the avatar need to meet their expectation for them to use the avatar-assisted therapy.
Similar content being viewed by others
References
Acharya, B., Hirachan, S., Mandel, J. S., et al. (2016). The mental health education gap among primary care providers in rural Nepal. Academic Psychiatry, 40, 667–671. https://doi.org/10.1007/s40596-016-0572-
Ajzen, I., & Fishbein, M. (1977). Attitude-behavior relations: A theoretical analysis and review of empirical research. Psychological Bulletin, 84(5), 888–918.
Andersson, G., Cuijpers, P., Carlbring, P., Riper, H., & Hedman, E. (2014). Guided Internet‐based vs. face‐to‐face cognitive behavior therapy for psychiatric and somatic disorders: A systematic review and meta‐analysis. World Psychiatry, 13(3), 288–295.
Arnberg, F. K., Linton, S. J., Hultcrantz, M., Heintz, E., & Jonsson, U. (2014). Internet-delivered psychological treatments for mood and anxiety disorders: A systematic review of their efficacy, safety, and cost-effectiveness. Plos One, 9(5). https://doi.org/10.1371/journal.pone.0098118
Bagozzi, R. P., & Yi, Y. (2012). Specification, evaluation, and interpretation of structural equation models. Journal of the Academy of Marketing Science, 40(1), 8–34.
Blut, M., Wang, C., Wünderlich, N. V., & Brock, C. (2021). Understanding anthropomorphism in service provision: A meta-analysis of physical robots, chatbots, and other AI. Journal of the Academy of Marketing Science, 1–27.
Bzdok, D., & Meyer-Lindenberg, A. (2018). Machine learning for precision psychiatry: Opportunities and challenges. Biological Psychiatry: Cognitive Neuroscience and Neuroimaging, 3(3), 223–230.
Carlbring, P., Andersson, G., Cuijpers, P., Riper, H., & Hedman-Lagerlöf, E. (2018). Internet-based vs. face-to-face cognitive behavior therapy for psychiatric and somatic disorders: An updated systematic review and meta-analysis. Cognitive Behaviour Therapy, 47(1), 1–18.
Crolic, C., Thomaz, F., Hadi, R., & Stephen, A. T. (2022). Blame the bot: Anthropomorphism and anger in customer–chatbot interactions. Journal of Marketing, 86(1), 132–148.
Davis, F. D. (1989). Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Quarterly, 319–340.
Deng, L., Turner, D. E., Gehling, R., & Prince, B. (2010). User experience, satisfaction, and continual usage intention of IT. European Journal of Information Systems, 19(1), 60–75.
De Visser, E. J., Monfort, S. S., McKendrick, R., Smith, M. A., McKnight, P. E., Krueger, F., & Parasuraman, R. (2016). Almost human: Anthropomorphism increases trust resilience in cognitive agents. Journal of Experimental Psychology: Applied, 22(3), 331.
Duffy, B. R. (2003). Anthropomorphism and the social robot. Robotics and Autonomous Systems, 42(3–4), 177–190.
Easton, C. J., Berbary, C. M., & Crane, C. A. (2018). Avatar and technology assisted platforms in the treatment of co-occurring addiction and IPV among male offenders. Advances in Dual Diagnosis, 11(3), 126–134.
Eastwood, J., Snook, B., & Luther, K. (2012). What people want from their professionals: Attitudes toward decision-making strategies. Journal of Behavioral Decision Making, 25(5), 458–468.
Fagan, M., Kilmon, C., & Pandey, V. (2012). Exploring the adoption of a virtual reality simulation: The role of perceived ease of use, perceived usefulness and personal innovativeness. Campus-Wide Information Systems, 29(2), 117–127.
Falconer, C. J., Davies, E. B., Grist, R., & Stallard, P. (2019). Innovations in practice: Avatar-based virtual reality in CAMHS talking therapy: Two exploratory case studies. Child and Adolescent Mental Health, 24(3), 283–287.
Gabbard, G. O., & Crisp-Han, H. (2017). The early career psychiatrist and the psychotherapeutic identity. Academic Psychiatry, 41(1), 30–34. https://doi.org/10.1007/s40596-016-0627-7
Goodhue, D. L., & Thompson, R. L. (1995). Task-technology fit and individual performance. MIS Quarterly, 213–236.
Graham, S., Depp, C., Lee, E. E., Nebeker, C., Tu, X., Kim, H. C., & Jeste, D. V. (2019). Artificial intelligence for mental health and mental illnesses: An overview. Current Psychiatry Reports, 21(11), 1–18.
Ha, Q. A., Chen, J. V., Uy, H. U., & Capistrano, E. P. (2021). Exploring the privacy concerns in using intelligent virtual assistants under perspectives of information sensitivity and anthropomorphism. International Journal of Human-computer Interaction, 37(6), 512–527.
Hair, J. F., Gabriel, M., & Patel, V. (2014). AMOS covariance-based structural equation modeling (CB-SEM): Guidelines on its application as a marketing research tool. Brazilian Journal of Marketing, 13(2), 44–55.
Hansen, N. B., Lambert, M. J., & Forman, E. M. (2002). The psychotherapy dose-response effect and its implications for treatment delivery services. Clinical Psychology: Science and Practice, 9(3), 329–343.
Hegner, S. M., Beldad, A. D., & Brunswick, G. J. (2019). In automatic we trust: Investigating the impact of trust, control, personality characteristics, and extrinsic and intrinsic motivations on the acceptance of autonomous vehicles. International Journal of Human-Computer Interaction, 35(19), 1769–1780.
Hoff, K. A., & Bashir, M. (2015). Trust in automation: Integrating empirical evidence on factors that influence trust. Human Factors, 57(3), 407–434.
Holden, R. J., Asan, O., Wozniak, E. M., et al. (2016). Nurses’ perceptions, acceptance, and use of a novel in-room pediatric ICU technology: Testing an expanded technology acceptance model. BMC Medical Informatics and Decision Making, 16, 145. https://doi.org/10.1186/s12911-016-0388-y
Hou, S. I., Charlery, S. A. R., & Roberson, K. (2014). Systematic literature review of Internet interventions across health behaviors. Health Psychology and Behavioral Medicine: an Open Access Journal, 2(1), 455–481.
Kaneshige, T., & Hong, D. (2018). Predictions 2019: This is the year to invest in humans, as backlash against chatbots and AI begins. Forrester, 8.
Kang, H. (2021). Sample size determination and power analysis using the G* Power software. Journal of Educational Evaluation for Health Professions, 18, 1–12.
Kaufman, E. A., McDonell, M. G., Cristofalo, M. A., & Ries, R. K. (2012). Exploring barriers to primary care for patients with severe mental illness: Frontline patient and provider accounts. Issues in Mental Health Nursing, 33(3), 172–180.
Keeffe, B., Subramanian, U., Tierney, W. M., Udris, E., Willems, J., McDonell, M., & Fihn, S. D. (2005). Provider response to computer-based care suggestions for chronic heart failure. Medical Care, 43(5), 461–465.
Kim, Y., & Sundar, S. S. (2012). Anthropomorphism of computers: Is it mindful or mindless? Computers in Human Behavior, 28(1), 241–250.
Knaevelsrud, C., & Maercker, A. (2007). Internet-based treatment for PTSD reduces distress and facilitates the development of a strong therapeutic alliance: A randomized controlled clinical trial. BMC Psychiatry, 7(1), 1–10.
Li, X., & Sung, Y. (2021). Anthropomorphism brings us closer: The mediating role of psychological distance in user–AI assistant interactions. Computers in Human Behavior, 118, 106680.
Lindell, M. K., & Whitney, D. J. (2001). Accounting for common method variance in cross-sectional research designs. Journal of Applied Psychology, 86(1), 114–121.
Lohr, S. (2016). IBM Is Counting on Its Bet on Watson, and Paying Big Money for It,” New York Times. Retrieved October 17, 2016, from https://www.nytimes.com/2016/10/17/technology/ibm-is-counting-on-its-bet-on-watson-and-paying-big-money-for-it.html
Lok, B., & Foster, A. E. (2019). Can virtual humans teach empathy? In Teaching empathy in healthcare (pp. 143–163). Springer, Cham.
Longoni, C., Bonezzi, A., & Morewedge, C. K. (2019). Resistance to medical artificial intelligence. Journal of Consumer Research, 46(4), 629–650.
McCormick, K. & Salcedo, J. (2017). Monte Carlo simulation and IBM SPSS bootstrapping. In SPSSreg Statistics for Data Analysis and Visualization (eds McCormick, K., and Salcedo, J.). https://doi.org/10.1002/9781119183426.ch2
Miller, D. D., Facp, C. M., & Brown, E. W. (2018). Artificial intelligence in medical practice: The question to the answer ? American Journal of Medicine, 131(2), 129–133. https://doi.org/10.1016/j.amjmed.2017.10.035
Moriuchi, E. (2021). An empirical study on anthropomorphism and engagement with disembodied AIs and consumers’ re-use behavior. Psychology & Marketing, 38(1), 21–42.
Müller, H., Lang, M., Wieser, B., Reihs, R., Pohn, B., Holzinger, A., & Zatloukal, K. (2019, July). Acceptance of virtual health avatars. In 2019 International Conference on Software Security and Assurance (ICSSA) (pp. 70–75). IEEE.
Nass, C., & Lee, K. M. (2001). Does computer-synthesized speech manifest personality? Experimental tests of recognition, similarity-attraction, and consistency-attraction. Journal of Experimental Psychology: Applied, 7(3), 171–181.
National Institute of Mental Health. (2022). Mental Health Information. Retrieved from https://www.nimh.nih.gov/health/statistics/mental-illness
Obermeyer, Z., & Emanuel, E. J. (2016). Predicting the future—Big data, machine learning, and clinical medicine. The New England Journal of Medicine, 375(13), 1216–1219.
Palan, S., & Schitter, C. (2018). Prolific. ac—A subject pool for online experiments. Journal of Behavioral and Experimental Finance, 17, 22–27.
Rozental, A., Kottorp, A., Forsström, D., Månsson, K., Boettcher, J., Andersson, G., & Carlbring, P. (2019). The negative effects questionnaire: Psychometric properties of an instrument for assessing negative effects in psychological treatments. Behavioural and Cognitive Psychotherapy, 47(5), 559–572.
Rozental, A., Magnusson, K., Boettcher, J., Andersson, G., & Carlbring, P. (2017). For better or worse: An individual patient data meta-analysis of deterioration among participants receiving Internet-based cognitive behavior therapy. Journal of Consulting and Clinical Psychology, 85(2), 160–177.
Safavi, K & Dare, F. (2018). Virtual health care could save the U.S. billions each year. Retrieved from https://hbr.org/2018/04/virtual-health-care-could-save-the-u-s-billions-each-year
Schanke, S., Burtch, G., & Ray, G. (2021). Estimating the impact of “humanizing” customer service chatbots. Information Systems Research, 32(3), 736–751.
Scherer, A., Wünderlich, N. V., & Von Wangenheim, F. (2015). The value of self-service: Long-term effects of technology-based self-service usage on customer retention. MIS Quarterly, 39(1), 177–200.
Siau, K., & Wang, W. (2020). Artificial intelligence (AI) ethics: Ethics of AI and ethical AI. Journal of Database Management (JDM), 31(2), 74–87. https://doi.org/10.4018/JDM.2020040105
The Lancent Global Health. (2020). Mental health matters. Retrieved from https://www.thelancet.com/journals/langlo/article/PIIS2214-109X(20)30432-0/fulltext
Venkatesh, V., Thong, J. Y., & Xu, X. (2012). Consumer acceptance and use of information technology: Extending the unified theory of acceptance and use of technology. MIS Quarterly, 157–178.
World Health Organization. (2022). Mental disorders. Retrieved from https://www.who.int/news-room/fact-sheets/detail/mental-disorders
Yokoi, R., Eguchi, Y., Fujita, T., & Nakayachi, K. (2021). Artificial intelligence is trusted less than a doctor in medical treatment decisions: Influence of perceived care and value similarity. International Journal of Human-Computer Interaction, 37(10), 981–990.
Zlotowski, J., Proudfoot, D., Yogeeswaran, K., & Bartneck, C. (2015). Opportunities and challenges in human–robot interaction. International Journal of Social Robotics, 7(3), 347–360.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Ethics Approval Statement
Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article. This study was approved by the RIT Institutional Review Board (IRB).
Conflict of Interest
The authors declare no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Moriuchi, E., Berbary, C. & Easton, C. Looking Through the Lenses of a Patient: An Empirical Study on the Factors Affecting Patients’ Intention to Use Avatar-assisted Therapy. J. technol. behav. sci. 8, 100–112 (2023). https://doi.org/10.1007/s41347-022-00298-8
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s41347-022-00298-8