Abstract
The diffusion of misinformation has garnered considerable attention in our society. As algorithms have been considered one of the major drivers behind the spread and amplification of misinformation, it is useful to understand the effects of these algorithms on misinformation sharing and the manner in which they spread it. This chapter examines the psychological, cognitive, and social factors involved in the processing of misinformation people receive through algorithms and artificial intelligence. Modeling cognitive processes has long been of interest for understanding user reasoning, and many theories from different fields have been formalized into cognitive models. Drawing on theoretical insights from information processing theory with the concept of diagnosticity, it examines how perceived normative values influence a user’s perceived diagnosticity and likelihood of sharing information and whether explainability further moderates this relationship. The findings showed that users with a high heuristic processing of normative values and positive diagnostic perception were more likely to proactively discern misinformation. Users with a high cognitive ability to understand information were more likely to discern it correctly and less likely to share misinformation online. When exposed to misinformation through algorithmic recommendations, users’ perceived diagnosticity of misinformation can be predicted accurately from their understanding of normative values. This perceived diagnosticity would then positively influence the accuracy and credibility of the misinformation. With this focus on misinformation processing, this chapter provides theoretical insights and relevant recommendations for firms to be more resilient in protecting themselves from the detrimental impact of misinformation.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Ahluwalia, R., Unnava, H. R., & Burnkrant, R. E. (2001). The moderating role of commitment on the spillover effect of marketing communications. Journal of Marketing Research, 38(4), 458–470. https://doi.org/10.1509/jmkr.38.4.458.18903
Ahmed, S., & Gil-Lopez, T. (2022). Engaging with vilifying stereotypes: The role of YouTube algorithmic use in perpetuating misinformation about Muslim congresswomen. Journalism & Mass Communication Quarterly. https://doi.org/10.1177/10776990221110113
Amazeeen, M., & Krishna, A. (2023). Processing vaccine misinformation: Recall and effects of source type on claim accuracy via perceived motivations and credibility. International Journal of Communication, 17, 560–582.
Barnoy, A., & Reich, Z. (2022). Trusting others: A Pareto distribution of source and message credibility among news reporters. Communication Research, 49(2), 196–220. https://doi.org/10.1177/0093650220911814
Borukhson, D., Lorenz-Spreen, P., & Ragni, M. (2022). When does an individual accept misinformation? An extended investigation through cognitive modeling. Computational Brain & Behavior, 5, 244–260. https://doi.org/10.1007/s42113-022-00136-3
Bryanov, K., Watson, B. K., Pingree, R. J., & Santia, M. (2020). Effects of partisan personalization in a news portal experiment. Public Opinion Quarterly, 84(S1), 216–235. https://doi.org/10.1093/poq/nfaa011
Chaiken, S. (1980). Heuristic versus systematic information processing and the use of source versus message cues in persuasion. Journal of Personality and Social Psychology, 39(5), 752–766. https://doi.org/10.1037/0022-3514.39.5.752
Chen, Z. F., & Cheng, Y. (2020). Consumer response to fake news about brands on social media: The effects of self-efficacy, media trust, and persuasion knowledge on brand trust. Journal of Product & Brand Management, 29(2), 188–198. https://doi.org/10.1108/JPBM-12-2018-2145
Cho, J., Ahmed, S., Hilbert, M., Liu, B., & Luu, J. (2020). Do search algorithms endanger democracy? An experimental investigation of algorithm effects on political polarization. Journal of Broadcasting & Electronic Media, 64(2), 150–172. https://doi.org/10.1080/08838151.2020.1757365
Clarke, C., & McComas, K. (2012). Seeking and processing influenza vaccine information: A study of health care workers at a large urban hospital. Health Communication, 27(3), 244–256. https://doi.org/10.1080/10410236.2011.578332
Cronbach, L. J. (1989). Construct validation after thirty years. In R. L. Linn (Ed.), Intelligence: Measurement, theory and public policy (pp. 147–171). University of Illinois Press.
Diakopoulos, N. (2016). Accountability in algorithmic decision making. Communications of the ACM, 59(2), 58–62. https://doi.org/10.1145/2844110
Diakopoulos, N., & Koliska, M. (2017). Algorithmic transparency in the news media. Digital Journalism, 5(7), 809–828. https://doi.org/10.1080/21670811.2016.1208053
Epstein, Z., Berinsky, A., Cole, R., Gully, A., Pennycook, G., & Rand, D. (2021). Developing an accuracy-prompt toolkit to reduce COVID-19 misinformation online. Harvard Kennedy School Misinformation Review, 2(3), 1–12. https://doi.org/10.37016/mr-2020-71
Gran, A., Booth, P., & Bucher, T. (2021). To be or not to be algorithm aware: A question of a new digital divide? Information, Communication & Society, 24(12), 1779–1796. https://doi.org/10.1080/1369118X.2020.1736124
Gunning, D., Stefik, M., Choi, J., Miller, T., Stumpf, S., & Yang, G. (2019). XAI: Explainable artificial intelligence. Science Robotics, 4(37), 7120. https://doi.org/10.1126/scirobotics.aay7120
Hair, J. F., Jr., Anderson, R. E., Tatham, R. L., & Black, W. C. (1995). Multivariate data analysis (3rd ed.). Macmillan Publishing Company.
Hermann, E. (2022). Artificial intelligence and mass personalization of communication content. New Media & Society, 24(5), 1258–1277. https://doi.org/10.1177/14614448211022702
Hwang, Y., & Jeong, S. (2021). Misinformation exposure and acceptance: The role of information seeking and processing. Health Communication, 23(1), 1–9. https://doi.org/10.1080/10410236.2021.1964187
Jahng, M. R. (2021). Is fake news the new social media crisis? Examining the public evaluation of crisis management for corporate organizations targeted in fake news. International Journal of Strategic Communication, 15(1), 18–36. https://doi.org/10.1080/1553118X.2020.1848842
Jiang, Z., & Benbasat, I. (2005). Virtual product experience: Effects of visual and functional control of products on perceived diagnosticity and flow in electronic shopping. Journal of Management Information Systems, 21(3), 111–147. https://doi.org/10.1080/07421222.2004.11045817
Joreskog, K., & Sorbom, D. (1996). LISREL 8: User’s reference guide. Scientific Software International.
Kim, H. K., Ahn, J., Atkinson, L., & Kahlor, L. A. (2020). Effects of COVID-19 misinformation on information seeking, avoidance, and processing: A multicountry comparative study. Science Communication, 42(5), 586–615. https://doi.org/10.1177/1075547020959670
Kim, H. K., & Tandoc, E. C., Jr. (2022). Consequences of online misinformation on COVID-19: Two potential pathways and disparity by eHealth literacy. Frontiers in Psychology, 13, 783909. https://doi.org/10.3389/fpsyg.2022.783909
Kreps, S., McCain, R., & Brundage, M. (2022). All the news that’s fit to fabricate: AI-generated text as a tool of media misinformation. Journal of Experimental Political Science, 9(1), 104–117. https://doi.org/10.1017/XPS.2020.37
Kwon, Y., Park, J., & Son, J.-Y. (2020). Accurately or accidentally? Recommendation agent and search experience in over-the-top (OTT) services. Internet Research, 31(2), 562–586. https://doi.org/10.1108/INTR-03-2020-0127
Lavigne, M., Bélanger, É., Nadeau, R., Daoust, J.-F., & Lachapelle, E. (2022). Hide and seek: The connection between false beliefs and perceptions of government transparency. Harvard Kennedy School Misinformation Review, 3(2). https://doi.org/10.37016/mr-2020-90
Lazer, D. M. J., Baum, M. A., Benkler, Y., Berinsky, A. J., Greenhill, K. M., Menczer, F., Metzger, M. J., Nyhan, B., Pennycook, G., Rothschild, D., Schudson, M., Sloman, S. A., Sunstein, C. R., Thorson, E. A., Watts, D. J., & Zittrain, J. (2018). The science of fake news. Science, 359(6380), 1094–1096. https://doi.org/10.1126/science.aao2998
Lewandowsky, S., & van der Linden, S. (2021). Countering misinformation and fake news through inoculation and prebunking. European Review of Social Psychology, 32(2), 348–384. https://doi.org/10.1080/10463283.2021.1876983
Lynch, J., Jr. (2006). Accessibility-diagnosticity and the multiple pathway anchoring and adjustment model. Journal of Consumer Research, 33(1), 25–27. https://doi.org/10.1086/504129
Metzger, M., Flanagin, A., & Nekmat, E. (2015). Comparative optimism in online credibility evaluation among parents and children. Journal of Broadcasting & Electronic Media, 59(3), 509–529. https://doi.org/10.1080/08838151.2015.1054995
Mhasawade, V., Zhao, Y., & Chunara, R. (2021). Machine learning and algorithmic fairness in public and population health. Nature Machine Intelligence, 3, 659–666. https://doi.org/10.1038/s42256-021-00373-4
Oh, H. J., & Lee, H. (2019). When do people verify and share health rumors on social media? The effects of message importance, health anxiety, and health literacy. Journal of Health Communication, 24(11), 837–847. https://doi.org/10.1080/10810730.2019.1677824
Peifer, J. T., & Meisinger, J. (2021). The value of explaining the process: How journalistic transparency and perceptions of news media importance can (sometimes) foster message credibility and engagement intentions. Journalism & Mass Communication Quarterly, 98(3), 828–853. https://doi.org/10.1177/10776990211012953
Pennycook, G., Cannon, T. D., & Rand, D. G. (2018). Prior exposure increases perceived accuracy of fake news. Journal of Experimental Psychology, 147(12), 1865–1880. https://doi.org/10.1037/xge0000465
Pennycook, G., & Rand, D. G. (2021). The psychology of fake news. Trends in Cognitive Sciences, 25(5), 388–402. https://doi.org/10.1016/j.tics.2021.02.007
Pennycook, G., & Rand, D. G. (2022). Accuracy prompts are a replicable and generalizable approach for reducing the spread of misinformation. Nature Communications, 13, 2333. https://doi.org/10.1038/s41467-022-30073-5
Rabb, N., Cowen, L., de Ruiter, J., & Scheutz, M. (2022). Cognitive cascades: How to model (and potentially counter) the spread of fake news. PLoS ONE, 17(1), e0261811. https://doi.org/10.1371/journal.pone.0261811
Rai, A. (2020). Explainable AI: From black box to glass box. Journal of the Academy of Marketing Science, 48, 137–141. https://doi.org/10.1007/s11747-019-00710-5
Schuetz, S. W., Sykes, T. A., & Venkatesh, V. (2021). Combating COVID-19 fake news on social media through fact checking: Antecedents and consequences. European Journal of Information Systems, 30(4), 376–388. https://doi.org/10.1080/0960085X.2021.1895682
Shin, D. (2021). The perception of humanness in conversational journalism. New Media & Society, 24(12), 2680–2704. https://doi.org/10.1177/1461444821993801
Shin, D., Zaid, B., Biocca, F., & Rasul, A. (2022). In platforms we trust? Unlocking the black-box of news algorithms through interpretable AI. Journal of Broadcasting & Electronic Media, 66(2), 235–256. https://doi.org/10.1080/08838151.2022.2057984
Sundar, S., Molina, M., & Cho, E. (2021). Seeing is believing: Is video modality more potent in spreading fake news via online messaging apps? Journal of Computer-Mediated Communication, 26(6), 301–319. https://doi.org/10.1093/jcmc/zmab010
Talwar, S., Dhir, A., Singh, D., Virk, G. S., & Salo, J. (2020). Sharing of fake news on social media: Application of the honeycomb framework and the third-person effect hypothesis. Journal of Retailing and Consumer Services, 57, 102197. https://doi.org/10.1016/j.jretconser.2020.102197
Tandoc, E. C., Jr. (2019). The facts of fake news: A research review. Sociology Compass, 13, 1–9. https://doi.org/10.1111/soc4.12724
Tully, M., Bode, L., & Vraga, E. K. (2020). Mobilizing users: Does exposure to misinformation and its correction affect users’ responses to a health misinformation post? Social Media + Society, 6(4). https://doi.org/10.1177/2056305120978377
Vraga, E. K., & Bode, L. (2020). Defining misinformation and understanding its bounded nature: Using expertise and evidence for describing misinformation. Political Communication, 37(1), 136–144. https://doi.org/10.1080/10584609.2020.1716500
Walter, N., & Tukachinsky, R. (2020). A meta-analytic examination of the continued influence of misinformation in the face of correction: How powerful is it, why does it happen, and how to stop it? Communication Research, 47, 155–177. https://doi.org/10.1177/0093650219854600
Wathen, C., & Burkell, J. (2002). Believe it or not: Factors influencing credibility on the Web. Journal of the American society for information science and technology, 53(2), 134–144.
Wilson, T. D., & Maceviciute, E. (2022). Information misbehavior: Modeling the motivations for the creation, acceptance and dissemination of misinformation. Journal of Documentation, 78(7), 485–505. https://doi.org/10.1108/JD-05-2022-0116
Wu, Y. (2018). How age affects journalists’ adoption of social media as an innovation. Journalism Practice, 13(5), 537–557. https://doi.org/10.1080/17512786.2018.1511821
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this chapter
Cite this chapter
Shin, D. (2024). Misinformation Processing Model: How Users Process Misinformation When Using Recommender Algorithms. In: Artificial Misinformation. Palgrave Macmillan, Cham. https://doi.org/10.1007/978-3-031-52569-8_5
Download citation
DOI: https://doi.org/10.1007/978-3-031-52569-8_5
Published:
Publisher Name: Palgrave Macmillan, Cham
Print ISBN: 978-3-031-52568-1
Online ISBN: 978-3-031-52569-8
eBook Packages: Literature, Cultural and Media StudiesLiterature, Cultural and Media Studies (R0)