Abstract
The use of AI in weapons systems raises numerous ethical issues. To date, work on weaponized AI has tended to be theoretical and normative in nature, consisting in critical policy analyses and ethical considerations, carried out by philosophers, legal scholars, and political scientists. However, adequately addressing the cultural and social dimensions of technology requires insights and methods from empirical moral and cultural psychology. To do so, this position piece describes the motivations for and sketches the nature of a normative, cultural psychology of weaponized AI. The motivations for this project include the increasingly global, cross-cultural and international, nature of technologies, and counter-intuitive nature of normative thoughts and behaviors. The nature of this project consists in developing standardized measures of AI ethical reasoning and intuitions, coupled with questions exploring the development of norms, administered and validated across different cultural groups and disciplinary contexts. The goal of this piece is not to provide a comprehensive framework for understanding the cultural facets and psychological dimensions of weaponized AI but, rather, to outline in broad terms the contours of an emerging research agenda.
Similar content being viewed by others
Notes
This characterization is based on (Bicchieri, 2016; Chudek & Henrich, 2011; Sripada & Stich, 2007). However, debate exists concerning the nature of norms/norm systems. See (Kelly & Setman, 2020) for a helpful overview. Understandings of norms in the field of international relations – in which one of the authors does most of her work – are quite different from the characterization provided above. See (Bode & Huelss, 2022) for an account oriented in scholarship on international relations.
The importance attached to values, one could argue, results from Western-centric, individualistic biases, where the pursuit of preferred states is given exaggerated importance in decision-making, rather than, for example, expectations about what others are doing or expect one to do (Bicchieri, 2016). One might argue that expectations about what others are doing or expect one to do is indicative of a preferred value, for example, harmony or community. However, this response supposes that individuals initially stand apart from/are different from groups. Independent conceptions of personhood such as these are in the cultural and historical minority. Most peoples, in most places, throughout most of time have conceived of themselves in interdependent terms, as thoroughly embedded in communities (Henrich, 2020; Nisbett, 2010).
In reality, this might not always be the case. For example, international law is affected by diplomatic negotiations, which always involve a degree of ambiguity.
References
Acharya, A. (2004). How ideas spread: whose norms matter? Norm localization and institutional change in asian regionalism. International Organization, 58(2), 239–275. https://doi.org/10.1017/S0020818304582024.
Bazerman, M. H., & Tenbrunsel, A. (2012). Blind spots: why we fail to do what’s right and what to do about it. Princeton: Princeton University Press.
Bebeau, M. J. (2002). The defining issues test and the four component model: contributions to professional education. Journal of Moral Education, 31(3), 271–295. https://doi.org/10.1080/0305724022000008115.
Bhuta, N., Beck, S., Geiss, R., Liu, H. Y., & Kress, C. (Eds.). (2016). Autonomous Weapons Systems: Law, Ethics, Policy. Cambridge: Cambridge University Press.
Bicchieri, C. (2016). Norms in the Wild. Oxford: Oxford University Press.
Bode, I., & Huelss, H. (2022). Autonomous Weapon Systems and International norms. Montreal: McGill-Queen’s University Press.
Chudek, M., & Henrich, J. (2011). Culture-gene coevolution, norm-psychology and the emergence of human prosociality. Trends in Cognitive Sciences, 15(5), 218–226. https://doi.org/10.1016/j.tics.2011.03.003.
Clancy, R. F., & Zhu, Q. (2021). Global Engineering Ethics: What? Why? How? and When? In ASEE Annual Conference and Exposition, Conference Proceedings.
Clancy, R. F., Zhu, Q., Martin, D. A., & Bombaerts, G. (2022). From Value- to Norm-sensitive Design? An Empirical and Intercultural Framework. ASEE Annual Conference and Exposition, Conference Proceedings.
Coeckelbergh, M. (2020). AI Ethics. Cambridge: MIT Press.
Crootof, R. (2015). The Killer Robots are Here: legal and policy implications. Cardozo Law Review, 36, 1837–1915.
Davidson, L. J., & Kelly, D. (2018). Minding the gap: Bias, Soft Structures, and the double life of social norms. Journal of Applied Philosophy, 1–20. https://doi.org/10.1111/japp.12351.
Dennis, M. J., & Clancy, R. F. (2022). Intercultural Ethics for Digital Well-Being: identifying problems and exploring solutions. Digital Society, 1(1), 1–16. https://doi.org/10.1007/s44206-022-00006-2.
Friedman, B., & Hendry, D. G. (2019). Value Sensitive Design: Shaping Technology with Moral Imagination. Cambridge: MIT Press.
Gelfand, M. J. (2018). Rule makers, Rule Breakers: tight and loose cultures and the secret signals that direct our lives. New York: Schribner.
Ghotbi, N., & Ho, M. T. (2021). Moral Awareness of College Students regarding Artificial Intelligence. Asian Bioethics Review, 13(4), 421–433. https://doi.org/10.1007/s41649-021-00182-2.
Graham, J., Nosek, B. A., Haidt, J., Iyer, R., Koleva, S., & Ditto, P. H. (2011). Mapping the Moral Domain. Journal of Personality and Social Psychology, 101(2), 366–385. https://doi.org/10.1037/a0021847.
Greene, J. D. (2014). Moral Tribes: emotion, reason, and the gap between us and them. New York: Penguin Books.
Greene, J. D., Cushman, F. A., Stewart, L. E., Lowenberg, K., Nystrom, L. E., & Cohen, J. D. (2009). Pushing moral buttons: the interaction between personal force and intention in moral judgment. Cognition, 111(3), 364–371. https://doi.org/10.1016/j.cognition.2009.02.001.
Hagendorff, T. (2020). The Ethics of AI Ethics: an evaluation of Guidelines. Minds and Machines, 30(1), 99–120. https://doi.org/10.1007/s11023-020-09517-8.
Haidt, J. (2012). The Righteous mind. New York: Vintage Press.
Henrich, J. (2015a). Culture and Social Behavior. Current Opinion in Behavioral Sciences, 3, 84–89. https://doi.org/10.1016/j.cobeha.2015.02.001.
Henrich, J. (2015b). The secret to our success: how culture is driving human evolution, domesticating our species, and making us smarter. Princeton: Princeton University Press.
Henrich, J. (2020). The WEIRDest people in the World: how the West Became psychologically peculiar and particularly prosperous. New York: Farrar, Straus and Giroux.
Henrich, J., Heine, S. J., & Norenzayan, A. (2010). The Weirdest People in the World? Behavioral and Brain Sciences, 33(2–3), 61–83. https://doi.org/10.1017/S0140525X0999152X.
Horowitz, M. C. (2016). The ethics & morality of robotic warfare: assessing the debate over autonomous weapons. Daedalus, 145(4), 25–36. https://doi.org/10.1162/DAED_a_00409.
Kelly, D., & Davis, T. (2018). Social norms and human normative psychology. Social Philosophy & Policy, 35(1), 54–76.
Kelly, D., & Setman, S. (2020). The Psychology of Normative Cognition. In E. N. Zalta (Ed.), The Stanford Encyclopedia of Philosophy. Retrieved from https://plato.stanford.edu/archives/fall2020/entries/psychology-normative-cognition/
Knafo, A., Schwartz, S. H., & Levine, R. V. (2009). Helping strangers is lower in embedded cultures. Journal of Cross-Cultural Psychology, 40(5), 875–879. https://doi.org/10.1177/0022022109339211.
Kulich, S. J., & Zhang, R. (2012). The multiple frames of “Chinese” values: from tradition to modernity and beyond. In M. H. Bond (Ed.), Oxford Handbook of chinese psychology (pp. 241–278). Oxford: Oxford University Press. https://doi.org/10.1093/oxfordhb/9780199541850.013.0017.
Leeman, R. F., Fischler, C., & Rozin, P. (2011). Medical doctors’ attitudes and beliefs about diet and health are more like those of their lay countrymen (France, Germany, Italy, UK and USA) than those of doctors in other countries. Appetite, 56, 558–563. https://doi.org/10.1016/j.appet.2011.01.022.
Luegenbiehl, H. C. (2010). Ethical principles for engineers in a global environment. In Van de I. Poel, & D. E. Goldberg (Eds.), Philosophy and Engineering: an emerging agenda (pp. 147–159). Dordrecht: Springer.
Luegenbiehl, H. C., & Clancy, R. F. (2017). Global Engineering Ethics. New York: Elsevier.
Mauri, D. (2022). Autonomous Weapons Systems and the Protection of the human person: an International Law Analysis. Northampton: Edward Elgar Publishing.
Müller, V. C. (2020). Ethics of Artificial Intelligence and Robotics. In Stanford Encyclopedia of Philosophy.
Narvaez, D., & Bock, T. (2002). Moral schemas and tacit judgement or how the defining issues test is supported by cognitive science. Journal of Moral Education, 31(3), 297–314. https://doi.org/10.1080/0305724022000008124.
Nisbett, R. E. (2010). The Geography of Thought: how Asians and Westerners think differently and why. New York: Free Press.
Ransohoff, K. J. (2011). Patients on the trolley track: the moral cognition of medical practitioners and public health professionals. Harvard University.
Reiber, C., & Garcia, J. R. (2010). Hooking up: gender differences, evolution, and pluralistic ignorance. Evolutionary Psychology. https://doi.org/10.1177/147470491000800307.
Rességuier, A., & Rodrigues, R. (2020). AI ethics should not remain toothless! A call to bring back the teeth of ethics. Big Data and Society, 7(2), https://doi.org/10.1177/2053951720942541.
Rest, J. R., Narvaez, D., Thoma, S. J., & Bebeau, M. J. (1999). DIT2: devising and testing a revised instrument of Moral Judgment. Journal of Educational Psychology, 91(4), 644–659. https://doi.org/10.1037/0022-0663.91.4.644.
Roberts, H., Cowls, J., Morley, J., Taddeo, M., Wang, V., & Floridi, L. (2021). The chinese approach to artificial intelligence: an analysis of policy, ethics, and regulation. AI and Society, 36(1), 59–77. https://doi.org/10.1007/s00146-020-00992-2.
Ryan, M., Antoniou, J., Brooks, L., Jiya, T., Macnish, K., & Stahl, B. (2021). Research and practice of AI Ethics: a Case Study Approach Juxtaposing Academic discourse with Organisational reality. Science and Engineering Ethics, 27(2), https://doi.org/10.1007/s11948-021-00293-x.
Seixas-Nunes, A. (2022). The legality and accountability of Autonomous Weapon Systems: a Humanitarian Law Perspective. New York: Cambridge University Press.
Shweder, R., Much, N. C., Mahapatra, M., & Park, L. (1997). The “big three” of morality (autonomy, community, divinity) and the “big three” explanations of suffering. In A. Brandt, & P. Rozin (Eds.), Morality and health (pp. 119–169). London: Routledge.
Smith, P. B. (2010). On the distinctiveness of chinese psychology; or: are we all Chinese?. In M. H. Bond (Ed.), Oxford Handbook of chinese psychology (pp. 699–710). Oxford: Oxford University Press.
Sripada, C. S., & Stich, S. (2007). A Framework for the psychology of norms. In P. Carruthers, S. Laurence, & S. Stich (Eds.), Innateness and the structure of the mind (II vol., pp. 280–301). New York: Oxford University Press.
Stich, S. (2017). The moral domain. In K. Gray, & J. Graham (Eds.), The Atlas of Moral psychology (pp. 547–555). New York: Guilford Press.
Talhelm, T., Zhang, X., Oishi, S., Shimin, C., Duan, D., Lan, X., & Kitayama, S. (2014). Large-scale Psychological Rice Versus Wheat Agriculture. Science, 344(May), 603–609. https://doi.org/10.1126/science.1246850.
Van de Poel, I. (2016). An ethical Framework for evaluating Experimental Technology. Science and Engineering Ethics, 22(3), 667–686. https://doi.org/10.1007/s11948-015-9724-3.
Van de Poel, I. (2017). Society as a laboratory to experiment with new technologies. In D. M. Bowman, E. Stokes, & A. Rip (Eds.), Embedding New Technologies into Society: A Regulatory, ethical and societal perspective. Pan Stanford Publishing. https://doi.org/10.1201/9781315379593.
Van de Poel, I., & Royakkers, L. (2011). Ethics, Technology, and Engineering: an introduction. Malden: Wiley-Blackwell.
Wiener, A. (2018). Contestation and Constitution of norms in Global International Relations. Cambridge: Cambridge University Press.
Wong, P. H. (2021). Global Engineering Ethics. In D. Michelfelder & N. Doorn (Eds.), Routledge Handbook of Philosophy of Engineering.
Zhu, Q., & Jesiek, B. (2017). Engineering Ethics in Global Context: Four Fundamental Approaches. In ASEE Annual Conference and Exposition. https://doi.org/10.18260/1-2--28252.
Zhu, Q., Martin, M., & Schinzinger, R. (2022). Ethics in Engineering (5th ed.). New York: McGraw-Hill.
Zimmermann, L. (2017). Global norms with a local Face: rule-of-Law Promotion and Norm-Translation. Cambridge: Cambridge University Press.
Funding
This material is based on work supported by the US National Science Foundation (NSF) under Grant No. 2124984. Professor Ingvild Bode’s contribution to this paper was funded by the European Research Council (ERC), under the European Union’s Horizon 2020 research and innovation program (grant agreement No. 852123).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Clancy, R., Bode, I. & Zhu, Q. The need for and nature of a normative, cultural psychology of weaponized AI (artificial intelligence). Ethics Inf Technol 25, 6 (2023). https://doi.org/10.1007/s10676-023-09680-3
Published:
DOI: https://doi.org/10.1007/s10676-023-09680-3