Abstract
This paper discusses means for promoting artificial intelligence (AI) that is designed to be safe and beneficial for society (or simply “beneficial AI”). The promotion of beneficial AI is a social challenge because it seeks to motivate AI developers to choose beneficial AI designs. Currently, the AI field is focused mainly on building AIs that are more capable, with little regard to social impacts. Two types of measures are available for encouraging the AI field to shift more toward building beneficial AI. Extrinsic measures impose constraints or incentives on AI researchers to induce them to pursue beneficial AI even if they do not want to. Intrinsic measures encourage AI researchers to want to pursue beneficial AI. Prior research focuses on extrinsic measures, but intrinsic measures are at least as important. Indeed, intrinsic factors can determine the success of extrinsic measures. Efforts to promote beneficial AI must consider intrinsic factors by studying the social psychology of AI research communities.
Similar content being viewed by others
Notes
On the extrinsic/intrinsic distinction, see e.g., Markowitz and Sharif (2012:246) and references therein.
See Sotala and Yampolskiy (2014, Sect. 3) for a review in the context of strong AI. Russell et al. (2015) also discuss a range of predominantly extrinsic measures. A notable exception to the focus on extrinsic measures is Russell’s emphasis on shifting “how practitioners define what they do” (Bohannon 2015:252).
Incentives can nonetheless provoke significant backlash. For example, in the United States, environmentalists have long pursued incentive-based policies such as taxes on pollution in order to appeal to industry interests that do not want constraints, yet industry has been largely successful at avoiding these incentive-based policies.
Incentives for completed technologies are less relevant for AIs that could be catastrophic because there may be no penalty that could adequately compensate for damages and, in the extreme case, no one alive to process the penalty.
Conversely, when beneficial AI groups are identified, rewards are to be applied, though this is less of a challenge because AI groups are likely to seek rewards, not dodge them.
Social sanctions are an extrinsic measure, specifically an incentive using a social penalty, though they can also cultivate certain social norms.
References
Amodei D, Olah C, Steinhardt J, Christiano P, Schulman J, Mané D (2016) Concrete problems in AI safety. arXiv:1606.06565
Armstrong S, Bostrom N, Shulman C (2016) Racing to the precipice: a model of artificial intelligence development. AI & Soc 31(2):201–206
Baum SD, Goertzel B, Goertzel TG (2011) How long until human-level AI? Results from an expert assessment. Technol Forecast Soc Change 78(1):185–195
Bohannon J (2015) Fears of an AI pioneer. Science 349(6245):252
Borrie J (2014) Humanitarian reframing of nuclear weapons and the logic of a ban. Int Aff 90(3):625–646
Cooter RD (2000) Three effects of social norms on law: expression, deterrence, and internalization. Oregon Law Rev 79:1–22
Deci EL (1971) Effects of externally mediated rewards on intrinsic motivation. J Pers Soc Psychol 18:105–115
Dickerson CA, Thibodeau R, Aronson E, Miller D (1992) Using cognitive dissonance to encourage water conservation. J Appl Soc Psychol 22:841–854
Eden AH, Moor JH, Soraker JH, Steinhart E (2013) Singularity hypotheses: a scientific and philosophical assessment. Springer, Berlin
Griffith E (2016) Who will build the next great car company? Fortune Magazine, 23 June. http://fortune.com/self-driving-cars-silicon-valley-detroit. Accessed 26 Sept 2016
Gurney JK (2013) Sue my car not me: products liability and accidents involving autonomous vehicles. Univ Ill J Law Technol Policy 2013(2):247–277
Joy B (2000) Why the future doesn’t need us. Wired 8(4):238–263
Krantz DH, Peterson N, Arora P, Milch K, Orlove B (2008) Individual values and social goals in environmental decision making. In: Smith JC, Connolly T, Son YJ (eds) Kugler T. Decision modeling and behavior in complex and uncertain environments New York, Springer, pp 165–198
Kunda Z (1990) The case for motivated reasoning. Psychol Bull 108(3):480–498
Lapinski MK, Rimal RN (2005) An explication of social norms. Commun Theory 15(2):127–147
Lewandowsky S, Cook J, Oberauer K, Brophy S, Lloyd EA, Marriott M (2015) Recurrent fury: conspiratorial discourse in the blogosphere triggered by research on the role of conspiracist ideation in climate denial. J Soc Political Psychol 3(1):142–178
Lin P, Abney K, Bekey GA (eds) (2011) Robot ethics: the ethical and social implications of robotics. MIT Press, Cambridge
Markowitz EM, Shariff AF (2012) Climate change and moral judgement. Nat Clim Change 2(4):243–247
McGinnis JO (2010) Accelerating AI. Northwest Univ Law Rev 104:366–381
Moses LB (2007) Recurring dilemmas: the law’s race to keep up with technological change. Univ Ill J Law Technol Policy 2007(2):239–285
Posner EA (2000) Law and social norms: the case of tax compliance. Va Law Rev 86:1781–1819
Posner RA (2004) Catastrophe: risk and response. Oxford University Press, Oxford
Russell S, Dewey D, Tegmark M (2015) Research priorities for robust and beneficial artificial intelligence. AI Mag 36(4):105–114
Schienke EW, Tuana N, Brown DA, Davis KJ, Keller K, Shortle JS, Stickler M, Baum SD (2009) The role of the NSF Broader Impacts Criterion in enhancing research ethics pedagogy. Soc Epistemol 23(3–4):317–336
Schienke EW, Baum SD, Tuana N, Davis KJ, Keller K (2011) Intrinsic ethics regarding integrated assessment models for climate management. Sci Eng Ethics 17(3):503–523
Schultz PW, Nolan JM, Cialdini RB, Goldstein NJ, Griskevicius V (2007) The constructive, destructive, and reconstructive power of social norms. Psychol Sci 18(5):429–434
Scruggs L, Benegal S (2012) Declining public concern about climate change: can we blame the great recession? Glob Environ Change 22(2):505–515
Shome D, Marx S (2009) The psychology of climate change communication: a guide for scientists, journalists, educators, political aides, and the interested public. Columbia University Center for Research on Environmental Decisions, New York
Shulman C (2009) Arms control and intelligence explosions. In: 7th European Conference on Computing and Philosophy (ECAP), Bellaterra, Spain, July 2–4
Sotala K, Yampolskiy RV (2014) Responses to catastrophic AGI risk: a survey. Physica Scripta 90(1):018001. doi:10.1088/0031-8949/90/1/018001
Stone J, Fernandez NC (2008) To practice what we preach: the use of hypocrisy and cognitive dissonance to motivate behavior change. Soc Pers Psychol Compass 2(2):1024–1051
Sunstein CR (1996) On the expressive function of law. Univ Pa Law Rev 144(5):2021–2053
Vohs KD, Mead NL, Goode MR (2006) The psychological consequences of money. Science 314:1154–1156
Wilson G (2013) Minimizing global catastrophic and existential risks from emerging technologies through international law. Va Environ Law J 31:307–364
Yampolskiy R, Fox J (2013) Safety engineering for artificial general intelligence. Topoi 32(2):217–226
Acknowledgments
This paper has benefited from conversations with many people including Dario Amodei, Miles Brundage, Sean Legassick, Richard Mallah, and Jaan Tallinn, and from time spent at the Columbia University Center for Research on Environmental Decisions. Tony Barrett, Steven Umbrello, and Peter Howe provided helpful comments on an earlier draft. Any errors or shortcomings in the paper are the author’s alone. Work on this paper was funded in part by a grant from the Future of Life Institute. The views in this paper are the author’s and are not necessarily the views of the Future of Life Institute or the Global Catastrophic Risk Institute.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Baum, S.D. On the promotion of safe and socially beneficial artificial intelligence. AI & Soc 32, 543–551 (2017). https://doi.org/10.1007/s00146-016-0677-0
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00146-016-0677-0