Abstract
Insanity is doing the same thing over and over and expecting a different result. “Friendly AI” (FAI) meets these criteria on four separate counts by expecting a good result after: 1) it not only puts all of humanity’s eggs into one basket but relies upon a totally new and untested basket, 2) it allows fear to dictate our lives, 3) it divides the universe into us vs. them, and finally 4) it rejects the value of diversity. In addition, FAI goal initialization relies on being able to correctly calculate a “Coherent Extrapolated Volition of Humanity” (CEV) via some as-yet-undiscovered algorithm. Rational Universal Benevolence (RUB) is based upon established game theory and evolutionary ethics and is simple, safe, stable, self-correcting, and sensitive to current human thinking, intuitions, and feelings. Which strategy would you prefer to rest the fate of humanity upon?
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Yudkowsky, E.: Artificial Intelligence as a Positive and Negative Factor in Global Risk. In: Bostrom, N., Cirkovic, M. (eds.) Global Catastrophic Risks, pp. 308–343. Oxford University Press Inc., New York (2008)
Omohundro, S.: The Basic AI Drives. In: Proceedings of the First Conference on Artificial General Intelligence, pp. 483–492. IOS Press, Amsterdam (2008)
Fox, J., Shulman, C.: Superintelligence Does Not Imply Benevolence. In: Mainzer, K. (ed.) ECAP 2010: VIII European Conference on Computing and Philosophy, pp. 456–462 (2010)
Sotala, K.: From Mostly Harmless to Civilization-Threatening: Pathways to Dangerous Artificial General Intelligences. In: Mainzer, K. (ed.) ECAP 2010: VIII European Conference on Computing and Philosophy, pp. 443–450 (2010)
Yudkowsky, E.: Creating Friendly AI 1.0: The Analysis and Design of Benevolent Goal Architectures, http://singinst.org/CFAI.html
Yudkowsky, E.: Coherent Extrapolated Volition, http://www.singinst.org/upload/CEV.html
Yudkowski, E.: Three Worlds Collide, http://robinhanson.typepad.com/files/three-worlds-collide.pdf
Gauthier, D.: Morals by Agreement. Oxford University Press, Oxford (1986)
Haidt, J., Kesebir, S.: Morality. In: Fiske, S., Gilbert, D., Lindzey, G. (eds.) Handbook of Social Psychology, 5th edn., pp. 797–832. Wiley, Hoboken (2010)
de Waal, F.: Primates and Philosophers: How Morality Evolved. Princeton University Press, Princeton (2006)
Trivers, R.: Deceit and self-deception. In: Robinson, M., Tiger, L. (eds.) Man and Beast Revisited. Smithsonian Press, Washington, DC (1991)
Haidt, J.: The Emotional Dog and Its Rational Tail: A Social Intuitionist Approach to Moral Judgment. Psychological Review 108, 813–814 (2001)
Hauser, M., Cushman, F., Young, L., Kang-Xing, R., Mikhail, J.: A Dissociation Between Moral Judgments and Justifications. Mind & Language 22(1), 1–27 (2007)
McLean, R., Fuentes-Hernandez, A., Greig, D., Hurst, L., Gudelj, I.: A Mixture of “Cheats” and “Co-operators” Can Enable Maximal Group Benefit. PLoS Biology 8(9) (2010)
Wright, R.: Nonzero: The Logic of Human Destiny. Pantheon, New York (2000)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2011 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Waser, M. (2011). Rational Universal Benevolence: Simpler, Safer, and Wiser Than “Friendly AI”. In: Schmidhuber, J., Thórisson, K.R., Looks, M. (eds) Artificial General Intelligence. AGI 2011. Lecture Notes in Computer Science(), vol 6830. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-22887-2_16
Download citation
DOI: https://doi.org/10.1007/978-3-642-22887-2_16
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-22886-5
Online ISBN: 978-3-642-22887-2
eBook Packages: Computer ScienceComputer Science (R0)