Abstract
We present what we call the Interpretation Problem, whereby any rule in symbolic form is open to infinite interpretation in ways that we might disapprove of and argue that any attempt to build morality into machines is subject to it. We show how the Interpretation Problem in Artificial Intelligence is an illustration of Wittgenstein’s general claim that no rule can contain the criteria for its own application, and that the risks created by this problem escalates in proportion to the degree to which a machine is causally connected to the world, in what we call the Law of Interpretative Exposure. Using games as an illustration, we attempt to define the structure of normative spaces and argue that any rule-following within a normative space is guided by values that are external to that space and which cannot themselves be represented as rules. In light of this, we categorise the types of mistakes an artificial moral agent could make into Mistakes of Intention and Instrumental Mistakes, and we propose ways of building morality into machines by getting them to interpret the rules we give in accordance with these external values, through explicit moral reasoning, the “Show, not Tell” paradigm, the adjustment of causal power and structure of the agent, and relational values, with the ultimate aim that the machine develop a virtuous character and that the impact of the Interpretation Problem is minimised.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Anderson, M., Anderson, S.L. (eds.) Machine ethics. CUP (2011)
Arendt, H.: The Human Condition. University of Chicago Press (1998)
Aristotle, Thompson J. A. K. (trans) Nicomachean Ethics. Penguin, London (1988)
Badea, C.: Have a break from making decisions, have a MARS: the multi-valued action reasoning system. In: arXiv:2109.03283 [cs] (2022)
Badea, C., Gilpin, L.H.: Establishing meta-decision-making for AI: an ontology of relevance, representation and reasoning. In: AAAI 2021 Fall Symposium FSS-21. arXiv:submit/4523302 [cs] (2021)
BBC: Artificial intelligence: Google’s AlphaGo beats Go master Lee Se-dol. BBC News Online. http://www.bbc.co.uk/news/technology-35785875 (2016)
Bolton, W., Badea, C., Georgiou, P., Holmes, A., Rawson, T.: Developing moral AI to support antimicrobial decision making. Nat. Mach. Intell (2022). https://doi.org/10.1038/s42256-022-00558-5
Bostrom, N.: Superintelligence: Oaths, Dangers, Strategies. OUP, Oxford (2014)
Havens, J.: The ethics of AI: how to stop your robot cooking your cat. The Guardian (2015). Retrieved 14 Dec 2017
Hindocha, S., Badea, C.: Moral exemplars for the virtuous machine: the clinician’s role in ethical artificial intelligence for healthcare. AI and Ethics 2, 167–175 (2021)
Jackson, P.: Introduction to Expert Systems, 3rd edn., p. 3. Addison-Wesley Longman Publishing Co., Inc, Boston, MA, USA (1998)
Krakovna, V., et al.: Specification gaming. DeepMind Blog (2020)
Le Roux, M.: Rise of the Machines: Keep an eye on AI, experts warn. Phys.org. https://phys.org/news/2016-03-machines-eye-ai-experts.html (2016)
Plato (trans: Lee, D) The Republic, Penguin (1987)
Post, B., Badea, C., Faisal, A., Brett, S.J.: Breaking bad news in the era of artificial intelligence and algorithmic medicine. AI Ethics (2022). https://doi.org/10.1007/s43681-022-00230-z
Russell, S.J., Norvig, P.: Artificial Intelligence: A Modern Approach, 2nd edn, p. 27, 32–58, pp. 968–972. Prentice Hall (2003). ISBN 0-13-790395-2
Ryle, G.: The Concept of Mind, pp. 74–80. Penguin, London (2000)
Searle: J. Speech Acts. Cambridge, CUP. (ch2) (1969)
Silver, D., et al.: Mastering the game of go with deep neural networks and tree research. Nature 529, 484–489 (2016)
Soares, N.: The Value Learning Problem. In: Ethics in Artificial Intelligence Workshop at IJCAI-16 (2016). Accessed 19 Sep 2022
Solon, O.: Oh the humanity! Poker computer […]. The Guardian. Accessed (2017)
Taylor, J., Yudkowsky, E., LaVictoire, P., Critch, A.: Alignment for advanced machine learning systems, p. 5. Machine Intelligence Research Institute (2016)
Wittgenstein, L.: Philosophical Investigations, 4th edn. Wiley-Blackwell (2009)
Wallach, W., Asaro, P., (eds.) Machine Ethics and Robot Ethics. Taylor&Francis (2016)
Wallach, W., Allen, C.: Moral Machines. OUP (2008)
Yudkowsky, E.: Artificial Intelligence as a positive and negative factor in global risk. In: Yudkowsky, E. (ed.) Global Catastrophic Risks. Oxford University Press (2008). https://doi.org/10.1093/oso/9780198570509.003.0021
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Badea, C., Artus, G. (2022). Morality, Machines, and the Interpretation Problem: A Value-based, Wittgensteinian Approach to Building Moral Agents. In: Bramer, M., Stahl, F. (eds) Artificial Intelligence XXXIX. SGAI-AI 2022. Lecture Notes in Computer Science(), vol 13652. Springer, Cham. https://doi.org/10.1007/978-3-031-21441-7_9
Download citation
DOI: https://doi.org/10.1007/978-3-031-21441-7_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-21440-0
Online ISBN: 978-3-031-21441-7
eBook Packages: Computer ScienceComputer Science (R0)