Abstract
Debates about the moral status of social robots (SRs) currently face a second-order, or metatheoretical impasse. On the one hand, moral individualists argue that the moral status of SRs depends on their possession of morally relevant properties. On the other hand, moral relationalists deny that we ought to attribute moral status on the basis of the properties that SRs instantiate, opting instead for other modes of reflection and critique. This paper develops and defends a pragmatic approach which aims to reconcile these two positions. The core of this proposal is that moral individualism and moral relationalism are best understood as distinct deliberative strategies for attributing moral status to SRs, and that both are worth preserving insofar as they answer to different kinds of practical problems that we face as moral agents.
Similar content being viewed by others
Notes
Following Kate Darling, I take a social robot to be “a physically embodied, autonomous agent that communicates and interacts with humans on a social level (Darling 2016, 2). Whether this definition includes chatbots and other large language models is an open question (given that these technologies are technically embodied in hardware). The pragmatic approach to moral status developed in this paper could, in principle, apply to these cases as well. Although I shall focus primarily on SRs given their centrality in recent debates.
For a general discussion of the potential impact of robots within our lives, see Bostrom (2014), Darling (2016; 2021), Nørskov (2016). For a discussion of the economic impact of integrating robots into the workplace, see Ford (2015), Danaher (2017). Some writers have considered features of human–robot relations, especially sexual and romantic relations with robots (Danaher 2019; Frank and Nyholm 2017; Jecker 2021a; McArthur 2017), but also friendship (Jecker 2021b; Marti 2010) and care-giving (Sharkey and Sharkey 2010). Since the European Parliament’s Committee on Legal Affairs issued a 2017 report proposing the creation of the category of “electronic personhood,” there has been considerable discussion of the legal status of social robots. For an overview of this debate, see Parviainen and Coeckelbergh (2021).
In addition to the parent–child relationship, other relationships that have been used to analogize the human–robot relation include the employee-employer relation, or the god-creature relation.
In this paper, I take the term moral status to be synonymous with ‘moral standing’, ‘moral considerability’ and ‘moral patienthood’.
James Rachels describes moral individualism as “a thesis about the justification of judgements concerning how individuals may be treated. The basic idea is that how an individual may be treated is to be determined, not by considering his group memberships, but by considering his own particular characteristics” (Rachels 2005, 173).
Some writers take the issue to be whether social robots will ever possess status-conferring properties, whereas others focus more on the question of whether social robots will likely soon possess those properties.
In admitting that social properties are status conferring, these authors are not thereby committing themselves to a relationalist view. As I shall explain below, MR is not (necessarily) the idea that moral status is conferred by virtue of relationships. Rather it consists in both a negative (anti-individualist) dimension as well as a positive dimension, which suggests that moral status ascriptions should be arrived at through various forms of critical reflection.
For a discussion of the likelihood that these technologies will be available in the (relatively) near future, see Bostrom (2014).
See for example Andreotta (2021), Mosakas (2021), Müller (2021), and Veliz (2021). For skeptical arguments grounded in the adverse social implications of ascribing moral status to SRs see Turkle (2011) and Bryson (2009). Relatedly, others have argued that, in principle, SRs are incapable of being moral agents (Sparrow 2021).
See also Mosakas (2021, 431).
For a discussion, see Chalmers (1995).
Andreotta does not find this line of argument convincing given its reliance on intuitions about cases for which we have no empirical support (Andreotta 2021, 27–8).
The first, “AI Consciousness Test” is meant to serve as a sufficient, but not necessary condition for determining consciousness (Schneider 2019, 50). It attempts to “challenge an AI with a series of increasingly demanding natural language interactions to see how readily it can grasp and use concepts based on the internal experiences we associate with consciousness” (51). For critical discussions of Schneider’s tests, see Andreotta (2021).
Neely argues that it is possible for AI to have interests even if they lack phenomenal consciousness, and that this suffices for their having moral status (Neely 2014).
Another example of the inferential indeterminacy would be the question of whether consciousness is conceptually separable from notions such as intelligence or rationality. Andreotta argues that these notions are independent of one another, such that it is possible to have an intelligent machine that is not phenomenally conscious (Andreotta 2021, 22–23). But one could envision someone who denied this claim of independence.
This idea is, however, not limited to metaethical constructivism, but has been developed in considerable detail within other areas of philosophy—notably feminist philosophy (Lindemann 2019, chapter 4) and pragmatism (Rorty 1989). It finds empirical support from social identity theory and self-categorization theory (Jenkins 2014).
The notion of “taking on face” is one that Gunkel and Coeckelbergh derive from Emmanuel Levinas.
In claiming that relationalists require a theory of error, I do not mean to claim that appealing to status-conferring properties to ground moral status is a universal practice. Indeed, traditional sub-Saharan African and contemporary Japanese societies do not rely on individualist intuitions (Jecker and Nakazawa 2022). My claim is that insofar as relationalists deny the legitimacy of individualist justifications, they owe an explanation not only of why these justifications are mistaken, but of how such a justificatory error became so prevalent—especially within post-enlightenment Western societies.
Constrained relationalism bears important similarities to other positions within the theoretical landscape. John Danaher has recently advanced a view called ethical behaviourism (EB), according to which an entity has moral status if it consistently behaves like other entities to which we ascribe moral status (Danaher 2020). Both EB and constrained relationalism are compatible with a range of views about how moral status attributions are justified (Danaher 2020, 2024). Moreover, both views prioritize normative and epistemological questions about moral status over metaphysical ones (Danaher 2020, 2027). But whereas EB focuses on justifying empirical inquiry into an entity’s behavior as a primary means of determining its moral status, constrained relationalism focuses on a broader range of deliberative strategies, including transcendental critique, sociolinguisitic analysis, phenomenological inquiry, and so on. Constrained relationalism also shares important affinities with pluriversal approaches to ethical thought (Reiter 2018). While a detailed comparison is beyond the scope of this paper, both positions are amenable to multiple legitimate methods and forms of inquiry in ethics. They are also resistant to the assumption that all ethical questions—including those about the limits of moral considerability—admit of a single answer that holds universally.
References
Andreotta, A. J. (2021). The hard problem of AI rights. AI & Society, 36, 19–32.
Bostrom, N. (2014). Superintelligence: Paths, dangers, strategies. Oxford University Press.
Bostrom, N., & Yudkowsky, E. (2014). The ethics of artificial intelligence. In K. Frankish & W. Ramsey (Eds.), The Cambridge handbook of artificial intelligence (pp. 316–334). Cambridge University Press.
Brey, P. (2008). Do we have moral duties towards information objects? Ethics and Information Technology, 10, 109–114.
Bryson, J. J. (2009). Robots should be slaves. In Y. Wilks (Ed.), Close engagements with artificial companions: Key social, psychological, ethical and design issues. John Benjamins Publishing Company.
Cappuccio, M. L., Peeters, A., & McDonald, W. (2019). Sympathy for Dolores: Moral consideration for robots based on virtue and recognition. Philosophy & Technology, 33(1), 9–31.
Chalmers, D. J. (1995). Facing up to the problem of consciousness. Journal of Consciousness Studies, 2(3), 200–219.
Coeckelbergh, M. (2010). Robot rights? Towards a social-relational justification of moral consideration. Ethics and Information Technology, 12, 209–221.
Coeckelbergh, M. (2012). Growing moral relations: Critique of moral status ascription. Palgrave Macmillan.
Coeckelbergh, M. (2014). The moral standing of machines: Towards a relational and non-Cartesian moral hermeneutics. Philosophy & Technology, 27(1), 61–77.
Coeckelbergh, M. (2018). Why care about robots? Empathy, moral standing, and the language of suffering. Kairos. Journal of Philosophy & Science, 20(1), 141–158.
Coeckelbergh, M. (2022a). Robot ethics. MIT Press.
Coeckelbergh, M. (2022b). The Ubuntu robot: Towards a relational conceptual framework for intercultural robotics. Science and Engineering Ethics, 28, 16.
Coeckelbergh, M., & Gunkel, D. J. (2014). Facing animals: A relational, other-oriented approach to moral standing. Journal of Agricultural and Environmental Ethics, 27, 715–733.
Danaher, J. (2017). Should we be thinking about sex robots? In J. Danaher & N. McArthur (Eds.), Robot sex: Social and ethical implications. MIT Press.
Danaher, J. (2019). The rise of the robots and the crisis of moral patiency. AI & Society, 34, 129–136.
Danaher, J. (2020). Welcoming robots into the moral circle: A defence of Ethical behaviourism. Science and Engineering Ethics, 26, 2023–2049.
Darling, K. (2016). Extending legal protection to social robots: The effects of anthropomorphism, empathy, and violent behavior towards robotic objects. In R. Calo, A. Michael Froomkin, & I. Kerr (Eds.), Robot law (pp. 213–231). Edward Elgar.
Darling, K. (2021). The new breed: What our history with animals reveals about our future with robots. Henry Holt & Company.
DeGrazia, D. (2008). Moral status as a matter of degree? The Southern Journal of Philosophy, 46(2), 181–198.
DiSilvestro, R. (2010). Human capacities and moral status. Springer.
Floridi, L. (1999). Information ethics: On the philosophical foundation of computer ethics. Ethics and Information Technology, 1, 33–52.
Floridi, L. (2002). On the intrinsic value of information objects and the infosphere. Ethics and Information Technology, 4, 287–304.
Ford, M. (2015). Rise of the robots: Technology and the threat of a jobless future. Basic Books.
Frank, L., & Nyholm, S. (2017). Robot sex and consent: Is consent to sex between a robot and a human conceivable, possible, and desirable? Artificial Intelligence and Law, 25, 305–323.
Gordon, J.-S. (2021). Artificial moral and legal personhood. AI & Society, 36, 457–471.
Gordon, J.-S. (2022a). Are superintelligent robots entitled to human rights? Ratio, 35, 181–193.
Gordon, J.-S. (2022b). The African relational account of social robots: A step back? Philosophy & Technology, 35, 49.
Gordon, J.-S., & Gunkel, D. J. (2022). Moral status and intelligent robots. The Southern Journal of Philosophy, 60(1), 88–117.
Gunkel, D. J. (2011). The machine question. MIT Press.
Gunkel, D. J. (2014). A vindication of the rights of machines. Philosophy & Technology, 27, 113–132.
Gunkel, D. J. (2018). The other question: Can and should robots have rights? Ethics and Information Technology, 20, 87–99.
Harman, E. (2003). The potentiality problem. Philosophical Studies, 114, 173–198.
Jaworska, A., & Tannenbaum, J. (2013). The grounds of moral status. In E. N. Zalta (Ed.), The Stanford Encyclopedia of Philosophy (Spring 2018 ed.) https://plato.stanford.edu/archives/spr2018/entries/grounds-moral-status/. Accessed 1 June 2020
Jecker, N. S. (2021a). Nothing to be ashamed of: Sex robots for older adults with disabilities. Journal of Medical Ethics, 47, 26–32.
Jecker, N. S. (2021b). You’ve got a friend in me: Sociable robots for older adults in an age of global pandemics. Ethics and Information Technology, 23(Suppl 1), S35–S43.
Jecker, N. S., & Nakazawa, E. (2022). "Bridging east-west differences in ethics guidance for AI and robotics. AI, 3(3), 764–777.
Jecker, N. S., Atiure, C. A., & Ajei, M. O. (2022a). The moral standing of social robots: Untapped insights from Africa. Philosophy & Technology, 35(2), 1–22.
Jecker, N. S., Atiure, C. A., & Ajei, M. O. (2022b). Two steps forward: An African relational account of moral standing. Philosophy & Technology, 35(2), 38.
Jenkins, R. (2014). Social identity (4th ed.). Routledge.
Kamm, F. M. (2007). Intricate ethics: Rights, responsibilities, and permissible harm. Oxford University Press.
Kenny, A. (1973). Wittgenstein. Harvard University Press.
Kittay, E. F. (2005). At the margins of moral personhood. Ethics, 116(1), 100–131.
Korsgaard, C. (1996). The sources of normativity. Cambridge University Press.
Korsgaard, C. (2009). Self-constitution: Agency, identity, and integrity. Oxford University Press.
Lindemann, H. (2019). An invitation to feminist ethics (2nd ed.). Oxford University Press.
Marti, P. (2010). Robot companions. Interaction Studies, 11(2), 220–226.
McArthur, N. (2017). The case for sexbots. In J. Danaher & N. McArthur (Eds.), Robot sex: Social and ethical implications. MIT Press.
McMahan, J. (2005). Our fellow creatures. The Journal of Ethics, 9(3/4), 353–380.
Mosakas, K. (2021). On the moral status of social robots: considering the consciousness criterion. AI & Society, 36, 429–443.
Müller, V. C. (2021). Is it time for robot rights? moral status in artificial entities. Ethics and Information Technology, 23, 579–587.
Neely, E. L. (2014). Machines and the moral community. Philosophy & Technology, 27(1), 97–111.
Nørskov, M. (2016). Social Robots: Boundaries. Routledge.
Parviainen, J., & Coeckelbergh, M. (2021). The political choreography of the Sophia robot: Beyond robot rights and citizenship to political performances for the social robotics market. AI & Society, 36, 715–724.
Rachels, J. (2005). Drawing lines. In C. R. Sunstein & M. C. Nussbaum (Eds.), Animal rights: Current debates and new directions. Oxford University Press.
Reiter, B. (2018). Introduction. In B. Reiter (Ed.), Constructing the pluriverse: The geopolitics of knowledge. Duke University Press.
Rorty, R. (1989). Contingency, irony, and solidarity. Cambridge University Press.
Sætra, H. S. (2021). Challenging the neo-anthropocentric relational approach to robot rights. Frontiers in Robotics and AI, 8, 744426.
Sagoff, M. (1984). Animal liberation and environmental ethics: Bad marriage, quick divorce. Osgoode Hall Law Journal, 22(2), 297–307.
Schneider, S. (2019). Artificial you: AI and the future of your mind. Princeton University Press.
Schwitzgebel, E., & Garza, M. (2015). A defense of the rights of artificial intelligences. Midwest Studies in Philosophy, 39(1), 98–119.
Sharkey, A., & Sharkey, N. (2010). Granny and the robots: Ethical issues in robot care for the elderly. Ethics and Information Technology, 14, 27–40.
Sparrow, R. (2021). Why machines cannot be moral. AI & Society, 36, 685–693.
Street, S. (2012). Coming to terms with contingency: Humean constructivism about practical reason. In J. Lenman & Y. Shemmer (Eds.), Constructivism in Practical Philosophy. Oxford University Press.
Tavani, H. T. (2018). Can social robots qualify for moral consideration? Reframing the question about robot rights. Information, 9(4), 73.
Turkle, S. (2011). Alone together: Why we expect more from technology and less from each other. Basic Books.
Véliz, C. (2021). Moral zombies: Why algorithms are not moral agents. AI & Society, 36, 487–497.
Wareham, C. S. (2021). Artificial intelligence and African conceptions of personhood. Ethics and Information Technology, 23(2), 127–136.
Warren, M. A. (1997). Moral status: Obligations to persons and other living things. Oxford University Press.
Acknowledgements
I am grateful to Colin Koopman for providing feedback on an earlier draft of this paper. I would also like to thank two anonymous reviewers for their exceptionally helpful comments.
Funding
The authors declare that no funds, grants, or other support were received during the preparation of this manuscript.
Author information
Authors and Affiliations
Contributions
N/a.
Corresponding author
Ethics declarations
Ethics Approval
N/A.
Consent to Participate
N/A.
Consent to Publish
N/A.
Competing Interests
The authors have no relevant financial or non-financial interests to disclose.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Showler, P. The Moral Status of Social Robots: A Pragmatic Approach. Philos. Technol. 37, 51 (2024). https://doi.org/10.1007/s13347-024-00737-9
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s13347-024-00737-9