Data Availability
Not applicable.
Notes
Although, according to Nick Bostrom, any superintelligent AI would be highly likely to include the acquisition of ever greater computational resources among its instrumental goals (Bostrom, 2014). If that is correct, then one might expect that a rogue superintelligence would be motivated to interfere with the implementation of the artificial successors envisaged by Lavazza and Vilaça, since they would require computational resources that it could put to its own use instead.
Some might suggest that even perceptual experiences could in principle be understood solely in terms of information processing and associated behavioural dispositions, with no necessary connection to phenomenal consciousness or “qualia”, so that even the philosophical “zombies” imagined by David Chalmers could have such experiences (Chalmers, 1996). Of course, the very possibility of such zombies is itself a contested issue in the philosophy of mind.
References
Bostrom, N. (2014). Superintelligence : Paths, Dangers. Oxford University Press.
Chalmers, D. J. (1996). The Conscious Mind : In Search of a Fundamental Theory. Oxford University Press.
Crisp, R. (2021). Well-Being. In E. N. Zalta (Ed.), The Stanford Encyclopedia of Philosophy (Winter 2021 edition), from https://plato.stanford.edu/entries/well-being/. Accessed 3 Mar 2024.
Lavazza, A., & Vilaça, M. (2024). Human extinction and AI: what we can learn from the ultimate threat. Philosophy and Technology, 37, 16. https://doi.org/10.1007/s13347-024-00706-2
Minsky, M. (1994). Will robots inherit the Earth. Scientific American, 271, 109–113.
Moravec, H. P. (1988). Mind Children : The Future of Robot and Human Intelligence. Harvard University Press.
Parfit, D. (1984). Reasons and Persons. Clarendon Press.
Russell, S. (2019). Human Compatible: Artificial Intelligence and the Problem of Control. Viking.
Schneider, S. (2019). Artificial You : AI and the Future of your Mind. Princeton University Press.
Funding
No funding was received to assist with the preparation of this manuscript.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Ethics Approval and Consent to Participate
Not applicable.
Consent for Publication
Not applicable.
Competing Interests
The author has no competing interests to declare that are relevant to the content of this article.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Erler, A. AI Successors Worth Creating? Commentary on Lavazza & Vilaça. Philos. Technol. 37, 40 (2024). https://doi.org/10.1007/s13347-024-00732-0
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s13347-024-00732-0