Notes
Moral particularism criticizes rule-based ethical theory, but its criticism is not that rules are not used in moral reasoning, but that one-size-fits all rules are practically bad. But rule-based or logic-based ethical theory never submits such a claim. See Smith and Dubbink (2011).
Gesammelte Schriften, ed. Akademie der Wissenschaften, Berlin: Reimer, later DeGruyter, 1910ff.) 5:161.33–6; translation by Paul Guyer (1992).
References
Arkoudas, K., Bringsjord, S., & Bello, P. (2005). Toward ethical robots via mechanized deontic logic. AAAI fall symposium on machine ethics (pp. 17–23). Menlo Park, CA: The AAAI Press.
Barocas, S., & Selbst, A. D. (2016). Big data’s disparate impact. California Law Review, 104(3), 671.
Bazerman, M. H. (2011). Bounded ethicality in negotiations. Negotiation and Conflict Management Research, 4(1), 8–11.
Bazerman, M. H., & Tenbrunsel, A. E. (2011). Blind spots: Why we fail to do what’s right and what to do about it. Princeton, NJ: Princeton University Press.
Bazerman, M. H., & Gino, F. (2012). Behavioral ethics: Toward a deeper understanding of moral judgment and dishonesty. Annual Review of Law and Social Science, 8, 85–104.
Bazerman, M. H., & Sezer, O. (2016). Bounded awareness: Implications for ethical decision making. Organizational Behavior and Human Decision Processes, 136, 95–105.
Bringsjord, S., & Taylor, J. (2012). The divine-command approach to robot ethics. In P. Lin, K. Abney, & G. A. Bekey (Eds.), Robot ethics: the ethical and social implications of robotics. MIT Press.
Bringsjord, S., Arkoudas, K., & Bello, P. (2006). Toward a general logicist methodology for engineering ethically correct robots. IEEE Intelligent Systems, 21(4), 38–44.
Chauvet, J.-M. (2018). The 30-year cycle in the AI debate. arXiv:1810.04053.
Chen, V., & Hooker, J. N. (2020). A just approach balancing Rawlsian leximax fairness and utilitarianism. In Proceedings of the AAAI/ACM conference on AI, ethics, and society, pp. 221–227.
Clark, A., & Chalmers, D. (1998). The extended mind. Analysis, 58(1), 7–19.
Crawford, K. (2013). The hidden biases in big data, Harvard Business Review, 1(4). (online version).
Daniels, N. (1980). Some methods of ethics and linguistics. Philosophical Studies, 37, 22–23.
Dastin, J. (2018). Amazon scraps secret AI recruiting tool that showed bias against women. London: Reuters.
Fodor, J. A., & Pylyshyn, Z. W. (1988). Connectionism and cognitive architecture: A critical analysis. Cognition, 28(1–2), 3–71.
Geffner, H. (2018). Model-free, model-based, and general intelligence. arXiv:1806.02308.
Govindarajulu, N. S., & Bringsjord, S. (2017). On automating the doctrine of double effect. arXiv:1703.08922.
Gunning, D. (2019). Explainable artificial intelligence (XAI). Arlington: Defense Advanced Research Projects Agency (DARPA).
Guyer, P. (1992). Introduction: The starry heavens and the moral law. In P. Guyer (Ed.), The Cambridge companion to Kant. Cambridge: Cambridge University Press.
Hansson, J. H. (2019). I applied for an apple card: What they offered was a sexist insult, Fast Company. https://www.fastcompany.com/90429224/i-applied-for-an-apple-card-what-theyoffered-was-a-sexist-insult
Hooker, J. N., & Kim, T. W. (2018). Toward non-intuition-based machine and artificial intelligence ethics: A deontological approach based on modal logic. In Proceedings of the 2018 AAAI/ACM conference on AI, ethics, and society, pp. 130–136.
Hooker, J., & Kim, T. W. (2019). Truly autonomous machines are ethical. AI Magazine, 40(4), 66–73.
Hooker, J. N., & Williams, H. P. (2012). Combining equity and utilitarianism in a mathematical programming model. Management Science, 58(9), 1682–1693.
Kahneman, D. (2011). Thinking, fast and slow. New York: Macmillan.
Johnson, C., & Kuipers, B. (2018). Socially-aware navigation using topological maps and social norm learning. In Proceedings of the 2018 AAAI/ACM conference on AI, ethics, and society, pp. 151–157.
Kim, T. W., & Mejia, S. (2019). From artificial intelligence to artificial wisdom: What Socrates teaches us. Computer, 52(10), 70–74.
Kim, T. W., Donaldson, T., & Hooker, J. (2019). Grounding value alignment with ethical principles. arXiv:1907.05447.
Lipton, Z. C. (2016). The mythos of model interpretability. In Proceedings of the 2016 ICML workshop on human interpretability of machine learning 96–100.
Lu, J. Lee, D., Kim, T. W., & Danks, D. (2020). Good explanation for algorithmic transparency. In AIES ’20: Proceedings of the AAAI/ACM conference on AI, ethics, and society.
McEwan, I. (2019). Machines like me. London: Jonathan Cape.
Michaelson, C. (2005). Dealing with swindlers and devils: Literature and business ethics. Journal of Business Ethics, 58(4), 359–373.
Michaelson, C. (2015). How reading novels can help management scholars cultivate ambiculturalism. Academy of Management Review, 40(1), 147–149.
Montavon, G., Samek, W., & Müller, K.-R. (2018). Methods for Interpreting and Understanding Deep Neural Networks. Digital Signal Processing, 73, 1–15.
Newell, A. (1981). The knowledge level: Presidential address. AI magazine, 2(2), 1.
Rawls, J. (1971). A theory of justice. Cambridge, MA: Harvard University Press.
Rossi, F., & Loreggia, A. (2019). Preferences and ethical priorities: Thinking fast and slow in AI. In Proceedings of the 18th international conference on autonomous agents and multiagent systems, 3–4.
Smith, J., & Dubbink, W. (2011). Understanding the role of moral principles in business ethics: A Kantian perspective. Business Ethics Quarterly, 21(2), 205–231.
Sezer, O., Gino, F., & Bazerman, M. H. (2015). Ethical blind spots: Explaining unintentional unethical behavior. Current Opinion in Psychology, 6, 77–81.
Smolensky, P. (1988). On the proper treatment of connectionism. Behavioral and Brain Sciences, 11(1), 1–23.
Turing, A. M. (1950). Computer machinery and intelligence. Mind, 59(236), 433–460.
Wallach, W., & Vallor, S. (2020). Moral machines. Ethics of artificial intelligence (p. 386). Oxford: Oxford University Press.
Wiggers, K. (2019). Amazon and national science foundation earmark $10 million for AI fairness research. San Francisco: VentureBeat.
Zhang, T., Fletcher, P. O., Gino, F., & Bazerman, M. H. (2015). Reducing bounded ethicality: How to help individuals notice and avoid unethical behavior. Organizational Dynamics, 44(4), 310–317.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The author above declares that he has no conflict of interest to report.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Kim, T.W. Flawed Like Us and the Starry Moral Law: Review of Machines Like Me by Ian McEwan. J Bus Ethics 170, 875–879 (2021). https://doi.org/10.1007/s10551-021-04758-8
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10551-021-04758-8