Skip to main content
Log in

When something goes wrong: Who is responsible for errors in ML decision-making?

  • Open Forum
  • Published:
AI & SOCIETY Aims and scope Submit manuscript

Abstract

Because of its practical advantages, machine learning (ML) is increasingly used for decision-making in numerous sectors. This paper demonstrates that the integral characteristics of ML, such as semi-autonomy, complexity, and non-deterministic modeling have important ethical implications. In particular, these characteristics lead to a lack of insight and lack of comprehensibility, and ultimately to the loss of human control over decision-making. Errors, which are bound to occur in any decision-making process, may lead to great harm and human rights violations. It is important to have a principled way of assigning responsibility for such errors. The integral characteristics of ML, however, pose serious difficulties in defining responsibility and regulating ML decision-making. First, we elaborate on these characteristics and their epistemic and ethical implications. We then analyze possible general strategies for resolving the assignment of moral responsibility and show that, due to the specific way in which ML functions, each potential solution is problematic, whether we assign responsibility to humans, machines, or using hybrid models. Then, we shift focus on an alternative approach that bypasses moral responsibility and attempts to define legal liability independently through solutions such as informed consent and the no-fault compensation system. Both of these solutions prove unsatisfactory because they leave too much room for potential abuses of ML decision-making. We conclude that both ethical and legal solutions are fraught with serious difficulties. These difficulties prompt us to re-weigh the costs and benefits of using ML for high-stake decisions.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

Data availability

Non applicable.

Notes

  1. We take artificial neural networks (ANNs) as the paradigmatic type of ML model. Some aspects of our analysis may, however, be relevant for other types of models that give rise to similar issues.

  2. There are different degrees of human involvement in supervised and unsupervised learning, and this implies different degrees of control over the learning process. Supervised learning is characterized by the use of labeled datasets, which requires human intervention to label the data appropriately. In this way humans ‘supervise’ machines to learn how to correctly classify data. In contrast, in unsupervised learning the machine discovers the underlying structures of unlabelled datasets on its own. Admittedly, even unsupervised modeling needs human intervention with regard to validating output variables to be able to learn from data. However, there is still a significant degree of (semi-)autonomy present in ML that is relevant for the epistemic consequences we discuss in 2.2.

  3. For example, ML might uncover a correlation between the number of physicians a patient visits, a patient’s access to transportation, and a patient's disease outcomes (Russ 2021; see also Wang et al. 2022).

  4. The degree of insight into ML models is of course not the same for an engineer who develops these models and for a person who is a complete layman. The engineer, unlike the layman, knows the general principles of functioning of the ML model and in that sense it can be said that for people developing the models they are ‘gray boxes’. However, when we talk about the blackboxness of models, we are referring to the lack of epistemic insight into the aspects of the working of the ML model which applies to experts as well, not just to laymen. Certain aspects of ML models’ functioning are not accessible to any human, and this is what we refer to by ‘blackboxness’ or ‘opacity’.

  5. There have, of course, been many attempts to make the information involved in ML decision-making explainable to humans, by constructing other models trained to produce explanations—the Explainable AI (xAI) Project. We discuss the xAI project and its limitations in Sect. 2.3.

  6. For more details on the differences between unpredictability and other epistemic obstacles such as unexplainability and incomprehensibility, see Yampolskiy (2020).

  7. For example, the legislative request most commonly cited in the xAI literature is the European Union’s General Data Protection Regulation (GDPR), which requests that the subjects of automated decision-making are provided with “meaningful information about the logic involved” in reaching the decision (GDPR, Article 12(2)(f)). The GDPR also states the right of the subjects to “obtain an explanation of the decision reached after such assessment and to challenge the decision” (GDPR, Recital 71). These are the only two mentions of anything related to explanations in this regulation, and these two formulations state two essentially different explanatory requests, one concerning the overall mechanism of the ML model, and the other concerning the path to the model’s reaching a particular decision.

  8. For example, explanatory tools commonly consist in surrogate models that solely attempt to capture the input–output trends of the opaque model they are intended to explain, but they employ entirely different features and are thereby not faithful to the original model’s computations (Rudin 2019).

  9. This understanding of moral agency is what in Moor’s terminology characterizes a ‘full’ moral agent—the only kind of moral agent that we can consider morally responsible. For the complete taxonomy of moral agency, which has become canonical in the literature on this topic, see Moor (2006). We discuss the prerequisites of moral agency in the context of ML in Sect. 4.1.

  10. Certain types of moral transgressions, such as lying in everyday life, are not legally regulated, nor are they expected to be. Giving a false promise to a friend is morally reprehensible, but we will not necessarily end up in court because of it. Of course, some cases of lying such as defamation and false testimony in court are legally regulated, but lying in ordinary daily life usually does not fall into these categories. There is also the possibility that some moral offenses are not legally regulated yet because they have only recently emerged, such as those made possible by the development of technology, but are expected to be regulated in the future.

  11. It may be objected that the internal processes of human decision-making are also inaccessible, and perhaps even more complex than ML decision-making. We cannot look into the heads of others (the so-called ‘problem of other minds’), so we turn to various social procedures developed for inferring the internal states of other humans (see Matthias 2004). These procedures do not make other minds completely transparent, but might provide some kind of insight about the thought processes, intentions and beliefs of others. Similarly, numerous xAI methods are being developed in attempts to gain insight into ML decision-making processes. So why would this make a problem for ML, but not human decision-making? The key difference is that in the case of human decision-making, the locus of responsibility is clear in most cases. In paradigmatic cases, the person who has made a particular decision is the one who is held responsible for it. In the context of ML, however, there are a number of obstacles that make it highly difficult to find the locus of responsibility for the consequences of the decisions. We discuss these obstacles in detail in Sect. 3.2.

  12. This ambiguity of assigning responsibility would not, of course, apply to cases of intentional biasing of data nor to cases of negligent or reckless use, if, for example, a company did not check the ML system for bias, or continued to use it even after bias is discovered.

  13. Having human experts keep track of the correctness of the ML decision-making may seem as a solution to problems of control and responsibility. However, as Matthias points out, “[w]ere it possible to supply every machine with a controlling human expert, nobody would need the machine in the first place” (2004, p. 177). Besides, it does not seem sensible to employ slower or less reliable systems such as humans to keep track of much more efficient and reliable systems and inspect the correctness of the processes. It would defeat the purpose of using ML decision-making in the first place.

  14. Predictability of errors is relevant if we adopt the risk/negligence model of responsibility (Fischer and Ravizza 1998; McKenna 2008; Lunney and Oliphant 2013). For a comparative analysis of different models of responsibility, see Yeung (2019).

  15. There is another direction taken in the literature that focuses on building a moral code into the machines. It is considered that this would prevent unethical machine decisions, as well as diminish harm and human rights violations (Anderson and Anderson 2007; Wallach and Allen 2009). However, this project faces several significant challenges. First, it needs to decide on a particular ethical theory: deontological ethics, virtue ethics, utilitarianism, or some other. Second, the chosen theory must be implementable in the machines, in the sense that it must be translatable into a language that allows computation, and it is still unclear whether this is a feasible task. Finally, even if building a moral code into the machines becomes possible, we still need to decide how to deal with errors if they occur. There is no reason to believe that the ethical machines would be completely infallible. It seems that we would still need a principled way of assigning responsibility and dealing with potential errors. It remains for future research to show how successful the project of building moral machines will be in meeting its challenges. Importantly, the topic of this paper is how to assign responsibility for the decisions of ML systems that are currently in use and that do not have any built-in moral code. The discussion might become different if machines with a built-in moral code are used in the future, depending on how exactly they would function.

  16. We will not enter into the controversy over which of the presented directions is the most adequate from the point of view of moral theory in general. We will only briefly present each of the possible directions and analyze the difficulties they face.

References

Download references

Acknowledgements

We would like to thank Nenad Filipović for the engaged discussion and helpful comments on the early versions of this paper.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Sanja Srećković.

Ethics declarations

Conflict of interest

On behalf of all authors, the corresponding author states that there is no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Berber, A., Srećković, S. When something goes wrong: Who is responsible for errors in ML decision-making?. AI & Soc (2023). https://doi.org/10.1007/s00146-023-01640-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s00146-023-01640-1

Keywords

Navigation