Abstract
The use of weapons without humans-in-the-loop in modern warfare has been a contentious issue for several decades, from land mines to more advanced systems like loitering munitions. With the emergence of artificial intelligence (AI), particularly machine learning (ML) technologies, the ethical difficulties in this complex field have increased. The challenges related to the adherence to International Humanitarian Law (IHL), or human dignity are compounded by ethical concerns related to AI, such as transparency, explainability, human agency, and autonomy.
In this chapter, we aim to provide a comprehensive overview of the main issues and current positions in the field of autonomous weapons and technological warfare. We will begin by clarifying the concept of autonomy in warfare, an area that still needs attention, as evidenced by the latest discussions within the United Nations (UN) Convention on Certain Conventional Weapons (CCW). We will also introduce the current legal basis in this field and the problems in its practical application and offer sound philosophical grounds to better understand this highly complex and multifaceted field.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
- 2.
See e.g., the 11 guiding principles adopted by the 2019 Meeting of the High Contracting Parties to the CCW, which summarize these meetings and provide some insight about these discussions (https://www.un.org/disarmament/the-convention-on-certain-conventional-weapons/background-on-laws-in-the-ccw/)
- 3.
As Paul Scharre has put it: “I am continually struck by how much the Terminator films influence debate on autonomous weapons. In nine out of ten serious conversations on autonomous weapons I have had, whether in the bowels of the Pentagon or the halls of the United Nations, someone invariably mentions the Terminator.” (Scharre 2018)
- 4.
The Guardian: “‘Kamikaze’ drones hit Kyiv despite Putin’s claim of no further strikes” https://www.theguardian.com/world/2022/oct/17/kyiv-hit-by-a-series-of-explosions-from-drone-attack
- 5.
- 6.
New Scientist: “US Air Force is giving military drones the ability to recognize faces”
- 7.
A similar concept (and equally ambiguous) has been introduced in the DoD Directive 3000.09 Autonomy in Weapon Systems: “appropriate level of human judgment” (DoD 2023).
- 8.
For dissenting, yet controversial, point of view based on the deterrence effect, see (Scharre 2018, p. 315): “Deploying an untested and unverified autonomous weapon would be even more of a deterrent, since one could convincingly say that its behavior was truly unpredictable.”
- 9.
Typical examples often cited to underpin unpredictability claims of AI systems are Microsoft Tay deployed in 2016 that soon started publishing offensive and racist comments, or Google image recognition software that classified in 2015 black people as “gorillas”.
- 10.
See e.g., Ronald Arkin’s position: “Arkin says the only ban he ‘could possibly support’ would be one limited to the very specific capability of ‘target generalization through machine learning.’ He would not want to see autonomous weapons that could learn on their own in the field in an unsupervised manner and generalize to new types of targets” (Scharre 2018).
- 11.
For the same reasons adduced herein Taddeo and Blanchard (2022) contend that the moral gambit would not be applicable to Lethal Autonomous Weapon Systems (LAWS).
- 12.
For a dissenting view see (Scharre 2018, p. 295): Considering the beyond-IHL principle of human dignity “would put militaries in the backwards position of accepting more battlefield harm and more deaths for an abstract concept.”
References
Arkin, R. 2009. Governing Lethal Behavior in Autonomous Robots. Taylor & Francis.
Arkin, Ronald Craig, Patrick Ulam, and Alan R. Wagner. 2012. Moral decision making in autonomous systems: Enforcement, moral emotions, dignity, trust, and deception. Proceedings of the IEEE 100 (3): 571–589. https://doi.org/10.1109/JPROC.2011.2173265.
Asaro, Peter. 2012. On banning autonomous weapon systems: Human rights, automation, and the dehumanization of lethal decision-making. International Review of the Red Cross 94 (886): 687–709. https://doi.org/10.1017/S1816383112000768.
Birnbacher, Dieter. 2016. “Are autonomous weapons systems a threat to human dignity?” In Autonomous weapons systems, edited by Nehal Bhuta, Susanne Beck, Robin Geiβ, Hin-Yan Liu, and Claus Kreβ, 105–121. Cambridge University Press https://doi.org/10.1017/CBO9781316597873.005.
Boshuijzen-van Burken, Christine. 2023. Value sensitive design for autonomous weapon systems – A primer. Ethics and Information Technology 25 (1): 11. https://doi.org/10.1007/s10676-023-09687-w.
Boulanin, Vincent, Neil Davison, Netta Goussac, and Moa Peldán Carlsson. 2020. Limits on autonomy in weapon systems: Identifying practical elements of human control. Stockholm International Peace Research Institute and the International Committee of the Red Cross.
Boutin, Bérénice. 2023. State responsibility in relation to military applications of artificial intelligence. Leiden Journal of International Law 36 (1): 133–150. https://doi.org/10.1017/S0922156522000607.
Cantrell, Hunter. 2022. Autonomous weapon systems and the claim-rights of innocents on the Battlefield. AI and Ethics 2 (4): 645–653. https://doi.org/10.1007/s43681-021-00119-3.
Chauhan, Krishna Deo Singh. 2022. From ‘what’ and ‘why’ to ‘how’: An imperative driven approach to mechanics of AI regulation. Global Jurist, December. https://doi.org/10.1515/gj-2022-0053.
Coeckelbergh, M. 2020. Introduction to Philosophy of Technology. Oxford University Press.
Crootof, Rebecca. 2016. War torts: Accountability for autonomous weapons. University of Pennsylvania Law Review 164 (6): 1347–1402.
de Fine Licht, Karl, and Jenny de Fine Licht. 2020. Artificial intelligence, transparency, and public decision-making. AI & SOCIETY 35 (4): 917–926. https://doi.org/10.1007/s00146-020-00960-w.
DoD. 2011. Unmanned systems integrated roadmap FY2011–2036.
———. 2017. Unmanned systems integrated roadmap FY2017–2042.
———. 2023. DoD DIRECTIVE 3000.09 Autonomy in Weapon Systems.
Ekelhof, Merel. 2019. Moving beyond semantics on autonomous weapons: Meaningful human control in operation. Global Policy 10 (3): 343–348. https://doi.org/10.1111/1758-5899.12665.
Gómez de Ágreda, Ángel. 2020. Ethics of autonomous weapons systems and its applicability to any AI systems. Telecommunications Policy 44 (6): 101953. https://doi.org/10.1016/j.telpol.2020.101953.
Horowitz, Michael C. 2016a. The ethics & morality of robotic warfare: Assessing the debate over autonomous weapons. Daedalus 145 (4): 25–36. https://doi.org/10.1162/DAED_a_00409.
———. 2016b. Why words matter: The real world consequences of defining autonomous weapons systems. Temple International & Comparative Law Journal. https://sites.temple.edu/ticlj/files/2017/02/30.1.Horowitz-TICLJ.pdf.
Horowitz, Michael C., and Paul Scharre. 2015. Meaningful human control in weapon systems: A Primer. 2015. https://www.cnas.org/publications/reports/meaningful-human-control-in-weapon-systems-a-primer.
HRW. 2018. Heed the call: A moral and legal imperative to ban killer robots. Human Rights Watch.
ICRC. 2016. Autonomous weapon systems implications of increasing autonomy. International Committee of the Red Cross, no. March.
———. n.d. Article 51. Protection of the civilian population. https://ihl-databases.icrc.org/en/ihl-treaties/api-1977/article-51.
Kahn, Lauren A. 2023. Risky incrementalism. Defense Al in the United States. Hamburg.
Kayser, Daan. 2023. Why a treaty on autonomous weapons is necessary and feasible. Ethics and Information Technology 25 (2): 25. https://doi.org/10.1007/s10676-023-09685-y.
Keulartz, J., Schermer, M., Korthals, M., and Swierstra, T. (2004). Ethics in Technological Culture: A Programmatic Proposal for a Pragmatist Approach. Science, Technology, & Human Values, 29 (1), 3–29. https://doi.org/10.1177/0162243903259188
Lin, Patrick. 2015. The right to life and the martens clause. Pravo – Teorija i Praksa. https://cyberlaw.stanford.edu/sites/default/files/publication/files/ccw_testimony.pdf.
Lin, Patrick, George Bekey, and Keith Abney. 2008. Autonomous military robotics: Risk, ethics, and design. Fort Belvoir: Defense Technical Information Center. https://doi.org/10.21236/ADA534697.
Lin, Patrick, Keith Abney, and George A. Bekey. 2014. Robot ethics: The ethical and social implications of robotics. London: MIT Press. https://doi.org/10.1007/s10677-015-9638-9.
May, Larry. 2015. Contingent pacifism. In Contingent pacifism, ed. Larry May. Cambridge University Press. https://doi.org/10.1017/CBO9781316344422.
McFarland, Tim. 2020. Autonomous weapon systems and the law of armed conflict: Compatibility with international humanitarian law. Cambridge: Cambridge University Press. https://doi.org/10.1017/9781108584654.
Montréal Institute for Learning Algorithms. 2018. Montréal declaration for a responsible development of artificial intelligence, 1–21.
Morgan, Forrest, Benjamin Boudreaux, Andrew Lohn, Mark Ashby, Christian Curriden, Kelly Klima, and Derek Grossman. 2020. Military applications of artificial intelligence: Ethical concerns in an uncertain world. RAND Corporation. https://doi.org/10.7249/RR3139-1.
Peterson, Martin. 2019. The value alignment problem: A geometric approach. Ethics and Information Technology 21 (1): 19–28. https://doi.org/10.1007/s10676-018-9486-0.
Pražák, Jakub. 2021. Dual-use conundrum: Towards the weaponization of outer space? Acta Astronautica 187: 397–405. https://doi.org/10.1016/j.actaastro.2020.12.051.
Roff, Heather M. 2014. The strategic robot problem: Lethal autonomous weapons in war. Journal of Military Ethics 13 (3): 211–227. https://doi.org/10.1080/15027570.2014.975010.
Roff, Heather M., and Richard Moyes. 2016. Meaningful human control, artificial intelligence and autonomous weapons. Article36.Org. Geneva.
Russell, Stuart. 2016. Should we fear supersmart robots? Scientific American 2, no. June.
Santoni de Sio, Filippo, and Jeroen van den Hoven. 2018. Meaningful human control over autonomous systems: A philosophical account. Frontiers in Robotics and AI 5. https://www.frontiersin.org/articles/10.3389/frobt.2018.00015.
Sari, Onur, and Sener Celik. 2021. Legal evaluation of the attacks caused by artificial intelligence-based lethal weapon systems within the context of rome statute. Computer Law & Security Review 42: 105564. https://doi.org/10.1016/j.clsr.2021.105564.
Scharre, Paul. 2018. Army of none: Autonomous weapons and the future of war. 1st ed. New York: W.W. Norton & Company.
Sehrawat, Vivek. 2017. Autonomous weapon system: Law of Armed Conflict (LOAC) and other legal challenges. Computer Law & Security Review 33 (1): 38–56. https://doi.org/10.1016/j.clsr.2016.11.001.
Taddeo, Mariarosaria, and Alexander Blanchard. 2021. A comparative analysis of the definitions of autonomous weapons, SSRN Scholarly Paper ID 3941214. Rochester: Social Science Research Network. https://doi.org/10.2139/ssrn.3941214.
———. 2022. Accepting moral responsibility for the actions of autonomous weapons systems—A moral gambit. Philosophy & Technology 35 (3): 78. https://doi.org/10.1007/s13347-022-00571-x.
Wood, Nathan Gabriel. 2023. Autonomous weapon systems and responsibility gaps: A taxonomy. Ethics and Information Technology 25 (1): 16. https://doi.org/10.1007/s10676-023-09690-1.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this chapter
Cite this chapter
del Valle, J.I., Moreno, M. (2023). Ethics of Autonomous Weapon Systems. In: Lara, F., Deckers, J. (eds) Ethics of Artificial Intelligence. The International Library of Ethics, Law and Technology, vol 41. Springer, Cham. https://doi.org/10.1007/978-3-031-48135-2_9
Download citation
DOI: https://doi.org/10.1007/978-3-031-48135-2_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-48134-5
Online ISBN: 978-3-031-48135-2
eBook Packages: Religion and PhilosophyPhilosophy and Religion (R0)