Zusammenfassung
Die Debatte über das Für und Wider von Künstlicher Intelligenz (KI) wird auf der einen Seite mit dem Argument der Optimierung menschlichen Handelns und Wirkens geführt. Auf der anderen Seite dienen überzogene Szenarien einer alles einnehmende Technologie als Gegenbeispiel. Dabei mangeln beide Argumentationsstränge häufig einer realistischen Einschätzung, Beobachtung und Analyse der Möglichkeiten und Grenzen von KI, inklusive der damit einhergehenden realen Risiken und Gefahren. Die Erwartungen an das ‚Können der KI‘ sind häufig eher illusorischer Natur und unter- bzw. überschätzen dadurch auch die Risiken. Bei genauerer Analyse wird ersichtlich, dass der Begriff KI in vielen Fällen irreführend ist – weder künstlich, noch intelligent – in welcher die Fehleinschätzung über das Können der KI begründet ist. In diesem Beitrag gehen wir dabei auf diese verzweigte Risiko-Debatte ein, analysieren die Aspekte der Künstlichkeit und Intelligenz der KI, bevor wir auf die unterschiedlichen Stränge der KI-Risiko Debatten eingehen – anhand von vier konkreten Einsatzszenarien.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Eine erwähnenswerte Ausnahme ist hier die in Frankreich gestartete Initiative BLOOM, bei der mehr als 1000 Forscher:innen im Bereich der KI-basierten Sprachprogramme an einem transparenten, nicht anglozentrischen Model arbeiten (Heikkilä 2022).
- 2.
Zur epistemischen und methodischen Problematik solcher als-ob Rekonstruktionen menschlichen Handelns siehe Schlicht (2002).
- 3.
Die soziologische Systemtheorie spricht hier von Komplexitätsreduktion und Heinrich Popitz (1976) vom Ordnungswert der Ordnung.
- 4.
Auch im Bereich des ML kommt das Prinzip der cognitive economy zum Einsatz. Siehe Warnett und McGonigle (2002).
- 5.
Literatur
Agostinho, Daniela, Annie Ring, Kristin Veel, Catherine D’Ignazio, und Nanna Bonde Thylstrup. 2019. Uncertain Archives: Approaching the Unknowns, Errors, and Vulnerabilities of Big Data through Cultural Theories of the Archive. Surveillance & Society 17 (3/4): 422–441.
Alfonseca, Manuel, Manuel Cebrian, Antonio Fernandez Anta, Lorenzo Coviello, Andres Abeliuk, und Iyad Rahwan. 2016. Superintelligence cannot be contained: Lessons from Computability Theory. doi: https://doi.org/10.48550/arXiv.1607.00913.
Anand, Pooja, Yashwant Singh, Arvind Selwal, Pradeep Kumar Singh, Raluca Andreea Felseghi, und Maria Simona Raboaca. 2020. IoVT: Internet of Vulnerable Things? Threat Architecture, Attack Surfaces, and Vulnerabilities in Internet of Things and Its Applications towards Smart Grids. Energies 13 (18): 4813. doi: https://doi.org/10.3390/en13184813.
Andrejevic, Mark. 2020. Shareable and unsharable knowledge. Big Data & Society 7 (1): 205395172093391. doi: https://doi.org/10.1177/2053951720933917.
Ariztía, Tomás. 2018. Consumer databases as practical accomplishments: The making of digital objects in three movements. Journal of Cultural Economy 11 (3). doi: https://doi.org/10.1080/17530350.2018.1435421.
Bauman, Zygmunt. 2015. Liquid Modernity. Cambridge, UK, Malden, MA: Polity Press; Blackwell.
Bauman, Zygmunt, und David Lyon. 2016. Liquid surveillance. A conversation, 2nd edn. Polity conversations series. Cambridge, Malden: Polity.
Beck, Ulrich. 1986. Risikogesellschaft. Auf dem Weg in eine andere Moderne. Berlin: Suhrkamp-Verlag.
Bostrom, Nick. 2005. In defense of posthuman dignity. Bioethics 19 (3): 202–214. doi: https://doi.org/10.1111/j.1467-8519.2005.00437.x.
Bostrom, Nick. 2014. Superintelligence. Paths, dangers, strategies. Oxford, England: Oxford University Press.
Bowker, Geoffrey C., und Susan Leigh Star. 1999. Sorting things out. Classification and its consequences, 1st edn. Inside technology. Cambridge, Mass.: MIT Press.
boyd, danah, und Kate Crawford. 2012. Critical Questions For Big Data: Provocations for a cultural, technological, and scholarly phenomenon. Information, Communication & Society 15 (5): 662–679. doi: https://doi.org/10.1080/1369118X.2012.678878.
Brown, Sheila. 2006. The criminology of hybrids. Theoretical Criminology 10 (2): 223–244. doi: https://doi.org/10.1177/1362480606063140.
Burrell, Jenna. 2016. How the machine ‘thinks’: Understanding opacity in machine learning algorithms. Big Data & Society 3 (1): 205395171562251. doi: https://doi.org/10.1177/2053951715622512.
Coletta, Claudio, und Rob Kitchin. 2017. Algorhythmic governance: Regulating the ‘heartbeat’ of a city using the Internet of Things. Big Data & Society 4 (2): 205395171774241. doi: https://doi.org/10.1177/2053951717742418.
Cowan, Nelson. 2010. The Magical Mystery Four: How is Working Memory Capacity Limited, and Why? Current directions in psychological science 19 (1): 51–57. doi: https://doi.org/10.1177/0963721409359277.
Crawford, Kate. 2021. Atlas of AI. Yale University Press.
David, Laurianne, Amol Thakkar, Rocío Mercado, und Ola Engkvist. 2020. Molecular representations in AI-driven drug discovery: a review und practical guide. Journal of cheminformatics 12 (1): 56. doi: https://doi.org/10.1186/s13321-020-00460-5.
Dreyfus, Hubert L. 1999. What computers still can’t do. A critique of artificial reason, 6th edn. Cambridge, Mass.: MIT Press.
Ducange, Pietro, Riccardo Pecori, und Paolo Mezzina. 2018. A glimpse on big data analytics in the framework of marketing strategies. Soft Computing 22 (1): 325–342. doi: https://doi.org/10.1007/s00500-017-2536-4.
Edwards, Lilian, und Michael Veale. 2017. Slave to the Algorithm?: Why a ‘Right to an Explanation’ Is Probably Not the Remedy You Are Looking For. 16 Duke Law & Technology Review 18. doi: https://doi.org/10.2139/ssrn.2972855.
Gallese, Vittorio, und Alvin Goldman. 1998. Mirror neurons and the simulation theory of mind-reading. Trends in Cognitive Sciences 2 (12): 493–501. doi: https://doi.org/10.1016/S1364-6613(98)01262-5.
Ghodselahi, Ahmad, und Ashkan Amirmadhi. 2011. Application of Artificial Intelligence Techniques for Credit Risk Evaluation. International Journal of Modeling and Optimization: 243–249. doi: https://doi.org/10.7763/IJMO.2011.V1.43.
Gibbs, Samuel. 2015. Apple co-founder Steve Wozniak says humans will be robots' pets. The Guardian, 25 Jun. https://www.theguardian.com/technology/2015/jun/25/apple-co-founder-steve-wozniak-says-humans-will-be-robots-pets. Zugriff 02 August 2022.
Giddens, Anthony. 1990. The consequences of modernity. Stanford, Calif.: Stanford Univ. Press.
Gillespie, Tarleton. 2014. The Relevance of Algorithms. In Media Technologies: Essays on Communication, Materiality, and Society, ed. Tarleton Gillespie, Pablo J. Boczkowski und Kirsten A. Foot, 167–194: MIT Press Scholarship Online.
Gillespie, Tarleton. 2016. Algorithm. In Digital keywords: A vocabulary of information society and culture, hrsg. Benjamin Peters, 18–30. Princeton studies in culture and technology. Princeton [New Jersey], Boston, Massachusetts: Princeton University Press; Credo Reference.
Gillespie, Tarleton. 2018. Custodians of the internet. Platforms, content moderation, and the hidden decisions that shape social media. New Haven, London: Yale University Press.
Goertzel, Ben. 2015. Superintelligence: Fears, Promises and Potentials. Journal of Evolution and Technology 25 (2): 55–87.
Hassani, Bertrand K. 2021. Societal bias reinforcement through machine learning: a credit scoring perspective. AI and Ethics 1 (3): 239–247. doi: https://doi.org/10.1007/s43681-020-00026-z.
Heikkilä, Melissa. 2022. Inside a radical new project to democratize AI. MIT Technology Review, Juli 12. https://www.technologyreview.com/2022/07/12/1055817/inside-a-radical-new-project-to-democratize-ai/, Zugriff 02 August 2022.
Hölscher, Christoph, Simon Büchner, Martin Brösamle, Tobias Meilinger, und Gerhard Strube. 2007. Signs and Maps: Cognitive Economy in the Use of External Aids for Indoor Navigation. Proceedings of the 29th Annual Conference of the Cognitive Science Society (CogSci 2007), 377–382 (2007).
Hutchins, Edwin. 2006. Cognition in the wild, 8th edn. A Bradford book. Cambridge, Mass.: MIT Press.
Jagtiani, Julapa, Lauren Lambie-Hanson, und Timothy Lambie-Hanson. 2021. Fintech Lending and Mortgage Credit Access. The Journal of FinTech 01 (01): 2050004. doi: https://doi.org/10.1142/S2705109920500042.
Jansen, Sue Curry. 2022. What Was Artificial Intelligence?. mediastudies.press.
Kennedy, Helen, Thomas Poell, und Jose van Dijck. 2015. Data and agency. Big Data & Society 2 (2): 205395171562156. doi: https://doi.org/10.1177/2053951715621569.
Kerr, Ian, und Jessica Earle. 2013. Prediction, Preemption, Presumption: How Big Data Threatens Big Picture Privacy. Stanford Law Review 66 (65): 65–72.
Kitchin, Rob. 2016. Thinking critically about and researching algorithms. Information, Communication & Society 20 (1): 14–29. doi: https://doi.org/10.1080/1369118X.2016.1154087.
Kosinski, Michal, David Stillwell, und Thore Graepel. 2013. Private traits and attributes are predictable from digital records of human behavior. Proceedings of the National Academy of Sciences of the United States of America 110 (15): 5802–5805. doi: https://doi.org/10.1073/pnas.1218772110.
Kreissl, Reinhard. 2014. Assessing security technology's impact: Old tools for new problems. Science und Engineering Ethics 20 (3): 659–673. doi: https://doi.org/10.1007/s11948-014-9529-9.
Langenbucher, Katja. 2020. Responsible A.I.-based Credit Scoring – A Legal Framework. European Business Law Review 31 (Issue 4): 527–572. doi: https://doi.org/10.54648/eulr2020022.
Lazer, David, Ryan Kennedy, Gary King, und Alessandro Vespignani. 2014. The Parable of Google Flu: Traps in Big Data Analysis. Science 343: 1203–1205.
Lyon, David. 2007. Surveillance, Security and Social Sorting: Emerging Research Priorities. International Criminal Justice Review 17 (3): 161–170. doi: https://doi.org/10.1177/1057567707306643.
Lyon, David. 2017. Surveillance Culture: Engagement, Exposure, and Ethics in Digital Modernity. International Journal of Communication 11 (0): 19.
Mackenzie, Adrian. 2012. More Parts Than Elements: How Databases Multiply. Environment and Planning D: Society and Space 30 (2): 335–350. doi: https://doi.org/10.1068/d6710.
Marcus, Gary. 2020. The Next Decade in AI: Four Steps Towards Robust Artificial Intelligence. https://doi.org/10.48550/arXiv.2002.06177
Marcus, Gary, und Ernest Davis. 2019. Rebooting AI: building artificial intelligence we can trust. Pantheon.
Matz, Sandra C., und Oded Netzer. 2017. Using Big Data as a window into consumers’ psychology. Current Opinion in Behavioral Sciences 18: 7–12. doi: https://doi.org/10.1016/j.cobeha.2017.05.009.
Mazzei, Matthew J., und David Noble. 2017. Big data dreams: A framework for corporate strategy. Business Horizons 60 (3): 405–414. doi: https://doi.org/10.1016/j.bushor.2017.01.010.
Mead, George Herbert. 2015. Mind, self, and society. The definitive edition. Chicago, London: The University of Chicago Press.
Miki, Takahiro, Joonho Lee, Jemin Hwangbo, Lorenz Wellhausen, Vladlen Koltun, und Marco Hutter. 2022. Learning robust perceptive locomotion for quadrupedal robots in the wild. Science robotics 7 (62): eabk2822. doi: https://doi.org/10.1126/scirobotics.abk2822.
Moorstedt, Michael. 2022. Sind Maschinen die besseren Menschen? Süddeutsche Zeitung, Juli 25. https://www.sueddeutsche.de/kultur/democratic-ai-gerechtigkeit-kuenstliche-intelligenz-netzkolumne-1.5626822. Zugriff 02 August 2022.
Musk, Elon, und Neuralink. 2019. An Integrated Brain-Machine Interface Platform With Thousands of Channels. Journal of medical Internet research 21 (10): e16194. doi: https://doi.org/10.2196/16194.
National Security Commission on AI. 2021. Final Report. https://www.nscai.gov/wp-content/uploads/2021/03/Full-Report-Digital-1.pdf, Zugriff 02 August 2022.
Neuralink, 2022. https://neuralink.com/. Zugriff 02 August 2022.
O’Connell, Mark. 2017. The Techno-Libertarians Praying for Dystopia. Intelligencer, April 30. https://nymag.com/intelligencer/2017/04/the-techno-libertarians-praying-for-dystopia.html. Zugriff 02 August 2022.
O’Neil, Cathy. 2016. Weapons of Math Destruction. How Big Data Increases Inequality and Threatens Democracy, 1st edn. New York: Crown.
Pasquale, Frank. 2015. The black box society. The secret algorithms that control money and information. Cambridge: Harvard University Press.
Perrow, Charles. 1999. Normal accidents: Living with high-risk technologies. Princeton University Press.
Philippon, Thomas. (2019). On Fintech and Financial Inclusion. NBER Working Paper No. 26330. Cambridge, MA. https://doi.org/10.3386/w26330
Pisarchik, Alexander N., Vladimir A. Maksimenko, und Alexander E. Hramov. 2019. From Novel Technology to Novel Applications: Comment on „An Integrated Brain-Machine Interface Platform With Thousands of Channels“ by Elon Musk and Neuralink. Journal of medical Internet research 21 (10): e16356. doi: https://doi.org/10.2196/16356.
Popitz, Heinrich. 1976. Prozesse der Machtbildung, 3rd edn. Recht und Staat in Geschichte und Gegenwart, 362/363. Tübingen: Mohr.
Reis, Joao, Paula Espirito Santo, und Nuno Melao. 2019. Impacts of Artificial Intelligence on Public Administration: A Systematic Literature Review. In 2019 14th Iberian Conference on Information Systems and Technologies (CISTI), 1–7. 2019 14th Iberian Conference on Information Systems and Technologies (CISTI), Coimbra, Portugal. 19/06/2019 – 22/06/2019: IEEE. doi: https://doi.org/10.23919/CISTI.2019.8760893.
Rumelhart, David E. 1975. Notes on a Schema for Stories. In Representation and Understanding, 211–236: Elsevier.
Sanders, Carrie, und Camie Condon. 2017. Crime analysis and cognitive effects: The practice of policing through flows of data. Global Crime 18 (3): 237–255. doi: https://doi.org/10.1080/17440572.2017.1323637.
Sanders, Carrie B., und James Sheptycki. 2017. Policing, crime and ‘big data’; towards a critique of the moral economy of stochastic governance. Crime, Law and Social Change 68 (1-2): 1–15. doi: https://doi.org/10.1007/s10611-016-9678-7.
Schlicht, Ekkehart (2002). Der Homo oeconomicus unter experimentellem Beschuß. Munich Discussion Paper. https://doi.org/10.5282/ubm/epub.70965
Schneider, Tanja, und Steve Woolgar. 2012. Technologies of ironic revelation: Enacting consumers in neuromarkets. Consumption Markets & Culture 15 (2): 169–189. doi: https://doi.org/10.1080/10253866.2012.654959.
Schütz, Alfred. 1972. Gesammelte Aufsätze. Dordrecht: Springer Netherlands.
Service, Robert. 2020. ‘The game has changed.’ AI triumphs at solving protein structures. Science. doi: https://doi.org/10.1126/science.abf9367. Zugriff 02 August 2022.
Smith, Craig. 2021. A.I. Here, There, Everywhere. The New York Times, Februar 23. https://www.nytimes.com/2021/02/23/technology/ai-innovation-privacy-seniors-education.html. Zugriff 21 August 2023.
Symons, John, und Ramón Alvarado. 2016. Can we trust Big Data?: Applying philosophy of science to software. Big Data & Society 3 (2): 1–17. doi: https://doi.org/10.1177/2053951716664747.
Todd, Peter M. und Gerd Gigerenzer. (2000). Précis of Simple heuristics that make us smart. The Behavioral and brain sciences, 23(5), 727–41; discussion 742–80. https://doi.org/10.1017/s0140525x00003447
Tomasello, Michael. 2010. Origins of human communication, 1st edn. A Bradford book. Cambridge, Mass., London: MIT Press.
van Dijck, Jose. 2014. Datafication, dataism and dataveillance: Big Data between scientific paradigm and ideology. Surveillance & Society 12 (2): 197–208.
Wachter, Sandra, Brent Mittelstadt, und Chris Russell. 2017. Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR. SSRN Electronic Journal.
Warnett, Lawrence, und Brendan McGonigle. 2002. Unsupervised navigation using an economy principle. https://core.ac.uk/download/pdf/85284.pdf.
Wang, Christina J. (2018). Technology, the Nature of Information, and FinTech Marketplace Lending. Current Policy Perspectives 18–3. Federal Reserve Bank of Boston. https://www.bostonfed.org/publications/current-policy-perspectives/2018/technology-nature-of-information-fintech-marketplace-lending.aspx. Zugriff 21 August 2023
Weidinger, Laura, Jonathan Uesato, Maribeth Rauh, Conor Griffin, Po-Sen Huang, John Mellor, Amelia Glaese, Myra Cheng, Borja Balle, Atoosa Kasirzadeh, Courtney Biles, Sasha Brown, Zac Kenton, Will Hawkins, Tom Stepleton, Abeba Birhane, Lisa Anne Hendricks, Laura Rimell, William Isaac, Julia Haas, Sean Legassick, Geoffrey Irving, und Iason Gabriel. 2022. Taxonomy of Risks posed by Language Models. In 2022 ACM Conference on Fairness, Accountability, and Transparency, 214–229. FAccT ‘22: 2022 ACM Conference on Fairness, Accountability, and Transparency, Seoul Republic of Korea. 21 06 2022 24 06 2022. New York, NY, USA: ACM. doi: https://doi.org/10.1145/3531146.3533088.
Wellner, Galit, und Tiran Rothman. 2020. Feminist AI: Can We Expect Our AI Systems to Become Feminist? Philosophy & Technology 33 (2): 191–205. doi: https://doi.org/10.1007/s13347-019-00352-z.
Weizenbaum, Joseph. 1966. ELIZA—a computer program for the study of natural language communication between man and machine. Communications of the ACM, 9 (1): 36-45.
Zeba, Gordana, Marina Dabić, Mirjana Čičak, Tugrul Daim, und Haydar Yalcin. 2021. Technology mining: Artificial intelligence in manufacturing. Technological Forecasting and Social Change 171: 120971. doi: https://doi.org/10.1016/j.techfore.2021.120971.
Zuboff, Shoshana. 2020. The age of surveillance capitalism. The fight for a human future at the new frontier of power. New York, NY: PublicAffairs.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 Der/die Autor(en), exklusiv lizenziert an Springer Fachmedien Wiesbaden GmbH, ein Teil von Springer Nature
About this chapter
Cite this chapter
Kreissl, R., von Laufenberg, R. (2024). Risiken und Gefahren der ‚Künstlichen‘ ‚Intelligenz‘. In: Heinlein, M., Huchler, N. (eds) Künstliche Intelligenz, Mensch und Gesellschaft. Springer VS, Wiesbaden. https://doi.org/10.1007/978-3-658-43521-9_10
Download citation
DOI: https://doi.org/10.1007/978-3-658-43521-9_10
Published:
Publisher Name: Springer VS, Wiesbaden
Print ISBN: 978-3-658-43520-2
Online ISBN: 978-3-658-43521-9
eBook Packages: Social Science and Law (German Language)