Zusammenfassung
Social Bots sind automatisiert gesteuerte Profile in sozialen Netzwerken, die menschliches Verhalten simulieren. Es gibt einfache und hochentwickelte Social Bots, sowie inaktive, die sich kaum von Lurkern unterscheiden lassen. Erforscht wurden Social Bots bisher hinblickend auf ihre Bedeutung für demokratische Online-Diskurse und für die COVID-19-Infodemie. Der Beitrag beleuchtet die politischen Konsequenzen und diskutiert ihr Framing, die ethische Perspektive auf Bots und gibt zwei konkrete Handlungsperspektiven.
Notes
- 1.
Einige Beispiele finden sich hier: https://realpython.com/twitter-bot-python-tweepy/, https://digitalinspiration.com/product/twitter-bots, https://devblog.axway.com/apis/api-builder-twitter-bot-example/ (Zugriff am 26.04.2021).
- 2.
SF Quake Bot: https://twitter.com/earthquakesSF (Zugriff 26.04.2021).
Literatur
Allem, J. P., & Ferrara, E. (2018, August 1). Could social bots pose a threat to public health? American Journal of Public Health. American Public Health Association Inc. https://doi.org/10.2105/AJPH.2018.304512
Alves de Lima Salge, C., & Berente, N. (2017). Is that social bot behaving unethically? | September 2017 | Communications of the ACM. Communications of the ACM, 60(9), 29–31. Retrieved from https://cacm.acm.org/magazines/2017/9/220438-is-that-social-bot-behaving-unethically/fulltext?mobile=false
Aro, J. (2016). The cyberspace war: Propaganda and trolling as warfare tools. European View, 15(1), 121–132. https://doi.org/10.1007/s12290-016-0395-5
Bessi, A., & Ferrara, E. (2016). Social bots distort the 2016 U.S. Presidential election online discussion. First Monday, 21(11). https://doi.org/10.5210/fm.v21i11.7090
Broniatowski, D. A., Jamison, A. M., Qi, S. H., AlKulaib, L., Chen, T., Benton, A., et al. (2018). Weaponized health communication: Twitter bots and Russian trolls amplify the vaccine debate. American Journal of Public Health, 108(10), 1378–1384. https://doi.org/10.2105/AJPH.2018.304567
Buchanan, M. (2020, September 1). Managing the infodemic. Nature Physics. Nature Research. https://doi.org/10.1038/s41567-020-01039-5
Chang, H.-C. H., Chen, E., Zhang, M., Muric, G., & Ferrara, E. (2021). Social bots and social media manipulation in 2020: The year in review. Retrieved from http://arxiv.org/abs/2102.08436
Cinelli, M., Cresci, S., Galeazzi, A., Quattrociocchi, W., & Tesconi, M. (2020). The limited reach of fake news on Twitter during 2019 European elections. PLOS One, 15(6), e0234689. https://doi.org/10.1371/journal.pone.0234689
Coleman, M. C. (2018). Bots, social capital, and the need for civility. Journal of Media Ethics: Exploring Questions of Media Morality, 33(3), 120–132. https://doi.org/10.1080/23736992.2018.1476149
Conover, M. D., Ratkiewicz, J., Francisco, M., Gonçalves, B., Flammini, A., & Menczer, F. (2011). Political polarization on Twitter. In Proceedings of the Fifth International AAAI Conference on Weblogs and Social Media (S. 89–96). Retrieved from www.aaai.org
Daub, A. (2020, April 13). The rise of the lurker. The New Republic. Retrieved from https://newrepublic.com/article/157274/rise-lurker-joanne-mcneil-book-review
Davis, C. A., Varol, O., Ferrara, E., Flammini, A., & Menczer, F. (2016). BotOrNot. In Proceedings of the 25th International Conference Companion on World Wide Web – WWW’16 Companion (S. 273–274). Association for Computing Machinery (ACM). https://doi.org/10.1145/2872518.2889302
dpa. (2016, October 21). Bundestagswahlkampf 2017: AfD setzt auf umstrittene Social Bots. Welt. Retrieved from https://www.welt.de/politik/deutschland/article158945108/AfD-schickt-Meinungsroboter-in-den-Wahlkampf.html
Edelmann, N. (2015). What is lurking? A literature review of research on lurking. In G. Riva, B. K. Wiederhold, & P. Cipresso (Hrsg.), The psychology of social networking Vol. 1: Personal experience in online communities (S. 159–174). De Gruyter Open.
Edelmann, N. (2017). Online lurking: Definitions, implications, and effects on E-participation. Talinn University of Technology.
Egli, P., & Rechensteiner, D. (2017). Social Bots und Meinungsbildung in der Demokratie. Aktuelle Juristische Praxis, 2, 249–258.
Elliott, N., Strohmenger, R., Parrish, M., Reitsma, R., Truog, D., Morill, R., et al. (2014). Introducing the new Social Technographics®: The first step toward social success. Retrieved from http://www.forrester.com/rb/Research/introducing_new_social_technographics%26%23174%3B/q/id/56291/t/2
Ferrara, E. (2017). Disinformation and social bot operations in the run up to the 2017 French Presidential Election. First Monday, 22(8). https://doi.org/10.2139/ssrn.2995809
Ferrara, E. (2020). View of What types of COVID-19 conspiracies are populated by Twitter bots? First Monday, 25(6).
Ferrara, E., Varol, O., Davis, C., Menczer, F., & Flammini, A. (2016). The rise of social bots. Communications of the ACM, 59(7), 96–104. https://doi.org/10.1145/2818717
Ferrara, E., Cresci, S., & Luceri, L. (2020). Misinformation, manipulation, and abuse on social media in the era of COVID-19. Journal of Computational Social Science, 3(2), 271–277.
Forelle, M. C., Howard, P. N., Monroy-Hernandez, A., & Savage, S. (2015). Political bots and the manipulation of public opinion in Venezuela. SSRN Electronic Journal. https://doi.org/10.2139/ssrn.2635800
Gehl, R. W., & Bakardjieva, M. (2017). Socialbots and their friends: Digital media and the automation of sociality. Routledge.
Green, M. (2021, April 6). 10 best places to buy Twitter followers (real & active) in 2021. HowSociable. Retrieved from https://howsociable.com/buy-twitter-followers/#faqs
Hegelich, S., & Janetzko, D. (2016). Are social bots on Twitter political actors? Empirical evidence from a Ukrainian social botnet. In Proceedings of the Tenth International AAAI Conference on Web and Social Media. ICWSM 2016. Retrieved from https://www.aaai.org/ocs/index.php/ICWSM/ICWSM16/paper/view/13015
Hepp, A. (2020). Artificial companions, social bots and work bots: Communicative robots as research objects of media and communication studies. Media, Culture & Society, 42(7–8), 1410–1426. https://doi.org/10.1177/0163443720916412
Howard, P. N., & Kollanyi, B. (2017). Bots, #Strongerin, and #Brexit: Computational propaganda during the UK-EU Referendum. SSRN Electronic Journal. https://doi.org/10.2139/ssrn.2798311
Isakovič, J., & Sulčič, A. (2011). Building web communities: An example methodology. In B. K. Daniel (Hrsg.), Handbook of research on methods and techniques for studying virtual communities: Paradigms and phenomena (S. 359–396). IGI Global.
Jensen, K. B. (2015). What’s social about social media? Social Media + Society, 1(1), 205630511557887. https://doi.org/10.1177/2056305115578874
Ji, Y., He, Y., Jiang, X., Cao, J., & Li, Q. (2016). Combating the evasion mechanisms of social bots. Computers and Security, 58, 230–249. https://doi.org/10.1016/j.cose.2016.01.007
Joinson, A. N. (2001). Knowing me, knowing you: Reciprocal self-disclosure in internet-based surveys. CyberPsychology & Behavior, 4(5), 587–591. https://doi.org/10.1089/109493101753235179
Kallmann, E. A., & Grillo, J. P. (1996). Ethical decision making and information technology. McGraw-Hill.
Keller, T. R., & Klinger, U. (2019). Social bots in election campaigns: Theoretical, empirical, and methodological implications. Political Communication, 36(1), 171–189. https://doi.org/10.1080/10584609.2018.1526238
Klinger, U. (2019). Bürger oder Bots? Automatisierte Kommunikation im Bundestagswahlkampf 2017. Digitale Desinformation (Bpb). Retrieved from https://www.bpb.de/gesellschaft/digitales/digitale-desinformation/290557/buerger-oder-bots-automatisierte-kommunikation-im-bundestagswahlkampf-2017
Klinger, U. (2020). Algorithmen, Bots und Trolle: Vom Ende der demokratischen Öffentlichkeit, wie wir sie kennen. In P. Massing, A. Kost, & M. Reiser (Hrsg.), Handbuch Demokratie (S. 271–280). Wochenschau Verlag.
Kollanyi, B. (2016). Where do bots come from? An analysis of bot codes shared on GitHub. International Journal of Communication, 10, 4932–4951.
Kollanyi, B., Howard, P. N., & Woolley, S. C. (2016). Bots and automation over Twitter during the U.S. Election. In COMPROP 2016.4 DATA MEMO. Retrieved from https://demtech.oii.ox.ac.uk/wp-content/uploads/sites/89/2016/11/Data-Memo-US-Election.pdf
Kovic, M., Rauchfleisch, A., Sele, M., & Caspar, C. (2018). Digital astroturfing in politics: Definition, typology, and countermeasures. Studies in Communication Sciences, 18(1), 69–85. https://doi.org/10.24434/j.scoms.2018.01.005
Liu, H., Morstatter, F., Tang, J., & Zafarani, R. (2016). The good, the bad, and the ugly: Uncovering novel research opportunities in social media mining. International Journal of Data Science and Analytics, 1(3–4), 137–143. https://doi.org/10.1007/s41060-016-0023-0
Livingstone, S. (2004). Media literacy and the challenge of new information and communication technologies. The Communication Review, 7(1), 3–14. https://doi.org/10.1080/10714420490280152
Martin, S. P., & Robinson, J. P. (2007). The income digital divide: Trends and predictions for levels of internet use. Social Problems, 54(1), 1–22. https://doi.org/10.1525/sp.2007.54.1.1
Martinez-Juarez, L. A., Sedas, A. C., Orcutt, M., & Bhopal, R. (2020, June 1). Governments and international institutions should urgently attend to the unjust disparities that COVID-19 is exposing and causing. EClinicalMedicine. Lancet Publishing Group. https://doi.org/10.1016/j.eclinm.2020.100376
Marx, J., Brünker, F., Mirbabaie, M., & Hochstrate, E. (2020). Conspiracy machines – The role of social bots during the COVID-19 infodemic. Retrieved from http://arxiv.org/abs/2012.09536
Matikainen, J. (2015). Motivations for content generation in social media. Participations: Journal of Audience & Reception Studies, 12(1), 41–58. Retrieved from http://www.participations.org/Volume12/Issue1/4.pdf
Mayer-Uellner, R. (2002). Das Schweigen der Lurker: politische Partizipation und soziale Kontrolle in Online-Diskussionsforen. Universität München.
McNeil, J. (2020). Lurking: How a person became a user. New York: MCD. Retrieved from https://www.amazon.com/Lurking-How-Person-Became-User/dp/0374194335
Metaxas, P. T., & Mustafaraj, E. (2012, October 26). Social media and the elections. Science. American Association for the Advancement of Science. https://doi.org/10.1126/science.1230456
Miner, A. S., Laranjo, L., & Kocaballi, A. B. (2020). Chatbots in the fight against the COVID-19 pandemic. Npj Digital Medicine, 3(1), 1–4. https://doi.org/10.1038/s41746-020-0280-0
Muller, M. (2012). Lurking as personal trait or situational disposition? Lurking and contributing in enterprise social media. Follow the Crowd. Retrieved from http://crowdresearch.org/blog/?p=2232
Murthy, D., Powell, A. B., Tinati, R., Anstead, N., Carr, L., & Carr Susan Halford Mark Weal, L. J. (2016). Bots and political influence: A sociotechnical investigation of social network capital. International Journal of Communication, 10, 4952–4971.
Nagy, P., & Neff, G. (2015). Imagined affordance: Reconstructing a keyword for communication theory. Social Media + Society, 1(2), 205630511560338. https://doi.org/10.1177/2056305115603385
Noelle-Neumann, E. (1980). Die Schweigespirale. Öffentliche Meinung – unsere soziale Haut. Piper.
Nonnecke, B., & Preece, J. (2003). Silent participants: Getting to know lurkers better. In C. Lueg & D. Fischer (Hrsg.), From Usenet to CoWebs: Interacting with social information spaces (S. 110–132). Springer. https://doi.org/10.1007/978-1-4471-0057-7_6
Plass-Fleßenkämper, B. (2016, December 13). Botswatch: Ehrenamtliche jagen Social Bots auf Twitter. GQ Magazin. Retrieved from https://www.gq-magazin.de/auto-technik/article/botswatch-ehrenamtliche-jagen-social-bots-auf-twitter
Porten-Cheé, P., & Eilders, C. (2015). Spiral of silence online: How online communication affects opinion climate perception and opinion expression regarding the climate change debate. Studies in Communication Sciences, 15(1), 143–150.
Ruane, E., Birhane, A., & Ventresque, A. (2019). Conversational AI: Social and ethical considerations. In AICS 2019: Proceedings of the 2019 International Conference on Artificial Intelligence and Computer Science. Wuhan.
Schuchard, R., Crooks, A., Stefanidis, A., & Croitoru, A. (2019). Bots in nets: Empirical comparative analysis of bot evidence in social networks. In Studies in Computational Intelligence (Bd. 813, S. 424–436). Springer Verlag. https://doi.org/10.1007/978-3-030-05414-4_34
Selwyn, N. (2006). Digital division or digital decision? A study of non-users and low-users of computers. Poetics, 34(4–5), 273–292. https://doi.org/10.1016/j.poetic.2006.05.003
Starbird, K. (2019). Disinformation’s spread: Bots, trolls and all of us. Nature, 571(7766), 449–450.
Stella, M., Ferrara, E., & De Domenico, M. (2018). Bots increase exposure to negative and inflammatory content in online social systems. Proceedings of the National Academy of Sciences of the United States of America, 115(49), 12435–12440. https://doi.org/10.1073/pnas.1803470115
Stieglitz, S., Brachten, F., Berthelé, D., Schlaus, M., Venetopoulou, C., & Veutgen, D. (2017). Do Social Bots (Still) Act different to humans? – Comparing metrics of social bots with those of humans. In G. Meiselwitz (Hrsg.), Social computing and social media (LNCS, Bd. 10282, S. 379–395). Springer International Publishing. https://doi.org/10.1007/978-3-319-58559-8_30
Takacs, R., & McCulloh, I. (2019). Dormant bots in social media: Twitter and the 2018 U.S. Senate Election. In Proceedings of the 2019 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, ASONAM 2019 (S. 796–800). Association for Computing Machinery, Inc. https://doi.org/10.1145/3341161.3343852
Takahashi, M., Fujimoto, M., & Yamasaki, N. (2007). Active lurking: Enhancing the value of in-house online communities through the related practices around the online communities (MIT Sloan School of Management Working Paper 4646-07, CCIWorking, pp. 1–11).
Uyheng, J., & Carley, K. M. (2020). Bots and online hate during the COVID-19 pandemic: Case studies in the United States and the Philippines. Journal of Computational Social Science, 3(2), 445–468. https://doi.org/10.1007/s42001-020-00087-4
Wagner, C., Mitter, S., Körner, C., & Strohmaier, M. (2012). When social bots attack: Modeling susceptibility of users in online social networks. In C. Wagner, S. Mitter, C. Körner, & M. Strohmaier (Hrsg.), 2nd Workshop on Making Sense of Microposts (S. 41–48). Retrieved from http://ceur-ws.org/Vol-838/paper_11.pdf
Wang, J., & Paschalidis, I. C. (2015). Botnet Detection using Social Graph Analysis. In 52nd Annual Allerton Conference on Communication, Control, and Computing, Allerton 2014 (pp. 393–400). Retrieved from http://arxiv.org/abs/1503.02337
Woolley, S. C. (2016). Automating power: Social bot interference in global politics. First Monday, 21(4). https://doi.org/10.5210/fm.v21i4.6161
Woolley, S. C., & Howard, P. N. (2017). Computational propaganda worldwide: Executive summary (Working Paper 2017.11). Retrieved from https://demtech.oii.ox.ac.uk/wp-content/uploads/sites/89/2017/06/Casestudies-ExecutiveSummary.pdf
Worden, K., Bullough, W. A., & Haywood, J. (2003). Smart technologies. World Scientific Publishing. https://doi.org/10.1142/4832
Ziems, C., He, B., Soni, S., & Kumar, S. (2020). Racism is a virus: Anti-Asian hate and counterhate in social media during the COVID-19 crisis. ArXiv. Retrieved from http://arxiv.org/abs/2005.12423
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 Springer Fachmedien Wiesbaden GmbH, ein Teil von Springer Nature
About this entry
Cite this entry
Hamm, A., Schossböck, J., Edelmann, N., Parycek, P. (2023). Social Bots im demokratischen Online-Diskurs: Ein Update zur Infodemie und Handlungsperspektiven. In: Kersting, N., Radtke, J., Baringhorst, S. (eds) Handbuch Digitalisierung und politische Beteiligung. Springer VS, Wiesbaden. https://doi.org/10.1007/978-3-658-31480-4_21-1
Download citation
DOI: https://doi.org/10.1007/978-3-658-31480-4_21-1
Received:
Accepted:
Published:
Publisher Name: Springer VS, Wiesbaden
Print ISBN: 978-3-658-31480-4
Online ISBN: 978-3-658-31480-4
eBook Packages: Springer Referenz Sozialwissenschaften und Recht