Abstract
At the start of 2021, Twitter launched a closed US pilot of Birdwatch, seeking to promote credible information online by giving users the opportunity to add context to misleading tweets. The pilot shows awareness of the importance of context, and the challenges, risks and vulnerabilities the system will face. But the mitigations against these vulnerabilities of Birdwatch can exacerbate wider societal vulnerabilities created by Birdwatch. This article examines how Twitter presents the Birdwatch system, outlines a taxonomy of potential sociotechnical vulnerabilities, and situates these risks within broader social issues. We highlight the importance of watching the watchers, not only in terms of those using and potentially manipulating Birdwatch, but also the way Twitter is developing the system and their wider decision-making processes that impact on public discourse.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
@Birdwatch: Private correspondence (Twitter DM), 29 March 2021 & 15 April 2021
Allen, J., Arechar, A.A., Pennycook G., Rand, D.G.: Scaling up fact-checking using the wisdom of crowds (2020). Preprint. Accessed 22 Apr 2021
Arun, C.: Rebalancing regulation of speech: hyper-local content on global web-based platforms (2018). Medium - Berkman Klein Center. Accessed 23 Apr 2021
Banchik, A.V.: Disappearing acts: content moderation and emergent practices to preserve at-risk human rights-related content. New Media Soc., 1–18 (2020)
Bell, E.J., Owen, T., Brown, P.D., Hauka, C., Rashidian, N.: The platform press: how silicon valley reengineered journalism. Tow Center Digit. J., 1–105 (2017)
Bender, E.M., Gebru, T., McMillan-Major, A., Shmitchell, S.: On the dangers of stochastic parrots: can language models be too big? In: FAccT 2021, pp. 610–623 (2021)
Benjamin, G.: From protecting to performing privacy. J. Sociotechnical Critique 1(1), 1–30 (2020)
Benjamin, R.: Race After Technology: Abolitionist Tools for the New Jim Code. Polity (2019)
Binns, R., Veale, M., Van Kleek, M., Shadbolt, N.: Like trainer, like bot? Inheritance of bias in algorithmic content moderation. In: Ciampaglia, G.L., Mashhadi, A., Yasseri, T. (eds.) SocInfo 2017. LNCS, vol. 10540, pp. 405–415. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-67256-4_32
Browne, S.: Dark Matters: On the Surveillance of Blackness. Duke UP (2015)
Buolamwini, J., Gebru, T.: Gender shades: intersectional accuracy disparities in commercial gender classification. In: FAT* 2018, pp. 77–91 (2018)
Citron, D.K., Solove, D.J.: Privacy harms. GWU Leg. Stud. 11, 1–56 (2021)
Cobbe, J., Singh, J.: Regulating recommending: motivations, considerations, and principles. EJLT 10(3), 1–49 (2019)
Coleman, K.: Introducing birdwatch, a community-based approach to misinformation (2021). Twitter Blog. Accessed 22 Apr 2021
Common, M.F.: Fear the reaper: how content moderation rules are enforced on social media. Int. Rev. Law, Comput. Technol. 34(2), 126–152 (2020)
D’Ignazio, C., Klein, L.: Data Feminism. MIT Press (2020)
Davidson, T., Bhattacharya, D., Weber, I.: Racial bias in hate speech and abusive language detection datasets. In: ACL 2019 ALW3, pp. 1–11 (2019)
Douek, E.: The Rise of Content Cartels. Knight First Amendment Institute at Columbia, pp. 1–51 (2020)
Douek, E.: Governing online speech: from “Posts-As-Trumps” to proportionality and probability. Colum. L. Rev. 121(1), 1–70 (2021)
Fan, J., Zhang A.X.: Digital juries: a civics-oriented approach to platform governance. In: CHI 2020, pp. 1–14 (2020)
Fisher, M.: Inside Facebook’s secret rulebook for global political speech (2018). New York Times. Accessed 23 Apr 2021
Gangadharan, S.: Context, research, refusal: perspectives on abstract problem-solving (2020). Our Data Bodies. Accessed 21 Apr 2021
Gebru, T., et al.: Datasheets for Datasets. In: FAT* 2018, pp. 1–17 (2018)
Geiger, R.S.: Bot-based collective blocklists in Twitter: the counterpublic moderation of harassment in a networked public space. ICS 19(6), 787–803 (2016)
Gerrard, Y.: Beyond the hashtag: circumventing content moderation on social media. New Media Soc. 20(12), 4492–4511 (2018)
Gillespie, T.: Custodians of the Internet: Platforms, Content Moderation, and the Hidden Decisions That Shape Social Media. Yale University Press (2018)
Gorwa, R.: The platform governance triangle: conceptualising the informal regulation of online content. IPR 8(2), 1–22 (2019)
Gorwa, R., Binns, R., Katzenbach, C.: Algorithmic content moderation: technical and political challenges in the automation of platform governance. BD &S 7(1), 1–15 (2020)
Gorwa, R., Guilbeault, D.: Unpacking the social media bot: a typology to guide research and policy. P &I 12(2), 225–248 (2020)
Gray, M., Suri, S.: Ghost Work: How to Stop Silicon Valley from Building a New Global Underclass. Houghton Mifflin Harcourt (2019)
Grimmelmann, J.: The virtues of moderation. YJoLT 17(42), 42–109 (2015)
Hamidi, F., Scheuerman, M.K., Branham, S.M.: Gender recognition or gender reductionism? The social implications of embedded gender recognition systems. In: CHI 2018, pp. 1–13 (2018)
Hanna, A., Park, T.M.: Against scale: provocations and resistances to scale thinking. In: CSCW 2020, pp. 1–4 (2020)
Hasinoff, A.A., Gibson A.D., Salehi, N.: The promise of restorative justice in addressing online harm (2020). Brookings Institute. Accessed 22 Apr 2021
Information Commissioner’s Office: Letter from the Information Commissioner ICO/O/ED/L/RTL/0181 (2020). ICO. Accessed 23 Apr 2021
Jhaver, S., Ghoshal, S., Bruckman, A., Gilbert, E.: Online harassment and content moderation: the case of blocklists. TOCHI 25(2), 1–33 (2018)
Jiang, J.A., Scheuerman, M.K., Fiesler, C., Brubaker, J.R.: Understanding international perceptions of the severity of harmful content online. PLoS One 16(8) (2021). e0256762
Keyes, O.: The misgendering machines: Trans/HCI implications of automatic gender recognition. In: CSCW 2018, pp. 1–22 (2018)
Krasodomski-Jones, A., Judson, E., Smith, J., Miller C., Jones, E.: Warring songs: information operations in the digital age (2019). Demos. Accessed 26 Apr 2021
Lampe C., Resnick, P.: Slash (dot) and burn: distributed moderation in a large online conversation space. In: CHI 2004, pp. 543–550 (2004)
Land, M.: Regulating private harms online: content regulation under human rights law. In: Jorgensen, R.F. (ed.) Human Rights in the Age of Platforms, pp. 285–315. MIT Press (2019)
Maddox, J.: Will the gamification of fact-checking work? Twitter seems to think so (2021). Medium - Start It Up. Accessed 22 Apr 2021
Maddox J., Malson, J.: Guidelines without lines, communities without borders: the marketplace of ideas and digital manifest destiny in social media platform policies. SM+S, 1–10 (2020)
Marwick, A.E., Boyd, D.: Networked privacy: how teenagers negotiate context in social media. NMS 16(7), 1051–1067 (2014)
Massanari, A.: Playful participatory culture: learning from reddit. AoIR (2013)
Matias, J.N.: The civic labor of volunteer moderators online. SM+S 5(2), 1–12 (2019)
Micallef, N., He, B., Kumar, S., Ahamad M., Memon, N.: The role of the crowd in countering misinformation: a case study of COVID-19 infodemic. IEEE BigData 2020, 1–10 (2020)
Mitchell, M., et al.: Model cards for model reporting. In: FAT* 2019, pp. 220–229 (2019)
West, S.M.: Censored, suspended, shadowbanned: user interpretations of content moderation on social media platforms. NMS 20(11), 4366–4383 (2018)
Nissenbaum, H.: A contextual approach to privacy online. Daedalus 140(4), 32–48 (2011)
Noble, S.: Algorithms of Oppression. NYU Press (2018)
Pennycock, G., Rand, D.G.: Fighting misinformation on social media using crowdsourced judgments of news source quality. PNAS 116(7), 2521–2526 (2019)
Reddit: Rainbow Six Siege players who use slurs are now getting instantly banned (2018). Reddit. Accessed 23 Apr 2021
Roberts, S.: Behind the Screen. Yale University Press (2019)
Sap, M., Card, D., Gabriel, S., Choi, Y., Smith, N.A.: The risk of racial bias in hate speech detection. ACL 57, 1668–1678 (2019)
Suzor, N.: Understanding content moderation systems: new methods to understand internet governance at scale, over time, and across platforms. In: Whalen, R. (ed.) In Computational Legal Studies, pp. 166–189. Edward Elgar (2020)
Tabassi, E., Burns, K.J., Hadjimichael, M., Molina-Markham, A.D., Sexton, J.T.: A taxonomy and terminology of adversarial machine learning (draft) (2019). NIST IR 8269. Accessed 21 Apr 2021
Tufekci, Z.: Big questions for social media big data: representativeness, validity and other methodological pitfalls. ICWSM 8(1), 505–514 (2014)
Turkewitz, N.: The week in tweets: the “Scale Doesn’t Scale” Ition feat. Emily Bell (2020). Medium. Accessed 22 Apr 2021
Turner, K., Wood, D., D’Ignazio, C.: The abuse and misogynoir playbook. In: Gupta, A., et al. (ed.) The State of AI Ethics Report January 2021. Montreal AI Ethics Institute, pp. 15–34 (2021)
Twitter (2021). Birdwatch Guide. Accessed 7 Sept 2021
UN Human Rights Council: Report of the independent international fact-finding mission on Myanmar (2018). OHCHR/A/HRC/39/64. Accessed 23 Apr 2021
Williams, J., Chowdhury, R.: Introducing our responsible machine learning initiative (2021). Twitter Blog. Accessed 23 Apr 2021
Wong, J.C.: How Facebook let fake engagement distort global politics: a whistleblower’s account (2021). The Guardian. Accessed 23 Apr 2021
York, J.C.: Syria’s Twitter spambots (2011). The Guardian. Accessed 23 Apr 2021
York, J.C., Zuckerman, E.: Moderating the public sphere. In: Jorgensen, R.J. (ed.) Human Rights in the Age of Platforms, pp. 137–162. MIT Press (2019)
Zannettou, S.: “I Won the Election!”: an empirical analysis of soft moderation interventions on Twitter. In: ICWSM 2021, pp. 1–13 (2021)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
A Description of the Taxonomy
A Description of the Taxonomy
Key
-
Scale [A] Individual, [B] Type of individual, [C] Community, [D] Public discourse, [E] Systemic/principles.
-
Timeframe [1] Immediate, [2] Short, [3] Mid, [4] Long, [5] Persistent.
Targets
-
Pilot: vulnerabilities during the US and subsequent pilots
-
Implementation: the transition to global context;
-
Tweets: user interactions and vectors for attacks;
-
Notes: user interactions and vectors for attacks;
-
Note ranking: user interactions and vectors for attacks;
-
Twitter management: internal decisions create/mitigate vulnerabilities;
-
Regulation: external (legislative) decisions prevent/permit vulnerabilities.
Attacks
-
A 1
Posts: tweets and notes, including many directly abusive tactics;
-
A 2
Followers: tools in extended abuse;
-
A 2
Bots: automating abuse to scale up coordinated attacks;
-
A 1
Text-as-image: an attack on whether the Birdwatch system is triggered;
-
A 2
Iterating tweets: as above;
-
A 1
Varying content or categories: as above;
-
A 1
Abusive or harmful content: as above;
-
C 3
Data: includes data poisoning of ranking system;
-
B 2
Algorithm: gives differential visibility or credibility to tweets/notes;
-
C 3
Third party: external attacks exploit vulnerabilities;
-
E 4
Design: internal flaws create vulnerabilities;
-
C 3
Injection: includes user interactions (likely) and breached security (less so);
-
C 3
Manipulation: includes user interactions and breached security;
-
E 4
Data structures: design flaws enable manipulation of data/the algorithm;
-
D 4
Faux transparency: data availability risks obscuring underlying structures;
-
D 4
External validation: public scrutiny and PR as tool for credibility;
-
D 5
Lobbying: pressure on regulators to prevent external constraints;
-
D 5
Loopholes: flaws in regulation (e.g. loose definitions) enable harmful design;
-
D 5
Self-regulation: Birdwatch is part of continued efforts to avoid regulation.
Harms
-
A 2
Coordinated attacks: combining attacks/accounts increases scale/impact;
-
B 3
Weaponisation: systematic targeting of certain groups/communities;
-
A 1
Abuse: effects (emotional, physical) against specific individuals(/groups);
-
B 3
Verification: shift for Twitter; harms marginalised groups with need for ID;
-
E 4
Policies/enforcement: precedent of unequal application/lack of context;
-
C 5
Access/exclusion: design, policies, implementation; method & type of harm;
-
D 3
Game rankings: vulnerabilities in practice and in credibility;
-
D 4
Ranking system: vulnerabilities to public discourse in Birdwatch design;
-
D 4
Avoiding moderation: not moderation; community not platform;
-
D 5
Avoiding regulation: visible action to placate regulators;
-
C 3
Exploitative labour: reliance on users; lack of protection; uneven burden;
-
E 5
Avoid scrutiny: systemic avoidance or deflection of external audit/criticism;
-
E 5
Systemic/narrative: structural impact on society; influence over debates;
-
E 4
Context shift: marginalisation of geospatial/cultural/etc. communities.
Rights and permissions
Copyright information
© 2022 Springer Nature Switzerland AG
About this paper
Cite this paper
Benjamin, G. (2022). Who Watches the Birdwatchers? Sociotechnical Vulnerabilities in Twitter’s Content Contextualisation. In: Parkin, S., Viganò, L. (eds) Socio-Technical Aspects in Security. STAST 2021. Lecture Notes in Computer Science, vol 13176. Springer, Cham. https://doi.org/10.1007/978-3-031-10183-0_1
Download citation
DOI: https://doi.org/10.1007/978-3-031-10183-0_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-10182-3
Online ISBN: 978-3-031-10183-0
eBook Packages: Computer ScienceComputer Science (R0)