Skip to main content

Who Watches the Birdwatchers? Sociotechnical Vulnerabilities in Twitter’s Content Contextualisation

  • Conference paper
  • First Online:
Socio-Technical Aspects in Security (STAST 2021)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 13176))

Included in the following conference series:

Abstract

At the start of 2021, Twitter launched a closed US pilot of Birdwatch, seeking to promote credible information online by giving users the opportunity to add context to misleading tweets. The pilot shows awareness of the importance of context, and the challenges, risks and vulnerabilities the system will face. But the mitigations against these vulnerabilities of Birdwatch can exacerbate wider societal vulnerabilities created by Birdwatch. This article examines how Twitter presents the Birdwatch system, outlines a taxonomy of potential sociotechnical vulnerabilities, and situates these risks within broader social issues. We highlight the importance of watching the watchers, not only in terms of those using and potentially manipulating Birdwatch, but also the way Twitter is developing the system and their wider decision-making processes that impact on public discourse.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 59.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 79.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. @Birdwatch: Private correspondence (Twitter DM), 29 March 2021 & 15 April 2021

    Google Scholar 

  2. Allen, J., Arechar, A.A., Pennycook G., Rand, D.G.: Scaling up fact-checking using the wisdom of crowds (2020). Preprint. Accessed 22 Apr 2021

  3. Arun, C.: Rebalancing regulation of speech: hyper-local content on global web-based platforms (2018). Medium - Berkman Klein Center. Accessed 23 Apr 2021

  4. Banchik, A.V.: Disappearing acts: content moderation and emergent practices to preserve at-risk human rights-related content. New Media Soc., 1–18 (2020)

    Google Scholar 

  5. Bell, E.J., Owen, T., Brown, P.D., Hauka, C., Rashidian, N.: The platform press: how silicon valley reengineered journalism. Tow Center Digit. J., 1–105 (2017)

    Google Scholar 

  6. Bender, E.M., Gebru, T., McMillan-Major, A., Shmitchell, S.: On the dangers of stochastic parrots: can language models be too big? In: FAccT 2021, pp. 610–623 (2021)

    Google Scholar 

  7. Benjamin, G.: From protecting to performing privacy. J. Sociotechnical Critique 1(1), 1–30 (2020)

    Google Scholar 

  8. Benjamin, R.: Race After Technology: Abolitionist Tools for the New Jim Code. Polity (2019)

    Google Scholar 

  9. Binns, R., Veale, M., Van Kleek, M., Shadbolt, N.: Like trainer, like bot? Inheritance of bias in algorithmic content moderation. In: Ciampaglia, G.L., Mashhadi, A., Yasseri, T. (eds.) SocInfo 2017. LNCS, vol. 10540, pp. 405–415. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-67256-4_32

    Chapter  Google Scholar 

  10. Browne, S.: Dark Matters: On the Surveillance of Blackness. Duke UP (2015)

    Google Scholar 

  11. Buolamwini, J., Gebru, T.: Gender shades: intersectional accuracy disparities in commercial gender classification. In: FAT* 2018, pp. 77–91 (2018)

    Google Scholar 

  12. Citron, D.K., Solove, D.J.: Privacy harms. GWU Leg. Stud. 11, 1–56 (2021)

    Google Scholar 

  13. Cobbe, J., Singh, J.: Regulating recommending: motivations, considerations, and principles. EJLT 10(3), 1–49 (2019)

    Google Scholar 

  14. Coleman, K.: Introducing birdwatch, a community-based approach to misinformation (2021). Twitter Blog. Accessed 22 Apr 2021

  15. Common, M.F.: Fear the reaper: how content moderation rules are enforced on social media. Int. Rev. Law, Comput. Technol. 34(2), 126–152 (2020)

    Google Scholar 

  16. D’Ignazio, C., Klein, L.: Data Feminism. MIT Press (2020)

    Google Scholar 

  17. Davidson, T., Bhattacharya, D., Weber, I.: Racial bias in hate speech and abusive language detection datasets. In: ACL 2019 ALW3, pp. 1–11 (2019)

    Google Scholar 

  18. Douek, E.: The Rise of Content Cartels. Knight First Amendment Institute at Columbia, pp. 1–51 (2020)

    Google Scholar 

  19. Douek, E.: Governing online speech: from “Posts-As-Trumps” to proportionality and probability. Colum. L. Rev. 121(1), 1–70 (2021)

    Google Scholar 

  20. Fan, J., Zhang A.X.: Digital juries: a civics-oriented approach to platform governance. In: CHI 2020, pp. 1–14 (2020)

    Google Scholar 

  21. Fisher, M.: Inside Facebook’s secret rulebook for global political speech (2018). New York Times. Accessed 23 Apr 2021

  22. Gangadharan, S.: Context, research, refusal: perspectives on abstract problem-solving (2020). Our Data Bodies. Accessed 21 Apr 2021

  23. Gebru, T., et al.: Datasheets for Datasets. In: FAT* 2018, pp. 1–17 (2018)

    Google Scholar 

  24. Geiger, R.S.: Bot-based collective blocklists in Twitter: the counterpublic moderation of harassment in a networked public space. ICS 19(6), 787–803 (2016)

    Google Scholar 

  25. Gerrard, Y.: Beyond the hashtag: circumventing content moderation on social media. New Media Soc. 20(12), 4492–4511 (2018)

    Article  Google Scholar 

  26. Gillespie, T.: Custodians of the Internet: Platforms, Content Moderation, and the Hidden Decisions That Shape Social Media. Yale University Press (2018)

    Google Scholar 

  27. Gorwa, R.: The platform governance triangle: conceptualising the informal regulation of online content. IPR 8(2), 1–22 (2019)

    Article  Google Scholar 

  28. Gorwa, R., Binns, R., Katzenbach, C.: Algorithmic content moderation: technical and political challenges in the automation of platform governance. BD &S 7(1), 1–15 (2020)

    Google Scholar 

  29. Gorwa, R., Guilbeault, D.: Unpacking the social media bot: a typology to guide research and policy. P &I 12(2), 225–248 (2020)

    Google Scholar 

  30. Gray, M., Suri, S.: Ghost Work: How to Stop Silicon Valley from Building a New Global Underclass. Houghton Mifflin Harcourt (2019)

    Google Scholar 

  31. Grimmelmann, J.: The virtues of moderation. YJoLT 17(42), 42–109 (2015)

    Google Scholar 

  32. Hamidi, F., Scheuerman, M.K., Branham, S.M.: Gender recognition or gender reductionism? The social implications of embedded gender recognition systems. In: CHI 2018, pp. 1–13 (2018)

    Google Scholar 

  33. Hanna, A., Park, T.M.: Against scale: provocations and resistances to scale thinking. In: CSCW 2020, pp. 1–4 (2020)

    Google Scholar 

  34. Hasinoff, A.A., Gibson A.D., Salehi, N.: The promise of restorative justice in addressing online harm (2020). Brookings Institute. Accessed 22 Apr 2021

  35. Information Commissioner’s Office: Letter from the Information Commissioner ICO/O/ED/L/RTL/0181 (2020). ICO. Accessed 23 Apr 2021

  36. Jhaver, S., Ghoshal, S., Bruckman, A., Gilbert, E.: Online harassment and content moderation: the case of blocklists. TOCHI 25(2), 1–33 (2018)

    Article  Google Scholar 

  37. Jiang, J.A., Scheuerman, M.K., Fiesler, C., Brubaker, J.R.: Understanding international perceptions of the severity of harmful content online. PLoS One 16(8) (2021). e0256762

  38. Keyes, O.: The misgendering machines: Trans/HCI implications of automatic gender recognition. In: CSCW 2018, pp. 1–22 (2018)

    Google Scholar 

  39. Krasodomski-Jones, A., Judson, E., Smith, J., Miller C., Jones, E.: Warring songs: information operations in the digital age (2019). Demos. Accessed 26 Apr 2021

  40. Lampe C., Resnick, P.: Slash (dot) and burn: distributed moderation in a large online conversation space. In: CHI 2004, pp. 543–550 (2004)

    Google Scholar 

  41. Land, M.: Regulating private harms online: content regulation under human rights law. In: Jorgensen, R.F. (ed.) Human Rights in the Age of Platforms, pp. 285–315. MIT Press (2019)

    Google Scholar 

  42. Maddox, J.: Will the gamification of fact-checking work? Twitter seems to think so (2021). Medium - Start It Up. Accessed 22 Apr 2021

  43. Maddox J., Malson, J.: Guidelines without lines, communities without borders: the marketplace of ideas and digital manifest destiny in social media platform policies. SM+S, 1–10 (2020)

    Google Scholar 

  44. Marwick, A.E., Boyd, D.: Networked privacy: how teenagers negotiate context in social media. NMS 16(7), 1051–1067 (2014)

    Google Scholar 

  45. Massanari, A.: Playful participatory culture: learning from reddit. AoIR (2013)

    Google Scholar 

  46. Matias, J.N.: The civic labor of volunteer moderators online. SM+S 5(2), 1–12 (2019)

    Google Scholar 

  47. Micallef, N., He, B., Kumar, S., Ahamad M., Memon, N.: The role of the crowd in countering misinformation: a case study of COVID-19 infodemic. IEEE BigData 2020, 1–10 (2020)

    Google Scholar 

  48. Mitchell, M., et al.: Model cards for model reporting. In: FAT* 2019, pp. 220–229 (2019)

    Google Scholar 

  49. West, S.M.: Censored, suspended, shadowbanned: user interpretations of content moderation on social media platforms. NMS 20(11), 4366–4383 (2018)

    Google Scholar 

  50. Nissenbaum, H.: A contextual approach to privacy online. Daedalus 140(4), 32–48 (2011)

    Article  Google Scholar 

  51. Noble, S.: Algorithms of Oppression. NYU Press (2018)

    Google Scholar 

  52. Pennycock, G., Rand, D.G.: Fighting misinformation on social media using crowdsourced judgments of news source quality. PNAS 116(7), 2521–2526 (2019)

    Article  Google Scholar 

  53. Reddit: Rainbow Six Siege players who use slurs are now getting instantly banned (2018). Reddit. Accessed 23 Apr 2021

  54. Roberts, S.: Behind the Screen. Yale University Press (2019)

    Google Scholar 

  55. Sap, M., Card, D., Gabriel, S., Choi, Y., Smith, N.A.: The risk of racial bias in hate speech detection. ACL 57, 1668–1678 (2019)

    Google Scholar 

  56. Suzor, N.: Understanding content moderation systems: new methods to understand internet governance at scale, over time, and across platforms. In: Whalen, R. (ed.) In Computational Legal Studies, pp. 166–189. Edward Elgar (2020)

    Google Scholar 

  57. Tabassi, E., Burns, K.J., Hadjimichael, M., Molina-Markham, A.D., Sexton, J.T.: A taxonomy and terminology of adversarial machine learning (draft) (2019). NIST IR 8269. Accessed 21 Apr 2021

  58. Tufekci, Z.: Big questions for social media big data: representativeness, validity and other methodological pitfalls. ICWSM 8(1), 505–514 (2014)

    Google Scholar 

  59. Turkewitz, N.: The week in tweets: the “Scale Doesn’t Scale” Ition feat. Emily Bell (2020). Medium. Accessed 22 Apr 2021

  60. Turner, K., Wood, D., D’Ignazio, C.: The abuse and misogynoir playbook. In: Gupta, A., et al. (ed.) The State of AI Ethics Report January 2021. Montreal AI Ethics Institute, pp. 15–34 (2021)

    Google Scholar 

  61. Twitter (2021). Birdwatch Guide. Accessed 7 Sept 2021

  62. UN Human Rights Council: Report of the independent international fact-finding mission on Myanmar (2018). OHCHR/A/HRC/39/64. Accessed 23 Apr 2021

  63. Williams, J., Chowdhury, R.: Introducing our responsible machine learning initiative (2021). Twitter Blog. Accessed 23 Apr 2021

  64. Wong, J.C.: How Facebook let fake engagement distort global politics: a whistleblower’s account (2021). The Guardian. Accessed 23 Apr 2021

  65. York, J.C.: Syria’s Twitter spambots (2011). The Guardian. Accessed 23 Apr 2021

  66. York, J.C., Zuckerman, E.: Moderating the public sphere. In: Jorgensen, R.J. (ed.) Human Rights in the Age of Platforms, pp. 137–162. MIT Press (2019)

    Google Scholar 

  67. Zannettou, S.: “I Won the Election!”: an empirical analysis of soft moderation interventions on Twitter. In: ICWSM 2021, pp. 1–13 (2021)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Garfield Benjamin .

Editor information

Editors and Affiliations

A Description of the Taxonomy

A Description of the Taxonomy

Key

  • Scale [A] Individual, [B] Type of individual, [C] Community, [D] Public discourse, [E] Systemic/principles.

  • Timeframe [1] Immediate, [2] Short, [3] Mid, [4] Long, [5] Persistent.

Targets

  • Pilot: vulnerabilities during the US and subsequent pilots

  • Implementation: the transition to global context;

  • Tweets: user interactions and vectors for attacks;

  • Notes: user interactions and vectors for attacks;

  • Note ranking: user interactions and vectors for attacks;

  • Twitter management: internal decisions create/mitigate vulnerabilities;

  • Regulation: external (legislative) decisions prevent/permit vulnerabilities.

Attacks

  1. A 1

    Posts: tweets and notes, including many directly abusive tactics;

  2. A 2

    Followers: tools in extended abuse;

  3. A 2

    Bots: automating abuse to scale up coordinated attacks;

  4. A 1

    Text-as-image: an attack on whether the Birdwatch system is triggered;

  5. A 2

    Iterating tweets: as above;

  6. A 1

    Varying content or categories: as above;

  7. A 1

    Abusive or harmful content: as above;

  8. C 3

    Data: includes data poisoning of ranking system;

  9. B 2

    Algorithm: gives differential visibility or credibility to tweets/notes;

  10. C 3

    Third party: external attacks exploit vulnerabilities;

  11. E 4

    Design: internal flaws create vulnerabilities;

  12. C 3

    Injection: includes user interactions (likely) and breached security (less so);

  13. C 3

    Manipulation: includes user interactions and breached security;

  14. E 4

    Data structures: design flaws enable manipulation of data/the algorithm;

  15. D 4

    Faux transparency: data availability risks obscuring underlying structures;

  16. D 4

    External validation: public scrutiny and PR as tool for credibility;

  17. D 5

    Lobbying: pressure on regulators to prevent external constraints;

  18. D 5

    Loopholes: flaws in regulation (e.g. loose definitions) enable harmful design;

  19. D 5

    Self-regulation: Birdwatch is part of continued efforts to avoid regulation.

Harms

  1. A 2

    Coordinated attacks: combining attacks/accounts increases scale/impact;

  2. B 3

    Weaponisation: systematic targeting of certain groups/communities;

  3. A 1

    Abuse: effects (emotional, physical) against specific individuals(/groups);

  4. B 3

    Verification: shift for Twitter; harms marginalised groups with need for ID;

  5. E 4

    Policies/enforcement: precedent of unequal application/lack of context;

  6. C 5

    Access/exclusion: design, policies, implementation; method & type of harm;

  7. D 3

    Game rankings: vulnerabilities in practice and in credibility;

  8. D 4

    Ranking system: vulnerabilities to public discourse in Birdwatch design;

  9. D 4

    Avoiding moderation: not moderation; community not platform;

  10. D 5

    Avoiding regulation: visible action to placate regulators;

  11. C 3

    Exploitative labour: reliance on users; lack of protection; uneven burden;

  12. E 5

    Avoid scrutiny: systemic avoidance or deflection of external audit/criticism;

  13. E 5

    Systemic/narrative: structural impact on society; influence over debates;

  14. E 4

    Context shift: marginalisation of geospatial/cultural/etc. communities.

Rights and permissions

Reprints and permissions

Copyright information

© 2022 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Benjamin, G. (2022). Who Watches the Birdwatchers? Sociotechnical Vulnerabilities in Twitter’s Content Contextualisation. In: Parkin, S., Viganò, L. (eds) Socio-Technical Aspects in Security. STAST 2021. Lecture Notes in Computer Science, vol 13176. Springer, Cham. https://doi.org/10.1007/978-3-031-10183-0_1

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-10183-0_1

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-10182-3

  • Online ISBN: 978-3-031-10183-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics