Abstract
As user-generated online content has been flourishing with both useful information and misinformation. One of the complexities surrounding such phenomena is its huge amounts of data and its associated difficulties to effectively moderate content, particularly as most initiatives are centralised and fraught with its intrinsic trust issues. One of the few examples using mainly a decentralised (i.e., community-driven) mechanism is Twitter’s Community Notes (once named as Birdwatch) experimental project. This paper thus is about testing the efficiency of such community-based content moderation mechanism and scenarios of interest aiming to better understanding how the users themselves better moderate online content. This is done through an agent-based approach and three conclusions are discussed in detail: (1) to some extent the community is able to fight against misinformation, (2) a Birdwatch-like mechanism can indeed boost the community’s content moderation ability, but there is a nontrivial trade-off between social influence and content timeliness and (3) a simple proposition, in the form of a reminder mechanism to users, cannot fulfil the task of improving the content moderation efficiency, which means a different approach to design is needed.
Similar content being viewed by others
Data availability
The code of agent-based model and the data generated from the model can be accessed by emailing the corresponding author: wangchenlong@gacrnd.com or yunming_wang@outlook.com.
Notes
The probability of a user posting a Tweet in each time step is smaller than 1. In this model, it is set to 0.2.
This amount was chosen based on the requirement for running Birdwatch note ranking algorithms and the computing resources the author could afford at the time of writing this paper. Each run requires from 30 to 50 min using one combination of the parameters on a computer with AMD Ryzen 7 5800H CPU and 16 GB RAM memory.
The author uses “decide” not to verification only for descriptive simplicity. Users usually are to inpatient or careless to carry out verification in reality. But the description does not affect the algorithm running in the model.
References
Åkerlund M (2021) Dog whistling far-right code words: the case of ‘culture enricher’ on the Swedish web. Inf Commun Soc. https://doi.org/10.1080/1369118X.2021.1889639
Allen J, Arechar AA, Pennycook G, Rand DG (2021) Scaling up fact-checking using the wisdom of crowds. Sci Adv 7(36):eabf4393. https://doi.org/10.1126/sciadv.abf4393
Allen J, Martel C, Rand DG (2022) Birds of a feather don’t fact-check each other: partisanship and the evaluation of news in Twitter’s Birdwatch crowdsourced fact-checking program. In: Proceedings of the 2022 CHI conference on human factors in computing systems. pp 1–19. https://doi.org/10.1145/3491102.3502040
Arkes HR, Hackett C, Boehm L (1989) The generality of the relation between familiarity and judged validity. J Behav Decis Mak 2(2):81–94. https://doi.org/10.1002/bdm.3960020203
Atreja S, Hemphill L, Resnick P (2022) What is the will of the people? Moderation preferences for misinformation. http://arxiv.org/abs/2202.00799
Becker J, Brackbill D, Centola D (2017) Network dynamics of social influence in the wisdom of crowds. Proc Natl Acad Sci. https://doi.org/10.1073/pnas.1615978114
Biron B (2022) Elon Musk said Twitter’s Birdwatch feature will be renamed ‘Community Notes’ and is aimed at ‘improving information accuracy’ amid growing content-moderation concerns. Business Insider. https://www.businessinsider.com/musk-renames-birdwatch-community-notes-touts-improving-accuracy-2022-11
Bloch-Wehba H (2021) Content moderation as surveillance (SSRN Scholarly Paper ID 3872915). Social Science Research Network. https://papers.ssrn.com/abstract=3872915
Borwankar S, Zheng J, Kannan KN (2022) Democratization of misinformation monitoring: the impact of Twitter’s Birdwatch program (SSRN Scholarly Paper 4236756). https://doi.org/10.2139/ssrn.4236756
Bozarth L, Im J, Quarles C, Budak C (2023) Wisdom of two crowds: misinformation moderation on reddit and how to improve this process—a case study of COVID-19. In: Proceedings of the ACM on human–computer interaction, 7(CSCW1), 155:1–155:33. https://doi.org/10.1145/3579631
Brashier NM, Marsh EJ (2020) Judging truth. Ann Rev Psychol. https://doi.org/10.1146/annurev-psych-010419-050807
Brashier NM, Schacter DL (2020) Aging in an era of fake news. Curr Dir Psychol Sci 29(3):316–323. https://doi.org/10.1177/0963721420915872
Bromell D (2022) Deplatforming and democratic legitimacy. In: Bromell D (ed) Regulating free speech in a digital age: hate, harm and the limits of censorship. Springer International Publishing, Berlin, pp 81–109
Chhabra A, Kaur R, Iyengar SRS (2020) Dynamics of edit war sequences in Wikipedia. In: Proceedings of the 16th International symposium on open collaboration. pp 1–10. https://doi.org/10.1145/3412569.3412585
Chuai Y, Tian H, Pröllochs N, Lenzini G (2023) The roll-out of community notes did not reduce engagement with misinformation on Twitter. arXiv:2307.07960. https://doi.org/10.48550/arXiv.2307.07960
Cisco (2022) Cisco Annual Internet Report. Cisco. https://www.cisco.com/c/en/us/solutions/executive-perspectives/annual-internet-report/index.html
Curran J, Fenton N, Freedman D (2012) Misunderstanding the Internet. Routledge, London
De Vynck G (2021) Australia is demanding tech giants pay for news. Google relented, Facebook didn’t. Washington Post. https://www.washingtonpost.com/technology/2021/02/17/google-pay-news-corp/
DiFonzo N, Beckstead JW, Stupak N, Walders K (2016) Validity judgments of rumors heard multiple times: the shape of the truth effect. Soc Infl 11(1):22–39. https://doi.org/10.1080/15534510.2015.1137224
Douek E (2022) Second wave content moderation institutional design: from rights to regulatory thinking. Soc Sci Res Netw. https://doi.org/10.2139/ssrn.4005326
Ecker UKH, Lewandowsky S, Cook J, Schmid P, Fazio LK, Brashier N, Kendeou P, Vraga EK, Amazeen MA (2022) The psychological drivers of misinformation belief and its resistance to correction. Nat Rev Psychol. https://doi.org/10.1038/s44159-021-00006-y
Flache A, Mäs M, Feliciani T, Chattoe-Brown E, Deffuant G, Huet S, Lorenz J (2017) Models of social influence: towards the next frontiers. J Artif Soc Soc Simul 20(4):2
Frey V, van de Rijt A (2021) Social influence undermines the wisdom of the crowd in sequential decision making. Manage Sci 67(7):4273–4286. https://doi.org/10.1287/mnsc.2020.3713
Gillespie T (2018) Custodians of the internet: Platforms, content moderation, and the hidden decisions that shape social media. Yale University Press, New Haven
Google (2020) White paper on information quality and content moderation. https://blog.google/documents/83/information_quality_content_moderation_white_paper.pdf/
Groh M, Epstein Z, Firestone C, Picard R (2022) Deepfake detection by human crowds, machines, and machine-informed crowds. Proc Natl Acad Sci 119(1):e2110013119. https://doi.org/10.1073/pnas.2110013119
Hettiachchi D, Goncalves J (2019) Towards effective crowd-powered online content moderation. In: Proceedings of the 31st Australian conference on human-computer-interaction. pp 342–346. https://doi.org/10.1145/3369457.3369491
How can I resolve a dispute with a moderator or moderator team? (2023) Reddit Help. https://support.reddithelp.com/hc/en-us/articles/205192355-How-can-I-resolve-a-dispute-with-a-moderator-or-moderator-team-
How do I become a moderator? (2022) Reddit Help. https://support.reddithelp.com/hc/en-us/articles/205192505-How-do-I-become-a-moderator-
Howe J (2006) The rise of crowdsourcing. Wired Mag 14:5
Hubley H (2022) Bad Speech, good evidence: content moderation in the context of open-source investigations. Int Crim Law Rev. https://doi.org/10.1163/15718123-bja10124
Jaki S, De Smedt T (2019) Right-wing German hate speech on Twitter: analysis and automatic detection. arXiv:1910.07518. arXiv. https://doi.org/10.48550/arXiv.1910.07518
Jhaver S, Appling DS, Gilbert E, Bruckman A (2019) Did you suspect the post would be removed?: understanding user reactions to content removals on Reddit. In: Proceedings of the ACM on human–computer interaction, 3(CSCW), 192:1–192:33. https://doi.org/10.1145/3359294
Lai V, Carton S, Bhatnagar R, Liao QV, Zhang Y, Tan C (2022) Human-AI collaboration via conditional delegation: a case study of content moderation. CHI Conf Hum Fact Comput Syst. https://doi.org/10.1145/3491102.3501999
Laird JE, Kinkade KR, Mohan S, Xu JZ (2012) cognitive robotics using the soar cognitive architecture. In: Workshops at the twenty-sixth AAAI conference on artificial intelligence. Workshops at the twenty-sixth AAAI conference on artificial intelligence. https://www.aaai.org/ocs/index.php/WS/AAAIW12/paper/view/5221
Lazer D, Friedman A (2007) The network structure of exploration and exploitation. Adm Sci Q 52(4):667
Lenaerts K, Gillis D, Waeyaert W (2022) Occupational safety and health risks of online content review work provided through digital labour platforms. In Case study. European agency for safety and health at work EU-OSHA. https://osha.europa.eu/en/publications/occupational-safety-and-health-risks-online-content-review-work-provided-through-digital-labour-platforms
Li G, Zhu H, Lu T, Ding X, Gu N (2015) Is it good to be like Wikipedia? Exploring the trade-offs of introducing collaborative editing model to Q&A sites. In: Proceedings of the 18th ACM conference on computer supported cooperative work & social computing. pp 1080–1091. https://doi.org/10.1145/2675133.2675155
Lorenz J, Rauhut H, Kittel B (2015) Majoritarian democracy undermines truth-finding in deliberative committees. Res Politics 2(2):2053168015582287. https://doi.org/10.1177/2053168015582287
Lorenz J, Rauhut H, Schweitzer F, Helbing D (2011) How social influence can undermine the wisdom of crowd effect. Proc Natl Acad Sci. https://doi.org/10.1073/pnas.1008636108
Martel C, Pennycook G, Rand DG (2020) Reliance on emotion promotes belief in fake news. Cogn Res Princ Impl. https://doi.org/10.1186/s41235-020-00252-3
Masullo Chen G, Muddiman A, Wilner T, Pariser E, Stroud NJ (2019) We should not get rid of incivility online. Soc Med Soc 5(3):2056305119862641. https://doi.org/10.1177/2056305119862641
Metz R (2022) Elon Musk thinks Twitter’s algorithm should be public. Here’s what that could mean | CNN Business. CNN. https://www.cnn.com/2022/04/19/tech/twitter-algorithm-open-source-elon-musk/index.html
Mourão RR, Robertson CT (2019) Fake news as discursive integration: an analysis of sites that publish false, misleading. Hyperpartis Sens Inf J Stud 20(14):2077–2095. https://doi.org/10.1080/1461670X.2019.1566871
Nadarevic L, Reber R, Helmecke AJ, Köse D (2020) Perceived truth of statements and simulated social media postings: An experimental investigation of source credibility, repeated exposure, and presentation format. Cogn Res Princ Impl 5(1):56. https://doi.org/10.1186/s41235-020-00251-4
Papadogiannakis E, Papadopoulos P, Markatos EP, Kourtellis N (2022) Who funds misinformation? A systematic analysis of the ad-related profit routines of fake news sites. http://arxiv.org/abs/2202.05079
Pariser E (2011) The filter bubble: what the internet is hiding from you. Penguin, London
Pennycook G, Cannon TD, Rand DG (2018) Prior exposure increases perceived accuracy of fake news. J Exp Psychol Gen 147(12):1865–1880. https://doi.org/10.1037/xge0000465
Pennycook G, Epstein Z, Mosleh M, Arechar AA, Eckles D, Rand DG (2021) Shifting attention to accuracy can reduce misinformation online. Nature. https://doi.org/10.1038/s41586-021-03344-2
Pennycook G, Rand DG (2021) The psychology of fake news. Trends Cogn Sci 25(5):388–402. https://doi.org/10.1016/j.tics.2021.02.007
Pröllochs N (2021) Community-based fact-checking on Twitter’s Birdwatch platform. arXiv:2104.07175. https://doi.org/10.48550/arXiv.2104.07175
Rathje S, Bavel JJV, Linden van der DS (2022) Accuracy and social motivations shape judgements of (mis)information. https://doi.org/10.31234/osf.io/hkqyv
Ribeiro MM, Ortellado P (2018) Fake news: what it is and how to deal with it. Int J Hum Rights 27:69
Ritter FE, Tehranchi F, Oury JD (2019) ACT-R: a cognitive architecture for modeling cognition. Wires Cognit Sci 10(3):e1488. https://doi.org/10.1002/wcs.1488
Roberts S (2016) Commercial content moderation: digital laborers’ dirty work. Media Studies Publications. https://ir.lib.uwo.ca/commpub/12
Roth Y, Pickles N (2020) Updating our approach to misleading information. https://blog.twitter.com/en_us/topics/product/2020/updating-our-approach-to-misleading-information
Saeed M, Traub N, Nicolas M, Demartini G, Papotti P (2022) Crowdsourced fact-checking at twitter: how does the crowd compare with experts?. In: Proceedings of the 31st ACM international conference on information & knowledge management. pp 1736–1746. https://doi.org/10.1145/3511808.3557279
Savolainen L (2022) The shadow banning controversy: perceived governance and algorithmic folklore. Media Cult Soc. https://doi.org/10.1177/01634437221077174
Shabayek S, Théro H, Almanla D, Vincent E (2022) Monitoring misinformation related interventions by Facebook, Twitter and YouTube: Methods and illustration. https://hal.archives-ouvertes.fr/hal-03662191
Steiger M, Bharucha TJ, Venkatagiri S, Riedl MJ, Lease M (2021) The psychological well-being of content moderators: the emotional labor of commercial moderation and avenues for improving support. In: Proceedings of the 2021 CHI conference on human factors in computing systems. pp 1–14. https://doi.org/10.1145/3411764.3445092
Stockmann D (2022) Tech companies and the public interest: the role of the state in governing social media platforms. Inf Commun Soc. https://doi.org/10.1080/1369118X.2022.2032796
Stryker R, Conway BA, Bauldry S, Kaul V (2022) Replication note: What is political incivility? Hum Commun Res 48(1):168–177. https://doi.org/10.1093/hcr/hqab017
Sumi R, Yasseri T, Rung A, Kornai A, Kertesz J (2011) Edit Wars in Wikipedia. In: 2011 IEEE Third international conference on privacy, security, risk and trust and 2011 IEEE third international conference on social computing. pp 724–727. https://doi.org/10.1109/PASSAT/SocialCom.2011.47
Sunstein CR (2001) Echo Chambers: bush v. gore, impeachment, and beyond. Princeton University Press, Princeton
Surowiecki J (2004) The wisdom of crowds: Why the many are smarter than the few and how collective wisdom shapes business, economies, societies and nations. Choice Rev Online. https://doi.org/10.5860/CHOICE.42-1645
Torres-Lugo C, Pote M, Nwala A, Menczer F (2022) Manipulating Twitter through deletions. arXiv:2203.13893. arXiv. https://doi.org/10.48550/arXiv.2203.13893
Twitter Inc (2021) Permanent suspension of @realDonaldTrump. https://blog.twitter.com/en_us/topics/company/2020/suspension
Twitter Inc (2022a) Downloading data | Birdwatch Guide. https://twitter.github.io/birdwatch/download-data/
Twitter Inc (2022b) Note ranking | Birdwatch Guide. https://twitter.github.io/birdwatch/ranking-notes/
Twitter Inc. (2022c) Overview | Birdwatch Guide. https://twitter.github.io/birdwatch/overview/
Walther S, McCoy A (2021) US extremism on telegram: fueling disinformation, conspiracy theories, and accelerationism. Perspect Terror 15(2):100–124
What’s a moderator? (2022) Reddit Help. https://support.reddithelp.com/hc/en-us/articles/204533859-What-s-a-moderator-
Wright L (2022) Automated platform governance through visibility and scale: on the transformational power of automoderator. Soc Med Soc 8(1):20563051221077020. https://doi.org/10.1177/20563051221077020
Wright S (2006) Government-run online discussion fora: moderation, censorship and the shadow of control. Br J Polit Int Relat 8(4):550–568. https://doi.org/10.1111/j.1467-856x.2006.00247.x
Young GK (2021) How much is too much: the difficulties of social media content moderation. Inf Commun Technol Law. https://doi.org/10.1080/13600834.2021.1905593
Zeng J, Kaye DBV (2022) From content moderation to visibility moderation: a case study of platform governance on TikTok. Policy Internet. https://doi.org/10.1002/poi3.287
Acknowledgements
Both authors thank participants to the internal seminars organised by the Geary Institute in UCD.
Funding
There is no specific funding attached to this research output. The first author (i.e., Chenlong Wang) is self-funded, and he earns salary by taking tutorial classes in UCD and salary earned in GAC R&D. The second author (i.e., Pablo Lucas) is supported by UCD employee salary.
Author information
Authors and Affiliations
Contributions
The first author is responsible for literature review, model design, coding, analysis, interpretation, and manuscript writing. The second author is working as the supervisor and offers help in model design and manuscript writing, proof reading.
Corresponding author
Ethics declarations
Conflict of interest
Both authors declare that they do not have any competing interests related to the topic covered by this manuscript.
Ethical approval
We confirm that this manuscript is original and has not been published elsewhere, nor is it currently under consideration for publication elsewhere.
Informed consent
The two authors agree the journal Evaluation to get access to the manuscript, and agree to give access to the subscribers of the journal.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Below is the link to the electronic supplementary material.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Wang, C., Lucas, P. Efficiency of Community-Based Content Moderation Mechanisms: A Discussion Focused on Birdwatch. Group Decis Negot (2024). https://doi.org/10.1007/s10726-024-09881-1
Accepted:
Published:
DOI: https://doi.org/10.1007/s10726-024-09881-1