Abstract
Online Social Networks (OSNs), once regarded as safe havens for sharing information and providing mutual support among groups of people, have become breeding grounds for spreading toxic behaviors, political propaganda, and radicalizing content. Toxic individuals often hide under the auspices of anonymity to create fruitless arguments and divert the attention of other users from the core objectives of a community. In this study, we examined five recurring forms of toxicity among the comments posted on pro- and anti-NATO channels on YouTube. We leveraged the YouTube Data API to collect video and comment data from eight channels. We then utilized Google’s Perspective API to assign toxic scores to each comment. Our analysis suggests that, on average, commenters on the anti-NATO channels are more likely to be more toxic than those on the pro-NATO channels. We further discovered that commenters on pro-NATO channels tend to use a mixture of toxic and innocuous comments. We generated word clouds to get an idea of word use frequency, as well as applied the Latent Dirichlet Allocation topic model to classify the comments into their overall topics. The topics extracted from the pro-NATO channels’ comments were primarily positive, such as “Alliance” and “United”; whereas, the topics extracted from anti-NATO channels’ comments were more geared towards geographical locations, such as “Russia”, and negative components such as “Profanity” and “Fake News”. By identifying and examining the toxic behaviors of commenters on YouTube, our analysis lends aid to the pressing need for understanding this toxicity.
Keywords
- Social network analysis
- Topic modeling
- Toxicity analysis
This is a preview of subscription content, access via your institution.
Buying options





References
Cheng, J., Bernstein, M., Danescu-Niculescu-Mizil, C., Leskovec, J.: Anyone can become a troll: causes of trolling behavior in online discussions. In: Proceedings of the 2017 ACM Conference on Computer Supported Cooperative Work and Social Computing - CSCW 2017, pp. 1217–1230. ACM Press, Portland (2017)
Lee, S.-H., Kim, H.-W.: Why people post benevolent and malicious comments online. Commun. ACM 58, 74–79 (2015)
Wulczyn, E., Thain, N., Dixon, L.: Ex machina: personal attacks seen at scale. arXiv:1610.08914 [cs] (2016)
Nobata, C., Tetreault, J., Thomas, A., Mehdad, Y., Chang, Y.: Abusive language detection in online user content. Presented at the (2016)
Martens, M., Shen, S., Iosup, A., Kuipers, F.: Toxicity detection in multiplayer online games. In: 2015 International Workshop on Network and Systems Support for Games (NetGames), pp. 1–6. IEEE, Zagreb (2015)
Online Harassment | Pew Research Center (2014). http://www.pewinternet.org/2014/10/22/online-harassment/
Chen, Y., Zhou, Y., Zhu, S., Xu, H.: Detecting offensive language in social media to protect adolescent online safety. In: 2012 International Conference on Privacy, Security, Risk and Trust and 2012 International Conference on Social Computing, pp. 71–80. IEEE, Amsterdam (2012)
Cao, Q., Yang, X., Yu, J., Palow, C.: Uncovering Large Groups of Active Malicious Accounts in Online Social Networks. Presented at the (2014)
Perspective. http://perspectiveapi.com/#/
Varjas, K., Talley, J., Meyers, J., Parris, L., Cutts, H.: High school students’ perceptions of motivations for cyberbullying: an exploratory study. West J. Emerg. Med. 11, 269–273 (2010)
Shachaf, P., Hara, N.: Beyond vandalism: Wikipedia trolls. J. Inf. Sci. 36, 357–370 (2010)
Warner, W., Hirschberg, J.: Detecting hate speech on the world wide web. In: Proceedings of the Second Workshop on Language in Social Media, pp. 19–26. Association for Computational Linguistics, Montréal (2012)
Suler, J.: The online disinhibition effect. Cyberpsychol. Behav. 7(3), 321–326 (2004)
Cheng, J., Danescu-Niculescu-Mizil, C., Leskovec, J.: Antisocial behavior in online discussion communities, 10
Yin, D., Xue, Z., Hong, L.: Detection of harassment on Web 2.0., 7
Sood, S.O., Antin, J., Churchill, E.F.: Using crowdsourcing to improve profanity detection, 6
Davidson, T., Warmsley, D., Macy, M., Weber, I.: Automated hate speech detection and the problem of offensive language. In: Proceedings of the Eleventh International AAAI Conference on Web and Social Media (ICWSM 2017) (2017)
Hosseini, H., Kannan, S., Zhang, B., Poovendran, R.: Deceiving Google’s perspective API built for detecting toxic comments. arXiv:1702.08138 [cs] (2017)
Gröndahl, T., Pajola, L., Juuti, M., Conti, M., Asokan, N.: All you need is “love”: evading hate-speech detection. arXiv:1808.09115 [cs] (2018)
Carley, K.M., Reminga, J.: ORA: Organization Risk Analyzer: Defense Technical Information Center. Fort Belvoir, VA (2004)
Responding to Cognitive Security Challenges | StratCom. https://www.stratcomcoe.org/responding-cognitive-security-challenges
Blei, D.M.: Latent Dirichlet allocation, 30
van der Maaten, L., Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res. 9, 2579–2605 (2008)
Acknowledgements
This research is funded in part by the U.S. National Science Foundation (IIS-1636933, ACI-1429160, and IIS-1110868), U.S. Office of Naval Research (N00014-10-1-0091, N00014-14-1-0489, N00014-15-P-1187, N00014-16-1-2016, N00014-16-1-2412, N00014-17-1-2605, N00014-17-1-2675), U.S. Air Force Research Lab, U.S. Army Research Office (W911NF-16-1-0189), U.S. Defense Advanced Research Projects Agency (W31P4Q-17-C-0059), Jerry L. Maulden/Entergy Endowment at the University of Arkansas at Little Rock and the Arkansas Research Alliance. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the funding organizations. The researchers gratefully acknowledge the support.
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Obadimu, A., Mead, E., Hussain, M.N., Agarwal, N. (2019). Identifying Toxicity Within YouTube Video Comment. In: Thomson, R., Bisgin, H., Dancy, C., Hyder, A. (eds) Social, Cultural, and Behavioral Modeling. SBP-BRiMS 2019. Lecture Notes in Computer Science(), vol 11549. Springer, Cham. https://doi.org/10.1007/978-3-030-21741-9_22
Download citation
DOI: https://doi.org/10.1007/978-3-030-21741-9_22
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-21740-2
Online ISBN: 978-3-030-21741-9
eBook Packages: Computer ScienceComputer Science (R0)