Abstract
Recent work on language technology has tried to recognize abusive language such as those containing hate speech and cyberbullying and enhance offensive language identification to moderate social media platforms. Most of these systems depend on machine learning models using a tagged dataset. Such models have been successful in detecting and eradicating negativity. However, an additional study has lately been conducted on the enhancement of free expression through social media. Instead of eliminating ostensibly unpleasant words, we created a multilingual dataset to recognize and encourage positivity in the comments, and we propose a novel custom deep network architecture, which uses a concatenation of embedding from T5-Sentence. We have experimented with multiple machine learning models, including SVM, logistic regression, K-nearest neighbour, decision tree, logistic neighbours, and we propose new CNN based model. Our proposed model outperformed all others with a macro F1-score of 0.75 for English, 0.62 for Tamil, and 0.67 for Malayalam.
Similar content being viewed by others
1 Introduction
Artificial intelligence is a rapidly evolving technology that significantly influences the global economy and society. Equality, Diversity, and Inclusion (EDI) have recently received a lot of attention, and an emphasis has been placed on protected classifications, including gender and race. Varied definitions of EDI may have different connotations depending on the context. EDI for an organization or workplace, for example, may have quite different priorities than EDI for a science community or a research subject, such as energy and AI. The emphasis may be on internal cultural changes in the workplace, with a small number of people serving as significant stakeholders. However, the larger academic community needs a much broader global perspective to define EDI and examine the entire society rather than individuals. The common core must be shared, and fair treatment must be assured, with equal opportunity for all and the abolition of all forms of discrimination. It began in 1960, but the definition of diversity has since expanded to include additional demographics such as lesbian, gay, bisexual, transgender, intersex, and queer/questioning (LGBTIQ+), women in science, engineering, technology, and management (STEM), and people with disabilities (Roberson et al. 2017).
Equality is defined as the fair and unbiased treatment of all individuals and groups. It is also expected that all groups would have an equal opportunity: the disadvantaged will have access to the same opportunities for advancement and accomplishment as their peers. Diversity is about being ’different’ and the manner in which it is expressed. It is often measured by the representation of people and groups from varied origins and viewpoints, but the essential idea is to acknowledge, accept, and celebrate such variety to stimulate creativity and innovation. Inclusion: An inclusive environment is one in which everyone can thrive and is valued. A diverse team has a number of points of view; inclusion ensures that those points are acknowledged and valued. In the research arena, it is equally crucial to have an inclusive approach to research and guarantee that research benefits all users, especially historically excluded communities (Roberson 2006; Shore et al. 2011; Xuan and Ocone 2022; Mehta et al. 2018). Nevertheless, the EDI for minority LGBTIQ+ or marginalized populations has not been considered with great urgency or importance compared to other topics or areas from the perspective of language technologies research (Cech and Waidzunas 2021). It is essential that language technologies are developed to consider the inclusion of all communities for social integration.
People have two images, one for the actual world they live in and another for the virtual world, such as the photos on social media platforms, where they are linked to close friends and converse with strangers in the virtual environment. Social networking services such as Instagram, Facebook, LinkedIn, and YouTube have become the default destination for individuals all around the world to spend their time. These social channels are utilized to not just share achievements but also request assistance in times of disaster. Many people’s lifestyles have been altered as a result of recent improvements in social media, and their everyday lives have been expanded with the virtual territory of the Internet and social networks. Social media platforms significantly impact users’ daily life. Users may post positive vibes, hope, or motivational information to provide positive proposals for peace or conquering problems such as COVID-19, conflict, or elections (Gowen et al. 2012; Yates et al. 2017; Wang and Jurgens 2018; Anderson et al. 2020). Several areas were affected worldwide, and the fear of losing a loved one caused the closure of even basic necessities such as schools, hospitals, and mental healthcare centers (Pérez-Escoda et al. 2020). Consequently, people were forced to look at online forums for their informational and emotional needs. In some areas, and for some people, online social networking was the only means of social connectivity and support during the COVID-19 pandemic (Elmer et al. 2020). As a result, individuals were driven to seek knowledge and emotional support from internet communities.
Online social networking provides a venue for network members to be in the know and to be known, both of which are more important with more social integration. Social inclusion is critical for the general well-being of all persons, particularly vulnerable individuals who are more vulnerable to social exclusion. A sense of belonging and community is an essential part of people’s mental health, influencing both their psychological and physical well-being (Rook and Charles 2017). People from marginalized groups, such as women in STEM fields, LGBTIQ+ people, people from racial minorities, and people with disabilities have been studied extensively, and it has been demonstrated that their online lives significantly impact their mental health (Chung 2013; Altszyler et al. 2018; Tortoreto et al. 2019). However, the contents from social media comments/posts may be negative/hateful/offensive/abusive, as there is no mediating authority.
Comments/posts on online social media have been analyzed to find and stem the spread of negativity using methods such as hate speech detection (Schmidt and Wiegand 2017), homophobia/transphobia detection (Chakravarthi et al. 2022c), offensive language identification (Zampieri et al. 2019a), and abusive language detection (Lee et al. 2018). However, according to Davidson et al. (2019), technologies developed for the detection of abusive language do not consider the potential biases of the dataset they are trained on. Because of persistent racial bias in the datasets, detection of abusive language is skewed. It may discriminate against one group more than another in some cases. This detrimentally influences people who are members of minority communities or are marginalized. As language is an essential component of communication, it should be accessible to everybody. A large internet community that uses language technology directly influences individuals worldwide, regardless of where they live. Instead of restricting an individual’s freedom of expression by eliminating unpleasant comments, we should direct our efforts to spreading optimism and encouraging others to do the same. On the other hand, hope speech detection should be carried out in conjunction with hate speech detection. Otherwise, hope speech recognition on its own may result in prejudice, and those who make harmful and damaging remarks on the Internet would continue to behave erratically.
Hence, we have shifted our study attention to the topic of hope speech. The promise, potential, support, comfort, recommendations, or inspiration offered to participants by their peers during moments of illness, stress, loneliness, and sadness are all typically connected with the term “hope” (Snyder et al. 2002). Psychologists, sociologists, and social workers in the Association of Hope have concluded that hope can also be a useful tool for saving people from committing suicide or harming themselves (Herrestad and Biong 2010). For example, the Hope Speech delivered by gay rights activist Harvey Milk on the steps of the San Francisco City Hall during a mass rally to celebrate California Gay Freedom Day on 25 June 1978Footnote 1 inspired millions to demand rights for EDI (Milk 1997). Recently, Palakodety et al. (2020a) investigated how to utilize hope speech from social media texts to defuse tensions between two nuclear weapon-possessing states (India and Pakistan) and help marginalized Rohingya refugees (Palakodety et al. 2020b). Additionally, they experimented with identifying the presence of hope against its absence. Although no past research has been done on hope speeches for women in STEM, LGBTIQ+ folks, racial minorities, or people with disabilities in general, we believe that this area needs research.
Furthermore, despite the fact that people from a variety of linguistic origins are exposed to online social media language, English continues to be at the forefront of current developments in language technology research. Recently, various investigations have been carried out on languages with a lot of resources, such as Arabic, German, Hindi, and Italian, to name a few. The majority of such research relies on monolingual corpora and does not analyze code-switched textual data (Sciullo et al. 1986). We introduce a dataset for hope speech identification not only in English but also in under-resourced code-switched Tamil (ISO 639-3: tam), Malayalam (ISO 639-3: mal), and Kannada (ISO 639-3: kan) languages. We have experimented with multiple machine learning models, including SVM, logistic regression, K-nearest neighbour, decision tree, and logistic neighbours. Our proposed model outperformed all the others with a macro F1-score of 0.75 for English, 0.62 for Tamil, and 0.67 for Malayalam.
-
We propose to encourage hope speech rather than take away an individual’s freedom of speech by detecting and removing a negative comment.
-
We apply the schema to create a multilingual hope speech dataset for EDI. This is a new large-scale dataset of English, Tamil (code-mixed), and Malayalam (code-mixed) YouTube comments with high-quality annotation of the target.
-
We have experimented with multiple machine learning models, including SVM, logistic regression, K-nearest neighbour, decision tree, logistic neighbours, and we propose new CNN based model. Our proposed model outperformed all the others with a macro F1-score of 0.75 for English, 0.62 for Tamil, and 0.67 for Malayalam.
2 Related works
When it comes to crawling social media data, there are many works on YouTube mining (Marrese-Taylor et al. 2017; Muralidhar et al. 2018), mainly focused on exploiting user comments. Krishna et al. (2013) did an opinion mining and trend analysis on YouTube comments. The researchers analyzed sentiments to identify their trends, seasonality, and forecasts; user sentiments were found to be well-correlated with the influence of real-world events. Severyn et al. (2014) systematically studied opinion mining targeting YouTube comments. The authors developed a comment corpus containing 35K manually labeled data for modelling the opinion polarity of the comments based on tree kernel models. Chakravarthi et al. (2020a), Chakravarthi et al. (2020b), Sampath et al. (2022), Chakravarthi et al. (2022b), and B et al. (2022b) collected comments from YouTube and created a manually annotated corpus for the sentiment analysis, emotional analysis, offensive language identification, and multimodal sentiment analysis of the under-resourced Tamil and Malayalam languages.
Methods to mitigate gender bias in natural language processing (NLP) have been extensively studied for the English language (Sun et al. 2019). Some studies have investigated gender bias beyond the English language using machine translation to French (Vanmassenhove et al. 2018) and other languages (Prates et al. 2020). Tatman (2017) studied the gender and dialect bias in automatically generated captions from YouTube. Technologies for abusive language (Waseem et al. 2017; Clarke and Grieve 2017), hate speech (Schmidt and Wiegand 2017; Ousidhoum et al. 2019), and offensive language detection (Nogueira dos Santos et al. 2018; Zampieri et al. 2019b; Sigurbergsson and Derczynski 2020) are being developed and applied without considering the potential biases (Davidson et al. 2019; Wiegand et al. 2019; Xia et al. 2020). However, current gender de-biasing methods in NLP are insufficient to de-bias other issues related to EDI in end-to-end systems of many language technology applications, which causes unrest and escalates the issues with EDI, as well as exacerbating inequality on digital platforms (Robinson et al. 2020).
Counter-narratives (i.e., informed textual responses) are another strategy that has received the attention of researchers recently (Chung et al. 2019; Tekiroğlu et al. 2020). A counter-narrative approach was proposed to weigh the right to freedom of speech and avoid over-blocking. Mathew et al. (2019) created and released a dataset for counter-speech using comments from YouTube. However, the core idea of directly intervening with textual responses escalates hostility even though it is advantageous to the writer to understand why their comment/post has been deleted or blocked and then favorably change the discourse and attitudes of their comments. Thus, we diverted our research to finding positive information such as hope and encouraging such activities.
Deep neural network models based on transformers have been used to detect abusive remarks on Bangla social media Aurpa et al. (2021), Lucky et al. (2021). Pre-training language architectures such as BERT (Bidirectional Encoder Representations from Transformers) and ELECTRA (Efficiency Learning an Encoder that Classifies Token Replacements Accurately) are used in conjunction. The authors created a unique dataset, which consists of 44,001 comments from a large number of different Facebook posts in the Bangla language.
Our work differs from the previous works in that we define hope speech for EDI, and we introduce a dataset for English, Tamil, and Malayalam on EDI of it. To the best of our knowledge, this is the first work to create a dataset for EDI in the under-resourced Tamil and Malayalam languages. We also created a dataset for the Kannada language Hande et al. (2021) in continuation of this research.
3 Hope speech
Hope is an upbeat state of mind based on a desire for positive outcomes in one’s life or the world at large, and it is both present- and future-oriented (Snyder et al. 2002). Inspirational talks about how people deal with and overcome adversity may also provide hope. Hope speech instills optimism and resilience, which beneficially impacts many parts of life, including (Youssef and Luthans 2007), college (Chang 1998), and other factors that put us at risk (Cover 2013). We define hope speech for our problem as “YouTube comments/posts that offer support, reassurance, suggestions, inspiration, and insight” (Chakravarthi 2020; Chakravarthi and Muralidaran 2021; Chakravarthi et al. 2022a).
The notion that one may uncover and become motivated to use routes to their desired goals is reflected in hope speech. Our approach seeks to shift the dominant mindset away from a focus on discrimination, loneliness, or the negative aspects of life toward a focus on creating confidence, support, and positive characteristics based on individual remarks. Thus, we provide instructions to annotators that if a comment/post meets the following conditions, it should be annotated as hope speech.
-
The comment contains inspiration provided to participants by their peers and others and/or offers support, reassurance, suggestions, and insight.
-
The comment promotes well-being and satisfaction (past), joy, sensual pleasures, and happiness (present).
-
The comment triggers constructive cognition about the future – optimism, hope, and faith.
-
The comment expresses love, courage, interpersonal skill, aesthetic sensibility, perseverance, forgiveness, tolerance, future-mindedness, praise for talents, and wisdom.
-
The comment promotes equality, diversity, and inclusion.
-
The comment brings out a survival story of gay, lesbian, or transgender individuals, women in science, or a COVID-19 survivor.
-
The comment talks about fairness in the industry. (e.g., [I do not think banning all apps is right; we should ban only the apps that are unsafe.]).
-
The comment explicitly talks about a hopeful future. (e.g., [We will survive these times.]).
-
The comment explicitly talks about and says no to division in any form.
Non-hope speech includes comments that do not exude positivity, such as the following:
-
The comment uses racially, ethnically, sexually, or nationally motivated slurs.
-
The comment promotes hate towards a minority.
-
The comment is very prejudiced and attacks people without thinking about the consequences.
-
The comment does not inspire hope in the reader’s mind.
4 Dataset construction
We concentrated on gathering information from comments on YouTubeFootnote 2, which is the most widely used platform in the world to comment and publicly express opinions about topics or videos. We didn’t include comments from LGBTIQ+ people’s personal coming out stories, as they contained personal information. For English, we gathered information on recent EDI themes such as women in STEM, LGBTIQ+ concerns, COVID-19, Black Lives Matter, United Kingdom (UK) vs China, the United States of America (USA), and Australia versus China. The information was gathered from recordings of individuals in English-speaking nations such as Australia, Canada, Ireland, the United Kingdom, the United States of America, and New Zealand.
We gathered data from India for Tamil and Malayalam on recent themes such as LGBTIQ+ concerns, COVID-19, women in STEM, the Indo-China war, and Dravidian affairs. India is a multilingual and multiracial country. In terms of linguistics, India is split into three major language families: Dravidian, Indo-Aryan, and Tibeto-Burman. The ongoing Indo-China border conflict has sparked online bigotry toward persons with East Asian characteristics, despite the fact that they are Indians from the North-Eastern regions. Similarly, in Tamil Nadu, the National Education Policy, which calls for the adoption of Sanskrit or Hindi, has exacerbated concerns about the linguistic autonomy of Dravidian languages. We used the YouTube comment scraperFootnote 3 to collect comments. From November 2019 to June 2020, we gathered data on the aforementioned subjects. We feel that our statistics will help reduce animosity and promote optimism. Our dataset was created as a multilingual resource to enable cross-lingual research and analysis. It includes hope speech in English, Tamil, and Malayalam, among other languages. The word cloud representation of the dataset is depicted in Fig. 1.
4.1 Ethical concerns
Data from social media, especially data concerning minorities such as the LGBTIQ+ community or women, is extremely sensitive. By eliminating personal information from the dataset, such as names but not celebrity names, we have taken great care to reduce the danger of revealing individual identity in the data. However, to investigate EDI, we needed to track information on race, gender, sexual orientation, ethnicity, and philosophical views. Annotators only viewed anonymized postings and promised not to contact the author of the remark. Only researchers who agreed to follow ethical norms had access to the dataset for research purposes. After a lengthy debate with our local EDI committee members, we opted not to ask the annotator for racial information.Footnote 4 Due to recent events, the EDI committee was firmly against the collection of racial information, based on the belief that it would split people on racial lines. Thus, we collected only the nationality of the annotators.
The information gathered via social media is particularly sensitive, especially pertaining to minorities, such as the LGBTIQ+ community or females. By removing personally identifiable information from the dataset, such as names but not celebrity names, we have made a great effort to minimize the possibility of individual identity being revealed in the data. It was necessary, however, in order to conduct an investigation into EDI to keep track of information about race and gender, sexual orientation, ethnicity, and philosophical viewpoints. Annotators were only allowed to examine anonymous remarks and were not allowed to contact the person who made the remark. Only researchers who agree to adhere to ethical standards will be granted access to the dataset for the purpose of conducting research. After a lengthy discussion with our local EDI committee members, we decided not to ask the annotator for racial information. We have defined the EDI community as people who are women in STEM fields, people who are members of the LGBTIQ+ community, people who are members of racial minorities, or people who are disabled for this study. EDI committee was vehemently opposed to the collection of race information because of previous occurrences, believing that it would cause individuals to be divided based on their racial origin. As a result, we only collected information on the nationality of annotators.
4.2 Annotation setup
After the data collection phase, we cleaned the data using LangdetectFootnote 5 to identify the language of the comments and removed comments that were not in the specified languages. However, owing to code-mixing at various levels, unintentional comments of other languages remained in the cleaned corpus of Tamil and Malayalam comments. Finally, based on our description from Section 3, we identified three groups, two of which are hope and non-hope, while the last (Other languages) was introduced to account for comments not in the required language. These classes were chosen because they provided a sufficient amount of generalization for describing the remarks in the EDI hope speech dataset.
4.3 Annotators
We set up Google forms to collect annotations from annotators, which you can see below. Each form was restricted to 100 comments, and each page was limited to ten comments to maintain the level of annotation. Gender, educational background, and preferred medium of instruction of the annotator were all obtained to understand the annotator’s diversity and minimize prejudice in the annotation process. Those who participated in the annotation process were informed that the comments might contain profanity and hateful content. They had the option of discontinuing annotation if they found the comments too hurtful or burdensome. We educated annotators by introducing them to YouTube videos on electronic data interchange (EDI).Footnote 6Footnote 7Footnote 8Footnote 9. Each form was annotated by at least three individuals. After the annotators marked the first form with 100 comments, the findings were manually validated as a warm-up phase. This strategy was utilized to help them gain a better knowledge of EDI and focus on the project. Following the initial stage of annotating their first form, a few annotators withdrew and their remarks were deleted. The annotators were requested to do another evaluation of the EDI videos and annotation guidelines. From Table 1, we can see the statistics of annotators. Annotators for the English language remarks came from Australia, Ireland, the United Kingdom, and the United States of America. We were able to obtain annotations in Tamil from persons from both India’s Tamil Nadu state and Sri Lanka. Graduate and post-graduate students constituted the majority of the annotators.
4.4 Inter-annotator agreement
We used the majority to aggregate the hope speech annotations from several annotators; the comments that did not get a majority in the first round were gathered and added to a second Google form to allow more annotators to contribute them. We calculated the inter-annotator agreement following the last round of annotation. We quantified the clarity of the annotation and reported on inter-annotator agreement using Krippendorff’s alpha. Krippendorff’s alpha applies to all these metrics. Our annotations achieved an agreement of 0.63, 0.76, and 0.85 for English, Tamil, and Malayalam, respectively, using the nominal measure. Previous research on sentiment analysis annotations and offensive language identification for Tamil and Malayalam in the code switched settings achieved 0.69 for Tamil, 0.87 for Tamil in sentiment analysis 0.74 for Tamil and 0.83 for Malayalam in offensive language. Our IAA values for hope speech are close to the previous research on sentiment analysis and offensive language identification in Dravidian languages.
4.5 Corpus statistics
Our dataset contains 59,354 YouTube comments, with 28,451 comments in English, 20,198 in Tamil, and 10,705 in Malayalam. Our dataset also includes 59,354 comments in other languages. The distribution of our dataset is depicted in Table 2. When tokenizing words and phrases in the comments, we used the nltk tool to obtain corpus statistics for use in research. Tamil and Malayalam have a broad vocabulary as a result of the various types of code-switching that take place.
Table 3 shows the distribution of the annotated dataset by the label in the reference tab: data distribution. Hence, the data is biased, with nearly all of the comments being classified as “not optimistic” (NOT). An automatic detection system that can manage imbalanced data is essential for being really successful in the age of user-generated content on internet platforms, which is increasingly popular. Using the fully annotated dataset, a train set, a development set, and a test set were produced.
A few samples from the dataset, together with their translations and hope speech class annotations, are shown below.
-
kashtam thaan. irundhaalum muyarchi seivom-It is indeed difficult. Let us try it out though. Hope speech.
-
uff. China mon vannallo-Phew! Here comes the Chinese guy Non-hope speech
-
paambu kari saappittu namma uyirai vaanguranunga-These guys (Chinese) eat snake meat and make our lives miserable Non-hope speech
5 Benchmark experiments
We presented our dataset utilizing a broad range of common classifiers on the dataset’s imbalanced parameters, and the results were quite promising. The experiment was conducted on the token frequency-inverse document frequency (TF-IDF) relationship between tokens and documents. To generate baseline classifiers, we utilized the sklearn package (https://scikit-learn.org/stable/) from the sklearn project. Alpha = 0.7 was used for the multinomial Naive Bayes model. We employed a grid search for the k-nearest neighbors (KNN), support vector machine (SVM), decision tree, logistic regression, and decision tree and logistic regression (Fig. 2).
We propose a novel custom deep network architecture, which will hereby be referred to as CNN, shown in Fig. 3, which uses a concatenation of embedding from T5-Sentence Ni et al. (2021). T5-Sentence is a sentence-embedding version of the T5 Raffel et al. (2019), which achieved state-of-the-art results on sentence representation learning benchmark SentGLUE, an extension using nine tasks from GLUE Wang et al. (2018). Moreover, the CNN uses Indic-BERTKakwani et al. (2020), which is a pre-trained language model based on ALBERT Lan et al. (2019), trained on 11 major language of Indian origin, which is part of the on-going Indic NLP research, as the input for the deep net.
Using the training dataset, we trained our models; the development dataset was used to fine-tune the hyper-parameters, and the models were assessed by predicting labels for the held-out test set, as shown in Table 4. The performance of the categorization was measured using a macro-averaged F1-Score, which was derived by averaging accuracy and recall over a large number of trials. The motive behind this decision is the uneven class distribution, which causes well-known measures of performance such as accuracy and the micro-average F1-Score to be less than accurately representative of the actual performance. As the overall performance of all classes is important, we also presented the weighted-precision, weighted-recall, and weighted F1-Score of the individual courses in addition to the overall performance. The three tables in this section provide the precision, recall, and F1-Score findings of the HopeEDI test set, employing baseline classifiers in conjunction with support from test data: Tables 5, 6, and 7. The visual representation of the macro-precision, macro-recall and macro-F1 score of the proposed novel architecture model is available in Fig. 4. Similarly, the weighted-precision, weighted-recall and weighted-F1 are visualized as Fig. 5.
As demonstrated, all of the models performed poorly because of an issue with class imbalance. Using the HopeEDI dataset, the SVM classifier had the worst performance, with a macro-average F1-Score of 0.32, 0.21, and 0.28 for English, Tamil, and Malayalam, respectively. The decision tree obtained a higher macro F1-Score for English and Malayalam than the logistic regression did; however, Tamil fared well in both tests. To eliminate non-intended language comments from our dataset, we applied language identification techniques. Other languages were annotated in some comments by annotators, although this was not the case in all of them. Another inconsistency was introduced into our dataset as a result of this. The majority of the macro scores were lower for English because of the “Other language” category. In the case of English, this could have been prevented by simply eliminating those comments from the dataset. However, this label was required for Tamil and Malayalam, as the comments in these languages were code-mixed and written in a script that was not native to the language (Latin script). The distribution of data for the Tamil language was roughly equal between the hope and non-hope classes.
The usefulness of our dataset was evaluated through the use of machine learning techniques, which we used in our trials. Because of its novel method of data collection and annotation, we believe that the HopeEDI dataset has the potential to revolutionize the field of language technology. We believe that it will open up new directions in the future for further research on positivity.
6 Task Description
We also organized a shared task to invite more researchers toward hope speech detection and benchmark the data in LTEDI 2021 and LTEDI 2022 workshops.
Overall, we received a total of 31, 31, and 30 submissions for English, Malayalam, and Tamil tasks at LTEDI 2021. It is interesting to note that the top-performing teams in all the three languages predominantly used XLM-RoBERTa to complete the shared task. One of the top-ranking teams for English used context-aware string embeddings for word representations and recurrent neural networks and pooled document embeddings for text representation. Among the other submissions, although Bi-LSTM was popular, other machine learning and deep learning models were used. However, they did not achieve good results compared to the RoBERTa-based models (Tables 7, 8, 9, 10, 11, 12, 13).
The top scores were 0.61, 0.85, and 0.93 for Tamil, Malayalam, and English. The range of scores was between 0.37 to 0.61, 0.49 to 0.85, and 0.61 to 0.93 for Tamil, Malayalam, and English datasets, respectively, at LTEDI 2021. It can be seen that the F1 scores of all the submissions on the Tamil dataset were considerably lower than those of Malayalam and English. It is not surprising that the English scores were better, because many approaches used variations of pre-trained transformer-based models trained on English data. Due to code-mixing at various levels, the scores are naturally lower for Malayalam and Tamil datasets. Among these two, the systems submitted performed poorly on Tamil data. The identification of the exact reasons for the bad performance in Tamil requires further research. However, one possible explanation for this could be that the distribution of ‘Hope_speech’ and ‘Non_hope_speech’ classes is starkly different from those of English and Malayalam. In the remaining two classes, the number of non-hope speech comments was significantly higher than hope speech comments.
The total of submissions received for the classification of English, Tamil, and Malayalam datasets for LTEDI 2022 shared task were 13,7, and 9, respectively. An ensemble of several machine learning classifiers, such as logistic regression, multinomial Naive Bayes, random forest, and support vector machines received the most votes out of all of the other submissions. On the other hand, we found that the performances of the machine learning classifiers that were utilized for this shared task were somewhat less than the baseline performances of the ML models employed the previous year. Although LSTM, BiLSTM, and CNN were utilized, the performance of these models was not as satisfactory as that of the transformer-based models.
7 Conclusion
As online content increases massively, it is necessary to encourage positivity, such as hope speech in online forums, to induce compassion and acceptable social behaviour. This paper presented the largest manually annotated dataset of hope speech detection in English, Tamil and Malayalam, consisting of 28,451, 20,198 and 10,705 comments, respectively. We propose a novel custom deep network architecture, which uses a concatenation of embedding from T5-Sentence. We have experimented with multiple machine learning models, including SVM, logistic regression, K-nearest neighbour, decision tree and logistic neighbours. Our proposed model outperformed all the models with a macro F1-score of 0.75 for English, 0.62 for Tamil, and 0.67 for Malayalam. We believe that this dataset will facilitate future research on encouraging positivity. We aim to promote research in hope speech and encourage positive content in online social media for EDI. In the future, we plan to extend the study by introducing a larger dataset with further fine-grained classification and content analysis.
Notes
We have considered women in the fields of STEM, people who belong to the LGBTIQ+ community, racial minorities, or people with disabilities as EDI community for this study.
References
Altszyler E, Berenstein AJ, Milne D, Calvo RA, Fernandez Slezak D (2018) Using contextual information for automatic triage of posts in a peer-support forum. In: Proceedings of the fifth workshop on computational linguistics and clinical psychology: from keyboard to clinic. Association for computational linguistics, New Orleans, LA, pp 57–68, https://doi.org/10.18653/v1/W18-0606, https://www.aclweb.org/anthology/W18-0606
Anderson RM, Heesterbeek H, Klinkenberg D, Hollingsworth TD (2020) How will country-based mitigation measures influence the course of the covid-19 epidemic? The Lancet 395(10228):931–934
Aurpa TT, Sadik R, Ahmed MS (2021) Abusive bangla comments detection on facebook using transformer-based deep learning models. Soc Netw Anal Min 12(1):24. https://doi.org/10.1007/s13278-021-00852-x
Awatramani V (2021) Hopeful NLP@LT-EDI-EACL2021: finding hope in YouTube comment section. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Online
Bharathi B, Srinivasan D, Varsha J, Durairaj T, Senthil KB (2022a) SSNCSE_NLP@LT-EDI-ACL2022: hope speech detection for equality, diversity and inclusion using sentence transformers. In: Proceedings of the second workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Dublin, Ireland, pp 218–222. https://doi.org/10.18653/v1/2022.ltedi-1.30, https://aclanthology.org/2022.ltedi-1.30
Premijith B, Chakravarthi BR, Subramanian M, Bharathi B, Soman KP, Dhanalaskhmi V, Sreelakshmi K, Pandian A, Kumaresan P (2022b) Findings of the shared task on multimodal sentiment analysis and troll meme classification in Dravidian languages. In: Proceedings of the Second Workshop on Speech and Language Technologies for Dravidian Languages, association for computational linguistics, Dublin, Ireland, pp 254–260. https://doi.org/10.18653/v1/2022.dravidianlangtech-1.39, https://aclanthology.org/2022.dravidianlangtech-1.39
Balouchzahi F, Aparna BK, Shashirekha HL (2021) MUCS@LT-EDI-EACL2021: CoHope-hope speech detection for equality, diversity, and inclusion in code-mixed texts. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Online. https://aclanthology.org/2021.ltedi-1.27
Balouchzahi F, Butt S, Sidorov G, Gelbukh A (2022) CIC@LT-EDI-ACL2022: Are transformers the only hope? Hope speech detection for Spanish and English comments. In: Proceedings of the second workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Dublin, Ireland, pp 206–211. https://doi.org/10.18653/v1/2022.ltedi-1.28, https://aclanthology.org/2022.ltedi-1.28
Cech EA, Waidzunas T (2021) Systemic inequalities for lgbtq professionals in stem. Sci Adv 7(3):eabe0933
Chakravarthi BR (2020) HopeEDI: a multilingual hope speech detection dataset for equality, diversity, and inclusion. In: Proceedings of the third workshop on computational modeling of people’s opinions, personality, and emotion’s in social media, association for computational linguistics, Barcelona, Spain (Online), pp 41–53. https://www.aclweb.org/anthology/2020.peoples-1.5
Chakravarthi BR, Muralidaran V (2021) Findings of the shared task on hope speech detection for equality, diversity, and inclusion. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Kyiv, pp 61–72. https://aclanthology.org/2021.ltedi-1.8
Chakravarthi BR, Jose N, Suryawanshi S, Sherly E, McCrae JP (2020a) A sentiment analysis dataset for code-mixed Malayalam-English. In: Proceedings of the 1st joint workshop on spoken language technologies for under-resourced languages (SLTU) and collaboration and computing for under-resourced languages (CCURL), European Language Resources association, Marseille, France, pp 177–184. https://www.aclweb.org/anthology/2020.sltu-1.25
Chakravarthi BR, Muralidaran V, Priyadharshini R, McCrae JP (2020b) Corpus creation for sentiment analysis in code-mixed Tamil-English text. In: Proceedings of the 1st joint workshop on spoken language technologies for under-resourced languages (SLTU) and collaboration and computing for under-resourced languages (CCURL), European Language Resources association, Marseille, France, pp 202–210. https://www.aclweb.org/anthology/2020.sltu-1.28
Chakravarthi BR, Muralidaran V, Priyadharshini R, Cn S, McCrae J, García MÁ, Jiménez-Zafra SM, Valencia-García R, Kumaresan P, Ponnusamy R, García-Baena D, García-Díaz J (2022a) Overview of the shared task on hope speech detection for equality, diversity, and inclusion. In: Proceedings of the second workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Dublin, Ireland, pp 378–388. https://doi.org/10.18653/v1/2022.ltedi-1.58, https://aclanthology.org/2022.ltedi-1.58
Chakravarthi BR, Priyadharshini R, Cn S, S S, Subramanian M, Shanmugavadivel K, Krishnamurthy P, Hande A, U Hegde S, Nayak R, Valli S (2022b) Findings of the shared task on multi-task learning in Dravidian languages. In: Proceedings of the second workshop on speech and language technologies for dravidian languages, association for computational linguistics, Dublin, Ireland, pp 286–291. https://doi.org/10.18653/v1/2022.dravidianlangtech-1.43, https://aclanthology.org/2022.dravidianlangtech-1.43
Chakravarthi BR, Priyadharshini R, Durairaj T, McCrae J, Buitelaar P, Kumaresan P, Ponnusamy R (2022c) Overview of the shared task on homophobia and transphobia detection in social media comments. In: Proceedings of the second workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Dublin, Ireland, pp 369–377. https://doi.org/10.18653/v1/2022.ltedi-1.57, https://aclanthology.org/2022.ltedi-1.57
Chang EC (1998) Hope, problem-solving ability, and coping in a college student population: some implications for theory and practice. J Clin Psychol 54(7):953–962. https://doi.org/10.1002/(SICI)1097-4679(199811)54:7<953::AID-JCLP9>3.0.CO;2-F
Chen S, Kong B (2021) cs-english@LT-EDI-EACL2021: hope speech detection based on fine-tuning AlBERT Model. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Online
Chinnappa D (2021) Multilingual hope speech detection for code-mixed and transliterated texts. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Online
Chung JE (2013) Social networking in online support groups for health: How online social networking benefits patients. J Health Commun 19(6):639–659. https://doi.org/10.1080/10810730.2012.757396
Chung YL, Kuzmenko E, Tekiroglu SS, Guerini M (2019) CONAN-COunter NArratives through nichesourcing: a multilingual dataset of responses to fight online hate speech. In: Proceedings of the 57th annual meeting of the association for computational linguistics. Association for computational linguistics, Florence, Italy, pp 2819–2829. https://doi.org/10.18653/v1/P19-1271, https://www.aclweb.org/anthology/P19-1271
Clarke I, Grieve J (2017) Dimensions of abusive language on twitter. In: Proceedings of the first workshop on abusive language online. Association for computational linguistics, Vancouver, BC, Canada, pp 1–10. https://doi.org/10.18653/v1/W17-3001, https://www.aclweb.org/anthology/W17-3001
Cover R (2013) Queer youth resilience: critiquing the discourse of hope and hopelessness in lgbt suicide representation. M/C Journal 16(5), http://www.journal.media-culture.org.au/index.php/mcjournal/article/view/702
Dave B, Bhat S, Majumder P (2021) IRNLP-DAIICT@LT-EDI-EACL2021: hope speech detection in code mixed text using TF-IDF char N-grams and MuRIL. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Online. https://aclanthology.org/2021.ltedi-1.15/
Davidson T, Bhattacharya D, Weber I (2019) Racial bias in hate speech and abusive language detection datasets. In: Proceedings of the third workshop on abusive language online, association for computational linguistics, Florence, Italy, pp 25–35. https://doi.org/10.18653/v1/W19-3504, https://www.aclweb.org/anthology/W19-3504
Dowlagar S, Mamidi R (2021) EDIOne@LT-EDI-EACL2021: pre-trained transformers with convolutional neural networks for hope speech detection. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Online. https://aclanthology.org/2021.ltedi-1.11/
Elmer T, Mepham K, Stadtfeld C (2020) Students under lockdown: comparisons of students’ social networks and mental health before and during the covid-19 crisis in Switzerland. Plos one 15(7):e0236337
Ghanghor NK, Ponnusamy R, Kumaresan PK, Priyadharshini R, Thavareesan S, Chakravarthi BR (2021) IIITK@LT-EDI-EACL2021: hope speech detection for equality, diversity, and inclusion in Tamil, Malayalam and English. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion, Online. https://aclanthology.org/2021.ltedi-1.30/
Gowda A, Balouchzahi F, Shashirekha H, Sidorov G (2022) MUCIC@LT-EDI-ACL2022: Hope speech detection using data re-sampling and 1D conv-LSTM. In: Proceedings of the second workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Dublin, Ireland, pp 161–166. https://doi.org/10.18653/v1/2022.ltedi-1.20, https://aclanthology.org/2022.ltedi-1.20
Gowen K, Deschaine M, Gruttadara D, Markey D (2012) Young adults with mental health conditions and social networking websites: seeking tools to build community. Psychiatr Rehabili J 35(3):245–250. https://doi.org/10.2975/35.3.2012.245.250
Gundapu S, Radhika M (2021) Autobots@LT-EDI-EACL2021: one world, one family: hope speech detection with BERT transformer model. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for Computational Linguistics, Kyiv, pp 143–148. https://aclanthology.org/2021.ltedi-1.21/
Hande A, Priyadharshini R, Sampath A, Thamburaj KP, Chandran P, Chakravarthi BR (2021) Hope speech detection in under-resourced kannada language. arXiv preprint arXiv:2108.04616
Herrestad H, Biong S (2010) Relational hopes: a study of the lived experience of hope in some patients hospitalized for intentional self-harm. Int J Qual Stud Health Well-being 5(1):4651. https://doi.org/10.3402/qhw.v5i1.4651 pMID: 20640026
Hossain E, Sharif O, Hoque MM (2021) NLP-CUET@LT-EDI-EACL2021: multilingual code-mixed hope speech detection using cross-lingual representation learner. In Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for Computational Linguistics, Kyiv, pp 168–174. https://aclanthology.org/2021.ltedi-1.25/
Huang B, Bai Y (2021) TEAM HUB@LT-EDI-EACL2021: hope speech detection based on pre-trained language model. In Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for Computational Linguistics, Kyiv, pp 122–127. https://aclanthology.org/2021.ltedi-1.17/
Junaida MK, Ajees AP (2021) KU_NLP@LT-EDI-EACL2021: A Multilingual Hope Speech Detection for Equality, Diversity, and Inclusion using Context Aware Embeddings. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for Computational Linguistics, Kyiv, pp 79–85. https://aclanthology.org/2021.ltedi-1.10/
Kakwani D, Kunchukuttan A, Golla S, NC G, Bhattacharyya A, Khapra MM, Kumar P (2020) IndicNLPSuite: monolingual corpora, evaluation benchmarks and pre-trained multilingual language models for Indian languages. In: Findings of EMNLP
Krishna A, Zambreno J, Krishnan S (2013) Polarity trend analysis of public sentiment on YouTube. In: Proceedings of the 19th international conference on management of data, computer society of India, Mumbai, Maharashtra, IND, COMAD ’13, p 125–128. https://dl.acm.org/doi/10.5555/2694476.2694505
Kumar A, Saumya S, Roy P (2022) SOA_NLP@LT-EDI-ACL2022: an ensemble model for hope speech detection from YouTube comments. In: Proceedings of the second workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Dublin, Ireland, pp 223–228. https://doi.org/10.18653/v1/2022.ltedi-1.31, https://aclanthology.org/2022.ltedi-1.31
Lan Z, Chen M, Goodman S, Gimpel K, Sharma P, Soricut R (2019) Albert: a lite bert for self-supervised learning of language representations. https://doi.org/10.48550/ARXIV.1909.11942, https://arxiv.org/abs/1909.11942
Lee Y, Yoon S, Jung K (2018) Comparative studies of detecting abusive language on twitter. In: Proceedings of the 2nd workshop on abusive language Online (ALW2), association for computational linguistics, Brussels, Belgium, pp 101–106. https://doi.org/10.18653/v1/W18-5113, https://www.aclweb.org/anthology/W18-5113
Lucky EAE, Sany MMH, Keya M, Khushbu SA, Noori SRH (2021) An attention on sentiment analysis of child abusive public comments towards bangla text and ml. In: 2021 12th international conference on computing communication and networking technologies (ICCCNT), pp 1–6. https://doi.org/10.1109/ICCCNT51525.2021.9580154
Mahajan K, Al-Hossami E, Shaikh S (2021) TeamUNCC@LT-EDI-EACL2021: hope speech detection using transfer learning with transformers. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Online. https://aclanthology.org/2021.ltedi-1.20/
Marrese-Taylor E, Balazs J, Matsuo Y (2017) Mining fine-grained opinions on closed captions of YouTube videos with an attention-RNN. In: Proceedings of the 8th workshop on computational approaches to subjectivity, sentiment and social media analysis, association for computational linguistics, Copenhagen, Denmark, pp 102–111. https://doi.org/10.18653/v1/W17-5213, https://www.aclweb.org/anthology/W17-5213
Mathew B, Saha P, Tharad H, Rajgaria S, Singhania P, Maity SK, Goyal P, Mukherjee A (2019) Thou shalt not hate: countering online hate speech. Proc Int AAAI Conf Web Soc Media 13(01):369–380. https://www.aaai.org/ojs/index.php/ICWSM/article/view/3237
Mehta G, Yam VW, Krief A, Hopf H, Matlin SA (2018) The chemical sciences and equality, diversity, and inclusion. Angewandte Chemie International Edition 57(45):14690–14698
Milk H (1997) The hope speech. A historical sourcebook of gay and lesbian politics, We are everywhere, pp 51–53
Muralidhar S, Nguyen L, Gatica-Perez D (2018) Words worth: verbal content and hirability impressions in YouTube video resumes. In: Proceedings of the 9th workshop on computational approaches to subjectivity, sentiment and social media analysis. Association for computational linguistics, Brussels, Belgium, pp 322–327. https://doi.org/10.18653/v1/W18-6247, https://www.aclweb.org/anthology/W18-6247
Muti A, Marchiori Manerba M, Korre K, Barrón-Cedeño A (2022) LeaningTower@LT-EDI-ACL2022: when hope and hate collide. In: Proceedings of the second workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Dublin, Ireland, pp 306–311. https://doi.org/10.18653/v1/2022.ltedi-1.46, https://aclanthology.org/2022.ltedi-1.46
Ni J, Ábrego GH, Constant N, Ma J, Hall KB, Cer D, Yang Y (2021) Sentence-t5: scalable sentence encoders from pre-trained text-to-text models. https://doi.org/10.48550/ARXIV.2108.08877, https://arxiv.org/abs/2108.08877
Nogueira dos Santos C, Melnyk I, Padhi I (2018) Fighting offensive language on social media with unsupervised text style transfer. In: Proceedings of the 56th annual meeting of the association for computational linguistics (Volume 2: Short Papers), Association for computational linguistics, Melbourne, Australia, pp 189–194. https://doi.org/10.18653/v1/P18-2031, https://www.aclweb.org/anthology/P18-2031
Ousidhoum N, Lin Z, Zhang H, Song Y, Yeung DY (2019) Multilingual and multi-aspect hate speech analysis. In: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP), Association for computational linguistics, Hong Kong, China, pp 4675–4684. https://doi.org/10.18653/v1/D19-1474, https://www.aclweb.org/anthology/D19-1474
Palakodety S, KhudaBukhsh AR, Carbonell JG (2020a) Hope speech detection: a computational analysis of the voice of peace. In: Proceedings of the 24th European conference on artificial intelligence-ECAI
Palakodety S, KhudaBukhsh AR, Carbonell JG (2020) Voice for the voiceless: active sampling to detect comments supporting the rohingyas. Proc AAAI Conf Artif Intell 34:454–462
Pérez-Escoda A, Jiménez-Narros C, Perlado-Lamo-de Espinosa M, Pedrero-Esteban LM (2020) Social networks’ engagement during the covid-19 pandemic in spain: Health media versus healthcare professionals. Int J Environ Res Public Health 17(14):5261
Prates MOR, Avelar PH, Lamb LC (2020) Assessing gender bias in machine translation: a case study with google translate. Neural Computing and Applications 32(10):6363–6381. https://doi.org/10.1007/s00521-019-04144-6
Puranik K, Hande A, Priyadharshini R, Thavareesan S, Chakravarthi BR (2021) IIITT@LT-EDI-EACL2021-hope speech detection: there is always hope in transformers. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Online. https://aclanthology.org/2021.ltedi-1.13/
Que Q (2021) Simon@LT-EDI-EACL2021: detecting hope speech with BERT. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Online https://aclanthology.org/2021.ltedi-1.26/
Raffel C, Shazeer N, Roberts A, Lee K, Narang S, Matena M, Zhou Y, Li W, Liu PJ (2019) Exploring the limits of transfer learning with a unified text-to-text transformer. CoRR abs/1910.10683, http://arxiv.org/abs/1910.10683, 1910.10683
Roberson Q, Ryan AM, Ragins BR (2017) The evolution and future of diversity at work. J Appl Psychol 102(3):483
Roberson QM (2006) Disentangling the meanings of diversity and inclusion in organizations. Group Organ Manag 31(2):212–236
Robinson L, Schulz J, Blank G, Ragnedda M, Ono H, Hogan B, Mesch GS, Cotten SR, Kretchmer SB, Hale TM, Drabowicz T, Yan P, Wellman B, Harper MG, Quan-Haase A, Dunn HS, Casilli AA, Tubaro P, Carvath R, Chen W, Wiest JB, Dodel M, Stern MJ, Ball C, Huang KT, Khilnani A (2020) Digital inequalities 2.0: legacy inequalities in the information age. First Monday 25(7), https://doi.org/10.5210/fm.v25i7.10842, https://firstmonday.org/ojs/index.php/fm/article/view/10842
Rook KS, Charles ST (2017) Close social ties and health in later life: strengths and vulnerabilities. Am Psychol 72(6):567–577
S A, Ramakrishnan A, Balaji A, D T, B SK (2021a) ssn-diBERTsity@LT-EDI-EACL2021: hope speech detection on multilingual YouTube comments via transformer based approach. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Online. https://aclanthology.org/2021.ltedi-1.12/
S T, Tasubilli RT, Sai Rahul K (2021b) Amrita@LT-EDI-EACL2021: hope speech detection on multilingual text. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Online. https://aclanthology.org/2021.ltedi-1.22/
Sampath A, Durairaj T, Chakravarthi BR, Priyadharshini R, Cn S, Shanmugavadivel K, Thavareesan S, Thangasamy S, Krishnamurthy P, Hande A, Benhur S, Ponnusamy K, Pandiyan S (2022) Findings of the shared task on emotion analysis in Tamil. In: Proceedings of the second workshop on speech and language technologies for dravidian languages. Association for computational linguistics, Dublin, Ireland, pp 279–285, https://doi.org/10.18653/v1/2022.dravidianlangtech-1.42, https://aclanthology.org/2022.dravidianlangtech-1.42
Saumya S, Mishra AK (2021) IIIT_DWD@LT-EDI-EACL2021: hope speech detection in YouTube multilingual comments. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for Computational Linguistics, Kyiv, pp 107–113. https://aclanthology.org/2021.ltedi-1.14/
Schmidt A, Wiegand M (2017) A survey on hate speech detection using natural language processing. In: Proceedings of the fifth international workshop on natural language processing for social media, association for computational linguistics, Valencia, Spain, pp 1–10, https://doi.org/10.18653/v1/W17-1101, https://www.aclweb.org/anthology/W17-1101
Sciullo AMD, Muysken P, Singh R (1986) Government and code-mixing. J Linguist 22(1):1–24. http://www.jstor.org/stable/4175815
Severyn A, Moschitti A, Uryupina O, Plank B, Filippova K (2014) Opinion mining on YouTube. In: Proceedings of the 52nd annual meeting of the association for computational linguistics (Volume 1: Long Papers), association for computational linguistics, Baltimore, Maryland, pp 1252–1261. https://doi.org/10.3115/v1/P14-1118, https://www.aclweb.org/anthology/P14-1118
Sharma M, Arora G (2021) Spartans@LT-EDI-EACL2021: inclusive speech detection using pretrained language models. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Online. https://aclanthology.org/2021.ltedi-1.28/
Shore LM, Randel AE, Chung BG, Dean MA, Holcombe Ehrhart K, Singh G (2011) Inclusion and diversity in work groups: a review and model for future research. J Manag 37(4):1262–1289
Sigurbergsson GI, Derczynski L (2020) Offensive language and hate speech detection for Danish. In: Proceedings of The 12th language resources and evaluation conference, European Language Resources Association, Marseille, France, pp 3498–3508, https://www.aclweb.org/anthology/2020.lrec-1.430
Snyder CR, Rand KL, Sigmon DR (2002) Hope theory: a member of the positive psychology family
Sun T, Gaut A, Tang S, Huang Y, ElSherief M, Zhao J, Mirza D, Belding E, Chang KW, Wang WY (2019) Mitigating gender bias in natural language processing: literature review. In: Proceedings of the 57th annual meeting of the association for computational linguistics, Florence, Italy, pp 1630–1640. https://doi.org/10.18653/v1/P19-1159, https://www.aclweb.org/anthology/P19-1159
Surana H, Chinagundi B (2022) giniUs @LT-EDI-ACL2022: Aasha: transformers based hope-EDI. In: Proceedings of the second workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Dublin, Ireland, pp 291–295. https://doi.org/10.18653/v1/2022.ltedi-1.43, https://aclanthology.org/2022.ltedi-1.43
Tatman R (2017) Gender and dialect bias in YouTube’s automatic captions. In: Proceedings of the first ACL workshop on ethics in natural language processing. association for computational linguistics, Valencia, Spain, pp 53–59. https://doi.org/10.18653/v1/W17-1606, https://www.aclweb.org/anthology/W17-1606
Tekiroğlu SS, Chung YL, Guerini M (2020) Generating counter narratives against online hate speech: Data and strategies. In: Proceedings of the 58th annual meeting of the association for computational linguistics, Online, pp 1177–1190. https://doi.org/10.18653/v1/2020.acl-main.110, https://www.aclweb.org/anthology/2020.acl-main.110
Tortoreto G, Stepanov E, Cervone A, Dubiel M, Riccardi G (2019) Affective behaviour analysis of on-line user interactions: are on-line support groups more therapeutic than twitter? In: Proceedings of the fourth social media mining for health applications (#SMM4H) workshop & shared task. Association for computational linguistics, Florence, Italy, pp 79–88. https://doi.org/10.18653/v1/W19-3211, https://www.aclweb.org/anthology/W19-3211
Upadhyay IS, E N, Wadhawan A, Mamidi R (2021) Hopeful Men@LT-EDI-EACL2021: hope speech detection using Indic transliteration and transformers. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for Computational Linguistics, Online. https://aclanthology.org/2021.ltedi-1.23/
Vanmassenhove E, Hardmeier C, Way A (2018) Getting gender right in neural machine translation. In: Proceedings of the 2018 conference on empirical methods in natural language processing, association for computational linguistics, Brussels, Belgium, pp 3003–3008. https://doi.org/10.18653/v1/D18-1334, https://www.aclweb.org/anthology/D18-1334
Vijayakumar P, S P, P A, S A, Sivanaiah R, Rajendram SM, T T M (2022) SSN_ARMM@ LT-EDI -ACL2022: Hope speech detection for equality, diversity, and inclusion using ALBERT model. In: Proceedings of the second workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Dublin, Ireland, pp 172–176. https://doi.org/10.18653/v1/2022.ltedi-1.22, https://aclanthology.org/2022.ltedi-1.22
Wang A, Singh A, Michael J, Hill F, Levy O, Bowman S (2018) Glue: a multi-task benchmark and analysis platform for natural language understanding. In: Proceedings of the 2018 EMNLP workshop BlackboxNLP: analyzing and interpreting neural networks for NLP, pp 353–355
Wang Z, Jurgens D (2018) It’s going to be okay: measuring access to support in online communities. In: Proceedings of the 2018 conference on empirical methods in natural language processing, association for computational linguistics, Brussels, Belgium, pp 33–45, https://doi.org/10.18653/v1/D18-1004, https://www.aclweb.org/anthology/D18-1004
Waseem Z, Davidson T, Warmsley D, Weber I (2017) Understanding abuse: a typology of abusive language detection subtasks. In: Proceedings of the first workshop on abusive language Online, association for computational linguistics, Vancouver, BC, Canada, pp 78–84, https://doi.org/10.18653/v1/W17-3012, https://www.aclweb.org/anthology/W17-3012
Wiegand M, Ruppenhofer J, Kleinbauer T (2019) Detection of abusive language: the problem of biased datasets. In: Proceedings of the 2019 conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), association for computational linguistics, Minneapolis, Minnesota, pp 602–608. https://doi.org/10.18653/v1/N19-1060, https://www.aclweb.org/anthology/N19-1060
Xia M, Field A, Tsvetkov Y (2020) Demoting racial bias in hate speech detection. In: Proceedings of the eighth international workshop on natural language processing for social media Association for computational linguistics, Online, pp 7–14. https://doi.org/10.18653/v1/2020.socialnlp-1.2, https://www.aclweb.org/anthology/2020.socialnlp-1.2
Xuan J, Ocone R (2022) The equality, diversity and inclusion in energy and Ai: call for actions. Energy AI 8:100152. https://doi.org/10.1016/j.egyai.2022.100152, https://www.sciencedirect.com/science/article/pii/S2666546822000131
Yates A, Cohan A, Goharian N (2017) Depression and self-harm risk assessment in online forums. In: Proceedings of the 2017 conference on empirical methods in natural language processing, association for computational linguistics, Copenhagen, Denmark, pp 2968–2978. https://doi.org/10.18653/v1/D17-1322, https://www.aclweb.org/anthology/D17-1322
Youssef CM, Luthans F (2007) Positive organizational behavior in the workplace: the impact of hope, optimism, and resilience. J Manag 33(5):774–800. https://doi.org/10.1177/0149206307305562
Zampieri M, Malmasi S, Nakov P, Rosenthal S, Farra N, Kumar R (2019a) Predicting the type and target of offensive posts in social media. In: Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, Volume 1 (Long and Short Papers), association for computational linguistics, Minneapolis, Minnesota, pp 1415–1420. https://doi.org/10.18653/v1/N19-1144, https://www.aclweb.org/anthology/N19-1144
Zampieri M, Malmasi S, Nakov P, Rosenthal S, Farra N, Kumar R (2019b) SemEval-2019 task 6: identifying and categorizing offensive language in social media (OffensEval). In: Proceedings of the 13th international workshop on semantic evaluation, association for computational linguistics, Minneapolis, Minnesota, USA, pp 75–86. https://doi.org/10.18653/v1/S19-2010, https://www.aclweb.org/anthology/S19-2010
Zhao Y, Tao X (2021) ZYJ@LT-EDI-EACL2021:XLM-RoBERTa-Based model with attention for hope speech detection. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for Computational Linguistics, Kyiv, pp 118–121. https://aclanthology.org/2021.ltedi-1.16/
Zhou S (2021) Zeus@LT-EDI-EACL2021: hope speech detection based on pre-training mode. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Online
Zhu Y (2022) LPS@LT-EDI-ACL2022:an ensemble approach about hope speech detection. In: Proceedings of the second workshop on language technology for equality, diversity and inclusion. Association for computational linguistics, Dublin, Ireland, pp 183–189. https://doi.org/10.18653/v1/2022.ltedi-1.24, https://aclanthology.org/2022.ltedi-1.24
Ziehe S, Pannach F, Krishnan A (2021) GCDH@LT-EDI-EACL2021: XLM-RoBERTa for hope speech detection in English, Malayalam, and Tamil. In: Proceedings of the first workshop on language technology for equality, diversity and inclusion. Association for Computational Linguistics, Kyiv, pp 132–135. https://aclanthology.org/2021.ltedi-1.19/
Acknowledgements
I thank Manoj Balaji for creating the word clouds, CNN model, and graphs for this work. The author Bharathi Raja Chakravarthi was supported in part by a research grant from Science Foundation Ireland (SFI) under Grant Number SFI/12/RC/2289_P2 (Insight_2) and Irish Research Council grant IRCLA/2017/129 (CARDAMOM-Comparative Deep Models of Language for Minority and Historical Languages) for his postdoctoral period at National University of Ireland, Galway.
Funding
Open Access funding provided by the IReL Consortium.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
On behalf of all authors, the corresponding author states that there is no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Chakravarthi, B.R. Hope speech detection in YouTube comments. Soc. Netw. Anal. Min. 12, 75 (2022). https://doi.org/10.1007/s13278-022-00901-z
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s13278-022-00901-z