Introduction

In recent years, the expression of hostile prejudice has increased in major online spaces, in particular social media such as Twitter. What is the mechanism behind the quick spread of online hate speech in this short period? It has been suggested that social norms, or shared rules of behavior, are one of the prime factors that affect individuals’ prejudiced attitudes and behavior [2, 7], and this has subsequently been confirmed by empirical studies in various settings [9, 31, 39].

Previous studies have reported that “social referents,” who gain comparatively more attention from other individuals, are particularly influential over social norms [35]. A recent series of studies focused on the effect of Donald Trump’s discourse on increases in prejudiced expressions and behavior. Trump’s political rise and his eventual victory in the 2016 presidential election in the United States increased the number of hate crimes and online xenophobic posts in the United States [11, 30] as well as racially biased attitudes in European countries [13].

The most compelling explanation of this phenomenon is a shift in social norms caused by Trump’s behavior, toward further acceptance of prejudiced behavior. Trump made prejudiced and violent expressions toward a range of groups during the election campaign. His victory signaled that a considerable number of individuals followed his ideas, which made him a stronger social referent. Despite the multiple empirical studies devoted to how politicians can impact prejudiced behavior by shifting social norms, the suggested mechanism is still under-investigated. Most studies demonstrate only a part of the mechanism: the connection of politicians’ behavior either with social norms [8, 32] or with individuals’ behavior [19]. As an exception, using survey experiments, [4] showed how politicians’ behavior changes both social norms and their behavior. However, their results are limited to an experimental setting, and their external validity may be questionable. Further, most studies in the area have focused on Donald Trump, while it is still unknown how the mechanism functions in other contexts. In addition, given there exists such an effect, its effect size is not well discussed in previous studies.

Given these research gaps, this study aims to provide evidence of the whole suggested mechanism using real behavior trace data. More specifically, this study demonstrates that a leading politician’s behavior signaling support for hate ideas increases hate speech on Twitter by shifting social norms. To quantify the social norm shift, this study utilizes each user’s adoption thresholds for expressing hate speech, that is, a minimum number of sources required to express hate publicly. In the context of Twitter, the number of sources would be the number of hate posts published by users who are connected to the focal user on a Twitter network. The idea of the adoption threshold is based on the characteristics of prejudiced behavior, which spreads through complex contagions requiring multiple sources of contact. Prejudiced behavior generally contrasts social norms; thus, it is accompanied by social costs such as disrespect, contempt, and repudiation from others. To adopt these types of behavior, individuals may require independent affirmation or reinforcement from multiple sources [6]. Since the level of required sources would depend on the degree of social norms, the change in the required sources, that is, the adoption thresholds can be used as a direct measure to infer social norm shift.

To provide empirical evidence, this study focuses on hate speech on Twitter in Japan toward Koreans. In Japan, nativist hate groups targeting Koreans have emerged in the late 2000s [15], and they have actively utilized social media to spread hate speech [40].Footnote 1 We show that the growing increase of anti-Korean posts on Twitter was magnified by a statement by a leading politician, Yuriko Koike, which supported the claims of hate groups in Japan.

Further, the study also compares the effect size of Koike’s announcement with that of a national security threat, a North Korean missile launch. Incidents concerning national security are another well-known factor eliciting prejudiced attitudes and behavior [5, 10, 17, 27]. This study takes advantage of the timing of these two events, which occurred in the same week. This situation enables us to measure their impacts in the same framework and compare them with each other.

The remainder of this paper proceeds as follows: The next section discusses the related literature on the expression of prejudice, based on which we elaborate a model of hate speech spread and derives a hypothesis. The third section introduces the two target events, Koike’s statement and North Korea’s missile launch, and the research strategies, including design and data collection. In particular, detection of hate speech posts on Twitter is crucial for this research; therefore, the fourth section begins by describing our method of detecting hate speech in detail. After presenting the analysis results, the last section will conclude this paper with some discussion.

Theory and hypothesis

Hate speech can be defined as speech intended to promote hatred on the basis of who the targets are, for example, based on their religion, ethnicity, national origin, or other identity factors [36, 41]. It has been suggested that social norms are one of the prime factors that alter attitudes and behavior. Accordingly, providing group norms makes individuals’ prejudiced attitudes draw closer to the group norms [39], while group norm promotion and social sanctioning reduce prejudiced behavior [31].

Recent studies have also focused on individual leading politicians’ behavior as a factor for prejudiced attitudes and behavior [4, 8, 13, 19, 32]. The primary mechanism is the capacity of a leading politician to shift social norms of prejudice. This is mainly because people with high status in a group tend to have more capability to affect group norms than other members do [31, 35]. These expectations align with political communication and public opinion studies, suggesting that elites are influential in affecting how people think about politics and that minorities tend to refrain from expressing their opinion [34, 42].

One of the most well-studied examples of this mechanism is Donald Trump’s US presidential election win in 2016, analyzed in multiple survey experiments: [4] demonstrated that Trump’s perceived popularity and his victory in the election increased individuals’ willingness to engage in public xenophobic behavior and the social acceptability of such behavior. Similarly, [8] demonstrated that Trump’s aggression increased respondents’ acceptability of prejudice toward the targeted groups after the 2016 presidential election.

However, the suggested mechanism has not yet been fully tested by previous studies. Most studies have focused on change in perception level but not in actual behavior. Although [4] showed change in behavior, the result was limited to a certain experimental setting, and we do not know whether it can be generalized to a more realistic setting.

This study aims to provide evidence for the suggested mechanism of hate speech on Twitter, a real social network. In particular, this study shows that a leading politician’s behavior signaling support for hate shifts social norms so that acceptance of prejudiced behavior increases and facilitates posting hate speech on Twitter. Regarding spreading behavior on social networks, it is well known that weak ties are more influential than close ties to spread information [14]. However, this weak ties theory was suggested based on the spread of information, such as job information. By contrast, single “weak ties” are presumably insufficient to spread prejudiced behavior because it is a socially undesirable act. It has been suggested that when collective behaviors are “costly, risky, or controversial, the willingness to participate may require independent affirmation or reinforcement from multiple sources” [6, p.703]; therefore, prejudiced behavior is more likely to spread in complex contagions with affirmation from multiple sources. In other words, it is reasonable to assume that the level of needed affirmation for hate speech directly reflects individuals’ perceptions about social norms on that matter. Thus, change in the level of needed affirmation for posting hate speech can be interpreted as a shifting social norm concerning prejudice.

To quantify needed affirmation from multiple sources, this study utilizes adoption thresholds, that is, the minimum number of sources required to publish hate speech. If we assume that prominent public figures’ gestures supporting hate speech should weaken the corresponding social norms, they should lead to lower adoption thresholds.

Hypothesis::

A leading politician’s behavior that signals support for hate groups’ ideas will increase the number of hate speech by decreasing the adoption thresholds of hate speech users.

In addition to testing the existence of the effect above, this study will also substantiate the effect sizes of interest by comparing the impact of a national security threat. Such a threatening event can also be considered to trigger hate speech toward out-groups perceived as responsible for the event. For example, the terrorist attacks of September 11, 2001, in the United States, triggered prejudiced behavior toward Muslims in Western countries [1, 20], created negative attitudes toward immigrants [17], and increased hate crimes against minorities [25]. Additionally, studies confirm that terror attacks increase segregation [10, 12] and impact attitudes of people in third countries [27]. Apart from terror attacks, diverse types of threat or perceived threat can affect attitudes: an increasing number of immigrants activates threat perception and opposition to immigrants [33]; territorial threat negatively impacts political tolerance [21]; security threat predicts exclusionist political attitudes toward minority groups [5].

Given these findings, we can expect that a threatening event related to national security may profoundly increase prejudiced behavior by lowering the adoption thresholds. The corresponding effect can provide a reference with which we can evaluate the effect size of our hypothesis above. To make use of this affordance, however, we must investigate a social situation in which both types of triggering factors can be observed at almost the same time.

Research design

This study is primarily interested in the effect of political leaders’ behavior on social norms concerning racist hate speech in the online sphere. To compare this effect with the effect of a threatening event, which has been well documented in previous research, we need a case where both incidents take place at almost the same time. Japan during Summer 2017 provides such a case.

While there are numerous online spheres in which hate speech can appear, this study will focus on Twitter, for the following reasons: In Japan, Twitter is one of the most used SNS communication tools, along with Facebook, and hate groups actively utilize Twitter to spread their racist beliefs to the larger public [40].Footnote 2 Another merit of Twitter is that we can observe user relationships in several ways, such as by who follows or retweets whom, which allows us to measure adoption thresholds of racist hate speech.

Hate speech in Japan, Summer 2017

Japan, during the Summer of 2017, experienced at least two events that affected hate speech behavior of Twitter users. The first event was provoked by Yuriko Koike, the Governor of Tokyo. On 24 August, media started to report Koike’s announcement that she would not send a memorial address to the annual memorial ceremony for the victims of a massacre in 1923, targeting Korean residents right after the Great Kanto Earthquake. Her decision was unexpected and quite surprising as prior governors of Tokyo had sent memorial addresses to the ceremony every year since 1973, when the ceremony had started, and Koike herself had sent one in a previous year.Footnote 3

With regard to her decision, Koike only stated that she would refrain from sending a memorial address as she found it appropriate to pay respect to all victims of the earthquake, not only Korean victims.Footnote 4 Although Koike did not state any further reasons why she canceled the address, it is highly plausible that hate groups perceived her decision as supporting their ideas. The hate groups had claimed that the number of victims of the disaster was exaggerated or that even the tragedy was fictional. In addition, given the meeting record at the Tokyo Metropolitan Assembly, it seems clear that there were pressures from radical right-wingersFootnote 5 behind her decision. At an assembly meeting in March 2017, five months prior to Koike’s announcement, a radical right-wing assembly member, Toshiaki Koga, stated that the number of Korean residents massacred in the disaster advertised by the ceremony organizer was baseless and that there were other opinions regarding the number which were far smaller. He claimed that the governor should reconsider sending a memorial address to the ceremony in future, otherwise the governor would be complicit in “distorting” historical facts. Koike responded that it was metropolitan bureaucrats who had sent out the address in the previous year as a customary procedure and that she would make an appropriate decision about sending the address after checking it carefully.Footnote 6 Five months later, she announced that she would not send a memorial address. On September 1, the day of the memorial ceremony, Koike indeed did not send a memorial address. This time lag between the announcement and the memorial ceremony helps us differentiate the effects of the two events.

The second event was a missile launch by North Korea over Japan on August 29. Although North Korea had launched many missiles before, this one was exceptional in terms of the threat level: It was the first ballistic weapon fired over Japan without any notification, and it fell close to the sea route connecting Japan and the US.Footnote 7 The Japanese prime minister at that time, Shinzo Abe, called this incident an “unprecedented threat” to Japan and called for an emergency meeting of the U.N. Security Council.Footnote 8 This incident also had a strong impact on Japanese residents. The residents living in 12 prefectures received alert messages via diverse channels, for example, cell phone messages, loudspeakers, and television, notifying them of the incident and telling them to seek a safe place.Footnote 9 Major media also reported this incident as a serious threat to Japan. Since the missile launched in the early morning, at about 6 am, this news became the top news of the day. Given these circumstances, it is highly plausible that this incident increased negative attitudes toward Koreans in Japan.Footnote 10

Measuring Adoption Threshold in Twitter

Here, an important task is to measure our dependent variable: the adoption threshold of hate speech users, that is, the minimum number of sources (i.e., hate user alters) that facilitate one to publish a hate speech post.

To measure a user’s adoption threshold at a specific period, we need to know about the user’s network beforehand. The study utilizes retweet relationships observed prior to a target event to capture the user’s network (hereafter, the base network). In particular, retweet networks for 7 days prior to the event (hereafter, the base period) are collected. This is a directed network:

$$\begin{aligned}G_{retweet} = (V,E)\\E = \{(i,j): i \ne j, i \text { retweet } j \text { in the base period}\}\end{aligned}$$

where V is the set of users in the collected data, and E is the set of edges. We do not differentiate between multiple retweets and one retweet from the same user.

In using retweet history to capture a user’s network, the study assumes that a user retweeting another user is also likely to follow or read the other user’s posts. This assumption may seem to be restrictive since a user can retweet others’ posts via a third user. Nevertheless, we stick to this assumption because the main interest of the study is not the absolute level of the threshold but the difference in the threshold level between days with and without contributing events. Even though the measure may be biased because of over-counting, the study expects the volume of bias to be more or less similar over the overall period and to not affect the temporal comparison.

The adoption threshold is measured by counting a hate speech user’s alters who published hate speech prior to the target user in the target period. Suppose there is a user A who published a hate post in the target period. Among A’s alters, who are detected by retweet relations in the base period, two have already published hate posts in the target period. In that case, A’s adoption threshold in this period is 2. We count each hate speech user’s adoption threshold when they publish a hate post for the first time in the period; in other words, we count the adoption threshold of each user only once even if they publish multiple hate posts. For a total adoption threshold for the target period, we take an average of all hate users’ threshold (hereafter, ADT).

The above measurement can be formalized in the following way. As described earlier, the base network is generated by using retweet relationships during the 7 days before the focusing events: Let \(V_T\) be a set of hate users detected in the target period. Let \(U = (u \in \{V \cap V_T\}, \le )\) be an ordered set that contains those who appear in the base network and are detected as hate users in the target period, where ordering is based on the time when their first hate posts are published. ADT is an average of each user’s adoption threshold in the period; therefore, ADT is,

$$\begin{aligned}\frac{1}{\vert U \vert }\sum _{u\in U}\vert \{ (u,w) \in E: u\ne w, w \in C \}\vert ,\end{aligned}$$

where \(C = \{ c \in U: c\le u \}\).Footnote 11

Data Collection

As described above, we focus on two events, respectively occurring events on August 25 and 29, 2017. To cover these events, the study collected Twitter data from August 14 to September 1; among them, 7 days of retweet history, from August 14 to August 20, will be used to generate the base network.

To obtain a retweet network of racist hate speech, the study first crawled tweets from Twitter’s standard search API using 15 keywords.Footnote 12 To obtain relevant tweets to the degree possible, we crawled tweets at two days after they were published.Footnote 13

Table 1 Keywords

Table 1 shows 15 keywords used to collect all tweets associated with racist harassment against Koreans. Keyword selection is based on [40] who investigated Japanese racism toward Koreans. The 15 keywords can be classified into four categories: General, Harassment, Diplomatic issues, and Anti–left wing. Harassment refers to words which are directly associated with racist harassment. Other keywords in three categories are chosen to gather all relevant tweets to this discourse.

Two general words used to refer to Korea (Chosen) and Koreans who live in Japan (Zainichi) were selected as keywords because they are often used with derogatory nuance in the online space.Footnote 14Anti–left wing words are used to criticize people who “defend" Korea(ns) (who are basically considered to be left-wingers), and Diplomatic issues covers the main actual topics of bilateral discussion between South Korea and Japan. By using these wide-scope keywords, it is likely that most tweets related to the racist harassment during the target period will have been collected. In total, over 19 days, 2,016,982 tweets were gathered from 257,698 distinct users.

The base period

The base network was generated based on the retweet history of 7 days, from August 14-20. This period covers two major events, occurring on August 14 and 15 respectively, which led to diplomatic friction between Japan and South Korea.

August 14 is the South Korean memorial day for the comfort women, who were forced to serve as sex slaves for the Imperial Japanese Army during World War II. South Korean and Japanese governments have different viewpoints concerning comfort women, and that has led to diplomatic conflicts multiple times. In particular, Japanese extremists have protested against the views and claims of South Korea.

The second event was the anniversary of the end of World War II, 15th August, which has caused diplomatic friction between Japan and the countries it formerly colonized, such as North and South Korea and China. Friction is mainly because of some cabinet ministers’ visits to the Imperial Shrine of Yasukuni. Yasukuni Shrine enshrines the war dead, including people who were convicted as war criminals in World War II. South and North Korea and China have criticized visits to the shrine by Japanese prime ministers and cabinet members since they considered the visits to be an affirmation of the righteousness of the past war. The right-wing extremists tend to strongly support visiting Yasukuni Shrine and to protest against criticism. In 2017, although no cabinet members visited the shrine out of consideration for the diplomatic relationships with Korea and China, Prime Minister Abe sent money to the shrine, and some major politicians, such as Tomomi Inada, who had resigned the office of the Ministry of Defence a month previously, visited the shrine.

Fig. 1
figure 1

Total number of tweets by hour. The lighter color shows number of tweets each hour; The darker color shows moving averages calculated every hour using a 24-hour time window. Numbers in the figure correspond to the numbers representing major events listed in Table 2

Table 2 Number of tweets and relevant events during the investigated period. The italic numbers corresponds to the numbers appeared in Fig. 1 and 2, which indicates the dates of the events

Given a spike in the number of tweets around the events, they are likely to have stimulated communication between users with negative attitudes toward Korea (see Table 2 and Fig. 1). This is useful for creating a base network since it allows us to observe more retweet relationships among users. Additionally, the base period covers a week; thus, it captures the relationships between users who are only active on particular days of the week.

The Target Period

The target period, during which adoption thresholds are measured, is from August 21 to September 1 and covers Koike’s announcement and North Korea’s missile launch. Estimation of the effect of these events premises that Twitter users are aware of the events. It is difficult to pinpoint when these effects start and end, as we cannot measure the awareness of each Twitter user. Therefore, we focus on the period during which these issues are most salient, as it is likely that the effect will be most pronounced during this time.

Although the media began reporting Koike’s announcement on the 24th, it appears that there was a time lag between the start date of reporting and its uptake on Twitter. Based on the keyword trends presented in Fig. 2, the event appears to have become a popular topic around the 26th and 27th, at which the frequency of the keyword “Koike” is at its highest level. Although the corresponding frequency was also at a higher level on the 24th, it was only for a short period. Top daily retweets also follow this trend. Out of the top 10 most retweeted tweets each day, 7 and 9 of them were related to Koike’s announcement on the 26th and 27th, respectively. In contrast, only 1, 0 and 4 were related to it on the 24th, 25th and 28th.Footnote 15

There are two possible explanations for this time lag. First, while some media outlets began reporting the news on the 24th, it only made it to the major morning newspapers on the following day (the 25th), and it was not a top news item.Footnote 16 Second, as of the 24th, Koike’s intentions and the context behind her decision were not clear. While there were speculations, the reporting mainly focused on the fact that she declined to send the address. Her intentions became more evident during the press conference on the 25th, where she addressed her decision and the reasons for that. After this press conference, on the 26th, major media reported Koike’s decision with more detailed context and interpretation, which is likely to have sparked controversies and made the topic more visible on Twitter.Footnote 17 The topmost retweeted tweets also illustrate the development of the issue. Initially, there was only one tweet ranked among the top 10 most retweeted tweets on the 24th, which only introduced the media coverage reporting Koike’s decision to decline sending a memorial address. In contrast, the top retweeted tweets from the 26th encompassed not only Koike’s decision but also its interpretation and users’ own opinions about the issue.

Given the trend of the keyword and top retweeted tweets, as well as the context of how the issue developed, we consider the 26th as the point when Koike’s announcement became salient on Twitter. Starting from the 26th, there are three peaks in the keyword trend for “Koike.” First, it heightened on the 26th at 11 a.m. and reached the second highest peak on the 26th at 9 p.m. It shows a tendency to fall after the third peak on the 27th at 9 a.m. Thus, the study focuses on the time between the 26th at 11 a.m. and the 27th at 9 a.m. to examine the effect of Koike’s announcement.

Fig. 2
figure 2

The number of tweets including keywords associated with political events (hourly). Abe refers to Shinzo Abe, the prime minister, and Inada to Tomomi Inada, former Ministry of Defense. Koike refers to Yuriko Koike, the Governor of Tokyo. Numbers in the figure correspond to the numbers representing major events listed in Table 2

By contrast, there was almost no time lag with the missile event on the 29th. Two hours after the missile launch, 29th at 8 am, the keyword “missile” heightened, and it reaches its highest peak at noon. The trend changed to fall after the third peak at 7 p.m. at the same day. The top retweets also follow this trend. Out of the top 10 most retweeted tweets, 8 of them were related to the missile launch on the 29th, while 0 and 4 were related on the 28th and 30th, respectively (Appendix C). Thus, we focus on the 29th between 8 a.m. and 7 p.m. to test the effect of the missile launch.

Analysis

In this section, after detecting hate speech posts using a machine learning algorithm, we evaluate the effect of the two targeted events on hate speech dissemination, using the average adoption threshold.

Detecting Hate Speech using a Support Vector Machine

In this section, we detect hate speech posts using a supervised machine learning algorithm, the support vector machine (SVM).Footnote 18 SVM is a popular machine learning algorithm that has demonstrated good performance in classification tasks. In particular, SVM can handle data with lots of features, such as text data. Further, by utilizing different Kerner functions, it allows flexible fit for diverse data structures.Footnote 19

This study utilizes linear Kerner since most text classification problems are linearly separable and it shows the best performance for data with many features [18, 22]. To improve performance, term frequency-inverse document frequency (tf-idf) values are used as weights, and Grid Search is used to find the best parameters (those which produce the highest performance). Before the machine learning process, each text is segmented into words, and the words are lemmatized to improve machine learning performance.Footnote 20

For the training algorithm, 2062 tweets were randomly chosen from the datasets and manually classified into two categories: hate speech (1) or not (0).Footnote 21 Tweet posts were classified as hate posts when they expressed harassment toward Koreans, that is, attacks, threats, or insults. Among 2062 tweets, 553 were identified as hate posts (26.8%).

K-fold cross-validation (k=10) was used to examine classification performance. The algorithm predicts 88% of tweets correctly. Precision and recall values provide information regarding false detection: 79% of the cases detected as hate posts were actual hate posts (precision) and 73% of actual hate posts were predicted to be hate posts (recall).Footnote 22

According to a manual inspection, most false-positive cases were in the gray zone between hate and non-hate speech. As we described before, only cases which expressed clear hate toward Korean were labeled as hate speech in the training set. By contrast, multiple tweets in the analysis were classified as hate posts because they contained offensive words, even though they did not apparently express hate toward Koreans. This type of false-positive group should not cause serious problems for the estimation of an adoption threshold. Even though they are in the false-positive group because of the strict coding standards of this study, it is also true that these posts are fairly close to hate speech. Furthermore, it can be assumed that this type of case behaves similarly to hate speech posts.

There are also cases classifying hate posts as non-hate posts (false-negatives). The current model often fails to classify posts when they express hate but do not use offensive words. Such cases are not desirable since we would not possibly observe all hate speech users. However, we can reasonably assume that these errors appear randomly over time. Therefore, bias in our inference about temporal dynamics of adoption threshold may be ignorable, even if the inference may be less efficient.

Measuring Average Adoption Thresholds

To examine the hypotheses concerning the effect of political leaders’ behavior on adoption thresholds and evaluate its effect size, we compare the average adoption threshold (hereafter, ADT) measures at each hour during the target period. Figure 3 shows the estimated ADT with 95% confidence interval. Each ADT is calculated by using a 24-hour time window starting at the corresponding time. For example, the ADT of August 21 at 0 a.m. is the average of adoption thresholds of those Twitter users who published their first hate tweets between August 21 0 a.m. and August 22 0 a.m. The reason we adopted 24 h time window is that it is natural to assume that the mechanism investigated here, posting tweets in response to others’ postings, occurs in 24-hours-rhythms. Since such circadian rhythms can vary according to individuals in terms of when they are active and inactive, we measure ADTs in each hour so that we can capture general tendency.

Fig. 3
figure 3

The ADTs (average adoption thresholds) and the number of the users expressing hate speech. Each data point is calculated using a 24-hour time window, starting with the data point time. For example, the ADT of August 21st at 0 a.m. is the average of adoption thresholds of haters appeared from August 21 at 0 a.m. to August 22 0 a.m. In the ADT plot, dots and vertical lines indicate the point estimates of ADT and their 95% confidence intervals. The data points subject to the effect of events are colored in red (Koike) and blue (Missile)

As discussed above, we focus on the period between the 26th at 11 a.m. and 27th at 9 a.m. as well as that between the 29th at 8 a.m. and at 7 p.m. The former is subject to the effect of Koike’s announcement and the latter to the effect of the missile event. However, we cannot directly translate these periods to the one in the ADTs, since the 24 h windows are used to measure ADTs. For example, the time window to measure the ADT at 9 a.m. on the 26th covers 22 h of the period subject to the Koike’s effect since the latter period begins at 11 a.m. of the same day. Thus, we consider the ADTs based on the period overlapped for more than 12 h with the event period as those which are subject to the corresponding event. In Fig. 3, those ADTs colored red are considered to be affected by the Koike’s event and those colored blue by the missile event. The other ADTs are considered to be a control period.

According to Fig. 3, both targeted events have lowered the adoption threshold level, which aligns with the hypothesis. Concerning the effect of Koike’s announcement, ADTs under its effect are between 3.5 and 4.0, while those in the previous period are mostly between 3.8 and 4.4. A closer look at the period under Koike’s effect reveals that ADTs generally fall on the 26th and remain at the lower level during the 27th and that they show less variance. By contrast, the other periods rather show the users’ 24-hour rhythms.

The ADTs under the missile event also seem to be lower than other control periods, but their pattern is different from those under the Koike’s event. While under the Koike’s effect it took almost a day to reach the lowest ADT level (26th 10 p.m.), the missile event hits the lowest level faster, in about half a day. This possibly reflects the speed of the news spreading to the public through the different media channels. While the missile news spread to almost all Japanese citizens in the couple of hours after the missile launched, Koike’s announcement spread to only some citizens, and it took some time to spread.Footnote 23

Additionally, the ADTs show more fluctuation under the missile event than Koike’s event. While ADTs under the missile event increased after hitting the lowest value, ADTs under Koike’s event remained in the lower range for a while after hitting the lowest value. The speed of the news spreading to Twitter users is a possible reason for this pattern. Since Koike’s announcement spread more slowly compared to the missile event, some users reacted later than others, which possibly contributed to this pattern. Another possible explanation is the distinct nature of these events. Unlike the missile launch, Koike’s announcement is more directly related to hate, as it could be interpreted as supporting the claims of hate groups. The effect of such a different nature of events should be carefully examined in future research.

Furthermore, we can observe a general tendency in that the number of hate users increases as the ADT decreases. The number of haters were heightened at both Koike and the missile event period.Footnote 24

Regression analysis

Having visually inspected the temporal development of the ADTs in Fig. 3, this section provides additional statistical evidence for the effect of leading politicians on hate speech behavior.

The dependent variable is ADTs measured each hour. The main independent variables are whether the corresponding ADT is subject to the targeted events or not. We included two event dummy variables, correspondingly: 1 indicates a corresponding ADT which is subject to the respective event, otherwise 0. Since each ADT was measured using a 24-h time window and was calculated each hour, there are 23 h time frame overlap between two consecutive ADTs, which causes auto-correlation. To remove this auto-correlation, a lagged dependent variable, ADTs at \(t-1\), is included. Additionally, dummy variables for each hour are included to control the circadian trend.

Table 3 The effects of two targeted events on the adoption threshold. Bold values indicate significance at the 5% level. The estimation result of the 23 h dummies are omitted. The complete results are available in Appendix B.1

According to Table 3, Koike’s announcement had a significantly negative effect on ADTs at 5% level, which corresponds to the hypothesis as well as the above visual inspection. One may wonder about the small effect size of \(-\)0.08. Here, we have to keep in mind that the effect size refers to only differences after one hour, whereas Koike’s announcement had an effect over almost two whole days, that is, 46 h. The accumulated effect for this period based on the regression results is \(-\)0.34, which results in increases of 2,029 hate posts and 522 hate users. This accounts for 5% of total hate posts and 4% of total hate users that appeared in the corresponding period.Footnote 25

The missile event also shows a significantly negative effect at the 5% level, while the effect size is almost half of that of the Koike’s event. The accumulated effect during the whole missile event is \(-\)0.25, which leads to the increase of 1,279 hate posts and 329 haters. This accounts for 3% of total hate posts and total hate users appeared in the missile event period. This smaller effect size might suggest that the Koike’s event had a stronger impact on lowering hate users ADTs than the missile event. However, the difference in the effect sizes is not significant (see Appendix B.2).

One may wonder whether Koike’s effect lasts even after August 27; if so, it will also affect our estimate of the missile event’s effect. While this possibility cannot be excluded, the extent of the Koike’s effect here should be small since Koike’s announcement did not appear in major Twitter topic on the 29th,Footnote 26 and the level of ADTs came back at a similar level to the pre-Koike period, before the missile event took place. Even if there was a small amount of effect of the Koike’s announcement that lasted, it should have resulted in a larger effect of the missile event. In other words, the result in Table 3 can be interpreted as a conservative result for comparison of two effect sizes.

Discussion

In this study, we hypothesized that a leading politician’s behavior signaling support for racist harassers will increase hate speech by shifting social norms. To examine the hypothesis and to quantify the social norm shift, the study utilized an adoption threshold measure using a Twitter retweet network. Our empirical study focused on the effect on hate speech spread of an announcement by Yuriko Koike, Governor of Tokyo. The results support the hypothesis: after her announcement concerning the memorial ceremony for Korean massacre victims, ADTs of hate speech users were significantly lowered. This led to increases of hate posts and hate users; Koike’s announcement was responsible for about 5% (2,029) of hate posts and 4% (522) of hate users appeared during the period.

Further, this study compared the effect of Koike’s announcement to that of North Korea’s missile launch, a national security threat. Among habitual users, who appeared both in the base and the target period, the average level of ADTs and the number of hate users were comparable in size. Meanwhile, the estimated effect of Koike’s announcement in the regression analysis was larger than that of the missile launch. This is a surprising result considering the characteristics of the two events. There were large differences in the number of people who accessed the news; while the missile event was headline news on the day, Koike’s announcement was not. This suggests that powerful people can have considerable capacity to guide other people’s behavior especially when social norms matter. Additionally, it is worth emphasizing that Koike’s announcement was not a direct attack on Koreans—rather, it signaled support for the radical right-wing discourse. It is reasonable to conjecture that politicians’ behavior which more directly attacks minorities could introduce wider hate speech spread than that demonstrated in this study.

However, there are also several limitations that should be acknowledged. Our threshold measurement, defined as the minimum number of sources required to motivate an individual to publish a hate speech post, operates under the assumption that users observe a certain number of hate posts from their alters before taking action themselves. This assumption implies a dependent relationship between prior hate speech posts from alters and subsequent posts by the observer. However, it is conceivable that some users do not adhere to this assumption. For example, some users may post hate speech independently, without seeing prior posts within their network. Consequently, our threshold measurement may incorporate some level of error due to these potential independent actions. Future studies addressing this issue will be necessary to more accurately measure adoption thresholds.

This study is designed to gauge aggregate-level shifts in social norms triggered by external events. With our primary focus on changes in social norms, we have assumed that the influence exerted by social networks remains constant throughout the period being studied, thereby not delving into the effects of networks. Nonetheless, a thorough investigation into how individual’s network influences their perception of social norms and their subsequent behaviors could significantly enhance our understanding of prejudiced behavior on social media. As important area for such study would involve analyzing the proportion of hate alters within an individual’s network and investigating the potential disproportionate influence these hate alters may exert.

This study posits that users’ psychological inclination to post hate speech remained constant during the investigated period. However, one may wonder whether the first event had influenced users’ psychological inclination, thereby affecting the estimated results in the second event. The most likely scenario is that experiencing the first event made users more comfortable with posting hate speech due to desensitization to hate speech. Desensitization is a psychological process that refers to a decrease in emotional or psychological sensitivity to a particular stimulus over time. Previous studies suggested that frequent exposure to hate speech leads to a lack of sensitivity to the usage of hate speech in everyday conversation and in online behavior through a process of desensitization [3, 38]. Although our core findings would remain unchanged even if such an effect existed, as it should have resulted even in a larger effect of the second event (i.e., the missile launch), it is worth considering in future research.

One may wonder whether the results of this study are specific to the Japanese case. One possible argument is that the Japanese group-centered culture is responsible for the results above. Such arguments seem to be unrealistic, for the following reasons: First, the absolute level of the adoption threshold is not at stake in the study. Japanese Twitter users may have a higher absolute adoption threshold because of their culture in comparison with those in other countries; however, the results above are about the relative change in the threshold level caused by certain events, which is less likely to be affected by such cultural factors. Second, our theoretical consideration is based on more general psychological mechanisms rather than culturally specific ones. There is no reason that these psychological mechanisms should operate only in Japan. Third, similar phenomena are also empirically observed around the world. As introduced before, it has been reported that hate speech, racist discourse, and hate crime strongly increased in the United States after Donald Trump took over the presidency.

Meanwhile, varied possible events in different contexts do not have to show the same size of effects on spread of hate speech behavior. The effects will be differently conditioned by their context. To obtain more general effects, further studies examining other, diverse cases will be needed.