Analysts of social media differ in their emphasis on the effects of message content versus social network structure. The balance of these factors may change substantially across time. When a major event occurs, initial independent reactions may give way to more social diffusion of interpretations of the event among different communities, including those committed to disinformation. Here, we explore these dynamics through a case study analysis of the Russian-language Twitter content emerging from Belarus before and after its presidential election of August 9, 2020. From these Russian-language tweets, we extracted a set of topics that characterize the social media data and construct networks to represent the sharing of these topics before and after the election. The case study in Belarus reveals how misinformation can be re-invigorated in discourse through the novelty of a major event. More generally, it suggests how audience networks can shift from influentials dispensing information before an event to a de-centralized sharing of information after it.
In the study of social media influence, theoretical perspectives range from behavioral economics to media exposure theory, social “influentials” and their networks, and random contagion dynamics (Aral and Walker 2012; Banerjee et al. 2013; Gleeson et al. 2014; Watts and Dodds 2007; Zajonc 2001). These approaches differ in how much they emphasize the content of messages versus their social networks or media (Acerbi 2020; Bentley et al. 2014; Bond et al. 2012; Carrignon et al. 2019; Lazer et al. 2014). Among those approaches that emphasize social influence, there are contrasting views over the importance of individual, exceptional agents. In one view, influence is heterogeneous and diffuse (Watts and Dodds 2007), such that ideas can emerge from virtually anyone in a network and be shared widely. In this view, there is little value in trying to identify key individuals who lead public information formation.
By contrast, the ‘influentials’ approach assumes that a relatively small fraction of opinion leaders exert influence on a comparatively large fraction of their peers in public opinion formation (Choi 2015; Dubois and Gaffney 2014; Van den Bulte and Joshi 2007). Influence, freely-conferred by followers, reflects the influencer’s own characteristics, such as prestige or status as highly informed, respected, or well-connected (Henrich and Broesch 2011; Watts and Dodds 2007). Since these qualities are judged by the followers, influencers rely on other agents. Represented as a ‘hub-and-spoke’ network structure, the influentials view has been presented, for example, for online terrorist networks, in which influential social media agents have a large number of followers who need not be well-connected to each other (Agarwal et al. 2017; Johnson et al. 2016; Weimann 2015).
The debate is central to the mitigation of online misinformation. On one hand, influential accounts may consistently and repeatedly amplify misinformation (Grinberg et al. 2019; EIPT 2020), such that the removal of one central ‘hub’ dramatically reduces the flow (Albert et al. 2000). On the other hand, misinformation may be diffused by large numbers of smaller-scale agents, such that the removal of smaller groups is needed to weaken the larger ones (Johnson et al. 2019). These models and strategies are not mutually exclusive (Hedström et al. 2000; Rivera et al. 2010); information networks are dynamic and may shift in time between centralized influentials versus diffusive sharing (Stopczynski et al. 2014).
Certain events likely catalyze this shift. Here, we refer to an “event” in the sense that most people would perceive—a national election or protest, for example. This pragmatic definition is subsumed by the more nuanced description of Wagner-Pacifici (2010), in that people perceive a specifiable event as having significant and durable transformational effects. In this sense, an event presents a break in the unremarkable continuity of everyday life (the uneventful world). Events become eventful as they are mediated by news reports, artistic media, and social media feeds (Wagner-Pacifici 2017).
Information about a planned event may initially be coordinated by influential “hubs” in the online network, but once that event has occurred, it may subsequently be shared in a more diffused network. We expect this as a regular pattern, as sharing—including information or gossip—is an evolved human behavior that establishes and maintains social connections (Hrdy 2009; Tomasello et al. 2005; Tomasello 2019; Dunbar 1998; Hess and Hagen 2006). In the U.S. in 2020, for example, older people shared more misinformation (regarding COVID-19) on social media than young people, and yet older people are actually less inclined than young people to believe that misinformation (Lazer et al. 2020).
Events can rejuvenate disinformation campaigns among their audiences, which risk saturation effects if the messages are unchanging. As novelty itself is attractive (O’Dwyer and Kandler 2017), it can be strategic to disseminate enduring themes and narratives in the fresh context of new events. Real-world events can provide new gloss on otherwise consistent broader themes of propaganda (Darczewska 2014; Gerber and Zavisca 2016). During and after the Ukraine crisis of 2014–2015, for example, Soviet-era symbols and narratives were shared on the Russian social media service VKontakte in ways that promoted neo-Soviet myths and nostalgia about World War II (Kozachenko 2019).
In terms of empirical patterns through time, we might expect a two-step process (Fig. 1). First, the lead-up to an event and the immediate response to it, and second, the social-sharing and discussion of the event in terms of long-running themes (Brock et al. 2014; Bentley and O’Brien 2017). Before an event, agents may be focused mainly on centralized information sources. Consequently, the pre-event attention network likely has a “hub-and-spoke” structure (Fig. 1, top left). Immediately after a consequential event is reported by such centralized sources, individuals respond independently to the news. Subsequently, as agents discuss the facts of the event with their peers, the network becomes more clustered and diffuse (Fig. 1, top right). Simultaneously, different clusters of correlated keywords/phrases/hashtags may emerge that contrast with the more centralized “broadcast” network of how the news itself was first disseminated.
Besides the diffusion of network interactions, Fig. 1 shows other patterns in the social media data we might expect to change after a significant event. Before the event, one null expectation is that if we rank words by their frequencies from most common to least common, then those frequencies will decline in a regular relationship known as “Zipf’s Law” (Fig. 1, middle left). After the event, we expect certain words (related to the event) to be emphasized beyond what is expected under Zipf’s Law, so we predict that the top few word frequencies will be unusually high, and then moving down the ranks will fall off abruptly (Fig. 1, middle right). This reflects increased copying of certain words and phrases across the diffused network, so we expect a decrease in a measure of information content known as Shannon entropy (Shannon 1949).
Another pattern to watch for on social media is the increased activity of non-human agents including bots, sockpuppets, news aggregators, or “useful idiots” without malicious intent (Broniatowski et al. 2018; Ferrara et al. 2016; Horne et al. 2019; Kumar et al. 2017; Starbird et al. 2018; Zannettou et al. 2019). We hypothesize that relevant bot activity changes after an event (Fig. 1, bottom). This would depend on the bot-designers strategy, but might generally be expected to be a burst of bot activity, with clusters of nearly identical bots surrounding certain nodes of the network (Johnson et al. 2016; Agarwal et al. 2017). Fake social media accounts, coordinated through common algorithms, often form star-like networks where certain accounts are linked by large numbers of follower accounts who are otherwise disconnected from each other (Johnson et al. 2016; Ruck e al. 2019). These “botnets” can emit synchronized bursts of identical tweets in disconnected networks, “star-like” network structures and other anomalous patterns involving specific hashtags and place names (Agarwal et al. 2017).
We explore these specific five dimensions (retweet networks, timelines, word frequencies, content, and bots) across a major event. Although all of these dimensions have been featured in studies of social media influence, they have not been studied as an ensemble across events. We thought that by examining the five dimensions, we would be able to better discern what might be impacting on the formations of the network structures and the shifts in the digital communication channels. Moreover, since each dimension is quickly and inexpensively quantifiable, we thought these measures may contribute to a toolkit for detecting when social networks are vulnerable to disinformation campaigns. To be robust, a predictive toolkit will require multiple dimensions with strong signals. By studying these dimensions together within a single context, we create a building block for future applied work.
If typical patterns in social media data can be identified, they could be used as null models for judging the magnitude effects of other events. They might also be used to contrast how artificial reactions to news, by bots and algorithms, reveal patterns distinct from human responses to events. The first step is to explore this premise on social media, with a hypothesized pivot point being a substantial event, such as a national election.
Case study: 2020 Belarusian presidential election
Here, we present some dynamics of pre- and post-event activity using a “natural experiment” among Russian-language Twitter accounts in Belarus, sampled over the 86 days from June 29 to September 23, 2020. We focus on the Belarusian presidential election of August 9, 2020, by monitoring Russian-language Twitter activity before and after this date. The team benefits from members who are native speakers of Russian, whose knowledge was crucial to the qualitative analysis—not just the literal meaning of Russian words in Twitter posts, but perspective in connection to a wider project at University of Tennessee on monitoring and measuring the effectiveness of Russian disinformation in Former Soviet Republics. The ability of the team to consider the results in terms of both content and pattern should reduce researcher bias in the analysis.
We provide brief context to the 2020 Belarusian presidential election (timeline in Supplementary Table S1). As president of independent Belarus, a former Soviet republic, since 1994, Alexander Lukashenko had solidified control over the government and public life, with restrictions on free speech and the press (IREX 2019). Although Belarusian elections between 1996 and 2020 have not been internationally recognized as legitimate (Freedom House 2020), over half of surveyed Belarusians have typically accepted and supported their president (Manaev 2016). In the lead-up to the 2020 presidential election, Belarusian authorities arrested and jailed two challenger candidates, popular blogger Sergei Tikhanovsky and banker Victor Babariko, while a third candidate, Valery Tsepkalo, fled the country. Subsequently, Tikhanovsky’s wife, Svetlana Tikhanovskaya, ran for president in his stead, gaining popularity with support from the campaigns of Babaroko and Tsepkalo. On July 30, over sixty thousand pro-Tikhanovskaya supporters gathered in a Minsk park. On August 9, Lukashenko was officially re-elected with 80.1% of the votes, over 10.1% for Tikhanovskaya. Amid reports of electoral violations (OSCE 2020; Nechepurenko and Higgins 2020), Tikhanovskaya refused to accept the results, fled the country on August 11 and hundreds of thousands of Belarusians protested. Violent clashes with riot police—омон in Russian—led to at least one protester being killed and hundreds forcibly detained (Chadwick 2020).
We monitored Russian-language Twitter content and copying networks before and after the 2020 Belarusian election. This project sought to increase methodological replication by gathering the social media sample using a commercially available technology: Salesforce’s Social Studio application (www.salesforce.com). This reduced the chance for researcher bias in selecting the sample which could also improve external validity. Using a keyword search in Social Studio, Russian-language Twitter data were collected from Belarus during an 86-day period, from June 29 to September 23, 2020 (Table 1). In total, 168,017 Twitter posts were collected. The topics and hashtags selected were based on our monitoring of Russian-language news sources and on content analysis of traditional media sources in Belarus.
Splitting the data into two sets, one before and one after the August election, we analyzed these data in terms of the patterns in Fig. 1. The native Russian-language speaker in our team directed the construction of our own “word bag” of Russian stop words. We processed the Twitter data by removing these stop words. Table 1 shows the keywords used in identifying Russian-language Twitter accounts (see also supplementary Tables S2 and S3). Despite not explicitly stated as part of our model, we also examined the conveyed sentiment of the Twitter data as a component of content by using vector lists of emotion words, or “word bags,” to determine normalized word frequencies, weighted inversely to document frequency. We computed the broad sentiment—positive, neutral, and negative—in Russian by using a neutral network classifier trained on a human-labeled dataset of 18,000 Russian-language news articles and a validation set of 6000 articles (Sakenovich and Zharmagambetov 2016).
Next, we applied unsupervised topic analysis using a Latent Dirichlet Allocation (LDA) topic model (Blei et al. 2003; Grün and Hornik 2018). The LDA topics are based on the co-occurrence of diagnostic two-word phrases across all tweets. Each tweet is considered a mixture of the topics based on frequency (or presence/absence) of the different word pairs. The number of topics is an adjustable parameter of LDA implemented in a Bayesian framework (Cao et al. 2009; Griffiths and Steyvers 2004; Murzintcev 2016). The ideal number of topics, k, is that which provides the model with the best topic prediction probability on a held-out sample of words. We ran topic models on all of our data combined, as well as the data collected using the keywords, for Tweets before versus after the August election. We used pyLDAvis (github.com/bmabey/pyLDAvis) to qualitatively explore the identified topics in each data split.
Finally, we visualized the content sharing networks before and after the election, using the text of each tweet and the timestamp of the tweet. Using a method developed previously (Horne et al. 2019), we built a network in a series of steps. First, we built a Term Frequency Inverse Document Frequency (TFIDF) matrix for each 5-day window in the dataset (Horne et al. 2019). Next, we visualized a network, where nodes represent tweets from different accounts and edges connect the tweets whose vector cosine similarity is greater than 0.85 (i.e., nearly the exact same text). The n tweets generate n completely connected networks of accounts retweeting each message. For each edge in each component, we ordered them by the timestamp by matching each copying tweet to an older tweet with highest cosine similarity. By creating directed edges from the original tweet to the copied tweet, this “pruning” process generated n trees (from what had been networks), one for each original source tweet. These trees were then aggregated into a network again, such that the nodes represent Twitter accounts and weighted, directed edges represent the normalized number of tweets copied from one account to another.
We examined the observed Twitter patterns while considering the idealized patterns illustrated in Fig. 1. Figure 2 shows the results of the LDA topic analysis for election-related tweets July–August 2020, with each topic characterized by 30 word frequencies. A principal component analysis of these topics represents, in the first two PCs, how related they are to each other.
The change in topics is complex, as each topic represents a unique list of the top 30 most salient words characterizing that topic. Nevertheless, a few broad changes are consistent with the predictions in Fig. 1. First, the majority of topics noticeably cluster together after the election (Fig. 2, right), illustrating the topics across Tweets are less distinctive and less diverse. While we used the same number of topics (parameter k) in both topic models for comparison, this heavy clustering post-election suggests fewer topics were discussed after the election than before it (a smaller k would provide a better model post-election). One of the words prominent in the tight cluster of topics after the election is “омон” (riot police) which was among the top three words of several topics with the post-election cluster (Fig. 2, right). In contrast, the topics before the election tended to be centered on different candidates, particularly “лукашенко” (Lukashenko), who is the top word in topics 1, 2, and 3 (Fig. 2, left). Overall, again the most notable change after the election is the clustering of topics around the left end of PC 2 (Fig. 2).
Figure 3 shows the top 20 words in the Tweets before versus after the election. We observe the expected increase in the frequency of the number one and two words after the election (Fig. 1), as well as change in the words occupying these top spots. Before the election, the most common word was the candidate, “бабарико” (Babariko), whereas after the election it was “омон” (riot police) and other words, including “жывебеларусь” (Belarus Live) and “нахуй” (f-word), which spiked on 11 August, just after the election. Notably, on several dates in summer 2020 (July 16, July 29 and August 11), the frequency of “омон” (“riot police”) spiked and then declined exponentially— as indicative of a response to a relevant event (Fig. 1).
Next, we observe timelines. Figure 4 shows timelines of the daily mentions of major political figures in the election-related tweets, namely “лукашенко” (Lukashenko), “Тиханóвская” (Tikhanovskaya) and ‘бабарико” (Babariko). Figure 4, left, shows the spike in mentions of these candidates after the election, followed by an exponential decline, as predicted (Figure 1). When we look at proportional mentions of the three candidates relative to one another (Fig. 4, right), we clearly see the effect of Lukashenko’s election victory. Before the election, mentions of Lukashenko decrease compared to his main opponent, Tikhanovskaya. After winning the election, Lukashenko predominates in terms of proportion of candidate mentions (Fig. 4 right), even as all mentions decline exponentially in raw counts (Fig. 4 left). As we will see in the network analysis, however, there were smaller sites after the election, that maintained support for Tikhanovskaya.
Following the examination of candidate mentions, we looked at the relative propensity to copy words through the correlations between the top words associated with each candidate. Our prediction (Fig. 1) is more copying after the event. As Fig. 5 shows, the correlations between each candidate and top words are much stronger after the election, consistent with our prediction. This likely reflects the effect of copying in the discussion of each candidate after the election, with a “signature” of distinct words associated with the candidate, in contrast to before the election when there was more reporting of news associated with each candidate. Notably, there is little overlap of word top lists between candidates either before or after the election (Fig. 5).
Similar to the pattern in candidate mentions (Fig. 4, left), we see a spike of candidate-related tweets after the election, followed by the expected exponential decline. Figure 6 shows the most productive Twitter accounts, in terms of number of tweets. As anticipated, the most productive accounts before the election were news sites, led by Tut.by, which is among Belarus’s five most popular Russian-language websites, with an estimated 3.5M users in 2020 including over 50% of Internet users in Belarus (IREX 2019). After the election, the most productive accounts included what appear to be bots (Pavelts, Artur Protska, and Patikoshka), given their high tweet rate and relatively few followers, pushing Tut.by down to rank number 10 during the post-election period (Fig. 6 right).
The changes in tweet-copying networks before versus after the election are visually striking. The election had considerable effect, along the lines predicted in Fig. 1, in how the copying networks change from hub-spoke to diffused. The networks were considerably more centralized before the election, with clear community structures. Afterward, they were de-centralized. Figures 7, 8, and 9 show these networks, which are representations of highly similar text. The networks before the election have well-defined “hubs,” such as Tut.by, which are likely the original (or close to original, due to Twitter API sampling) producers of wide-spreading tweets. The nodes facilitating the spread, however, may not be obvious in these network representations, due to the potential for coordinated cooperation among accounts (Starbird et al. 2018; Horne et al. 2019; Golovchenko et al. 2020; Wilson and Starbird 2020).
Finally, there was little change in average sentiment among the tweets, which averaged neutral in sentiment before and after the election (Supplementary Figs. S4 and S5). This may reflect limitations in our sentiment analysis. First, we may be losing signal by averaging across such a large number of tweets, ultimately compressing any extremely negative or positive sentiment in the data. Second, automatic sentiment detection in the Russian language is understudied. In this work, we used a Russian sentiment classifier that was trained on language from news articles. While this classifier may work on many tweets in our dataset, other tweets may contain language that is much more informal than the language used in news articles. Hence, we may also be misclassifying some extremely positive or negative tweets that use informal or slang language. Further investigation into these limitations is left to future work.
Qualitatively consistent with our model in Fig. 1, we observed a number of changes in the Twitter network before and after the 2020 presidential election in Belarus. The changes included the diffusion of copying networks, the spike and exponential decline of event-related terms, the over-use of most frequent words and the emergence of distinct clusters of words associated with each of the main candidates. The sentiment analysis showed negligible change after the election, which may reflect averaging a large sample. Twitter bot activity appears to have increased after the election. As a general explanation for these changes, we propose that a major event, such as the election, catalyzes a shift from a more “broadcast” pattern, where attention is focused on major news providers, to a more de-centralized process of discussing the news with peers. The initial pattern we have identified is in line with the findings from a study by Lin et al. (2014) that during major planned events capturing shared media and public attention Twitter activity tends to concentrate on elite users such as news providers with large numbers of followers. While retweeting behavior increases during media events, the content being retweeted are produced by the elites (Lin et al. 2014).
In Belarus, 80% of the population has Internet access to independent information, while government exerts control and censorship over the traditional mass media (IREX 2019; Freedom House 2020). In Belarus, TUT.BY has been the country’s most popular independent news source, providing detailed coverage of the election and the protests. As an influential source, TUT.BY helped facilitate the response that would later diffuse to other smaller sources. Surveyed afterward, multiple participants in the 2020 mass protests in Belarus stated they were motivated by news posted by TUT.BY (IREX 2019; Freedom House 2020)Footnote 1. After the election and this immediate response, the role of TUT.BY as an influential hub gave way to a de-centralized social media network. Presumably, as users shared their views about the election and protests, their discussions began to sort into distinct clusters, de-contextualizing the facts of the event in accordance with their respective worldviews (Rutten 2013).
Assuming it is not a sampling issue, bot activity was stronger after, rather than before the August 9th election. If borne out by other studies, this would indicate that disinformation is used to leverage the socially divisive effects of an event, rather than to cause the event itself (cf. Shirky 2011). This strategy can be effective whichever way the event (e.g., an election, a protest) turns out. For example, English-language twitter bots released before the U.S. 2016 presidential election by the Internet Research Agency (IRA), infamous for its bots and disinformation campaigns worldwide (Connell and Vogler 2017; Chivvis 2017), spiked in response to divisive events exacerbating racial and/or political tensions (Ruck e al. 2019). More recently, “anti-woke” content on YouTube has increased by engaging the existing individual preferences of a small but stable percentage of consumers of far-right content, rather than being caused by YouTube recommendations (Homa Hosseinmardi et al. 2021). If it follows that state-sponsored campaigns often aim to aggravate long-standing divisions, rather than try to create new ones, their general activity may be predictable even if the events themselves are unpredictable.
Rather than trying to cause events, state-sponsored disinformation can leverage de-centralized social network communications after an event by fueling existing divisions. The outcome of the August 9th presidential election in Belarus was never in doubt, as national surveys show president Lukashenko’s previous elections since 1994 were all largely accepted by the Belarusian public (Manaev 2016). Given this expectation, state authorities benefit by deploying social media bots to nudge post-discourse toward vilifying and marginalizing opposing voices, including political rivals (in this case Tikhanovsky, Babariko, and Tsepkalo). This “nudge” strategy leverages existing social media users and does not require overt “control of the information environment,” (cf. Reuter and Szakonyi 2015; Tucker et al. 2017). The state can hold more drastic measures in reserve when needed, such as blocking or shutting down independent online news organizations, social media platforms, and search engines for Belarus-based internet users (Stratcom 202).
In this process, the veracity of information may be less important than the social function of sharing it. This may underlie a strategy, observed previously, of Russian bots exploiting politically or socially divisive events (Ruck e al. 2019). Often large in number, bots might be designed as widespread, low-profile catalysts of social sharing. Although the Russian government had taken a cautious public stand on the Belarusian election, it may have directed bots to confuse the message of mass protests against Lukashenko. The Kremlin’s Internet Research Agency (IRA), for example, has programmed social media bots for targeted disinformation campaigns abroad (Chivvis 2017; Connell and Vogler 2017; Ruck e al. 2019).
While our theoretical framework (Fig. 1) invoked a single, planned event, such as an election, it could apply to micro-events within a broader change, such as social movement that advances in punctuated bursts (Horne et al. 2016). These granular events themselves may become incorporated into a larger disinformation campaign (Arif et al. 2018). For this reason, there is need for more fine-grained observation. As we observed negligible sentiment change, future studies should explore sentiments at closer scale. A genuine lack of sentiment change would be notable, as anger and fear are commonly known levers of propaganda (Berger and Milkman 2012; Wollebæk et al. 2019). It may not be easy to change baseline sentiment in a long-term campaign. Another need for granularity involves the constantly evolving use of social media platforms. While this study monitored Twitter as readily available data, future studies will need to monitor multiple social media, as most users have multiple platforms. Twitter represents a small fraction of followers and users of TUT.BY, which has (as of early 2021) over a million users or followers on YouTube, and hundreds of thousands more on Instagram, Telegram, Vkontakte, Facebook, and Odnoklassniki (IREX 2019; Stratcom 202).
It will also be fruitful to observe how users move between platforms when marginalized or blocked by authorities (Johnson et al. 2019). To counter discourse from the 2020 protests, the Belarusian government blocked access of Belarusian internet users to independent online news organizations, social media platforms, search engines, and mobile internet (Stratcom 202). In December 2020, TUT.BY was deprived of its media status by Belarussian authorities. In response, hundreds of thousands of Belarusian social media users have moved to less conspicuous social media platforms, including Telegram (Stratcom 202).
Similarly, Arif et al. (2018) examined both the retweet networks and content flowing across those networks to examine IRA operations during Black Lives Matter discourse on Twitter, finding that information operations use fictitious identities to influence and shape social divisions. These inauthentic accounts, presented as authentic, leveraged an ongoing set of current events (various Black Lives Matters protests in the U.S. at the time) to influence public opinion on the events themselves, as well as opinions on related events, like the 2016 U.S. Presidential Election (Starbird 2018).
Future research can qualitatively asses individual messages within the dynamics described here. State-sponsored disinformation campaigns often seek to deepen cultural divides and undermine trust in democratic institutions (DHS 2019). This can include reviving nationalist mythologies to demonize foreigners (Kozachenko 2019). On the topic of Crimea’s incorporation, for example, bloggers on LiveJournal leveraged deep-rooted divisions between patriots and liberals in Russian society partly through inconsistent use of word connotations relating to the same event (Kravchenko and Valiulina 2020).
Social media, which have been used for a decade around the world to inform, organize, and coordinate mass opposition to authoritarian governmental control (Reuter and Szakonyi 2015; Shirky 2011; Tucker et al. 2017), are increasingly used to misinform public audiences as counter-measures. Social media disinformation in post-communist countries involves producers, facilitators, and counter-acting online communities (Koinova 2009; Lyons 2006). The case study in Belarus indicates how the effects of influential opinion leaders, versus peer-to-peer sharing, are in dynamic balance in spreading disinformation. The novelty of a major event can lead audience networks to shift from influentials dispensing information before an event to a de-centralized sharing of information after it.
By considering the network dynamics of how information spreads (Badawy et al. 2018; Freelon et al. 2020; Vosoughi et al. 2018; Watts and Dodds 2007), we can expand upon current studies of the effects of disinformation campaigns, which tend to focus the content of propagandist messages, narratives, and themes (Paul and Matthews 2016; Lucas and Pomeranzev 2016). Overall, this case study contributes to our knowledge of the value of using null models for judging the magnitude effects for events. During prolonged events that may be ongoing, comparative models can be captured throughout the event to find moments of clear transition in dynamics, much like a phase transition in classic network traffic models (Ohira and Sawatari 1998).
In the future, use of such baselines to facilitate pattern detection could also be used to look for disruption in a network to uncover events that may be occurring but are not covered by the news, such as increased activity among pernicious actors who are trying to cloak their activities. We envisage a toolbox that allows social media activity to be used as an “Early Warning System” to help reveal what is not easily seen. This could furthermore be the start of a toolkit for assessing the reaction to an event as being artificial or human, and ultimately how we might use this to uncover invisible events by seeing the “vapor trail” they leave behind.
The data used in this paper have been deposited on Harvard Dataverse at: https://doi.org/10.7910/DVN/FOJPBT.
On December 3, 2020, Belarusian authorities deprived the TUT.BY web portal of its legal status of mass media organization.
Acerbi A (2020) Cultural evolution in the digital age. Oxford University Press, Oxford
Agarwal N, Al-khateeb S, Galeano R, Goolsby R (2017) Examining the use of Botnets and their evolution in propaganda dissemination. Def Strateg Commun 2:87–112
Albert R, Jeong H, Barabcsi AL (2000) Error and attack tolerance of complex networks. Nature 406:378–382
Aral S, Walker D (2012) Identifying influential and susceptible members of social networks. Science 337:337–341
Arif A, Stewart LG, Starbird K (2018) Acting the part: examining information operations within# BlackLivesMatter discourse. Proc ACM Hum-Comput Interact 2:1–27
Badawy A, Ferrara E, Lerman K (2018) Analyzing the digital traces of political manipulation: the 2016 Russian interference Twitter campaign. In: ASONAM ’18: proceedings of the 2018 IEEE/ACM international conference on advances in social networks analysis, pp 258–265
Banerjee A, Chandrasekhar AG, Duflo E, Jackson MO (2013) The diffusion of microfinance. Science 341:1236498
Bentley RA, O’Brien MJ (2017) The acceleration of cultural change: from ancestors to algorithms. M.I.T. Press, Cambridge
Bentley RA, O’Brien MJ, Brock WA (2014) Mapping collective behavior in the big-data era. Behav Brain Sci 37:63–119
Berger J, Milkman KL (2012) What makes online content viral? J Mark Res 49:192–205
Blei DM, Ng AY, Jordan MI (2003) Latent Dirichlet allocation. J Mach Learn Res 3:993–1022
Bond RM, Fariss CJ, Jones JJ, Kramer AD, Marlow C, Settle JE, Fowler JH (2012) A 61-million-person experiment in social influence and political mobilization. Nature 489:295–298
Brock WA, Bentley RA, O’Brien MJ, Caiado CCS (2014) Estimating a path through a map of decision making. PLoS ONE 9(11):e111022
Broniatowski DA, Jamison AM, Qi S, AlKulaib L, Chen T, Benton A, Quinn SC, Dredze M (2018) Weaponized health communication: Twitter Bots and Russian trolls amplify the vaccine debate. Am J Public Health 108(10):1378–1384
Cao J, Xia T, Li J, Zhang Y, Tang S (2009) A density-based method for adaptive LDA model selection. Neurocomputing 72(7–9):1775–1781
Carrignon S, Ruck DJ, Bentley RA (2019) Modelling rapid online cultural transmission: evaluating neutral models on Twitter data with approximate Bayesian computation. Palgrave Commun 5:1–9
Chadwick L (2020) Belarus protests: 50th day of protests marked by massive demonstration and 500 arrested. EuroNews, September 28, 2020. https://www.euronews.com/2020/09/27/belarus-protests-how-did-we-get-here
Chivvis CS (2017) Understanding Russian “hybrid warfare” and what can be done about it. Testimony before the House Armed Services Committee, 04/22/17 (CT-468). https://www.rand.org/content/dam/rand/pubs/testimonies/CT400/CT468/RAND_CT468.pdf
Choi S (2015) The two-step flow of communication in Twitter-Based public forums. Soc Sci Comput Rev 33(6):696–711
Chuang J, Ramage D, Manning C, Heer J (2012) Interpretation and trust: designing model-driven visualizations for text analysis. In: CHI ’12: Proceedings of the SIGCHI conference on human factors in computing systems, pp 443–452. https://doi.org/10.1145/2207676.2207738
Connell M, Vogler S (2017) Russia’s approach to cyber warfare. CNA analysis https://www.cna.org/cna_files/pdf/DOP-2016-U-014231-1Rev.pdf
Darczewska J (2014) The anatomy of Russia information warfare. The Crimea operation, a case study. Point of view 42. Ośrodek Studiów Wschodnich im. Marka Karpia. Centre for Eastern Studies
Dubois E, Gaffney D (2014) The multiple facets of influence: identifying political influential and opinion leaders on Twitter. Am Behav Sci 58(10):1260–1277
Dunbar R (1998) Grooming, gossip and the evolution of language. Harvard University Press, Cambridge
EIPT, Election Integrity Partnership Team (2020) Repeat offenders: voting misinformation on Twitter in the 2020 United States election. https://www.eipartnership.net/rapid-response/repeat-offenders
Ferrara E, Varol O, Davis C, Menczer F, Flammini A (2016) The rise of social bots. Commun ACM 59(7):96–104
Freedom House (2020) Belarus. https://freedomhouse.org/country/belarus/freedom-world/2020
Freelon D, Marwick A, Kreiss D (2020) False equivalencies: online activism from left to right. Science 369(6508):1197–1201
Gerber TP, Zavisca J (2016) Does Russian propaganda work? Wash Q 39(2):79–98
Gleeson JP, Cellai D, Onnela J-P, Porter MA, Reed-Tsochas F (2014) A simple generative model of collective online behaviour. Proc Natl Acad Sci 111:10411–10415
Golovchenko Y, Buntain C, Eady G, Brown MA, Tucker JA (2020) Cross-platform state propaganda: Russian trolls on Twitter and YouTube during the 2016 US presidential election. Int J Press/Politics 25(3):357–389
Griffiths TL, Steyvers M (2004) Finding scientific topics. Proc Natl Acad Sci 101:5228–5235
Grinberg N, Joseph K, Friedland L, Swire-Thompson B, Lazer D (2019) Fake news on Twitter during the 2016 U.S. presidential election. Science 363:374–378
Grün B, Hornik Kurt (2018) Topicmodels: an R package for fitting topic models
Hedström P, Sandell R, Stern C (2000) Meso-level networks and the diffusion of social movements. Am J Sociol 106:145–172
Henrich J, Broesch J (2011) On the nature of cultural transmission networks. Philos Trans R Soc B 2011(366):1139–1148
Hess NH, Hagen EH (2006) Psychological adaptations for assessing gossip veracity. Hum Nat 17:337–354
Horne BD, Adal S, Chan K (2016) Impact of message sorting on access to novel information in networks. In: IEEE/ACM international conference on advances in social networks analysis and mining (ASONAM), pp. 647–653
Horne BD, Nørregaard J, Adalı S (2019) Different spirals of sameness: a study of content sharing in mainstream and alternative media. Proc Int AAAI Conf Web Soc Media 13(1): 257–266.
Homa Hosseinmardi H, Ghasemian A, Clauset A, Mobius M, Rothschild DM, Watts DJ (2021) Examining the consumption of radical content on YouTube. Proc Natl Acad Sci 118(32):e2101967118
Hrdy S.B (2009) Mothers and others: the evolutionary origins of mutual understanding. Harvard University Press, Cambridge
IREX, International Research & Exchanges Board (2019) Media sustainability index 2019: Belarus. https://www.irex.org/sites/default/files/pdf/media-sustainability-index-europe-eurasia-2019-belarus.pdf
Johnson NF, Zheng M, Vorobyeva Y, Gabriel A, Qi H, Velasquez N, Manrique P, Johnson D, Restrepo E, Song C, Wuchty S (2016) New online ecology of adversarial aggregates: ISIS and beyond. Science 352(6292):1459–1463
Johnson NF, Leahy R, Restrepo NJ, Johnson Restrepo N, Velasquez N, Zheng M, Manrique P, Devkota P, Wuchty S (2019) Hidden resilience and adaptive dynamics of the global online hate ecology. Nature 573:261–265
Koinova M (2009) Diasporas and democratization in the post-communist world. Communist Post-Communist Stud 42:41–64
Kozachenko I (2019) Fighting for the soviet union 2.0: digital nostalgia and national belonging in the context of the Ukrainian crisis. Communist and Post-Communist Stud 52:1–10
Kravchenko E, Valiulina T (2020) Social antinomies of linguistic consciousness: Russian blogosphere debates Crimea’s incorporation. Communist Post-Communist Stud 53(3):157–171
Kumar S, Cheng J, Leskovec J, Subrahmanian VS (2017) An army of me: sockpuppets in online discussion communities. In: WWW ’17: proceedings of the 26th international conference on world wide web, pp. 857–866
Lazer D, Kennedy R, King G, Vespignani A (2014) The parable of google flu: traps in big data analysis. Science 343:1203–1205
Lazer D, Ruck DJ, Quintana A, Shugars S, Joseph K et al (2020) The state of the nation: a 50-state COVID-19 survey. Report 18: COVID-19 fake news on Twitter
Lin Y, Keegan B, Margolin D, Lazer D (2014) Rising tides or rising stars? Dynamics of shared attention on Twitter during media events. PLoS ONE 9(5):e94093
Lucas E, Pomeranzev P (2016) Winning the information war: techniques and counter-strategies to Russian propaganda in Central and Eastern Europe. CEPA Information Warfare Project
Lyons T (2006) Diasporas and homeland conflict. In: Kahler M, Walter B (eds) Territoriality and conflict in an era of globalization. Cambridge University Press, Cambridge, pp 111–130
Manaev O (2016) Public opinion polling in authoritarian state: the case of Belarus. In: Bachmann K, Gieseke J (eds) The silent majority in the authoritarian states opinion poll research in eastern and south-eastern Europe’. Peter Lang, Bern, pp 79–98
Murzintcev N (2016) Package “idatuning”. https://cran.r-project.org/web/packages/ldatuning/ldatuning.pdf
Stratcom NATO (2020) NATO strategic communications centre of excellence. Belarus protests, information control and technological censorship
Nechepurenko I, Higgins A (2020) Belarus says longtime leader is re-elected in vote critics call rigged. The New York Times. https://www.nytimes.com/2020/08/09/world/europe/belarus-election-lukashenko.html
Ohira Toru, Sawatari Ryusuke (1998) Phase transition in a computer network traffic model. Phys Rev E 58(1):193
OSCE, Organization for Security and Co-operation in Europe (2020) Elections in Belarus. https://www.osce.org/odihr/elections/belarus
O’Dwyer JP, Kandler A (2017) Inferring processes of cultural transmission: the critical role of rare variants in distinguishing neutrality from novelty biases. Philos Trans R Soc B 372:20160426
Paul C, Matthews M (2016) The Russian “Firehose of Falsehood’’ propaganda model. RAND Corporation, Santa Monica
Reuter OJ, Szakonyi D (2015) Online social media and political awareness in authoritarian regimes. Br J Political Sci 45(1):29–51
Rivera MT, Soderstrom SB, Uzzi B (2010) Dynamics of dyads in social networks: assortative, relational, and proximity mechanisms. Ann Rev Sociol 36(1):91–115
Ruck DJ, Rice NM, Borycz J, Bentley RA (2019) Internet research agency Twitter activity predicted 2016 U.S. election polls. First Monday 24(7). https://doi.org/10.5210/fm.v24i7.10107
Rutten E (2013) Why digital memory studies should not overlook eastern Europe’s memory wars. In: Blacker U, Etkind A, Fedor J (eds) Memory and theory in eastern Europe. Palgrave Macmillan, New York, pp 219–231
Sakenovich NS, Zharmagambetov AS (2016) On one approach of solving sentiment analysis task for Kazakh and Russian languages using deep learning. In: ICCCI 2016: computational collective intelligence, 9876 LNCS, pp. 537–545
Shannon CE (1949) Communication theory of secrecy systems. Bell Syst Tech J 28(4):656–715
Shirky C (2011) The political power of social media technology, the public sphere, and political change. Foreign Aff 90(1):28–41
Sievert C, Shirley K (2014) LDAvis: a method for visualizing and interpreting topics. In: Proceedings of the workshop on interactive language learning, visualization, and interfaces, pp 63-70. https://doi.org/10.3115/v1/W14-3110
Starbird K (2018) The surprising nuance behind the Russian troll strategy. https://medium.com/s/story/the-trolls-within-how-russian-information-operations-infiltrated-online-communities-691fb969b9e4
Starbird K, Arif A, Wilson T, Van Koevering K, Yefimova K, Scarnecchia D (2018) Ecosystem or echo-system? Exploring content sharing across alternative media domains. ICWSM, pp 365–374
Stopczynski A, Sekara V, Sapiezynski P, Cuttone A, Madsen MM, Larsen JE, Lehmann S (2014) Measuring large-scale social networks with high resolution. PLoS ONE 9(4):e95978
Tomasello M (2019) Becoming human: a theory of ontogeny. Belknap Press, Cambridge
Tomasello M, Carpenter M, Call J, Behne T, Moll H (2005) Understanding and sharing intentions: the origins of cultural cognition. Behav Brain Sci 28(5):675–727
Tucker JA, Theocharis Y, Roberts ME, Barberá P (2017) From liberation to turmoil: social media and democracy. J Democr 28(4):46–59
DHS, U.S. Department of Homeland Security (2019) Public-private analytic exchange program (AEP). Combatting targeted disinformation campaigns: a whole-of-society issue. https://www.dhs.gov/sites/default/files/publications/ia/ia_combatting-targeted-disinformation-campaigns.pdf
Van den Bulte C, Joshi YV (2007) New product diffusion with influentials and imitators. Market Sci 26:400–421
Vosoughi S, Roy D, Aral S (2018) The spread of true and false news online. Science 359:1146–1151
Wagner-Pacifici R (2010) Theorizing the restlessness of events. Am J Sociol 115(5):1351–1386
Wagner-Pacifici R (2017) What Is an event? University of Chicago Press, Chicago
Watts DJ, Dodds PS (2007) Influentials, networks, and public opinion formation. J Consum Res 34:441–458
Weimann G (2015) Terrorism in cyberspace: the next generation. Columbia University Press, New York
Wilson T, Starbird K (2020) Cross-platform disinformation campaigns: lessons learned and next steps. Harvard Kennedy school misinformation review, (1)1. https://misinforeview.hks.harvard.edu/article/cross-platform-disinformation-campaigns
Wollebæk D, Karlsen RH, Steen-Johnsen K, Enjolras B (2019) Anger, fear, and echo chambers: the emotional basis for online behavior. Social Media and Society, April 2019, pp 1–14
Zajonc RB (2001) Mere exposure: a gateway to the subliminal. Curr Dir Psychol Sci 10(6):224–228
Zannettou S, Sirivianos M, Blackburn J, Kourtellis N (2019) The web of false information: rumors, fake news, hoaxes, clickbait, and various other shenanigans. J Data Inf Qual 11(3):1–37
This research was supported through the Minerva Research Initiative in partnership with the Office of Naval Research. It is part of a larger University of Tennessee interdisciplinary project titled Monitoring the Content and Measuring the Effectiveness of Russian Disinformation and Propaganda Campaigns in Selected Former Soviet Union States (Grant #: N000142012618).
Conflict of Interest
The authors declare that there is no conflict of interest.
Not applicable; all data used for this project were publicly available aggregated data.
Not applicable; all data used for this project were publicly available aggregated data.
Below is the link to the electronic supplementary material.
About this article
Cite this article
Rice, N.M., Horne, B.D., Luther, C.A. et al. Monitoring event-driven dynamics on Twitter: a case study in Belarus. SN Soc Sci 2, 36 (2022). https://doi.org/10.1007/s43545-022-00330-x
- Computational social science
- Twitter analysis