I will present a dynamic, stochastic, and discrete agent-based model which rests mainly on graph theory and local optimizationa. As Epstein argues, agent-based models are an ‘analytical-computational approach to non-equilibrium social systems’ [15]. For this reason, they are likely to be the most powerful tool available for theorizing about opinion dynamics, both in political and other social contexts. And indeed, agent-based models have been relatively often used to study opinion formation [16]-[20] while their prevalence is still comparatively low in political science in general [21],[22].
Agent-based models follow a ‘generative paradigm’ [15] because they serve to model agents’ individual behavior in a bottom-up fashion and then evaluate the resulting macro structure against the benchmark of empirical observations or theoretical expectations of this macro structure. For this reason, agent-based models have been called ‘thought experiments’ in the literature [23].
This is precisely the research design employed here. Some simple behavioral rules are designed about how agents decide which claims they want to contribute to the discourse. Given each mechanism, how will the behavior of the agents affect the macro structure of the policy-domain-wide political discourse? Moreover, how do these separately plausible assumptions have to be combined to yield artificial discourses which are compatible with the theoretical properties of discourses outlined above? Social network analysis [5] is employed as a measurement model of the resulting discourse.
The model presented hereafter is modular insofar as various kinds of structural or individual effects can be easily plugged into it or omitted from it. This design principle allows to start off with a baseline model of completely random behavior, then testing of effects of any single behavioral mechanism, and later on combining these mechanisms by attaching weights to each of them.
Definitions and basic setup
In each round of the discourse, an actor makes a statement by publicly choosing and announcing a concept. Let A={a1,…,a
i
,…,a
m
} be the set of actors in the discourse, with i being the index of the evaluating actor (ego) and i′ being the index of an arbitrary alter. Every single actor is associated with a randomly assigned ideal point on a single ideological dimension (e.g., leftist versus rightist ideology). While drawing the ideological ideal points of actors from a continuous, bimodal random distribution would be more compatible with conventional political science models, the model described here rather fixes the ideal points of actors at the extreme points of the ideological dimension, that is, with Φ={0;1}, in order to yield a clean bipolar ideology effect. This dichotomous ideology is exogenously given and fixed over time. In theoretical terms, it corresponds to what Sabatier calls ‘deep core beliefs’ [9] - unalterable normative ideal points which are the source of much of the ideological conflict that is observable in political discourses. These ideology constants work like in spatial models of politics where the ideal points of agents are exogenously givenb. Actors are also exactly one out of two types: an interest group or a governmental actor, irrespective of their ideology score. The actor type is captured by a dummy variable θ(a
i
)={0;1}, which equals 1 iff a
i
is a governmental actor.
Furthermore, let C={c1,…,c
j
,…,c
n
} be the set of concepts, or claims, in the discourse. A random ideological score ϕ(c)∼U(0,1) is associated with each c, with one extreme of this dimension representing leftist values and the other end representing rightist values. This ideological dimension is the same as for the actors, but the ideological scores of the concepts are drawn from a continuous uniform distribution.
Each actor shall have a publicly visible history of the o last statements a
i
has made in the media. Let H
i
={hi,1,…,hi,t,…,hi,o} be the history of actor i with concepts at time t. By default, the history has o=5 rounds with o being the most recent round, o−1 the second-most recent round, etc. The current decision about which concept to choose is being made at o+1 and thus does not yet belong to the history.
In each round, exactly one actor makes a statement. The actor is chosen randomly with a probability of κ that a governmental actor is chosen. By default, there are five leftist governmental actors, five rightist governmental actors, 15 leftist interest groups, and 15 rightist interest groups, and κ=0.6. These proportions roughly correspond to empirical discourses in various policy domains [1],[2],[10]. By default, C comprises eight different normative concepts. This number is sufficiently large to yield meaningful results, but it is low enough to ensure a reasonable computing time.
Actors make their concept choices based on a behavioral mechanism or based on a custom blend of eight possible mechanisms. In each round, actor i computes an aggregated score for each concept, which is based on various criteria and then adopts the concept with the highest score. The details of these criteria or mechanisms and the aggregation via a utility function are described in the following paragraphs.
Exogenous ideology
Ego assigns a score to c
j
which corresponds to the similarity between his or her ideology and the ideology of c
j
:
(1)
This mechanism represents the classic view that interest groups have a rather fixed set of preferences and try to pull the outcome of policy-making into their ideological directions. Actors make normative statements in the discourse which are close to their own ideal point [24]. It should be straightforward to assume that interest groups will rather express statements that support their ideology than statements running against their own ideology.
Endogenous ideology
Ideology does not have to be exogenously given. It can be conceptualized as being path-dependent such that ego assesses whether a concept is ideologically compatible with the concepts ego has named before. Ego computes the average (or total) similarity between the concepts in his or her history and c
j
:
(2)
Concept popularity or preferential attachment to concepts
For a given actor i and a given concept j, let M be the set of alters i′ who have announced concept j in the latest round, o.
(3)
Ego assigns the number of alters in the set as the score:
In other words, ego chooses a concept that is currently popular. This may be one of the drivers of ‘political waves’ [25] or ‘issue attention cycles’ [26]; topics or claims within a debate come and go in ‘waves’ because actors jump on the bandwagon once another actor has started to use a concept in the discourse.
A politician or a party that just ignores claims brought up by others will not be re-elected, and similarly a government department will be criticized if it ignores important facts of the problem it is in charge of.
In the literature on social networks, the choice of already popular nodes is known as ‘preferential attachment’ and has been described as one of the building blocks of network formation [27].
Actor similarity or coalition formation
The actor similarity score follows from the prediction of the Advocacy Coalition Framework [9] and other interest group theories [28] that interest groups are organized in coalitions. According to the former theory, actors learn more easily from other actors if they are in the same coalition. A coalition, however, is not a formal arrangement. It is rather constituted by the similarity of ‘belief systems’ between actors [9]. Consequently, actors identify other actors who are in the same coalition by measuring their overall belief similarity and then adopt their beliefs.
More precisely, for any given concept j, an actor first considers the histories of all other actors (that is, their last five concepts) who recently chose j and computes their similarities to his or her own history list by counting the co-occurrences of concepts. Actor i chooses the concept which maximizes the fit between actor i and actors i′ who recently chose j. This method directly implements the idea of adaptation from actors who have similar belief systems and is thus a device of discursive coalition formation. Given Equation 3, ego assigns the following score to concept j:
(5)
where is the size of the intersection of the histories of a
i
and and where the whole numerator captures the total similarity of a
i
to all who announced c
j
in the last round. The function therefore assigns a value to c
j
which is equivalent to the average similarity of the history of a
i
to other actors’ histories who recently announced c
j
and thus captures the degree to which c
j
was recently chosen by similar actors (i.e., actors from the same coalition) and operationalizes coalition formation.
Concept similarity or following the wisdom of the crowds
Define G
j
as the set of alters who contain concept c
j
in their history:
(6)
Let be the similarity between two concepts as measured by the number of actors who have both concepts in their histories:
(7)
Ego sums up the similarities between all of his or her past concepts and concept j. The result is the concept similarity score of c
j
:
(8)
In other words, a co-occurrence matrix is created which contains the similarities between all concepts. Two concepts are more similar, the more often an actor refers to both of them in his or her history. For every concept in the discourse, the method then adds up the similarities between this concept and each item in the history list of the current actor. This results in the total similarity between a concept and the current actor, based on both his or her history and the evaluations of all other actors.
One rationale for this method is the tendency of subjects to deny information that does not conform to their own beliefs [29]. Concepts should be as compatible to one’s own prior statements as possible, and other actors’ use of concepts may be an important data source for evaluating how similar two concepts actually are.
At the same time, political actors strive for new arguments and concepts in support of their claims ([30], pp. 290). Actors try to position themselves in the discourse, avoid being isolated with their claims, and they closely observe which other concepts are both popular and close to their own position. If they adopt such a claim, it is more likely that other actors will refer to them and in turn adopt their other concepts.
Actor’s history or self-consistency
Ego assigns a positive score to concept j if ego named the concept in his or her history, otherwise 0. The more recent the statement, the higher the score:
(9)
This mechanism stems from social psychology. Individuals (and organizations as collective actors) try to maximize consistency with their previous statements. This results in an actor’s own latest statement scoring highest, the one before second-highest, etc. Concepts which are not in the actor’s history receive the minimum score.
This method is based on cognitive dissonance theory and its descendants [31],[32]: Aronson’s self-consistency theory implies that actors strive for consistent views of themselves in order to avoid dissonance [29]. Cialdini et al. define consistency as the ‘tendency to base one’s responses to incoming stimuli on the implications of existing (prior entry) variables, such as previous expectancies, commitments and choices’ [33]. Actors weight their own previous statements higher than other concepts because they want to maintain and communicate a coherent image of themselves, leading to path dependence of an actor’s statement choice.
Rare concepts or agenda setting
Particularly interest groups find it worthwhile to revive a sleeping discourse. If there is a concept that has not been actively discussed for quite some time, it proves to be a good opportunity to pick up this concept and re-introduce it into the discourse in order to pull the debate into a certain direction. Actors may thus have a tendency to make claims which are not on many actors’ agenda in order to act as an agenda setter. Especially those rare concepts are attractive which are close to the actor’s own ideal point. Therefore, it makes sense to combine the mechanism of choosing rare concepts with the ideology mechanism described above.
The agenda-setting score is the number of alters who have concept c
j
in their history divided by the number of alters in the discourse, subtracted from 1.
(10)
Based on Equation 6, this strategy resembles inverse concept popularity (but concerning all rounds in the history, not just the latest round).
Government coherence
As an extension of Equation 6, let Bi,j be the set of alters who hold concept c
j
in their history and who are governmental actors.
(11)
Government coherence is the number of alters who are governmental actors and who have c
j
in their history:
For every concept, the government coherence method counts by how many (other) governmental actors the concept was previously chosen, i.e., in how many (other) governmental actors’ history lists it is present. The method serves two different purposes for interest groups and governmental actors.
For governmental actors, this in an important procedure because they usually share common goals, despite potential ideological differences due to coalition governments. In most political systems, the various government departments are tied together by common objectives, parties, coalition contracts, or presidents. The aim of the method is to unify governmental actors by aligning them with each other.
For interest groups, this method defines the degree to which they adhere to the collective ideal point of the government. In consensual or corporatist political systems, the importance of this mechanism for interest groups should be high, while it should be low in majoritarian or pressure-pluralist systems [34].
Normalization
Each of these variables, or evaluation functions, yields a score that an actor assigns to a concept. Their ranges (in the statistical sense) differ substantially. However, they should be on the same scale in order to be comparable in a meaningful way. The following procedure thus converts the list of scores into a ranking list with 1 being the lowest rank and n (the number of concepts) being the highest or best rank. Two items with the same frequency (that is, ties) should be assigned the lower of the two rank positions. For example, if there are three concepts with the same score, and there are two items with a lower and three items with a higher score than the three items under consideration, the three items should all be assigned rank 3.
Let f(c
j
) be an evaluation function that assigns a score to concept j. Furthermore, let
(13)
be an n-tuple of assigned scores such that ∀l<n:d
l
≤dl+1. Then, let be a surjection with
(14)
Finally, (g∘f)(c
j
) is the normalized value of f(c
j
).
A fictitious example: Assume there are four concepts with different scores resulting from the concept popularity evaluation function, CP
i
(c1)=14, CP
i
(c2)=3, CP
i
(c3)=14, and CP
i
(c4)=19. Then, D is the ordered set {3,14,14,19}. The normalization function assigns the values {1,2,2,4} because the first item matches the first rule of the surjection, the second and fourth items the second rule, and the third item the third rule of the g function.
Utility functions
Assuming that there are p=8 evaluation functions, the utility of actor i for choosing concept j is defined as
(15)
with being an arbitrary weight. As mentioned above, there are governmental actors and interest groups. The utility function with the β weights introduced in Equation 15 is only applicable to interest groups. A second utility function for governmental actors exists where the β weights are replaced by γ weights. In all other regards, the functions are identical. This allows for a subclass of model specifications where there is in fact only a single actor type, that is, in the case where ∀k:β
k
=γ
k
. Governmental actors thus maximize the following utility function:
(16)
In either case, actors choose an optimal concept which maximizes their utility:
(17)
If there are several optimal concepts, the actor chooses from a discrete uniform distribution:
(18)
This step ensures that there is always exactly one concept per actor and round.
The discourse model is run for several thousand time steps. At each step, an actor is selected according to the probability rule described above, and this actor makes a publicly visible statement according to the outcome of the utility function, thereby updating his or her history of concepts.
Measurement
As shown in the literature on ‘discourse networks’ [1],[2],[35],[36], network analysis can be employed to study empirical aspects of political discourse like the shape, stability, and coherence of discourse or advocacy coalitions, cleavage lines in a policy domain, diversity of arguments, and the degree of polarization of a discourse. A discourse can be operationalized as follows. If A is the set of actors in a discourse and C denotes the set of concepts in the discourse, then a bipartite graph Gaff=(A,C,Eaff) with edges eaff(a,c)∈Eaff can be constructed. The aff superscript indicates that this is an affiliation network, or a bipartite graph. The bipartite graph can be converted into a one-mode projection, or ‘co-occurrence’ network, Ga=(A,Ea), where the superscript a denotes a network composed only of actors. Formally, this can be achieved by considering neighbors of actors in the bipartite graph. The set of neighbors is defined as the collection of concepts which are incident to an actor, that is, . Accordingly, edge weights between actors in Ga are computed as , which amounts to the number of concepts two actors share. The resulting network provides a cross-sectional map of the discursive landscape of political actors. Ties between actors show their discursive similarity; the absence of ties between actors or groups of actors represents discursive dissimilarity. The full array of network-analytic methods can be used to describe the discourse in a precise way, e.g., the degree of polarization between groups of actors, the number of components or clusters as instances of discourse coalitions or advocacy coalitions, or the change of the discursive network structure over time (and thus discursive equilibria).
This measurement model is used to analyze the simulation outcomes as follows. After each new statement, the collective histories of all actors are visualized as an actor × actor co-occurrence network over all concept histories of actors. Beyond visual inspection, five particular statistics are used to analyze the resulting networks: a new measure of ideological polarization, betweenness centralization, the number of components as a share of the initial number of concepts, the proportion of concepts still alive, and the concept replacement rate in the histories of all agents. These measures are discussed below.
Each of the five indices is calculated after a single new step in the discourse. Every particular configuration of the utility functions (in terms of the β and γ weights) of interest is simulated 100 times in order to guarantee that the results are reliable. Since there are 100 simulation runs, there is a random sample of 100 observations for each measure per time step. The mean values of the 100 simulations are plotted as a time series, and this is done for each of the five indices (see Figure 1). Dashed lines around the time series lines represent the 95% confidence interval. Ninety-five out of the 100 simulations lie within these boundaries. The procedure is repeated for several different configurations of the utility functions.
While the agent-based model was implemented in the programming language Java, additional capabilities from the JGraphTc and RepastJd add-on libraries are used for the analysis.
Betweenness centralization
The primary structural feature of interest is whether everybody aligns with everybody else, or whether the network tends to be composed of factions or ‘coalitions’ [9],[11]. Several studies have shown that empirical discourses tend to be composed of distinct coalitions with few bridging ties or nodes [1],[2],[35]. Betweenness centralization [37] captures this aspect by measuring the tendency of a vertex to act as a bridge between many other vertices. The notion of betweenness is operationalized by the number of shortest paths (geodesics, g
jk
) the vertex is situated on (g
jk
(n
i
)), standardized by the number of dyads not involving the vertex for which betweenness centrality is being calculated ([4], pp. 190):
(19)
Centralization (in contrast to centrality) is a network-level index which sums up the differences between the highest centrality value found in the network, denoted as C B′(n∗), and the centrality values of all other nodes, C B′(n
i
), and divides this sum by the maximum sum that is theoretically possible:
(20)
Centralization thus captures the tendency of a network to have few (in the extreme case: one) very central and many peripheral actors ([4], pp. 177). Applied to the problem at hand, betweenness centralization measures the tendency of a network to have very few vertices that interconnect distinct factions in the network. The drawbacks of this measure are that betweenness centralization is zero if the factions lose their interconnection completely and that the factions do not necessarily correspond to ideological cliques. Therefore, another measure of ideological polarization is introduced below.
Ideological polarization
Ideological polarization is a variant of assortative mixing by scalar properties [38]. It measures the degree to which vertices with a similar ideology score are connected and dissimilar actors are disconnected and thus to what extent the whole network is polarized with regard to the nodal attribute ‘ideology’. Polarized networks exhibit two ideological clusters.
To compute nodal attribute polarization, three equations are necessary. First, the sum of absolute ideological differences between non-connected (separated) actors has to be computed:
(21)
Second, calculate the sum of absolute ideological differences between connected actors, but this time, the difference has to be multiplied by the edge weight each time:
(22)
Note that the vertical bars denote absolute values in the first case and cardinality of the set in the second case. Finally, ideological polarization can be measured as
(23)
These equations define the polarization measure between 0 and 1. Values of 0.5 do not show any association, values close to 1.0 a strong positive association, and values close to 0.0 a strong negative association.
In contrast to betweenness centralization, the polarization score remains high if two distinct components are completely separated, and the measure captures only ideological polarization, not any other polarization tendencies such as endogenous coalition formation. On the other hand, betweenness centralization does not become obsolete because it is still a useful measure in situations where polarization occurs along other dimensions.
In comparison to assortative mixing by scalar properties [38], this ideological polarization measure is compatible with valued graphs as employed here, it is applicable in cases where a maximum of two coalitions or communities is possible, it scales between 0 and 1 (rather than −1 and +1), and it has a simpler formulation.
Number of components
A component is a subgraph that is not connected to the remaining network. If the discourse gets so bi- or multipolar that two or more separate components exist, there is no more common ground between different factions. This may occasionally happen in real-world policy debates, but the situation should be reversible - usually after a couple of rounds.
To obtain a standardized index between zero and one, the number of components is divided by the initial number of concepts in the discourse because this corresponds to the maximum number of components possible.
Proportion of concepts still alive
The fourth index measures the integrity of the discourse by counting how many concepts are still alive (that is, mentioned by at least one actor in his or her latest round). In a healthy and ongoing debate, at least two thirds of the initial ideas should still be present after a substantial amount of time. This integrity measure is defined as follows:
(24)
Concept attrition is high if this integrity index is low.
Number of recent concept changes
The following replacement index captures the number of recent concept changes by counting the number of actors whose latest concept differs from his or her concept in the previous round, divided by the number of actors in total:
(25)
The replacement index thus measures in how far the discourse is in motion and can serve as an early indicator of whether the discourse is in equilibrium or not. The goal is to obtain a replacement level that is neither close to 0 nor close to 1 in order to yield a steady-state but out-of-equilibrium discourse as described at the beginning of the article.