1 Introduction

For many years, researchers have resorted to various latent variable models to measure political participation (e.g. Marsh, 1974; van Deth, 1986; Teorell, Torcal and Montero 2007; García-Albacete, 2014; Talò & Mannarini, 2015; Oser, 2017; Theocharis & van Deth, 2018). Measurement models of this kind allow to link observable indicators with the latent construct of political participation, account for the measurement error, and to check if the chosen model and, consequently, the chosen conceptualization have any resemblance in the data (cf. Fariss et al., 2020).

If we entertain the idea of political participation as a latent variable, one of the key decisions to make is about what conceptualization and measurement model to choose. In the literature, the dominant view is that the latent construct of political participation should be treated as a continuous variable (García-Albacete, 2014; Marsh, 1974, 1977; Milbrath, 1965; Quaranta, 2013; van Deth, 1986; Verba et al., 1978; Vráblíková, 2014). According to this view, we would assume that people have various propensities to participate, and depending on these propensities, they choose different forms of participation. Measurement models of political participation as a continuous latent variable vary in complexity from simple additive indices to complex factor analytic models, such as the bifactor model (Koc, 2021). In an alternative conceptualization, participation is not conceived of as some kind of latent continuum. Researchers favouring this approach try to identify relatively homogenous groups of people, where the criterion of homogeneity is based on participation repertoires (Oser, Hooghe & Marien, 2013; Jeroense & Spierings, 2022; Johann et al., 2020; Oser, 2017, 2021; Quaranta, 2018), e. g., people having a high probability of participating in all forms, those who choose only electoral participation or people refraining from any participation (cf. Oser, 2017). To define these homogenous groups, scholars use Latent Class Analysis (LCA) and treat political participation as a categorical latent variable.

However, both of these approaches are based on strong assumptions and have their limitations. In the ‘participation as a continuous-variable’ case, all individuals are assumed to come from the same homogenous population, and any differences between individuals to arise only from different propensities on the latent continuum. Hence, there are no subpopulations with distinct participation repertoires. With ‘participation as a categorical variable’, we ignore the within-group quantitative differences and we assume that people in each group participate to the same extent.

In this article, I go beyond the dichotomy of continuous vs. categorical and investigate the possibility of political participation exhibiting simultaneously features of both of these conceptualizations. Specifically, I propose to divide people into two latent classes: those who refrain from participating at all (politically disengaged) and those who participate to a varying degree. I argue that we should distinguish these two latent classes based on the frequently used data—European Social Survey—that show that most people do not participate in any non-electoral form of political participation, and substantive considerations, i.e. the theory of political apathy and alienation. To test the proposal, I utilize zero-inflated models which assume that there are two different distributions, or processes, contributing to the all-zero answers (i.e., non-participating in any form): one where subjects always produce zeros (“structural zeros” or “excessive zeros”), and the other, where subjects reported all-zero answers during the period of interest or at the time of assessment ("sampling zeros”) (Feng, 2021). That is, a person may report not participating in any non-electoral form because they belong to the latent class of politically disengaged and does not participate at all or because the person did not participate in the particular period that the interviewer asked about.

The analytical procedure consists of two steps. First, I estimate the proportions of the two latent classes in the population and investigate the heterogeneity within the class of those who participate using a Zero-Inflated Item Response Theory (ZI-IRT) model. I compare the fit of this model with a regular IRT model. Second, I estimate two regression models—a Beta-Binomial and Zero-Inflated Beta-Binomial—which closely mirror the data generating process of the two measurement models, with five frequently used antecedents of political participation: gender, age, education, political interest, and political efficacy. This allows me to check whether accounting for the zero-inflation matters for explanatory models. Additionally, with the Zero-Inflated Beta-Binomial, I will be able to find out if those antecedents of participation have the same effect on the participation intensity and the probability of belonging to the latent class of politically disengaged.

The results show that the ZI-IRT model fits the data better than the model not accounting for the zero-inflation. When taking the zero-inflation into account, parameters of the measurement model change: item difficulty and discrimination coefficients shrink, and the mass of the distribution of latent scores moves to the right (people seem to be more politically active). Moreover, accounting for the zero-inflation matters when we regress political participation on other variables: the fit of the Zero-Inflated Beta-Binomial model is substantially better and the regression coefficients of the model are different. Also, the zero-inflated part reveals that out of the five predictors, political interest and education have an impact on the probability of belonging to the latent class of politically disengaged, while political efficacy—a strong predictor of participation intensity—has no effect on the probability of belonging to the latent class of disengaged. Overall, this study provides suggestive evidence for the superiority of the approach distinguishing two latent classes while modeling political participation—the disengaged and the participating to a varying degree—and that we can and ought to model two separate processes—the intensity of participation and the probability of belonging to the latent class of politically disengaged.

2 Existing Perspectives on Political Participation as a Latent Variable

Following van Deth (2014), for a phenomenon to be counted as political participation, it needs to be a voluntary behavior conducted by individuals acting as citizens and not as professionals, that fulfills at least one of the following conditions: (1) it is located within the realm of politics, government, or the state; (2) it is targeted at a subject from one of these realms or aimed at solving a community problem; (3) it is placed in a political context (e.g., dressing up as money-eating zombies and walking outside the New York stock exchange during the height of the global financial crisis; see van Deth and Theocharis (2018)) or it is used to express political aims and intentions (van Deth, 2014; van Deth & Theocharis, 2018). This approach to conceptualizing political participation allows us, on the one hand, to capture old and new forms of participation and, on the other hand, to avoid falling into the trap of building a “concept of everything” (van Deth, 2001; van Deth & Theocharis, 2018).

2.1 Political Participation as a Continuous Variable

Thinking about political participation as a continuum has a long and well-established tradition in political behavior research. Probably the most famous early examples of this approach can be found in the works of Milbrath (1965) and Marsh (1974). According to Milbrath (1965), political participation is cumulative and forms a ranking of behavior types that is based on resources needed to enact them (cf. Brady et al., 1995). Different forms of participation require different amounts of time and energy, with the least resources needed for easier activities such as voting, and significantly more resources required for more demanding acts such as taking part in demonstrations. In Milbrath’s (1965) work, the easiest acts are labeled as “spectator activities”, and the most difficult ones— “gladiator” activities. The fact that political participation, according to Milbrath (1965), is cumulative means that people participating in more difficult forms would very likely participate in easier acts, with the opposite not being true. While Milbrath considered more “conventional” forms of participation, Marsh (1974) presents similar reasoning while focusing on the citizens’ protest potential and finds that it is positively related to their socioeconomic status. The idea that political participation can be understood cumulatively and that participation acts can be ordered according to the costs they impose on the individual has been discussed later also by other authors (e.g., García-Albacete, 2014; van Deth, 1986).

To translate the conceptualization of political participation as a continuum into a measurement model, researchers initially used either Guttman scaling (e.g., Marsh, 1974) or Mokken Scale Analysis (e.g., García-Albacete, 2014; van Deth, 1986). Technical details of the two techniques are not important for the discussion in this article.Footnote 1 What is important though is that in the case of both models (1) every subject is assumed to have a certain, unknown value on the latent variable of political participation—which we can call participation potential, participation intensity, or a propensity to participate—and (2) that for each participation form, the probability of participating in that form increases with that latent variable (van Deth, 1986). Furthermore, forms of political participation, or acts, are thought to vary in terms of their difficulty,Footnote 2 which corresponds to the resources needed to perform those acts. It should be mentioned that these assumptions are made also when political participation is measured with more complicated cumulative latent variable models like in Quaranta (2013), Vráblíková (2014), or Koc (2021).

2.2 Political Participation as a Categorical Variable

In the approach where participation is treated as a categorical variable, we would not think of participation to be cumulative and of participation acts to form some kind of ranking. There would be no latent propensity to participate that explains performing acts of various difficulty. Instead, we would assume a person-centered approach, where the key question would be about how individual actors engage in various political acts and simultaneously refrain from political participation (Oser, 2017). The level of analysis would be shifted from political acts to political actors (Oser, Hooghe & Marien, 2013; Oser, 2017). In this approach, we would try to classify people into homogenous groups, depending on how these people choose to participate, in what forms (Oser, 2021). The set of forms people choose to participate in would be called repertoires (Oser, 2017).

For that classification into groups, researchers would use Latent Class Analysis. LCA provides us with two kinds of information that is useful if the purpose of the analysis is classification: conditional item probabilities and class probabilities (Collins & Lanza, 2010). First, the conditional item probabilities are class-specific and provide information about the probability that an individual in that class will provide a positive answer to a specific item (Clark et al., 2013). In the research on political participation, the conditional item probabilities indicate the likelihood that respondents in each participant type (a homogenous group of individuals distinguished based on the forms they use) engage in each of the political acts (Oser, 2021). Second, the class probabilities specify the relative size of each class, or the proportion of the population that is in a particular class (Clark et al., 2013; Collins & Lanza, 2010). The optimal number of latent classes would be chosen using information criteria, model comparison, and the inspection of conditional item probabilities and class probabilities (Collins & Lanza, 2010).

2.3 Limitations

From the measurement perspective, both approaches have their limitations, which were briefly mentioned in the introduction. First, with cumulative latent variable models, such as Guttman, Confirmatory Factor Analysis (CFA), and various parametric and non-parametric Item Response Theory (IRT) models, we assume that all individuals come from the same homogenous population and the only differences between individuals are due to different levels on the latent trait(s). In the case of political participation, this means that all citizens participate allegedly in the same way, which is highly implausible. We know, for example, that young people participate differently than adults and that these differences are not only about the levels of participation but also about the forms-young people choose different forms than adults (García-Albacete, 2014; Quintelier, 2007). In instances, like the one with age groups, where the source of heterogeneity is observable, standard extensions to CFA and IRT—Multiple Group models—can be used. But once the source of heterogeneity is unobservable these standard approaches fail and cannot account for the violation of the population homogeneity assumption (Lubke & Muthén, 2005). A way to circumvent the problem is to resort to LCA, which is meant to model unobserved sources of population heterogeneity (Collins & Lanza, 2010). But with LCA we assume that there are no quantitative differences within the classes, which does not seem to be convincing. If we take the classification provided by Oser (2017), who distinguishes four types of participants—“Disengaged”, “All-around activists”, “High-voting engaged”, and “Mainstream participants”—and focus on “Mainstream participants”, it is very likely the among people with that participant type we have individuals who are more and less active. This can be said also about the other types, except for those who are disengaged. Hence, with LCA we ignore the fact that people of a given participant type may participate to a different extent.

3 Analytical Strategy

3.1 Motivation for Zero-Inflated Models

One way to overcome the problems presented by the existing approaches is to resort to zero-inflated models. Using this kind of models can be motivated by data and substantive considerations. When it comes to the data perspective, we should look at the proportion of people who report non-participation in forms other than voting. For example, for countries participating in rounds 8 and 9 of the European Social Survey (ESS 2016; ESS 2018), on average (for all countries), approx. 50% of respondents reported that they had not participated in any of the non-electoral forms listed in the questionnaire, with the maximum of 87% in Hungary and the minimum of 18% in Sweden. Figure 1 shows the proportions for all countries participating in both rounds. Concentrating on non-electoral forms of participation makes sense since researchers often do not include voting in their composite measures of participation (e.g., Hooghe & Marien, 2013; Li, 2021; Marien et al., 2010; Oser & Hooghe, 2018; Theocharis & van Deth, 2018; Vráblíková, 2014) because voting is a very frequent activity and therefore an exception from other indicators (Parry, Moyser and Day 1992).

Fig. 1
figure 1

Proportions of all-zero answers for items measuring political participation across countries for rounds 8 and 9 of ESS. Average proportions across the two rounds are depicted above the bars for each country. Note Calculations are based on the respondents with complete information on political participation items. AT-Austria, BE-Belgium, CH-Switzerland, CZ-Czechia, DE-Germany, EE-Estonia, ES-Spain, FI-Finland, FR-France, GB-United Kingdom, HU-Hungary, IE-Ireland, IT-Italy, LT-Lithuania, NL-Netherlands, NO-Norway, PL-Poland, PT-Portugal, SE-Sweden, SI-Slovenia

From the substantive point of view, we might want to use zero-inflated models to account for the existence of a subpopulation of people who do not participate at all, described by Amnå and Ekman (2014) as "unengaged" and "disillusioned" citizens. In the literature on political participation, we could find at least two explanations for why people would refrain from participating altogether—political alienation and political apathy (Aberbach, 1969; Abramson & Aldrich, 1982; Dahl et al., 2018; Farthing, 2010; Finifter, 1970; Fox, 2015). Being alienated and apathetic could stop a person from enacting any form of political participation (Dahl et al., 2018).

These data and substantive considerations should result in modeling decisions. When it comes to measurement models, we should relax the assumption of population homogeneity and take into consideration the possibility that there are two subpopulations: those who do not participate at all (politically disengaged), and those who participate to various degrees. Furthermore, in the subpopulation of politically active, we should allow for quantitative differences among individuals. By doing so, we can overcome the limitations of the classical approaches to modeling political participation as a latent variable.

Given that we often want to know not only what the structure of political participation is, but also how it is related to other variables, the explanatory model should take into account the existence of the two subpopulations as well. As a result, we could include variables that would predict whether an observation is likely to end up in the latent class of politically disengaged (people whose count of political activities is always 0) and variables that would predict the intensity of participation.

Now, I will describe measurement and explanatory models that are well-suited for accounting for the excess of zeros in the data.

3.2 Zero-Inflated IRT

To account for the excess of zeros, methodologists came up with the idea of Zero-Inflated IRT models (Finkelman et al., 2011; Magnus & Liu, 2018; Wall et al., 2015). To better understand their mechanics, I will introduce an equation for the Zero-Inflated Graded Response Model, an IRT model which is meant to accommodate the zero-inflation for ordinal indicators, including binary ones. For the parametrization of the model proposed by Magnus and Garnier-Villarreal (2022), the probability function for the Zero-Inflated Graded Response Model could be written as:

$$p\left( {y_{ij} |\alpha , \theta_{i} ,\varvec{ \kappa },\lambda } \right) = \left\{ {\begin{array}{*{20}l} { \lambda + \left( {1 - \lambda } \right) \times {\text{OrderedLogit}}\left( {0|\alpha \theta_{i} , {\varvec{\kappa}}} \right)\quad {\text{if }}\sum y_{i} = { }0{ }} \hfill \\ {\left( {1 - \lambda } \right) \times {\text{OrderedLogit}}\left( {y_{ij} |\alpha \theta_{i} ,\varvec{ \kappa }} \right)\quad {\text{if }}\sum y_{i} > { }0} \hfill \\ \end{array} } \right.$$
(1)

where \({y}_{ij}\) denotes a response to an item j by a person i; \(\alpha\) is a slope for each indicator; \({\theta }_{i}\) denotes an individual’s score on the latent variable of political participation; \({\varvec{\kappa}}\) is a vector of thresholds’ parameter; \(\lambda\) is the proportion of excessive all-zeros, i.e. individuals belonging to the class of politically disengaged. Item thresholds are often understood in terms of difficulty, i.e. what value on the latent variable one needs to have to have a 50% probability of responding above the given threshold. In the binary case, it simplifies to have a 50% probability for endorsing an item (for example, where one needs to fall on the participation intensity continuum to have 50% probability of endorsing the item ‘taking part in demonstrations’). Slopes represent item discrimination that provides information on how well the item can differentiate among individuals located at different points on the latent continuum (Ayala, 2009) (for example, how well ‘taking part in demonstrations’ discriminates between people having rather high level on the participation intensity continuum).

From Eq. 1 we can read off an important feature of the Zero-Inflated IRT model: The all-zero answers are due to two processes. That is, some all-zeros are contributed by the measurement model, \(\mathrm{OrderedLogit}\left(0|\alpha {\theta }_{i},{\varvec{\kappa}}\right)\) in the equation above, and some by a Bernoulli process, with mixing probability \(\lambda\). In other words, some of the all-zero answers originate from the latent class of people participating to a varying degree (the measurement model part), and the remaining ones come from those belonging to the latent class of politically disengaged, that is individuals who always respond with 0 to each item (cf. Magnus & Garnier-Villarreal, 2022). By contrast, the non-all-zero answers are only due to the measurement model.

Given the observations we made about the data generating process, it should be clear that estimating a Zero-Inflated IRT model is not the same as estimating a regular IRT model after discarding all-zero answers. The latter approach would very likely lead to biased estimates since “throwing out the all-zeros is the same as selecting who to fit the IRT model to conditional on their score” (Wall et al., 2015, p. 595). Moreover, simply reading off the percentage of people who have all-zero answers in the sample will not provide us with the proportion of people belonging to the latent class of politically disengaged.

Lastly, Eq. 1 highlights the major difference between the ZI-IRT approach and the LCA approach. Like in the LCA approach, we employ latent classes to model the unknown population heterogeneity (disengaged and those participating to a varying degree). However, the ZI-IRT model also allows us to model heterogeneity within the class of politically active. This is represented by a continuous latent variable of political participation and is quantified using each individual’s score on that variable, \({\theta }_{i}\). Therefore, with the ZI-IRT model, we effectively combine two perspectives of modeling participation as a latent variable—it is both categorical (capturing unobserved population heterogeneity with latent classes) and continuous (estimating the individuals’ scores denoting varying participation intensity levels within the class of politically active).

3.3 Zero-Inflated Beta-Binomial

To build an explanatory model for political participation, I treat political participation as an observable. I do so for two reasons. First, because this is the standard practice among students of political participation (García-Albacete, 2014; Kostelka, 2014; Letki, 2004; Oser & Hooghe, 2018). Second, estimating the zero-inflated measurement model with predictors would necessitate rather advanced programming skills in Stan (Carpenter et al., 2017), a general Bayesian statistical software program, potentially rendering the approach inaccessible for most applied researchers.

To show that accounting for the subpopulation of disengaged makes sense while regressing political participation on external variables, we should find a regression model that would closely mirror the data generating process as specified in the Zero-Inflated Graded Response Model. A very good candidate for that would be the Zero-Inflated Beta-Binomial model.

Zero-Inflated Beta-Binomial is an extension of the Beta-Binomial model. Beta-binomial model is used to model grouped binary data, that is a count of 1s out of a fixed number of trails, when we believe that the data might be correlated or that the probability of endorsing 1s across the trails is not constant (King, 1998). These are very plausible assumptions for modeling participation as a count of forms one participated in. We could write the probability function for Beta-Binomial as:

$$Beta\;Binomial\;\left( {n{|}N,\mu \phi , \left( {1 - \mu } \right)\phi } \right) = \left( {\begin{array}{*{20}c} n \\ N \\ \end{array} } \right)\frac{{B\left( {n + \mu \phi , N - n + \left( {1 - \mu } \right)\phi } \right)}}{{B\left( {\mu \phi , \left( {1 - \mu } \right)\phi } \right)}}$$
(2)

where \(n\) denotes the number of 1s (endorsements or successes) out of \(N\) trails, \(\mu\) is the expected average success probability, \(\phi\) is a measure for how much variation exists in \(\mu\) across the binary random variables, and \(B()\) denotes the beta function. What we usually want to predict is \(\mu\), which in our case can be understood as the expected average probability of participating.

The probability function for Zero-Inflated Beta-Binomial is a simple extension of Eq. 2, and could be written in the following way:

$$p\left( {n{|}N,\mu \phi , \left( {1 - \mu } \right)\phi ,\lambda } \right) = \left\{ {\begin{array}{*{20}l} {\lambda + \left( {1 - \lambda } \right) \times Beta\,Binomial\left( {0{|}N,\mu \phi , \left( {1 - \mu } \right)\phi } \right) \quad {\text{if }} n = 0} \hfill \\ {\left( {1 - \lambda } \right) \times Beta\,Binomial\left( {n{|}N,\mu \phi , \left( {1 - \mu } \right)\phi } \right)\quad {\text{if }} n > 0} \hfill \\ \end{array} } \right.$$
(3)

with \(\lambda\) being the proportion of excessive zeros, i.e. individuals belonging to the latent class of politically disengaged. The observations we have made about the data generating process with respect to the Zero-Inflated IRT model hold also here: (1) some zeros are contributed by the Beta-Binomial model, and the rest by a Bernoulli process; (2) non-zero answers are only due to the Beta-Binomial model. Again, we cannot treat all all-zero answers as coming from the individuals belonging to the latent class of politically disengaged.

3.4 Data and Cases

To investigate the latent structure of political participation, I use 8th and 9th round of the European Social Survey (ESS 2016; ESS 2018). I use both rounds of ESS because the participation data matrix is very sparse, with few cases in the co-occurrence cells. Using only one round led to nonsensical parameter values for the zero-inflated model and convergence problems. Using two consecutive waves for each country does not pose a problem since the scale of political participation, as measured by the ESS items, retains high levels of measurement invariance within countries across waves (Koc & Pokropek, 2022). ESS battery of participation items has been frequently used in the research on political participation (e.g., García-Albacete, 2014; Kostelka, 2014) and includes numerous forms. From the analyzed battery of items, I remove voting due to its special character and the practice of constructing composite measures without that item.

The items that measure political participation were recoded into dichotomous variables, with values 1 “participation” and 0 “no participation”. Refusals, “No answer”, and “Don’t know” answers were recoded as missing values. Besides, all interviewees-minors were excluded from the analysis as they could not participate in some participation forms for legal reasons. A full list of indicators can be found in Table 1.

Table 1 Item and item labels

Additionally, for some regression models, which I will discuss later in detail, I use five variables as predictors: gender, age, years of education, political interest, and political efficacy. Political efficacy is constructed as an index out of four variables measuring efficacy: (1) political system allows people to have a say in what government does; (2) political system allows people to have an influence on politics; (3) able to take an active role in a political group; (4) confident in own ability to participate in politics. The values of the index range from 0 to 1. Again, refusals, “No answer”, and “Don’t know” answers were recoded as missing values. Also, I removed all interviewees-minors and individuals who reported more than 30 years of formal education as such values are highly unrealistic.

3.5 Estimation and Model Comparison

To fit all the models, I use Stan. For the measurement models, I use the code provided by Magnus and Garnier-Villarreal (2022) and weakly informative priors (see Appendix for a detailed model specification). I run four chains, each with 2000 iterations, using the first 1000 iterations as a warm-up. Hence, inferences are based on the posterior distributions from 4000 draws. I set the adapt_delta argument to 0.99 to prevent divergent transitions after the warm-up phase.

For Beta-Binomial and Zero-Inflated Beta-Binomial regression models, I use multilevel versions of the said models to account for the fact that the observations are nested within countries. Additionally, I add a dummy variable denoting the ESS round to every model. Four models are estimated: (1) Beta-Binomial with the round dummy; (2) Beta-Binomial with the round dummy and a set of five frequently used antecedents of political participation: gender, age, years of education, political interest, and political efficacy; (3) Zero-Inflated Beta-Binomial where the expected average probability of participating is regressed on the set of five predictors and the round dummy, and the probability of belonging to the latent class of politically disengaged is regressed on the round dummy; (4) Zero-Inflated Beta-Binomial where the average probability of participating and the probability of belonging to the latent class of politically disengaged are regressed on the set of five predictors and the round dummy. In all the models, \(\phi\) parameters have random intercepts. In model 3, the probability of belonging to the latent class of politically disengaged has a random intercept and is regressed on the round dummy.

To fit the regression models I use brms (Bürkner, 2017), an R package that has Stan as its backend. I run four chains, each with 5000 iterations, using the first 2500 iterations as a warm-up. Hence, inferences are based on the posterior distributions from 10,000 draws. I set the adapt_delta argument to 0.99 to prevent divergent transitions after the warm-up phase. I use weakly informative priors, which are described in the Appendix.

To compare the fit of the measurement and regression models, I follow the common practice in Bayesian statistics and use leave-one-out cross-validation and the difference in expected log pointwise predictive density (ELPD) (Lambert, 2018; McElreath, 2020). A model is thought to have superior out-of-sample predictive accuracy over another one when the difference in ELPD values is at least larger than twice the estimated standard error of the difference (Vehtari et al., 2017).

Additionally, for the measurement models, I compare difficulty and discrimination parameters, and the distribution of the \(\theta\) parameters, which denote the individual scores on the latent variable of political participation. To be more precise, I compare the individual mean values of \(\theta\) s. For regression models, I compare the regression coefficients using predicted probabilities.

4 Results

4.1 Measurement Models

Table 2 displays the best fitting model and the difference in ELPD between this model and the alternative one, along with the standard error of this difference. In all countries, the zero-inflated version has a better fit. The difference in ELPD is much larger than twice the estimated standard error of the difference (except for Sweden, Norway, and Finland, the difference is of two orders of magnitude). This means that we should strongly prefer the zero-inflated model over the model not accounting for the zero-inflation.

Table 2 Comparison of out-of-sample predictive accuracy for the regular and zero-inflated IRT models

What are the consequences of accounting for the zero-inflation for the model parameters? The estimates of item discrimination move to the left (to the zero) by 0.46 and the estimates of item difficulty by 0.08, on average. By looking at Table 3, we can better understand how well the items measure the latent trait of political participation after taking into account the existence of a subpopulation of politically disengaged.Footnote 3 It is worth inspecting the properties of the items as most students of political participation concentrate on the properties of the whole scale or latent factors (e.g., Koc, 2021; Ohme et al., 2018; Talò & Mannarini, 2015).

Table 3 Item discrimination and item difficulty estimates for the regular and zero-inflated IRT models

First, all item difficulty estimates are positive, meaning that to have a 50% chance of endorsing the items one needs to have an above-average level on the latent trait of political participation. On average, the most difficult items are working for a party or action group (average difficulty: 2.23), taking part in public demonstrations (2.10), and badging (1.82). Second, the greatest discriminatory power is provided by the item “working for a political party or action group”. That is, this item best detects differences between individuals on the latent trait of political participation. One obvious implication of the observations we have made would be that these items provide very little information about persons with below-average levels on the latent variable of participation.

Another question that we can answer by comparing the regular and Zero-Inflated IRT models is that about the distribution of the latent trait of participation in the population. Figure 2 shows the densities representing average individual scores on the latent variable of political participation. As a result of modeling the excess of zeros, the mass of the distribution of the latent trait moves to the right, meaning that after taking into consideration the subpopulation of disengaged, the picture of the intensity of participation in the population changes—people seem to be more active. The size of this move is a function of the proportion of the latent class of disengaged.

Fig. 2
figure 2

Densities of average individual scores on the latent variable of political participation. Note Distributions in the top row have different x and y axes 

Figure 3 reveals that new democracies, along with Italy, feature the highest proportions of politically disengaged citizens. In Hungary, Lithuania, Poland, and Italy, the disengaged make up more than 50% of the population. The distribution of the latent trait in these countries shows high kurtosis, marked by a distinct peak and narrow spread, indicating minimal variance in the intensity of participation. In stark contrast, Scandinavian countries such as Finland, Norway, and Sweden have the smallest proportions of disengaged citizens. After accounting for the latent class of disengaged citizens in these countries, the mass of the distribution shifts only slightly. Furthermore, these Scandinavian countries exhibit a larger spread in the distribution of political participation, indicating more variance compared to the other countries analyzed in the study.

Fig. 3
figure 3

Proportions of the latent class of politically disengaged

Nevertheless, even after modeling the excess of zeros, the majority of the population (representing 51% of the distribution's mass) falls below zero on the latent trait of participation and the distribution of this trait remains heavily skewed to the right in almost all countries. The notable exceptions are Finland and Sweden, where the majority neither falls below zero, nor is the distribution significantly skewed. Lastly, we observe that the proportions of the latent class of disengaged, as estimated by the Zero-Inflated IRT, do not match the proportions from Fig. 1.

4.2 Regression Models

Similarly to measurement models, regression models accounting for the zero-inflation have a better fit than the models not accounting for the excess of zeros, which is shown in Table 4. Zero-Inflated Beta-Binomial with predictors for both parts of the model has the best fit, followed by ZIBB with predictors only for the average probability of participating. ZIBB with predictors only for the average probability of participating has a better fit than the Beta-Binomial model with predictors (ELPD difference: − 57.3; SE: 11.8). Hence, judging by how well the models fit the data, zero-inflated models are to be preferred. They seem to better reflect the data generating process.

Table 4 Comparison of out-of-sample predictive accuracy for zero-inflated and regular beta-binomial model

Figure 4 shows partial predicted probabilities for ZIBB with predictors for both parts (Model 4) and Beta-Binomial (Model 2) with the same predictors. To calculate the probabilities, values of covariates were either set to their mean value or, in the case of political interest, ESS round, and gender, to “Quite interested”, round 9, and “Women”, respectively.

Fig. 4
figure 4

Partial predicted probabilities for Beta-Binomial and Zero-Inflated Beta-Binomial (μ; expected average probability of participating). Note To calculate partial predicted probabilities, covariates were either set to their mean value, or, in case of political interest, ESS rounds, and gender, to “quite interested”, 9, and “women”, respectively. The ribbons and error bars denote 95% credible intervals

First, the predicted probabilities are systematically higher for ZIBB than for Beta-Binomial. For years of education, the line representing the impact of the variable on political participation is less steep if estimated with the ZIBB model. Overall, however, the tendencies and relative differences remain largely the same. The directions of the predictions in both models are in line with the expectations from the literature: political interest, political efficacy, and education have a positive impact on the probability of participation, the probability decreases with age, and women are slightly more likely to participate than men.

What is special about the ZIBB model is that, unlike the regular Beta-Binomial, it offers the possibility of modeling the zero-inflation, i.e. the probability of ending up in the latent class of politically disengaged. Figure 5 shows partial predicted probabilities which were calculated in the same way as previously. We can see that two variables predict well the probability of belonging to the class of disengaged: political interest and education. Interestingly, political efficacy—a strong predictor of the probability of participating—has no effect on belonging to the class of disengaged. The same can be said about age. In the case of gender, at first sight, we would say that there is no effect. A more precise approach involving calculating marginal effects (contrasts) shows that the effect of being a woman is extremely small, somewhere between 0.3 and 2.8 percentage points, with 1.3 points as the mean. It is relatively easy to explain why political interest predicts belonging to the class of disengaged—lack of interest is the key measure of political apathy (Dahl et al., 2018). It is more difficult to explain the role of education as this variable often captures several different things, such as skills, knowledge, social status, and social network centrality (Persson, 2013). Also, interpreting the non-effect of political efficacy is challenging given that these models are mainly predictive. However, this non-effect aligns with correlational studies that use voting turnout as a dependent variable (Dahl et al., 2018). Still, such dramatic change should provoke future research to single out and investigate the mechanism of political efficacy on the probability of belonging to the latent class of politically disengaged. Full regression tables can be found in the Online Supplementary Materials (ESM 2).

Fig. 5
figure 5

Partial predicted probabilities for Zero-inflated beta-binomial (\(\lambda\); probability of belonging to the latent class of politically disengaged). Note To calculate partial predicted probabilities, covariates were either set to their mean value, or, in case of political interest, ESS rounds, and gender, to “quite interested”, 9, and “women”, respectively. The ribbons and error bars denote 95% credible intervals

5 Discussion

In this paper, I argued that political participation exhibits features of categorical and continuous latent variables, i.e. that it makes sense to (1) distinguish two latent classes: a class of politically disengaged, with people refraining from participation altogether, and a class composed of people who participate to a varying degree; and (2) to allow for quantitative differences in the class of politically active. To verify those claims, I used ESS data and analyzed them with zero-inflated measurement and regression models.

Existing studies on modeling political participation as a latent variable consider the possibility of political participation being either continuous or categorical. There has been no approach, to the best knowledge of the author of this work, that would go beyond this dichotomy. Modeling participation as a continuous latent variable rests upon a very strong assumption of population homogeneity, which as the existing studies show (Oser, Hooghe & Marien, 2013; Oser, 2017, 2021) is not plausible. Accounting for the population heterogeneity by the means of LCA is also problematic because with LCA we assume that there are no quantitative differences within the homogenous groups.

To overcome the limitations of the existing approaches, I used a Zero-Inflated IRT model and a regression model closely resembling the data generating process implied by that IRT model—Zero-Inflated Beta-Binomial. Judging by the out-of-sample predictive accuracy, zero-inflated models that account for the latent class of politically disengaged are superior to the models not accounting for the existence of that class. Accounting for the zero-inflation showed to have an impact on various model estimates: item difficulty and discrimination estimates decreased, the distribution of the latent trait in the population moved to the right, and predicated probabilities for the participation intensity part of the regression model increased. Also, thanks to the Zero-Inflated Beta-Binomial model, I showed that one can and should model the probability of belonging to the class of politically disengaged and that the predictors of belonging to that class do not function in the same way as they would in predicting the participation intensity. Considering all the evidence, not accounting for the excess of zeros, attributed to the latent class of politically disengaged, casts doubt on the measurement validity. This doubt arises because we (1) fail to translate the theoretical concept into the measurement specification (translation error; Fariss et al., 2020; Trochim & Donnelly, 2007), and (2) leave the mechanism leading to the zero-inflation unmodeled.

Moreover, the proposed methodology outlined in this study could be extended to a variety of contexts beyond the ESS data and European countries. Particularly in regions where democracy is fragile or in transitional stages, such as some parts of Latin America, this methodology might provide valuable insights into the nuances of political participation. It could also be applicable to data from other large-scale surveys like the World Values Survey, broadening the range of possible analyses and outcomes.

Yet, this research suffers from several limitations. First, the measurement models assume normality for the latent variables, which as the density plots suggest-may not be true. Future research should consider other population distributions for the latent trait, such as skew-normal (Azzalini, 2014). Second, even though ESS has eight participation items it is still relatively few and it would be beneficial to compare models with a larger battery of items. Third, the predictors used for the model comparison cannot be interpreted causally. For that, we would need to single out causal mechanisms and ponder on the conditions of identification (see, Keele et al., 2020). Only then we would be able to understand, for example, why political efficacy does not have any effect on belonging to the latent class of politically disengaged. Finally, unlike the LCA approach, the ZI-IRT model proposed in this study does not permit differentiation beyond two latent classes. While modeling the continuous trait and accommodating more than two latent classes might be possible in a full Factor Mixture Modelling approach, this presents additional challenges due to the complexity of fitting these models (see, Clark et al., 2013). Nevertheless, this is an avenue that future research could explore.