Abstract
Scholars have made considerable strides in evaluating and improving the external validity of experimental research. However, little attention has been paid to a crucial aspect of external validity – the topic of study. Researchers frequently develop a general theory and hypotheses (e.g., about policy attitudes), then conduct a study on a specific topic (e.g., environmental attitudes). Yet, the results may vary depending on the topic chosen. In this paper, we develop the idea of topic sampling – rather than studying a single topic, we randomly sample many topics from a defined population. As an application, we combine topic sampling with a classic survey experiment design on partisan cues. Using a hierarchical model, we efficiently estimate the effect of partisan cues for each policy, showing that the size of the effect varies considerably, and predictably, across policies. We conclude with advice on implementing our approach and using it to improve theory testing.
Similar content being viewed by others
Notes
In a recent study, Barber and Pope (2019) study 10 issues at once, providing perhaps the most generalizable findings. Yet, the study was restricted to 10 topics on which former President Trump publicly took stances on both sides of the issue. Moreover, their analysis focuses only on the average effect, while analyses in the supplementary materials suggest meaningful but unexamined heterogeneity in effect size across issues.
We used 2016 rather than a later year because 2017 polls were still being added to the Roper database at the time.
Specifically, the terms were “favor or oppose or for or against or should or approve or support.” Diagnostic checks suggested that this set of terms included virtually all policy attitudes measured in this time period.
Questions that asked about the same policy but used different question wording were considered redundant.
Of the 48 policies we sampled (see below), 40% also showed up three years later in the 2019 Roper population, suggesting considerable stability in the population over time.
To validate our coding, we compared assigned values to partisan differences in the data among untreated respondents. In 77% of the cases, we observed a statistically significant difference in the expected direction. For the remaining 23% of cases, there was no significant partisan difference. Thus, our coding reliably mapped onto partisan differences.
This study was reviewed and approved by the London School of Economics Research Ethics Committee. All respondents gave informed consent before participating in the study.
Using the average of posterior simulations with Stan (Carpenter et al. 2017) produces nearly identical estimates of the awareness of the parties’ positions on each issue.
This possibility of investigating many moderators underscores the importance of pre-registration. We also note that the topic sampling should be described in the pre-registration document just as a researcher would describe the sampling of participants.
References
Arceneaux, K. (2007). Can partisan cues diminish democratic accountability? Political Behavior, 30(2), 139–160.
Bakker, B. N., & Lelkes, Y., and Ariel Malka (2020). Understanding partisan cue receptivity: Tests of predictions from the Bounded rationality and expressive utility perspectives. Journal of Politics, 82(3), 1061–1077.
Bansak, K., Hainmueller, J., Hopkins, D. J., & Teppei Yamamoto. (2021). “Conjoint Survey Experiments.” Ch. 2 in Advances in Experimental Political Science, eds. Druckman, J. N., & Green, D. P. (p. 19). Cambridge: Cambridge University Press.
Barabas, J., and Jennifer Jerit (2010). Are Survey experiments externally valid? American Political Science Review, 104(02), 226–242.
Barber, M., and Jeremy C. Pope (2019). Does Party Trump ideology? Disentangling Party and ideology in America. American Political Science Review, 113(1), 38–54.
Bates, D., Mächler, M., Bolker, B., & Walker, S. (2015). Fitting linear mixed-effects models using Lme4. Journal of Statistical Software, 67(1), 1–48.
Berinsky, A. J., Gregory, A., Huber, & Lenz, G. S. (2012). Evaluating online labor markets for experimental research: Amazon.Com’s mechanical Turk. Political Analysis, 20(3), 351–368.
Brandt, M. J., Jarret, T., & Crawford (2019). Studying a heterogeneous array of Target Groups can help us understand prejudice. Current Directions in Psychological Science, 28(3), 292–298.
Brutger, R. (2022). “Abstraction and Detail in Experimental Design.” American Journal of Political Science,
Bullock, W., Imai, K., & Shapiro, J. (2011). Statistical Analysis of Endorsement Experiments: Measuring Support for Militant Groups in Pakistan. Political Analysis, 19(4), 363–384.
Bullock, J. G. (2011). “Elite Influence on Public Opinion in an Informed Electorate.” American Political Science Review, 105(03), 496–515.
Bürkner, P. C. (2017). Brms: An R Package for bayesian Multilevel Models using Stan. Journal of Statistical Software, 80(1), 1–28.
Butler, D. M., and Jonathan Homola (2017). An empirical justification for the Use of racially distinctive names to signal race in experiments. Political Analysis, 25(1), 122–130.
Carpenter, B., Gelman, A., Hoffman, M. D., Lee, D., Goodrich, B., Betancourt, M., Brubaker, M., Guo, J., Li, P., & Riddell, A. (2017). “Stan: A Probabilistic Programming Language.” Journal of Statistical Software, 76(1): 1–32.
Cameron, A., Colin, J. B., Gelbach, & Miller, D. L. (2011). Robust inference with Multiway Clustering. Journal of Business & Economic Statistics, 29(2), 238–249.
Chong, D., and James N. Druckman (2007). Framing Public Opinion in competitive democracies. American Political Science Review, 101(04), 637–655.
Chong, D., and James N. Druckman (2010). Dynamic public opinion: Communication Effects over Time. American Political Science Review, 104(04), 663–680.
Chong, D., and James N. Druckman (2012). Counterframing Effects. The Journal of Politics, 75(01), 1–16.
Chung, Y., et al. (2013). “A Nondegenerate Penalized Likelihood Estimator for Variance Parameters in Multilevel Models.” Psychometrika, 78(4), 685–709.
Clifford, S., Sheagley, G., & Piston, S. (2021). Increasing Precision without Altering Treatment Effects: Repeated measures designs in Survey experiments. American Political Science Review, 115(3), 1048–1065.
Coppock, A., & Green, D. P. (2015). Assessing the correspondence between experimental results obtained in the lab and field: A review of recent Social Science Research. Political Science Research and Methods, 3(01), 113–131.
Coppock, A., Leeper, T. J., & Mullinix, K. J. (2018). “Generalizability of Heterogeneous Treatment Effect Estimates across Samples.” Proceedings of the National Academy of Sciences: 201808083.
Ditto, P. H., et al. (2019). At least Bias is bipartisan: A Meta-Analytic comparison of partisan Bias in Liberals and Conservatives. Perspectives on Psychological Science, 14(2), 273–291.
Druckman, J. N., Thomas, J., & Leeper (2012). Learning more from political communication experiments: Pretreatment and its Effects. American Journal of Political Science, 56(4), 875–896.
Franco, A., & Malhotra, N., and Gabor Simonovits (2014). Publication Bias in the Social Sciences: Unlocking the file drawer. Science (New York N Y), 345(6203), 1502–1505.
Goodrich, B., Gabry, J., Ali, I., & Brilleman, S. (2020). rstanarm: Bayesian applied regression modeling via Stan. R package version 2.21.1. https://mc-stan.org/rstanarm
Gelman, A., and Jennifer Hill (2007). Data Analysis using regression and Multilevel/Hierarchical models. New York: Cambridge University Press.
Gelman, A., Jennifer Hill, and Masanao Yajima (2012). Why we (usually) don’t have to worry about multiple comparisons. Journalof Research on Educational Effectiveness, 5(2), 189–211.
Gilens, M. (2001). Political ignorance and collective policy preferences. American Political Science Review, 95(2), 379–396.
Green, D. P., & Kern, H. L. (2012). Modeling Heterogeneous Treatment Effects in Survey experiments with bayesian additive regression trees. Public Opinion Quarterly, 76(3), 491–511.
Grieco, J. M., Gelpi, C., & Reifler, J., and Peter D. Feaver (2011). Let’s get a second opinion: International Institutions and American Public support for War1. International Studies Quarterly, 55(2), 563–583.
Guess, A., & Coppock, A. (2020). Does Counter-Attitudinal Information Cause Backlash? Results from Three Large Survey Experiments. British Journal of Political Science, 50(4), 1497–1515.
Guisinger, A., and Elizabeth N. Saunders (2017). Mapping the Boundaries of Elite Cues: How elites shape Mass Opinion across International Issues. International Studies Quarterly, 61(2), 425–441.
Hainmueller, J., & Hopkins, D. J., and Teppei Yamamoto (2014). Causal inference in Conjoint Analysis: Understanding multidimensional choices via stated preference experiments. Political Analysis, 22(1), 1–30.
Jerit, J. (2009). How predictive appeals affect policy opinions. American Journal of Political Science, 53(2), 411–426.
Jerit, J. (2012). Partisan perceptual bias and the information environment. The Journal of Politics, 74(03), 672–684.
Jerit, J., & Barabas, J. (2013). Comparing contemporaneous laboratory and field experiments on media effects. Public Opinion Quarterly, 77(1), 256–282.
Johnston, C. D., Lavine, H., & Federico, C. M. (2017). Open versus closed: Personality, identity, and the politics of redistribution. Cambridge: Cambridge University Press.
Johnston, C. D. (2015). Personality dispositions and political preferences across hard and easy issues. Political Psychology, 36(1), 35–53.
Kam, C. D., & Trussler, M. J. (2016). At the nexus of observational and experimental research: Theory, specification, and analysis of experiments with heterogeneous treatment effects. Political Behavior, 39, 1–27.
Kam, C. D. (2005). Who Toes the Party Line? Cues, Values, and Individual Differences. Political Behavior, 27, 163–182.
Kay, M. (2019). “Tidybayes: Tidy Data and Geoms for Bayesian Models.” https://doi.org/10.5281/zenodo.1308151
Kuklinski, J. H., et al. (2000). Misinformation and the currency of democratic citizenship. The Journal of Politics, 62(3), 790–816.
Lenz, G. S. (2009). Learning and opinion change, not priming: Reconsidering the priming hypothesis. American Journal of Political Science, 53(4), 821–837.
Lindner, N. M., & Nosek, B. A. (2009). Alienable Speech: Ideological variations in the application of Free-Speech Principles. Political Psychology, 30(1), 67–92.
McDermott, R. (2002). Experimental methodology in Political Science. Political Analysis, 10(4), 325–342.
Miller, J. M., & Krosnick, J. A. (2004). Threat as a motivator of political activism: A field experiment. Political Psychology, 25(4), 507–523.
Nicholson, S. P. (2012). “Polarizing Cues.” American Journal of Political Science, 56(1), 52–66.
Nyhan, B., and Jason Reifler (2010). When Corrections fail: The persistence of political Misperceptions. Political Behavior, 32(2), 303–330.
Ryan, T. J. (2014). “Reconsidering Moral Issues in Politics.” The Journal of Politics, 76(2), 1–18.
Ryan, T. J. (2017). No compromise: Political consequences of moralized attitudes. American Journal of Political Science, 61(2), 409–423.
Simas, E. N., & Milita, K., and John Barry Ryan (2021). Ambiguous rhetoric and legislative accountability. Journal of Politics, 83(4), 1695–1705.
Slothuus, R. (2016). Assessing the influence of political parties on public opinion: The challenge from pretreatment effects. Political Communication, 33(2), 302–327.
Swire-Thompson, Briony, J., DeGutis, & Lazer, D. (2020). Searching for the backfire effect: Measurement and design considerations. Journal of Applied Research in Memory and Cognition, 9(3), 286–299.
Tappin, B. M. (2022). “Estimating the Between-Issue Variability in Party Elite Cue Effects.” Public Opinion Quarterly, 86(4), 862–885
Tavits, M. (2007). Principle vs. pragmatism: Policy shifts and political competition. American Journal of Political Science, 51(1), 151–165.
Tesler, M. (2015). Priming Predispositions and changing policy positions: An account of when Mass Opinion is primed or changed. American Journal of Political Science, 59(4), 806–824.
Vehtari, A., & Gelman, A. (2017). Practical bayesian model evaluation using leave-one-out cross-validation and WAIC. Statistics and Computing, 27, 1413–1432.
Wells, G. L. (2016). Stimulus sampling and social psychological experimentation. Personality and Social Psychology Bulletin, 25(9), 1115–1125.
Wood, T., & Porter, E. (2019). “The elusive backfire effect: mass attitudes’ steadfast factual adherence.” Political Behavior, 41, 135–163.
Acknowledgements
The authors would like to thank Brandon de la Cuesta and Brendan Nyhan for helpful comments.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
This study was funded by the Danish Council for Independent Research award DFF–4003-00192B. The authors declare they have no financial interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Below is the link to the electronic supplementary material.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Clifford, S., Leeper, T.J. & Rainey, C. Generalizing Survey Experiments Using Topic Sampling: An Application to Party Cues. Polit Behav (2023). https://doi.org/10.1007/s11109-023-09870-1
Accepted:
Published:
DOI: https://doi.org/10.1007/s11109-023-09870-1