Abstract
The explanatory/pragmatic-trial distinction enjoys a burgeoning philosophical and medical literature and a significant contingent of support among philosophers and healthcare stakeholders as an important way to assess the design and results of randomized controlled trials. A major motivation has been the need to provide relevant, generalizable data to drive healthcare decisions. While talk of pragmatic and explanatory trials could be seen as convenient shorthand, the distinction can also be seen as harboring deeper issues related to inferential strategies used to evaluate causal claims regarding medical treatments. A comprehensive, critical analysis of the distinction and underlying epistemological framework upon which the distinction is based, particularly with respect to treatment effectiveness, has yet to be forthcoming. I provide this, analyzing the distinction’s relationship to generalizability and cognate distinctions between ideal conditions and real-world practice, internal and external validity, and efficacy and effectiveness. I also analyze recent philosophical work that relies on the explanatory/pragmatic-trial distinction and that advocates for more pragmatic trials. I conclude that as an organizing principle for trial-design decisions and trial evaluation, the explanatory/pragmatic-trial distinction is conceptually problematic and not as useful as its proponents seem to think. Since some pragmatic-trial features can be inimical to establishing treatment effectiveness, pragmatic-trial features should not be conflated with pragmatic trials’ avowed goals. If the distinction is to be useful, it and some associated concepts, including generalizability, should be reformulated, lest they continue to underlie a medical epistemology that could contribute to methodologically flawed and potentially unethical advice for the design and interpretation of trials.
Similar content being viewed by others
Notes
I henceforth simply refer to pragmatic or explanatory RCTs as pragmatic or explanatory trials, based on the assumption that all explanatory and pragmatic trials are RCTs. This is not a universally accepted assumption, though (e.g. Porzsolt et al., 2015). To insist on randomization as excluding whether a trial is pragmatic, however, collapses the distinction between RCTs and non-randomized observational studies in a way many authors (e.g. Thorpe et al., 2010) might be uncomfortable with and elevates randomization to a level that does not do justice to the many other trial features that determine how closely “real-world” conditions are approximated. The explanatory/pragmatic-trial distinction is best thought of as a multidimensional continuum (whereby trials can be somewhat explanatory and somewhat pragmatic and as having some explanatory-trial and pragmatic-trial features), although it could be viewed as a dichotomy when, for example, contrasting a very explanatory trial with a very pragmatic trial. To be clear, my purpose in this article is not to argue against it being a simplistic distinction, and in fact many authors (e.g. Borgerson 2013; Schwartz and Lellouch 1967; Thorpe et al., 2009) explicitly see the distinction as not a simplistic one.
The status of blinding in pragmatic trials is not universally agreed upon: Williams et al. (2015) think pragmatic trials should always involve blinding when possible, Dal-Ré et al. (2018) think pragmatic trials should avoid blinding of treatment assignment, Thorpe et al. (2009) think blinding is not inconsistent with a pragmatic trial, while other authors (Maningat and Breslow 2011; Ware and Hamel 2011) think blinding is often omitted in pragmatic trials.
A treatment’s effectiveness supervenes on the set of its true treatment effectiveness claims. In practice it is rarely if ever known for certain what a treatment’s full set of true treatment effectiveness claims is. Knowledge of a treatment’s effectiveness is in most cases partial and subject to some level of uncertainty, given the fallibility of medical knowledge. Actual treatment effectiveness claims are at best well-confirmed, empirically well-supported, practically reliable, etc. I thank Bert Leuridan for these points in this footnote, and for suggesting that I pay close attention to not conflate the ontic with the epistemic.
How to amalgamate evidence informing a treatment effectiveness claim, and how to evaluate the evidence’s warrant, are of course crucial questions but are beyond the scope of this article.
Hey and Weijer (2013, p. 3), for example, note that no single well-designed trial can provide sufficient evidence of effectiveness. This is not to say that not viewing effectiveness as comprising a prediction but instead as an extrapolatable trial result entails the view that single studies or types of studies establish effectiveness, but it does seem to make it more likely, especially when stakeholders use RCTs as dispute-resolution mechanisms (Deaton & Cartwright, 2018).
In this respect even something as simple as, “This treatment will likely make me feel better” may provide a relevant notion of effectiveness for a patient in clinical practice. Such a prediction does not match any outcome measure from a study (RCT or otherwise), yet it may be all a patient is able to fathom, and could suffice as the key part of an argument for a useful treatment effectiveness claim.
Fuller (2013, p. 644) offers a better description of generalizability as asking what useful inferences can be made from RCT results.
Many thanks to Bert Leuridan for suggesting I fully expand the typology beyond the abbreviated version I originally had.
I thank Jonathan Fuller for suggesting the third type.
Alternatively, a different term than “generalizability” could be used, or a qualification added (e.g. “generalizability for treatment effectiveness claims”). I retain “generalizability” for prudential reasons, and unless otherwise indicated use it henceforth throughout this article according to the reformulated meaning I propose.
Reiss’s (2015) pragmatist theory of evidence (which does not pertain to pragmatic trials) also looks at the question(s) to be answered by a trial and what evidence could warrant a successful outcome. Similarly, Cartwright (2010, p. 69) thinks, “the best way to evaluate effectiveness claims is by the construction of a causal model, where information about the behaviour of C in an RCT is only one small part of the information needed to construct the model.”
Although this was not the case when Schwartz and Lellouch wrote, the recent growth of interest in “learning healthcare systems” (of which pragmatic trials benefit and benefit from; Tuzzio and Larson 2019) is partly based on the idea that all clinician–patient encounters can in principle elicit information that bears on causal hypotheses. This can contribute to the implementation of trial results into clinical practice and aid both understanding and decision. I thank an anonymous reviewer for this journal for this point.
And guidance sought from whether criteria (such as those articulated by Hey and Weijer 2016) are satisfied for when placebo controls are ethically allowed.
Notwithstanding this, causation in clinical care and in studies of its tools is complex and in many cases opaque, likely affected by factors not included or studied in a trial, such as molecular and genetic components. I thank an anonymous reviewer for this journal for this point.
Nonetheless, comparative effectiveness trials may indeed contribute to better inferences to treatment effectiveness claims than placebo-controlled trials, in addition to often being ethically mandated.
This should not be misinterpreted as the claim that placebo-controlled trials are superior to active-controlled trials.
References
Alphs, L. D., & Bossie, C. A. (2016). ASPECT-R-A tool to rate the pragmatic and explanatory characteristics of a clinical trial design. Innovations in Clinical Neuroscience, 13(1–2), 15–26.
Anderson, J. A. (2006). The ethics and science of placebo-controlled trials: Assay sensitivity and the Duhem-Quine thesis. Journal of Medicine and Philosophy, 31(1), 65–81.
Aronson, J. K., La Caze, A., Kelly, M. P., Parkkinen, V. P., & Williamson, J. (2018). The use of mechanistic evidence in drug approval. Journal of Evaluation in Clinical Practice, 24(5), 1166–1176.
Ashcroft, R. (2002). What is clinical effectiveness? Studies in History and Philosophy of Science Part C: Studies in History and Philosophy of Biological and Biomedical Sciences, 33, 219–233.
Bluhm, R., & Borgerson, K. (2018). An epistemic argument for research-practice integration in medicine. Journal of Medicine and Philosophy, 43(4), 469–484.
Borgerson, K. (2005). Evidence-based alternative medicine? Perspectives in Biology and Medicine, 48(4), 502–515.
Borgerson, K. (2013). Are explanatory trials ethical? Shifting the burden of justification in clinical trial design. Theoretical Medicine and Bioethics, 34(4), 293–308.
Brænd, A. M., Staand, J., & Klovning, A. (2017). Clinical drug trials in general practice: How well are external validity issues reported? BMC Family Practice, 18, 113.
Caffo, O. (2001). Radiosensitization with chemotherapeutic agents. Lung Cancer, S81-S90.
Candelaria, M., Garcia-Arias, A., Cetina, L., & Dueñas-Gonzalez, A. (2006). Radiosensitization with chemotherapeutic agents. Cisplatin and beyond. Radiation Oncology, 1, 15.
Campbell, D. T. (1957). Factors relevant to the validity of experiments in social settings. Psychological Bulletin, 54(4), 297–312.
Cartwright, N. (2010). What are randomised controlled trials good for? Philosophical Studies, 147, 59–70.
Cartwright, N. (2011). Predicting what will happen when we act. What counts for warrant? Preventive Medicine, 53, 221–224.
Cartwright, N. (2012). Presidential Address: Will this policy work for you? Predicting effectiveness better: How philosophy helps. Philosophy of Science, 79, 973–989.
Cartwright, N., & Munro, E. (2010). The limitations of randomized controlled trials in predicting effectiveness. Journal of Evaluation in Clinical Practice, 16(2), 260–266.
Cartwright, N., & Stegenga, J. (2011). A theory of evidence for evidence-based policy. Proceedings of the British Academy, 171, 289–319.
Cherkin, D. C., Eisenberg, D., Sherman, K. J., Barlow, W., Kaptchuk, T. J., Street, J., et al. (2001). Randomized trial comparing traditional Chinese medical acupuncture, therapeutic massage, and self-care education for chronic low back pain. Archives of Internal Medicine, 161(8), 1081–1088.
Coppock, A., & Green, D. P. (2015). Assessing the correspondence between experimental results obtained in the lab and field: A review of recent social science research. Political Science Research and Methods, 3(1), 113–131.
Corrigan-Curay, J., Sacks, L., & Woodcock, J. (2018). Real-world evidence and real-world data for evaluating drug safety and effectiveness. Journal of the American Medical Association, 320(9), 867–868.
Dahlen, S. E., Dahlen, B., & Drazen, J. M. (2011). Asthma treatment guidelines meet the real world. New England Journal of Medicine, 364(18), 1769–1770.
Dal-Ré, R., Janiaud, P., & Ioannidis, J. P. A. (2018). Real-world evidence: How pragmatic are randomized controlled trials labeled as pragmatic? BMC Medicine, 16(1), 49.
Deaton, A., & Cartwright, N. (2018). Understanding and misunderstanding randomized controlled trials. Social Science and Medicine, 210, 2–21.
Dekkers, O. M., Bossuyt, P. M., & Vandenbroucke, J. P. (2017). How trial results are intended to be used: Is PRECIS-2 a step forward? Journal of Clinical Epidemiology, 84, 25–26.
DeMaria, A. N. (2008). Clinical trials and clinical judgment. Journal of the American College of Cardiology, 51(11), 1120–1122.
Dincer, F., & Linde, K. (2003). Sham interventions in randomized clinical trials of acupuncture—a review. Complementary Therapies in Medicine, 11(4), 235–242.
Doll, R., Peto, R., Boreham, J., & Sutherland, I. (2004). Mortality in relation to smoking: 50 years’ observations on male British doctors. BMJ, 328, 1519.
Duan, N., Kravitz, R. L., & Schmid, C. H. (2013). Single-patient (n-of-1) trials: A pragmatic clinical decision methodology for patient-centered comparative effectiveness research. Journal of Clinical Epidemiology, 66(8 Suppl), S21-28.
Fahed, R., Finitsis, S., Khoury, N., Deschaintre, Y., Daneault, N., Gioia, L., et al. (2018). A randomized pragmatic care trial on endovascular acute stroke interventions (EASI): Criticisms, responses, and ethics of integrating research and clinical care. Trials, 19(1), 508.
Fisher, A. J., Medaglia, J. D., & Jeronimus, B. F. (2018). Lack of group-to-individual generalizability is a threat to human subjects research. PNAS, 115(27), E6106–E6115.
Freedland, K. E., Mohr, D. C., Davidson, K. W., & Schwartz, J. E. (2011). Usual and unusual care: Existing practice control groups in randomized controlled trials of behavioral interventions. Psychosomatic Medicine, 73(4), 323–335.
Fuller, J. (2013). Rationality and the generalization of randomized controlled trial evidence. Journal of Evaluation in Clinical Practice, 19, 644–647.
Fuller, J. (2021). The myth and fallacy of simple extrapolation in medicine. Synthese, 198, 2919–2939.
Gartlehner, G., Hansen, R. A., Nissman, D., Lohr, K. N., & Carey, T. S. (2006). A simple and valid tool distinguished efficacy from effectiveness studies. Journal of Clinical Epidemiology, 59(10), 1040–1048.
Godwin, M., Ruhland, L., Casson, I., MacDonald, S., Delva, D., Birtwhistle, R., et al. (2003). Pragmatic controlled clinical trials in primary care: The struggle between external and internal validity. BMC Medical Research Methodology, 3, 28.
Hansen, K., & Kappel, K. (2010). The proper role of evidence in complementary/alternative medicine. Journal of Medicine and Philosophy, 35(1), 7–18.
Hey, S. P. (2015). What theories are tested in clinical trials? Philosophy of Science, 82, 1318–1329.
Hey, S. P., & Weijer, C. (2013). Assay sensitivity and the epistemic contexts of clinical trials. Perspectives in Biology and Medicine, 56(1), 1–17.
Hey, S. P., & Weijer, C. (2016). What questions can a placebo answer? Monash Bioethics Review, 34, 23–36.
Holman, B. (2019). Philosophers on drugs. Synthese, 196, 4363–4390.
Horn, A. R., Weijer, C., Grimshaw, J., Brehaut, J., Fergusson, D., Goldstein, C. E., & Taljaard, M. (2018). An ethical analysis of the SUPPORT trial: Addressing challenges posed by a pragmatic comparative effectiveness randomized controlled trial. Kennedy Institute of Ethics Journal, 28(1), 85–118.
Howick, J. (2009). Questioning the methodologic superiority of “placebo” over “active” controlled trials. American Journal of Bioethics, 9(9), 34–48.
Howick, J. (2019). Exploring the asymmetrical relationship between the power of finance bias and evidence. Perspectives in Biology and Medicine, 62(1), 159–187.
Howick, J., Friedemann, C., Tsakok, M., Watson, R., Tsakok, T., Thomas, J., et al. (2013). Are treatments more effective than placebos? A systematic review and meta-analysis. PLoS One, 8(5), e62599.
Howick, J., & Hoffmann, T. (2018). How placebo characteristics can influence estimates of intervention effects in trials. CMAJ, 190(30), E908–E911.
Jiménez-Buedo, M. (2011). Conceptual tools for assessing experiments: Some well-entrenched confusions regarding the internal/external validity distinction. Journal of Economic Methodology, 18(3), 271–282.
Jiménez-Buedo, M., & Miller, L. M. (2010). Why a trade-off? The relationship between the external and internal validity of experiments. Theoria, 69, 301–321.
Kalkman, S., van Thiel, G. J. M. W., Grobbee, D. E., Meinecke, A-K., Zuidgeest, M. G. P., van Delden, J. J. M. on behalf of Work Package 3 of the IMI GetReal Consortium. (2016). Stakeholders’ views on the ethical challenges of pragmatic trials investigating pharmaceutical drugs. Trials, 17(1), 419.
Kalkman, S., van Thiel, G., van der Graaf, R., Zuidgeest, M., Goetz, I., Grobbee, D., et al. (2017). The social value of pragmatic trials. Bioethics, 31(2), 136–143.
Karanicolas, P. J., Montori, V. M., Devereaux, P. J., Schunemann, H., & Guyatt, G. H. (2009a). A new 'mechanistic-practical" framework for designing and interpreting randomized trials. Journal of Clinical Epidemiology, 62(5), 479–484.
Karanicolas, P. J., Montori, V. M., Devereaux, P. J., Schunemann, H., & Guyatt, G. H. (2009b). The practicalists’ response. Journal of Clinical Epidemiology, 62(5), 489–494.
Karanicolas, P. J., Montori, V. M., Schunemann, H., & Guyatt, G. H. (2009c). “Pragmatic” clinical trials: From whose perspective? Evidence-Based Medicine, 14(5), 130–131.
Kent, D. M., & Kitsios, G. (2009). Against pragmatism: On efficacy, effectiveness and the real world. Trials, 10, 48.
Koppenaal, T., Linmans, J., Knottnerus, J. A., & Spigt, M. (2011). Pragmatic vs. explanatory: An adaptation of the PRECIS tool helps to judge the applicability of systematic reviews for daily practice. Journal of Clinical Epidemiology, 64(10), 1095–1101.
Kravitz, R. L., Duan, N., & Braslow, J. (2004). Evidence-based medicine, heterogeneity of treatment effects, and the trouble with averages. Milbank Quarterly, 82(4), 661–687.
Kvols, L. K. (2005). Radiation sensitizers: A selective review of molecules targeting DNA and non-DNA targets. Journal of Nuclear Medicine, 46, 187S-190S.
Lilienfeld, S. O., Ritschel, L. A., Lynn, S. J., Cautin, R. L., & Latzman, R. D. (2014). Why ineffective psychotherapies appear to work: A taxonomy of causes of spurious therapeutic effectiveness. Perspectives on Psychological Science, 9(4), 355–387.
Lillie, E. O., Patay, B., Diamant, J., Issell, B., Topol, E. J., & Schork, N. J. (2011). The n-of-1 clinical trial: The ultimate strategy for individualizing medicine? Personalized Medicine, 8(2), 161–173.
Loudon, K., Treweek, S., Sullivan, F., Donnan, P., Thorpe, K. E., & Zwarenstein, M. (2015). The PRECIS-2 tool: designing trials that are fit for purpose. BMJ, 350, h2147.
Lucas, J. W. (2003). Theory-testing, generalization, and the problem of external validity. Sociological Theory, 21(3), 236–253.
Maclure, M. (2009). Explaining pragmatic trials to pragmatic policy-makers. CMAJ, 180(10), 1001–1003.
MacPherson, H., Thomas, K., Walters, S., & Fitter, M. (2001). The York acupuncture safety study: Prospective survey of 34 000 treatments by traditional acupuncturists. BMJ, 323(7311), 486–487.
Maningat, P., & Breslow, J. L. (2011). Needed: Pragmatic clinical trials for statin-intolerant patients. New England Journal of Medicine, 365(24), 2250–2251.
Marcellesi, A. (2015). External validity: Is there still a problem? Philosophy of Science, 82, 1308–1317.
Metge, C. J. (2011). What comes after producing the evidence? The importance of external validity to translating science to practice. Clinical Therapeutics, 33(5), 578–580.
Mohr, D. C., Ho, J., Hart, T. L., Baron, K. G., Berendsen, M., Beckner, V., et al. (2014). Control condition design and implementation features in controlled trials: A meta-analysis of trials evaluating psychotherapy for depression. Translational Behavioral Medicine, 4(4), 407–423.
Mommaerts, J. L., & Devroey, D. (2013). From “does it work?” to “what is ‘it’?”: Implications for voodoo, psychotherapy, pop-psychology, regular, and alternative medicine. Perspectives in Biology and Medicine, 56(2), 274–288.
Musial, F. (2019). Acupuncture for the treatment of pain - a mega-placebo? Frontiers in Neuroscience, 13, 1110.
Navarro, D. F., Tempini, N., & Teira, D. (2021). The trade-off between impartiality and freedom in the 21st Century Cures Act. Philosophy of Medicine, 2(1).
Nieuwenhuis, J. B., Irving, E., Rengerink, K. O., Lloyd, E., Goetz, I., Grobbee, D. E., et al. (2016). Pragmatic trial design elements showed a different impact on trial interpretation and feasibility than explanatory elements. Journal of Clinical Epidemiology, 77, 95–100.
Oxman, A. D., Lombard, C., Treweek, S., Gagnier, J. J., Maclure, M., & Zwarenstein, M. (2009a). A pragmatic resolution. Journal of Clinical Epidemiology, 62(5), 495–498.
Oxman, A. D., Lombard, C., Treweek, S., Gagnier, J. J., Maclure, M., & Zwarenstein, M. (2009b). Why we will remain pragmatists: Four problems with the impractical mechanistic framework and a better solution. Journal of Clinical Epidemiology, 62(5), 485–488.
Parkkinen, V. P., Wallmann, C., Wilde, M., Clarke, B., Illari, P., Kelly, M. P., et al. (2018). Evaluating Evidence of Mechanisms in Medicine: Principles and Procedures. Springer International Publishing.
Patsopoulos, N. A. (2011). A pragmatic view on pragmatic trials. Dialogues in Clinical Neuroscience, 13(2), 217–224.
Pawson, R. (2019a). The “pragmatic trial”: An essentially contested concept? Journal of Evaluation in Clinical Practice, 25(6), 943–954.
Pawson, R. (2019b). The shrinking scope of pragmatic trials: A methodological reflection on their domain of applicability. Journal of Clinical Epidemiology, 107, 71–76.
Pawson, R. (2019c). Pragmatic trials and implementation science: Grounds for divorce? BMC Medical Research Methodology, 19, 176.
Porzsolt, F., Rocha, N. G., Toledo-Arruda, A. C., Thomaz, T. G., Moraes, C., Bessa-Guerra, T. R., et al. (2015). Efficacy and effectiveness trials have different goals, use different tools, and generate different messages. Pragmatic and Observational Research, 6, 47–54.
Post, P. N., de Beer, H., & Guyatt, G. H. (2013). How to generalize efficacy results of randomized trials: Recommendations based on a systematic review of possible approaches. Journal of Evaluation in Clinical Practice, 19(4), 638–643.
Price, D., Musgrave, S. D., Shepstone, L., Hillyer, E. V., Sims, E. J., Gilbert, R. F., et al. (2011). Leukotriene antagonists as first-line or add-on asthma-controller therapy. New England Journal of Medicine, 364(18), 1695–1707.
Reiss, J. (2015). A pragmatist theory of evidence. Philosophy of Science, 82, 341–362.
Reiss, J. (2019). Against external validity. Synthese, 196(8), 3103–3121.
Rothwell, P. M. (2005). External validity of randomised controlled trials: “to whom do the results of this trial apply?” Lancet, 365(9453), 82–93.
Sackett, D. L. (2011). Explanatory and pragmatic clinical trials: A primer and application to a recent asthma trial. Polskie Archiwum Medycyny Wewnetrznej, 121(7–8), 259–263.
Sacristán, J. A., & Dilla, T. (2017). Generalizability in pragmatic trials. Journal of the American Medical Association, 317(1), 87–88.
Sacristán, J. A., & Dilla, T. (2018). Pragmatic trials revisited: Applicability is about individualization. Journal of Clinical Epidemiology, 99, 164–166.
Schwartz, D., & Lellouch, J. (1967). Explanatory and pragmatic attitudes in therapeutical trials. Journal of Chronic Diseases, 20(8), 637–648.
Sedgwick, P. (2014). Explanatory trials versus pragmatic trials. BMJ, 349, g6694.
Selker, H. P., Eichler, H. G., Stockbridge, N. L., McElwee, N. E., Dere, W. H., Cohen, T., et al. (2019). Efficacy and effectiveness too trials: Clinical trial designs to generate evidence on efficacy and on effectiveness in wide practice. Clinical Pharmacology and Therapeutics, 105(4), 857–866.
Senn, S. (2004). Individual response to treatment: Is it a valid assumption? BMJ, 329(7472), 966–968.
Shadish, W. R., Cook, T. D., & Campbell, D. T. (2002). Experimental and Quasi-Experimental Designs for Generalized Causal Inference. Houghton.
Stegenga, J. (2015). Measuring effectiveness. Studies in History and Philosophy of Science Part C: Studies in History and Philosophy of Biological and Biomedical Sciences, 54, 62–71.
Stegenga, J. (2018). Medical Nihilism. Oxford University Press.
Tassinari, M., & Roberti di Sarsina, P. (2015). Evidence-based medicine and acupuncture: Old bias for new perspectives in clinical context. Journal of Evaluation in Clinical Practice, 21(6), 1035–1037.
Thomas, K. J., MacPherson, H., Thorpe, L., Brazier, J., Fitter, M., Campbell, M. J., et al. (2006). Randomised controlled trial of a short course of traditional acupuncture compared with usual care for persistent non-specific low back pain. BMJ, 333(7569), 623.
Thorpe, K. E., Oxman, A. D., Treweek, S., & Furberg, C. D. (2010). Pragmatic trials are randomized and may use a placebo. Journal of Clinical Epidemiology, 63, 694–695.
Thorpe, K. E., Zwarenstein, M., Oxman, A. D., Treweek, S., Furberg, C. D., Altman, D. G., et al. (2009). A pragmatic-explanatory continuum indicator summary (PRECIS): A tool to help trial designers. Journal of Clinical Epidemiology, 62(5), 464–475.
Tosh, G., Soares-Weiser, K., & Adams, C. E. (2011). Pragmatic vs explanatory trials: The pragmascope tool to help measure differences in protocols of mental health randomized controlled trials. Dialogues in Clinical Neuroscience, 13(2), 209–215.
Tresker, S. (2022). Treatment effectiveness and the Russo–Williamson Thesis, EBM+, and Bradford Hill's viewpoints. International Studies in the Philosophy of Science, 1–28. https://doi.org/10.1080/02698595.2022.2054396
Treweek, S., & Zwarenstein, M. (2009). Making trials matter: Pragmatic and explanatory trials and the problem of applicability. Trials, 10, 37.
Tunis, S. R., Stryer, D. B., & Clancy, C. M. (2003). Practical clinical trials: Increasing the value of clinical research for decision making in clinical and health policy. Journal of the American Medical Association, 290(12), 1624–1632.
Tuzzio, L., & Larson, E. B. (2019). The promise of pragmatic clinical trials embedded in learning health systems. eGEMs, 7(1), 1–4.
Ware, J. H., & Hamel, M. B. (2011). Pragmatic trials—guides to better patient care? New England Journal of Medicine, 364(18), 1685–1687.
Williams, H. C., Burden-Teh, E., & Nunn, A. J. (2015). What is a pragmatic clinical trial? The Journal of Investigative Dermatology, 135(6), 1–3.
Zuidgeest, M. G., Goetz, I., & Grobbee, D. E. (2017). PRECIS-2 in perspective: what is next for pragmatic trials? Journal of Clinical Epidemiology, 84, 22–24.
Zwarenstein, M., & Treweek, S. (2009). What kind of randomized trials do we need? CMAJ, 180(10), 998–1000.
Zwarenstein, M., Treweek, S., Gagnier, J. J., Altman, D. G., Tunis, S., Haynes, B., et al. (2008). Improving the reporting of pragmatic trials: an extension of the CONSORT statement. BMJ, 337, a2390.
Zwarenstein, M., Treweek, S., & Loudon, K. (2017). PRECIS-2 helps researchers design more applicable RCTs while CONSORT Extension for Pragmatic Trials helps knowledge users decide whether to apply them. Journal of Clinical Epidemiology, 84, 27–29.
Acknowledgements
I am deeply indebted to three anonymous reviewers for this journal whose expertise, close reading of the manuscript, and highly insightful comments helped me avoid many mistakes, clarify my claims, and overall present a stronger argument. This work would not have been the same without their input. Many thanks to Bert Leuridan, Jonathan Fuller, Andreas De Block, and María Jiménez-Buedo for helpful reviews of the final or almost-final versions.
Funding
The author was supported by the Fonds Voor Wetenschappelijk Onderzoek – Vlaanderen (FWO; Research Foundation – Flanders; 1130819N) during the writing of this article.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflicts of interest
None.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Tresker, S. Treatment effectiveness, generalizability, and the explanatory/pragmatic-trial distinction. Synthese 200, 316 (2022). https://doi.org/10.1007/s11229-022-03517-0
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11229-022-03517-0