The recent paper in Trials by Porter and colleagues highlights the utility of applying a critical realism approach in randomised trials, an approach central to the Medical Research Council’s (MRC) Framework for the Development and Evaluation of Complex Healthcare Interventions. The MRC framework offers a pragmatic step towards a more open systems approach that bridges randomised evaluation with social context and human agency in an effort to improve the generalisability of trial outcomes.
The MRC framework has contributed to the proliferation of a more open systems approach in health research; however, the broader acceptance of the realist approach to health research does not seem to be emulated by norms in research fund allocation, which largely prioritises laboratory-based research.
This commentary is simply a plea, to those who make the strategic decisions regarding allocation of research funding, to support all phases of health intervention research in complex systems that contribute to the development of effective, translational and sustainable interventions in the promotion of health.
The dissemination over a decade ago of the Medical Research Council’s (MRC) Framework for the Development and Evaluation of Complex Healthcare Interventions , and appeals by researchers for more research to be conducted on interventions in complex systems  has contributed to the proliferation of research in more open systems, i.e. in ‘real-world’ settings (for an example, see the INCLUSIVE study – ). The MRC framework advocates a mixed-method approach, including: qualitative research – such as process evaluation of implementation of an intervention and of contextual factors that could lead to variability in outcomes ; causal modelling approaches, that can provide information on the process and outcome of an intervention and that may lead to design changes prior to implementation, e.g. ; and a range of experimental methods for evaluating effectiveness, with preference given to randomised designs (randomised controlled trials – RCTs), wherever possible or appropriate, for optimisation of internal validity . The mixed-method approach is an attempt to supplement the randomised trial with procedures embedded in realism that take the complexity of social context and human volition into account, and to shift the focus from identifying causality to explaining causality, or in other words, to explaining how intervention components interact with different populations and in different contexts . Moore and Evans  demonstrate the importance of context by describing how an intervention nearly two decades ago, that effectively reduced smoking in young people by prohibiting smoking on school premises and thereby challenging norms surrounding smoking behaviour, may not prove as effective today. Norms surrounding smoking that existed 15–20 years ago (e.g. high prevalence rates and social acceptance) have changed, and consequently, so have the mechanisms that predict smoking behaviour in young people .
The combination of positivist and realist approaches found in the MRC framework, which Porter et al.  and others  refer to as realist RCTs, has given rise to debate [7, 9,10,11,12,13]. Central to the debate is whether randomised and realist evaluation can be consolidated. While critics argue that the realist RCT design found in the MRC framework does not fully take complexity into account , the MRC framework does offer a pragmatic step towards a more open systems (real-world) approach that bridges, to a degree, randomised evaluation with social context and human agency . The recent paper in Trials by Porter and colleagues highlighting the utility of applying a critical realism approach  and the ongoing debate exploring the utility between realist and randomised evaluations [9, 12] are welcome and important steps towards the development and refinement of a usable open systems approach.
A plea for more research funding aimed at supporting health interventions in complex systems
While incremental steps are being made as evidenced by the MRC complex framework, the broader adoption of a more open systems approach in health research does not seem to have penetrated the orthodoxy of applying a traditional positivist epistemological and ontological (bio-medical) approach to evaluating research and, by extension, to deciding how research funds are allocated [15, 16]. For example, an overview of research-fund allocation from the Medical Research Council, The Wellcome Trust, British Heart Foundation, and Cancer Research UK, showed that across the four research-funding agencies, 85% of funds on average were allocated to laboratory-based research . A recent review of funding trends within diabetes research illustrated that the proportion of funded studies that included social context and human volition was small compared to the proportion of funded bio-medical studies, with an estimated mean ratio of 17:1 . The uptake of a non-linear, or real-world conception of causality by funding agencies seems, therefore, to be a slow burn. There is no doubt that the closed linear system approach, so pervasive in bio-medical and health research, has been effective in solving health-related and other issues. However, researchers, including those within bio-medicine, are increasingly asking whether the documented efficacy of an intervention, gained from identifying causality in a closed part of a larger or open system (traditional reductionist approach), is robust enough to reproduce the recorded effects reliably in real-world settings [18, 19]; or if successionist models of causality alone are adequate in solving the complex health issues of our time [20,21,22,23].
It is estimated that as much as 85% of research investment can be categorised as waste (i.e. not benefitting society/patients) due largely to avoidable weaknesses in research design and production [24,25,26,27]. Lack of external validity in the experimental research designs employed is one of the main reasons why laboratory-based and randomised clinical trial outcomes often fail to translate into benefits for the patient, as the robustness of probabilities in different contexts (what works best for whom and in which setting) is uncertain [12, 18, 19, 21, 28]. Laboratory and clinical trials that do not take higher-level concepts, such as social context and human volition, into account are, therefore, vulnerable when attempting to replicate intervention effects in real-world settings. As stated by Rothwell: “Government expenditure should provide value for money, and medical research is no exception” .
The MRC framework is an attempt to increase the external validity of evaluations while still preserving internal validity. The availability of more research funds that support health interventions and practice-orientated research in complex systems may result in more robust and sustainable interventions and lead to reduced waste in research investment.
Since the introduction of the MRC framework, complex interventions are increasingly being developed in an attempt to solve the continually expanding burden of health-related issues. This commentary is, therefore, a plea to policy-makers, and those who make the strategic decisions regarding allocation of research funding, to support to a greater degree all phases of real-world trials that take into account social context and human volition as advocated by the MRC framework, rather than to automatically allocate research funds on the basis of received wisdom.
Medical Research Council
Randomised controlled trials
Craig P, Dieppe P, Macintyre S, Michie S, Nazareth I, Petticrew M. Developing and evaluating complex interventions: new guidance. Medical Research Council: Prepared on behalf of the Medical Research Council; 2006. https://www.mrc.ac.uk/documents/pdf/complex-interventions-guidance/
Rickles D. Causality in complex interventions. Med Health Care Philos. 2009;12:77–90. doi:10.1007/s11019-008-9140-4.
Bonell C, Allen E, Christie D, Elbourne D, Fletcher A, Grieve R, et al. Initiating change locally in bullying and aggression through the school environment (INCLUSIVE): study protocol for a cluster randomised controlled trial. Trials. 2014;15:1–14.
Moore G, Audrey S, Barker M, Bond L, Bonell C, Hardeman W, et al. Process evaluation of complex interventions: Medical Research Council guidance. BMJ. 2015;350:h1258. doi:10.1136/bmj.h1258.
Hardeman W, Sutton S, Griffin S, Johnston M, White A, Wareham NJ, et al. A causal modelling approach to the development of theory-based behaviour change programmes for trial evaluation. Health Educ Res. 2005;20:676–87.
Craig P, Dieppe P, Macintyre S, Michie S, Nazareth I, Petticrew M. Developing and evaluating complex interventions: the new Medical Research Council guidance. BMJ. 2008;337:a1655.
Porter S, McConnell T, Reid J. The possibility of critical realist randomised controlled trials. Trials. 2017;18:133. doi:10.1186/s13063-017-1855-1.
Moore GF, Evans RE. What theory, for whom and in which context? Reflections on the application of theory in the development and evaluation of complex population health interventions. SSM-Population Health. 2017;3:132–5.
Bonell C, Fletcher A, Morton M, Lorenc T, Moore L. Realist randomised controlled trials: a new approach to evaluating complex public health interventions. Social science medicine. 2012;75:2299–306. doi:10.1016/j.socscimed.2012.08.032.
Van Belle S, Wong G, Westhorp G, Pearson M, Emmel N, Manzano A, et al. Can ‘realist’ randomised controlled trials be genuinely realist? Trials. 2016;17:313. doi:10.1186/s13063-016-1407-0.
Jamal F, Fletcher A, Shackleton N, Elbourne D, Viner R, Bonell C. The three stages of building and testing mid-level theories in a realist RCT: a theoretical and methodological case-example. Trials. 2015;16:466. doi:10.1186/s13063-015-0980-y.
Marchal B, Westhorp G, Wong G, Van Belle S, Greenhalgh T, Kegels G, et al. Realist RCTs of complex interventions—an oxymoron. Social Science Medicine. 2013;94:124–8. doi:10.1016/j.socscimed.2013.06.025.
Bonell C, Warren E, Fletcher A, Viner R. Realist trials and the testing of context-mechanism-outcome configurations: a response to Van Belle et al. Trials. 2016;17:478. doi:10.1186/s13063-016-1613-9.
Fletcher A, Jamal F, Moore G, Evans R, Murphy S, Bonell C. Realist complex intervention science: applying realist principles across all phases of the Medical Research Council framework for developing and evaluating complex interventions. Evaluation. 2016;22:286–303. doi:10.1177/1356389016652743.
Rutter H, Glonti K. Towards a new model of evidence for public health. Lancet. 2016;388:S7. doi:10.1016/S0140-6736(16)32243-7.
Jones A, Vallis M, Cooke D, Pouwer F. Review of research grant allocation to psychosocial studies in diabetes research. Diabetic Med. 2016;33:1673–6. doi:10.1111/dme.13255.
Rothwell P. Funding for practice-oriented clinical research. Lancet. 2006;368:262–6. doi:10.1016/S0140-6736(06)69010-7.
Rothwell P. Commentary: external validity of results of randomized trials: disentangling a complex concept. Int J Epidemiol. 2010;39:94–6. doi:10.1093/ije/dyp305.
Rothwell P. External validity of randomised controlled trials: ‘to whom do the results of this trial apply?’. Lancet. 2005;365:82–93. doi:10.1016/S0140-6736(04)17670-8.
Van Regenmortel MHV. Reductionism and complexity in molecular biology. EMBO Rep. 2004;5:1016–20.
Federoff H, Gostin L. Evolving from reductionism to holism: is there a future for systems medicine? JAMA: the Journal of the American Medical Association. 2009;302:994–6. doi:10.1001/jama.2009.1264.
Strange K. Revisiting the Krogh Principle in the post-genome era: Caenorhabditis elegans as a model system for integrative physiology research. J Exp Biol. 2007;210:1622–31. doi:10.1242/jeb.000125.
Levi-Montalcini R, Calissano P. The scientific challenge of the 21st century: from a reductionist to a holistic approach via systems biology. BMC Neuroscience. 2006;7:S1. doi:10.1186/1471-2202-7-S1-S1.
Chalmers I, Glasziou P. Avoidable waste in the production and reporting of research evidence. Lancet. 2009;374:86–9. doi:10.1016/S0140-6736(09)60329-9.
Yordanov Y, Dechartres A, Porcher R, Boutron I, Altman D, Ravaud P. Avoidable waste of research related to inadequate methods in clinical trials. BMJ. 2015;350:h809. doi:10.1136/bmj.h809.
Ioannidis JPA, Greenland S, Hlatky M, Khoury M, Macleod M, Moher D, et al. Increasing value and reducing waste in research design, conduct, and analysis. Lancet. 2014;383:166–75. doi:10.1016/S0140-6736(13)62227-8.
Heneghan C, Goldacre B, Mahtani K. Why clinical trial outcomes fail to translate into benefits for patients. Trials. 2017;18:122. doi:10.1186/s13063-017-1870-2.
Steckler A, McLeroy K. The importance of external validity. Am J Public Health. 2008;98:9–10. doi:10.2105/AJPH.2007.126847.
No sources of funding to declare.
Availability of data and materials
Not applicable (no empirical data involved)
Ethics approval and consent to participate
Not applicable (as this is not an empirical study, it involved no participants and does not require ethics approval).
Consent for publication
Not applicable (no individual data included).
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Hansen, A.B.G., Jones, A. Advancing ‘real-world’ trials that take account of social context and human volition. Trials 18, 531 (2017). https://doi.org/10.1186/s13063-017-2286-8
- Complex interventions
- External validity
- Research funding