Background

With a growing emphasis on evidence-based practice in the clinical setting, health services are developing increasing numbers of clinical guidelines to direct more efficient clinical practice. Defined as “systematically developed statements to assist practitioner and patient decisions about appropriate health care for specific clinical circumstances”, clinical guidelines delineate appropriate and inappropriate care [1]. The overall goal of developing clinical guidelines is to improve the quality of care provided to patients by increasing clinical efficiency reducing inappropriate practices, thereby improving patient outcomes [2, 3]. They can also be a resource to inform patients regarding their clinicians’ decisions, and drive public policy by drawing attention to areas of need [4]. In addition, clinical guidelines can afford benefits to health care professionals as they can improve the quality of clinical decisions, support quality improvement activities and highlight gaps in the evidence, thus encouraging further research [4].

Despite the growing number of guidelines, the success of their use to change or introduce evidence-based health practices, otherwise known as implementation, is variable [5]. It has been acknowledged that the mere existence of clinical guidelines will not necessarily result in their implementation [6]. Research identifying determinants of guideline implementation across various settings has been extensively undertaken across the globe and has highlighted a myriad of factors that can impact the successful use of clinical guidelines [7,8,9,10,11,12]. These include individual health professional factors such as knowledge, awareness, self-efficacy, expectancy of positive outcomes, attitudes and intention [2, 7,8,9,10,11,12,13,14], as well as patient factors, such as applicability of guidelines to patients, patient preferences and behaviour [7, 8, 11]. Other factors such as organisational and environmental factors, as well as guideline complexity, can also potentially impact a clinician’s decision or ability to adhere to clinical guidelines [7,8,9,10,11,12, 14]. Therefore, guideline implementation strategies seek to address these potential barriers to compliance, and facilitate the application of new guidelines into practice [15].

Although research has focussed on evaluating which implementation strategies are the most effective in changing practice [16, 17], the evidence from these reviews were inconclusive, and often included studies that were most common in acute care or general practice settings [16, 17]. It has been acknowledged that these findings may not be applicable to other settings, such as the dental setting, which research suggests could be one of the clinical areas with lowest guideline compliance [18]. The dental setting has unique contextual factors that may impact the uptake of new guidelines, and these need to be considered when developing guideline implementation strategies [17, 19,20,21]. For example, the size of dental practices is highly variable, with some practitioners operating alone, and other practices involving more than 10 practitioners. This may have an influence on the success of guideline implementation, with evidence suggesting larger practices are more likely to comply with clinical guidelines [19]. With the push for dental practitioners to expand their scope of practice to include addressing health issues such as tobacco cessation [22], diabetes [23] and childhood obesity [24], it is essential to identify effective guideline implementation strategies specifically in the dental setting. With no studies synthesising the available findings in this area, the aim of this systematic review was to explore the effectiveness of various guideline implementation strategies in improving dental practitioners’ adherence to any clinical guidelines. This review specifically focussed on identifying changes in dental practitioners’ adherence, that is behaviour change, to any clinical guidelines they may have to follow, both dental and non-dental.

Methods

Research design

A systematic review was conducted according to the framework developed by Khan et al. [25] and reported according to the preferred reporting items for systematic reviews and meta-analyses (PRISMA) statement (Additional file 1) [26]. This approach was chosen as it appropriately addressed the study aims to systematically synthesise the existing evidence regarding the effectiveness of each type of implementation strategy. Several study authors are experienced in systematic review methodologies, having published multiple systematic reviews across various fields [27,28,29,30].

Searches

Databases including Scopus, CINAHL, MEDLINE, ProQuest, Embase, Cochrane, PsycINFO and Web of Science, as well as Google Scholar, were extensively searched from October 2018 to April 2019 (Additional file 2). With the assistance of a librarian, a combination of Boolean operators, truncations and Medical Subject Headings (MeSH) were used to develop individualised search strategies according to the indexing terms of each database. These search strategies incorporated key words such as guideline, recommendation, consensus, implementation, dissemination, translation, strategy, approach, intervention, and dentist. Upon identification of relevant articles, the reference lists and any cited references were manually searched for further relevant literature. This search strategy was deemed to meet the PRESS Checklist for Elements for the Peer Review of Electronic Search Strategies [31] (Additional file 3).

Study inclusion and exclusion criteria

All articles that were relevant to the study aims and published in the searched databases from inception up to 7 April 2019 were eligible for inclusion in this review. Inclusion and exclusion criteria have been presented in Table 1 according to the Population, Intervention, Control, Outcome (PICO) framework. Included studies followed experimental or quasi-experimental designs, including randomised controlled trials, pretest-posttest studies, interrupted time series studies and non-equivalent groups studies. Observational studies such as cross-sectional surveys, case-control studies and cohort studies were excluded from this review.

Table 1 Inclusion and exclusion criteria according to the PICO framework

Terminology

There are numerous health professionals that provide dental care to individuals. Among the most well-known are dentists, who can practice in general dentistry, or go on to specialise in various areas, including endodontics, periodontics, orthodontics and special needs dentistry [32,33,34]. However other dental practitioners work as part of the dental team to provide direct care to individuals, including dental assistants, dental therapists and dental hygienists [33, 35, 36]. In this review, the term dental practice was used to encompass all of the above professions, and dental practitioner encompassed the care providers working in these professions. Previous systematic reviews have commonly categorised implementation strategies into single interventions, which utilise a sole strategy, such as audit and feedback, education or reminders, and multifaceted interventions, which utilise multiple strategies concurrently [16, 17]. These terms were also used to classify implementation strategies in this review.

Study quality assessment

The risk of bias and quality of each study was assessed using the Joanna Briggs Institute critical appraisal tools for both randomised controlled trials and quasi-experimental designs [37, 38] (Tables 2 and 3). This was initially performed by the first author and then independently reviewed by a second author (LR or DM). In the instance of any discrepancies in assessment, a third author (LR or DM) was involved to achieve consensus. This third author was not able to identify who provided each assessment. The appraisal tools were used to develop a score for each article, as a percentage of the number of met criteria out of the total number of applicable criteria. Through consensus among the authors, cut-off values were established prior to scoring, whereby studies with a score of less than 30% were excluded from analysis, 30–59% were considered poor quality, 60–79% were considered moderate quality and greater than 80% were considered high quality [54].

Table 2 Summaries of included studies
Table 3 Effectiveness of each implementation strategy

Screening

A systematic screening process was conducted whereby initial search results were screened for eligibility first by title, then by abstract, and finally by full text. Titles and abstracts of identified records were screened independently by two authors (AV and DM), and a third author (AG) was invited to achieve consensus when discrepancies arose. For records that were not excluded by title or abstract, full texts were obtained and independently screened by the same investigators (AV and DM), and consensus was achieved by a third investigator (AG) where required.

Data extraction strategy

Data from included studies was extracted by the first author using a data extraction tool (Additional file 4), which was validated through consensus with the research team, and included fields such as author, year, location, aims, study design, population and eligibility criteria, intervention/implementation strategies used and outcomes, measured by change in proportion of guideline-adherent practice where possible.

Data synthesis and presentation

Due to high heterogeneity of included interventions, as well as the small number of articles meeting the eligibility criteria within each type of intervention, it was decided that a narrative synthesis of study findings would be undertaken. Studies were categorised by type of implementation strategy according to the Effective Practice and Organisation of Care (EPOC) taxonomy of health systems interventions [55]. This taxonomy was systematically and iteratively developed by the Cochrane EPOC Review Group to allow the classification of health systems interventions into categories based on conceptual or practical similarities [56]. Throughout the development process of this taxonomy, it was applied to various reviews of health systems interventions of high relevance to developing countries, making this a tool that could be relevant to a variety of settings [56]. In contrast, many other taxonomies, such as the ERIC taxonomy, only included panellists from specific geographic locations in their development, which could limit their applicability to studies conducted elsewhere [57]. Thus, the EPOC taxonomy was deemed an appropriate tool for the classification of the interventions used in this review. A process was adopted whereby initial classification was conducted by the first author (AV) who examined the interventions described within each study allocated these interventions to the EPOC category with the best matching definition. This categorisation was peer checked by other authors (LR, DM and AG). In the case where interventions consisted of components from multiple EPOC categories, these were classified as multifaceted interventions. All data was presented as a qualitative review.

Registration

This systematic review was registered with the International Prospective Register of Systematic Reviews (PROSPERO), registration ID CRD42018093023.

Results

Search results

All database searches returned results, yielding a total of 3493 records. Following the removal of 1652 duplicates, titles and abstracts of records were screened, resulting in the exclusion of a further 1772 articles. Of the 96 full-text articles assessed, reasons for exclusion included the incorrect intervention (n = 19), incorrect study design (n = 14), not in the dental setting (n = 4), incorrect outcome (n = 2), the control group was not exposed to the guidelines (n = 10), no English full text available (n = 1) and no formal guidelines were in place (n = 3). A total of 15 studies were included in this review [39,40,41,42,43,44,45,46,47,48,49,50,51,52,53]. See Fig. 1 for further detail regarding the search and screening process.

Fig. 1
figure 1

PRISMA flow chart for the systematic review

Study characteristics

Of the included studies, seven studies were from the UK [39,40,41,42, 48, 50, 52], six studies were from the USA [43,44,45,46, 51, 53], one study was from Finland [47] and one study was from the Netherlands [49]. Included studies were published between 2004 and 2018. A total of seven included studies were randomised controlled trials (RCTs) [39, 42, 44,45,46,47, 53], of which six were cluster randomised [39, 42, 44, 46, 47, 53] and one was individually randomised [45]. The remaining eight studies followed quasi-experimental designs, with seven being pretest-posttest quasi-experimental studies [40, 41, 43, 48, 49, 51, 52] and one being a non-equivalent groups design [50]. Out of all 15 studies, six involved implementation of interdisciplinary guidelines related to tobacco cessation into the dental setting [43, 44, 46, 47, 49, 51], and the rest explored the implementation of general dental guidelines [39,40,41,42, 45, 48, 50, 52, 53]. Implementation strategies employed in each study were classified into the following categories: audit and feedback (n = 4) [39,40,41,42], reminders (n = 4) [42,43,44,45], education (n = 3) [46,47,48], patient-mediated interventions (n = 2) [45, 49], pay for performance (n = 1) [50] and multifaceted interventions (n = 5) [42, 47, 51,52,53]. Three studies [42, 45, 47] were classified into multiple categories, as they included both multifaceted and single intervention arms. Details of the included studies can be seen in Table 2, and a summary of the effectiveness of implementation strategies in each category can be seen in Table 3.

Quality assessment

The quality assessments of included studies are shown in Tables 4 and 5. All 15 studies received an acceptable quality assessment (score of 30% or higher) for inclusion in the review. Of the seven included RCTs, two were assessed to be of high quality [45, 46], four were of moderate quality [39, 42, 44, 47] and one was of poor quality [53] (Table 4). On average, RCTs had a score of 72.1%. The most common areas of methodological weakness in included RCTs were lack of blinding of both participants and those administering the intervention. This was expected due to the nature of the interventions making blinding difficult and unfeasible at times. Despite this, one RCT was able to implement a study design that permitted blinding [46]. Three of the eight included quasi-experimental studies was of high quality [49,50,51], with four being of moderate quality [40, 41, 43, 52], and one being of poor quality [48] (Table 5). Quasi-experimental studies had an average score of 74.08%. Weaknesses among these studies included limited use of control groups and lack of reliability of measures.

Table 4 Quality assessment of included randomised controlled trials
Table 5 Quality assessment of included quasi-experimental studies

Single interventions

The 12 studies classified into single intervention categories involved interventions classified as audit and feedback (n = 4) [39,40,41,42], reminders (n = 4) [42,43,44,45], education (n = 3) [46,47,48], patient-mediated interventions (n = 2) [45, 49] and pay for performance (n = 1) [50].

Audit and feedback

Four studies explored the effectiveness of audit and feedback interventions on the implementation of guidelines in dental practice [39,40,41,42]. This category included two cluster RCTs [39, 42] and three pretest-posttest quasi-experimental studies [40, 41, 49]. The study by Zahaboyoun et al. found audit and feedback to significantly increase compliance with both metronidazole and amoxicillin prescription guidelines [40]. Similarly, the study by Elouafkaoui et al. found a significant decrease in antibiotic prescription following an audit and feedback intervention [39]. Simons and Williams reported mixed results, where audit and feedback increased adherence to some, but not all, endodontic guidelines; however, the statistical significance of these changes was not evaluated [41]. Conversely, the cluster RCT by Bahrami et al. [42] showed no significant improvement in guideline adherence to unerupted third molar guidelines following an audit and feedback intervention.

Reminders

An additional four studies investigated the effect of reminder strategies on guideline implementation [42,43,44,45]. Of these studies, three were RCTs [42, 44, 45] and one was a quasi-experimental study [43]. The quasi-experimental study by Montini et al. found that a computer decision support system significantly improved not only tobacco screening rates, but also rates of advising, referring and prescribing nicotine replacement therapy for tobacco users [43]. Similarly, the computer decision support system used by Rindal et al. also significantly improved clinicians’ adherence to tobacco cessation guidelines [44]. However, while increasing guideline adherence for the first 6 months, the study by Friction et al. found that computer alerts to providers did not cause sustained change in adherence to guidelines for patients with medically complex conditions [45]. In addition, the computer-aided learning and decision support strategy used by Bahrami et al. showed no significant effect on guideline adherence [42].

Education

Three studies evaluated the effectiveness of education strategies on improving guideline adherence [46,47,48], two of which were RCTs [46, 47]. The cluster RCT by Walsh et al. found that although both self-study and workshop-based education strategies improved self-reported adherence to tobacco cessation guidelines, there was higher reported adherence in the workshop intervention [46]. In addition, the pretest-posttest quasi-experimental study by Chopra et al. found that extensive training and education caused an increase in adherence to antimicrobial prescribing guidelines, despite having no statistical evaluation of this effect [48]. However, Amemori et al. found mixed results, concluding that an education package consisting of lectures, interactive sessions, multimedia demonstrations and a role play session resulted in a significant increase in provision of tobacco cessation counselling, but not tobacco prevention counselling [47].

Patient-mediated interventions

Two studies explored the use of patient-mediated interventions as guideline implementation strategies. Rosseel et al. found that patient-mediated feedback increased the proportion of patients reporting having received guideline-adherent information, advice and support regarding tobacco cessation; however, this declined after 6 months [49]. In addition, the study by Friction et al. found that notifying patients to ask for a review of care during their visit caused an increase in adherence to guidelines for patients with medically complex conditions in the first 6 months; however, this was not sustained [45].

Pay for performance

Finally, one non-equivalent groups quasi-experimental study by Gnich et al. evaluated the effectiveness of fee-for-service as a single intervention to improve guideline implementation [50]. The investigators found that guideline-adherent fluoride varnish application rates increased rapidly among practitioners that had received fee-for-service during the intervention, when compared to practitioners who had already been receiving fee-for-service prior to the intervention.

Multifaceted interventions

A total of five studies tested the effectiveness of a multifaceted intervention and were therefore included in this category [42, 47, 51,52,53]. This category included three RCT studies [42, 47, 53], with the remaining two studies following pretest-posttest quasi-experimental study designs [51, 52]. The multifaceted interventions utilised in these studies varied and comprised of a combination of two or more implementation strategies including education, audit and feedback, fee-for-service or decision support (see Table 2). All five studies in this category involved education in combination with other strategies as part of the multifaceted intervention. Four studies combined education with reminders [42, 51,52,53], two utilised audit and feedback [42, 51], and one employed a pay for performance strategy [47].

Although results seen within each study varied, three studies [47, 51, 53] showed a significant increase in some component of guideline implementation following the multifaceted intervention. These studies [47, 51, 53], which explored the implementation of multicomponent tobacco cessation guidelines, highlighted mixed results, with multifaceted interventions causing a significant improvement in adherence to some guideline components, but no change was seen in other components. Similar to their single intervention, Amemori et al. [47] highlighted that their education plus fee-for-service intervention caused a significant increase in tobacco cessation counselling; however, no change was seen in tobacco prevention counselling. Similarly, the multicomponent intervention utilised by Shelley et al. [51], which involved a chart system, training, protocols and referral feedback, resulted in a significant increase in providing advice, assessments and assistance to tobacco users; however, rates of tobacco use screening remained the same. Finally, the local consensus process, multi-modal reminders for patients and practitioners and interactive educational workshop utilised by Isaacson et al. caused no significant change in patient-reported adherence to oral cancer screening guidelines aside from patient agreement with the statement “the dentist/hygienist told me that I was being screened for oral cancer” [53].

The multifaceted interventions in the remaining two studies [42, 52] did not exhibit significant changes in adherence for various reasons. One study by Afuakwah and Welbury [52] indicated that a pro forma, aide memoire and staff training had a positive effect on adherence to guidelines regarding documentation of caries risk assessments. However, this study did not compute any inferential statistics; therefore, the significance of this change could not be ascertained. The study by Bahrami et al. did compute inferential statistics to determine the effect of a multifaceted intervention involving computer-aided learning with decision support plus audit and feedback on adherence; however, no significant improvement was detected.

As seen in Fig. 2, overall, a slightly higher proportion of studies involving multifaceted interventions reported improvements in guideline adherence when compared to single interventions. Among single interventions, studies classified into the “reminders” and “education” categories reported larger improvements in guideline adherence. In addition, all studies involving interdisciplinary guidelines reported some effect of the implementation strategies, on guideline adherence, whereas over 20% of studies involving dental guidelines reported no effect.

Fig. 2
figure 2

A diagrammatic representation of the effect of each study intervention on improvement in adherence, grouped by intervention category and guideline type

Discussion

With existing research highlighting dental practice to have some of the lowest guideline compliance in the health sector, it is evident that the implementation of clinical guidelines in this setting remains a challenge [18,19,20,21]. As no published reviews summarise the evidence regarding implementation strategies in the dental setting, this review aimed to evaluate the current research in this area and identify effective guideline implementation strategies. This review identified a total of 15 studies that investigated the effectiveness of implementation strategies in the dental setting, with 13 studies investigating single implementation strategies and 5 studies involving multifaceted implementation strategies. There were studies classified across the categories of audit and feedback [39,40,41], reminders [43, 44], education [46,47,48], patient-mediated interventions [49], pay for performance [50] and multifaceted interventions [47, 51,52,53] that reported a significant increase in adherence to some or all guideline components.

Within the single interventions, the effectiveness of audit and feedback interventions was variable, with the three successful studies showing slightly smaller effect sizes than studies in other categories [39,40,41]. Systematic reviews of implementation strategies in other health care settings have also reported audit and feedback interventions to have variable outcomes on guideline adherence [17, 58]. Ivers et al. [58] reported that audit and feedback interventions may be more effective when baseline performance is low; when the source of feedback is a supervisor or colleague; when they are provided more than once; when they are delivered in both verbal and written formats; and when they include both explicit targets and action plans. Hypotheses developed by Colquhoun et al have reinforced these findings, also suggesting that involvement of recipients in the design of the audit and feedback process could improve the effectiveness of these interventions [59].

Three of four studies within this review using reminders reported guideline adherence to improve following intervention [43,44,45], with higher improvements indicating reminders may be a more effective strategy in the dental setting than audit and feedback. Other reviews have also highlighted reminders as a promising strategy to improve guideline adherence [17, 60,61,62,63], and one review has even highlighted the potential for this strategy to positively impact on patient outcomes [64]. It is suggested that computerised decision support systems may be more effective in improving compliance; however, they may also lengthen consultation times and may be more stressful for clinicians to use [17]. Further, studies evaluating reminder systems developed specifically by the organisations where implementation will occur have shown larger effects on improving practice than the adoption of existing reminder systems [62].

Similar to other health care settings [17, 60, 61], this review found education to be an effective implementation strategy, with three of four included studies in this category reporting improvement in guideline adherence [46,47,48]. However, the intervention consisting of passive education strategies, that is, in the “educational materials” EPOC category, resulted in lower effects when compared with other interventions [46]. This is consistent with other literature, which highlight passive education strategies to be largely unsuccessful in improving guideline adherence [17].

Only one out of the two studies in the patient-reported feedback demonstrated a significant improvement in guideline adherence, indicating some potential variability in the effectiveness of this implementation strategy; however, more research is required to confirm this [49]. Reviews in other health care settings have shown these types of interventions to be more effective [17, 65].

Although the sole study investigating a pay-for-service intervention in this review reported a moderate significant effect on guideline adherence [50], interestingly, this type of implementation strategy was less often explored in other settings with inconclusive results found regarding its effectiveness [17, 60, 61]. The effectiveness of this strategy was hypothesised to be due to its ability to normalise or validate the responsibility of dental care practitioners in performing the service.

Of the five studies exploring multifaceted interventions in this review, four highlighted positive effects on guideline adherence [47, 51,52,53]. Previous studies in other settings have shown huge success of multifaceted interventions, showing them to be much more effective than single implementation strategies alone, making them one of the most frequently researched and recommended approaches for guideline implementation [17, 61]. Despite this, there is little evidence regarding the best components to include in these strategies, as there does not seem to be a relationship between number or type of components used and effectiveness of the strategy [17].

Despite this evidence of positive effects across all types of implementation strategies, some studies only found partial effects on implementation, and further there were some studies that did not show evidence of effect on guideline adherence. Many of the studies that showed only partial effects had some characteristics in common, the first of which was complexity of guidelines. All five studies that only found partial improvements in guideline adherence involved complex guidelines with multiple criteria for compliance, often combined with interdisciplinary practice [41, 46, 47, 51, 53]. This is supported by previous research in other settings, with previous reviews highlighting guideline complexity as the single most frequently cited guideline characteristic affecting guideline implementation [66].

Further, the two studies that highlighted no effect of their implementation strategies on guideline adherence had some methodological limitations that may have impacted on the effects of the implementation strategies. Firstly, Bahrami et al. acknowledged a high baseline compliance in their study, which may have produced a ceiling effect, meaning that no greater improvement was possible following the intervention [42]. In addition, this is the oldest study included in this review, conducted in 2004, yet it heavily involved the use of computer decision support systems as part of the implementation strategies, despite the fact that computer systems were not commonly used in patient care at that time. This is reinforced by the fact that the authors specifically developed the computer software package for the purpose of this study, contained on a stand-alone laptop [42]. As a result, clinicians would be required to specifically use this laptop when decision support was required, rather than the decision support being embedded into computer patient records as was commonly done in more recent studies [43, 44], potentially reducing compliance to this decision support intervention. The second study that did not find overall significant improvements was Friction et al., which used the frequency of accessing an online computer decision support tool as an indicator of guideline adherence [45]. The limitation of this proxy, acknowledged by the authors, was that there was the potential that the more the decision support tool was used, the more clinicians may begin to learn the guidelines, and as a result may not have needed to refer to the decision support repeatedly [45]. This is reinforced by the trend in the data that increases in guideline adherence were seen at 6 months, but regressed to baseline at 12 months [45].

In summary, this study has highlighted that implementation strategies such as audit and feedback, reminders, education, patient-mediated interventions, pay for performance and multifaceted interventions have all had some success in the dental setting, with reminders, education, pay for performance and multifaceted interventions showing the most promise. Further research is required to provide more high-quality evidence regarding the effectiveness of each strategy type and gain an understanding of aspects of each type of strategy that may increase the success of guideline implementation. Despite some promising findings in this review, it has several limitations that should be considered in the interpretation of results. Firstly, the quality of evidence identified by this review varied, with most of the 15 included studies being of poor or moderate quality, and only five studies being deemed high quality. In addition, two included studies did not conduct statistical tests for the positive changes they identified following their implementation strategies, one of which only involved four clinicians, further impacting the interpretability of the study findings [48, 52]. Further, due to the heterogeneity of this data, resulting from the variability of implementation strategies used in each study, meta-analysis was not possible in this review, and as a result, the effects of each strategy could not be quantitatively compared. A large number of included studies were published in the UK and USA, which could limit generalisability of findings. In addition, although rigorous search strategies were used, there is a chance that not all relevant studies were identified, and due to a paucity of research in this field, relatively few studies were ultimately included. Finally, this study excluded studies for which full texts or English translations could not be obtained, which may have potentially introduced some bias into the results of the review. Nonetheless, this study is the first systematic review of implementation strategies in the dental setting and has provided significant insight into which strategies may be most effective for the implementation of guidelines in this sector.

Conclusions

This study has confirmed findings in other settings that implementation strategies such as audit and feedback, reminders, education, patient-mediated interventions and multifaceted interventions may be effective in improving guideline adherence in the dental setting. It has highlighted that interventions such as education, reminders and multifaceted interventions may be the most effective in this setting, and it has identified pay for performance as a potentially effective strategy that has previously been inconclusive in other settings. Although some included studies showed equivocal findings or no effects on guideline adherence, key strategies were identified that could be utilised in the implementation of any future dental guidelines, as well as considerations that should be taken into account in the use of these strategies. This information is particularly relevant in light of the increased need and focus on role expansion of dental professionals into other areas like childhood obesity. This review highlights the need for further, high-quality research to be conducted in this setting, to gain a better understanding of the conditions under which each strategy works best. Increasing the number of studies using rigorous methods within each strategy category will allow heterogeneity of findings to be reduced, therefore enabling meta-analyses to be conducted.