Introduction

Computerized cognitive training (CCT) is receiving increasing attention as a potential approach to prevent cognitive decline and dementia in older adults. Meta-analyses of published trials have indicated that CCT can improve cognitive functioning immediately in the post-training period in cognitively healthy older adults (Kelly et al., 2014; Lampit et al., 2014b). Evidence supporting the long-term benefit of cognitive training is limited as few studies have included long-term follow-up. Findings from the landmark ACTIVE trial suggest that the gains in cognitive function, particularly from training that focused on speed of processing, may translate into sustained maintenance of daily functioning, reduced risk of driving cessation and reduced incidence of dementia 10 years after training (Edwards et al., 2016, 2017; Rebok et al., 2014; Ross et al., 2016).

Lampit et al. (2014b, 2020) have published the two most comprehensive reviews to date of the effectiveness of CCT trials in cognitively healthy older adults (aged ≥ 60 years). Their first meta-analysis of 51 trials indicated a small but statistically significant effect on global cognitive functioning favouring CCT over the control group (Hedges’ g = 0.22, 95% confidence interval [CI] = 0.15–0.29) (Lampit et al., 2014b). An update of this review included a total of 90 papers and showed similar efficacy (Hedges’ g = 0.18, CI = 0.14–0.23) (Lampit et al., 2020). The majority of the trials included in these reviews were tightly controlled, group-based trials with relatively small numbers of participants (n < 50), and most employed high intensity protocols to deliver the CCT intervention, particularly in terms of the time commitment required for participants to complete the intervention (Lampit et al., 2014b, 2020).

To make CCT feasible to implement for a wide audience, translation of the evidence from these lab-based trials into large-scale community interventions is needed. This requires greater understanding about the precise nature and delivery of CCT intervention that can achieve maximum benefits for brain health. Motivation, engagement, and expectations are believed to influence the effectiveness of CCT (Boot et al., 2013; Foroughi et al., 2016). These factors can be influenced by behaviour change techniques (BCTs), which are the ‘active ingredients’ in an intervention designed to bring about the desired behaviour change (Michie et al., 2015). In the context of CCT, BCTs are strategies put in place to encourage adherence to the cognitive training protocol, namely, the frequency and time spent training. For examples of BCTs used in the context of CCT, please refer to the supplemental file. BCTs may influence efficacy of interventions indirectly via their influence on adherence. BCTs may also influence efficacy directly by enhancing participant’s engagement while training. Meta-analyses of lifestyle behaviour interventions have indicated that BCTs related to social support, goal setting, and self-monitoring are associated with improved adoption of diet and physical activity behaviours, while problem-solving techniques may be important for supporting sustained long-term behaviour change (Cradock et al., 2017; Lara et al., 2014; Olander et al., 2013).

To date, little attention has been given to the potential role of BCTs in the adoption and maintenance of cognitive training behaviours. It is unclear to what extent BCTs are being incorporated in CCT intervention design, which BCTs are used (if any), or whether there is any evidence supporting the usefulness of BCTs to improve the efficacy of CCT programs.

This review is the first to synthesise evidence from published trials about the use of behavioural strategies in CCT interventions. The aims of this review were to (a) describe BCTs used in CCT trials in cognitively healthy older adults, and (b) explore whether specific BCTs are associated with improved adherence and efficacy. The current review extends the recent review of 90 trials evaluating the efficacy of CCT in older adults without cognitive decline (Lampit et al., 2020) by conducting a comprehensive meta-analysis of BCTs involved these trials. The findings provide valuable information relevant to the design and implementation of future large-scale CCT interventions.

Methods

This review was registered in PROSPERO (CRD42017071112) and was conducted in accordance with the PRIMSA guidelines. Note, however, that this review extends the previously published update by Lampit et al. (2020). Therefore, where relevant, we refer to that review to avoid unnecessary duplication.

Search Strategy

The search was conducted in Medline, Embase, and PsycINFO and in August 2019. For the current review, all full texts were retrieved of the 90 studies included in the 2020 review that published results from randomised controlled trials of effects of CCT on one or more cognitive outcomes in healthy (i.e., no major cognitive, neurological, psychiatric, or sensory impairments) older adults (≥ 60 years) (Lampit et al., 2020). Studies were included if they involved CCT either as a single intervention or as part of a multidomain strategy, provided that the CCT intervention consisted of at least 50% of the total intervention load. The review did not exclude studies based on their delivery techniques or CCT content to allow a comparison of the efficacy of different approaches.

Data Extraction and Coding of Behaviour Change Techniques

The following data were extracted from each of the 90 studies: (a) the instructed total duration to train, (b) the average time spent training, (c) the number of people who completed the study, (d) the number of people who dropped out, (e) reasons for drop out, (f) the number of people completing minimum required duration of training, (g) reasons for non-adherence, and (h) BCTs built into the intervention to support adoption or maintenance of new cognitive training behaviours. In addition, we used the published effect sizes (standard mean difference calculated as Hedges’ g) and variance for each of the studies (Lampit et al., 2020). The Hedges’ g values were used as a measure of efficacy and were calculated by pooling the results from all reported cognitive tests to reflect overall cognitive function (Lampit et al., 2020). For each study, data were extracted by a minimum of two investigators (GP, IB, SG, JF, CME, AS) and discrepancies were solved after discussion. If needed, a third investigator was consulted.

Coding of BCTs was based on the Behaviour Change Technique Taxonomy (BCTT v1) (Michie et al., 2015). This taxonomy describes 93 distinct BCTs divided into 16 clusters and has become the standard for classifying and reporting BCTs in the behaviour change literature. Only BCTs that targeted uptake of CCT behaviours in the active intervention group were coded. For four BCTs on monitoring and feedback, we specified whether the participants were monitored and whether feedback was given by a person or a computer. This distinction is not made in the BCTT, but we felt this mode of delivery could potentially modify the association between the BCT and the efficacy of CCT. For each paper, a minimum of two investigators (GP, IB, SG, JF, CME, AS) coded all papers on presence of BCTs. If needed, a third investigator was consulted to solve disagreements between the coders. All coders completed the BCTT training (https://www.bct-taxonomy.com/) and are certified BCT coders.

As generally few data were available on adherence, we had to adopt a pragmatic definition of adherence based on available information. Adherence was defined as either (in order of priority): [1] the percentage of average duration spent training relative to the instructed duration to train, [2] the percentage of participants in the study that met the studies’ criterion for adherence, or [3] the percentage of participants that completed the intervention.

Risk of Bias and Study Quality

For a detailed description of the risk of bias in the trials, please refer to our previous publication (Lampit et al., 2020). Briefly, using the Risk of Bias 2.0 tool (Sterne et al., 2019) studies with high risk of bias or some concerns in the domains ‘bias due to missing outcome data’ or ‘bias in measurement of the outcome’ were considered as having a high risk of bias (n = 36) or some concerns (n = 29), respectively. As the published information on risk of bias and methodological quality was deemed sufficient, no additional quality scoring was done for the current review.

Data Synthesis

First, the number and types of BCTs used across the 90 studies were described. Second, a three-level meta-regression (including the levels study, comparison, and outcome) was done including the BCT as a covariate and Hedges’ g as the outcome reflecting measure of efficacy of CCT over and above control. The models were repeated for each of the BCTs in the Taxonomy that were used in three or more studies. The regression coefficient reflects the difference in Hedges’ g between studies that did or did not use that BCT. Presented are the differences in Hedges’ g (Δg) and 95% confidence intervals (CI). Third, adherence rates and reasons for non-adherence were described. For each of the BCTs, median adherence rates were compared between studies that used or did not use that BCT using Wilcoxon rank sum test. The level of statistical significance was set at p < 0.05. Although, no cut-points or norms have been established for what constitutes a clinically meaningful difference in Hedges’ g values. We therefore applied the arbitrary cut-points of Δg < -0.10 and > 0.10, which is one standard deviation difference between subgroups and more than 50% difference in Hedge’s g relative to the overall efficacy across the 90 studies (g = 0.18).

Results

Full texts were retrieved for all 90 studies. The 90 studies included a total of 117 subgroups and 1201 comparisons. The studies varied in sample size from 20 to 1398 and in mean age from 60.7 to 85.8 years (Table 1).

Table 1 Study characteristics

The overall agreement in coding of BCTs between the two coders was high (99%), and the agreement in BCTs selected by at least one of the coders was 76%. Eighty-eight of the 90 studies reported using at least one BCT. The median number of BCTs used was 3 (interquartile range [IQR] = 2—5). Of the 93 BCTs in the taxonomy, 34 BCTs were identified in at least one study (Table 2). The most commonly used BCTs were ‘graded tasks’ (n = 62), ‘instruction on how to perform the behaviour’ (n = 51), ‘feedback on outcomes of behaviour’ (n = 29), and ‘adding objects to the environment ‘ (n = 19).

Table 2 Meta-analyses of the efficacy of computerized cognitive training interventions on cognitive function stratified by use of behaviour change techniques

We were able to reproduce the overall estimated effect size as reported by Lampit et al. (i.e., Hedges’ g = 0.18, CI = 0.14—0.23). No BCTs were associated with a statistically significant greater efficacy (Fig. 1; Table 2). However, Δg exceeded the cut-point for clinical relevance (> 0.10) for the BCTs ‘monitoring of outcomes of behaviour without feedback – by a person’, ‘feedback on outcomes of behaviour – by a person’ and ‘credible source’. The BCTs ‘feedback on behaviour – by a person’ (Δg = -0.19, CI = -0.31, -0.07) and ‘non-specific reward’ (Δg = -0.19, CI = -0.34, -0.05) were associated with lower efficacy (Fig. 1; Table 2). In addition, Δg exceeded the cut-point for clinical relevance (< -0.10) for the BCTs ‘monitoring of behaviour by others without feedback – by a computer’, ‘monitoring of outcomes of behaviour without feedback – by a computer’, ‘demonstration of the behaviour’ and ‘verbal persuasion about capability’. Supplemental file 1 lists examples for each of these BCTs.

Fig. 1
figure 1

Meta-regression of computerized cognitive training interventions comparing the efficacy (expressed as Hedge’s g, x-axis) between studies that did and dit not use each of the behaviour change techniques (BCT, y-axis). The difference in Hedges’ g (Δg) reflects the difference in efficacy of the studies that did versus studies that did not include that BCT. A positive Δg reflects that the efficacy was higher in the studies that did use that specific BCT. A negative Δg reflects that the efficacy was lower in the studies that did use that specific BCT. Presented are the results for those BCTs that were used in at least three studies. Detailed results for all BCTs are presented in Table 2

Data on adherence were available for 52 studies (Table 1). Across these studies, the median adherence was high (90%, IQR = 81–95). Twenty-five studies reported the average time spent doing the cognitive training, which ranged from 75 to 100% of the total time participants were instructed to train. Forty-four studies reported the reasons for drop-out or non-adherence in the intervention group. The reported reasons for drop-out or non-adherence were excessive time commitment (eight papers), health problems (25 papers), lack of interest or motivation (11 papers), holiday or travel (five papers), lack of time (six papers), non-adherence (six papers), and disliking training (two papers) (Table 3). Studies that used the BCTs ‘self-monitoring of behaviour’, ‘monitoring of behaviour by others without feedback’ (p = 0.05) and ‘self-monitoring of outcomes of behaviour’ (p = 0.05) reported higher adherence rates than studies that did not use these BCTs (p < 0.001) (Table 4). Studies that used the BCT ‘graded tasks’ had lower adherence rates than studies that did not use that BCT (p < 0.001).

Table 3 Reasons for drop-out and non-adherence
Table 4 Comparing adherence in intervention groups between studies using and not using behaviour change techniquesa

Discussion

Main Findings

The presented results provide first evidence that specific BCTs can improve adherence to and negatively or positively affect efficacy of CCT programs in cognitively healthy older adults. Nearly all studies used at least one BCT. While 34 out of a potential 93 different BCTs were used across the studies, there was substantial variation in the BCTs that were used between studies. No BCTs were statistically significantly associated with positive effects on efficacy, while two BCTs were associated with negative effects on efficacy. Three BCTs were associated with positive effects on adherence and one with negative effects on adherence. As the average duration of the CCT interventions was 7.3 ± 4.2 (range 1.5–26) weeks, the results and discussion reflect the short-term adoption phase of CCT rather than the long-term maintenance (> 6 months) phase of CCT.

Few statistically significant differences in efficacy were found for any of the BCTs. This may either be explained by a true absence of associations or lack of statistical power. Moreover, many tests were done and Type 1 error cannot be ruled out, but it is yet unclear how best to account for multiple testing in meta-analyses (Bender et al., 2008). Also, a potential true absence of associations could be due to individual variation in preferences for BCTs, resulting no overall benefits on a group level. Although a large number of studies was included in the meta-regressions, some BCTs were used in only a small number of studies, resulting in wide confidence intervals. While these results need to be interpreted with caution, it is important to also consider the results in the context of clinical relevance. Seven BCTs exceed the threshold for clinical relevance (Δg of < -0.10 or > 0.10), suggesting that these BCTs may be interesting targets to explore further in future research to boost the efficacy of CCT interventions (Fig. 1, Table 2). These BCTs are discussed further below.

The contrasting directions of associations for the BCTs around feedback and monitoring may either reflect the lack of statistical power or suggest a complex interplay between what is monitored and how. Monitoring of the outcomes of behaviour (i.e., game performance) might be beneficial, while monitoring of the behaviour itself (i.e., time spent training) might be counter effective. Moreover, monitoring and feedback may translate to better efficacy if provided by a person but not if provided by a computer. Many CCT programs have an inbuilt option to track progress and time spent exercising. This feature is often used by researchers to monitor adherence to the training program. It is also used to send reminders to participants if adherence is dropping (i.e., prompting). Interestingly, ‘self-monitoring of the behaviour’, ‘feedback on behaviour’ ‘monitoring of the behaviour by others without feedback’ and (to a lesser extent) ‘prompts or cues’ appear to be associated with better adherence (Table 4), but not with better efficacy. This suggests that monitoring of behaviour and prompting may help to increase awareness of time spent training, but not necessarily be beneficial for improving the quality or intensity of training. While ‘feedback on outcomes of behaviour’ may be associated with better efficacy if the feedback was delivered by a person, ‘monitoring of behaviour by others without feedback’ by a computer tended to be associated with poorer efficacy (Fig. 1). Awareness of being monitored without receiving feedback may be perceived as stressful. Mixed methods research may provide insight into how the use of these BCTs influences experience, participation and engagement.

The positive direction of associations of ‘credible source’ and ‘feedback on performance by a person’ with efficacy suggest that personal contact is important. In contrast, there was a non-significant tendency toward lower efficacy in studies that provided verbal persuasion about capability (n = 3, Δg = -0.17, CI = -0.43, 0.10). Verbal persuasion involves telling the participant that they can successfully perform the training and arguing against self-doubts. Collectively, these findings suggest that personal contact is important, but that the type and form of information require further thought. Co-creation of CCT programs may be a good step forward to design programs that better align with the participants preferences and needs and avoid stress.

The BCT ‘non-specific rewards’ was associated with lower efficacy (Fig. 1, Table 2), but did not affect adherence (Table 4). If a reward is performed well, then the amount of effort participants invest is influenced by the expected reward. The non-specific nature of the reward may dilute the efficacy or may be de-motivating compared to a more specific, tangible award. Examples of the types of rewards that were offered included access to the game for three additional months post-intervention (Bozoki et al., 2013), a fun fact on the screen (Anguera et al., 2013), or points or animations based on their performance (Mayas et al., 2014). The lack of a positive association with adherence may be because these rewards were mostly linked to outcomes (i.e., performance during training) rather than behaviour (i.e., time spent training). It may be that the use of rewards alone, in absence of other BCTs, is not sufficient to stimulate adherence. It could also be that the type of reward is not meaningful to the participants. Note that some studies offered a financial reward for completion of the pre- and post-intervention assessments. As this reward was offered to improve adherence to the data collection, it was not coded as a BCT to improve adherence to the training program. The potential adverse effects of this BCT further emphasises the need to tailor the BCTs to the target population.

Three commonly used BCTs were ‘graded tasks’ (n = 62)’, ‘instruction on how to perform the behaviour’ (n = 51) and ‘adding objects to the environment ‘ (n = 19). One could argue that these three BCTs are essential to achieve any training effect. Participants need to receive instructions on the required behaviour to be able to adhere to the protocol (‘instruction on how to perform the behaviour’). The BCT ‘graded tasks’ was considered present if the program involved adaptation to changes in the participants level of cognitive function. The ability of the program to constantly adapt to the individual’s progress during training is fundamental to the efficacy of CCT (Lampit et al., 2014b). The BCT ‘adding objects to the environment’ was coded if studies provided software or hardware, or access to an online program. In accordance with the taxonomy manual (Michie et al., 2015), these three BCTs were coded only if the studies specifically described using these techniques in their papers. While not all studies described using these BCTs, it is likely that most, if not all, studies would have applied these BCTs in some form. This may explain why no beneficial effects are observed in adherence or efficacy when comparing studies coded as using these BCTs with those coded as not using these BCTs. Moreover, ‘graded tasks’ was associated with poorer adherence, which might be due to losing motivation when reaching a level that is perceived as too high. Detailed reporting of training protocols and consistent use of terminology would facilitate more accurate estimation of the effectiveness of these BCTs on adherence and efficacy.

Previous reviews of BCTs used in lifestyle interventions identified ‘goal setting (outcome)’ or ‘action planning’ as important (Cradock et al., 2017; Lara et al., 2014; Olander et al., 2013). In the current review, only four studies used ‘action planning’ (i.e., detailed planning of performance of the behaviour, e.g. frequency, duration, intensity), which had similar efficacy as the studies that did not use this BCT (Fig. 1). Two studies used ‘goal setting (outcome)’ (i.e., setting or agreeing on a goal defined in terms of the behaviour to be achieved). One of these studies had a markedly higher efficacy (Hedges’ g = 0.48, CI = 0.28–0.69) and the other study had a markedly lower efficacy (Hedges’ g = -0.24, CI = -0.55, 0.07) than the pooled efficacy across all studies (Hedges’ g = 0.18, CI = 0.14–0.23). More studies incorporating goal setting into CCT interventions are required for meaningful meta-analysis.

For some studies the definition of adherence is based on study completion, whereas for other studies the definition of adherence is based on time spent training. Drop-out during the intervention period and adherence to the protocol were generally poorly reported. The two definitions were combined to have sufficient number of papers for meaningful analyses. However, it is possible that BCTs affect drop-out differently than time spent training. Many studies present results from completers only. Some studies describe low adherence as a reason for exclusion from the analyses. This means that both the adherence and efficacy of CCT are likely overestimated. The high average adherence and narrow IQR (90%, IQR = 81–95) illustrate that there is little variation, which reduces the likelihood of finding statistically significant differences. A survey among 831 brain trainers identified participant characteristics associated with training time and performance. The survey found that being open to experience and having a positive attitude towards cognitively demanding situations were associated with longer training continuation (Double & Birney, 2016). In addition, being open to experience and believing intelligence to be modifiable were associated with better game performance (Double & Birney, 2016). Future research may explore how BCTs can tap into these personality traits and metacognitive beliefs to boost adherence and efficacy.

A surprising finding in the Lampit review was that interventions with lower frequency (1–3 session per week) had better efficacy than interventions with higher frequency (> 3 sessions per week) (Lampit et al., 2020). The current review of adherence and BCTs provides new insights that may explain this finding. First, all six studies with low frequency interventions were group-based interventions (Bottiroli & Cavallini, 2009; Peng et al., 2012; Rasmusson et al., 1999; Vance et al., 2007; Wolinsky et al., 2011) and four were supervised (Bottiroli & Cavallini, 2009; Vance et al., 2007; Wang et al., 2011; Wolinsky et al., 2011). Group-based interventions had better efficacy than home-based interventions. It may be that the apparent better outcomes of low frequency interventions are driven by the group-based, supervised nature of the interventions. This finding is consistent with other behavioural interventions (e.g., physical activity) and the challenge is to transition from supervised group settings to unsupervised individual training while maintaining adherence and efficacy. Second, four of the studies used a BCT that involved personal contact, including social support (Bottiroli & Cavallini, 2009; Wolinsky et al., 2011), credible source (Rasmusson et al., 1999; Wolinsky et al., 2011) and feedback on the outcomes of the behaviour (Wang et al., 2011). This may have boosted the efficacy of these studies. Third, interventions with lower frequency may be easier to fit in to the daily routine than interventions with higher frequency. Subsequently, adherence may be better in lower frequency interventions. However, the six studies provided little information on adherence and no strong conclusion can be drawn.

Strengths and Limitations

Strengths of this review include the large number of studies and detailed information on efficacy (data were collected for all domains of cognitive function) and the BCTs were coded by a minimum of two trained coders with high agreement between coders. The following limitations need to be considered. First, this review included studies that predominantly involved unimodal CCT interventions. Multimodal interventions, typically a combination of CCT and exercise, were considered only if the CCT component consisted at least 50% of the full intervention. However, the increase in multimodal interventions indicates that researchers believe greater gains in cognitive function (or less decline) may be obtained from CCT in combination with wider risk factor management than from CCT alone. In multimodal interventions, the role of BCTs is even more crucial than in CCT only interventions as participants need to be encouraged to adhere to each of the intervention components. As use of BCTs is more integrated in lifestyle interventions than in CCT interventions, it is likely that use of BCTs are more common in multimodal interventions than in CCT only interventions. In a review of multimodal interventions, it may be challenging to tease out which BCTs were important for which intervention component. A critical motivation for this work is to inform design of future interventions. This review will be particularly relevant to inform which BCTs may be useful to promote CCT either as a single intervention or as part of a multimodal intervention. Further research is required to examine which BCTs are specifically effective in optimising adherence to CCT in single and multimodal interventions, and whether these differ from BCTs used in lifestyle interventions. Second, this review focused on BCTs used in trials evaluating the efficacy of CCT in healthy older adults. Hence, the current findings may not be generalizable to other types of cognitive training programs (e.g., paper-based cognitive training or multicomponent interventions) or other subgroups (e.g., older adults with mild cognitive impairment). Finally, the results were pooled across domains of cognitive function, creating an overall estimate of efficacy in terms of global cognitive function. Too few studies would be available for meaningful meta-analyses of the less-frequently used BCTs and of the less frequently measured cognitive domains. Moreover, there were no reasons to believe a priori that the influence of BCTs on efficacy may be different for domains of cognitive function, though future research would have to confirm this.

Conclusion

The presented results provide first evidence that BCTs may influence both adherence to and efficacy of CCT programs in cognitively healthy older adults. The BCTs that appear to positively influence efficacy of CCT programs include ‘credible source’, ‘information about social and environmental consequences’ and ‘feedback on outcomes of behaviour’ (but only when the feedback is delivered by a person). The BCTs that appear to negatively influence efficacy include BCTs related to monitoring of and feedback on behaviours and outcomes of behaviour, particularly when provided by a computer. However, few associations were statistically significant and further research is required to verify the current findings. Moreover, future research should involve the target group in the design of CCT programs and application of BCTs to ensure the use of BCTs is aligned with their preferences and needs.