Abstract
The Cognitive Failures Questionnaire (CFQ) has been widely used as a measure of subjective cognitive function in everyday life for decades. However, the evidence on how it relates to objective performance on executive function tasks is mixed. One possible reason for these mixed results is that the CFQ has selective relationships with some aspects of executive function and not others. Here, therefore, we classified tasks according to an influential framework of executive functions—switching, updating, inhibition, and we also considered the Sustained Attention to Response Task (SART) as a category because it was custom designed to gauge cognitive failures. We synthesized a large body of available evidence and performed four Bayesian meta-analyses on the relationship between CFQ scores and objective performance on executive function tasks in these four categories. Results suggested that CFQ scores were associated with objective performance on SART (18 effect sizes, μ = −.19, BF10 = 18.03, i.e., 18.03 times more evidence of a relationship versus no relationship), updating working memory (49 effect sizes, μ = −.06, BF10 = 17.80), and inhibition tasks (41 effect sizes, μ = −.07, BF10 = 15.40), whereas there was not definitive evidence regarding switching (34 effect sizes, μ = −.06, BF10 = .50, i.e., two times greater evidence for no relationship). This suggests that subjective cognitive function can predict objective performance on at least some executive function tasks. We discuss methodological and theoretical factors that constrain the maximum observable correlation and consider the relative insights that subjective measures versus task performance provide.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Introduction
Cognitive failures—such as forgetting appointments, daydreaming when you ought to be listening, and forgetting what you came to the shops to buy—are common human experiences. However, people differ in the frequency with which they experience these cognitive failures, and these differences are measured by the Cognitive Failures Questionnaire (CFQ; Broadbent et al., 1982). The CFQ has widespread currency in experimental, clinical, and applied work, as reflected in the fact that the paper introducing the CFQ has been cited more than 3,500 times at the time of this writing. The CFQ scores have been found to be associated with meaningful outcomes, such as that those who experience greater cognitive failures are more prone to work accidents and car accidents (Larson et al., 1997; Wallace & Vodanovich, 2003) and have impaired work ability (Gajewski et al., 2023). They also show more difficulty regulating their emotions (Robins et al., 2012) and have lower life satisfaction (Haslam et al., 2008).
Some studies have found that those who experience greater cognitive failures perform worse on tasks that measure attentional control or executive functioning (e.g., Forster & Lavie, 2007; Friedman & Miyake, 2004; Goodhew & Edwards, 2023; Manly et al., 1999; McVay & Kane, 2009). However, not all studies find such relationships (e.g., Carrigan & Barkus, 2016; Keilp et al., 2013; Richter et al., 2015; Ryskin et al., 2015). Some of this variance may be due to the specific type of task used. For example, studies that measure performance on the Sustained Attention to Response Task (SART) appear to find relationships with CFQ scores reasonably consistently (e.g., Linden et al., 2005; Manly et al., 1999; Robertson et al., 1997; Smilek et al., 2010). In contrast, for other tasks, such as Stroop, the findings are more mixed, with some studies finding a relationship between performance and CFQ scores (e.g., Adjorlolo, 2016; Vom Hofe et al., 1998) and others not (e.g., Gajewski et al., 2011; Gunduz et al., 2022; Keilp et al., 2013). Similarly, some studies find a relationship between performance on the Trail Making Test B and CFQ scores (e.g., Gajewski et al., 2011; Miskowiak et al., 2023), whereas others do not (e.g., Baker et al., 2018; Bellens et al., 2022). These differences could reflect the CFQ having selectively stronger associations with some aspects of executive function than others. Further, some studies have found no relationship between CFQ scores and memory when it is measured via tasks that simply require the retention of information (e.g., Castellon et al., 2004; Rodgers, 2000). But retention of information is not an executive function, whereas updating working memory—which requires the processing and updating of information in memory—is an executive function, and tasks that gauge updating working memory have been found to correlate with CFQ scores (e.g., Chang et al., 2021; but see Ryskin et al., 2015). However, attempting to draw such conclusions from isolated comparisons is fraught because these studies all differ in other ways which could have affected the outcomes, such as the number of participants in the study. Further, evidence for small effects may only be revealed when data are collated across multiple studies. This is where meta-analysis is a particularly powerful and useful tool, because it can synthesize a large body of available findings to assess whether there is overall evidence for an effect, and whether this depends on factors such as task type.
Here, therefore, we performed meta-analyses on the large body of available evidence on the relationship between CFQ scores and objective performance on executive function tasks to assess whether there is a robust relationship, and whether this differed according to different aspects of executive function. In particular, we applied an influential model of executive functions that identifies switching,Footnote 1 working memory updating, and inhibition, as three distinct components of executive functions (Miyake et al., 2000). In addition, we considered the association between CFQ scores and performance on a particular task—the SART—because this task was specifically developed to gauge the same type of cognitive errors as gauged by the CFQ (Robertson et al., 1997). That is, in the SART, a dominant response set is developed by responding with a particular key press to most stimuli, but then having to withhold a response when an infrequently occurring stimulus appears. Errors of commission—responding when a response should be withheld—are thought to reflect the same type of everyday cognitive lapses captured by the CFQ (Manly et al., 1999; Robertson et al., 1997). Therefore, we measured the strength of the relationship between CFQ scores and objective performance where cognitive tasks were classified into one of four categories: switching tasks, working memory updating tasks, inhibition tasks, and SART.
Method
The search for relevant works was conducted on 7 July 2023. Two databases were used: PsycInfo and Scopus. In PsycInfo, title, abstract, and key concepts were searched with the following terms: (‘cognitive failures’ OR ‘CFQ’ OR ‘subjective cogniti*’) AND (‘executive function’ OR ‘working memory’ OR ‘inhibition’ OR ‘inhibitory control’ OR ‘shifting’ OR ‘switching’ OR ‘updating’ OR ‘SART’ OR ‘sustained attention’). This yielded 432 results. In Scopus, title, abstract, and keywords were searched with the same terms as PsycInfo, which yielded 827 results. This means that a total of 1,259 works were initially identified (see Fig. 1). From these, there were 965 unique works after duplicates were removed. In the process of assessing these 965 for eligibility, an additional 10 works were identified (six from authors and four from additional references), as described below. This means that a total of 975 works were ultimately assessed for eligibility.
For studies that contained both CFQ and one or more eligible performance measures, but did not report the correlation, the corresponding author was emailed on the address listed on the paper to ask for this information. If this did not elicit a response, then a search was done to attempt to identify their current email address, and they were contacted on this if available, and/or the current address of other authors. In addition, corresponding authors of recent works (published from 2014 onwards) who reported the relevant correlation(s) in their work were contacted see if they had other published or unpublished work that would meet eligibility criteria. An additional six works were identified this way (of which four met eligibility criteria). In addition, an older meta-analysis on CFQ–SART associations (Smilek et al., 2010) and a qualitative review that treated cognitive function as unitary (Carrigan & Barkus, 2016) were identified amongst the N = 965 records. While these reviews were not eligible (although Smilek et al., 2010, also included an original experiment that was), their identified references were checked against the 965 records. From this, four additional older works that were not identified in the original search were identified and all assessed as eligible for inclusion. A flow chart of this process is shown in Fig. 1.
The eligibility criteria were as follows:
-
1.
The work must be an original study (as opposed to a meta-analysis or review), but could be published or unpublished, and could be a thesis.
-
2.
The sample size must be greater than 1 (i.e., case studies not included).
-
3.
If the same data were presented in a thesis and a published paper, then only the published paper was counted as eligible.
-
4.
The study must be with adults and could include clinical groups. This is because correlations can be constrained if there is insufficient range in a variable such as cognitive failures. Excluding clinical groups (e.g., those with ADHD) could unduly constrain the range of CFQ scores.
-
5.
CFQ scores must be provided by the person performing the objective task (i.e., self-rated CFQ scores). This is because we were interested in the CFQ as a measure of an individual’s subjective cognitive function.
-
6.
While most studies used the original 25-item CFQ, other variants of the CFQ were eligible if they included original items from the CFQ. The role of CFQ type was assessed via statistical analysis.
-
7.
The study must report an association (or the data were available so that an association could be calculated, or authors provide the association on request) between CFQ scores and objective performance scores on an eligible task (see below). While correlation coefficients were preferred, regression coefficients were included if that was all that was available, as were tests of participants categorised as low versus high scorers on the CFQ.
-
8.
Where a study explicitly only reported significant correlation coefficients and omitted nonsignificant correlation coefficients, if the nonsignificant correlation coefficients could not be obtained on request, then these significant coefficients were not included so as not to introduce bias toward significant results (two studies).
-
9.
If the study was assessing an intervention, then correlation(s) between CFQ and performance that were assessed at baseline (preintervention) only were eligible.
-
10.
Where two works used the same participants (e.g., a later follow-up), the earlier study only was considered eligible because this would have the maximum N size (smaller at follow-up due to drop-out).
-
11.
Where a work had multiple potential relevant measures in a single category (e.g., three different measures of updating working memory), then the average effect sizes for each of these measures was computed, unless the authors had already computed a composite of the relevant measures (Lange & Süß, 2014; McVay & Kane, 2009; Unsworth et al., 2012), in which case this was used.
-
12.
Where a work had eligible effect sizes in different categories (e.g., a measure of updating working memory and a measure of inhibition), all eligible effect sizes were included (i.e., a single study could contribute effect sizes to multiple categories).
Where studies reported correlations for separate groups (e.g., patients vs. controls), or multiple experiments with different participants, these were treated as separate effect sizes. Since these are with different samples, this does not undermine the independence of effect sizes. However, it was also common for a given study to contribute multiple effect sizes in different categories (see eligibility criterion #12). This is because many studies had participants complete a battery of executive function tasks in addition to the CFQ. Had we restricted each study to only contributing a single effect size, then we would have had to make arbitrary choices about which to include, and we would have been discarding large amounts of data. We reasoned that it was better to use all this available data. However, consequently, the effect sizes in different categories are not truly independent of one another. For example, the same participants may have contributed an effect size to the switching and inhibition categories. For this reason, we performed a meta-analysis separately on each of the four categories of executive function tasks (i.e., switching, updating working memory, inhibition, SART).
For Category 1, we identified tasks that operationalise switching of task sets or goals (see Table 1). This includes the classic Wisconsin Card Sorting Test (WCST; Berg, 1948), which measures the switching component of executive function (Miyake et al., 2000). The preferred metric was perseverative errors, but if unavailable then overall accuracy metrics were used. The intra-extra dimensional set shift offered in CANTAB® (Cambridge Cognition) was considered equivalent to the WCST, although we were not able to obtain any the required correlation coefficients for studies that used this measure, and we have therefore omitted it from Table 1. Trail Making Test B (Reitan, 1955), where participants have to switch between using letters and numbers to determine where to draw a line, also measures switching and was therefore eligible (Arbuthnott & Frank, 2000). Trail Making Test B scores reflect the time taken to complete the task. Other task-switching paradigms were also eligible, which were defined as where participants were performing one task (e.g., classifying digits as odd/even) and then must switch to a new task (e.g., classifying digits as larger or smaller than 5). Switch cost or absolute accuracy and/or RT in switch conditions (used whichever available, or average of accuracy and RT if both available).
For Category 2, we included tasks that measured working memory capacity, as opposed to mere retention of information. That is, only tasks that required participants to update and/or perform some type of mental manipulation of the to-be-remembered material were included. Tasks that simply required the retention of information were not eligible. Eligible tasks included reading span, operation span (OSPAN), and symmetry span tasks (Conway et al., 2005; Daneman & Carpenter, 1980; Redick et al., 2012; Shah & Miyake, 1996; Turner & Engle, 1989). These are complex span tasks that require participants to juggle two tasks and update the list of to-be-remembered material, and such tasks have been linked to the updating component of working memory (Miyake et al., 2000). The complex span used in one study (Guye et al., 2020) was considered equivalent to symmetry span because it involved memorising the positions of red squares presented in a grid, where the presentation of the squares was interleaved by another task (judging orientation of Ts and Ls). The computation span used in one study (Montgomery & Fisk, 2007) was considered equivalent to OSPAN because participants were solving arithmetic problems while also simultaneously remembering the second digit of each presented problem. Dot span was also considered eligible as a complex span when it was aggregated with reading span, operation span, and other updating tasks (Lange & Süß, 2014).
Digit span backwards (rather than forwards) has been linked to working memory (Hale et al., 2002; Reynolds, 1997; Yoshimura et al., 2023), and digit span sequencing may be especially linked to working memory (Meyer & Reynolds, 2018; Werheid et al., 2002). Similarly, there is evidence that letter–number sequencing reflects working memory (Crowe, 2000; Haut et al., 2000). Therefore, digit span backwards, digit span sequencing, and letter–number sequencing were eligible for Category 2, whereas digit span forwards was not. In a similar vein, tasks that required participants reverse or otherwise change the sequence of remembered information were considered eligible, such as the Corsi block tapping reverse (Ahmad et al., 2022; Könen & Karbach, 2020), visual working memory backwards (Kitahata et al., 2017), and dot matrix and spatial span backwards (Peers et al., 2022).
The n-back requires online processing and updating of information and was therefore considered eligible for Category 2. Although there have been some questions about its relationship with other measures of working memory (e.g., Jaeggi et al., 2010; Redick & Lindsey, 2013), it has been suggested to capture working memory function and to detect subtle changes in cognitive function that other tasks were not able to detect (Miller et al., 2009), and therefore it was considered eligible. Where more than one n was available (e.g., 2-back and 1-back), the highest n available was used to obtain the relevant effect size. Further, the Paced Auditory Serial Addition Test (PASAT; Bloomfield et al., 2010; Nikravesh et al., 2017), and its visual counterpart, the Paced Visual Serial Addition Task (PVST; Krabbe et al., 2017; Nagels et al., 2005), require online mental arithmetic with the to-be-remembered information, and were therefore considered eligible for Category 2.
For Category 3, tasks that gauge inhibition were eligible. Specifically, tasks that require inhibiting salient information including Stroop (MacLeod, 1991; Stroop, 1935), Simon (Hommel, 2011; Simon & Rudell, 1967), flanker (Burgoyne et al., 2023; Eriksen & Eriksen, 1974), and antisaccade (Everling & Fischer, 1998; Hutton & Ettinger, 2006) were eligible for Category 3. Any other task where participants were responding to targets and concurrent nontargets could produce interference were also eligible. For all, the preferred measure was interference (i.e., difference in performance between incongruent and congruent conditions), but if unavailable, then performance in the incongruent condition itself was used. Whichever of accuracy or RT was available was used. If both were available, then the average of the two effect sizes (one for accuracy, one for RT) was used.
Tasks that require inhibiting a dominant response, such as go/no-go (Gomez et al., 2007; Wright et al., 2014) and were also eligible for Category 3. For go/no-go tasks accuracy only was eligible because this provides direct insight into the success of inhibiting the response. Commission errors were preferred, but overall accuracy was also eligible. Continuous performance tasks were considered equivalent to go/no-go tasks only if they included the requirement to withhold a response to certain stimuli (Wright et al., 2014). Here, again, the focus was on accuracy. Stop-signal tasks were also eligible, in which case stop-signal RT was used (Verbruggen & Logan, 2008).
Finally, Category 4 had only a single eligible task: The SART (Robertson et al., 1997; Vallesi et al., 2021). Here, the preferred outcome was commission errors, but if only an overall sensitivity or accuracy measure was available, then this was used. It is possible that the SART belongs in Category 3 along with other response inhibition tasks. However, given the intrinsic conceptual relationship between SART and CFQ, we initially assessed it as a separate category, and assessed during data-analysis whether it could be productively combined with Category 3.
All performance measures were coded such that higher scores reflected better performance on the task. Given that higher CFQ scores reflect greater cognitive failures (i.e., poorer cognitive function), this means that a negative correlation was expected between CFQ scores and performance measures. A bivariate correlation was the preferred effect size, but if only regression coefficients or partial correlations were available, then these were used, and +.05 was added to positive coefficients and −.05 was subtracted from negative coefficients to correct for the reduction in coefficient magnitude relative to bivariate correlations (Vaerenbergh, 2018). Where an F-value or t-value for higher and lower CFQ groups was all that was available, this F-value was converted to a Cohen’s d effect size using an online calculator (Lenhard & Lenhard, 2022), and then converted to a correlation coefficient (Lin, n.d.). All standard errors were calculated using 95% confidence intervals using an online tool provided by Psychometrica (Lenhard & Lenhard, 2014). As stipulated in Berkhout et al. (2024), effect sizes and standard errors were transformed to Fisher’s z to correct for the bounded nature of correlation coefficients.
Results
Eligible effect sizes and participant characteristics
A total of 88 works, spanning publication years from 1987 to 2024, contributed a total of 151 effect sizes met the above eligibility criteria and were included in the meta-analysis.
In Category 1, there were 37 effect sizes from 37 studies (Adjorlolo, 2016; Amidi et al., 2015; Baker et al., 2018; Bassel et al., 2002; Bellens et al., 2022; Berende et al., 2019; Costa et al., 2023; Das-Smaal et al., 1993; Dawson et al., 2014; Dey et al., 2019; Finn & McDonald, 2015; Friedman & Miyake, 2004; Gajewski & Falkenstein, 2012, 2014, 2015a; Gajewski et al., 2011, 2022; Gajewski, Thönes, et al., 2020b; Goodman et al., 2022; Guariglia et al., 2020; Gustavson et al., 2015; Hagen et al., 2020; Jozwiak et al., 2017; Keilp et al., 2013; Koso et al., 2012; Kramer et al., 1994; Kreitz et al., 2015; Lange & Süß, 2014; Lapadula et al., 2020; Miskowiak et al., 2021, 2022, 2023; Shakeel & Goghari, 2017; Slana Ozimič & Repovš, 2020; Steinbusch et al., 2017; Vom Hofe et al., 1998; Woods et al., 2024).Footnote 2,Footnote 3N = 6,092 participants contributed data to this category, including students, general community adults, healthy older adults (50+), older adults with mild or subjective cognitive impairment, older adults with vascular risk factors, older adults with amnesic mild cognitive impairment, twins, people previously hospitalised due to coronavirus disease (COVID-19) infection, cardiac arrest survivors, breast cancer survivors with cognitive complaints, combat veterans with posttraumatic stress disorder (PTSD), and people with: chronic back pain, long COVID, persistent symptoms attributed to Lyme disease, testicular cancer, Major Depressive Disorder (MDD), traumatic brain injury (TBI), and Parkinson’s disease. For this and all subsequent categories, the list of the specific task measures derived from each study and the participant type for each are provided in the Supplementary Material.
In Category 2, there were 46 references (Ahmad et al., 2022; Amidi et al., 2015; Bellens et al., 2022; Bertens et al., 2016; Bloomfield et al., 2010; Chang et al., 2021; Dawson et al., 2014; Dean & Sterr, 2013; Donohoe et al., 2009; Friedman & Miyake, 2004; Gajewski & Falkenstein, 2012, 2014, 2015a; Gajewski et al., 2011, 2022; Gajewski, Thönes, et al., 2020b; Gokal et al., 2018; Guariglia et al., 2020; Gustavson et al., 2015; Guye et al., 2020; Hagen et al., 2020; Hitchcott et al., 2017; Jensen et al., 2022; Jozwiak et al., 2017; Judah et al., 2014; Kim et al., 2014; Kitahata et al., 2017; Könen & Karbach, 2020; Krabbe et al., 2017; Kreitz et al., 2015; Lange & Süß, 2014; McVay & Kane, 2009; Miskowiak et al., 2022; Molteni et al., 2008; Montgomery & Fisk, 2007; Peers et al., 2022; Pollina et al., 1992; Richter et al., 2015; Rodriguez et al., 2013; Rossiter et al., 2006; Ryskin et al., 2015; Shakeel & Goghari, 2017; Torenvliet et al., 2022; Unsworth et al., 2012; Welhaf & Kane, 2024; Woods et al., 2024).Footnote 4 These yielded 52 effect sizes, because five references had two separate effect sizes for different groups (e.g., clinical vs. control; Ahmad et al., 2022; Dean & Sterr, 2013; Krabbe et al., 2017; Rossiter et al., 2006; Torenvliet et al., 2022), and one reference reported two separate studies that were both eligible (Kreitz et al., 2015). N = 6,690 individuals contributed data to this category, including students, students with attention-deficit/hyperactivity disorder (ADHD), general community adults, healthy older adults (50+), older adults with mild or subjective cognitive impairment, twins, breast cancer survivors with cognitive complaints, stroke survivors, people who use Methylenedioxymethamphetamine (MDMA), and people with: work-related stress, stress-related exhaustion, long COVID, tinnitus, testicular cancer, breast cancer, MDD, autism, schizophrenia, TBI, vestibular loss, and Parkinson’s disease.
Category 3 included 41 references (Adjorlolo, 2016; Baker et al., 2018; Bellens et al., 2022; Berende et al., 2019; Costa et al., 2023; Dey et al., 2019; Forster & Lavie, 2007; Friedman & Miyake, 2004; Gajewski & Falkenstein, 2012, 2014, 2015a; Gajewski et al., 2011, 2022; Gajewski, Thönes, et al., 2020b; Gokal et al., 2018; Gunduz et al., 2022; Gustavson et al., 2015; Hagen et al., 2020; Horvat & Tement, 2020; Hsieh et al., 2022; Ishigami & Klein, 2009; Jozwiak et al., 2017; Judah et al., 2014; Kalpidou et al., 2021; Keilp et al., 2013; Könen & Karbach, 2020; Krabbe et al., 2017; Kramer et al., 1994; Kreitz et al., 2015; Lopresti et al., 2022, 2023; Polsinelli et al., 2020; Ryskin et al., 2015; Slagboom et al., 2021; Slana Ozimič & Repovš, 2020; Swick & Ashley, 2017; Tipper & Baylis, 1987; Vom Hofe et al., 1998; Welhaf & Kane, 2024; Woltering et al., 2013; Woods et al., 2024). These yielded 44 effect sizes because two references had two separate eligible studies (Ishigami & Klein, 2009; Kreitz et al., 2015), and one reference had two discrete groups of participants (Krabbe et al., 2017). N = 6,058 contributed data to this category, including students, students with low- and high-trait anxiety, students with ADHD, general community adults, healthy older adults (50+), people 40–75 with self-reported subjective cognitive issues, older adults with mild or subjective cognitive impairment, older adults with vascular risk factors, twins, breast cancer survivors with cognitive complaints, combat veterans with PTSD, and people with: chronic back pain, stress-related exhaustion, persistent symptoms attributed to Lyme disease, hypopituitarism, breast cancer, MDD, TBI, and Parkinson’s disease.
In Category 4, sixteen references (Bloomfield et al., 2010; Dockree et al., 2006; Donohoe et al., 2009; Farrin et al., 2003; Gokal et al., 2018; Koso et al., 2012; Larue et al., 2010; Linden et al., 2005; Manly et al., 1999; McVay & Kane, 2009; Righi et al., 2009; Robertson et al., 1997; Smilek et al., 2010; Wallace et al., 2001; Whyte et al., 2006; Zanesco et al., 2019) contributed 18 effect sizes, because two references had two discrete groups (Dockree et al., 2006; Whyte et al., 2006). N = 1,448 individuals contributed data to this category, including students, general community adults, military personnel, combat veterans with PTSD, and people with: work burnout, breast cancer, schizophrenia, and TBI.
Bayesian meta-analysis: All eligible effects
Values were rounded to two decimal places for calculations. Analysis was performed in JASP (Version 0.14.1). We performed four random-effects Bayesian meta-analyses with default priors, one for each of the four categories of executive function tasks. Bayes factors reflected the evidence for the alternative hypothesis (that CFQ scores and performance were associated, no direction assumed) divided by the evidence for the null hypothesis (that CFQ scores and performance were not related). Therefore, values greater than 1 reflect greater evidence for the alternative hypothesis while values less than 1 reflect greater evidence for the null hypothesis. The strength of evidence provided by different Bayes factors were interpreted consistent with the recommendations in van Doorn et al. (2021). That is, a BF10 greater than 3 reflects moderate evidence for alternative hypothesis (i.e., 3× evidence for alternative vs. null), while a BF10 less than 1/3 reflects moderate evidence for the null (i.e., 3× more evidence for null vs alternative). Values less than 3 and greater than .33 reflect an indeterminant outcome, or weak evidence for the alternative and null hypotheses respectively. BF10 > 10 and < .1 reflect strong evidence for the alternative and null hypotheses, respectively.
The first analysis included variants of the CFQ and all types of effect size. The results of this first analysis are provided in Table 2. In subsequent sensitivity analyses, we assessed the influence of removing outliers, restricting only to studies that used the original 25-item CFQ (Broadbent et al., 1982), and restricting only to studies that provided bivariate correlations. Note for all studies included in the meta-analyses, information about the individual studies including the specific measures they used are provided in theSupplementary Material.
Table 2 shows that all four categories produced negative mean effects (i.e., correlations), consistent with individuals more prone to cognitive failures demonstrating worse objective task performance. The BF10 values indicated that there was only weak evidence for the effects for switching and inhibition. In contrast, there was strong support for the effects for Updating Working Memory and SART.
Bayesian meta-analysis: Outliers excluded
Despite the diversity of eligible tasks for the updating working memory category, there was minimal variance in the effect sizes in this category. However, given the large heterogeneity present in many of the categories, boxplots of effect sizes were used to identify outliers within each category. Based on this, N = 3 effect sizes were excluded from Category 1, N = 3 from Category 2, N = 3 from Category 3, and none from Category 4. It appeared that most if not all these effect size outliers were derived from studies with lower sample sizes, and thus a sampling error likely contributed to them being outlier values. That is, for Category 1, these outlier effect sizes all arose from sample sizes smaller than the average for the category (N = 36, N = 12, and N = 70; Bassel et al., 2002; Miskowiak et al., 2022; Vom Hofe et al., 1998), where the mean was N = 165, median = 91. Similarly, for Category 2, these outlier effect sizes also all arose from sample sizes smaller than the mean for the category (N = 14, N = 25, and N = 37, where mean N = 129, median = 75; Ahmad et al., 2022; Chang et al., 2021). Further, for Category 3, these outlier effect sizes were also all derived from sample sizes smaller than average for the category (N = 62, N = 25, and N = 32, where mean N = 138, median N = 69; Costa et al., 2023; Krabbe et al., 2017; Tipper & Baylis, 1987). The study with N = 62 in Category 3 also derived from (a) a factor of the CFQ rather than the full original CFQ, (b) from regression weights rather than bivariate correlations, and (c) Stroop incongruent trials rather than interference scores (Costa et al., 2023). These factors may have also contributed to making that effect size an outlier.
The results of the meta-analyses following these exclusions are provided in Table 3. The forest plots are large and thus are provided in the Supplementary Material. Table 3 shows that once outliers were excluded, now inhibition also showed strong evidence for a mean effect, along with updating working memory and SART. Switching now showed weak evidence in favour of the null hypothesis of no relationship. There appeared to be substantive reduction in heterogeneity in some categories once these small number of outliers were excluded (e.g., now confidence intervals around τ for both switching and inhibition did not overlap their confidence intervals from previous analysis). Further, as discussed above, these outlier effect sizes appeared to be products of studies with smaller sample sizes producing unreliable estimates of effect size. Therefore, this analysis with the outliers excluded was considered the primary results, and all subsequent analyses were performed with these outliers excluded to provide representative and robust estimates of effect sizes.
It could be argued that SART contains elements of response inhibition, akin to that captured by tasks such as go/no-go in Category 3. To test this, we examined the effect of combining Category 3 and Category 4. While the mean effect was clear (μ = −.11, [−.15, −.07], BF10 = 3,321.19), the combined τ was halfway between the two individual τs suggesting that there was no synergistic benefit to combining them, the confidence intervals for τ still overlapped that of both original estimates, suggesting that there was no clear reduction in heterogeneity from combining them, and there was greater evidence for heterogeneity than for either category in isolation (τ = .11, [.08, .15], BF10 = 6.334e6). Therefore, they were treated as separate categories for the remaining analyses.
To test for potential publication bias, funnel plots were generated via classical meta-analysis with restricted ML method examined, and rank correlation test for plot asymmetry was performed. None of the plots suggested publication bias, and none of the rank tests were significant (p-values ≥ .176), indicating no evidence of publication bias, except for inhibition, which was significant (p = .047). This was such that negative correlations may have been overrepresented. However, additional analyses on subsets of studies below eliminated evidence for publication bias in this category, but there was still clear evidence for the effect size for this category. Thus, the presence of a reliable effect is not attributable to publication bias.
Bayesian meta-analysis: Outliers excluded and restricted to original CFQ
In the previous analyses, variants of the CFQ were eligible for inclusion. While most studies used the original CFQ, some did not. Specifically, in switching, four out of 31 used a variant of the CFQ, in updating working memory, eight out of 49 used variants, in inhibition, five out of 41 used variants, and in SART, three out of 18 used a variant. Given the relatively small number of studies using variants, to assess whether CFQ type affected the results, the next analysis was focussed exclusively on studies that used the original CFQ only (not variants of the CFQ). The results for all categories are shown in Table 4.
Here, there was strong evidence for the mean effect for inhibition and moderate evidence for the presence of an effect for updating working memory and SART. There was still only weak support for switching. Further, in this analysis, there was no longer evidence for publication bias in Category 3 (inhibition) (p = .139).
Bayesian meta-analysis: Outliers excluded and restricted to bivariate correlations only
The most accurate effect size derivation is from bivariate correlations. Regression coefficients require an adjustment for the other variables in the analysis, and this adjustment is likely going to be an imperfect correction when the true bivariate relationship is not known. Therefore, to check that such corrections had not distorted the outcome, next we performed an additional analysis deriving effect sizes only from bivariate correlations, not partial correlations, regression coefficients, or group-difference statistics. We returned to considering all CFQ types for these analyses. The results are shown in Table 5.
Here, there was moderate evidence for the presence of an effect for SART, and strong evidence for updating working memory and inhibition. There was weak evidence in favour of the null hypothesis for switching.
Bayesian meta-analysis: Reducing heterogeneity for switching
Finally, we assessed whether the heterogeneity in switching could be reduced by focusing only on studies that used the same outcome measure. The Trail Making Test B was the most commonly used metric in this category, and therefore we focussed the analysis exclusively on this category. In addition, to promote homogeneity, the analysis was focused only on original CFQ items and only bivariate correlations. This left N = 19 studies for analysis. Although there was minimal change in τ value itself, the strength of evidence for heterogeneity was now substantively reduced, (τ = .12, [.06, .20], BF10 = 49.77), but there was still only weak evidence for the switching effect size (μ = −.10, [−.17, −.03], BF10 = 1.85). This means that although we reduced the magnitude of evidence for it (from BF10 = 19,406.53 to BF10 = 49.77), we were not able to eliminate the strong evidence for heterogeneity within the switching condition. The SART was another category which consistently showed strong evidence for heterogeneity across analyses. This suggests that there are likely additional variables that moderate the strength of the observed relationship in these categories. As discussed below, measurement reliability is one such candidate.
Results summary
To summarise, in every analysis, there was evidence that multiple aspects of objective performance on executive function were related to CFQ scores. These effects were such that individuals more prone to subjective cognitive failures (i.e., higher CFQ scores) were associated with poorer objective performance on these executive function tasks. With respect to the specific categories, there was not consistent evidence that switching was related to CFQ scores. Instead, the evidence for this category percolated around an indeterminant zone of weak evidence for either the null or alternative hypothesis. There was consistently moderate or strong evidence for updating working memory and SART being related to CFQ scores. Once outliers were removed, there was also consistently strong evidence for inhibition. There was some potential evidence for publication bias for inhibition, but even in analyses where this evidence was eliminated (i.e., focus on original CFQ only), there was still clear evidence for a robust effect size for inhibition.
Discussion
Here, we used Bayesian meta-analyses to assess the relationship between scores on the CFQ and objective performance on executive function tasks that fell into one of four categories. Three categories were derived from an influential taxonomy of executive function (i.e., switching, updating working memory, and inhibition; Miyake et al., 2000), and an additional one specifically for SART was included because this task was purpose-made to detect cognitive failures (Robertson et al., 1997). We assessed the robustness of effects across multiple analytic choices. There was consistently moderate or strong evidence for CFQ scores being related to performance on updating working memory tasks and SART, and once outliers were removed, there was also strong evidence for tasks gauging inhibition. In contrast, there was never definitive evidence for the relationship between CFQ score and performance on tasks that are designed to measure switching.
These results indicate that people do have a degree of insight into their cognitive function, and how they subjectively experience it playing out in everyday life is systematically related to how it influences objective performance on tasks designed to measure some aspects of executive function. However, an important qualification is that the category of task matters: This relationship was only robust for certain types of tasks and not others. Consistent with SART being developed to detect the absent-minded lapses in attention that are reflected in the content of the CFQ, there was strong evidence for a CFQ–SART relationships. Indeed, numerically at least, SART consistently produced the strongest relationships. This is consistent with recent evidence that the CFQ shows reliable associations with low-prevalence visual search (Goodhew & Edwards, 2023; Thomson & Goodhew, 2021)—a special kind of visual search tasks where targets occur infrequently and which results in an elevated target miss rate (Horowitz, 2017; Wolfe et al., 2005). While such studies are not pure SART tasks and thus were not included in the present meta-analysis, the fact that both SART and low-prevalence search performance correlates with CFQ suggests that there may be some shared variance between SART and low-prevalence visual search. We suggest that this is likely because they both develop participants’ expectations about a particular type of outcome being the most common, and then infrequently require deviating from this expectation, and it is this flexibility that individuals prone to cognitive failures appear to have difficulty with. This is consistent with some CFQ items reflecting inflexibility in adapting to changing context, such as failing to hear people speaking to you when doing something else. Here, the person has focussed their attention in one way, and is inefficient and/or ineffective in changing this focus as required.
Further, the results indicated that individuals prone to cognitive failures have difficulty juggling the demands of complex span tasks including updating and performing operations on the contents of their working memory. Working memory facilitates effective and strategic regulation of attention in the pursuit of goals (e.g., Bleckley et al., 2003; Burgoyne et al., 2023; Robison & Unsworth, 2017). In this light, it is interesting that multiple CFQ items refer to issues with goal pursuit, including a failure to initiate goals (e.g., leaving important letters unanswered for days), failure to properly execute a goal (e.g., throwing away the thing you meant to keep and keeping the thing you meant to throw away), and losing track of goals as the context changes (e.g., forgetting why you went from one part of the house to another, and forgetting what you came to the shops to buy). Given the function of working memory, in conjunction with the negative relationship between updating working memory and CFQ scores observed in the meta-analysis, it is likely that poorer ability to update working memory is implicated in these issues with initiating, maintaining, and successfully executing goals that are apparent in those prone to cognitive failures.
In a similar vein, the present findings indicate that individuals prone to cognitive failures also have difficulty inhibiting salient stimuli or inhibiting a response. This is consistent with CFQ items reflecting failures to apply attention to important sources of information (e.g., failing to notice signposts on the road, daydreaming when you ought to be listening to something, and failing to see what you want in the supermarket even though it is there), and failing to inhibit responses (e.g., losing one’s temper and regretting it). This suggests that inhibition is an important aspect of a person’s overall cognitive function in everyday life.
What individuals prone to cognitive failures did not appear to have clear difficulty doing is switching tasks or attentional sets. Across our analyses, the Bayes factors never deviated from the weak/indeterminant zone of evidence to provide definitive support for the presence (or absence) of this relationship. This seems surprising, given that CFQ scores were related to all other aspects of executive function. It is possible that the CFQ is truly unrelated to switching. None of the items refer explicitly to difficulty starting a new task after doing a different one. Although some refer to the opposite—accidentally switching tasks, such as Do you start doing one thing at home and get distracted into doing something else (unintentionally)?—this is different from a carryover where a previous task negatively impacts a new one. Alternatively, it is possible that CFQ scores are related to switching performance but that the methodology of the studies in the meta-analysis obscured this relationship. One possible methodological explanation for this is that since most effect sizes in this category were derived from the Trail Making Test B, this measure may be not optimally sensitive to individual differences in switching. Indeed, the Trail Making Test B derives from the neuropsychological literature, and therefore it may be better suited to detecting gross cognitive impairment following brain injury rather than making subtle differentiation between task-switching ability. It may be that CFQ scores would be robustly related to switching task performance if more sensitive measurements of switching were used that were better able to finely differentiate between individuals’ switching capacities. All we can say at this point is that there is not definitive evidence that CFQ scores are related (or not related) to switching performance.
While here we focussed on the dissociable components of executive function (switching, updating, and inhibition), there is also evidence of a common shared variance between all executive functions (e.g., Miyake & Friedman, 2012). The shared (unity) versus dissociable (diversity) variance of executive functions can be gauged via latent variable modelling. This approach addresses the ‘task impurity’ problem, which is that while some of the variance in performance on a specific executive function task reflects a person’s executive function, task-specific factors that do not gauge executive function also contribute variance to performance (e.g., colour-naming ability in Stroop). Having participants complete multiple measures of each of the dissociable components of executive function allows the shared versus dissociable components of executive function to be distinguished from one another and from task-specific variance via latent variable modelling (Miyake & Friedman, 2012; Miyake et al., 2000). Unfortunately, this was not feasible to do for this meta-analysis, given that most studies did not have the requisite measures and/or individual-level data available. However, we think this would be a valuable future research direction to further elucidate how the CFQ relates to performance on executive function tasks. In particular, for switching, it would be informative to determine whether it is related to CFQ scores when multiple (potentially more sensitive) measures of switching are used, and switching is quantified via latent variable modelling. If multiple measures of other aspects of executive function were also included, then it would also inform whether CFQ is most strongly related to general executive function rather than the dissociable components, which is quite possible given the breadth of cognitive domains captured by the CFQ. In this light, it is interesting that higher CFQ scores are associated with difficulties in emotion regulation (Robins et al., 2012) including reduced use of reappraisal emotion regulation strategy (Goodhew & Edwards, 2024b), and reappraisal ability has been shown to have selective relationships with general executive function (Toh & Yang, 2022).
It is also important to acknowledge that none of the effect size estimates here were large. In the following sections, we discuss some methodological as well as theoretical reasons that could explain this. When there is divergence between subjective measures and measures derived from objective performance on experimental tasks, subjective measures are often blamed. But it is important to consider reasons that objective measures may contribute to such divergence. Here, we will consider four such reasons. The first is measurement reliability. The CFQ typically shows good measurement reliability. In contrast, many experimental tasks designed to assess executive function—especially those that were first designed to show large-magnitude effects in the whole sample—can have poor measurement reliability when it comes to consistently differentiating between individuals (Friedman & Gustavson, 2022; Goodhew & Edwards, 2019; Hedge et al., 2018). This is because an effect that is experienced strongly by most or all in the sample typically has low between-participant variability, and between-participant variability is a prerequisite for individual differences reliability. Indeed, tasks such as the flanker task and Stroop task, both of which were eligible for the inhibition category, have been critiqued for having poor measurement reliability (Hedge et al., 2018).Footnote 5 This is important because the measurement reliability of both the subjective and experimental measure will ultimately constrain the maximum correlation that can be observed between them (e.g., Dang et al., 2020; Parsons et al., 2019; Spearman, 1910), and therefore poor measurement reliability of experimental measures may have led to an underestimation of the magnitude of the association between CFQ scores and objective performance on at least some of the task categories here. It is so rare for the measurement reliability of experimental tasks to be explicitly reported that we could not perform any systematic correction for measurement reliability here, and thus it remains a viable cause of effect size underestimation. In this light, it is impressive that despite these issues, we still observed robust relationships between CFQ and objective performance on multiple aspects of executive function.
Second, subjective measures also typically aggregate information over a far larger times cale (e.g., CFQ asks about experiences in the previous 6 months), whereas objective performance on executive function tasks is usually assessed at a single point in time (Duckworth & Kern, 2011). This means that not only is the experimental measure more limited in its scope but it is also potentially more affected by state factors (e.g., sleep deprivation the previous night) that muddy estimates of a person’s trait executive function ability. These factors may contribute to the poor measurement reliability of some experimental measures, thus limiting the maximum observable correlation.
Third, subjective measures often capture a broader range of cognitive function, in contrast to experimental tasks, which typically reflect just a narrow snapshot of cognitive function, often with impoverished or artificial stimuli (Carrigan & Barkus, 2016; Dang et al., 2020; Toplak et al., 2013). For example, the CFQ asks about failures across a diverse array of aspects of executive function (e.g., prioritising the wrong information stream, succumbing to distracting information, failing to initiate goals, failing to maintain goals across contexts), whereas most experimental executive function tasks operationalise just a single aspect (e.g., Stroop task assess succumbing to distracting information). Further, while the CFQ collates information across a variety of contexts (e.g., driving, social settings, shopping), behavioural measures are derived from an experimental context with impoverished stimuli in artificial set-ups which often do not resemble the stimuli or tasks of everyday life (e.g., most adults do not spend significant time naming colours). These differences mean that we would not necessarily expect a high degree of convergence between the subjective and experimental measures, and the subjective measures may provide the more meaningful insight.
Fourth, while performance-based measures likely gauge optimal or maximal function where the experimental context provides clear structure and requirements, subjective measures poll information about typical everyday usage in less structured contexts (Dang et al., 2020; Friedman & Gustavson, 2022; Toplak et al., 2013). It is possible that executive function deficits may be more pronounced in typical usage in the relatively unstructured contexts of everyday life. In this vein, it is interesting to note that subjective cognitive decline in older adults appears to precede and predict objectively measurable decline in cognitive performance in the years ahead (Mitchell et al., 2014).
Of course, however, subjective measures are vulnerable to several key sources of measurement error, including self-presentation biases and failures of metacognitive insight. There are some situations in which these sources of error will fatally undermine the validity of a subjective measure (e.g., contexts where there is overwhelming motivation to present as capable, or where metacognitive insight is severely impaired; Friedman & Gustavson, 2022). However, one key advantage of the CFQ over other measures of subjective cognitive function, such as the Attentional Control Scale (ACS; Clarke & Todd, 2021; Derryberry & Reed, 2002; Judah et al., 2014), is that the item design minimises the role of metacognitive insight by focusing on the outcomes of cognitive function, which may improve its validity (Goodhew & Edwards, 2024b; Thomson & Goodhew, 2021). That is, the ACS requires participants to engage in sophisticated metacognition and comparison with others to answer questions such as “it is easy for me to alternate between two tasks.” There is clear evidence that humans experience a task-switching cost (e.g., Monsell, 2003), so an individual with astutely tuned metacognitive insight would appreciate this cost and thus may not endorse a statement about finding it easy to switch between tasks even though they are relatively good at it. To endorse this item, they need to know how their task-switching cost (or subjective feelings in relation to task switching) compares with that of other individuals, and such information is likely difficult to ascertain. If they do not endorse such items, then their ACS score would underestimate their true attentional control ability. In contrast, the CFQ simply requires participants to remember instances of failures of attentional control, such as bumping into people and forgetting appointments. Of course, a person with severe memory impairment may forget instances of forgetting. But provided that overall memory is not severely degraded, CFQ items are likely less demanding of metacognitive insight than other subjective cognition measures like the ACS. Reducing the contribution of metacognitive insight to scores on a subjective cognitive measure will (all other things being equal) reduce error variance and thus increase their validity. In this light, it is interesting that CFQ scores have been shown to predict low prevalence visual search performance when ACS do not (Thomson & Goodhew, 2021).
Ultimately, subjective and objective performance measures may gauge different aspects of cognitive function, and therefore, the greatest utility may be obtained by considering the information offered by both types of measures, rather than relying on either in isolation (e.g., Duckworth & Seligman, 2005; Friedman et al., 2020; Toplak et al., 2013). It may be particularly revealing to consider when and how they converge versus diverge. However, if a researcher faces time and resource constraints, then in many instances subjective measures are going to provide a more reliable and practical-to-obtain estimate (Duckworth & Kern, 2011). The CFQ in particular has a lot to offer.
However, the CFQ is many decades old, and many of the items no longer have the same contemporary relevance (e.g., referring to physical newspapers), and it lacks items that capture many modern instances of cognitive failures (e.g., forgetting passwords). We recently developed a contemporary measure of cognitive failures—the CFQ 2.0 (Goodhew & Edwards, 2024a). It consists of 15 items that are a mix of original CFQ items and new contemporary items. We found that the CFQ 2.0 outperformed the original CFQ in explaining variance in objective performance in low-prevalence visual search (Goodhew & Edwards, 2024a). Researchers should consider using this updated CFQ so that the CFQ can continue to maintain and even enhance its validity.
In conclusion, we performed Bayesian meta-analyses on the relationship between CFQ scores and objective performance on executive function tasks in four categories—switching, updating working memory, inhibition, and SART. There was robust evidence for the relationship between CFQ scores and performance on SART, updating working memory, and inhibition tasks, whereas there was not definitive evidence either for or against the relationship between CFQ and performance on switching tasks. This could be because the primary method of measuring switching in the studies included was the Trail Making Test B, which may be more suited to detecting gross dysfunction than subtle individual differences in switching ability. Altogether, this suggests that subjective cognitive function is associated with objective performance on executive function tasks, but it is important to consider the type of cognitive task when examining this association. It is also important to consider the different information about cognitive function provided by subjective measures like the CFQ compared with measures derived from performance on experimental tasks. At least in some cases, the CFQ can offer richer and more reliable insight into cognitive function.
Availability of data and material
Some additional data are provided in the Supplementary Material. The full raw dataset is available upon request.
Code availability
Not applicable.
Notes
This is also known as “shifting” (the two terms are used interchangeably in the literature). Here, we prefer the term “switching” because this maps onto task-switching, which is one of the core aspects of this aspect of executive function (at least to us, task-shifting sounds peculiar).
For Jozwiak et al. (2017) for this and subsequent categories, correlation was based on a subset of original data that was available for reanalysis.
For Gajewski and Falkenstein (2015a) that was identified in original search, when providing the correlations an author clarified that same sample is used in Gajewski and Falkenstein (2015b) and correlations reported here are from both papers. For simplicity, we count this as a single reference and cite Gajewski and Falkenstein (2015a) throughout.
For Kim et al. (2014), an author clarified that they were providing the correlation based on a series of studies in the laboratory all with Digit Span backwards and on students with ADHD, including but not limited to Kim et al. (2014). For simplicity, we count this as a single reference, since there was a single effect size resulting.
Hedge et al. (2018) attributed the poor measurement reliability to low between-participant variability. In contrast, Gajewski, Getzmann et al. (2020a), Gajewski, Thönes et al. (2020b) observed large between-participant variability in Stroop interference across different age groups, although they did not report estimates of measurement reliability. More broadly, the measurement reliability of any measure depends on an interaction between specific contextual variables (e.g., number of trials used) and the characteristics (e.g., traits) of the people completing the task, and thus there is not a single measurement reliability value for the Stroop task. However, there is clear evidence that at least in some instances, its measurement reliability can be low, and there is reason to think that this is true for many of studies considered here. Interested readers are referred to Burgoyne et al. (2023) for a modified Stroop task that is highly reliable.
References
Adjorlolo, S. (2016). Ecological validity of executive function tests in moderate traumatic brain injury in Ghana. The Clinical Neuropsychologist, 30(sup1), 1517–1537. https://doi.org/10.1080/13854046.2016.1172667
Ahmad, M., Bola, L., Boutabla, A., King, S., Lewis, R. F., & Chari, D. A. (2022). Visuospatial cognitive dysfunction in patients with vestibular loss. Otology and Neurotology, 43(10), e1140–e1147. https://doi.org/10.1097/mao.0000000000003696
Amidi, A., Wu, L. M., Agerbæk, M., Larsen, P. L., Pedersen, A. D., Mehlsen, M., Larsen, L., & Zachariae, R. (2015). Cognitive impairment and potential biological and psychological correlates of neuropsychological performance in recently orchiectomized testicular cancer patients. Psychooncology, 24(9), 1174–1180. https://doi.org/10.1002/pon.3804
Arbuthnott, K., & Frank, J. (2000). Trail Making Test, Part B as a measure of executive control: Validation using a set-switching paradigm. Journal of Clinical and Experimental Neuropsychology, 22(4), 518–528. https://doi.org/10.1076/1380-3395(200008)22:4;1-0;FT518
Baker, K. S., Gibson, S. J., Georgiou-Karistianis, N., & Giummarra, M. J. (2018). Relationship between self-reported cognitive difficulties, objective neuropsychological test performance and psychological distress in chronic pain. European Journal of Pain, 22(3), 601–613. https://doi.org/10.1002/ejp.1151
Bassel, C., Rourke, S. B., Halman, M. H., & Smith, M. L. (2002). Working memory performance predicts subjective cognitive complaints in HIV infection. Neuropsychology, 16(3), 400–410. https://doi.org/10.1037/0894-4105.16.3.400
Bellens, A., Roelant, E., Sabbe, B., Peeters, M., & van Dam, P. A. (2022). Evaluation of a new online cognitive assessment tool in breast cancer survivors with cognitive impairment: A prospective cohort study. Supportive Care in Cancer, 30(1), 21–31. https://doi.org/10.1007/s00520-021-06397-1
Berende, A., Agelink van Rentergem, J., Evers, A. W. M., Ter Hofstede, H. J. M., Vos, F. J., Kullberg, B. J., & Kessels, R. P. C. (2019). Cognitive impairments in patients with persistent symptoms attributed to Lyme disease. BMC Infectious Diseases, 19(1), 833. https://doi.org/10.1186/s12879-019-4452-y
Berg, E. A. (1948). A simple objective technique for measuring flexibility in thinking. The Journal of General Psychology, 39(1), 15–22. https://doi.org/10.1080/00221309.1948.9918159
Berkhout, S. W., Haaf, J. M., Gronau, Q. F., Heck, D. W., & Wagenmakers, E.-J. (2024). A tutorial on Bayesian model-averaged meta-analysis in JASP. Behavior Research Methods, 56(3), 1260–1282. https://doi.org/10.3758/s13428-023-02093-6
Bertens, D., Kessels, R. P., Boelen, D. H., & Fasotti, L. (2016). Transfer effects of errorless goal management training on cognitive function and quality of life in brain-injured persons. NeuroRehabilitation, 38(1), 79–84. https://doi.org/10.3233/nre-151298
Bleckley, M. K., Durso, F. T., Crutchfield, J. M., Engle, R. W., & Khanna, M. M. (2003). Individual differences in working memory capacity predict visual attention allocation. Psychonomic Bulletin & Review, 10(4), 884–889. https://doi.org/10.3758/BF03196548
Bloomfield, I. L., Espie, C. A., & Evans, J. J. (2010). Do sleep difficulties exacerbate deficits in sustained attention following traumatic brain injury? Journal of the International Neuropsychological Society, 16(1), 17–25. https://doi.org/10.1017/s1355617709990798
Broadbent, D. E., Cooper, P. F., FitzGerald, P., & Parkes, K. R. (1982). The Cognitive Failures Questionnaire (CFQ) and its correlates. British Journal of Clinical Psychology, 21(1), 1–16. https://doi.org/10.1111/j.2044-8260.1982.tb01421.x
Burgoyne, A. P., Tsukahara, J. S., Mashburn, C. A., Pak, R., & Engle, R. W. (2023). Nature and measurement of attention control. Journal of Experimental Psychology: General, 152(8), 2369–2402. https://doi.org/10.1037/xge0001408
Carrigan, N., & Barkus, E. (2016). A systematic review of cognitive failures in daily life: Healthy populations. Neuroscience & Biobehavioral Reviews, 63, 29–42. https://doi.org/10.1016/j.neubiorev.2016.01.010
Castellon, S. A., Ganz, P. A., Bower, J. E., Petersen, L., Abraham, L., & Greendale, G. A. (2004). Neurocognitive performance in breast cancer survivors exposed to adjuvant chemotherapy and tamoxifen. Journal of Clinical and Experimental Neuropsychology, 26(7), 955–969. https://doi.org/10.1080/13803390490510905
Chang, L.-H., Chen, P.-Y., Wang, J., Shih, B.-H., Tseng, Y.-H., & Mao, H.-F. (2021). High-ecological cognitive intervention to improve cognitive skills and cognitive–functional performance for older adults with mild cognitive impairment. The American Journal of Occupational Therapy, 75(5), 7505205050. https://doi.org/10.5014/ajot.2021.041996
Clarke, P. J. F., & Todd, J. (2021). Lessons unlearned: A conceptual review and meta-analysis of the relationship between the Attention Control Scale and objective attention control. Cognition and Emotion, 35(8), 1447–1459. https://doi.org/10.1080/02699931.2021.1987861
Conway, A. R. A., Kane, M. J., Bunting, M. F., Hambrick, D. Z., Wilhelm, O., & Engle, R. W. (2005). Working memory span tasks: A methodological review and user’s guide. Psychonomic Bulletin & Review, 12(5), 769–786. https://doi.org/10.3758/bf03196772
Costa, A. N., Nowakowski, L. M., McCrae, C. S., Cowan, N., & Curtis, A. F. (2023). Discrepancies in Objective and Subjective Cognition in Middle-Aged and Older Adults: Does Personality Matter? Gerontology and Geriatric Medicine, 9, 23337214221146664. https://doi.org/10.1177/23337214221146663
Crowe, S. F. (2000). Does the letter number sequencing task measure anything more than digit span? Assessment, 7(2), 113–117. https://doi.org/10.1177/107319110000700202
Daneman, M., & Carpenter, P. A. (1980). Individual differences in working memory and reading. Journal of Verbal Learning and Verbal Behavior, 19(4), 450–466. https://doi.org/10.1016/S0022-5371(80)90312-6
Dang, J., King, K. M., & Inzlicht, M. (2020). Why are self-report and behavioral measures weakly correlated? Trends in Cognitive Sciences, 24(4), 267–269. https://doi.org/10.1016/j.tics.2020.01.007
Das-Smaal, E. A., de Jong, P. F., & Koopmans, J. R. (1993). Working memory, attentional regulation and the star counting test. Personality and Individual Differences, 14(6), 815–824. https://doi.org/10.1016/0191-8869(93)90094-J
Dawson, D., Richardson, J., Troyer, A., Binns, M., Clark, A., Polatajko, H., Winocur, G., Hunt, A., & Bar, Y. (2014). An occupation-based strategy training approach to managing age-related executive changes: a pilot randomized controlled trial. Clinical Rehabilitation, 28(2), 118–127. https://doi.org/10.1177/0269215513492541
Dean, P. J., & Sterr, A. (2013). Long-term effects of mild traumatic brain injury on cognitive performance. Frontiers in Human Neuroscience, 7, 30. https://doi.org/10.3389/fnhum.2013.00030
Derryberry, D., & Reed, M. A. (2002). Anxiety-related attentional biases and their regulation by attentional control. Journal of Abnormal Psychology, 111(2), 225–236. https://doi.org/10.1037/0021-843X.111.2.225
Dey, A. K., Stamenova, V., Bacopulos, A., Jeyakumar, N., Turner, G. R., Black, S. E., & Levine, B. (2019). Cognitive heterogeneity among community-dwelling older adults with cerebral small vessel disease. Neurobiology of Aging, 77, 183–193. https://doi.org/10.1016/j.neurobiolaging.2018.12.011
Dockree, P. M., Bellgrove, M. A., O’Keeffe, F. M., Moloney, P., Aimola, L., Carton, S., & Robertson, I. H. (2006). Sustained attention in traumatic brain injury (TBI) and healthy controls: Enhanced sensitivity with dual-task load. Experimental Brain Research, 168(1), 218–229. https://doi.org/10.1007/s00221-005-0079-x
Donohoe, G., Hayden, J., McGlade, N., O’Gráda, C., Burke, T., Barry, S., Behan, C., Dinan, T. G., O’Callaghan, E., Gill, M., & Corvin, A. P. (2009). Is “clinical” insight the same as “cognitive” insight in schizophrenia? Journal of the International Neuropsychological Society, 15(3), 471–475. https://doi.org/10.1017/s1355617709090559
Duckworth, A. L., & Kern, M. L. (2011). A meta-analysis of the convergent validity of self-control measures. Journal of Research in Personality, 45(3), 259–268. https://doi.org/10.1016/j.jrp.2011.02.004
Duckworth, A. L., & Seligman, M. E. P. (2005). Self-discipline outdoes IQ in predicting academic performance of adolescents. Psychological Science, 16(12), 939–944. https://doi.org/10.1111/j.1467-9280.2005.01641.x
Eriksen, B. A., & Eriksen, C. W. (1974). Effects of noise letters upon the identification of a target letter in a nonsearch task. Perception & Psychophysics, 16(1), 143–149. https://doi.org/10.3758/bf03203267
Everling, S., & Fischer, B. (1998). The antisaccade: A review of basic research and clinical studies. Neuropsychologia, 36(9), 885–899. https://doi.org/10.1016/S0028-3932(98)00020-7
Farrin, L., Hull, L., Unwin, C., Wykes, T., & David, A. (2003). Effects of depressed mood on objective and subjective measures of attention. The Journal of Neuropsychiatry and Clinical Neurosciences, 15(1), 98–104. https://doi.org/10.1176/jnp.15.1.98
Finn, M., & McDonald, S. (2015). Repetition-lag training to improve recollection memory in older people with amnestic mild cognitive impairment: A randomized controlled trial. Aging, Neuropsychology, and Cognition, 22(2), 244–258. https://doi.org/10.1080/13825585.2014.915918
Forster, S., & Lavie, N. (2007). High perceptual load makes everybody equal: Eliminating individual differences in distractibility with load. Psychological Science, 18(5), 377–381. https://doi.org/10.1111/j.1467-9280.2007.01908.x
Friedman, N. P., & Gustavson, D. E. (2022). Do rating and task measures of control abilities assess the same thing? Current Directions in Psychological Science, 31(3), 262–271. https://doi.org/10.1177/09637214221091824
Friedman, N. P., Hatoum, A. S., Gustavson, D. E., Corley, R. P., Hewitt, J. K., & Young, S. E. (2020). Executive functions and impulsivity are genetically distinct and independently predict psychopathology: Results from two adult twin studies. Clinical Psychological Science, 8(3), 519–538. https://doi.org/10.1177/2167702619898814
Friedman, N. P., & Miyake, A. (2004). The relations among inhibition and interference control functions: A latent-variable analysis. Journal of Experimental Psychology: General, 133(1), 101–135. https://doi.org/10.1037/0096-3445.133.1.101
Gajewski, P. D., & Falkenstein, M. (2012). Training-induced improvement of response selection and error detection in aging assessed by task switching: Effects of cognitive, physical, and relaxation training. Frontiers in Human Neuroscience, 6, 130. https://doi.org/10.3389/fnhum.2012.00130
Gajewski, P. D., & Falkenstein, M. (2014). Age-related effects on ERP and oscillatory EEG-dynamics in a 2-back task. Journal of Psychophysiology, 28(3), 162–177. https://doi.org/10.1027/0269-8803/a000123
Gajewski, P. D., & Falkenstein, M. (2015a). Lifelong physical activity and executive functions in older age assessed by memory based task switching. Neuropsychologia, 73, 195–207. https://doi.org/10.1016/j.neuropsychologia.2015.04.031
Gajewski, P. D., & Falkenstein, M. (2015b). Long-term habitual physical activity is associated with lower distractibility in a Stroop interference task in aging: Behavioral and ERP evidence. Brain and Cognition, 98, 87–101.
Gajewski, P. D., Hengstler, J. G., Golka, K., Falkenstein, M., & Beste, C. (2011). The Met-allele of the BDNF Val66Met polymorphism enhances task switching in elderly. Neurobiology of Aging, 32(12), 2327–2319. https://doi.org/10.1016/j.neurobiolaging.2011.06.010
Gajewski, P. D., Falkenstein, M., Thönes, S., & Wascher, E. (2020a). Stroop task performance across the lifespan: High cognitive reserve in older age is associated with enhanced proactive and reactive interference control. NeuroImage, 207, 116430. https://doi.org/10.1016/j.neuroimage.2019.116430
Gajewski, P. D., Thönes, S., Falkenstein, M., Wascher, E., & Getzmann, S. (2020b). Multidomain cognitive training transfers to attentional and executive functions in healthy older adults. Frontiers in Human Neuroscience, 14, 586963. https://doi.org/10.3389/fnhum.2020.586963
Gajewski, P. D., Getzmann, S., Bröde, P., Burke, M., Cadenas, C., Capellino, S., Claus, M., Genç, E., Golka, K., Hengstler, J. G., Kleinsorge, T., Marchan, R., Nitsche, M. A., Reinders, J., van Thriel, C., Watzl, C., & Wascher, E. (2022). Impact of biological and lifestyle factors on cognitive aging and work ability in the dortmund vital study: Protocol of an interdisciplinary, cross-sectional, and longitudinal study [Protocol]. JMIR Research Protocol, 11(3), e32352. https://doi.org/10.2196/32352
Gajewski, P. D., Rieker, J. A., Athanassiou, G., Bröde, P., Claus, M., Golka, K., Hengstler, J. G., Kleinsorge, T., Nitsche, M. A., Reinders, J., Tisch, A., Watzl, C., Wascher, E., & Getzmann, S. (2023). A systematic analysis of biological, sociodemographic, psychosocial, and lifestyle factors contributing to work ability across the working life span: Cross-sectional study. JMIR Formative Research, 7, e40818. https://doi.org/10.2196/40818
Gokal, K., Munir, F., Ahmed, S., Kancherla, K., & Wallis, D. (2018). Does walking protect against decline in cognitive functioning among breast cancer patients undergoing chemotherapy? Results from a small randomised controlled trial. PLOS ONE, 13(11), e0206874. https://doi.org/10.1371/journal.pone.0206874
Gomez, P., Ratcliff, R., & Perea, M. (2007). A model of the go/no-go task. Journal of Experimental Psychology: General, 136(3), 389–413. https://doi.org/10.1037/0096-3445.136.3.389
Goodhew, S. C., & Edwards, M. (2019). Translating experimental paradigms into individual-differences research: Contributions, challenges, and practical recommendations. Consciousness and Cognition, 69, 14–25. https://doi.org/10.1016/j.concog.2019.01.008
Goodhew, S. C., & Edwards, M. (2023). Subjective assessments of cognition and affect and their relationship with objective performance: Individuals with high levels of cognitive failures or negative affect miss more rare visual targets. Journal of Experimental Psychology: General, 152(3), 749–762. https://doi.org/10.1037/xge0001294
Goodhew, S. C., & Edwards, M. (2024a). The Cognitive Failures Questionnaire 2.0. Personality and Individual Differences, 218, 112472. https://doi.org/10.1016/j.paid.2023.112472
Goodhew, S. C., & Edwards, M. (2024b). Elevated cognitive failures in trait anxiety. Personality and Individual Differences, 216, 112418. https://doi.org/10.1016/j.paid.2023.112418
Goodman, Z. T., Timpano, K. R., Llabre, M. M., & Bainter, S. A. (2022). Revisiting the factor structure and construct validity of the Cognitive Failures Questionnaire. Psychological Assessment, 34(7), 671–683. https://doi.org/10.1037/pas0001127
Guariglia, P., Giaimo, F., Palmiero, M., & Piccardi, L. (2020). Normative data and validation of the Italian translation of the Working Memory Questionnaire (WMQ). Applied Neuropsychology: Adult, 27(4), 376–389. https://doi.org/10.1080/23279095.2018.1552147
Gunduz, H., Gunduz, T., & Ozkan Ceylan, A. (2022). High bladder pressure reduces the ability to filter out interference from distractors in low perceptual load condition. The Quarterly Journal of Experimental Psychology, 75(12), 2219–2231. https://doi.org/10.1177/17470218211068828
Gustavson, D. E., Miyake, A., Hewitt, J. K., & Friedman, N. P. (2015). Understanding the cognitive and genetic underpinnings of procrastination: Evidence for shared genetic influences with goal management and executive function abilities. Journal of Experimental Psychology: General, 144(6), 1063–1079. https://doi.org/10.1037/xge0000110
Guye, S., Röcke, C., Martin, M., & von Bastian, C. C. (2020). Functional ability in everyday life: Are associations with an engaged lifestyle mediated by working memory? Journals of Gerontology: Series B Psychological Sciences and Social Sciences, 75(9), 1873–1883. https://doi.org/10.1093/geronb/gbz056
Hagen, B. I., Lau, B., Joormann, J., Småstuen, M. C., Landrø, N. I., & Stubberud, J. (2020). Goal management training as a cognitive remediation intervention in depression: A randomized controlled trial. Journal of Affective Disorders, 275, 268–277. https://doi.org/10.1016/j.jad.2020.07.015
Hale, J. B., Hoeppner, J.-A.B., & Fiorello, C. A. (2002). Analyzing digit span components for assessment of attention processes. Journal of Psychoeducational Assessment, 20(2), 128–143. https://doi.org/10.1177/073428290202000202
Haslam, C., Holme, A., Haslam, S. A., Iyer, A., Jetten, J., & Williams, W. H. (2008). Maintaining group memberships: Social identity continuity predicts well-being after stroke. Neuropsychological Rehabilitation, 18(5/6), 671–691. https://doi.org/10.1080/09602010701643449
Haut, M. W., Kuwabara, H., Leach, S., & Arias, R. G. (2000). Neural activation during performance of number-letter sequencing. Applied Neuropsychology, 7(4), 237–242. https://doi.org/10.1207/S15324826AN0704_5
Hedge, C., Powell, G., & Sumner, P. (2018). The reliability paradox: Why robust cognitive tasks do not produce reliable individual differences. Behavior Research Methods, 50(3), 1166–1186. https://doi.org/10.3758/s13428-017-0935-1
Hitchcott, P. K., Fastame, M. C., Langiu, D., & Penna, M. P. (2017). Cognitive failures in late adulthood: The role of age, social context and depressive symptoms. PLOS ONE, 12(12), e0189683. https://doi.org/10.1371/journal.pone.0189683
Hommel, B. (2011). The Simon effect as tool and heuristic. Acta Psychologica, 136(2), 189–202. https://doi.org/10.1016/j.actpsy.2010.04.011
Horowitz, T. S. (2017). Prevalence in visual search: From the clinic to the lab and back again. Japanese Psychological Research, 59(2), 65–108. https://doi.org/10.1111/jpr.12153
Horvat, M., & Tement, S. (2020). Self-reported cognitive difficulties and cognitive functioning in relation to emotional exhaustion: Evidence from two studies. Stress Health, 36(3), 350–364. https://doi.org/10.1002/smi.2930
Hsieh, M. T., Lu, H., Chen, L. F., Liu, C. Y., Hsu, S. C., & Cheng, C. H. (2022). Cancellation but not restraint ability is modulated by trait anxiety: An event-related potential and oscillation study using go-nogo and stop-signal tasks. Journal of Affective Disorders, 299, 188–195. https://doi.org/10.1016/j.jad.2021.11.066
Hutton, S. B., & Ettinger, U. (2006). The antisaccade task as a research tool in psychopathology: A critical review. Psychophysiology, 43(3), 302–313. https://doi.org/10.1111/j.1469-8986.2006.00403.x
Ishigami, Y., & Klein, R. M. (2009). Are individual differences in absentmindedness correlated with individual differences in attention? Journal of Individual Differences, 30(4), 220–237. https://doi.org/10.1027/1614-0001.30.4.220
Jaeggi, S. M., Buschkuehl, M., Perrig, W. J., & Meier, B. (2010). The concurrent validity of the N-back task as a working memory measure. Memory, 18(4), 394–412. https://doi.org/10.1080/09658211003702171
Jensen, J. H., Miskowiak, K. W., Purdon, S. E., Thomsen, J. F., & Eller, N. H. (2022). Screening for cognitive impairment among patients with work-related stress complaints in Denmark: validation and evaluation of objective and self-report tools. Scandinavian Journal of Work, Environment & Health, 48(1), 71–80. https://doi.org/10.5271/sjweh.3990
Jozwiak, N., Postuma, R. B., Montplaisir, J., Latreille, V., Panisset, M., Chouinard, S., Bourgouin, P. A., & Gagnon, J. F. (2017). REM sleep behavior disorder and cognitive impairment in Parkinson’s disease. Sleep, 40(8), zsx101. https://doi.org/10.1093/sleep/zsx101
Judah, M. R., Grant, D. M., Mills, A. C., & Lechner, W. V. (2014). Factor structure and validation of the Attentional Control Scale. Cognition & Emotion, 28(3), 433–451. https://doi.org/10.1080/02699931.2013.835254
Kalpidou, M. M., Volungis, A. M., & Bates, C. (2021). Mediators between adversity and well-being of college students. Journal of Adult Development, 28(4), 286–298. https://doi.org/10.1007/s10804-021-09372-6
Keilp, J. G., Gorlyn, M., Russell, M., Oquendo, M. A., Burke, A. K., Harkavy-Friedman, J., & Mann, J. J. (2013). Neuropsychological function and suicidal behavior: Attention control, memory and executive dysfunction in suicide attempt. Psychological Medicine, 43(3), 539–551. https://doi.org/10.1017/s0033291712001419
Kim, S., Liu, Z., Glizer, D., Tannock, R., & Woltering, S. (2014). Adult ADHD and working memory: Neural evidence of impaired encoding. Clinical Neurophysiology, 125(8), 1596–1603. https://doi.org/10.1016/j.clinph.2013.12.094
Kitahata, R., Nakajima, S., Uchida, H., Hayashida, T., Takahashi, M., Nio, S., Hirano, J., Nagaoka, M., Suzuki, T., Jinno, H., Kitagawa, Y., & Mimura, M. (2017). Self-rated cognitive functions following chemotherapy in patients with breast cancer: A 6-month prospective study. Neuropsychiatric Disease and Treatment, 13, 2489–2496. https://doi.org/10.2147/ndt.S141408
Könen, T., & Karbach, J. (2020). Self-reported cognitive failures in everyday life: A closer look at their relation to personality and cognitive performance. Assessment, 27(5), 982–995. https://doi.org/10.1177/1073191118786800
Koso, M., Sarač-Hadžihalilović, A., & Hansen, S. (2012). Neuropsychological performance, psychiatric symptoms, and everyday cognitive failures in Bosnian ex-servicemen with posttraumatic stress disorder. Review of Psychology, 19(2), 131–139.
Krabbe, D., Ellbin, S., Nilsson, M., Jonsdottir, I. H., & Samuelsson, H. (2017). Executive function and attention in patients with stress-related exhaustion: perceived fatigue and effect of distraction. Stress, 20(4), 333–340. https://doi.org/10.1080/10253890.2017.1336533
Kramer, A. F., Humphrey, D. G., Larish, J. F., Logan, G. D., & Strayer, D. L. (1994). Aging and inhibition: Beyond a unitary view of inhibitory processing in attention. Psychology and Aging, 9(4), 491–512. https://doi.org/10.1037/0882-7974.9.4.491
Kreitz, C., Furley, P., Memmert, D., & Simons, D. J. (2015). Inattentional Blindness and Individual Differences in Cognitive Abilities. PLOS ONE, 10(8), e0134675. https://doi.org/10.1371/journal.pone.0134675
Lange, S., & Süß, H.-M. (2014). Measuring slips and lapses when they occur—Ambulatory assessment in application to cognitive failures. Consciousness and Cognition, 24, 1–11. https://doi.org/10.1016/j.concog.2013.12.008
Lapadula, G., Bernasconi, D. P., Bai, F., Focà, E., Di Biagio, A., Bonora, S., Castelli, F., Squillace, N., Bandera, A., Monforte, A. d. A., Migliorino, G. M., Gori, A., for the SWEAR Study Group. (2020). Switching from efavirenz to rilpivirine improves sleep quality and self-perceived cognition but has no impact on neurocognitive performances. AIDS, 34(1), 53–61. https://doi.org/10.1097/QAD.0000000000002377
Larson, G. E., Alderton, D. L., Neideffer, M., & Underhill, E. (1997). Further evidence on dimensionality and correlates of the Cognitive Failures Questionnaire. British Journal of Psychology, 88(1), 29–38. https://doi.org/10.1111/j.2044-8295.1997.tb02618.x
Larue, G. S., Rakotonirainy, A., & Pettitt, A. N. (2010). Real-time performance modelling of a Sustained Attention to Response Task. Ergonomics, 53(10), 1205–1216. https://doi.org/10.1080/00140139.2010.512984
Lenhard, W., & Lenhard, A. (2014). Hypothesis tests for comparing correlations. Psychometrica. https://doi.org/10.13140/RG.2.1.2954.1367
Lenhard, W., & Lenhard, A. (2022). Computation of effect sizes. Psychometrica. https://doi.org/10.13140/RG.2.2.17823.92329
Lin, H. (n.d.). Effect size converter. https://www.escal.site/
Linden, D. V. D., Keijsers, G. P. J., Eling, P., & Schaijk, R. V. (2005). Work stress and attentional difficulties: An initial study on burnout and cognitive failures. Work & Stress, 19(1), 23–36. https://doi.org/10.1080/02678370500065275
Lopresti, A. L., Smith, S. J., & Drummond, P. D. (2022). The effects of lutein and zeaxanthin supplementation on cognitive function in adults with self-reported mild cognitive complaints: A randomized, double-blind, placebo-controlled study. Frontiers in Nutrition, 9, 843512. https://doi.org/10.3389/fnut.2022.843512
Lopresti, A. L., Smith, S. J., Pouchieu, C., Pourtau, L., Gaudout, D., Pallet, V., & Drummond, P. D. (2023). Effects of a polyphenol-rich grape and blueberry extract (Memophenol™) on cognitive function in older adults with mild cognitive impairment: A randomized, double-blind, placebo-controlled study. Frontiers in Psychology, 14, 1144231. https://doi.org/10.3389/fpsyg.2023.1144231
MacLeod, C. (1991). Half a century of research on the Stroop effect: An integrative review. Psychological Bulletin, 109(2), 163–203. https://doi.org/10.1037/0033-2909.109.2.163
Manly, T., Robertson, I. H., Galloway, M., & Hawkins, K. (1999). The absent mind: Further investigations of sustained attention to response. Neuropsychologia, 37(6), 661–670. https://doi.org/10.1016/S0028-3932(98)00127-4
McVay, J. C., & Kane, M. J. (2009). Conducting the train of thought: Working memory capacity, goal neglect, and mind wandering in an executive-control task. Journal of Experimental Psychology: Learning, Memory, and Cognition, 35(1), 196–204. https://doi.org/10.1037/a0014104
Meyer, E. M., & Reynolds, M. R. (2018). Scores in space: Multidimensional scaling of the WISC-V. Journal of Psychoeducational Assessment, 36(6), 562–575. https://doi.org/10.1177/0734282917696935
Miller, K. M., Price, C. C., Okun, M. S., Montijo, H., & Bowers, D. (2009). Is the N-back task a valid neuropsychological measure for assessing working memory? Archives of Clinical Neuropsychology, 24(7), 711–717. https://doi.org/10.1093/arclin/acp063
Miskowiak, K. W., Johnsen, S., Sattler, S. M., Nielsen, S., Kunalan, K., Rungby, J., Lapperre, T., & Porsberg, C. M. (2021). Cognitive impairments four months after COVID-19 hospital discharge: Pattern, severity and association with illness variables. European Neuropsychopharmacology, 46, 39–48. https://doi.org/10.1016/j.euroneuro.2021.03.019
Miskowiak, K. W., Bech, J. L., Henriksen, A. C., Johnsen, S., Podlekareva, D., & Marner, L. (2022). Cerebral metabolic rate of glucose and cognitive tests in long COVID patients. Brain Sciences, 13(1), 23. https://doi.org/10.3390/brainsci13010023
Miskowiak, K. W., Pedersen, J. K., Gunnarsson, D. V., Roikjer, T. K., Podlekareva, D., Hansen, H., Dall, C. H., & Johnsen, S. (2023). Cognitive impairments among patients in a long-COVID clinic: Prevalence, pattern and relation to illness severity, work function and quality of life. Journal of Affective Disorders, 324, 162–169. https://doi.org/10.1016/j.jad.2022.12.122
Mitchell, A. J., Beaumont, H., Ferguson, D., Yadegarfar, M., & Stubbs, B. (2014). Risk of dementia and mild cognitive impairment in older people with subjective memory complaints: meta-analysis. Acta Psychiatrica Scandinavica, 130(6), 439–451. https://doi.org/10.1111/acps.12336
Miyake, A., & Friedman, N. P. (2012). The nature and organization of individual differences in executive functions: Four general conclusions. Current Directions in Psychological Science, 21(1), 8–14. https://doi.org/10.1177/0963721411429458
Miyake, A., Friedman, N. P., Emerson, M. J., Witzki, A. H., Howerter, A., & Wager, T. D. (2000). The unity and diversity of executive functions and their contributions to complex “frontal lobe” tasks: A latent variable analysis. Cognitive Psychology, 41(1), 49–100. https://doi.org/10.1006/cogp.1999.0734
Molteni, E., Butti, M., Bianchi, A. M., & Reni, G. (2008). Activation of the prefrontal cortex during a visual n-back working memory task with varying memory load: A near infrared spectroscopy study. Annual International Conference of the IEEE Engineering in Medicine and Biology Society, 2008, 4024–4027. https://doi.org/10.1109/iembs.2008.4650092
Monsell, S. (2003). Task switching. Trends in Cognitive Sciences, 7(3), 134–140. https://doi.org/10.1016/S1364-6613(03)00028-7
Montgomery, C., & Fisk, J. E. (2007). Everyday memory deficits in ecstasy-polydrug users. Journal of psychopharmacology, 21(7), 709–717. https://doi.org/10.1177/0269881107077220
Nagels, G., Geentjens, L., Kos, D., Vleugels, L., D’hooghe, M. B., Asch, P. V., Vuylsteke, K., & De Deyn, P. P. (2005). Paced visual serial addition test in multiple sclerosis. Clinical Neurology and Neurosurgery, 107(3), 218–222. https://doi.org/10.1016/j.clineuro.2004.11.016
Nikravesh, M., Jafari, Z., Mehrpour, M., Kazemi, R., AmiriShavaki, Y., Hossienifar, S., & Azizi, M. P. (2017). The paced auditory serial addition test for working memory assessment: Psychometric properties. Medical Journal of The Islamic Republic of Iran, 31(1), 349–354. https://doi.org/10.14196/mjiri.31.61
Parsons, S., Kruijt, A.-W., & Fox, E. (2019). Psychological science needs a standard practice of reporting the reliability of cognitive-behavioral measurements. Advances in Methods and Practices in Psychological Science, 2(4), 378–395. https://doi.org/10.1177/2515245919879695
Peers, P. V., Punton, S. F., Murphy, F. C., Watson, P., Bateman, A., Duncan, J., Astle, D. E., Hampshire, A., & Manly, T. (2022). A randomized control trial of the effects of home-based online attention training and working memory training on cognition and everyday function in a community stroke sample. Neuropsychological Rehabilitation, 32(10), 2603–2627. https://doi.org/10.1080/09602011.2021.1972817
Pollina, L. K., Greene, A. L., Tunick, R. H., & Puckett, J. M. (1992). Dimensions of everyday memory in young adulthood. British Journal of Psychology, 83(3), 305–321. https://doi.org/10.1111/j.2044-8295.1992.tb02443.x
Polsinelli, A. J., Kaszniak, A. W., Glisky, E. L., & Ashish, D. (2020). Effects of a brief, online, focused attention mindfulness training on cognition in older adults: A randomized controlled trial. Mindfulness, 11(5), 1182–1193. https://doi.org/10.1007/s12671-020-01329-2
Redick, T. S., Broadway, J. M., Meier, M. E., Kuriakose, P. S., Unsworth, N., Kane, M. J., & Engle, R. W. (2012). Measuring working memory capacity with automated complex span tasks. European Journal of Psychological Assessment, 28(3), 164–171. https://doi.org/10.1027/1015-5759/a000123
Redick, T. S., & Lindsey, D. R. B. (2013). Complex span and n-back measures of working memory: A meta-analysis. Psychonomic Bulletin & Review, 20(6), 1102–1113. https://doi.org/10.3758/s13423-013-0453-9
Reitan, R. M. (1955). The relation of the Trail Making Test to organic brain damage. Journal of Consulting Psychology, 19(5), 393–394. https://doi.org/10.1037/h0044509
Reynolds, C. R. (1997). Forward and backward memory span should not be combined for clinical analysis. Archives of Clinical Neuropsychology, 12(1), 29–40. https://doi.org/10.1016/S0887-6177(96)00015-7
Richter, K. M., Mödden, C., Eling, P., & Hildebrandt, H. (2015). Working memory training and semantic structuring improves remembering future events, not past events. Neurorehabilitation and Neural Repair, 29(1), 33–40. https://doi.org/10.1177/1545968314527352
Righi, S., Mecacci, L., & Viggiano, M. P. (2009). Anxiety, cognitive self-evaluation and performance: ERP correlates. Journal of Anxiety Disorders, 23(8), 1132–1138. https://doi.org/10.1016/j.janxdis.2009.07.018
Robertson, I. H., Manly, T., Andrade, J., Baddeley, B. T., & Yiend, J. (1997). “Oops!”: Performance correlates of everyday attentional failures in traumatic brain injured and normal subjects. Neuropsychologia, 35(6), 747–758. https://doi.org/10.1016/S0028-3932(97)00015-8
Robins, C. J., Keng, S.-L., Ekblad, A. G., & Brantley, J. G. (2012). Effects of mindfulness-based stress reduction on emotional experience and expression: A randomized controlled trial. Journal of Clinical Psychology, 68(1), 117–131. https://doi.org/10.1002/jclp.20857
Robison, M. K., & Unsworth, N. (2017). Individual differences in working memory capacity predict learned control over attentional capture. Journal of Experimental Psychology: Human Perception & Performance, 43(11), 1912–1924. https://doi.org/10.1037/xhp0000419
Rodgers, J. (2000). Cognitive performance amongst recreational users of “ecstasy.” Psychopharmacology, 151(1), 19–24. https://doi.org/10.1007/s002130000467
Rodriguez, C., Ruggero, C. J., Callahan, J. L., Kilmer, J. N., Boals, A., & Banks, J. B. (2013). Does risk for bipolar disorder heighten the disconnect between objective and subjective appraisals of cognition? Journal of Affective Disorders, 148(2/3), 400–405. https://doi.org/10.1016/j.jad.2012.06.029
Rossiter, S., Stevens, C., & Walker, G. (2006). Tinnitus and its effect on working memory and attention. Journal of Speech, Language, and Hearing Research, 49(1), 150–160. https://doi.org/10.1044/1092-4388(2006/012)
Ryskin, R. A., Benjamin, A. S., Tullis, J., & Brown-Schmidt, S. (2015). Perspective-taking in comprehension, production, and memory: An individual differences approach. Journal of Experimental Psychology: General, 144(5), 898–915. https://doi.org/10.1037/xge0000093
Shah, P., & Miyake, A. (1996). The separability of working memory resources for spatial thinking and language processing: An individual differences approach. Journal of Experimental Psychology: General, 125(1), 4–27. https://doi.org/10.1037/0096-3445.125.1.4
Shakeel, M. K., & Goghari, V. M. (2017). Measuring fluid intelligence in healthy older adults. Journal of Aging Research, 2017, 8514582. https://doi.org/10.1155/2017/8514582
Simon, J. R., & Rudell, A. P. (1967). Auditory S-R compatibility: The effect of an irrelevant cue on information processing. Journal of Applied Psychology, 51(3), 300–304. https://doi.org/10.1037/h0020586
Slagboom, T. N. A., Deijen, J. B., Van Bunderen, C. C., Knoop, H. A., & Drent, M. L. (2021). Impaired neuropsychological functioning in patients with hypopituitarism. Endocrinology, Diabetes & Metabolism, 4(1), e00165. https://doi.org/10.1002/edm2.165
SlanaOzimič, A., & Repovš, G. (2020). Visual working memory capacity is limited by two systems that change across lifespan. Journal of Memory and Language, 112, 104090. https://doi.org/10.1016/j.jml.2020.104090
Smilek, D., Carriere, J. S., & Cheyne, J. A. (2010). Failures of sustained attention in life, lab, and brain: Ecological validity of the SART. Neuropsychologia, 48(9), 2564–2570. https://doi.org/10.1016/j.neuropsychologia.2010.05.002
Spearman, C. (1910). Correlation calculated from faulty data. British Journal of Psychology, 1904–1920, 3(3), 271–295. https://doi.org/10.1111/j.2044-8295.1910.tb00206.x
Steinbusch, C. V. M., van Heugten, C. M., Rasquin, S. M. C., Verbunt, J. A., & Moulaert, V. R. M. (2017). Cognitive impairments and subjective cognitive complaints after survival of cardiac arrest: A prospective longitudinal cohort study. Resuscitation, 120, 132–137. https://doi.org/10.1016/j.resuscitation.2017.08.007
Stroop, J. R. (1935). Studies of interference in serial verbal reactions. Journal of Experimental Psychology, 18(6), 643–662. https://doi.org/10.1037/h0054651
Swick, D., & Ashley, V. (2017). Enhanced attentional bias variability in post-traumatic stress disorder and its relationship to more general impairments in cognitive control. Scientific Reports, 7(1), 14559–14559. https://doi.org/10.1038/s41598-017-15226-7
Thomson, K. J., & Goodhew, S. C. (2021). The relationship between the subjective experience of real-world cognitive failures and objective target-detection performance in visual search. Cognition, 217, 104914. https://doi.org/10.1016/j.cognition.2021.104914
Tipper, S. P., & Baylis, G. C. (1987). Individual differences in selective attention: The relation of priming and interference to cognitive failure. Personality and Individual Differences, 8(5), 667–675. https://doi.org/10.1016/0191-8869(87)90064-X
Toh, W. X., & Yang, H. (2022). Common executive function predicts reappraisal ability but not frequency. Journal of Experimental Psychology: General, 151(3), 643–664. https://doi.org/10.1037/xge0001099
Toplak, M. E., West, R. F., & Stanovich, K. E. (2013). Practitioner review: Do performance-based measures and ratings of executive function assess the same construct? Journal of Child Psychology and Psychiatry, 54(2), 131–143. https://doi.org/10.1111/jcpp.12001
Torenvliet, C., Groenman, A. P., Radhoe, T. A., Agelink van Rentergem, J. A., Van der Putten, W. J., & Geurts, H. M. (2022). Parallel age-related cognitive effects in autism: A cross-sectional replication study. Autism Research, 15(3), 507–518. https://doi.org/10.1002/aur.2650
Turner, M. L., & Engle, R. W. (1989). Is working memory capacity task dependent? Journal of Memory and Language, 28(2), 127–154. https://doi.org/10.1016/0749-596X(89)90040-5
Unsworth, N., Brewer, G. A., & Spillers, G. J. (2012). Variation in cognitive failures: An individual differences investigation of everyday attention and memory failures. Journal of Memory and Language, 67(1), 1–16. https://doi.org/10.1016/j.jml.2011.12.005
Vaerenbergh, Y. V. (2018). Effect size calculation in meta analysis. https://www.youtube.com/watch?v=CF05obuNis4
Vallesi, A., Tronelli, V., Lomi, F., & Pezzetta, R. (2021). Age differences in sustained attention tasks: A meta-analysis. Psychonomic Bulletin & Review, 28(6), 1755–1775. https://doi.org/10.3758/s13423-021-01908-x
van Doorn, J., van den Bergh, D., Böhm, U., Dablander, F., Derks, K., Draws, T., Etz, A., Evans, N. J., Gronau, Q. F., Haaf, J. M., Hinne, M., Kucharský, Š, Ly, A., Marsman, M., Matzke, D., Gupta, A. R. K. N., Sarafoglou, A., Stefan, A., Voelkel, J. G., & Wagenmakers, E.-J. (2021). The JASP guidelines for conducting and reporting a Bayesian analysis. Psychonomic Bulletin & Review, 28(3), 813–826. https://doi.org/10.3758/s13423-020-01798-5
Verbruggen, F., & Logan, G. D. (2008). Response inhibition in the stop-signal paradigm. Trends in Cognitive Sciences, 12(11), 418–424. https://doi.org/10.1016/j.tics.2008.07.005
Vom Hofe, A., Mainemarre, G., & Vannier, L.-C. (1998). Sensitivity to everyday failures and cognitive inhibition: Are they related? European Review of Applied Psychology/Revue Européenne de Psychologie Appliquée, 48(1), 49–56.
Wallace, J. C., Kass, S. J., & Stanny, C. (2001). Predicting performance in ‘Go’ situations: A new use for the Cognitive Failures Questionnaire? North American Journal of Psychology, 3(3), 481–490.
Wallace, J. C., & Vodanovich, S. J. (2003). Can Accidents and Industrial Mishaps Be Predicted? Further Investigation into the Relationship Between Cognitive Failure and Reports of Accidents. Journal of Business and Psychology, 17(4), 503–514. https://doi.org/10.1023/a:1023452218225
Welhaf, M. S., & Kane, M. J. (2024). A nomothetic span approach to the construct validation of sustained attention consistency: Re-analyzing two latent-variable studies of performance variability and mind-wandering self-reports. Psychological Research, 88(1), 39–80. https://doi.org/10.1007/s00426-023-01820-0
Werheid, K., Hoppe, C., Thöne, A., Müller, U., Müngersdorf, M., & von Cramon, D. Y. (2002). The Adaptive Digit Ordering Test: Clinical application, reliability, and validity of a verbal working memory test. Archives of Clinical Neuropsychology, 17(6), 547–565.
Whyte, J., Grieb-Neff, P., Gantz, C., & Polansky, M. (2006). Measuring sustained attention after traumatic brain injury: Differences in key findings from the Sustained Attention to Response Task (SART). Neuropsychologia, 44(10), 2007–2014. https://doi.org/10.1016/j.neuropsychologia.2006.02.012
Wolfe, J. M., Horowitz, T. S., & Kenner, N. M. (2005). Rare items often missed in visual searches. Nature, 435(7041), 439–440. https://doi.org/10.1038/435439a
Woltering, S., Liu, Z., Rokeach, A., & Tannock, R. (2013). Neurophysiological differences in inhibitory control between adults with ADHD and their peers. Neuropsychologia, 51(10), 1888–1895. https://doi.org/10.1016/j.neuropsychologia.2013.06.023
Woods, D., Pebler, P., Johnson, D. K., Herron, T., Hall, K., Blank, M., Geraci, K., Williams, G., Chok, J., Lwi, S., Curran, B., Schendel, K., Spinelli, M., & Baldo, J. (2024). The California Cognitive Assessment Battery (CCAB). Frontiers in Human Neuroscience, 17, 1305529. https://doi.org/10.3389/fnhum.2023.1305529
Wright, L., Lipszyc, J., Dupuis, A., Thayapararajah, S. W., & Schachar, R. (2014). Response inhibition and psychopathology: A meta-analysis of go/no-go task performance. Journal of Abnormal Psychology, 123(2), 429–439. https://doi.org/10.1037/a0036295
Yoshimura, T., Osaka, M., Osawa, A., & Maeshima, S. (2023). The classical backward digit span task detects changes in working memory but is unsuitable for classifying the severity of dementia. Applied Neuropsychology: Adult, 30(5), 528–534. https://doi.org/10.1080/23279095.2021.1961774
Zanesco, A. P., Denkova, E., Rogers, S. L., MacNulty, W. K., & Jha, A. P. (2019). Mindfulness training as cognitive training in high-demand cohorts: An initial study in elite military servicemembers. In N. Srinivasan (Ed.), Progress in brain research (pp. 323–354). ScienceDirect. https://doi.org/10.1016/bs.pbr.2018.10.001
Funding
Open Access funding enabled and organized by CAUL and its Member Institutions.
Author information
Authors and Affiliations
Contributions
Goodhew: conceptualization, methodology, software, investigation, formal analysis, writing – original draft, writing - review and editing, Edwards: conceptualization, methodology, writing – review and editing.
Corresponding author
Ethics declarations
Conflicts of interest/competing interests
Not applicable.
Ethics approval
Not applicable.
Consent to participate
Not applicable.
Consent for publication
Not applicable.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Below is the link to the electronic supplementary material.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Goodhew, S.C., Edwards, M. A meta-analysis on the relationship between subjective cognitive failures as measured by the cognitive failures questionnaire (CFQ) and objective performance on executive function tasks. Psychon Bull Rev (2024). https://doi.org/10.3758/s13423-024-02573-6
Accepted:
Published:
DOI: https://doi.org/10.3758/s13423-024-02573-6