Our ability to predict future events benefits many tasks. For example, visual search efficiency is increased when the locations of key elements in a display are repeated across trials (Chun & Jiang, 1998) and when target objects are semantically congruent with their surroundings (Biederman, Mezzanotte, & Rabinowitz, 1982). Similarly, validly cueing the location of a future target significantly enhances both perceptual and response-related target processing, whereas the opposite is true for targets presented at unexpected locations (Henderson, 1991; Nakayama & Mackeben, 1989; Posner, 1980; Posner, Snyder, & Davidson, 1980; Yeshurun & Carrasco, 1999). These same patterns of performance carry over to real-world scenarios, such as motor vehicle operation (Engström, Aust, & Vistrom, 2010; Ho, Tan, & Spence, 2005).

In addition to knowing where a target may appear, and what it might be, knowing when a target will appear is also of great value. For example, a seminal study by Coull and Nobre (1998) found that cues that accurately predicted the onset of a target benefited detection in a similar way to a cue indicating the target’s spatial location. Correa, Lupiáñez, and Tudela (2005) observed increased discrimination sensitivity for targets presented at a cued temporal interval relative to targets presented at unexpected times. Finally, validly cueing a target’s onset can allow it to escape the negative consequences of inattention (e.g., Hilkenmeier & Scharlau, 2010; Martens & Johnson, 2005; Shen & Alain, 2011; Tang, Badcock, & Visser, 2014; Visser, Tang, Badcock, & Enns, 2014).

Although these studies have provided evidence that temporal cues can help participants overcome resource limitations, it is still unclear whether participants can generate predictions about when targets will appear on the basis of temporal regularities present in the task environment. Much of the previous research in this area has been done in the context of the rapid serial visual presentation (RSVP; e.g., Broadbent & Broadbent, 1987; Potter, 1976; Raymond, Shapiro, & Arnell, 1992) paradigm. Here, stimuli are presented rapidly (~10 Hz) at a central, fixated location and participants are instructed to detect or identify one or more targets (for variations, see Visser, Zuvic, Bischof, & Di Lollo, 1999; Ward, Duncan, & Shapiro, 1997). When the RSVP stream contains two targets, performance on the first (T1) is typically quite good, whereas performance on the second (T2) is poor at intertarget intervals (lags) of less than about 500 ms. This so-called attentional blink (AB) arises from an inability to adequately allocate processing resources to T2 as a result of the need to process T1 (e.g., Chun & Potter, 1995; Di Lollo, Kawahara, Ghorashi, & Enns, 2005; Olivers & Meeter, 2008; Raymond et al., 1992; Taatgen, Juvina, Schipper, Borst, & Martens, 2009; Wyble, Bowman, & Nieuwenstein, 2009; see Dux & Marois, 2009, and Martens & Wyble, 2010, for reviews).

Several studies have shown that conveying explicit information about the expected onset of T2 can reduce the AB deficit. For example, Hilkenmeier and Scharlau (2010) found that T2 accuracy increased when T1 identity signaled the most likely T1–T2 lag, whereas Martens and Johnson (2005; see also Visser et al., 2014) found similar results using a cue presented prior to each trial. Shen and Alain (2011) showed that instructing participants at the beginning of a block of trials to focus attention at a specific time interval following T1 strongly benefited T2 accuracy when it appeared at that interval. Finally, both Choi and colleagues (Choi, Chang, Shibata, Sasaki, & Watanabe, 2012; Choi & Watanabe, 2014) and Tang et al. (2014) showed a reduction in the AB following a training regimen consisting of several hundred trials presented at a constant intertarget lag with a uniquely colored T2, whereas Nieuwenstein, Chun, van der Lubbe, and Hooge (2005) found that identification of a unique, red T2 was facilitated by the presentation of an earlier red item in the RSVP stream.

To date, however, the evidence that participants can derive processing benefits from temporally predictable events in the AB task has been equivocal. On the one hand, Badcock, Badcock, Fletcher, and Hogben (2013) found that the AB was ameliorated when T1 was preceded by a distractor stream of a predictable duration. This implies that being able to reliably predict T1 onset aids T2 processing. On the other hand, Martens and Johnson (2005) failed to find benefits of presenting a series of trials at a constant intertarget lag, and Visser et al. (2014) obtained a benefit under these conditions only when participants were explicitly alerted to its presence. This implies that a reliably predictive task environment is, by itself, insufficient to allow participants to make predictions about when T2 will be presented. Consistent with this conjecture, in their second experiment, Hilkenmeier and Scharlau (2010) found greater temporal cuing benefits when participants were informed that T1 identity predicted T1–T2 lag than when participants were unaware of this contingency. The dependence of these effects on explicit instructions is surprising when considered with other evidence that humans are highly sensitive to statistical patterns and regularities, and that they are able to use this information without instruction in many tasks to selectively process task-relevant features (Turk-Browne, Scholl, Chun, & Johnson, 2009; Zhao, Al-Aidroos, & Turk-Browne, 2013).

To investigate this issue, we capitalized on a phenomenon known as the foreperiod effect (Gabay & Henik, 2008, 2010; Niemi & Näätänen, 1981), in which observers respond more quickly and accurately to a target stimulus when it is preceded by a warning signal that can be used to predict target onset. This improvement appears to arise at multiple levels, including the motor system (Burle, Tandonnet, & Hasbroucq, 2010; Sanders, 1998), response selection mechanisms (Hackley, 2009; Los & Schut, 2008), and perception (Bausenhart, Rolke, & Ulrich, 2008; Correa et al., 2005; Rolke & Hofmann, 2007). A foreperiod arises in most conventional AB experiments because all trials contain two targets that are presented with equal frequency at each potential lag. As a result, probabilistic information about the occurrence of T2 is contained in the time that has elapsed since the onset of T1. That is, presentation of T1 serves as a warning that T2 is upcoming, with the likelihood of T2 appearing increasing directly with the passage of time since T1. To illustrate, if there are 25 trials at each of four lags, the probability of the target appearing at the first lag is .25 (25/100); however, given that it does not appear at the first lag, the probability that the target will appear in the second lag increases to .33 (25/100 – 25), the probability at the third lag increases to .50 [25/(100 – 25 – 25)], and if the target has not appeared by the third lag, the probability that it will appear at the longest lag is 1 [25/(100 – 25 – 25 – 25)]. This is referred to as an “aging” distribution with regard to lag (Näätänen, 1970).

In Experiment 1, we tested whether participants are sensitive to this temporal cue for T2 by comparing performance on an “aging” distribution of trials (all lags equally frequent)—which is the distribution of lags across trials conventionally used in the AB task—with performance on a “nonaging” distribution of trials (Näätänen, 1970) in which the number of trials at each lag decreased by half with each step in lag. With this modification, the probability of T2 onset remains constant as the time elapsed since T1 increases. Thus, if participants are sensitive to the temporal information given by the aging distribution then the AB should be larger (i.e., T2 accuracy should be poorer) in the nonaging condition relative to the aging condition, with this difference being greatest at the longest lag, where the greatest disparity in temporal information is provided by the passage of time from T1 onset.

Although this manipulation of trial frequency across lags allowed us to vary the temporal information provided by the passage of time following T1, it also introduced a second possible source of temporal information, by altering the a priori probability of trials occurring at a given lag. If participants are sensitive to this information, they should be able to predict the likely lag that T2 will be presented at on each trial, in a manner similar to that of the participants in Martens and Johnson (2005) and Visser et al. (2014), who received every trial at the same lag. Thus, we would expect that the AB should be smaller in the nonaging than in the aging condition at the shortest lag, whereas the opposite pattern should be found at the longest lag, reflecting the fact that these lags have the largest disparity in a priori probabilities between conditions.

Experiment 1

Method

Participants

Twenty undergraduate students (eight female, 12 male; mean age 20.15 years) were recruited in exchange for partial credit toward course completion. All provided informed consent prior to participating, had normal or corrected-to-normal vision and were naïve about the purpose of the experiment.

Apparatus and stimuli

The stimuli were presented on a 19-in. (viewing size: 18 in.) Dell M992 monitor running at a refresh rate of 100 Hz, attached to a Pentium computer running Presentation software (Version 16.20; Neurobehavioral Systems), located in a dimly lit room. The software also recorded responses from the computer keyboard.

All stimuli subtended a visual angle of approximately 1° at a viewing distance of 60 cm. Targets were dark gray (RGB: 70, 70, 70) and consisted of all letters of the English alphabet except I, O, Q, and Z, which were omitted due to their structural similarity to the digits 1, 0, 2, and 7. Distractors were light gray (RGB: 250, 250, 250) and consisted of the digits 1–9. Targets and distractors were presented in 28-point Arial font.

Procedure

Participants completed two blocks of trials in counterbalanced order. In the “aging” condition, a conventional AB paradigm was presented, with 60 trials at each of four lags. In the “nonaging” condition, the proportion of trials decreased as lag increased, such that 128 trials were presented at lag 2, 64 trials at lag 3, 32 trials at lag 5, and 16 trials at lag 7. In both conditions, 16 “catch” trials were also included, in which the second target was omitted and replaced with a distractor digit. This yielded a total of 256 trials in both conditions. Given this distribution of trials, in the aging condition, the probability that T2 would be presented (given that it had not already appeared) at lag 2 was .23, at lag 2 was .31, at lag 5 was .44, and at lag 7 was .79; in the nonaging condition, the probability that T2 would be presented (given that it had not already appeared) was .50 across all lags. Participants were informed that some trials would contain only one target and to respond appropriately, but they were not told about variations in the numbers of trials per lag.

Each trial began with a fixation cross at the center of the screen. When participants were fixated on this cross, they pressed the space bar to initiate an RSVP stream. Each item in the stream was presented for 30 ms and separated by a 60-ms blank interstimulus interval. The RSVP began with a randomly determined five to eight distractor digits. The digits were chosen randomly, with the proviso that the same digit could not appear consecutively. Next, the first target letter was presented, followed by one, two, four, or six additional distractor digits, and then the second target letter. This yielded intertarget stimulus onset asynchronies of 180, 270, 450, and 630 ms (lags 2, 3, 5, and 7, respectively). Targets were chosen randomly, with the proviso that they were different letters. The last item in the RSVP was always a distractor digit. Following the disappearance of this digit, a 200-ms blank screen was presented, and then participants were prompted to type in the letters that they had seen during the trial. Participants were instructed to enter the letters in any order, to guess if they were sure that a letter had appeared but they did not know its identity, or to press the space bar if they thought that the second target had not been presented. After making their responses, the fixation cross reappeared, and participants could initiate the next trial.

Results

The main finding was that T2 accuracy was significantly greater at the longest lag (lag 7) in the aging than in the nonaging condition. However, T2 accuracy was not significantly greater in the nonaging than in the aging condition at the shortest lag (lag 2). This pattern of results suggests that participants were sensitive to the temporal cues provided by the statistical patterns in the aging conditions (i.e., the foreperiod effect), but not to differences in the a priori probabilities of trials at each lag, as had been found in earlier studies by Martens and Johnson (2005) and Visser et al. (2014).

The mean target accuracy was first calculated separately for each lag in each condition. Only performance on the last 80 % of trials was considered, on the grounds that differences in probability could only be established after exposure to a number of trials. That said, the pattern of results reported below was also found when all trials were considered in the analysis. As has been customary in many studies of the AB, second-target accuracy was calculated only on trials in which T1 had been identified correctly, in an effort to ensure that the first target had been attended (Raymond et al., 1992). Preliminary examination of the data showed that one participant had an overall T1 accuracy more than 2.5 standard deviations below the group mean, and thus the data from this participant were omitted from further analyses. This left a sample of 19 participants.

T1 performance

Mean T1 accuracy scores, shown in Table 1, were submitted to a Condition (aging, nonaging) × Lag (2, 3, 5, 7) within-subjects analysis of variance (ANOVA). The analysis revealed no significant main effects or interactions (ps > .76, η 2s < .022), indicating that neither trial distribution nor lag influenced T1 accuracy.

Table 1 Mean percentage accuracies for T1 and T2 | T1, separated by condition and lag in Experiment 1

T2 performance

In order to examine the impact of condition on the AB, we first calculated an index of AB magnitude separately for each lag by subtracting T2 | T1 scores from T1 scores (as is shown in Table 1). These means are illustrated in Fig. 1. We chose to analyze AB magnitude rather than conditional T2 accuracy on the grounds that it better isolates the contribution of attention to T2 performance (although the same pattern of results was obtained if conditional T2 accuracy was analyzed). Conditional T2 accuracy reflects the specific costs associated with attending to T1 as well as more general perceptual and cognitive abilities required to perceive items presented in rapid succession. Because the impact of these general abilities can be estimated by considering T1 accuracy, it then follows that calculating a difference score between T1 and T2 at each lag will isolate the contribution of attentional availability to T2 performance. We submitted these difference scores to a Condition × Lag within-subjects ANOVA, which yielded a significant main effect of lag, F(3, 54) = 15.80, p < .001, η 2 = .47, indicating robust ABs in both conditions, a marginally significant main effect of condition, F(1, 18) = 3.42, p = .081, η 2 = .16, and a nonsignificant Lag × Condition interaction, F(3, 54) = 1.49, p > .22, η 2 < .08.

Fig. 1
figure 1

Magnitudes of the attentional blink deficit (AB, calculated by subtracting T2 | T1 from T1 accuracy separately at each lag) in Experiment 1, plotted as a function of intertarget lag. Solid circle symbols represent AB magnitudes in the aging condition, and solid square symbols represent AB magnitudes in the nonaging condition. Error bars represent one standard error of the mean

Although the interaction was nonsignificant, on the basis of past literature, we hypothesized in the introduction that a significant difference in T2 accuracy across conditions should be found at the shortest lag, due to the large change in the a priori probability of T2 across conditions, and at the longest lag, due to the foreperiod effect. For this reason, and because differences specific to these two lags might have been obscured in the context of the overall ANOVA that included all four lags run in the experiment, we performed additional analyses specifically comparing performance in the nonaging and aging conditions at lags 2 and 7. As is suggested by examination of Fig. 1, AB magnitude was significantly greater in the nonaging than in the aging condition at lag 7, t(18) = 2.82, p = .011. However, this difference was not significant at lag 2, t(18) = 0.84, p > .41.

The lack of a significant difference in performance between conditions at lag 2 suggests that participants were insensitive to changes in the a priori probability of trials at particular lags. This result parallels earlier findings by Martens and Johnson (2005) and Visser et al. (2014), who both showed that AB magnitude was unaffected by presenting a block of trials at a consistent lag when participants were uninformed of this contingency. On the other hand, the pattern of performance at lag 7 suggests that participants are sensitive to statistical pattern information that arises from using equal numbers of trials at each lag. When this information was eliminated by reducing the number of trials at lag 7 in the nonaging condition, the AB magnitude increased significantly.

That said, it is important to note that the greater AB magnitude in the nonaging condition at lag 7 could have arisen either due to the elimination of statistical pattern information or because of a reduction in the a priori probability of trials occurring at this lag. Thus, it is not yet clear whether the performance differences at lag 7 are evidence for participants’ sensitivity to the temporal information provided by statistical patterns in the aging condition or to reductions in the a priori probability of lag-7 trials in the nonaging condition. To address this question, in Experiment 2, we replicated the paradigm used in Experiment 1, but eliminated catch trials. This change equated the temporal information provided by the passage of time following T1 at lag 7 (in both conditions, the probability of T2 appearing at lag 7, given that it had not appeared at an earlier lag, was 1.0), while maintaining differences in the a priori probabilities of lag 7 trials.

The predictions were straightforward. If the results at lag 7 in Experiment 1 were indicative of sensitivity to temporal information in aged distributions, then the difference in AB magnitudes between the aging and nonaging conditions should be eliminated in Experiment 2. On the other hand, if the differences at lag 7 in Experiment 1 were due to sensitivity to differences in the a priori probability of trials at a given lag, then AB magnitude should remain greater in the nonaging condition in Experiment 2.

Experiment 2

Method

Participants

Twenty undergraduate students (13 female, seven male; mean age 18.38 years) were recruited in exchange for partial credit toward course completion. All provided informed consent prior to participating, had normal or corrected-to-normal vision, and were naïve as to the purpose of the experiment. None had participated in Experiment 1.

Apparatus and stimuli

The apparatus and stimuli were identical to those of Experiment 1, except that stimuli were presented on a 24-in. BenQ XL2420TE monitor.

Procedure

The procedure was similar to that of Experiment 1, except that catch trials were eliminated, yielding a total of 240 trials in each condition. As in Experiment 1, 60 trials were presented at each lag in the aging condition. Given this distribution of trials, the probability that T2 would be presented at lag 2 was .25, at lag 3 was .33, at lag 5 was .50, and at lag 7 was 1.0. In the nonaging condition, 124 trials were presented at lag 2, 64 trials at lag 3, 32 trials at lag 5, and 20 trials at lag 7. Given this distribution of trials, the probability that T2 would be presented at lag 2 (given that it had not already appeared) was .52, at lag 3 was .51, at lag 5 was .52, and at lag 7 was 1.0. Note that the a priori probability of T2 being presented at lag 7 was .25 in the aging condition and .083 in the nonaging condition. This closely matched the a priori probability of lag-7 trials in the aging and nonaging conditions in Experiment 1, as well as the magnitude of the probability difference between these conditions.

Results

The main finding was that AB magnitudes did not differ between conditions at any lag. This pattern of results suggests that participants are sensitive to the statistical patterns inherent in aging distributions and that this sensitivity drove the performance differences seen at lag 7 in Experiment 1. Additionally, as in Experiment 1, despite similar differences in a priori probabilities between the aging and nonaging conditions at lag 2, we again found no differences in AB magnitude.

As in Experiment 1, the mean target accuracy was first calculated separately for each lag in each condition, considering only the last 80 % of trials. Preliminary examination of the data showed that one participant had an overall T1 accuracy more than 2.5 standard deviations below the group mean, and thus the data from this participant were omitted from further analyses. This left a sample of 19 participants.

T1 performance

Mean T1 accuracy scores, shown in Table 2, were submitted to a Condition × Lag within-subjects ANOVA. The analysis revealed no significant main effects or interactions (ps > .23, η 2s < .08), indicating that neither trial distribution nor lag influenced T1 accuracy.

Table 2 Mean percentage accuracies for T1 and T2 | T1, separated by condition and lag in Experiment 2

T2 performance

AB magnitude scores were submitted to a Condition × Lag within-subjects ANOVA. As is suggested by examination of these scores in Fig. 2, this analysis yielded only a significant main effect of lag, F(3, 54) = 14.67, p < .001, η 2 = .45, indicating robust ABs in both conditions. No other main effects or interactions were significant (ps > .26, η 2s < .07). Follow-up t tests comparing performance between the conditions separately at lags 2 and 7 also were nonsignificant [lag 2, t(18) = 0.27, p > .78; lag 7, t(18) = 1.07, p > .29].

Fig. 2
figure 2

Magnitudes of the attentional blink deficit (AB, calculated by subtracting T2 | T1 from T1 accuracy separately at each lag) in Experiment 2, plotted as a function of intertarget lag. Solid circle symbols represent AB magnitudes in the aging condition, and solid square symbols represent AB magnitudes in the nonaging condition. Error bars represent one standard error of the mean

Considered together with the findings from Experiment 1, these results provide clear evidence that participants are sensitive to the temporal information given by the passage of time since T1 in the aging condition. This indicates that participants can derive temporal cues from statistical patterns and can use this information to ameliorate the AB, in much the same way that participants are sensitive to other statistical patterns that communicate information about the likely location and identity of upcoming sensory inputs (Turk-Browne et al., 2009).

The results from Experiment 2 also bolster the evidence obtained in Experiment 1 and in earlier studies suggesting that participants are insensitive to temporal information derived from differences in the a priori probabilities of trials occurring at particular lags. One possible explanation for this insensitivity is that the differences in probability are not sufficiently salient to be recognized. Support for this option has come from Visser et al. (2014), who found that informing participants about the presence of blocks of trials at a consistent lag increased T2 accuracy. Another possibility is that probability-linked effects are too small to be detected using a measure of T2 accuracy. In light of this option, a more sensitive alternative measure might be T2 response times (RTs). The rationale for this suggestion is twofold. First, RTs largely avoid potential problems with scale constraints—in particular, ceiling effects. This is particularly salient here, because T2 accuracy was relatively high across lags. Second, previous studies had suggested that temporal information can significantly benefit response-related processes (e.g., Coull & Nobre, 1998). Thus, any temporal information accrued from consideration of the a priori probability of trials at a particular lag might be more likely to benefit performance when T2 requires a speeded judgement.

To test this second possibility, we modified the AB task to require a speeded two-alternative forced choice discrimination for T2 (presented without a trailing mask), allowing us to estimate T2 processing time directly (Visser, 2007). This task is similar to those used by Jolicœur and Dell’Acqua (1998) and Zuvic, Visser, and Di Lollo (2000), who both found an AB reflected in a decrease in T2 RTs with increasing lag. In order to bolster our statistical power, we also conducted a power analysis (Faul, Erdfelder, Lang, & Buchner, 2007) using the effect sizes and intercorrelations amongst levels of the repeated measures variables obtained in Experiment 1, in order to estimate the sample size required for approximately .95 power. This yielded an estimate of 44 participants. Of chief interest was whether using a more sensitive RT measure, combined with a larger sample size, would yield evidence for changes in AB magnitude reflective of the a priori probability of trials occurring at a given lag.

Experiment 3

Method

Participants

Forty-five undergraduate students (30 female, 15 male; mean age 19.62 years) were recruited in exchange for partial credit toward course completion. All provided informed consent prior to participating, had normal or corrected-to-normal vision, and were naïve as to the purpose of the experiment. None had participated in Experiment 1 or 2.

Apparatus and stimuli

The apparatus and stimuli were identical to those of Experiment 2, except that some participants viewed stimuli on a 19-in. Acer AC716 monitor.

Procedure

The procedure was identical to that of Experiment 1, with three exceptions. First, T2 was equally likely to be a G or C on each trial, whereas T1 was additionally constrained to exclude G or C. Second, the last distractor was eliminated, so that each trial ended with T2. Third, participants were instructed to respond as quickly and accurately as possible to the appearance of T2 by pressing an appropriately marked key on the keyboard and then to respond to T1 at their leisure when prompted. On trials on which T2 was omitted, participants were asked to wait for the T1 prompt without responding. This prompt appeared 2,000 ms following the presentation of T2 if no key was pressed.

Results

The main finding was a crossover interaction in RTs, such that they were faster at lag 2 in the nonaging condition (over the aging condition), and faster at lag 7 in the aging condition (over the nonaging condition). The faster nonaging RTs at lag 2 indicate sensitivity to the a priori probability of T2 occurring at this lag. The faster aging RTs at lag 7 provide converging evidence that the distribution of trials in this condition provides a reliable statistical pattern that can be used to aid T2 processing as the time since T1 onset increases.

As in Experiments 1 and 2, the mean target accuracy and median T2 RTs were calculated separately for each lag in each condition, using only the final 80 % of trials. Mean T1 and T2 | T1 accuracy scores are shown in Table 3. The data from four participants whose overall T1 accuracy scores were more than 2.5 standard deviations below the group mean were omitted from further analysis. The data from one participant whose overall T2 accuracy score in the aging condition was more than four standard deviations below the group mean were also omitted.

Table 3 Mean percentage accuracies for T1 and T2 | T1, separated by condition and lag in Experiment 3

T1 performance

Mean T1 accuracy scores were submitted to a Condition × Lag within-subjects ANOVA. The analysis did not reveal any significant main effects or interactions (ps > .21, η 2s < .04), indicating no effect of trial distribution or lag on T1 accuracy.

T2 performance (accuracy)

As can be seen in Table 3, mean T2 | T1 accuracy scores were much higher than in previous experiments, due to the omission of the mask after T2. These means were submitted to Condition × Lag within-subjects ANOVA that revealed only a main effect of lag, F(3, 117) = 3.04, p = .032, η 2 = .07, consistent with small ABs across both conditions. No other significant main effects or interactions (ps > .10, η 2s < .06) were obtained.

T2 performance (adjusted RTs)

Trials on which RTs were less than 200 ms or greater than 2,000 ms were excluded from the analysis, since they indicated that participants had either anticipated target onset or failed to respond to T2 before the prompt requesting T1 identity had been displayed. This resulted in the omission of 1.10 % of trials across all conditions. In order to examine the impact of trial probability on the AB, we calculated adjusted RTs separately at each lag by dividing the median RTs by the proportion of correct responses. The purpose of adjusted RTs is to provide a more sensitive, composite index of target perceptibility that takes into account both response speed and criterion (Chambers, Stokes, Janko, & Mattingley, 2006; Townsend & Ashby, 1983). This overcomes the conventional requirement to consider accuracy and RTs separately, and the associated uncertainties about which is the superior indicator of perceptual sensitivity. Notably, the pattern of adjusted RTs reported below was mirrored when considering RTs alone.

We submitted the adjusted RTs to a Condition × Lag within-subjects ANOVA. The main effect of condition was not significant, F(1, 39) = 0.04, p > .83, η 2 < .01. However, as is evidenced in Fig. 3, we found a main effect of lag, F(3, 117) = 39.51, p < .001, η 2 = .50, consistent with an overall drop in adjusted RTs with increasing lag. This is the signature of the AB deficit seen in previous research (Jolicœur & Dell’Acqua, 1998; Zuvic et al., 2000). Moreover, there was a significant Condition × Lag interaction, F(3, 117) = 6.65, p < .001, η 2 = .15. As in earlier analyses, we followed up the ANOVA by comparing performance across the aging and nonaging conditions at lags 2 and 7. Examination of Fig. 2 suggests that performance was faster in the nonaging condition than in the aging condition at lag 2, but that this pattern reversed at lag 7. Consistent with this impression, significant differences in adjusted RTs emerged at lag 2, t(39) = 1.73, p = .046 (one-tailed), and at lag 7, t(39) = 2.03, p = .049 (two-tailed). This pattern replicates the findings of Experiment 1, showing that participants are sensitive to the temporal cues provided by statistical patterns arising in the aging condition. The results also show evidence that participants are sensitive to the temporal information provided by differences in the numbers of trials presented at each lag across the experiment. However, even when we boosted the likelihood that these cues would be helpful by requiring a speeded response to T2, the improvements were quite modest.

Fig. 3
figure 3

Mean adjusted response times (RTs) in Experiment 3, plotted as a function of intertarget lag. Solid circle symbols represent performance in the aging condition, and solid square symbols represent performance in the nonaging condition. Error bars represent one within-subjects standard error of the mean, calculated using the Cousineau–Morey method (Cousineau, 2005; Morey, 2008)

General discussion

A number of studies have examined the interaction between temporal cues and attentional limits using the AB paradigm, in which processing of a second target is impaired by the requirement to attend to an immediately preceding item. This work has shown reliable effects of explicit temporal cues (e.g., Hilkenmeier & Scharlau, 2010; Martens & Johnson, 2005), as well as salient events within the task that indicate the likely appearance of T2 (e.g., Nieuwenstein et al., 2005). However, results have been mixed when participants have not received salient temporal cues and/or no explicit instruction (Badcock et al., 2013; Martens & Johnson, 2005; Tang et al., 2014). This study has clarified the apparent lack of consistency on this issue by demonstrating that participants can use the predictive temporal information contained in the statistical structure of a trial block in order to overcome limitations of attention.

Here, we investigated the source of these conflicting results by examining whether AB task performance is influenced by a foreperiod effect (Niemi & Näätänen, 1981). We compared a standard AB task condition (in which all potential intertarget lags occurred with equal frequencies) with one in which trials at shorter lags were more frequent. In the standard design, the factor of Lag occurs with an “aging” distribution, such that T2 is increasingly likely to occur as time passes from the onset of T1. This was compared with a “nonaging” distribution, in which the frequency of T2 trials at each lag was reduced by approximately half as lag increased (Näätänen, 1970). This difference in frequency between conditions also produced variations in the a priori probability that a trial at a particular lag would be presented, similar to earlier experiments in which performance on blocks of trials presented at randomly varying lags was compared to performance on blocks of trials presented at a constant lag (Martens & Johnson, 2005; Visser et al., 2014). Thus, in our experiment, we were able to assess participants’ sensitivity to two different types of temporal information embedded in the trial structure of the AB task.

On the whole, the present results are in line with previous studies showing that target performance is enhanced by the presence of a consistent interval prior to T1 (Badcock et al., 2013), as well as a broader literature suggesting humans are highly sensitive to statistical patterns and environmental regularities (Turk-Browne et al., 2009; Zhao et al., 2013). The results of Experiments 1 and 3 revealed a strong foreperiod effect at the longest lag in the aging condition, suggesting that participants could make some predictions about T2 onset on the basis of the temporal information afforded by the distribution of lags within a condition. With a more sensitive RT measure of T2 processing in Experiment 3, the results showed that participants could also benefit from temporal predictions based on the a priori probability of different lags across a block of trials. This indicates that the temporal information concerning lag that is contained in a block of trials can be used effectively even when these regularities are not communicated explicitly (Visser et al., 2014).

However, this conclusion does come with caveats. First, the fact that variations in the relative frequencies of an intertarget lag on T2 performance in Experiment 3 were detectable only with a more sensitive RT measure and larger sample size suggests that this type of temporal information leads to relatively subtle effects that may vary with the sensitivity of the performance measure and the nature of the underlying processes that it taps. Second, it is clear that participants are differentially sensitive to sources of temporal information contained in the statistical patterns present in the AB task. Whereas participants reliably showed foreperiod effects, probability variations were much more modest. Considering these results and past demonstrations of the benefits of explicitly informing participants about the existence of temporal information, it seems reasonable to conclude that explicit information likely boosts temporal cuing benefits but are not necessary for such effects to be found. Additionally, it is clear that future work will be needed to examine the interaction between participants’ knowledge about the presence of temporal information in a task and the type of temporal information in order to gain a more complete picture of how and when such information is used.

An obvious question is why temporal cue information arising from aging distributions was considerably stronger than those arising from variations in a priori probability. Some explanation for this difference may lie in the fact that foreperiod effects are mediated by multiple mechanisms, only one of which was likely to have been engaged by variations in a priori probability. The foreperiod effect has been suggested to arise from (a) increased alertness induced by the warning signal, in combination with increasing certainty about target appearance as time passes (Bertelson, 1967; Coull & Nobre, 1998); (b) deployment of attentional resources at the expected target interval (Correa, Lupiáñez, Milliken, & Tudela, 2004; Coull & Nobre, 1998); and (c) trace conditioning, wherein the warning signal initiates a preparatory RT locked to target appearance (Los & Heslenfeld, 2005; Los & Van den Heuvel, 2001). Notably, the operation of the first and last of these mechanisms requires a high degree of temporal certainty. However, this is unlikely to result from changes in a priori probability. For example, lag-2 trials in the nonaging condition still only occurred on approximately half of the trials, and this was little different from the corresponding lag in the aging condition (see below). As a result, temporal cues arising from variations in a priori probability are less likely than those arising from aging distributions to induce increased alertness or trace conditioning. In turn, cuing benefits are likely to be smaller in the former than in the latter case.

Another issue addressed by the present results concerns the degree to which the a priori probabilities of an event must differ in order for this information to serve as a temporal cue. In Martens and Johnson (2005), two lags were used and the cue was either present or absent. Thus, the a priori probability of a trial presented at the shortest lag in their experiment was either 1.0, with the cue present, or .50, when the cue was absent—a difference of .50. The present results augment this initial finding by showing that temporal information can also be gleaned from smaller variations in target predictability. For example, in Experiment 3, the a priori probability of a lag-2 trial was approximately .23 in the aging condition, and .50 in the nonaging condition—a difference about half that in Martens and Johnson. Future investigations into this issue would benefit from looking beyond group means, as we have done here, in order to investigate individual differences in sensitivity to the probability information available in the environment. This would provide a valuable new perspective on our ability to use the predictive information available from task structure to guide perception.

To the extent that participants can predict the onset of T2, how does this benefit performance? A variety of suggestions have been made in this regard. Martens and Johnson (2005) argued that when T2 was presented at a cued interval, it gained additional weighting in a postperceptual competition with adjacent items (particularly T1) for high-level resources. Alternatively, Visser et al. (2014) proposed that cueing T2 onset allows more efficient suppression of distractors by increasing the efficiency of a perceptual filtering process (Ghorashi, Zuvic, Visser, & Di Lollo, 2003; Visser, Bischof, & Di Lollo, 2004) that screens out distracting inputs prior to engagement with postperceptual resources. Although the present study did not address this issue directly, we believe the present results are consistent with distractor suppression, because the finding of a foreperiod effect at long lags in Experiments 1 and 3 likely occurred after T1 processing had already been completed. This leaves more effective suppression of the stream of distractors as a viable mechanism, in addition to the possibility that temporal cuing may benefit T2 in the competition for attentional resources with T1 at short lags.

As we noted in the introduction, all theories of the AB attribute the deficit to the requirement to process T1. However, there are important differences between these accounts in terms of the putative mechanisms underlying this relationship. According to bottleneck accounts, the AB arises because the processing resources allocated to T1 are unavailable for processing T2 when it is presented at shorter lags (e.g., Chun & Potter, 1995; Jolicœur & Dell’Acqua, 1998). On the other hand, according to distractor-based accounts, presentation of nontarget items following T1 initiates mechanisms designed to suppress perceptual processing in order to prevent interference with T1 (e.g., Olivers & Meeter, 2008; Taatgen et al., 2009; Wyble et al., 2009). This interferes with T2 processing at short intertarget intervals.

The present results indicate that limits on T2 processing can be overcome by virtue of the temporal information provided by statistical patterns in the AB task, even when these patterns are not explicitly communicated to participants. In past studies, it has been suggested that temporal cues reduce postperceptual competition between targets and distractors by boosting the strength of the target representation (Martens & Johnson, 2005) and/or by allowing perceptual resources to be focused on the expected time of T2 arrival, thereby avoiding selection of distractors interposed between T1 and T2. From these perspectives, the present results would seem most consistent with distractor-based accounts of the AB in suggesting that temporal cues reduce or eliminate the deleterious effects of post-T1 distraction.

A final important point to be made regards the implications of the foreperiod effects found here on conventional estimates of the duration of the AB. More than 20 years of studies have indicated that the the impact of T1 identification on the processing of T2 in neurotypical populations extends for approximately 500–700 ms. On this basis, the negative impact of stimulus processing on perception is conventionally assumed to cease after about 700 ms. However, almost without exception, the past studies upon which this conclusion has been based used intertarget lags that occurred with equal frequencies across experimental trials (i.e., an aging distribution with respect to lag). The evidence shown here clearly indicates that this practice leads to a foreperiod effect that has boosted accuracy at the longest lag in these experiments. In turn, this implies that T1 processing may still have been ongoing at the longest lag in past experiments (usually lag 7), but the deleterious effects of this processing may have been masked by participants’ ability to accurately predict the onset time for T2 at longer intertarget intervals. The upshot is that current estimates of the duration of “the blink” are likely to underestimate the true duration that perception is impacted by the processing of prior stimuli.