Abstract
Most models of language comprehension assume that the linguistic system is able to pre-activate phonological information. However, the evidence for phonological prediction is mixed and controversial. In this study, we implement a paradigm that capitalizes on the fact that foreign speakers usually make phonological errors. We investigate whether speaker identity (native vs. foreign) is used to make specific phonological predictions. Fifty-two participants were recruited to read sentence frames followed by a last spoken word which was uttered by either a native or a foreign speaker. They were required to perform a lexical decision on the last spoken word, which could be either semantically predictable or not. Speaker identity (native vs. foreign) may or may not be cued by the face of the speaker. We observed that the face cue is effective in speeding up the lexical decision when the word is predictable, but it is not effective when the word is not predictable. This result shows that speech prediction takes into account the phonological variability between speakers, suggesting that it is possible to pre-activate in a detailed and specific way the phonological representation of a predictable word.
Similar content being viewed by others
Introduction
Classic models of language comprehension have long recognized the significant role of prediction. Traditionally, it was believed that the parser could predict the part of speech (grammatical category) of forthcoming words in a sentence (Kimball, 1975). Within the garden path model, this information is used to predict the minimal grammatical completion of the fragment read so far (Staub & Clifton, 2006). However, there has been ongoing controversy surrounding whether and how the system also predicts specific words, at least in highly constraining contexts. Initially, skepticism surrounded this notion, as it was considered wasteful to predict incoming information, given the large number of sensible sentence continuations (Forster, 1981; Jackendoff, 2002; see also Van Petten & Luka, 2012). Nevertheless, a large body of empirical evidence has now dispelled this skepticism, embracing the idea that language comprehension involves context-based pre-activation of upcoming words, which facilitates the bottom-up processing of linguistic input (Altmann & Mirković, 2009; Dell & Chang, 2014; Kutas et al., 2011; Pickering & Garrod, 2007, 2013).
Despite the widespread acceptance that people implicitly predict upcoming linguistic information, the underlying processes are not yet fully understood. The debate focuses on the nature of the processes and the representations involved in prediction, as well as the circumstances under which prediction occurs. Traditionally it was assumed that the pre-activation of upcoming words arises from the passive spreading of activation between pre-existing representations, which are (partially) activated during the processing of the context (Anderson, 1983; Collins & Loftus, 1975; Hutchison, 2003; Huettig et al., 2022; McRae et al., 1997). More recent models of language comprehension do not consider spreading of activation as the only mechanism involved and highlight the pro-active aspect of internal prediction generation (Federmeier, 2007; Huettig, 2015; Kuperberg & Jaeger, 2016; Pickering & Gambi, 2018; Pickering & Garrod, 2013). In this context, prediction-by-production models have received particular attention. These models propose that prediction during comprehension can be implemented by using representations and mechanisms that are also employed in language production (Huettig, 2015; Pickering & Gambi, 2018; Pickering & Garrod, 2013).
Another issue in the language prediction literature concerns when and to what extent higher-level internal information can be used to pre-activate upcoming information. It has been proposed that predictive processes come with a cost and may only be implemented when the context is highly constraining and when sufficient resources and time are available (Pickering & Gambi, 2018), or when predictions are particularly useful (Kuperberg & Jaeger, 2016). Additionally, prediction models usually assume that highly constraining contexts allow for the pre-activation of sub-lexical information. However, while there is solid evidence in favor of semantic (Altmann & Kamide, 1999, 2007; Chambers et al., 2002; Federmeier & Kutas, 1999; Kamide et al., 2003; Metusalem et al., 2012; Paczynski & Kuperberg, 2012) and syntactic (Crocker, 2000; Kimball, 1975; Levy, 2008; Lewis, 2000; Staub & Clifton, 2006; Traxler, 2014; Traxler et al., 1998; van Gompel et al., 2005) pre-activation, the evidence for prediction of the phonological form of the upcoming word is not fully consistent (DeLong et al., 2005; Heilbron et al., 2022; Ito et al., 2016, 2017, 2018, 2020; Ito & Sakai, 2021; Martin et al., 2013; Nicenboim et al., 2020; Nieuwland et al., 2018).
The ERP study by DeLong et al. (2005) was the first to provide evidence of phonological prediction. They investigated the modulation of the N400 amplitude elicited by an indefinite article which could agree or disagree with the phonological form of the semantically predictable following noun. The results showed the N400 modulation at the presentation of the article, as a function of its expectancy. This indicates that participants predicted the phonological form of the word, leading to increased negativity when the article mismatches with this prediction. Despite the strong theoretical impact of the DeLong et al. (2005) study, subsequent research has tried to replicate the N400 modulation on the pre-target article with mixed results (Ito et al., 2017, 2020; Martin et al., 2013; Nicenboim et al., 2020; Nieuwland et al., 2018).
Another relevant challenge for phonological prediction is the acoustic variability of speech. Speakers vary in the physical realization of speech sounds (Liberman et al., 1967). This is particularly evident in non-native speech, which deviates not only phonetically but frequently also at the phonological level (Best et al., 2001; Clopper et al., 2005; Flege, 1988). Listeners, however, demonstrate the ability to adapt to non-native speech (Bradlow & Bent, 2008; Clarke & Garrett, 2004; Clopper & Bradlow, 2008; Maye et al., 2008), and imitation of a foreign accent seems to facilitate such perceptual adaptation (Adank et al., 2010).
The present study aims to provide evidence in favor of prediction at the phonological level by exposing listeners to misspelled target words uttered by a foreign-accented speaker. Participants were asked to read sentence frames in which the final word was uttered by either a native or a foreign speaker. The foreign speaker’s words contained a consistent phonological error in the first phoneme of the word. Before the experiment, participants were familiarized with both native and foreign speakers. The speaker’s face was presented along with a 1-min audio clip in which each speaker introduced themselves. Experimental trials consisted of reading a sentence frame that missed a last word to be grammatical. The last word was presented acoustically, and participants were required to perform an auditory lexical decision task on the spoken word. The last word could be either predictable or not based on the preceding sentential context. A further critical manipulation was that the speaker’s accent could be either predictable or not, as the written sentences were presented in association with the speaker’s face or with a neutral visual stimulus (see Fig. 1).
One key feature of the present paradigm is the choice to present the sentential context in written form. This was driven by the theoretical requirement of dissociating the effect of the prediction of foreign-accented phonology from the possible cognitive load increase associated with processing the sentence context with an unusual (foreign) phonology. Previous studies often confounded these two effects by presenting both the context and the target words with a native or a foreign accent (Brunellière & Soto-Faraco, 2013; Porretta et al., 2020; Romero-Rivas et al., 2016). In contrast, written form presentation ensures that the cognitive load associated with the processing of the context is matched for both the native and the non-native speaker.
The use of the speaker face cue during reading is expected to facilitate the perception and recognition of the target word. Specifically, we hypothesize a facilitatory effect of the face cue in the lexical decision response times only when the context is highly constraining, as this should allow the prediction of the semantic and phonological features of the target word. If the effect is to be attributed to prediction, no effect of face manipulation on response times should emerge in the low-constraint condition. We are also interested in investigating whether the use of information about the speaker’s accent to predict upcoming speech differs between the native (expecting standard phonology) and the foreign accent (expecting a specific deviant phonology).
Method
Participants
Fifty-four adults (40 females, meanage = 23.80 years, SD = 2.97) were recruited. Participants were native Italian speakers with no history of neurological, language-related, or psychiatric disorders. Two participants were excluded from the analyses due to a recruitment error (one person had participated in the stimulus norming cloze probability test, and the other was not a native Italian speaker). The final sample consisted of 52 participants (40 females, mean age = 23.67 years, SD = 2.96). Since no reliable estimate of the effect sizes of our interest was available, the sample size was determined a priori based on the recommendation that regression analyses should include five to ten observations per variable to give an acceptable estimate of regression coefficients (Hair et al., 2010; Tabachnick & Fidell, 1989). It is important to note that the total number of observations in linear mixed-effects models includes both the number of participants and the number of observations nested within each participant per variable (Bates et al., 2015). The research adhered to the principles outlined in the Declaration of Helsinki. Participants provided their informed consent before participating in the experiment. The research protocol was approved by the Ethics Committee for Psychological Research of the University of Padova (protocol number: 5181).
Materials
The materials are available via the Open Science Framework (OSF) repository of the current project (https://osf.io/n42x9/). The target stimuli consisted of 64 spoken words (mean length = 6.23 phonemes, SD = 1.97) and 64 spoken non-words (mean length = 6.45 phonemes, SD = 1.98) beginning with the phonemes /r/, /p/ and /k/. These three phonemes were never present in any other position of words and non-words. Given that in the foreign accent condition participants were asked to discriminate mispronounced words and non-words, non-words had no phonological neighbors and could be easily identified as such. Each target word was preceded by a written sentential context that could be either High constraining (example in 1a) or Low constraining (example in 1b) toward the target word.
-
1a)
Nello stagno gracidano le rane
In-the pond croak the frogs
‘In the pond the frogs croak’
-
1b)
Al ristorante ho mangiato le rane
At-the restaurant have eaten the frogs
‘At the restaurant I ate frogs’
To determine the constraining level, an online sentence completion questionnaire was administered to 22 participants, who did not take part in the experiment. They were instructed to complete each sentence frame with the first word that came to mind. The sentence constraint was operationalized as the proportion of total responses involving the most frequent continuation (HC constraint: mean = 0.94, SD = 0.07; LC constraint: mean = 0.18, SD = 0.08). The target word in High-Constraint sentence frames was the most frequent continuation. The target word in Low-Constraint sentences was always a semantically plausible continuation. The sentence frames varied in length (mean = 9.00 words, SD = 2.00, range = 4–14), but their length was matched between conditions (HC sentences: mean = 9.14 words, SD = 2.12, range = 4–14; LC sentences: mean = 8.86 words, SD = 1.88, range= 4–13; p = .429). An equal number of written sentences, matched for length (HC sentences: mean = 9.31, SD = 2.05, range= 5–15; LC sentences: mean = 8.83, SD = 1.90, range= 5–15; p = .167) and level of constraint (HC constraint: mean = 0.90, SD = 0.11; LC constraint: mean = 0.22, SD = 0.08) to the sentences preceding words, were used to precede the nonwords.
The speech stimuli were spoken by an artificial voice with a native accent in one condition and a voice with a foreign accent in the other condition. All stimuli of the native and foreign conditions were synthesized using the Microsoft Azure text-to-speech service. Microsoft Azure allows to synthesize speech stimuli using the prebuilt neural voices of speakers of different languages. We used two different speakers for the native and foreign-accented conditions. For the native-accented voice, the prebuilt neural voice of the Italian speaker Fabiola was selected. For the foreign-accented voice, the prebuilt neural voice of the Indian speaker Neerja was selected. The two speakers were selected with the aim of making participants associate a given face stimuli with a foreign/native speaker voice. To control for participant’s familiarity with different foreign accents, the foreign accent was artificially created by modifying three phonemes (/r/, /p/ and /k/) produced by the foreign-accented voice as /l/, /b/ and /ɢ/, respectively. The novel foreign accent was created by manipulating the place or manner of articulation of the target phonemes. This choice was aimed at obtaining a foreign accent that was easily perceived by the listener but at the same time sounded natural. To minimize the impact of the phonetic variability between the native and the foreign voice, all stimuli were synthesized starting from the IPA encoding of the words and non-words in Italian. The phonological manipulation of the foreign-accented speaker was implemented by changing the target phonemes in the IPA encoding of the stimuli (e.g., the target stimulus was /kˈaldo/ for the native speaker voice and /ɢ'aldo/ for the foreign speaker voice). In this way, both the native and the foreign speakers received the same phonetic sequences as input, thus, there were no other phonetic approximations or differences between the voices except for the phonological manipulation. The foreign-accented voice mispronounced the initial phoneme of all target stimuli. The mispronounced words did not correspond to any existing Italian word. The phonological manipulation was implemented to synthesize both the experimental stimuli and the familiarization speech of the foreign speaker. The native speaker voice and the foreign-accented voice differ in prosody, but this was only perceivable in the familiarization phase where participants were exposed to lengthy sections of text.
Procedure and design
The experiment was carried out using Psychopy (Peirce et al., 2019). Participants were tested in a quiet room while wearing headphones. In the first familiarization phase, participants viewed a picture of the speaker’s face and listened to a one-minute speech in which the speaker introduced herself. Two different speeches of the same length were prepared, each associated with a different face. These speeches were synthesized either with the native-accented voice or with the foreign-accented voice. The computer screen displayed a face, either an Indian-looking female face (for the foreign-accented speaker) or an Italian-looking female face (for the native speaker). Both speeches were presented to participants in two parts of 30 s each, alternating between speakers. Then the experimental stage followed. The participants were instructed to read the sentence frames displayed on the screen and to judge whether the subsequent auditory target was a word or not. The auditory target could be pronounced by the familiarized native speaker or the familiarized foreign-accented speaker. The speaker’s face was presented 2,500 ms before the sentence, 4.5 cm below the center of the screen, and remained visible throughout the whole trial. In half of the trials, the face was replaced by a control stimulus, which was a scrambled version of the faces of the two speakers. Both the face and the control stimulus were 10 cm wide and 10 cm high. The presentation of each sentence started with a fixation point appearing 4.5 cm above the center of the screen for 50 ms. The sentence frames were then presented phrase-by-phrase, with each phrase presented for 800 ms, followed by a 150-ms inter-phrase interval. The auditory target stimuli (word or nonword) were presented 800 ms after the presentation of the last phrase of the sentence. Participants were instructed to categorize the spoken targets as words or non-words by pressing the ‘M’ or ‘C’ keys on the keyboard with their left and right index. They were asked to respond to words with the index finger of their dominant hand. In the case of foreign accent, participants were explicitly asked to accept mispronounced words as real words. Response times were recorded from the presentation of the target stimulus for a maximum time of 2,000 ms after the end of the stimulus. To encourage participants to read the sentences, a comprehension question requiring a forced choice yes / no response was presented in 10% of trials (20% of trials with words as target) after the lexical decision. The participants used the same keys as in the lexical decision, using the finger of the dominant hand to provide the “Yes” response. Before starting the experimental session, participants completed 12 practice trials that were not part of the experimental materials. The session lasted about 45 min.
Each participant was presented with 256 trials, 128 ending with a word and 128 with a non-word. Target words and non-words appeared in High and Low constraining sentence frames. The experimental material was divided into two blocks, ensuring that each word or non-word appeared only once per block. Within each block, 64 speech stimuli (32 words, 32 non-words) were spoken by the native speaker and 64 by the foreign speaker. The assignment of speech stimuli to the native or foreign speaker was counterbalanced between blocks. The speaker’s accent was either cued or not cued by the speaker’s face, resulting in 16 observations per cell. The order of the two blocks was counterbalanced between participants and the order of the trials within blocks was randomized. To ensure that each stimulus was presented in both native and foreign accents, and with or without the speaker’s face, four experimental lists were created, in which materials were rotated between conditions in a Latin square design. Participants were randomly assigned to one of these lists.
Statistical analyses
The statistical analyses were performed using the statistical software R (R Core Team, 2023). The complete dataset and analyses scripts can be found in the OSF repository (https://osf.io/n42x9/). First, a preliminary accuracy check was performed, demonstrating that all participants achieved an accuracy level above 80% in both the lexical decision and comprehension questions. Only responses to words were considered (see Online Supplementary Materials for non-word data). Response accuracy was analyzed using generalized linear mixed-effects models (binomial distribution with logit link). Response times (RTs) of correct responses were log-transformed and analyzed using linear mixed-effects models (Gaussian distribution). All models were fitted with the lme4 package (Bates et al., 2015). Responses faster than 150 ms and slower than 2,500 ms were excluded from the analyses (0.007% of total observations). To find the best-fitting model for our data, we used a hierarchical model comparison approach (Heinze et al., 2018). The model comparison was based on the Akaike Information Criterion (AIC) and especially delta AIC and AIC weight as indexes of the goodness of fit. The AIC and AIC weight gives information on the models’ relative evidence (i.e., likelihood and parsimony), therefore the model with the lowest AIC and the highest AIC weight is to be preferred (Wagenmakers & Farrell, 2004). For both accuracy and RTs, model comparison included a null model with Participant and Item as random intercepts to account for participant-specific variability and item-specific idiosyncrasies (Baayen et al., 2008). Random slopes were not included due to the failure of tested models to converge. Predictor’s order was established giving priority to the main effects over interactions and to the effects (i.e., Accent and Constraint) with substantial support in the literature (Faust & Kravetz, 1998; Federmeier et al., 2007; Floccia et al., 2006, 2009; Munro & Derwing, 1995). Face and the related interaction terms were added subsequently in order to determine to what extent this variable increased the fitting of the model. Thus, inclusion of predictors followed this order: (i) Accent (Native vs. Foreign); (ii) Constraint (HC vs. LC); (iii) Face (Face vs. No Face); (iv) The two-way interaction between Constraint*Accent; (v) The two-way interaction between Constraint*Face; (vi) The three-way interaction between Constraint*Accent*Face.
Outliers were identified using the outlierTest function of the car package (Fox & Weisberg, 2019) and removed (0.0006 % of model observations). Sum coding was used as contrast coding in order to estimate main effects (Brehm & Alday, 2022). Post hoc comparisons were performed using the contrast function of the emmeans package (Lenth et al., 2023). P-values were adjusted using Bonferroni correction (Bonferroni, 1936).
Results
Accuracy
Descriptive statistics for accuracy according to conditions are reported in Table 1.
As shown in Table 2, model comparison indicates that the best-fitting model (lower delta AIC and higher AIC weight) for accuracy is Model 3:
Model estimates for the best-fitting model for accuracy are reported in Table 3. The effect of Accent indicates that correct responses are less likely for foreign compared to native accent. The effect of Constraint indicates that correct responses are more likely in High-Constraint compared to Low-Constraint sentences. Finally, the effect of Face indicates that correct responses are more likely when the speaker’s face is present compared to when it is not present.
Response times
Descriptive statistics for response times (ms) according to conditions are reported in Table 4.
As shown in Table 5, model comparison indicates that the best fitting model (lower delta AIC and higher AIC weight) for response times is Model 5:
Model estimates for the best-fitting model for response times are reported in Table 6. The interaction between Constraint*Accent indicates that the Constraint effect, namely faster RTs for HC sentences compared to LC sentences, is larger for the foreign compared to the native accent. Importantly, the interaction between Constraint*Face indicates that cueing the speaker’s face is associated with an increased Constraint effect (Fig. 2).
Post hoc comparisons have shown that cueing the speaker’s face is associated with faster RTs for HC sentences (p < .001) but not for LC sentences (p = .709). We found no evidence that this effect is modulated by the speaker’s accent (Native or Foreign) since the three-way interaction between Constraint*Accent*Face did not improve the model fit.
Discussion
In the present study, we aimed to investigate the role of context information in predicting upcoming words, focusing on whether predictions occur at the phonological level of representation. To do so, we took advantage of the fact that foreign speakers often exhibit phonological errors and examined whether the prediction system is flexible enough to account for the phonological variability between speakers. In our experimental paradigm, the speaker’s face provides a cue to the phonological properties of the upcoming word before its presentation. The results showed that cueing the speaker’s face speeded up RTs only when the word was predictable. This facilitation seems to occur regardless of whether the face cued a foreign or a native accent. These results provide compelling evidence supporting the involvement of phonological representations in prediction and suggest that predictions rely on flexible and finely tuned processes capable of accommodating interindividual phonological variability. Previous experiments have shown that the prediction system is able to adapt to the specific speaker. For example, the extent to which comprehenders rely on predictive processing appears to be influenced by the reliability of the speaker (Brothers et al., 2017, 2019). Our findings extend this flexibility to predictions encompassing not only the semantic content of the speech but also the phonological form of the words.
Previous evidence indicated that phonological predictions were present when the sentential context was produced by a native speaker, but not when it was produced by a foreign speaker (Brunellière & Soto-Faraco, 2013). This pattern was attributed to less precise priors for prediction or a lower frequency of occurrence of the phonological variants in the mental lexicon when dealing with unfamiliar phonological contexts (Connine et al., 2008). However, our experiment showed that cueing the speaker's accent facilitated the recognition of words regardless of whether the accent was native or foreign. A possible explanation for these seemingly conflicting results could be due to the greater cognitive load associated with the processing of the sentential context when pronounced by a non-native speaker (Adank et al., 2009; Cristia et al., 2012; Floccia et al., 2006, 2009; Porretta et al., 2016, 2020). It has been shown that the availability of cognitive resources impacts predictive processes (Ding et al., 2023; Huettig & Mani, 2016; Otten & Van Berkum, 2009), and this occurs even in the comprehension of the second language (Ito et al., 2018). Thus, the reduction of phonological prediction found by Brunellière and Soto-Faraco (2013) is not incompatible with our results, but it just reveals a very different aspect of the interplay between predictive processing and non-standard speech. When the system does not have to deal with the uncertainty associated with the decoding of the speech of a non-native speaker very precise phonological predictions of the forthcoming speech can be made.
Our data also shed light on the processes involved in generating predictions. Cueing the speaker's face resulted in faster RTs only for predictable words, suggesting that this effect cannot be solely attributed to the priming of talker-specific representations (Creel et al., 2008; Creel & Bregman, 2011; Goldinger, 1996; Nygaard & Pisoni, 1998; Palmeri et al., 1993; Remez et al., 1997). Rather, our result appears to be specific to prediction processes based on sentential constraints. In this case, predictions seem to be generated through mechanisms that exploit all available linguistic and extralinguistic elements to anticipate the input, resulting in predictions that are tailored to the speaker. This result can hardly be explained by the spreading of activation among phonological abstract representations stored in long-term memory. The foreign speaker in our study produced words with phonological errors, and a pre-existing lexical representation of such phonological forms (namely lexemes) should not be available or at least it should be much less available compared to the phonological form of words spoken with standard phonology. Prediction-by-production accounts offer a valuable framework for understanding how listeners actively use the information about the speaker's phonological categories to generate predictions. According to these models, prediction during comprehension is supported by language production representations and mechanisms (Huettig, 2015; Pickering & Gambi, 2018; Pickering & Garrod, 2007, 2013), allowing listeners to generate predictions at different levels of representation, including the phonological level. Neurophysiology experiments using a paradigm similar to the one implemented here (Gastaldon et al., 2020, 2023; Piai et al., 2015), in which high- and low-constraining sentence frames were followed by a picture to be produced or by a target word to be perceived, showed very similar brain activations between production and comprehension. These results suggest that predicting the last word of high constraint sentences is at least partially subserved by the language production network. Although the behavioral paradigm implemented here does not explicitly support a prediction-by-production framework, our data are clearly compatible with this view. Covert imitation, a mechanism often emphasized by prediction-by-production models (Pickering & Gambi, 2018; Pickering & Garrod, 2013), could explain the flexibility of the prediction system. According to Pickering and Gambi (2018), covert imitation allows the transformation of comprehension representations into production representations. This mechanism could allow the system to generate predictions constrained by both the preceding sentential context and the speaker's accent.
The notion that covert imitation may adapt to the speaker’s accent aligns with findings demonstrating a direct relationship between overt speech imitation and speech comprehension. For instance, Adank et al. (2010) showed that participants who were instructed to overtly imitate a foreign accent demonstrated improved comprehension of foreign-accented sentences from background noise. Moreover, research in the field of speech perception has investigated the mechanisms by which listeners adapt to different speakers (Bradlow & Bent, 2008; Clarke & Garrett, 2004; Clopper & Bradlow, 2008; Maye et al., 2008; Weatherholtz & Jaeger, 2016). The system implicitly tracks and learns speaker-specific properties to optimally process the variations present in the environment and help listeners cope with talker variability (Kleinschmidt & Jaeger, 2015). For example, Nygaard et al. (1994) found that new words were recognized more accurately when produced by familiar speakers compared to new speakers. From this perspective, speech perception can be influenced by expectations about the speaker, such as their dialect background (Hay, Nolan et al., 2006a; Niedzielski, 1999), ethnicity (Casasanto, 2008), age (Drager, 2011; Hay, Warren et al., 2006b; Walker & Hay, 2011), and socio-economic status (Hay, Warren et al., 2006b). In our study, we propose that the perceptual adaptation to the speaker also extends to the predictive processes based on sentential constraint, possibly relying on production mechanisms.
Pickering and Garrod (2013) hypothesized that comprehenders prioritize prediction when they can predict accurately as in the case in which they can identify with the speaker. Therefore, in our experiment participants were expected to engage more in phonological prediction for the native speaker than for the foreign speaker. The results did not corroborate this expectation. A possible explanation might be associated with the systematic nature of our manipulation. The foreign-accented speech entailed altering three phonemes, and all target words began with one of these phonemes. Participants might have accurately anticipated the target phonology due to this structured pattern. Moreover, despite it is known that people tend to identify more with an in-group member, our paradigm was not aimed at manipulating this variable (likelihood of identification with the speaker) and thus it is a clearly suboptimal way to test this specific hypothesis of the Pickering and Garrod (2013) proposal.
The literature includes other prediction accounts proposing that comprehenders can actively predict upcoming speech without necessarily involving the production system. For instance, Kuperberg and Jaeger introduced a multi-representational hierarchical generative model (Kuperberg, 2016; Kuperberg & Jaeger, 2016) in which comprehenders rely upon internal generative models – a set of hierarchically organized internal representations - to probabilistically pre-activate information at multiple levels of representation. This pre-activation maximizes the probability of accurately recognizing the incoming information. Internal representations are built using both linguistic and non-linguistic information, and they may also include knowledge of the speaker’s sound structure (Connine et al., 1991; Szostak & Pitt, 2013). Listeners may learn different generative models corresponding to different statistical environments (Kleinschmidt & Jaeger, 2015), enabling them to consider the phonological variability between speakers when predicting upcoming words.
To conclude, comprehenders not only exhibit rapid adaptation to non-native speakers but also exploit the flexibility of the perceptual system to predict the upcoming speech even when it contains phonological errors. This provides valuable insights into both the level(s) of representation and the processes involved in generating predictions. Our results strongly support the notion that linguistic prediction involves the pre-activation of phonological representations, clearly showing that linguistic prediction processes go beyond the mere spreading of activation between long-term stored representations. Further research using not only a behavioral methodology (e.g., neurophysiology, neurostimulation or patient study) is warranted to gain a deeper understanding of the mechanisms involved in generating predictions at a sub-lexical level of representation and to determine the relative weight of these mechanisms. It would be also crucial to investigate the extent to which the effects found in the present study generalize to more ecological conversational settings, where participants engage in listening to contextual sentences. Finally, to better define what are the boundary conditions of the speaker effect in phonological predictions, it would be necessary to further develop paradigms aimed at controlling/manipulating the cognitive load requirements.
Data availability
The complete dataset and materials (including experimental materials and Supplementary materials) can be found in the Open Science Framework (OSF) repository (https://osf.io/n42x9/).
Code availability
The analyses script can be found in the OSF repository (https://osf.io/n42x9/).
References
Adank, P., Evans, B. G., Stuart-Smith, J., & Scott, S. K. (2009). Comprehension of familiar and unfamiliar native accents under adverse listening conditions. Journal of Experimental Psychology: Human Perception and Performance, 35(2), 520–529. https://doi.org/10.1037/a0013552
Adank, P., Hagoort, P., & Bekkering, H. (2010). Imitation Improves Language Comprehension. Psychological Science, 21(12), 1903–1909. https://doi.org/10.1177/0956797610389192
Altmann, G. T. M., & Kamide, Y. (1999). Incremental interpretation at verbs: restricting the domain of subsequent reference. Cognition, 73(3), 247–264. https://doi.org/10.1016/S0010-0277(99)00059-1
Altmann, G. T. M., & Kamide, Y. (2007). The real-time mediation of visual attention by language and world knowledge: Linking anticipatory (and other) eye movements to linguistic processing. Journal of Memory and Language, 57(4), 502–518. https://doi.org/10.1016/j.jml.2006.12.004
Altmann, G. T. M., & Mirković, J. (2009). Incrementality and Prediction in Human Sentence Processing. Cognitive Science, 33(4), 583–609. https://doi.org/10.1111/j.1551-6709.2009.01022.x
Anderson, J. R. (1983). The architecture of cognition. Psychology Press. https://doi.org/10.4324/9781315799438
Baayen, R. H., Davidson, D. J., & Bates, D. M. (2008). Mixed-effects modeling with crossed random effects for subjects and items. Journal of Memory and Language, 59(4), 390–412. https://doi.org/10.1016/j.jml.2007.12.005
Bates, D., Maechler, M., Bolker, B., Walker, S., Christensen, R. H. B., Singmann, H., & Krivitsky, P. (2015). Package “lme4.” Convergence, 12(1), 2.
Best, C. T., McRoberts, G. W., & Goodell, E. (2001). Discrimination of non-native consonant contrasts varying in perceptual assimilation to the listener’s native phonological system. The Journal of the Acoustical Society of America, 109(2), 775–794. https://doi.org/10.1121/1.1332378
Bonferroni, C. (1936). Teoria statistica delle classi e calcolo delle probabilita. Pubblicazioni Del R Istituto Superiore Di Scienze Economiche e Commericiali Di Firenze, 8, 3–62.
Bradlow, A. R., & Bent, T. (2008). Perceptual adaptation to non-native speech. Cognition, 106(2), 707–729. https://doi.org/10.1016/j.cognition.2007.04.005
Brehm, L., & Alday, P. M. (2022). Contrast coding choices in a decade of mixed models. Journal of Memory and Language, 125, 104334. https://doi.org/10.1016/j.jml.2022.104334
Brothers, T., Swaab, T. Y., & Traxler, M. J. (2017). Goals and strategies influence lexical prediction during sentence comprehension. Journal of Memory and Language, 93, 203–216. https://doi.org/10.1016/j.jml.2016.10.002
Brothers, T., Dave, S., Hoversten, L. J., Traxler, M. J., & Swaab, T. Y. (2019). Flexible predictions during listening comprehension: Speaker reliability affects anticipatory processes. Neuropsychologia, 135, 107225. https://doi.org/10.1016/j.neuropsychologia.2019.107225
Brunellière, A., & Soto-Faraco, S. (2013). The speakers’ accent shapes the listeners’ phonological predictions during speech perception. Brain and Language, 125(1), 82–93. https://doi.org/10.1016/j.bandl.2013.01.007
Casasanto, L. S. (2008). Does Social Information Influence Sentence Processing? Proceedings of the Annual Meeting of the Cognitive Science Society, 30. Retrieved from https://escholarship.org/uc/item/8dc2t2gf
Chambers, C. G., Tanenhaus, M. K., Eberhard, K. M., Filip, H., & Carlson, G. N. (2002). Circumscribing Referential Domains during Real-Time Language Comprehension. Journal of Memory and Language, 47(1), 30–49. https://doi.org/10.1006/jmla.2001.2832
Clarke, C. M., & Garrett, M. F. (2004). Rapid adaptation to foreign-accented English. The Journal of the Acoustical Society of America, 116(6), 3647–3658. https://doi.org/10.1121/1.1815131
Clopper, C. G., & Bradlow, A. R. (2008). Perception of Dialect Variation in Noise: Intelligibility and Classification. Language and Speech, 51(3), 175–198. https://doi.org/10.1177/0023830908098539
Clopper, C. G., Pisoni, D. B., & de Jong, K. (2005). Acoustic characteristics of the vowel systems of six regional varieties of American English. The Journal of the Acoustical Society of America, 118(3), 1661–1676. https://doi.org/10.1121/1.2000774
Collins, A. M., & Loftus, E. F. (1975). A spreading-activation theory of semantic processing. Psychological Review, 82(6), 407–428. https://doi.org/10.1037/0033-295X.82.6.407
Connine, C. M., Blasko, D. G., & Hall, M. (1991). Effects of subsequent sentence context in auditory word recognition: Temporal and linguistic constrainst. Journal of Memory and Language, 30(2), 234–250. https://doi.org/10.1016/0749-596X(91)90005-5
Connine, C. M., Ranbom, L. J., & Patterson, D. J. (2008). Processing variant forms in spoken word recognition: The role of variant frequency. Perception & Psychophysics, 70(3), 403–411. https://doi.org/10.3758/PP.70.3.403
Creel, S. C., & Bregman, M. R. (2011). How Talker Identity Relates to Language Processing. Language and Linguistics Compass, 5(5), 190–204. https://doi.org/10.1111/j.1749-818X.2011.00276.x
Creel, S. C., Aslin, R. N., & Tanenhaus, M. K. (2008). Heeding the voice of experience: The role of talker variation in lexical access. Cognition, 106(2), 633–664. https://doi.org/10.1016/j.cognition.2007.03.013
Cristia, A., Seidl, A., Vaughn, C., Schmale, R., Bradlow, A., & Floccia, C. (2012). Linguistic Processing of Accented Speech Across the Lifespan. Frontiers in Psychology, 3. https://doi.org/10.3389/fpsyg.2012.00479
Crocker, M. W. (2000). Wide-Coverage Probabilistic Sentence Processing. Journal of Psycholinguistic Research, 29(6), 647–669. https://doi.org/10.1023/A:1026560822390
Dell, G. S., & Chang, F. (2014). The P-chain: relating sentence production and its disorders to comprehension and acquisition. Philosophical Transactions of the Royal Society B: Biological Sciences, 369(1634), 20120394. https://doi.org/10.1098/rstb.2012.0394
DeLong, K. A., Urbach, T. P., & Kutas, M. (2005). Probabilistic word pre-activation during language comprehension inferred from electrical brain activity. Nature Neuroscience, 8(8), 1117–1121. https://doi.org/10.1038/nn1504
Ding, J., Zhang, Y., Liang, P., & Li, X. (2023). Modulation of working memory capacity on predictive processing during language comprehension. Language, Cognition and Neurosicence, 38(8), 1133–1152. https://doi.org/10.1080/23273798.2023.2212819
Drager, K. (2011). Speaker Age and Vowel Perception. Language and Speech, 54(1), 99–121. https://doi.org/10.1177/0023830910388017
Faust, M., & Kravetz, S. (1998). Levels of sentence constraint and lexical decision in the two hemispheres. Brain and Language, 62(2), 149–162. https://doi.org/10.1006/brln.1997.1892
Federmeier, K. D. (2007). Thinking ahead: The role and roots of prediction in language comprehension. Psychophysiology, 44(4), 491–505. https://doi.org/10.1111/j.1469-8986.2007.00531.x
Federmeier, K. D., & Kutas, M. (1999). A rose by any other name: Long-Term memory structure and sentence processing. Journal of Memory and Language, 41(4), 469–495. https://doi.org/10.1006/jmla.1999.2660
Federmeier, K. D., Wlotko, E. W., De Ochoa-Dewald, E., & Kutas, M. (2007). Multiple effects of sentential constraint on word processing. Brain Research, 1146, 75–84. https://doi.org/10.1016/j.brainres.2006.06.101
Flege, J. E. (1988). Factors affecting degree of perceived foreign accent in English sentences. The Journal of the Acoustical Society of America, 84(1), 70–79. https://doi.org/10.1121/1.396876
Floccia, C., Goslin, J., Girard, F., & Konopczynski, G. (2006). Does a regional accent perturb speech processing? Journal of Experimental Psychology: Human Perception and Performance, 32(5), 1276–1293. https://doi.org/10.1037/0096-1523.32.5.1276
Floccia, C., Butler, J., Goslin, J., & Ellis, L. (2009). Regional and foreign accent processing in english: Can listeners adapt? Journal of Psycholinguistic Research, 38(4), 379–412. https://doi.org/10.1007/s10936-008-9097-8
Forster, K. I. (1981). Priming and the effects of sentence and lexical contexts on naming time: Evidence for autonomous lexical processing. The Quarterly Journal of Experimental Psychology Section A, 33(4), 465–495. https://doi.org/10.1080/14640748108400804
Fox, J., & Weisberg, S. (2019). An R Companion to Applied Regression (3rd ed.). Sage.
Gastaldon, S., Arcara, G., Navarrete, E., & Peressotti, F. (2020). Commonalities in alpha and beta neural desynchronizations during prediction in language comprehension and production. Cortex, 133, 328–345. https://doi.org/10.1016/j.cortex.2020.09.026
Gastaldon, S., Busan, P., Arcara, G., & Peressotti, F. (2023). Inefficient speech-motor control affects predictive speech comprehension: Atypical electrophysiological correlates in stuttering. Cerebral Cortex, 33(11), 6834–6851. https://doi.org/10.1093/cercor/bhad004
Goldinger, S. D. (1996). Words and voices: Episodic traces in spoken word identification and recognition memory. Journal of Experimental Psychology: Learning, Memory, and Cognition, 22(5), 1166–1183. https://doi.org/10.1037/0278-7393.22.5.1166
Hair, J. F., Black, W. C., Babin, B. J., & Anderson, R. E. (2010). Multivariate Data Analysis (7th ed.). Pearson.
Hay, J., Warren, P., & Drager, K. (2006). Factors influencing speech perception in the context of a merger-in-progress. Journal of Phonetics, 34(4), 458–484. https://doi.org/10.1016/j.wocn.2005.10.001
Hay, J., Nolan, A., & Drager, K. (2006a). From fush to feesh: Exemplar priming in speech perception. The Linguistic Review, 23(3). https://doi.org/10.1515/TLR.2006.014
Heilbron, M., Armeni, K., Schoffelen, J.-M., Hagoort, P., & de Lange, F. P. (2022). A hierarchy of linguistic predictions during natural language comprehension. Proceedings of the National Academy of Sciences, 119(32). https://doi.org/10.1073/pnas.2201968119
Heinze, G., Wallisch, C., & Dunkler, D. (2018). Variable selection - A review and recommendations for the practicing statistician. Biometrical Journal, 60(3), 431–449. https://doi.org/10.1002/bimj.201700067
Huettig, F. (2015). Four central questions about prediction in language processing. Brain Research, 1626, 118–135. https://doi.org/10.1016/j.brainres.2015.02.014
Huettig, F., & Mani, N. (2016). Individual differences in working memory and processing speed predict anticipatory spoken language processing in the visual world. Language, Cognition and Neuroscience, 31(1), 80–93. https://doi.org/10.1080/23273798.2015.1047459
Huettig, F., Audring, J., & Jackendoff, R. (2022). A parallel architecture perspective on pre-activation and prediction in language processing. Cognition, 224, 105050. https://doi.org/10.1016/j.cognition.2022.105050
Hutchison, K. A. (2003). Is semantic priming due to association strength or feature overlap? A microanalytic review. Psychonomic Bulletin & Review, 10(4), 785–813. https://doi.org/10.3758/BF03196544
Ito, A., & Sakai, H. (2021). Everyday language exposure shapes prediction of specific words in listening comprehension: A visual world eye-tracking study. Frontiers in Psychology, 12. https://doi.org/10.3389/fpsyg.2021.607474
Ito, A., Corley, M., Pickering, M. J., Martin, A. E., & Nieuwland, M. S. (2016). Predicting form and meaning: Evidence from brain potentials. Journal of Memory and Language, 86, 157–171. https://doi.org/10.1016/j.jml.2015.10.007
Ito, A., Martin, A. E., & Nieuwland, M. S. (2017). How robust are prediction effects in language comprehension? Failure to replicate article-elicited N400 effects. Language, Cognition and Neuroscience, 32(8), 954–965. https://doi.org/10.1080/23273798.2016.1242761
Ito, A., Pickering, M. J., & Corley, M. (2018). Investigating the time-course of phonological prediction in native and non-native speakers of English: A visual world eye-tracking study. Journal of Memory and Language, 98, 1–11. https://doi.org/10.1016/j.jml.2017.09.002
Ito, A., Gambi, C., Pickering, M. J., Fuellenbach, K., & Husband, E. M. (2020). Prediction of phonological and gender information: An event-related potential study in Italian. Neuropsychologia, 136, 107291. https://doi.org/10.1016/j.neuropsychologia.2019.107291
Jackendoff, R. (2002). Foundations of Language. Oxford University PressOxford. https://doi.org/10.1093/acprof:oso/9780198270126.001.0001
Kamide, Y., Altmann, G. T. M., & Haywood, S. L. (2003). The time-course of prediction in incremental sentence processing: Evidence from anticipatory eye movements. Journal of Memory and Language, 49(1), 133–156. https://doi.org/10.1016/S0749-596X(03)00023-8
Kimball, J. (1975). Predictive analysis and over-the-top parsing. In J. Kimball (Ed.), Syntax and semantics (4th ed., pp. 155–179). Academic Press.
Kleinschmidt, D. F., & Jaeger, T. F. (2015). Robust speech perception: Recognize the familiar, generalize to the similar, and adapt to the novel. Psychological Review, 122(2), 148–203. https://doi.org/10.1037/a0038695
Kuperberg, G. R. (2016). Separate streams or probabilistic inference? What the N400 can tell us about the comprehension of events. Language, Cognition and Neuroscience, 31(5), 602–616. https://doi.org/10.1080/23273798.2015.1130233
Kuperberg, G. R., & Jaeger, T. F. (2016). What do we mean by prediction in language comprehension? Language, Cognition and Neuroscience, 31(1), 32–59. https://doi.org/10.1080/23273798.2015.1102299
Kutas, M., DeLong, K. A., & Smith, N. J. (2011). A Look around at What Lies Ahead: Prediction and Predictability in Language Processing. Predictions in the Brain (pp. 190–207). Oxford University Press. https://doi.org/10.1093/acprof:oso/9780195395518.003.0065
Lenth, R. V., Bolker, B., Buerkner, P., Gine-Vázquez, I., Herve, M., Jung, M., Love, J., Miguez, F., Riebl, H., Singmann, H. emmeans: Estimated Marginal Means, aka Least-Squares Means (Online). https://cran.r-project.org/web/packages/emmeans/index.html [2023 Jun 15].
Levy, R. (2008). Expectation-based syntactic comprehension. Cognition, 106(3), 1126–1177. https://doi.org/10.1016/j.cognition.2007.05.006
Lewis, R. L. (2000). Falsifying Serial and Parallel Parsing Models: Empirical Conundrums and An Overlooked Paradigm. Journal of Psycholinguistic Research, 29(2), 241–248. https://doi.org/10.1023/A:1005105414238
Liberman, A. M., Cooper, F. S., Shankweiler, D. P., & Studdert-Kennedy, M. (1967). Perception of the speech code. Psychological Review, 74(6), 431–461. https://doi.org/10.1037/h0020279
Martin, C. D., Thierry, G., Kuipers, J.-R., Boutonnet, B., Foucart, A., & Costa, A. (2013). Bilinguals reading in their second language do not predict upcoming words as native readers do. Journal of Memory and Language, 69(4), 574–588. https://doi.org/10.1016/j.jml.2013.08.001
Maye, J., Aslin, R. N., & Tanenhaus, M. K. (2008). The Weckud Wetch of the Wast: Lexical Adaptation to a Novel Accent. Cognitive Science, 32(3), 543–562. https://doi.org/10.1080/03640210802035357
McRae, K., de Sa, V. R., & Seidenberg, M. S. (1997). On the nature and scope of featural representations of word meaning. Journal of Experimental Psychology: General, 126(2), 99–130. https://doi.org/10.1037/0096-3445.126.2.99
Metusalem, R., Kutas, M., Urbach, T. P., Hare, M., McRae, K., & Elman, J. L. (2012). Generalized event knowledge activation during online sentence comprehension. Journal of Memory and Language, 66(4), 545–567. https://doi.org/10.1016/j.jml.2012.01.001
Munro, M. J., & Derwing, T. G. (1995). Processing time, accent and comprehensibility in the perception of native and foreign-accented speech. Language and Speech, 38, 289–306. https://doi.org/10.1177/002383099503800305
Nicenboim, B., Vasishth, S., & Rösler, F. (2020). Are words pre-activated probabilistically during sentence comprehension? Evidence from new data and a Bayesian random-effects meta-analysis using publicly available data. Neuropsychologia, 142, 107427. https://doi.org/10.1016/j.neuropsychologia.2020.107427
Niedzielski, N. (1999). The Effect of Social Information on the Perception of Sociolinguistic Variables. Journal of Language and Social Psychology, 18(1), 62–85. https://doi.org/10.1177/0261927X99018001005
Nieuwland, M. S., Politzer-Ahles, S., Heyselaar, E., Segaert, K., Darley, E., Kazanina, N., Von Grebmer Zu Wolfsthurn, S., Bartolozzi, F., Kogan, V., Ito, A., Mézière, D., Barr, D. J., Rousselet, G. A., Ferguson, H. J., Busch-Moreno, S., Fu, X., Tuomainen, J., Kulakova, E., Husband, E. M., …, Huettig, F. (2018). Large-scale replication study reveals a limit on probabilistic prediction in language comprehension. ELife, 7. https://doi.org/10.7554/eLife.33468
Nygaard, L. C., & Pisoni, D. B. (1998). Talker-specific learning in speech perception. Perception & Psychophysics, 60(3), 355–376. https://doi.org/10.3758/BF03206860
Nygaard, L. C., Sommers, M. S., & Pisoni, D. B. (1994). Speech Perception as a Talker-Contingent Process. Psychological Science, 5(1), 42–46. https://doi.org/10.1111/j.1467-9280.1994.tb00612.x
Otten, M., & Van Berkum, J. J. A. (2009). Does working memory capacity affect the ability to predict upcoming words in discourse? Brain Research, 1291, 92–101. https://doi.org/10.1016/j.brainres.2009.07.042
Paczynski, M., & Kuperberg, G. R. (2012). Multiple influences of semantic memory on sentence processing: Distinct effects of semantic relatedness on violations of real-world event/state knowledge and animacy selection restrictions. Journal of Memory and Language, 67(4), 426–448. https://doi.org/10.1016/j.jml.2012.07.003
Palmeri, T. J., Goldinger, S. D., & Pisoni, D. B. (1993). Episodic encoding of voice attributes and recognition memory for spoken words. Journal of Experimental Psychology: Learning, Memory, and Cognition, 19(2), 309–328. https://doi.org/10.1037/0278-7393.19.2.309
Peirce, J., Gray, J. R., Simpson, S., MacAskill, M., Höchenberger, R., Sogo, H., Kastman, E., & Lindeløv, J. K. (2019). PsychoPy2: Experiments in behavior made easy. Behavior Research Methods, 51(1), 195–203. https://doi.org/10.3758/s13428-018-01193-y
Piai, V., Roelofs, A., Rommers, J., & Maris, E. (2015). Beta oscillations reflect memory and motor aspects of spoken word production. Human Brain Mapping, 36(7), 2767–2780. https://doi.org/10.1002/hbm.22806
Pickering, M. J., & Gambi, C. (2018). Predicting while comprehending language: A theory and review. Psychological Bulletin, 144(10), 1002–1044. https://doi.org/10.1037/bul0000158
Pickering, M. J., & Garrod, S. (2007). Do people use language production to make predictions during comprehension? Trends in Cognitive Sciences, 11(3), 105–110. https://doi.org/10.1016/j.tics.2006.12.002
Pickering, M. J., & Garrod, S. (2013). An integrated theory of language production and comprehension. Behavioral and Brain Sciences, 36(4), 329–347. https://doi.org/10.1017/S0140525X12001495
Porretta, V., Tucker, B. V., & Järvikivi, J. (2016). The influence of gradient foreign accentedness and listener experience on word recognition. Journal of Phonetics, 58, 1–21. https://doi.org/10.1016/j.wocn.2016.05.006
Porretta, V., Buchanan, L., & Järvikivi, J. (2020). When processing costs impact predictive processing: The case of foreign-accented speech and accent experience. Attention, Perception, & Psychophysics, 82(4), 1558–1565. https://doi.org/10.3758/s13414-019-01946-7
R Core Team. (2023). R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria. https://www.R-project.org
Remez, R. E., Fellowes, J. M., & Rubin, P. E. (1997). Talker identification based on phonetic information. Journal of Experimental Psychology: Human Perception and Performance, 23(3), 651–666. https://doi.org/10.1037/0096-1523.23.3.651
Romero-Rivas, C., Martin, C. D., & Costa, A. (2016). Foreign-accented speech modulates linguistic anticipatory processes. Neuropsychologia, 85, 245–255. https://doi.org/10.1016/j.neuropsychologia.2016.03.022
Staub, A., & Clifton, C. (2006). Syntactic prediction in language comprehension: Evidence from either...or. Journal of Experimental Psychology: Learning, Memory, and Cognition, 32(2), 425–436. https://doi.org/10.1037/0278-7393.32.2.425
Szostak, C. M., & Pitt, M. A. (2013). The prolonged influence of subsequent context on spoken word recognition. Attention, Perception, & Psychophysics, 75(7), 1533–1546. https://doi.org/10.3758/s13414-013-0492-3
Tabachnick, B. G., & Fidell, L. S. (1989). Using multivariate statistics (2nd ed.). Harper & Row.
Traxler, M. J. (2014). Trends in syntactic parsing: anticipation, Bayesian estimation, and good-enough parsing. Trends in Cognitive Sciences, 18(11), 605–611. https://doi.org/10.1016/j.tics.2014.08.001
Traxler, M. J., Pickering, M. J., & Clifton, C. (1998). Adjunct Attachment Is Not a Form of Lexical Ambiguity Resolution. Journal of Memory and Language, 39(4), 558–592. https://doi.org/10.1006/jmla.1998.2600
van Gompel, R. P. G., Pickering, M. J., Pearson, J., & Liversedge, S. P. (2005). Evidence against competition during syntactic ambiguity resolution. Journal of Memory and Language, 52(2), 284–307. https://doi.org/10.1016/j.jml.2004.11.003
Van Petten, C., & Luka, B. J. (2012). Prediction during language comprehension: Benefits, costs, and ERP components. International Journal of Psychophysiology, 83(2), 176–190. https://doi.org/10.1016/j.ijpsycho.2011.09.015
Wagenmakers, E.-J., & Farrell, S. (2004). AIC model selection using Akaike weights. Psychonomic Bulletin & Review, 11(1), 192–196. https://doi.org/10.3758/BF03206482
Walker, A., & Hay, J. (2011). Congruence between ‘word age’ and ‘voice age’ facilitates lexical access. Laboratory Phonology, 2(1). https://doi.org/10.1515/labphon.2011.007
Weatherholtz, K., & Jaeger, T. F. (2016). Speech Perception and Generalization Across Talkers and Accents. Oxford University Press. https://doi.org/10.1093/acrefore/9780199384655.013.95
Funding
Open access funding provided by Università degli Studi di Padova within the CRUI-CARE Agreement. MS was supported by a PhD grant from the University of Padova (2022–2025).
Author information
Authors and Affiliations
Contributions
MS, FP e FV equally contributed to the conceptualization of the study. MS collected the data, performed the analyses and drafted the manuscript. LC contributed to data collection and served in a supporting role for formal analysis, methodology, and writing-review, editing. FV contributed to the writing of the manuscript. FP coordinated the teamwork and contributed to the writing of the manuscript. All authors provided critical feedback during all phases of the work and read and approved the final version of the manuscript.
Corresponding authors
Ethics declarations
Conflicts of interest/Competing interest
The authors declare no competing interests.
Ethics approval
The research adhered to the principles outlined in the Declaration of Helsinki. The research protocol was approved by the Ethics Committee for Psychological Research of the University of Padova (protocol number: 5181).
Consent to participate
Participants provided their informed consent before participating in the experiment. They also consent to the use of their data (collected anonymously) for research aims.
Consent to publish
Not applicable.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Below is the link to the electronic supplementary material.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Sala, M., Vespignani, F., Casalino, L. et al. I know how you’ll say it: evidence of speaker-specific speech prediction. Psychon Bull Rev (2024). https://doi.org/10.3758/s13423-024-02488-2
Accepted:
Published:
DOI: https://doi.org/10.3758/s13423-024-02488-2