Listeners identified spoken words, letters, and numbers and the spatial location of these utterances in three listening conditions as a function of the number of simultaneously presented utterances. The three listening conditions were a normal listening condition, in which the sounds were presented over seven possible loudspeakers to a listener seated in a sound-deadened listening room; a one-headphone listening condition, in which a single microphone that was placed in the listening room delivered the sounds to a single headphone worn by the listener in a remote room; and a stationary KEMAR listening condition, in which binaural recordings from an acoustic manikin placed in the listening room were delivered to a listener in the remote room. The listeners were presented one, two, or three simultaneous utterances. The results show that utterance identification was better in the normal listening condition than in the one-headphone condition, with the KEMAR listening condition yielding intermediate levels of performance. However, the differences between listening in the normal and in the one-headphone conditions were much smaller when two, rather than three, utterances were presented at a time. Localization performance was good for both the normal and the KEMAR listening conditions and at chance for the one-headphone condition. The results suggest that binaural processing is probably more important for solving the “cocktail party” problem when there are more than two concurrent sound sources.
Blauert, J. (1983).Spatial hearing. Cambridge, MA: MIT Press.
Bregman, A. S. (1990).Auditory scene analysis. Cambridge, MA: MIT Press.
Cherry, C. (1953). Some experiments on the recognition of speech, with one and two ears.Journal of the Acoustical Society of America,25, 975–981.
Jones, M., &Yee, W. (1993). Attending to auditory events: The role of temporal organization. In S. McAdams & E. Bigand (Eds.),Thinking in sound (pp. 69–106). Oxford: Oxford University Press, Clarendon Press.
Mills, A. W. (1958). On the minimum audible angle.Journal of the Acoustical Society of America,30, 237–243.
Shackleton, T. M., &Meddis, R. (1992). The role of interaural time difference and fundamental frequency difference in identification of concurrent vowel pairs.Journal of the Acoustical Society of America,91, 3579–3581.
Wenzel, E. M., Arruda, M., Kistler, D. J., &Wightman, F. L. (1993). Localization using nonindividualized head-related transfer functions.Journal of the Acoustical Society of America,94, 111–123.
Wightman, F. L., &Kistler, D. J. (1993). Sound localization. In W. A. Yost, A. N. Popper, & R. R. Fay (Eds.),Auditory psychophysics (pp. 155–193). New York: Springer-Verlag.
Yost, W. A. (1992a). Auditory image perception and analysis.Hearing Research,56, 8–19.
Yost, W. A. (1992b). Auditory perception and sound source determination.Current Directions in Psychological Sciences,1(6), 15–19.
Yost, W. A. (in press). The cocktail party problem: 40 years later. In R. H. Gilkey & T. R. Anderson (Eds.),Binaural and spatial hearing. Mahwah, NJ: Erlbaum.
This research was supported by grants from the National Institute on Deafness and Other Communication Disorders and the Air Force Office of Scientific Research.
About this article
Cite this article
Yost, W.A., Dye, R.H. & Sheft, S. A simulated “cocktail party” with up to three sound sources. Perception & Psychophysics 58, 1026–1036 (1996). https://doi.org/10.3758/BF03206830
- Sound Source
- Listening Condition
- Localization Task
- Cocktail Party
- Male Talker