Audio-Visual Speech-Turn Detection and Tracking

  • Israel D. Gebru
  • Silèye Ba
  • Georgios Evangelidis
  • Radu HoraudEmail author
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 9237)


Speaker diarization is an important component of multi-party dialog systems in order to assign speech-signal segments among participants. Diarization may well be viewed as the problem of detecting and tracking speech turns. It is proposed to address this problem by modeling the spatial coincidence of visual and auditory observations and by combining this coincidence model with a dynamic Bayesian formulation that tracks the identity of the active speaker. Speech-turn tracking is formulated as a latent-variable temporal graphical model and an exact inference algorithm is proposed. We describe in detail an audio-visual discriminative observation model as well as a state-transition model. We also describe an implementation of a full system composed of multi-person visual tracking, sound-source localization and the proposed online diarization technique. Finally we show that the proposed method yields promising results with two challenging scenarios that were carefully recorded and annotated.


Speaker diarization Audio-visual fusion Sound-source localization Multi-person tracking Temporal graphical models 


  1. 1.
    Anguera Miro, X., Bozonnet, S., Evans, N., Fredouille, C., Friedland, G., Vinyals, O.: Speaker diarization: A review of recent research. IEEE Trans. Audio Speech Lang. Process. 20(2), 356–370 (2012)CrossRefGoogle Scholar
  2. 2.
    Bae, S.H., Yoon, K.J.: Robust online multi-object tracking based on tracklet confidence and online discriminative appearance learning. In: Computer Vision and Pattern Recognition, pp. 1218–1225 (2014)Google Scholar
  3. 3.
    Deleforge, A., Horaud, R., Schechner, Y.Y., Girin, L.: Co-localization of audio sources in images using binaural features and locally-linear regression. IEEE Trans. Audio Speech Lang. Process. 23(4), 718–731 (2015)CrossRefGoogle Scholar
  4. 4.
    Gatica-Perez, D., Lathoud, G., Odobez, J.M., McCowan, I.: Audiovisual probabilistic tracking of multiple speakers in meetings. IEEE Trans. Audio Speech Lang. Process. 15(2), 601–616 (2007)CrossRefGoogle Scholar
  5. 5.
    Kidron, E., Schechner, Y.Y., Elad, M.: Cross-modal localization via sparsity. IEEE Trans. Signal Process. 55(4), 1390–1404 (2007)MathSciNetCrossRefGoogle Scholar
  6. 6.
    Naqvi, S., Yu, M., Chambers, J.: A multimodal approach to blind source separation of moving sources. IEEE J. Sel. Top. Signal Process. 4(5), 895–910 (2010)CrossRefGoogle Scholar
  7. 7.
    Noulas, A., Englebienne, G., Krose, B.J.A.: Multimodal speaker diarization. IEEE Trans. Pattern Anal. Mach. Intell. 34(1), 79–93 (2012)CrossRefGoogle Scholar
  8. 8.
    Potamianos, G., Neti, C., Gravier, G., Garg, A., Senior, A.W.: Recent advances in the automatic recognition of audiovisual speech. Proc. IEEE 91(9), 1306–1326 (2003)CrossRefGoogle Scholar
  9. 9.
    Sohn, J., Kim, N.S., Sung, W.: A statistical model-based voice activity detection. IEEE Signal Process. Lett. 6(1), 1–3 (1999)CrossRefGoogle Scholar

Copyright information

© Springer International Publishing Switzerland 2015

Authors and Affiliations

  • Israel D. Gebru
    • 1
  • Silèye Ba
    • 1
  • Georgios Evangelidis
    • 1
  • Radu Horaud
    • 1
    Email author
  1. 1.INRIA Grenoble Rhône-AlpesMontbonnot Saint-MartinFrance

Personalised recommendations