Abstract
A great deal of inner variabilities such as emotion and stress are largely missing from traditional speaker recognition system. The direct result is that the recognition system is easily disturbed when the enrollment and the authentication are made under different emotional state. Reynolds [1] proposed a new normalization technique called feature mapping. This technique achieved big successes in channel robust speaker verification. We extend the mapping idea to develop a feature variety training approach for affective-insensitive speaker recognition.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Reynolds, D.A.: Channel robust speaker verification via feature mapping. In: ICASSP 2003, vol. 2, pp. 53–56 (2003)
Wu, T., Yang, Y.C., Wu, Z.H., Li, D.D., MASC,: A Speech Corpus in Mandarin for Emotion Analysis and Affective Speaker Recognition, The IEEE Odyssey, 1–59 (2006)
Bonastre, J.F., Wils, F., Meignier, S.: ALIZE, a free toolkit for speaker recognition. In: IEEE International Conference on Acoustics, Speech, and Signal Processing. In (ICASSP 2005), March 18-23, vol. 1, pp. 737–740 (2005)
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 2007 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Li, D., Yang, Y. (2007). Affect-Insensitive Speaker Recognition by Feature Variety Training. In: Paiva, A.C.R., Prada, R., Picard, R.W. (eds) Affective Computing and Intelligent Interaction. ACII 2007. Lecture Notes in Computer Science, vol 4738. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-74889-2_78
Download citation
DOI: https://doi.org/10.1007/978-3-540-74889-2_78
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-74888-5
Online ISBN: 978-3-540-74889-2
eBook Packages: Computer ScienceComputer Science (R0)