An Artificial Neural Networks Model by Using Wavelet Analysis for Speaker Recognition

Conference paper
Part of the Advances in Intelligent Systems and Computing book series (AISC, volume 340)


An Artificial Neural Networks Model by using Wavelet Analysis for Speaker Recognition has been presented in this paper. The wavelet analysis was used to extract the features. These extracted features were trained using Artificial Neural Networks with popular Back Propagation Learning Algorithm. In this analysis of testing, the speakers speak out the same set of words, with these set words the features were extracted and fed into the training of the neural network. The neural network notifies the identity of the speaker. In order to test the system, the voice data of the speakers were recorded. The experiments were carried out by using 800 data sets of total 40 individual speakers. For each of these speakers, 20 speech signals were used for training. All these signals were used for training, validation and testing. This approach reveals that the overall performance of system is 95 %.


Artificial neural networks Speaker recognition Wavelet analysis Back propagation 


  1. 1.
    Daqrouq, K.: Wavelet entropy and neural network for text-independent speaker identification. Eng. Appl. Artif. Intell. 24(5), 796–802 (2011)CrossRefGoogle Scholar
  2. 2.
    Adami, A.G., Barone, D.A.C.: A speaker identification system using a model of artificial neural networks for an elevator application. Inf. Sci. 138(1–4), 1–5 (2001)Google Scholar
  3. 3.
    Kinnunen,T., Sidoroff, I., Tuononen, M., Fränti, P.: Comparison of clustering methods: a case study of text-independent speaker modeling. Patt. Recogn. Lett. 32(13), 1604–1617 (2011)Google Scholar
  4. 4.
    Ganchev, T.D., Fakotakis, N.D.: Generalized locally recurrent probabilistic neural networks with application to text independent speaker verification. Neurocomputing 70(7–9), 1424–1438 (2007)Google Scholar
  5. 5.
    Tomi, K., Lib, H.: An overview of text-independent speaker recognition: from features to supervectors. Speech Commun. 52(1), 12–40 (2010)CrossRefGoogle Scholar
  6. 6.
    Solera-Ureña, R., Martín-Iglesias, D., Gallardo-Antolín, A., Peláez-Moreno, C., Díaz-de-María, F.: Robust ASR using support vector machines. Speech Commun. 49(4), 253–267 (2007)CrossRefGoogle Scholar
  7. 7.
    Huenupán, F., Yoma, N.B., Molina, C., Garretón. C.: Confidence based multiple classifier fusion in speaker verification. Patt. Recogn. Lett. 29(7), 957–966 (2008)Google Scholar
  8. 8.
    Lina, S.-Y., Guhb, R.-S., Shiuec, Y.-R.: Effective recognition of control chart patterns in auto correlated data using a support vector machine based approach. Comput. Ind. Eng. 61(4), 1123–1134 (2011)CrossRefGoogle Scholar
  9. 9.
    Moattar, M.H., Homayounpour, M.M.: Variational conditional random fields for online speaker detection and tracking. Speech Commun. 54(6), 763–780 (2012)CrossRefGoogle Scholar
  10. 10.
    Nemati, S., Basiri, M.E.: Text-independent speaker verification using ant colony optimization-based selected features. Expert Syst. Appl. 38(1), 620–630 (2011)Google Scholar
  11. 11.
    Chen, L., Mao, X., Xue, Y., Lee, L.C.: Speech emotion recognition: features and classification models. Digital Signal Process. 22(6), 1154–1160 (2012) Google Scholar
  12. 12.
    Gajšek, R., Mihelič, F., Dobrišek, S.: Speaker state recognition using an HMM-based feature extraction method. Comput. Speech Lang. 27(1), 135–150 (2013)Google Scholar
  13. 13.
    Bennani, Y., Gallinari, P.: Neural networks for discrimination and modelization of speakers. Speech Commun. 17(1–2), 159–175 (1995)Google Scholar
  14. 14.
    Mak, M.W., Allen, W.G., Sexton, G.G.: Speaker identification using multilayer perceptrons and radial basis function networks. Neurocomputing 6(1), 99–117 (1994)Google Scholar
  15. 15.
    Thévenaz, P., Hügli, H.: Usefulness of the LPC-residue in text-independent speaker verification. Speech Commun. 17(1–2), 145–157 (1995)Google Scholar
  16. 16.
    Gong, Y.: Speech recognition in noisy environments: a survey. Speech Commun. 16(3), 261–291 (1995)CrossRefGoogle Scholar

Copyright information

© Springer India 2015

Authors and Affiliations

  1. 1.Department of Computer Science and Engineering, GITAM Institute of TechnologyGITAM UniversityRushikonda, VisakhapatnamIndia

Personalised recommendations