Abstract
When conducting voice interactions, it is found that loud noises have terrible effects on the recognition of industrial robots. Some industrial machines like motors and fans, tend to produce loud noises when working, placing great obstacles to the voice interactions. Therefore, an end-to-end full convolutional network model, called Res-Unet, is proposed to solve this problem. The difference between this network and other conventional ones is the accession of residual networks to the encoder and the decoder. It has increased the convergence and complexity of the network and improved the expression ability of the network. In experiments on the decrease of noises caused by industrial machines, it is proved that the performance of Res-Unet was better than other networks, in respect of speech enhancement.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Liu, W.J., Nie, S., Liang, S., et al.: Deep learning based speech separation technology and its developments. Acta Automatica Sinica 42(6), 819–833 (2016)
Wang, Y., Wang, D.L.: Towards scaling up classification-based speech separation. IEEE Trans. Audio Speech Lang. Process. 21(7), 1381–1390 (2013)
Narayanan, A., Wang, D.L.: Ideal ratio mask estimation using deep neural networks for robust speech recognition. In: 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 7092–7096. IEEE (2013)
Park, S.R., Lee, J.: A fully convolutional neural network for speech enhancement. arXiv preprint arXiv:1609.07132 (2016)
Jansson, A., Humphrey, E., Montecchio, N., et al.: Singing voice separation with deep u-net convolutional networks (2017)
Oord, A., Dieleman, S., Zen, H., et al.: Wavenet: a generative model for raw audio. arXiv preprint arXiv:1609.03499 (2016)
Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234–241. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-24574-4_28
Panayotov, V., Chen, G., Povey, D., et al.: Librispeech: an ASR corpus based on public domain audio books. In: 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5206–5210. IEEE (2015)
Kingma, D.P., Adam, B.J.: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)
He, K., Zhang, X., Ren, S., et al.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016)
Huber, P.J.: Robust estimation of a location parameter. In: Kotz, S., Johnson, N.L. (eds.) Breakthroughs in Statistics, pp. 492–518. Springer, New York (1992). https://doi.org/10.1007/978-1-4612-4380-9_35
Piczak, K.J.: ESC: dataset for environmental sound classification. In: Proceedings of the 23rd ACM International Conference on Multimedia, pp. 1015–1018 (2015)
Wang, Y., Narayanan, A., Wang, D.L.: On training targets for supervised speech separation. IEEE/ACM Trans. Audio Speech Lang. Process. 22(12), 1849–1858 (2014)
He, K., Zhang, X,. Ren, S., et al.: Delving deep into rectifiers: surpassing human-level performance on imagenet classification. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1026–1034 (2015)
Rix, A.W., Beerends, J.G., Hollier, M.P., et al.: Perceptual evaluation of speech quality (PESQ)-a new method for speech quality assessment of telephone networks and codecs. In: 2001 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings (Cat. No. 01CH37221), vol. 2, pp. 749–752. IEEE (2001)
Taal, C.H., Hendriks, R.C., Heusdens, R., et al.: A short-time objective intelligibility measure for time-frequency weighted noisy speech. In: 2010 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 4214–4217. IEEE (2010)
Cohen, I., Berdugo, B.: Speech enhancement for non-stationary noise environments. Signal Process. 81(11), 2403–2418 (2001)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 Springer Nature Switzerland AG
About this paper
Cite this paper
Pu, Y., Yu, H. (2022). ResUnet: A Fully Convolutional Network for Speech Enhancement in Industrial Robots. In: Fujita, H., Fournier-Viger, P., Ali, M., Wang, Y. (eds) Advances and Trends in Artificial Intelligence. Theory and Practices in Artificial Intelligence. IEA/AIE 2022. Lecture Notes in Computer Science(), vol 13343. Springer, Cham. https://doi.org/10.1007/978-3-031-08530-7_4
Download citation
DOI: https://doi.org/10.1007/978-3-031-08530-7_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-08529-1
Online ISBN: 978-3-031-08530-7
eBook Packages: Computer ScienceComputer Science (R0)