Abstract
Machine learning , like neural network methods, has been implemented in the natural language processing of virtually every domain. With speech utilizations like vocabulary to speech organization, coalescence challenge that has been adequately immune directly toward successful machine attainments learning approaches is fundamentals normalization. Considering example, in this application it must be determined that 123 is verbalized in signatures as one hundred and twenty-three but in sovereign potentate Ave as one twenty-three. Modern industrial systems for this role are heavily dependent on hand-recorded penned double speak-specific stratification. We introduce neural interconnection miniatures well-known regard text notarization for as a streak to progression problem, where input is admission taking in token in history, and turn out gain would be that token’s verbalization.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Zhang H, Sproat R, Ng AH, Stahlberg F, Peng X, Gorman K, Roark B (2019) Neural models of text normalization for speech applications. Comput Linguist 45(2):293–337
Allauzen C, Riley M (2012) A pushdown transducer extension for the OpenFst library. In: CIAA, Porto, pp 66–77
Allen J, Hunnicutt SM, Klatt D (1987) From text to speech: the MITalk system. Cambridge University Press, Cambridge
Arthur P, Neubig G, Nakamura S (2016) Incorporating discrete translation lexicons into neural machine translation. In: EMNLP, Austin, TX, pp 1557–1567
Arik S, Chrzanowski M, Coates A, Diamos G, Gibiansky A, Kang Y, Li X, Miller J, Ng A, Raiman J, Sengupta S, Shoeybi M (2017) Deep voice: real-time neural text-to-speech. ArXiv: 1702.07825
Aw AT, Lee LH (2012) Personalized normalization for a multilingual chat system. In: ACL, Jeju Island, pp 31–36
Bahdanau D, Cho K, Bengio Y (2015) Neural machine translation by jointly learning to align and translate. In: ICLR, San Diego, CA
Beaufort R, Roekhaut S, Cougnon L-A, Fairon C (2010) A hybrid rule/model-based finite-state framework for normalizing SMS messages. In: ACL, Uppsala, pp 770–779
Chen MX, Firat O, Bapna A, Johnson M, Macherey W, Foster G, Jones L, Parmar N, Schuster M, Chen Z, Wu Y, Hughes M (2018) The best of both worlds: combining recent advances in neural machine translation. CoRR, abs/1804.09849
Chiu C-C, Sainath TN, Wu Y, Prabhavalkar R, Nguyen P, Chen Z, Kannan A, Weiss RJ, Rao K, Gonina E, Jaitly N, Li B, Chorowski J, Bacchiani M (2017) State-of-the-art speech recognition with sequence-to-sequence models. ArXiv: 1712.01769
Cho K, van Merriënboer B, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using RNN encoder-decoder for statistical machine translation. In: EMNLP, Doha, pp 1724–1734
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Varalakshmi, P.N.K., Kallimani, J.S. (2022). Text Normalization Through Neural Models in Generating Text Summary for Various Speech Synthesis Applications. In: Agrawal, D.P., Nedjah, N., Gupta, B.B., Martinez Perez, G. (eds) Cyber Security, Privacy and Networking. Lecture Notes in Networks and Systems, vol 370. Springer, Singapore. https://doi.org/10.1007/978-981-16-8664-1_18
Download citation
DOI: https://doi.org/10.1007/978-981-16-8664-1_18
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-16-8663-4
Online ISBN: 978-981-16-8664-1
eBook Packages: EngineeringEngineering (R0)