Abstract
Recently, artificial intelligence-based machine translation has been much improved over the traditional methods. A machine translator is very useful for translating text or speech from one language to another. Machine translators have replaced the word mechanism in one language for words in another with verbatim translations. However, a good translation should be employed a both a sentence and a word that have complete meaning in accordance with the context of relevant sentence. In this paper, we studied on English – Vietnamese translation using deep learning methods including Recurrent Neural Network (RNN), Long Short-Term Memory (LSTM), Gated Recurrent Units (GRU), Attention, and Transformer. The deep learning-based machine translators were compared based on the test accuracy of results translation. It was found that best deep learning-based machine translator model was the Attention mechanism, achieving 98.8% accuracy. The Transformer yielded second rank or 98.5% accuracy.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Tsai, C.-W., Lai, C.-F., Chao, H.-C., Vasilakos, A.V.: Big data analytics: a survey. J. Big Data, 1–3, 6–7 (2015)
Kambatla, K., Kollias, G., Kumar, V., Grama, A.: Trends in big data analytics, 1–3, 8–9 (2014)
Phan-Vu, H.-H., Tran, V.-T., Nguyen, V.-N., Dang, H.-V., Do, P.-T.: Neural machine translation between Vietnamese and English: an empiracal study. J. Comput. Sci. Cybern., 1–15 (2018)
Amidi, A., Amidi, S.: Recurrent neural networks cheatsheet. https://stanford.edu/~shervine/teaching/cs-230/cheatsheet-recurrent-neural-networks. Accessed 20 Dec 2020
Sak, H., Senior, A., Beaufays, F.: Long short-term memory recurrent neural network architectures for large scale acoustic modeling. Interspeech, 1–3 (2014)
Xu, S., Li, J., Liu, K., Wu, L.: A parallel GRU recurrent network model and its application to multi-channel time-varying signal classification. IEEE Access, 1–5 (2019)
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N. Kaiser, L., Polosukhin, I.: Attention is all you need. https://arxiv.org/abs/1706.03762. Accessed 20 Dec 2020
Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. https://arxiv.org/abs/1810.04805. Accessed 20 Dec 2020
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Minh, T.N., Meesad, P., Nguyen Ha, H.C. (2021). English-Vietnamese Machine Translation Using Deep Learning. In: Meesad, P., Sodsee, D.S., Jitsakul, W., Tangwannawit, S. (eds) Recent Advances in Information and Communication Technology 2021. IC2IT 2021. Lecture Notes in Networks and Systems, vol 251. Springer, Cham. https://doi.org/10.1007/978-3-030-79757-7_10
Download citation
DOI: https://doi.org/10.1007/978-3-030-79757-7_10
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-79756-0
Online ISBN: 978-3-030-79757-7
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)