Skip to main content
Log in

Speech synthesis system based on big data and evaluation of Japanese language feeling

  • Original Article
  • Published:
International Journal of System Assurance Engineering and Management Aims and scope Submit manuscript

Abstract

With the progress and development of modern technology, people have put forward higher standards for speech synthesis technology. Because speech synthesis technology has already been applied to various fields and industries in people’s lives at this stage, speech synthesis technology faces such high standards. Some are overwhelmed. The main problem now is that the synthesized speech is not natural enough compared to human speech, the pronunciation is not standard enough, and the emotional color is lacking. The main basis of this paper is the continuous self-learning and optimized speech synthesis technology in the Markov model, analyzes its main algorithms, and mainly optimizes its input text analysis and output speech synthesis algorithms. Language sense is a kind of resonance with a language when it is used for a long time, so it is basically based on the proficiency of a language. Therefore, the quality of a language can directly reflect the proficiency of a language, and it is also an important guarantee for smooth communication. Among them, Japanese is an important medium for carrying Japanese culture. The study of Japanese vocabulary will play a very important role in the study of Japanese language sense, and it has certain significance. In daily Japanese learning, consciously paying attention to language sense can accelerate the formation of language sense. Therefore, when you have a better sense of language, you can quickly and directly resonate with the language. So in the process of learning Japanese, if you can have a better sense of Japanese, then Japanese learning will advance by leaps and bounds. Therefore, this paper studies the speech synthesis system and Japanese language sense and its evaluation.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8

Similar content being viewed by others

References

  • Bettayeb N, Guerti M (2021) Speech synthesis system for the holy quran recitation. Int Arab J Inf Technol 18(1):8–15

    Google Scholar 

  • Boschi V, Catricala E, Consonni M, Chesi C, Moro A, Cappa SF (2017) Connected speech in neurodegenerative language disorders: a review. Front Psychol 8:269

    Article  Google Scholar 

  • Garcia Rodrigues J, Villasante S, Sousa Pinto I (2022) Non-material nature’s contributions to people from a marine protected area support multiple dimensions of human well-being. Sustain Sci 17(3):793–808

    Article  Google Scholar 

  • Kayte S, Mundada M, Kayte DC (2015) Speech synthesis system for marathi accent using festvox. Int J Comput Appl 130(6):38–42

    Google Scholar 

  • Mattheyses W, Verhelst W (2015) Audiovisual speech synthesis: an overview of the state-of-the-art. Speech Commun 66:182–217

    Article  Google Scholar 

  • Mor B, Garhwal S, Kumar A (2021) A systematic review of hidden Markov models and their applications. Arch Comput Methods Eng 28:1429–1448

    Article  MathSciNet  Google Scholar 

  • Morise M, Yokomori F, Ozawa K (2016) WORLD: a vocoder-based high-quality speech synthesis system for real-time applications. IEICE Trans Inf Syst 99(7):1877–1884

    Article  Google Scholar 

  • Ning Y, He S, Wu Z, Xing C, Zhang LJ (2019) A review of deep learning based speech synthesis. Appl Sci 9(19):4050

    Article  Google Scholar 

  • Ren F, Bao Y (2020) A review on human-computer interaction and intelligent robots. Int J Inform Technol Decis Mak 19(01):5–47

    Article  Google Scholar 

  • Rose K, Eldridge S, Chapin L (2015) The internet of things: an overview. Internet Soc (ISOC) 80:1–50

    Google Scholar 

  • Schnell S (2020) Vision, voice, and technology: is there a global open government trend? Adm Soc 52(10):1593–1620

    Article  Google Scholar 

  • Wang X, Takaki S, Yamagishi J (2019) Neural source-filter waveform models for statistical parametric speech synthesis. IEEE/ACM Trans Audio, Speech,  Lang Process 28:402–415

  • Warner C, Dupuy B (2018) Moving toward multiliteracies in foreign language teaching: past and present perspectives… and beyond. Foreign Lang Annals 51(1):116–128

    Article  Google Scholar 

  • Wood SN, Pya N, Safken B (2016) Smoothing parameter and model selection for general smooth models. J Am Stat Assoc 111(516):1548–1563

    Article  MathSciNet  Google Scholar 

  • Xiong W, Droppo J, Huang X et al (2017) Toward human parity in conversational speech recognition. IEEE/ACM Trans Audio Speech Lang Process 25(12):2410–2423

    Article  Google Scholar 

Download references

Funding

This paper was supported by (1) Ministry of Education Project Name: Japanese Listening Course Teaching Reform Based on “Tanzhou Classroom” Network Platform ,Project Number: 201802182015 ; (2) Ministry of Education Project Name: Reform and Construction of the Practice Base of Professional Literacy MOS International Certification Project Based on Tanzhou Education Platform, Project Number: 201902144066.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yanli Peng.

Ethics declarations

Conflict of interest

The author reports no conflicts of interest.

Ethical approval

This article does not contain any studies with human participants performed by any of the authors.

Informed consent

Not applicable.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Peng, Y. Speech synthesis system based on big data and evaluation of Japanese language feeling. Int J Syst Assur Eng Manag (2023). https://doi.org/10.1007/s13198-023-02154-1

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s13198-023-02154-1

Keywords

Navigation