Skip to main content

Exploring the Factors Aiding Speech-to-Text Emotional Restoration

  • Conference paper
  • First Online:
Design, User Experience, and Usability: Design for Diversity, Well-being, and Social Development (HCII 2021)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 12780))

Included in the following conference series:

Abstract

In recent years, with the development of artificial intelligence technology, speech recognition technology can perform high-precision interpretation and transcription on voices in various complex environments, improving typing efficiency. However, the text obtained by speech translation is only composed of text and simple punctuation, which hinders the real emotion expression of users. The pale translated text hinders the formation of context, affects the emotional transmission of semantics, and lead to a poor user experience when users communicate with others. Based on user experience and emotion, this article discusses the factors that assist the speech-to-text emotional restoration. Through the qualitative and quantitative study, this research compares four emotional effects of information texts composed by different elements: emoticon, punctuation, interjections, and speech-to-text function of WeChat, and further studies the factors that assist speech-to-text emotion restoration. The research results reveal that emoticon and punctuation have a positive effect on the speech-to-text emotional restoration. The addition of the above two factors can restore the emotional effect of speech in text mode with lower loss, fully improve the user experience in mobile communication, and make the online communication smoother.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Shneiderman, B.: The limits of speech recognition. Commun. ACM 43(9), 63–65 (2000)

    Article  Google Scholar 

  2. Basapur S, Xu S, Ahlenius M, et al.: User expectations from dictation on mobile devices. In: International Conference on Human-Computer Interaction, pp. 217–225 (2007)

    Google Scholar 

  3. Ruan, S., Wobbrock, J.O., Liou, K., et al.: Comparing speech and keyboard text entry for short messages in two languages on touchscreen phones. Proc. ACM Interact. Mob. Wearable Ubiquit. Technol. 1(4), 1–23 (2018)

    Article  Google Scholar 

  4. Kumar, A., Paek, T., Lee, B.: Voice typing: a new speech interaction model for dictation on touchscreen devices. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 2277–2286 (2012)

    Google Scholar 

  5. Karat, C.M., Halverson, C., Horn, D., et al.: Patterns of entry and correction in large vocabulary continuous speech recognition systems. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 568–575 (1999)

    Google Scholar 

  6. Kwon, O.W., Chan, K., Hao, J., et al.: Emotion recognition by speech signals. In: Eighth European Conference on Speech Communication and Technology (2003)

    Google Scholar 

  7. Ekman, P.: Facial expression and emotion. Am. Psychol. 48(4), 384 (1993)

    Google Scholar 

  8. Morency, L.P., Mihalcea, R., Doshi, P.: Towards multimodal sentiment analysis: harvesting opinions from the web. In: Proceedings of the 13th International Conference on Multimodal Interfaces, pp. 169–176 (2011)

    Google Scholar 

  9. Yadollahi, A., Shahraki, A.G., Zaiane, O.R.: Current state of text sentiment analysis from opinion to emotion mining. ACM Comput. Surv. (CSUR) 50(2), 1–33 (2017)

    Google Scholar 

  10. Report of WeChat data (2019). https://mp.weixin.qq.com/s/gi_3xSDWBie-fgg76XXJCg

  11. Report of WeChat data (2018). https://support.weixin.qq.com/cgi-bin/mmsupport-bin/getopendays

  12. Busso, C., Narayanan, S.S.: Joint analysis of the emotional fingerprint in the face and speech: a single subject study. In: 2007 IEEE 9th Workshop on Multimedia Signal Processing, pp. 43–47 (2007)

    Google Scholar 

  13. Arya, A., Jefferies, L.N., Enns, J.T., et al.: Facial actions as visual cues for personality. Comput. Anim. Virtual Worlds 17(3–4), 371–382 (2006)

    Article  Google Scholar 

  14. Huang, A.H., Yen, D.C., Zhang, X.: Exploring the potential effects of emoticons. Inf. Manage. 45(7), 466–473 (2008)

    Google Scholar 

  15. Kalra, A., Karahalios, K.: TextTone: expressing emotion through text. In: IFIP Conference on Human-Computer Interaction, pp. 966–969 (2005)

    Google Scholar 

  16. Motley, M.T.: Facial affect and verbal context in conversation: facial expression as interjection. Hum. Commun. Res. 20(1), 3–40 (1993)

    Article  Google Scholar 

  17. Bailey, D.V., Dürmuth, M., Paar, C.: “Typing” passwords with voice recognition: how to authenticate to Google Glass. In: Proceedings of the Symposium on Usable Privacy and Security, pp. 1–2 (2014)

    Google Scholar 

Download references

Acknowledgments

We thank the Foundation for Young Talents in Higher Education of Guangdong, China [Project Batch No. 2020WQNCX061] for the research support. Part of the study was supported by Shenzhen Educational Science Planning Project (zdfz20015).

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Chen, X., Deng, Q. (2021). Exploring the Factors Aiding Speech-to-Text Emotional Restoration. In: Soares, M.M., Rosenzweig, E., Marcus, A. (eds) Design, User Experience, and Usability: Design for Diversity, Well-being, and Social Development. HCII 2021. Lecture Notes in Computer Science(), vol 12780. Springer, Cham. https://doi.org/10.1007/978-3-030-78224-5_29

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-78224-5_29

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-78223-8

  • Online ISBN: 978-3-030-78224-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics