Abstract
Aphasia and other communication disorders affect a person’s daily life, leading to isolation and lack of self-confidence, affecting independence, and hindering the ability to express themselves easily, including asking for help. Even though assistive technology for these disorders already exists, solutions rely mostly on a graphical output and touch, gaze, or brain-activated input modalities, which do not provide all the necessary features to cover all periods of the day (e.g., night-time). In the scope of the AAL APH-ALARM project, we aim at providing communication support to users with speech difficulties (mainly aphasics), while lying in bed. Towards this end, we propose a system based on gesture recognition using a radar deployed, for example, in a wall of the bedroom. A first prototype was implemented and used to evaluate gesture recognition, relying on radar data and transfer learning. The initial results are encouraging, indicating that using a radar can be a viable option to enhance the communication of people with speech difficulties, in the in-bed scenario.
This work was supported by EU and national funds through the Portuguese Foundation for Science and Technology (FCT), in the context of the AAL APH-ALARM project (AAL/0006/2019), and funding to the research unit IEETA (UIDB/00127/2020).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Ahmed, S., Kallu, K.D., Ahmed, S., Cho, S.H.: Hand gestures recognition using radar sensors for human-computer-interaction: a review. Remote Sens. 13(3), 527 (2021)
Elsahar, Y., Hu, S., Bouazza-Marouf, K., Kerr, D., Mansor, A.: Augmentative and Alternative Communication (AAC) advances: a review of configurations for individuals with a speech disability. Sensors 19(8), 1911 (2019)
Hazra, S., Santra, A.: Robust gesture recognition using millimetric-wave radar system. IEEE Sens. Lett. 2(4), 1–4 (2018)
Ishak, K., Appenrodt, N., Dickmann, J., Waldschmidt, C.: Human gesture classification for autonomous driving applications using radars. In: IEEE MTT-S International Conference on Microwaves for Intelligent Mobility (ICMIM), pp. 1–4, November 2020
Jiang, S., Kang, P., Song, X., Lo, B., Shull, P.B.: Emerging wearable interfaces and algorithms for hand gesture recognition: a survey. In: IEEE Reviews in Biomedical Engineering, p. 1 (2021)
Keras: Keras applications. https://keras.io/api/applications/
Wang, T., et al.: A survey on vision-based hand gesture recognition. In: Basu, A., Berretti, S. (eds.) Smart Multimedia, pp. 219–231. Springer, Cham (2018)
Yasen, M., Jusoh, S.: A systematic review on hand gesture recognition techniques, challenges and applications. Peer J. Comput. Sci. 5, e218 (2019)
Yu, M., Kim, N., Jung, Y., Lee, S.: A frame detection method for real-time hand gesture recognition systems using CW-radar. Sensors 20(8), 2321 (2020)
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 ICST Institute for Computer Sciences, Social Informatics and Telecommunications Engineering
About this paper
Cite this paper
Santana, L. et al. (2022). Radar-Based Gesture Recognition Towards Supporting Communication in Aphasia: The Bedroom Scenario. In: Hara, T., Yamaguchi, H. (eds) Mobile and Ubiquitous Systems: Computing, Networking and Services. MobiQuitous 2021. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering, vol 419. Springer, Cham. https://doi.org/10.1007/978-3-030-94822-1_30
Download citation
DOI: https://doi.org/10.1007/978-3-030-94822-1_30
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-94821-4
Online ISBN: 978-3-030-94822-1
eBook Packages: Computer ScienceComputer Science (R0)