Abstract
Analyzing human activity through wearable sensors can assist applications connected to context—vigilance and health care. The proposed approach utilizes convolutional and recurrent modeling to express the space–time-dependent constraints from information provided by numerous sensors. We construct a neural network layout, a combination of convolutional neural network and recurrent neural network, which focuses on spatiotemporal characteristics provided by sensor time series data with short listing and learning crucial points by executing a self-attentive procedure. We exhibit validation of our designed strategy on the WISDM dataset and indicate that the self-attention technique achieves a remarkable amendment in presentation over deep neural networks through the union consisting of convolutional and recurrent network architectures. We also present that our designed algorithm offers a statistic performance improvement over other designed approaches for the information (dataset) under consideration. The presented technique allows precise interpretation of activity from numerous body part sensors through various time intervals. We evaluated our technique through various classification methods namely: MLP Classifier, gradient boosting, Random Forest Classifier, Conv2D–LSTM, and self-attention mechanism (proposed). The simulation results show that our proposed method achieved the highest accuracy of 89.58% out of all other methods. We compared the effects of the designed architecture scheme with a baseline constructed through ConvNN and RecNN models, and with earlier designed approaches. We figured out that our self-attention model outperformed the baseline model with high difference among all the considered metrics.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Heikenfeld J, Jajack A, Rogers J, Gutruf P, Tian L, Pan T, Li R, Khine M, Kim J, Wang J (2018) Wearable sensors: modalities, challenges, and prospects. Lab Chip 18(2):217–248
Fiore L, Fehr D, Bodor R, Drenner A, Somasundaram G, Papanikolopoulos N (2008) Multi-camera human activity monitoring. J Intell Rob Syst 52(1):5–43
Bonato P (2010) Wearable sensors and systems. IEEE Eng Med Biol Mag 29(3):25–36
Banos O, Garcia R, Holgado-Terriza JA, Damas M, Pomares H, Rojas I, Saez A, Villalonga C (2014) mHealthDroid: a novel framework for agile development of mobile health applications. In: International workshop on ambient assisted living. Springer, Cham, pp 91–98
Ordóñez FJ, Roggen D (2016) Deep convolutional and LSTM recurrent neural networks for multimodal wearable activity recognition. Sensors 16(1):115
Yang J, Nguyen MN, San PP, Li XL, Krishnaswamy S (2015) Deep convolutional neural networks on multichannel time series for human activity recognition. In: Twenty-fourth international joint conference on artificial intelligence
Putra PU, Shima K, Shimatani K (2022) A deep neural network model for multi-view human activity recognition. PlOS ONE 17(1):e0262181
Weiss GM, Yoneda K, Hayajneh T (2019) Smartphone and smartwatch-based biometrics using activities of daily living. IEEE Access 7:133190–133202
Zhang S, Li Y, Zhang S, Shahabi F, Xia S, Deng Y, Alshurafa N (2022) Deep learning in human activity recognition with wearable sensors: a review on advances. Sensors 22(4):1476
Ferrari A, Micucci D, Mobilio M, Napoletano P (2022) Deep learning and model personalization in sensor-based human activity recognition. J Reliable Intell Environ, pp 1–13
Basak H, Kundu R, Singh PK, Ijaz MF, Wozniak M, Sarkar R (2022) A union of deep learning and swarm-based optimization for 3D human action recognition. Sci Rep 12(1):1–17
Li Y, Wang L (2022) Human activity recognition based on residual network and BiLSTM. Sensors 22(2):635
Habib-Ur-Rehman K, Gorji A, Bourdoux A, Pollin S, Sahli H (2022) Multi-view CNNLSTM architecture for radar-based human activity recognition. IEEE Access 10:24509–24519
Chollet F (2018) Deep learning mit python und keras: das praxis-handbuch vom entwickler der keras-bibliothek. MITP-Verlags GmbH Co, KG
Singh SP, Sharma MK, Lay-Ekuakille A, Gangwar D, Gupta S (2020) Deep ConvLSTM with self-attention for human activity decoding using wearable sensors. IEEE Sens J 21(6):8575–8582
Singh SP, Urooj S (2015) Localized radon polar harmonic transform (LRPHT) based rotation invariant analysis of textured images. Int J Syst Dyn Appl (IJSDA) 4(2):21–41
Singh SP, Urooj S (2015) Wavelet packets based spectral estimation of textured images. In: 2015 IEEE International conference on computational intelligence communication technology. IEEE, pp 651–654
Singh SP, Urooj S (2019) A new computational framework for fast computation of a class of polar harmonic transforms. J Signal Proc Syst 91(8):915–922
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Diwakar, S., Dwivedi, D., Singh, S.P., Sharma, M. (2023). Self-attention-based Human Activity Detection Using Wearable Sensors. In: Rani, A., Kumar, B., Shrivastava, V., Bansal, R.C. (eds) Signals, Machines and Automation. SIGMA 2022. Lecture Notes in Electrical Engineering, vol 1023. Springer, Singapore. https://doi.org/10.1007/978-981-99-0969-8_66
Download citation
DOI: https://doi.org/10.1007/978-981-99-0969-8_66
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-0968-1
Online ISBN: 978-981-99-0969-8
eBook Packages: EnergyEnergy (R0)