Abstract
Human Activity Recognition (HAR) enables context-aware user experiences where mobile apps can alter content and interactions depending on user activities. Hence, smartphones have become valuable for HAR as they allow large, and diversified data collection. Although previous work in HAR managed to detect simple activities (i.e., sitting, walking, running) with good accuracy using inertial sensors (i.e., accelerometer), the recognition of complex daily activities remains an open problem, specially in remote work/study settings when people are more sedentary. Moreover, understanding the everyday activities of a person can support the creation of applications that aim to support their well-being. This paper investigates the recognition of complex activities exclusively using smartphone accelerometer data. We used a large smartphone sensing dataset collected from over 600 users in five countries during the pandemic and showed that deep learning-based, binary classification of eight complex activities (sleeping, eating, watching videos, online communication, attending a lecture, sports, shopping, studying) can be achieved with AUROC scores up to 0.76 with partially personalized models. This shows encouraging signs toward assessing complex activities only using phone accelerometer data in the post-pandemic world.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Final model of diversity. https://www.internetofus.eu/wp-content/uploads/sites/38/2021/03/D1.3-Final-model-of-diversity.pdf
Ann, O.C., Theng, L.B.: Human activity recognition: a review. In: 2014 IEEE International Conference on Control System, Computing and Engineering (ICCSCE 2014), pp. 389–393 (2014). https://doi.org/10.1109/ICCSCE.2014.7072750
Arif, M., Bilal, M., Kattan, A., Ahamed, S.I.: Better physical activity classification using smartphone acceleration sensor. J. Med. Syst. 38(9), 1–10 (2014). https://doi.org/10.1007/s10916-014-0095-0
Bae, S., et al.: Detecting drinking episodes in young adults using smartphone-based sensors. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 1(2), 1–36 (2017)
Coorevits, L., Coenen, T.: The rise and fall of wearable fitness trackers, August 2016
Giunchiglia, F., et al.: A worldwide diversity pilot on daily routines and social practices (2020). University of Trento Technical report. No. #DISI-2001-DS-0 (23), 36–44, April 2021
Guvensan, M.A., Dusun, B., Can, B., Turkmen, H.I.: A novel segment-based approach for improving classification performance of transport mode detection. Sensors 18(1) (2018). https://doi.org/10.3390/s18010087, https://www.mdpi.com/1424-8220/18/1/87
Hassan, M.M., Uddin, M.Z., Mohamed, A., Almogren, A.: A robust human activity recognition system using smartphone sensors and deep learning. Futur. Gener. Comput. Syst. 81, 307–313 (2018). https://doi.org/10.1016/j.future.2017.11.029, https://www.sciencedirect.com/science/article/pii/S0167739X17317351
Jobanputra, C., Bavishi, J., Doshi, N.: Human activity recognition: a survey. Procedia Comput. Sci. 155, 698–703 (2019). https://doi.org/10.1016/j.procs.2019.08.100, https://www.sciencedirect.com/science/article/pii/S1877050919310166, the 16th International Conference on Mobile Systems and Pervasive Computing (MobiSPC 2019), The 14th International Conference on Future Networks and Communications (FNC-2019), The 9th International Conference on Sustainable Energy Information Technology
Laput, G., Harrison, C.: Sensing Fine-Grained Hand Activity with Smartwatches, pp. 1–13. Association for Computing Machinery, New York, NY, USA (2019). https://doi.org/10.1145/3290605.3300568
Likamwa, R., Liu, Y., Lane, N., Zhong, L.: Moodscope: building a mood sensor from smartphone usage patterns, June 2013. https://doi.org/10.1145/2462456.2464449
Mea, V.D., Quattrin, O., Parpinel, M.: A feasibility study on smartphone accelerometer-based recognition of household activities and influence of smartphone position. Inform. Health Soc. Care 42(4), 321–334 (2017). https://doi.org/10.1080/17538157.2016.1255214, pMID: 28005434
Meegahapola, L., Bangamuarachchi, W., Chamantha, A., Ruiz-Correa, S., Perera, I., Gatica-Perez, D.: Sensing eating events in context: a smartphone-only approach. IEEE Access 10(ARTICLE) (2022)
Meegahapola, L., Gatica-Perez, D.: Smartphone sensing for the well-being of young adults: a review. IEEE Access 9, 3374–3399 (2020). https://doi.org/10.1109/ACCESS.2020.3045935
Mohamed, R., Zainudin, M.N.S., Perumal, T., Mustapha, N.: Multi-label classification for physical activity recognition from various accelerometer sensor positions (2018)
Plötz, T., Guan, Y.: Deep learning for human activity recognition in mobile computing. Computer 51(5), 50–59 (2018). https://doi.org/10.1109/MC.2018.2381112
Ranasinghe, S., Machot, F.A., Mayr, H.C.: A review on applications of activity recognition systems with regard to performance and evaluation. Int. J. Distrib. Sens. Netw. 12(8), 1550147716665520 (2016). https://doi.org/10.1177/1550147716665520
Servia-Rodríguez, S., Rachuri, K.K., Mascolo, C., Rentfrow, P.J., Lathia, N., Sandstrom, G.M.: Mobile sensing at the service of mental well-being: a large-scale longitudinal study. In: Proceedings of the 26th International Conference on World Wide Web, pp. 103–112 (2017)
Shoaib, M., Bosch, S., Incel, O.D., Scholten, H., Havinga, P.J.M.: Complex human activity recognition using smartphone and wrist-worn motion sensors. Sensors 16(4) (2016). https://doi.org/10.3390/s16040426, https://www.mdpi.com/1424-8220/16/4/426
Straczkiewicz, M., James, P., Onnela, J.P.: A systematic review of smartphone-based human activity recognition for health research (2021)
Straczkiewicz, M., Glynn, N., Harezlak, J.: On placement, location and orientation of wrist-worn tri-axial accelerometers during free-living measurements, May 2019. https://doi.org/10.3390/s19092095
Wu, W., Dasgupta, S., Ramirez, E.E., Peterson, C., Norman, G.J.: Classification accuracies of physical activities using smartphone motion sensors. J. Med. Internet Res. 14(5), e130 (2012). https://doi.org/10.2196/jmir.2208, http://www.jmir.org/2012/5/e130/
Zhu, C., Sheng, W.: Wearable sensor-based hand gesture and daily activity recognition for robot-assisted living. IEEE Trans. Syst. Man Cybern. Part A Syst. Humans 41(3), 569–573 (2011). https://doi.org/10.1109/TSMCA.2010.2093883
Acknowledgement
This work was funded by the European Union’s Horizon 2020 WeNet project, under grant agreement 823783. We also thank Darshana Rathnayake (Singapore Management University, Singapore) for discussions.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 ICST Institute for Computer Sciences, Social Informatics and Telecommunications Engineering
About this paper
Cite this paper
Bouton-Bessac, E., Meegahapola, L., Gatica-Perez, D. (2023). Your Day in Your Pocket: Complex Activity Recognition from Smartphone Accelerometers. In: Tsanas, A., Triantafyllidis, A. (eds) Pervasive Computing Technologies for Healthcare. PH 2022. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering, vol 488. Springer, Cham. https://doi.org/10.1007/978-3-031-34586-9_17
Download citation
DOI: https://doi.org/10.1007/978-3-031-34586-9_17
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-34585-2
Online ISBN: 978-3-031-34586-9
eBook Packages: Computer ScienceComputer Science (R0)