Abstract
In today’s world, where technology is advancing every single day, new methodologies are being developed, and are brought in everyday use making our lives simpler, faster, safer, and powerful. Similarly, Human Activity Recognition (HAR) is getting more popular with all the revolutions made in the technologies. Sensor Network Technology is used in industrial applications, smart homes and system. A massive amount of data can be obtained from these sensors which are linked to the human body. Recognition of Human Activities using these sensors, and wearable technologies has been actively studied. Behavior Recognition seeks to distinguish one or more people’s activities and goals through a collection of observations on the actions and environmental conditions of the person. Health surveillance, aged treatment, and plenty of other domains can be used to automatically understand the behavioral context. An existing dataset consisting of 10 subjects (5 females, 5 males) is being used in the paper, which incorporates both young and old volunteers between 19 and 60 years of old with weights ranging from 55 to 85 kg. The dataset reflects motion data collected when subjects are engaged in 11 separate (static and dynamic) smart home activities: computer usage (1 min), telephone conversation (1 min), vacuum cleaning (1 min), book reading (1 min), TV watching (1 min), ironing (1 min), walking (1 min), exercise (1 min), cooking (1 min), drinking (20 times), hair brushing (1 min) (20 times). Most of the activities are similar because of the multi sensor environment which makes it more difficult. Using three tri axial IMU (inertial measurement unit), Magnetometer, Accelerometer, Gyroscope sensors attached to the subject area of the hand, chest, and thigh, using Machine Learning we introduced a better model to prognosticate the human activity. We have applied various machine learning classification algorithms like Random Forest Classifier, K-Nearest Neighbors, Decision Tree, Multi-layer Perceptron Classifier, Extra Tree Classifier, Ensemble Extra Trees Classifier, Label Propagation and Label Spreading. The experimental results are tabulated and analyzed, and might be effectively accustomed to recognize human activities in terms of efficiency and accuracy.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Kwapisz, Jennifer R., Weiss, Gary M., Moore, Samuel A.: Activity recognition using cell phone accelerometers. ACM SIGKDD Explor. Newslett. 12(2), 74–82 (2010)
Anguita, D., Ghio, A., Oneto, L., Parra, X., Reyes-Ortiz, J.L.: Human activity recognition on smartphones using a multiclass hardware friendly support vector machine. Springer International Workshop on Ambient Assisted Living Lecture notes in Computer Science, vol. 7657, pp. 216–223 (2012)
Varkey, J., Pompili, D., Walls, T.: Human motion recognition using a wireless sensor-based wearable system. In: Proceedings of Ubiquitous Computing, pp. 897–910 (2012)
Jalal, A., Khan, M.A., Hasan, A.S.: Wearable sensor based human behavior understanding and recognition in daily life for smart environments. In: Proceedings of IEEE Conference on FIT (2018)
Wu, J., Sun, L., Jafari, R.: A wearable system for recognizing American sign language in real-time using IMU and surface EMG sensors. In: Proceedings of IEEE Journal of Biomedical and Health Informatics, pp. 1281–1290 (2016)
Roobini, S., FenilaNaomi, J.: Smartphone sensor based human activity recognition using deep learning models. Int. J. Recent Technol. Eng. 8(1), ISSN: 2277-3878
Randhawa, P., Shanthagiri, V., Kumar, A., Yadav, V.: Human activity detection using machine learning methods from wearable sensors (2020). https://doi.org/10.1108/sr-02-2020-0027
D’souza, W.T., Kavitha, R.: Human activity recognition using accelerometer and gyroscope sensors (2017). https://doi.org/10.21817/ijet/2017/v9i2/170902134
Ronao, C.A., Cho, S.-B.: Human activity recognition with smartphone sensors using deep learning neural networks. Expert Syst. Appl. 59, 235–244 (2016)
Bayat, A., Pomplun, M., Tran, D.A.: A study on human activity recognition using accelerometer data from smartphones. In: The 11th International Conference on Mobile Systems and Pervasive Computing (MobiSPC-2014)
Shoaib, M., Scholten, H., Havinga, P.J.M.: Towards physical activity recognition using smartphone sensors. In: 2013 IEEE 10th International Conference on Ubiquitous Intelligence &Computing and 2013 IEEE 10th International Conference on Autonomic & Trusted Computing, pp. 80–87 (2013)
Anguita, D., Ghio, A., Oneto, L., Parra, X., Reyes-Ortiz, J.L.: A public domain dataset for human activity recognition using smartphone’s. In: ESANN 2013 proceedings, European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning. Bruges (Belgium) (2013)
Mannini, A., Sabatini, A.M.: Machine learning methods or classifying human physical activity from on body accelerometers. Sensors 2010(10), 1154–1175 (2010)
Rosati, S., Balestra, G., Knaflitz, M.: Comparison of different sets of features for human activity recognition by wearable sensors. Sensors 18(12), 4189 (2018)
Krishnan, N.C., Panchanathan, S.: Analysis of low resolution accelerometer data for continuous human activity recognition. ICASSP (2008)
Tapia, E.M., Intille, S.S. et al.: Real-time recognition of physical activities and their intensities using wireless accelerometers and a heart rate monitor. In: Proceedings of the 2007 11th IEEE International Symposium on Wearable Computers, 1–4 (2007)
Kunze, K., Barry, M., Heinz, E.A., Lukowicz, P., Majoe, D., Gutknecht, J.: Towards recognizing Tai Chi—an initial experiment using wearable sensors (2006)
Kumar, M., Shenbagaraman, V.M., Ghosh, A.: Predictive data analysis for energy management of a smart factory leading to sustainability. In: Favorskaya, M.N., Mekhilef, S., Pandey, R.K., Singh, N. (eds.) Innovations in Electrical and Electronic Engineering. Springer, pp. 765–773 [ISBN 978-981-15-4691-4] (2020)
Ravi, N., Dandekar, N., Mysore, P., Littman, M.L.: Activity recognition from accelerometer data. In: IAAI’05 Proceedings of the 17th Conference on Innovative Applications of Artificial Intelligence, Vol. 3, pp. 1541–1546, Pittsburg, Pennsylvania (2005)
Mandal, S., Biswas, S., Balas, V.E., Shaw, R.N., Ghosh, A.: Motion prediction for autonomous vehicles from Lyft dataset using deep learning. In: 2020 IEEE 5th International Conference on Computing Communication and Automation (ICCCA), Greater Noida, India, pp. 768–773 (2020). https://doi.org/10.1109/iccca49541.2020.9250790
Mandal, S., Balas, V.E., Shaw, R.N., Ghosh, A.: Prediction analysis of idiopathic pulmonary fibrosis progression from OSIC dataset. In: 2020 IEEE International Conference on Computing, Power and Communication Technologies (GUCON), Greater Noida, India, pp. 861–865 (2020). https://doi.org/10.1109/gucon48875.2020.9231239
AIR University Dataset: Intelligent Media—Wearable Smart Home Activities (IM-WSHA) Dataset
Zhu, X., Ghahramani, Z.: Learning from labeled and unlabeled data with label propagation. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, Technical Report. CMU-CALD-02–107 (2002)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Palimkar, P., Bajaj, V., Mal, A.K., Shaw, R.N., Ghosh, A. (2022). Unique Action Identifier by Using Magnetometer, Accelerometer and Gyroscope: KNN Approach. In: Bianchini, M., Piuri, V., Das, S., Shaw, R.N. (eds) Advanced Computing and Intelligent Technologies. Lecture Notes in Networks and Systems, vol 218. Springer, Singapore. https://doi.org/10.1007/978-981-16-2164-2_48
Download citation
DOI: https://doi.org/10.1007/978-981-16-2164-2_48
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-16-2163-5
Online ISBN: 978-981-16-2164-2
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)