Abstract
With the rapid increase in digital technology, most research areas are involved in human activity recognition, which can help to analyze the activities of patients. A novel approach for human activity recognition in egocentric video has been invoked in this research article. Generally, only the objects are identified, but the actions are not recognized. With this motivation and new trends, this paper presents an efficient technique to recognize the activities. In our approach, first the various activity dataset is trained, and the feature vector values are stored for various activities, which are applied to the testing inputs. Here, we use a filtering technique, i.e., a median filter followed by a segmentation method using watershed and feature extraction, such as a Histogram of Oriented Gradient (HOG), Color and GiST and a combination of all Features. Features are reduced using a genetic algorithm, and classification is done using Support Vector Machine (SVM) and a Random Forest classifier. The experimental results demonstrate that the Random Forest classifier outperformed the SVM classifier.
Similar content being viewed by others
References
Bieniek A, Moga A (2000) An efficient watershed algorithm based on connected components. Pattern Recogn 33:907–916
Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection’ Computer Vis ion and Pattern Recognition, CVPR, IEEE Computer Society Conference, vol 1, p 886–893
de San Roman PP, Benois-Pineau J, Domenger J-P, de Rugy A, Paclet F, Cataert D (2017) Saliency Driven Object recognition in egocentric videos with deep CNN: toward application in assistance to Neuroprostheses. Comput Vis Image Underst 164:82–91
Fathi A, Farhadi A, Rehg JM (2011) Understanding egocentric activities. In: ICCV, Washington, DC, USA, p 407–414
Fong S, Liu K, Cho K (2016) Improvised methods for tackling big data stream mining challenges: case study of human activity recognition. J Supercomput 72(10):3927–3959
Gemmell J, Bell G, Lueder R (2006) Mylifebits: a personal database for everything. Commun ACM 49(1):88–95
Hassan M, Ahmad T (2014) A Review on Human Actions Recognition Using Vision Based Techniques. J Image Graph 2(1):28–32
Hori T, Aizawa K (2003) Context-based video retrieval system for the life-log applications. In: SIGMM international workshop on multimedia information retrieval, p 31–38
Jacobson L, Kanber B (2015) Genetic Algorithms in Java Basics. Apress, New York
Knoop S, Vacek S, Dillmann R, Brännström S, Christensen HI (2005) Extraction, evaluation, selection and classification of motion features for human activity recognition. Universität Karlsruhe, Internal Report
Kuncheva LI (2014) Combining Pattern Classifiers: Methods and Algorithms, 2nd edn. Wiley, Hoboken
Li C, Lin M, Yang LT (2014) Integrating the enriched feature with machine learning algorithms for human movement and fall detection. J Supercomput 67(3):854–865
Maurer U, Smailagic A, Siewiorek D, Deisher M (2006) Activity recognition and monitoring using multiple sensors on different body positions. In: Proc. Int. Workshop Wearable and Implantable Body Sensor Networks. https://doi.org/10.1109/BSN.2006.6
Niebles JC, Han B, Fei-Fei L (2010) Efficient extraction of human motion volumes by tracking. In: IEEE Computer Society Conference on Computer Vision, San Francisco, CA, USA, pp 655–692
OpenCV-Introduction to Support Vector Machines http://docs.opencv.org/2.4/doc/tutorials/ml/introduction_to_svm/introduction_to_svm.html. Accessed 22 July 2016
Ortis A, Farinella GM, D’Amico V, Addesso L, Torrisi G, Battiato S (2017) Organizing egocentric videos of daily living activities. Pattern Recogn 72:207–218
Rachmadi RF, Ketut Eddy Purnama I Large Scale Scene Classification Using Gist Feature. Institut Teknologi Sepuluh Nopember Surabaya Indonesia 60111
Pirsiavash H, Ramanan D (2012) Detecting activities of daily living in first-person camera views. In: CVPR, Providence, RI, USA, p 2847–2854
Salman N (2006) Image Segmentation Based on Watershed and Edge Detection Techniques. Int Arab J Inf Technol 3(2):104–110
Sanal Kumar KP, Bhavani R (2016) Analysis of SVM and kNN Classifiers For Egocentric Activity Recognition. In: Proceedings of the International Conference on Informatics and Analytics (Pondicherry, India: ACM) August 25–26, 2016
Sanal Kumar KP, Bhavani R (2017) Human activity recognition in egocentric video using PNN, SVM, kNN and SVM+kNN classifiers. Cluster Computing
Sanal Kumar KP, Bhavani R (2017) Activity Recognition in Egocentric video using SVM, kNN and Combined SVMkNN Classifiers. IOP Conf Series Mater Sci Eng 225:012226
Solomon C, Breckon T (2010) Fundamentals of Digital Image Processing. Wiley, Hoboken, pp 1–18
Song S, Chandrasekhar V, Cheung N, Narayan S, Li L, Lim J (2014) Activity Recognition in Egocentric Life-logging Videos, Computer Vision ACCV 2014 Workshops, Singapore, Nov 2014
Suresha M, Shilpa NA, Soumya B (2012) Apples Grading based on SVM Classifier. In: National Conference on Advanced Computing and Communications, April 2012
Tukey JW (1974) Nonlinear (Nonsuperposable). Methods for Smoothing Data. Conference Record EASCON, pp 673–685
Vaca-Castano G, Das S, Sousa JP, Lobo ND, Shah M (2016) Improved scene identification and object detection on egocentric vision of daily activities. Comp Vis Image Underst
Wang H, Schmid C (2013) Action Recognition with Improved Trajectories ICCV '13 Proceedings of the 2013 I.E. International Conference on Computer Vision, p 3551–3558 December 01–08, 2013
Wang X, Gao L, Song J, Zhen X, Sebe N, Shen HT (2017) Deep Appearance and Motion Learning for Egocentric Activity Recognition. Neurocomputing
Wu J, Wei Z, Chang Y (June 2010) Color and Texture Feature For Content Based Image Retrieval. Int J Digit Content Technol Appl 4(3):43–49
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Sanal Kumar, K.P., Bhavani, R. Human activity recognition in egocentric video using HOG, GiST and color features. Multimed Tools Appl 79, 3543–3559 (2020). https://doi.org/10.1007/s11042-018-6034-1
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-018-6034-1