Abstract
A success progress of pose estimation approaches motivates the activity recognition used in CCTV-based surveillance systems. In this paper, a method is proposed for recognizing interactive activities between two human objects. Based on articulated joint coordinates obtained from a pose estimation algorithm, the distance and direction feature are extracted from objects to describe both the spatial and temporal relation. The multiclass Support Vector Machine is finally employed for activity classification task. Compared with existing methods using the public interaction dataset, the proposed method outperforms in overall classification accuracy.
Keywords
- Interactive activity recognition
- Articulated-pose feature
- Spatio-temporal relation
This is a preview of subscription content, access via your institution.
Buying options
Preview
Unable to display preview. Download preview PDF.
References
Qiu, Q., Jiang, Z., Chellappa, R.: Sparse dictionary-based representation and recognition of action attributes. In: 2011 IEEE International Conference on Computer Vision (ICCV), pp. 707–714 (2011)
Matikainen, P., Hebert, M., Sukthankar, R.: Representing pairwise spatial and temporal relations for action recognition. In: European Conference on Computer Vision: Part I (ECCV 2010), pp. 508–521 (2010)
Meng, L., Qing, L., Yang, P., Miao, J., Chen, X., Metaxas, D.N.: Activity recognition based on semantic spatial relation. In: 2012 21st International Conference on Pattern Recognition (ICPR), pp. 609–612 (2012)
Moghimi, M., Azagra, P., Montesano, L., Murillo, A.C., Belongie, S.: Experiments on an RGB-D wearable vision system for egocentric activity recognition. In: 2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 611–617 (2014)
Yang, Y., Ramanan, D.: Articulated Human Detection with Flexible Mixtures of Parts. IEEE Transactions on Pattern Analysis and Machine Intelligence 35(12), 2878–2890 (2013)
Cai, J.X., Tang, X., Feng, G.: Learning pose dictionary for human action recognition. In: 2014 22nd International Conference on Pattern Recognition (ICPR), pp. 381–386 (2014)
Ryoo, M.S., Aggarwal, J.K.: Spatio-temporal relationship match: video structure comparison for recognition of complex human activities. In: 2009 IEEE 12th International Conference on Computer Vision, pp. 1593–1600 (2009)
Delaitre, V., Laptev, I., Sivic, J.: Recognizing human actions in still images: a study of bag-of-features and part-based representations. In: British Machine Vision Conference, pp. 97.1–97.11 (2010)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer Science+Business Media Singapore
About this paper
Cite this paper
Huynh-The, T., Bui, DM., Lee, S., Yoon, Y. (2015). Interactive Activity Recognition Using Articulated-Pose Features on Spatio-Temporal Relation. In: Park, DS., Chao, HC., Jeong, YS., Park, J. (eds) Advances in Computer Science and Ubiquitous Computing. Lecture Notes in Electrical Engineering, vol 373. Springer, Singapore. https://doi.org/10.1007/978-981-10-0281-6_50
Download citation
DOI: https://doi.org/10.1007/978-981-10-0281-6_50
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-10-0280-9
Online ISBN: 978-981-10-0281-6
eBook Packages: Computer ScienceComputer Science (R0)