Abstract
Recognizing activities of daily living is useful for ambient assisted living. In this regard, the use of wearable cameras is a promising technology. In this paper, we propose a novel approach for recognizing activities of daily living using egocentric viewpoint video clips. First, in every frame the appearing objects are detected and labelled depending if they are being used or not by the subject. Later, the video clip is divided into spatio temporal bins created with an object centric cut. Finally, a support vector machine classifier is computed using a spatio-temporal flexible kernel between video clips. The validity of the proposed method has been proved by conducting experiments in the ADL dataset. Results confirm the suitability of using the space-time location of objects as information for the classification of activities using an egocentric viewpoint.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
INE. Mujeres y hombres en España. Technical report, Catálogo de publicaciones de la Administración General del Estado (2014)
Nguyen, T.-H.-C., Nebel, J.-C., Florez-Revuelta, F.: Recognition of activities of daily living with egocentric vision: a review. Sensors 16(1), 72 (2016)
Hodges, S., Berry, E., Wood, K.: Sensecam: a wearable camera that stimulates and rehabilitates autobiographical memory. Memory 19(7), 685–696 (2011). PMID: 21995708
Fathi, A., Farhadi, A., Rehg, J.M.: Understanding egocentric activities. In: 2011 IEEE International Conference on Computer Vision (ICCV), pp. 407–414. IEEE, November 2011
Pirsiavash, H., Ramanan, D.: Detecting activities of daily living in first-person camera views. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE (2012)
McCandless, T., Grauman, K.: Object-centric spatio-temporal pyramids for egocentric activity recognition. In: British Machine Vision Conference, BMVC 2013, Bristol, UK, 9–13 September 2013
Matsuo, K., Yamada, K., Ueno, S., Naito, S.: An attention-based activity recognition for egocentric video. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR Workshops 2014, Columbus, OH, USA, 23–28 June 2014, pp. 565–570 (2014)
Wang, H., Schmid, C.: Action recognition with improved trajectories. In: IEEE International Conference on Computer Vision (ICCV) (2013)
Lazebnik, S., Schmid, C., Ponce, J.: Beyond bags of features: spatial pyramid matching for recognizing natural scene categories. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), vol. 2, pp. 2169–2178 (2006)
Rodriguez, M., Orrite, C., Medrano, C., Makris, D.: A time flexible kernel framework for video-based activity recognition. Image Vis. Comput. 48, 26–36 (2016)
Jebara, T., Kondor, R., Howard, A.: Probability product kernels. J. Mach. Learn. Res. 5, 819–844 (2004)
Choi, J., Wang, Z., Lee, S.-C., Jeon, W.J.: A spatio-temporal pyramid matching for video retrieval. Comput. Vis. Image Underst. 117(6), 660–669 (2013)
Acknowledgments
This work was partially supported by Spanish Grant TIN2013- 45312-R (MINECO), Gobierno de Aragon and the European Social Found.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Rodriguez, M., Orrite, C., Medrano, C. (2017). Space-Time Flexible Kernel for Recognizing Activities from Wearable Cameras. In: Alexandre, L., Salvador Sánchez, J., Rodrigues, J. (eds) Pattern Recognition and Image Analysis. IbPRIA 2017. Lecture Notes in Computer Science(), vol 10255. Springer, Cham. https://doi.org/10.1007/978-3-319-58838-4_56
Download citation
DOI: https://doi.org/10.1007/978-3-319-58838-4_56
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-58837-7
Online ISBN: 978-3-319-58838-4
eBook Packages: Computer ScienceComputer Science (R0)