Abstract
Event searching in videos is important task for video genre classification, event detection, content-based video searching, etc. Being able to search for a particular event without going through the whole video saves lots of time for the users. It also helps in indexing the videos based on an event in it, so that users can retrieve videos with same event efficiently. The proposed approach is three-step approach. The approach uses the map-reduce concept to decrease the overall processing time for a video using deep learning. As the number of frames in a video can be large but events are limited so to solve this issue, we have used Eratosthenes sieve based key-frame extraction technique for extracting key-frames that describe the video efficiently, then from the key-frames, we have extracted the events in them, and then indexing of the events is done after finding event boundaries.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Dean, J., Ghemawat, S.: MapReduce: simplified data processing on large clusters. Commun. ACM 51(1), 107–113 (2008)
Kumar, K., et al.: ESUMM: Event SUMMarization on scale-free network. IETE Tech. Rev. (2018)
Szegedy, C., et al.: Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016)
Krishan, et al.: Deep Event Learning boosT-up Approach: DELTA. Multimed. Tools Appl. (2018)
Vinyals, O., et al.: Show and tell: a neural image caption generator. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE (2015)
Kumar, et al.: F-DES: fast and deep event summarization. IEEE Trans. Multimed. (2017)
Narkiewicz, W.: Eratosthenes’ sieve method. Number Theory. World Scientific Publishing Company (1984)
Abhay, et al.: An automated hierarchical framework for player recognition in sports image. In: ICVIP, pp. 1–10 (2017)
Hodosh, M., Young, P., Hockenmaier, J.: Framing image description as a ranking task: data, models and evaluation metrics. J. Artif. Intell. Res. 47, 853–899 (2013)
Singh, G., et al.: PICS: a novel technique for video summarization. MISP (2017)
Zhuang, Y., Rui, Y., Huang, T.S., Mehrotra, S.: Adaptive key frame extraction using unsupervised clustering. In: Proceedings of the International Conference on Image Processing, vol. 1, pp. 866–870. IEEE (1998)
Kumar, K., et al.: Key-lectures: keyframes extraction in video lectures. MISP. Springer (2017)
Brunelli, R., Mich, O., Modena, C.M.: A survey on the automatic indexing of video data. J. Vis. Commun. Image Represent. 10(2), 78–112 (1999)
Kumar, K., et al.: D-CAD: deep and crowded anomaly detection. In: ICCCT (2017)
Mundur, P., Rao, Y., Yesha, Y.: Keyframe-based video summarization using Delaunay clustering. Int. J. Digit. Libr. 6(2), 219–232 (2006)
Krishan, et al.: V-LESS: a video from linear event summaries. In: CVIP (2017)
de Avila, S.E.F., Lopes, A.P.B., et al.: VSUMM: a mechanism designed to produce static video summaries and a novel evaluation method. Pattern Recognit. Lett. 32(1), 56–68 (2011)
Krishan, et al.: SOMES: an efficient SOM technique for Event Summarization in multi-view surveillance videos. In: ICACNI (2017)
Karpathy, A., Li, F.-F.: Deep visual-semantic alignments for generating image descriptions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2015)
Goodfellow, I., et al.: Deep Learning, vol. 1. MIT Press, Cambridge (2016)
Kumar, et al.: Event BAGGING: a novel event summarization approach in multi-view surveillance videos. In: IEEE IESC (2017)
de Avila, S.E.F., Lopes, A.P.B., et al.: VSUMM: a mechanism designed to produce static video summaries and a novel evaluation method. Pattern Recognit. Lett. 32(1), 56–68 (2011)
Furini, M., Geraci, F., Montangero, M., Pellegrini, M.: STIMO: still and moving video storyboard for the web scenario. Multimed. Tools Appl. 46(1), 47–69 (2010)
Video open project storyboard. https://open-video.org/results.php?size=extralarge. Accessed May 2016
Cong, Y., Yuan, J., Luo, J.: Towards scalable summarization of consumer videos via sparse dictionary selection. IEEE Trans. Multimed. 14(1), 66–75 (2012)
Guan, G., Wang, Z., Lu, S., Da Deng, J., Feng, D.: Keypoint based keyframe selection. IEEE Trans. Circuits Syst. Video Technol. 23(4), 729–734 (2013)
Mei, S., Guan, G., et al.: Video summarization via minimum sparse reconstruction. Pattern Recognit. 48(2), 522–533 (2015)
Kumar, K., et al.: Eratosthenes sieve based key-frame extraction technique for event summarization in videos. Multimed. Tools Appl. 77(6), 7383–7404 (2018)
Kumar, K., et al.: Equal partition based clustering approach for event summarization in videos. In: IEEE SITIS, pp. 119–126 (2016)
Shetty, R.: Natural Language Description of Images and Videos. Machine Learning and Data Mining (2016)
http://www.image-net.org/challenges/LSVRC/2015/results. Accessed 16 June 2018
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Solanki, A., Bamrara, R., Kumar, K., Singh, N. (2020). VEDL: A Novel Video Event Searching Technique Using Deep Learning. In: Pant, M., Sharma, T., Verma, O., Singla, R., Sikander, A. (eds) Soft Computing: Theories and Applications. Advances in Intelligent Systems and Computing, vol 1053. Springer, Singapore. https://doi.org/10.1007/978-981-15-0751-9_83
Download citation
DOI: https://doi.org/10.1007/978-981-15-0751-9_83
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-15-0750-2
Online ISBN: 978-981-15-0751-9
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)