Abstract
A method was developed to enable users of video sharing websites to easily retrieve video scenes relevant to their interests. The system analyzes both text and non-text aspects of a user’s comment and then retrieves and displays relevant scenes along with attached comments. The text analysis works in tandem with non-text features, namely, the selected area and temporal duration associated with user comments. In this way, our system supports a better-organized retrieval of scenes that have been commented on with a higher degree of relevancy than conventional methods, such as using matching keywords. We describe our method and the relation between the scenes and discuss a prototype system.
Similar content being viewed by others
References
Allen JF (1983) Maintaining knowledge about temporal intervals. In: Communications of the ACM, vol 26, pp 832–843
Baluja S, Seth R, Sivakumar D, Jing Y, Yagnik J, Kumar S, Ravichandran D, Aly M (2008) Video suggestion and discovery for YouTube: taking random walks through the view graph. In: Proc. of the 17th international world wide web conference (WWW2008), pp 895–904
Braudy L, Cohen M (2004) Film theory and criticism. Oxford University Press, Oxford
Dao MS, Babaguchi N (2008) Sports event detection using temporal patterns mining and web-casting text. In: Proc. of the 1st ACM workshop on analysis and retrieval of events/actions and workflows in video streams (AREA2008), vol 26, pp 33–40
Fukino N, Ma Q, Sumiya K, Tanaka K (2003) Generating football video summery using news article. In: Proc. of the 14th data engineering workshop (DEWS2003), vol 8-P-03 (in Japanese)
Gong Y (1999) An accurate and robust method for detecting video shot boundaries. In: Proc. of IEEE international conference on multimedia computing and systems (ICMCS’99), vol 1, pp 850–854
GoogleVideo (2010) http://video.google.com/. Accessed 22 Apr 2010
Karpenko A, Aarabi P (2008) Tiny videos: non-parametric content-based video retrieval and recognition. In: Proc. of the tenth IEEE international symposium on multimedia (ISM 2008), pp 619–624
Kimura T, Sumiya K, Tanaka H (2005) A video editing support system using users’ gazes. In: Proc. of IEEE Pacific Rim conference on communications, computers and signal processing (PACRIM2005), pp 149–152
Kitayama D, Oda N, Sumiya K (2008) Organizing user comments in a social video sharing system by temporal duration and pointing region. In: Proc. of international workshop on information-explosion and next generation search (INGS2008), pp 55–58
Masuda T, Yamamoto D, Ohira S, Nagao K (2008) Video scene retrieval using online video annotation. In: Lecture notes on artificial intelligence. Springer, Heidelberg
Miura K, Yamada I, Sumiyoshi H, Yagi N (2006) Automatic generation of a multimedia encyclopedia from tv programs by using closed captions and detecting principal video objects. In: Proc. of the eighth IEEE international symposium on multimedia, pp 873–880
Miyamori H, Nakamura S, Tanaka K (2005) Generation of views of TV content using TV viewers’ perspectives expressed in live chats on the web. In: Proc. of the 13th annual ACM international conference on multimedia (ACM Multimedia2005), pp 853–861
Nakamura S, Shimizu M, Tanaka K (2008) Can social annotation support users in evaluating the trustworthiness of video clips?. In: Proc. of the 2nd ACM workshop on information credibility on the web (WICOW’08), pp 59–62
NicoNicoDouga. http://www.nicovideo.jp/
Pradhan S, Tajima K, Tanaka K (1999) A query model for retrieving relevant intervals within a video stream. In: Proc. of IEEE international conference on multimedia computing and systems, vol 2, pp 788–792
Saraceno C, Leonardi R (1997) Identification of successive correlated camera shots using audio and video information. In: Proc. of international conference on image processing (ICIP’97), vol 3, pp 116–119
Shen E Y-T, Lieberman H, Davenport G (2009) What’s next?: emergent storytelling from video collections. In: Proc. of the 27th international conference on human factors in computing systems, pp 809–818
Su J-H, Huang Y-T, Tseng VS (2008) Efficient content-based video retrieval by mining temporal patterns. In: Proc. of the 9th international workshop on multimedia data mining: held in conjunction with the ACM SIGKDD 2008, pp 36–42
Sundaram H, Chang S-F (2000) Determining computable scenes in films and their structures using audio-visual memory models. In: Proc. of the eighth ACM international conference on multimedia, pp 95–104
Tahaghoghi SMM, Williams HE, Thom JA, Volkmer T (2005) Video cut detection using frame windows. In: Proc. of the twenty-eighth Australasian conference on computer science (ACSC ’05), pp 193–199
Uehara H, Yoshida K (2005) Annotating TV drama based on viewer dialogue—analysis of viewers’ attention generated on an internet bulletin board. In: Proc. of IEEE/IPSJ international symposium on applications and the internet (SAINT2005), pp 334–340
Wang J, Chua T-S (2008) A framework for video scene boundary detection. In: Proc. of the tenth ACM international conference on multimedia, pp 243–246
Wu S, Chen Y (2007) Mining nonambiguous temporal patterns for interval-based events. In: IEEE transactions on knowledge and data engineering, vol 19, pp 742–758
Wu X, Takimoto M, Satoh S, Adachi J (2008) Scene duplicate detection based on the pattern of discontinuities in feature point trajectories. In: Proc. of the 16th ACM international conference on multimedia (MM’08), pp 51–60
Yamamoto D, Nagao K (2004) iVAS: web-based video annotation system and its applications. In: Proc. of the3rd international semantic web conference
Yamamoto D, Masuda T, Ohira S, Nagao K (2008) Video scene annotation based on web social activities. In: IEEE multimedia, pp 22–32
Yoshitaka A, Miyake M (2001) Scene detection by audio-visual features. In: Proc. of IEEE international conference on multimedia and expo (ICME2001), vol 3, pp 48–51
YouTube. http://www.youtube.com/
Zanetti S, Zelnik-Manor L, Perona P (2008) A walk through the web’s video clips. In: Proc. of IEEE computer society conference on computer vision and pattern recognition workshops (CVPRW ’08), pp 1–8
Acknowledgements
This research was supported in part by a Grant-in-Aid for Scientific Research (B)(2) 20300039 and Grant-in-Aid for JSPS Fellows 21.197 from the Ministry of Education, Culture, Sports, Science, and Technology of Japan.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Wakamiya, S., Kitayama, D. & Sumiya, K. Scene extraction system for video clips using attached comment interval and pointing region. Multimed Tools Appl 54, 7–25 (2011). https://doi.org/10.1007/s11042-010-0531-1
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-010-0531-1