Abstract
We propose a novel scheme to address video concept learning by leveraging social media, one that includes the selection of web training data and the transfer of subspace learning within a unified framework. Due to the existence of cross-domain incoherence resulting from the mismatch of data distributions, how to select sufficient positive training samples from scattered and diffused social media resources is a challenging problem in the training of effective concept detectors. In this paper, given a concept, the coherent positive samples from web images for further concept learning are selected based on the degree of image coherence. Then, by exploiting both the selected dataset and video keyframes, we train a robust concept classifier by means of a transfer subspace learning method. Experiment results demonstrate that the proposed approach can achieve constant overall improvement despite cross-domain incoherence.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Wessel, K.: TRECVID-2007 High-Level Feature task:Overview, http://wwwnlpir.nist.gov/projects/tvpubs/tv7.slides/
Jinqiao, W., et al.: IVA-NLPR-IA-CAS TRECVID 2009: High LevelFeatures Extraction, http://www-nlpir.nist.gov/projects/tvpubs/tv.pubs.org.html#2009
Sun, Y.: A Novel Region-based Approach to Visual Concept Modeling using Web Images. In: ACM Multimedia (2008)
Jinhui, T.: To construct optimal training set for video annotation. In: ACM Multimedia (2006)
Su, Y.: Cross-database age estimation based on transfer learning. In: ICASSP 2010, pp. 1270–1273 (2010)
Fergus, R.: Learning Object Categories from Google’s Image Search. In: ICCV, vol. 2, pp. 1816–1823 (2005)
Kennedy, L.S.: Generating diverse and representative image search results for landmarks. In: WWW, pp. 297–306 (2008)
Baudat, G., et al.: Feature vector selection and projection using kernels. Neurocomputing 55(1-2), 21–38 (2003)
Yang, J.: Cross-Domain Video Concept Detection using Adaptive SVMs. In: ACM Multimedia 2007, pp. 188–197 (2007)
Chang, S.: Columbia University/VIREO-CityU/IRIT TRECVID2008 High-Level Feature Extraction and Interactive Video Search, http://www-nlpir.nist.gov/projects/tvpubs/tv8.papers/
Bay, H., Ess, A., Tuytelaars, T., Gool, L.V.: Surf: Speeded up robust features. Computer Vision and Image Understanding 110(3), 346–359 (2008)
Borth, D., Ulges, A., Breuel, T.M.: Automatic concept-to-query mapping for web-based concept detector training. In: ACM Multimedia 2011, New York, NY, USA, pp. 1453–1456 (2011)
Donoho, D.: For most large underdetermined systems of linear equations the minimal l1-norm solution is also the sparsest solution. Comm. Pure and Applied Math. 59(6), 797–826 (2006)
Efron, T.J.I., Bradley, H., Tibshirani, R.: Least angle regression.  Annals of Statistics 32(2), 407–499 (2004)
Huiskes, M.J., Thomee, B., Lew, M.S.: New trends and ideas in visual concept detection: the mir flickr retrieval evaluation initiative. In: MIR 2010, New York, NY, USA, pp. 527–536 (2010)
Jiang, Y.-G., Yang, J., Ngo, C.-W., Hauptmann, A.G.: Representations of keypoint-based semantic concept detection: A comprehensive study. IEEE Transactions on Multimedia 12, 42–53 (2010)
Lowe, D.G.: Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision 60(2), 91–110 (2004)
Mairal, J., Bach, F., Ponce, J., Sapiro, G.: Online learning for matrix factorization and sparse coding. J. Mach. Learn. Res. 11, 19–60 (2010)
Over, P., Awad, G., Rose, R.T., Fiscus, J.G., Kraaij, W., Smeaton, A.F.: Trecvid 2008 - goals, tasks, data, evaluation mechanisms and metrics. In: TRECVID Workshop (2008)
Ramirez, I., Sprechmann, P., Sapiro, G.: Classification and Clustering via Dictionary Learning with Structured Incoherence and Shared Features. In: CVPR 2010, pp. 3501–3508 (June 2010)
Sun, Y., Kojima, A.: A novel method for semantic video concept learning using web images. In: ACM Multimedia 2011, New York, NY, USA, pp. 1081–1084 (2011)
Sun, Y., Shimada, S., Taniguchi, Y., Kojima, A.: A novel region-based approach to visual concept modeling using web images. In: ACM Multimedia 2008, New York, NY, USA, pp. 635–638 (2008)
Tang, S., Li, J.-T., Li, M., Xie, C., Liu, Y.-Z., Tao, K., Xu, S.-X.: TRECVID 2008 High-Level Feature Extraction By MCG-ICT-CAS. In: Proc. TRECVID 2008 Workshop, Gaithesburg, USA (November 2008)
Tang, S., Zheng, Y.-T., Wang, Y., Chua, T.-S.: Sparse ensemble learning for concept detection. IEEE Transactions on Multimedia 14(1) (2012)
Zhu, S., Wang, G., Ngo, C.-W., Jiang, Y.-G.: On the sampling of web images for learning visual concept classifiers. In: CIVR 2010, New York, NY, USA, pp. 50–57 (2010)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Sun, Y., Sudo, K., Taniguchi, Y. (2015). Cross-Domain Concept Detection with Dictionary Coherence by Leveraging Web Images. In: He, X., Luo, S., Tao, D., Xu, C., Yang, J., Hasan, M.A. (eds) MultiMedia Modeling. MMM 2015. Lecture Notes in Computer Science, vol 8936. Springer, Cham. https://doi.org/10.1007/978-3-319-14442-9_47
Download citation
DOI: https://doi.org/10.1007/978-3-319-14442-9_47
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-14441-2
Online ISBN: 978-3-319-14442-9
eBook Packages: Computer ScienceComputer Science (R0)