Advertisement

International Journal of Computer Vision

, Volume 119, Issue 2, pp 110–144 | Cite as

Robust Tracking via Locally Structured Representation

  • Yao SuiEmail author
  • Li Zhang
Article

Abstract

Representation method is critical to visual tracking. A robust representation describes the target accurately, leading to good tracking performance. In this work, a novel representation is proposed, which is designed to be simultaneously low-rank and joint sparse for the local patches within a target region. In this representation, the subspace structure is exploited by the low-rank constraint to reflect the global information of all the patches, and the sparsity structure is captured by the joint sparsity restriction to describe the locally intimate relationship between the neighboring patches. Importantly, to make the representation computationally applicable to visual tracking, a novel fast algorithm based on greedy strategy is proposed, and the performance analysis of this algorithm is also presented. Thus, the tracking in this work is formulated as a locally low-rank and joint sparse matching problem within particle filtering framework. A large number of experimental results show that the tracking drift problem is effectively alleviated in various challenging situations by using the proposed representation method. Both qualitative and quantitative evaluations demonstrate that the proposed tracker performs favorably against many other state-of-the-art trackers. Benefitting from the good adaptive capability of the representation, all the parameters of the proposed tracking algorithm are fixed in all the experiments.

Keywords

Visual tracking Low-rank approximation Sparse representation Greedy algorithm Appearance model 

Notes

Acknowledgments

This work was supported by the National Natural Science Foundation of China (NSFC) under Grant 61172125 and Grant 61132007.

References

  1. Adam, A., Rivlin, E., & Shimshoni, I. (2006). Robust fragments-based tracking using the integral histogram. IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), 1, 798–805.Google Scholar
  2. Arulampalam, M., Maskell, S., Gordon, N., & Clapp, T. (2002). A tutorial on particle filters for online nonlinear/non-Gaussian Bayesian tracking. IEEE Transactions on Signal Processing (TSP), 50(2), 174–188.CrossRefGoogle Scholar
  3. Babenko, B., Member, S., Yang, M. H., & Member, S. (2011). Robust object tracking with online multiple instance learning. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 33(8), 1619–1632.CrossRefGoogle Scholar
  4. Bach, F., Jenatton, R., Mairal, J., & Obozinski, G. (2011). Convex optimization with sparsity-inducing norms. Optimization for Machine Learning, 19–53.Google Scholar
  5. Beck, A., & Teboulle, M. (2009). A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM Journal on Imaging Sciences, 2(1), 183–202.MathSciNetCrossRefzbMATHGoogle Scholar
  6. Cai, J., Candès, E., & Shen, Z. (2010). A singular value thresholding algorithm for matrix completion. SIAM Journal on Optimization, 20(4), 1956–1982.MathSciNetCrossRefzbMATHGoogle Scholar
  7. Candes, E., & Plan, Y. (2010). Matrix completion with noise. Proceedings of the IEEE, 98(6), 925–936.CrossRefGoogle Scholar
  8. Candes, E., Li, X., Ma, Y., & Wright, J. (2009). Robust principal component analysis?. arXiv:0912.3599v1.
  9. Coates, A., & Ng, A. (2011). The importance of encoding versus training with sparse coding and vector quantization. In International Conference on Machine Learning (ICML).Google Scholar
  10. Dinh, T. B., Vo, N., & Medioni, G. (2011). Context tracker: Exploring supporters and distracters in unconstrained environments. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), IEEE (pp. 1177–1184).Google Scholar
  11. Grabner, H., & Bischof, H. (2006). On-line boosting and vision. IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), 1, 260–267.Google Scholar
  12. Grabner, H., Leistner, C., & Bischof, H. (2008). Semi-supervised on-line boosting for robust tracking. European Conference on Computer Vision (ECCV), 813399, 234–247.Google Scholar
  13. Hager, G. D., & Belhumeur, P. N. (1996). Real-time tracking of image regions with changes in geometry and illumination. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 403–410).Google Scholar
  14. Hare, S., Saffari, A., & Torr, P. (2011). Struck: Structured output tracking with kernels. In IEEE International Conference on Computer Vision (ICCV) (pp. 263–270).Google Scholar
  15. Henriques, F., Caseiro, R., Martins, P., & Batista, J. (2012). Exploiting the circulant structure of tracking-by-detection with kernels. In European Conference on Computer Vision (ECCV) (pp. 702–715).Google Scholar
  16. Henriques, J. F., Caseiro, R., Martins, P., & Batista, J. (2015). High-speed tracking with kernelized correlation filters. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI) (pp. 1–14).Google Scholar
  17. Hong, Z., Mei, X., Prokhorov, D., & Tao, D. (2013). Tracking via robust multi-task multi-view joint sparse representation. In IEEE International Conference on Computer Vision (ICCV), IEEE (pp. 649–656).Google Scholar
  18. Hue, C., Vermaak, J., & Gangnet, M. (2002). Color-based probabilistic tracking. In European Conference on Computer Vision (ECCV) (pp. 661–675).Google Scholar
  19. Isard, M. (1998). CONDENSATION—Conditional density propagation for visual tracking. International Journal of Computer Vision (IJCV), 29(1), 5–28.CrossRefGoogle Scholar
  20. Jia, X., Lu, H., & Yang, M. H. (2012). Visual tracking via adaptive structural local sparse appearance model. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 1822–1829).Google Scholar
  21. Kalal, Z., Matas, J., & Mikolajczyk, K. (2010). P-N learning: Bootstrapping binary classifiers by structural constraints. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), IEEE (pp. 49–56).Google Scholar
  22. Kalal, Z., Mikolajczyk, K., & Matas, J. (2012). Tracking-learning-detection. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 34(7), 1409–1422.CrossRefGoogle Scholar
  23. Kriegmant, D. J., Engineering, E., & Haven, N. (1996). What is the set of images of an object under all possible lighting conditions? In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 270–277).Google Scholar
  24. Kwon, J., & Lee, K. (2010). Visual tracking decomposition. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 1269–1276).Google Scholar
  25. Kwon, J., & Lee, K. M. (2011). Tracking by Sampling Trackers. In IEEE International Conference on Computer Vision (ICCV) (pp. 1195–1202).Google Scholar
  26. Li, X., Dick, A., Shen, C., van den Hengel, A., & Wang, H. (2013). Incremental learning of 3d-dct compact representations for robust visual tracking. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 35(4), 863–881.CrossRefGoogle Scholar
  27. Lin, Z., Chen, M., & Ma, Y. (2010). The augmented lagrange multiplier method for exact recovery of corrupted low-rank matrices. UIUC Technical Report arXiv:1009.5055v2.
  28. Liu, B., Yang, L., Huang, J., & Meer, P. (2010). Robust and fast collaborative tracking with two stage sparse optimization. In European Conference on Computer Vision (ECCV) (pp. 1–14).Google Scholar
  29. Liu, B., Huang, J., Yang, L., & Kulikowsk, C. (2011). Robust tracking using local sparse appearance model and K-selection. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), IEEE (pp. 1313–1320).Google Scholar
  30. Liu, B., Huang, J., Kulikowski, C., & Yang, L. (2013). Robust visual tracking using local sparse appearance model and k-selection. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 35(12), 2968–2981.CrossRefGoogle Scholar
  31. Mei, X., & Ling, H. (2009). Robust visual tracking using L1 minimization. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 1436–1443).Google Scholar
  32. Mei, X., & Ling, H. (2011). Robust visual tracking and vehicle classification via sparse representation. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 33(11), 2259–2272.CrossRefGoogle Scholar
  33. Mei, X., Ling, H., & Wu, Y. (2011). Minimum error bounded efficient l1 tracker with occlusion detection. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 1257–1264).Google Scholar
  34. Oron, S., Bar-Hillel, A., Levi, D., & Avidan, S. (2012). Locally orderless tracking. IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), 1, 1940–1947.Google Scholar
  35. Ross, D. A., Lim, J., Lin, R. S., & Yang, M. H. (2007). Incremental learning for robust visual tracking. International Journal of Computer Vision (IJCV), 77(1–3), 125–141.Google Scholar
  36. Sevilla-Lara, L., & Learned-Miller, E. (2012). Distribution fields for tracking. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), IEEE (pp. 1910–1917).Google Scholar
  37. Smeulders, A. W. M., Chu, D. M., Cucchiara, R., Calderara, S., Dehghan, A., & Shah, M. (2014). Visual tracking: An experimental survey. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 36(7), 1442–1468.CrossRefGoogle Scholar
  38. Sui, Y., & Zhang, L. (2015). Visual tracking via locally structured gaussian process regression. IEEE Signal Processing Letters (SPL), 22(9), 1331–1335.CrossRefGoogle Scholar
  39. Sui, Y., Tang, Y., & Zhang, L. (2015a). Discriminative low-rank tracking. In IEEE International Conference on Computer Vision (ICCV) (pp. 3002–3010).Google Scholar
  40. Sui, Y., Zhang, S., & Zhang, L. (2015b). Robust visual tracking via sparsity induced subspace learning. IEEE Transactions on Image Processing (TIP), 48(12), 4686–4700.MathSciNetCrossRefGoogle Scholar
  41. Sui, Y., Zhao, X., Zhang, S., Yu, X., Zhao, S., & Zhang, L. (2015c). Self-expressive tracking. Pattern Recognition (PR), 48(9), 2872–2884.CrossRefGoogle Scholar
  42. Wang, D., & Lu, H. (2012). Object tracking via 2DPCA and L1-regularization. IEEE Signal Processing Letters, 19(11), 711–714.CrossRefGoogle Scholar
  43. Wang, D., & Lu, H. (2014). Visual tracking via probability continuous outlier model. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR).Google Scholar
  44. Wang, D., Lu, H., & Yang, M. H. (2013a). Least soft-thresold squares tracking. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 2371–2378).Google Scholar
  45. Wang, D., Lu, H., & Yang, M. H. (2013b). Online object tracking with sparse prototypes. IEEE Transactions on Image Processing (TIP), 22(1), 314–325.MathSciNetCrossRefGoogle Scholar
  46. Wright, J., Ma, Y., Mairal, J., & Sapiro, G. (2010). Sparse representation for computer vision and pattern recognition. Proceedings of The IEEE, 98(6), 1031–1044.CrossRefGoogle Scholar
  47. Wu, Y., Lim, J., & Yang, M. H. (2013). Online object tracking: A benchmark. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 2411–2418).Google Scholar
  48. Xiangyuan, L., Ma, A. J., Yuen, P. C. (2014). Multi-cue visual tracking using robust feature-level fusion based on joint sparse representation. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR).Google Scholar
  49. Yilmaz, A., Javed, O., & Shah, M. (2006). Object tracking. ACM Computing Surveys, 38(4), 13–57.CrossRefGoogle Scholar
  50. Yuan, X., & Yan, S. (2010). Visual classification with multi-task joint sparse representation. EEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), 21, 4349–4360.MathSciNetGoogle Scholar
  51. Zhang, K., Zhang, L., Yang, M. H. (2012a) Real-time compressive tracking. In European Conference on Computer Vision (ECCV) (pp. 866–879).Google Scholar
  52. Zhang, T., Ghanem, B., & Liu, S. (2012b) Robust visual tracking via multi-task sparse learning. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 2042–2049).Google Scholar
  53. Zhang, T., Ghanem, B., Liu, S., & Ahuja, N. (2012c) Low-rank sparse learning for robust visual tracking. In European Conference on Computer Vision (ECCV) (pp. 470–484).Google Scholar
  54. Zhong, W., Lu, H., & Yang, M. H. (2012). Robust object tracking via sparsity-based collaborative model. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 1838–1845).Google Scholar

Copyright information

© Springer Science+Business Media New York 2016

Authors and Affiliations

  1. 1.Department of Electronic EngineeringTsinghua UniversityBeijingChina

Personalised recommendations