Advertisement

Visual Cognition–Inspired Multi-View Vehicle Re-Identification via Laplacian-Regularized Correlative Sparse Ranking

  • Aihua Zheng
  • Jiacheng Dong
  • Xianmin Lin
  • Lidan Liu
  • Bo JiangEmail author
  • Bin Luo
Article
  • 21 Downloads

Abstract

Vehicle re-identification has gradually gained attention and widespread applications. However, most of the existing methods learn the discriminative features for identities by single-feature channel only. It is worth noting that visual cognition of the human eyes is a multi-channel system which usually seeks a sparse representation. Therefore, integrating the multi-view information in sparse representation is a natural way to boost computer vision tasks in challenging scenarios. In this paper, we propose to mine multi-view deep features via Laplacian-regularized correlative sparse ranking for vehicle re-identification. Specifically, first, we employ multiple baseline networks to generate features. Then, we explore the feature correlation via enforcing the correlation term into the multi-view Laplacian sparse ranking framework. The original rankings are obtained by the reconstruction coefficients between the probe and gallery. Finally, we utilize a re-ranking technique to further boost performance. Experimental results on public benchmark VeRi-776 and VehicleID datasets demonstrate that our approach outperforms state-of-the-art approaches. The Laplacian-regularized correlative sparse ranking as a general framework can be used in any multi-view feature fusion and will obtain more competitive results.

Keywords

Vehicle re-identification Laplacian-regularized correlative sparse ranking Multi-view Deep feature 

Notes

Funding Information

This research is supported in part by the National Natural Science Foundation of China (61976002, 61602001, 61671018 and 61860206004), the Open Project Program of the National Laboratory of Pattern Recognition (NLPR) (201900046), and the Natural Science Foundation of Anhui Higher Education Institutions of China (KJ2019A0033); Open fund for Discipline Construction, Institute of Physical Science and Information Technology, Anhui University.

Compliance with Ethical Standards

Conflict of interest

The authors declare that they have no conflict of interest.

Ethical approval

This article does not contain any studies with human participants or animals performed by any of the authors.

References

  1. 1.
    Yan Y, Ren J, Zhao H, Sun G, Wang Z, Zheng J, Marshall S, Soraghan J. Cognitive fusion of thermal and visible imagery for effective detection and tracking of pedestrians in videos. Cogn Comput. 2017:1–11.Google Scholar
  2. 2.
    Zhao C, Li X, Ren J, Marshall S. Improved sparse representation using adaptive spatial support for effective target detection in hyperspectral imagery. Int J Remote Sens. 2013:8669–8684.CrossRefGoogle Scholar
  3. 3.
    Gegenfurtner KR. Cortical mechanisms of colour vision. Nat Rev Neurosc. 2003:563.CrossRefGoogle Scholar
  4. 4.
    Ravello CR, Perrinet LU, Escobar MJ, Palacios AG. Speed-selectivity in retinal ganglion cells is sharpened by broad spatial frequency, naturalistic stimuli. Scientific reports. 2019:456.Google Scholar
  5. 5.
    Yang ZX, Tang L, Zhang K, Wong PK. Multi-view cnn feature aggregation with elm auto-encoder for 3d shape recognition. Cogn Comput. 2018:1–14.Google Scholar
  6. 6.
    Xing J, Niu Z, Huang J, Hu W, Yan S. Towards robust and accurate multi-view and partially-occluded face alignment. IEEE Trans Pattern Anal Mach Intell. 2018:1–1.Google Scholar
  7. 7.
    Niu L, Li W, Xu D, Cai J. An exemplar-based multi-view domain generalization framework for visual recognition. IEEE Trans Neural Netw Learn Sys. 2018:259–272.Google Scholar
  8. 8.
    Tulsiani S, Efros AA, Malik J. Multi-view consistency as supervisory signal for learning shape and pose prediction. IEEE conference on computer vision and pattern recognition; 2018. p. 2897–2905.Google Scholar
  9. 9.
    You X, Xu J, Yuan W, Jing XY, Tao D, Zhang T. Multi-view common component discriminant analysis for cross-view classification. Pattern Recognit. 2019:1.Google Scholar
  10. 10.
    Zhang H, Patel VM. Convolutional sparse and low-rank coding-based image decomposition. IEEE Trans Image Process. 2018:1–1.Google Scholar
  11. 11.
    De K, Masilamani V. A no-reference image quality measure for blurred and compressed images using sparsity features. Cogn Comput. 2018:1–11.Google Scholar
  12. 12.
    Qi Y, Qin L, Zhang J, Zhang S, Huang Q, Yang MH. Structure-aware local sparse coding for visual tracking. IEEE Trans Image Process. 2018:1–1.Google Scholar
  13. 13.
    Zhang T, Xu C, Yang MH. Robust structural sparse tracking. IEEE Trans Pattern Anal Mach Intell. 2019:473–486.CrossRefGoogle Scholar
  14. 14.
    Zeng N, Zhang H, Song B, Liu W, Li Y, Dobaie AM. Facial expression recognition via learning deep sparse autoencoders. Neurocomputing. 2018:643–649.CrossRefGoogle Scholar
  15. 15.
    He R, Zheng WS, Hu BG, Kong XW. Two-stage nonnegative sparse representation for large-scale face recognition. IEEE Trans Neural Netw Learn Sys. 2013:35–46.Google Scholar
  16. 16.
    He R, Zheng WS, Tan T, Sun Z. Half-quadratic-based iterative minimization for robust sparse representation. IEEE Trans Pattern Anal Mach Intell. 2014:261–275.Google Scholar
  17. 17.
    Yao Y, Guo P, Xin X, Jiang Z. Image fusion by hierarchical joint sparse representation. Cogn Comput. 2014:281–292.CrossRefGoogle Scholar
  18. 18.
    Sarfraz MS, Schumann A, Eberle A, Stiefelhagen R. 2017. A pose-sensitive embedding for person re-identification with expanded cross neighborhood re-ranking. arXiv:1711.10378.
  19. 19.
    Sun D, Liu L, Zheng A, Jiang B, Luo B. Visual cognition inspired vehicle re-identification via correlative sparse ranking with multi-view deep features. International conference on brain inspired cognitive systems; 2018. p. 54–63.Google Scholar
  20. 20.
    Chen YC, Zhu X, Zheng WS, Lai JH. Person re-identification by camera correlation aware feature augmentation. IEEE Trans Pattern Anal Mach Intell. 2018:392–408.CrossRefGoogle Scholar
  21. 21.
    Li X, Wu A, Zheng WS. 2018. Adversarial open-world person re-identification. arXiv:1807.10482.
  22. 22.
    Zheng L, Yang Y, Hauptmann AG. 2016. Person re-identification: Past, present and future. arXiv:1610.02984.
  23. 23.
    Liu X, Liu W, Ma H, Fu H. Large-scale vehicle re-identification in urban surveillance videos. IEEE International Conference on Multimedia and Expo; 2016. p. 1–6.Google Scholar
  24. 24.
    Liu H, Tian Y, Yang Y, Pang L, Huang T. Deep relative distance learning: Tell the difference between similar vehicles. IEEE conference on computer vision and pattern recognition; 2016 . p. 2167–2175.Google Scholar
  25. 25.
    Yang L, Luo P, Chen CL, Tang X. A large-scale car dataset for fine-grained categorization and verification. IEEE conference on computer vision and pattern recognition; 2015. p. 3973–3981.Google Scholar
  26. 26.
    Guo H, Zhao C, Liu Z, Wang J, Lu H. Learning coarse-to-fine structured feature embedding for vehicle re-identification. Association for the advancement of artificial intelligence; 2018 . p. 1–8.Google Scholar
  27. 27.
    Zapletal D, Herout A. Vehicle re-identification for automatic video traffic surveillance. IEEE conference on computer vision and pattern recognition workshops; 2016. p. 25–31.Google Scholar
  28. 28.
    Zhang Y, Liu D, Zha ZJ. Improving triplet-wise training of convolutional neural network for vehicle re-identification. IEEE international conference on multimedia and expo; 2017 . p. 1386–1391.Google Scholar
  29. 29.
    Kanacı A, Zhu X, Gong S. Vehicle reidentification by fine-grained cross-level deep learning. British machine vision conference; 2017. p. 1–6.Google Scholar
  30. 30.
    Zhu J, Du Y, Hu Y, Zheng L, Cai C. Vrsdnet: vehicle re-identification with a shortly and densely connected convolutional neural network. Multimedia Tools and Applications. 2018:1–15.Google Scholar
  31. 31.
    Shen Y, Xiao T, Li H, Yi S, Wang X. Learning deep neural networks for vehicle re-id with visual-spatio-temporal path proposals. IEEE international conference on computer vision; 2017. p. 1918–1927.Google Scholar
  32. 32.
    Wang Z, Tang L, Liu X, Yao Z, Yi S, Shao J, Yan J, Wang S, Li H, Wang X. Orientation invariant feature embedding and spatial temporal regularization for vehicle re-identification. IEEE conference on computer vision and pattern recognition; 2017. p. 379–387.Google Scholar
  33. 33.
    Liu X, Liu W, Mei T, Ma H. 2016. A deep learning-based approach to progressive vehicle re-identification for urban surveillance.Google Scholar
  34. 34.
    Li Y, Li Y, Yan H, Liu J. Deep joint discriminative learning for vehicle re-identification and retrieval. IEEE international conference on image processing; 2017. p. 395–399.Google Scholar
  35. 35.
    Zhou Y, Shao L. Viewpoint-aware attentive multi-view inference for vehicle re-identification. Proceedings of the IEEE conference on computer vision and pattern recognition; 2018 . p. 6489–6498.Google Scholar
  36. 36.
    Jia Y, Salzmann M, Darrell T. Factorized latent spaces with structured sparsity. Advances in Neural Information Processing Systems. 2010:982–990.Google Scholar
  37. 37.
    Liu W, Tao D, Cheng J, Tang Y. Multiview hessian discriminative sparse coding for image annotation. IEEE conference on computer vision and pattern recognition. 2014:50–60.CrossRefGoogle Scholar
  38. 38.
    Han Y, Wu F, Tao D, Shao J, Zhuang Y, Jiang J. Sparse unsupervised dimensionality reduction for multiple view data. IEEE Trans. Circuits Syst. Video Techno. 2012:1485–1496.Google Scholar
  39. 39.
    Yu J, Rui Y, Tao D. Click prediction for web image reranking using multimodal sparse coding. IEEE Trans Image Process. 2014:2019–2032.Google Scholar
  40. 40.
    Wu F, Zhou Y, Yang Y, Tang S, Zhang Y, Zhuang Y. Sparse multi-modal hashing. IEEE Transactions on Multimedia. 2014:427–439.CrossRefGoogle Scholar
  41. 41.
    Lan X, Ma AJ, Yuen PC. Multi-cue visual tracking using robust feature-level fusion based on joint sparse representation. IEEE conference on computer vision and pattern recognition; 2014 . p. 1194–1201.Google Scholar
  42. 42.
    Chen X, Ma H, Wan J, Li B, Xia T. Multi-view 3d object detection network for autonomous driving. IEEE conference on computer vision and pattern recognition; 2017. p. 6526– 6534.Google Scholar
  43. 43.
    Rubino C, Crocco M, Bue AD. 3d object localisation from multi-view image detections. IEEE Trans Pattern Anal Mach Intell. 2017:1–1.Google Scholar
  44. 44.
    Lin Y, Zheng L, Zheng Z, Wu Y, Yang Y. 2017. Improving person re-identification by attribute and identity learning. arXiv:1703.07220.
  45. 45.
    Zheng M, Bu J, Chen C, Wang C, Zhang L, Qiu G, Cai D. Graph regularized sparse coding for image representation. IEEE Transactions on Image Processing A Publication of the IEEE Signal Processing Society. 2011:1327.Google Scholar
  46. 46.
    Jiang B, Ding C, Tang J, Luo B. Image representation and learning with graph-laplacian tucker tensor decomposition. IEEE Trans Cybern. 2018:1–10.Google Scholar
  47. 47.
    Jiang B, Ding C, Luo B, Tang J. Graph-laplacian pca: Closed-form solution and robustness. IEEE conference on computer vision and pattern recognition; 2013. p. 3492–3498.Google Scholar
  48. 48.
    Parikh N, Boyd S, et al. Proximal algorithms. Foundations and Trends in Optimization 2014:127–239.CrossRefGoogle Scholar
  49. 49.
    Liao S, Hu Y, Zhu X, Li SZ. Person re-identification by local maximal occurrence representation and metric learning. IEEE conference on computer vision and pattern recognition; 2015 . p. 2197–2206.Google Scholar
  50. 50.
    Zheng L, Shen L, Tian L, Wang S, Wang J, Tian Q. Scalable person re-identification: a benchmark. IEEE international conference on computer vision; 2015. p. 1116–1124.Google Scholar
  51. 51.
    Yang L, Luo P, Change Loy C, Tang X. A large-scale car dataset for fine-grained categorization and verification. IEEE conference on computer vision and pattern recognition; 2015 . p. 3973–3981.Google Scholar
  52. 52.
    Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems; 2012. p. 1097– 1105.Google Scholar
  53. 53.
    Liu X, Liu W, Mei T, Ma H. Provid: Progressive and multimodal vehicle reidentification for large-scale urban surveillance. IEEE Transactions on Multimedia. 2018:645–658.CrossRefGoogle Scholar
  54. 54.
    Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H. 2017. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv:1704.04861.
  55. 55.
    Szegedy C, Ioffe S, Vanhoucke V, Alemi AA. Inception-v4, inception-resnet and the impact of residual connections on learning. Association for the advancement of artificial intelligence; 2017. p. 1.Google Scholar
  56. 56.
    He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. IEEE conference on computer vision and pattern recognition; 2016. p. 770–778.Google Scholar

Copyright information

© Springer Science+Business Media, LLC, part of Springer Nature 2019

Authors and Affiliations

  • Aihua Zheng
    • 1
  • Jiacheng Dong
    • 1
  • Xianmin Lin
    • 1
  • Lidan Liu
    • 1
  • Bo Jiang
    • 1
    • 2
    Email author
  • Bin Luo
    • 1
  1. 1.Key Laboratory of Intelligent Computing and Signal Processing of Ministry of Education, School of Computer Science and TechnologyAnhui UniversityHefeiChina
  2. 2.Institute of Physical Science and Information TechnologyAnhui UniversityHefeiChina

Personalised recommendations