Skip to main content
Log in

Multi-cascaded attention and overlapping part features network for person re-identification

  • Original Paper
  • Published:
Signal, Image and Video Processing Aims and scope Submit manuscript

Abstract

Person re-identification (Person re-ID) is a challenging task in the field of computer vision. Generally, re-ID is regarded as an image retrieval problem; this task aims at searching a query person across multiple non-overlapping cameras. Re-ID is an important research direction after facial recognition that has many application scenarios, such as intelligent security, intelligent homing system, and unmanned supermarket. In practice, the identification would be deteriorated by the pedestrian postures, occlusions, and lighting conditions. Cropped images of pedestrians are often obtained from automatic detectors; this type of detection introduces two types of errors: part missing and excessive background. In previous works, fine-grained information has been proved to be useful for pedestrian retrieval. In this paper, we observe that multiple attention blocks can perform long-range multi-hop communication. To tackle the problem of excessive background, multistage cascaded attention blocks are introduced to put more focus on the information about human body. On the other hand, the part-level features have been proven to be effective against high quality of feature representation, but considering traditional vertical segmentation brings part inconsistency and lost some detailed semantic clues; the concept of overlapping features has been proposed to overcome this problem. Experiments on two large-scale re-ID datasets show that our method improves the learned representation of the feature embeddings and achieved competitive results.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

References

  1. Su, C., Zhang, S., Xing, J., Gao, W., Tian, Q.: Deep attributes driven multi-camera person re-identification. In: ECCV, pp. 475–491 (2016)

  2. Liu, H., Feng, J., Qi, M., Jiang, J., Yan, S.: End-to-end comparative attention networks for person re-identification. IEEE Trans. Image Process. 26(7), 3492–3506 (2017)

    Article  MathSciNet  Google Scholar 

  3. Wei, L., Zhang, S., Gao, W., Tian, Q.: Person transfer gan to bridge domain gap for person re-identification. In: CVPR, pp. 79–88 (2018)

  4. Cheng, D., Gong, Y., Zhou, S., Wang, J., Zheng, N.: Person re-identification by multi-channel parts-based cnn with improved triplet loss function. In: CVPR, pp. 1335–1344 (2016)

  5. Su, C., Li, J., Zhang, S., Xing, J., Gao, W., Tian, Q.: Pose-driven deep convolutional model for person re-identification. In: ICCV, pp. 3960–3969 (2017)

  6. Chen, B., Deng, W., Hu, J.: Mixed highorder attention network for person re-identification. arXiv:1908.05819 (2019)

  7. Chen, T., Ding, S., Xie, J., Yuan, Y., Chen, W., Yang, Y., Ren, Z., Wang, Z.: Abdnet: attentive but diverse person re-identification. In: ICCV, pp. 8351–8361 (2019)

  8. Mignon, J., Mignon, A., Jurie, F., Pcca.: a new approach for distance learning from sparse pairwise constraints. In: CVPR (2012)

  9. Zheng, L., Shen, L., Tian, L., Wang, S., Wang, J., Tian, Q.: Scalable person reidentification: a benchmark. In: ICCV (2015)

  10. Lin, Y., Zheng, L., Zheng, Z., Wu, Y., Yang, Y.: Improving person re-identification by attribute and identity learning. arXiv:1703.07220 (2017)

  11. Zhao, H., Tian, M., Sun, S., Shao, J., Yan, J., Yi, S., Wang, X., Tang, X.: Spindle net: person re-identification with human body region guided feature decomposition and fusion. In: CVPR, pp. 1077–1085 (2017)

  12. Sun, Y., Zheng, L., Yang, Y., Tian, Q., Wang, S.: Person retrieval with refined part pooling. In: ECCV, Beyond Part Models (2018)

  13. Wang, G., Yuan, Y., Chen, X., Li, J., Zhou, X.: Learning discriminative features with multiple granularities for person re-identification. In: ACM International Conference on Multimedia, pp. 274–282 (2018)

  14. Chen, Y., Zhu, X., Gong, S.: Instance-guided context rendering for cross-domain person re-identification. In: ICCV, pp. 232–242 (2019)

  15. Kaiwei, Z.: Hierarchical clustering with hard-batch triplet loss for person re-identification. In: CVPR (2020)

  16. Zhong, Z., Zheng, L., Luo, Z., Li, S., Yang, Y.: Exemplar memory for domain adaptive person re-identification. In: CVPR, Invariance matters (2019)

  17. Yu, H.X., et al.: Unsupervised person re-identification by soft multilabel learning. In: CVPR (2019)

  18. Wang, D., Zhang, S.: Unsupervised person re-identification via multi-label classification. In: CVPR, pp. 10978–10987 (2020)

  19. Zhai, Y., Lu, S.: Augmented discriminative clustering for domain adaptive person re-identification. In: CVPR, AD-Cluster (2020)

  20. Xu, M., et al.: Object tracking based on learning collaborative representation with adaptive weight. In: Signal Image and Video Processing (2020)

  21. Fan, S., et al.: Correction to: high-speed tracking based on multi-CF filters and attention mechanism. SIViP 14(3), 637–637 (2020)

    Article  Google Scholar 

  22. Havangi, R.: Intelligent adaptive unscented particle filter with application in target tracking. SIViP 14(2), 1–9 (2020)

    Google Scholar 

  23. Lan, X., Zhang, W., Zhang, S., Jain, D. K., Zhou, H.: Robust multi-modality anchor graph-based label prediction for rgb-infrared tracking. IEEE Trans. Ind. Inf. 99 (2019)

  24. Jaderberg, M., Simonyan, K., Zisserman, A., Kavukcuoglu, K.: Spatial transformer networks. In: NIPS, pp. 2017–2025 (2015)

  25. Woo, S., Park, J., Lee, J.-Y., Kweon, S.: Cbam: convolutional block attention module. In: ECCV, pp. 3–19 (2018)

  26. Wang, X., Girshick, R., Gupta, A., He, K.: Non-local neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7794–7803 (2018)

  27. Tay, C.P., Roy, S., Yap. K.H.: Attribute attention network for person re-identifications. In: CVPR, AANet (2019)

  28. Xia, B.N., Gong, Y., Zhang, Y., Poellabauer, C.: Second-order nonlocal attention networks for person re-identification. In: ICCV, pp. 3760–3769 (2019)

  29. Wang, Z., Luo, S., Sun, H., Pan, H., Yin, J.: An efficient non-local attention network for video-based person re-identification. In: ICIT (2019)

  30. Okay, A.E., AlGhamdi, M., Westendorp, R., Abdel-Mottaleb, M.: Multi-resolution overlapping stripes network for person re-identification. In: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (2020)

  31. Hermans, A., Beyer, L., Leibe, B.: In defense of the triplet loss for person re-identification. arXiv:1703.07737 (2017)

  32. Zhao, L., Li, X., Wang, J., Zhuang, Y.: Deeply-learned part-aligned representations for person reidentification. In: ICCV (2017)

  33. Zheng, Z., Zheng, L., Yang, Y.: Pedestrian alignment network for large-scale person re-identification. In: IEEE Transactions on Circuits and Systems for Video Technology (2018)

  34. Li, W., Zhu, X., Gong, S.: Person reidentification by deep joint learning of multi-loss classification. In: IJCAI (2017)

  35. Chen, Y., Zhu, X., Gong, S.: Person reidentification by deep learning multi-scale representations. In: ICCV (2017)

  36. Chang, X., Hospedales, T.M., Xiang, T.: Multi-level factorisation net for person re-identification. In: CVPR, pp. 2109–2118 (2018)

  37. Si, J., Zhang, H., Li, C.-G., Kuen, J., Kong, X., Kot, A.C., Wang, G.: Dual attention matching network for context-aware feature sequence based person re-identification. In: CVPR, pp. 5363–5372 (2018)

  38. Shen, Y., Li, H., Xiao, T., Yi, S., Chen, D., Wang, X.: Deep group-shuffling random walk for person re-identification. In: CVPR, pp. 2265–2274 (2018)

  39. Sun, Y., Xu, Q., Li, Y., Zhang, C., Li, Y., Wang, S., Sun, J.: Perceive where to focus: learning visibility-aware part-level features for partial person reidentification. In: CVPR, pp. 393–402 (2019)

  40. Huang, H., Li, D., Zhang, Z., Chen, X., Huang, K.: Adversarially occluded samples for person re-identification. In: CVPR (2018)

  41. Saquib Sarfraz, M., Schumann, A., Eberle, A., Stiefelhagen, R..: A pose-sensitive embedding for person re-identification with expanded cross neighborhood re-ranking. In: CVPR, pp. 420–429 (2018)

  42. Almazan, J., Gajic, B., Murray, N., Larlus, D.: Re-ID done right: towards good practices for person re-identification. arXiv:1801.05339 (2018)

  43. Chen, D., Xu, D., Li, H., Sebe, N., Wang, X.: Group consistent similarity learning via deep CRF for person re-identification. In: CVPR, pp. 8649–8658 (2018)

  44. Sun, Y., Zheng, L., Deng, W., Wang, S.: SVDNet for pedestrian retrieval. In: ICCV, pp. 2590–2600 (2017)

  45. Li, W., Zhu, X., Gong, S..: Harmonious attention network for person re-identification. In: CVPR (2018)

Download references

Acknowledgements

This work was supported by the National Natural Science Foundation of China (Grant No. U1833115).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xin Zhang.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, X., Hou, M., Deng, X. et al. Multi-cascaded attention and overlapping part features network for person re-identification. SIViP 16, 1525–1532 (2022). https://doi.org/10.1007/s11760-021-02106-x

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11760-021-02106-x

Keywords

Navigation