Skip to main content
Log in

Multi-granularity feature utilization network for cross-modality visible-infrared person re-identification

  • Focus
  • Published:
Soft Computing Aims and scope Submit manuscript

Abstract

Cross-modality visible-infrared person re-identification (VI-ReID) aims to recognize images with the same identity between visible modality and infrared modality, which is a very challenging task because it not only includes the troubles of variations between cross-cameras in traditional person ReID, but also suffers from the huge differences between two modalities. Some existing VI-ReID methods extract the modality-shared information from global features through single-stream or double-stream networks, while ignoring the complementarity between fine-grained and coarse-grained information. To solve this problem, our paper designs a multi-granularity feature utilization network (MFUN) to make up for the lack of shared features between different modalities by promoting the complementarity between coarse-grained features and fine-grained features. Firstly, in order to learn fine-grained shared features better, we design a local feature constraint module, which uses both hard-mining triplet loss and heterogeneous center loss to constrain local features in the common subspace, so as to better promote intra-class compactness and inter-class differences at the level of sample and class center. Then, our method uses a multi-modality feature aggregation module for global features to fuse the information of two modalities to narrow the modality gap. Through the combination of these two modules, visible and infrared image features can be better fused, thus alleviating the problem of modality discrepancy and supplementing the lack of modality-shared information. Extensive experimental results on RegDB and SYSU-MM01 datasets fully prove that our proposed MFUN outperforms the state-of-the-art solutions. Our code is available at https://github.com/ZhangYinyinzzz/MFUN.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8

Similar content being viewed by others

Data availability

Enquiries about data availability should be directed to the authors.

References

Download references

Acknowledgements

This research is supported in part by the National Natural Science Foundation of China under Grant 62172231 and U20B2065 and by the Natural Science Foundation of Jiangsu Province of China under Grant BK20220107 and BK20211539; this research is also supported in part by the Engineering Research Center of Digital Forensics, Ministry of Education.

Funding

Funding is provided by National Natural Science Foundation of China (Grant Nos. 62172231, U20B2065) and Natural Science Foundation of Jiangsu Province (Grant Nos. BK20220107, BK20211539).

Author information

Authors and Affiliations

Authors

Contributions

All authors contributed to the conceptualization and methodology. The writing—original draft, writing—review editing and visualization were performed by Guoqing Zhang and Yinyin Zhang. Investigation and data curation were performed by Yuhao Chen and Hongwei Zhang. Supervision was performed by Yuhui Zheng, and all authors commented on previous versions of the manuscript. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Guoqing Zhang.

Ethics declarations

Conflict of interest

All the authors declare no conflict of interest.

Human and animal rights

This study does not contain any studies with human participants or animals performed by any of the authors.

Ethical approval

Any of the authors’ investigations with human participants or animals are not included in this article.

Informed consent

Informed consent was obtained from all individual participants included in the study.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, G., Zhang, Y., Chen, Y. et al. Multi-granularity feature utilization network for cross-modality visible-infrared person re-identification. Soft Comput (2023). https://doi.org/10.1007/s00500-023-08321-7

Download citation

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s00500-023-08321-7

Keywords

Navigation