Abstract
Fashion retrieval is a challenging task of finding an exact match for fashion items contained within an image. Difficulties arise from the fine-grained nature of clothing items, very large intra-class and inter-class variance. Additionally, query and source images for the task usually come from different domains - street and catalogue photos, respectively. Due to these differences, a significant gap in quality, lighting, contrast, background clutter and item presentation exists. As a result, fashion retrieval is an active field of research both in academia and the industry. Inspired by recent advancements in person re-identification research, we adapt leading ReID models to fashion retrieval tasks. We introduce a simple baseline model for fashion retrieval, significantly outperforming previous state-of-the-art results, despite a much simpler architecture. We conduct in-depth experiments on Street2Shop and DeepFashion datasets. Finally, we propose a cross-domain (cross-dataset) evaluation method to test the robustness of fashion retrieval models.
This work was supported by the EU co-funded Smart Growth Operational Programme 2014–2020 (project no. POIR.01.01.01-00-0695/19).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Chen, W., Chen, X., Zhang, J., Huang, K.: Beyond triplet loss: a deep quadruplet network for person re-identification. In: CVPR. IEEE (2017)
Dodds, E., Nguyen, H., Herdade, S., Culpepper, J., Kae, A., Garrigues, P.: Learning embeddings for product visual search with triplet loss and online sampling. arXiv preprint arXiv:1810.04652 (2018)
Kuang, Z., et al.: Fashion retrieval via graph reasoning networks on a similarity pyramid. In: ICCV. IEEE (2019)
Kucer, M., Murray, N.: A detect-then-retrieve model for multi-domain fashion item retrieval. In: CVPR Workshops. IEEE (2019)
Luo, H., et al.: A strong baseline and batch normalization neck for deep person re-identification. IEEE Trans. Multimed. 22, 2597–2609 (2019)
Schroff, F., Kalenichenko, D., Philbin, J.: FaceNet: a unified embedding for face recognition and clustering. In: CVPR. IEEE (2015)
Wang, G., Lai, J., Huang, P., Xie, X.: Spatial-temporal person re-identification. In: AAAI. AAAI Press (2019)
Wen, Y., Zhang, K., Li, Z., Qiao, Y.: A discriminative feature learning approach for deep face recognition. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9911, pp. 499–515. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46478-7_31
Xingang, P., Ping Luo, J.S., Tang, X.: Two at once: enhancing learning and generalization capacities via IBN-Net. In: ECCV (2018)
Zhou, K., Yang, Y., Cavallaro, A., Xiang, T.: Learning generalisable omni-scale representations for person re-identification. arXiv preprint arXiv:1910.06827 (2019)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Wieczorek, M., Michalowski, A., Wroblewska, A., Dabrowski, J. (2020). A Strong Baseline for Fashion Retrieval with Person Re-identification Models. In: Yang, H., Pasupa, K., Leung, A.CS., Kwok, J.T., Chan, J.H., King, I. (eds) Neural Information Processing. ICONIP 2020. Communications in Computer and Information Science, vol 1332. Springer, Cham. https://doi.org/10.1007/978-3-030-63820-7_33
Download citation
DOI: https://doi.org/10.1007/978-3-030-63820-7_33
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-63819-1
Online ISBN: 978-3-030-63820-7
eBook Packages: Computer ScienceComputer Science (R0)