Abstract
Fashion retrieval is a challenging task of finding an exact match for fashion items contained within an image. Difficulties arise from the fine-grained nature of clothing items, very large intra-class and inter-class variance. Additionally, query and source images for the task usually come from different domains - street and catalogue photos, respectively. Due to these differences, a significant gap in quality, lighting, contrast, background clutter and item presentation exists. As a result, fashion retrieval is an active field of research both in academia and the industry. Inspired by recent advancements in person re-identification research, we adapt leading ReID models to fashion retrieval tasks. We introduce a simple baseline model for fashion retrieval, significantly outperforming previous state-of-the-art results, despite a much simpler architecture. We conduct in-depth experiments on Street2Shop and DeepFashion datasets. Finally, we propose a cross-domain (cross-dataset) evaluation method to test the robustness of fashion retrieval models.
This work was supported by the EU co-funded Smart Growth Operational Programme 2014–2020 (project no. POIR.01.01.01-00-0695/19).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Chen, W., Chen, X., Zhang, J., Huang, K.: Beyond triplet loss: a deep quadruplet network for person re-identification. In: CVPR. IEEE (2017)
Dodds, E., Nguyen, H., Herdade, S., Culpepper, J., Kae, A., Garrigues, P.: Learning embeddings for product visual search with triplet loss and online sampling. arXiv preprint arXiv:1810.04652 (2018)
Kuang, Z., et al.: Fashion retrieval via graph reasoning networks on a similarity pyramid. In: ICCV. IEEE (2019)
Kucer, M., Murray, N.: A detect-then-retrieve model for multi-domain fashion item retrieval. In: CVPR Workshops. IEEE (2019)
Luo, H., et al.: A strong baseline and batch normalization neck for deep person re-identification. IEEE Trans. Multimed. 22, 2597–2609 (2019)
Schroff, F., Kalenichenko, D., Philbin, J.: FaceNet: a unified embedding for face recognition and clustering. In: CVPR. IEEE (2015)
Wang, G., Lai, J., Huang, P., Xie, X.: Spatial-temporal person re-identification. In: AAAI. AAAI Press (2019)
Wen, Y., Zhang, K., Li, Z., Qiao, Y.: A discriminative feature learning approach for deep face recognition. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9911, pp. 499–515. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46478-7_31
Xingang, P., Ping Luo, J.S., Tang, X.: Two at once: enhancing learning and generalization capacities via IBN-Net. In: ECCV (2018)
Zhou, K., Yang, Y., Cavallaro, A., Xiang, T.: Learning generalisable omni-scale representations for person re-identification. arXiv preprint arXiv:1910.06827 (2019)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Wieczorek, M., Michalowski, A., Wroblewska, A., Dabrowski, J. (2020). A Strong Baseline for Fashion Retrieval with Person Re-identification Models. In: Yang, H., Pasupa, K., Leung, A.CS., Kwok, J.T., Chan, J.H., King, I. (eds) Neural Information Processing. ICONIP 2020. Communications in Computer and Information Science, vol 1332. Springer, Cham. https://doi.org/10.1007/978-3-030-63820-7_33
Download citation
DOI: https://doi.org/10.1007/978-3-030-63820-7_33
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-63819-1
Online ISBN: 978-3-030-63820-7
eBook Packages: Computer ScienceComputer Science (R0)