Abstract
Cloth-changing person re-identification (CC-ReID) aims to retrieve specific pedestrians in a cloth-changing scenario. Its main challenge is to disentangle the clothing-related and clothing-unrelated features. Most existing approaches force the model to learn clothing-unrelated features by changing the color of the clothes. However, due to the lack of ground truth, these methods inevitably introduce noise, which destroys the discriminative features and leads to an uncontrollable disentanglement process. In this paper, we propose a new person re-identification network called features reconstruction disentanglement ReID (FRD-ReID), which can controllably decouple the clothing-unrelated and clothing-related features. Specifically, we first introduce the human parsing mask as the ground truth of the reconstruction process. At the same time, we propose the far away attention (FAA) mechanism and the person contour attention (PCA) mechanism for clothing-unrelated features and pedestrian contour features to improve the feature reconstruction efficiency. In the testing phase, we directly discard the clothing-related features for inference, which leads to a controllable disentanglement process. We conducted extensive experiments on the PRCC, LTCC, and Vc-Clothes datasets and demonstrated that our method outperforms existing state-of-the-art methods.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Chen, Z., Ge, Y.: Part-attention based model make occluded person re-identification stronger. arXiv preprint arXiv:2404.03443 (2024)
Chen, Z., Ge, Y., Zhang, J., et al.: Multi-branch person re-identification net. In: 2023 International Conference on Image Processing, Computer Vision and Machine Learning (ICICML), pp. 1104–1107. IEEE (2023)
Ge, Y., Zhang, J., Chen, Z., et al.: End-to-end person search based on content awareness. In: 2023 International Conference on Image Processing, Computer Vision and Machine Learning (ICICML), pp. 1108–1111. IEEE (2023)
Chen, Z., Ge, Y.: Occluded cloth-changing person re-identification. arXiv preprint arXiv:2403.08557 (2024)
Selvaraju, R.R., et al.: Grad-CAM: visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 618–626 (2017)
Chen, J., et al.: Deep shape-aware person re-identification for overcoming moderate clothing changes. TMM 24, 4285–4300 (2021)
Tu, M., et al.: Multi-granularity mutual learning network for object re-identification. TIST 23(9), 15178–15189 (2022)
Liu, H., et al.: End-to-end comparative attention networks for person re-identification. IEEE Trans. Image Process. 26(7), 3492–3506 (2017)
Ge, Y., Niu, K., Chen, Z., et al.: Lightweight traffic sign recognition model based on dynamic feature extraction. In: Huang, D.S., Premaratne, P., Yuan, C. (eds.) Applied Intelligence. ICAI 2023. Communications in Computer and Information Science, vol. 2014, pp. 339–350. Springer, Singapore (2023). https://doi.org/10.1007/978-981-97-0903-8_32
Zhang, J., et al.: An efficient convolutional multi-scale vision transformer for image classification. In: 2023 International Conference on Image Processing, Computer Vision and Machine Learning (ICICML), pp. 344–347 IEEE (2023)
Xu, W., et al.: Adversarial feature disentanglement for long-term person re-identification. In: IJCAI (2021)
Li, Y.-J., et al.: Learning shape representations for clothing variations in person re-identification. arXiv preprint arXiv:2003.07340 (2020)
Ge, Y.H., et al.: Semantically enhanced attention map-driven occluded person re-identification. Electron. Lett. 60, e13217 (2024). https://doi.org/10.1049/ell2.13217
Qian, X., et al.: Long-term cloth-changing person re-identification. In: Proceedings of the Asian Conference on Computer Vision (2020)
Yang, Q., et al.: Person re-identification by contour sketch under moderate clothing change. IEEE Trans. Pattern Anal. Mach. Intell. 43(6), 2029–2046 (2019)
Wan, F., et al.: When person re-identification meets changing clothes. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (2020)
He, K., et al.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016)
Li, W., et al.: Harmonious attention network for person re-identification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2018)
Sun, Y., Zheng, L., Yang, Y., Tian, Q., Wang, S.: Beyond part models: person retrieval with refined part pooling (and a strong convolutional baseline). In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11208, pp. 501–518. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01225-0_30
Zhou, K., et al.: Omni-scale feature learning for person re-identification. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (2019)
Hou, R., et al.: Interaction-and-aggregation network for person re-identification. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2019)
Jin, X., et al.: Cloth-changing person re-identification from a single image with gait prediction and regularization. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2022)
Yang, Z., et al.: Good is bad: causality inspired cloth-debiasing for cloth-changing person re-identification. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2023)
He, W., et al.: Retrieve anyone: a general-purpose person re-identification task with instructions. arXiv preprint arXiv:2306.07520 (2023)
Yuan, L., et al.: Tokens-to-Token ViT: training vision transformers from scratch on ImageNet. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (2021)
Jaderberg, M., Simonyan, K., Zisserman, A.: Spatial transformer networks. In: Advances in Neural Information Processing Systems, vol. 28 (2015)
Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: CBAM: convolutional block attention module. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 3–19. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01234-2_1
Zhang, Y., Li, K., Li, K., Wang, L., Zhong, B., Fu, Y.: Image super-resolution using very deep residual channel attention networks. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 294–310. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01234-2_18
Acknowledgments
Supported by Promoting the Classification and Development of Colleges and Universities-Student Innovation and Entrepreneurship Training Programme Project-School of Computer (5112410852).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Chen, Z., Ge, Y., Yue, Q. (2024). Features Reconstruction Disentanglement Cloth-Changing Person Re-identification. In: Huang, DS., Zhang, X., Guo, J. (eds) Advanced Intelligent Computing Technology and Applications. ICIC 2024. Lecture Notes in Computer Science, vol 14866. Springer, Singapore. https://doi.org/10.1007/978-981-97-5594-3_33
Download citation
DOI: https://doi.org/10.1007/978-981-97-5594-3_33
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-97-5593-6
Online ISBN: 978-981-97-5594-3
eBook Packages: Computer ScienceComputer Science (R0)