Discover the SciOpen Platform and Achieve Your Research Goals with Ease.
Search articles, authors, keywords, DOl and etc.
RGB-Infrared person re-IDentification (re-ID) aims to match RGB and infrared (IR) images of the same person. However, the modality discrepancy between RGB and IR images poses a significant challenge for re-ID. To address this issue, this paper proposes a Proxy-based Embedding Alignment (PEA) method to align the RGB and IR modalities in the embedding space. PEA introduces modality-specific identity proxies and leverages the sample-to-proxy relations to learn the model. Specifically, PEA focuses on three types of alignments: intra-modality alignment, inter-modality alignment, and cycle alignment. Intra-modality alignment aims to align sample features and proxies of the same identity within a modality. Inter-modality alignment aims to align sample features and proxies of the same identity across different modalities. Cycle alignment requires that a proxy is aligned with itself after tracing it along a cross-modality cycle (e.g., IR→RGB→IR). By integrating these alignments into the training process, PEA effectively mitigates the impact of modality discrepancy and learns discriminative features across modalities. We conduct extensive experiments on several RGB-IR re-ID datasets, and the results show that PEA outperforms current state-of-the-art methods. Notably, on SYSU-MM01 dataset, PEA achieves 71.0% mAP under the multi-shot setting of the indoor-search protocol, surpassing the best-performing method by 7.2%.
J. Lu, H. Wan, P. Li, X. Zhao, N. Ma, and Y. Gao, Exploring high-order spatio-temporal correlations from skeleton for person re-identification, IEEE Trans. Image Process., vol. 32, pp. 949–963, 2023.
M. Ye, X. Lan, Z. Wang, and P. C. Yuen, Bi-directional center-constrained top-ranking for visible thermal person re-identification, IEEE Trans. Inf. Forensics Secur., vol. 15, pp. 407–419, 2020.
N. Huang, J. Liu, Y. Luo, Q. Zhang, and J. Han, Exploring modality-shared appearance features and modality-invariant relation features for cross-modality person Re-IDentification, Pattern Recognit., vol. 135, p. 109145, 2023.
D. T. Nguyen, H. G. Hong, K. W. Kim, and K. R. Park, Person recognition system based on a combination of body images from visible light and thermal cameras, Sensors, vol. 17, no. 3, p. 605, 2017.
H. M. Hu, W. Fang, G. Zeng, Z. Hu, and B. Li, A person re-identification algorithm based on pyramid color topology feature, Multimed. Tools Appl., vol. 76, no. 24, pp. 26633–26646, 2017.
M. Pang, Y. M. Cheung, Q. Shi, and M. Li, Iterative dynamic generic learning for face recognition from a contaminated single-sample per person, IEEE Trans. Neural Netw. Learn. Syst., vol. 32, no. 4, pp. 1560–1574, 2021.
J. Huo, Y. Gao, Y. Shi, W. Yang, and H. Yin, Heterogeneous face recognition by margin-based cross-modality metric learning, IEEE Trans. Cybern., vol. 48, no. 6, pp. 1814–1826, 2018.
L. Lin, G. Wang, W. Zuo, X. Feng, and L. Zhang, Cross-domain visual matching via generalized similarity measure and feature learning, IEEE Trans. Pattern Anal. Mach. Intell., vol. 39, no. 6, pp. 1089–1102, 2017.
C. Peng, X. Gao, N. Wang, and J. Li, Graphical representation for heterogeneous face recognition, IEEE Trans, Pattern Anal. Mach. Intell., vol. 39, no. 2, pp. 301–312, 2017.
B. Yang, J. Chen, X. Ma, and M. Ye, Translation, Association and augmentation: Learning cross-modality re-identification from single-modality annotation, IEEE Transactions on Image Processing, vol. 32, pp. 5099–5113, 2023.
Y. Zhu, Z. Yang, L. Wang, S. Zhao, X. Hu, and D. Tao, Hetero-center loss for cross-modality person re-identification, Neurocomputing, vol. 386, pp. 97–109, 2020.
L. Wu, Y. Wang, and L. Shao, Cycle-consistent deep generative hashing for cross-modal retrieval, IEEE Trans. Image Process., vol. 28, no. 4, pp. 1602–1612, 2019.
D. Zhang, Z. Zhang, Y. Ju, C. Wang, Y. Xie, and Y. Qu, Dual mutual learning for cross-modality person re-identification, IEEE Trans. Circuits Syst. Video Technol., vol. 32, no. 8, pp. 5361–5373, 2022.
Z. Feng, J. Lai, and X. Xie, Learning modality-specific representations for visible-infrared person re-identification, IEEE Trans. Image Process., vol. 29, pp. 579–590, 2020.
Y. Zhou, R. Li, Y. Sun, K. Dong, and S. Li, Knowledge self-distillation for visible-infrared cross-modality person re-identification, Appl. Intell., vol. 52, no. 9, pp. 10617–10631, 2022.
L. Van Der Maaten, Accelerating t-SNE using tree-based algorithms, J. Mach. Learn. Res., vol. 15, no. 1, pp. 3221–3245, 2014.
The articles published in this open access journal are distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/).