TY - JOUR
T1 - RSPH: Robust self-paced hashing for cross-modal retrieval
AU - Zhang, Donglin
AU - Hu, Zhikai
AU - Wu, Xiao-Jun
AU - Kittler, Josef
N1 - This work was supported by National Natural Science Foundation of China (62202204), the Fundamental Research Funds for the Central Universities (JUSRP123032), and the National Key Research and Development Program of China (2023YFF1105102).
Publisher Copyright:
© 2025 Elsevier Ltd
PY - 2025/7/9
Y1 - 2025/7/9
N2 - Recently, hashing technology has attracted considerable attention in cross-modal retrieval due to its high retrieval efficiency and low storage overhead. Therefore, many cross-modal hashing (CMH) methods have been proposed and performed well. However, most existing CMH methods treat all samples equally without discrimination, which may lead to suboptimal local minima. Besides, most existing approaches usually leverage modality-specific information when learning hashing functions, which cannot effectively capture both intra-and inter-modality correlations in multimodal data. To mitigate these problems, a novel Robust Self-Paced hashing (RSPH) approach is developed for cross-modal retrieval. Concretely, in the proposed RSPH, the multimodal data are progressively integrated into the hashing learning process from easy to hard, helping to avoid poor local minima. Besides, the ℓ
2,1-norm is employed in the proposed method to further reduce the impact of outliers and noise. To fully exploit multimodal information, this method introduces a collaboration projection scheme that captures both modality-specific properties and cross-modal correlations, resulting in more effective hash functions. Experimental results on three benchmark datasets validate our RSPH can achieve state-of-the-art performance, demonstrating its efficacy.
AB - Recently, hashing technology has attracted considerable attention in cross-modal retrieval due to its high retrieval efficiency and low storage overhead. Therefore, many cross-modal hashing (CMH) methods have been proposed and performed well. However, most existing CMH methods treat all samples equally without discrimination, which may lead to suboptimal local minima. Besides, most existing approaches usually leverage modality-specific information when learning hashing functions, which cannot effectively capture both intra-and inter-modality correlations in multimodal data. To mitigate these problems, a novel Robust Self-Paced hashing (RSPH) approach is developed for cross-modal retrieval. Concretely, in the proposed RSPH, the multimodal data are progressively integrated into the hashing learning process from easy to hard, helping to avoid poor local minima. Besides, the ℓ
2,1-norm is employed in the proposed method to further reduce the impact of outliers and noise. To fully exploit multimodal information, this method introduces a collaboration projection scheme that captures both modality-specific properties and cross-modal correlations, resulting in more effective hash functions. Experimental results on three benchmark datasets validate our RSPH can achieve state-of-the-art performance, demonstrating its efficacy.
KW - Retrieval
KW - Hashing
KW - Multimodal data
KW - Cross-modal
UR - https://www.scopus.com/pages/publications/105010555314
U2 - 10.1016/j.patcog.2025.112072
DO - 10.1016/j.patcog.2025.112072
M3 - Journal article
SN - 0031-3203
VL - 171, Part A
JO - Pattern Recognition
JF - Pattern Recognition
M1 - 112072
ER -