Back to Search Start Over

Discriminative latent semantics-preserving similarity embedding hashing for cross-modal retrieval.

Authors :
Chen, Yongfeng
Tan, Junpeng
Yang, Zhijing
Cheng, Yongqiang
Chen, Ruihan
Source :
Neural Computing & Applications. Jun2024, Vol. 36 Issue 18, p10655-10680. 26p.
Publication Year :
2024

Abstract

Recently, there has been a significant increase in interest in cross-modal hashing technology. For hash code learning, most previous supervision methods use label information to create a similarity matrix in a straightforward manner. However, there are still the following challenges: (1) The asymmetric similarity matrix method only considers the similarity of labels, and the discriminant constraint in Hamming space is ignored; (2) there are optimization errors between the cross-modal semantic correlation of the Hamming space and the nonlinearity of the feature space; and (3) the cross-modal hash matrix is in a dynamic state during the optimization process, and the hash code is easily disturbed and there is bit uncertainty. To this end, we propose the Discriminative Latent Semantics-preserving Similarity Embedding Cross-modal Hashing (DLSSECH) method. Specifically, to reduce the quantization error, we introduce a non-asymmetric similarity decomposition based on orthogonal rotation bias. It can decompose the bitwise correlation of the learned hash code to capture more discriminative semantic information and compact hash code. In addition, to capture the cross-modal semantic correlation of nonlinear feature transformation and reduce quantization error, we propose the latent correlation error matrices based on orthogonal rotation decomposition. The model maintains the maximum difference in semantic dependencies between the projected data and its semantic representation. Finally, we use a sparse common hash matrix and non-asymmetric similarity decomposition factors to solve the uncertainty of the dynamic changes of the hash code. The effectiveness of DLSSECH has demonstrated through experiments on four cross-modal retrieval datasets, where it outperformed some state-of-the-art methods. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
09410643
Volume :
36
Issue :
18
Database :
Academic Search Index
Journal :
Neural Computing & Applications
Publication Type :
Academic Journal
Accession number :
177560461
Full Text :
https://doi.org/10.1007/s00521-024-09616-y