Back to Search Start Over

Hypergraph clustering based multi-label cross-modal retrieval.

Authors :
Guo, Shengtang
Zhang, Huaxiang
Liu, Li
Liu, Dongmei
Lu, Xu
Li, Liujian
Source :
Journal of Visual Communication & Image Representation. Aug2024, Vol. 103, pN.PAG-N.PAG. 1p.
Publication Year :
2024

Abstract

Most existing cross-modal retrieval methods face challenges in establishing semantic connections between different modalities due to inherent heterogeneity among them. To establish semantic connections between different modalities and align relevant semantic features across modalities, so as to fully capture important information within the same modality, this paper considers the superiority of hypergraph in representing higher-order relationships, and proposes an image-text retrieval method based on hypergraph clustering. Specifically, we construct hypergraphs to capture feature relationships within image and text modalities, as well as between image and text. This allows us to effectively model complex relationships between features of different modalities and explore the semantic connectivity within and across modalities. To compensate for potential semantic feature loss during the construction of the hypergraph neural network, we design a weight-adaptive coarse and fine-grained feature fusion module for semantic supplementation. Comprehensive experimental results on three common datasets demonstrate the effectiveness of the proposed method. • A hypergraph cluster module is proposed to model modal relationships. • A fusion module dynamically learns weights is proposed. • The experiments conducted on three datasets prove the effectiveness of our method. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
10473203
Volume :
103
Database :
Academic Search Index
Journal :
Journal of Visual Communication & Image Representation
Publication Type :
Academic Journal
Accession number :
179420904
Full Text :
https://doi.org/10.1016/j.jvcir.2024.104258