Back to Search
Start Over
Hypergraph clustering based multi-label cross-modal retrieval.
- Source :
-
Journal of Visual Communication & Image Representation . Aug2024, Vol. 103, pN.PAG-N.PAG. 1p. - Publication Year :
- 2024
-
Abstract
- Most existing cross-modal retrieval methods face challenges in establishing semantic connections between different modalities due to inherent heterogeneity among them. To establish semantic connections between different modalities and align relevant semantic features across modalities, so as to fully capture important information within the same modality, this paper considers the superiority of hypergraph in representing higher-order relationships, and proposes an image-text retrieval method based on hypergraph clustering. Specifically, we construct hypergraphs to capture feature relationships within image and text modalities, as well as between image and text. This allows us to effectively model complex relationships between features of different modalities and explore the semantic connectivity within and across modalities. To compensate for potential semantic feature loss during the construction of the hypergraph neural network, we design a weight-adaptive coarse and fine-grained feature fusion module for semantic supplementation. Comprehensive experimental results on three common datasets demonstrate the effectiveness of the proposed method. • A hypergraph cluster module is proposed to model modal relationships. • A fusion module dynamically learns weights is proposed. • The experiments conducted on three datasets prove the effectiveness of our method. [ABSTRACT FROM AUTHOR]
- Subjects :
- *HYPERGRAPHS
*GRAPH theory
*SEMANTICS
*IMAGE retrieval
*MULTIMEDIA systems
Subjects
Details
- Language :
- English
- ISSN :
- 10473203
- Volume :
- 103
- Database :
- Academic Search Index
- Journal :
- Journal of Visual Communication & Image Representation
- Publication Type :
- Academic Journal
- Accession number :
- 179420904
- Full Text :
- https://doi.org/10.1016/j.jvcir.2024.104258