Back to Search Start Over

Efficient Cross-Modal Retrieval via Deep Binary Hashing and Quantization

Authors :
Shi, Yang
Chung, Young-joo
Source :
BMVC 2021
Publication Year :
2022

Abstract

Cross-modal retrieval aims to search for data with similar semantic meanings across different content modalities. However, cross-modal retrieval requires huge amounts of storage and retrieval time since it needs to process data in multiple modalities. Existing works focused on learning single-source compact features such as binary hash codes that preserve similarities between different modalities. In this work, we propose a jointly learned deep hashing and quantization network (HQ) for cross-modal retrieval. We simultaneously learn binary hash codes and quantization codes to preserve semantic information in multiple modalities by an end-to-end deep learning architecture. At the retrieval step, binary hashing is used to retrieve a subset of items from the search space, then quantization is used to re-rank the retrieved items. We theoretically and empirically show that this two-stage retrieval approach provides faster retrieval results while preserving accuracy. Experimental results on the NUS-WIDE, MIR-Flickr, and Amazon datasets demonstrate that HQ achieves boosts of more than 7% in precision compared to supervised neural network-based compact coding models.<br />Comment: Accepted at BMVC 2021

Details

Database :
arXiv
Journal :
BMVC 2021
Publication Type :
Report
Accession number :
edsarx.2202.10232
Document Type :
Working Paper