Back to Search Start Over

Attentive Cross-Modal Fusion Network for RGB-D Saliency Detection

Authors :
Di Liu
Kao Zhang
Zhenzhong Chen
Source :
IEEE Transactions on Multimedia. 23:967-981
Publication Year :
2021
Publisher :
Institute of Electrical and Electronics Engineers (IEEE), 2021.

Abstract

In this paper, an attentive cross-modal fusion (ACMF) network is proposed for RGB-D salient object detection. The proposed method selectively fuses features in a cross-modal manner and uses a fusion refinement module to fuse output features from different resolutions. Our attentive cross-modal fusion network is built based on residual attention. In each level of ResNet output, both the RGB and depth features are turned into an identity map and a weighted attention map. The identity map is reweighted by the attention map of the paired modality. Moreover, the lower level features with higher resolution are adopted to refine the boundary of detected targets. The entire architecture can be trained end-to-end. The proposed ACMF is compared with state-of-the-art methods on eight recent datasets. The results demonstrate that our model can achieve advanced performance on RGB-D salient object detection.

Details

ISSN :
19410077 and 15209210
Volume :
23
Database :
OpenAIRE
Journal :
IEEE Transactions on Multimedia
Accession number :
edsair.doi...........e74cfb212a3f4cbfd4fd3daee4607a93
Full Text :
https://doi.org/10.1109/tmm.2020.2991523