Back to Search Start Over

S $^3$ Net: Self-Supervised Self-Ensembling Network for Semi-Supervised RGB-D Salient Object Detection

Authors :
Xin Yang
C.L. Philip Chen
Carola-Bibiane Schönlieb
Ping Li
Xiaoqiang Wang
Lei Zhu
Zhang Qing
Weiming Wang
Source :
IEEE Transactions on Multimedia. 25:676-689
Publication Year :
2023
Publisher :
Institute of Electrical and Electronics Engineers (IEEE), 2023.

Abstract

RGB-D salient object detection aims to detect visually distinctive objects or regions from a pair of the RGB image and the depth image. State-of-the-art RGB-D saliency detectors are mainly based on convolutional neural networks but almost suffer from an intrinsic limitation relying on the labeled data, thus degrading detection accuracy in complex cases. In this work, we present a self-supervised self-ensembling network (S $^3$ Net) for semi-supervised RGB-D salient object detection by leveraging the unlabeled data and exploring a self-supervised learning mechanism. To be specific, we first build a self-guided convolutional neural network (SG-CNN) as a baseline model by developing a series of three-layer cross-model feature fusion (TCF) modules to leverage complementary information among depth and RGB modalities and formulating an auxiliary task that predicts a self-supervised image rotation angle. After that, to further explore the knowledge from unlabeled data, we assign SG-CNN to a student network and a teacher network, and encourage the saliency predictions and self-supervised rotation predictions from these two networks to be consistent on the unlabeled data. Experimental results on seven widely-used benchmark datasets demonstrate that our network quantitatively and qualitatively outperforms the state-of-the-art methods.

Details

ISSN :
19410077 and 15209210
Volume :
25
Database :
OpenAIRE
Journal :
IEEE Transactions on Multimedia
Accession number :
edsair.doi...........3c541dfc84f17e2df954e5f100252a20
Full Text :
https://doi.org/10.1109/tmm.2021.3129730