Back to Search
Start Over
Self-Supervised Pretraining for Stereoscopic Image Super-Resolution With Parallax-Aware Masking
- Source :
- IEEE Transactions on Broadcasting; 2024, Vol. 70 Issue: 2 p482-491, 10p
- Publication Year :
- 2024
-
Abstract
- Most existing learning-based methods for stereoscopic image super-resolution rely on a great number of high-resolution stereoscopic images as labels. To alleviate the problem of data dependency, this paper proposes a self-supervised pretraining-based method for stereoscopic image super-resolution (SelfSSR). Specifically, to develop a self-supervised pretext task for stereoscopic images, a parallax-aware masking strategy (PAMS) is designed to adaptively mask matching areas of the left and right views. With PAMS, the network is encouraged to effectively predict missing information of input images. Besides, a cross-view Transformer module (CVTM) is presented to aggregate the intra-view and inter-view information simultaneously for stereoscopic image reconstruction. Meanwhile, the cross-attention map learned by CVTM is utilized to guide the masking process in PAMS. Comparative results on four datasets show that the proposed SelfSSR achieves state-of-the-art performance by using only 10% of labeled training data.
Details
- Language :
- English
- ISSN :
- 00189316 and 15579611
- Volume :
- 70
- Issue :
- 2
- Database :
- Supplemental Index
- Journal :
- IEEE Transactions on Broadcasting
- Publication Type :
- Periodical
- Accession number :
- ejs66619914
- Full Text :
- https://doi.org/10.1109/TBC.2024.3382960