Back to Search
Start Over
Denoising-Based Multiscale Feature Fusion for Remote Sensing Image Captioning.
- Source :
- IEEE Geoscience & Remote Sensing Letters; Mar2021, Vol. 18 Issue 3, p436-440, 5p
- Publication Year :
- 2021
-
Abstract
- With the benefits from deep learning technology, generating captions for remote sensing images has become achievable, and great progress has been made in this field in the recent years. However, a large-scale variation of remote sensing images, which would lead to errors or omissions in feature extraction, still limits the further improvement of caption quality. To address this problem, we propose a denoising-based multi-scale feature fusion (DMSFF) mechanism for remote sensing image captioning in this letter. The proposed DMSFF mechanism aggregates multiscale features with the denoising operation at the stage of visual feature extraction. It can help the encoder–decoder framework, which is widely used in image captioning, to obtain the denoising multiscale feature representation. In experiments, we apply the proposed DMSFF in the encoder–decoder framework and perform the comparative experiments on two public remote sensing image captioning data sets including UC Merced (UCM)-captions and Sydney-captions. The experimental results demonstrate the effectiveness of our method. [ABSTRACT FROM AUTHOR]
Details
- Language :
- English
- ISSN :
- 1545598X
- Volume :
- 18
- Issue :
- 3
- Database :
- Complementary Index
- Journal :
- IEEE Geoscience & Remote Sensing Letters
- Publication Type :
- Academic Journal
- Accession number :
- 148970225
- Full Text :
- https://doi.org/10.1109/LGRS.2020.2980933