Back to Search Start Over

Deep networks under scene-level supervision for multi-class geospatial object detection from remote sensing images

Authors :
Yongjun Zhang
Yansheng Li
Xin Huang
Alan L. Yuille
Source :
ISPRS Journal of Photogrammetry and Remote Sensing. 146:182-196
Publication Year :
2018
Publisher :
Elsevier BV, 2018.

Abstract

Due to its many applications, multi-class geospatial object detection has attracted increasing research interest in recent years. In the literature, existing methods highly depend on costly bounding box annotations. Based on the observation that scene-level tags provide important cues for the presence of objects, this paper proposes a weakly supervised deep learning (WSDL) method for multi-class geospatial object detection using scene-level tags only. Compared to existing WSDL methods which take scenes as isolated ones and ignore the mutual cues between scene pairs when optimizing deep networks, this paper exploits both the separate scene category information and mutual cues between scene pairs to sufficiently train deep networks for pursuing the superior object detection performance. In the first stage of our training method, we leverage pair-wise scene-level similarity to learn discriminative convolutional weights by exploiting the mutual information between scene pairs. The second stage utilizes point-wise scene-level tags to learn class-specific activation weights. While considering that the testing remote sensing image generally covers a large region and may contain a large number of objects from multiple categories with large size variations, a multi-scale scene-sliding-voting strategy is developed to calculate the class-specific activation maps (CAM) based on the aforementioned weights. Finally, objects can be detected by segmenting the CAM. The deep networks are trained on a seemingly unrelated remote sensing image scene classification dataset. Additionally, the testing phase is conducted on a publicly open multi-class geospatial object detection dataset. The experimental results demonstrate that the proposed deep networks dramatically outperform the state-of-the-art methods.

Details

ISSN :
09242716
Volume :
146
Database :
OpenAIRE
Journal :
ISPRS Journal of Photogrammetry and Remote Sensing
Accession number :
edsair.doi...........54e30090bc11a21976e3d8d08c2c4f3f
Full Text :
https://doi.org/10.1016/j.isprsjprs.2018.09.014