283 results
Search Results
2. SSAformer: Spatial–Spectral Aggregation Transformer for Hyperspectral Image Super-Resolution.
- Author
-
Wang, Haoqian, Zhang, Qi, Peng, Tao, Xu, Zhongjie, Cheng, Xiangai, Xing, Zhongyang, and Li, Teng
- Subjects
- *
TRANSFORMER models , *HIGH resolution imaging , *CONVOLUTIONAL neural networks , *REMOTE sensing , *ENVIRONMENTAL monitoring , *SPECTRAL imaging , *IMAGE reconstruction algorithms - Abstract
The hyperspectral image (HSI) distinguishes itself in material identification through its exceptional spectral resolution. However, its spatial resolution is constrained by hardware limitations, prompting the evolution of HSI super-resolution (SR) techniques. Single HSI SR endeavors to reconstruct high-spatial-resolution HSI from low-spatial-resolution inputs, and recent progress in deep learning-based algorithms has significantly advanced the quality of reconstructed images. However, convolutional methods struggle to extract comprehensive spatial and spectral features. Transformer-based models have yet to harness long-range dependencies across both dimensions fully, thus inadequately integrating spatial and spectral data. To solve the above problem, in this paper, we propose a new HSI SR method, SSAformer, which merges the strengths of CNNs and Transformers. It introduces specially designed attention mechanisms for HSI, including spatial and spectral attention modules, and overcomes the previous challenges in extracting and amalgamating spatial and spectral information. Evaluations on benchmark datasets show that SSAformer surpasses contemporary methods in enhancing spatial details and preserving spectral accuracy, underscoring its potential to expand HSI's utility in various domains, such as environmental monitoring and remote sensing. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. Changes in the Water Area of an Inland River Terminal Lake (Taitma Lake) Driven by Climate Change and Human Activities, 2017–2022.
- Author
-
Zi, Feng, Wang, Yong, Lu, Shanlong, Ikhumhen, Harrison Odion, Fang, Chun, Li, Xinru, Wang, Nan, and Kuang, Xinya
- Subjects
- *
ENDORHEIC lakes , *WATER resources development , *CONVOLUTIONAL neural networks , *LAKES , *DEEP learning , *CLIMATE change - Abstract
Constructed from a dataset capturing the seasonal and annual water body distribution of the lower Qarqan River in the Taitma Lake area from 2017 to 2022, and combined with the meteorological and hydraulic engineering data, the spatial and temporal change patterns of the Taitma Lake watershed area were determined. Analyses were conducted using Planetscope (PS) satellite images and a deep learning model. The results revealed the following: ① Deep learning-based water body extraction provides significantly greater accuracy than the conventional water body index approach. With an impressive accuracy of up to 96.0%, UPerNet was found to provide the most effective extraction results among the three convolutional neural networks (U-Net, DeeplabV3+, and UPerNet) used for semantic segmentation; ② Between 2017 and 2022, Taitma Lake's water area experienced a rapid decrease, with the distribution of water predominantly shifting towards the east–west direction more than the north–south. The shifts between 2017 and 2020 and between 2020 and 2022 were clearly discernible, with the latter stage (2020–2022) being more significant than the former (2017–2020); ③ According to observations, Taitma Lake's changing water area has been primarily influenced by human activity over the last six years. Based on the research findings of this paper, it was observed that this study provides a valuable scientific basis for water resource allocation aiming to balance the development of water resources in the middle and upper reaches of the Tarim and Qarqan Rivers, as well as for the ecological protection of the downstream Taitma Lake. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. MEA-EFFormer: Multiscale Efficient Attention with Enhanced Feature Transformer for Hyperspectral Image Classification.
- Author
-
Sun, Qian, Zhao, Guangrui, Fang, Yu, Fang, Chenrong, Sun, Le, and Li, Xingying
- Subjects
- *
IMAGE recognition (Computer vision) , *CONVOLUTIONAL neural networks , *DEEP learning , *TRANSFORMER models , *FEATURE extraction - Abstract
Hyperspectral image classification (HSIC) has garnered increasing attention among researchers. While classical networks like convolution neural networks (CNNs) have achieved satisfactory results with the advent of deep learning, they are confined to processing local information. Vision transformers, despite being effective at establishing long-distance dependencies, face challenges in extracting high-representation features for high-dimensional images. In this paper, we present the multiscale efficient attention with enhanced feature transformer (MEA-EFFormer), which is designed for the efficient extraction of spectral–spatial features, leading to effective classification. MEA-EFFormer employs a multiscale efficient attention feature extraction module to initially extract 3D convolution features and applies effective channel attention to refine spectral information. Following this, 2D convolution features are extracted and integrated with local binary pattern (LBP) spatial information to augment their representation. Then, the processed features are fed into a spectral–spatial enhancement attention (SSEA) module that facilitates interactive enhancement of spectral–spatial information across the three dimensions. Finally, these features undergo classification through a transformer encoder. We evaluate MEA-EFFormer against several state-of-the-art methods on three datasets and demonstrate its outstanding HSIC performance. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. Object-Based Semi-Supervised Spatial Attention Residual UNet for Urban High-Resolution Remote Sensing Image Classification.
- Author
-
Lu, Yuanbing, Li, Huapeng, Zhang, Ce, and Zhang, Shuqing
- Subjects
- *
CONVOLUTIONAL neural networks , *DISTRIBUTION (Probability theory) , *WILCOXON signed-rank test , *DEEP learning , *LAND cover - Abstract
Accurate urban land cover information is crucial for effective urban planning and management. While convolutional neural networks (CNNs) demonstrate superior feature learning and prediction capabilities using image-level annotations, the inherent mixed-category nature of input image patches leads to classification errors along object boundaries. Fully convolutional neural networks (FCNs) excel at pixel-wise fine segmentation, making them less susceptible to heterogeneous content, but they require fully annotated dense image patches, which may not be readily available in real-world scenarios. This paper proposes an object-based semi-supervised spatial attention residual UNet (OS-ARU) model. First, multiscale segmentation is performed to obtain segments from a remote sensing image, and segments containing sample points are assigned the categories of the corresponding points, which are used to train the model. Then, the trained model predicts class probabilities for all segments. Each unlabeled segment's probability distribution is compared against those of labeled segments for similarity matching under a threshold constraint. Through label propagation, pseudo-labels are assigned to unlabeled segments exhibiting high similarity to labeled ones. Finally, the model is retrained using the augmented training set incorporating the pseudo-labeled segments. Comprehensive experiments on aerial image benchmarks for Vaihingen and Potsdam demonstrate that the proposed OS-ARU achieves higher classification accuracy than state-of-the-art models, including OCNN, 2OCNN, and standard OS-U, reaching an overall accuracy (OA) of 87.83% and 86.71%, respectively. The performance improvements over the baseline methods are statistically significant according to the Wilcoxon Signed-Rank Test. Despite using significantly fewer sparse annotations, this semi-supervised approach still achieves comparable accuracy to the same model under full supervision. The proposed method thus makes a step forward in substantially alleviating the heavy sampling burden of FCNs (densely sampled deep learning models) to effectively handle the complex issue of land cover information identification and classification. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
6. Prediction of Sea Surface Temperature Using U-Net Based Model.
- Author
-
Ren, Jing, Wang, Changying, Sun, Ling, Huang, Baoxiang, Zhang, Deyu, Mu, Jiadong, and Wu, Jianqiang
- Subjects
- *
OCEAN temperature , *CONVOLUTIONAL neural networks - Abstract
Sea surface temperature (SST) is a key parameter in ocean hydrology. Currently, existing SST prediction methods fail to fully utilize the potential spatial correlation between variables. To address this challenge, we propose a spatiotenporal UNet (ST-UNet) model based on the UNet model. In particular, in the encoding phase of ST-UNet, we use parallel convolution with different kernel sizes to efficiently extract spatial features, and use ConvLSTM to capture temporal features based on the utilization of spatial features. Atrous Spatial Pyramid Pooling (ASPP) module is placed at the bottleneck of the network to further incorporate the multi-scale features, allowing the spatial features to be fully utilized. The final prediction is then generated in the decoding stage using parallel convolution with different kernel sizes similar to the encoding stage. We conducted a series of experiments on the Bohai Sea and Yellow Sea SST data set, as well as the South China Sea SST data set, using SST data from the past 35 days to predict SST data for 1, 3, and 7 days in the future. The model was trained using data spanning from 2010 to 2021, with data from 2022 being utilized to assess the model's predictive performance. The experimental results show that the model proposed in this research paper achieves excellent results at different prediction scales in both sea areas, and the model consistently outperforms other methods. Specifically, in the Bohai Sea and Yellow Sea sea areas, when the prediction scales are 1, 3, and 7 days, the MAE of ST-UNet outperforms the best results of the other three compared models by 17%, 12%, and 2%, and the MSE by 16%, 18%, and 9%, respectively. In the South China Sea, when the prediction ranges are 1, 3, and 7 days, the MAE of ST-UNet is 27%, 18%, and 3% higher than the best of the other three compared models, and the MSE is 46%, 39%, and 16% higher, respectively. Our results highlight the effectiveness of the ST-UNet model in capturing spatial correlations and accurately predicting SST. The proposed model is expected to improve marine hydrographic studies. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. Ship Detection with Deep Learning in Optical Remote-Sensing Images: A Survey of Challenges and Advances.
- Author
-
Zhao, Tianqi, Wang, Yongcheng, Li, Zheng, Gao, Yunxiao, Chen, Chi, Feng, Hao, and Zhao, Zhikang
- Subjects
- *
DEEP learning , *REMOTE-sensing images , *OPTICAL remote sensing , *OPTICAL images , *CONVOLUTIONAL neural networks , *TRANSFORMER models , *FEATURE extraction - Abstract
Ship detection aims to automatically identify whether there are ships in the images, precisely classifies and localizes them. Regardless of whether utilizing early manually designed methods or deep learning technology, ship detection is dedicated to exploring the inherent characteristics of ships to enhance recall. Nowadays, high-precision ship detection plays a crucial role in civilian and military applications. In order to provide a comprehensive review of ship detection in optical remote-sensing images (SDORSIs), this paper summarizes the challenges as a guide. These challenges include complex marine environments, insufficient discriminative features, large scale variations, dense and rotated distributions, large aspect ratios, and imbalances between positive and negative samples. We meticulously review the improvement methods and conduct a detailed analysis of the strengths and weaknesses of these methods. We compile ship information from common optical remote sensing image datasets and compare algorithm performance. Simultaneously, we compare and analyze the feature extraction capabilities of backbones based on CNNs and Transformer, seeking new directions for the development in SDORSIs. Promising prospects are provided to facilitate further research in the future. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. Forest Aboveground Biomass Estimation Using Multisource Remote Sensing Data and Deep Learning Algorithms: A Case Study over Hangzhou Area in China.
- Author
-
Tian, Xin, Li, Jiejie, Zhang, Fanyi, Zhang, Haibo, and Jiang, Mi
- Subjects
- *
DEEP learning , *BIOMASS estimation , *MACHINE learning , *MULTISPECTRAL imaging , *REMOTE sensing , *FOREST biomass , *CONVOLUTIONAL neural networks , *SYNTHETIC aperture radar - Abstract
The accurate estimation of forest aboveground biomass is of great significance for forest management and carbon balance monitoring. Remote sensing instruments have been widely applied in forest parameters inversion with wide coverage and high spatiotemporal resolution. In this paper, the capability of different remote-sensed imagery was investigated, including multispectral images (GaoFen-6, Sentinel-2 and Landsat-8) and various SAR (Synthetic Aperture Radar) data (GaoFen-3, Sentinel-1, ALOS-2), in aboveground forest biomass estimation. In particular, based on the forest inventory data of Hangzhou in China, the Random Forest (RF), Convolutional Neural Network (CNN) and Convolutional Neural Networks Long Short-Term Memory Networks (CNN-LSTM) algorithms were deployed to construct the forest biomass estimation models, respectively. The estimate accuracies were evaluated under the different configurations of images and methods. The results show that for the SAR data, ALOS-2 has a higher biomass estimation accuracy than the GaoFen-3 and Sentinel-1. Moreover, the GaoFen-6 data is slightly worse than Sentinel-2 and Landsat-8 optical data in biomass estimation. In contrast with the single source, integrating multisource data can effectively enhance accuracy, with improvements ranging from 5% to 10%. The CNN-LSTM generally performs better than CNN and RF, regardless of the data used. The combination of CNN-LSTM and multisource data provided the best results in this case and can achieve the maximum R2 value of up to 0.74. It was found that the majority of the biomass values in the study area in 2018 ranged from 60 to 90 Mg/ha, with an average value of 64.20 Mg/ha. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. CroplandCDNet: Cropland Change Detection Network for Multitemporal Remote Sensing Images Based on Multilayer Feature Transmission Fusion of an Adaptive Receptive Field.
- Author
-
Wu, Qiang, Huang, Liang, Tang, Bo-Hui, Cheng, Jiapei, Wang, Meiqi, and Zhang, Zixuan
- Subjects
- *
CONVOLUTIONAL neural networks , *CHANGE-point problems , *FARMS , *MARKOV random fields , *REMOTE-sensing images , *FEATURE extraction - Abstract
Dynamic monitoring of cropland using high spatial resolution remote sensing images is a powerful means to protect cropland resources. However, when a change detection method based on a convolutional neural network employs a large number of convolution and pooling operations to mine the deep features of cropland, the accumulation of irrelevant features and the loss of key features will lead to poor detection results. To effectively solve this problem, a novel cropland change detection network (CroplandCDNet) is proposed in this paper; this network combines an adaptive receptive field and multiscale feature transmission fusion to achieve accurate detection of cropland change information. CroplandCDNet first effectively extracts the multiscale features of cropland from bitemporal remote sensing images through the feature extraction module and subsequently embeds the receptive field adaptive SK attention (SKA) module to emphasize cropland change. Moreover, the SKA module effectively uses spatial context information for the dynamic adjustment of the convolution kernel size of cropland features at different scales. Finally, multiscale features and difference features are transmitted and fused layer by layer to obtain the content of cropland change. In the experiments, the proposed method is compared with six advanced change detection methods using the cropland change detection dataset (CLCD). The experimental results show that CroplandCDNet achieves the best F1 and OA at 76.04% and 94.47%, respectively. Its precision and recall are second best of all models at 76.46% and 75.63%, respectively. Moreover, a generalization experiment was carried out using the Jilin-1 dataset, which effectively verified the reliability of CroplandCDNet in cropland change detection. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
10. AIDB-Net: An Attention-Interactive Dual-Branch Convolutional Neural Network for Hyperspectral Pansharpening.
- Author
-
Sun, Qian, Sun, Yu, and Pan, Chengsheng
- Subjects
- *
CONVOLUTIONAL neural networks , *DEEP learning - Abstract
Despite notable advancements achieved on Hyperspectral (HS) pansharpening tasks through deep learning techniques, previous methods are inherently constrained by convolution or self-attention intrinsic defects, leading to limited performance. In this paper, we proposed an Attention-Interactive Dual-Branch Convolutional Neural Network (AIDB-Net) for HS pansharpening. Our model purely consists of convolutional layers and simultaneously inherits the strengths of both convolution and self-attention, especially the modeling of short- and long-range dependencies. Specially, we first extract, tokenize, and align the hyperspectral image (HSI) and panchromatic image (PAN) by Overlapping Patch Embedding Blocks. Then, we specialize a novel Spectral-Spatial Interactive Attention which is able to globally interact and fuse the cross-modality features. The resultant token-global similarity scores can guide the refinement and renewal of the textural details and spectral characteristics within HSI features. By deeply combined these two paradigms, our AIDB-Net significantly improve the pansharpening performance. Moreover, with the acceleration by the convolution inductive bias, our interactive attention can be trained without large scale dataset and achieves competitive time cost with its counterparts. Compared with the state-of-the-art methods, our AIDB-Net makes 5.2%, 3.1%, and 2.2% improvement on PSNR metric on three public datasets, respectively. Comprehensive experiments quantitatively and qualitatively demonstrate the effectiveness and superiority of our AIDB-Net. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
11. A Lightning Classification Method Based on Convolutional Encoding Features.
- Author
-
Zhu, Shunxing, Zhang, Yang, Fan, Yanfeng, Sun, Xiubin, Zheng, Dong, Zhang, Yijun, Lyu, Weitao, Zhang, Huiyi, and Wang, Jingxuan
- Subjects
- *
CONVOLUTIONAL neural networks , *RANDOM forest algorithms , *THUNDERSTORMS - Abstract
At present, for business lightning positioning systems, the classification of lightning discharge types is mostly based on lightning pulse signal features, and there is still a lot of room for improvement. We propose a lightning discharge classification method based on convolutional encoding features. This method utilizes convolutional neural networks to extract encoding features, and uses random forests to classify the extracted encoding features, achieving high accuracy discrimination for various lightning discharge events. Compared with traditional multi-parameter-based methods, the new method proposed in this paper has the ability to identify multiple lightning discharge events and does not require precise detailed feature engineering to extract individual pulse parameters. The accuracy of this method for identifying lightning discharge types in intra-cloud flash (IC), cloud-to-ground flash (CG), and narrow bipolar events (NBEs) is 97%, which is higher than that of multi-parameter methods. Moreover, our method can complete the classification task of lightning signals at a faster speed. Under the same conditions, the new method only requires 28.2 µs to identify one pulse, while deep learning-based methods require 300 µs. This method has faster recognition speed and higher accuracy in identifying multiple discharge types, which can better meet the needs of real-time business positioning. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
12. Hyperspectral Image Classification Based on Mutually Guided Image Filtering.
- Author
-
Zhan, Ying, Hu, Dan, Yu, Xianchuan, and Wang, Yufeng
- Subjects
- *
IMAGE recognition (Computer vision) , *ARTIFICIAL neural networks , *CONVOLUTIONAL neural networks , *FEATURE extraction , *GENERATIVE adversarial networks , *HYPERSPECTRAL imaging systems , *REMOTE sensing - Abstract
Hyperspectral remote sensing images (HSIs) have both spectral and spatial characteristics. The adept exploitation of these attributes is central to enhancing the classification accuracy of HSIs. In order to effectively utilize spatial and spectral features to classify HSIs, this paper proposes a method for the spatial feature extraction of HSIs based on a mutually guided image filter (muGIF) and combined with the band-distance-grouped principal component. Firstly, aiming at the problem that previously guided image filtering cannot effectively deal with the inconsistent information structure between the guided and target information, a method for extracting spatial features using muGIF is proposed. Then, aiming at the problem of the information loss caused by a single principal component as a guided image in the traditional GIF-based spatial–spectral classification, a spatial feature-extraction framework based on the band-distance-grouped principal component is proposed. The method groups the bands according to the band distance and extracts the principal components of each set of band subsets as the guide map of the current band subset to filter the HSIs. A deep convolutional neural network model and a generative adversarial network model for the filtered HSIs are constructed and then trained using samples for HSIs' spatial–spectral classification. Experiments show that compared with the traditional methods and several popular spatial–spectral HSI classification methods based on a filter, the proposed methods based on muGIF can effectively extract the spatial–spectral features and improve the classification accuracy of HSIs. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
13. Remote Sensing for Maritime Monitoring and Vessel Identification.
- Author
-
Salerno, Emanuele, Di Paola, Claudio, and Lo Duca, Angelica
- Subjects
- *
DEEP learning , *REMOTE sensing , *CONVOLUTIONAL neural networks , *SURVEILLANCE radar , *SYNTHETIC aperture radar , *INFORMATION technology , *PATTERN recognition systems - Abstract
This document explores the significance of remote sensing in monitoring maritime activities and identifying vessels. It emphasizes the need for surveillance to ensure safety, security, and emergency management, given the increasing number of vessels worldwide. The document highlights the use of technologies like the Automatic Identification System (AIS) and remote sensing in situations where collaborative systems are not reliable. It also discusses the integration of data from different sensors and the application of data science techniques for a comprehensive assessment of maritime traffic. The document concludes by summarizing research papers on ship detection, tracking, and classification using various sensors and data processing techniques. [Extracted from the article]
- Published
- 2024
- Full Text
- View/download PDF
14. Multi-View Scene Classification Based on Feature Integration and Evidence Decision Fusion.
- Author
-
Zhou, Weixun, Shi, Yongxin, and Huang, Xiao
- Subjects
- *
FEATURE extraction , *IMAGE recognition (Computer vision) , *IMAGE fusion , *CONVOLUTIONAL neural networks , *DEEP learning - Abstract
Leveraging multi-view remote sensing images in scene classification tasks significantly enhances the precision of such classifications. This approach, however, poses challenges due to the simultaneous use of multi-view images, which often leads to a misalignment between the visual content and semantic labels, thus complicating the classification process. In addition, as the number of image viewpoints increases, the quality problem for remote sensing images further limits the effectiveness of multi-view image classification. Traditional scene classification methods predominantly employ SoftMax deep learning techniques, which lack the capability to assess the quality of remote sensing images or to provide explicit explanations for the network's predictive outcomes. To address these issues, this paper introduces a novel end-to-end multi-view decision fusion network specifically designed for remote sensing scene classification. The network integrates information from multi-view remote sensing images under the guidance of image credibility and uncertainty, and when the multi-view image fusion process encounters conflicts, it greatly alleviates the conflicts and provides more reasonable and credible predictions for the multi-view scene classification results. Initially, multi-scale features are extracted from the multi-view images using convolutional neural networks (CNNs). Following this, an asymptotic adaptive feature fusion module (AAFFM) is constructed to gradually integrate these multi-scale features. An adaptive spatial fusion method is then applied to assign different spatial weights to the multi-scale feature maps, thereby significantly enhancing the model's feature discrimination capability. Finally, an evidence decision fusion module (EDFM), utilizing evidence theory and the Dirichlet distribution, is developed. This module quantitatively assesses the uncertainty in the multi-perspective image classification process. Through the fusing of multi-perspective remote sensing image information in this module, a rational explanation for the prediction results is provided. The efficacy of the proposed method was validated through experiments conducted on the AiRound and CV-BrCT datasets. The results show that our method not only improves single-view scene classification results but also advances multi-view remote sensing scene classification results by accurately characterizing the scene and mitigating the conflicting nature of the fusion process. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
15. Remote Sensing Extraction of Lakes on the Tibetan Plateau Based on the Google Earth Engine and Deep Learning.
- Author
-
Pang, Yunxuan, Yu, Junchuan, Xi, Laidian, Ge, Daqing, Zhou, Ping, Hou, Changhong, He, Peng, and Zhao, Liu
- Subjects
- *
DEEP learning , *CONVOLUTIONAL neural networks , *REMOTE sensing , *WATER management , *LAKES , *WATER supply - Abstract
Lakes are an important component of global water resources. In order to achieve accurate lake extractions on a large scale, this study takes the Tibetan Plateau as the study area and proposes an Automated Lake Extraction Workflow (ALEW) based on the Google Earth Engine (GEE) and deep learning in response to the problems of a low lake identification accuracy and low efficiency in complex situations. It involves pre-processing massive images and creating a database of examples of lake extraction on the Tibetan Plateau. A lightweight convolutional neural network named LiteConvNet is constructed that makes it possible to obtain spatial–spectral features for accurate extractions while using less computational resources. We execute model training and online predictions using the Google Cloud platform, which leads to the rapid extraction of lakes over the whole Tibetan Plateau. We assess LiteConvNet, along with thresholding, traditional machine learning, and various open-source classification products, through both visual interpretation and quantitative analysis. The results demonstrate that the LiteConvNet model may greatly enhance the precision of lake extraction in intricate settings, achieving an overall accuracy of 97.44%. The method presented in this paper demonstrates promising capabilities in extracting lake information on a large scale, offering practical benefits for the remote sensing monitoring and management of water resources in cloudy and climate-differentiated regions. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
16. Ship Formation Identification with Spatial Features and Deep Learning for HFSWR.
- Author
-
Wang, Jiaqi, Liu, Aijun, Yu, Changjun, and Ji, Yuanzheng
- Subjects
- *
DEEP learning , *CONVOLUTIONAL neural networks , *MACHINE learning , *SHIP models , *SHIPS - Abstract
Ship detection has been an area of focus for high-frequency surface wave radar (HFSWR). The detection and identification of ship formation have proven significant in early warning, while studies on the formation identification are limited due to the complex background and low resolution of HFSWR. In this paper, we first establish a spatial distribution model of ship formation in HFSWR. Then, we propose a cascade identification algorithm of ship formation in the clutter edge. The proposed algorithm includes a preprocessing stage and a two-stage formation identification stage. The Faster R-CNN is introduced in the preprocessing stage to locate the clutter regions. In the first stage, we propose an extremum detector based on connected regions to extract suspicious regions. The suspicious regions contain ship formations, single-ship targets, and false targets. In the second stage, we design a network connected by a convolutional neural network (CNN) and an extreme learning machine (ELM) to identify two densely distributed ship formations from inhomogeneous clutter and single-ship targets. The experimental results based on the factual HFSWR background demonstrate that the proposed cascade identification algorithm is superior to the extremum detector combined with the classical CNN algorithm for ship formation identification. Meanwhile, the proposed algorithm performs well in weak formation and deformed formation identification. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
17. Spatial-Spectral BERT for Hyperspectral Image Classification.
- Author
-
Ashraf, Mahmood, Zhou, Xichuan, Vivone, Gemine, Chen, Lihui, Chen, Rong, and Majdard, Reza Seifi
- Subjects
- *
IMAGE recognition (Computer vision) , *LANGUAGE models , *DEEP learning , *TRANSFORMER models , *CONVOLUTIONAL neural networks , *SPECTRAL imaging - Abstract
Several deep learning and transformer models have been recommended in previous research to deal with the classification of hyperspectral images (HSIs). Among them, one of the most innovative is the bidirectional encoder representation from transformers (BERT), which applies a distance-independent approach to capture the global dependency among all pixels in a selected region. However, this model does not consider the local spatial-spectral and spectral sequential relations. In this paper, a dual-dimensional (i.e., spatial and spectral) BERT (the so-called D2BERT) is proposed, which improves the existing BERT model by capturing more global and local dependencies between sequential spectral bands regardless of distance. In the proposed model, two BERT branches work in parallel to investigate relations among pixels and spectral bands, respectively. In addition, the layer intermediate information is used for supervision during the training phase to enhance the performance. We used two widely employed datasets for our experimental analysis. The proposed D2BERT shows superior classification accuracy and computational efficiency with respect to some state-of-the-art neural networks and the previously developed BERT model for this task. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
18. Hyperspectral Image Super-Resolution Based on Feature Diversity Extraction.
- Author
-
Zhang, Jing, Zheng, Renjie, Wan, Zekang, Geng, Ruijing, Wang, Yi, Yang, Yu, Zhang, Xuepeng, and Li, Yunsong
- Subjects
- *
DEEP learning , *FEATURE extraction , *IMAGE reconstruction algorithms , *HIGH resolution imaging , *CONVOLUTIONAL neural networks - Abstract
Deep learning is an important research topic in the field of image super-resolution. Problematically, the performance of existing hyperspectral image super-resolution networks is limited by feature learning for hyperspectral images. Nevertheless, the current algorithms exhibit some limitations in extracting diverse features. In this paper, we address limitations to existing hyperspectral image super-resolution networks, focusing on feature learning challenges. We introduce the Channel-Attention-Based Spatial–Spectral Feature Extraction network (CSSFENet) to enhance hyperspectral image feature diversity and optimize network loss functions. Our contributions include: (a) a convolutional neural network super-resolution algorithm incorporating diverse feature extraction to enhance the network's diversity feature learning by elevating the matrix rank, (b) a three-dimensional (3D) feature extraction convolution module, the Channel-Attention-Based Spatial–Spectral Feature Extraction Module (CSSFEM), to boost the network's performance in both the spatial and spectral domains, (c) a feature diversity loss function designed based on the image matrix's singular value to maximize element independence, and (d) a spatial–spectral gradient loss function introduced based on space and spectrum gradient values to enhance the reconstructed image's spatial–spectral smoothness. In contrast to existing hyperspectral super-resolution algorithms, we used four evaluation indexes, PSNR, mPSNR, SSIM, and SAM, and our method showed superiority during testing with three common hyperspectral datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
19. Self-Supervised Convolutional Neural Network Learning in a Hybrid Approach Framework to Estimate Chlorophyll and Nitrogen Content of Maize from Hyperspectral Images.
- Author
-
Gallo, Ignazio, Boschetti, Mirco, Rehman, Anwar Ur, and Candiani, Gabriele
- Subjects
- *
CONVOLUTIONAL neural networks , *BLENDED learning , *MACHINE learning , *SUPERVISED learning , *CHLOROPHYLL - Abstract
The new generation of available (i.e., PRISMA, ENMAP, DESIS) and future (i.e., ESA-CHIME, NASA-SBG) spaceborne hyperspectral missions provide unprecedented data for environmental and agricultural monitoring, such as crop trait assessment. This paper focuses on retrieving two crop traits, specifically Chlorophyll and Nitrogen content at the canopy level (CCC and CNC), starting from hyperspectral images acquired during the CHIME-RCS project, exploiting a self-supervised learning (SSL) technique. SSL is a machine learning paradigm that leverages unlabeled data to generate valuable representations for downstream tasks, bridging the gap between unsupervised and supervised learning. The proposed method comprises pre-training and fine-tuning procedures: in the first stage, a de-noising Convolutional Autoencoder is trained using pairs of noisy and clean CHIME-like images; the pre-trained Encoder network is utilized as-is or fine-tuned in the second stage. The paper demonstrates the applicability of this technique in hybrid approach methods that combine Radiative Transfer Modelling (RTM) and Machine Learning Regression Algorithm (MLRA) to set up a retrieval schema able to estimate crop traits from new generation space-born hyperspectral data. The results showcase excellent prediction accuracy for estimating CCC (R2 = 0.8318; RMSE = 0.2490) and CNC (R2 = 0.9186; RMSE = 0.7908) for maize crops from CHIME-like images without requiring further ground data calibration. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
20. Deep Learning for Earthquake Disaster Assessment: Objects, Data, Models, Stages, Challenges, and Opportunities.
- Author
-
Jia, Jing and Ye, Wenjie
- Subjects
- *
DEEP learning , *CONVOLUTIONAL neural networks , *EARTHQUAKES , *GENERATIVE adversarial networks , *RECURRENT neural networks , *IMAGE recognition (Computer vision) - Abstract
Earthquake Disaster Assessment (EDA) plays a critical role in earthquake disaster prevention, evacuation, and rescue efforts. Deep learning (DL), which boasts advantages in image processing, signal recognition, and object detection, has facilitated scientific research in EDA. This paper analyses 204 articles through a systematic literature review to investigate the status quo, development, and challenges of DL for EDA. The paper first examines the distribution characteristics and trends of the two categories of EDA assessment objects, including earthquakes and secondary disasters as disaster objects, buildings, infrastructure, and areas as physical objects. Next, this study analyses the application distribution, advantages, and disadvantages of the three types of data (remote sensing data, seismic data, and social media data) mainly involved in these studies. Furthermore, the review identifies the characteristics and application of six commonly used DL models in EDA, including convolutional neural network (CNN), multi-layer perceptron (MLP), recurrent neural network (RNN), generative adversarial network (GAN), transfer learning (TL), and hybrid models. The paper also systematically details the application of DL for EDA at different times (i.e., pre-earthquake stage, during-earthquake stage, post-earthquake stage, and multi-stage). We find that the most extensive research in this field involves using CNNs for image classification to detect and assess building damage resulting from earthquakes. Finally, the paper discusses challenges related to training data and DL models, and identifies opportunities in new data sources, multimodal DL, and new concepts. This review provides valuable references for scholars and practitioners in related fields. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
21. Despeckling of SAR Images Using Residual Twin CNN and Multi-Resolution Attention Mechanism.
- Author
-
Pongrac, Blaž and Gleich, Dušan
- Subjects
- *
CONVOLUTIONAL neural networks , *SYNTHETIC aperture radar , *SPECKLE interference - Abstract
The despeckling of synthetic aperture radar images using two different convolutional neural network architectures is presented in this paper. The first method presents a novel Siamese convolutional neural network with a dilated convolutional network in each branch. Recently, attention mechanisms have been introduced to convolutional networks to better model and recognize features. Therefore, we propose a novel design for a convolutional neural network using an attention mechanism for an encoder–decoder-type network. The framework consists of a multiscale spatial attention network to improve the modeling of semantic information at different spatial levels and an additional attention mechanism to optimize feature propagation. Both proposed methods are different in design but they provide comparable despeckling results in subjective and objective measurements in terms of correlated speckle noise. The experimental results are evaluated on both synthetically generated speckled images and real SAR images. The methods proposed in this paper are able to despeckle SAR images and preserve SAR features. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
22. A CatBoost-Based Model for the Intensity Detection of Tropical Cyclones over the Western North Pacific Based on Satellite Cloud Images.
- Author
-
Zhong, Wei, Zhang, Deyuan, Sun, Yuan, and Wang, Qian
- Subjects
- *
TROPICAL cyclones , *REMOTE-sensing images , *CONVOLUTIONAL neural networks , *STANDARD deviations , *BRIGHTNESS temperature - Abstract
A CatBoost-based intelligent tropical cyclone (TC) intensity-detecting model was built to quantify the intensity of TCs over the Western North Pacific (WNP) with the cloud-top brightness temperature (CTBT) data of Fengyun-2F (FY-2F) and Fengyun-2G (FY-2G) and the best-track data of the China Meteorological Administration (CMA-BST) in recent years (2015–2018). The CatBoost-based model was featured with the greedy strategy of combination, the ordering principle in optimizing the possible gradient bias and prediction shift problems, and the oblivious tree in fast scoring. Compared with the previous studies based on the pure convolutional neural network (CNN) models, the CatBoost-based model exhibited better skills in detecting the TC intensity with the root mean square error (RMSE) of 3.74 m s−1. In addition to the three mentioned model features, there are also two reasons for the model's design. On one hand, the CatBoost-based model used the method of introducing prior physical factors (e.g., the structure and shape of the cloud, deep convections, and background fields) into its training process. On the other hand, the CatBoost-based model expanded the dataset size from 2342 to 13,471 samples through hourly interpolations of the original dataset. Furthermore, this paper investigated the errors of this model in detecting the different categories of TC intensity. The results showed that the deep learning-based TC intensity-detecting model proposed in this paper has systematic biases, namely, the overestimation (underestimation) of intensities in TCs which were weaker (stronger) than at the typhoon level, and the errors of the model in detecting weaker (stronger) TCs were smaller (larger). This implies that more factors than the CTBT should be included to further reduce the errors in detecting strong TCs. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
23. HyperSFormer: A Transformer-Based End-to-End Hyperspectral Image Classification Method for Crop Classification.
- Author
-
Xie, Jiaxing, Hua, Jiajun, Chen, Shaonan, Wu, Peiwen, Gao, Peng, Sun, Daozong, Lyu, Zhendong, Lyu, Shilei, Xue, Xiuyun, and Lu, Jianqiang
- Subjects
- *
IMAGE recognition (Computer vision) , *TRANSFORMER models , *RECURRENT neural networks , *CONVOLUTIONAL neural networks , *CROP yields - Abstract
Crop classification of large-scale agricultural land is crucial for crop monitoring and yield estimation. Hyperspectral image classification has proven to be an effective method for this task. Most current popular hyperspectral image classification methods are based on image classification, specifically on convolutional neural networks (CNNs) and recurrent neural networks (RNNs). In contrast, this paper focuses on methods based on semantic segmentation and proposes a new transformer-based approach called HyperSFormer for crop hyperspectral image classification. The key enhancement of the proposed method is the replacement of the encoder in SegFormer with an improved Swin Transformer while keeping the SegFormer decoder. The entire model adopts a simple and uniform transformer architecture. Additionally, the paper introduces the hyper patch embedding (HPE) module to extract spectral and local spatial information from the hyperspectral images, which enhances the effectiveness of the features used as input for the model. To ensure detailed model processing and achieve end-to-end hyperspectral image classification, the transpose padding upsample (TPU) module is proposed for the model's output. In order to address the problem of insufficient and imbalanced samples in hyperspectral image classification, the paper designs an adaptive min log sampling (AMLS) strategy and a loss function that incorporates dice loss and focal loss to assist model training. Experimental results using three public hyperspectral image datasets demonstrate the strong performance of HyperSFormer, particularly in the presence of imbalanced sample data, complex negative samples, and mixed sample classes. HyperSFormer outperforms state-of-the-art methods, including fast patch-free global learning (FPGA), a spectral–spatial-dependent global learning framework (SSDGL), and SegFormer, by at least 2.7% in the mean intersection over union (mIoU). It also improves the overall accuracy and average accuracy values by at least 0.9% and 0.3%, respectively, and the kappa coefficient by at least 0.011. Furthermore, ablation experiments were conducted to determine the optimal hyperparameter and loss function settings for the proposed method, validating the rationality of these settings and the fusion loss function. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
24. DAFCNN: A Dual-Channel Feature Extraction and Attention Feature Fusion Convolution Neural Network for SAR Image and MS Image Fusion.
- Author
-
Luo, Jiahao, Zhou, Fang, Yang, Jun, and Xing, Mengdao
- Subjects
- *
IMAGE fusion , *DEEP learning , *CONVOLUTIONAL neural networks , *FEATURE extraction , *MACHINE learning , *SYNTHETIC aperture radar , *SPATIAL ability - Abstract
In the field of image fusion, spatial detail blurring and color distortion appear in synthetic aperture radar (SAR) images and multispectral (MS) during the traditional fusion process due to the difference in sensor imaging mechanisms. To solve this problem, this paper proposes a fusion method for SAR images and MS images based on a convolutional neural network. In order to make use of the spatial information and different scale feature information of high-resolution SAR image, a dual-channel feature extraction module is constructed to obtain a SAR image feature map. In addition, different from the common direct addition strategy, an attention-based feature fusion module is designed to achieve spectral fidelity of the fused images. In order to obtain better spectral and spatial retention ability of the network, an unsupervised joint loss function is designed to train the network. In this paper, the Sentinel 1 SAR images and Landsat 8 MS images are used as datasets for experiments. The experimental results show that the proposed algorithm has better performance in quantitative and visual representation when compared with traditional fusion methods and deep learning algorithms. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
25. Spatiotemporal Prediction of Ionospheric Total Electron Content Based on ED-ConvLSTM.
- Author
-
Li, Liangchao, Liu, Haijun, Le, Huijun, Yuan, Jing, Shan, Weifeng, Han, Ying, Yuan, Guoming, Cui, Chunjie, and Wang, Junling
- Subjects
- *
RECURRENT neural networks , *GLOBAL Positioning System , *CONVOLUTIONAL neural networks , *MAGNETIC storms , *DEEP learning , *PREDICTION models - Abstract
Total electron content (TEC) is a vital parameter for describing the state of the ionosphere, and precise prediction of TEC is of great significance for improving the accuracy of the Global Navigation Satellite System (GNSS). At present, most deep learning prediction models just consider TEC temporal variation, while ignoring the impact of spatial location. In this paper, we propose a TEC prediction model, ED-ConvLSTM, which combines convolutional neural networks with recurrent neural networks to simultaneously consider spatiotemporal features. Our ED-ConvLSTM model is built based on the encoder-decoder architecture, which includes two modules: encoder module and decoder module. Each module is composed of ConvLSTM cells. The encoder module is used to extract the spatiotemporal features from TEC maps, while the decoder module converts spatiotemporal features into predicted TEC maps. We compared the predictive performance of our model with two traditional time series models: LSTM, GRU, a spatiotemporal mode1 ConvGRU, and the TEC daily forecast product C1PG provided by CODE on a total of 135 grid points in East Asia (10°–45°N, 90°–130°E). The experimental results show that the prediction error indicators MAE, RMSE, MAPE, and prediction similarity index SSIM of our model are superior to those of the comparison models in high, normal, and low solar activity years. The paper also analyzed the predictive performance of each model monthly. The experimental results indicate that the predictive performance of each model is influenced by the monthly mean of TEC. The ED-ConvLSTM model proposed in this paper is the least affected and the most stable by the monthly mean of TEC. Additionally, the paper compared the predictive performance of each model during two magnetic storm periods when TEC changes sharply. The results indicate that our ED-ConvLSTM model is least affected during magnetic storms and its predictive performance is superior to those of the comparative models. This paper provides a more stable and high-performance TEC spatiotemporal prediction model. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
26. A Review of Hyperspectral Image Super-Resolution Based on Deep Learning.
- Author
-
Chen, Chi, Wang, Yongcheng, Zhang, Ning, Zhang, Yuxi, and Zhao, Zhikang
- Subjects
- *
DEEP learning , *HIGH resolution imaging , *COMPUTER vision , *GENERATIVE adversarial networks , *CONVOLUTIONAL neural networks - Abstract
Hyperspectral image (HSI) super-resolution (SR) is a classical computer vision task that aims to accomplish the conversion of images from lower to higher resolutions. With the booming development of deep learning (DL) technology, more and more researchers are dedicated to the research of image SR techniques based on DL and have made remarkable progress. However, no scholar has provided a comprehensive review of the field. As a response, in this paper we aim to supply a comprehensive summary of the DL-based SR techniques for HSI, including upsampling frameworks, upsampling methods, network design, loss functions, representative works with different strategies, and future directions, in which we design several sets of comparative experiments for the advantages and limitations of two-dimensional convolution and three-dimensional convolution in the field of HSI SR and analyze the experimental results in depth. In addition, the paper also briefly discusses the secondary foci such as common datasets, evaluation metrics, and traditional SR algorithms. To the best of our knowledge, this paper is the first review on DL-based HSI SR. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
27. Cross-Hole GPR for Soil Moisture Estimation Using Deep Learning.
- Author
-
Pongrac, Blaž, Gleich, Dušan, Malajner, Marko, and Sarjaš, Andrej
- Subjects
- *
SOIL moisture , *DEEP learning , *SOIL moisture measurement , *TRANSMITTING antennas , *CONVOLUTIONAL neural networks , *ANTENNAS (Electronics) - Abstract
This paper presents the design of a high-voltage pulse-based radar and a supervised data processing method for soil moisture estimation. The goal of this research was to design a pulse-based radar to detect changes in soil moisture using a cross-hole approach. The pulse-based radar with three transmitting antennas was placed into a 12 m deep hole, and a receiver with three receive antennas was placed into a different hole separated by 100 m from the transmitter. The pulse generator was based on a Marx generator with an LC filter, and for the receiver, the high-frequency data acquisition card was used, which can acquire signals using 3 Gigabytes per second. Used borehole antennas were designed to operate in the wide frequency band to ensure signal propagation through the soil. A deep regression convolutional network is proposed in this paper to estimate volumetric soil moisture using time-sampled signals. A regression convolutional network is extended to three dimensions to model changes in wave propagation between the transmitted and received signals. The training dataset was acquired during the period of 73 days of acquisition between two boreholes separated by 100 m. The soil moisture measurements were acquired at three points 25 m apart to provide ground truth data. Additionally, water was poured into several specially prepared boreholes between transmitter and receiver antennas to acquire additional dataset for training, validation, and testing of convolutional neural networks. Experimental results showed that the proposed system is able to detect changes in the volumetric soil moisture using Tx and Rx antennas. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
28. Intelligent Environment-Adaptive GNSS/INS Integrated Positioning with Factor Graph Optimization.
- Author
-
Li, Zhengdao, Lee, Pin-Hsun, Hung, Tsz Hin Marcus, Zhang, Guohao, and Hsu, Li-Ta
- Subjects
- *
GLOBAL Positioning System , *CONVOLUTIONAL neural networks , *DEEP learning , *INTELLIGENT transportation systems , *STANDARD deviations , *INERTIAL navigation systems - Abstract
Global navigation satellite systems (GNSSs) applied to intelligent transport systems in urban areas suffer from multipath and non-line-of-sight (NLOS) effects due to the signal reflections from high-rise buildings, which seriously degrade the accuracy and reliability of vehicles in real-time applications. Accordingly, the integration between GNSS and inertial navigation systems (INSs) could be utilized to improve positioning performance. However, the fixed GNSS solution uncertainty of the conventional integration method cannot determine the fluctuating GNSS reliability in fast-changing urban environments. This weakness becomes solvable using a deep learning model for sensing the ambient environment intelligently, and it can be further mitigated using factor graph optimization (FGO), which is capable of generating robust solutions based on historical data. This paper mainly develops the adaptive GNSS/INS loosely coupled system on FGO, along with the fixed-gain Kalman filter (KF) and adaptive KF (AKF) being taken as comparisons. The adaptation is aided by a convolutional neural network (CNN), and the feasibility is verified using data from different grades of receivers. Compared with the integration using fixed-gain KF, the proposed adaptive FGO (AFGO) maintains the 100% positioning availability and reduces the overall 2D positioning error by up to 70% in the aspects of both root mean square error (RMSE) and standard deviation (STD). [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
29. Self-Adaptive-Filling Deep Convolutional Neural Network Classification Method for Mountain Vegetation Type Based on High Spatial Resolution Aerial Images.
- Author
-
Li, Shiou, Fei, Xianyun, Chen, Peilong, Wang, Zhen, Gao, Yajun, Cheng, Kai, Wang, Huilong, and Zhang, Yuanzhi
- Subjects
- *
CONVOLUTIONAL neural networks , *MOUNTAIN plants , *VEGETATION classification , *SPATIAL resolution , *REMOTE sensing - Abstract
The composition and structure of mountain vegetation are complex and changeable, and thus urgently require the integration of Object-Based Image Analysis (OBIA) and Deep Convolutional Neural Networks (DCNNs). However, while integration technology studies are continuing to increase, there have been few studies that have carried out the classification of mountain vegetation by combining OBIA and DCNNs, for it is difficult to obtain enough samples to trigger the potential of DCNNs for mountain vegetation type classification, especially using high-spatial-resolution remote sensing images. To address this issue, we propose a self-adaptive-filling method (SAF) to incorporate the OBIA method to improve the performance of DCNNs in mountain vegetation type classification using high-spatial-resolution aerial images. Using this method, SAF technology was employed to produce enough regular sample data for DCNNs by filling the irregular objects created by image segmenting using interior adaptive pixel blocks. Meanwhile, non-sample segmented image objects were shaped into different regular rectangular blocks via SAF. Then, the classification result was defined by voting combining the DCNN performance. Compared to traditional OBIA methods, SAF generates more samples for the DCNN and fully utilizes every single pixel of the DCNN input. We design experiments to compare them with traditional OBIA and semantic segmentation methods, such as U-net, MACU-net, and SegNeXt. The results show that our SAF-DCNN outperforms traditional OBIA in terms of accuracy and it is similar to the accuracy of the best performing method in semantic segmentation. However, it reduces the common pretzel phenomenon of semantic segmentation (black and white noise generated in classification). Overall, the SAF-based OBIA using DCNNs, which is proposed in this paper, is superior to other commonly used methods for vegetation classification in mountainous areas. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
30. Your Input Matters—Comparing Real-Valued PolSAR Data Representations for CNN-Based Segmentation.
- Author
-
Hochstuhl, Sylvia, Pfeffer, Niklas, Thiele, Antje, Hammer, Horst, and Hinz, Stefan
- Subjects
- *
DEEP learning , *CONVOLUTIONAL neural networks , *SYNTHETIC aperture radar , *IMAGE segmentation , *FEATURE selection , *OPTICAL images - Abstract
Inspired by the success of Convolutional Neural Network (CNN)-based deep learning methods for optical image segmentation, there is a growing interest in applying these methods to Polarimetric Synthetic Aperture Radar (PolSAR) data. However, effectively utilizing well-established real-valued CNNs for PolSAR image segmentation requires converting complex-valued data into real-valued representations. This paper presents a systematic comparison of 14 different real-valued representations used as CNN input in the literature. These representations encompass various approaches, including the use of coherency matrix elements, hand-crafted feature vectors, polarimetric features based on target decomposition, and combinations of these methods. The goal is to assess the impact of the choice of PolSAR data representation on segmentation performance and identify the most suitable representation. Four test configurations are employed to achieve this, involving different CNN architectures (U-Net with ResNet-18 or EfficientNet backbone) and PolSAR data acquired in different frequency bands (S- and L-band). The results emphasize the importance of selecting an appropriate real-valued representation for CNN-based PolSAR image segmentation. This study's findings reveal that combining multiple polarimetric features can potentially enhance segmentation performance but does not consistently improve the results. Therefore, when employing this approach, careful feature selection becomes crucial. In contrast, using coherency matrix elements with amplitude and phase representation consistently achieves high segmentation performance across different test configurations. This representation emerges as one of the most suitable approaches for CNN-based PolSAR image segmentation. Notably, it outperforms the commonly used alternative approach of splitting the coherency matrix elements into real and imaginary parts. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
31. Geological Hazard Identification and Susceptibility Assessment Based on MT-InSAR.
- Author
-
Lu, Zhaowei, Yang, Honglei, Zeng, Wei, Liu, Peng, and Wang, Yuedong
- Subjects
- *
CONVOLUTIONAL neural networks , *RECURRENT neural networks , *SYNTHETIC aperture radar , *DEFORMATION of surfaces , *SYNTHETIC apertures , *SUPPORT vector machines - Abstract
Geological hazards often occur in mountainous areas and are sudden and hidden, so it is important to identify and assess geological hazards. In this paper, the western mountainous area of Beijing was selected as the study area. We conducted research on landslides, collapses, and unstable slopes in the study area. The surface deformation of the study area was monitored by multi-temporal interferometric synthetic aperture radar (MT-InSAR), using a combination of multi-looking point selection and permanent scatterer (PS) point selection methods. Random forest (RF), support vector machine (SVM), convolutional neural network (CNN), and recurrent neural network (RNN) models were selected for the assessment of geological hazard susceptibility. Sixteen geological hazard-influencing factors were collected, and their information values were calculated using their features. Multicollinearity analysis with the relief-F method was used to calculate the correlation and importance of the factors for factor selection. The results show that the deformation rate along the line-of-sight (LOS) direction is between −44 mm/year and 28 mm/year. A total of 60 geological hazards were identified by combining surface deformation with optical imagery and other data, including 7 collapses, 25 unstable slopes, and 28 landslides. Forty-eight of the identified geological hazards are not recorded in the Beijing geological hazards list. The most effective model in the study area was RF. The percentage of geological hazard susceptibility zoning in the study area is as follows: very low susceptibility 27.40%, low susceptibility 28.06%, moderate susceptibility 21.19%, high susceptibility 13.80%, very high susceptibility 9.57%. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
32. LRTransDet: A Real-Time SAR Ship-Detection Network with Lightweight ViT and Multi-Scale Feature Fusion.
- Author
-
Feng, Kunyu, Lun, Li, Wang, Xiaofeng, and Cui, Xiaoxin
- Subjects
- *
TRANSFORMER models , *SYNTHETIC aperture radar , *CONVOLUTIONAL neural networks , *DEEP learning - Abstract
In recent years, significant strides have been made in the field of synthetic aperture radar (SAR) ship detection through the application of deep learning techniques. These advanced methods have substantially improved the accuracy of ship detection. Nonetheless, SAR images present distinct challenges, including complex backgrounds, small ship targets, and noise interference, thereby rendering the detectors particularly demanding. In this paper, we introduce LRTransDet, a real-time SAR ship detector. LRTransDet leverages a lightweight vision transformer (ViT) and a multi-scale feature fusion neck to address these challenges effectively. First, our model implements a lightweight backbone that combines convolutional neural networks (CNNs) and transformers, thus enabling it to simultaneously capture both local and global features from input SAR images. Moreover, we boost the model's efficiency by incorporating the faster weighted feature fusion (Faster-WF2) module and coordinate attention (CA) mechanism within the feature fusion neck. These components optimize computational resources while maintaining the model's performance. To overcome the challenge of detecting small ship targets in SAR images, we refine the original loss function and use the normalized Wasserstein distance (NWD) metric and the intersection over union (IoU) scheme. This combination improves the detector's ability to efficiently detect small targets. To prove the performance of our proposed model, we conducted experiments on four challenging datasets (the SSDD, the SAR-Ship Dataset, the HRSID, and the LS-SSDD-v1.0). The results demonstrate that our model surpasses both general object detectors and state-of-the-art SAR ship detectors in terms of detection accuracy (97.8% on the SSDD and 93.9% on the HRSID) and speed (74.6 FPS on the SSDD and 75.8 FPS on the HRSID), all while demanding 3.07 M parameters. Additionally, we conducted a series of ablation experiments to illustrate the impact of the EfficientViT, the Faster-WF2 module, the CA mechanism, and the NWD metric on multi-scale feature fusion and detection performance. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
33. Completing 3D Point Clouds of Thin Corn Leaves for Phenotyping Using 3D Gridding Convolutional Neural Networks.
- Author
-
Zhang, Ying, Su, Wei, Tao, Wancheng, Li, Ziqian, Huang, Xianda, Zhang, Ziyue, and Xiong, Caisen
- Subjects
- *
CONVOLUTIONAL neural networks , *POINT cloud , *CROPS - Abstract
Estimating the complete 3D points of crop plants from incomplete points is vital for phenotyping and smart agriculture management. Compared with the completion of regular man-made objects such as airplanes, chairs, and desks, the completion of corn plant points is more difficult for thin, curled, and irregular corn leaves. This study focuses on MSGRNet+OA, which is based on GRNet, to complete a 3D point cloud of thin corn plants. The developed MSGRNet+OA was accompanied by gridding, multi-scale 3DCNN, gridding reverse, cubic feature sampling, and offset-attention. In this paper, we propose the introduction of a 3D grid as an intermediate representation to regularize the unorganized point cloud, use multi-scale predictive fusion to utilize global information at different scales, and model the geometric features by adding offset-attention to compute the point position offsets. These techniques enable the network to exhibit good adaptability and robustness in dealing with irregular and varying point cloud structures. The accuracy assessment results show that the accuracy of completion using MSGRNet+OA is superlative, with a CD (×10−4) of 1.258 and an F-Score@1% of 0.843. MSGRNet+OA is the most effective when compared with other networks (PCN, shape inversion, the original GRNet, SeedFormer, and PMP-Net++), and it improves the accuracy of the CD (×10−4)/F-Score@1% with −15.882/0.404, −15.96/0.450, −0.181/0.018, −1.852/0.274, and −1.471/0.203, respectively. These results reveal that the developed MSGRNet+OA can be used to complete a 3D point cloud of thin corn leaves for phenotyping. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
34. A Comprehensive Survey on SAR ATR in Deep-Learning Era.
- Author
-
Li, Jianwei, Yu, Zhentao, Yu, Lu, Cheng, Pu, Chen, Jie, and Chi, Cheng
- Subjects
- *
DEEP learning , *CONVOLUTIONAL neural networks , *SUPERVISED learning , *GENERATIVE adversarial networks , *AUTOMATIC target recognition , *DATA augmentation - Abstract
Due to the advantages of Synthetic Aperture Radar (SAR), the study of Automatic Target Recognition (ATR) has become a hot topic. Deep learning, especially in the case of a Convolutional Neural Network (CNN), works in an end-to-end way and has powerful feature-extracting abilities. Thus, researchers in SAR ATR also seek solutions from deep learning. We review the related algorithms with regard to SAR ATR in this paper. We firstly introduce the commonly used datasets and the evaluation metrics. Then, we introduce the algorithms before deep learning. They are template-matching-, machine-learning- and model-based methods. After that, we introduce mainly the SAR ATR methods in the deep-learning era (after 2017); those methods are the core of the paper. The non-CNNs and CNNs, that is, those used in SAR ATR, are summarized at the beginning. We found that researchers tend to design specialized CNN for SAR ATR. Then, the methods to solve the problem raised by limited samples are reviewed. They are data augmentation, Generative Adversarial Networks (GAN), electromagnetic simulation, transfer learning, few-shot learning, semi-supervised learning, metric leaning and domain knowledge. After that, the imbalance problem, real-time recognition, polarimetric SAR, complex data and adversarial attack are also reviewed. The principles and problems of them are also introduced. Finally, the future directions are conducted. In this part, we point out that the dataset, CNN architecture designing, knowledge-driven, real-time recognition, explainable and adversarial attack should be considered in the future. This paper gives readers a quick overview of the current state of the field. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
35. A Multilevel Spatial and Spectral Feature Extraction Network for Marine Oil Spill Monitoring Using Airborne Hyperspectral Image.
- Author
-
Wang, Jian, Li, Zhongwei, Yang, Junfang, Liu, Shanwei, Zhang, Jie, and Li, Shibao
- Subjects
- *
OIL spills , *CONVOLUTIONAL neural networks , *FEATURE extraction , *MULTISPECTRAL imaging , *ARTIFICIAL neural networks , *REMOTE sensing - Abstract
Marine oil spills can cause serious damage to marine ecosystems and biological species, and the pollution is difficult to repair in the short term. Accurate oil type identification and oil thickness quantification are of great significance for marine oil spill emergency response and damage assessment. In recent years, hyperspectral remote sensing technology has become an effective means to monitor marine oil spills. The spectral and spatial features of oil spill images at different levels are different. To accurately identify oil spill types and quantify oil film thickness, and perform better extraction of spectral and spatial features, a multilevel spatial and spectral feature extraction network is proposed in this study. First, the graph convolutional neural network and graph attentional neural network models were used to extract spectral and spatial features in non-Euclidean space, respectively, and then the designed modules based on 2D expansion convolution, depth convolution, and point convolution were applied to extract feature information in Euclidean space; after that, a multilevel feature fusion method was developed to fuse the obtained spatial and spectral features in Euclidean space in a complementary way to obtain multilevel features. Finally, the multilevel features were fused at the feature level to obtain the oil spill information. The experimental results show that compared with CGCNN, SSRN, and A2S2KResNet algorithms, the accuracy of oil type identification and oil film thickness classification of the proposed method in this paper is improved by 12.82%, 0.06%, and 0.08% and 2.23%, 0.69%, and 0.47%, respectively, which proves that the method in this paper can effectively extract oil spill information and identify different oil spill types and different oil film thicknesses. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
36. SquconvNet: Deep Sequencer Convolutional Network for Hyperspectral Image Classification.
- Author
-
Li, Bing, Wang, Qi-Wen, Liang, Jia-Hong, Zhu, En-Ze, and Zhou, Rong-Qian
- Subjects
- *
DEEP learning , *CONVOLUTIONAL neural networks , *COMPUTER vision - Abstract
The application of Transformer in computer vision has had the most significant influence of all the deep learning developments over the past five years. In addition to the exceptional performance of convolutional neural networks (CNN) in hyperspectral image (HSI) classification, Transformer has begun to be applied to HSI classification. However, for the time being, Transformer has not produced satisfactory results in HSI classification. Recently, in the field of image classification, the creators of Sequencer have proposed a Sequencer structure that substitutes the Transformer self-attention layer with a BiLSTM2D layer and achieves satisfactory results. As a result, this paper proposes a unique network called SquconvNet, that combines CNN with Sequencer block to improve hyperspectral classification. In this paper, we conducted rigorous HSI classification experiments on three relevant baseline datasets to evaluate the performance of the proposed method. The experimental results show that our proposed method has clear advantages in terms of classification accuracy and stability. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
37. Cloud Removal from Satellite Images Using a Deep Learning Model with the Cloud-Matting Method.
- Author
-
Ma, Deying, Wu, Renzhe, Xiao, Dongsheng, and Sui, Baikai
- Subjects
- *
REMOTE-sensing images , *DEEP learning , *CONTROL groups , *OPTICAL remote sensing , *CONVOLUTIONAL neural networks , *OPTICAL limiting - Abstract
Clouds seriously limit the application of optical remote sensing images. In this paper, we remove clouds from satellite images using a novel method that considers ground surface reflections and cloud top reflections as a linear mixture of image elements from the perspective of image superposition. We use a two-step convolutional neural network to extract the transparency information of clouds and then recover the ground surface information of thin cloud regions. Given the poor balance of the generated samples, this paper also improves the binary Tversky loss function and applies it on multi-classification tasks. The model was validated on the simulated dataset and ALCD dataset, respectively. The results show that this model outperformed other control group experiments in cloud detection and removal. The model better locates the clouds in images with cloud matting, which is built based on cloud detection. In addition, the model successfully recovers the surface information of the thin cloud region when thick and thin clouds coexist, and it does not damage the original image's information. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
38. Acoustic Impedance Inversion from Seismic Imaging Profiles Using Self Attention U-Net.
- Author
-
Tao, Liurong, Ren, Haoran, and Gu, Zhiwei
- Subjects
- *
ACOUSTIC impedance , *IMAGING systems in seismology , *CONVOLUTIONAL neural networks , *INVERSION (Geophysics) , *INVERSE problems , *DEEP learning , *NONLINEAR equations - Abstract
Seismic impedance inversion is a vital way of geological interpretation and reservoir investigation from a geophysical perspective. However, it is inevitably an ill-posed problem due to the noise or the band-limited characteristic of seismic data. Artificial neural network have been used to solve nonlinear inverse problems in recent years. This research obtained an acoustic impedance profile by feeding seismic profile and background impedance into a well-trained self-attention U-Net. The U-Net got convergence by appropriate iteration, and the output predicted the impedance profiles in the test. To value the quality of predicted profiles from different perspectives, e.g., correlation, regression, and similarity, we used four kinds of indexes. At the same time, our results were predicted by conventional methods (e.g., deconvolution with recursive inversion, and TV regularization) and a 1D neural network was calculated in contrast. Self-attention U-Net showed to be robust to noise and does not require prior knowledge. Furthermore, spatial continuity is also better than deconvolution, regularization, and 1D deep learning methods in contrast. The U-Net in this paper is a type of full convolutional neural network, so there are no limits to the shape of the input. Based on this, a large impedance profile can be predicted by U-Net, which is trained by a patchy training dataset. In addition, this paper applied the proposed method to the field data obtained by the Ceduna survey without any label. The predictions prove that this well-trained network could be generalized from synthetic data to field data. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
39. Remote Sensing Crop Recognition by Coupling Phenological Features and Off-Center Bayesian Deep Learning.
- Author
-
Wu, Yongchuang, Wu, Penghai, Wu, Yanlan, Yang, Hui, and Wang, Biao
- Subjects
- *
REMOTE sensing , *DEEP learning , *RECURRENT neural networks , *CONVOLUTIONAL neural networks , *AREA measurement - Abstract
Obtaining accurate and timely crop area information is crucial for crop yield estimates and food security. Because most existing crop mapping models based on remote sensing data have poor generalizability, they cannot be rapidly deployed for crop identification tasks in different regions. Based on a priori knowledge of phenology, we designed an off-center Bayesian deep learning remote sensing crop classification method that can highlight phenological features, combined with an attention mechanism and residual connectivity. In this paper, we first optimize the input image and input features based on a phenology analysis. Then, a convolutional neural network (CNN), recurrent neural network (RNN), and random forest classifier (RFC) were built based on farm data in northeastern Inner Mongolia and applied to perform comparisons with the method proposed here. Then, classification tests were performed on soybean, maize, and rice from four measurement areas in northeastern China to verify the accuracy of the above methods. To further explore the reliability of the method proposed in this paper, an uncertainty analysis was conducted by Bayesian deep learning to analyze the model's learning process and model structure for interpretability. Finally, statistical data collected in Suibin County, Heilongjiang Province, over many years, and Shandong Province in 2020 were used as reference data to verify the applicability of the methods. The experimental results show that the classification accuracy of the three crops reached 90.73% overall and the average F1 and IOU were 89.57% and 81.48%, respectively. Furthermore, the proposed method can be directly applied to crop area estimations in different years in other regions based on its good correlation with official statistics. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
40. GSDerainNet: A Deep Network Architecture Based on a Gaussian Shannon Filter for Single Image Deraining.
- Author
-
Yao, Yanji, Shi, Zhimin, Hu, Huiwen, Li, Jing, Wang, Guocheng, and Liu, Lintao
- Subjects
- *
CONVOLUTIONAL neural networks , *RAINFALL , *SEVERE storms , *DEEP learning , *WEATHER - Abstract
With the continuous advancement of target detection technology in remote sensing, target detection technology in images captured by drones has performed well. However, object detection in drone imagery is still a challenge under rainy conditions. Rain is a common severe weather condition, and rain streaks often degrade the image quality of sensors. The main issue of rain streaks removal from a single image is to prevent over smoothing (or underclearing) phenomena. Aiming at the above problems, this paper proposes a deep learning (DL)-based rain streaks removal framework called GSDerainNet, which properly formulates the single image rain streaks removal problem; rain streaks removal is aptly described as a Gaussian Shannon (GS) filter-based image decomposition problem. The GS filter is a novel filter proposed by us, which consists of a parameterized Gaussian function and a scaled Shannon function. Two-dimensional GS filters exhibit high stability and effectiveness in dividing an image into low- and high-frequency parts. In our framework, an input image is first decomposed into a low-frequency part and a high-frequency part by using the GS filter. Rain streaks are located in the high-frequency part. We extract and separate the rain features of the high-frequency part through a deep convolutional neural network (CNN). The experimental results obtained on synthetic data and real data show that the proposed method can better suppress the morphological artifacts caused by filtering. Compared with state-of-the-art single image rain streaks removal methods, the proposed method retains finer image object structures while removing rain streaks. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
41. A New Architecture of a Complex-Valued Convolutional Neural Network for PolSAR Image Classification.
- Author
-
Ren, Yihui, Jiang, Wen, and Liu, Ying
- Subjects
- *
CONVOLUTIONAL neural networks , *IMAGE recognition (Computer vision) , *DEEP learning , *SYNTHETIC aperture radar , *REAL numbers , *COMPLEX numbers , *SYNTHETIC apertures , *NONLINEAR oscillators - Abstract
Polarimetric synthetic aperture radar (PolSAR) image classification has been an important area of research due to its wide range of applications. Traditional machine learning methods were insufficient in achieving satisfactory results before the advent of deep learning. Results have significantly improved with the widespread use of deep learning in PolSAR image classification. However, the challenge of reconciling the complex-valued inputs of PolSAR images with the real-valued models of deep learning remains unsolved. Current complex-valued deep learning models treat complex numbers as two distinct real numbers, providing limited assistance in PolSAR image classification results. This paper proposes a novel, complex-valued deep learning approach for PolSAR image classification to address this issue. The approach includes amplitude-based max pooling, complex-valued nonlinear activation, and a cross-entropy loss function based on complex-valued probability. Amplitude-based max pooling reduces computational effort while preserving the most valuable complex-valued features. Complex-valued nonlinear activation maps feature into a high-dimensional complex-domain space, producing the most discriminative features. The complex-valued cross-entropy loss function computes the classification loss using the complex-valued model output and dataset labels, resulting in more accurate and robust classification results. The proposed method was applied to a shallow CNN, deep CNN, FCN, and SegNet, and its effectiveness was verified on three public datasets. The results showed that the method achieved optimal classification results on any model and dataset. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
42. Deep Learning for Remote Sensing Image Scene Classification: A Review and Meta-Analysis.
- Author
-
Thapa, Aakash, Horanont, Teerayut, Neupane, Bipul, and Aryal, Jagannath
- Subjects
- *
DEEP learning , *IMAGE recognition (Computer vision) , *CONVOLUTIONAL neural networks , *GENERATIVE adversarial networks , *TRANSFORMER models , *DISTANCE education , *REMOTE sensing - Abstract
Remote sensing image scene classification with deep learning (DL) is a rapidly growing field that has gained significant attention in the past few years. While previous review papers in this domain have been confined to 2020, an up-to-date review to show the progression of research extending into the present phase is lacking. In this review, we explore the recent articles, providing a thorough classification of approaches into three main categories: Convolutional Neural Network (CNN)-based, Vision Transformer (ViT)-based, and Generative Adversarial Network (GAN)-based architectures. Notably, within the CNN-based category, we further refine the classification based on specific methodologies and techniques employed. In addition, a novel and rigorous meta-analysis is performed to synthesize and analyze the findings from 50 peer-reviewed journal articles to provide valuable insights in this domain, surpassing the scope of existing review articles. Our meta-analysis shows that the most adopted remote sensing scene datasets are AID (41 articles) and NWPU-RESISC45 (40). A notable paradigm shift is seen towards the use of transformer-based models (6) starting from 2021. Furthermore, we critically discuss the findings from the review and meta-analysis, identifying challenges and future opportunities for improvement in this domain. Our up-to-date study serves as an invaluable resource for researchers seeking to contribute to this growing area of research. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
43. Exploiting Temporal–Spatial Feature Correlations for Sequential Spacecraft Depth Completion.
- Author
-
Liu, Xiang, Wang, Hongyuan, Chen, Xinlong, Chen, Weichun, and Xie, Zhengyou
- Subjects
- *
CONVOLUTIONAL neural networks , *OPTICAL images - Abstract
The recently proposed spacecraft three-dimensional (3D) structure recovery method based on optical images and LIDAR has enhanced the working distance of a spacecraft's 3D perception system. However, the existing methods ignore the richness of temporal features and fail to capture the temporal coherence of consecutive frames. This paper proposes a sequential spacecraft depth completion network (S2DCNet) for generating accurate and temporally consistent depth prediction results, and it can fully exploit temporal–spatial coherence in sequential frames. Specifically, two parallel convolution neural network (CNN) branches were first adopted to extract the features latent in different inputs. The gray image features and the depth features were hierarchically encapsulated into unified feature representations through fusion modules. In the decoding stage, the convolutional long short-term memory (ConvLSTM) networks were embedded with the multi-scale scheme to capture the feature spatial–temporal distribution variation, which could reflect the past state and generate more accurate and temporally consistent depth maps. In addition, a large-scale dataset was constructed, and the experiments revealed the outstanding performance of the proposed S2DCNet, achieving a mean absolute error of 0.192 m within the region of interest. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
44. Polarimetric Synthetic Aperture Radar Image Semantic Segmentation Network with Lovász-Softmax Loss Optimization.
- Author
-
Guo, Rui, Zhao, Xiaopeng, Zuo, Guanzhong, Wang, Ying, and Liang, Yi
- Subjects
- *
SYNTHETIC aperture radar , *IMAGE segmentation , *MICROWAVE remote sensing , *SYNTHETIC apertures , *CONVOLUTIONAL neural networks , *DEEP learning - Abstract
The deep learning technique has already been successfully applied in the field of microwave remote sensing. Especially, convolutional neural networks have demonstrated remarkable effectiveness in synthetic aperture radar (SAR) image semantic segmentation. In this paper, a Lovász-softmax loss optimization SAR net (LoSARNet) is proposed which optimizes the semantic segmentation metric intersection over union (IOU) instead of using the traditional cross-entropy loss. Meanwhile, making use of the advantages of the dual-path structure, the network extracts feature through the spatial path (SP) and the context path (CP) to achieve a balance between efficiency and accuracy. Aiming at a polarimetric SAR (PolSAR) image, the proposed network is conducted on the PolSAR datasets for terrain segmentation. Compared to the typical dual-path network, which is the bilateral segmentation network (BiSeNet), the proposed LoSARNet can obtain better mean intersection over union (MIOU). And the proposed network also shows the highest evaluation index and the best performance when compared with several typical networks. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
45. A Fourier Frequency Domain Convolutional Neural Network for Remote Sensing Crop Classification Considering Global Consistency and Edge Specificity.
- Author
-
Song, Binbin, Min, Songhan, Yang, Hui, Wu, Yongchuang, and Wang, Biao
- Subjects
- *
CONVOLUTIONAL neural networks , *REMOTE sensing , *HEBBIAN memory , *CROPS , *DEEP learning , *CROP growth - Abstract
The complex remote sensing image acquisition conditions and the differences in crop growth create many crop classification challenges. Frequency decomposition enables the capture of the feature information in an image that is difficult to discern. Frequency domain filters can strengthen or weaken specific frequency components to enhance the interclass differences among the different crops and can reduce the intraclass variations within the same crops, thereby improving crop classification accuracy. In concurrence with the Fourier frequency domain learning strategy, we propose a convolutional neural network called the Fourier frequency domain convolutional (FFDC) net, which transforms feature maps from the spatial domain to the frequency spectral domain. In this network, the dynamic frequency filtering components in the frequency spectral domain are used to separate the feature maps into low-frequency and high-frequency components, and the strength and distribution of the different frequency components are automatically adjusted to suppress the low-frequency information variations within the same crop, enhancing the overall consistency of the crops. Simultaneously, it is also used to strengthen the high-frequency information differences among the different crops to widen the interclass differences and to achieve high-precision remote sensing crop classification. In the test areas, which are randomly selected in multiple farms located far from the sampling area, we compare our method with other methods. The results demonstrate that the frequency-domain learning approach better mitigates issues, such as incomplete crop extractions and fragmented boundaries, which leads to higher classification accuracy and robustness. This paper applies frequency-domain deep learning to remote sensing crop classification, highlighting a novel and effective solution that supports agricultural management decisions and planning. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
46. EddyDet: A Deep Framework for Oceanic Eddy Detection in Synthetic Aperture Radar Images.
- Author
-
Zhang, Di, Gade, Martin, Wang, Wensheng, and Zhou, Haoran
- Subjects
- *
SYNTHETIC aperture radar , *SYNTHETIC apertures , *CONVOLUTIONAL neural networks , *EDDIES , *MANUAL labor - Abstract
This paper presents a deep framework EddyDet to automatically detect oceanic eddies in Synthetic Aperture Radar (SAR) images. The EddyDet has been developed using the Mask Region with Convolutional Neural Networks (Mask RCNN) framework, incorporating two new branches: Edge Head and Mask Intersection over Union (IoU) Head. The Edge Head can learn internal texture information implicitly, and the Mask IoU Head improves the quality of predicted masks. A SAR dataset for Oceanic Eddy Detection (SOED) is specifically constructed to evaluate the effectiveness of the EddyDet model in detecting oceanic eddies. We demonstrate that the EddyDet is capable of achieving acceptable eddy detection results under the condition of limited training samples, which outperforms a Mask RCNN baseline in terms of average precision. The combined Edge Head and Mask IoU Head have the ability to describe the characteristics of eddies more correctly, while the EddyDet shows great potential in practice use accurately and time efficiently, saving manual labor to a large extent. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
47. Global–Local Information Fusion Network for Road Extraction: Bridging the Gap in Accurate Road Segmentation in China.
- Author
-
Wang, Xudong, Cai, Yujie, He, Kang, Wang, Sheng, Liu, Yan, and Dong, Yusen
- Subjects
- *
INFORMATION networks , *CONVOLUTIONAL neural networks , *DATA mining , *DEEP learning , *BRIDGES , *ENVIRONMENTAL monitoring - Abstract
Road extraction is crucial in urban planning, rescue operations, and military applications. Compared to traditional methods, using deep learning for road extraction from remote sensing images has demonstrated unique advantages. However, previous convolutional neural networks (CNN)-based road extraction methods have had limited receptivity and failed to effectively capture long-distance road features. On the other hand, transformer-based methods have good global information-capturing capabilities, but face challenges in extracting road edge information. Additionally, existing excellent road extraction methods lack validation for the Chinese region. To address these issues, this paper proposes a novel road extraction model called the global–local information fusion network (GLNet). In this model, the global information extraction (GIE) module effectively integrates global contextual relationships, the local information extraction (LIE) module accurately captures road edge information, and the information fusion (IF) module combines the output features from both global and local branches to generate the final extraction results. Further, a series of experiments on two different Chinese road datasets with geographic robustness demonstrate that our model outperforms the state-of-the-art deep learning models for road extraction tasks in China. On the CHN6-CUG dataset, the overall accuracy (OA) and intersection over union (IoU) reach 97.49% and 63.27%, respectively, while on the RDCME dataset, OA and IoU reach 98.73% and 84.97%, respectively. These research results hold significant implications for road traffic, humanitarian rescue, and environmental monitoring, particularly in the context of the Chinese region. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
48. TTNet: A Temporal-Transform Network for Semantic Change Detection Based on Bi-Temporal Remote Sensing Images.
- Author
-
Jiang, Liangcun, Li, Feng, Huang, Li, Peng, Feifei, and Hu, Lei
- Subjects
- *
REMOTE sensing , *CONVOLUTIONAL neural networks , *LAND cover , *IMAGE analysis - Abstract
Semantic change detection (SCD) holds a critical place in remote sensing image interpretation, as it aims to locate changing regions and identify their associated land cover classes. Presently, post-classification techniques stand as the predominant strategy for SCD due to their simplicity and efficacy. However, these methods often overlook the intricate relationships between alterations in land cover. In this paper, we argue that comprehending the interplay of changes within land cover maps holds the key to enhancing SCD's performance. With this insight, a Temporal-Transform Module (TTM) is designed to capture change relationships across temporal dimensions. TTM selectively aggregates features across all temporal images, enhancing the unique features of each temporal image at distinct pixels. Moreover, we build a Temporal-Transform Network (TTNet) for SCD, comprising two semantic segmentation branches and a binary change detection branch. TTM is embedded into the decoder of each semantic segmentation branch, thus enabling TTNet to obtain better land cover classification results. Experimental results on the SECOND dataset show that TTNet achieves enhanced performance when compared to other benchmark methods in the SCD task. In particular, TTNet elevates mIoU accuracy by a minimum of 1.5% in the SCD task and 3.1% in the semantic segmentation task. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
49. Learning-Based Optimization of Hyperspectral Band Selection for Classification.
- Author
-
Ayna, Cemre Omer, Mdrafi, Robiulhossain, Du, Qian, and Gurbuz, Ali Cafer
- Subjects
- *
DEEP learning , *FEATURE selection , *SPECTRAL sensitivity , *IMAGE recognition (Computer vision) , *SUPERVISED learning , *CONVOLUTIONAL neural networks - Abstract
Hyperspectral sensors acquire spectral responses from objects with a large number of narrow spectral bands. The large volume of data may be costly in terms of storage and computational requirements. In addition, hyperspectral data are often information-wise redundant. Band selection intends to overcome these limitations by selecting a small subset of spectral bands that provide more information or better performance for particular tasks. However, existing band selection techniques do not directly maximize the task-specific performance, but rather utilize hand-crafted metrics as a proxy to the final goal of performance improvement. In this paper, we propose a deep learning (DL) architecture composed of a constrained measurement learning network for band selection, followed by a classification network. The proposed joint DL architecture is trained in a data-driven manner to optimize the classification loss along band selection. In this way, the proposed network directly learns to select bands that enhance the classification performance. Our evaluation results with Indian Pines (IP) and the University of Pavia (UP) datasets show that the proposed constrained measurement learning-based band selection approach provides higher classification accuracy compared to the state-of-the-art supervised band selection methods for the same number of bands selected. The proposed method shows 89.08 % and 97.78 % overall accuracy scores for IP and UP respectively, being 1.34 % and 2.19 % higher than the second-best method. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
50. CNN and Transformer Fusion for Remote Sensing Image Semantic Segmentation.
- Author
-
Chen, Xin, Li, Dongfen, Liu, Mingzhe, and Jia, Jiaru
- Subjects
- *
TRANSFORMER models , *CONVOLUTIONAL neural networks , *DATA mining , *DEEP learning , *INFORMATION modeling , *IMAGE segmentation - Abstract
Semantic segmentation of remote sensing images has been widely used in environmental protection, geological disaster discovery, and natural resource assessment. With the rapid development of deep learning, convolutional neural networks (CNNs) have dominated semantic segmentation, relying on their powerful local information extraction capabilities. Due to the locality of convolution operation, it can be challenging to obtain global context information directly. However, Transformer has excellent potential in global information modeling. This paper proposes a new hybrid convolutional and Transformer semantic segmentation model called CTFuse, which uses a multi-scale convolutional attention module in the convolutional part. CTFuse is a serial structure composed of a CNN and a Transformer. It first uses convolution to extract small-size target information and then uses Transformer to embed large-size ground target information. Subsequently, we propose a spatial and channel attention module in convolution to enhance the representation ability for global information and local features. In addition, we also propose a spatial and channel attention module in Transformer to improve the ability to capture detailed information. Finally, compared to other models used in the experiments, our CTFuse achieves state-of-the-art results on the International Society of Photogrammetry and Remote Sensing (ISPRS) Vaihingen and ISPRS Potsdam datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.