71 results on '"Yongtao Yu"'
Search Results
2. CCapFPN: A Context-Augmented Capsule Feature Pyramid Network for Pavement Crack Detection
- Author
-
Yongjun Zhang, Dilong Li, Haiyan Guan, Shenghua Jin, Yongtao Yu, and Changhui Yu
- Subjects
Computer science ,business.industry ,Mechanical Engineering ,Deep learning ,Pattern recognition ,Context (language use) ,Network topology ,Computer Science Applications ,Feature (computer vision) ,Automotive Engineering ,Fuse (electrical) ,Pyramid (image processing) ,Artificial intelligence ,business ,Representation (mathematics) ,Intelligent transportation system - Abstract
Periodically monitoring the pavement conditions is of great importance to many intelligent transportation activities. Timely and correctly identifying the distresses or anomalies on pavement surfaces can help to smooth traffic flows and avoid potential threats to pavement securities. In this paper, we develop a novel context-augmented capsule feature pyramid network (CCapFPN) to detect cracks from pavement images. The CCapFPN adopts vectorial capsules to represent high-level, intrinsic, and salient features of cracks. By designing a feature pyramid architecture, the CCapFPN can fuse different levels and different scales of capsule features to provide a high-resolution, semantically strong feature representation for accurate crack detection. To take advantage of the context properties, a context-augmented module is embedded into each stage of the CCapFPN to rapidly enlarge the receptive field. The CCapFPN performs effectively and efficiently in processing pavement images of diverse conditions and detecting cracks of different topologies. Quantitative evaluations show that an overall performance with a precision, a recall, and an F-score of 0.9200, 0.9149, and 0.9174, respectively, were achieved on the test datasets. Comparative studies with some existing deep learning and edge based crack detection methods also confirm the superior performance of the CCapFPN in crack detection tasks.
- Published
- 2022
3. MarkCapsNet: Road Marking Extraction From Aerial Images Using Self-Attention-Guided Capsule Network
- Author
-
Changhui Yu, Zuojun Liu, Xiaoling Jiang, Lanfang Wang, Jun Wang, Chao Liu, Yongjun Zhang, Yinyin Li, and Yongtao Yu
- Subjects
Computer science ,business.industry ,Extraction (chemistry) ,Self attention ,Capsule ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,Geotechnical Engineering and Engineering Geology ,business - Published
- 2022
4. Building Extraction From Remote Sensing Imagery With a High-Resolution Capsule Network
- Author
-
Chao Liu, Yahong Zhang, Yongtao Yu, Junyong Gao, Mingxin Jiang, Shenghua Jin, Haiyan Zhang, and Xiaoling Jiang
- Subjects
Computer science ,Feature (computer vision) ,Remote sensing (archaeology) ,Encoding (memory) ,Shadow ,Extraction (military) ,Electrical and Electronic Engineering ,Architecture ,Resolution (logic) ,Geotechnical Engineering and Engineering Geology ,Image (mathematics) ,Remote sensing - Abstract
The up-to-date and accurate building database serves as an important prerequisite to many applications. However, caused by the issues of shape and size variations, texture and distribution diversities, and occlusion and shadow covers of buildings in remote sensing images, it is still challenging to well guarantee the integrity and accuracy of the extracted building instances. This letter proposes a high-resolution capsule network (HR-CapsNet) to conduct building extraction. First, designed with an HR-CapsNet architecture assisted by multiresolution feature propagation and fusion, the HR-CapsNet can provide semantically strong and spatially accurate feature representations to promote the pixel-wise building extraction accuracy. In addition, integrated with an efficient capsule feature attention module, the HR-CapsNet can attend to channel-wise informative and class-specific spatial features to boost the feature encoding quality. Quantitative evaluations, visual inspections, and comparative experiments on two large remote sensing image datasets demonstrate that the HR-CapsNet provides a feasible and competitive solution to building extraction tasks.
- Published
- 2022
5. Land Cover Classification of Multispectral LiDAR Data With an Efficient Self-Attention Capsule Network
- Author
-
Yahong Zhang, Lanfang Wang, Haiyan Zhang, Chao Liu, Jonathan Li, Haiyan Guan, Shangbing Gao, and Yongtao Yu
- Subjects
Land use ,Channel (digital image) ,business.industry ,Computer science ,Multispectral image ,Ranging ,Pattern recognition ,Land cover ,Geotechnical Engineering and Engineering Geology ,Lidar ,Feature (computer vision) ,Robustness (computer science) ,Artificial intelligence ,Electrical and Electronic Engineering ,business - Abstract
Periodically conducting land cover mapping plays a vital role in monitoring the status and changes of the land use. The up-to-date and accurate land use database serves importantly for a wide range of applications. This letter constructs an efficient self-attention capsule network (ESA-CapsNet) for land cover classification of multispectral light detection and ranging (LiDAR) data. First, formulated with a novel capsule encoder-decoder architecture, the ESA-CapsNet performs promisingly in extracting high-level, informative, and strong feature semantics for pixel-wise land cover classification by using the five types of rasterized feature images. Furthermore, designed with a novel capsule-based attention module, the channel and spatial feature encodings are comprehensively exploited to boost the feature saliency and robustness. The ESA-CapsNet is evaluated on two multispectral LiDAR data sets and achieves an advantageous performance with the overall accuracy, average accuracy, and kappa coefficient of over 98.42%, 95.15%, and 0.9776, respectively. Comparative experiments with the existing methods also demonstrate the effectiveness and applicability of the ESA-CapsNet in land cover classification tasks.
- Published
- 2022
6. CS-CapsFPN: A Context-Augmentation and Self-Attention Capsule Feature Pyramid Network for Road Network Extraction from Remote Sensing Imagery
- Author
-
Jonathan Li, Shenghua Jin, Yongjun Zhang, Haiyan Guan, E. Tang, Shaozhang Xiao, Changhui Yu, Yongtao Yu, and Jun Wang
- Subjects
Computer science ,Remote sensing (archaeology) ,Feature (computer vision) ,Self attention ,General Earth and Planetary Sciences ,Context (language use) ,Pyramid (image processing) ,Remote sensing ,Network model - Abstract
The information-accurate road network database is greatly significant and provides essential input to many transportation-related activities. Recently, remote sensing images have been an important ...
- Published
- 2021
7. Capsule Feature Pyramid Network for Building Footprint Extraction From High-Resolution Aerial Imagery
- Author
-
Changhui Yu, Haiyan Guan, Shenghua Jin, Lanfang Wang, Yongfeng Ren, Yongtao Yu, and Dilong Li
- Subjects
business.industry ,Computer science ,Feature extraction ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,0211 other engineering and technologies ,Signed distance function ,02 engineering and technology ,Geotechnical Engineering and Engineering Geology ,Footprint ,Data set ,Feature (computer vision) ,Extraction (military) ,Computer vision ,Pyramid (image processing) ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,Aerial image ,021101 geological & geomatics engineering - Abstract
Building footprint extraction plays an important role in a wide range of applications. However, due to size and shape diversities, occlusions, and complex scenarios, it is still challenging to accurately extract building footprints from aerial images. This letter proposes a capsule feature pyramid network (CapFPN) for building footprint extraction from aerial images. Taking advantage of the properties of capsules and fusing different levels of capsule features, the CapFPN can extract high-resolution, intrinsic, and semantically strong features, which perform effectively in improving the pixel-wise building footprint extraction accuracy. With the use of signed distance maps as ground truths, the CapFPN can extract solid building regions free of tiny holes. Quantitative evaluations on an aerial image data set show that a precision, recall, intersection-over-union (IoU), and F-score of 0.928, 0.914, 0.853, and 0.921, respectively, are obtained. Comparative studies with six existing methods confirm the superior performance of the CapFPN in accurately extracting building footprints.
- Published
- 2021
8. Capsule-Based Networks for Road Marking Extraction and Classification From Mobile LiDAR Point Clouds
- Author
-
Lingfei Ma, José Marcato Junior, Wesley Nunes Gonçalves, Ying Li, Jonathan Li, Yongtao Yu, and Michael A. Chapman
- Subjects
050210 logistics & transportation ,business.industry ,Computer science ,Mechanical Engineering ,Deep learning ,05 social sciences ,Feature extraction ,Point cloud ,Pattern recognition ,Image segmentation ,Thresholding ,Computer Science Applications ,Robustness (computer science) ,Inverse distance weighting ,0502 economics and business ,Automotive Engineering ,Softmax function ,Artificial intelligence ,business - Abstract
Accurate road marking extraction and classification play a significant role in the development of autonomous vehicles (AVs) and high-definition (HD) maps. Due to point density and intensity variations from mobile laser scanning (MLS) systems, most of the existing thresholding-based extraction methods and rule-based classification methods cannot deliver high efficiency and remarkable robustness. To address this, we propose a capsule-based deep learning framework for road marking extraction and classification from massive and unordered MLS point clouds. This framework mainly contains three modules. Module I is first implemented to segment road surfaces from 3D MLS point clouds, followed by an inverse distance weighting (IDW) interpolation method for 2D georeferenced image generation. Then, in Module II, a U-shaped capsule-based network is constructed to extract road markings based on the convolutional and deconvolutional capsule operations. Finally, a hybrid capsule-based network is developed to classify different types of road markings by using a revised dynamic routing algorithm and large-margin Softmax loss function. A road marking dataset containing both 3D point clouds and manually labeled reference data is built from three types of road scenes, including urban roads, highways, and underground garages. The proposed networks were accordingly evaluated by estimating robustness and efficiency using this dataset. Quantitative evaluations indicate the proposed extraction method can deliver 94.11% in precision, 90.52% in recall, and 92.43% in F1-score, respectively, while the classification network achieves an average of 3.42% misclassification rate in different road scenes.
- Published
- 2021
9. OA-CapsNet: A One-Stage Anchor-Free Capsule Network for Geospatial Object Detection from Remote Sensing Imagery
- Author
-
Yongtao Yu, Changhui Yu, Fenfen Li, Chao Liu, Jonathan Li, Shenghua Jin, Haiyan Guan, Dilong Li, and Junyong Gao
- Subjects
Geospatial analysis ,010504 meteorology & atmospheric sciences ,Computer science ,Orientation (computer vision) ,0211 other engineering and technologies ,One stage ,02 engineering and technology ,computer.software_genre ,01 natural sciences ,Object detection ,Remote sensing (archaeology) ,General Earth and Planetary Sciences ,Scale (map) ,computer ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences ,Remote sensing - Abstract
Object detection from remote sensing images serves as an important prerequisite to many applications. However, caused by scale and orientation variations, appearance and distribution diversities, o...
- Published
- 2021
10. Multi-Scale Point-Wise Convolutional Neural Networks for 3D Object Segmentation From LiDAR Point Clouds in Large-Scale Environments
- Author
-
Yongtao Yu, Lingfei Ma, Ying Li, Michael A. Chapman, Weikai Tan, and Jonathan Li
- Subjects
Conditional random field ,050210 logistics & transportation ,Artificial neural network ,Computer science ,business.industry ,Mechanical Engineering ,05 social sciences ,Feature extraction ,Point cloud ,Pattern recognition ,Solid modeling ,Convolutional neural network ,Computer Science Applications ,Robustness (computer science) ,0502 economics and business ,Automotive Engineering ,Segmentation ,Artificial intelligence ,business - Abstract
Although significant improvement has been achieved in fully autonomous driving and semantic high-definition map (HD) domains, most of the existing 3D point cloud segmentation methods cannot provide high representativeness and remarkable robustness. The principally increasing challenges remain in completely and efficiently extracting high-level 3D point cloud features, specifically in large-scale road environments. This paper provides an end-to-end feature extraction framework for 3D point cloud segmentation by using dynamic point-wise convolutional operations in multiple scales. Compared to existing point cloud segmentation methods that are commonly based on traditional convolutional neural networks (CNNs), our proposed method is less sensitive to data distribution and computational powers. This framework mainly includes four modules. Module I is first designed to construct a revised 3D point-wise convolutional operation. Then, a U-shaped downsampling-upsampling architecture is proposed to leverage both global and local features in multiple scales in Module II. Next, in Module III, high-level local edge features in 3D point neighborhoods are further extracted by using an adaptive graph convolutional neural network based on the K-Nearest Neighbor (KNN) algorithm. Finally, in Module IV, a conditional random field (CRF) algorithm is developed for postprocessing and segmentation result refinement. The proposed method was evaluated on three large-scale LiDAR point cloud datasets in both urban and indoor environments. The experimental results acquired by using different point cloud scenarios indicate our method can achieve state-of-the-art semantic segmentation performance in feature representativeness, segmentation accuracy, and technical robustness.
- Published
- 2021
11. A Cascaded Deep Convolutional Network for Vehicle Logo Recognition From Frontal and Rear Images of Vehicles
- Author
-
Changhui Yu, Yongtao Yu, Haiyan Guan, and Dilong Li
- Subjects
050210 logistics & transportation ,Logo recognition ,Computer science ,business.industry ,Mechanical Engineering ,05 social sciences ,Feature extraction ,Quantitative Evaluations ,Logo ,Computer Science Applications ,Support vector machine ,Robustness (computer science) ,Test set ,0502 economics and business ,Automotive Engineering ,Computer vision ,Artificial intelligence ,business ,License - Abstract
Vehicle logo recognition provides an important supplement to vehicle make and model analysis. Some of the existing vehicle logo recognition methods depend on the detection of license plates to roughly locate vehicle logo regions using prior knowledge. The vehicle logo recognition performance is greatly affected by the license plate detection techniques. This paper presents a cascaded deep convolutional network for directly recognizing vehicle logos without depending on the existence of license plates. This is a two-stage processing framework composed of a region proposal network and a convolutional capsule network. First, potential region proposals that might contain vehicle logos are generated by the region proposal network. Then, the convolutional capsule network classifies these region proposals into the background and different types of vehicle logos. We have evaluated the proposed framework on a large test set towards vehicle logo recognition. Quantitative evaluations show that a detection rate, a recognition rate, and an overall performance of 0.987, 0.994, and 0.981, respectively, are achieved. Comparative studies with the Faster R-CNN and other three existing methods also confirm that the proposed method performs effectively and robustly in recognizing vehicle logos of various conditions.
- Published
- 2021
12. A self-attention capsule feature pyramid network for water body extraction from remote sensing imagery
- Author
-
Haiyan Guan, Shaozhang Xiao, Wenhao Wang, Lanfang Wang, Changhui Yu, Zuojun Liu, Yuting Yao, Lv Chang, Yongtao Yu, and Dilong Li
- Subjects
animal structures ,010504 meteorology & atmospheric sciences ,Computer science ,Self attention ,0211 other engineering and technologies ,02 engineering and technology ,01 natural sciences ,Water body ,Remote sensing (archaeology) ,Feature (computer vision) ,embryonic structures ,General Earth and Planetary Sciences ,Extraction (military) ,sense organs ,Pyramid (image processing) ,skin and connective tissue diseases ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences ,Remote sensing - Abstract
Timely and accurately measuring surface water bodies and monitoring their conditions and changes are greatly important to a wide range of environmental and social activities. Recently, with the dev...
- Published
- 2020
13. Land-cover classification of multispectral LiDAR data using CNN with optimized hyper-parameters
- Author
-
Yating Chen, Haiyan Guan, José Marcato Junior, Jonathan Li, Wesley Nunes Gonçalves, Yongtao Yu, and Suoyan Pan
- Subjects
010504 meteorology & atmospheric sciences ,Contextual image classification ,Computer science ,business.industry ,Multispectral image ,0211 other engineering and technologies ,Pattern recognition ,02 engineering and technology ,01 natural sciences ,Convolutional neural network ,Atomic and Molecular Physics, and Optics ,Object detection ,Computer Science Applications ,Lidar ,Preprocessor ,Artificial intelligence ,Computers in Earth Sciences ,business ,Engineering (miscellaneous) ,Feature learning ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences ,Geometric data analysis - Abstract
Multispectral LiDAR (Light Detection And Ranging) is characterized of the completeness and consistency of its spectrum and spatial geometric data, which provides a new data source for land-cover classification. In recent years, the convolutional neural network (CNN), compared with traditional machine learning methods, has made a series of breakthroughs in image classification, object detection, and image semantic segmentation due to its stronger feature learning and feature expression abilities. However, traditional CNN models suffer from some issues, such as a large number of layers, leading to higher computational cost. To address this problem, we propose a CNN-based multi-spectral LiDAR land-cover classification framework and analyze its optimal parameters to improve classification accuracy. This framework starts with the preprocessing of multi-spectral 3D LiDAR data into 2D images. Next, a CNN model is constructed with seven fundamental functional layers, and its hyper-parameters are comprehensively discussed and optimized. The constructed CNN model with the optimized hyper-parameters was tested on the Titan multi-spectral LiDAR data, which include three wavelengths of 532 nm, 1064 nm, and 1550 nm. Extensive experiments demonstrated that the constructed CNN with the optimized hyper-parameters is feasible for multi-spectral LiDAR land-cover classification tasks. Compared with the classical CNN models (i.e., AlexNet, VGG16 and ResNet50) and our previous studies, our constructed CNN model with the optimized hyper-parameters is superior in computational performance and classification accuracies.
- Published
- 2020
14. A Hybrid Capsule Network for Land Cover Classification Using Multispectral LiDAR Data
- Author
-
Jonathan Li, Tiannan Gu, Haiyan Guan, Lingfei Ma, Yongtao Yu, Lanfang Wang, and Dilong Li
- Subjects
Lidar ,Computer science ,Feature (computer vision) ,Multispectral image ,Feature extraction ,0211 other engineering and technologies ,Hyperspectral imaging ,02 engineering and technology ,Land cover ,Electrical and Electronic Engineering ,Geotechnical Engineering and Engineering Geology ,021101 geological & geomatics engineering ,Remote sensing - Abstract
Land cover mapping is an effective way to quantify land resources and monitor their changes. It plays an important role in a wide range of applications. This letter proposes a hybrid capsule network for land cover classification using multispectral light detection and ranging (LiDAR) data. First, the multispectral LiDAR data were rasterized into a set of feature images to exploit the geometrical and spectral properties of different types of land covers. Then, a hybrid capsule network composed of an encoder network and a decoder network is trained to extract both high-level local and global entity-oriented capsule features for accurate land cover classification. Quantitative classification evaluations on two data sets show that the overall accuracy, average accuracy, and kappa coefficient of over 97.89%, 94.54%, and 0.9713, respectively, are obtained. Comparative studies with five existing methods confirm that the proposed method performs robustly and accurately in land cover classification using the multispectral LiDAR data.
- Published
- 2020
15. A Convolutional Capsule Network for Traffic-Sign Recognition Using Mobile LiDAR Data With Digital Images
- Author
-
Daifeng Peng, Aixia Li, Haiyan Guan, Jonathan Li, Yongtao Yu, Jianyong Lu, and Yufu Zang
- Subjects
business.industry ,Computer science ,Feature extraction ,0211 other engineering and technologies ,Boltzmann machine ,Point cloud ,Pattern recognition ,02 engineering and technology ,Geotechnical Engineering and Engineering Geology ,Convolutional neural network ,Digital image ,Lidar ,Traffic sign recognition ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,021101 geological & geomatics engineering - Abstract
Traffic-sign recognition plays an important role in road transportation systems. This letter presents a novel two-stage method for detecting and recognizing traffic signs from mobile Light Detection and Ranging (LiDAR) point clouds and digital images. First, traffic signs are detected from mobile LiDAR point cloud data according to their geometrical and spectral properties, which have been fully studied in our previous work. Afterward, the traffic-sign patches are obtained by projecting the detected points onto the registered digital images. To improve the performance of traffic-sign recognition, we apply a convolutional capsule network to the traffic-sign patches to classify them into different types. We have evaluated the proposed framework on data sets acquired by a RIEGL VMX-450 system. Quantitative evaluations show that a recognition rate of 0.957 is achieved. Comparative studies with the convolutional neural network (CNN) and our previous supervised Gaussian–Bernoulli deep Boltzmann machine (GB-DBM) classifier also confirm that the proposed method performs effectively and robustly in recognizing traffic signs of various types and conditions.
- Published
- 2020
16. 3-D Feature Matching for Point Cloud Object Extraction
- Author
-
Haiyan Guan, Tai-Yue Chen, Dilong Li, Jonathan Li, Shenghua Jin, Cheng Wang, and Yongtao Yu
- Subjects
business.industry ,Computer science ,Feature extraction ,0211 other engineering and technologies ,Point cloud ,Pattern recognition ,02 engineering and technology ,Function (mathematics) ,Geotechnical Engineering and Engineering Geology ,Object (computer science) ,Feature (computer vision) ,Point (geometry) ,Artificial intelligence ,Affine transformation ,Electrical and Electronic Engineering ,business ,021101 geological & geomatics engineering ,Test data - Abstract
Effective object extraction plays an important role in many point cloud-based applications. This letter proposes a 3-D feature matching framework for point cloud object extraction. To determine the optimal affine transformation parameters for each template feature point, a convex dissimilarity function and the locally affine-invariant geometric constraints are designed to construct the overall objective function. The 3-D feature matching framework is integrated into a point cloud object extraction workflow. Extraction results on six test data sets show that average completeness, correctness, quality, and $F_{1}$ -measure of 0.96, 0.97, 0.93, and 0.96, respectively, are obtained in extracting light poles, vehicles, and palm trees. Comparative studies also confirm that the proposed method performs effectively and robustly, and exhibits superior or compatible performance over the other compared methods.
- Published
- 2020
17. Orientation guided anchoring for geospatial object detection from remote sensing imagery
- Author
-
Aixia Li, Yongtao Yu, Haiyan Guan, Dilong Li, Tiannan Gu, and E. Tang
- Subjects
Geospatial analysis ,010504 meteorology & atmospheric sciences ,Orientation (computer vision) ,Computer science ,Feature extraction ,0211 other engineering and technologies ,02 engineering and technology ,computer.software_genre ,Object (computer science) ,01 natural sciences ,Convolutional neural network ,Atomic and Molecular Physics, and Optics ,Object detection ,Computer Science Applications ,Pyramid (image processing) ,Computers in Earth Sciences ,Engineering (miscellaneous) ,Subnetwork ,computer ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences ,Remote sensing - Abstract
Object detection from remote sensing imagery plays a significant role in a wide range of applications, including urban planning, intelligent transportation systems, ecology and environment analysis, etc. However, scale variations, orientation variations, illumination changes, and partial occlusions, as well as image qualities, bring great challenges for accurate geospatial object detection. In this paper, we propose an efficient orientation guided anchoring based geospatial object detection network based on convolutional neural networks. To handle objects of varying sizes, the feature extraction subnetwork extracts a pyramid of semantically strong features at different scales. Based on orientation guided anchoring, the anchor generation subnetwork generates a small set of high-quality, oriented anchors as object proposals. After orientation region of interest pooling, objects of interest are detected from the object proposals through the object detection subnetwork. The proposed method has been tested on a large geospatial object detection dataset. Quantitative evaluations show that an overall completeness, correctness, quality, and F1-measure of 0.9232, 0.9648, 0.8931, and 0.9435, respectively, are obtained. In addition, the proposed method achieves a processing speed of 8 images per second on a GPU on the cloud computing platform. Comparative studies with the existing object detection methods also demonstrate the advantageous detection accuracy and computational efficiency of our proposed method.
- Published
- 2020
18. Road Manhole Cover Delineation Using Mobile Laser Scanning Point Cloud Data
- Author
-
Dilong Li, Cheng Wang, Jonathan Li, Yongtao Yu, Haiyan Guan, and Chunhua Jin
- Subjects
Computer science ,business.industry ,Feature extraction ,0211 other engineering and technologies ,Point cloud ,02 engineering and technology ,Image segmentation ,Geotechnical Engineering and Engineering Geology ,Kernel (image processing) ,Road surface ,Georeference ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,Classifier (UML) ,021101 geological & geomatics engineering - Abstract
Periodical road manhole cover measurement is extremely important to ensure road safety and reduce traffic disasters. This letter proposes an effective method for delineating road manhole covers from mobile laser scanning point cloud data. To improve processing efficiency, first, road surface points are segmented and rasterized into georeferenced intensity images. Then, object-oriented patches are generated through superpixel segmentation and further fed to a convolutional capsule network classifier for manhole cover detection. Finally, manhole covers are accurately delineated through a marked point process of disks. Quantitative evaluations on three data sets show that an average completeness, correctness, quality, and ${F} _{1}$ -measure of 0.965, 0.961, 0.929, and 0.963, respectively, are obtained. Comparative studies with three existing methods confirm that the proposed method performs superiorly in delineating manhole covers of varying conditions and on complex road surface environments.
- Published
- 2020
19. GGM-Net: Graph Geometric Moments Convolution Neural Network for Point Cloud Shape Classification
- Author
-
Haiyan Guan, Hanyun Wang, Deren Li, Xin Shen, Dilong Li, and Yongtao Yu
- Subjects
010504 meteorology & atmospheric sciences ,General Computer Science ,Computer science ,Geometric moments ,0211 other engineering and technologies ,General Engineering ,Point cloud ,graph geometric moments convolution neural network ,02 engineering and technology ,shape classification ,01 natural sciences ,Convolutional neural network ,Convolution ,Robustness (computer science) ,Feature (machine learning) ,Graph (abstract data type) ,General Materials Science ,Point (geometry) ,lcsh:Electrical engineering. Electronics. Nuclear engineering ,Representation (mathematics) ,Algorithm ,lcsh:TK1-9971 ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences ,point cloud - Abstract
Geometric feature acts as an important role in point cloud shape classification tasks. Previous methods have proved that the geometric information of point clouds effectively improves the classification accuracy. Mo-Net firstly introduced geometric moments into point cloud shape classification, which, to fit the form of second order geometric moments, extends the number of input channels from three to nine. Unfortunately, similar to PointNet, Mo-Net cannot capture the local structures. To address this issue, we propose a graph geometric moments convolution neural network (GGM-Net), which learns local geometric features from geometric moments representation of a local point set. The core module of the GGM-Net is to learn features from geometric moments (termed as GGM convolution). Specifically, the GGM convolution learns point features and local features from the first and second order geometric moments of points and its local neighbors, respectively, and then combines these features by using an addition operation. In this way, a geometrical local representation about points is obtained, which leads to much surface geometry awareness and robustness. Equipped with the GGM convolution, GGM-Net, a simple end-to-end architecture, is developed to achieve a competitive accuracy on the benchmark dataset ModelNet40 and perform more efficiently in terms of memory and computational complexity.
- Published
- 2020
20. Vehicle Detection From High-Resolution Remote Sensing Imagery Using Convolutional Capsule Networks
- Author
-
Tiannan Gu, Dilong Li, Haiyan Guan, Shenghua Jin, and Yongtao Yu
- Subjects
Standard test image ,Kernel (image processing) ,Computer science ,Vehicle detection ,Feature extraction ,0211 other engineering and technologies ,High resolution ,02 engineering and technology ,Electrical and Electronic Engineering ,Geotechnical Engineering and Engineering Geology ,021101 geological & geomatics engineering ,Test data ,Remote sensing - Abstract
Vehicle detection plays an important role in a variety of traffic-related applications. However, due to the scale and orientation variations and partial occlusions of vehicles, it is still challengeable to accurately detect vehicles from remote sensing images. This letter proposes a convolutional capsule network for detecting vehicles from high-resolution remote sensing images. First, a test image is segmented into superpixels to generate meaningful and nonredundant patches. Then, these patches are input to a convolutional capsule network to label them into vehicles or the background. Finally, nonmaximum suppression is adopted to eliminate repetitive detections. Quantitative evaluations on four test data sets show that average completeness, correctness, quality, and F1-measure of 0.93, 0.97, 0.90, and 0.95, respectively, are obtained. Comparative studies with three existing methods confirm that the proposed method effectively performs in detecting vehicles of various conditions.
- Published
- 2019
21. Multispectral LiDAR Point Cloud Classification Using SE-PointNet++
- Author
-
Hanyun Wang, Yufu Zang, Haiyan Guan, Peiran Zhao, Zhuangwei Jing, Yongtao Yu, Dilong Li, and Jonathan Li
- Subjects
010504 meteorology & atmospheric sciences ,Computer science ,Science ,Multispectral image ,PointNet++ ,0211 other engineering and technologies ,Cognitive neuroscience of visual object recognition ,Point cloud ,squeeze and excitation ,point cloud classification ,multispectral LiDAR ,Ranging ,02 engineering and technology ,01 natural sciences ,Lidar ,General Earth and Planetary Sciences ,Segmentation ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences ,Remote sensing ,Geometric data analysis ,Block (data storage) - Abstract
A multispectral light detection and ranging (LiDAR) system, which simultaneously collects spatial geometric data and multi-wavelength intensity information, opens the door to three-dimensional (3-D) point cloud classification and object recognition. Because of the irregular distribution property of point clouds and the massive data volume, point cloud classification directly from multispectral LiDAR data is still challengeable and questionable. In this paper, a point-wise multispectral LiDAR point cloud classification architecture termed as SE-PointNet++ is proposed via integrating a Squeeze-and-Excitation (SE) block with an improved PointNet++ semantic segmentation network. PointNet++ extracts local features from unevenly sampled points and represents local geometrical relationships among the points through multi-scale grouping. The SE block is embedded into PointNet++ to strengthen important channels to increase feature saliency for better point cloud classification. Our SE-PointNet++ architecture has been evaluated on the Titan multispectral LiDAR test datasets and achieved an overall accuracy, a mean Intersection over Union (mIoU), an F1-score, and a Kappa coefficient of 91.16%, 60.15%, 73.14%, and 0.86, respectively. Comparative studies with five established deep learning models confirmed that our proposed SE-PointNet++ achieves promising performance in multispectral LiDAR point cloud classification tasks.
- Published
- 2021
22. Aeroplane detection from high-resolution remotely sensed imagery using bag-of-visual-words based hough forests
- Author
-
Yan Yuan, Haiyan Guan, Tiannan Gu, Yongtao Yu, and Dilong Li
- Subjects
010504 meteorology & atmospheric sciences ,business.industry ,Computer science ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,0211 other engineering and technologies ,High resolution ,02 engineering and technology ,01 natural sciences ,Bag-of-words model in computer vision ,Salient ,Computer Science::Computer Vision and Pattern Recognition ,General Earth and Planetary Sciences ,Computer vision ,Artificial intelligence ,business ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences - Abstract
This paper presents a rotation-invariant method for detecting aeroplanes from high-resolution remotely sensed images. First, a superpixel-based strategy is proposed to generate salient and ...
- Published
- 2019
23. 3D-CNN BASED TREE SPECIES CLASSIFICATION USING MOBILE LIDAR DATA
- Author
-
Li Dengyun, Wanqian Yan, Yongtao Yu, Haiyan Guan, and Jonathan Li
- Subjects
lcsh:Applied optics. Photonics ,business.industry ,Computer science ,lcsh:T ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,lcsh:TA1501-1820 ,Terrain ,Pattern recognition ,computer.software_genre ,Convolutional neural network ,lcsh:Technology ,Euclidean distance ,Tree (data structure) ,Voxel ,lcsh:TA1-2040 ,Segmentation ,Artificial intelligence ,Cluster analysis ,business ,lcsh:Engineering (General). Civil engineering (General) ,Tree species ,computer - Abstract
Our work addresses the problem of classifying tree species from mobile LiDAR data. The work is a two step-wise strategy, including tree segmentation and tree species classification. In the tree segmentation step, a voxel-based upward growing filtering is proposed to remove terrain points from the mobile laser scanning data. Then, individual trees are segmented via a Euclidean distance clustering approach and Voxel-based Normalized Cut (VNCut) segmentation approach. In the tree species classification, a voxel-based 3D convolutional neural network (3D-CNN) model is developed based on intensity information. A road section data acquired by a RIEGL VMX-450 system are selected for evaluating the proposed tree classification method. Qualitative analysis shows that our algorithm achieves a good performance.
- Published
- 2019
24. A Comparative Land-Cover Classification Feature Study of Learning Algorithms: DBM, PCA, and RF Using Multispectral LiDAR Data
- Author
-
Yongtao Yu, Jonathan Li, Suoyan Pan, Daifeng Peng, and Haiyan Guan
- Subjects
Atmospheric Science ,010504 meteorology & atmospheric sciences ,business.industry ,Computer science ,Deep learning ,Feature extraction ,Multispectral image ,0211 other engineering and technologies ,Pattern recognition ,02 engineering and technology ,Land cover ,01 natural sciences ,Random forest ,Lidar ,Feature (computer vision) ,Personal computer ,Artificial intelligence ,Computers in Earth Sciences ,business ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences - Abstract
Multispectral LiDAR, characterization of completeness, and consistency of spectrum and spatial geometric data provide a new data source for land cover classification. However, how to choose the optimal features for a given set of land covers is an open problem for effective land cover classification. To address this problem, we propose a comparative scheme, which investigates a popular deep learning (deep Boltzmann machine, DBM) model for high-level feature representation and widely used machine learning methods for low-level feature extraction and selection [principal component analysis (PCA) and random forest (RF)] in land cover classification. The comparative study was conducted on the multispectral LiDAR point clouds, acquired by a Teledyne Optech's Titan airborne system. The deep learning-based high-level feature representation experimental results showed that, on an ordinary personal computer or workstation, this method required larger training samples and more computational complexity than the machine learning-based low-level feature extraction and selection methods. However, our comparative experiments demonstrated that the classification accuracies of the DBM-based method were higher than those of the RF-based and PCA-based methods using multispectral LiDAR data.
- Published
- 2019
25. DA-CapsUNet: A Dual-Attention Capsule U-Net for Road Extraction from Remote Sensing Imagery
- Author
-
Haiyan Guan, Yongtao Yu, and Yongfeng Ren
- Subjects
road extraction ,010504 meteorology & atmospheric sciences ,Exploit ,Computer science ,capsule network ,capsule U-Net ,capsule attention network ,remote sensing imagery ,deep learning ,0211 other engineering and technologies ,02 engineering and technology ,01 natural sciences ,Encoding (memory) ,Representation (mathematics) ,lcsh:Science ,Image resolution ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences ,Remote sensing ,business.industry ,Deep learning ,Perspective (graphical) ,Feature (computer vision) ,Fuse (electrical) ,General Earth and Planetary Sciences ,lcsh:Q ,Artificial intelligence ,business - Abstract
The up-to-date and information-accurate road database plays a significant role in many applications. Recently, with the improvement in image resolutions and quality, remote sensing images have provided an important data source for road extraction tasks. However, due to the topology variations, spectral diversities, and complex scenarios, it is still challenging to realize fully automated and highly accurate road extractions from remote sensing images. This paper proposes a novel dual-attention capsule U-Net (DA-CapsUNet) for road region extraction by combining the advantageous properties of capsule representations and the powerful features of attention mechanisms. By constructing a capsule U-Net architecture, the DA-CapsUNet can extract and fuse multiscale capsule features to recover a high-resolution and semantically strong feature representation. By designing the multiscale context-augmentation and two types of feature attention modules, the DA-CapsUNet can exploit multiscale contextual properties at a high-resolution perspective and generate an informative and class-specific feature encoding. Quantitative evaluations on a large dataset showed that the DA-CapsUNet provides a competitive road extraction performance with a precision of 0.9523, a recall of 0.9486, and an F-score of 0.9504, respectively. Comparative studies with eight recently developed deep learning methods also confirmed the applicability and superiority or compatibility of the DA-CapsUNet in road extraction tasks.
- Published
- 2020
- Full Text
- View/download PDF
26. AUTOMATIC TRAFFIC SIGN DETECTION AND RECOGNITION USING MOBILE LIDAR DATA WITH DIGITAL IMAGES
- Author
-
Yongtao Yu, Haiyan Guan, Jonathan Li, and Li Dengyun
- Subjects
lcsh:Applied optics. Photonics ,050210 logistics & transportation ,lcsh:T ,Computer science ,business.industry ,05 social sciences ,Point cloud ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,lcsh:TA1501-1820 ,02 engineering and technology ,lcsh:Technology ,Traffic sign detection ,Digital image ,Mobile lidar ,lcsh:TA1-2040 ,0502 economics and business ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,lcsh:Engineering (General). Civil engineering (General) ,business ,Network model - Abstract
This paper presents a traffic sign detection and recognition method from mobile LiDAR data and digital images for intelligent transportation-related applications. The traffic sign detection and recognition method includes two steps: traffic sign interest regions are first extracted from mobile LiDRA data. Next, traffic signs are identified from digital images simultaneously collected from the multi-sensor mobile LiDAR systems via a convolutional capsule network model. The experimental results demonstrate that the proposed method obtains a promising, reliable, and high performance in both detecting traffic signs in 3-D point clouds and recognizing traffic signs on 2-D images.
- Published
- 2020
27. Research on the Relationship Between the Surface Temperature of CPU and Its Utilization Based on Infrared Testing Technology
- Author
-
Zhiwei Fu, Xiaoqiang Wang, Jun Luo, Yongtao Yu, Yu Sun, and Yiwen Long
- Subjects
Surface (mathematics) ,Very large scale integrated circuits ,Surface fitting ,Infrared ,Computer science ,CPU time ,Thermal protection ,Central processing unit ,Computer Science::Operating Systems ,Temperature measurement ,Computational science - Abstract
As a kind of very large scale integrated circuit, the surface temperature of CPU is closely affected by its utilization. However, there relatively lacks quantitative analysis on the relationship between the surface temperature and CPU utilization at present. In this study the CPU utilization is controlled by a load tool running on the kernels, and the number of kernels get into operation is gradually increased from the initial to the steady-state threshold, meanwhile the surface temperature of CPU die without thermal protection is collected based on the infrared testing technology. Then the utilization rates of each core under different load levels are obtained, and the model coefficients about the surface temperature and CPU utilization are fitted according to the data collected, so that the quantitative relationship between the surface temperature and CPU utilization is analyzed.
- Published
- 2020
28. A self-adaptive mean shift tree-segmentation method Using UAV LiDAR data
- Author
-
Lin Cao, Wanqian Yan, Yongtao Yu, Haiyan Guan, Jianyong Lu, and Cheng Li
- Subjects
010504 meteorology & atmospheric sciences ,Computer science ,Science ,0211 other engineering and technologies ,02 engineering and technology ,Kernel Bandwidth ,01 natural sciences ,self-adaptive kernel bandwidth ,tree segmentation ,mean shift ,UAV LiDAR ,Bandwidth (computing) ,Segmentation ,Point (geometry) ,Mean-shift ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences ,business.industry ,Pattern recognition ,Ranging ,uav lidar ,Tree (data structure) ,Lidar ,General Earth and Planetary Sciences ,Artificial intelligence ,business - Abstract
Unmanned aerial vehicles using light detection and ranging (UAV LiDAR) with high spatial resolution have shown great potential in forest applications because they can capture vertical structures of forests. Individual tree segmentation is the foundation of many forest research works and applications. The tradition fixed bandwidth mean shift has been applied to individual tree segmentation and proved to be robust in tree segmentation. However, the fixed bandwidth-based segmentation methods are not suitable for various crown sizes, resulting in omission or commission errors. Therefore, to increase tree-segmentation accuracy, we propose a self-adaptive bandwidth estimation method to estimate the optimal kernel bandwidth automatically without any prior knowledge of crown size. First, from the global maximum point, we divide the three-dimensional (3D) space into a set of angular sectors, for each of which a canopy surface is simulated and the potential tree crown boundaries are identified to estimate average crown width as the kernel bandwidth. Afterwards, we use a mean shift with the automatically estimated kernel bandwidth to extract individual tree points. The method is iteratively implemented within a given area until all trees are segmented. The proposed method was tested on the 7 plots acquired by a Velodyne 16E LiDAR system, including 3 simple plots and 4 complex plots, and 95% and 80% of trees were correctly segmented, respectively. Comparative experiments show that our method contributes to the improvement of both segmentation accuracy and computational efficiency.
- Published
- 2020
29. Design and implementation of airborne hyperspectral data processing platform compatible with intelligent processing algorithms
- Author
-
Jiang Tian, Kecheng Gong, Hao Hao, Yongtao Yu, Lixiong Zhang, and Peng Yuanxi
- Subjects
Data processing ,Computer science ,Real-time computing ,Hyperspectral imaging - Published
- 2020
30. Semantic Labeling of Mobile LiDAR Point Clouds via Active Learning and Higher Order MRF
- Author
-
Huan Luo, Jonathan Li, Dawei Zai, Cheng Wang, Chenglu Wen, Ziyi Chen, and Yongtao Yu
- Subjects
Markov random field ,Training set ,Point (typography) ,Computer science ,Active learning (machine learning) ,business.industry ,0211 other engineering and technologies ,Point cloud ,Statistical model ,02 engineering and technology ,Machine learning ,computer.software_genre ,Semantics ,Active learning ,0202 electrical engineering, electronic engineering, information engineering ,General Earth and Planetary Sciences ,020201 artificial intelligence & image processing ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,computer ,021101 geological & geomatics engineering - Abstract
Using mobile Light Detection and Ranging point clouds to accomplish road scene labeling tasks shows promise for a variety of applications. Most existing methods for semantic labeling of point clouds require a huge number of fully supervised point cloud scenes, where each point needs to be manually annotated with a specific category. Manually annotating each point in point cloud scenes is labor intensive and hinders practical usage of those methods. To alleviate such a huge burden of manual annotation, in this paper, we introduce an active learning method that avoids annotating the whole point cloud scenes by iteratively annotating a small portion of unlabeled supervoxels and creating a minimal manually annotated training set. In order to avoid the biased sampling existing in traditional active learning methods, a neighbor-consistency prior is exploited to select the potentially misclassified samples into the training set to improve the accuracy of the statistical model. Furthermore, lots of methods only consider short-range contextual information to conduct semantic labeling tasks, but ignore the long-range contexts among local variables. In this paper, we use a higher order Markov random field model to take into account more contexts for refining the labeling results, despite of lacking fully supervised scenes. Evaluations on three data sets show that our proposed framework achieves a high accuracy in labeling point clouds although only a small portion of labels is provided. Moreover, comparative experiments demonstrate that our proposed framework is superior to traditional sampling methods and exhibits comparable performance to those fully supervised models.
- Published
- 2018
31. Robust Traffic-Sign Detection and Classification Using Mobile LiDAR Data With Digital Images
- Author
-
Wanqian Yan, Liang Zhong, Dilong Li, Yongtao Yu, and Haiyan Guan
- Subjects
050210 logistics & transportation ,Atmospheric Science ,Computer science ,business.industry ,05 social sciences ,Feature extraction ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,0211 other engineering and technologies ,Boltzmann machine ,Point cloud ,02 engineering and technology ,Digital image ,Lidar ,Mobile lidar ,0502 economics and business ,Computer vision ,Artificial intelligence ,Computers in Earth Sciences ,business ,Classifier (UML) ,Image resolution ,021101 geological & geomatics engineering - Abstract
This study aims at building a robust method for detecting and classifying traffic signs from mobile LiDAR point clouds and digital images. First, this method detects traffic signs from mobile LiDAR point clouds with regard to a prior knowledge of road width, pole height, reflectance, geometrical structure, and traffic-sign size. Then, traffic-sign images are segmented by projecting the detected traffic-sign points onto the digital images. Afterward, the segmented traffic-sign images are normalized for automatic classification with a given image size. Finally, a traffic-sign classifier is proposed based on a supervised Gaussian–Bernoulli deep Boltzmann machine model. We evaluated the proposed method using datasets acquired by a RIEGL VMX-450 system. The traffic-sign detection accuracy of 86.8% was achieved; through parameter sensitivity analysis, the overall performance of traffic-sign classification achieved a recognition rate of 93.3%. The computational performance showed that our method provides a promising solution to traffic-sign detection and classification using mobile LiDAR point clouds and digital images.
- Published
- 2018
32. LF-LDA
- Author
-
Liang Shi, Jialin Ma, Yongtao Yu, Yongjun Zhang, Bolun Chen, and Zijian Wang
- Subjects
Topic model ,Computer science ,business.industry ,02 engineering and technology ,computer.software_genre ,ComputingMethodologies_PATTERNRECOGNITION ,Hardware and Architecture ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer ,Software ,Natural language processing - Abstract
This article describes how text documents are a major data structure in the era of big data. With the explosive growth of data, the number of documents with multi-labels has increased dramatically. The popular multi-label classification technology, which is usually employed to handle multinomial text documents, is sensitive to the noise terms of text documents. Therefore, there still exists a huge room for multi-label classification of text documents. This article introduces a supervised topic model, named labeled LDA with function terms (LF-LDA), to filter out the noisy function terms from text documents, which can help to improve the performance of multi-label classification of text documents. The article also shows the derivation of the Gibbs Sampling formulas in detail, which can be generalized to other similar topic models. Based on the textual data set RCV1-v2, the article compared the proposed model with other two state-of-the-art multi-label classifiers, Tuned SVM and labeled LDA, on both Macro-F1 and Micro-F1 metrics. The result shows that LF-LDA outperforms them and has the lowest variance, which indicates the robustness of the LF-LDA classifier.
- Published
- 2018
33. Sparse anchoring guided high-resolution capsule network for geospatial object detection from remote sensing imagery
- Author
-
Yongjun Zhang, Hao Qiang, Yongtao Yu, E. Tang, Mingxin Jiang, Jonathan Li, Changhui Yu, and Jun Wang
- Subjects
Global and Planetary Change ,Network architecture ,Geospatial analysis ,Channel (digital image) ,Orientation (computer vision) ,Computer science ,Cognitive neuroscience of visual object recognition ,Management, Monitoring, Policy and Law ,computer.software_genre ,Object detection ,Feature (computer vision) ,Shadow ,Computers in Earth Sciences ,computer ,Earth-Surface Processes ,Remote sensing - Abstract
As the optical remote sensing techniques keep developing with a rapid pace, remote sensing images are positively considered in many fields. Accordingly, a great number of algorithms have been exploited for remote sensing image interpretation purposes. Thereinto, object recognition acts as an important ingredient to many applications. However, to achieve highly accurate object recognition is still challengeable caused by the orientation and size diversities, spatial distribution and density variations, shape and aspect ratio irregularities, occlusion and shadow impacts, as well as complex texture and surrounding environment changes. In this paper, a sparse anchoring guided high-resolution capsule network (SAHR-CapsNet) is proposed for geospatial object detection based on remote sensing images. First, formulated with the multibranch high-resolution capsule network architecture assisted by multiscale feature propagation and fusion, the SAHR-CapsNet can extract semantically strong and spatially accurate feature semantics at multiple scales. Second, integrated with the efficient capsule-based self-attention module, the SAHR-CapsNet functions promisingly to attend to target-specific spatial features and informative channel features. Finally, adopted with the capsule-based sparse anchoring network, the SAHR-CapsNet performs efficiently in generating a fixed number of lightweight, high-quality sparse region proposals. Quantitative assessments and comparative analyses on two challenging remote sensing image datasets demonstrate the applicability and effectiveness of the developed SAHR-CapsNet for geospatial object detection applications.
- Published
- 2021
34. Airborne multispectral LiDAR point cloud classification with a feature Reasoning-based graph convolution network
- Author
-
José Marcato Junior, Haiyan Guan, Kyle Gao, Dilong Li, Yongtao Yu, Peiran Zhao, Jonathan Li, and Hanyun Wang
- Subjects
Global and Planetary Change ,business.industry ,Computer science ,Multispectral image ,Pattern recognition ,Management, Monitoring, Policy and Law ,Lidar point cloud ,Convolution ,Feature (computer vision) ,Graph (abstract data type) ,Artificial intelligence ,Computers in Earth Sciences ,business ,Earth-Surface Processes - Published
- 2021
35. Collective topical PageRank: a model to evaluate the topic-dependent academic impact of scientific papers
- Author
-
Yongjun Zhang, Jialin Ma, Yongtao Yu, Bolun Chen, and Zijian Wang
- Subjects
Topic model ,Computer science ,media_common.quotation_subject ,General Social Sciences ,02 engineering and technology ,Area of interest ,Library and Information Sciences ,Data science ,Computer Science Applications ,law.invention ,PageRank ,law ,020204 information systems ,Reading (process) ,Academic writing ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Pagerank algorithm ,media_common - Abstract
With the explosive growth of academic writing, it is difficult for researchers to find significant papers in their area of interest. In this paper, we propose a pipeline model, named collective topical PageRank, to evaluate the topic-dependent impact of scientific papers. First, we fit the model to a correlation topic model based on the textual content of papers to extract scientific topics and correlations. Then, we present a modified PageRank algorithm, which incorporates the venue, the correlations of the scientific topics, and the publication year of each paper into a random walk to evaluate the paper’s topic-dependent academic impact. Our experiments showed that the model can effectively identify significant papers as well as venues for each scientific topic, recommend papers for further reading or citing, explore the evolution of scientific topics, and calculate the venues’ dynamic topic-dependent academic impact.
- Published
- 2017
36. A self-attention based faster R-CNN for polyp detection from colonoscopy images
- Author
-
Bolun Chen, Jing-Jing Wan, Yongtao Yu, Tai-Yue Chen, and Min Ji
- Subjects
medicine.diagnostic_test ,Computer science ,business.industry ,media_common.quotation_subject ,Self attention ,Feature extraction ,Biomedical Engineering ,Process (computing) ,Colonoscopy ,Health Informatics ,Pattern recognition ,digestive system diseases ,Feature (computer vision) ,Signal Processing ,medicine ,Key (cryptography) ,Quality (business) ,Artificial intelligence ,business ,Abstraction (linguistics) ,media_common - Abstract
At present, the incidence rate of colorectal cancer (CRC) is increasing year by year. It has always affected people's physical and mental health and quality of life. How to improve the detection ability of polyp plays a key role in colonoscopy. In order to solve these problems, in this paper, we first enhance the contrast of the input image by well distinguishing the foreground from the background in order to improve the saliency of the polyp regions. Then, we feed the enhanced image into an improved Faster R-CNN architecture comprised three processing modules for feature extraction, region proposal generation, and polyp detection, respectively. In order to further improve the quality, as well as the feature abstraction capability of the feature maps produced by the feature extraction network, we append an attention module to attend to the useful feature channels and weaken the contributions of the helpless feature channels. The experimental results demonstrate that the accuracy of the proposed polyp detection network is greatly improved compared with the existing algorithms, and the network not only can accurately identify polyps of varying sizes and conditions in single polyp images, but also can achieve excellent performance in handling multiple polyp images. This paper will be greatly helpful to alleviate the missed diagnosis of clinicians in the process of endoscopic examination and disease treatment, as well as providing effective assistance for the early diagnosis, treatment and prevention of the CRC, which is also of great significance to the clinical work of physicians.
- Published
- 2021
37. Projector undistortion for high-accuracy fringe projection profilometry
- Author
-
Long Xu, Jian Wang, Yongtao Yu, Liping Zhou, and Yaping Cao
- Subjects
Optics ,Projector ,law ,business.industry ,Fringe projection profilometry ,Computer science ,Applied Mathematics ,business ,Instrumentation ,Engineering (miscellaneous) ,law.invention - Published
- 2021
38. Characteristic extraction of rolling bearing compound faults of aero-engine
- Author
-
Zhigang Feng, Yongtao Yu, Jiajing Huang, and Mingyue Yu
- Subjects
0209 industrial biotechnology ,Computer science ,Cyclostationary process ,lcsh:Mechanical engineering and machinery ,aero-engine ,02 engineering and technology ,compound faults ,Fault (power engineering) ,Signal ,law.invention ,symbols.namesake ,020901 industrial engineering & automation ,law ,Control theory ,0202 electrical engineering, electronic engineering, information engineering ,lcsh:TJ1-1570 ,General Materials Science ,single-channel ,Signal processing ,Bearing (mechanical) ,Mechanical Engineering ,020208 electrical & electronic engineering ,Autocorrelation ,Wavelet transform ,rolling bearing ,vibration signal ,symbols ,Hilbert transform - Abstract
Rolling bearing’s fault mode usually shows compound faults in aero-engine. The compound faults characteristics are more complex than single one, and many signal analysis methods have rather great limitation for compound fault characteristic extraction which leads to the difficulty to monitor the running state of rolling bearing in aero-engine. Based on above analysis, a method of combining wavelet transform with cyclostationary theory, autocorrelation function and Hilbert transform is proposed and applied to extract characteristic frequency of rolling bearing from compound faults mode only according to single-channel vibration acceleration signal of aero-engine. Meanwhile, a consideration is given to the influence of sensor installation position, compound fault types in the extraction of compound faults characteristics. The result indicates that the proposed new method can effectively monitor rolling bearing running state in four different compound fault modes just according to single-channel vibration acceleration signal no matter sensors are installed in horizontal or vertical direction.
- Published
- 2017
39. Aeroplane detection in very-high-resolution images using deep feature representation and rotation-invariant Hough forests
- Author
-
Haiyan Guan, Dilong Li, Liang Zhong, and Yongtao Yu
- Subjects
Superpixel segmentation ,Very high resolution ,Correctness ,Computer science ,Machine vision ,business.industry ,Deep learning ,0211 other engineering and technologies ,Pattern recognition ,02 engineering and technology ,0202 electrical engineering, electronic engineering, information engineering ,General Earth and Planetary Sciences ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,Invariant (mathematics) ,business ,Encoder ,021101 geological & geomatics engineering - Abstract
This letter proposes a processing chain for detecting aeroplanes from very high-resolution (VHR) remotely sensed images with the fusion of deep feature representation and rotation-invariant Hough forests. First, superpixel segmentation is used to generate meaningful and non-redundant patches. Second, deep learning techniques are exploited to construct a multi-layer feature encoder for representing high-order features of patches. Third, a set of multi-scale rotation-invariant Hough forests are trained to detect aeroplanes of varying orientations and sizes. Experiments show that the proposed method is a promising solution for detecting aeroplanes from VHR remotely sensed images, with a completeness, correctness, and F-measure of 0.956, 0.970, and 0.963, respectively. Comparative studies with four existing methods also demonstrate that the proposed method outperforms the other existing methods in accurately detecting aeroplanes of varying appearances, orientations, and sizes.
- Published
- 2017
40. Rapid Localization and Extraction of Street Light Poles in Mobile LiDAR Point Clouds: A Supervoxel-Based Approach
- Author
-
Chenglu Wen, Cheng Wang, Fan Wu, Jingjing Wang, Yongtao Yu, Yulan Guo, and Jonathan Li
- Subjects
050210 logistics & transportation ,business.industry ,Computer science ,Mechanical Engineering ,05 social sciences ,Feature extraction ,0211 other engineering and technologies ,Point cloud ,02 engineering and technology ,Computer Science Applications ,Random forest ,Support vector machine ,Lidar ,0502 economics and business ,Automotive Engineering ,Trajectory ,Preprocessor ,Segmentation ,Computer vision ,Artificial intelligence ,business ,021101 geological & geomatics engineering - Abstract
This paper presents a supervoxel-based approach for automated localization and extraction of street light poles in point clouds acquired by a mobile LiDAR system. The method consists of five steps: preprocessing, localization, segmentation, feature extraction, and classification. First, the raw point clouds are divided into segments along the trajectory, the ground points are removed, and the remaining points are segmented into supervoxels. Then, a robust localization method is proposed to accurately identify the pole-like objects. Next, a localization-guided segmentation method is proposed to obtain pole-like objects. Subsequently, the pole features are classified using the support vector machine and random forests. The proposed approach was evaluated on three datasets with 1,055 street light poles and 701 million points. Experimental results show that our localization method achieved an average recall value of 98.8%. A comparative study proved that our method is more robust and efficient than other existing methods for localization and extraction of street light poles.
- Published
- 2017
41. Identifying Subscribers in Freemium E-commerce Model Based on Support Vector Classification
- Author
-
Bo Wang, Yongtao Yu, and Xiaodan Yu
- Subjects
Service (business) ,Computer science ,business.industry ,Information technology ,02 engineering and technology ,E-commerce ,Freemium ,Task (project management) ,Support vector machine ,World Wide Web ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,General Earth and Planetary Sciences ,020201 artificial intelligence & image processing ,business ,General Environmental Science - Abstract
Advances in information technologies have brought many changes to our lives. Finding free music through online platform rather than buying hard copies offline is one of the most significant changes. Recent researches suggest that as users becoming more engaged with the online content-provider platform, they are more willing to pay for the service or premium service. This study addresses the need of freemium e-commerce identifying potential subscribers. In specific, we propose a novel method, namely support vector classification (SVC), to categorize content viewers into potential subscriber and non-potential subscribers. Our method provides satisfied prediction result and the experiment showed that SVC is a superior method in this kind of task.
- Published
- 2017
42. Bag of Contextual-Visual Words for Road Scene Object Detection From Mobile Laser Scanning Data
- Author
-
Cheng Wang, Chenglu Wen, Jonathan Li, Haiyan Guan, and Yongtao Yu
- Subjects
050210 logistics & transportation ,Vocabulary ,Similarity (geometry) ,business.industry ,Computer science ,Mechanical Engineering ,media_common.quotation_subject ,05 social sciences ,0211 other engineering and technologies ,Point cloud ,02 engineering and technology ,Object (computer science) ,Object detection ,Mobile laser scanning ,Computer Science Applications ,Feature (computer vision) ,0502 economics and business ,Automotive Engineering ,Computer vision ,Visual Word ,Artificial intelligence ,business ,021101 geological & geomatics engineering ,media_common - Abstract
This paper proposes a novel algorithm for detecting road scene objects (e.g., light poles, traffic signposts, and cars) from 3-D mobile-laser-scanning point cloud data for transportation-related applications. To describe local abstract features of point cloud objects, a contextual visual vocabulary is generated by integrating spatial contextual information of feature regions. Objects of interest are detected based on the similarity measures of the bag of contextual-visual words between the query object and the segmented semantic objects. Quantitative evaluations on two selected data sets show that the proposed algorithm achieves an average recall, precision, quality, and F-score of 0.949, 0.970, 0.922, and 0.959, respectively, in detecting light poles, traffic signposts, and cars. Comparative studies demonstrate the superior performance of the proposed algorithm over other existing methods.
- Published
- 2016
43. Affine-Function Transformation-Based Object Matching for Vehicle Detection from Unmanned Aerial Vehicle Imagery
- Author
-
Wanqian Yan, Daifeng Peng, Shuang Cao, Haiyan Guan, and Yongtao Yu
- Subjects
Matching (statistics) ,Correctness ,superpixel segmentation ,Computer science ,0211 other engineering and technologies ,02 engineering and technology ,Image (mathematics) ,remote sensing imagery ,0202 electrical engineering, electronic engineering, information engineering ,unmanned aerial vehicle ,Computer vision ,lcsh:Science ,021101 geological & geomatics engineering ,object matching ,Orientation (computer vision) ,business.industry ,Thresholding ,Transformation (function) ,General Earth and Planetary Sciences ,vehicle detection ,020201 artificial intelligence & image processing ,lcsh:Q ,Affine transformation ,Artificial intelligence ,Scale (map) ,business - Abstract
Vehicle detection from remote sensing images plays a significant role in transportation related applications. However, the scale variations, orientation variations, illumination variations, and partial occlusions of vehicles, as well as the image qualities, bring great challenges for accurate vehicle detection. In this paper, we present an affine-function transformation-based object matching framework for vehicle detection from unmanned aerial vehicle (UAV) images. First, meaningful and non-redundant patches are generated through a superpixel segmentation strategy. Then, the affine-function transformation-based object matching framework is applied to a vehicle template and each of the patches for vehicle existence estimation. Finally, vehicles are detected and located after matching cost thresholding, vehicle location estimation, and multiple response elimination. Quantitative evaluations on two UAV image datasets show that the proposed method achieves an average completeness, correctness, quality, and F1-measure of 0.909, 0.969, 0.883, and 0.938, respectively. Comparative studies also demonstrate that the proposed method achieves compatible performance with the Faster R-CNN and outperforms the other eight existing methods in accurately detecting vehicles of various conditions.
- Published
- 2019
44. An Automated Hierarchical Approach for Three-Dimensional Segmentation of Single Trees Using UAV LiDAR Data
- Author
-
Wanqian Yan, Haiyan Guan, Sha Gao, Lin Cao, Jianyong Lu, and Yongtao Yu
- Subjects
Correctness ,010504 meteorology & atmospheric sciences ,Computer science ,Science ,0211 other engineering and technologies ,Point cloud ,02 engineering and technology ,single tree ,computer.software_genre ,01 natural sciences ,mean shift ,UAV LiDAR ,Voxel ,Segmentation ,Mean-shift ,improved normalized cut ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences ,business.industry ,segmentation ,Pattern recognition ,Tree (data structure) ,Lidar ,General Earth and Planetary Sciences ,Data pre-processing ,Artificial intelligence ,business ,computer - Abstract
Forests play a key role in terrestrial ecosystems, and the variables extracted from single trees can be used in various fields and applications for evaluating forest production and assessing forest ecosystem services. In this study, we developed an automated hierarchical single-tree segmentation approach based on the high density three-dimensional (3D) Unmanned Aerial Vehicle (UAV) point clouds. First, this approach obtains normalized non-ground UAV points in data preprocessing; then, a voxel-based mean shift algorithm is used to roughly classify the non-ground UAV points into well-detected and under-segmentation clusters. Moreover, potential tree apices for each under-segmentation cluster are obtained with regard to profile shape curves and finally input to the normalized cut segmentation (NCut) algorithm to segment iteratively the under-segmentation cluster into single trees. We evaluated the proposed method using datasets acquired by a Velodyne 16E LiDAR system mounted on a multi-rotor UAV. The results showed that the proposed method achieves the average correctness, completeness, and overall accuracy of 0.90, 0.88, and 0.89, respectively, in delineating single trees. Comparative analysis demonstrated that our method provided a promising solution to reliable and robust segmentation of single trees from UAV LiDAR data with high point cloud density.
- Published
- 2018
45. Automated Detection of Three-Dimensional Cars in Mobile Laser Scanning Point Clouds Using DBM-Hough-Forests
- Author
-
Haiyan Guan, Jonathan Li, Cheng Wang, and Yongtao Yu
- Subjects
Computer science ,Orientation (computer vision) ,business.industry ,0211 other engineering and technologies ,Point cloud ,Centroid ,02 engineering and technology ,Solid modeling ,Object detection ,Hough transform ,law.invention ,law ,Feature (computer vision) ,0202 electrical engineering, electronic engineering, information engineering ,General Earth and Planetary Sciences ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,021101 geological & geomatics engineering - Abstract
This paper presents an automated algorithm for rapidly and effectively detecting cars directly from large-volume 3-D point clouds. Rather than using low-order descriptors, a multilayer feature generation model is created to obtain high-order feature representations for 3-D local patches through deep learning techniques. To handle cars with different levels of incompleteness caused by data acquisition ways and occlusions, a hierarchical visibility estimation model is developed to augment Hough voting. Considering scale and orientation variations in the azimuth direction, a set of multiscale Hough forests is constructed to rotationally cast votes to estimate cars' centroids. Quantitative assessments show that the proposed algorithm achieves average completeness, correctness, quality, and $F_{1}$ -measure of 0.94, 0.96, 0.90, and 0.95, respectively, in detecting 3-D cars. Comparative studies also demonstrate that the proposed algorithm outperforms the other four existing algorithms in accurately and completely detecting 3-D cars from large-scale 3-D point clouds.
- Published
- 2016
46. Patch-Based Semantic Labeling of Road Scene Using Colorized Mobile LiDAR Point Clouds
- Author
-
Hanyun Wang, Zhipeng Cai, Jonathan Li, Cheng Wang, Huan Luo, Ziyi Chen, Chenglu Wen, and Yongtao Yu
- Subjects
050210 logistics & transportation ,Digital mapping ,Computer science ,business.industry ,Mechanical Engineering ,05 social sciences ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,0211 other engineering and technologies ,Point cloud ,Markov process ,Context (language use) ,02 engineering and technology ,Computer Science Applications ,symbols.namesake ,Lidar ,0502 economics and business ,Automotive Engineering ,symbols ,Graph (abstract data type) ,Computer vision ,Artificial intelligence ,Mobile telephony ,business ,Intelligent transportation system ,021101 geological & geomatics engineering - Abstract
Semantic labeling of road scenes using colorized mobile LiDAR point clouds is of great significance in a variety of applications, particularly intelligent transportation systems. However, many challenges, such as incompleteness of objects caused by occlusion, overlapping between neighboring objects, interclass local similarities, and computational burden brought by a huge number of points, make it an ongoing open research area. In this paper, we propose a novel patch-based framework for labeling road scenes of colorized mobile LiDAR point clouds. In the proposed framework, first, three-dimensional (3-D) patches extracted from point clouds are used to construct a 3-D patch-based match graph structure (3D-PMG), which transfers category labels from labeled to unlabeled point cloud road scenes efficiently. Then, to rectify the transferring errors caused by local patch similarities in different categories, contextual information among 3-D patches is exploited by combining 3D-PMG with Markov random fields. In the experiments, the proposed framework is validated on colorized mobile LiDAR point clouds acquired by the RIEGL VMX-450 mobile LiDAR system. Comparative experiments show the superior performance of the proposed framework for accurate semantic labeling of road scenes.
- Published
- 2016
47. Pole-Like Road Object Detection in Mobile LiDAR Data via Supervoxel and Bag-of-Contextual-Visual-Words Representation
- Author
-
Jonathan Li, Haiyan Guan, Yongtao Yu, and Pengfei Liu
- Subjects
010504 meteorology & atmospheric sciences ,Computer science ,business.industry ,0211 other engineering and technologies ,02 engineering and technology ,Geotechnical Engineering and Engineering Geology ,Object (computer science) ,01 natural sciences ,Object detection ,Visualization ,Lidar ,Feature (computer vision) ,Viola–Jones object detection framework ,Segmentation ,Computer vision ,Visual Word ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences - Abstract
This letter addresses the problem of detecting pole-like road objects (including light poles and traffic signposts) from mobile light detection and ranging (LiDAR) data for transportation-related applications. The method consists of two consecutive stages: training and pole-like object detection. At the training stage, a contextual visual vocabulary is created from the feature regions generated from a training data set by supervoxel segmentation. At the pole-like object detection stage, a bag-of-contextual-visual-words representation is generated for each semantic object segmented from mobile LiDAR data. The experimental results show that the proposed method achieves correctness, omission, and commission of 88.9%, 11.1%, and 2.8%, respectively, in detecting pole-like road objects. Computational complexity analysis demonstrates that our method provides a promising and effective solution to rapid and accurate detection of pole-like objects from large volumes of mobile LiDAR data.
- Published
- 2016
48. Bag-of-visual-phrases and hierarchical deep models for traffic sign detection and recognition in mobile laser scanning data
- Author
-
Haiyan Guan, Huan Luo, Chenglu Wen, Jonathan Li, Cheng Wang, and Yongtao Yu
- Subjects
050210 logistics & transportation ,business.industry ,Computer science ,05 social sciences ,Boltzmann machine ,Point cloud ,02 engineering and technology ,Atomic and Molecular Physics, and Optics ,Computer Science Applications ,Traffic sign detection ,Hierarchical classifier ,Task (computing) ,0502 economics and business ,0202 electrical engineering, electronic engineering, information engineering ,Feature (machine learning) ,Traffic sign recognition ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,Computers in Earth Sciences ,business ,Engineering (miscellaneous) ,Encoder - Abstract
This paper presents a novel algorithm for detection and recognition of traffic signs in mobile laser scanning (MLS) data for intelligent transportation-related applications. The traffic sign detection task is accomplished based on 3-D point clouds by using bag-of-visual-phrases representations; whereas the recognition task is achieved based on 2-D images by using a Gaussian-Bernoulli deep Boltzmann machine-based hierarchical classifier. To exploit high-order feature encodings of feature regions, a deep Boltzmann machine-based feature encoder is constructed. For detecting traffic signs in 3-D point clouds, the proposed algorithm achieves an average recall, precision, quality, and F-score of 0.956, 0.946, 0.907, and 0.951, respectively, on the four selected MLS datasets. For on-image traffic sign recognition, a recognition accuracy of 97.54% is achieved by using the proposed hierarchical classifier. Comparative studies with the existing traffic sign detection and recognition methods demonstrate that our algorithm obtains promising, reliable, and high performance in both detecting traffic signs in 3-D point clouds and recognizing traffic signs on 2-D images.
- Published
- 2016
49. Rotation-and-scale-invariant airplane detection in high-resolution satellite images based on deep-Hough-forests
- Author
-
Dawei Zai, Yongtao Yu, Haiyan Guan, and Zheng Ji
- Subjects
business.product_category ,Scale (ratio) ,Computer science ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,0211 other engineering and technologies ,02 engineering and technology ,Hough transform ,law.invention ,Airplane ,law ,0202 electrical engineering, electronic engineering, information engineering ,Computer vision ,Computers in Earth Sciences ,Engineering (miscellaneous) ,021101 geological & geomatics engineering ,Orientation (computer vision) ,business.industry ,Deep learning ,Centroid ,Atomic and Molecular Physics, and Optics ,Computer Science Applications ,Feature (computer vision) ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,Rotation (mathematics) - Abstract
This paper proposes a rotation-and-scale-invariant method for detecting airplanes from high-resolution satellite images. To improve feature representation capability, a multi-layer feature generation model is created to produce high-order feature representations for local image patches through deep learning techniques. To effectively estimate airplane centroids, a Hough forest model is trained to learn mappings from high-order patch features to the probabilities of an airplane being present at specific locations. To handle airplanes with varying orientations, patch orientation is defined and integrated into the Hough forest to augment Hough voting. The scale invariance is achieved by using a set of scale factors embedded in the Hough forest. Quantitative evaluations on the images collected from Google Earth service show that the proposed method achieves a completeness, correctness, quality, and F1-measure of 0.968, 0.972, 0.942, and 0.970, respectively, in detecting airplanes with arbitrary orientations and sizes. Comparative studies also demonstrate that the proposed method outperforms the other three existing methods in accurately and completely detecting airplanes in high-resolution remotely sensed images.
- Published
- 2016
50. Automated Detection of Urban Road Manhole Covers Using Mobile Laser Scanning Data
- Author
-
Zheng Ji, Yongtao Yu, and Haiyan Guan
- Subjects
Digital mapping ,Computer science ,business.industry ,Mechanical Engineering ,Feature extraction ,Point cloud ,Image segmentation ,Computer Science Applications ,Random forest ,Road surface ,Automotive Engineering ,Computer vision ,Algorithm design ,Artificial intelligence ,business ,Interpolation - Abstract
This paper proposes a novel framework for automated detection of urban road manhole covers using mobile laser scanning (MLS) data. First, to narrow searching regions and reduce the computational complexity, road surface points are segmented from a raw point cloud via a curb-based road surface segmentation approach and rasterized into a georeferenced intensity image through inverse distance weighted interpolation. Then, a supervised deep learning model is developed to construct a multilayer feature generation model for depicting high-order features of local image patches. Next, a random forest model is trained to learn mappings from high-order patch features to the probabilities of the existence of urban road manhole covers centered at specific locations. Finally, urban road manhole covers are detected from georeferenced intensity images based on the multilayer feature generation model and random forest model. Quantitative evaluations show that the proposed algorithm achieves an average completeness, correctness, quality, and $\mathrm{F}_1$ -measure of 0.955, 0.959, 0.917, and 0.957, respectively, in detecting urban road manhole covers from georeferenced intensity images. Comparative studies demonstrate the advantageous performance of the proposed algorithm over other existing methods for rapid and automated detection of urban road manhole covers using MLS data.
- Published
- 2015
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.