1,123 results on '"lightweight network"'
Search Results
2. LGCANet: lightweight hand pose estimation network based on HRNet.
- Author
-
Pan, Xiaoying, Li, Shoukun, Wang, Hao, Wang, Beibei, and Wang, Haoyi
- Subjects
- *
FEATURE extraction , *DEEP learning , *COMPUTER vision , *APPLICATION software , *VIRTUAL reality , *COMPUTATIONAL complexity , *AUTONOMOUS vehicles - Abstract
Hand pose estimation is a fundamental task in computer vision with applications in virtual reality, gesture recognition, autonomous driving, and virtual surgery. Keypoint detection often relies on deep learning methods and high-resolution feature map representations to achieve accurate detection. The HRNet framework serves as the basis, but it presents challenges in terms of extensive parameter count and demanding computational complexity due to high-resolution representations. To mitigate these challenges, we propose a lightweight keypoint detection network called LGCANet (Lightweight Ghost-Coordinate Attention Network). This network primarily consists of a lightweight feature extraction head for initial feature extraction and multiple lightweight foundational network modules called GCAblocks. GCAblocks introduce linear transformations to generate redundant feature maps while concurrently considering inter-channel relationships and long-range positional information using a coordinate attention mechanism. Validation on the RHD dataset and the COCO-WholeBody-Hand dataset shows that LGCANet reduces the number of parameters by 65.9% and GFLOPs by 72.6% while preserving the accuracy and improves the detection speed. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. A Highly Efficient and Lightweight Detection Method for Steel Surface Defect.
- Author
-
Xu, Changyu, Li, Jie, and Li, Xianguo
- Subjects
- *
SURFACE defects , *LIGHTWEIGHT steel , *STEEL , *FEATURE extraction - Abstract
The detection of steel surface defects is of great significance to steel production. In order to better meet the requirements of accuracy, real-time, and lightweight model, this paper proposes a highly efficient and lightweight steel surface defect detection method based on YOLOv5n. Firstly, ODMobileNetV2 composed of MobileNetV2 and ODConv is used as the backbone to improve the defect feature extraction capability. Secondly, GSConv is utilized in the neck to achieve deep information fusion through channel concatenation and shuffling, enhancing the ability of feature fusion. Finally, this paper proposes a spatial-channel reconstruction block (SCRB) designed to suppress redundant features and improve the representation ability of defect features through feature separation and reconstruction. Experimental results show that this method achieves 84.1% mAP and 109 FPS on the NEU-DET dataset, and 72.9% mAP and 110.1 FPS on the GC10-DET dataset, enabling accurate and efficient detection. Furthermore, the number of parameters is only 5.04M, which has a significant lightweight advantage. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. Fast-Activated Minimal Gated Unit: Lightweight Processing and Feature Recognition for Multiple Mechanical Impact Signals.
- Author
-
Wang, Wenrui, Han, Dong, Duan, Xinyi, Yong, Yaxin, Wu, Zhengqing, Ma, Xiang, Zhang, He, and Dai, Keren
- Subjects
- *
MULTIBODY systems , *WAVELET transforms , *DYNAMICAL systems , *COMPUTATIONAL complexity , *SIGNALS & signaling - Abstract
Multiple dynamic impact signals are widely used in a variety of engineering scenarios and are difficult to identify accurately and quickly due to the signal adhesion phenomenon caused by nonlinear interference. To address this problem, an intelligent algorithm combining wavelet transforms with lightweight neural networks is proposed. First, the features of multiple impact signals are analyzed by establishing a transfer model for multiple impacts in multibody dynamical systems, and interference is suppressed using wavelet transformation. Second, a lightweight neural network, i.e., fast-activated minimal gated unit (FMGU), is elaborated for multiple impact signals, which can reduce computational complexity and improve real-time performance. Third, the experimental results show that the proposed method maintains excellent feature recognition results compared to gate recurrent unit (GRU) and long short-term memory (LSTM) networks under all test datasets with varying impact speeds, while its metrics for computational complexity are 50% lower than those of the GRU and LSTM. Therefore, the proposed method is of great practical value for weak hardware application platforms that require the accurate identification of multiple dynamic impact signals in real time. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. LTSCD-YOLO: A Lightweight Algorithm for Detecting Typical Satellite Components Based on Improved YOLOv8.
- Author
-
Tang, Zixuan, Zhang, Wei, Li, Junlin, Liu, Ran, Xu, Yansong, Chen, Siyu, Fang, Zhiyue, and Zhao, Fuchenglong
- Subjects
- *
SPACE environment , *EXTRATERRESTRIAL resources , *ALGORITHMS , *GENERALIZATION , *NECK - Abstract
Typical satellite component detection is an application-valuable and challenging research field. Currently, there are many algorithms for detecting typical satellite components, but due to the limited storage space and computational resources in the space environment, these algorithms generally have the problem of excessive parameter count and computational load, which hinders their effective application in space environments. Furthermore, the scale of datasets used by these algorithms is not large enough to train the algorithm models well. To address the above issues, this paper first applies YOLOv8 to the detection of typical satellite components and proposes a Lightweight Typical Satellite Components Detection algorithm based on improved YOLOv8 (LTSCD-YOLO). Firstly, it adopts the lightweight network EfficientNet-B0 as the backbone network to reduce the model's parameter count and computational load; secondly, it uses a Cross-Scale Feature-Fusion Module (CCFM) at the Neck to enhance the model's adaptability to scale changes; then, it integrates Partial Convolution (PConv) into the C2f (Faster Implementation of CSP Bottleneck with two convolutions) module and Re-parameterized Convolution (RepConv) into the detection head to further achieve model lightweighting; finally, the Focal-Efficient Intersection over Union (Focal-EIoU) is used as the loss function to enhance the model's detection accuracy and detection speed. Additionally, a larger-scale Typical Satellite Components Dataset (TSC-Dataset) is also constructed. Our experimental results show that LTSCD-YOLO can maintain high detection accuracy with minimal parameter count and computational load. Compared to YOLOv8s, LTSCD-YOLO improved the mean average precision (mAP50) by 1.50% on the TSC-Dataset, reaching 94.5%. Meanwhile, the model's parameter count decreased by 78.46%, the computational load decreased by 65.97%, and the detection speed increased by 17.66%. This algorithm achieves a balance between accuracy and light weight, and its generalization ability has been validated on real images, making it effectively applicable to detection tasks of typical satellite components in space environments. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
6. 城市低空小型无人机目标实时高精度检测算法.
- Author
-
崔勇强, 黄谦, 高雪, 梅涛, 白迪, and 王晓磊
- Abstract
Copyright of Journal of Computer Engineering & Applications is the property of Beijing Journal of Computer Engineering & Applications Journal Co Ltd. and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
- Full Text
- View/download PDF
7. YOLOv5s-BiPCNeXt, a Lightweight Model for Detecting Disease in Eggplant Leaves.
- Author
-
Xie, Zhedong, Li, Chao, Yang, Zhuang, Zhang, Zhen, Jiang, Jiazhuo, and Guo, Hongyu
- Subjects
INTERPOLATION algorithms ,PLANT diseases ,PLANT identification ,COMPUTATIONAL complexity ,DISEASE mapping ,EGGPLANT - Abstract
Ensuring the healthy growth of eggplants requires the precise detection of leaf diseases, which can significantly boost yield and economic income. Improving the efficiency of plant disease identification in natural scenes is currently a crucial issue. This study aims to provide an efficient detection method suitable for disease detection in natural scenes. A lightweight detection model, YOLOv5s-BiPCNeXt, is proposed. This model utilizes the MobileNeXt backbone to reduce network parameters and computational complexity and includes a lightweight C3-BiPC neck module. Additionally, a multi-scale cross-spatial attention mechanism (EMA) is integrated into the neck network, and the nearest neighbor interpolation algorithm is replaced with the content-aware feature recombination operator (CARAFE), enhancing the model's ability to perceive multidimensional information and extract multiscale disease features and improving the spatial resolution of the disease feature map. These improvements enhance the detection accuracy for eggplant leaves, effectively reducing missed and incorrect detections caused by complex backgrounds and improving the detection and localization of small lesions at the early stages of brown spot and powdery mildew diseases. Experimental results show that the YOLOv5s-BiPCNeXt model achieves an average precision (AP) of 94.9% for brown spot disease, 95.0% for powdery mildew, and 99.5% for healthy leaves. Deployed on a Jetson Orin Nano edge detection device, the model attains an average recognition speed of 26 FPS (Frame Per Second), meeting real-time requirements. Compared to other algorithms, YOLOv5s-BiPCNeXt demonstrates superior overall performance, accurately detecting plant diseases under natural conditions and offering valuable technical support for the prevention and treatment of eggplant leaf diseases. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. RS-Xception: A Lightweight Network for Facial Expression Recognition.
- Author
-
Liao, Liefa, Wu, Shouluan, Song, Chao, and Fu, Jianglong
- Subjects
EMOTION recognition ,FACIAL expression ,ARTIFICIAL intelligence ,SENTIMENT analysis ,ACTING education - Abstract
Facial expression recognition (FER) utilizes artificial intelligence for the detection and analysis of human faces, with significant applications across various scenarios. Our objective is to deploy the facial emotion recognition network on mobile devices and extend its application to diverse areas, including classroom effect monitoring, human–computer interaction, specialized training for athletes (such as in figure skating and rhythmic gymnastics), and actor emotion training. Recent studies have employed advanced deep learning models to address this task, though these models often encounter challenges like subpar performance and an excessive number of parameters that do not align with the requirements of FER for embedded devices. To tackle this issue, we have devised a lightweight network structure named RS-Xception, which is straightforward yet highly effective. Drawing on the strengths of ResNet and SENet, this network integrates elements from the Xception architecture. Our models have been trained on FER2013 datasets and demonstrate superior efficiency compared to conventional network models. Furthermore, we have assessed the model's performance on the CK+, FER2013, and Bigfer2013 datasets, achieving accuracy rates of 97.13%, 69.02%, and 72.06%, respectively. Evaluation on the complex RAF-DB dataset yielded an accuracy rate of 82.98%. The incorporation of transfer learning notably enhanced the model's accuracy, with a performance of 75.38% on the Bigfer2013 dataset, underscoring its significance in our research. In conclusion, our proposed model proves to be a viable solution for precise sentiment detection and estimation. In the future, our lightweight model may be deployed on embedded devices for research purposes. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. Lightweight Road Damage Detection Network Based on YOLOv5.
- Author
-
Jingwei Zhao, Ye Tao, Zhixian Zhang, Chao Huang, and Wenhua Cui
- Abstract
The field of computer vision has experienced rapid progress owing to deep learning. The importance of road damage detection in ensuring traffic safety and reducing road maintenance costs is becoming increasingly evident. For detecting road damage, the YOLOv5 algorithm provides a reliable and effective method. However, YOLOv5 still requires a significant amount of computation. This paper proposes a lightweight network for detecting road damage that improves upon the YOLOv5 model in four ways. The algorithm accurately identifies and classifies different types of road damage, while simultaneously reducing the number of parameters and required computations. First, lightweight processing of the model is achieved. The Ghost module and Ghost Bottleneck are employed to construct the novel GBS module and C3Ghost, which replace the existing CBS and C3 modules. Second, the CIoU loss function is transformed into SIoU to improve the precision of target box regression. Furthermore, the original upsampling module is replaced by CARAFE to improve the model's semantic adaptability and receptive field. Finally, the CBAM attention mechanism is employed to concentrate on crucial feature information. The experiment's findings present that, in comparison to the baseline model, the upgraded model has 41.8% fewer parameters. Additionally, there has been a 43.8% reduction in floating-point computation and an improvement of 0.2% in detection accuracy. [ABSTRACT FROM AUTHOR]
- Published
- 2024
10. 基于注意力和重构特征融合的轻量级煤矿安全帽检测方法.
- Author
-
董彦强, 程德强, 张云鹤, 寇旗旗, and 张皓翔
- Abstract
Copyright of Journal of Computer Engineering & Applications is the property of Beijing Journal of Computer Engineering & Applications Journal Co Ltd. and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
- Full Text
- View/download PDF
11. Image super‐resolution via dynamic network.
- Author
-
Tian, Chunwei, Zhang, Xuanyu, Zhang, Qi, Yang, Mingming, and Ju, Zhaojie
- Subjects
HIGH resolution imaging ,CONVOLUTIONAL neural networks ,DIGITAL technology - Abstract
Convolutional neural networks depend on deep network architectures to extract accurate information for image super‐resolution. However, obtained information of these convolutional neural networks cannot completely express predicted high‐quality images for complex scenes. A dynamic network for image super‐resolution (DSRNet) is presented, which contains a residual enhancement block, wide enhancement block, feature refinement block and construction block. The residual enhancement block is composed of a residual enhanced architecture to facilitate hierarchical features for image super‐resolution. To enhance robustness of obtained super‐resolution model for complex scenes, a wide enhancement block achieves a dynamic architecture to learn more robust information to enhance applicability of an obtained super‐resolution model for varying scenes. To prevent interference of components in a wide enhancement block, a refinement block utilises a stacked architecture to accurately learn obtained features. Also, a residual learning operation is embedded in the refinement block to prevent long‐term dependency problem. Finally, a construction block is responsible for reconstructing high‐quality images. Designed heterogeneous architecture can not only facilitate richer structural information, but also be lightweight, which is suitable for mobile digital devices. Experimental results show that our method is more competitive in terms of performance, recovering time of image super‐resolution and complexity. The code of DSRNet can be obtained at https://github.com/hellloxiaotian/DSRNet. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
12. A Lightweight YOLOv8 Model for Apple Leaf Disease Detection.
- Author
-
Gao, Lijun, Zhao, Xing, Yue, Xishen, Yue, Yawei, Wang, Xiaoqiang, Wu, Huanhuan, and Zhang, Xuedong
- Subjects
MOBILE apps ,APPLE growing ,PLANT diseases ,COMPUTATIONAL complexity ,ALGORITHMS - Abstract
China holds the top position globally in apple production and consumption. Detecting diseases during the planting process is crucial for increasing yields and promoting the rapid development of the apple industry. This study proposes a lightweight algorithm for apple leaf disease detection in natural environments, which is conducive to application on mobile and embedded devices. Our approach modifies the YOLOv8n framework to improve accuracy and efficiency. Key improvements include replacing conventional Conv layers with GhostConv and parts of the C2f structure with C3Ghost, reducing the model's parameter count, and enhancing performance. Additionally, we integrate a Global attention mechanism (GAM) to improve lesion detection by more accurately identifying affected areas. An improved Bi-Directional Feature Pyramid Network (BiFPN) is also incorporated for better feature fusion, enabling more effective detection of small lesions in complex environments. Experimental results show a 32.9% reduction in computational complexity and a 39.7% reduction in model size to 3.8 M, with performance metrics improving by 3.4% to a mAP@0.5 of 86.9%. Comparisons with popular models like YOLOv7-Tiny, YOLOv6, YOLOv5s, and YOLOv3-Tiny demonstrate that our YOLOv8n–GGi model offers superior detection accuracy, the smallest size, and the best overall performance for identifying critical apple diseases. It can serve as a guide for implementing real-time crop disease detection on mobile and embedded devices. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
13. A Lightweight CER-YOLOv5s Algorithm for Detection of Construction Vehicles at Power Transmission Lines.
- Author
-
Yu, Pingping, Yan, Yuting, Tang, Xinliang, Shang, Yan, and Su, He
- Subjects
ELECTRIC lines ,FEATURE extraction ,PYRAMIDS ,ALGORITHMS - Abstract
In the context of power-line scenarios characterized by complex backgrounds and diverse scales and shapes of targets, and addressing issues such as large model parameter sizes, insufficient feature extraction, and the susceptibility to missing small targets in engineering-vehicle detection tasks, a lightweight detection algorithm termed CER-YOLOv5s is firstly proposed. The C3 module was restructured by embedding a lightweight Ghost bottleneck structure and convolutional attention module, enhancing the model's ability to extract key features while reducing computational costs. Secondly, an E-BiFPN feature pyramid network is proposed, utilizing channel attention mechanisms to effectively suppress background noise and enhance the model's focus on important regions. Bidirectional connections were introduced to optimize the feature fusion paths, improving the efficiency of multi-scale feature fusion. At the same time, in the feature fusion part, an ERM (enhanced receptive module) was added to expand the receptive field of shallow feature maps through multiple convolution repetitions, enhancing the global information perception capability in relation to small targets. Lastly, a Soft-DIoU-NMS suppression algorithm is proposed to improve the candidate box selection mechanism, addressing the issue of suboptimal detection of occluded targets. The experimental results indicated that compared with the baseline YOLOv5s algorithm, the improved algorithm reduced parameters and computations by 27.8% and 31.9%, respectively. The mean average precision (mAP) increased by 2.9%, reaching 98.3%. This improvement surpasses recent mainstream algorithms and suggests stronger robustness across various scenarios. The algorithm meets the lightweight requirements for embedded devices in power-line scenarios. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
14. A Efficient and Accurate UAV Detection Method Based on YOLOv5s.
- Author
-
Feng, Yunsong, Wang, Tong, Jiang, Qiangfu, Zhang, Chi, Sun, Shaohang, and Qian, Wangjiahe
- Subjects
FEATURE extraction ,DRONE aircraft ,ALGORITHMS ,NECK - Abstract
Due to the limited computational resources of portable devices, target detection models for drone detection face challenges in real-time deployment. To enhance the detection efficiency of low, slow, and small unmanned aerial vehicles (UAVs), this study introduces an efficient drone detection model based on YOLOv5s (EDU-YOLO), incorporating lightweight feature extraction and balanced feature fusion modules. The model employs the ShuffleNetV2 network and coordinate attention mechanisms to construct a lightweight backbone network, significantly reducing the number of model parameters. It also utilizes a bidirectional feature pyramid network and ghost convolutions to build a balanced neck network, enriching the model's representational capacity. Additionally, a new loss function, EloU, replaces CIoU to improve the model's positioning accuracy and accelerate network convergence. Experimental results indicate that, compared to the YOLOv5s algorithm, our model only experiences a minimal decrease in mAP by 1.1%, while reducing GFLOPs from 16.0 to 2.2 and increasing FPS from 153 to 188. This provides a substantial foundation for networked optoelectronic detection of UAVs and similar slow-moving aerial targets, expanding the defensive perimeter and enabling earlier warnings. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
15. TinyCount: an efficient crowd counting network for intelligent surveillance.
- Author
-
Lee, Hyeonbeen and Lee, Jangho
- Abstract
Crowd counting, the task of estimating the total number of people in an image, is essential for intelligent surveillance. Integrating a well-trained crowd counting network into edge devices, such as intelligent CCTV systems, enables its application across various domains, including the prevention of crowd collapses and urban planning. For a model to be embedded in edge devices, it requires robust performance, reduced parameter count, and faster response times. This study proposes a lightweight and powerful model called TinyCount, which has only 60k parameters. The proposed TinyCount is a fully convolutional network consisting of a feature extract module (FEM) for robust and rapid feature extraction, a scale perception module (SPM) for scale variation perception and an upsampling module (UM) that adjusts the feature map to the same size as the original image. TinyCount demonstrated competitive performance across three representative crowd counting datasets, despite utilizing approximately 3.33 to 271 times fewer parameters than other crowd counting approaches. The proposed model achieved relatively fast inference times by leveraging the MobileNetV2 architecture with dilated and transposed convolutions. The application of SEblock and findings from existing studies further proved its effectiveness. Finally, we evaluated the proposed TinyCount on multiple edge devices, including the Raspberry Pi 4, NVIDIA Jetson Nano, and NVIDIA Jetson AGX Xavier, to demonstrate its potential for practical applications. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
16. Lightweight improved residual network for efficient inverse tone mapping.
- Author
-
Xue, Liqi, Xu, Tianyi, Song, Yongbao, Liu, Yan, Zhang, Lei, Zhen, Xiantong, and Xu, Jun
- Subjects
IMAGE reconstruction ,IMAGE reconstruction algorithms ,HIGH dynamic range imaging ,EVERYDAY life - Abstract
The display devices like HDR10 televisions are increasingly prevalent in our daily life for visualizing high dynamic range (HDR) images. But the majority of media images on the internet remain in 8-bit standard dynamic range (SDR) format. Therefore, converting SDR images to HDR ones by inverse tone mapping (ITM) is crucial to unlock the full potential of abundant media images. However, existing ITM methods are usually developed with complex network architectures requiring huge computational costs. In this paper, we propose a lightweight Improved Residual Network (IRNet) by enhancing the power of popular residual block for efficient ITM. Specifically, we propose a new Improved Residual Block (IRB) to extract and fuse multi-layer features for fine-grained HDR image reconstruction. Experiments on three benchmark datasets demonstrate that our IRNet achieves state-of-the-art performance on both the ITM and joint SR-ITM tasks. The code, models and data will be publicly available at https://github.com/ThisisVikki/ITM-baseline. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
17. LSRN-AED: lightweight super-resolution network based on asymmetric encoder–decoder.
- Author
-
Huang, Shuying, Li, Wei, Yang, Yong, Wan, Weiguo, and Lai, Houzeng
- Subjects
- *
ARTIFICIAL neural networks , *FEATURE extraction , *TRANSFORMER models , *HIGH resolution imaging - Abstract
Due to limited memory and computing resources, the application of deep neural networks on embedded and mobile devices is still a great challenge. To tackle this problem, this paper proposes a lightweight super-resolution network based on asymmetric encoder–decoder (LSRN-AED), which achieves better performance while reducing model parameters and computation. On the basis of rethinking the roles of encoder and decoder, an asymmetric encoder–decoder (AED) composed of complex encoders and simple decoders is designed to achieve feature extraction and reconstruction. Here, the decoder only adopts one inverted residual block, which can reduce the computational cost of the model and the redundancy of mapping features. For the encoder, inspired by the Transformer structure, an epiphany encoder is designed to realize the feature extraction and representation. In the encoder, a multi-way epiphany attention module (MEAM) is constructed, in which inverted residual blocks are used to replace traditional residual blocks to extract features and reduce model complexity. To realize the selection and enhancement of spatial features, an epiphany attention block (EAB) is designed by exploiting depth-wise convolutions which can learn the significant spatial information of the feature maps. Experimental results demonstrate that the proposed LSRN-AED can achieve better performance at lower parameter cost and outperform some existing state-of-the-art lightweight models. For example, compared to the advanced SMSR method, the proposed LSRN-AED has better evaluation metrics while reducing the number of parameters by 45%, 44%, and 44%, and FLOPs by 44%, 42%, and 41% on the × 2/3/4 SR tasks, respectively. The code will be published on GitHub after our paper is accepted for publication. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
18. Deep Learning-Based Dynamic Region of Interest Autofocus Method for Grayscale Image.
- Author
-
Wang, Yao, Wu, Chuan, Gao, Yunlong, and Liu, Huiying
- Subjects
- *
DEEP learning , *GRAYSCALE model , *COST effectiveness - Abstract
In the field of autofocus for optical systems, although passive focusing methods are widely used due to their cost-effectiveness, fixed focusing windows and evaluation functions in certain scenarios can still lead to focusing failures. Additionally, the lack of datasets limits the extensive research of deep learning methods. In this work, we propose a neural network autofocus method with the capability of dynamically selecting the region of interest (ROI). Our main work is as follows: first, we construct a dataset for automatic focusing of grayscale images; second, we transform the autofocus issue into an ordinal regression problem and propose two focusing strategies: full-stack search and single-frame prediction; and third, we construct a MobileViT network with a linear self-attention mechanism to achieve automatic focusing on dynamic regions of interest. The effectiveness of the proposed focusing method is verified through experiments, and the results show that the focusing MAE of the full-stack search can be as low as 0.094, with a focusing time of 27.8 ms, and the focusing MAE of the single-frame prediction can be as low as 0.142, with a focusing time of 27.5 ms. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
19. ELA-Net: An Efficient Lightweight Attention Network for Skin Lesion Segmentation.
- Author
-
Nie, Tianyu, Zhao, Yishi, and Yao, Shihong
- Subjects
- *
SKIN imaging , *FEATURE extraction , *IMAGE segmentation , *DATA mining , *MEDICAL equipment , *IMAGE processing - Abstract
In clinical conditions limited by equipment, attaining lightweight skin lesion segmentation is pivotal as it facilitates the integration of the model into diverse medical devices, thereby enhancing operational efficiency. However, the lightweight design of the model may face accuracy degradation, especially when dealing with complex images such as skin lesion images with irregular regions, blurred boundaries, and oversized boundaries. To address these challenges, we propose an efficient lightweight attention network (ELANet) for the skin lesion segmentation task. In ELANet, two different attention mechanisms of the bilateral residual module (BRM) can achieve complementary information, which enhances the sensitivity to features in spatial and channel dimensions, respectively, and then multiple BRMs are stacked for efficient feature extraction of the input information. In addition, the network acquires global information and improves segmentation accuracy by putting feature maps of different scales through multi-scale attention fusion (MAF) operations. Finally, we evaluate the performance of ELANet on three publicly available datasets, ISIC2016, ISIC2017, and ISIC2018, and the experimental results show that our algorithm can achieve 89.87%, 81.85%, and 82.87% of the mIoU on the three datasets with a parametric of 0.459 M, which is an excellent balance between accuracy and lightness and is superior to many existing segmentation methods. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
20. Enhanced Hybrid Vision Transformer with Multi-Scale Feature Integration and Patch Dropping for Facial Expression Recognition.
- Author
-
Li, Nianfeng, Huang, Yongyuan, Wang, Zhenyan, Fan, Ziyao, Li, Xinyuan, and Xiao, Zhiguo
- Subjects
- *
TRANSFORMER models , *FACIAL expression , *CONVOLUTIONAL neural networks , *FEATURE extraction - Abstract
Convolutional neural networks (CNNs) have made significant progress in the field of facial expression recognition (FER). However, due to challenges such as occlusion, lighting variations, and changes in head pose, facial expression recognition in real-world environments remains highly challenging. At the same time, methods solely based on CNN heavily rely on local spatial features, lack global information, and struggle to balance the relationship between computational complexity and recognition accuracy. Consequently, the CNN-based models still fall short in their ability to address FER adequately. To address these issues, we propose a lightweight facial expression recognition method based on a hybrid vision transformer. This method captures multi-scale facial features through an improved attention module, achieving richer feature integration, enhancing the network's perception of key facial expression regions, and improving feature extraction capabilities. Additionally, to further enhance the model's performance, we have designed the patch dropping (PD) module. This module aims to emulate the attention allocation mechanism of the human visual system for local features, guiding the network to focus on the most discriminative features, reducing the influence of irrelevant features, and intuitively lowering computational costs. Extensive experiments demonstrate that our approach significantly outperforms other methods, achieving an accuracy of 86.51% on RAF-DB and nearly 70% on FER2013, with a model size of only 3.64 MB. These results demonstrate that our method provides a new perspective for the field of facial expression recognition. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
21. Task-Sensitive Efficient Feature Extraction Network for Oriented Object Detection in Remote Sensing Images.
- Author
-
Liu, Zhe, He, Guiqing, Dong, Liheng, Jing, Donglin, and Zhang, Haixi
- Subjects
- *
OBJECT recognition (Computer vision) , *REMOTE sensing , *FEATURE extraction , *CONVOLUTIONAL neural networks , *REMOTE-sensing images - Abstract
The widespread application of convolutional neural networks (CNNs) has led to significant advancements in object detection. However, challenges remain in achieving efficient and precise extraction of critical features when applying typical CNN-based methods to remote sensing detection tasks: (1) The convolutional kernels sliding horizontally in the backbone are misaligned with the features of arbitrarily oriented objects. Additionally, the detector shares the features extracted from the backbone, but the classification task requires orientation-invariant features while the regression task requires orientation-sensitive features. The inconsistency in feature requirements makes it difficult for the detector to extract the critical features required for each task. (2) The use of deeper convolutional structures can improve the detection accuracy, but it also results in substantial convolutional computations and feature redundancy, leading to inefficient feature extraction. To address this issue, we propose a Task-Sensitive Efficient Feature Extraction Network (TFE-Net). Specifically, we propose a special mixed fast convolution module for constructing an efficient network architecture that employs cheap transform operations to replace some of the convolution operations, generating more features with fewer parameters and computation resources. Next, we introduce the task-sensitive detection module, which first aligns the convolutional features with the targets using adaptive dynamic convolution based on the orientation of the targets. The task-sensitive feature decoupling mechanism is further designed to extract orientation-sensitive features and orientation-invariant features from the aligned features and feed them into the regression and classification branches, respectively, which provide the critical features needed for different tasks, thus improving the detection performance comprehensively. In addition, in order to make the training process more stable, we propose a balanced loss function to balance the gradients generated by different samples. Extensive experiments demonstrate that our proposed TFE-Net can achieve superior performance and obtain an effective balance between detection speed and accuracy on DOTA, UCAS-AOD, and HRSC2016. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
22. Recognition of coal and gangue under low illumination based on SG-YOLO model.
- Author
-
Shang, Deyong, Yang, Zhiyuan, and Lv, Zhibin
- Subjects
- *
COAL , *RECOGNITION (Psychology) , *LIGHTING , *FEATURE extraction , *LEARNING ability - Abstract
For the low illumination and dust in the coal and gangue identification site environment, which leads to poor recognition, an improved lightweight low-illumination gangue recognition algorithm based on YOLOv5s model is proposed: SG-YOLO algorithm. The original backbone network is replaced by GhostNet, a lightweight network, to optimize the feature extraction structure, reduce the model parameters, and decrease the computational power of the model; the SimAM attention mechanism module is introduced in the head part of the model to enhance the learning ability of coal and gangue features. Experiments show that compared with the YOLOv5s model, the improved model has a mAP of 97.0% on the gangue dataset, which improved by 1.4%. The size of the model is compressed to 55% of the original. The number of parameters is reduced by 47.6%, and the computational effort is reduced by 49.4%. Meanwhile, the recognition accuracy of the improved SG-YOLO model for coal and gangue under low illumination is 96.5% and 98.5% respectively, which effectively improves the recognition accuracy of coal rain gangue under low illumination environment. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
23. FireYOLO-Lite: Lightweight Forest Fire Detection Network with Wide-Field Multi-Scale Attention Mechanism.
- Author
-
Sheng, Sha, Liang, Zhengyin, Xu, Wenxing, Wang, Yong, and Su, Jiangdan
- Subjects
FEATURE extraction ,FOREST fires ,DEEP learning ,ALGORITHMS ,DETECTORS - Abstract
A lightweight forest fire detection model based on YOLOv8 is proposed in this paper in response to the problems existing in traditional sensors for forest fire detection. The performance of traditional sensors is easily constrained by hardware computing power, and their adaptability in different environments needs improvement. To balance the accuracy and speed of fire detection, the GhostNetV2 lightweight network is adopted to replace the backbone network for feature extraction of YOLOv8. The Ghost module is utilized to replace traditional convolution operations, conducting feature extraction independently in different dimensional channels, significantly reducing the complexity of the model while maintaining excellent performance. Additionally, an improved CPDCA channel priority attention mechanism is proposed, which extracts spatial features through dilated convolution, thereby reducing computational overhead and enabling the model to focus more on fire targets, achieving more accurate detection. In response to the problem of small targets in fire detection, the Inner IoU loss function is introduced. By adjusting the size of the auxiliary bounding boxes, this function effectively enhances the convergence effect of small target detection, further reducing missed detections, and improving overall detection accuracy. Experimental results indicate that, compared with traditional methods, the algorithm proposed in this paper significantly improves the average precision and FPS of fire detection while maintaining a smaller model size. Through experimental analysis, compared with YOLOv3-tiny, the average precision increased by 5.9% and the frame rate reached 285.3 FPS when the model size was only 4.9 M; compared with Shufflenet, the average precision increased by 2.9%, and the inference speed tripled. Additionally, the algorithm effectively addresses false positives, such as cloud and reflective light, further enhancing the detection of small targets and reducing missed detections. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
24. A Forest Fire Smoke Monitoring System Based on a Lightweight Neural Network for Edge Devices.
- Author
-
Huang, Jingwen, Yang, Huizhou, Liu, Yunfei, and Liu, Han
- Subjects
COMPUTER vision ,ARTIFICIAL intelligence ,FOREST fires ,RUNNING speed ,ARTIFICIAL vision ,DEEP learning - Abstract
Forest resources are one of the indispensable resources of the earth, which are the basis for the survival and development of human society. With the swift advancements in computer vision and artificial intelligence technology, the utilization of deep learning for smoke detection has achieved remarkable results. However, the existing deep learning models have poor performance in forest scenes and are difficult to deploy because of numerous parameters. Hence, we introduce an optimized forest fire smoke monitoring system for embedded edge devices based on a lightweight deep learning model. The model makes full use of the multi-scale variable attention mechanism of Transformer architecture to strengthen the ability of image feature extraction. Considering the needs of application scenarios, we propose an improved lightweight network model LCNet for feature extraction, which can reduce the parameters and enhance detecting ability. In order to improve running speed, a simple semi-supervised label knowledge distillation scheme is used to enhance the overall detection capability. Finally, we design and implement a forest fire smoke detection system on an embedded device, including the Jetson NX hardware platform, high-definition camera, and detection software system. The lightweight model is transplanted to the embedded edge device to achieve rapid forest fire smoke detection. Also, an asynchronous processing framework is designed to make the system highly available and robust. The improved model reduces three-fourths of the parameters and increases speed by 3.4 times with similar accuracy to the original model. This demonstrates that our system meets the precision demand and detects smoke in time. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
25. Multi-Feature Fusion Recognition and Localization Method for Unmanned Harvesting of Aquatic Vegetables.
- Author
-
Guan, Xianping, Shi, Longyuan, Yang, Weiguang, Ge, Hongrui, Wei, Xinhua, and Ding, Yuhan
- Subjects
OBJECT recognition (Computer vision) ,RECOGNITION (Psychology) ,FEATURE extraction ,IMAGE recognition (Computer vision) ,DEEP learning - Abstract
The vision-based recognition and localization system plays a crucial role in the unmanned harvesting of aquatic vegetables. After field investigation, factors such as illumination, shading, and computational cost have become the main difficulties restricting the identification and positioning of Brasenia schreberi. Therefore, this paper proposes a new lightweight detection method, YOLO-GS, which integrates feature information from both RGB and depth images for recognition and localization tasks. YOLO-GS employs the Ghost convolution module as a replacement for traditional convolution and innovatively introduces the C3-GS, a cross-stage module, to effectively reduce parameters and computational costs. With the redesigned detection head structure, its feature extraction capability in complex environments has been significantly enhanced. Moreover, the model utilizes Focal EIoU as the regression loss function to mitigate the adverse effects of low-quality samples on gradients. We have developed a data set of Brasenia schreberi that covers various complex scenarios, comprising a total of 1500 images. The YOLO-GS model, trained on this dataset, achieves an average accuracy of 95.7%. The model size is 7.95 MB, with 3.75 M parameters and a 9.5 GFLOPS computational cost. Compared to the original YOLOv5s model, YOLO-GS improves recognition accuracy by 2.8%, reduces the model size and parameter number by 43.6% and 46.5%, and offers a 39.9% reduction in computational requirements. Furthermore, the positioning errors of picking points are less than 5.01 mm in the X direction, 3.65 mm in the Y direction, and 1.79 mm in the Z direction. As a result, YOLO-GS not only excels with high recognition accuracy but also exhibits low computational demands, enabling precise target identification and localization in complex environments so as to meet the requirements of real-time harvesting tasks. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
26. UAV forest fire detection based on lightweight YOLOv5 model.
- Author
-
Zhou, Mengdong, Wu, Lei, Liu, Shuai, and Li, Jianjun
- Subjects
FOREST fires ,FOREST fire prevention & control ,WILDFIRE prevention ,DRONE aircraft - Abstract
In recent years, the frequent occurrence of forest fires has caused serious impact on the environment and economy. Fire detection has become a hot research direction. Despite the remarkable achievements, the unmanned aerial vehicle (UAV) still has some problems such as insufficient precision and excessive parameters. In order to improve the application ability of UAV in forest fire prevention and control, a lightweight target detection model based on YOLOv5 is proposed. The model is based on the overall structure of YOLOv5, MobileNetV3 is used as the backbone network, and semi-supervised knowledge distillation (SSLD) is used for training to improve the convergence speed and accuracy of the model. The final model size was reduced by 94.1% from 107.6 MB to 6.3 MB. mAP0.5 increased by 0.8% and mAP0.95 increased by 2.6%. The improved lightweight YOLOv5 model has fewer parameters and less computation, which confirms that MobileNetV3 has an excellent effect on the compression of model memory, and the semi-supervised knowledge distillation method is beneficial to improve the accuracy of the model. In the future, the accuracy of the model and the detection rate of the covered flame should be further improved. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
27. A Lightweight Network with Dual Encoder and Cross Feature Fusion for Cement Pavement Crack Detection.
- Author
-
Zhong Qu, Guoqing Mu, and Bin Yuan
- Subjects
CRACKING of pavements ,DENTAL cements ,DEEP learning ,CONVOLUTIONAL neural networks ,FEATURE extraction ,RECOMMENDER systems ,INFORMATION filtering - Abstract
Automatic crack detection of cement pavement chiefly benefits from the rapid development of deep learning, with convolutional neural networks (CNN) playing an important role in this field. However, as the performance of crack detection in cement pavement improves, the depth and width of the network structure are significantly increased, which necessitates more computing power and storage space. This limitation hampers the practical implementation of crack detection models on various platforms, particularly portable devices like small mobile devices. To solve these problems, we propose a dual-encoder-based network architecture that focuses on extracting more comprehensive fracture feature information and combines cross-fusion modules and coordinated attention mechanisms formore efficient feature fusion. Firstly, we use small channel convolution to construct shallow feature extractionmodule (SFEM) to extract low-level feature information of cracks in cement pavement images, in order to obtainmore information about cracks in the shallowfeatures of images. In addition,we construct large kernel atrous convolution (LKAC) to enhance crack information, which incorporates coordination attention mechanism for non-crack information filtering, and large kernel atrous convolution with different cores, using different receptive fields to extract more detailed edge and context information. Finally, the three-stage feature map outputs from the shallow feature extraction module is cross-fused with the two-stage feature map outputs from the large kernel atrous convolution module, and the shallow feature and detailed edge feature are fully fused to obtain the final crack prediction map. We evaluate our method on three public crack datasets: DeepCrack, CFD, and Crack500. Experimental results on theDeepCrack dataset demonstrate the effectiveness of our proposed method compared to state-of-the-art crack detection methods, which achieves Precision (P) 87.2%, Recall (R) 87.7%, and F-score (F1) 87.4%. Thanks to our lightweight crack detectionmodel, the parameter count of the model in real-world detection scenarios has been significantly reduced to less than 2M. This advancement also facilitates technical support for portable scene detection. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
28. Deep Dynamic Weights for Underwater Image Restoration.
- Author
-
Awan, Hafiz Shakeel Ahmad and Mahmood, Muhammad Tariq
- Subjects
CONVOLUTIONAL neural networks ,STANDARD deviations ,IMAGE intensifiers ,ATTENUATION of light ,DEEP learning - Abstract
Underwater imaging presents unique challenges, notably color distortions and reduced contrast due to light attenuation and scattering. Most underwater image enhancement methods first use linear transformations for color compensation and then enhance the image. We observed that linear transformation for color compensation is not suitable for certain images. For such images, non-linear mapping is a better choice. This paper introduces a unique underwater image restoration approach leveraging a streamlined convolutional neural network (CNN) for dynamic weight learning for linear and non-linear mapping. In the first phase, a classifier is applied that classifies the input images as Type I or Type II. In the second phase, we use the Deep Line Model (DLM) for Type-I images and the Deep Curve Model (DCM) for Type-II images. For mapping an input image to an output image, the DLM creatively combines color compensation and contrast adjustment in a single step and uses deep lines for transformation, whereas the DCM employs higher-order curves. Both models utilize lightweight neural networks that learn per-pixel dynamic weights based on the input image's characteristics. Comprehensive evaluations on benchmark datasets using metrics like peak signal-to-noise ratio (PSNR) and root mean square error (RMSE) affirm our method's effectiveness in accurately restoring underwater images, outperforming existing techniques. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
29. A lightweight algorithm for small traffic sign detection based on improved YOLOv5s.
- Author
-
Cai, Kunhui, Yang, Jingmin, Ren, Jinghui, and Zhang, Wenjie
- Abstract
With the rise of deep learning technology, significant progress has been made in object detection. Traffic sign detection is a research hotspot for object detection tasks. However, due to small size of traffic signs, there is room for further improvement in the comprehensive performance of the existing technology. In this paper, we propose a lightweight network based on yolov5s to achieve real-time localization and classification of small traffic signs. First, we improve the bottleneck transformers with 3 convolution (Bot3) module to enhance the backbone network's ability to extract features from small targets, improving the accuracy while reducing the number of parameters and giga floating-point operations per second (GFLOPs). Second, we introduce ghost convolution (GhostConv) to obtain redundant feature maps with cheap operations to further improve the model's efficiency. Finally, we use soft non-maximum suppression (Soft-NMS) in the detection phase to improve the model accuracy again without additional computational overhead for training. According to the tests on the Tsinghua-Tencent 100 K (TT100K) dataset, the proposed method outperforms the original YOLOv5s in small traffic sign detection, with an increase of 8.7% in m A P 50 , a reduction of 22.5% in parameter count, and a 17.2% reduction in computational complexity. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
30. Ultrafast‐and‐Ultralight ConvNet‐Based Intelligent Monitoring System for Diagnosing Early‐Stage Mpox Anytime and Anywhere.
- Author
-
Yue, Yubiao, Shi, Xiaoqiang, Qin, Li, Zhang, Xinyue, Xu, Jialong, Zheng, Zipei, Li, Zhenzhang, and Li, Yang
- Abstract
Due to the absence of more efficient diagnostic tools, the spread of mpox continues to be unchecked. Although related studies have demonstrated the high efficiency of deep learning models in diagnosing mpox, key aspects such as model inference speed and parameter size have always been overlooked. Herein, an ultrafast and ultralight network named Fast‐MpoxNet is proposed. Fast‐MpoxNet, with only 0.27 m parameters, can process input images at 68 frames per second (FPS) on the CPU. To detect subtle image differences and optimize model parameters better, Fast‐MpoxNet incorporates an attention‐based feature fusion module and a multiple auxiliary losses enhancement strategy. Experimental results indicate that Fast‐MpoxNet, utilizing transfer learning and data augmentation, produces 98.40% classification accuracy for four classes on the mpox dataset. Furthermore, its Recall for early‐stage mpox is 93.65%. Most importantly, an application system named Mpox‐AISM V2 is developed, suitable for both personal computers and smartphones. Mpox‐AISM V2 can rapidly and accurately diagnose mpox and can be easily deployed in various scenarios to offer the public real‐time mpox diagnosis services. This work has the potential to mitigate future mpox outbreaks and pave the way for developing real‐time diagnostic tools in the healthcare field. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
31. Efficient and Lightweight Neural Network for Hard Hat Detection.
- Author
-
He, Chenxi, Tan, Shengbo, Zhao, Jing, Ergu, Daji, Liu, Fangyao, Ma, Bo, and Li, Jianjun
- Subjects
HELMETS ,SAFETY hats ,COMPUTER vision ,VIDEO surveillance ,COMPUTER engineering ,FEATURE extraction - Abstract
Electric power operation, as one of the key fields in the world, faces particularly prominent safety issues. Ensuring the safety of operators has become the most fundamental requirement in power operation. However, there are some safety hazards in power construction. These hazards are mainly due to weak safety awareness among staff and the failure to standardize the wearing of safety helmets. In order to effectively address this situation, technical means such as video surveillance technology and computer vision technology can be utilized to monitor whether staff are wearing helmets and provide timely feedback. Such measures will greatly enhance the safety level of power operation. This paper proposes an improved lightweight helmet detection algorithm named YOLO-M3C. The algorithm first replaces the YOLOv5s backbone network with MobileNetV3, successfully reducing the model size from 13.7 MB to 10.2 MB, thereby increasing the model's detection speed from 42.0 frames per second to 55.6 frames per second. Then, the CA attention mechanism is introduced into the backbone network to enhance the feature extraction capability of the model. Finally, in order to further improve the detection recall rate and accuracy of the model, a knowledge distillation of the model was carried out. The experimental results show that, compared with the original YOLOv5s algorithm, the average accuracy of the improved YOLO-M3C algorithm is improved by 0.123, and the recall rate is the same. These results verify that the algorithm YOLO-M3C has excellent performance in target detection and recognition, which can improve accuracy and confidence, while reducing false detection and missing detection, and effectively meet the needs of helmet-wearing detection. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
32. LEO navigation observables extraction using CLOCFC network
- Author
-
Zhisen Wang, Hu Lu, and Zhiang Bian
- Subjects
Signals of opportunity ,Low earth orbit satellite communication ,Instantaneous Doppler positioning ,Lightweight network ,CFC network ,Medicine ,Science - Abstract
Abstract In case of mitigate the reliance of aviation users on the Global Navigation Satellite System (GNSS) in an increasingly interference-prone environment, utilizing opportunistic signals from Low-Earth Orbit (LEO) for navigation and positioning is an alternative approach. However, LEO satellite SOPs are not intended for navigation. Therefore, it is necessary to design methods to extract navigation observables from these signals. In this paper, we proposed a lightweight deep learning model with a two-branch structure called CLOCFC, designed to extract navigation observables. Furthermore, we have established a low Earth orbit satellite signal dataset by using ORBCOMM constellation signals as the input to the model and Doppler frequency as the label for the model. The results show that CLOCFC, as a lightweight model, demonstrates a significantly faster convergence rate and higher accuracy in navigation observables extraction compared to other models (ResNet, Swin Transformer, and Clo Transformer). In CLOCFC, we introduce the CFC module, a kind of Liquid Neural Network, to enhance the information acquisition capability through the spatiotemporal information in the data sequence. Finally, we have also conducted extensive experiments with the Doppler shift extraction of LEO satellites as an example, under various noise and resolution conditions, demonstrating the superiority of the CLOCFC.
- Published
- 2024
- Full Text
- View/download PDF
33. Image super‐resolution via dynamic network
- Author
-
Chunwei Tian, Xuanyu Zhang, Qi Zhang, Mingming Yang, and Zhaojie Ju
- Subjects
CNN ,dynamic network ,image super‐resolution ,lightweight network ,Computational linguistics. Natural language processing ,P98-98.5 ,Computer software ,QA76.75-76.765 - Abstract
Abstract Convolutional neural networks depend on deep network architectures to extract accurate information for image super‐resolution. However, obtained information of these convolutional neural networks cannot completely express predicted high‐quality images for complex scenes. A dynamic network for image super‐resolution (DSRNet) is presented, which contains a residual enhancement block, wide enhancement block, feature refinement block and construction block. The residual enhancement block is composed of a residual enhanced architecture to facilitate hierarchical features for image super‐resolution. To enhance robustness of obtained super‐resolution model for complex scenes, a wide enhancement block achieves a dynamic architecture to learn more robust information to enhance applicability of an obtained super‐resolution model for varying scenes. To prevent interference of components in a wide enhancement block, a refinement block utilises a stacked architecture to accurately learn obtained features. Also, a residual learning operation is embedded in the refinement block to prevent long‐term dependency problem. Finally, a construction block is responsible for reconstructing high‐quality images. Designed heterogeneous architecture can not only facilitate richer structural information, but also be lightweight, which is suitable for mobile digital devices. Experimental results show that our method is more competitive in terms of performance, recovering time of image super‐resolution and complexity. The code of DSRNet can be obtained at https://github.com/hellloxiaotian/DSRNet.
- Published
- 2024
- Full Text
- View/download PDF
34. Lightweight Neural Networks
- Author
-
Li, Bin and Li, Bin
- Published
- 2024
- Full Text
- View/download PDF
35. An Enhanced MobileNet with Multi-scale Attention Aggregation for DR Classification
- Author
-
Xi, Heran, Ji, Hongxu, Hu, Yang, Li, Jinbao, Zhu, Jinghua, Goos, Gerhard, Series Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Zhang, Wenjie, editor, Tung, Anthony, editor, Zheng, Zhonglong, editor, Yang, Zhengyi, editor, Wang, Xiaoyang, editor, and Guo, Hongjie, editor
- Published
- 2024
- Full Text
- View/download PDF
36. Infrared Image Super-Resolution via Lightweight Information Split Network
- Author
-
Liu, Shijie, Yan, Kang, Qin, Feiwei, Wang, Changmiao, Ge, Ruiquan, Zhang, Kai, Huang, Jie, Peng, Yong, Cao, Jin, Goos, Gerhard, Series Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Huang, De-Shuang, editor, Chen, Wei, editor, and Pan, Yijie, editor
- Published
- 2024
- Full Text
- View/download PDF
37. Light-UNet: An Efficient Segmentation Network for Medical Image
- Author
-
Zhang, Yue, Xu, Chao, Zhang, Zhifan, Wang, Jianjun, Goos, Gerhard, Series Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Huang, De-Shuang, editor, Si, Zhanjun, editor, and Guo, Jiayang, editor
- Published
- 2024
- Full Text
- View/download PDF
38. DSCVSR: A Lightweight Video Super-Resolution for Arbitrary Magnification
- Author
-
Hong, Zixuan, Cao, Weipeng, Xu, Zhiwu, Ming, Zhong, Cao, Chuqing, Zheng, Liang, Goos, Gerhard, Series Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Cao, Cungeng, editor, Chen, Huajun, editor, Zhao, Liang, editor, Arshad, Junaid, editor, Asyhari, Taufiq, editor, and Wang, Yonghao, editor
- Published
- 2024
- Full Text
- View/download PDF
39. Selected Partially Labeled Learning for Abdominal Organ and Pan-Cancer Segmentation
- Author
-
Zhu, Yuntao, Zou, Liwen, Li, Linyao, Wen, Pengxu, Hartmanis, Juris, Founding Editor, van Leeuwen, Jan, Series Editor, Hutchison, David, Editorial Board Member, Kanade, Takeo, Editorial Board Member, Kittler, Josef, Editorial Board Member, Kleinberg, Jon M., Editorial Board Member, Kobsa, Alfred, Series Editor, Mattern, Friedemann, Editorial Board Member, Mitchell, John C., Editorial Board Member, Naor, Moni, Editorial Board Member, Nierstrasz, Oscar, Series Editor, Pandu Rangan, C., Editorial Board Member, Sudan, Madhu, Series Editor, Terzopoulos, Demetri, Editorial Board Member, Tygar, Doug, Editorial Board Member, Weikum, Gerhard, Series Editor, Vardi, Moshe Y, Series Editor, Goos, Gerhard, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Ma, Jun, editor, and Wang, Bo, editor
- Published
- 2024
- Full Text
- View/download PDF
40. LightNet+: Boosted Light-Weighted Network for Smoke Semantic Segmentation
- Author
-
Li, Kang, Wang, Chunmei, Meng, Chunli, Yuan, Feiniu, Filipe, Joaquim, Editorial Board Member, Ghosh, Ashish, Editorial Board Member, Zhou, Lizhu, Editorial Board Member, Zhai, Guangtao, editor, Zhou, Jun, editor, Ye, Long, editor, Yang, Hua, editor, An, Ping, editor, and Yang, Xiaokang, editor
- Published
- 2024
- Full Text
- View/download PDF
41. Second-Order Channel Attention Multi-scale Grouped Convolution LSTM Networks for Automatic Modulation Recognition
- Author
-
Liu, Xin, Zhang, Jiashu, Angrisani, Leopoldo, Series Editor, Arteaga, Marco, Series Editor, Chakraborty, Samarjit, Series Editor, Chen, Jiming, Series Editor, Chen, Shanben, Series Editor, Chen, Tan Kay, Series Editor, Dillmann, Rüdiger, Series Editor, Duan, Haibin, Series Editor, Ferrari, Gianluigi, Series Editor, Ferre, Manuel, Series Editor, Jabbari, Faryar, Series Editor, Jia, Limin, Series Editor, Kacprzyk, Janusz, Series Editor, Khamis, Alaa, Series Editor, Kroeger, Torsten, Series Editor, Li, Yong, Series Editor, Liang, Qilian, Series Editor, Martín, Ferran, Series Editor, Ming, Tan Cher, Series Editor, Minker, Wolfgang, Series Editor, Misra, Pradeep, Series Editor, Mukhopadhyay, Subhas, Series Editor, Ning, Cun-Zheng, Series Editor, Nishida, Toyoaki, Series Editor, Oneto, Luca, Series Editor, Panigrahi, Bijaya Ketan, Series Editor, Pascucci, Federica, Series Editor, Qin, Yong, Series Editor, Seng, Gan Woon, Series Editor, Speidel, Joachim, Series Editor, Veiga, Germano, Series Editor, Wu, Haitao, Series Editor, Zamboni, Walter, Series Editor, Tan, Kay Chen, Series Editor, Dong, Jian, editor, Zhang, Long, editor, and Cheng, Deqiang, editor
- Published
- 2024
- Full Text
- View/download PDF
42. FSANet: A Lightweight Network for Tobacco Grouping Using Multi-scale Convolution and Attention Mechanism
- Author
-
Su, Yongzhou, Hou, Kaihu, Long, Jie, Gai, Xiaolei, Zhang, Yiwu, Zhang, Xiaowei, Filipe, Joaquim, Editorial Board Member, Ghosh, Ashish, Editorial Board Member, Prates, Raquel Oliveira, Editorial Board Member, Zhou, Lizhu, Editorial Board Member, Jin, Hai, editor, Pan, Yi, editor, and Lu, Jianfeng, editor
- Published
- 2024
- Full Text
- View/download PDF
43. Deep Tiny Network for Recognition-Oriented Face Image Quality Assessment
- Author
-
Peng, Baoyun, Liu, Min, Zhang, Zhaoning, Xu, Kai, Li, Dongsheng, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Zhang, Fang-Lue, editor, and Sharf, Andrei, editor
- Published
- 2024
- Full Text
- View/download PDF
44. Research on Semantic Segmentation Algorithm for Autonomous Driving Based on Improved DeepLabv3+
- Author
-
Qin, Jun, Xu, Chunsen, Ai, Yong, Zhang, Huili, Cheng, Yong, Angrisani, Leopoldo, Series Editor, Arteaga, Marco, Series Editor, Chakraborty, Samarjit, Series Editor, Chen, Jiming, Series Editor, Chen, Shanben, Series Editor, Chen, Tan Kay, Series Editor, Dillmann, Rüdiger, Series Editor, Duan, Haibin, Series Editor, Ferrari, Gianluigi, Series Editor, Ferre, Manuel, Series Editor, Jabbari, Faryar, Series Editor, Jia, Limin, Series Editor, Kacprzyk, Janusz, Series Editor, Khamis, Alaa, Series Editor, Kroeger, Torsten, Series Editor, Li, Yong, Series Editor, Liang, Qilian, Series Editor, Martín, Ferran, Series Editor, Ming, Tan Cher, Series Editor, Minker, Wolfgang, Series Editor, Misra, Pradeep, Series Editor, Mukhopadhyay, Subhas, Series Editor, Ning, Cun-Zheng, Series Editor, Nishida, Toyoaki, Series Editor, Oneto, Luca, Series Editor, Panigrahi, Bijaya Ketan, Series Editor, Pascucci, Federica, Series Editor, Qin, Yong, Series Editor, Seng, Gan Woon, Series Editor, Speidel, Joachim, Series Editor, Veiga, Germano, Series Editor, Wu, Haitao, Series Editor, Zamboni, Walter, Series Editor, Zhang, Junjie James, Series Editor, Tan, Kay Chen, Series Editor, Wang, Wei, editor, Mu, Jiasong, editor, Liu, Xin, editor, and Na, Zhenyu Na, editor
- Published
- 2024
- Full Text
- View/download PDF
45. GoatPose: A Lightweight and Efficient Network with Attention Mechanism
- Author
-
Sun, Yaxuan, Wang, Annan, Wu, Shengxi, Filipe, Joaquim, Editorial Board Member, Ghosh, Ashish, Editorial Board Member, Prates, Raquel Oliveira, Editorial Board Member, Zhou, Lizhu, Editorial Board Member, Luo, Biao, editor, Cheng, Long, editor, Wu, Zheng-Guang, editor, Li, Hongyi, editor, and Li, Chaojie, editor
- Published
- 2024
- Full Text
- View/download PDF
46. Lightweight Infrared and Visible Image Fusion Based on Attention Mechanism and Receptive Field Enhancement
- Author
-
Liu, Ting, Zhang, Yuxin, Fan, Yunsheng, Luo, Peiqi, Wang, Guofeng, Angrisani, Leopoldo, Series Editor, Arteaga, Marco, Series Editor, Chakraborty, Samarjit, Series Editor, Chen, Jiming, Series Editor, Chen, Shanben, Series Editor, Chen, Tan Kay, Series Editor, Dillmann, Rüdiger, Series Editor, Duan, Haibin, Series Editor, Ferrari, Gianluigi, Series Editor, Ferre, Manuel, Series Editor, Jabbari, Faryar, Series Editor, Jia, Limin, Series Editor, Kacprzyk, Janusz, Series Editor, Khamis, Alaa, Series Editor, Kroeger, Torsten, Series Editor, Li, Yong, Series Editor, Liang, Qilian, Series Editor, Martín, Ferran, Series Editor, Ming, Tan Cher, Series Editor, Minker, Wolfgang, Series Editor, Misra, Pradeep, Series Editor, Mukhopadhyay, Subhas, Series Editor, Ning, Cun-Zheng, Series Editor, Nishida, Toyoaki, Series Editor, Oneto, Luca, Series Editor, Panigrahi, Bijaya Ketan, Series Editor, Pascucci, Federica, Series Editor, Qin, Yong, Series Editor, Seng, Gan Woon, Series Editor, Speidel, Joachim, Series Editor, Veiga, Germano, Series Editor, Wu, Haitao, Series Editor, Zamboni, Walter, Series Editor, Zhang, Junjie James, Series Editor, Tan, Kay Chen, Series Editor, Wang, Wei, editor, Liu, Xin, editor, Na, Zhenyu, editor, and Zhang, Baoju, editor
- Published
- 2024
- Full Text
- View/download PDF
47. A Lightweight and Real-Time Network for Unmanned Aerial Vehicle Object Tracking
- Author
-
Jin, Qiuyu, Wang, Wenzheng, Wang, Ban, Wang, Xing, Sun, Zhiliang, Sun, Haotian, Angrisani, Leopoldo, Series Editor, Arteaga, Marco, Series Editor, Chakraborty, Samarjit, Series Editor, Chen, Jiming, Series Editor, Chen, Shanben, Series Editor, Chen, Tan Kay, Series Editor, Dillmann, Rüdiger, Series Editor, Duan, Haibin, Series Editor, Ferrari, Gianluigi, Series Editor, Ferre, Manuel, Series Editor, Jabbari, Faryar, Series Editor, Jia, Limin, Series Editor, Kacprzyk, Janusz, Series Editor, Khamis, Alaa, Series Editor, Kroeger, Torsten, Series Editor, Li, Yong, Series Editor, Liang, Qilian, Series Editor, Martín, Ferran, Series Editor, Ming, Tan Cher, Series Editor, Minker, Wolfgang, Series Editor, Misra, Pradeep, Series Editor, Mukhopadhyay, Subhas, Series Editor, Ning, Cun-Zheng, Series Editor, Nishida, Toyoaki, Series Editor, Oneto, Luca, Series Editor, Panigrahi, Bijaya Ketan, Series Editor, Pascucci, Federica, Series Editor, Qin, Yong, Series Editor, Seng, Gan Woon, Series Editor, Speidel, Joachim, Series Editor, Veiga, Germano, Series Editor, Wu, Haitao, Series Editor, Zamboni, Walter, Series Editor, Zhang, Junjie James, Series Editor, Tan, Kay Chen, Series Editor, and Chinese Institute of Command and Control, editor
- Published
- 2024
- Full Text
- View/download PDF
48. GSUNet: A Brain Tumor Segmentation Method Based on 3D Ghost Shuffle U-Net
- Author
-
Hong, JiXuan, Xie, JingJing, He, XueQin, Yang, ChenHui, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Rudinac, Stevan, editor, Hanjalic, Alan, editor, Liem, Cynthia, editor, Worring, Marcel, editor, Jónsson, Björn Þór, editor, Liu, Bei, editor, and Yamakata, Yoko, editor
- Published
- 2024
- Full Text
- View/download PDF
49. Research on Deep Learning-Based Lightweight Object Grasping Algorithm for Robots
- Author
-
Zhao, Yancheng, Wei, Tianxu, Du, Baoshuai, Zhao, Jingbo, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Sheng, Bin, editor, Bi, Lei, editor, Kim, Jinman, editor, Magnenat-Thalmann, Nadia, editor, and Thalmann, Daniel, editor
- Published
- 2024
- Full Text
- View/download PDF
50. ELFLN: An Efficient Lightweight Facial Landmark Network Based on Hybrid Knowledge Distillation
- Author
-
Chen, Shidong, Wang, Yalun, Bian, Huicong, Lu, Qin, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Liu, Qingshan, editor, Wang, Hanzi, editor, Ma, Zhanyu, editor, Zheng, Weishi, editor, Zha, Hongbin, editor, Chen, Xilin, editor, Wang, Liang, editor, and Ji, Rongrong, editor
- Published
- 2024
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.