989 results on '"yolov7"'
Search Results
2. Highway Obstacle Recognition Based on Improved YOLOv7 and Defogging Algorithm
- Author
-
Fan, Mingliang, Liu, Jing, Yu, Jiaming, Akan, Ozgur, Editorial Board Member, Bellavista, Paolo, Editorial Board Member, Cao, Jiannong, Editorial Board Member, Coulson, Geoffrey, Editorial Board Member, Dressler, Falko, Editorial Board Member, Ferrari, Domenico, Editorial Board Member, Gerla, Mario, Editorial Board Member, Kobayashi, Hisashi, Editorial Board Member, Palazzo, Sergio, Editorial Board Member, Sahni, Sartaj, Editorial Board Member, Shen, Xuemin, Editorial Board Member, Stan, Mircea, Editorial Board Member, Jia, Xiaohua, Editorial Board Member, Zomaya, Albert Y., Editorial Board Member, Chen, Xiang, editor, Wang, Xijun, editor, Lin, Shangjing, editor, and Liu, Jing, editor
- Published
- 2025
- Full Text
- View/download PDF
3. Gun Detection Using Yolov7
- Author
-
Rizwana, Shaik, Tomer, Vikas, Singh, Prabhishek, Diwakar, Manoj, Yamsani, Nagendar, Kacprzyk, Janusz, Series Editor, Gomide, Fernando, Advisory Editor, Kaynak, Okyay, Advisory Editor, Liu, Derong, Advisory Editor, Pedrycz, Witold, Advisory Editor, Polycarpou, Marios M., Advisory Editor, Rudas, Imre J., Advisory Editor, Wang, Jun, Advisory Editor, Goar, Vishal, editor, Kuri, Manoj, editor, Kumar, Rajesh, editor, and Senjyu, Tomonobu, editor
- Published
- 2025
- Full Text
- View/download PDF
4. Small Target Underwater Sonar Image Target Detection Based on Adaptive Global Feature Enhancement Network
- Author
-
Zheng, Kun, Chen, Zhe, Tang, Jianxun, Chaw, Jun Kit, Akan, Ozgur, Editorial Board Member, Bellavista, Paolo, Editorial Board Member, Cao, Jiannong, Editorial Board Member, Coulson, Geoffrey, Editorial Board Member, Dressler, Falko, Editorial Board Member, Ferrari, Domenico, Editorial Board Member, Gerla, Mario, Editorial Board Member, Kobayashi, Hisashi, Editorial Board Member, Palazzo, Sergio, Editorial Board Member, Sahni, Sartaj, Editorial Board Member, Shen, Xuemin, Editorial Board Member, Stan, Mircea, Editorial Board Member, Jia, Xiaohua, Editorial Board Member, Zomaya, Albert Y., Editorial Board Member, and Wang, Junyi, editor
- Published
- 2025
- Full Text
- View/download PDF
5. Pedestrian Fall Detection Algorithm Based on Improved YOLOv7
- Author
-
Wang, Fei, Zhang, Yunchu, Zhang, Xinyi, Liu, Yiming, Ghosh, Ashish, Editorial Board Member, Zhou, Lizhu, Editorial Board Member, Zhang, Haijun, editor, Li, Xianxian, editor, Hao, Tianyong, editor, Meng, Weizhi, editor, Wu, Zhou, editor, and He, Qian, editor
- Published
- 2025
- Full Text
- View/download PDF
6. Safety Helmet-Wearing Detection Method Fusing Pose Estimation
- Author
-
Zhang, Xinyi, Zhang, Yunchu, Liu, Yiming, Wang, Fei, Ghosh, Ashish, Editorial Board Member, Zhou, Lizhu, Editorial Board Member, Zhang, Haijun, editor, Li, Xianxian, editor, Hao, Tianyong, editor, Meng, Weizhi, editor, Wu, Zhou, editor, and He, Qian, editor
- Published
- 2025
- Full Text
- View/download PDF
7. Deformable attention mechanism-based YOLOv7 structure for lung nodule detection.
- Author
-
Liu, Yu and Ao, Yongcai
- Subjects
- *
PULMONARY nodules , *DEEP learning , *LUNG cancer , *OVERALL survival , *COMPUTED tomography , *LUNGS - Abstract
Early detection of lung nodules is essential for lung cancer screening and improving patient survival rates. Traditional object detection networks such as YOLO and Faster R-CNN have shown promising results in detecting lung nodules but often lack sufficient integration of extracted features to enhance accuracy and efficiency. Moreover, these methods typically do not retain the spatial information of lung nodules from the original CT images. To overcome these limitations, a novel lung nodule detection algorithm based on YOLOv7 is introduced. Firstly, to better preserve essential features and minimize interference from irrelevant background noise, a deformable attention module for feature fusion has been designed. Additionally, maximum intensity projection is employed to create projection images at various intensities, thereby enriching the spatial background information that is often missing in single CT slices. Thirdly, the WIoU loss function is utilized to replace the original YOLOv7 loss function, aiming to reduce the influence of low-quality samples on the gradient within the dataset. The proposed model was validated using the publicly available LUNA16 dataset and achieved a recall rate of 94.40% and an AP value of 95.39%. These results demonstrate the enhanced precision and efficiency of lung nodule detection. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. YOLOv7 for brain tumour detection using morphological transfer learning model.
- Author
-
Pandey, Sanat Kumar and Bhandari, Ashish Kumar
- Subjects
- *
BRAIN tumors , *MAGNETIC resonance imaging , *ARTIFICIAL intelligence , *DEEP learning , *COMPUTER-aided design - Abstract
An accurate diagnosis of a brain tumour in its early stages is required to improve the possibility of survival for cancer patients. Due to the structural complexity of the brain, it has become very difficult and tedious for neurologists and radiologists to diagnose brain tumours in the initial stages with the help of various common manual approaches to tumour diagnosis. To improve the performance of the diagnosis, some computer-aided diagnosis-based systems are developed with the concepts of artificial intelligence. In this proposed manuscript, we analyse various computer-aided design (CAD)-based approaches and design a modern approach with ideas of transfer learning over deep learning on magnetic resonance imaging (MRI). In this study, we apply a transfer learning approach with the object detection model YOLO (You Only Look Once) and analyse the MRI dataset with the various modified versions of YOLO. After the analysis, we propose an object detection model based on the modified YOLOv7 with a morphological filtering approach to reach an efficient and accurate diagnosis. To enhance the performance accuracy of this suggested model, we also analyse the various versions of YOLOv7 models and find that the proposed model having the YOLOv7-E6E object detection technique gives the optimum value of performance indicators as precision, recall, F1, and mAP@50 as 1, 0.92, 0.958333, and 0.974, respectively. The value of mAP@50 improves to 0.992 by introducing a morphological filtering approach before the object detection technique. During the complete analysis of the suggested model, we use the BraTS 2021 dataset. The BraTS 2021 dataset has brain MR images from the RSNA-MICCAI brain tumour radiogenetic competition, and the complete dataset is labelled using the online tool MakeSense AI. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. An improved YOLOv7 model based on Swin Transformer and Trident Pyramid Networks for accurate tomato detection.
- Author
-
Liu, Guoxu, Zhang, Yonghui, Liu, Jun, Liu, Deyong, Chen, Chunlei, Li, Yujie, Zhang, Xiujie, and Touko Mbouembe, Philippe Lyonel
- Abstract
Accurate fruit detection is crucial for automated fruit picking. However, real-world scenarios, influenced by complex environmental factors such as illumination variations, occlusion, and overlap, pose significant challenges to accurate fruit detection. These challenges subsequently impact the commercialization of fruit harvesting robots. A tomato detection model named YOLO-SwinTF, based on YOLOv7, is proposed to address these challenges. Integrating Swin Transformer (ST) blocks into the backbone network enables the model to capture global information by modeling long-range visual dependencies. Trident Pyramid Networks (TPN) are introduced to overcome the limitations of PANet's focus on communication-based processing. TPN incorporates multiple self-processing (SP) modules within existing top-down and bottom-up architectures, allowing feature maps to generate new findings for communication. In addition, Focaler-IoU is introduced to reconstruct the original intersection-over-union (IoU) loss to allow the loss function to adjust its focus based on the distribution of difficult and easy samples. The proposed model is evaluated on a tomato dataset, and the experimental results demonstrated that the proposed model's detection recall, precision, F
1 score, and AP reach 96.27%, 96.17%, 96.22%, and 98.67%, respectively. These represent improvements of 1.64%, 0.92%, 1.28%, and 0.88% compared to the original YOLOv7 model. When compared to other state-of-the-art detection methods, this approach achieves superior performance in terms of accuracy while maintaining comparable detection speed. In addition, the proposed model exhibits strong robustness under various lighting and occlusion conditions, demonstrating its significant potential in tomato detection. [ABSTRACT FROM AUTHOR]- Published
- 2024
- Full Text
- View/download PDF
10. Research on Deep Learning Detection Model for Pedestrian Objects in Complex Scenes Based on Improved YOLOv7.
- Author
-
Hu, Jun, Zhou, Yongqi, Wang, Hao, Qiao, Peng, and Wan, Wenwei
- Abstract
Objective: Pedestrian detection is very important for the environment perception and safety action of intelligent robots and autonomous driving, and is the key to ensuring the safe action of intelligent robots and auto assisted driving. Methods: In response to the characteristics of pedestrian objects occupying a small image area, diverse poses, complex scenes and severe occlusion, this paper proposes an improved pedestrian object detection method based on the YOLOv7 model, which adopts the Convolutional Block Attention Module (CBAM) attention mechanism and Deformable ConvNets v2 (DCNv2) in the two Efficient Layer Aggregation Network (ELAN) modules of the backbone feature extraction network. In addition, the detection head is replaced with a Dynamic Head (DyHead) detector head with an attention mechanism; unnecessary background information around the pedestrian object is also effectively excluded, making the model learn more concentrated feature representations. Results: Compared with the original model, the log-average miss rate of the improved YOLOv7 model is significantly reduced in both the Citypersons dataset and the INRIA dataset. Conclusions: The improved YOLOv7 model proposed in this paper achieved good performance improvement in different pedestrian detection problems. The research in this paper has important reference significance for pedestrian detection in complex scenes such as small, occluded and overlapping objects. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
11. 基于轻量化YOLOv7 的织物疵点检测算法研究.
- Author
-
赵英宝, 刘姝含, 黄丽敏, and 武晓晶
- Abstract
Copyright of Cotton Textile Technology is the property of Cotton Textile Technology Editorial Office and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
12. YOLOv7-P: a lighter and more effective UAV aerial photography object detection algorithm.
- Author
-
Sun, Fengxi, He, Ning, Wang, Xin, Liu, Hongfei, and Zou, Yuxiang
- Abstract
Because of the special way an unmanned aerial vehicle (UAV) acquires aerial photography, UAV images have the characteristics of large coverage area, complex background, and a large proportion of small targets, which exacerbate the difficulty of object detection. Additionally, UAV-based aerial image detection needs to meet lightweight and real-time capabilities. To address these issues, this paper proposes a lightweight model YOLOv7-P that is based on YOLOv7 but has a stronger detection capability for small targets. First, partial convolution (PConv) is used to reduce redundant parameters and computation in YOLOv7. Second, an optimal combination of detection heads is determined that can significantly improve the detection performance of small objects. Third, a novel lightweight convolution called PConv-wide is proposed to replace RepConv in the network, thus simplifying the network without affecting detection accuracy. Finally, the normalized wasserstein distance loss is reasonably combined with the complete intersection over union loss to further improve the sensitivity of the network to small targets. The proposed YOLOv7-P model strikes a delicate balance between precision and parameter count. Compared with the baseline YOLOv7 network, it reduces parameter count by 47.1% without increasing computational complexity and boosts AP50 by 8% and mAP by 5.4% on the VisDrone dataset. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
13. 基于YOLOv7 改进的夜间樱桃检测方法:YOLOv7-Cherry.
- Author
-
盖荣丽, 孔祥宙, 秦山, and 魏凯
- Abstract
Copyright of Journal of Computer Engineering & Applications is the property of Beijing Journal of Computer Engineering & Applications Journal Co Ltd. and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
- Full Text
- View/download PDF
14. Targeted weed management of Palmer amaranth using robotics and deep learning (YOLOv7).
- Author
-
Balabantaray, Amlan, Behera, Shaswati, Liew, CheeTown, Chamara, Nipuna, Singh, Mandeep, Jhala, Amit J., Pitla, Santosh, Asaduzzaman, Asad (Md), and Everman, Wesley
- Subjects
AGRICULTURAL robots ,WEED control ,OBJECT recognition (Computer vision) ,INDUSTRIAL robots ,COMPUTER vision ,DEEP learning - Abstract
Effective weed management is a significant challenge in agronomic crops which necessitates innovative solutions to reduce negative environmental impacts and minimize crop damage. Traditional methods often rely on indiscriminate herbicide application, which lacks precision and sustainability. To address this critical need, this study demonstrated an AI-enabled robotic system, Weeding robot, designed for targeted weed management. Palmer amaranth (Amaranthus palmeri S. Watson) was selected as it is the most troublesome weed in Nebraska. We developed the full stack (vision, hardware, software, robotic platform, and AI model) for precision spraying using YOLOv7, a state-of-the-art object detection deep learning technique. The Weeding robot achieved an average of 60.4% precision and 62% recall in real-time weed identification and spot spraying with the developed gantry-based sprayer system. The Weeding robot successfully identified Palmer amaranth across diverse growth stages in controlled outdoor conditions. This study demonstrates the potential of AI-enabled robotic systems for targeted weed management, offering a more precise and sustainable alternative to traditional herbicide application methods. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
15. Improved YOLOv7 Electric Work Safety Belt Hook Suspension State Recognition Algorithm Based on Decoupled Head.
- Author
-
Xie, Xiaona, Chang, Zhengwei, Lan, Zhongxiao, Chen, Mingju, and Zhang, Xingyue
- Subjects
INDUSTRIAL safety ,FEATURE extraction ,DEEP learning ,COMPUTER performance ,NETWORK performance - Abstract
Safety is the eternal theme of power systems. In view of problems such as time-consuming and poor real-time performance in the correct use of seat belt hooks by manual supervision operators in the process of power operation, this paper proposes an improved YOLOv7 seat belt hook suspension state recognition algorithm. Firstly, the feature extraction part of the YOLOv7 backbone network is improved, and the M-Spatial Pyramid Pooling Concurrent Spatial Pyramid Convolution (M-SPPCSPC) feature extraction module is constructed to replace the Spatial Pyramid Pooling Concurrent Spatial Pyramid Convolution (SPPCSPC) module of the backbone network, which reduces the amount of computation and improves the detection speed of the backbone network while keeping the sensory field of the backbone network unchanged. Second, a decoupled head, which realizes the confidence and regression frames separately, is introduced to alleviate the negative impact of the conflict between the classification and regression tasks, consequently improving the network detection accuracy and accelerating the network convergence. Ultimately, a dynamic non-monotonic focusing mechanism is introduced in the output layer, and the Wise Intersection over Union (WioU) loss function is used to reduce the competitiveness of high-quality anchor frames while reducing the harmful gradient generated by low-quality anchor frames, which ultimately improves the overall performance of the detection network. The experimental results show that the mean Average Precision (mAP@0.5) value of the improved network reaches 81.2%, which is 7.4% higher than that of the original YOLOv7, therefore achieving better detection results for multiple-state recognition of hooks. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
16. An improved YOLOv7 model based on Swin Transformer and Trident Pyramid Networks for accurate tomato detection.
- Author
-
Guoxu Liu, Yonghui Zhang, Jun Liu, Deyong Liu, Chunlei Chen, Yujie Li, Xiujie Zhang, and Touko Mbouembe, Philippe Lyonel
- Subjects
TRANSFORMER models ,FRUIT harvesting ,FRUIT ,PYRAMIDS ,COMMERCIALIZATION - Abstract
Accurate fruit detection is crucial for automated fruit picking. However, real-world scenarios, influenced by complex environmental factors such as illumination variations, occlusion, and overlap, pose significant challenges to accurate fruit detection. These challenges subsequently impact the commercialization of fruit harvesting robots. A tomato detection model named YOLO-SwinTF, based on YOLOv7, is proposed to address these challenges. Integrating Swin Transformer (ST) blocks into the backbone network enables the model to capture global information by modeling long-range visual dependencies. Trident Pyramid Networks (TPN) are introduced to overcome the limitations of PANet's focus on communication-based processing. TPN incorporates multiple self-processing (SP) modules within existing top-down and bottom-up architectures, allowing feature maps to generate new findings for communication. In addition, Focaler-IoU is introduced to reconstruct the original intersection-over-union (IoU) loss to allow the loss function to adjust its focus based on the distribution of difficult and easy samples. The proposed model is evaluated on a tomato dataset, and the experimental results demonstrated that the proposed model's detection recall, precision, F1 score, and AP reach 96.27%, 96.17%, 96.22%, and 98.67%, respectively. These represent improvements of 1.64%, 0.92%, 1.28%, and 0.88% compared to the original YOLOv7 model. When compared to other state-of-the-art detection methods, this approach achieves superior performance in terms of accuracy while maintaining comparable detection speed. In addition, the proposed model exhibits strong robustness under various lighting and occlusion conditions, demonstrating its significant potential in tomato detection. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
17. A Candy Defect Detection Method Based on StyleGAN2 and Improved YOLOv7 for Imbalanced Data.
- Author
-
Li, Xingyou, Xue, Sheng, Li, Zhenye, Fang, Xiaodong, Zhu, Tingting, and Ni, Chao
- Subjects
GENERATIVE adversarial networks ,COMPUTER vision ,CANDY industry ,CONVEYOR belts ,FOOD quality ,DEEP learning - Abstract
Quality management in the candy industry is a vital part of food quality management. Defective candies significantly affect subsequent packaging and consumption, impacting the efficiency of candy manufacturers and the consumer experience. However, challenges exist in candy defect detection on food production lines due to the small size of the targets and defects, as well as the difficulty of batch sampling defects from automated production lines. A high-precision candy defect detection method based on deep learning is proposed in this paper. Initially, pseudo-defective candy images are generated based on Style Generative Adversarial Network-v2 (StyleGAN2), thereby enhancing the authenticity of these synthetic defect images. Following the separation of the background based on the color characteristics of the defective candies on the conveyor belt, a GAN is utilized for negative sample data enhancement. This effectively reduces the impact of data imbalance between complete and defective candies on the model's detection performance. Secondly, considering the challenges brought by the small size and random shape of candy defects to target detection, the efficient target detection method YOLOv7 is improved. The Spatial Pyramid Pooling Fast Cross Stage Partial Connection (SPPFCSPC) module, the C3C2 module, and the global attention mechanism are introduced to enhance feature extraction precision. The improved model achieves a 3.0% increase in recognition accuracy and a 3.7% increase in recall rate while supporting real-time recognition scenery. This method not only enhances the efficiency of food quality management but also promotes the application of computer vision and deep learning in industrial production. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
18. APD-YOLOv7: Enhancing Sustainable Farming through Precise Identification of Agricultural Pests and Diseases Using a Novel Diagonal Difference Ratio IOU Loss.
- Author
-
Li, Jianwen, Liu, Shutian, Chen, Dong, Zhou, Shengbang, and Li, Chuanqi
- Abstract
The diversity and complexity of the agricultural environment pose significant challenges for the collection of pest and disease data. Additionally, pest and disease datasets often suffer from uneven distribution in quantity and inconsistent annotation standards. Enhancing the accuracy of pest and disease recognition remains a challenge for existing models. We constructed a representative agricultural pest and disease dataset, FIP6Set, through a combination of field photography and web scraping. This dataset encapsulates key issues encountered in existing agricultural pest and disease datasets. Referencing existing bounding box regression (BBR) loss functions, we reconsidered their geometric features and proposed a novel bounding box similarity comparison metric, DDRIoU, suited to the characteristics of agricultural pest and disease datasets. By integrating the focal loss concept with the DDRIoU loss, we derived a new loss function, namely Focal-DDRIoU loss. Furthermore, we modified the network structure of YOLOV7 by embedding the MobileViTv3 module. Consequently, we introduced a model specifically designed for agricultural pest and disease detection in precision agriculture. We conducted performance evaluations on the FIP6Set dataset using mAP75 as the evaluation metric. Experimental results demonstrate that the Focal-DDRIoU loss achieves improvements of 1.12%, 1.24%, 1.04%, and 1.50% compared to the GIoU, DIoU, CIoU, and EIoU losses, respectively. When employing the GIoU, DIoU, CIoU, EIoU, and Focal-DDRIoU loss functions, the adjusted network structure showed enhancements of 0.68%, 0.68%, 0.78%, 0.60%, and 0.56%, respectively, compared to the original YOLOv7. Furthermore, the proposed model outperformed the mainstream YOLOv7 and YOLOv5 models by 1.86% and 1.60%, respectively. The superior performance of the proposed model in detecting agricultural pests and diseases directly contributes to reducing pesticide misuse, preventing large-scale pest and disease outbreaks, and ultimately enhancing crop yields. These outcomes strongly support the promotion of sustainable agricultural development. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
19. Object Detection Based on Improved YOLOv7 for UAV Aerial Image.
- Author
-
CUI Liqun and CAO Huawei
- Subjects
OBJECT recognition (Computer vision) ,FEATURE extraction ,AERIAL photography ,IMAGE reconstruction ,PROBLEM solving - Abstract
An improved YOLOv7 aerial image object detection algorithm is proposed to solve the problems of low detection accuracy caused by mesoscale changes, small targets and dense occlusion in UAV aerial images. Firstly, a weighted sampling module with joint dynamic convolution is designed to capture features from multiple dimensions and improve the feature extraction ability of the model. Secondly, add a shallow feature detection head to retain more detailed information and enhance the ability to utilize small target features. Then, a multi- scale feature aggregation module (C2-Res2Block) with residual structure is constructed in the feature fusion part to make the model fuse rich multi-scale information. Finally, the MPDIoU measure is used to replace the traditional IOU to calculate the boundary regression loss and improve the localization ability of the model to the densely occluding target. Experiments on UAV aerial photography data set VisDrone2019 show that the improved algorithm is 4.3 percentage points higher than the original model on mAP@0.5, 2.4 percentage points on mAP@0.5: 0.95, the number of parameters is reduced by 6.81x10
6 , and the detection accuracy is higher than the current mainstream object detection algorithms. It effectively improves the detection accuracy of UAV aerial images, and obviously improves the false detection and missing detection of aerial objects. [ABSTRACT FROM AUTHOR]- Published
- 2024
- Full Text
- View/download PDF
20. 基于改进 YOLOv7 的无人机航拍视频西瓜计数方法.
- Author
-
殷慧军, 王宝丽, 景运革, 李菊霞, 王鹏岭, 权高翔, and 孙婷婷
- Subjects
- *
FEATURE extraction , *FORECASTING methodology , *LEARNING ability , *AGRICULTURE , *MELONS , *WATERMELONS - Abstract
To address the difficulties in manual counting for the uneven distribution and severe occlusion of watermelons in natural environments, this study utilizes drones and smartphones to collect videos and images, combined with manual annotation to establish a dataset for Sanbai melons and Ningxia selenium sand melons. A watermelon video automatic counting method based on the YOLOv7-GCSF model and an improved DeepSORT algorithm is proposed. The lightweight YOLOv7 model with GhostConv is enhanced with GBS modules, G-ELAN modules, and G-SPPCSPC modules to increase the model’s detection speed. Some ELAN modules are replaced with the C2f module from YOLOv8 to reduce redundant information. The SimAM attention mechanism is introduced into the MP module of the feature fusion layer to construct the MP-SimAM module, which is used to enhance the model's feature extraction capability. The CIoU loss function is replaced with the fasterconverging, lower-loss Focal EIoU loss function to increase the model's convergence speed. In video tracking and counting, a mask collision line mechanism is proposed for more accurate counting of Sanbai melons and Ningxia selenium sand melons. The results show that in terms of object detection: the four improvements to the YOLOv7-GCSF model have all enhanced the model’s performance to some extent. Specifically, compared to the YOLOv7 model, the construction of the MP-SimAM module increased accuracy by 1.5 percentage points, indicating a greater focus on Sanbai melons and Ningxia selenium sand melons. The addition of GhostConv reduced the model size by 28.1MB, demonstrating that the construction of GBS, G-ELAN, and G-SPPCSPC modules effectively reduced the model size and improved detection speed. The incorporation of the C2f module reduced the model's floating-point operations (FLOPs) by 77.5 billion, indicating that the model has eliminated most of the redundant information. The addition of the Focal EIoU loss function significantly increased the model’s convergence speed, indicating further enhancement of the model's learning ability. The improved YOLOv7-GCSF model achieved an accuracy (P) of 94.2% and a mean average precision (mAP0.5) of 98.2%, which is 5.0, 2.3, 21.9, and 14.9 percentage points higher in accuracy and 3.7, 0.3, 4.6, and 9.3 percentage points higher in mean average precision compared to YOLOv5, YOLOv7, Faster RCNN, and SSD, respectively. In terms of model lightweighting, the YOLOv7-GCSF model has seen a decrease of 1.18M and 0.11M in the number of parameters compared to the YOLOv4-Ghostnet and YOLOv7-Slimneck models, respectively. Compared to the original YOLOv7, the YOLOv7-GCSF model has reduced the parameter count and model size by 0.57M and 18.88MB, respectively. In terms of object tracking: the improved DeepSORT multi-object tracking accuracy is 91.2%, and the multi-object tracking precision is 89.6%, which is 5.0 and 13.7 percentage points higher in tracking accuracy and 3.7 and 13.1 percentage points higher in tracking precision compared to Tracktor and SORT, respectively. Comparing the improved model with manual counting results, the determination coefficient for the counting results of Sanbai melons and Ningxia selenium sand melons is 0.93, the average counting accuracy is 96.3%, and the average absolute error is 0.77, indicating that the error between the improved model and manual counting is small. This approach, by enabling effective counting of watermelons in agricultural fields, provides a technical methodology for the forecasting of watermelon yields. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
21. Vehicle detection algorithm for foggy based on improved AOD-Net.
- Author
-
Zhang, Liyan, Zhao, Jianing, Lang, Zhengang, and Fang, Liu
- Subjects
- *
DEEP learning , *SIGNAL-to-noise ratio , *MOTOR vehicle driving , *INTELLIGENT transportation systems , *HAZE , *ALGORITHMS - Abstract
To strengthen the safety monitoring of foggy road traffic and maintain the safety of vehicle driving on foggy roads, image dehazing algorithms are used to improve the clarity of road images detected in foggy environments, thereby improving the detection ability and monitoring efficiency of intelligent transportation systems for vehicle targets. Due to the low accuracy of vehicle detection and serious problem of missed detections in haze environments, this paper proposes an improved All-in-One Dehazing Network (AOD-Net) algorithm for detecting foggy vehicles, which adds batch normalization (BN) layers after each layer of convolution in AOD-Net, accelerating the convergence of the model and controlling overfitting. To enhance image detail information, an effective pyramid-shaped PSA attention module is embedded to extract richer feature information, enrich model representation, and improve the loss function to a multi-scale structural similarity (MS-SSIM) + L1 mixed loss function, thereby improving the quality, brightness, and contrast of dehazing images. Compared with current image dehazing algorithms, the dehazing quality of our algorithm is superior to other dehazing algorithms, such as dark channel prior (DCP), Dehaze-Net, and Fusion Feature Attention Network (FFA-Net). Compared with AOD-Net, the improved algorithm has increased the peak signal-to-noise ratio by 3.23 dB. At the same time, after the improved AOD-Net image dehazing processing, YOLOv7 object detection was performed and experimentally validated on a real foggy dataset. The results showed that compared with the previous method, it had better recognition performance in foggy detection and recognition, and higher detection accuracy for vehicles. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
22. Las-yolo: a lightweight detection method based on YOLOv7 for small objects in airport surveillance.
- Author
-
Zhou, Wentao, Cai, Chengtao, Wu, Kejun, Li, Chenming, and Gao, Biqin
- Subjects
- *
OBJECT recognition (Computer vision) , *AIRPORT safety , *MODEL airplanes , *COMPUTER training , *AIRPLANES - Abstract
The civil aviation transportation has sustained rapid growth, which poses significant challenges in ensuring airport safety and efficiency of use. Persons and vehicles are tiny targets in airport surveillance. Existing detection methods are difficult to detect accurately. Enhancing small target detection by only the method of adding enhancement modules inevitably leads to increased network parameters. To address the above issues, this article proposes a lightweight airport surveillance detection based on YOLOv7 named LAS-YOLO. Firstly, we design the lightweight basic module, which significantly reduces network parameters while retaining certain local features. Secondly, we replace the SPPCSPC module with the spatial pyramid pooling-fast module with fewer parameters, further reducing the quantity of network parameters. Finally, the attention mechanism and small object detection layer are introduced to enhance small object detection accuracy. The efficient channel attention module is selected among three attention methods by experiments. We simulate the application process of object detection methods in airport surveillance, training on high-performance computers and testing on lower-performance computer. This article verifies the performance of the proposed method on the public ASS dataset consisting of the airport surface surveillance dataset (ASS1) and panoramic surveillance dataset (ASS2). The experiment shows that the parameters of LAS-YOLO are 12.5 M, which is 34.2 % of the original model. The mean average precision is 89.8 % on the ASS1. This proposed method enhances the average precision for airplane and vehicle detection by 14.5 % and 22.7 % compared to YOLOv7 on the ASS2. In order to reflect the robustness of the model in airport surveillance, we conduct another experiment using airplane data from the ROSD. The experiment demonstrates the superiority of the proposed method over other models in airport surveillance. Code is available at https://zenodo.org/records/10969930. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
23. 基于改进 YOLOv7 的小目标焊点缺陷检测算法.
- Author
-
刘兆龙, 曹 伟, and 高军伟
- Abstract
Copyright of Chinese Journal of Liquid Crystal & Displays is the property of Chinese Journal of Liquid Crystal & Displays and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
- Full Text
- View/download PDF
24. 基于改进YOLOv7 的织物疵点小目标检测算法.
- Author
-
陈泽纯, 林富生, 张庆, 宋志峰, 刘泠杉, and 余联庆
- Subjects
TRANSFORMER models ,INFORMATION networks ,INDUSTRIAL applications ,ALGORITHMS ,DETECTORS - Abstract
Copyright of Cotton Textile Technology is the property of Cotton Textile Technology Editorial Office and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
25. Mcan-YOLO: An Improved Forest Fire and Smoke Detection Model Based on YOLOv7.
- Author
-
Liu, Hongying, Zhu, Jun, Xu, Yiqing, and Xie, Ling
- Subjects
FOREST fires ,WILDLIFE resources ,MODEL validation ,SMOKE ,NECK - Abstract
Forest fires pose a significant threat to forest resources and wildlife. To balance accuracy and parameter efficiency in forest fire detection, this study proposes an improved model, Mcan-YOLO, based on YOLOv7. In the Neck section, the asymptotic feature pyramid network (AFPN) was employed to effectively capture multi-scale information, replacing the traditional module. Additionally, the content-aware reassembly of features (CARAFE) replaced the conventional upsampling method, further reducing the number of parameters. The normalization-based attention module (NAM) was integrated after the ELAN-T module to enhance the recognition of various fire smoke features, and the Mish activation function was used to optimize model convergence. A real fire smoke dataset was constructed using the mean structural similarity (MSSIM) algorithm for model training and validation. The experimental results showed that, compared to YOLOv7-tiny, Mcan-YOLO improved precision by 4.6%, recall by 6.5%, and mAP50 by 4.7%, while reducing the number of parameters by 5%. Compared with other mainstream algorithms, Mcan-YOLO achieved better precision with fewer parameters. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
26. PDC-YOLO: A Network for Pig Detection under Complex Conditions for Counting Purposes.
- Author
-
He, Peitong, Zhao, Sijian, Pan, Pan, Zhou, Guomin, and Zhang, Jianhua
- Subjects
SWINE farms ,FOOD supply ,ERROR rates ,AGRICULTURE ,SWINE - Abstract
Pigs play vital roles in the food supply, economic development, agricultural recycling, bioenergy, and social culture. Pork serves as a primary meat source and holds extensive applications in various dietary cultures, making pigs indispensable to human dietary structures. Manual pig counting, a crucial aspect of pig farming, suffers from high costs and time-consuming processes. In this paper, we propose the PDC-YOLO network to address these challenges, dedicated to detecting pigs in complex farming environments for counting purposes. Built upon YOLOv7, our model incorporates the SPD-Conv structure into the YOLOv7 backbone to enhance detection under varying lighting conditions and for small-scale pigs. Additionally, we replace the neck of YOLOv7 with AFPN to efficiently fuse features of different scales. Furthermore, the model utilizes rotated bounding boxes for improved accuracy. Achieving a mAP of 91.97%, precision of 95.11%, and recall of 89.94% on our collected pig dataset, our model outperforms others. Regarding technical performance, PDC-YOLO exhibits an error rate of 0.002 and surpasses manual counting significantly in speed. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
27. Lightweight wildfire smoke monitoring algorithm based on unmanned aerial vehicle vision.
- Author
-
Li, Guanyi, Cheng, Pengle, Li, Yong, and Huang, Ying
- Abstract
Forest fires have a serious impact on people's living environment. Currently, drones enable rapid detection of forest fires. Due to the limited processing capabilities of onboard drones, the accuracy of smoke detection algorithms is low, and the processing speed is slow. This paper proposes an early wildfire smoke detection system designed for unmanned aerial vehicle (UAV) images, leveraging a modified YOLOv7 model, termed YOLOv7-MS(Modified Smoke). A dataset of more than 4,000 wildfire images was curated using existing UAV imagery. Our approach introduces several advancements. First, we propose a novel 3FIoU loss function to enhance stability and expedite convergence during training. Second, we optimize the backbone network by employing the FasterNet technique to reduce the number of parameters and increase the detection speed. Third, we address information loss and quality degradation problems by implementing the Asymptotic Feature Pyramid Network (AFPN) to counter indirect interactions between non-adjacent layers. Finally, we integrate a three-dimensional attention mechanism into the network to enhance focus on the target. Experimental findings showcase the efficacy of our YOLOv7-MS method, achieving a mean Average Precision (mAP) of 79.3% while maintaining a frame rate of 175 fps, outperforming other one-stage object detectors when evaluated on customized image datasets and public datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
28. Improved Architecture and Training Strategies of YOLOv7 for Remote Sensing Image Object Detection.
- Author
-
Zhao, Dewei, Shao, Faming, Liu, Qiang, Zhang, Heng, Zhang, Zihan, and Yang, Li
- Subjects
- *
OBJECT recognition (Computer vision) , *REMOTE sensing , *FEATURE extraction , *NETWORK performance , *ALGORITHMS - Abstract
The technology for object detection in remote sensing images finds extensive applications in production and people's lives, and improving the accuracy of image detection is a pressing need. With that goal, this paper proposes a range of improvements, rooted in the widely used YOLOv7 algorithm, after analyzing the requirements and difficulties in the detection of remote sensing images. Specifically, we strategically remove some standard convolution and pooling modules from the bottom of the network, adopting stride-free convolution to minimize the loss of information for small objects in the transmission. Simultaneously, we introduce a new, more efficient attention mechanism module for feature extraction, significantly enhancing the network's semantic extraction capabilities. Furthermore, by adding multiple cross-layer connections in the network, we more effectively utilize the feature information of each layer in the backbone network, thereby enhancing the network's overall feature extraction capability. During the training phase, we introduce an auxiliary network to intensify the training of the underlying network and adopt a new activation function and a more efficient loss function to ensure more effective gradient feedback, thereby elevating the network performance. In the experimental results, our improved network achieves impressive mAP scores of 91.2% and 80.8% on the DIOR and DOTA version 1.0 remote sensing datasets, respectively. These represent notable improvements of 4.5% and 7.0% over the original YOLOv7 network, significantly enhancing the efficiency of detecting small objects in particular. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
29. High-Precision Mango Orchard Mapping Using a Deep Learning Pipeline Leveraging Object Detection and Segmentation.
- Author
-
Afsar, Muhammad Munir, Bakhshi, Asim Dilawar, Iqbal, Muhammad Shahid, Hussain, Ejaz, and Iqbal, Javed
- Subjects
- *
OBJECT recognition (Computer vision) , *ORCHARD management , *STANDARD deviations , *MANGO , *CROP yields - Abstract
Precision agriculture-based orchard management relies heavily on the accurate delineation of tree canopies, especially for high-value crops like mangoes. Traditional GIS and remote sensing methods, such as Object-Based Imagery Analysis (OBIA), often face challenges due to overlapping canopies, complex tree structures, and varied light conditions. This study aims to enhance the accuracy of mango orchard mapping by developing a novel deep-learning approach that combines fine-tuned object detection and segmentation techniques. UAV imagery was collected over a 65-acre mango orchard in Multan, Pakistan, and processed into an RGB orthomosaic with a 3 cm ground sampling distance. The You Only Look Once (YOLOv7) framework was trained on an annotated dataset to detect individual mango trees. The resultant bounding boxes were used as prompts for the segment anything model (SAM) for precise delineation of canopy boundaries. Validation against ground truth data of 175 manually digitized trees showed a strong correlation ( R 2 = 0.97), indicating high accuracy and minimal bias. The proposed method achieved a mean absolute percentage error (MAPE) of 4.94% and root mean square error (RMSE) of 80.23 sq ft against manually digitized tree canopies with an average size of 1290.14 sq ft. The proposed approach effectively addresses common issues such as inaccurate bounding boxes and over- or under-segmentation of tree canopies. The enhanced accuracy can substantially assist in various downstream tasks such as tree location mapping, canopy volume estimation, health monitoring, and crop yield estimation. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
30. GHA-Inst: a real-time instance segmentation model utilizing YOLO detection framework.
- Author
-
Dong, Chengang, Tang, Yuhao, and Zhang, Liyan
- Subjects
- *
DEEP learning , *NECK , *NOISE , *VIDEOS - Abstract
The real-time instance segmentation task based on deep learning aims to accurately identify and distinguish all instance objects from images or videos. However, due to the existence of problems such as mutual occlusion between instances, limitations in model receptive fields, etc., achieving accurate and real-time segmentation continues to pose a formidable challenge. To alleviate the aforementioned issues, this paper proposes a real-time instance segmentation method based on a dual-branch structure, called GHA-Inst. Specifically, we made improvements to the feature fusion module (Neck) and output end (Head) of the YOLOv7-seg real-time instance segmentation framework to mitigate the accuracy reduction caused by feature loss and reduce the interference of background noise on the model. Secondly, we introduced a Global Hybrid-Domain Attention (GHA) module to improve the model's focus on significant information while retaining more original spatial features, alleviate incomplete segmentation caused by instance occlusion, and improve the quality of generated masks. Finally, our method achieved competitive results on multiple metrics of the MS COCO 2017 and KINS open-source datasets. Compared with the YOLOv7-seg baseline model, GHA-Inst improved the average precision (AP) by 3.4% and 2.6% on the two datasets, respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
31. YOLO-FNC: An Improved Method for Small Object Detection in Remote Sensing Images Based on YOLOv7.
- Author
-
Lanxue Dang, Gang Liu, Yan-e Hou, and Hongyu Han
- Subjects
OBJECT recognition (Computer vision) ,REMOTE sensing ,ALGORITHMS - Abstract
The detection algorithms of small objects in remote sensing images is often challenging due to the complex background and limited pixels. This can lead to reduced accuracy in detection and an increased number of missed small objects. So this paper introduces YOLOFNC, an enhanced network based on YOLOv7. To improve the model's ability to capture features of small objects, an enhanced C3-Faster module based on the C3 module is designed and integrated into the YOLOv7 network. This module helps extract more features related to small objects. Additionally, we employ Normalized Wasserstein Distance (NWD) fusion GIoU as a novel loss function to refine the accuracy of network optimization weights and the small object regression framework. Furthermore, a coordinated attention (CA) mechanism is incorporated at strategic locations in the model to reduce redundant information in the feature layer and prevent the loss of important small object features. we conduct comparison experiments between YOLO-FNC and other commonly used object detection algorithms on DIOR, AITOD, and VisDrone datasets. The experimental results show that YOLO-FNC achieves 84.4% mAP on the DIOR dataset, 35.9% mAP on the AI-TOD dataset, and 52.6% mAP on the VisDrone dataset. Compared to YOLOv7 and other remote sensing object detection models, YOLO-FNC demonstrates better performance in object detection. [ABSTRACT FROM AUTHOR]
- Published
- 2024
32. 基于 YOLOv7-CA-BiFPN 的路面缺陷检测.
- Author
-
高敏 and 李元
- Abstract
Copyright of Computer Measurement & Control is the property of Magazine Agency of Computer Measurement & Control and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
- Full Text
- View/download PDF
33. 基于改进 YOLOv7 的水下小目标 检测算法研究.
- Author
-
杜锋
- Abstract
Copyright of Computer Measurement & Control is the property of Magazine Agency of Computer Measurement & Control and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
- Full Text
- View/download PDF
34. 基于改进 YOLOv7 的玉米作物害虫检测研究.
- Author
-
宫妍, 程俊杰, 王凯, and 李玉
- Abstract
Copyright of Computer Measurement & Control is the property of Magazine Agency of Computer Measurement & Control and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
- Full Text
- View/download PDF
35. 多尺度特征融合的铁轨异物入侵检测研究.
- Author
-
王楠, 侯涛, and 牛宏侠
- Abstract
Copyright of Journal of Xi'an Jiaotong University is the property of Editorial Office of Journal of Xi'an Jiaotong University and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
- Full Text
- View/download PDF
36. 基于YOLOv7-R 的多车辆目标识别.
- Author
-
李 珣, 伍荣兴, 周慧龙, 刘 欣, 高 涵, and 王文杰
- Abstract
Copyright of Journal of Southeast University / Dongnan Daxue Xuebao is the property of Journal of Southeast University Editorial Office and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
- Full Text
- View/download PDF
37. EGS-YOLO: A Fast and Reliable Safety Helmet Detection Method Modified Based on YOLOv7.
- Author
-
Han, Jianfeng, Li, Zhiwei, Cui, Guoqing, and Zhao, Jingxuan
- Subjects
SAFETY hats ,INDUSTRIAL safety ,BUILDING sites ,LINEAR operators ,COMPUTATIONAL complexity - Abstract
Wearing safety helmets at construction sites is a major measure to prevent safety accidents, so it is essential to supervise and ensure that workers wear safety helmets. This requires a high degree of real-time performance. We improved the network structure based on YOLOv7. To enhance real-time performance, we introduced GhostModule after comparing various modules to create a new efficient structure that generates more feature mappings with fewer linear operations. SE blocks were introduced after comparing several attention mechanisms to highlight important information in the image. The EIOU loss function was introduced to speed up the convergence of the model. Eventually, we constructed the efficient model EGS-YOLO. EGS-YOLO achieves a mAP of 91.1%, 0.2% higher than YOLOv7, and the inference time is 13.3% faster than YOLOv7 at 3.9 ms (RTX 3090). The parameters and computational complexity are reduced by 37.3% and 33.8%, respectively. The enhanced real-time performance while maintaining the original high precision can meet actual detection requirements. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
38. Dense Small Object Detection Based on an Improved YOLOv7 Model.
- Author
-
Chen, Xun, Deng, Linyi, Hu, Chao, Xie, Tianyi, and Wang, Chengqi
- Subjects
COMPUTER vision ,FEATURE extraction ,MATHEMATICAL optimization ,ALGORITHMS - Abstract
Detecting small and densely packed objects in images remains a significant challenge in computer vision. Existing object detection methods often exhibit low accuracy and frequently miss detection when identifying dense small objects and require larger model parameters. This study introduces a novel detection framework designed to address these limitations by integrating advanced feature fusion and optimization techniques. Our approach focuses on enhancing both detection accuracy and parameter efficiency. The approach was evaluated on the open-source VisDrone2019 data set and compared with mainstream algorithms. Experimental results demonstrate a 70.2% reduction in network parameters and a 6.3% improvement in mAP@0.5 over the original YOLOv7 algorithm. These results demonstrate that the enhanced model surpasses existing algorithms in detecting small objects. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
39. Uav identification based on improved YOLOv7 under foggy condition.
- Author
-
He, Xin, Fan, Kuangang, and Xu, Zhitao
- Abstract
One-stage algorithm can be effectively used in normal conditions, showing excellent performance on unmanned aerial vehicle (UAV) detection. However, when facing inclement weathers, such as foggy environment, it cannot give a satisfactory outcome we crave for. At the same time, UAV, a tiny object, only contains a few pixels in images and is hided in the fog, causing object obscurity. Concerned about these premises, an improved YOLOv7 is proposed to focus on UAV detection in foggy situation. We adopt BiFormer, a novel dynamic sparse attention through bi-level routing to achieve a flexible distribution of calculation with content awareness, and CL, combined loss function for replacing original IoU metric, to overcome these challenges. At last, Content-Aware ReAssembly of Features (CARAFE) is integrated to the network, aggregating contextual information within a large receptive field. According to this task, we built a new dataset for fog detection (UAV-FG) in which objects are covered by fog, and amount of experiments on UAV-FG datasets verify the effectiveness of our design. Compared with YOLOv7, our method shows consistent and substantial gains (23.21%, 14.35% improvement in mAP@0.5 and mAP@0.5:0.95, respectively) with negligible computational overhead and also satisfies the requirement of real-time detection. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
40. CF-YOLO: a capable forest fire identification algorithm founded on YOLOv7 improvement.
- Author
-
Liu, Wanjie, Shen, Zirui, and Xu, Sheng
- Abstract
Forest fire is an ecological catastrophe with great damage and rapid spread, which inflicts significant damage upon the ecological balance of the forests and poses a threat to human well-being. Given the current problems of low forest fire recognition accuracy and weak local detection, an improved forest fire detection algorithm Catch Fire YOLO-based neural networks (CF-YOLO) based on YOLOv7 model is studied. In global information processing, the plug-and-play coordinate attention mechanism is introduced into the YOLOv7 model, which enhances the visual depiction of the receptive field, while aggregate features along different spatial directions to improve the depiction of the focal interest. We present the three parallel max-pooling operations in the SPPCSPC module of the Neck to a serial mode, where the output of each pooling is used as the next pooling input. In local information processing, we prepare a feature fusion module to replace the partial high-efficiency layer aggregation network (ELAN), so that the network further improves the detection accuracy while speeding up the calculation. The proposed model was trained and verified on a forest fire dataset, the experimental results demonstrate an improved detection capability, especially for small targets, and can meet the requirements of edge deployment in forest fire scenarios. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
41. 基于并联堆叠模型的织物疵点检测算法.
- Author
-
周星亚, 孙红蕊, 宋 荣, and 夏克尔·赛塔尔
- Abstract
Copyright of Advanced Textile Technology is the property of Zhejiang Sci-Tech University Magazines and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
- Full Text
- View/download PDF
42. 结合Transformer和SimAM轻量化 路面损伤检测算法.
- Author
-
杨杰, 蒋严宣, and 熊欣燕
- Abstract
Copyright of Journal of Railway Science & Engineering is the property of Journal of Railway Science & Engineering Editorial Office and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
- Full Text
- View/download PDF
43. Detection and Tracking of Low-Frame-Rate Water Surface Dynamic Multi-Target Based on the YOLOv7-DeepSORT Fusion Algorithm.
- Author
-
Han, Xingcheng, Fu, Shiwen, and Han, Junxuan
- Subjects
TRACKING algorithms ,SUBMARINES (Ships) ,EUCLIDEAN distance ,FEATURE extraction ,SAILING ships ,TRACKING radar - Abstract
This study aims to address the problem in tracking technology in which targeted cruising ships or submarines sailing near the water surface are tracked at low frame rates or with some frames missing in the video image, so that the tracked targets have a large gap between frames, leading to a decrease in tracking accuracy and inefficiency. Thus, in this study, we proposed a water surface dynamic multi-target tracking algorithm based on the fusion of YOLOv7 and DeepSORT. The algorithm first introduces the super-resolution reconstruction network. The network can eliminate the interference of clouds and waves in images to improve the quality of tracking target images and clarify the target characteristics in the image. Then, the shuffle attention module is introduced into YOLOv7 to enhance the feature extraction ability of the target features in the recognition network. Finally, Euclidean distance matching is introduced into the cascade matching of the DeepSORT algorithm to replace the distance matching of IOU to improve the target tracking accuracy. Simulation results showed that the algorithm proposed in this study has a good tracking effect, with an improvement of 9.4% in the improved YOLOv7 model relative to the mAP50-95 value and an improvement of 13.1% in the tracking accuracy in the DeepSORT tracking network compared with the SORT tracking accuracy. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
44. CC-De-YOLO: A Multiscale Object Detection Method for Wafer Surface Defect
- Author
-
Ma Jianhong, Zhang Tao, Ma Xiaoyan, and Tian Hui
- Subjects
surface defect detection on wafers ,yolov7 ,coordinate attention ,carevc ,idetect_decoupled ,Electronic computers. Computer science ,QA75.5-76.95 - Abstract
Surface defect detection on wafers is crucial for quality control in semiconductor manufacturing. However, the complexity of defect spatial features, including mixed defect types, large scale differences, and overlapping, results in low detection accuracy. In this paper, we propose a CC-De-YOLO model, which is based on the YOLOv7 backbone network. Firstly, the coordinate attention is inserted into the main feature extraction network. Coordinate attention decomposes channel attention into two one-dimensional feature coding processes, which are aggregated along both horizontal and vertical spatial directions to enhance the network’s sensitivity to orientation and position. Then, the nearest neighbor interpolation in the upsampling part is replaced by the CAR-EVC module, which predicts the upsampling kernel from the previous feature map and integrates semantic information into the feature map. Two residual structures are used to capture long-range semantic dependencies and improve feature representation capability. Finally, an efficient decoupled detection head is used to separate classification and regression tasks for better defect classification. To evaluate our model’s performance, we established a wafer surface defect dataset containing six typical defect categories. The experimental results show that the CCDe-YOLO model achieves 91.0% mAP@0.5 and 46.2% mAP@0.5:0.95, with precision of 89.5% and recall of 83.2%. Compared with the original YOLOv7 model and other object detection models, CC-De-YOLO performs better. Therefore, our proposed method meets the accuracy requirements for wafer surface defect detection and has broad application prospects. The dataset containing surface defect data on wafers is currently publicly available on GitHub (https://github.com/ztao3243/Wafer-Datas.git).
- Published
- 2024
- Full Text
- View/download PDF
45. Camellia oleifera trunks detection and identification based on improved YOLOv7.
- Author
-
Wang, Haorui, Liu, Yang, Luo, Hong, Luo, Yuanyin, Zhang, Yuyan, Long, Fei, and Li, Lijun
- Abstract
Summary: Camellia oleifera typically thrives in unstructured environments, making the identification of its trunks crucial for advancing agricultural robots towards modernization and sustainability. Traditional target detection algorithms, however, fall short in accurately identifying Camellia oleifera trunks, especially in scenarios characterized by small targets and poor lighting. This article introduces an enhanced trunk detection algorithm for Camellia oleifera based on an improved YOLOv7 model. This model incorporates dynamic snake convolution instead of standard convolutions to bolster its feature extraction capabilities. It integrates more contextual information, thus enhancing the model's generalization ability across various scenes. Additionally, coordinate attention is introduced to refine the model's spatial feature representation, amplifying the network's focus on essential target region features, which in turn boosts detection accuracy and robustness. This feature selectively strengthens response levels across different channels, prioritizing key attributes for classification and localization. Moreover, the original coordinate loss function of YOLOv7 is replaced with EIoU loss, further enhancing the model's robustness and convergence speed. Experimental results demonstrate a recall rate of 96%, a mean average precision (mAP) of 87.9%, an F1 score of 0.87, and a detection speed of 18 milliseconds per frame. When compared with other models like Faster‐RCNN, YOLOv3, ScaledYOLOv4, YOLOv5, and the original YOLOv7, our improved model shows mAP increases of 8.1%, 7.0%, 7.5%, and 6.6% respectively. Occupying only 70.8 MB, our model requires 9.8 MB less memory than the original YOLOv7. This model not only achieves high accuracy and detection efficiency but is also easily deployable on mobile devices, providing a robust foundation for future intelligent harvesting technologies. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
46. Advancing jasmine tea production: YOLOv7‐based real‐time jasmine flower detection.
- Author
-
Zhou, Hanlin, Luo, Jianlong, Ye, Qiuping, Leng, Wenjun, Qin, Jingfeng, Lin, Jing, Xie, Xiaoyu, Sun, Yilan, Huang, Shiguo, and Pang, Jie
- Subjects
- *
MACHINE learning , *CONVOLUTIONAL neural networks , *IMAGE recognition (Computer vision) , *INDUSTRIAL costs , *JASMINE - Abstract
Background: To produce jasmine tea of excellent quality, it is crucial to select jasmine flowers at their optimal growth stage during harvesting. However, achieving this goal remains a challenge due to environmental and manual factors. This study addresses this issue by classifying different jasmine flowers based on visual attributes using the YOLOv7 algorithm, one of the most advanced algorithms in convolutional neural networks. Results: The mean average precision (mAP value) for detecting jasmine flowers using this model is 0.948, and the accuracy for five different degrees of openness of jasmine flowers, namely small buds, buds, half‐open, full‐open and wiltered, is 87.7%, 90.3%, 89%, 93.9% and 86.4%, respectively. Meanwhile, other ways of processing the images in the dataset, such as blurring and changing the brightness, also increased the credibility of the algorithm. Conclusion: This study shows that it is feasible to use deep learning algorithms for distinguishing jasmine flowers at different growth stages. This study can provide a reference for jasmine production estimation and for the development of intelligent and precise flower‐picking applications to reduce flower waste and production costs. © 2024 Society of Chemical Industry. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
47. MDA-YOLO Person: a 2D human pose estimation model based on YOLO detection framework.
- Author
-
Dong, Chengang, Tang, Yuhao, and Zhang, Liyan
- Subjects
- *
BODY image , *HUMAN body , *POSE estimation (Computer vision) , *ARCHAEOLOGICAL human remains , *PERSONAL names , *DETECTORS - Abstract
Human pose estimation aims to locate and predict the key points of the human body in images or videos. Due to the challenges of capturing complex spatial relationships and handling different body scales, accurate estimation of human pose remains challenging. Our work proposes a real-time human pose estimation method based on the anchor-assisted YOLOv7 framework, named MDA-YOLO Person. In this study, we propose the Keypoint Augmentation Strategies (KAS) to overcome the challenges faced in human pose estimation and improve the model's ability to accurately predict keypoints. Furthermore, we introduce the Anchor Adjustment Module (AAM) as a replacement for the original YOLOv7's detection head. By adjusting the parameters associated with the detector's anchors, we achieve an increased recall rate and enhance the completeness of the pose estimation. Additionally, we incorporate the Multi-Scale Dual-Head Attention (MDA) module, which effectively models the weights of both channel and spatial dimensions at multiple scales, enabling the model to focus on more salient feature information. As a result, our approach outperforms other methods, as demonstrated by the promising results obtained on two large-scale public datasets. MDA-YOLO Person outperforms the baseline model YOLOv7-pose on both MS COCO 2017 and CrowdPose datasets, with improvements of 2.2% and 3.7% in precision and recall on MS COCO 2017, and 1.9% and 3.5% on CrowdPose, respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
48. YOLOv7-GCM: a detection algorithm for creek waste based on improved YOLOv7 model.
- Author
-
Qin, Jianhua, Zhou, Honglan, Yi, Huaian, Ma, Luyao, Nie, Jianhan, and Huang, Tingting
- Abstract
To enhance the cleanliness of creek environments, quadruped robots can be utilized to detect for creek waste. The continuous changes in the water environment significantly reduce the accuracy of image detection when using quadruped robots for image acquisition. In order to improve the accuracy of quadruped robots in waste detection, this article proposed a detection model called YOLOv7-GCM model for creek waste. The model integrated a global attention mechanism (GAM) into the YOLOv7 model, which achieved accurate waste detection in ever-changing backgrounds and underwater conditions. A content-aware reassembly of features (CARAFE) replaced a up-sampling of the YOLOv7 model to achieve more accurate and efficient feature reconstruction. A minimum point distance intersection over union (MPDIOU) loss function replaced the CIOU loss function of the YOLOv7 model to more accurately measure the similarity between target boxes and predictive boxes. After the aforementioned improvements, the YOLOv7-GCM model was obtained. A quadruped robot to patrol the creek and collect images of creek waste. Finally, the YOLOv7-GCM model was trained on the creek waste dataset. The outcomes of the experiment show that the precision rate of the YOLOv7-GCM model has increased by 4.2% and the mean average precision (mAP@0.5) has accumulated by 2.1%. The YOLOv7-GCM model provides a new method for identifying creek waste, which may help promote efficient waste management. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
49. Traffic signs detection and prohibitor signs recognition in Morocco road scene.
- Author
-
Taouqi, Imane, Klilou, Abdessamad, Chaji, Kebir, and Arsalane, Assia
- Subjects
CONVOLUTIONAL neural networks ,TRAFFIC monitoring ,DRIVER assistance systems ,TRAFFIC signs & signals ,TRAFFIC safety - Abstract
Traffic sign detection is a crucial aspect of advanced driver assistance systems (ADAS) for academic research and the automotive industry. seeing that accurate and timely detection of traffic signs (TS) is essential for ensuring the safety of driving. However, TS detection methods encounter challenges like slow detection speed and a lack of robustness in complex environments. This paper suggests addressing these limitations by proposing the use of the you only look one version 7 (YOLOv7) network to detect and recognize TS in road scenes. Furthermore, the k-means++ algorithm is used to acquire anchor boxes. Additionally, a tiny version of YOLOv7 is used to take advantage of its real-time and low model size, which are required for real-time hardware implementation. So, we conducted an experiment using our proprietary Morocco dataset. According to the experimental results, YOLOv7 achieves 85% in terms of mean average precision (mAP) at 0.5 for all classes. And YOLOv7-tiny obtains 90% in the same term. Afterward, a recognition system for the prohibitive class using the convolutional neural network (CNN) is trained and integrated inside the YOLOv7 algorithm; its model achieves an accuracy of 99%, which leads to a good specification of the prohibitive sign meaning. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
50. An improved insulator self-explosion detection method based on group-level pruning for the YOLOv7-tiny algorithm.
- Author
-
You, Xilai, Ma, Jianqiao, and Yang, Guangze
- Abstract
With the construction of intelligent grids, unmanned aerial vehicle have been widely employed to inspect transmission lines. The inspection process generates a large amount of data, which requires a lightweight model to reduce computational overhead. Here, we propose an improved model based on YOLOv7-tiny with group-level pruning to reduce the model size, which achieves a balance between detection accuracy and speed. Firstly, we replace the activation function with the Funnel activation function to optimize the activation domain dynamically. Second, we introduce a lightweight DFC attention mechanism to enhance the ability of backbone to extract long-range features. Finally, we use adaptively spatial feature fusion network to reduce semantic degradation during feature fusion. We group the parameters according to their dependencies and use a consistent sparse approach to obtain parameter importance. The redundant parameter groups were pruned to achieve model light-weighting. Experimental results show that the improved model achieves 95.6% detection accuracy after pruning. Compared with YOLOv7-tiny, the computational complexity is reduced by 53% and the processing speed is increased by 48.1% to 73 frames per second. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.