14 results
Search Results
2. An acoustic imaging recognition based cutting tools wear state prediction method.
- Author
-
Hou, Junjian, Zhang, Bingyu, Zhong, Yudong, Zhao, Dengfeng, He, Wenbin, and Zhou, Fang
- Abstract
Online monitoring of cutting tools wear is an important component of advanced manufacturing technology, which can greatly improve the processing efficiency and reduce the production cost. In this paper, a cutting tools wear state prediction method based on acoustic imaging recognition is developed. By applying the advantages of the functional generalized inverse beamforming method in the sound field reconstruction, the acoustic signal is used as the carrier to reconstruct the three-dimensional space radiated sound field. And then, slice the reconstructed sound field image and input it into the convolutional neural network model as a sample, to process and classify the image and mines the feature information related to state from the sound field image. By incorporating amplitude and phase information of the sound field, the presented method utilizes spatial domain mapping to accurately identify the noise source and address challenges such as low recognition rate and difficult diagnosis under weak fault conditions. Furthermore, the paper also demonstrates the recognition of sound field states through a fault experiment in sound box simulation, based on these theories. And the recognition of sound field states is achieved through a simulation fault experiment conducted on the sound box, thereby validating the feasibility of the state monitoring method based on pattern recognition of sound and image. Finally, the experimental object is selected as the four-edge carbide milling cutter, and the cutting tools wear state is monitored by integrating sound field reconstruction techniques with convolution feature extraction methods to validate the robustness of the proposed approach. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. The ringed residual u-net with non-natural regions feature for image splicing forgery detection and localization.
- Author
-
Wang, Qi and Lu, TongWei
- Subjects
- *
CONVOLUTIONAL neural networks , *FORGERY , *FEATURE extraction , *PETRI nets , *LOCALIZATION (Mathematics) , *PERFORMANCE standards - Abstract
Recently, with the emergence of many image editing tools (photoshop, Topaz studio, etc.), the authenticity of images has been severely challenged. However, the performance of some existing traditional feature extraction methods and detection methods based on convolutional neural network (CNN) is poor, and the information provided by the features extracted from the network is limited and single. In this paper, an end-to-end ringed residual U-Net is proposed to detect image splicing forgery by blending features of non-natural regions. Some regions with significant differences from the image background are defined as non-natural regions(such as the irregular border at the splicing of images). In this paper, a feature enhancement module for non-natural regions is constructed, which the image through the pooling of four different scales, and these features are then combined with the original image and input to the backbone network for processing, aiming to highlight regions of the image that differ significantly from the background. Therefore, after adding the feature enhancement module for non-natural regions to the end-to-end ring residual U-Net, more attention will be paid to the tampering regions in the feature extraction stage, image manipulation detection and localization will also become more accurate. Compared with some mainstream methods, this method achieves better performance on the three standard datasets(CASIA2.0, NIST2016, COLUMBIA). In addition, it has excellent robustness under JPEG compression attack and noise corruption attack. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. Pure large kernel convolutional neural network transformer for medical image registration.
- Author
-
Fang, Zhao and Cao, Wenming
- Subjects
- *
CONVOLUTIONAL neural networks , *IMAGE registration , *TRANSFORMER models , *DIAGNOSTIC imaging , *PIXELS , *MEDICAL registries , *IMAGE analysis - Abstract
Deformable medical image registration is a fundamental and critical task in medical image analysis. Recently, deep learning-based methods have rapidly developed and have shown impressive results in deformable image registration. However, existing approaches still suffer from limitations in registration accuracy or generalization performance. To address these challenges, in this paper, we propose a pure convolutional neural network module (CVTF) to implement hierarchical transformers and enhance the registration performance of medical images. CVTF has a larger convolutional kernel, providing a larger global effective receptive field, which can improve the network's ability to capture long-range dependencies. In addition, we introduce the spatial interaction attention (SIA) module to compute the interrelationship between the target feature pixel points and all other points in the feature map. This helps to improve the semantic understanding of the model by emphasizing important features and suppressing irrelevant ones. Based on the proposed CVTF and SIA, we construct a novel registration framework named PCTNet. We applied PCTNet to generate displacement fields and register medical images, and we conducted extensive experiments and validation on two public datasets, OASIS and LPBA40. The experimental results demonstrate the effectiveness and generality of our method, showing significant improvements in registration accuracy and generalization performance compared to existing methods. Our code has been available at https://github.com/fz852/PCTNet. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. A hybrid approach using CNN and active contour model for automated segmentation of macular edema.
- Author
-
George, Neetha, Ramachandran, Sivakumar, and Jiji, C.V.
- Abstract
Macula is the part of retina responsible for sharp and clear vision. Macular edema is caused by the accumulation of intraretinal fluid (IRF) in the macula, which is further distinguished by the compromised integrity of the blood-retinal barrier, particularly evident in the retinal vasculature. This results in swelling, that may lead to vision impairment and is the dominant sign of several ocular diseases, including age-related macular degeneration, diabetic retinopathy, etc. Quantitative analysis of the fluid regions in macular edema helps in ascertaining the severity as well as the response to treatment of the diseases. Optical coherence tomography (OCT) is a major tool used by ophthalmologists for visualizing edema. The prevalent practice for diagnosing and treating macular edema involves measuring Central Retinal Thickness (CRT). Segmenting the IRF in OCT images offers the potential for a more accurate and better quantification of macular edema. This paper proposes a novel method combining convolutional neural network (CNN) and active contour model for segmenting the IRF to ascertain the severity of macular edema. The IRF region is initially segmented using an encoder-decoder architecture. Contour evolution is then performed on this segmented image to demarcate the IRF boundaries. The advantage of the method is that it does not require precisely labeled images for training the CNN. A comparison of the experimental results with models employing CNN alone and with other state-of-the art methods demonstrates the superior performance and consistency of the proposed method. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
6. GAMNet: A deep learning approach for precise gesture identification.
- Author
-
Amrutha Raj, V. and Malu, G.
- Abstract
Deep learning has gained popularity across several industries, including object recognition and classification. In the case of Convolutional Neural Networks (CNN), the first layers extract the most noticeable elements, such as shape and margin. As the model progresses, it learns to extract more complex features such as texture and color; conversely, skeleton features encompass significant locations (joints) that do not naturally align with the grid-like architecture intended for these networks. This study emphasizes the importance of structural features in enhancing the performance of deep learning models. It introduces the Gesture Analysis Module Network (GAMNet), which computes abstract structural values within the architecture for feature extraction, prioritization, and classification. These values go through a rigorous evaluation process along with the cutting-edge deep learning model, CNN, and result in intermediate representations, leading to better performance in gesture analysis. An automated dance gesture identification system can address the challenges of recognizing hand movements in unpredictable lighting, varied backgrounds, noise, and changing camera angles. Despite these challenges, GAMNet performed remarkably well, surpassing renowned models like VGGNet, ResNet, EfficientNet, and CNN, achieving a classification accuracy of 96.80%, even in challenging image circumstances. This paper highlights how GAMNet can revolutionize the world of classical Indian dance, opening up new opportunities for research and development in this field. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. Single image dehazing using attention layers in convolutional neural networks.
- Author
-
Rao, Vishisht Srihari, Vinay, P., and Uma, D.
- Abstract
A hazy image is characterized by atmospheric conditions that reduce the image’s clarity and contrast, thereby making it less visible. This degradation in image quality can hinder the performance of advanced computer vision tasks such as object detection and identifying open spaces which need to perform with high accuracy in important real world applications such as security surveillance and autonomous driving. In the recent past, the use of deep learning in image processing tasks have shown a remarkable improvement in performance, in particular, Convolutional Neural Networks (CNNs) perform superior to any other type of neural network in image related tasks. In this paper, we propose the addition of Channel Attention and Pixel Attention layers to four state-of-the-art CNNs, namely, GMAN, U-Net, 123-CEDH and DMPHN, used for the task of image dehazing. We show that the addition of these layers yields a non-trivial improvement on the quality of the dehazed images which we show qualitatively with examples and quantitatively by obtaining PSNR and SSIM scores of 28.63 and 0.959 respectively. Through the experiments, we show that the addition of the mentioned attention layers to the GMAN architecture yields the best results. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. Neuron image segmentation based on convolution and BN fusion and multi-input feature fusion.
- Author
-
He, Fuyun, Feng, Huiling, and Tang, Xiaohu
- Subjects
- *
DEEP learning , *IMAGE segmentation , *RETINAL blood vessels , *CONVOLUTIONAL neural networks , *NEURONS , *STEM cells - Abstract
The segmentation of neuronal morphology in electron microscopy images is crucial for the analysis and understanding of neuronal function. However, most of the existing segmentation methods are not suitable for challenging datasets where the neuronal structure is contaminated by noise or has interrupted parts. In this paper, we propose a segmentation method based on deep learning to determine the location information of neurons and reduce the influence of image noise in the data. Specifically, we adapt our neuron dataset based on UNet by using convolution with BN fusion and multi-input feature fusion. The method is named REDAFNet. The model simplifies the model structure and enhances the generalization ability by fusing the convolution layer and BN layer. The noise interference in the data was reduced by multi-input feature fusion, and the ability to understand and express the data was enhanced. The method takes a neuron image as input and its pixel segmentation map as output. Experimental results show that the segmentation accuracy of the proposed method is 91.96%, 93.86% and 80.25% on the ISBI2012 dataset, U-RISC retinal neuron dataset and N2DH-GOWT1 stem cell dataset, respectively. Compared with the existing segmentation methods, the proposed method can extract more complete feature information and achieve more accurate segmentation. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. LogCSS: Log anomaly detection based on BERT-CNN with context-semantics-statistics features.
- Author
-
Li, Zhongliang, Tu, Xuezhen, Gao, Hong, Huang, Shiyue, and Ma, Zongmin
- Subjects
- *
DEEP learning , *ANOMALY detection (Computer security) , *CONVOLUTIONAL neural networks , *ARTIFICIAL intelligence - Abstract
With the development of artificial intelligence, deep-learning-based log anomaly detection proves to be an important research topic. In this paper, we propose LogCSS, a novel log anomaly detection framework based on the Context-Semantics-Statistics Convolutional Neural Network (CSSCNN). It is the first model that uses BERT (Bidirectional Encoder Representation from Transformers) and CNN (Convolutional Neural Network) to extract the semantic, temporal, and correlational features of the logs. We combine the features with the statistic information of log templates for the classification model to improve the accuracy. We also propose a technique, DOOT (Deals with the Out-Of-Templates), for online template matching. The experimental research shows that our framework improves the average F1 score of the six best algorithms in the industry by more than 5% on the open-source dataset HDFS, and improves the average F1 score of the six best algorithms in the industry by more than 8% on the BGL dataset, LogCSS also performs better than other similar methods on our own constructed dataset. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
10. Towards lightweight military object detection.
- Author
-
Li, Zhigang, Nian, Wenhao, Sun, Xiaochuan, and Li, Shujie
- Subjects
- *
OBJECT recognition (Computer vision) , *CONVOLUTIONAL neural networks , *FEATURE extraction - Abstract
Military object military object detection technology serves as the foundation and critical component for reconnaissance and command decision-making, playing a significant role in information-based and intelligent warfare. However, many existing military object detection models focus on exploring deeper and more complex architectures, which results in models with a large number of parameters. This makes them unsuitable for inference on mobile or resource-constrained combat equipment, such as combat helmets and reconnaissance Unmanned Aerial Vehicles (UAVs). To tackle this problem, this paper proposes a lightweight detection framework. A CSP-GhostnetV2 module is proposed in our method to make the feature extraction network more lightweight while extracting more effective information. Furthermore, to fuse multiscale information in low-computational scenarios, GSConv and the proposed CSP-RepGhost are used to form a lightweight feature aggregation network. The experimental results demonstrate that our proposed lightweight model has significant advantages in detection accuracy and efficiency compared to other detection algorithms. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
11. Remote sensing image classification method based on improved ShuffleNet convolutional neural network.
- Author
-
Li, Ziqi, Su, Yuxuan, Zhang, Yonghong, Yin, Hefeng, Sun, Jun, and Wu, Xiaojun
- Subjects
- *
CONVOLUTIONAL neural networks , *REMOTE sensing , *SPINE , *DEEP learning , *IMAGE recognition (Computer vision) , *IMAGE processing - Abstract
As a list of remotely sensed data sources is available, the effective processing of remote sensing images is of great significance in practical applications in various fields. This paper proposes a new lightweight network to solve the problem of remote sensing image processing by using the method of deep learning. Specifically, the proposed model employs ShuffleNet V2 as the backbone network, appropriately increases part of the convolution kernels to improve the classification accuracy of the network, and uses the maximum overlapping pooling layer to enhance the detailed features of the input images. Finally, Squeeze and Excitation (SE) blocks are introduced as the attention mechanism to improve the architecture of the network. Experimental results based on several multisource data show that our proposed network model has a good classification effect on the test samples and can achieve more excellent classification performance than some existing methods, with an accuracy of 91%, and can be used for the classification of remote sensing images. Our model not only has high accuracy but also has faster training speed compared with large networks and can greatly reduce computation costs. The demo code of our proposed method will be available at https://github.com/li-zi-qi. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
12. Text summarization using modified generative adversarial network1.
- Author
-
Srivastava, Jyoti, Srivastava, Ashish Kumar, Muthu Kumar, B., and Anandaraj, S.P.
- Subjects
- *
TEXT summarization , *AUTOMATIC summarization , *CONVOLUTIONAL neural networks , *GENERATIVE adversarial networks , *RESEARCH personnel , *TEXT messages - Abstract
Text summarizing (TS) takes key information from a source text and condenses it for the user while retaining the primary material. When it comes to text summaries, the most difficult problem is to provide broad topic coverage and diversity in a single summary. Overall, text summarization addresses the fundamental need to distill large volumes of information into more manageable and digestible forms, making it a crucial technology in the era of information abundance. It benefits individuals, businesses, researchers, and various other stakeholders by enhancing efficiency and comprehension in dealing with textual data. In this paper, proposed a novel Modified Generative adversarial network (MGAN) for summarize the text. The proposed model involves three stages namely pre-processing, Extractive summarization, and summary generation. In the first Phase, the Text similarity dataset is pre-processed using Lowering Casing, Tokenization, Lemmatization, and, Stop Word Removal. In the second Phase, the Extractive summarization is done in three steps Generating similarity metrics, Sentence Ranking, and Sentence Extractive. In the third stage, a generative adversarial network (GAN) employs summary generation to jointly train the discriminative model D and the generative model G. To classify texts and annotate their syntax, Generative Model G employs a convolutional neural network called Bidirectional Gated Recursive Unit (CNN-BiGRU). The performance analysis of the proposed MGAN is calculated based on the parameters like accuracy, specificity, Recall, and Precision metrics. The proposed MGAN achieves an accuracy range of 99%. The result shows that the proposed MGAN improves the overall accuracy better than 9%, 6.5% and 5.4% is DRM, LSTM, and CNN respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
13. Artificial intelligence auxiliary diagnosis and treatment system for breast cancer in developing countries.
- Author
-
Li, Wenxiu, Gou, Fangfang, and Wu, Jia
- Subjects
- *
ARTIFICIAL intelligence , *BREAST cancer , *CANCER diagnosis , *ELECTRONIC health records ,DEVELOPING countries - Abstract
BACKGROUND: In many developing countries, a significant number of breast cancer patients are unable to receive timely treatment due to a large population base, high patient numbers, and limited medical resources. OBJECTIVE: This paper proposes a breast cancer assisted diagnosis system based on electronic medical records. The goal of this system is to address the limitations of existing systems, which primarily rely on structured electronic records and may miss crucial information stored in unstructured records. METHODS: The proposed approach is a breast cancer assisted diagnosis system based on electronic medical records. The system utilizes breast cancer enhanced convolutional neural networks with semantic initialization filters (BC-INIT-CNN). It extracts highly relevant tumor markers from unstructured medical records to aid in breast cancer staging diagnosis and effectively utilizes the important information present in unstructured records. RESULTS: The model's performance is assessed using various evaluation metrics. Such as accuracy, ROC curves, and Precision-Recall curves. Comparative analysis demonstrates that the BC-INIT-CNN model outperforms several existing methods in terms of accuracy and computational efficiency. CONCLUSIONS: The proposed breast cancer assisted diagnosis system based on BC-INIT-CNN showcases the potential to address the challenges faced by developing countries in providing timely treatment to breast cancer patients. By leveraging unstructured medical records and extracting relevant tumor markers, the system enables accurate staging diagnosis and enhances the utilization of valuable information. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
14. Knee cartilage MR images segmentation based on multi-dimensional hybrid convolutional neural network.
- Author
-
Chen, Wenda and Shi, Cao
- Subjects
- *
CONVOLUTIONAL neural networks , *ARTIFICIAL neural networks , *IMAGE segmentation , *MAGNETIC resonance imaging , *KNEE , *CARTILAGE - Abstract
Accurate segmentation of knee cartilage in MR images is crucial for early diagnosis and treatment of knee conditions. Manual segmentation is time-consuming, leading researchers to explore automatic deep learning methods. However, the choice between 2D and 3D networks for organ segmentation remains debated. In this paper, we propose a hybrid 2D and 3D deep neural network approach, named UVNet, which combines the strengths of both techniques to enhance segmentation performance. Within this network structure, the 3D segmentation network serves as the backbone for feature extraction, while the 2D segmentation network functions as an information supplement network. Local and global MIP images are generated by employing various maximum intensity projection modes of knee MRI volumes as input for the information supplement network. By constructing a local and global MIP feature fusion module, the supplementary information obtained from the 2D segmentation network is fully integrated into the backbone network. We assess the quality of the proposed method using the Osteoarthritis Initiative (OAI) dataset and the 2010 Grand Challenge Knee Image Segmentation (SKI-10) dataset, comparing it to the Baseline Network and other advanced 2D and 3D segmentation methods. The experiments demonstrate that UVNet achieves competitive performance in the aforementioned two cartilage segmentation tasks. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.